1 | // |
---|---|
2 | // Copyright 2016 The ANGLE Project Authors. All rights reserved. |
3 | // Use of this source code is governed by a BSD-style license that can be |
4 | // found in the LICENSE file. |
5 | // |
6 | // vk_utils: |
7 | // Helper functions for the Vulkan Renderer. |
8 | // |
9 | |
10 | #include "libANGLE/renderer/vulkan/vk_utils.h" |
11 | |
12 | #include "libANGLE/Context.h" |
13 | #include "libANGLE/Display.h" |
14 | #include "libANGLE/renderer/vulkan/BufferVk.h" |
15 | #include "libANGLE/renderer/vulkan/ContextVk.h" |
16 | #include "libANGLE/renderer/vulkan/DisplayVk.h" |
17 | #include "libANGLE/renderer/vulkan/RendererVk.h" |
18 | #include "libANGLE/renderer/vulkan/ResourceVk.h" |
19 | #include "libANGLE/renderer/vulkan/android/vk_android_utils.h" |
20 | #include "libANGLE/renderer/vulkan/vk_mem_alloc_wrapper.h" |
21 | |
22 | namespace angle |
23 | { |
24 | egl::Error ToEGL(Result result, EGLint errorCode) |
25 | { |
26 | if (result != angle::Result::Continue) |
27 | { |
28 | egl::Error error = std::move(*egl::Display::GetCurrentThreadErrorScratchSpace()); |
29 | error.setCode(errorCode); |
30 | return error; |
31 | } |
32 | else |
33 | { |
34 | return egl::NoError(); |
35 | } |
36 | } |
37 | } // namespace angle |
38 | |
39 | namespace rx |
40 | { |
41 | namespace |
42 | { |
43 | // Pick an arbitrary value to initialize non-zero memory for sanitization. Note that 0x3F3F3F3F |
44 | // as float is about 0.75. |
45 | constexpr int kNonZeroInitValue = 0x3F; |
46 | |
47 | VkImageUsageFlags GetStagingBufferUsageFlags(vk::StagingUsage usage) |
48 | { |
49 | switch (usage) |
50 | { |
51 | case vk::StagingUsage::Read: |
52 | return VK_BUFFER_USAGE_TRANSFER_DST_BIT; |
53 | case vk::StagingUsage::Write: |
54 | return VK_BUFFER_USAGE_TRANSFER_SRC_BIT; |
55 | case vk::StagingUsage::Both: |
56 | return (VK_BUFFER_USAGE_TRANSFER_DST_BIT | VK_BUFFER_USAGE_TRANSFER_SRC_BIT); |
57 | default: |
58 | UNREACHABLE(); |
59 | return 0; |
60 | } |
61 | } |
62 | |
63 | bool FindCompatibleMemory(const VkPhysicalDeviceMemoryProperties &memoryProperties, |
64 | const VkMemoryRequirements &memoryRequirements, |
65 | VkMemoryPropertyFlags requestedMemoryPropertyFlags, |
66 | VkMemoryPropertyFlags *memoryPropertyFlagsOut, |
67 | uint32_t *typeIndexOut) |
68 | { |
69 | for (size_t memoryIndex : angle::BitSet32<32>(memoryRequirements.memoryTypeBits)) |
70 | { |
71 | ASSERT(memoryIndex < memoryProperties.memoryTypeCount); |
72 | |
73 | if ((memoryProperties.memoryTypes[memoryIndex].propertyFlags & |
74 | requestedMemoryPropertyFlags) == requestedMemoryPropertyFlags) |
75 | { |
76 | *memoryPropertyFlagsOut = memoryProperties.memoryTypes[memoryIndex].propertyFlags; |
77 | *typeIndexOut = static_cast<uint32_t>(memoryIndex); |
78 | return true; |
79 | } |
80 | } |
81 | |
82 | return false; |
83 | } |
84 | |
85 | angle::Result FindAndAllocateCompatibleMemory(vk::Context *context, |
86 | vk::MemoryAllocationType memoryAllocationType, |
87 | const vk::MemoryProperties &memoryProperties, |
88 | VkMemoryPropertyFlags requestedMemoryPropertyFlags, |
89 | VkMemoryPropertyFlags *memoryPropertyFlagsOut, |
90 | const VkMemoryRequirements &memoryRequirements, |
91 | const void *extraAllocationInfo, |
92 | uint32_t *memoryTypeIndexOut, |
93 | vk::DeviceMemory *deviceMemoryOut) |
94 | { |
95 | VkDevice device = context->getDevice(); |
96 | |
97 | ANGLE_TRY(memoryProperties.findCompatibleMemoryIndex( |
98 | context, memoryRequirements, requestedMemoryPropertyFlags, (extraAllocationInfo != nullptr), |
99 | memoryPropertyFlagsOut, memoryTypeIndexOut)); |
100 | |
101 | VkMemoryAllocateInfo allocInfo = {}; |
102 | allocInfo.sType = VK_STRUCTURE_TYPE_MEMORY_ALLOCATE_INFO; |
103 | allocInfo.pNext = extraAllocationInfo; |
104 | allocInfo.memoryTypeIndex = *memoryTypeIndexOut; |
105 | allocInfo.allocationSize = memoryRequirements.size; |
106 | |
107 | // Add the new allocation for tracking. |
108 | RendererVk *renderer = context->getRenderer(); |
109 | renderer->getMemoryAllocationTracker()->setPendingMemoryAlloc( |
110 | allocType: memoryAllocationType, size: allocInfo.allocationSize, memoryTypeIndex: *memoryTypeIndexOut); |
111 | |
112 | // If the initial allocation fails, it is possible to retry the allocation after cleaning the |
113 | // garbage. |
114 | VkResult result; |
115 | bool anyBatchCleaned = false; |
116 | uint32_t batchesWaitedAndCleaned = 0; |
117 | |
118 | do |
119 | { |
120 | result = deviceMemoryOut->allocate(device, allocInfo); |
121 | if (result != VK_SUCCESS) |
122 | { |
123 | ANGLE_TRY(renderer->finishOneCommandBatchAndCleanup(context, &anyBatchCleaned)); |
124 | |
125 | if (anyBatchCleaned) |
126 | { |
127 | batchesWaitedAndCleaned++; |
128 | } |
129 | } |
130 | } while (result != VK_SUCCESS && anyBatchCleaned); |
131 | |
132 | if (batchesWaitedAndCleaned > 0) |
133 | { |
134 | INFO() << "Initial allocation failed. Waited for "<< batchesWaitedAndCleaned |
135 | << " commands to finish and free garbage | Allocation result: " |
136 | << ((result == VK_SUCCESS) ? "SUCCESS": "FAIL"); |
137 | } |
138 | |
139 | ANGLE_VK_CHECK(context, result == VK_SUCCESS, result); |
140 | |
141 | renderer->onMemoryAlloc(allocType: memoryAllocationType, size: allocInfo.allocationSize, memoryTypeIndex: *memoryTypeIndexOut, |
142 | handle: deviceMemoryOut->getHandle()); |
143 | |
144 | return angle::Result::Continue; |
145 | } |
146 | |
147 | template <typename T> |
148 | angle::Result AllocateAndBindBufferOrImageMemory(vk::Context *context, |
149 | vk::MemoryAllocationType memoryAllocationType, |
150 | VkMemoryPropertyFlags requestedMemoryPropertyFlags, |
151 | VkMemoryPropertyFlags *memoryPropertyFlagsOut, |
152 | const VkMemoryRequirements &memoryRequirements, |
153 | const void *extraAllocationInfo, |
154 | const VkBindImagePlaneMemoryInfoKHR *extraBindInfo, |
155 | T *bufferOrImage, |
156 | uint32_t *memoryTypeIndexOut, |
157 | vk::DeviceMemory *deviceMemoryOut); |
158 | |
159 | template <> |
160 | angle::Result AllocateAndBindBufferOrImageMemory(vk::Context *context, |
161 | vk::MemoryAllocationType memoryAllocationType, |
162 | VkMemoryPropertyFlags requestedMemoryPropertyFlags, |
163 | VkMemoryPropertyFlags *memoryPropertyFlagsOut, |
164 | const VkMemoryRequirements &memoryRequirements, |
165 | const void *extraAllocationInfo, |
166 | const VkBindImagePlaneMemoryInfoKHR *extraBindInfo, |
167 | vk::Image *image, |
168 | uint32_t *memoryTypeIndexOut, |
169 | vk::DeviceMemory *deviceMemoryOut) |
170 | { |
171 | const vk::MemoryProperties &memoryProperties = context->getRenderer()->getMemoryProperties(); |
172 | |
173 | ANGLE_TRY(FindAndAllocateCompatibleMemory(context, memoryAllocationType, memoryProperties, |
174 | requestedMemoryPropertyFlags, memoryPropertyFlagsOut, |
175 | memoryRequirements, extraAllocationInfo, |
176 | memoryTypeIndexOut, deviceMemoryOut)); |
177 | |
178 | if (extraBindInfo) |
179 | { |
180 | VkBindImageMemoryInfoKHR bindInfo = {}; |
181 | bindInfo.sType = VK_STRUCTURE_TYPE_BIND_IMAGE_MEMORY_INFO; |
182 | bindInfo.pNext = extraBindInfo; |
183 | bindInfo.image = image->getHandle(); |
184 | bindInfo.memory = deviceMemoryOut->getHandle(); |
185 | bindInfo.memoryOffset = 0; |
186 | |
187 | ANGLE_VK_TRY(context, image->bindMemory2(context->getDevice(), bindInfo)); |
188 | } |
189 | else |
190 | { |
191 | ANGLE_VK_TRY(context, image->bindMemory(context->getDevice(), *deviceMemoryOut)); |
192 | } |
193 | |
194 | return angle::Result::Continue; |
195 | } |
196 | |
197 | template <> |
198 | angle::Result AllocateAndBindBufferOrImageMemory(vk::Context *context, |
199 | vk::MemoryAllocationType memoryAllocationType, |
200 | VkMemoryPropertyFlags requestedMemoryPropertyFlags, |
201 | VkMemoryPropertyFlags *memoryPropertyFlagsOut, |
202 | const VkMemoryRequirements &memoryRequirements, |
203 | const void *extraAllocationInfo, |
204 | const VkBindImagePlaneMemoryInfoKHR *extraBindInfo, |
205 | vk::Buffer *buffer, |
206 | uint32_t *memoryTypeIndexOut, |
207 | vk::DeviceMemory *deviceMemoryOut) |
208 | { |
209 | ASSERT(extraBindInfo == nullptr); |
210 | |
211 | const vk::MemoryProperties &memoryProperties = context->getRenderer()->getMemoryProperties(); |
212 | |
213 | ANGLE_TRY(FindAndAllocateCompatibleMemory(context, memoryAllocationType, memoryProperties, |
214 | requestedMemoryPropertyFlags, memoryPropertyFlagsOut, |
215 | memoryRequirements, extraAllocationInfo, |
216 | memoryTypeIndexOut, deviceMemoryOut)); |
217 | ANGLE_VK_TRY(context, buffer->bindMemory(context->getDevice(), *deviceMemoryOut, 0)); |
218 | return angle::Result::Continue; |
219 | } |
220 | |
221 | template <typename T> |
222 | angle::Result AllocateBufferOrImageMemory(vk::Context *context, |
223 | vk::MemoryAllocationType memoryAllocationType, |
224 | VkMemoryPropertyFlags requestedMemoryPropertyFlags, |
225 | VkMemoryPropertyFlags *memoryPropertyFlagsOut, |
226 | const void *extraAllocationInfo, |
227 | T *bufferOrImage, |
228 | uint32_t *memoryTypeIndexOut, |
229 | vk::DeviceMemory *deviceMemoryOut, |
230 | VkDeviceSize *sizeOut) |
231 | { |
232 | // Call driver to determine memory requirements. |
233 | VkMemoryRequirements memoryRequirements; |
234 | bufferOrImage->getMemoryRequirements(context->getDevice(), &memoryRequirements); |
235 | |
236 | ANGLE_TRY(AllocateAndBindBufferOrImageMemory( |
237 | context, memoryAllocationType, requestedMemoryPropertyFlags, memoryPropertyFlagsOut, |
238 | memoryRequirements, extraAllocationInfo, nullptr, bufferOrImage, memoryTypeIndexOut, |
239 | deviceMemoryOut)); |
240 | |
241 | *sizeOut = memoryRequirements.size; |
242 | |
243 | return angle::Result::Continue; |
244 | } |
245 | |
246 | // Unified layer that includes full validation layer stack |
247 | constexpr char kVkKhronosValidationLayerName[] = "VK_LAYER_KHRONOS_validation"; |
248 | constexpr char kVkStandardValidationLayerName[] = "VK_LAYER_LUNARG_standard_validation"; |
249 | const char *kVkValidationLayerNames[] = { |
250 | "VK_LAYER_GOOGLE_threading", "VK_LAYER_LUNARG_parameter_validation", |
251 | "VK_LAYER_LUNARG_object_tracker", "VK_LAYER_LUNARG_core_validation", |
252 | "VK_LAYER_GOOGLE_unique_objects"}; |
253 | |
254 | } // anonymous namespace |
255 | |
256 | const char *VulkanResultString(VkResult result) |
257 | { |
258 | switch (result) |
259 | { |
260 | case VK_SUCCESS: |
261 | return "Command successfully completed"; |
262 | case VK_NOT_READY: |
263 | return "A fence or query has not yet completed"; |
264 | case VK_TIMEOUT: |
265 | return "A wait operation has not completed in the specified time"; |
266 | case VK_EVENT_SET: |
267 | return "An event is signaled"; |
268 | case VK_EVENT_RESET: |
269 | return "An event is unsignaled"; |
270 | case VK_INCOMPLETE: |
271 | return "A return array was too small for the result"; |
272 | case VK_SUBOPTIMAL_KHR: |
273 | return "A swapchain no longer matches the surface properties exactly, but can still be " |
274 | "used to present to the surface successfully"; |
275 | case VK_ERROR_OUT_OF_HOST_MEMORY: |
276 | return "A host memory allocation has failed"; |
277 | case VK_ERROR_OUT_OF_DEVICE_MEMORY: |
278 | return "A device memory allocation has failed"; |
279 | case VK_ERROR_INITIALIZATION_FAILED: |
280 | return "Initialization of an object could not be completed for implementation-specific " |
281 | "reasons"; |
282 | case VK_ERROR_DEVICE_LOST: |
283 | return "The logical or physical device has been lost"; |
284 | case VK_ERROR_MEMORY_MAP_FAILED: |
285 | return "Mapping of a memory object has failed"; |
286 | case VK_ERROR_LAYER_NOT_PRESENT: |
287 | return "A requested layer is not present or could not be loaded"; |
288 | case VK_ERROR_EXTENSION_NOT_PRESENT: |
289 | return "A requested extension is not supported"; |
290 | case VK_ERROR_FEATURE_NOT_PRESENT: |
291 | return "A requested feature is not supported"; |
292 | case VK_ERROR_INCOMPATIBLE_DRIVER: |
293 | return "The requested version of Vulkan is not supported by the driver or is otherwise " |
294 | "incompatible for implementation-specific reasons"; |
295 | case VK_ERROR_TOO_MANY_OBJECTS: |
296 | return "Too many objects of the type have already been created"; |
297 | case VK_ERROR_FORMAT_NOT_SUPPORTED: |
298 | return "A requested format is not supported on this device"; |
299 | case VK_ERROR_SURFACE_LOST_KHR: |
300 | return "A surface is no longer available"; |
301 | case VK_ERROR_NATIVE_WINDOW_IN_USE_KHR: |
302 | return "The requested window is already connected to a VkSurfaceKHR, or to some other " |
303 | "non-Vulkan API"; |
304 | case VK_ERROR_OUT_OF_DATE_KHR: |
305 | return "A surface has changed in such a way that it is no longer compatible with the " |
306 | "swapchain"; |
307 | case VK_ERROR_INCOMPATIBLE_DISPLAY_KHR: |
308 | return "The display used by a swapchain does not use the same presentable image " |
309 | "layout, or is incompatible in a way that prevents sharing an image"; |
310 | case VK_ERROR_VALIDATION_FAILED_EXT: |
311 | return "The validation layers detected invalid API usage"; |
312 | case VK_ERROR_INVALID_SHADER_NV: |
313 | return "Invalid Vulkan shader was generated"; |
314 | case VK_ERROR_OUT_OF_POOL_MEMORY: |
315 | return "A pool memory allocation has failed"; |
316 | case VK_ERROR_FRAGMENTED_POOL: |
317 | return "A pool allocation has failed due to fragmentation of the pool's memory"; |
318 | case VK_ERROR_INVALID_EXTERNAL_HANDLE: |
319 | return "An external handle is not a valid handle of the specified type"; |
320 | default: |
321 | return "Unknown vulkan error code"; |
322 | } |
323 | } |
324 | |
325 | bool GetAvailableValidationLayers(const std::vector<VkLayerProperties> &layerProps, |
326 | bool mustHaveLayers, |
327 | VulkanLayerVector *enabledLayerNames) |
328 | { |
329 | |
330 | ASSERT(enabledLayerNames); |
331 | for (const auto &layerProp : layerProps) |
332 | { |
333 | std::string layerPropLayerName = std::string(layerProp.layerName); |
334 | |
335 | // Favor unified Khronos layer, but fallback to standard validation |
336 | if (layerPropLayerName == kVkKhronosValidationLayerName) |
337 | { |
338 | enabledLayerNames->push_back(value: kVkKhronosValidationLayerName); |
339 | continue; |
340 | } |
341 | else if (layerPropLayerName == kVkStandardValidationLayerName) |
342 | { |
343 | enabledLayerNames->push_back(value: kVkStandardValidationLayerName); |
344 | continue; |
345 | } |
346 | |
347 | for (const char *validationLayerName : kVkValidationLayerNames) |
348 | { |
349 | if (layerPropLayerName == validationLayerName) |
350 | { |
351 | enabledLayerNames->push_back(value: validationLayerName); |
352 | break; |
353 | } |
354 | } |
355 | } |
356 | |
357 | if (enabledLayerNames->size() == 0) |
358 | { |
359 | // Generate an error if the layers were explicitly requested, warning otherwise. |
360 | if (mustHaveLayers) |
361 | { |
362 | ERR() << "Vulkan validation layers are missing."; |
363 | } |
364 | else |
365 | { |
366 | WARN() << "Vulkan validation layers are missing."; |
367 | } |
368 | |
369 | return false; |
370 | } |
371 | |
372 | return true; |
373 | } |
374 | |
375 | namespace vk |
376 | { |
377 | const char *gLoaderLayersPathEnv = "VK_LAYER_PATH"; |
378 | const char *gLoaderICDFilenamesEnv = "VK_ICD_FILENAMES"; |
379 | |
380 | VkImageAspectFlags GetDepthStencilAspectFlags(const angle::Format &format) |
381 | { |
382 | return (format.depthBits > 0 ? VK_IMAGE_ASPECT_DEPTH_BIT : 0) | |
383 | (format.stencilBits > 0 ? VK_IMAGE_ASPECT_STENCIL_BIT : 0); |
384 | } |
385 | |
386 | VkImageAspectFlags GetFormatAspectFlags(const angle::Format &format) |
387 | { |
388 | VkImageAspectFlags dsAspect = GetDepthStencilAspectFlags(format); |
389 | // If the image is not depth stencil, assume color aspect. Note that detecting color formats |
390 | // is less trivial than depth/stencil, e.g. as block formats don't indicate any bits for RGBA |
391 | // channels. |
392 | return dsAspect != 0 ? dsAspect : VK_IMAGE_ASPECT_COLOR_BIT; |
393 | } |
394 | |
395 | // Context implementation. |
396 | Context::Context(RendererVk *renderer) : mRenderer(renderer), mPerfCounters{} {} |
397 | |
398 | Context::~Context() {} |
399 | |
400 | VkDevice Context::getDevice() const |
401 | { |
402 | return mRenderer->getDevice(); |
403 | } |
404 | |
405 | const angle::FeaturesVk &Context::getFeatures() const |
406 | { |
407 | return mRenderer->getFeatures(); |
408 | } |
409 | |
410 | // MemoryProperties implementation. |
411 | MemoryProperties::MemoryProperties() : mMemoryProperties{} {} |
412 | |
413 | void MemoryProperties::init(VkPhysicalDevice physicalDevice) |
414 | { |
415 | ASSERT(mMemoryProperties.memoryTypeCount == 0); |
416 | vkGetPhysicalDeviceMemoryProperties(physicalDevice, &mMemoryProperties); |
417 | ASSERT(mMemoryProperties.memoryTypeCount > 0); |
418 | } |
419 | |
420 | void MemoryProperties::destroy() |
421 | { |
422 | mMemoryProperties = {}; |
423 | } |
424 | |
425 | bool MemoryProperties::hasLazilyAllocatedMemory() const |
426 | { |
427 | for (uint32_t typeIndex = 0; typeIndex < mMemoryProperties.memoryTypeCount; ++typeIndex) |
428 | { |
429 | const VkMemoryType &memoryType = mMemoryProperties.memoryTypes[typeIndex]; |
430 | if ((memoryType.propertyFlags & VK_MEMORY_PROPERTY_LAZILY_ALLOCATED_BIT) != 0) |
431 | { |
432 | return true; |
433 | } |
434 | } |
435 | return false; |
436 | } |
437 | |
438 | angle::Result MemoryProperties::findCompatibleMemoryIndex( |
439 | Context *context, |
440 | const VkMemoryRequirements &memoryRequirements, |
441 | VkMemoryPropertyFlags requestedMemoryPropertyFlags, |
442 | bool isExternalMemory, |
443 | VkMemoryPropertyFlags *memoryPropertyFlagsOut, |
444 | uint32_t *typeIndexOut) const |
445 | { |
446 | ASSERT(mMemoryProperties.memoryTypeCount > 0 && mMemoryProperties.memoryTypeCount <= 32); |
447 | |
448 | // Find a compatible memory pool index. If the index doesn't change, we could cache it. |
449 | // Not finding a valid memory pool means an out-of-spec driver, or internal error. |
450 | // TODO(jmadill): Determine if it is possible to cache indexes. |
451 | // TODO(jmadill): More efficient memory allocation. |
452 | if (FindCompatibleMemory(memoryProperties: mMemoryProperties, memoryRequirements, requestedMemoryPropertyFlags, |
453 | memoryPropertyFlagsOut, typeIndexOut)) |
454 | { |
455 | return angle::Result::Continue; |
456 | } |
457 | |
458 | // We did not find a compatible memory type. If the caller wanted a host visible memory, just |
459 | // return the memory index with fallback, guaranteed, memory flags. |
460 | if (requestedMemoryPropertyFlags & VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT) |
461 | { |
462 | // The Vulkan spec says the following - |
463 | // There must be at least one memory type with both the |
464 | // VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT and VK_MEMORY_PROPERTY_HOST_COHERENT_BIT |
465 | // bits set in its propertyFlags |
466 | constexpr VkMemoryPropertyFlags fallbackMemoryPropertyFlags = |
467 | VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT | VK_MEMORY_PROPERTY_HOST_COHERENT_BIT; |
468 | |
469 | if (FindCompatibleMemory(memoryProperties: mMemoryProperties, memoryRequirements, requestedMemoryPropertyFlags: fallbackMemoryPropertyFlags, |
470 | memoryPropertyFlagsOut, typeIndexOut)) |
471 | { |
472 | return angle::Result::Continue; |
473 | } |
474 | } |
475 | |
476 | // We did not find a compatible memory type. When importing external memory, there may be |
477 | // additional restrictions on memoryType. Find the first available memory type that Vulkan |
478 | // driver decides being compatible with external memory import. |
479 | if (isExternalMemory) |
480 | { |
481 | if (FindCompatibleMemory(memoryProperties: mMemoryProperties, memoryRequirements, requestedMemoryPropertyFlags: 0, memoryPropertyFlagsOut, |
482 | typeIndexOut)) |
483 | { |
484 | return angle::Result::Continue; |
485 | } |
486 | } |
487 | |
488 | // TODO(jmadill): Add error message to error. |
489 | context->handleError(result: VK_ERROR_INCOMPATIBLE_DRIVER, __FILE__, ANGLE_FUNCTION, __LINE__); |
490 | return angle::Result::Stop; |
491 | } |
492 | |
493 | // StagingBuffer implementation. |
494 | StagingBuffer::StagingBuffer() : mSize(0) {} |
495 | |
496 | void StagingBuffer::destroy(RendererVk *renderer) |
497 | { |
498 | VkDevice device = renderer->getDevice(); |
499 | mBuffer.destroy(device); |
500 | mAllocation.destroy(allocator: renderer->getAllocator()); |
501 | mSize = 0; |
502 | } |
503 | |
504 | angle::Result StagingBuffer::init(Context *context, VkDeviceSize size, StagingUsage usage) |
505 | { |
506 | VkBufferCreateInfo createInfo = {}; |
507 | createInfo.sType = VK_STRUCTURE_TYPE_BUFFER_CREATE_INFO; |
508 | createInfo.flags = 0; |
509 | createInfo.size = size; |
510 | createInfo.usage = GetStagingBufferUsageFlags(usage); |
511 | createInfo.sharingMode = VK_SHARING_MODE_EXCLUSIVE; |
512 | createInfo.queueFamilyIndexCount = 0; |
513 | createInfo.pQueueFamilyIndices = nullptr; |
514 | |
515 | VkMemoryPropertyFlags preferredFlags = 0; |
516 | VkMemoryPropertyFlags requiredFlags = |
517 | VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT | VK_MEMORY_PROPERTY_HOST_COHERENT_BIT; |
518 | |
519 | RendererVk *renderer = context->getRenderer(); |
520 | const Allocator &allocator = renderer->getAllocator(); |
521 | |
522 | uint32_t memoryTypeIndex = 0; |
523 | ANGLE_VK_TRY(context, |
524 | allocator.createBuffer(createInfo, requiredFlags, preferredFlags, |
525 | renderer->getFeatures().persistentlyMappedBuffers.enabled, |
526 | &memoryTypeIndex, &mBuffer, &mAllocation)); |
527 | mSize = static_cast<size_t>(size); |
528 | |
529 | // Wipe memory to an invalid value when the 'allocateNonZeroMemory' feature is enabled. The |
530 | // invalid values ensures our testing doesn't assume zero-initialized memory. |
531 | if (renderer->getFeatures().allocateNonZeroMemory.enabled) |
532 | { |
533 | ANGLE_TRY(InitMappableAllocation(context, allocator, &mAllocation, size, kNonZeroInitValue, |
534 | requiredFlags)); |
535 | } |
536 | |
537 | return angle::Result::Continue; |
538 | } |
539 | |
540 | void StagingBuffer::release(ContextVk *contextVk) |
541 | { |
542 | contextVk->addGarbage(object: &mBuffer); |
543 | contextVk->addGarbage(object: &mAllocation); |
544 | } |
545 | |
546 | void StagingBuffer::collectGarbage(RendererVk *renderer, const QueueSerial &queueSerial) |
547 | { |
548 | GarbageList garbageList; |
549 | garbageList.emplace_back(args: GetGarbage(obj: &mBuffer)); |
550 | garbageList.emplace_back(args: GetGarbage(obj: &mAllocation)); |
551 | |
552 | ResourceUse use(queueSerial); |
553 | renderer->collectGarbage(use, sharedGarbage: std::move(garbageList)); |
554 | } |
555 | |
556 | angle::Result InitMappableAllocation(Context *context, |
557 | const Allocator &allocator, |
558 | Allocation *allocation, |
559 | VkDeviceSize size, |
560 | int value, |
561 | VkMemoryPropertyFlags memoryPropertyFlags) |
562 | { |
563 | uint8_t *mapPointer; |
564 | ANGLE_VK_TRY(context, allocation->map(allocator, &mapPointer)); |
565 | memset(s: mapPointer, c: value, n: static_cast<size_t>(size)); |
566 | |
567 | if ((memoryPropertyFlags & VK_MEMORY_PROPERTY_HOST_COHERENT_BIT) == 0) |
568 | { |
569 | allocation->flush(allocator, offset: 0, size); |
570 | } |
571 | |
572 | allocation->unmap(allocator); |
573 | |
574 | return angle::Result::Continue; |
575 | } |
576 | |
577 | angle::Result AllocateBufferMemory(Context *context, |
578 | vk::MemoryAllocationType memoryAllocationType, |
579 | VkMemoryPropertyFlags requestedMemoryPropertyFlags, |
580 | VkMemoryPropertyFlags *memoryPropertyFlagsOut, |
581 | const void *extraAllocationInfo, |
582 | Buffer *buffer, |
583 | uint32_t *memoryTypeIndexOut, |
584 | DeviceMemory *deviceMemoryOut, |
585 | VkDeviceSize *sizeOut) |
586 | { |
587 | return AllocateBufferOrImageMemory(context, memoryAllocationType, requestedMemoryPropertyFlags, |
588 | memoryPropertyFlagsOut, extraAllocationInfo, bufferOrImage: buffer, |
589 | memoryTypeIndexOut, deviceMemoryOut, sizeOut); |
590 | } |
591 | |
592 | angle::Result AllocateImageMemory(Context *context, |
593 | vk::MemoryAllocationType memoryAllocationType, |
594 | VkMemoryPropertyFlags memoryPropertyFlags, |
595 | VkMemoryPropertyFlags *memoryPropertyFlagsOut, |
596 | const void *extraAllocationInfo, |
597 | Image *image, |
598 | uint32_t *memoryTypeIndexOut, |
599 | DeviceMemory *deviceMemoryOut, |
600 | VkDeviceSize *sizeOut) |
601 | { |
602 | return AllocateBufferOrImageMemory(context, memoryAllocationType, requestedMemoryPropertyFlags: memoryPropertyFlags, |
603 | memoryPropertyFlagsOut, extraAllocationInfo, bufferOrImage: image, |
604 | memoryTypeIndexOut, deviceMemoryOut, sizeOut); |
605 | } |
606 | |
607 | angle::Result AllocateImageMemoryWithRequirements( |
608 | Context *context, |
609 | vk::MemoryAllocationType memoryAllocationType, |
610 | VkMemoryPropertyFlags memoryPropertyFlags, |
611 | const VkMemoryRequirements &memoryRequirements, |
612 | const void *extraAllocationInfo, |
613 | const VkBindImagePlaneMemoryInfoKHR *extraBindInfo, |
614 | Image *image, |
615 | uint32_t *memoryTypeIndexOut, |
616 | DeviceMemory *deviceMemoryOut) |
617 | { |
618 | VkMemoryPropertyFlags memoryPropertyFlagsOut = 0; |
619 | return AllocateAndBindBufferOrImageMemory(context, memoryAllocationType, requestedMemoryPropertyFlags: memoryPropertyFlags, |
620 | memoryPropertyFlagsOut: &memoryPropertyFlagsOut, memoryRequirements, |
621 | extraAllocationInfo, extraBindInfo, image, |
622 | memoryTypeIndexOut, deviceMemoryOut); |
623 | } |
624 | |
625 | angle::Result AllocateBufferMemoryWithRequirements(Context *context, |
626 | MemoryAllocationType memoryAllocationType, |
627 | VkMemoryPropertyFlags memoryPropertyFlags, |
628 | const VkMemoryRequirements &memoryRequirements, |
629 | const void *extraAllocationInfo, |
630 | Buffer *buffer, |
631 | VkMemoryPropertyFlags *memoryPropertyFlagsOut, |
632 | uint32_t *memoryTypeIndexOut, |
633 | DeviceMemory *deviceMemoryOut) |
634 | { |
635 | return AllocateAndBindBufferOrImageMemory(context, memoryAllocationType, requestedMemoryPropertyFlags: memoryPropertyFlags, |
636 | memoryPropertyFlagsOut, memoryRequirements, |
637 | extraAllocationInfo, extraBindInfo: nullptr, buffer, |
638 | memoryTypeIndexOut, deviceMemoryOut); |
639 | } |
640 | |
641 | angle::Result InitShaderModule(Context *context, |
642 | ShaderModule *shaderModule, |
643 | const uint32_t *shaderCode, |
644 | size_t shaderCodeSize) |
645 | { |
646 | VkShaderModuleCreateInfo createInfo = {}; |
647 | createInfo.sType = VK_STRUCTURE_TYPE_SHADER_MODULE_CREATE_INFO; |
648 | createInfo.flags = 0; |
649 | createInfo.codeSize = shaderCodeSize; |
650 | createInfo.pCode = shaderCode; |
651 | |
652 | ANGLE_VK_TRY(context, shaderModule->init(context->getDevice(), createInfo)); |
653 | return angle::Result::Continue; |
654 | } |
655 | |
656 | gl::TextureType Get2DTextureType(uint32_t layerCount, GLint samples) |
657 | { |
658 | if (layerCount > 1) |
659 | { |
660 | if (samples > 1) |
661 | { |
662 | return gl::TextureType::_2DMultisampleArray; |
663 | } |
664 | else |
665 | { |
666 | return gl::TextureType::_2DArray; |
667 | } |
668 | } |
669 | else |
670 | { |
671 | if (samples > 1) |
672 | { |
673 | return gl::TextureType::_2DMultisample; |
674 | } |
675 | else |
676 | { |
677 | return gl::TextureType::_2D; |
678 | } |
679 | } |
680 | } |
681 | |
682 | GarbageObject::GarbageObject() : mHandleType(HandleType::Invalid), mHandle(VK_NULL_HANDLE) {} |
683 | |
684 | GarbageObject::GarbageObject(HandleType handleType, GarbageHandle handle) |
685 | : mHandleType(handleType), mHandle(handle) |
686 | {} |
687 | |
688 | GarbageObject::GarbageObject(GarbageObject &&other) : GarbageObject() |
689 | { |
690 | *this = std::move(other); |
691 | } |
692 | |
693 | GarbageObject &GarbageObject::operator=(GarbageObject &&rhs) |
694 | { |
695 | std::swap(x&: mHandle, y&: rhs.mHandle); |
696 | std::swap(x&: mHandleType, y&: rhs.mHandleType); |
697 | return *this; |
698 | } |
699 | |
700 | // GarbageObject implementation |
701 | // Using c-style casts here to avoid conditional compile for MSVC 32-bit |
702 | // which fails to compile with reinterpret_cast, requiring static_cast. |
703 | void GarbageObject::destroy(RendererVk *renderer) |
704 | { |
705 | ANGLE_TRACE_EVENT0("gpu.angle", "GarbageObject::destroy"); |
706 | VkDevice device = renderer->getDevice(); |
707 | switch (mHandleType) |
708 | { |
709 | case HandleType::Semaphore: |
710 | vkDestroySemaphore(device, (VkSemaphore)mHandle, nullptr); |
711 | break; |
712 | case HandleType::CommandBuffer: |
713 | // Command buffers are pool allocated. |
714 | UNREACHABLE(); |
715 | break; |
716 | case HandleType::Event: |
717 | vkDestroyEvent(device, (VkEvent)mHandle, nullptr); |
718 | break; |
719 | case HandleType::Fence: |
720 | vkDestroyFence(device, (VkFence)mHandle, nullptr); |
721 | break; |
722 | case HandleType::DeviceMemory: |
723 | vkFreeMemory(device, (VkDeviceMemory)mHandle, nullptr); |
724 | break; |
725 | case HandleType::Buffer: |
726 | vkDestroyBuffer(device, (VkBuffer)mHandle, nullptr); |
727 | break; |
728 | case HandleType::BufferView: |
729 | vkDestroyBufferView(device, (VkBufferView)mHandle, nullptr); |
730 | break; |
731 | case HandleType::Image: |
732 | vkDestroyImage(device, (VkImage)mHandle, nullptr); |
733 | break; |
734 | case HandleType::ImageView: |
735 | vkDestroyImageView(device, (VkImageView)mHandle, nullptr); |
736 | break; |
737 | case HandleType::ShaderModule: |
738 | vkDestroyShaderModule(device, (VkShaderModule)mHandle, nullptr); |
739 | break; |
740 | case HandleType::PipelineLayout: |
741 | vkDestroyPipelineLayout(device, (VkPipelineLayout)mHandle, nullptr); |
742 | break; |
743 | case HandleType::RenderPass: |
744 | vkDestroyRenderPass(device, (VkRenderPass)mHandle, nullptr); |
745 | break; |
746 | case HandleType::Pipeline: |
747 | vkDestroyPipeline(device, (VkPipeline)mHandle, nullptr); |
748 | break; |
749 | case HandleType::DescriptorSetLayout: |
750 | vkDestroyDescriptorSetLayout(device, (VkDescriptorSetLayout)mHandle, nullptr); |
751 | break; |
752 | case HandleType::Sampler: |
753 | vkDestroySampler(device, (VkSampler)mHandle, nullptr); |
754 | break; |
755 | case HandleType::DescriptorPool: |
756 | vkDestroyDescriptorPool(device, (VkDescriptorPool)mHandle, nullptr); |
757 | break; |
758 | case HandleType::Framebuffer: |
759 | vkDestroyFramebuffer(device, (VkFramebuffer)mHandle, nullptr); |
760 | break; |
761 | case HandleType::CommandPool: |
762 | vkDestroyCommandPool(device, (VkCommandPool)mHandle, nullptr); |
763 | break; |
764 | case HandleType::QueryPool: |
765 | vkDestroyQueryPool(device, (VkQueryPool)mHandle, nullptr); |
766 | break; |
767 | case HandleType::Allocation: |
768 | vma::FreeMemory(allocator: renderer->getAllocator().getHandle(), allocation: (VmaAllocation)mHandle); |
769 | break; |
770 | default: |
771 | UNREACHABLE(); |
772 | break; |
773 | } |
774 | |
775 | renderer->onDeallocateHandle(handleType: mHandleType); |
776 | } |
777 | |
778 | void MakeDebugUtilsLabel(GLenum source, const char *marker, VkDebugUtilsLabelEXT *label) |
779 | { |
780 | static constexpr angle::ColorF kLabelColors[6] = { |
781 | angle::ColorF(1.0f, 0.5f, 0.5f, 1.0f), // DEBUG_SOURCE_API |
782 | angle::ColorF(0.5f, 1.0f, 0.5f, 1.0f), // DEBUG_SOURCE_WINDOW_SYSTEM |
783 | angle::ColorF(0.5f, 0.5f, 1.0f, 1.0f), // DEBUG_SOURCE_SHADER_COMPILER |
784 | angle::ColorF(0.7f, 0.7f, 0.7f, 1.0f), // DEBUG_SOURCE_THIRD_PARTY |
785 | angle::ColorF(0.5f, 0.8f, 0.9f, 1.0f), // DEBUG_SOURCE_APPLICATION |
786 | angle::ColorF(0.9f, 0.8f, 0.5f, 1.0f), // DEBUG_SOURCE_OTHER |
787 | }; |
788 | |
789 | int colorIndex = source - GL_DEBUG_SOURCE_API; |
790 | ASSERT(colorIndex >= 0 && static_cast<size_t>(colorIndex) < ArraySize(kLabelColors)); |
791 | |
792 | label->sType = VK_STRUCTURE_TYPE_DEBUG_UTILS_LABEL_EXT; |
793 | label->pNext = nullptr; |
794 | label->pLabelName = marker; |
795 | kLabelColors[colorIndex].writeData(data: label->color); |
796 | } |
797 | |
798 | angle::Result SetDebugUtilsObjectName(ContextVk *contextVk, |
799 | VkObjectType objectType, |
800 | uint64_t handle, |
801 | const std::string &label) |
802 | { |
803 | RendererVk *renderer = contextVk->getRenderer(); |
804 | |
805 | VkDebugUtilsObjectNameInfoEXT objectNameInfo = {}; |
806 | objectNameInfo.sType = VK_STRUCTURE_TYPE_DEBUG_UTILS_OBJECT_NAME_INFO_EXT; |
807 | objectNameInfo.objectType = objectType; |
808 | objectNameInfo.objectHandle = handle; |
809 | objectNameInfo.pObjectName = label.c_str(); |
810 | |
811 | if (vkSetDebugUtilsObjectNameEXT) |
812 | { |
813 | ANGLE_VK_TRY(contextVk, |
814 | vkSetDebugUtilsObjectNameEXT(renderer->getDevice(), &objectNameInfo)); |
815 | } |
816 | return angle::Result::Continue; |
817 | } |
818 | |
819 | // ClearValuesArray implementation. |
820 | ClearValuesArray::ClearValuesArray() : mValues{}, mEnabled{} {} |
821 | |
822 | ClearValuesArray::~ClearValuesArray() = default; |
823 | |
824 | ClearValuesArray::ClearValuesArray(const ClearValuesArray &other) = default; |
825 | |
826 | ClearValuesArray &ClearValuesArray::operator=(const ClearValuesArray &rhs) = default; |
827 | |
828 | void ClearValuesArray::store(uint32_t index, |
829 | VkImageAspectFlags aspectFlags, |
830 | const VkClearValue &clearValue) |
831 | { |
832 | ASSERT(aspectFlags != 0); |
833 | |
834 | // We do this double if to handle the packed depth-stencil case. |
835 | if ((aspectFlags & VK_IMAGE_ASPECT_STENCIL_BIT) != 0) |
836 | { |
837 | // Ensure for packed DS we're writing to the depth index. |
838 | ASSERT(index == kUnpackedDepthIndex || |
839 | (index == kUnpackedStencilIndex && aspectFlags == VK_IMAGE_ASPECT_STENCIL_BIT)); |
840 | |
841 | storeNoDepthStencil(index: kUnpackedStencilIndex, clearValue); |
842 | } |
843 | |
844 | if (aspectFlags != VK_IMAGE_ASPECT_STENCIL_BIT) |
845 | { |
846 | storeNoDepthStencil(index, clearValue); |
847 | } |
848 | } |
849 | |
850 | void ClearValuesArray::storeNoDepthStencil(uint32_t index, const VkClearValue &clearValue) |
851 | { |
852 | mValues[index] = clearValue; |
853 | mEnabled.set(pos: index); |
854 | } |
855 | |
856 | gl::DrawBufferMask ClearValuesArray::getColorMask() const |
857 | { |
858 | return gl::DrawBufferMask(mEnabled.bits() & kUnpackedColorBuffersMask); |
859 | } |
860 | |
861 | // ResourceSerialFactory implementation. |
862 | ResourceSerialFactory::ResourceSerialFactory() : mCurrentUniqueSerial(1) {} |
863 | |
864 | ResourceSerialFactory::~ResourceSerialFactory() {} |
865 | |
866 | uint32_t ResourceSerialFactory::issueSerial() |
867 | { |
868 | uint32_t newSerial = ++mCurrentUniqueSerial; |
869 | // make sure serial does not wrap |
870 | ASSERT(newSerial > 0); |
871 | return newSerial; |
872 | } |
873 | |
874 | #define ANGLE_DEFINE_GEN_VK_SERIAL(Type) \ |
875 | Type##Serial ResourceSerialFactory::generate##Type##Serial() \ |
876 | { \ |
877 | return Type##Serial(issueSerial()); \ |
878 | } |
879 | |
880 | ANGLE_VK_SERIAL_OP(ANGLE_DEFINE_GEN_VK_SERIAL) |
881 | |
882 | void ClampViewport(VkViewport *viewport) |
883 | { |
884 | // 0-sized viewports are invalid in Vulkan. |
885 | ASSERT(viewport); |
886 | if (viewport->width == 0.0f) |
887 | { |
888 | viewport->width = 1.0f; |
889 | } |
890 | if (viewport->height == 0.0f) |
891 | { |
892 | viewport->height = 1.0f; |
893 | } |
894 | } |
895 | |
896 | void ApplyPipelineCreationFeedback(Context *context, const VkPipelineCreationFeedback &feedback) |
897 | { |
898 | const bool cacheHit = |
899 | (feedback.flags & VK_PIPELINE_CREATION_FEEDBACK_APPLICATION_PIPELINE_CACHE_HIT_BIT) != 0; |
900 | |
901 | angle::VulkanPerfCounters &perfCounters = context->getPerfCounters(); |
902 | |
903 | if (cacheHit) |
904 | { |
905 | ++perfCounters.pipelineCreationCacheHits; |
906 | perfCounters.pipelineCreationTotalCacheHitsDurationNs += feedback.duration; |
907 | } |
908 | else |
909 | { |
910 | ++perfCounters.pipelineCreationCacheMisses; |
911 | perfCounters.pipelineCreationTotalCacheMissesDurationNs += feedback.duration; |
912 | } |
913 | } |
914 | |
915 | size_t MemoryAllocInfoMapKey::hash() const |
916 | { |
917 | return angle::ComputeGenericHash(key: *this); |
918 | } |
919 | } // namespace vk |
920 | |
921 | #if !defined(ANGLE_SHARED_LIBVULKAN) |
922 | // VK_EXT_debug_utils |
923 | PFN_vkCreateDebugUtilsMessengerEXT vkCreateDebugUtilsMessengerEXT = nullptr; |
924 | PFN_vkDestroyDebugUtilsMessengerEXT vkDestroyDebugUtilsMessengerEXT = nullptr; |
925 | PFN_vkCmdBeginDebugUtilsLabelEXT vkCmdBeginDebugUtilsLabelEXT = nullptr; |
926 | PFN_vkCmdEndDebugUtilsLabelEXT vkCmdEndDebugUtilsLabelEXT = nullptr; |
927 | PFN_vkCmdInsertDebugUtilsLabelEXT vkCmdInsertDebugUtilsLabelEXT = nullptr; |
928 | PFN_vkSetDebugUtilsObjectNameEXT vkSetDebugUtilsObjectNameEXT = nullptr; |
929 | |
930 | // VK_KHR_get_physical_device_properties2 |
931 | PFN_vkGetPhysicalDeviceProperties2KHR vkGetPhysicalDeviceProperties2KHR = nullptr; |
932 | PFN_vkGetPhysicalDeviceFeatures2KHR vkGetPhysicalDeviceFeatures2KHR = nullptr; |
933 | PFN_vkGetPhysicalDeviceMemoryProperties2KHR vkGetPhysicalDeviceMemoryProperties2KHR = nullptr; |
934 | |
935 | // VK_KHR_external_semaphore_fd |
936 | PFN_vkImportSemaphoreFdKHR vkImportSemaphoreFdKHR = nullptr; |
937 | |
938 | // VK_EXT_host_query_reset |
939 | PFN_vkResetQueryPoolEXT vkResetQueryPoolEXT = nullptr; |
940 | |
941 | // VK_EXT_transform_feedback |
942 | PFN_vkCmdBindTransformFeedbackBuffersEXT vkCmdBindTransformFeedbackBuffersEXT = nullptr; |
943 | PFN_vkCmdBeginTransformFeedbackEXT vkCmdBeginTransformFeedbackEXT = nullptr; |
944 | PFN_vkCmdEndTransformFeedbackEXT vkCmdEndTransformFeedbackEXT = nullptr; |
945 | PFN_vkCmdBeginQueryIndexedEXT vkCmdBeginQueryIndexedEXT = nullptr; |
946 | PFN_vkCmdEndQueryIndexedEXT vkCmdEndQueryIndexedEXT = nullptr; |
947 | PFN_vkCmdDrawIndirectByteCountEXT vkCmdDrawIndirectByteCountEXT = nullptr; |
948 | |
949 | // VK_KHR_get_memory_requirements2 |
950 | PFN_vkGetBufferMemoryRequirements2KHR vkGetBufferMemoryRequirements2KHR = nullptr; |
951 | PFN_vkGetImageMemoryRequirements2KHR vkGetImageMemoryRequirements2KHR = nullptr; |
952 | |
953 | // VK_KHR_bind_memory2 |
954 | PFN_vkBindBufferMemory2KHR vkBindBufferMemory2KHR = nullptr; |
955 | PFN_vkBindImageMemory2KHR vkBindImageMemory2KHR = nullptr; |
956 | |
957 | // VK_KHR_external_fence_capabilities |
958 | PFN_vkGetPhysicalDeviceExternalFencePropertiesKHR vkGetPhysicalDeviceExternalFencePropertiesKHR = |
959 | nullptr; |
960 | |
961 | // VK_KHR_external_fence_fd |
962 | PFN_vkGetFenceFdKHR vkGetFenceFdKHR = nullptr; |
963 | PFN_vkImportFenceFdKHR vkImportFenceFdKHR = nullptr; |
964 | |
965 | // VK_KHR_external_semaphore_capabilities |
966 | PFN_vkGetPhysicalDeviceExternalSemaphorePropertiesKHR |
967 | vkGetPhysicalDeviceExternalSemaphorePropertiesKHR = nullptr; |
968 | |
969 | // VK_KHR_sampler_ycbcr_conversion |
970 | PFN_vkCreateSamplerYcbcrConversionKHR vkCreateSamplerYcbcrConversionKHR = nullptr; |
971 | PFN_vkDestroySamplerYcbcrConversionKHR vkDestroySamplerYcbcrConversionKHR = nullptr; |
972 | |
973 | // VK_KHR_create_renderpass2 |
974 | PFN_vkCreateRenderPass2KHR vkCreateRenderPass2KHR = nullptr; |
975 | |
976 | # if defined(ANGLE_PLATFORM_FUCHSIA) |
977 | // VK_FUCHSIA_imagepipe_surface |
978 | PFN_vkCreateImagePipeSurfaceFUCHSIA vkCreateImagePipeSurfaceFUCHSIA = nullptr; |
979 | # endif |
980 | |
981 | # if defined(ANGLE_PLATFORM_ANDROID) |
982 | PFN_vkGetAndroidHardwareBufferPropertiesANDROID vkGetAndroidHardwareBufferPropertiesANDROID = |
983 | nullptr; |
984 | PFN_vkGetMemoryAndroidHardwareBufferANDROID vkGetMemoryAndroidHardwareBufferANDROID = nullptr; |
985 | # endif |
986 | |
987 | # if defined(ANGLE_PLATFORM_GGP) |
988 | PFN_vkCreateStreamDescriptorSurfaceGGP vkCreateStreamDescriptorSurfaceGGP = nullptr; |
989 | # endif |
990 | |
991 | # define GET_INSTANCE_FUNC(vkName) \ |
992 | do \ |
993 | { \ |
994 | vkName = reinterpret_cast<PFN_##vkName>(vkGetInstanceProcAddr(instance, #vkName)); \ |
995 | ASSERT(vkName); \ |
996 | } while (0) |
997 | |
998 | # define GET_DEVICE_FUNC(vkName) \ |
999 | do \ |
1000 | { \ |
1001 | vkName = reinterpret_cast<PFN_##vkName>(vkGetDeviceProcAddr(device, #vkName)); \ |
1002 | ASSERT(vkName); \ |
1003 | } while (0) |
1004 | |
1005 | // VK_KHR_shared_presentable_image |
1006 | PFN_vkGetSwapchainStatusKHR vkGetSwapchainStatusKHR = nullptr; |
1007 | |
1008 | // VK_EXT_extended_dynamic_state |
1009 | PFN_vkCmdBindVertexBuffers2EXT vkCmdBindVertexBuffers2EXT = nullptr; |
1010 | PFN_vkCmdSetCullModeEXT vkCmdSetCullModeEXT = nullptr; |
1011 | PFN_vkCmdSetDepthBoundsTestEnableEXT vkCmdSetDepthBoundsTestEnableEXT = nullptr; |
1012 | PFN_vkCmdSetDepthCompareOpEXT vkCmdSetDepthCompareOpEXT = nullptr; |
1013 | PFN_vkCmdSetDepthTestEnableEXT vkCmdSetDepthTestEnableEXT = nullptr; |
1014 | PFN_vkCmdSetDepthWriteEnableEXT vkCmdSetDepthWriteEnableEXT = nullptr; |
1015 | PFN_vkCmdSetFrontFaceEXT vkCmdSetFrontFaceEXT = nullptr; |
1016 | PFN_vkCmdSetPrimitiveTopologyEXT vkCmdSetPrimitiveTopologyEXT = nullptr; |
1017 | PFN_vkCmdSetScissorWithCountEXT vkCmdSetScissorWithCountEXT = nullptr; |
1018 | PFN_vkCmdSetStencilOpEXT vkCmdSetStencilOpEXT = nullptr; |
1019 | PFN_vkCmdSetStencilTestEnableEXT vkCmdSetStencilTestEnableEXT = nullptr; |
1020 | PFN_vkCmdSetViewportWithCountEXT vkCmdSetViewportWithCountEXT = nullptr; |
1021 | |
1022 | // VK_EXT_extended_dynamic_state2 |
1023 | PFN_vkCmdSetDepthBiasEnableEXT vkCmdSetDepthBiasEnableEXT = nullptr; |
1024 | PFN_vkCmdSetLogicOpEXT vkCmdSetLogicOpEXT = nullptr; |
1025 | PFN_vkCmdSetPatchControlPointsEXT vkCmdSetPatchControlPointsEXT = nullptr; |
1026 | PFN_vkCmdSetPrimitiveRestartEnableEXT vkCmdSetPrimitiveRestartEnableEXT = nullptr; |
1027 | PFN_vkCmdSetRasterizerDiscardEnableEXT vkCmdSetRasterizerDiscardEnableEXT = nullptr; |
1028 | |
1029 | // VK_KHR_fragment_shading_rate |
1030 | PFN_vkGetPhysicalDeviceFragmentShadingRatesKHR vkGetPhysicalDeviceFragmentShadingRatesKHR = nullptr; |
1031 | PFN_vkCmdSetFragmentShadingRateKHR vkCmdSetFragmentShadingRateKHR = nullptr; |
1032 | |
1033 | // VK_GOOGLE_display_timing |
1034 | PFN_vkGetPastPresentationTimingGOOGLE vkGetPastPresentationTimingGOOGLE = nullptr; |
1035 | |
1036 | void InitDebugUtilsEXTFunctions(VkInstance instance) |
1037 | { |
1038 | GET_INSTANCE_FUNC(vkCreateDebugUtilsMessengerEXT); |
1039 | GET_INSTANCE_FUNC(vkDestroyDebugUtilsMessengerEXT); |
1040 | GET_INSTANCE_FUNC(vkCmdBeginDebugUtilsLabelEXT); |
1041 | GET_INSTANCE_FUNC(vkCmdEndDebugUtilsLabelEXT); |
1042 | GET_INSTANCE_FUNC(vkCmdInsertDebugUtilsLabelEXT); |
1043 | GET_INSTANCE_FUNC(vkSetDebugUtilsObjectNameEXT); |
1044 | } |
1045 | |
1046 | void InitGetPhysicalDeviceProperties2KHRFunctions(VkInstance instance) |
1047 | { |
1048 | GET_INSTANCE_FUNC(vkGetPhysicalDeviceProperties2KHR); |
1049 | GET_INSTANCE_FUNC(vkGetPhysicalDeviceFeatures2KHR); |
1050 | GET_INSTANCE_FUNC(vkGetPhysicalDeviceMemoryProperties2KHR); |
1051 | } |
1052 | |
1053 | void InitTransformFeedbackEXTFunctions(VkDevice device) |
1054 | { |
1055 | GET_DEVICE_FUNC(vkCmdBindTransformFeedbackBuffersEXT); |
1056 | GET_DEVICE_FUNC(vkCmdBeginTransformFeedbackEXT); |
1057 | GET_DEVICE_FUNC(vkCmdEndTransformFeedbackEXT); |
1058 | GET_DEVICE_FUNC(vkCmdBeginQueryIndexedEXT); |
1059 | GET_DEVICE_FUNC(vkCmdEndQueryIndexedEXT); |
1060 | GET_DEVICE_FUNC(vkCmdDrawIndirectByteCountEXT); |
1061 | } |
1062 | |
1063 | // VK_KHR_sampler_ycbcr_conversion |
1064 | void InitSamplerYcbcrKHRFunctions(VkDevice device) |
1065 | { |
1066 | GET_DEVICE_FUNC(vkCreateSamplerYcbcrConversionKHR); |
1067 | GET_DEVICE_FUNC(vkDestroySamplerYcbcrConversionKHR); |
1068 | } |
1069 | |
1070 | // VK_KHR_create_renderpass2 |
1071 | void InitRenderPass2KHRFunctions(VkDevice device) |
1072 | { |
1073 | GET_DEVICE_FUNC(vkCreateRenderPass2KHR); |
1074 | } |
1075 | |
1076 | # if defined(ANGLE_PLATFORM_FUCHSIA) |
1077 | void InitImagePipeSurfaceFUCHSIAFunctions(VkInstance instance) |
1078 | { |
1079 | GET_INSTANCE_FUNC(vkCreateImagePipeSurfaceFUCHSIA); |
1080 | } |
1081 | # endif |
1082 | |
1083 | # if defined(ANGLE_PLATFORM_ANDROID) |
1084 | void InitExternalMemoryHardwareBufferANDROIDFunctions(VkInstance instance) |
1085 | { |
1086 | GET_INSTANCE_FUNC(vkGetAndroidHardwareBufferPropertiesANDROID); |
1087 | GET_INSTANCE_FUNC(vkGetMemoryAndroidHardwareBufferANDROID); |
1088 | } |
1089 | # endif |
1090 | |
1091 | # if defined(ANGLE_PLATFORM_GGP) |
1092 | void InitGGPStreamDescriptorSurfaceFunctions(VkInstance instance) |
1093 | { |
1094 | GET_INSTANCE_FUNC(vkCreateStreamDescriptorSurfaceGGP); |
1095 | } |
1096 | # endif // defined(ANGLE_PLATFORM_GGP) |
1097 | |
1098 | void InitExternalSemaphoreFdFunctions(VkInstance instance) |
1099 | { |
1100 | GET_INSTANCE_FUNC(vkImportSemaphoreFdKHR); |
1101 | } |
1102 | |
1103 | void InitHostQueryResetFunctions(VkDevice device) |
1104 | { |
1105 | GET_DEVICE_FUNC(vkResetQueryPoolEXT); |
1106 | } |
1107 | |
1108 | // VK_KHR_get_memory_requirements2 |
1109 | void InitGetMemoryRequirements2KHRFunctions(VkDevice device) |
1110 | { |
1111 | GET_DEVICE_FUNC(vkGetBufferMemoryRequirements2KHR); |
1112 | GET_DEVICE_FUNC(vkGetImageMemoryRequirements2KHR); |
1113 | } |
1114 | |
1115 | // VK_KHR_bind_memory2 |
1116 | void InitBindMemory2KHRFunctions(VkDevice device) |
1117 | { |
1118 | GET_DEVICE_FUNC(vkBindBufferMemory2KHR); |
1119 | GET_DEVICE_FUNC(vkBindImageMemory2KHR); |
1120 | } |
1121 | |
1122 | // VK_KHR_external_fence_capabilities |
1123 | void InitExternalFenceCapabilitiesFunctions(VkInstance instance) |
1124 | { |
1125 | GET_INSTANCE_FUNC(vkGetPhysicalDeviceExternalFencePropertiesKHR); |
1126 | } |
1127 | |
1128 | // VK_KHR_external_fence_fd |
1129 | void InitExternalFenceFdFunctions(VkInstance instance) |
1130 | { |
1131 | GET_INSTANCE_FUNC(vkGetFenceFdKHR); |
1132 | GET_INSTANCE_FUNC(vkImportFenceFdKHR); |
1133 | } |
1134 | |
1135 | // VK_KHR_external_semaphore_capabilities |
1136 | void InitExternalSemaphoreCapabilitiesFunctions(VkInstance instance) |
1137 | { |
1138 | GET_INSTANCE_FUNC(vkGetPhysicalDeviceExternalSemaphorePropertiesKHR); |
1139 | } |
1140 | |
1141 | // VK_KHR_shared_presentable_image |
1142 | void InitGetSwapchainStatusKHRFunctions(VkDevice device) |
1143 | { |
1144 | GET_DEVICE_FUNC(vkGetSwapchainStatusKHR); |
1145 | } |
1146 | |
1147 | // VK_EXT_extended_dynamic_state |
1148 | void InitExtendedDynamicStateEXTFunctions(VkDevice device) |
1149 | { |
1150 | GET_DEVICE_FUNC(vkCmdBindVertexBuffers2EXT); |
1151 | GET_DEVICE_FUNC(vkCmdSetCullModeEXT); |
1152 | GET_DEVICE_FUNC(vkCmdSetDepthBoundsTestEnableEXT); |
1153 | GET_DEVICE_FUNC(vkCmdSetDepthCompareOpEXT); |
1154 | GET_DEVICE_FUNC(vkCmdSetDepthTestEnableEXT); |
1155 | GET_DEVICE_FUNC(vkCmdSetDepthWriteEnableEXT); |
1156 | GET_DEVICE_FUNC(vkCmdSetFrontFaceEXT); |
1157 | GET_DEVICE_FUNC(vkCmdSetPrimitiveTopologyEXT); |
1158 | GET_DEVICE_FUNC(vkCmdSetScissorWithCountEXT); |
1159 | GET_DEVICE_FUNC(vkCmdSetStencilOpEXT); |
1160 | GET_DEVICE_FUNC(vkCmdSetStencilTestEnableEXT); |
1161 | GET_DEVICE_FUNC(vkCmdSetViewportWithCountEXT); |
1162 | } |
1163 | |
1164 | // VK_EXT_extended_dynamic_state2 |
1165 | void InitExtendedDynamicState2EXTFunctions(VkDevice device) |
1166 | { |
1167 | GET_DEVICE_FUNC(vkCmdSetDepthBiasEnableEXT); |
1168 | GET_DEVICE_FUNC(vkCmdSetLogicOpEXT); |
1169 | GET_DEVICE_FUNC(vkCmdSetPatchControlPointsEXT); |
1170 | GET_DEVICE_FUNC(vkCmdSetPrimitiveRestartEnableEXT); |
1171 | GET_DEVICE_FUNC(vkCmdSetRasterizerDiscardEnableEXT); |
1172 | } |
1173 | |
1174 | // VK_KHR_fragment_shading_rate |
1175 | void InitFragmentShadingRateKHRInstanceFunction(VkInstance instance) |
1176 | { |
1177 | GET_INSTANCE_FUNC(vkGetPhysicalDeviceFragmentShadingRatesKHR); |
1178 | } |
1179 | |
1180 | void InitFragmentShadingRateKHRDeviceFunction(VkDevice device) |
1181 | { |
1182 | GET_DEVICE_FUNC(vkCmdSetFragmentShadingRateKHR); |
1183 | } |
1184 | |
1185 | // VK_GOOGLE_display_timing |
1186 | void InitGetPastPresentationTimingGoogleFunction(VkDevice device) |
1187 | { |
1188 | GET_DEVICE_FUNC(vkGetPastPresentationTimingGOOGLE); |
1189 | } |
1190 | |
1191 | # undef GET_INSTANCE_FUNC |
1192 | # undef GET_DEVICE_FUNC |
1193 | |
1194 | #endif // !defined(ANGLE_SHARED_LIBVULKAN) |
1195 | |
1196 | #define ASSIGN_FROM_CORE(vkName, EXT) \ |
1197 | do \ |
1198 | { \ |
1199 | /* The core entry point must be present */ \ |
1200 | ASSERT(vkName != nullptr); \ |
1201 | vkName##EXT = vkName; \ |
1202 | } while (0) |
1203 | |
1204 | void InitGetPhysicalDeviceProperties2KHRFunctionsFromCore() |
1205 | { |
1206 | ASSIGN_FROM_CORE(vkGetPhysicalDeviceProperties2, KHR); |
1207 | ASSIGN_FROM_CORE(vkGetPhysicalDeviceFeatures2, KHR); |
1208 | ASSIGN_FROM_CORE(vkGetPhysicalDeviceMemoryProperties2, KHR); |
1209 | } |
1210 | |
1211 | void InitExternalFenceCapabilitiesFunctionsFromCore() |
1212 | { |
1213 | ASSIGN_FROM_CORE(vkGetPhysicalDeviceExternalFenceProperties, KHR); |
1214 | } |
1215 | |
1216 | void InitExternalSemaphoreCapabilitiesFunctionsFromCore() |
1217 | { |
1218 | ASSIGN_FROM_CORE(vkGetPhysicalDeviceExternalSemaphoreProperties, KHR); |
1219 | } |
1220 | |
1221 | void InitSamplerYcbcrKHRFunctionsFromCore() |
1222 | { |
1223 | ASSIGN_FROM_CORE(vkCreateSamplerYcbcrConversion, KHR); |
1224 | ASSIGN_FROM_CORE(vkDestroySamplerYcbcrConversion, KHR); |
1225 | } |
1226 | |
1227 | void InitGetMemoryRequirements2KHRFunctionsFromCore() |
1228 | { |
1229 | ASSIGN_FROM_CORE(vkGetBufferMemoryRequirements2, KHR); |
1230 | ASSIGN_FROM_CORE(vkGetImageMemoryRequirements2, KHR); |
1231 | } |
1232 | |
1233 | void InitBindMemory2KHRFunctionsFromCore() |
1234 | { |
1235 | ASSIGN_FROM_CORE(vkBindBufferMemory2, KHR); |
1236 | ASSIGN_FROM_CORE(vkBindImageMemory2, KHR); |
1237 | } |
1238 | |
1239 | #undef ASSIGN_FROM_CORE |
1240 | |
1241 | GLenum CalculateGenerateMipmapFilter(ContextVk *contextVk, angle::FormatID formatID) |
1242 | { |
1243 | const bool formatSupportsLinearFiltering = contextVk->getRenderer()->hasImageFormatFeatureBits( |
1244 | format: formatID, featureBits: VK_FORMAT_FEATURE_SAMPLED_IMAGE_FILTER_LINEAR_BIT); |
1245 | const bool hintFastest = contextVk->getState().getGenerateMipmapHint() == GL_FASTEST; |
1246 | |
1247 | return formatSupportsLinearFiltering && !hintFastest ? GL_LINEAR : GL_NEAREST; |
1248 | } |
1249 | |
1250 | // Return the log of samples. Assumes |sampleCount| is a power of 2. The result can be used to |
1251 | // index an array based on sample count. See for example TextureVk::PerSampleCountArray. |
1252 | size_t PackSampleCount(GLint sampleCount) |
1253 | { |
1254 | if (sampleCount == 0) |
1255 | { |
1256 | sampleCount = 1; |
1257 | } |
1258 | |
1259 | // We currently only support up to 16xMSAA. |
1260 | ASSERT(sampleCount <= VK_SAMPLE_COUNT_16_BIT); |
1261 | ASSERT(gl::isPow2(sampleCount)); |
1262 | return gl::ScanForward(bits: static_cast<uint32_t>(sampleCount)); |
1263 | } |
1264 | |
1265 | namespace gl_vk |
1266 | { |
1267 | |
1268 | VkFilter GetFilter(const GLenum filter) |
1269 | { |
1270 | switch (filter) |
1271 | { |
1272 | case GL_LINEAR_MIPMAP_LINEAR: |
1273 | case GL_LINEAR_MIPMAP_NEAREST: |
1274 | case GL_LINEAR: |
1275 | return VK_FILTER_LINEAR; |
1276 | case GL_NEAREST_MIPMAP_LINEAR: |
1277 | case GL_NEAREST_MIPMAP_NEAREST: |
1278 | case GL_NEAREST: |
1279 | return VK_FILTER_NEAREST; |
1280 | default: |
1281 | UNIMPLEMENTED(); |
1282 | return VK_FILTER_MAX_ENUM; |
1283 | } |
1284 | } |
1285 | |
1286 | VkSamplerMipmapMode GetSamplerMipmapMode(const GLenum filter) |
1287 | { |
1288 | switch (filter) |
1289 | { |
1290 | case GL_LINEAR_MIPMAP_LINEAR: |
1291 | case GL_NEAREST_MIPMAP_LINEAR: |
1292 | return VK_SAMPLER_MIPMAP_MODE_LINEAR; |
1293 | case GL_LINEAR: |
1294 | case GL_NEAREST: |
1295 | case GL_NEAREST_MIPMAP_NEAREST: |
1296 | case GL_LINEAR_MIPMAP_NEAREST: |
1297 | return VK_SAMPLER_MIPMAP_MODE_NEAREST; |
1298 | default: |
1299 | UNIMPLEMENTED(); |
1300 | return VK_SAMPLER_MIPMAP_MODE_MAX_ENUM; |
1301 | } |
1302 | } |
1303 | |
1304 | VkSamplerAddressMode GetSamplerAddressMode(const GLenum wrap) |
1305 | { |
1306 | switch (wrap) |
1307 | { |
1308 | case GL_REPEAT: |
1309 | return VK_SAMPLER_ADDRESS_MODE_REPEAT; |
1310 | case GL_MIRRORED_REPEAT: |
1311 | return VK_SAMPLER_ADDRESS_MODE_MIRRORED_REPEAT; |
1312 | case GL_CLAMP_TO_BORDER: |
1313 | return VK_SAMPLER_ADDRESS_MODE_CLAMP_TO_BORDER; |
1314 | case GL_CLAMP_TO_EDGE: |
1315 | return VK_SAMPLER_ADDRESS_MODE_CLAMP_TO_EDGE; |
1316 | case GL_MIRROR_CLAMP_TO_EDGE_EXT: |
1317 | return VK_SAMPLER_ADDRESS_MODE_MIRROR_CLAMP_TO_EDGE; |
1318 | default: |
1319 | UNIMPLEMENTED(); |
1320 | return VK_SAMPLER_ADDRESS_MODE_MAX_ENUM; |
1321 | } |
1322 | } |
1323 | |
1324 | VkRect2D GetRect(const gl::Rectangle &source) |
1325 | { |
1326 | return {.offset: {.x: source.x, .y: source.y}, |
1327 | .extent: {.width: static_cast<uint32_t>(source.width), .height: static_cast<uint32_t>(source.height)}}; |
1328 | } |
1329 | |
1330 | VkPrimitiveTopology GetPrimitiveTopology(gl::PrimitiveMode mode) |
1331 | { |
1332 | switch (mode) |
1333 | { |
1334 | case gl::PrimitiveMode::Triangles: |
1335 | return VK_PRIMITIVE_TOPOLOGY_TRIANGLE_LIST; |
1336 | case gl::PrimitiveMode::Points: |
1337 | return VK_PRIMITIVE_TOPOLOGY_POINT_LIST; |
1338 | case gl::PrimitiveMode::Lines: |
1339 | return VK_PRIMITIVE_TOPOLOGY_LINE_LIST; |
1340 | case gl::PrimitiveMode::LineStrip: |
1341 | return VK_PRIMITIVE_TOPOLOGY_LINE_STRIP; |
1342 | case gl::PrimitiveMode::TriangleFan: |
1343 | return VK_PRIMITIVE_TOPOLOGY_TRIANGLE_FAN; |
1344 | case gl::PrimitiveMode::TriangleStrip: |
1345 | return VK_PRIMITIVE_TOPOLOGY_TRIANGLE_STRIP; |
1346 | case gl::PrimitiveMode::LineLoop: |
1347 | return VK_PRIMITIVE_TOPOLOGY_LINE_STRIP; |
1348 | case gl::PrimitiveMode::LinesAdjacency: |
1349 | return VK_PRIMITIVE_TOPOLOGY_LINE_LIST_WITH_ADJACENCY; |
1350 | case gl::PrimitiveMode::LineStripAdjacency: |
1351 | return VK_PRIMITIVE_TOPOLOGY_LINE_STRIP_WITH_ADJACENCY; |
1352 | case gl::PrimitiveMode::TrianglesAdjacency: |
1353 | return VK_PRIMITIVE_TOPOLOGY_TRIANGLE_LIST_WITH_ADJACENCY; |
1354 | case gl::PrimitiveMode::TriangleStripAdjacency: |
1355 | return VK_PRIMITIVE_TOPOLOGY_TRIANGLE_STRIP_WITH_ADJACENCY; |
1356 | case gl::PrimitiveMode::Patches: |
1357 | return VK_PRIMITIVE_TOPOLOGY_PATCH_LIST; |
1358 | default: |
1359 | UNREACHABLE(); |
1360 | return VK_PRIMITIVE_TOPOLOGY_POINT_LIST; |
1361 | } |
1362 | } |
1363 | |
1364 | VkPolygonMode GetPolygonMode(const gl::PolygonMode polygonMode) |
1365 | { |
1366 | switch (polygonMode) |
1367 | { |
1368 | case gl::PolygonMode::Point: |
1369 | return VK_POLYGON_MODE_POINT; |
1370 | case gl::PolygonMode::Line: |
1371 | return VK_POLYGON_MODE_LINE; |
1372 | case gl::PolygonMode::Fill: |
1373 | return VK_POLYGON_MODE_FILL; |
1374 | default: |
1375 | UNREACHABLE(); |
1376 | return VK_POLYGON_MODE_FILL; |
1377 | } |
1378 | } |
1379 | |
1380 | VkCullModeFlagBits GetCullMode(const gl::RasterizerState &rasterState) |
1381 | { |
1382 | if (!rasterState.cullFace) |
1383 | { |
1384 | return VK_CULL_MODE_NONE; |
1385 | } |
1386 | |
1387 | switch (rasterState.cullMode) |
1388 | { |
1389 | case gl::CullFaceMode::Front: |
1390 | return VK_CULL_MODE_FRONT_BIT; |
1391 | case gl::CullFaceMode::Back: |
1392 | return VK_CULL_MODE_BACK_BIT; |
1393 | case gl::CullFaceMode::FrontAndBack: |
1394 | return VK_CULL_MODE_FRONT_AND_BACK; |
1395 | default: |
1396 | UNREACHABLE(); |
1397 | return VK_CULL_MODE_NONE; |
1398 | } |
1399 | } |
1400 | |
1401 | VkFrontFace GetFrontFace(GLenum frontFace, bool invertCullFace) |
1402 | { |
1403 | // Invert CW and CCW to have the same behavior as OpenGL. |
1404 | switch (frontFace) |
1405 | { |
1406 | case GL_CW: |
1407 | return invertCullFace ? VK_FRONT_FACE_CLOCKWISE : VK_FRONT_FACE_COUNTER_CLOCKWISE; |
1408 | case GL_CCW: |
1409 | return invertCullFace ? VK_FRONT_FACE_COUNTER_CLOCKWISE : VK_FRONT_FACE_CLOCKWISE; |
1410 | default: |
1411 | UNREACHABLE(); |
1412 | return VK_FRONT_FACE_CLOCKWISE; |
1413 | } |
1414 | } |
1415 | |
1416 | VkSampleCountFlagBits GetSamples(GLint sampleCount, bool limitSampleCountTo2) |
1417 | { |
1418 | if (limitSampleCountTo2) |
1419 | { |
1420 | // Limiting samples to 2 allows multisampling to work while reducing |
1421 | // how much graphics memory is required. This makes ANGLE nonconformant |
1422 | // (GLES 3.0+ requires 4 samples minimum) but gives low memory systems a |
1423 | // better chance of running applications. |
1424 | sampleCount = std::min(a: sampleCount, b: 2); |
1425 | } |
1426 | |
1427 | switch (sampleCount) |
1428 | { |
1429 | case 0: |
1430 | UNREACHABLE(); |
1431 | return VK_SAMPLE_COUNT_1_BIT; |
1432 | case 1: |
1433 | return VK_SAMPLE_COUNT_1_BIT; |
1434 | case 2: |
1435 | return VK_SAMPLE_COUNT_2_BIT; |
1436 | case 4: |
1437 | return VK_SAMPLE_COUNT_4_BIT; |
1438 | case 8: |
1439 | return VK_SAMPLE_COUNT_8_BIT; |
1440 | case 16: |
1441 | return VK_SAMPLE_COUNT_16_BIT; |
1442 | case 32: |
1443 | return VK_SAMPLE_COUNT_32_BIT; |
1444 | default: |
1445 | UNREACHABLE(); |
1446 | return VK_SAMPLE_COUNT_FLAG_BITS_MAX_ENUM; |
1447 | } |
1448 | } |
1449 | |
1450 | VkComponentSwizzle GetSwizzle(const GLenum swizzle) |
1451 | { |
1452 | switch (swizzle) |
1453 | { |
1454 | case GL_ALPHA: |
1455 | return VK_COMPONENT_SWIZZLE_A; |
1456 | case GL_RED: |
1457 | return VK_COMPONENT_SWIZZLE_R; |
1458 | case GL_GREEN: |
1459 | return VK_COMPONENT_SWIZZLE_G; |
1460 | case GL_BLUE: |
1461 | return VK_COMPONENT_SWIZZLE_B; |
1462 | case GL_ZERO: |
1463 | return VK_COMPONENT_SWIZZLE_ZERO; |
1464 | case GL_ONE: |
1465 | return VK_COMPONENT_SWIZZLE_ONE; |
1466 | default: |
1467 | UNREACHABLE(); |
1468 | return VK_COMPONENT_SWIZZLE_IDENTITY; |
1469 | } |
1470 | } |
1471 | |
1472 | VkCompareOp GetCompareOp(const GLenum compareFunc) |
1473 | { |
1474 | switch (compareFunc) |
1475 | { |
1476 | case GL_NEVER: |
1477 | return VK_COMPARE_OP_NEVER; |
1478 | case GL_LESS: |
1479 | return VK_COMPARE_OP_LESS; |
1480 | case GL_EQUAL: |
1481 | return VK_COMPARE_OP_EQUAL; |
1482 | case GL_LEQUAL: |
1483 | return VK_COMPARE_OP_LESS_OR_EQUAL; |
1484 | case GL_GREATER: |
1485 | return VK_COMPARE_OP_GREATER; |
1486 | case GL_NOTEQUAL: |
1487 | return VK_COMPARE_OP_NOT_EQUAL; |
1488 | case GL_GEQUAL: |
1489 | return VK_COMPARE_OP_GREATER_OR_EQUAL; |
1490 | case GL_ALWAYS: |
1491 | return VK_COMPARE_OP_ALWAYS; |
1492 | default: |
1493 | UNREACHABLE(); |
1494 | return VK_COMPARE_OP_ALWAYS; |
1495 | } |
1496 | } |
1497 | |
1498 | VkStencilOp GetStencilOp(GLenum compareOp) |
1499 | { |
1500 | switch (compareOp) |
1501 | { |
1502 | case GL_KEEP: |
1503 | return VK_STENCIL_OP_KEEP; |
1504 | case GL_ZERO: |
1505 | return VK_STENCIL_OP_ZERO; |
1506 | case GL_REPLACE: |
1507 | return VK_STENCIL_OP_REPLACE; |
1508 | case GL_INCR: |
1509 | return VK_STENCIL_OP_INCREMENT_AND_CLAMP; |
1510 | case GL_DECR: |
1511 | return VK_STENCIL_OP_DECREMENT_AND_CLAMP; |
1512 | case GL_INCR_WRAP: |
1513 | return VK_STENCIL_OP_INCREMENT_AND_WRAP; |
1514 | case GL_DECR_WRAP: |
1515 | return VK_STENCIL_OP_DECREMENT_AND_WRAP; |
1516 | case GL_INVERT: |
1517 | return VK_STENCIL_OP_INVERT; |
1518 | default: |
1519 | UNREACHABLE(); |
1520 | return VK_STENCIL_OP_KEEP; |
1521 | } |
1522 | } |
1523 | |
1524 | VkLogicOp GetLogicOp(const GLenum logicOp) |
1525 | { |
1526 | // GL's logic op values are 0x1500 + op, where op is the same value as Vulkan's VkLogicOp. |
1527 | return static_cast<VkLogicOp>(logicOp - GL_CLEAR); |
1528 | } |
1529 | |
1530 | void GetOffset(const gl::Offset &glOffset, VkOffset3D *vkOffset) |
1531 | { |
1532 | vkOffset->x = glOffset.x; |
1533 | vkOffset->y = glOffset.y; |
1534 | vkOffset->z = glOffset.z; |
1535 | } |
1536 | |
1537 | void GetExtent(const gl::Extents &glExtent, VkExtent3D *vkExtent) |
1538 | { |
1539 | vkExtent->width = glExtent.width; |
1540 | vkExtent->height = glExtent.height; |
1541 | vkExtent->depth = glExtent.depth; |
1542 | } |
1543 | |
1544 | VkImageType GetImageType(gl::TextureType textureType) |
1545 | { |
1546 | switch (textureType) |
1547 | { |
1548 | case gl::TextureType::_2D: |
1549 | case gl::TextureType::_2DArray: |
1550 | case gl::TextureType::_2DMultisample: |
1551 | case gl::TextureType::_2DMultisampleArray: |
1552 | case gl::TextureType::CubeMap: |
1553 | case gl::TextureType::CubeMapArray: |
1554 | case gl::TextureType::External: |
1555 | return VK_IMAGE_TYPE_2D; |
1556 | case gl::TextureType::_3D: |
1557 | return VK_IMAGE_TYPE_3D; |
1558 | default: |
1559 | // We will need to implement all the texture types for ES3+. |
1560 | UNIMPLEMENTED(); |
1561 | return VK_IMAGE_TYPE_MAX_ENUM; |
1562 | } |
1563 | } |
1564 | |
1565 | VkImageViewType GetImageViewType(gl::TextureType textureType) |
1566 | { |
1567 | switch (textureType) |
1568 | { |
1569 | case gl::TextureType::_2D: |
1570 | case gl::TextureType::_2DMultisample: |
1571 | case gl::TextureType::External: |
1572 | return VK_IMAGE_VIEW_TYPE_2D; |
1573 | case gl::TextureType::_2DArray: |
1574 | case gl::TextureType::_2DMultisampleArray: |
1575 | return VK_IMAGE_VIEW_TYPE_2D_ARRAY; |
1576 | case gl::TextureType::_3D: |
1577 | return VK_IMAGE_VIEW_TYPE_3D; |
1578 | case gl::TextureType::CubeMap: |
1579 | return VK_IMAGE_VIEW_TYPE_CUBE; |
1580 | case gl::TextureType::CubeMapArray: |
1581 | return VK_IMAGE_VIEW_TYPE_CUBE_ARRAY; |
1582 | default: |
1583 | // We will need to implement all the texture types for ES3+. |
1584 | UNIMPLEMENTED(); |
1585 | return VK_IMAGE_VIEW_TYPE_MAX_ENUM; |
1586 | } |
1587 | } |
1588 | |
1589 | VkColorComponentFlags GetColorComponentFlags(bool red, bool green, bool blue, bool alpha) |
1590 | { |
1591 | return (red ? VK_COLOR_COMPONENT_R_BIT : 0) | (green ? VK_COLOR_COMPONENT_G_BIT : 0) | |
1592 | (blue ? VK_COLOR_COMPONENT_B_BIT : 0) | (alpha ? VK_COLOR_COMPONENT_A_BIT : 0); |
1593 | } |
1594 | |
1595 | VkShaderStageFlags GetShaderStageFlags(gl::ShaderBitSet activeShaders) |
1596 | { |
1597 | VkShaderStageFlags flags = 0; |
1598 | for (const gl::ShaderType shaderType : activeShaders) |
1599 | { |
1600 | flags |= kShaderStageMap[shaderType]; |
1601 | } |
1602 | return flags; |
1603 | } |
1604 | |
1605 | void GetViewport(const gl::Rectangle &viewport, |
1606 | float nearPlane, |
1607 | float farPlane, |
1608 | bool invertViewport, |
1609 | bool clipSpaceOriginUpperLeft, |
1610 | GLint renderAreaHeight, |
1611 | VkViewport *viewportOut) |
1612 | { |
1613 | viewportOut->x = static_cast<float>(viewport.x); |
1614 | viewportOut->y = static_cast<float>(viewport.y); |
1615 | viewportOut->width = static_cast<float>(viewport.width); |
1616 | viewportOut->height = static_cast<float>(viewport.height); |
1617 | viewportOut->minDepth = gl::clamp01(x: nearPlane); |
1618 | viewportOut->maxDepth = gl::clamp01(x: farPlane); |
1619 | |
1620 | // Say an application intends to draw a primitive (shown as 'o' below), it can choose to use |
1621 | // different clip space origin. When clip space origin (shown as 'C' below) is switched from |
1622 | // lower-left to upper-left, primitives will be rendered with its y-coordinate flipped. |
1623 | |
1624 | // Rendered content will differ based on whether it is a default framebuffer or a user defined |
1625 | // framebuffer. We modify the viewport's 'y' and 'h' accordingly. |
1626 | |
1627 | // clip space origin is lower-left |
1628 | // Expected draw in GLES default framebuffer user defined framebuffer |
1629 | // (0,H) (0,0) (0,0) |
1630 | // + +-----------+ (W,0) +-----------+ (W,0) |
1631 | // | | | C----+ |
1632 | // | | | | | (h) |
1633 | // | +----+ | +----+ | | O | |
1634 | // | | O | | | O | (-h) | +----+ |
1635 | // | | | | | | | |
1636 | // | C----+ | C----+ | |
1637 | // +-----------+ (W,0) + + |
1638 | // (0,0) (0,H) (0,H) |
1639 | // y' = H - h y' = y |
1640 | |
1641 | // clip space origin is upper-left |
1642 | // Expected draw in GLES default framebuffer user defined framebuffer |
1643 | // (0,H) (0,0) (0,0) |
1644 | // + +-----------+ (W,0) +-----------+ (W,0) |
1645 | // | | | +----+ |
1646 | // | | | | O | (-h) |
1647 | // | C----+ | C----+ | | | |
1648 | // | | | | | | (h) | C----+ |
1649 | // | | O | | | O | | |
1650 | // | +----+ | +----+ | |
1651 | // +-----------+ (W,0) + + |
1652 | // (0,0) (0,H) (0,H) |
1653 | // y' = H - (y + h) y' = y + H |
1654 | |
1655 | if (clipSpaceOriginUpperLeft) |
1656 | { |
1657 | if (invertViewport) |
1658 | { |
1659 | viewportOut->y = static_cast<float>(renderAreaHeight - (viewport.height + viewport.y)); |
1660 | } |
1661 | else |
1662 | { |
1663 | viewportOut->y = static_cast<float>(viewport.height + viewport.y); |
1664 | viewportOut->height = -viewportOut->height; |
1665 | } |
1666 | } |
1667 | else |
1668 | { |
1669 | if (invertViewport) |
1670 | { |
1671 | viewportOut->y = static_cast<float>(renderAreaHeight - viewport.y); |
1672 | viewportOut->height = -viewportOut->height; |
1673 | } |
1674 | } |
1675 | } |
1676 | |
1677 | void GetExtentsAndLayerCount(gl::TextureType textureType, |
1678 | const gl::Extents &extents, |
1679 | VkExtent3D *extentsOut, |
1680 | uint32_t *layerCountOut) |
1681 | { |
1682 | extentsOut->width = extents.width; |
1683 | extentsOut->height = extents.height; |
1684 | |
1685 | switch (textureType) |
1686 | { |
1687 | case gl::TextureType::CubeMap: |
1688 | extentsOut->depth = 1; |
1689 | *layerCountOut = gl::kCubeFaceCount; |
1690 | break; |
1691 | |
1692 | case gl::TextureType::_2DArray: |
1693 | case gl::TextureType::_2DMultisampleArray: |
1694 | case gl::TextureType::CubeMapArray: |
1695 | extentsOut->depth = 1; |
1696 | *layerCountOut = extents.depth; |
1697 | break; |
1698 | |
1699 | default: |
1700 | extentsOut->depth = extents.depth; |
1701 | *layerCountOut = 1; |
1702 | break; |
1703 | } |
1704 | } |
1705 | |
1706 | vk::LevelIndex GetLevelIndex(gl::LevelIndex levelGL, gl::LevelIndex baseLevel) |
1707 | { |
1708 | ASSERT(baseLevel <= levelGL); |
1709 | return vk::LevelIndex(levelGL.get() - baseLevel.get()); |
1710 | } |
1711 | |
1712 | } // namespace gl_vk |
1713 | |
1714 | namespace vk_gl |
1715 | { |
1716 | void AddSampleCounts(VkSampleCountFlags sampleCounts, gl::SupportedSampleSet *setOut) |
1717 | { |
1718 | // The possible bits are VK_SAMPLE_COUNT_n_BIT = n, with n = 1 << b. At the time of this |
1719 | // writing, b is in [0, 6], however, we test all 32 bits in case the enum is extended. |
1720 | for (size_t bit : angle::BitSet32<32>(sampleCounts & kSupportedSampleCounts)) |
1721 | { |
1722 | setOut->insert(v: static_cast<GLuint>(1 << bit)); |
1723 | } |
1724 | } |
1725 | |
1726 | GLuint GetMaxSampleCount(VkSampleCountFlags sampleCounts) |
1727 | { |
1728 | GLuint maxCount = 0; |
1729 | for (size_t bit : angle::BitSet32<32>(sampleCounts & kSupportedSampleCounts)) |
1730 | { |
1731 | maxCount = static_cast<GLuint>(1 << bit); |
1732 | } |
1733 | return maxCount; |
1734 | } |
1735 | |
1736 | GLuint GetSampleCount(VkSampleCountFlags supportedCounts, GLuint requestedCount) |
1737 | { |
1738 | for (size_t bit : angle::BitSet32<32>(supportedCounts & kSupportedSampleCounts)) |
1739 | { |
1740 | GLuint sampleCount = static_cast<GLuint>(1 << bit); |
1741 | if (sampleCount >= requestedCount) |
1742 | { |
1743 | return sampleCount; |
1744 | } |
1745 | } |
1746 | |
1747 | UNREACHABLE(); |
1748 | return 0; |
1749 | } |
1750 | |
1751 | gl::LevelIndex GetLevelIndex(vk::LevelIndex levelVk, gl::LevelIndex baseLevel) |
1752 | { |
1753 | return gl::LevelIndex(levelVk.get() + baseLevel.get()); |
1754 | } |
1755 | } // namespace vk_gl |
1756 | } // namespace rx |
1757 |
Definitions
- ToEGL
- kNonZeroInitValue
- GetStagingBufferUsageFlags
- FindCompatibleMemory
- FindAndAllocateCompatibleMemory
- AllocateAndBindBufferOrImageMemory
- AllocateAndBindBufferOrImageMemory
- AllocateBufferOrImageMemory
- kVkKhronosValidationLayerName
- kVkStandardValidationLayerName
- kVkValidationLayerNames
- VulkanResultString
- GetAvailableValidationLayers
- gLoaderLayersPathEnv
- gLoaderICDFilenamesEnv
- GetDepthStencilAspectFlags
- GetFormatAspectFlags
- Context
- ~Context
- getDevice
- getFeatures
- MemoryProperties
- init
- destroy
- hasLazilyAllocatedMemory
- findCompatibleMemoryIndex
- StagingBuffer
- destroy
- init
- release
- collectGarbage
- InitMappableAllocation
- AllocateBufferMemory
- AllocateImageMemory
- AllocateImageMemoryWithRequirements
- AllocateBufferMemoryWithRequirements
- InitShaderModule
- Get2DTextureType
- GarbageObject
- GarbageObject
- GarbageObject
- operator=
- destroy
- MakeDebugUtilsLabel
- SetDebugUtilsObjectName
- ClearValuesArray
- ~ClearValuesArray
- ClearValuesArray
- operator=
- store
- storeNoDepthStencil
- getColorMask
- ResourceSerialFactory
- ~ResourceSerialFactory
- issueSerial
- ClampViewport
- ApplyPipelineCreationFeedback
- hash
- InitGetPhysicalDeviceProperties2KHRFunctionsFromCore
- InitExternalFenceCapabilitiesFunctionsFromCore
- InitExternalSemaphoreCapabilitiesFunctionsFromCore
- InitSamplerYcbcrKHRFunctionsFromCore
- InitGetMemoryRequirements2KHRFunctionsFromCore
- InitBindMemory2KHRFunctionsFromCore
- CalculateGenerateMipmapFilter
- PackSampleCount
- GetFilter
- GetSamplerMipmapMode
- GetSamplerAddressMode
- GetRect
- GetPrimitiveTopology
- GetPolygonMode
- GetCullMode
- GetFrontFace
- GetSamples
- GetSwizzle
- GetCompareOp
- GetStencilOp
- GetLogicOp
- GetOffset
- GetExtent
- GetImageType
- GetImageViewType
- GetColorComponentFlags
- GetShaderStageFlags
- GetViewport
- GetExtentsAndLayerCount
- GetLevelIndex
- AddSampleCounts
- GetMaxSampleCount
- GetSampleCount
Learn more about Flutter for embedded and desktop on industrialflutter.com