| 1 | // |
| 2 | // Copyright (c) 2017-2022 Advanced Micro Devices, Inc. All rights reserved. |
| 3 | // |
| 4 | // Permission is hereby granted, free of charge, to any person obtaining a copy |
| 5 | // of this software and associated documentation files (the "Software"), to deal |
| 6 | // in the Software without restriction, including without limitation the rights |
| 7 | // to use, copy, modify, merge, publish, distribute, sublicense, and/or sell |
| 8 | // copies of the Software, and to permit persons to whom the Software is |
| 9 | // furnished to do so, subject to the following conditions: |
| 10 | // |
| 11 | // The above copyright notice and this permission notice shall be included in |
| 12 | // all copies or substantial portions of the Software. |
| 13 | // |
| 14 | // THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR |
| 15 | // IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, |
| 16 | // FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE |
| 17 | // AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER |
| 18 | // LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, |
| 19 | // OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN |
| 20 | // THE SOFTWARE. |
| 21 | // |
| 22 | |
| 23 | #ifndef AMD_VULKAN_MEMORY_ALLOCATOR_H |
| 24 | #define AMD_VULKAN_MEMORY_ALLOCATOR_H |
| 25 | |
| 26 | /** \mainpage Vulkan Memory Allocator |
| 27 | |
| 28 | <b>Version 3.0.1 (2022-05-26)</b> |
| 29 | |
| 30 | Copyright (c) 2017-2022 Advanced Micro Devices, Inc. All rights reserved. \n |
| 31 | License: MIT |
| 32 | |
| 33 | <b>API documentation divided into groups:</b> [Modules](modules.html) |
| 34 | |
| 35 | \section main_table_of_contents Table of contents |
| 36 | |
| 37 | - <b>User guide</b> |
| 38 | - \subpage quick_start |
| 39 | - [Project setup](@ref quick_start_project_setup) |
| 40 | - [Initialization](@ref quick_start_initialization) |
| 41 | - [Resource allocation](@ref quick_start_resource_allocation) |
| 42 | - \subpage choosing_memory_type |
| 43 | - [Usage](@ref choosing_memory_type_usage) |
| 44 | - [Required and preferred flags](@ref choosing_memory_type_required_preferred_flags) |
| 45 | - [Explicit memory types](@ref choosing_memory_type_explicit_memory_types) |
| 46 | - [Custom memory pools](@ref choosing_memory_type_custom_memory_pools) |
| 47 | - [Dedicated allocations](@ref choosing_memory_type_dedicated_allocations) |
| 48 | - \subpage memory_mapping |
| 49 | - [Mapping functions](@ref memory_mapping_mapping_functions) |
| 50 | - [Persistently mapped memory](@ref memory_mapping_persistently_mapped_memory) |
| 51 | - [Cache flush and invalidate](@ref memory_mapping_cache_control) |
| 52 | - \subpage staying_within_budget |
| 53 | - [Querying for budget](@ref staying_within_budget_querying_for_budget) |
| 54 | - [Controlling memory usage](@ref staying_within_budget_controlling_memory_usage) |
| 55 | - \subpage resource_aliasing |
| 56 | - \subpage custom_memory_pools |
| 57 | - [Choosing memory type index](@ref custom_memory_pools_MemTypeIndex) |
| 58 | - [Linear allocation algorithm](@ref linear_algorithm) |
| 59 | - [Free-at-once](@ref linear_algorithm_free_at_once) |
| 60 | - [Stack](@ref linear_algorithm_stack) |
| 61 | - [Double stack](@ref linear_algorithm_double_stack) |
| 62 | - [Ring buffer](@ref linear_algorithm_ring_buffer) |
| 63 | - \subpage defragmentation |
| 64 | - \subpage statistics |
| 65 | - [Numeric statistics](@ref statistics_numeric_statistics) |
| 66 | - [JSON dump](@ref statistics_json_dump) |
| 67 | - \subpage allocation_annotation |
| 68 | - [Allocation user data](@ref allocation_user_data) |
| 69 | - [Allocation names](@ref allocation_names) |
| 70 | - \subpage virtual_allocator |
| 71 | - \subpage debugging_memory_usage |
| 72 | - [Memory initialization](@ref debugging_memory_usage_initialization) |
| 73 | - [Margins](@ref debugging_memory_usage_margins) |
| 74 | - [Corruption detection](@ref debugging_memory_usage_corruption_detection) |
| 75 | - \subpage opengl_interop |
| 76 | - \subpage usage_patterns |
| 77 | - [GPU-only resource](@ref usage_patterns_gpu_only) |
| 78 | - [Staging copy for upload](@ref usage_patterns_staging_copy_upload) |
| 79 | - [Readback](@ref usage_patterns_readback) |
| 80 | - [Advanced data uploading](@ref usage_patterns_advanced_data_uploading) |
| 81 | - [Other use cases](@ref usage_patterns_other_use_cases) |
| 82 | - \subpage configuration |
| 83 | - [Pointers to Vulkan functions](@ref config_Vulkan_functions) |
| 84 | - [Custom host memory allocator](@ref custom_memory_allocator) |
| 85 | - [Device memory allocation callbacks](@ref allocation_callbacks) |
| 86 | - [Device heap memory limit](@ref heap_memory_limit) |
| 87 | - <b>Extension support</b> |
| 88 | - \subpage vk_khr_dedicated_allocation |
| 89 | - \subpage enabling_buffer_device_address |
| 90 | - \subpage vk_ext_memory_priority |
| 91 | - \subpage vk_amd_device_coherent_memory |
| 92 | - \subpage general_considerations |
| 93 | - [Thread safety](@ref general_considerations_thread_safety) |
| 94 | - [Versioning and compatibility](@ref general_considerations_versioning_and_compatibility) |
| 95 | - [Validation layer warnings](@ref general_considerations_validation_layer_warnings) |
| 96 | - [Allocation algorithm](@ref general_considerations_allocation_algorithm) |
| 97 | - [Features not supported](@ref general_considerations_features_not_supported) |
| 98 | |
| 99 | \section main_see_also See also |
| 100 | |
| 101 | - [**Product page on GPUOpen**](https://gpuopen.com/gaming-product/vulkan-memory-allocator/) |
| 102 | - [**Source repository on GitHub**](https://github.com/GPUOpen-LibrariesAndSDKs/VulkanMemoryAllocator) |
| 103 | |
| 104 | \defgroup group_init Library initialization |
| 105 | |
| 106 | \brief API elements related to the initialization and management of the entire library, especially #VmaAllocator object. |
| 107 | |
| 108 | \defgroup group_alloc Memory allocation |
| 109 | |
| 110 | \brief API elements related to the allocation, deallocation, and management of Vulkan memory, buffers, images. |
| 111 | Most basic ones being: vmaCreateBuffer(), vmaCreateImage(). |
| 112 | |
| 113 | \defgroup group_virtual Virtual allocator |
| 114 | |
| 115 | \brief API elements related to the mechanism of \ref virtual_allocator - using the core allocation algorithm |
| 116 | for user-defined purpose without allocating any real GPU memory. |
| 117 | |
| 118 | \defgroup group_stats Statistics |
| 119 | |
| 120 | \brief API elements that query current status of the allocator, from memory usage, budget, to full dump of the internal state in JSON format. |
| 121 | See documentation chapter: \ref statistics. |
| 122 | */ |
| 123 | |
| 124 | |
| 125 | #ifdef __cplusplus |
| 126 | extern "C" { |
| 127 | #endif |
| 128 | |
| 129 | #ifndef VULKAN_H_ |
| 130 | #include <vulkan/vulkan.h> |
| 131 | #endif |
| 132 | |
| 133 | // Define this macro to declare maximum supported Vulkan version in format AAABBBCCC, |
| 134 | // where AAA = major, BBB = minor, CCC = patch. |
| 135 | // If you want to use version > 1.0, it still needs to be enabled via VmaAllocatorCreateInfo::vulkanApiVersion. |
| 136 | #if !defined(VMA_VULKAN_VERSION) |
| 137 | #if defined(VK_VERSION_1_3) |
| 138 | #define VMA_VULKAN_VERSION 1003000 |
| 139 | #elif defined(VK_VERSION_1_2) |
| 140 | #define VMA_VULKAN_VERSION 1002000 |
| 141 | #elif defined(VK_VERSION_1_1) |
| 142 | #define VMA_VULKAN_VERSION 1001000 |
| 143 | #else |
| 144 | #define VMA_VULKAN_VERSION 1000000 |
| 145 | #endif |
| 146 | #endif |
| 147 | |
| 148 | #if defined(__ANDROID__) && defined(VK_NO_PROTOTYPES) && VMA_STATIC_VULKAN_FUNCTIONS |
| 149 | extern PFN_vkGetInstanceProcAddr vkGetInstanceProcAddr; |
| 150 | extern PFN_vkGetDeviceProcAddr vkGetDeviceProcAddr; |
| 151 | extern PFN_vkGetPhysicalDeviceProperties vkGetPhysicalDeviceProperties; |
| 152 | extern PFN_vkGetPhysicalDeviceMemoryProperties vkGetPhysicalDeviceMemoryProperties; |
| 153 | extern PFN_vkAllocateMemory vkAllocateMemory; |
| 154 | extern PFN_vkFreeMemory vkFreeMemory; |
| 155 | extern PFN_vkMapMemory vkMapMemory; |
| 156 | extern PFN_vkUnmapMemory vkUnmapMemory; |
| 157 | extern PFN_vkFlushMappedMemoryRanges vkFlushMappedMemoryRanges; |
| 158 | extern PFN_vkInvalidateMappedMemoryRanges vkInvalidateMappedMemoryRanges; |
| 159 | extern PFN_vkBindBufferMemory vkBindBufferMemory; |
| 160 | extern PFN_vkBindImageMemory vkBindImageMemory; |
| 161 | extern PFN_vkGetBufferMemoryRequirements vkGetBufferMemoryRequirements; |
| 162 | extern PFN_vkGetImageMemoryRequirements vkGetImageMemoryRequirements; |
| 163 | extern PFN_vkCreateBuffer vkCreateBuffer; |
| 164 | extern PFN_vkDestroyBuffer vkDestroyBuffer; |
| 165 | extern PFN_vkCreateImage vkCreateImage; |
| 166 | extern PFN_vkDestroyImage vkDestroyImage; |
| 167 | extern PFN_vkCmdCopyBuffer vkCmdCopyBuffer; |
| 168 | #if VMA_VULKAN_VERSION >= 1001000 |
| 169 | extern PFN_vkGetBufferMemoryRequirements2 vkGetBufferMemoryRequirements2; |
| 170 | extern PFN_vkGetImageMemoryRequirements2 vkGetImageMemoryRequirements2; |
| 171 | extern PFN_vkBindBufferMemory2 vkBindBufferMemory2; |
| 172 | extern PFN_vkBindImageMemory2 vkBindImageMemory2; |
| 173 | extern PFN_vkGetPhysicalDeviceMemoryProperties2 vkGetPhysicalDeviceMemoryProperties2; |
| 174 | #endif // #if VMA_VULKAN_VERSION >= 1001000 |
| 175 | #endif // #if defined(__ANDROID__) && VMA_STATIC_VULKAN_FUNCTIONS && VK_NO_PROTOTYPES |
| 176 | |
| 177 | #if !defined(VMA_DEDICATED_ALLOCATION) |
| 178 | #if VK_KHR_get_memory_requirements2 && VK_KHR_dedicated_allocation |
| 179 | #define VMA_DEDICATED_ALLOCATION 1 |
| 180 | #else |
| 181 | #define VMA_DEDICATED_ALLOCATION 0 |
| 182 | #endif |
| 183 | #endif |
| 184 | |
| 185 | #if !defined(VMA_BIND_MEMORY2) |
| 186 | #if VK_KHR_bind_memory2 |
| 187 | #define VMA_BIND_MEMORY2 1 |
| 188 | #else |
| 189 | #define VMA_BIND_MEMORY2 0 |
| 190 | #endif |
| 191 | #endif |
| 192 | |
| 193 | #if !defined(VMA_MEMORY_BUDGET) |
| 194 | #if VK_EXT_memory_budget && (VK_KHR_get_physical_device_properties2 || VMA_VULKAN_VERSION >= 1001000) |
| 195 | #define VMA_MEMORY_BUDGET 1 |
| 196 | #else |
| 197 | #define VMA_MEMORY_BUDGET 0 |
| 198 | #endif |
| 199 | #endif |
| 200 | |
| 201 | // Defined to 1 when VK_KHR_buffer_device_address device extension or equivalent core Vulkan 1.2 feature is defined in its headers. |
| 202 | #if !defined(VMA_BUFFER_DEVICE_ADDRESS) |
| 203 | #if VK_KHR_buffer_device_address || VMA_VULKAN_VERSION >= 1002000 |
| 204 | #define VMA_BUFFER_DEVICE_ADDRESS 1 |
| 205 | #else |
| 206 | #define VMA_BUFFER_DEVICE_ADDRESS 0 |
| 207 | #endif |
| 208 | #endif |
| 209 | |
| 210 | // Defined to 1 when VK_EXT_memory_priority device extension is defined in Vulkan headers. |
| 211 | #if !defined(VMA_MEMORY_PRIORITY) |
| 212 | #if VK_EXT_memory_priority |
| 213 | #define VMA_MEMORY_PRIORITY 1 |
| 214 | #else |
| 215 | #define VMA_MEMORY_PRIORITY 0 |
| 216 | #endif |
| 217 | #endif |
| 218 | |
| 219 | // Defined to 1 when VK_KHR_external_memory device extension is defined in Vulkan headers. |
| 220 | #if !defined(VMA_EXTERNAL_MEMORY) |
| 221 | #if VK_KHR_external_memory |
| 222 | #define VMA_EXTERNAL_MEMORY 1 |
| 223 | #else |
| 224 | #define VMA_EXTERNAL_MEMORY 0 |
| 225 | #endif |
| 226 | #endif |
| 227 | |
| 228 | // Define these macros to decorate all public functions with additional code, |
| 229 | // before and after returned type, appropriately. This may be useful for |
| 230 | // exporting the functions when compiling VMA as a separate library. Example: |
| 231 | // #define VMA_CALL_PRE __declspec(dllexport) |
| 232 | // #define VMA_CALL_POST __cdecl |
| 233 | #ifndef VMA_CALL_PRE |
| 234 | #define VMA_CALL_PRE |
| 235 | #endif |
| 236 | #ifndef VMA_CALL_POST |
| 237 | #define VMA_CALL_POST |
| 238 | #endif |
| 239 | |
| 240 | // Define this macro to decorate pointers with an attribute specifying the |
| 241 | // length of the array they point to if they are not null. |
| 242 | // |
| 243 | // The length may be one of |
| 244 | // - The name of another parameter in the argument list where the pointer is declared |
| 245 | // - The name of another member in the struct where the pointer is declared |
| 246 | // - The name of a member of a struct type, meaning the value of that member in |
| 247 | // the context of the call. For example |
| 248 | // VMA_LEN_IF_NOT_NULL("VkPhysicalDeviceMemoryProperties::memoryHeapCount"), |
| 249 | // this means the number of memory heaps available in the device associated |
| 250 | // with the VmaAllocator being dealt with. |
| 251 | #ifndef VMA_LEN_IF_NOT_NULL |
| 252 | #define VMA_LEN_IF_NOT_NULL(len) |
| 253 | #endif |
| 254 | |
| 255 | // The VMA_NULLABLE macro is defined to be _Nullable when compiling with Clang. |
| 256 | // see: https://clang.llvm.org/docs/AttributeReference.html#nullable |
| 257 | #ifndef VMA_NULLABLE |
| 258 | #ifdef __clang__ |
| 259 | #define VMA_NULLABLE _Nullable |
| 260 | #else |
| 261 | #define VMA_NULLABLE |
| 262 | #endif |
| 263 | #endif |
| 264 | |
| 265 | // The VMA_NOT_NULL macro is defined to be _Nonnull when compiling with Clang. |
| 266 | // see: https://clang.llvm.org/docs/AttributeReference.html#nonnull |
| 267 | #ifndef VMA_NOT_NULL |
| 268 | #ifdef __clang__ |
| 269 | #define VMA_NOT_NULL _Nonnull |
| 270 | #else |
| 271 | #define VMA_NOT_NULL |
| 272 | #endif |
| 273 | #endif |
| 274 | |
| 275 | // If non-dispatchable handles are represented as pointers then we can give |
| 276 | // then nullability annotations |
| 277 | #ifndef VMA_NOT_NULL_NON_DISPATCHABLE |
| 278 | #if defined(__LP64__) || defined(_WIN64) || (defined(__x86_64__) && !defined(__ILP32__) ) || defined(_M_X64) || defined(__ia64) || defined (_M_IA64) || defined(__aarch64__) || defined(__powerpc64__) |
| 279 | #define VMA_NOT_NULL_NON_DISPATCHABLE VMA_NOT_NULL |
| 280 | #else |
| 281 | #define VMA_NOT_NULL_NON_DISPATCHABLE |
| 282 | #endif |
| 283 | #endif |
| 284 | |
| 285 | #ifndef VMA_NULLABLE_NON_DISPATCHABLE |
| 286 | #if defined(__LP64__) || defined(_WIN64) || (defined(__x86_64__) && !defined(__ILP32__) ) || defined(_M_X64) || defined(__ia64) || defined (_M_IA64) || defined(__aarch64__) || defined(__powerpc64__) |
| 287 | #define VMA_NULLABLE_NON_DISPATCHABLE VMA_NULLABLE |
| 288 | #else |
| 289 | #define VMA_NULLABLE_NON_DISPATCHABLE |
| 290 | #endif |
| 291 | #endif |
| 292 | |
| 293 | #ifndef VMA_STATS_STRING_ENABLED |
| 294 | #define VMA_STATS_STRING_ENABLED 1 |
| 295 | #endif |
| 296 | |
| 297 | //////////////////////////////////////////////////////////////////////////////// |
| 298 | //////////////////////////////////////////////////////////////////////////////// |
| 299 | // |
| 300 | // INTERFACE |
| 301 | // |
| 302 | //////////////////////////////////////////////////////////////////////////////// |
| 303 | //////////////////////////////////////////////////////////////////////////////// |
| 304 | |
| 305 | // Sections for managing code placement in file, only for development purposes e.g. for convenient folding inside an IDE. |
| 306 | #ifndef _VMA_ENUM_DECLARATIONS |
| 307 | |
| 308 | /** |
| 309 | \addtogroup group_init |
| 310 | @{ |
| 311 | */ |
| 312 | |
| 313 | /// Flags for created #VmaAllocator. |
| 314 | typedef enum VmaAllocatorCreateFlagBits |
| 315 | { |
| 316 | /** \brief Allocator and all objects created from it will not be synchronized internally, so you must guarantee they are used from only one thread at a time or synchronized externally by you. |
| 317 | |
| 318 | Using this flag may increase performance because internal mutexes are not used. |
| 319 | */ |
| 320 | VMA_ALLOCATOR_CREATE_EXTERNALLY_SYNCHRONIZED_BIT = 0x00000001, |
| 321 | /** \brief Enables usage of VK_KHR_dedicated_allocation extension. |
| 322 | |
| 323 | The flag works only if VmaAllocatorCreateInfo::vulkanApiVersion `== VK_API_VERSION_1_0`. |
| 324 | When it is `VK_API_VERSION_1_1`, the flag is ignored because the extension has been promoted to Vulkan 1.1. |
| 325 | |
| 326 | Using this extension will automatically allocate dedicated blocks of memory for |
| 327 | some buffers and images instead of suballocating place for them out of bigger |
| 328 | memory blocks (as if you explicitly used #VMA_ALLOCATION_CREATE_DEDICATED_MEMORY_BIT |
| 329 | flag) when it is recommended by the driver. It may improve performance on some |
| 330 | GPUs. |
| 331 | |
| 332 | You may set this flag only if you found out that following device extensions are |
| 333 | supported, you enabled them while creating Vulkan device passed as |
| 334 | VmaAllocatorCreateInfo::device, and you want them to be used internally by this |
| 335 | library: |
| 336 | |
| 337 | - VK_KHR_get_memory_requirements2 (device extension) |
| 338 | - VK_KHR_dedicated_allocation (device extension) |
| 339 | |
| 340 | When this flag is set, you can experience following warnings reported by Vulkan |
| 341 | validation layer. You can ignore them. |
| 342 | |
| 343 | > vkBindBufferMemory(): Binding memory to buffer 0x2d but vkGetBufferMemoryRequirements() has not been called on that buffer. |
| 344 | */ |
| 345 | VMA_ALLOCATOR_CREATE_KHR_DEDICATED_ALLOCATION_BIT = 0x00000002, |
| 346 | /** |
| 347 | Enables usage of VK_KHR_bind_memory2 extension. |
| 348 | |
| 349 | The flag works only if VmaAllocatorCreateInfo::vulkanApiVersion `== VK_API_VERSION_1_0`. |
| 350 | When it is `VK_API_VERSION_1_1`, the flag is ignored because the extension has been promoted to Vulkan 1.1. |
| 351 | |
| 352 | You may set this flag only if you found out that this device extension is supported, |
| 353 | you enabled it while creating Vulkan device passed as VmaAllocatorCreateInfo::device, |
| 354 | and you want it to be used internally by this library. |
| 355 | |
| 356 | The extension provides functions `vkBindBufferMemory2KHR` and `vkBindImageMemory2KHR`, |
| 357 | which allow to pass a chain of `pNext` structures while binding. |
| 358 | This flag is required if you use `pNext` parameter in vmaBindBufferMemory2() or vmaBindImageMemory2(). |
| 359 | */ |
| 360 | VMA_ALLOCATOR_CREATE_KHR_BIND_MEMORY2_BIT = 0x00000004, |
| 361 | /** |
| 362 | Enables usage of VK_EXT_memory_budget extension. |
| 363 | |
| 364 | You may set this flag only if you found out that this device extension is supported, |
| 365 | you enabled it while creating Vulkan device passed as VmaAllocatorCreateInfo::device, |
| 366 | and you want it to be used internally by this library, along with another instance extension |
| 367 | VK_KHR_get_physical_device_properties2, which is required by it (or Vulkan 1.1, where this extension is promoted). |
| 368 | |
| 369 | The extension provides query for current memory usage and budget, which will probably |
| 370 | be more accurate than an estimation used by the library otherwise. |
| 371 | */ |
| 372 | VMA_ALLOCATOR_CREATE_EXT_MEMORY_BUDGET_BIT = 0x00000008, |
| 373 | /** |
| 374 | Enables usage of VK_AMD_device_coherent_memory extension. |
| 375 | |
| 376 | You may set this flag only if you: |
| 377 | |
| 378 | - found out that this device extension is supported and enabled it while creating Vulkan device passed as VmaAllocatorCreateInfo::device, |
| 379 | - checked that `VkPhysicalDeviceCoherentMemoryFeaturesAMD::deviceCoherentMemory` is true and set it while creating the Vulkan device, |
| 380 | - want it to be used internally by this library. |
| 381 | |
| 382 | The extension and accompanying device feature provide access to memory types with |
| 383 | `VK_MEMORY_PROPERTY_DEVICE_COHERENT_BIT_AMD` and `VK_MEMORY_PROPERTY_DEVICE_UNCACHED_BIT_AMD` flags. |
| 384 | They are useful mostly for writing breadcrumb markers - a common method for debugging GPU crash/hang/TDR. |
| 385 | |
| 386 | When the extension is not enabled, such memory types are still enumerated, but their usage is illegal. |
| 387 | To protect from this error, if you don't create the allocator with this flag, it will refuse to allocate any memory or create a custom pool in such memory type, |
| 388 | returning `VK_ERROR_FEATURE_NOT_PRESENT`. |
| 389 | */ |
| 390 | VMA_ALLOCATOR_CREATE_AMD_DEVICE_COHERENT_MEMORY_BIT = 0x00000010, |
| 391 | /** |
| 392 | Enables usage of "buffer device address" feature, which allows you to use function |
| 393 | `vkGetBufferDeviceAddress*` to get raw GPU pointer to a buffer and pass it for usage inside a shader. |
| 394 | |
| 395 | You may set this flag only if you: |
| 396 | |
| 397 | 1. (For Vulkan version < 1.2) Found as available and enabled device extension |
| 398 | VK_KHR_buffer_device_address. |
| 399 | This extension is promoted to core Vulkan 1.2. |
| 400 | 2. Found as available and enabled device feature `VkPhysicalDeviceBufferDeviceAddressFeatures::bufferDeviceAddress`. |
| 401 | |
| 402 | When this flag is set, you can create buffers with `VK_BUFFER_USAGE_SHADER_DEVICE_ADDRESS_BIT` using VMA. |
| 403 | The library automatically adds `VK_MEMORY_ALLOCATE_DEVICE_ADDRESS_BIT` to |
| 404 | allocated memory blocks wherever it might be needed. |
| 405 | |
| 406 | For more information, see documentation chapter \ref enabling_buffer_device_address. |
| 407 | */ |
| 408 | VMA_ALLOCATOR_CREATE_BUFFER_DEVICE_ADDRESS_BIT = 0x00000020, |
| 409 | /** |
| 410 | Enables usage of VK_EXT_memory_priority extension in the library. |
| 411 | |
| 412 | You may set this flag only if you found available and enabled this device extension, |
| 413 | along with `VkPhysicalDeviceMemoryPriorityFeaturesEXT::memoryPriority == VK_TRUE`, |
| 414 | while creating Vulkan device passed as VmaAllocatorCreateInfo::device. |
| 415 | |
| 416 | When this flag is used, VmaAllocationCreateInfo::priority and VmaPoolCreateInfo::priority |
| 417 | are used to set priorities of allocated Vulkan memory. Without it, these variables are ignored. |
| 418 | |
| 419 | A priority must be a floating-point value between 0 and 1, indicating the priority of the allocation relative to other memory allocations. |
| 420 | Larger values are higher priority. The granularity of the priorities is implementation-dependent. |
| 421 | It is automatically passed to every call to `vkAllocateMemory` done by the library using structure `VkMemoryPriorityAllocateInfoEXT`. |
| 422 | The value to be used for default priority is 0.5. |
| 423 | For more details, see the documentation of the VK_EXT_memory_priority extension. |
| 424 | */ |
| 425 | VMA_ALLOCATOR_CREATE_EXT_MEMORY_PRIORITY_BIT = 0x00000040, |
| 426 | |
| 427 | VMA_ALLOCATOR_CREATE_FLAG_BITS_MAX_ENUM = 0x7FFFFFFF |
| 428 | } VmaAllocatorCreateFlagBits; |
| 429 | /// See #VmaAllocatorCreateFlagBits. |
| 430 | typedef VkFlags VmaAllocatorCreateFlags; |
| 431 | |
| 432 | /** @} */ |
| 433 | |
| 434 | /** |
| 435 | \addtogroup group_alloc |
| 436 | @{ |
| 437 | */ |
| 438 | |
| 439 | /// \brief Intended usage of the allocated memory. |
| 440 | typedef enum VmaMemoryUsage |
| 441 | { |
| 442 | /** No intended memory usage specified. |
| 443 | Use other members of VmaAllocationCreateInfo to specify your requirements. |
| 444 | */ |
| 445 | VMA_MEMORY_USAGE_UNKNOWN = 0, |
| 446 | /** |
| 447 | \deprecated Obsolete, preserved for backward compatibility. |
| 448 | Prefers `VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT`. |
| 449 | */ |
| 450 | VMA_MEMORY_USAGE_GPU_ONLY = 1, |
| 451 | /** |
| 452 | \deprecated Obsolete, preserved for backward compatibility. |
| 453 | Guarantees `VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT` and `VK_MEMORY_PROPERTY_HOST_COHERENT_BIT`. |
| 454 | */ |
| 455 | VMA_MEMORY_USAGE_CPU_ONLY = 2, |
| 456 | /** |
| 457 | \deprecated Obsolete, preserved for backward compatibility. |
| 458 | Guarantees `VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT`, prefers `VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT`. |
| 459 | */ |
| 460 | VMA_MEMORY_USAGE_CPU_TO_GPU = 3, |
| 461 | /** |
| 462 | \deprecated Obsolete, preserved for backward compatibility. |
| 463 | Guarantees `VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT`, prefers `VK_MEMORY_PROPERTY_HOST_CACHED_BIT`. |
| 464 | */ |
| 465 | VMA_MEMORY_USAGE_GPU_TO_CPU = 4, |
| 466 | /** |
| 467 | \deprecated Obsolete, preserved for backward compatibility. |
| 468 | Prefers not `VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT`. |
| 469 | */ |
| 470 | VMA_MEMORY_USAGE_CPU_COPY = 5, |
| 471 | /** |
| 472 | Lazily allocated GPU memory having `VK_MEMORY_PROPERTY_LAZILY_ALLOCATED_BIT`. |
| 473 | Exists mostly on mobile platforms. Using it on desktop PC or other GPUs with no such memory type present will fail the allocation. |
| 474 | |
| 475 | Usage: Memory for transient attachment images (color attachments, depth attachments etc.), created with `VK_IMAGE_USAGE_TRANSIENT_ATTACHMENT_BIT`. |
| 476 | |
| 477 | Allocations with this usage are always created as dedicated - it implies #VMA_ALLOCATION_CREATE_DEDICATED_MEMORY_BIT. |
| 478 | */ |
| 479 | VMA_MEMORY_USAGE_GPU_LAZILY_ALLOCATED = 6, |
| 480 | /** |
| 481 | Selects best memory type automatically. |
| 482 | This flag is recommended for most common use cases. |
| 483 | |
| 484 | When using this flag, if you want to map the allocation (using vmaMapMemory() or #VMA_ALLOCATION_CREATE_MAPPED_BIT), |
| 485 | you must pass one of the flags: #VMA_ALLOCATION_CREATE_HOST_ACCESS_SEQUENTIAL_WRITE_BIT or #VMA_ALLOCATION_CREATE_HOST_ACCESS_RANDOM_BIT |
| 486 | in VmaAllocationCreateInfo::flags. |
| 487 | |
| 488 | It can be used only with functions that let the library know `VkBufferCreateInfo` or `VkImageCreateInfo`, e.g. |
| 489 | vmaCreateBuffer(), vmaCreateImage(), vmaFindMemoryTypeIndexForBufferInfo(), vmaFindMemoryTypeIndexForImageInfo() |
| 490 | and not with generic memory allocation functions. |
| 491 | */ |
| 492 | VMA_MEMORY_USAGE_AUTO = 7, |
| 493 | /** |
| 494 | Selects best memory type automatically with preference for GPU (device) memory. |
| 495 | |
| 496 | When using this flag, if you want to map the allocation (using vmaMapMemory() or #VMA_ALLOCATION_CREATE_MAPPED_BIT), |
| 497 | you must pass one of the flags: #VMA_ALLOCATION_CREATE_HOST_ACCESS_SEQUENTIAL_WRITE_BIT or #VMA_ALLOCATION_CREATE_HOST_ACCESS_RANDOM_BIT |
| 498 | in VmaAllocationCreateInfo::flags. |
| 499 | |
| 500 | It can be used only with functions that let the library know `VkBufferCreateInfo` or `VkImageCreateInfo`, e.g. |
| 501 | vmaCreateBuffer(), vmaCreateImage(), vmaFindMemoryTypeIndexForBufferInfo(), vmaFindMemoryTypeIndexForImageInfo() |
| 502 | and not with generic memory allocation functions. |
| 503 | */ |
| 504 | VMA_MEMORY_USAGE_AUTO_PREFER_DEVICE = 8, |
| 505 | /** |
| 506 | Selects best memory type automatically with preference for CPU (host) memory. |
| 507 | |
| 508 | When using this flag, if you want to map the allocation (using vmaMapMemory() or #VMA_ALLOCATION_CREATE_MAPPED_BIT), |
| 509 | you must pass one of the flags: #VMA_ALLOCATION_CREATE_HOST_ACCESS_SEQUENTIAL_WRITE_BIT or #VMA_ALLOCATION_CREATE_HOST_ACCESS_RANDOM_BIT |
| 510 | in VmaAllocationCreateInfo::flags. |
| 511 | |
| 512 | It can be used only with functions that let the library know `VkBufferCreateInfo` or `VkImageCreateInfo`, e.g. |
| 513 | vmaCreateBuffer(), vmaCreateImage(), vmaFindMemoryTypeIndexForBufferInfo(), vmaFindMemoryTypeIndexForImageInfo() |
| 514 | and not with generic memory allocation functions. |
| 515 | */ |
| 516 | VMA_MEMORY_USAGE_AUTO_PREFER_HOST = 9, |
| 517 | |
| 518 | VMA_MEMORY_USAGE_MAX_ENUM = 0x7FFFFFFF |
| 519 | } VmaMemoryUsage; |
| 520 | |
| 521 | /// Flags to be passed as VmaAllocationCreateInfo::flags. |
| 522 | typedef enum VmaAllocationCreateFlagBits |
| 523 | { |
| 524 | /** \brief Set this flag if the allocation should have its own memory block. |
| 525 | |
| 526 | Use it for special, big resources, like fullscreen images used as attachments. |
| 527 | */ |
| 528 | VMA_ALLOCATION_CREATE_DEDICATED_MEMORY_BIT = 0x00000001, |
| 529 | |
| 530 | /** \brief Set this flag to only try to allocate from existing `VkDeviceMemory` blocks and never create new such block. |
| 531 | |
| 532 | If new allocation cannot be placed in any of the existing blocks, allocation |
| 533 | fails with `VK_ERROR_OUT_OF_DEVICE_MEMORY` error. |
| 534 | |
| 535 | You should not use #VMA_ALLOCATION_CREATE_DEDICATED_MEMORY_BIT and |
| 536 | #VMA_ALLOCATION_CREATE_NEVER_ALLOCATE_BIT at the same time. It makes no sense. |
| 537 | */ |
| 538 | VMA_ALLOCATION_CREATE_NEVER_ALLOCATE_BIT = 0x00000002, |
| 539 | /** \brief Set this flag to use a memory that will be persistently mapped and retrieve pointer to it. |
| 540 | |
| 541 | Pointer to mapped memory will be returned through VmaAllocationInfo::pMappedData. |
| 542 | |
| 543 | It is valid to use this flag for allocation made from memory type that is not |
| 544 | `HOST_VISIBLE`. This flag is then ignored and memory is not mapped. This is |
| 545 | useful if you need an allocation that is efficient to use on GPU |
| 546 | (`DEVICE_LOCAL`) and still want to map it directly if possible on platforms that |
| 547 | support it (e.g. Intel GPU). |
| 548 | */ |
| 549 | VMA_ALLOCATION_CREATE_MAPPED_BIT = 0x00000004, |
| 550 | /** \deprecated Preserved for backward compatibility. Consider using vmaSetAllocationName() instead. |
| 551 | |
| 552 | Set this flag to treat VmaAllocationCreateInfo::pUserData as pointer to a |
| 553 | null-terminated string. Instead of copying pointer value, a local copy of the |
| 554 | string is made and stored in allocation's `pName`. The string is automatically |
| 555 | freed together with the allocation. It is also used in vmaBuildStatsString(). |
| 556 | */ |
| 557 | VMA_ALLOCATION_CREATE_USER_DATA_COPY_STRING_BIT = 0x00000020, |
| 558 | /** Allocation will be created from upper stack in a double stack pool. |
| 559 | |
| 560 | This flag is only allowed for custom pools created with #VMA_POOL_CREATE_LINEAR_ALGORITHM_BIT flag. |
| 561 | */ |
| 562 | VMA_ALLOCATION_CREATE_UPPER_ADDRESS_BIT = 0x00000040, |
| 563 | /** Create both buffer/image and allocation, but don't bind them together. |
| 564 | It is useful when you want to bind yourself to do some more advanced binding, e.g. using some extensions. |
| 565 | The flag is meaningful only with functions that bind by default: vmaCreateBuffer(), vmaCreateImage(). |
| 566 | Otherwise it is ignored. |
| 567 | |
| 568 | If you want to make sure the new buffer/image is not tied to the new memory allocation |
| 569 | through `VkMemoryDedicatedAllocateInfoKHR` structure in case the allocation ends up in its own memory block, |
| 570 | use also flag #VMA_ALLOCATION_CREATE_CAN_ALIAS_BIT. |
| 571 | */ |
| 572 | VMA_ALLOCATION_CREATE_DONT_BIND_BIT = 0x00000080, |
| 573 | /** Create allocation only if additional device memory required for it, if any, won't exceed |
| 574 | memory budget. Otherwise return `VK_ERROR_OUT_OF_DEVICE_MEMORY`. |
| 575 | */ |
| 576 | VMA_ALLOCATION_CREATE_WITHIN_BUDGET_BIT = 0x00000100, |
| 577 | /** \brief Set this flag if the allocated memory will have aliasing resources. |
| 578 | |
| 579 | Usage of this flag prevents supplying `VkMemoryDedicatedAllocateInfoKHR` when #VMA_ALLOCATION_CREATE_DEDICATED_MEMORY_BIT is specified. |
| 580 | Otherwise created dedicated memory will not be suitable for aliasing resources, resulting in Vulkan Validation Layer errors. |
| 581 | */ |
| 582 | VMA_ALLOCATION_CREATE_CAN_ALIAS_BIT = 0x00000200, |
| 583 | /** |
| 584 | Requests possibility to map the allocation (using vmaMapMemory() or #VMA_ALLOCATION_CREATE_MAPPED_BIT). |
| 585 | |
| 586 | - If you use #VMA_MEMORY_USAGE_AUTO or other `VMA_MEMORY_USAGE_AUTO*` value, |
| 587 | you must use this flag to be able to map the allocation. Otherwise, mapping is incorrect. |
| 588 | - If you use other value of #VmaMemoryUsage, this flag is ignored and mapping is always possible in memory types that are `HOST_VISIBLE`. |
| 589 | This includes allocations created in \ref custom_memory_pools. |
| 590 | |
| 591 | Declares that mapped memory will only be written sequentially, e.g. using `memcpy()` or a loop writing number-by-number, |
| 592 | never read or accessed randomly, so a memory type can be selected that is uncached and write-combined. |
| 593 | |
| 594 | \warning Violating this declaration may work correctly, but will likely be very slow. |
| 595 | Watch out for implicit reads introduced by doing e.g. `pMappedData[i] += x;` |
| 596 | Better prepare your data in a local variable and `memcpy()` it to the mapped pointer all at once. |
| 597 | */ |
| 598 | VMA_ALLOCATION_CREATE_HOST_ACCESS_SEQUENTIAL_WRITE_BIT = 0x00000400, |
| 599 | /** |
| 600 | Requests possibility to map the allocation (using vmaMapMemory() or #VMA_ALLOCATION_CREATE_MAPPED_BIT). |
| 601 | |
| 602 | - If you use #VMA_MEMORY_USAGE_AUTO or other `VMA_MEMORY_USAGE_AUTO*` value, |
| 603 | you must use this flag to be able to map the allocation. Otherwise, mapping is incorrect. |
| 604 | - If you use other value of #VmaMemoryUsage, this flag is ignored and mapping is always possible in memory types that are `HOST_VISIBLE`. |
| 605 | This includes allocations created in \ref custom_memory_pools. |
| 606 | |
| 607 | Declares that mapped memory can be read, written, and accessed in random order, |
| 608 | so a `HOST_CACHED` memory type is required. |
| 609 | */ |
| 610 | VMA_ALLOCATION_CREATE_HOST_ACCESS_RANDOM_BIT = 0x00000800, |
| 611 | /** |
| 612 | Together with #VMA_ALLOCATION_CREATE_HOST_ACCESS_SEQUENTIAL_WRITE_BIT or #VMA_ALLOCATION_CREATE_HOST_ACCESS_RANDOM_BIT, |
| 613 | it says that despite request for host access, a not-`HOST_VISIBLE` memory type can be selected |
| 614 | if it may improve performance. |
| 615 | |
| 616 | By using this flag, you declare that you will check if the allocation ended up in a `HOST_VISIBLE` memory type |
| 617 | (e.g. using vmaGetAllocationMemoryProperties()) and if not, you will create some "staging" buffer and |
| 618 | issue an explicit transfer to write/read your data. |
| 619 | To prepare for this possibility, don't forget to add appropriate flags like |
| 620 | `VK_BUFFER_USAGE_TRANSFER_DST_BIT`, `VK_BUFFER_USAGE_TRANSFER_SRC_BIT` to the parameters of created buffer or image. |
| 621 | */ |
| 622 | VMA_ALLOCATION_CREATE_HOST_ACCESS_ALLOW_TRANSFER_INSTEAD_BIT = 0x00001000, |
| 623 | /** Allocation strategy that chooses smallest possible free range for the allocation |
| 624 | to minimize memory usage and fragmentation, possibly at the expense of allocation time. |
| 625 | */ |
| 626 | VMA_ALLOCATION_CREATE_STRATEGY_MIN_MEMORY_BIT = 0x00010000, |
| 627 | /** Allocation strategy that chooses first suitable free range for the allocation - |
| 628 | not necessarily in terms of the smallest offset but the one that is easiest and fastest to find |
| 629 | to minimize allocation time, possibly at the expense of allocation quality. |
| 630 | */ |
| 631 | VMA_ALLOCATION_CREATE_STRATEGY_MIN_TIME_BIT = 0x00020000, |
| 632 | /** Allocation strategy that chooses always the lowest offset in available space. |
| 633 | This is not the most efficient strategy but achieves highly packed data. |
| 634 | Used internally by defragmentation, not recomended in typical usage. |
| 635 | */ |
| 636 | VMA_ALLOCATION_CREATE_STRATEGY_MIN_OFFSET_BIT = 0x00040000, |
| 637 | /** Alias to #VMA_ALLOCATION_CREATE_STRATEGY_MIN_MEMORY_BIT. |
| 638 | */ |
| 639 | VMA_ALLOCATION_CREATE_STRATEGY_BEST_FIT_BIT = VMA_ALLOCATION_CREATE_STRATEGY_MIN_MEMORY_BIT, |
| 640 | /** Alias to #VMA_ALLOCATION_CREATE_STRATEGY_MIN_TIME_BIT. |
| 641 | */ |
| 642 | VMA_ALLOCATION_CREATE_STRATEGY_FIRST_FIT_BIT = VMA_ALLOCATION_CREATE_STRATEGY_MIN_TIME_BIT, |
| 643 | /** A bit mask to extract only `STRATEGY` bits from entire set of flags. |
| 644 | */ |
| 645 | VMA_ALLOCATION_CREATE_STRATEGY_MASK = |
| 646 | VMA_ALLOCATION_CREATE_STRATEGY_MIN_MEMORY_BIT | |
| 647 | VMA_ALLOCATION_CREATE_STRATEGY_MIN_TIME_BIT | |
| 648 | VMA_ALLOCATION_CREATE_STRATEGY_MIN_OFFSET_BIT, |
| 649 | |
| 650 | VMA_ALLOCATION_CREATE_FLAG_BITS_MAX_ENUM = 0x7FFFFFFF |
| 651 | } VmaAllocationCreateFlagBits; |
| 652 | /// See #VmaAllocationCreateFlagBits. |
| 653 | typedef VkFlags VmaAllocationCreateFlags; |
| 654 | |
| 655 | /// Flags to be passed as VmaPoolCreateInfo::flags. |
| 656 | typedef enum VmaPoolCreateFlagBits |
| 657 | { |
| 658 | /** \brief Use this flag if you always allocate only buffers and linear images or only optimal images out of this pool and so Buffer-Image Granularity can be ignored. |
| 659 | |
| 660 | This is an optional optimization flag. |
| 661 | |
| 662 | If you always allocate using vmaCreateBuffer(), vmaCreateImage(), |
| 663 | vmaAllocateMemoryForBuffer(), then you don't need to use it because allocator |
| 664 | knows exact type of your allocations so it can handle Buffer-Image Granularity |
| 665 | in the optimal way. |
| 666 | |
| 667 | If you also allocate using vmaAllocateMemoryForImage() or vmaAllocateMemory(), |
| 668 | exact type of such allocations is not known, so allocator must be conservative |
| 669 | in handling Buffer-Image Granularity, which can lead to suboptimal allocation |
| 670 | (wasted memory). In that case, if you can make sure you always allocate only |
| 671 | buffers and linear images or only optimal images out of this pool, use this flag |
| 672 | to make allocator disregard Buffer-Image Granularity and so make allocations |
| 673 | faster and more optimal. |
| 674 | */ |
| 675 | VMA_POOL_CREATE_IGNORE_BUFFER_IMAGE_GRANULARITY_BIT = 0x00000002, |
| 676 | |
| 677 | /** \brief Enables alternative, linear allocation algorithm in this pool. |
| 678 | |
| 679 | Specify this flag to enable linear allocation algorithm, which always creates |
| 680 | new allocations after last one and doesn't reuse space from allocations freed in |
| 681 | between. It trades memory consumption for simplified algorithm and data |
| 682 | structure, which has better performance and uses less memory for metadata. |
| 683 | |
| 684 | By using this flag, you can achieve behavior of free-at-once, stack, |
| 685 | ring buffer, and double stack. |
| 686 | For details, see documentation chapter \ref linear_algorithm. |
| 687 | */ |
| 688 | VMA_POOL_CREATE_LINEAR_ALGORITHM_BIT = 0x00000004, |
| 689 | |
| 690 | /** Bit mask to extract only `ALGORITHM` bits from entire set of flags. |
| 691 | */ |
| 692 | VMA_POOL_CREATE_ALGORITHM_MASK = |
| 693 | VMA_POOL_CREATE_LINEAR_ALGORITHM_BIT, |
| 694 | |
| 695 | VMA_POOL_CREATE_FLAG_BITS_MAX_ENUM = 0x7FFFFFFF |
| 696 | } VmaPoolCreateFlagBits; |
| 697 | /// Flags to be passed as VmaPoolCreateInfo::flags. See #VmaPoolCreateFlagBits. |
| 698 | typedef VkFlags VmaPoolCreateFlags; |
| 699 | |
| 700 | /// Flags to be passed as VmaDefragmentationInfo::flags. |
| 701 | typedef enum VmaDefragmentationFlagBits |
| 702 | { |
| 703 | /* \brief Use simple but fast algorithm for defragmentation. |
| 704 | May not achieve best results but will require least time to compute and least allocations to copy. |
| 705 | */ |
| 706 | VMA_DEFRAGMENTATION_FLAG_ALGORITHM_FAST_BIT = 0x1, |
| 707 | /* \brief Default defragmentation algorithm, applied also when no `ALGORITHM` flag is specified. |
| 708 | Offers a balance between defragmentation quality and the amount of allocations and bytes that need to be moved. |
| 709 | */ |
| 710 | VMA_DEFRAGMENTATION_FLAG_ALGORITHM_BALANCED_BIT = 0x2, |
| 711 | /* \brief Perform full defragmentation of memory. |
| 712 | Can result in notably more time to compute and allocations to copy, but will achieve best memory packing. |
| 713 | */ |
| 714 | VMA_DEFRAGMENTATION_FLAG_ALGORITHM_FULL_BIT = 0x4, |
| 715 | /** \brief Use the most roboust algorithm at the cost of time to compute and number of copies to make. |
| 716 | Only available when bufferImageGranularity is greater than 1, since it aims to reduce |
| 717 | alignment issues between different types of resources. |
| 718 | Otherwise falls back to same behavior as #VMA_DEFRAGMENTATION_FLAG_ALGORITHM_FULL_BIT. |
| 719 | */ |
| 720 | VMA_DEFRAGMENTATION_FLAG_ALGORITHM_EXTENSIVE_BIT = 0x8, |
| 721 | |
| 722 | /// A bit mask to extract only `ALGORITHM` bits from entire set of flags. |
| 723 | VMA_DEFRAGMENTATION_FLAG_ALGORITHM_MASK = |
| 724 | VMA_DEFRAGMENTATION_FLAG_ALGORITHM_FAST_BIT | |
| 725 | VMA_DEFRAGMENTATION_FLAG_ALGORITHM_BALANCED_BIT | |
| 726 | VMA_DEFRAGMENTATION_FLAG_ALGORITHM_FULL_BIT | |
| 727 | VMA_DEFRAGMENTATION_FLAG_ALGORITHM_EXTENSIVE_BIT, |
| 728 | |
| 729 | VMA_DEFRAGMENTATION_FLAG_BITS_MAX_ENUM = 0x7FFFFFFF |
| 730 | } VmaDefragmentationFlagBits; |
| 731 | /// See #VmaDefragmentationFlagBits. |
| 732 | typedef VkFlags VmaDefragmentationFlags; |
| 733 | |
| 734 | /// Operation performed on single defragmentation move. See structure #VmaDefragmentationMove. |
| 735 | typedef enum VmaDefragmentationMoveOperation |
| 736 | { |
| 737 | /// Buffer/image has been recreated at `dstTmpAllocation`, data has been copied, old buffer/image has been destroyed. `srcAllocation` should be changed to point to the new place. This is the default value set by vmaBeginDefragmentationPass(). |
| 738 | VMA_DEFRAGMENTATION_MOVE_OPERATION_COPY = 0, |
| 739 | /// Set this value if you cannot move the allocation. New place reserved at `dstTmpAllocation` will be freed. `srcAllocation` will remain unchanged. |
| 740 | VMA_DEFRAGMENTATION_MOVE_OPERATION_IGNORE = 1, |
| 741 | /// Set this value if you decide to abandon the allocation and you destroyed the buffer/image. New place reserved at `dstTmpAllocation` will be freed, along with `srcAllocation`, which will be destroyed. |
| 742 | VMA_DEFRAGMENTATION_MOVE_OPERATION_DESTROY = 2, |
| 743 | } VmaDefragmentationMoveOperation; |
| 744 | |
| 745 | /** @} */ |
| 746 | |
| 747 | /** |
| 748 | \addtogroup group_virtual |
| 749 | @{ |
| 750 | */ |
| 751 | |
| 752 | /// Flags to be passed as VmaVirtualBlockCreateInfo::flags. |
| 753 | typedef enum VmaVirtualBlockCreateFlagBits |
| 754 | { |
| 755 | /** \brief Enables alternative, linear allocation algorithm in this virtual block. |
| 756 | |
| 757 | Specify this flag to enable linear allocation algorithm, which always creates |
| 758 | new allocations after last one and doesn't reuse space from allocations freed in |
| 759 | between. It trades memory consumption for simplified algorithm and data |
| 760 | structure, which has better performance and uses less memory for metadata. |
| 761 | |
| 762 | By using this flag, you can achieve behavior of free-at-once, stack, |
| 763 | ring buffer, and double stack. |
| 764 | For details, see documentation chapter \ref linear_algorithm. |
| 765 | */ |
| 766 | VMA_VIRTUAL_BLOCK_CREATE_LINEAR_ALGORITHM_BIT = 0x00000001, |
| 767 | |
| 768 | /** \brief Bit mask to extract only `ALGORITHM` bits from entire set of flags. |
| 769 | */ |
| 770 | VMA_VIRTUAL_BLOCK_CREATE_ALGORITHM_MASK = |
| 771 | VMA_VIRTUAL_BLOCK_CREATE_LINEAR_ALGORITHM_BIT, |
| 772 | |
| 773 | VMA_VIRTUAL_BLOCK_CREATE_FLAG_BITS_MAX_ENUM = 0x7FFFFFFF |
| 774 | } VmaVirtualBlockCreateFlagBits; |
| 775 | /// Flags to be passed as VmaVirtualBlockCreateInfo::flags. See #VmaVirtualBlockCreateFlagBits. |
| 776 | typedef VkFlags VmaVirtualBlockCreateFlags; |
| 777 | |
| 778 | /// Flags to be passed as VmaVirtualAllocationCreateInfo::flags. |
| 779 | typedef enum VmaVirtualAllocationCreateFlagBits |
| 780 | { |
| 781 | /** \brief Allocation will be created from upper stack in a double stack pool. |
| 782 | |
| 783 | This flag is only allowed for virtual blocks created with #VMA_VIRTUAL_BLOCK_CREATE_LINEAR_ALGORITHM_BIT flag. |
| 784 | */ |
| 785 | VMA_VIRTUAL_ALLOCATION_CREATE_UPPER_ADDRESS_BIT = VMA_ALLOCATION_CREATE_UPPER_ADDRESS_BIT, |
| 786 | /** \brief Allocation strategy that tries to minimize memory usage. |
| 787 | */ |
| 788 | VMA_VIRTUAL_ALLOCATION_CREATE_STRATEGY_MIN_MEMORY_BIT = VMA_ALLOCATION_CREATE_STRATEGY_MIN_MEMORY_BIT, |
| 789 | /** \brief Allocation strategy that tries to minimize allocation time. |
| 790 | */ |
| 791 | VMA_VIRTUAL_ALLOCATION_CREATE_STRATEGY_MIN_TIME_BIT = VMA_ALLOCATION_CREATE_STRATEGY_MIN_TIME_BIT, |
| 792 | /** Allocation strategy that chooses always the lowest offset in available space. |
| 793 | This is not the most efficient strategy but achieves highly packed data. |
| 794 | */ |
| 795 | VMA_VIRTUAL_ALLOCATION_CREATE_STRATEGY_MIN_OFFSET_BIT = VMA_ALLOCATION_CREATE_STRATEGY_MIN_OFFSET_BIT, |
| 796 | /** \brief A bit mask to extract only `STRATEGY` bits from entire set of flags. |
| 797 | |
| 798 | These strategy flags are binary compatible with equivalent flags in #VmaAllocationCreateFlagBits. |
| 799 | */ |
| 800 | VMA_VIRTUAL_ALLOCATION_CREATE_STRATEGY_MASK = VMA_ALLOCATION_CREATE_STRATEGY_MASK, |
| 801 | |
| 802 | VMA_VIRTUAL_ALLOCATION_CREATE_FLAG_BITS_MAX_ENUM = 0x7FFFFFFF |
| 803 | } VmaVirtualAllocationCreateFlagBits; |
| 804 | /// Flags to be passed as VmaVirtualAllocationCreateInfo::flags. See #VmaVirtualAllocationCreateFlagBits. |
| 805 | typedef VkFlags VmaVirtualAllocationCreateFlags; |
| 806 | |
| 807 | /** @} */ |
| 808 | |
| 809 | #endif // _VMA_ENUM_DECLARATIONS |
| 810 | |
| 811 | #ifndef _VMA_DATA_TYPES_DECLARATIONS |
| 812 | |
| 813 | /** |
| 814 | \addtogroup group_init |
| 815 | @{ */ |
| 816 | |
| 817 | /** \struct VmaAllocator |
| 818 | \brief Represents main object of this library initialized. |
| 819 | |
| 820 | Fill structure #VmaAllocatorCreateInfo and call function vmaCreateAllocator() to create it. |
| 821 | Call function vmaDestroyAllocator() to destroy it. |
| 822 | |
| 823 | It is recommended to create just one object of this type per `VkDevice` object, |
| 824 | right after Vulkan is initialized and keep it alive until before Vulkan device is destroyed. |
| 825 | */ |
| 826 | VK_DEFINE_HANDLE(VmaAllocator) |
| 827 | |
| 828 | /** @} */ |
| 829 | |
| 830 | /** |
| 831 | \addtogroup group_alloc |
| 832 | @{ |
| 833 | */ |
| 834 | |
| 835 | /** \struct VmaPool |
| 836 | \brief Represents custom memory pool |
| 837 | |
| 838 | Fill structure VmaPoolCreateInfo and call function vmaCreatePool() to create it. |
| 839 | Call function vmaDestroyPool() to destroy it. |
| 840 | |
| 841 | For more information see [Custom memory pools](@ref choosing_memory_type_custom_memory_pools). |
| 842 | */ |
| 843 | VK_DEFINE_HANDLE(VmaPool) |
| 844 | |
| 845 | /** \struct VmaAllocation |
| 846 | \brief Represents single memory allocation. |
| 847 | |
| 848 | It may be either dedicated block of `VkDeviceMemory` or a specific region of a bigger block of this type |
| 849 | plus unique offset. |
| 850 | |
| 851 | There are multiple ways to create such object. |
| 852 | You need to fill structure VmaAllocationCreateInfo. |
| 853 | For more information see [Choosing memory type](@ref choosing_memory_type). |
| 854 | |
| 855 | Although the library provides convenience functions that create Vulkan buffer or image, |
| 856 | allocate memory for it and bind them together, |
| 857 | binding of the allocation to a buffer or an image is out of scope of the allocation itself. |
| 858 | Allocation object can exist without buffer/image bound, |
| 859 | binding can be done manually by the user, and destruction of it can be done |
| 860 | independently of destruction of the allocation. |
| 861 | |
| 862 | The object also remembers its size and some other information. |
| 863 | To retrieve this information, use function vmaGetAllocationInfo() and inspect |
| 864 | returned structure VmaAllocationInfo. |
| 865 | */ |
| 866 | VK_DEFINE_HANDLE(VmaAllocation) |
| 867 | |
| 868 | /** \struct VmaDefragmentationContext |
| 869 | \brief An opaque object that represents started defragmentation process. |
| 870 | |
| 871 | Fill structure #VmaDefragmentationInfo and call function vmaBeginDefragmentation() to create it. |
| 872 | Call function vmaEndDefragmentation() to destroy it. |
| 873 | */ |
| 874 | VK_DEFINE_HANDLE(VmaDefragmentationContext) |
| 875 | |
| 876 | /** @} */ |
| 877 | |
| 878 | /** |
| 879 | \addtogroup group_virtual |
| 880 | @{ |
| 881 | */ |
| 882 | |
| 883 | /** \struct VmaVirtualAllocation |
| 884 | \brief Represents single memory allocation done inside VmaVirtualBlock. |
| 885 | |
| 886 | Use it as a unique identifier to virtual allocation within the single block. |
| 887 | |
| 888 | Use value `VK_NULL_HANDLE` to represent a null/invalid allocation. |
| 889 | */ |
| 890 | VK_DEFINE_NON_DISPATCHABLE_HANDLE(VmaVirtualAllocation); |
| 891 | |
| 892 | /** @} */ |
| 893 | |
| 894 | /** |
| 895 | \addtogroup group_virtual |
| 896 | @{ |
| 897 | */ |
| 898 | |
| 899 | /** \struct VmaVirtualBlock |
| 900 | \brief Handle to a virtual block object that allows to use core allocation algorithm without allocating any real GPU memory. |
| 901 | |
| 902 | Fill in #VmaVirtualBlockCreateInfo structure and use vmaCreateVirtualBlock() to create it. Use vmaDestroyVirtualBlock() to destroy it. |
| 903 | For more information, see documentation chapter \ref virtual_allocator. |
| 904 | |
| 905 | This object is not thread-safe - should not be used from multiple threads simultaneously, must be synchronized externally. |
| 906 | */ |
| 907 | VK_DEFINE_HANDLE(VmaVirtualBlock) |
| 908 | |
| 909 | /** @} */ |
| 910 | |
| 911 | /** |
| 912 | \addtogroup group_init |
| 913 | @{ |
| 914 | */ |
| 915 | |
| 916 | /// Callback function called after successful vkAllocateMemory. |
| 917 | typedef void (VKAPI_PTR* PFN_vmaAllocateDeviceMemoryFunction)( |
| 918 | VmaAllocator VMA_NOT_NULL allocator, |
| 919 | uint32_t memoryType, |
| 920 | VkDeviceMemory VMA_NOT_NULL_NON_DISPATCHABLE memory, |
| 921 | VkDeviceSize size, |
| 922 | void* VMA_NULLABLE pUserData); |
| 923 | |
| 924 | /// Callback function called before vkFreeMemory. |
| 925 | typedef void (VKAPI_PTR* PFN_vmaFreeDeviceMemoryFunction)( |
| 926 | VmaAllocator VMA_NOT_NULL allocator, |
| 927 | uint32_t memoryType, |
| 928 | VkDeviceMemory VMA_NOT_NULL_NON_DISPATCHABLE memory, |
| 929 | VkDeviceSize size, |
| 930 | void* VMA_NULLABLE pUserData); |
| 931 | |
| 932 | /** \brief Set of callbacks that the library will call for `vkAllocateMemory` and `vkFreeMemory`. |
| 933 | |
| 934 | Provided for informative purpose, e.g. to gather statistics about number of |
| 935 | allocations or total amount of memory allocated in Vulkan. |
| 936 | |
| 937 | Used in VmaAllocatorCreateInfo::pDeviceMemoryCallbacks. |
| 938 | */ |
| 939 | typedef struct VmaDeviceMemoryCallbacks |
| 940 | { |
| 941 | /// Optional, can be null. |
| 942 | PFN_vmaAllocateDeviceMemoryFunction VMA_NULLABLE pfnAllocate; |
| 943 | /// Optional, can be null. |
| 944 | PFN_vmaFreeDeviceMemoryFunction VMA_NULLABLE pfnFree; |
| 945 | /// Optional, can be null. |
| 946 | void* VMA_NULLABLE pUserData; |
| 947 | } VmaDeviceMemoryCallbacks; |
| 948 | |
| 949 | /** \brief Pointers to some Vulkan functions - a subset used by the library. |
| 950 | |
| 951 | Used in VmaAllocatorCreateInfo::pVulkanFunctions. |
| 952 | */ |
| 953 | typedef struct VmaVulkanFunctions |
| 954 | { |
| 955 | /// Required when using VMA_DYNAMIC_VULKAN_FUNCTIONS. |
| 956 | PFN_vkGetInstanceProcAddr VMA_NULLABLE vkGetInstanceProcAddr; |
| 957 | /// Required when using VMA_DYNAMIC_VULKAN_FUNCTIONS. |
| 958 | PFN_vkGetDeviceProcAddr VMA_NULLABLE vkGetDeviceProcAddr; |
| 959 | PFN_vkGetPhysicalDeviceProperties VMA_NULLABLE vkGetPhysicalDeviceProperties; |
| 960 | PFN_vkGetPhysicalDeviceMemoryProperties VMA_NULLABLE vkGetPhysicalDeviceMemoryProperties; |
| 961 | PFN_vkAllocateMemory VMA_NULLABLE vkAllocateMemory; |
| 962 | PFN_vkFreeMemory VMA_NULLABLE vkFreeMemory; |
| 963 | PFN_vkMapMemory VMA_NULLABLE vkMapMemory; |
| 964 | PFN_vkUnmapMemory VMA_NULLABLE vkUnmapMemory; |
| 965 | PFN_vkFlushMappedMemoryRanges VMA_NULLABLE vkFlushMappedMemoryRanges; |
| 966 | PFN_vkInvalidateMappedMemoryRanges VMA_NULLABLE vkInvalidateMappedMemoryRanges; |
| 967 | PFN_vkBindBufferMemory VMA_NULLABLE vkBindBufferMemory; |
| 968 | PFN_vkBindImageMemory VMA_NULLABLE vkBindImageMemory; |
| 969 | PFN_vkGetBufferMemoryRequirements VMA_NULLABLE vkGetBufferMemoryRequirements; |
| 970 | PFN_vkGetImageMemoryRequirements VMA_NULLABLE vkGetImageMemoryRequirements; |
| 971 | PFN_vkCreateBuffer VMA_NULLABLE vkCreateBuffer; |
| 972 | PFN_vkDestroyBuffer VMA_NULLABLE vkDestroyBuffer; |
| 973 | PFN_vkCreateImage VMA_NULLABLE vkCreateImage; |
| 974 | PFN_vkDestroyImage VMA_NULLABLE vkDestroyImage; |
| 975 | PFN_vkCmdCopyBuffer VMA_NULLABLE vkCmdCopyBuffer; |
| 976 | #if VMA_DEDICATED_ALLOCATION || VMA_VULKAN_VERSION >= 1001000 |
| 977 | /// Fetch "vkGetBufferMemoryRequirements2" on Vulkan >= 1.1, fetch "vkGetBufferMemoryRequirements2KHR" when using VK_KHR_dedicated_allocation extension. |
| 978 | PFN_vkGetBufferMemoryRequirements2KHR VMA_NULLABLE vkGetBufferMemoryRequirements2KHR; |
| 979 | /// Fetch "vkGetImageMemoryRequirements2" on Vulkan >= 1.1, fetch "vkGetImageMemoryRequirements2KHR" when using VK_KHR_dedicated_allocation extension. |
| 980 | PFN_vkGetImageMemoryRequirements2KHR VMA_NULLABLE vkGetImageMemoryRequirements2KHR; |
| 981 | #endif |
| 982 | #if VMA_BIND_MEMORY2 || VMA_VULKAN_VERSION >= 1001000 |
| 983 | /// Fetch "vkBindBufferMemory2" on Vulkan >= 1.1, fetch "vkBindBufferMemory2KHR" when using VK_KHR_bind_memory2 extension. |
| 984 | PFN_vkBindBufferMemory2KHR VMA_NULLABLE vkBindBufferMemory2KHR; |
| 985 | /// Fetch "vkBindImageMemory2" on Vulkan >= 1.1, fetch "vkBindImageMemory2KHR" when using VK_KHR_bind_memory2 extension. |
| 986 | PFN_vkBindImageMemory2KHR VMA_NULLABLE vkBindImageMemory2KHR; |
| 987 | #endif |
| 988 | #if VMA_MEMORY_BUDGET || VMA_VULKAN_VERSION >= 1001000 |
| 989 | PFN_vkGetPhysicalDeviceMemoryProperties2KHR VMA_NULLABLE vkGetPhysicalDeviceMemoryProperties2KHR; |
| 990 | #endif |
| 991 | #if VMA_VULKAN_VERSION >= 1003000 |
| 992 | /// Fetch from "vkGetDeviceBufferMemoryRequirements" on Vulkan >= 1.3, but you can also fetch it from "vkGetDeviceBufferMemoryRequirementsKHR" if you enabled extension VK_KHR_maintenance4. |
| 993 | PFN_vkGetDeviceBufferMemoryRequirements VMA_NULLABLE vkGetDeviceBufferMemoryRequirements; |
| 994 | /// Fetch from "vkGetDeviceImageMemoryRequirements" on Vulkan >= 1.3, but you can also fetch it from "vkGetDeviceImageMemoryRequirementsKHR" if you enabled extension VK_KHR_maintenance4. |
| 995 | PFN_vkGetDeviceImageMemoryRequirements VMA_NULLABLE vkGetDeviceImageMemoryRequirements; |
| 996 | #endif |
| 997 | } VmaVulkanFunctions; |
| 998 | |
| 999 | /// Description of a Allocator to be created. |
| 1000 | typedef struct VmaAllocatorCreateInfo |
| 1001 | { |
| 1002 | /// Flags for created allocator. Use #VmaAllocatorCreateFlagBits enum. |
| 1003 | VmaAllocatorCreateFlags flags; |
| 1004 | /// Vulkan physical device. |
| 1005 | /** It must be valid throughout whole lifetime of created allocator. */ |
| 1006 | VkPhysicalDevice VMA_NOT_NULL physicalDevice; |
| 1007 | /// Vulkan device. |
| 1008 | /** It must be valid throughout whole lifetime of created allocator. */ |
| 1009 | VkDevice VMA_NOT_NULL device; |
| 1010 | /// Preferred size of a single `VkDeviceMemory` block to be allocated from large heaps > 1 GiB. Optional. |
| 1011 | /** Set to 0 to use default, which is currently 256 MiB. */ |
| 1012 | VkDeviceSize preferredLargeHeapBlockSize; |
| 1013 | /// Custom CPU memory allocation callbacks. Optional. |
| 1014 | /** Optional, can be null. When specified, will also be used for all CPU-side memory allocations. */ |
| 1015 | const VkAllocationCallbacks* VMA_NULLABLE pAllocationCallbacks; |
| 1016 | /// Informative callbacks for `vkAllocateMemory`, `vkFreeMemory`. Optional. |
| 1017 | /** Optional, can be null. */ |
| 1018 | const VmaDeviceMemoryCallbacks* VMA_NULLABLE pDeviceMemoryCallbacks; |
| 1019 | /** \brief Either null or a pointer to an array of limits on maximum number of bytes that can be allocated out of particular Vulkan memory heap. |
| 1020 | |
| 1021 | If not NULL, it must be a pointer to an array of |
| 1022 | `VkPhysicalDeviceMemoryProperties::memoryHeapCount` elements, defining limit on |
| 1023 | maximum number of bytes that can be allocated out of particular Vulkan memory |
| 1024 | heap. |
| 1025 | |
| 1026 | Any of the elements may be equal to `VK_WHOLE_SIZE`, which means no limit on that |
| 1027 | heap. This is also the default in case of `pHeapSizeLimit` = NULL. |
| 1028 | |
| 1029 | If there is a limit defined for a heap: |
| 1030 | |
| 1031 | - If user tries to allocate more memory from that heap using this allocator, |
| 1032 | the allocation fails with `VK_ERROR_OUT_OF_DEVICE_MEMORY`. |
| 1033 | - If the limit is smaller than heap size reported in `VkMemoryHeap::size`, the |
| 1034 | value of this limit will be reported instead when using vmaGetMemoryProperties(). |
| 1035 | |
| 1036 | Warning! Using this feature may not be equivalent to installing a GPU with |
| 1037 | smaller amount of memory, because graphics driver doesn't necessary fail new |
| 1038 | allocations with `VK_ERROR_OUT_OF_DEVICE_MEMORY` result when memory capacity is |
| 1039 | exceeded. It may return success and just silently migrate some device memory |
| 1040 | blocks to system RAM. This driver behavior can also be controlled using |
| 1041 | VK_AMD_memory_overallocation_behavior extension. |
| 1042 | */ |
| 1043 | const VkDeviceSize* VMA_NULLABLE VMA_LEN_IF_NOT_NULL("VkPhysicalDeviceMemoryProperties::memoryHeapCount" ) pHeapSizeLimit; |
| 1044 | |
| 1045 | /** \brief Pointers to Vulkan functions. Can be null. |
| 1046 | |
| 1047 | For details see [Pointers to Vulkan functions](@ref config_Vulkan_functions). |
| 1048 | */ |
| 1049 | const VmaVulkanFunctions* VMA_NULLABLE pVulkanFunctions; |
| 1050 | /** \brief Handle to Vulkan instance object. |
| 1051 | |
| 1052 | Starting from version 3.0.0 this member is no longer optional, it must be set! |
| 1053 | */ |
| 1054 | VkInstance VMA_NOT_NULL instance; |
| 1055 | /** \brief Optional. The highest version of Vulkan that the application is designed to use. |
| 1056 | |
| 1057 | It must be a value in the format as created by macro `VK_MAKE_VERSION` or a constant like: `VK_API_VERSION_1_1`, `VK_API_VERSION_1_0`. |
| 1058 | The patch version number specified is ignored. Only the major and minor versions are considered. |
| 1059 | It must be less or equal (preferably equal) to value as passed to `vkCreateInstance` as `VkApplicationInfo::apiVersion`. |
| 1060 | Only versions 1.0, 1.1, 1.2, 1.3 are supported by the current implementation. |
| 1061 | Leaving it initialized to zero is equivalent to `VK_API_VERSION_1_0`. |
| 1062 | */ |
| 1063 | uint32_t vulkanApiVersion; |
| 1064 | #if VMA_EXTERNAL_MEMORY |
| 1065 | /** \brief Either null or a pointer to an array of external memory handle types for each Vulkan memory type. |
| 1066 | |
| 1067 | If not NULL, it must be a pointer to an array of `VkPhysicalDeviceMemoryProperties::memoryTypeCount` |
| 1068 | elements, defining external memory handle types of particular Vulkan memory type, |
| 1069 | to be passed using `VkExportMemoryAllocateInfoKHR`. |
| 1070 | |
| 1071 | Any of the elements may be equal to 0, which means not to use `VkExportMemoryAllocateInfoKHR` on this memory type. |
| 1072 | This is also the default in case of `pTypeExternalMemoryHandleTypes` = NULL. |
| 1073 | */ |
| 1074 | const VkExternalMemoryHandleTypeFlagsKHR* VMA_NULLABLE VMA_LEN_IF_NOT_NULL("VkPhysicalDeviceMemoryProperties::memoryTypeCount" ) pTypeExternalMemoryHandleTypes; |
| 1075 | #endif // #if VMA_EXTERNAL_MEMORY |
| 1076 | } VmaAllocatorCreateInfo; |
| 1077 | |
| 1078 | /// Information about existing #VmaAllocator object. |
| 1079 | typedef struct VmaAllocatorInfo |
| 1080 | { |
| 1081 | /** \brief Handle to Vulkan instance object. |
| 1082 | |
| 1083 | This is the same value as has been passed through VmaAllocatorCreateInfo::instance. |
| 1084 | */ |
| 1085 | VkInstance VMA_NOT_NULL instance; |
| 1086 | /** \brief Handle to Vulkan physical device object. |
| 1087 | |
| 1088 | This is the same value as has been passed through VmaAllocatorCreateInfo::physicalDevice. |
| 1089 | */ |
| 1090 | VkPhysicalDevice VMA_NOT_NULL physicalDevice; |
| 1091 | /** \brief Handle to Vulkan device object. |
| 1092 | |
| 1093 | This is the same value as has been passed through VmaAllocatorCreateInfo::device. |
| 1094 | */ |
| 1095 | VkDevice VMA_NOT_NULL device; |
| 1096 | } VmaAllocatorInfo; |
| 1097 | |
| 1098 | /** @} */ |
| 1099 | |
| 1100 | /** |
| 1101 | \addtogroup group_stats |
| 1102 | @{ |
| 1103 | */ |
| 1104 | |
| 1105 | /** \brief Calculated statistics of memory usage e.g. in a specific memory type, heap, custom pool, or total. |
| 1106 | |
| 1107 | These are fast to calculate. |
| 1108 | See functions: vmaGetHeapBudgets(), vmaGetPoolStatistics(). |
| 1109 | */ |
| 1110 | typedef struct VmaStatistics |
| 1111 | { |
| 1112 | /** \brief Number of `VkDeviceMemory` objects - Vulkan memory blocks allocated. |
| 1113 | */ |
| 1114 | uint32_t blockCount; |
| 1115 | /** \brief Number of #VmaAllocation objects allocated. |
| 1116 | |
| 1117 | Dedicated allocations have their own blocks, so each one adds 1 to `allocationCount` as well as `blockCount`. |
| 1118 | */ |
| 1119 | uint32_t allocationCount; |
| 1120 | /** \brief Number of bytes allocated in `VkDeviceMemory` blocks. |
| 1121 | |
| 1122 | \note To avoid confusion, please be aware that what Vulkan calls an "allocation" - a whole `VkDeviceMemory` object |
| 1123 | (e.g. as in `VkPhysicalDeviceLimits::maxMemoryAllocationCount`) is called a "block" in VMA, while VMA calls |
| 1124 | "allocation" a #VmaAllocation object that represents a memory region sub-allocated from such block, usually for a single buffer or image. |
| 1125 | */ |
| 1126 | VkDeviceSize blockBytes; |
| 1127 | /** \brief Total number of bytes occupied by all #VmaAllocation objects. |
| 1128 | |
| 1129 | Always less or equal than `blockBytes`. |
| 1130 | Difference `(blockBytes - allocationBytes)` is the amount of memory allocated from Vulkan |
| 1131 | but unused by any #VmaAllocation. |
| 1132 | */ |
| 1133 | VkDeviceSize allocationBytes; |
| 1134 | } VmaStatistics; |
| 1135 | |
| 1136 | /** \brief More detailed statistics than #VmaStatistics. |
| 1137 | |
| 1138 | These are slower to calculate. Use for debugging purposes. |
| 1139 | See functions: vmaCalculateStatistics(), vmaCalculatePoolStatistics(). |
| 1140 | |
| 1141 | Previous version of the statistics API provided averages, but they have been removed |
| 1142 | because they can be easily calculated as: |
| 1143 | |
| 1144 | \code |
| 1145 | VkDeviceSize allocationSizeAvg = detailedStats.statistics.allocationBytes / detailedStats.statistics.allocationCount; |
| 1146 | VkDeviceSize unusedBytes = detailedStats.statistics.blockBytes - detailedStats.statistics.allocationBytes; |
| 1147 | VkDeviceSize unusedRangeSizeAvg = unusedBytes / detailedStats.unusedRangeCount; |
| 1148 | \endcode |
| 1149 | */ |
| 1150 | typedef struct VmaDetailedStatistics |
| 1151 | { |
| 1152 | /// Basic statistics. |
| 1153 | VmaStatistics statistics; |
| 1154 | /// Number of free ranges of memory between allocations. |
| 1155 | uint32_t unusedRangeCount; |
| 1156 | /// Smallest allocation size. `VK_WHOLE_SIZE` if there are 0 allocations. |
| 1157 | VkDeviceSize allocationSizeMin; |
| 1158 | /// Largest allocation size. 0 if there are 0 allocations. |
| 1159 | VkDeviceSize allocationSizeMax; |
| 1160 | /// Smallest empty range size. `VK_WHOLE_SIZE` if there are 0 empty ranges. |
| 1161 | VkDeviceSize unusedRangeSizeMin; |
| 1162 | /// Largest empty range size. 0 if there are 0 empty ranges. |
| 1163 | VkDeviceSize unusedRangeSizeMax; |
| 1164 | } VmaDetailedStatistics; |
| 1165 | |
| 1166 | /** \brief General statistics from current state of the Allocator - |
| 1167 | total memory usage across all memory heaps and types. |
| 1168 | |
| 1169 | These are slower to calculate. Use for debugging purposes. |
| 1170 | See function vmaCalculateStatistics(). |
| 1171 | */ |
| 1172 | typedef struct VmaTotalStatistics |
| 1173 | { |
| 1174 | VmaDetailedStatistics memoryType[VK_MAX_MEMORY_TYPES]; |
| 1175 | VmaDetailedStatistics memoryHeap[VK_MAX_MEMORY_HEAPS]; |
| 1176 | VmaDetailedStatistics total; |
| 1177 | } VmaTotalStatistics; |
| 1178 | |
| 1179 | /** \brief Statistics of current memory usage and available budget for a specific memory heap. |
| 1180 | |
| 1181 | These are fast to calculate. |
| 1182 | See function vmaGetHeapBudgets(). |
| 1183 | */ |
| 1184 | typedef struct VmaBudget |
| 1185 | { |
| 1186 | /** \brief Statistics fetched from the library. |
| 1187 | */ |
| 1188 | VmaStatistics statistics; |
| 1189 | /** \brief Estimated current memory usage of the program, in bytes. |
| 1190 | |
| 1191 | Fetched from system using VK_EXT_memory_budget extension if enabled. |
| 1192 | |
| 1193 | It might be different than `statistics.blockBytes` (usually higher) due to additional implicit objects |
| 1194 | also occupying the memory, like swapchain, pipelines, descriptor heaps, command buffers, or |
| 1195 | `VkDeviceMemory` blocks allocated outside of this library, if any. |
| 1196 | */ |
| 1197 | VkDeviceSize usage; |
| 1198 | /** \brief Estimated amount of memory available to the program, in bytes. |
| 1199 | |
| 1200 | Fetched from system using VK_EXT_memory_budget extension if enabled. |
| 1201 | |
| 1202 | It might be different (most probably smaller) than `VkMemoryHeap::size[heapIndex]` due to factors |
| 1203 | external to the program, decided by the operating system. |
| 1204 | Difference `budget - usage` is the amount of additional memory that can probably |
| 1205 | be allocated without problems. Exceeding the budget may result in various problems. |
| 1206 | */ |
| 1207 | VkDeviceSize budget; |
| 1208 | } VmaBudget; |
| 1209 | |
| 1210 | /** @} */ |
| 1211 | |
| 1212 | /** |
| 1213 | \addtogroup group_alloc |
| 1214 | @{ |
| 1215 | */ |
| 1216 | |
| 1217 | /** \brief Parameters of new #VmaAllocation. |
| 1218 | |
| 1219 | To be used with functions like vmaCreateBuffer(), vmaCreateImage(), and many others. |
| 1220 | */ |
| 1221 | typedef struct VmaAllocationCreateInfo |
| 1222 | { |
| 1223 | /// Use #VmaAllocationCreateFlagBits enum. |
| 1224 | VmaAllocationCreateFlags flags; |
| 1225 | /** \brief Intended usage of memory. |
| 1226 | |
| 1227 | You can leave #VMA_MEMORY_USAGE_UNKNOWN if you specify memory requirements in other way. \n |
| 1228 | If `pool` is not null, this member is ignored. |
| 1229 | */ |
| 1230 | VmaMemoryUsage usage; |
| 1231 | /** \brief Flags that must be set in a Memory Type chosen for an allocation. |
| 1232 | |
| 1233 | Leave 0 if you specify memory requirements in other way. \n |
| 1234 | If `pool` is not null, this member is ignored.*/ |
| 1235 | VkMemoryPropertyFlags requiredFlags; |
| 1236 | /** \brief Flags that preferably should be set in a memory type chosen for an allocation. |
| 1237 | |
| 1238 | Set to 0 if no additional flags are preferred. \n |
| 1239 | If `pool` is not null, this member is ignored. */ |
| 1240 | VkMemoryPropertyFlags preferredFlags; |
| 1241 | /** \brief Bitmask containing one bit set for every memory type acceptable for this allocation. |
| 1242 | |
| 1243 | Value 0 is equivalent to `UINT32_MAX` - it means any memory type is accepted if |
| 1244 | it meets other requirements specified by this structure, with no further |
| 1245 | restrictions on memory type index. \n |
| 1246 | If `pool` is not null, this member is ignored. |
| 1247 | */ |
| 1248 | uint32_t memoryTypeBits; |
| 1249 | /** \brief Pool that this allocation should be created in. |
| 1250 | |
| 1251 | Leave `VK_NULL_HANDLE` to allocate from default pool. If not null, members: |
| 1252 | `usage`, `requiredFlags`, `preferredFlags`, `memoryTypeBits` are ignored. |
| 1253 | */ |
| 1254 | VmaPool VMA_NULLABLE pool; |
| 1255 | /** \brief Custom general-purpose pointer that will be stored in #VmaAllocation, can be read as VmaAllocationInfo::pUserData and changed using vmaSetAllocationUserData(). |
| 1256 | |
| 1257 | If #VMA_ALLOCATION_CREATE_USER_DATA_COPY_STRING_BIT is used, it must be either |
| 1258 | null or pointer to a null-terminated string. The string will be then copied to |
| 1259 | internal buffer, so it doesn't need to be valid after allocation call. |
| 1260 | */ |
| 1261 | void* VMA_NULLABLE pUserData; |
| 1262 | /** \brief A floating-point value between 0 and 1, indicating the priority of the allocation relative to other memory allocations. |
| 1263 | |
| 1264 | It is used only when #VMA_ALLOCATOR_CREATE_EXT_MEMORY_PRIORITY_BIT flag was used during creation of the #VmaAllocator object |
| 1265 | and this allocation ends up as dedicated or is explicitly forced as dedicated using #VMA_ALLOCATION_CREATE_DEDICATED_MEMORY_BIT. |
| 1266 | Otherwise, it has the priority of a memory block where it is placed and this variable is ignored. |
| 1267 | */ |
| 1268 | float priority; |
| 1269 | } VmaAllocationCreateInfo; |
| 1270 | |
| 1271 | /// Describes parameter of created #VmaPool. |
| 1272 | typedef struct VmaPoolCreateInfo |
| 1273 | { |
| 1274 | /** \brief Vulkan memory type index to allocate this pool from. |
| 1275 | */ |
| 1276 | uint32_t memoryTypeIndex; |
| 1277 | /** \brief Use combination of #VmaPoolCreateFlagBits. |
| 1278 | */ |
| 1279 | VmaPoolCreateFlags flags; |
| 1280 | /** \brief Size of a single `VkDeviceMemory` block to be allocated as part of this pool, in bytes. Optional. |
| 1281 | |
| 1282 | Specify nonzero to set explicit, constant size of memory blocks used by this |
| 1283 | pool. |
| 1284 | |
| 1285 | Leave 0 to use default and let the library manage block sizes automatically. |
| 1286 | Sizes of particular blocks may vary. |
| 1287 | In this case, the pool will also support dedicated allocations. |
| 1288 | */ |
| 1289 | VkDeviceSize blockSize; |
| 1290 | /** \brief Minimum number of blocks to be always allocated in this pool, even if they stay empty. |
| 1291 | |
| 1292 | Set to 0 to have no preallocated blocks and allow the pool be completely empty. |
| 1293 | */ |
| 1294 | size_t minBlockCount; |
| 1295 | /** \brief Maximum number of blocks that can be allocated in this pool. Optional. |
| 1296 | |
| 1297 | Set to 0 to use default, which is `SIZE_MAX`, which means no limit. |
| 1298 | |
| 1299 | Set to same value as VmaPoolCreateInfo::minBlockCount to have fixed amount of memory allocated |
| 1300 | throughout whole lifetime of this pool. |
| 1301 | */ |
| 1302 | size_t maxBlockCount; |
| 1303 | /** \brief A floating-point value between 0 and 1, indicating the priority of the allocations in this pool relative to other memory allocations. |
| 1304 | |
| 1305 | It is used only when #VMA_ALLOCATOR_CREATE_EXT_MEMORY_PRIORITY_BIT flag was used during creation of the #VmaAllocator object. |
| 1306 | Otherwise, this variable is ignored. |
| 1307 | */ |
| 1308 | float priority; |
| 1309 | /** \brief Additional minimum alignment to be used for all allocations created from this pool. Can be 0. |
| 1310 | |
| 1311 | Leave 0 (default) not to impose any additional alignment. If not 0, it must be a power of two. |
| 1312 | It can be useful in cases where alignment returned by Vulkan by functions like `vkGetBufferMemoryRequirements` is not enough, |
| 1313 | e.g. when doing interop with OpenGL. |
| 1314 | */ |
| 1315 | VkDeviceSize minAllocationAlignment; |
| 1316 | /** \brief Additional `pNext` chain to be attached to `VkMemoryAllocateInfo` used for every allocation made by this pool. Optional. |
| 1317 | |
| 1318 | Optional, can be null. If not null, it must point to a `pNext` chain of structures that can be attached to `VkMemoryAllocateInfo`. |
| 1319 | It can be useful for special needs such as adding `VkExportMemoryAllocateInfoKHR`. |
| 1320 | Structures pointed by this member must remain alive and unchanged for the whole lifetime of the custom pool. |
| 1321 | |
| 1322 | Please note that some structures, e.g. `VkMemoryPriorityAllocateInfoEXT`, `VkMemoryDedicatedAllocateInfoKHR`, |
| 1323 | can be attached automatically by this library when using other, more convenient of its features. |
| 1324 | */ |
| 1325 | void* VMA_NULLABLE pMemoryAllocateNext; |
| 1326 | } VmaPoolCreateInfo; |
| 1327 | |
| 1328 | /** @} */ |
| 1329 | |
| 1330 | /** |
| 1331 | \addtogroup group_alloc |
| 1332 | @{ |
| 1333 | */ |
| 1334 | |
| 1335 | /// Parameters of #VmaAllocation objects, that can be retrieved using function vmaGetAllocationInfo(). |
| 1336 | typedef struct VmaAllocationInfo |
| 1337 | { |
| 1338 | /** \brief Memory type index that this allocation was allocated from. |
| 1339 | |
| 1340 | It never changes. |
| 1341 | */ |
| 1342 | uint32_t memoryType; |
| 1343 | /** \brief Handle to Vulkan memory object. |
| 1344 | |
| 1345 | Same memory object can be shared by multiple allocations. |
| 1346 | |
| 1347 | It can change after the allocation is moved during \ref defragmentation. |
| 1348 | */ |
| 1349 | VkDeviceMemory VMA_NULLABLE_NON_DISPATCHABLE deviceMemory; |
| 1350 | /** \brief Offset in `VkDeviceMemory` object to the beginning of this allocation, in bytes. `(deviceMemory, offset)` pair is unique to this allocation. |
| 1351 | |
| 1352 | You usually don't need to use this offset. If you create a buffer or an image together with the allocation using e.g. function |
| 1353 | vmaCreateBuffer(), vmaCreateImage(), functions that operate on these resources refer to the beginning of the buffer or image, |
| 1354 | not entire device memory block. Functions like vmaMapMemory(), vmaBindBufferMemory() also refer to the beginning of the allocation |
| 1355 | and apply this offset automatically. |
| 1356 | |
| 1357 | It can change after the allocation is moved during \ref defragmentation. |
| 1358 | */ |
| 1359 | VkDeviceSize offset; |
| 1360 | /** \brief Size of this allocation, in bytes. |
| 1361 | |
| 1362 | It never changes. |
| 1363 | |
| 1364 | \note Allocation size returned in this variable may be greater than the size |
| 1365 | requested for the resource e.g. as `VkBufferCreateInfo::size`. Whole size of the |
| 1366 | allocation is accessible for operations on memory e.g. using a pointer after |
| 1367 | mapping with vmaMapMemory(), but operations on the resource e.g. using |
| 1368 | `vkCmdCopyBuffer` must be limited to the size of the resource. |
| 1369 | */ |
| 1370 | VkDeviceSize size; |
| 1371 | /** \brief Pointer to the beginning of this allocation as mapped data. |
| 1372 | |
| 1373 | If the allocation hasn't been mapped using vmaMapMemory() and hasn't been |
| 1374 | created with #VMA_ALLOCATION_CREATE_MAPPED_BIT flag, this value is null. |
| 1375 | |
| 1376 | It can change after call to vmaMapMemory(), vmaUnmapMemory(). |
| 1377 | It can also change after the allocation is moved during \ref defragmentation. |
| 1378 | */ |
| 1379 | void* VMA_NULLABLE pMappedData; |
| 1380 | /** \brief Custom general-purpose pointer that was passed as VmaAllocationCreateInfo::pUserData or set using vmaSetAllocationUserData(). |
| 1381 | |
| 1382 | It can change after call to vmaSetAllocationUserData() for this allocation. |
| 1383 | */ |
| 1384 | void* VMA_NULLABLE pUserData; |
| 1385 | /** \brief Custom allocation name that was set with vmaSetAllocationName(). |
| 1386 | |
| 1387 | It can change after call to vmaSetAllocationName() for this allocation. |
| 1388 | |
| 1389 | Another way to set custom name is to pass it in VmaAllocationCreateInfo::pUserData with |
| 1390 | additional flag #VMA_ALLOCATION_CREATE_USER_DATA_COPY_STRING_BIT set [DEPRECATED]. |
| 1391 | */ |
| 1392 | const char* VMA_NULLABLE pName; |
| 1393 | } VmaAllocationInfo; |
| 1394 | |
| 1395 | /** \brief Parameters for defragmentation. |
| 1396 | |
| 1397 | To be used with function vmaBeginDefragmentation(). |
| 1398 | */ |
| 1399 | typedef struct VmaDefragmentationInfo |
| 1400 | { |
| 1401 | /// \brief Use combination of #VmaDefragmentationFlagBits. |
| 1402 | VmaDefragmentationFlags flags; |
| 1403 | /** \brief Custom pool to be defragmented. |
| 1404 | |
| 1405 | If null then default pools will undergo defragmentation process. |
| 1406 | */ |
| 1407 | VmaPool VMA_NULLABLE pool; |
| 1408 | /** \brief Maximum numbers of bytes that can be copied during single pass, while moving allocations to different places. |
| 1409 | |
| 1410 | `0` means no limit. |
| 1411 | */ |
| 1412 | VkDeviceSize maxBytesPerPass; |
| 1413 | /** \brief Maximum number of allocations that can be moved during single pass to a different place. |
| 1414 | |
| 1415 | `0` means no limit. |
| 1416 | */ |
| 1417 | uint32_t maxAllocationsPerPass; |
| 1418 | } VmaDefragmentationInfo; |
| 1419 | |
| 1420 | /// Single move of an allocation to be done for defragmentation. |
| 1421 | typedef struct VmaDefragmentationMove |
| 1422 | { |
| 1423 | /// Operation to be performed on the allocation by vmaEndDefragmentationPass(). Default value is #VMA_DEFRAGMENTATION_MOVE_OPERATION_COPY. You can modify it. |
| 1424 | VmaDefragmentationMoveOperation operation; |
| 1425 | /// Allocation that should be moved. |
| 1426 | VmaAllocation VMA_NOT_NULL srcAllocation; |
| 1427 | /** \brief Temporary allocation pointing to destination memory that will replace `srcAllocation`. |
| 1428 | |
| 1429 | \warning Do not store this allocation in your data structures! It exists only temporarily, for the duration of the defragmentation pass, |
| 1430 | to be used for binding new buffer/image to the destination memory using e.g. vmaBindBufferMemory(). |
| 1431 | vmaEndDefragmentationPass() will destroy it and make `srcAllocation` point to this memory. |
| 1432 | */ |
| 1433 | VmaAllocation VMA_NOT_NULL dstTmpAllocation; |
| 1434 | } VmaDefragmentationMove; |
| 1435 | |
| 1436 | /** \brief Parameters for incremental defragmentation steps. |
| 1437 | |
| 1438 | To be used with function vmaBeginDefragmentationPass(). |
| 1439 | */ |
| 1440 | typedef struct VmaDefragmentationPassMoveInfo |
| 1441 | { |
| 1442 | /// Number of elements in the `pMoves` array. |
| 1443 | uint32_t moveCount; |
| 1444 | /** \brief Array of moves to be performed by the user in the current defragmentation pass. |
| 1445 | |
| 1446 | Pointer to an array of `moveCount` elements, owned by VMA, created in vmaBeginDefragmentationPass(), destroyed in vmaEndDefragmentationPass(). |
| 1447 | |
| 1448 | For each element, you should: |
| 1449 | |
| 1450 | 1. Create a new buffer/image in the place pointed by VmaDefragmentationMove::dstMemory + VmaDefragmentationMove::dstOffset. |
| 1451 | 2. Copy data from the VmaDefragmentationMove::srcAllocation e.g. using `vkCmdCopyBuffer`, `vkCmdCopyImage`. |
| 1452 | 3. Make sure these commands finished executing on the GPU. |
| 1453 | 4. Destroy the old buffer/image. |
| 1454 | |
| 1455 | Only then you can finish defragmentation pass by calling vmaEndDefragmentationPass(). |
| 1456 | After this call, the allocation will point to the new place in memory. |
| 1457 | |
| 1458 | Alternatively, if you cannot move specific allocation, you can set VmaDefragmentationMove::operation to #VMA_DEFRAGMENTATION_MOVE_OPERATION_IGNORE. |
| 1459 | |
| 1460 | Alternatively, if you decide you want to completely remove the allocation: |
| 1461 | |
| 1462 | 1. Destroy its buffer/image. |
| 1463 | 2. Set VmaDefragmentationMove::operation to #VMA_DEFRAGMENTATION_MOVE_OPERATION_DESTROY. |
| 1464 | |
| 1465 | Then, after vmaEndDefragmentationPass() the allocation will be freed. |
| 1466 | */ |
| 1467 | VmaDefragmentationMove* VMA_NULLABLE VMA_LEN_IF_NOT_NULL(moveCount) pMoves; |
| 1468 | } VmaDefragmentationPassMoveInfo; |
| 1469 | |
| 1470 | /// Statistics returned for defragmentation process in function vmaEndDefragmentation(). |
| 1471 | typedef struct VmaDefragmentationStats |
| 1472 | { |
| 1473 | /// Total number of bytes that have been copied while moving allocations to different places. |
| 1474 | VkDeviceSize bytesMoved; |
| 1475 | /// Total number of bytes that have been released to the system by freeing empty `VkDeviceMemory` objects. |
| 1476 | VkDeviceSize bytesFreed; |
| 1477 | /// Number of allocations that have been moved to different places. |
| 1478 | uint32_t allocationsMoved; |
| 1479 | /// Number of empty `VkDeviceMemory` objects that have been released to the system. |
| 1480 | uint32_t deviceMemoryBlocksFreed; |
| 1481 | } VmaDefragmentationStats; |
| 1482 | |
| 1483 | /** @} */ |
| 1484 | |
| 1485 | /** |
| 1486 | \addtogroup group_virtual |
| 1487 | @{ |
| 1488 | */ |
| 1489 | |
| 1490 | /// Parameters of created #VmaVirtualBlock object to be passed to vmaCreateVirtualBlock(). |
| 1491 | typedef struct VmaVirtualBlockCreateInfo |
| 1492 | { |
| 1493 | /** \brief Total size of the virtual block. |
| 1494 | |
| 1495 | Sizes can be expressed in bytes or any units you want as long as you are consistent in using them. |
| 1496 | For example, if you allocate from some array of structures, 1 can mean single instance of entire structure. |
| 1497 | */ |
| 1498 | VkDeviceSize size; |
| 1499 | |
| 1500 | /** \brief Use combination of #VmaVirtualBlockCreateFlagBits. |
| 1501 | */ |
| 1502 | VmaVirtualBlockCreateFlags flags; |
| 1503 | |
| 1504 | /** \brief Custom CPU memory allocation callbacks. Optional. |
| 1505 | |
| 1506 | Optional, can be null. When specified, they will be used for all CPU-side memory allocations. |
| 1507 | */ |
| 1508 | const VkAllocationCallbacks* VMA_NULLABLE pAllocationCallbacks; |
| 1509 | } VmaVirtualBlockCreateInfo; |
| 1510 | |
| 1511 | /// Parameters of created virtual allocation to be passed to vmaVirtualAllocate(). |
| 1512 | typedef struct VmaVirtualAllocationCreateInfo |
| 1513 | { |
| 1514 | /** \brief Size of the allocation. |
| 1515 | |
| 1516 | Cannot be zero. |
| 1517 | */ |
| 1518 | VkDeviceSize size; |
| 1519 | /** \brief Required alignment of the allocation. Optional. |
| 1520 | |
| 1521 | Must be power of two. Special value 0 has the same meaning as 1 - means no special alignment is required, so allocation can start at any offset. |
| 1522 | */ |
| 1523 | VkDeviceSize alignment; |
| 1524 | /** \brief Use combination of #VmaVirtualAllocationCreateFlagBits. |
| 1525 | */ |
| 1526 | VmaVirtualAllocationCreateFlags flags; |
| 1527 | /** \brief Custom pointer to be associated with the allocation. Optional. |
| 1528 | |
| 1529 | It can be any value and can be used for user-defined purposes. It can be fetched or changed later. |
| 1530 | */ |
| 1531 | void* VMA_NULLABLE pUserData; |
| 1532 | } VmaVirtualAllocationCreateInfo; |
| 1533 | |
| 1534 | /// Parameters of an existing virtual allocation, returned by vmaGetVirtualAllocationInfo(). |
| 1535 | typedef struct VmaVirtualAllocationInfo |
| 1536 | { |
| 1537 | /** \brief Offset of the allocation. |
| 1538 | |
| 1539 | Offset at which the allocation was made. |
| 1540 | */ |
| 1541 | VkDeviceSize offset; |
| 1542 | /** \brief Size of the allocation. |
| 1543 | |
| 1544 | Same value as passed in VmaVirtualAllocationCreateInfo::size. |
| 1545 | */ |
| 1546 | VkDeviceSize size; |
| 1547 | /** \brief Custom pointer associated with the allocation. |
| 1548 | |
| 1549 | Same value as passed in VmaVirtualAllocationCreateInfo::pUserData or to vmaSetVirtualAllocationUserData(). |
| 1550 | */ |
| 1551 | void* VMA_NULLABLE pUserData; |
| 1552 | } VmaVirtualAllocationInfo; |
| 1553 | |
| 1554 | /** @} */ |
| 1555 | |
| 1556 | #endif // _VMA_DATA_TYPES_DECLARATIONS |
| 1557 | |
| 1558 | #ifndef _VMA_FUNCTION_HEADERS |
| 1559 | |
| 1560 | /** |
| 1561 | \addtogroup group_init |
| 1562 | @{ |
| 1563 | */ |
| 1564 | |
| 1565 | /// Creates #VmaAllocator object. |
| 1566 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaCreateAllocator( |
| 1567 | const VmaAllocatorCreateInfo* VMA_NOT_NULL pCreateInfo, |
| 1568 | VmaAllocator VMA_NULLABLE* VMA_NOT_NULL pAllocator); |
| 1569 | |
| 1570 | /// Destroys allocator object. |
| 1571 | VMA_CALL_PRE void VMA_CALL_POST vmaDestroyAllocator( |
| 1572 | VmaAllocator VMA_NULLABLE allocator); |
| 1573 | |
| 1574 | /** \brief Returns information about existing #VmaAllocator object - handle to Vulkan device etc. |
| 1575 | |
| 1576 | It might be useful if you want to keep just the #VmaAllocator handle and fetch other required handles to |
| 1577 | `VkPhysicalDevice`, `VkDevice` etc. every time using this function. |
| 1578 | */ |
| 1579 | VMA_CALL_PRE void VMA_CALL_POST vmaGetAllocatorInfo( |
| 1580 | VmaAllocator VMA_NOT_NULL allocator, |
| 1581 | VmaAllocatorInfo* VMA_NOT_NULL pAllocatorInfo); |
| 1582 | |
| 1583 | /** |
| 1584 | PhysicalDeviceProperties are fetched from physicalDevice by the allocator. |
| 1585 | You can access it here, without fetching it again on your own. |
| 1586 | */ |
| 1587 | VMA_CALL_PRE void VMA_CALL_POST vmaGetPhysicalDeviceProperties( |
| 1588 | VmaAllocator VMA_NOT_NULL allocator, |
| 1589 | const VkPhysicalDeviceProperties* VMA_NULLABLE* VMA_NOT_NULL ppPhysicalDeviceProperties); |
| 1590 | |
| 1591 | /** |
| 1592 | PhysicalDeviceMemoryProperties are fetched from physicalDevice by the allocator. |
| 1593 | You can access it here, without fetching it again on your own. |
| 1594 | */ |
| 1595 | VMA_CALL_PRE void VMA_CALL_POST vmaGetMemoryProperties( |
| 1596 | VmaAllocator VMA_NOT_NULL allocator, |
| 1597 | const VkPhysicalDeviceMemoryProperties* VMA_NULLABLE* VMA_NOT_NULL ppPhysicalDeviceMemoryProperties); |
| 1598 | |
| 1599 | /** |
| 1600 | \brief Given Memory Type Index, returns Property Flags of this memory type. |
| 1601 | |
| 1602 | This is just a convenience function. Same information can be obtained using |
| 1603 | vmaGetMemoryProperties(). |
| 1604 | */ |
| 1605 | VMA_CALL_PRE void VMA_CALL_POST vmaGetMemoryTypeProperties( |
| 1606 | VmaAllocator VMA_NOT_NULL allocator, |
| 1607 | uint32_t memoryTypeIndex, |
| 1608 | VkMemoryPropertyFlags* VMA_NOT_NULL pFlags); |
| 1609 | |
| 1610 | /** \brief Sets index of the current frame. |
| 1611 | */ |
| 1612 | VMA_CALL_PRE void VMA_CALL_POST vmaSetCurrentFrameIndex( |
| 1613 | VmaAllocator VMA_NOT_NULL allocator, |
| 1614 | uint32_t frameIndex); |
| 1615 | |
| 1616 | /** @} */ |
| 1617 | |
| 1618 | /** |
| 1619 | \addtogroup group_stats |
| 1620 | @{ |
| 1621 | */ |
| 1622 | |
| 1623 | /** \brief Retrieves statistics from current state of the Allocator. |
| 1624 | |
| 1625 | This function is called "calculate" not "get" because it has to traverse all |
| 1626 | internal data structures, so it may be quite slow. Use it for debugging purposes. |
| 1627 | For faster but more brief statistics suitable to be called every frame or every allocation, |
| 1628 | use vmaGetHeapBudgets(). |
| 1629 | |
| 1630 | Note that when using allocator from multiple threads, returned information may immediately |
| 1631 | become outdated. |
| 1632 | */ |
| 1633 | VMA_CALL_PRE void VMA_CALL_POST vmaCalculateStatistics( |
| 1634 | VmaAllocator VMA_NOT_NULL allocator, |
| 1635 | VmaTotalStatistics* VMA_NOT_NULL pStats); |
| 1636 | |
| 1637 | /** \brief Retrieves information about current memory usage and budget for all memory heaps. |
| 1638 | |
| 1639 | \param allocator |
| 1640 | \param[out] pBudgets Must point to array with number of elements at least equal to number of memory heaps in physical device used. |
| 1641 | |
| 1642 | This function is called "get" not "calculate" because it is very fast, suitable to be called |
| 1643 | every frame or every allocation. For more detailed statistics use vmaCalculateStatistics(). |
| 1644 | |
| 1645 | Note that when using allocator from multiple threads, returned information may immediately |
| 1646 | become outdated. |
| 1647 | */ |
| 1648 | VMA_CALL_PRE void VMA_CALL_POST vmaGetHeapBudgets( |
| 1649 | VmaAllocator VMA_NOT_NULL allocator, |
| 1650 | VmaBudget* VMA_NOT_NULL VMA_LEN_IF_NOT_NULL("VkPhysicalDeviceMemoryProperties::memoryHeapCount" ) pBudgets); |
| 1651 | |
| 1652 | /** @} */ |
| 1653 | |
| 1654 | /** |
| 1655 | \addtogroup group_alloc |
| 1656 | @{ |
| 1657 | */ |
| 1658 | |
| 1659 | /** |
| 1660 | \brief Helps to find memoryTypeIndex, given memoryTypeBits and VmaAllocationCreateInfo. |
| 1661 | |
| 1662 | This algorithm tries to find a memory type that: |
| 1663 | |
| 1664 | - Is allowed by memoryTypeBits. |
| 1665 | - Contains all the flags from pAllocationCreateInfo->requiredFlags. |
| 1666 | - Matches intended usage. |
| 1667 | - Has as many flags from pAllocationCreateInfo->preferredFlags as possible. |
| 1668 | |
| 1669 | \return Returns VK_ERROR_FEATURE_NOT_PRESENT if not found. Receiving such result |
| 1670 | from this function or any other allocating function probably means that your |
| 1671 | device doesn't support any memory type with requested features for the specific |
| 1672 | type of resource you want to use it for. Please check parameters of your |
| 1673 | resource, like image layout (OPTIMAL versus LINEAR) or mip level count. |
| 1674 | */ |
| 1675 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaFindMemoryTypeIndex( |
| 1676 | VmaAllocator VMA_NOT_NULL allocator, |
| 1677 | uint32_t memoryTypeBits, |
| 1678 | const VmaAllocationCreateInfo* VMA_NOT_NULL pAllocationCreateInfo, |
| 1679 | uint32_t* VMA_NOT_NULL pMemoryTypeIndex); |
| 1680 | |
| 1681 | /** |
| 1682 | \brief Helps to find memoryTypeIndex, given VkBufferCreateInfo and VmaAllocationCreateInfo. |
| 1683 | |
| 1684 | It can be useful e.g. to determine value to be used as VmaPoolCreateInfo::memoryTypeIndex. |
| 1685 | It internally creates a temporary, dummy buffer that never has memory bound. |
| 1686 | */ |
| 1687 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaFindMemoryTypeIndexForBufferInfo( |
| 1688 | VmaAllocator VMA_NOT_NULL allocator, |
| 1689 | const VkBufferCreateInfo* VMA_NOT_NULL pBufferCreateInfo, |
| 1690 | const VmaAllocationCreateInfo* VMA_NOT_NULL pAllocationCreateInfo, |
| 1691 | uint32_t* VMA_NOT_NULL pMemoryTypeIndex); |
| 1692 | |
| 1693 | /** |
| 1694 | \brief Helps to find memoryTypeIndex, given VkImageCreateInfo and VmaAllocationCreateInfo. |
| 1695 | |
| 1696 | It can be useful e.g. to determine value to be used as VmaPoolCreateInfo::memoryTypeIndex. |
| 1697 | It internally creates a temporary, dummy image that never has memory bound. |
| 1698 | */ |
| 1699 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaFindMemoryTypeIndexForImageInfo( |
| 1700 | VmaAllocator VMA_NOT_NULL allocator, |
| 1701 | const VkImageCreateInfo* VMA_NOT_NULL pImageCreateInfo, |
| 1702 | const VmaAllocationCreateInfo* VMA_NOT_NULL pAllocationCreateInfo, |
| 1703 | uint32_t* VMA_NOT_NULL pMemoryTypeIndex); |
| 1704 | |
| 1705 | /** \brief Allocates Vulkan device memory and creates #VmaPool object. |
| 1706 | |
| 1707 | \param allocator Allocator object. |
| 1708 | \param pCreateInfo Parameters of pool to create. |
| 1709 | \param[out] pPool Handle to created pool. |
| 1710 | */ |
| 1711 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaCreatePool( |
| 1712 | VmaAllocator VMA_NOT_NULL allocator, |
| 1713 | const VmaPoolCreateInfo* VMA_NOT_NULL pCreateInfo, |
| 1714 | VmaPool VMA_NULLABLE* VMA_NOT_NULL pPool); |
| 1715 | |
| 1716 | /** \brief Destroys #VmaPool object and frees Vulkan device memory. |
| 1717 | */ |
| 1718 | VMA_CALL_PRE void VMA_CALL_POST vmaDestroyPool( |
| 1719 | VmaAllocator VMA_NOT_NULL allocator, |
| 1720 | VmaPool VMA_NULLABLE pool); |
| 1721 | |
| 1722 | /** @} */ |
| 1723 | |
| 1724 | /** |
| 1725 | \addtogroup group_stats |
| 1726 | @{ |
| 1727 | */ |
| 1728 | |
| 1729 | /** \brief Retrieves statistics of existing #VmaPool object. |
| 1730 | |
| 1731 | \param allocator Allocator object. |
| 1732 | \param pool Pool object. |
| 1733 | \param[out] pPoolStats Statistics of specified pool. |
| 1734 | */ |
| 1735 | VMA_CALL_PRE void VMA_CALL_POST vmaGetPoolStatistics( |
| 1736 | VmaAllocator VMA_NOT_NULL allocator, |
| 1737 | VmaPool VMA_NOT_NULL pool, |
| 1738 | VmaStatistics* VMA_NOT_NULL pPoolStats); |
| 1739 | |
| 1740 | /** \brief Retrieves detailed statistics of existing #VmaPool object. |
| 1741 | |
| 1742 | \param allocator Allocator object. |
| 1743 | \param pool Pool object. |
| 1744 | \param[out] pPoolStats Statistics of specified pool. |
| 1745 | */ |
| 1746 | VMA_CALL_PRE void VMA_CALL_POST vmaCalculatePoolStatistics( |
| 1747 | VmaAllocator VMA_NOT_NULL allocator, |
| 1748 | VmaPool VMA_NOT_NULL pool, |
| 1749 | VmaDetailedStatistics* VMA_NOT_NULL pPoolStats); |
| 1750 | |
| 1751 | /** @} */ |
| 1752 | |
| 1753 | /** |
| 1754 | \addtogroup group_alloc |
| 1755 | @{ |
| 1756 | */ |
| 1757 | |
| 1758 | /** \brief Checks magic number in margins around all allocations in given memory pool in search for corruptions. |
| 1759 | |
| 1760 | Corruption detection is enabled only when `VMA_DEBUG_DETECT_CORRUPTION` macro is defined to nonzero, |
| 1761 | `VMA_DEBUG_MARGIN` is defined to nonzero and the pool is created in memory type that is |
| 1762 | `HOST_VISIBLE` and `HOST_COHERENT`. For more information, see [Corruption detection](@ref debugging_memory_usage_corruption_detection). |
| 1763 | |
| 1764 | Possible return values: |
| 1765 | |
| 1766 | - `VK_ERROR_FEATURE_NOT_PRESENT` - corruption detection is not enabled for specified pool. |
| 1767 | - `VK_SUCCESS` - corruption detection has been performed and succeeded. |
| 1768 | - `VK_ERROR_UNKNOWN` - corruption detection has been performed and found memory corruptions around one of the allocations. |
| 1769 | `VMA_ASSERT` is also fired in that case. |
| 1770 | - Other value: Error returned by Vulkan, e.g. memory mapping failure. |
| 1771 | */ |
| 1772 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaCheckPoolCorruption( |
| 1773 | VmaAllocator VMA_NOT_NULL allocator, |
| 1774 | VmaPool VMA_NOT_NULL pool); |
| 1775 | |
| 1776 | /** \brief Retrieves name of a custom pool. |
| 1777 | |
| 1778 | After the call `ppName` is either null or points to an internally-owned null-terminated string |
| 1779 | containing name of the pool that was previously set. The pointer becomes invalid when the pool is |
| 1780 | destroyed or its name is changed using vmaSetPoolName(). |
| 1781 | */ |
| 1782 | VMA_CALL_PRE void VMA_CALL_POST vmaGetPoolName( |
| 1783 | VmaAllocator VMA_NOT_NULL allocator, |
| 1784 | VmaPool VMA_NOT_NULL pool, |
| 1785 | const char* VMA_NULLABLE* VMA_NOT_NULL ppName); |
| 1786 | |
| 1787 | /** \brief Sets name of a custom pool. |
| 1788 | |
| 1789 | `pName` can be either null or pointer to a null-terminated string with new name for the pool. |
| 1790 | Function makes internal copy of the string, so it can be changed or freed immediately after this call. |
| 1791 | */ |
| 1792 | VMA_CALL_PRE void VMA_CALL_POST vmaSetPoolName( |
| 1793 | VmaAllocator VMA_NOT_NULL allocator, |
| 1794 | VmaPool VMA_NOT_NULL pool, |
| 1795 | const char* VMA_NULLABLE pName); |
| 1796 | |
| 1797 | /** \brief General purpose memory allocation. |
| 1798 | |
| 1799 | \param allocator |
| 1800 | \param pVkMemoryRequirements |
| 1801 | \param pCreateInfo |
| 1802 | \param[out] pAllocation Handle to allocated memory. |
| 1803 | \param[out] pAllocationInfo Optional. Information about allocated memory. It can be later fetched using function vmaGetAllocationInfo(). |
| 1804 | |
| 1805 | You should free the memory using vmaFreeMemory() or vmaFreeMemoryPages(). |
| 1806 | |
| 1807 | It is recommended to use vmaAllocateMemoryForBuffer(), vmaAllocateMemoryForImage(), |
| 1808 | vmaCreateBuffer(), vmaCreateImage() instead whenever possible. |
| 1809 | */ |
| 1810 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaAllocateMemory( |
| 1811 | VmaAllocator VMA_NOT_NULL allocator, |
| 1812 | const VkMemoryRequirements* VMA_NOT_NULL pVkMemoryRequirements, |
| 1813 | const VmaAllocationCreateInfo* VMA_NOT_NULL pCreateInfo, |
| 1814 | VmaAllocation VMA_NULLABLE* VMA_NOT_NULL pAllocation, |
| 1815 | VmaAllocationInfo* VMA_NULLABLE pAllocationInfo); |
| 1816 | |
| 1817 | /** \brief General purpose memory allocation for multiple allocation objects at once. |
| 1818 | |
| 1819 | \param allocator Allocator object. |
| 1820 | \param pVkMemoryRequirements Memory requirements for each allocation. |
| 1821 | \param pCreateInfo Creation parameters for each allocation. |
| 1822 | \param allocationCount Number of allocations to make. |
| 1823 | \param[out] pAllocations Pointer to array that will be filled with handles to created allocations. |
| 1824 | \param[out] pAllocationInfo Optional. Pointer to array that will be filled with parameters of created allocations. |
| 1825 | |
| 1826 | You should free the memory using vmaFreeMemory() or vmaFreeMemoryPages(). |
| 1827 | |
| 1828 | Word "pages" is just a suggestion to use this function to allocate pieces of memory needed for sparse binding. |
| 1829 | It is just a general purpose allocation function able to make multiple allocations at once. |
| 1830 | It may be internally optimized to be more efficient than calling vmaAllocateMemory() `allocationCount` times. |
| 1831 | |
| 1832 | All allocations are made using same parameters. All of them are created out of the same memory pool and type. |
| 1833 | If any allocation fails, all allocations already made within this function call are also freed, so that when |
| 1834 | returned result is not `VK_SUCCESS`, `pAllocation` array is always entirely filled with `VK_NULL_HANDLE`. |
| 1835 | */ |
| 1836 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaAllocateMemoryPages( |
| 1837 | VmaAllocator VMA_NOT_NULL allocator, |
| 1838 | const VkMemoryRequirements* VMA_NOT_NULL VMA_LEN_IF_NOT_NULL(allocationCount) pVkMemoryRequirements, |
| 1839 | const VmaAllocationCreateInfo* VMA_NOT_NULL VMA_LEN_IF_NOT_NULL(allocationCount) pCreateInfo, |
| 1840 | size_t allocationCount, |
| 1841 | VmaAllocation VMA_NULLABLE* VMA_NOT_NULL VMA_LEN_IF_NOT_NULL(allocationCount) pAllocations, |
| 1842 | VmaAllocationInfo* VMA_NULLABLE VMA_LEN_IF_NOT_NULL(allocationCount) pAllocationInfo); |
| 1843 | |
| 1844 | /** \brief Allocates memory suitable for given `VkBuffer`. |
| 1845 | |
| 1846 | \param allocator |
| 1847 | \param buffer |
| 1848 | \param pCreateInfo |
| 1849 | \param[out] pAllocation Handle to allocated memory. |
| 1850 | \param[out] pAllocationInfo Optional. Information about allocated memory. It can be later fetched using function vmaGetAllocationInfo(). |
| 1851 | |
| 1852 | It only creates #VmaAllocation. To bind the memory to the buffer, use vmaBindBufferMemory(). |
| 1853 | |
| 1854 | This is a special-purpose function. In most cases you should use vmaCreateBuffer(). |
| 1855 | |
| 1856 | You must free the allocation using vmaFreeMemory() when no longer needed. |
| 1857 | */ |
| 1858 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaAllocateMemoryForBuffer( |
| 1859 | VmaAllocator VMA_NOT_NULL allocator, |
| 1860 | VkBuffer VMA_NOT_NULL_NON_DISPATCHABLE buffer, |
| 1861 | const VmaAllocationCreateInfo* VMA_NOT_NULL pCreateInfo, |
| 1862 | VmaAllocation VMA_NULLABLE* VMA_NOT_NULL pAllocation, |
| 1863 | VmaAllocationInfo* VMA_NULLABLE pAllocationInfo); |
| 1864 | |
| 1865 | /** \brief Allocates memory suitable for given `VkImage`. |
| 1866 | |
| 1867 | \param allocator |
| 1868 | \param image |
| 1869 | \param pCreateInfo |
| 1870 | \param[out] pAllocation Handle to allocated memory. |
| 1871 | \param[out] pAllocationInfo Optional. Information about allocated memory. It can be later fetched using function vmaGetAllocationInfo(). |
| 1872 | |
| 1873 | It only creates #VmaAllocation. To bind the memory to the buffer, use vmaBindImageMemory(). |
| 1874 | |
| 1875 | This is a special-purpose function. In most cases you should use vmaCreateImage(). |
| 1876 | |
| 1877 | You must free the allocation using vmaFreeMemory() when no longer needed. |
| 1878 | */ |
| 1879 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaAllocateMemoryForImage( |
| 1880 | VmaAllocator VMA_NOT_NULL allocator, |
| 1881 | VkImage VMA_NOT_NULL_NON_DISPATCHABLE image, |
| 1882 | const VmaAllocationCreateInfo* VMA_NOT_NULL pCreateInfo, |
| 1883 | VmaAllocation VMA_NULLABLE* VMA_NOT_NULL pAllocation, |
| 1884 | VmaAllocationInfo* VMA_NULLABLE pAllocationInfo); |
| 1885 | |
| 1886 | /** \brief Frees memory previously allocated using vmaAllocateMemory(), vmaAllocateMemoryForBuffer(), or vmaAllocateMemoryForImage(). |
| 1887 | |
| 1888 | Passing `VK_NULL_HANDLE` as `allocation` is valid. Such function call is just skipped. |
| 1889 | */ |
| 1890 | VMA_CALL_PRE void VMA_CALL_POST vmaFreeMemory( |
| 1891 | VmaAllocator VMA_NOT_NULL allocator, |
| 1892 | const VmaAllocation VMA_NULLABLE allocation); |
| 1893 | |
| 1894 | /** \brief Frees memory and destroys multiple allocations. |
| 1895 | |
| 1896 | Word "pages" is just a suggestion to use this function to free pieces of memory used for sparse binding. |
| 1897 | It is just a general purpose function to free memory and destroy allocations made using e.g. vmaAllocateMemory(), |
| 1898 | vmaAllocateMemoryPages() and other functions. |
| 1899 | It may be internally optimized to be more efficient than calling vmaFreeMemory() `allocationCount` times. |
| 1900 | |
| 1901 | Allocations in `pAllocations` array can come from any memory pools and types. |
| 1902 | Passing `VK_NULL_HANDLE` as elements of `pAllocations` array is valid. Such entries are just skipped. |
| 1903 | */ |
| 1904 | VMA_CALL_PRE void VMA_CALL_POST vmaFreeMemoryPages( |
| 1905 | VmaAllocator VMA_NOT_NULL allocator, |
| 1906 | size_t allocationCount, |
| 1907 | const VmaAllocation VMA_NULLABLE* VMA_NOT_NULL VMA_LEN_IF_NOT_NULL(allocationCount) pAllocations); |
| 1908 | |
| 1909 | /** \brief Returns current information about specified allocation. |
| 1910 | |
| 1911 | Current paramteres of given allocation are returned in `pAllocationInfo`. |
| 1912 | |
| 1913 | Although this function doesn't lock any mutex, so it should be quite efficient, |
| 1914 | you should avoid calling it too often. |
| 1915 | You can retrieve same VmaAllocationInfo structure while creating your resource, from function |
| 1916 | vmaCreateBuffer(), vmaCreateImage(). You can remember it if you are sure parameters don't change |
| 1917 | (e.g. due to defragmentation). |
| 1918 | */ |
| 1919 | VMA_CALL_PRE void VMA_CALL_POST vmaGetAllocationInfo( |
| 1920 | VmaAllocator VMA_NOT_NULL allocator, |
| 1921 | VmaAllocation VMA_NOT_NULL allocation, |
| 1922 | VmaAllocationInfo* VMA_NOT_NULL pAllocationInfo); |
| 1923 | |
| 1924 | /** \brief Sets pUserData in given allocation to new value. |
| 1925 | |
| 1926 | The value of pointer `pUserData` is copied to allocation's `pUserData`. |
| 1927 | It is opaque, so you can use it however you want - e.g. |
| 1928 | as a pointer, ordinal number or some handle to you own data. |
| 1929 | */ |
| 1930 | VMA_CALL_PRE void VMA_CALL_POST vmaSetAllocationUserData( |
| 1931 | VmaAllocator VMA_NOT_NULL allocator, |
| 1932 | VmaAllocation VMA_NOT_NULL allocation, |
| 1933 | void* VMA_NULLABLE pUserData); |
| 1934 | |
| 1935 | /** \brief Sets pName in given allocation to new value. |
| 1936 | |
| 1937 | `pName` must be either null, or pointer to a null-terminated string. The function |
| 1938 | makes local copy of the string and sets it as allocation's `pName`. String |
| 1939 | passed as pName doesn't need to be valid for whole lifetime of the allocation - |
| 1940 | you can free it after this call. String previously pointed by allocation's |
| 1941 | `pName` is freed from memory. |
| 1942 | */ |
| 1943 | VMA_CALL_PRE void VMA_CALL_POST vmaSetAllocationName( |
| 1944 | VmaAllocator VMA_NOT_NULL allocator, |
| 1945 | VmaAllocation VMA_NOT_NULL allocation, |
| 1946 | const char* VMA_NULLABLE pName); |
| 1947 | |
| 1948 | /** |
| 1949 | \brief Given an allocation, returns Property Flags of its memory type. |
| 1950 | |
| 1951 | This is just a convenience function. Same information can be obtained using |
| 1952 | vmaGetAllocationInfo() + vmaGetMemoryProperties(). |
| 1953 | */ |
| 1954 | VMA_CALL_PRE void VMA_CALL_POST vmaGetAllocationMemoryProperties( |
| 1955 | VmaAllocator VMA_NOT_NULL allocator, |
| 1956 | VmaAllocation VMA_NOT_NULL allocation, |
| 1957 | VkMemoryPropertyFlags* VMA_NOT_NULL pFlags); |
| 1958 | |
| 1959 | /** \brief Maps memory represented by given allocation and returns pointer to it. |
| 1960 | |
| 1961 | Maps memory represented by given allocation to make it accessible to CPU code. |
| 1962 | When succeeded, `*ppData` contains pointer to first byte of this memory. |
| 1963 | |
| 1964 | \warning |
| 1965 | If the allocation is part of a bigger `VkDeviceMemory` block, returned pointer is |
| 1966 | correctly offsetted to the beginning of region assigned to this particular allocation. |
| 1967 | Unlike the result of `vkMapMemory`, it points to the allocation, not to the beginning of the whole block. |
| 1968 | You should not add VmaAllocationInfo::offset to it! |
| 1969 | |
| 1970 | Mapping is internally reference-counted and synchronized, so despite raw Vulkan |
| 1971 | function `vkMapMemory()` cannot be used to map same block of `VkDeviceMemory` |
| 1972 | multiple times simultaneously, it is safe to call this function on allocations |
| 1973 | assigned to the same memory block. Actual Vulkan memory will be mapped on first |
| 1974 | mapping and unmapped on last unmapping. |
| 1975 | |
| 1976 | If the function succeeded, you must call vmaUnmapMemory() to unmap the |
| 1977 | allocation when mapping is no longer needed or before freeing the allocation, at |
| 1978 | the latest. |
| 1979 | |
| 1980 | It also safe to call this function multiple times on the same allocation. You |
| 1981 | must call vmaUnmapMemory() same number of times as you called vmaMapMemory(). |
| 1982 | |
| 1983 | It is also safe to call this function on allocation created with |
| 1984 | #VMA_ALLOCATION_CREATE_MAPPED_BIT flag. Its memory stays mapped all the time. |
| 1985 | You must still call vmaUnmapMemory() same number of times as you called |
| 1986 | vmaMapMemory(). You must not call vmaUnmapMemory() additional time to free the |
| 1987 | "0-th" mapping made automatically due to #VMA_ALLOCATION_CREATE_MAPPED_BIT flag. |
| 1988 | |
| 1989 | This function fails when used on allocation made in memory type that is not |
| 1990 | `HOST_VISIBLE`. |
| 1991 | |
| 1992 | This function doesn't automatically flush or invalidate caches. |
| 1993 | If the allocation is made from a memory types that is not `HOST_COHERENT`, |
| 1994 | you also need to use vmaInvalidateAllocation() / vmaFlushAllocation(), as required by Vulkan specification. |
| 1995 | */ |
| 1996 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaMapMemory( |
| 1997 | VmaAllocator VMA_NOT_NULL allocator, |
| 1998 | VmaAllocation VMA_NOT_NULL allocation, |
| 1999 | void* VMA_NULLABLE* VMA_NOT_NULL ppData); |
| 2000 | |
| 2001 | /** \brief Unmaps memory represented by given allocation, mapped previously using vmaMapMemory(). |
| 2002 | |
| 2003 | For details, see description of vmaMapMemory(). |
| 2004 | |
| 2005 | This function doesn't automatically flush or invalidate caches. |
| 2006 | If the allocation is made from a memory types that is not `HOST_COHERENT`, |
| 2007 | you also need to use vmaInvalidateAllocation() / vmaFlushAllocation(), as required by Vulkan specification. |
| 2008 | */ |
| 2009 | VMA_CALL_PRE void VMA_CALL_POST vmaUnmapMemory( |
| 2010 | VmaAllocator VMA_NOT_NULL allocator, |
| 2011 | VmaAllocation VMA_NOT_NULL allocation); |
| 2012 | |
| 2013 | /** \brief Flushes memory of given allocation. |
| 2014 | |
| 2015 | Calls `vkFlushMappedMemoryRanges()` for memory associated with given range of given allocation. |
| 2016 | It needs to be called after writing to a mapped memory for memory types that are not `HOST_COHERENT`. |
| 2017 | Unmap operation doesn't do that automatically. |
| 2018 | |
| 2019 | - `offset` must be relative to the beginning of allocation. |
| 2020 | - `size` can be `VK_WHOLE_SIZE`. It means all memory from `offset` the the end of given allocation. |
| 2021 | - `offset` and `size` don't have to be aligned. |
| 2022 | They are internally rounded down/up to multiply of `nonCoherentAtomSize`. |
| 2023 | - If `size` is 0, this call is ignored. |
| 2024 | - If memory type that the `allocation` belongs to is not `HOST_VISIBLE` or it is `HOST_COHERENT`, |
| 2025 | this call is ignored. |
| 2026 | |
| 2027 | Warning! `offset` and `size` are relative to the contents of given `allocation`. |
| 2028 | If you mean whole allocation, you can pass 0 and `VK_WHOLE_SIZE`, respectively. |
| 2029 | Do not pass allocation's offset as `offset`!!! |
| 2030 | |
| 2031 | This function returns the `VkResult` from `vkFlushMappedMemoryRanges` if it is |
| 2032 | called, otherwise `VK_SUCCESS`. |
| 2033 | */ |
| 2034 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaFlushAllocation( |
| 2035 | VmaAllocator VMA_NOT_NULL allocator, |
| 2036 | VmaAllocation VMA_NOT_NULL allocation, |
| 2037 | VkDeviceSize offset, |
| 2038 | VkDeviceSize size); |
| 2039 | |
| 2040 | /** \brief Invalidates memory of given allocation. |
| 2041 | |
| 2042 | Calls `vkInvalidateMappedMemoryRanges()` for memory associated with given range of given allocation. |
| 2043 | It needs to be called before reading from a mapped memory for memory types that are not `HOST_COHERENT`. |
| 2044 | Map operation doesn't do that automatically. |
| 2045 | |
| 2046 | - `offset` must be relative to the beginning of allocation. |
| 2047 | - `size` can be `VK_WHOLE_SIZE`. It means all memory from `offset` the the end of given allocation. |
| 2048 | - `offset` and `size` don't have to be aligned. |
| 2049 | They are internally rounded down/up to multiply of `nonCoherentAtomSize`. |
| 2050 | - If `size` is 0, this call is ignored. |
| 2051 | - If memory type that the `allocation` belongs to is not `HOST_VISIBLE` or it is `HOST_COHERENT`, |
| 2052 | this call is ignored. |
| 2053 | |
| 2054 | Warning! `offset` and `size` are relative to the contents of given `allocation`. |
| 2055 | If you mean whole allocation, you can pass 0 and `VK_WHOLE_SIZE`, respectively. |
| 2056 | Do not pass allocation's offset as `offset`!!! |
| 2057 | |
| 2058 | This function returns the `VkResult` from `vkInvalidateMappedMemoryRanges` if |
| 2059 | it is called, otherwise `VK_SUCCESS`. |
| 2060 | */ |
| 2061 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaInvalidateAllocation( |
| 2062 | VmaAllocator VMA_NOT_NULL allocator, |
| 2063 | VmaAllocation VMA_NOT_NULL allocation, |
| 2064 | VkDeviceSize offset, |
| 2065 | VkDeviceSize size); |
| 2066 | |
| 2067 | /** \brief Flushes memory of given set of allocations. |
| 2068 | |
| 2069 | Calls `vkFlushMappedMemoryRanges()` for memory associated with given ranges of given allocations. |
| 2070 | For more information, see documentation of vmaFlushAllocation(). |
| 2071 | |
| 2072 | \param allocator |
| 2073 | \param allocationCount |
| 2074 | \param allocations |
| 2075 | \param offsets If not null, it must point to an array of offsets of regions to flush, relative to the beginning of respective allocations. Null means all ofsets are zero. |
| 2076 | \param sizes If not null, it must point to an array of sizes of regions to flush in respective allocations. Null means `VK_WHOLE_SIZE` for all allocations. |
| 2077 | |
| 2078 | This function returns the `VkResult` from `vkFlushMappedMemoryRanges` if it is |
| 2079 | called, otherwise `VK_SUCCESS`. |
| 2080 | */ |
| 2081 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaFlushAllocations( |
| 2082 | VmaAllocator VMA_NOT_NULL allocator, |
| 2083 | uint32_t allocationCount, |
| 2084 | const VmaAllocation VMA_NOT_NULL* VMA_NULLABLE VMA_LEN_IF_NOT_NULL(allocationCount) allocations, |
| 2085 | const VkDeviceSize* VMA_NULLABLE VMA_LEN_IF_NOT_NULL(allocationCount) offsets, |
| 2086 | const VkDeviceSize* VMA_NULLABLE VMA_LEN_IF_NOT_NULL(allocationCount) sizes); |
| 2087 | |
| 2088 | /** \brief Invalidates memory of given set of allocations. |
| 2089 | |
| 2090 | Calls `vkInvalidateMappedMemoryRanges()` for memory associated with given ranges of given allocations. |
| 2091 | For more information, see documentation of vmaInvalidateAllocation(). |
| 2092 | |
| 2093 | \param allocator |
| 2094 | \param allocationCount |
| 2095 | \param allocations |
| 2096 | \param offsets If not null, it must point to an array of offsets of regions to flush, relative to the beginning of respective allocations. Null means all ofsets are zero. |
| 2097 | \param sizes If not null, it must point to an array of sizes of regions to flush in respective allocations. Null means `VK_WHOLE_SIZE` for all allocations. |
| 2098 | |
| 2099 | This function returns the `VkResult` from `vkInvalidateMappedMemoryRanges` if it is |
| 2100 | called, otherwise `VK_SUCCESS`. |
| 2101 | */ |
| 2102 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaInvalidateAllocations( |
| 2103 | VmaAllocator VMA_NOT_NULL allocator, |
| 2104 | uint32_t allocationCount, |
| 2105 | const VmaAllocation VMA_NOT_NULL* VMA_NULLABLE VMA_LEN_IF_NOT_NULL(allocationCount) allocations, |
| 2106 | const VkDeviceSize* VMA_NULLABLE VMA_LEN_IF_NOT_NULL(allocationCount) offsets, |
| 2107 | const VkDeviceSize* VMA_NULLABLE VMA_LEN_IF_NOT_NULL(allocationCount) sizes); |
| 2108 | |
| 2109 | /** \brief Checks magic number in margins around all allocations in given memory types (in both default and custom pools) in search for corruptions. |
| 2110 | |
| 2111 | \param allocator |
| 2112 | \param memoryTypeBits Bit mask, where each bit set means that a memory type with that index should be checked. |
| 2113 | |
| 2114 | Corruption detection is enabled only when `VMA_DEBUG_DETECT_CORRUPTION` macro is defined to nonzero, |
| 2115 | `VMA_DEBUG_MARGIN` is defined to nonzero and only for memory types that are |
| 2116 | `HOST_VISIBLE` and `HOST_COHERENT`. For more information, see [Corruption detection](@ref debugging_memory_usage_corruption_detection). |
| 2117 | |
| 2118 | Possible return values: |
| 2119 | |
| 2120 | - `VK_ERROR_FEATURE_NOT_PRESENT` - corruption detection is not enabled for any of specified memory types. |
| 2121 | - `VK_SUCCESS` - corruption detection has been performed and succeeded. |
| 2122 | - `VK_ERROR_UNKNOWN` - corruption detection has been performed and found memory corruptions around one of the allocations. |
| 2123 | `VMA_ASSERT` is also fired in that case. |
| 2124 | - Other value: Error returned by Vulkan, e.g. memory mapping failure. |
| 2125 | */ |
| 2126 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaCheckCorruption( |
| 2127 | VmaAllocator VMA_NOT_NULL allocator, |
| 2128 | uint32_t memoryTypeBits); |
| 2129 | |
| 2130 | /** \brief Begins defragmentation process. |
| 2131 | |
| 2132 | \param allocator Allocator object. |
| 2133 | \param pInfo Structure filled with parameters of defragmentation. |
| 2134 | \param[out] pContext Context object that must be passed to vmaEndDefragmentation() to finish defragmentation. |
| 2135 | \returns |
| 2136 | - `VK_SUCCESS` if defragmentation can begin. |
| 2137 | - `VK_ERROR_FEATURE_NOT_PRESENT` if defragmentation is not supported. |
| 2138 | |
| 2139 | For more information about defragmentation, see documentation chapter: |
| 2140 | [Defragmentation](@ref defragmentation). |
| 2141 | */ |
| 2142 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaBeginDefragmentation( |
| 2143 | VmaAllocator VMA_NOT_NULL allocator, |
| 2144 | const VmaDefragmentationInfo* VMA_NOT_NULL pInfo, |
| 2145 | VmaDefragmentationContext VMA_NULLABLE* VMA_NOT_NULL pContext); |
| 2146 | |
| 2147 | /** \brief Ends defragmentation process. |
| 2148 | |
| 2149 | \param allocator Allocator object. |
| 2150 | \param context Context object that has been created by vmaBeginDefragmentation(). |
| 2151 | \param[out] pStats Optional stats for the defragmentation. Can be null. |
| 2152 | |
| 2153 | Use this function to finish defragmentation started by vmaBeginDefragmentation(). |
| 2154 | */ |
| 2155 | VMA_CALL_PRE void VMA_CALL_POST vmaEndDefragmentation( |
| 2156 | VmaAllocator VMA_NOT_NULL allocator, |
| 2157 | VmaDefragmentationContext VMA_NOT_NULL context, |
| 2158 | VmaDefragmentationStats* VMA_NULLABLE pStats); |
| 2159 | |
| 2160 | /** \brief Starts single defragmentation pass. |
| 2161 | |
| 2162 | \param allocator Allocator object. |
| 2163 | \param context Context object that has been created by vmaBeginDefragmentation(). |
| 2164 | \param[out] pPassInfo Computed informations for current pass. |
| 2165 | \returns |
| 2166 | - `VK_SUCCESS` if no more moves are possible. Then you can omit call to vmaEndDefragmentationPass() and simply end whole defragmentation. |
| 2167 | - `VK_INCOMPLETE` if there are pending moves returned in `pPassInfo`. You need to perform them, call vmaEndDefragmentationPass(), |
| 2168 | and then preferably try another pass with vmaBeginDefragmentationPass(). |
| 2169 | */ |
| 2170 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaBeginDefragmentationPass( |
| 2171 | VmaAllocator VMA_NOT_NULL allocator, |
| 2172 | VmaDefragmentationContext VMA_NOT_NULL context, |
| 2173 | VmaDefragmentationPassMoveInfo* VMA_NOT_NULL pPassInfo); |
| 2174 | |
| 2175 | /** \brief Ends single defragmentation pass. |
| 2176 | |
| 2177 | \param allocator Allocator object. |
| 2178 | \param context Context object that has been created by vmaBeginDefragmentation(). |
| 2179 | \param pPassInfo Computed informations for current pass filled by vmaBeginDefragmentationPass() and possibly modified by you. |
| 2180 | |
| 2181 | Returns `VK_SUCCESS` if no more moves are possible or `VK_INCOMPLETE` if more defragmentations are possible. |
| 2182 | |
| 2183 | Ends incremental defragmentation pass and commits all defragmentation moves from `pPassInfo`. |
| 2184 | After this call: |
| 2185 | |
| 2186 | - Allocations at `pPassInfo[i].srcAllocation` that had `pPassInfo[i].operation ==` #VMA_DEFRAGMENTATION_MOVE_OPERATION_COPY |
| 2187 | (which is the default) will be pointing to the new destination place. |
| 2188 | - Allocation at `pPassInfo[i].srcAllocation` that had `pPassInfo[i].operation ==` #VMA_DEFRAGMENTATION_MOVE_OPERATION_DESTROY |
| 2189 | will be freed. |
| 2190 | |
| 2191 | If no more moves are possible you can end whole defragmentation. |
| 2192 | */ |
| 2193 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaEndDefragmentationPass( |
| 2194 | VmaAllocator VMA_NOT_NULL allocator, |
| 2195 | VmaDefragmentationContext VMA_NOT_NULL context, |
| 2196 | VmaDefragmentationPassMoveInfo* VMA_NOT_NULL pPassInfo); |
| 2197 | |
| 2198 | /** \brief Binds buffer to allocation. |
| 2199 | |
| 2200 | Binds specified buffer to region of memory represented by specified allocation. |
| 2201 | Gets `VkDeviceMemory` handle and offset from the allocation. |
| 2202 | If you want to create a buffer, allocate memory for it and bind them together separately, |
| 2203 | you should use this function for binding instead of standard `vkBindBufferMemory()`, |
| 2204 | because it ensures proper synchronization so that when a `VkDeviceMemory` object is used by multiple |
| 2205 | allocations, calls to `vkBind*Memory()` or `vkMapMemory()` won't happen from multiple threads simultaneously |
| 2206 | (which is illegal in Vulkan). |
| 2207 | |
| 2208 | It is recommended to use function vmaCreateBuffer() instead of this one. |
| 2209 | */ |
| 2210 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaBindBufferMemory( |
| 2211 | VmaAllocator VMA_NOT_NULL allocator, |
| 2212 | VmaAllocation VMA_NOT_NULL allocation, |
| 2213 | VkBuffer VMA_NOT_NULL_NON_DISPATCHABLE buffer); |
| 2214 | |
| 2215 | /** \brief Binds buffer to allocation with additional parameters. |
| 2216 | |
| 2217 | \param allocator |
| 2218 | \param allocation |
| 2219 | \param allocationLocalOffset Additional offset to be added while binding, relative to the beginning of the `allocation`. Normally it should be 0. |
| 2220 | \param buffer |
| 2221 | \param pNext A chain of structures to be attached to `VkBindBufferMemoryInfoKHR` structure used internally. Normally it should be null. |
| 2222 | |
| 2223 | This function is similar to vmaBindBufferMemory(), but it provides additional parameters. |
| 2224 | |
| 2225 | If `pNext` is not null, #VmaAllocator object must have been created with #VMA_ALLOCATOR_CREATE_KHR_BIND_MEMORY2_BIT flag |
| 2226 | or with VmaAllocatorCreateInfo::vulkanApiVersion `>= VK_API_VERSION_1_1`. Otherwise the call fails. |
| 2227 | */ |
| 2228 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaBindBufferMemory2( |
| 2229 | VmaAllocator VMA_NOT_NULL allocator, |
| 2230 | VmaAllocation VMA_NOT_NULL allocation, |
| 2231 | VkDeviceSize allocationLocalOffset, |
| 2232 | VkBuffer VMA_NOT_NULL_NON_DISPATCHABLE buffer, |
| 2233 | const void* VMA_NULLABLE pNext); |
| 2234 | |
| 2235 | /** \brief Binds image to allocation. |
| 2236 | |
| 2237 | Binds specified image to region of memory represented by specified allocation. |
| 2238 | Gets `VkDeviceMemory` handle and offset from the allocation. |
| 2239 | If you want to create an image, allocate memory for it and bind them together separately, |
| 2240 | you should use this function for binding instead of standard `vkBindImageMemory()`, |
| 2241 | because it ensures proper synchronization so that when a `VkDeviceMemory` object is used by multiple |
| 2242 | allocations, calls to `vkBind*Memory()` or `vkMapMemory()` won't happen from multiple threads simultaneously |
| 2243 | (which is illegal in Vulkan). |
| 2244 | |
| 2245 | It is recommended to use function vmaCreateImage() instead of this one. |
| 2246 | */ |
| 2247 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaBindImageMemory( |
| 2248 | VmaAllocator VMA_NOT_NULL allocator, |
| 2249 | VmaAllocation VMA_NOT_NULL allocation, |
| 2250 | VkImage VMA_NOT_NULL_NON_DISPATCHABLE image); |
| 2251 | |
| 2252 | /** \brief Binds image to allocation with additional parameters. |
| 2253 | |
| 2254 | \param allocator |
| 2255 | \param allocation |
| 2256 | \param allocationLocalOffset Additional offset to be added while binding, relative to the beginning of the `allocation`. Normally it should be 0. |
| 2257 | \param image |
| 2258 | \param pNext A chain of structures to be attached to `VkBindImageMemoryInfoKHR` structure used internally. Normally it should be null. |
| 2259 | |
| 2260 | This function is similar to vmaBindImageMemory(), but it provides additional parameters. |
| 2261 | |
| 2262 | If `pNext` is not null, #VmaAllocator object must have been created with #VMA_ALLOCATOR_CREATE_KHR_BIND_MEMORY2_BIT flag |
| 2263 | or with VmaAllocatorCreateInfo::vulkanApiVersion `>= VK_API_VERSION_1_1`. Otherwise the call fails. |
| 2264 | */ |
| 2265 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaBindImageMemory2( |
| 2266 | VmaAllocator VMA_NOT_NULL allocator, |
| 2267 | VmaAllocation VMA_NOT_NULL allocation, |
| 2268 | VkDeviceSize allocationLocalOffset, |
| 2269 | VkImage VMA_NOT_NULL_NON_DISPATCHABLE image, |
| 2270 | const void* VMA_NULLABLE pNext); |
| 2271 | |
| 2272 | /** \brief Creates a new `VkBuffer`, allocates and binds memory for it. |
| 2273 | |
| 2274 | \param allocator |
| 2275 | \param pBufferCreateInfo |
| 2276 | \param pAllocationCreateInfo |
| 2277 | \param[out] pBuffer Buffer that was created. |
| 2278 | \param[out] pAllocation Allocation that was created. |
| 2279 | \param[out] pAllocationInfo Optional. Information about allocated memory. It can be later fetched using function vmaGetAllocationInfo(). |
| 2280 | |
| 2281 | This function automatically: |
| 2282 | |
| 2283 | -# Creates buffer. |
| 2284 | -# Allocates appropriate memory for it. |
| 2285 | -# Binds the buffer with the memory. |
| 2286 | |
| 2287 | If any of these operations fail, buffer and allocation are not created, |
| 2288 | returned value is negative error code, `*pBuffer` and `*pAllocation` are null. |
| 2289 | |
| 2290 | If the function succeeded, you must destroy both buffer and allocation when you |
| 2291 | no longer need them using either convenience function vmaDestroyBuffer() or |
| 2292 | separately, using `vkDestroyBuffer()` and vmaFreeMemory(). |
| 2293 | |
| 2294 | If #VMA_ALLOCATOR_CREATE_KHR_DEDICATED_ALLOCATION_BIT flag was used, |
| 2295 | VK_KHR_dedicated_allocation extension is used internally to query driver whether |
| 2296 | it requires or prefers the new buffer to have dedicated allocation. If yes, |
| 2297 | and if dedicated allocation is possible |
| 2298 | (#VMA_ALLOCATION_CREATE_NEVER_ALLOCATE_BIT is not used), it creates dedicated |
| 2299 | allocation for this buffer, just like when using |
| 2300 | #VMA_ALLOCATION_CREATE_DEDICATED_MEMORY_BIT. |
| 2301 | |
| 2302 | \note This function creates a new `VkBuffer`. Sub-allocation of parts of one large buffer, |
| 2303 | although recommended as a good practice, is out of scope of this library and could be implemented |
| 2304 | by the user as a higher-level logic on top of VMA. |
| 2305 | */ |
| 2306 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaCreateBuffer( |
| 2307 | VmaAllocator VMA_NOT_NULL allocator, |
| 2308 | const VkBufferCreateInfo* VMA_NOT_NULL pBufferCreateInfo, |
| 2309 | const VmaAllocationCreateInfo* VMA_NOT_NULL pAllocationCreateInfo, |
| 2310 | VkBuffer VMA_NULLABLE_NON_DISPATCHABLE* VMA_NOT_NULL pBuffer, |
| 2311 | VmaAllocation VMA_NULLABLE* VMA_NOT_NULL pAllocation, |
| 2312 | VmaAllocationInfo* VMA_NULLABLE pAllocationInfo); |
| 2313 | |
| 2314 | /** \brief Creates a buffer with additional minimum alignment. |
| 2315 | |
| 2316 | Similar to vmaCreateBuffer() but provides additional parameter `minAlignment` which allows to specify custom, |
| 2317 | minimum alignment to be used when placing the buffer inside a larger memory block, which may be needed e.g. |
| 2318 | for interop with OpenGL. |
| 2319 | */ |
| 2320 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaCreateBufferWithAlignment( |
| 2321 | VmaAllocator VMA_NOT_NULL allocator, |
| 2322 | const VkBufferCreateInfo* VMA_NOT_NULL pBufferCreateInfo, |
| 2323 | const VmaAllocationCreateInfo* VMA_NOT_NULL pAllocationCreateInfo, |
| 2324 | VkDeviceSize minAlignment, |
| 2325 | VkBuffer VMA_NULLABLE_NON_DISPATCHABLE* VMA_NOT_NULL pBuffer, |
| 2326 | VmaAllocation VMA_NULLABLE* VMA_NOT_NULL pAllocation, |
| 2327 | VmaAllocationInfo* VMA_NULLABLE pAllocationInfo); |
| 2328 | |
| 2329 | /** \brief Creates a new `VkBuffer`, binds already created memory for it. |
| 2330 | |
| 2331 | \param allocator |
| 2332 | \param allocation Allocation that provides memory to be used for binding new buffer to it. |
| 2333 | \param pBufferCreateInfo |
| 2334 | \param[out] pBuffer Buffer that was created. |
| 2335 | |
| 2336 | This function automatically: |
| 2337 | |
| 2338 | -# Creates buffer. |
| 2339 | -# Binds the buffer with the supplied memory. |
| 2340 | |
| 2341 | If any of these operations fail, buffer is not created, |
| 2342 | returned value is negative error code and `*pBuffer` is null. |
| 2343 | |
| 2344 | If the function succeeded, you must destroy the buffer when you |
| 2345 | no longer need it using `vkDestroyBuffer()`. If you want to also destroy the corresponding |
| 2346 | allocation you can use convenience function vmaDestroyBuffer(). |
| 2347 | */ |
| 2348 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaCreateAliasingBuffer( |
| 2349 | VmaAllocator VMA_NOT_NULL allocator, |
| 2350 | VmaAllocation VMA_NOT_NULL allocation, |
| 2351 | const VkBufferCreateInfo* VMA_NOT_NULL pBufferCreateInfo, |
| 2352 | VkBuffer VMA_NULLABLE_NON_DISPATCHABLE* VMA_NOT_NULL pBuffer); |
| 2353 | |
| 2354 | /** \brief Destroys Vulkan buffer and frees allocated memory. |
| 2355 | |
| 2356 | This is just a convenience function equivalent to: |
| 2357 | |
| 2358 | \code |
| 2359 | vkDestroyBuffer(device, buffer, allocationCallbacks); |
| 2360 | vmaFreeMemory(allocator, allocation); |
| 2361 | \endcode |
| 2362 | |
| 2363 | It it safe to pass null as buffer and/or allocation. |
| 2364 | */ |
| 2365 | VMA_CALL_PRE void VMA_CALL_POST vmaDestroyBuffer( |
| 2366 | VmaAllocator VMA_NOT_NULL allocator, |
| 2367 | VkBuffer VMA_NULLABLE_NON_DISPATCHABLE buffer, |
| 2368 | VmaAllocation VMA_NULLABLE allocation); |
| 2369 | |
| 2370 | /// Function similar to vmaCreateBuffer(). |
| 2371 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaCreateImage( |
| 2372 | VmaAllocator VMA_NOT_NULL allocator, |
| 2373 | const VkImageCreateInfo* VMA_NOT_NULL pImageCreateInfo, |
| 2374 | const VmaAllocationCreateInfo* VMA_NOT_NULL pAllocationCreateInfo, |
| 2375 | VkImage VMA_NULLABLE_NON_DISPATCHABLE* VMA_NOT_NULL pImage, |
| 2376 | VmaAllocation VMA_NULLABLE* VMA_NOT_NULL pAllocation, |
| 2377 | VmaAllocationInfo* VMA_NULLABLE pAllocationInfo); |
| 2378 | |
| 2379 | /// Function similar to vmaCreateAliasingBuffer(). |
| 2380 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaCreateAliasingImage( |
| 2381 | VmaAllocator VMA_NOT_NULL allocator, |
| 2382 | VmaAllocation VMA_NOT_NULL allocation, |
| 2383 | const VkImageCreateInfo* VMA_NOT_NULL pImageCreateInfo, |
| 2384 | VkImage VMA_NULLABLE_NON_DISPATCHABLE* VMA_NOT_NULL pImage); |
| 2385 | |
| 2386 | /** \brief Destroys Vulkan image and frees allocated memory. |
| 2387 | |
| 2388 | This is just a convenience function equivalent to: |
| 2389 | |
| 2390 | \code |
| 2391 | vkDestroyImage(device, image, allocationCallbacks); |
| 2392 | vmaFreeMemory(allocator, allocation); |
| 2393 | \endcode |
| 2394 | |
| 2395 | It it safe to pass null as image and/or allocation. |
| 2396 | */ |
| 2397 | VMA_CALL_PRE void VMA_CALL_POST vmaDestroyImage( |
| 2398 | VmaAllocator VMA_NOT_NULL allocator, |
| 2399 | VkImage VMA_NULLABLE_NON_DISPATCHABLE image, |
| 2400 | VmaAllocation VMA_NULLABLE allocation); |
| 2401 | |
| 2402 | /** @} */ |
| 2403 | |
| 2404 | /** |
| 2405 | \addtogroup group_virtual |
| 2406 | @{ |
| 2407 | */ |
| 2408 | |
| 2409 | /** \brief Creates new #VmaVirtualBlock object. |
| 2410 | |
| 2411 | \param pCreateInfo Parameters for creation. |
| 2412 | \param[out] pVirtualBlock Returned virtual block object or `VMA_NULL` if creation failed. |
| 2413 | */ |
| 2414 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaCreateVirtualBlock( |
| 2415 | const VmaVirtualBlockCreateInfo* VMA_NOT_NULL pCreateInfo, |
| 2416 | VmaVirtualBlock VMA_NULLABLE* VMA_NOT_NULL pVirtualBlock); |
| 2417 | |
| 2418 | /** \brief Destroys #VmaVirtualBlock object. |
| 2419 | |
| 2420 | Please note that you should consciously handle virtual allocations that could remain unfreed in the block. |
| 2421 | You should either free them individually using vmaVirtualFree() or call vmaClearVirtualBlock() |
| 2422 | if you are sure this is what you want. If you do neither, an assert is called. |
| 2423 | |
| 2424 | If you keep pointers to some additional metadata associated with your virtual allocations in their `pUserData`, |
| 2425 | don't forget to free them. |
| 2426 | */ |
| 2427 | VMA_CALL_PRE void VMA_CALL_POST vmaDestroyVirtualBlock( |
| 2428 | VmaVirtualBlock VMA_NULLABLE virtualBlock); |
| 2429 | |
| 2430 | /** \brief Returns true of the #VmaVirtualBlock is empty - contains 0 virtual allocations and has all its space available for new allocations. |
| 2431 | */ |
| 2432 | VMA_CALL_PRE VkBool32 VMA_CALL_POST vmaIsVirtualBlockEmpty( |
| 2433 | VmaVirtualBlock VMA_NOT_NULL virtualBlock); |
| 2434 | |
| 2435 | /** \brief Returns information about a specific virtual allocation within a virtual block, like its size and `pUserData` pointer. |
| 2436 | */ |
| 2437 | VMA_CALL_PRE void VMA_CALL_POST vmaGetVirtualAllocationInfo( |
| 2438 | VmaVirtualBlock VMA_NOT_NULL virtualBlock, |
| 2439 | VmaVirtualAllocation VMA_NOT_NULL_NON_DISPATCHABLE allocation, VmaVirtualAllocationInfo* VMA_NOT_NULL pVirtualAllocInfo); |
| 2440 | |
| 2441 | /** \brief Allocates new virtual allocation inside given #VmaVirtualBlock. |
| 2442 | |
| 2443 | If the allocation fails due to not enough free space available, `VK_ERROR_OUT_OF_DEVICE_MEMORY` is returned |
| 2444 | (despite the function doesn't ever allocate actual GPU memory). |
| 2445 | `pAllocation` is then set to `VK_NULL_HANDLE` and `pOffset`, if not null, it set to `UINT64_MAX`. |
| 2446 | |
| 2447 | \param virtualBlock Virtual block |
| 2448 | \param pCreateInfo Parameters for the allocation |
| 2449 | \param[out] pAllocation Returned handle of the new allocation |
| 2450 | \param[out] pOffset Returned offset of the new allocation. Optional, can be null. |
| 2451 | */ |
| 2452 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaVirtualAllocate( |
| 2453 | VmaVirtualBlock VMA_NOT_NULL virtualBlock, |
| 2454 | const VmaVirtualAllocationCreateInfo* VMA_NOT_NULL pCreateInfo, |
| 2455 | VmaVirtualAllocation VMA_NULLABLE_NON_DISPATCHABLE* VMA_NOT_NULL pAllocation, |
| 2456 | VkDeviceSize* VMA_NULLABLE pOffset); |
| 2457 | |
| 2458 | /** \brief Frees virtual allocation inside given #VmaVirtualBlock. |
| 2459 | |
| 2460 | It is correct to call this function with `allocation == VK_NULL_HANDLE` - it does nothing. |
| 2461 | */ |
| 2462 | VMA_CALL_PRE void VMA_CALL_POST vmaVirtualFree( |
| 2463 | VmaVirtualBlock VMA_NOT_NULL virtualBlock, |
| 2464 | VmaVirtualAllocation VMA_NULLABLE_NON_DISPATCHABLE allocation); |
| 2465 | |
| 2466 | /** \brief Frees all virtual allocations inside given #VmaVirtualBlock. |
| 2467 | |
| 2468 | You must either call this function or free each virtual allocation individually with vmaVirtualFree() |
| 2469 | before destroying a virtual block. Otherwise, an assert is called. |
| 2470 | |
| 2471 | If you keep pointer to some additional metadata associated with your virtual allocation in its `pUserData`, |
| 2472 | don't forget to free it as well. |
| 2473 | */ |
| 2474 | VMA_CALL_PRE void VMA_CALL_POST vmaClearVirtualBlock( |
| 2475 | VmaVirtualBlock VMA_NOT_NULL virtualBlock); |
| 2476 | |
| 2477 | /** \brief Changes custom pointer associated with given virtual allocation. |
| 2478 | */ |
| 2479 | VMA_CALL_PRE void VMA_CALL_POST vmaSetVirtualAllocationUserData( |
| 2480 | VmaVirtualBlock VMA_NOT_NULL virtualBlock, |
| 2481 | VmaVirtualAllocation VMA_NOT_NULL_NON_DISPATCHABLE allocation, |
| 2482 | void* VMA_NULLABLE pUserData); |
| 2483 | |
| 2484 | /** \brief Calculates and returns statistics about virtual allocations and memory usage in given #VmaVirtualBlock. |
| 2485 | |
| 2486 | This function is fast to call. For more detailed statistics, see vmaCalculateVirtualBlockStatistics(). |
| 2487 | */ |
| 2488 | VMA_CALL_PRE void VMA_CALL_POST vmaGetVirtualBlockStatistics( |
| 2489 | VmaVirtualBlock VMA_NOT_NULL virtualBlock, |
| 2490 | VmaStatistics* VMA_NOT_NULL pStats); |
| 2491 | |
| 2492 | /** \brief Calculates and returns detailed statistics about virtual allocations and memory usage in given #VmaVirtualBlock. |
| 2493 | |
| 2494 | This function is slow to call. Use for debugging purposes. |
| 2495 | For less detailed statistics, see vmaGetVirtualBlockStatistics(). |
| 2496 | */ |
| 2497 | VMA_CALL_PRE void VMA_CALL_POST vmaCalculateVirtualBlockStatistics( |
| 2498 | VmaVirtualBlock VMA_NOT_NULL virtualBlock, |
| 2499 | VmaDetailedStatistics* VMA_NOT_NULL pStats); |
| 2500 | |
| 2501 | /** @} */ |
| 2502 | |
| 2503 | #if VMA_STATS_STRING_ENABLED |
| 2504 | /** |
| 2505 | \addtogroup group_stats |
| 2506 | @{ |
| 2507 | */ |
| 2508 | |
| 2509 | /** \brief Builds and returns a null-terminated string in JSON format with information about given #VmaVirtualBlock. |
| 2510 | \param virtualBlock Virtual block. |
| 2511 | \param[out] ppStatsString Returned string. |
| 2512 | \param detailedMap Pass `VK_FALSE` to only obtain statistics as returned by vmaCalculateVirtualBlockStatistics(). Pass `VK_TRUE` to also obtain full list of allocations and free spaces. |
| 2513 | |
| 2514 | Returned string must be freed using vmaFreeVirtualBlockStatsString(). |
| 2515 | */ |
| 2516 | VMA_CALL_PRE void VMA_CALL_POST vmaBuildVirtualBlockStatsString( |
| 2517 | VmaVirtualBlock VMA_NOT_NULL virtualBlock, |
| 2518 | char* VMA_NULLABLE* VMA_NOT_NULL ppStatsString, |
| 2519 | VkBool32 detailedMap); |
| 2520 | |
| 2521 | /// Frees a string returned by vmaBuildVirtualBlockStatsString(). |
| 2522 | VMA_CALL_PRE void VMA_CALL_POST vmaFreeVirtualBlockStatsString( |
| 2523 | VmaVirtualBlock VMA_NOT_NULL virtualBlock, |
| 2524 | char* VMA_NULLABLE pStatsString); |
| 2525 | |
| 2526 | /** \brief Builds and returns statistics as a null-terminated string in JSON format. |
| 2527 | \param allocator |
| 2528 | \param[out] ppStatsString Must be freed using vmaFreeStatsString() function. |
| 2529 | \param detailedMap |
| 2530 | */ |
| 2531 | VMA_CALL_PRE void VMA_CALL_POST vmaBuildStatsString( |
| 2532 | VmaAllocator VMA_NOT_NULL allocator, |
| 2533 | char* VMA_NULLABLE* VMA_NOT_NULL ppStatsString, |
| 2534 | VkBool32 detailedMap); |
| 2535 | |
| 2536 | VMA_CALL_PRE void VMA_CALL_POST vmaFreeStatsString( |
| 2537 | VmaAllocator VMA_NOT_NULL allocator, |
| 2538 | char* VMA_NULLABLE pStatsString); |
| 2539 | |
| 2540 | /** @} */ |
| 2541 | |
| 2542 | #endif // VMA_STATS_STRING_ENABLED |
| 2543 | |
| 2544 | #endif // _VMA_FUNCTION_HEADERS |
| 2545 | |
| 2546 | #ifdef __cplusplus |
| 2547 | } |
| 2548 | #endif |
| 2549 | |
| 2550 | #endif // AMD_VULKAN_MEMORY_ALLOCATOR_H |
| 2551 | |
| 2552 | //////////////////////////////////////////////////////////////////////////////// |
| 2553 | //////////////////////////////////////////////////////////////////////////////// |
| 2554 | // |
| 2555 | // IMPLEMENTATION |
| 2556 | // |
| 2557 | //////////////////////////////////////////////////////////////////////////////// |
| 2558 | //////////////////////////////////////////////////////////////////////////////// |
| 2559 | |
| 2560 | // For Visual Studio IntelliSense. |
| 2561 | #if defined(__cplusplus) && defined(__INTELLISENSE__) |
| 2562 | #define VMA_IMPLEMENTATION |
| 2563 | #endif |
| 2564 | |
| 2565 | #ifdef VMA_IMPLEMENTATION |
| 2566 | #undef VMA_IMPLEMENTATION |
| 2567 | |
| 2568 | #if defined(__GNUC__) && !defined(__clang__) |
| 2569 | #pragma GCC diagnostic push |
| 2570 | #pragma GCC diagnostic ignored "-Wunused-variable" |
| 2571 | #pragma GCC diagnostic ignored "-Wunused-parameter" |
| 2572 | #pragma GCC diagnostic ignored "-Wmissing-field-initializers" |
| 2573 | #pragma GCC diagnostic ignored "-Wparentheses" |
| 2574 | #pragma GCC diagnostic ignored "-Wimplicit-fallthrough" |
| 2575 | #elif defined(__clang__) |
| 2576 | #pragma clang diagnostic push |
| 2577 | #pragma clang diagnostic ignored "-Wunused-variable" |
| 2578 | #pragma clang diagnostic ignored "-Wunused-parameter" |
| 2579 | #pragma clang diagnostic ignored "-Wmissing-field-initializers" |
| 2580 | #pragma clang diagnostic ignored "-Wparentheses" |
| 2581 | #pragma clang diagnostic ignored "-Wimplicit-fallthrough" |
| 2582 | #pragma clang diagnostic ignored "-Wnullability-completeness" |
| 2583 | #endif |
| 2584 | |
| 2585 | #include <cstdint> |
| 2586 | #include <cstdlib> |
| 2587 | #include <cstring> |
| 2588 | #include <utility> |
| 2589 | #include <type_traits> |
| 2590 | |
| 2591 | #ifdef _MSC_VER |
| 2592 | #include <intrin.h> // For functions like __popcnt, _BitScanForward etc. |
| 2593 | #endif |
| 2594 | #if __cplusplus >= 202002L || _MSVC_LANG >= 202002L // C++20 |
| 2595 | #include <bit> // For std::popcount |
| 2596 | #endif |
| 2597 | |
| 2598 | /******************************************************************************* |
| 2599 | CONFIGURATION SECTION |
| 2600 | |
| 2601 | Define some of these macros before each #include of this header or change them |
| 2602 | here if you need other then default behavior depending on your environment. |
| 2603 | */ |
| 2604 | #ifndef _VMA_CONFIGURATION |
| 2605 | |
| 2606 | /* |
| 2607 | Define this macro to 1 to make the library fetch pointers to Vulkan functions |
| 2608 | internally, like: |
| 2609 | |
| 2610 | vulkanFunctions.vkAllocateMemory = &vkAllocateMemory; |
| 2611 | */ |
| 2612 | #if !defined(VMA_STATIC_VULKAN_FUNCTIONS) && !defined(VK_NO_PROTOTYPES) |
| 2613 | #define VMA_STATIC_VULKAN_FUNCTIONS 1 |
| 2614 | #endif |
| 2615 | |
| 2616 | /* |
| 2617 | Define this macro to 1 to make the library fetch pointers to Vulkan functions |
| 2618 | internally, like: |
| 2619 | |
| 2620 | vulkanFunctions.vkAllocateMemory = (PFN_vkAllocateMemory)vkGetDeviceProcAddr(device, "vkAllocateMemory"); |
| 2621 | |
| 2622 | To use this feature in new versions of VMA you now have to pass |
| 2623 | VmaVulkanFunctions::vkGetInstanceProcAddr and vkGetDeviceProcAddr as |
| 2624 | VmaAllocatorCreateInfo::pVulkanFunctions. Other members can be null. |
| 2625 | */ |
| 2626 | #if !defined(VMA_DYNAMIC_VULKAN_FUNCTIONS) |
| 2627 | #define VMA_DYNAMIC_VULKAN_FUNCTIONS 1 |
| 2628 | #endif |
| 2629 | |
| 2630 | #ifndef VMA_USE_STL_SHARED_MUTEX |
| 2631 | // Compiler conforms to C++17. |
| 2632 | #if __cplusplus >= 201703L |
| 2633 | #define VMA_USE_STL_SHARED_MUTEX 1 |
| 2634 | // Visual studio defines __cplusplus properly only when passed additional parameter: /Zc:__cplusplus |
| 2635 | // Otherwise it is always 199711L, despite shared_mutex works since Visual Studio 2015 Update 2. |
| 2636 | #elif defined(_MSC_FULL_VER) && _MSC_FULL_VER >= 190023918 && __cplusplus == 199711L && _MSVC_LANG >= 201703L |
| 2637 | #define VMA_USE_STL_SHARED_MUTEX 1 |
| 2638 | #else |
| 2639 | #define VMA_USE_STL_SHARED_MUTEX 0 |
| 2640 | #endif |
| 2641 | #endif |
| 2642 | |
| 2643 | /* |
| 2644 | Define this macro to include custom header files without having to edit this file directly, e.g.: |
| 2645 | |
| 2646 | // Inside of "my_vma_configuration_user_includes.h": |
| 2647 | |
| 2648 | #include "my_custom_assert.h" // for MY_CUSTOM_ASSERT |
| 2649 | #include "my_custom_min.h" // for my_custom_min |
| 2650 | #include <algorithm> |
| 2651 | #include <mutex> |
| 2652 | |
| 2653 | // Inside a different file, which includes "vk_mem_alloc.h": |
| 2654 | |
| 2655 | #define VMA_CONFIGURATION_USER_INCLUDES_H "my_vma_configuration_user_includes.h" |
| 2656 | #define VMA_ASSERT(expr) MY_CUSTOM_ASSERT(expr) |
| 2657 | #define VMA_MIN(v1, v2) (my_custom_min(v1, v2)) |
| 2658 | #include "vk_mem_alloc.h" |
| 2659 | ... |
| 2660 | |
| 2661 | The following headers are used in this CONFIGURATION section only, so feel free to |
| 2662 | remove them if not needed. |
| 2663 | */ |
| 2664 | #if !defined(VMA_CONFIGURATION_USER_INCLUDES_H) |
| 2665 | #include <cassert> // for assert |
| 2666 | #include <algorithm> // for min, max |
| 2667 | #include <mutex> |
| 2668 | #else |
| 2669 | #include VMA_CONFIGURATION_USER_INCLUDES_H |
| 2670 | #endif |
| 2671 | |
| 2672 | #ifndef VMA_NULL |
| 2673 | // Value used as null pointer. Define it to e.g.: nullptr, NULL, 0, (void*)0. |
| 2674 | #define VMA_NULL nullptr |
| 2675 | #endif |
| 2676 | |
| 2677 | #if defined(__ANDROID_API__) && (__ANDROID_API__ < 16) |
| 2678 | #include <cstdlib> |
| 2679 | static void* vma_aligned_alloc(size_t alignment, size_t size) |
| 2680 | { |
| 2681 | // alignment must be >= sizeof(void*) |
| 2682 | if(alignment < sizeof(void*)) |
| 2683 | { |
| 2684 | alignment = sizeof(void*); |
| 2685 | } |
| 2686 | |
| 2687 | return memalign(alignment, size); |
| 2688 | } |
| 2689 | #elif defined(__APPLE__) || defined(__ANDROID__) || (defined(__linux__) && defined(__GLIBCXX__) && !defined(_GLIBCXX_HAVE_ALIGNED_ALLOC)) |
| 2690 | #include <cstdlib> |
| 2691 | |
| 2692 | #if defined(__APPLE__) |
| 2693 | #include <AvailabilityMacros.h> |
| 2694 | #endif |
| 2695 | |
| 2696 | static void* vma_aligned_alloc(size_t alignment, size_t size) |
| 2697 | { |
| 2698 | // Unfortunately, aligned_alloc causes VMA to crash due to it returning null pointers. (At least under 11.4) |
| 2699 | // Therefore, for now disable this specific exception until a proper solution is found. |
| 2700 | //#if defined(__APPLE__) && (defined(MAC_OS_X_VERSION_10_16) || defined(__IPHONE_14_0)) |
| 2701 | //#if MAC_OS_X_VERSION_MAX_ALLOWED >= MAC_OS_X_VERSION_10_16 || __IPHONE_OS_VERSION_MAX_ALLOWED >= __IPHONE_14_0 |
| 2702 | // // For C++14, usr/include/malloc/_malloc.h declares aligned_alloc()) only |
| 2703 | // // with the MacOSX11.0 SDK in Xcode 12 (which is what adds |
| 2704 | // // MAC_OS_X_VERSION_10_16), even though the function is marked |
| 2705 | // // availabe for 10.15. That is why the preprocessor checks for 10.16 but |
| 2706 | // // the __builtin_available checks for 10.15. |
| 2707 | // // People who use C++17 could call aligned_alloc with the 10.15 SDK already. |
| 2708 | // if (__builtin_available(macOS 10.15, iOS 13, *)) |
| 2709 | // return aligned_alloc(alignment, size); |
| 2710 | //#endif |
| 2711 | //#endif |
| 2712 | |
| 2713 | // alignment must be >= sizeof(void*) |
| 2714 | if(alignment < sizeof(void*)) |
| 2715 | { |
| 2716 | alignment = sizeof(void*); |
| 2717 | } |
| 2718 | |
| 2719 | void *pointer; |
| 2720 | if(posix_memalign(&pointer, alignment, size) == 0) |
| 2721 | return pointer; |
| 2722 | return VMA_NULL; |
| 2723 | } |
| 2724 | #elif defined(_WIN32) |
| 2725 | static void* vma_aligned_alloc(size_t alignment, size_t size) |
| 2726 | { |
| 2727 | return _aligned_malloc(size, alignment); |
| 2728 | } |
| 2729 | #else |
| 2730 | static void* vma_aligned_alloc(size_t alignment, size_t size) |
| 2731 | { |
| 2732 | return aligned_alloc(alignment: alignment, size: size); |
| 2733 | } |
| 2734 | #endif |
| 2735 | |
| 2736 | #if defined(_WIN32) |
| 2737 | static void vma_aligned_free(void* ptr) |
| 2738 | { |
| 2739 | _aligned_free(ptr); |
| 2740 | } |
| 2741 | #else |
| 2742 | static void vma_aligned_free(void* VMA_NULLABLE ptr) |
| 2743 | { |
| 2744 | free(ptr: ptr); |
| 2745 | } |
| 2746 | #endif |
| 2747 | |
| 2748 | // If your compiler is not compatible with C++11 and definition of |
| 2749 | // aligned_alloc() function is missing, uncommeting following line may help: |
| 2750 | |
| 2751 | //#include <malloc.h> |
| 2752 | |
| 2753 | // Normal assert to check for programmer's errors, especially in Debug configuration. |
| 2754 | #ifndef VMA_ASSERT |
| 2755 | #ifdef NDEBUG |
| 2756 | #define VMA_ASSERT(expr) |
| 2757 | #else |
| 2758 | #define VMA_ASSERT(expr) assert(expr) |
| 2759 | #endif |
| 2760 | #endif |
| 2761 | |
| 2762 | // Assert that will be called very often, like inside data structures e.g. operator[]. |
| 2763 | // Making it non-empty can make program slow. |
| 2764 | #ifndef VMA_HEAVY_ASSERT |
| 2765 | #ifdef NDEBUG |
| 2766 | #define VMA_HEAVY_ASSERT(expr) |
| 2767 | #else |
| 2768 | #define VMA_HEAVY_ASSERT(expr) //VMA_ASSERT(expr) |
| 2769 | #endif |
| 2770 | #endif |
| 2771 | |
| 2772 | #ifndef VMA_ALIGN_OF |
| 2773 | #define VMA_ALIGN_OF(type) (__alignof(type)) |
| 2774 | #endif |
| 2775 | |
| 2776 | #ifndef VMA_SYSTEM_ALIGNED_MALLOC |
| 2777 | #define VMA_SYSTEM_ALIGNED_MALLOC(size, alignment) vma_aligned_alloc((alignment), (size)) |
| 2778 | #endif |
| 2779 | |
| 2780 | #ifndef VMA_SYSTEM_ALIGNED_FREE |
| 2781 | // VMA_SYSTEM_FREE is the old name, but might have been defined by the user |
| 2782 | #if defined(VMA_SYSTEM_FREE) |
| 2783 | #define VMA_SYSTEM_ALIGNED_FREE(ptr) VMA_SYSTEM_FREE(ptr) |
| 2784 | #else |
| 2785 | #define VMA_SYSTEM_ALIGNED_FREE(ptr) vma_aligned_free(ptr) |
| 2786 | #endif |
| 2787 | #endif |
| 2788 | |
| 2789 | #ifndef VMA_COUNT_BITS_SET |
| 2790 | // Returns number of bits set to 1 in (v) |
| 2791 | #define VMA_COUNT_BITS_SET(v) VmaCountBitsSet(v) |
| 2792 | #endif |
| 2793 | |
| 2794 | #ifndef VMA_BITSCAN_LSB |
| 2795 | // Scans integer for index of first nonzero value from the Least Significant Bit (LSB). If mask is 0 then returns UINT8_MAX |
| 2796 | #define VMA_BITSCAN_LSB(mask) VmaBitScanLSB(mask) |
| 2797 | #endif |
| 2798 | |
| 2799 | #ifndef VMA_BITSCAN_MSB |
| 2800 | // Scans integer for index of first nonzero value from the Most Significant Bit (MSB). If mask is 0 then returns UINT8_MAX |
| 2801 | #define VMA_BITSCAN_MSB(mask) VmaBitScanMSB(mask) |
| 2802 | #endif |
| 2803 | |
| 2804 | #ifndef VMA_MIN |
| 2805 | #define VMA_MIN(v1, v2) ((std::min)((v1), (v2))) |
| 2806 | #endif |
| 2807 | |
| 2808 | #ifndef VMA_MAX |
| 2809 | #define VMA_MAX(v1, v2) ((std::max)((v1), (v2))) |
| 2810 | #endif |
| 2811 | |
| 2812 | #ifndef VMA_SWAP |
| 2813 | #define VMA_SWAP(v1, v2) std::swap((v1), (v2)) |
| 2814 | #endif |
| 2815 | |
| 2816 | #ifndef VMA_SORT |
| 2817 | #define VMA_SORT(beg, end, cmp) std::sort(beg, end, cmp) |
| 2818 | #endif |
| 2819 | |
| 2820 | #ifndef VMA_DEBUG_LOG |
| 2821 | #define VMA_DEBUG_LOG(format, ...) |
| 2822 | /* |
| 2823 | #define VMA_DEBUG_LOG(format, ...) do { \ |
| 2824 | printf(format, __VA_ARGS__); \ |
| 2825 | printf("\n"); \ |
| 2826 | } while(false) |
| 2827 | */ |
| 2828 | #endif |
| 2829 | |
| 2830 | // Define this macro to 1 to enable functions: vmaBuildStatsString, vmaFreeStatsString. |
| 2831 | #if VMA_STATS_STRING_ENABLED |
| 2832 | static inline void VmaUint32ToStr(char* VMA_NOT_NULL outStr, size_t strLen, uint32_t num) |
| 2833 | { |
| 2834 | snprintf(s: outStr, maxlen: strLen, format: "%u" , static_cast<unsigned int>(num)); |
| 2835 | } |
| 2836 | static inline void VmaUint64ToStr(char* VMA_NOT_NULL outStr, size_t strLen, uint64_t num) |
| 2837 | { |
| 2838 | snprintf(s: outStr, maxlen: strLen, format: "%llu" , static_cast<unsigned long long>(num)); |
| 2839 | } |
| 2840 | static inline void VmaPtrToStr(char* VMA_NOT_NULL outStr, size_t strLen, const void* ptr) |
| 2841 | { |
| 2842 | snprintf(s: outStr, maxlen: strLen, format: "%p" , ptr); |
| 2843 | } |
| 2844 | #endif |
| 2845 | |
| 2846 | #ifndef VMA_MUTEX |
| 2847 | class VmaMutex |
| 2848 | { |
| 2849 | public: |
| 2850 | void Lock() { m_Mutex.lock(); } |
| 2851 | void Unlock() { m_Mutex.unlock(); } |
| 2852 | bool TryLock() { return m_Mutex.try_lock(); } |
| 2853 | private: |
| 2854 | std::mutex m_Mutex; |
| 2855 | }; |
| 2856 | #define VMA_MUTEX VmaMutex |
| 2857 | #endif |
| 2858 | |
| 2859 | // Read-write mutex, where "read" is shared access, "write" is exclusive access. |
| 2860 | #ifndef VMA_RW_MUTEX |
| 2861 | #if VMA_USE_STL_SHARED_MUTEX |
| 2862 | // Use std::shared_mutex from C++17. |
| 2863 | #include <shared_mutex> |
| 2864 | class VmaRWMutex |
| 2865 | { |
| 2866 | public: |
| 2867 | void LockRead() { m_Mutex.lock_shared(); } |
| 2868 | void UnlockRead() { m_Mutex.unlock_shared(); } |
| 2869 | bool TryLockRead() { return m_Mutex.try_lock_shared(); } |
| 2870 | void LockWrite() { m_Mutex.lock(); } |
| 2871 | void UnlockWrite() { m_Mutex.unlock(); } |
| 2872 | bool TryLockWrite() { return m_Mutex.try_lock(); } |
| 2873 | private: |
| 2874 | std::shared_mutex m_Mutex; |
| 2875 | }; |
| 2876 | #define VMA_RW_MUTEX VmaRWMutex |
| 2877 | #elif defined(_WIN32) && defined(WINVER) && WINVER >= 0x0600 && !defined(__MINGW32__) |
| 2878 | // Use SRWLOCK from WinAPI. |
| 2879 | // Minimum supported client = Windows Vista, server = Windows Server 2008. |
| 2880 | class VmaRWMutex |
| 2881 | { |
| 2882 | public: |
| 2883 | VmaRWMutex() { InitializeSRWLock(&m_Lock); } |
| 2884 | void LockRead() { AcquireSRWLockShared(&m_Lock); } |
| 2885 | void UnlockRead() { ReleaseSRWLockShared(&m_Lock); } |
| 2886 | bool TryLockRead() { return TryAcquireSRWLockShared(&m_Lock) != FALSE; } |
| 2887 | void LockWrite() { AcquireSRWLockExclusive(&m_Lock); } |
| 2888 | void UnlockWrite() { ReleaseSRWLockExclusive(&m_Lock); } |
| 2889 | bool TryLockWrite() { return TryAcquireSRWLockExclusive(&m_Lock) != FALSE; } |
| 2890 | private: |
| 2891 | SRWLOCK m_Lock; |
| 2892 | }; |
| 2893 | #define VMA_RW_MUTEX VmaRWMutex |
| 2894 | #else |
| 2895 | // Less efficient fallback: Use normal mutex. |
| 2896 | class VmaRWMutex |
| 2897 | { |
| 2898 | public: |
| 2899 | void LockRead() { m_Mutex.Lock(); } |
| 2900 | void UnlockRead() { m_Mutex.Unlock(); } |
| 2901 | bool TryLockRead() { return m_Mutex.TryLock(); } |
| 2902 | void LockWrite() { m_Mutex.Lock(); } |
| 2903 | void UnlockWrite() { m_Mutex.Unlock(); } |
| 2904 | bool TryLockWrite() { return m_Mutex.TryLock(); } |
| 2905 | private: |
| 2906 | VMA_MUTEX m_Mutex; |
| 2907 | }; |
| 2908 | #define VMA_RW_MUTEX VmaRWMutex |
| 2909 | #endif // #if VMA_USE_STL_SHARED_MUTEX |
| 2910 | #endif // #ifndef VMA_RW_MUTEX |
| 2911 | |
| 2912 | /* |
| 2913 | If providing your own implementation, you need to implement a subset of std::atomic. |
| 2914 | */ |
| 2915 | #ifndef VMA_ATOMIC_UINT32 |
| 2916 | #include <atomic> |
| 2917 | #define VMA_ATOMIC_UINT32 std::atomic<uint32_t> |
| 2918 | #endif |
| 2919 | |
| 2920 | #ifndef VMA_ATOMIC_UINT64 |
| 2921 | #include <atomic> |
| 2922 | #define VMA_ATOMIC_UINT64 std::atomic<uint64_t> |
| 2923 | #endif |
| 2924 | |
| 2925 | #ifndef VMA_DEBUG_ALWAYS_DEDICATED_MEMORY |
| 2926 | /** |
| 2927 | Every allocation will have its own memory block. |
| 2928 | Define to 1 for debugging purposes only. |
| 2929 | */ |
| 2930 | #define VMA_DEBUG_ALWAYS_DEDICATED_MEMORY (0) |
| 2931 | #endif |
| 2932 | |
| 2933 | #ifndef VMA_MIN_ALIGNMENT |
| 2934 | /** |
| 2935 | Minimum alignment of all allocations, in bytes. |
| 2936 | Set to more than 1 for debugging purposes. Must be power of two. |
| 2937 | */ |
| 2938 | #ifdef VMA_DEBUG_ALIGNMENT // Old name |
| 2939 | #define VMA_MIN_ALIGNMENT VMA_DEBUG_ALIGNMENT |
| 2940 | #else |
| 2941 | #define VMA_MIN_ALIGNMENT (1) |
| 2942 | #endif |
| 2943 | #endif |
| 2944 | |
| 2945 | #ifndef VMA_DEBUG_MARGIN |
| 2946 | /** |
| 2947 | Minimum margin after every allocation, in bytes. |
| 2948 | Set nonzero for debugging purposes only. |
| 2949 | */ |
| 2950 | #define VMA_DEBUG_MARGIN (0) |
| 2951 | #endif |
| 2952 | |
| 2953 | #ifndef VMA_DEBUG_INITIALIZE_ALLOCATIONS |
| 2954 | /** |
| 2955 | Define this macro to 1 to automatically fill new allocations and destroyed |
| 2956 | allocations with some bit pattern. |
| 2957 | */ |
| 2958 | #define VMA_DEBUG_INITIALIZE_ALLOCATIONS (0) |
| 2959 | #endif |
| 2960 | |
| 2961 | #ifndef VMA_DEBUG_DETECT_CORRUPTION |
| 2962 | /** |
| 2963 | Define this macro to 1 together with non-zero value of VMA_DEBUG_MARGIN to |
| 2964 | enable writing magic value to the margin after every allocation and |
| 2965 | validating it, so that memory corruptions (out-of-bounds writes) are detected. |
| 2966 | */ |
| 2967 | #define VMA_DEBUG_DETECT_CORRUPTION (0) |
| 2968 | #endif |
| 2969 | |
| 2970 | #ifndef VMA_DEBUG_GLOBAL_MUTEX |
| 2971 | /** |
| 2972 | Set this to 1 for debugging purposes only, to enable single mutex protecting all |
| 2973 | entry calls to the library. Can be useful for debugging multithreading issues. |
| 2974 | */ |
| 2975 | #define VMA_DEBUG_GLOBAL_MUTEX (0) |
| 2976 | #endif |
| 2977 | |
| 2978 | #ifndef VMA_DEBUG_MIN_BUFFER_IMAGE_GRANULARITY |
| 2979 | /** |
| 2980 | Minimum value for VkPhysicalDeviceLimits::bufferImageGranularity. |
| 2981 | Set to more than 1 for debugging purposes only. Must be power of two. |
| 2982 | */ |
| 2983 | #define VMA_DEBUG_MIN_BUFFER_IMAGE_GRANULARITY (1) |
| 2984 | #endif |
| 2985 | |
| 2986 | #ifndef VMA_DEBUG_DONT_EXCEED_MAX_MEMORY_ALLOCATION_COUNT |
| 2987 | /* |
| 2988 | Set this to 1 to make VMA never exceed VkPhysicalDeviceLimits::maxMemoryAllocationCount |
| 2989 | and return error instead of leaving up to Vulkan implementation what to do in such cases. |
| 2990 | */ |
| 2991 | #define VMA_DEBUG_DONT_EXCEED_MAX_MEMORY_ALLOCATION_COUNT (0) |
| 2992 | #endif |
| 2993 | |
| 2994 | #ifndef VMA_SMALL_HEAP_MAX_SIZE |
| 2995 | /// Maximum size of a memory heap in Vulkan to consider it "small". |
| 2996 | #define VMA_SMALL_HEAP_MAX_SIZE (1024ull * 1024 * 1024) |
| 2997 | #endif |
| 2998 | |
| 2999 | #ifndef VMA_DEFAULT_LARGE_HEAP_BLOCK_SIZE |
| 3000 | /// Default size of a block allocated as single VkDeviceMemory from a "large" heap. |
| 3001 | #define VMA_DEFAULT_LARGE_HEAP_BLOCK_SIZE (256ull * 1024 * 1024) |
| 3002 | #endif |
| 3003 | |
| 3004 | /* |
| 3005 | Mapping hysteresis is a logic that launches when vmaMapMemory/vmaUnmapMemory is called |
| 3006 | or a persistently mapped allocation is created and destroyed several times in a row. |
| 3007 | It keeps additional +1 mapping of a device memory block to prevent calling actual |
| 3008 | vkMapMemory/vkUnmapMemory too many times, which may improve performance and help |
| 3009 | tools like RenderDOc. |
| 3010 | */ |
| 3011 | #ifndef VMA_MAPPING_HYSTERESIS_ENABLED |
| 3012 | #define VMA_MAPPING_HYSTERESIS_ENABLED 1 |
| 3013 | #endif |
| 3014 | |
| 3015 | #ifndef VMA_CLASS_NO_COPY |
| 3016 | #define VMA_CLASS_NO_COPY(className) \ |
| 3017 | private: \ |
| 3018 | className(const className&) = delete; \ |
| 3019 | className& operator=(const className&) = delete; |
| 3020 | #endif |
| 3021 | |
| 3022 | #define VMA_VALIDATE(cond) do { if(!(cond)) { \ |
| 3023 | VMA_ASSERT(0 && "Validation failed: " #cond); \ |
| 3024 | return false; \ |
| 3025 | } } while(false) |
| 3026 | |
| 3027 | /******************************************************************************* |
| 3028 | END OF CONFIGURATION |
| 3029 | */ |
| 3030 | #endif // _VMA_CONFIGURATION |
| 3031 | |
| 3032 | |
| 3033 | static const uint8_t VMA_ALLOCATION_FILL_PATTERN_CREATED = 0xDC; |
| 3034 | static const uint8_t VMA_ALLOCATION_FILL_PATTERN_DESTROYED = 0xEF; |
| 3035 | // Decimal 2139416166, float NaN, little-endian binary 66 E6 84 7F. |
| 3036 | static const uint32_t VMA_CORRUPTION_DETECTION_MAGIC_VALUE = 0x7F84E666; |
| 3037 | |
| 3038 | // Copy of some Vulkan definitions so we don't need to check their existence just to handle few constants. |
| 3039 | static const uint32_t VK_MEMORY_PROPERTY_DEVICE_COHERENT_BIT_AMD_COPY = 0x00000040; |
| 3040 | static const uint32_t VK_MEMORY_PROPERTY_DEVICE_UNCACHED_BIT_AMD_COPY = 0x00000080; |
| 3041 | static const uint32_t VK_BUFFER_USAGE_SHADER_DEVICE_ADDRESS_BIT_COPY = 0x00020000; |
| 3042 | static const uint32_t VK_IMAGE_CREATE_DISJOINT_BIT_COPY = 0x00000200; |
| 3043 | static const int32_t VK_IMAGE_TILING_DRM_FORMAT_MODIFIER_EXT_COPY = 1000158000; |
| 3044 | static const uint32_t VMA_ALLOCATION_INTERNAL_STRATEGY_MIN_OFFSET = 0x10000000u; |
| 3045 | static const uint32_t VMA_ALLOCATION_TRY_COUNT = 32; |
| 3046 | static const uint32_t VMA_VENDOR_ID_AMD = 4098; |
| 3047 | |
| 3048 | // This one is tricky. Vulkan specification defines this code as available since |
| 3049 | // Vulkan 1.0, but doesn't actually define it in Vulkan SDK earlier than 1.2.131. |
| 3050 | // See pull request #207. |
| 3051 | #define VK_ERROR_UNKNOWN_COPY ((VkResult)-13) |
| 3052 | |
| 3053 | |
| 3054 | #if VMA_STATS_STRING_ENABLED |
| 3055 | // Correspond to values of enum VmaSuballocationType. |
| 3056 | static const char* VMA_SUBALLOCATION_TYPE_NAMES[] = |
| 3057 | { |
| 3058 | "FREE" , |
| 3059 | "UNKNOWN" , |
| 3060 | "BUFFER" , |
| 3061 | "IMAGE_UNKNOWN" , |
| 3062 | "IMAGE_LINEAR" , |
| 3063 | "IMAGE_OPTIMAL" , |
| 3064 | }; |
| 3065 | #endif |
| 3066 | |
| 3067 | static VkAllocationCallbacks VmaEmptyAllocationCallbacks = |
| 3068 | { VMA_NULL, VMA_NULL, VMA_NULL, VMA_NULL, VMA_NULL, VMA_NULL }; |
| 3069 | |
| 3070 | |
| 3071 | #ifndef _VMA_ENUM_DECLARATIONS |
| 3072 | |
| 3073 | enum VmaSuballocationType |
| 3074 | { |
| 3075 | VMA_SUBALLOCATION_TYPE_FREE = 0, |
| 3076 | VMA_SUBALLOCATION_TYPE_UNKNOWN = 1, |
| 3077 | VMA_SUBALLOCATION_TYPE_BUFFER = 2, |
| 3078 | VMA_SUBALLOCATION_TYPE_IMAGE_UNKNOWN = 3, |
| 3079 | VMA_SUBALLOCATION_TYPE_IMAGE_LINEAR = 4, |
| 3080 | VMA_SUBALLOCATION_TYPE_IMAGE_OPTIMAL = 5, |
| 3081 | VMA_SUBALLOCATION_TYPE_MAX_ENUM = 0x7FFFFFFF |
| 3082 | }; |
| 3083 | |
| 3084 | enum VMA_CACHE_OPERATION |
| 3085 | { |
| 3086 | VMA_CACHE_FLUSH, |
| 3087 | VMA_CACHE_INVALIDATE |
| 3088 | }; |
| 3089 | |
| 3090 | enum class VmaAllocationRequestType |
| 3091 | { |
| 3092 | Normal, |
| 3093 | TLSF, |
| 3094 | // Used by "Linear" algorithm. |
| 3095 | UpperAddress, |
| 3096 | EndOf1st, |
| 3097 | EndOf2nd, |
| 3098 | }; |
| 3099 | |
| 3100 | #endif // _VMA_ENUM_DECLARATIONS |
| 3101 | |
| 3102 | #ifndef _VMA_FORWARD_DECLARATIONS |
| 3103 | // Opaque handle used by allocation algorithms to identify single allocation in any conforming way. |
| 3104 | VK_DEFINE_NON_DISPATCHABLE_HANDLE(VmaAllocHandle); |
| 3105 | |
| 3106 | struct VmaMutexLock; |
| 3107 | struct VmaMutexLockRead; |
| 3108 | struct VmaMutexLockWrite; |
| 3109 | |
| 3110 | template<typename T> |
| 3111 | struct AtomicTransactionalIncrement; |
| 3112 | |
| 3113 | template<typename T> |
| 3114 | struct VmaStlAllocator; |
| 3115 | |
| 3116 | template<typename T, typename AllocatorT> |
| 3117 | class VmaVector; |
| 3118 | |
| 3119 | template<typename T, typename AllocatorT, size_t N> |
| 3120 | class VmaSmallVector; |
| 3121 | |
| 3122 | template<typename T> |
| 3123 | class VmaPoolAllocator; |
| 3124 | |
| 3125 | template<typename T> |
| 3126 | struct VmaListItem; |
| 3127 | |
| 3128 | template<typename T> |
| 3129 | class VmaRawList; |
| 3130 | |
| 3131 | template<typename T, typename AllocatorT> |
| 3132 | class VmaList; |
| 3133 | |
| 3134 | template<typename ItemTypeTraits> |
| 3135 | class VmaIntrusiveLinkedList; |
| 3136 | |
| 3137 | // Unused in this version |
| 3138 | #if 0 |
| 3139 | template<typename T1, typename T2> |
| 3140 | struct VmaPair; |
| 3141 | template<typename FirstT, typename SecondT> |
| 3142 | struct VmaPairFirstLess; |
| 3143 | |
| 3144 | template<typename KeyT, typename ValueT> |
| 3145 | class VmaMap; |
| 3146 | #endif |
| 3147 | |
| 3148 | #if VMA_STATS_STRING_ENABLED |
| 3149 | class VmaStringBuilder; |
| 3150 | class VmaJsonWriter; |
| 3151 | #endif |
| 3152 | |
| 3153 | class VmaDeviceMemoryBlock; |
| 3154 | |
| 3155 | struct VmaDedicatedAllocationListItemTraits; |
| 3156 | class VmaDedicatedAllocationList; |
| 3157 | |
| 3158 | struct VmaSuballocation; |
| 3159 | struct VmaSuballocationOffsetLess; |
| 3160 | struct VmaSuballocationOffsetGreater; |
| 3161 | struct VmaSuballocationItemSizeLess; |
| 3162 | |
| 3163 | typedef VmaList<VmaSuballocation, VmaStlAllocator<VmaSuballocation>> VmaSuballocationList; |
| 3164 | |
| 3165 | struct VmaAllocationRequest; |
| 3166 | |
| 3167 | class VmaBlockMetadata; |
| 3168 | class VmaBlockMetadata_Linear; |
| 3169 | class VmaBlockMetadata_TLSF; |
| 3170 | |
| 3171 | class VmaBlockVector; |
| 3172 | |
| 3173 | struct VmaPoolListItemTraits; |
| 3174 | |
| 3175 | struct VmaCurrentBudgetData; |
| 3176 | |
| 3177 | class VmaAllocationObjectAllocator; |
| 3178 | |
| 3179 | #endif // _VMA_FORWARD_DECLARATIONS |
| 3180 | |
| 3181 | |
| 3182 | #ifndef _VMA_FUNCTIONS |
| 3183 | |
| 3184 | /* |
| 3185 | Returns number of bits set to 1 in (v). |
| 3186 | |
| 3187 | On specific platforms and compilers you can use instrinsics like: |
| 3188 | |
| 3189 | Visual Studio: |
| 3190 | return __popcnt(v); |
| 3191 | GCC, Clang: |
| 3192 | return static_cast<uint32_t>(__builtin_popcount(v)); |
| 3193 | |
| 3194 | Define macro VMA_COUNT_BITS_SET to provide your optimized implementation. |
| 3195 | But you need to check in runtime whether user's CPU supports these, as some old processors don't. |
| 3196 | */ |
| 3197 | static inline uint32_t VmaCountBitsSet(uint32_t v) |
| 3198 | { |
| 3199 | #if __cplusplus >= 202002L || _MSVC_LANG >= 202002L // C++20 |
| 3200 | return std::popcount(v); |
| 3201 | #else |
| 3202 | uint32_t c = v - ((v >> 1) & 0x55555555); |
| 3203 | c = ((c >> 2) & 0x33333333) + (c & 0x33333333); |
| 3204 | c = ((c >> 4) + c) & 0x0F0F0F0F; |
| 3205 | c = ((c >> 8) + c) & 0x00FF00FF; |
| 3206 | c = ((c >> 16) + c) & 0x0000FFFF; |
| 3207 | return c; |
| 3208 | #endif |
| 3209 | } |
| 3210 | |
| 3211 | static inline uint8_t VmaBitScanLSB(uint64_t mask) |
| 3212 | { |
| 3213 | #if defined(_MSC_VER) && defined(_WIN64) |
| 3214 | unsigned long pos; |
| 3215 | if (_BitScanForward64(&pos, mask)) |
| 3216 | return static_cast<uint8_t>(pos); |
| 3217 | return UINT8_MAX; |
| 3218 | #elif defined __GNUC__ || defined __clang__ |
| 3219 | return static_cast<uint8_t>(__builtin_ffsll(mask)) - 1U; |
| 3220 | #else |
| 3221 | uint8_t pos = 0; |
| 3222 | uint64_t bit = 1; |
| 3223 | do |
| 3224 | { |
| 3225 | if (mask & bit) |
| 3226 | return pos; |
| 3227 | bit <<= 1; |
| 3228 | } while (pos++ < 63); |
| 3229 | return UINT8_MAX; |
| 3230 | #endif |
| 3231 | } |
| 3232 | |
| 3233 | static inline uint8_t VmaBitScanLSB(uint32_t mask) |
| 3234 | { |
| 3235 | #ifdef _MSC_VER |
| 3236 | unsigned long pos; |
| 3237 | if (_BitScanForward(&pos, mask)) |
| 3238 | return static_cast<uint8_t>(pos); |
| 3239 | return UINT8_MAX; |
| 3240 | #elif defined __GNUC__ || defined __clang__ |
| 3241 | return static_cast<uint8_t>(__builtin_ffs(mask)) - 1U; |
| 3242 | #else |
| 3243 | uint8_t pos = 0; |
| 3244 | uint32_t bit = 1; |
| 3245 | do |
| 3246 | { |
| 3247 | if (mask & bit) |
| 3248 | return pos; |
| 3249 | bit <<= 1; |
| 3250 | } while (pos++ < 31); |
| 3251 | return UINT8_MAX; |
| 3252 | #endif |
| 3253 | } |
| 3254 | |
| 3255 | static inline uint8_t VmaBitScanMSB(uint64_t mask) |
| 3256 | { |
| 3257 | #if defined(_MSC_VER) && defined(_WIN64) |
| 3258 | unsigned long pos; |
| 3259 | if (_BitScanReverse64(&pos, mask)) |
| 3260 | return static_cast<uint8_t>(pos); |
| 3261 | #elif defined __GNUC__ || defined __clang__ |
| 3262 | if (mask) |
| 3263 | return 63 - static_cast<uint8_t>(__builtin_clzll(mask)); |
| 3264 | #else |
| 3265 | uint8_t pos = 63; |
| 3266 | uint64_t bit = 1ULL << 63; |
| 3267 | do |
| 3268 | { |
| 3269 | if (mask & bit) |
| 3270 | return pos; |
| 3271 | bit >>= 1; |
| 3272 | } while (pos-- > 0); |
| 3273 | #endif |
| 3274 | return UINT8_MAX; |
| 3275 | } |
| 3276 | |
| 3277 | static inline uint8_t VmaBitScanMSB(uint32_t mask) |
| 3278 | { |
| 3279 | #ifdef _MSC_VER |
| 3280 | unsigned long pos; |
| 3281 | if (_BitScanReverse(&pos, mask)) |
| 3282 | return static_cast<uint8_t>(pos); |
| 3283 | #elif defined __GNUC__ || defined __clang__ |
| 3284 | if (mask) |
| 3285 | return 31 - static_cast<uint8_t>(__builtin_clz(mask)); |
| 3286 | #else |
| 3287 | uint8_t pos = 31; |
| 3288 | uint32_t bit = 1UL << 31; |
| 3289 | do |
| 3290 | { |
| 3291 | if (mask & bit) |
| 3292 | return pos; |
| 3293 | bit >>= 1; |
| 3294 | } while (pos-- > 0); |
| 3295 | #endif |
| 3296 | return UINT8_MAX; |
| 3297 | } |
| 3298 | |
| 3299 | /* |
| 3300 | Returns true if given number is a power of two. |
| 3301 | T must be unsigned integer number or signed integer but always nonnegative. |
| 3302 | For 0 returns true. |
| 3303 | */ |
| 3304 | template <typename T> |
| 3305 | inline bool VmaIsPow2(T x) |
| 3306 | { |
| 3307 | return (x & (x - 1)) == 0; |
| 3308 | } |
| 3309 | |
| 3310 | // Aligns given value up to nearest multiply of align value. For example: VmaAlignUp(11, 8) = 16. |
| 3311 | // Use types like uint32_t, uint64_t as T. |
| 3312 | template <typename T> |
| 3313 | static inline T VmaAlignUp(T val, T alignment) |
| 3314 | { |
| 3315 | VMA_HEAVY_ASSERT(VmaIsPow2(alignment)); |
| 3316 | return (val + alignment - 1) & ~(alignment - 1); |
| 3317 | } |
| 3318 | |
| 3319 | // Aligns given value down to nearest multiply of align value. For example: VmaAlignUp(11, 8) = 8. |
| 3320 | // Use types like uint32_t, uint64_t as T. |
| 3321 | template <typename T> |
| 3322 | static inline T VmaAlignDown(T val, T alignment) |
| 3323 | { |
| 3324 | VMA_HEAVY_ASSERT(VmaIsPow2(alignment)); |
| 3325 | return val & ~(alignment - 1); |
| 3326 | } |
| 3327 | |
| 3328 | // Division with mathematical rounding to nearest number. |
| 3329 | template <typename T> |
| 3330 | static inline T VmaRoundDiv(T x, T y) |
| 3331 | { |
| 3332 | return (x + (y / (T)2)) / y; |
| 3333 | } |
| 3334 | |
| 3335 | // Divide by 'y' and round up to nearest integer. |
| 3336 | template <typename T> |
| 3337 | static inline T VmaDivideRoundingUp(T x, T y) |
| 3338 | { |
| 3339 | return (x + y - (T)1) / y; |
| 3340 | } |
| 3341 | |
| 3342 | // Returns smallest power of 2 greater or equal to v. |
| 3343 | static inline uint32_t VmaNextPow2(uint32_t v) |
| 3344 | { |
| 3345 | v--; |
| 3346 | v |= v >> 1; |
| 3347 | v |= v >> 2; |
| 3348 | v |= v >> 4; |
| 3349 | v |= v >> 8; |
| 3350 | v |= v >> 16; |
| 3351 | v++; |
| 3352 | return v; |
| 3353 | } |
| 3354 | |
| 3355 | static inline uint64_t VmaNextPow2(uint64_t v) |
| 3356 | { |
| 3357 | v--; |
| 3358 | v |= v >> 1; |
| 3359 | v |= v >> 2; |
| 3360 | v |= v >> 4; |
| 3361 | v |= v >> 8; |
| 3362 | v |= v >> 16; |
| 3363 | v |= v >> 32; |
| 3364 | v++; |
| 3365 | return v; |
| 3366 | } |
| 3367 | |
| 3368 | // Returns largest power of 2 less or equal to v. |
| 3369 | static inline uint32_t VmaPrevPow2(uint32_t v) |
| 3370 | { |
| 3371 | v |= v >> 1; |
| 3372 | v |= v >> 2; |
| 3373 | v |= v >> 4; |
| 3374 | v |= v >> 8; |
| 3375 | v |= v >> 16; |
| 3376 | v = v ^ (v >> 1); |
| 3377 | return v; |
| 3378 | } |
| 3379 | |
| 3380 | static inline uint64_t VmaPrevPow2(uint64_t v) |
| 3381 | { |
| 3382 | v |= v >> 1; |
| 3383 | v |= v >> 2; |
| 3384 | v |= v >> 4; |
| 3385 | v |= v >> 8; |
| 3386 | v |= v >> 16; |
| 3387 | v |= v >> 32; |
| 3388 | v = v ^ (v >> 1); |
| 3389 | return v; |
| 3390 | } |
| 3391 | |
| 3392 | static inline bool VmaStrIsEmpty(const char* pStr) |
| 3393 | { |
| 3394 | return pStr == VMA_NULL || *pStr == '\0'; |
| 3395 | } |
| 3396 | |
| 3397 | /* |
| 3398 | Returns true if two memory blocks occupy overlapping pages. |
| 3399 | ResourceA must be in less memory offset than ResourceB. |
| 3400 | |
| 3401 | Algorithm is based on "Vulkan 1.0.39 - A Specification (with all registered Vulkan extensions)" |
| 3402 | chapter 11.6 "Resource Memory Association", paragraph "Buffer-Image Granularity". |
| 3403 | */ |
| 3404 | static inline bool VmaBlocksOnSamePage( |
| 3405 | VkDeviceSize resourceAOffset, |
| 3406 | VkDeviceSize resourceASize, |
| 3407 | VkDeviceSize resourceBOffset, |
| 3408 | VkDeviceSize pageSize) |
| 3409 | { |
| 3410 | VMA_ASSERT(resourceAOffset + resourceASize <= resourceBOffset && resourceASize > 0 && pageSize > 0); |
| 3411 | VkDeviceSize resourceAEnd = resourceAOffset + resourceASize - 1; |
| 3412 | VkDeviceSize resourceAEndPage = resourceAEnd & ~(pageSize - 1); |
| 3413 | VkDeviceSize resourceBStart = resourceBOffset; |
| 3414 | VkDeviceSize resourceBStartPage = resourceBStart & ~(pageSize - 1); |
| 3415 | return resourceAEndPage == resourceBStartPage; |
| 3416 | } |
| 3417 | |
| 3418 | /* |
| 3419 | Returns true if given suballocation types could conflict and must respect |
| 3420 | VkPhysicalDeviceLimits::bufferImageGranularity. They conflict if one is buffer |
| 3421 | or linear image and another one is optimal image. If type is unknown, behave |
| 3422 | conservatively. |
| 3423 | */ |
| 3424 | static inline bool VmaIsBufferImageGranularityConflict( |
| 3425 | VmaSuballocationType suballocType1, |
| 3426 | VmaSuballocationType suballocType2) |
| 3427 | { |
| 3428 | if (suballocType1 > suballocType2) |
| 3429 | { |
| 3430 | VMA_SWAP(suballocType1, suballocType2); |
| 3431 | } |
| 3432 | |
| 3433 | switch (suballocType1) |
| 3434 | { |
| 3435 | case VMA_SUBALLOCATION_TYPE_FREE: |
| 3436 | return false; |
| 3437 | case VMA_SUBALLOCATION_TYPE_UNKNOWN: |
| 3438 | return true; |
| 3439 | case VMA_SUBALLOCATION_TYPE_BUFFER: |
| 3440 | return |
| 3441 | suballocType2 == VMA_SUBALLOCATION_TYPE_IMAGE_UNKNOWN || |
| 3442 | suballocType2 == VMA_SUBALLOCATION_TYPE_IMAGE_OPTIMAL; |
| 3443 | case VMA_SUBALLOCATION_TYPE_IMAGE_UNKNOWN: |
| 3444 | return |
| 3445 | suballocType2 == VMA_SUBALLOCATION_TYPE_IMAGE_UNKNOWN || |
| 3446 | suballocType2 == VMA_SUBALLOCATION_TYPE_IMAGE_LINEAR || |
| 3447 | suballocType2 == VMA_SUBALLOCATION_TYPE_IMAGE_OPTIMAL; |
| 3448 | case VMA_SUBALLOCATION_TYPE_IMAGE_LINEAR: |
| 3449 | return |
| 3450 | suballocType2 == VMA_SUBALLOCATION_TYPE_IMAGE_OPTIMAL; |
| 3451 | case VMA_SUBALLOCATION_TYPE_IMAGE_OPTIMAL: |
| 3452 | return false; |
| 3453 | default: |
| 3454 | VMA_ASSERT(0); |
| 3455 | return true; |
| 3456 | } |
| 3457 | } |
| 3458 | |
| 3459 | static void VmaWriteMagicValue(void* pData, VkDeviceSize offset) |
| 3460 | { |
| 3461 | #if VMA_DEBUG_MARGIN > 0 && VMA_DEBUG_DETECT_CORRUPTION |
| 3462 | uint32_t* pDst = (uint32_t*)((char*)pData + offset); |
| 3463 | const size_t numberCount = VMA_DEBUG_MARGIN / sizeof(uint32_t); |
| 3464 | for (size_t i = 0; i < numberCount; ++i, ++pDst) |
| 3465 | { |
| 3466 | *pDst = VMA_CORRUPTION_DETECTION_MAGIC_VALUE; |
| 3467 | } |
| 3468 | #else |
| 3469 | // no-op |
| 3470 | #endif |
| 3471 | } |
| 3472 | |
| 3473 | static bool VmaValidateMagicValue(const void* pData, VkDeviceSize offset) |
| 3474 | { |
| 3475 | #if VMA_DEBUG_MARGIN > 0 && VMA_DEBUG_DETECT_CORRUPTION |
| 3476 | const uint32_t* pSrc = (const uint32_t*)((const char*)pData + offset); |
| 3477 | const size_t numberCount = VMA_DEBUG_MARGIN / sizeof(uint32_t); |
| 3478 | for (size_t i = 0; i < numberCount; ++i, ++pSrc) |
| 3479 | { |
| 3480 | if (*pSrc != VMA_CORRUPTION_DETECTION_MAGIC_VALUE) |
| 3481 | { |
| 3482 | return false; |
| 3483 | } |
| 3484 | } |
| 3485 | #endif |
| 3486 | return true; |
| 3487 | } |
| 3488 | |
| 3489 | /* |
| 3490 | Fills structure with parameters of an example buffer to be used for transfers |
| 3491 | during GPU memory defragmentation. |
| 3492 | */ |
| 3493 | static void VmaFillGpuDefragmentationBufferCreateInfo(VkBufferCreateInfo& outBufCreateInfo) |
| 3494 | { |
| 3495 | memset(s: &outBufCreateInfo, c: 0, n: sizeof(outBufCreateInfo)); |
| 3496 | outBufCreateInfo.sType = VK_STRUCTURE_TYPE_BUFFER_CREATE_INFO; |
| 3497 | outBufCreateInfo.usage = VK_BUFFER_USAGE_TRANSFER_SRC_BIT | VK_BUFFER_USAGE_TRANSFER_DST_BIT; |
| 3498 | outBufCreateInfo.size = (VkDeviceSize)VMA_DEFAULT_LARGE_HEAP_BLOCK_SIZE; // Example size. |
| 3499 | } |
| 3500 | |
| 3501 | |
| 3502 | /* |
| 3503 | Performs binary search and returns iterator to first element that is greater or |
| 3504 | equal to (key), according to comparison (cmp). |
| 3505 | |
| 3506 | Cmp should return true if first argument is less than second argument. |
| 3507 | |
| 3508 | Returned value is the found element, if present in the collection or place where |
| 3509 | new element with value (key) should be inserted. |
| 3510 | */ |
| 3511 | template <typename CmpLess, typename IterT, typename KeyT> |
| 3512 | static IterT VmaBinaryFindFirstNotLess(IterT beg, IterT end, const KeyT& key, const CmpLess& cmp) |
| 3513 | { |
| 3514 | size_t down = 0, up = (end - beg); |
| 3515 | while (down < up) |
| 3516 | { |
| 3517 | const size_t mid = down + (up - down) / 2; // Overflow-safe midpoint calculation |
| 3518 | if (cmp(*(beg + mid), key)) |
| 3519 | { |
| 3520 | down = mid + 1; |
| 3521 | } |
| 3522 | else |
| 3523 | { |
| 3524 | up = mid; |
| 3525 | } |
| 3526 | } |
| 3527 | return beg + down; |
| 3528 | } |
| 3529 | |
| 3530 | template<typename CmpLess, typename IterT, typename KeyT> |
| 3531 | IterT VmaBinaryFindSorted(const IterT& beg, const IterT& end, const KeyT& value, const CmpLess& cmp) |
| 3532 | { |
| 3533 | IterT it = VmaBinaryFindFirstNotLess<CmpLess, IterT, KeyT>( |
| 3534 | beg, end, value, cmp); |
| 3535 | if (it == end || |
| 3536 | (!cmp(*it, value) && !cmp(value, *it))) |
| 3537 | { |
| 3538 | return it; |
| 3539 | } |
| 3540 | return end; |
| 3541 | } |
| 3542 | |
| 3543 | /* |
| 3544 | Returns true if all pointers in the array are not-null and unique. |
| 3545 | Warning! O(n^2) complexity. Use only inside VMA_HEAVY_ASSERT. |
| 3546 | T must be pointer type, e.g. VmaAllocation, VmaPool. |
| 3547 | */ |
| 3548 | template<typename T> |
| 3549 | static bool VmaValidatePointerArray(uint32_t count, const T* arr) |
| 3550 | { |
| 3551 | for (uint32_t i = 0; i < count; ++i) |
| 3552 | { |
| 3553 | const T iPtr = arr[i]; |
| 3554 | if (iPtr == VMA_NULL) |
| 3555 | { |
| 3556 | return false; |
| 3557 | } |
| 3558 | for (uint32_t j = i + 1; j < count; ++j) |
| 3559 | { |
| 3560 | if (iPtr == arr[j]) |
| 3561 | { |
| 3562 | return false; |
| 3563 | } |
| 3564 | } |
| 3565 | } |
| 3566 | return true; |
| 3567 | } |
| 3568 | |
| 3569 | template<typename MainT, typename NewT> |
| 3570 | static inline void VmaPnextChainPushFront(MainT* mainStruct, NewT* newStruct) |
| 3571 | { |
| 3572 | newStruct->pNext = mainStruct->pNext; |
| 3573 | mainStruct->pNext = newStruct; |
| 3574 | } |
| 3575 | |
| 3576 | // This is the main algorithm that guides the selection of a memory type best for an allocation - |
| 3577 | // converts usage to required/preferred/not preferred flags. |
| 3578 | static bool FindMemoryPreferences( |
| 3579 | bool isIntegratedGPU, |
| 3580 | const VmaAllocationCreateInfo& allocCreateInfo, |
| 3581 | VkFlags bufImgUsage, // VkBufferCreateInfo::usage or VkImageCreateInfo::usage. UINT32_MAX if unknown. |
| 3582 | VkMemoryPropertyFlags& outRequiredFlags, |
| 3583 | VkMemoryPropertyFlags& outPreferredFlags, |
| 3584 | VkMemoryPropertyFlags& outNotPreferredFlags) |
| 3585 | { |
| 3586 | outRequiredFlags = allocCreateInfo.requiredFlags; |
| 3587 | outPreferredFlags = allocCreateInfo.preferredFlags; |
| 3588 | outNotPreferredFlags = 0; |
| 3589 | |
| 3590 | switch(allocCreateInfo.usage) |
| 3591 | { |
| 3592 | case VMA_MEMORY_USAGE_UNKNOWN: |
| 3593 | break; |
| 3594 | case VMA_MEMORY_USAGE_GPU_ONLY: |
| 3595 | if(!isIntegratedGPU || (outPreferredFlags & VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT) == 0) |
| 3596 | { |
| 3597 | outPreferredFlags |= VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT; |
| 3598 | } |
| 3599 | break; |
| 3600 | case VMA_MEMORY_USAGE_CPU_ONLY: |
| 3601 | outRequiredFlags |= VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT | VK_MEMORY_PROPERTY_HOST_COHERENT_BIT; |
| 3602 | break; |
| 3603 | case VMA_MEMORY_USAGE_CPU_TO_GPU: |
| 3604 | outRequiredFlags |= VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT; |
| 3605 | if(!isIntegratedGPU || (outPreferredFlags & VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT) == 0) |
| 3606 | { |
| 3607 | outPreferredFlags |= VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT; |
| 3608 | } |
| 3609 | break; |
| 3610 | case VMA_MEMORY_USAGE_GPU_TO_CPU: |
| 3611 | outRequiredFlags |= VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT; |
| 3612 | outPreferredFlags |= VK_MEMORY_PROPERTY_HOST_CACHED_BIT; |
| 3613 | break; |
| 3614 | case VMA_MEMORY_USAGE_CPU_COPY: |
| 3615 | outNotPreferredFlags |= VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT; |
| 3616 | break; |
| 3617 | case VMA_MEMORY_USAGE_GPU_LAZILY_ALLOCATED: |
| 3618 | outRequiredFlags |= VK_MEMORY_PROPERTY_LAZILY_ALLOCATED_BIT; |
| 3619 | break; |
| 3620 | case VMA_MEMORY_USAGE_AUTO: |
| 3621 | case VMA_MEMORY_USAGE_AUTO_PREFER_DEVICE: |
| 3622 | case VMA_MEMORY_USAGE_AUTO_PREFER_HOST: |
| 3623 | { |
| 3624 | if(bufImgUsage == UINT32_MAX) |
| 3625 | { |
| 3626 | VMA_ASSERT(0 && "VMA_MEMORY_USAGE_AUTO* values can only be used with functions like vmaCreateBuffer, vmaCreateImage so that the details of the created resource are known." ); |
| 3627 | return false; |
| 3628 | } |
| 3629 | // This relies on values of VK_IMAGE_USAGE_TRANSFER* being the same VK_BUFFER_IMAGE_TRANSFER*. |
| 3630 | const bool deviceAccess = (bufImgUsage & ~(VK_BUFFER_USAGE_TRANSFER_DST_BIT | VK_BUFFER_USAGE_TRANSFER_SRC_BIT)) != 0; |
| 3631 | const bool hostAccessSequentialWrite = (allocCreateInfo.flags & VMA_ALLOCATION_CREATE_HOST_ACCESS_SEQUENTIAL_WRITE_BIT) != 0; |
| 3632 | const bool hostAccessRandom = (allocCreateInfo.flags & VMA_ALLOCATION_CREATE_HOST_ACCESS_RANDOM_BIT) != 0; |
| 3633 | const bool hostAccessAllowTransferInstead = (allocCreateInfo.flags & VMA_ALLOCATION_CREATE_HOST_ACCESS_ALLOW_TRANSFER_INSTEAD_BIT) != 0; |
| 3634 | const bool preferDevice = allocCreateInfo.usage == VMA_MEMORY_USAGE_AUTO_PREFER_DEVICE; |
| 3635 | const bool preferHost = allocCreateInfo.usage == VMA_MEMORY_USAGE_AUTO_PREFER_HOST; |
| 3636 | |
| 3637 | // CPU random access - e.g. a buffer written to or transferred from GPU to read back on CPU. |
| 3638 | if(hostAccessRandom) |
| 3639 | { |
| 3640 | if(!isIntegratedGPU && deviceAccess && hostAccessAllowTransferInstead && !preferHost) |
| 3641 | { |
| 3642 | // Nice if it will end up in HOST_VISIBLE, but more importantly prefer DEVICE_LOCAL. |
| 3643 | // Omitting HOST_VISIBLE here is intentional. |
| 3644 | // In case there is DEVICE_LOCAL | HOST_VISIBLE | HOST_CACHED, it will pick that one. |
| 3645 | // Otherwise, this will give same weight to DEVICE_LOCAL as HOST_VISIBLE | HOST_CACHED and select the former if occurs first on the list. |
| 3646 | outPreferredFlags |= VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT | VK_MEMORY_PROPERTY_HOST_CACHED_BIT; |
| 3647 | } |
| 3648 | else |
| 3649 | { |
| 3650 | // Always CPU memory, cached. |
| 3651 | outRequiredFlags |= VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT | VK_MEMORY_PROPERTY_HOST_CACHED_BIT; |
| 3652 | } |
| 3653 | } |
| 3654 | // CPU sequential write - may be CPU or host-visible GPU memory, uncached and write-combined. |
| 3655 | else if(hostAccessSequentialWrite) |
| 3656 | { |
| 3657 | // Want uncached and write-combined. |
| 3658 | outNotPreferredFlags |= VK_MEMORY_PROPERTY_HOST_CACHED_BIT; |
| 3659 | |
| 3660 | if(!isIntegratedGPU && deviceAccess && hostAccessAllowTransferInstead && !preferHost) |
| 3661 | { |
| 3662 | outPreferredFlags |= VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT | VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT; |
| 3663 | } |
| 3664 | else |
| 3665 | { |
| 3666 | outRequiredFlags |= VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT; |
| 3667 | // Direct GPU access, CPU sequential write (e.g. a dynamic uniform buffer updated every frame) |
| 3668 | if(deviceAccess) |
| 3669 | { |
| 3670 | // Could go to CPU memory or GPU BAR/unified. Up to the user to decide. If no preference, choose GPU memory. |
| 3671 | if(preferHost) |
| 3672 | outNotPreferredFlags |= VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT; |
| 3673 | else |
| 3674 | outPreferredFlags |= VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT; |
| 3675 | } |
| 3676 | // GPU no direct access, CPU sequential write (e.g. an upload buffer to be transferred to the GPU) |
| 3677 | else |
| 3678 | { |
| 3679 | // Could go to CPU memory or GPU BAR/unified. Up to the user to decide. If no preference, choose CPU memory. |
| 3680 | if(preferDevice) |
| 3681 | outPreferredFlags |= VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT; |
| 3682 | else |
| 3683 | outNotPreferredFlags |= VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT; |
| 3684 | } |
| 3685 | } |
| 3686 | } |
| 3687 | // No CPU access |
| 3688 | else |
| 3689 | { |
| 3690 | // GPU access, no CPU access (e.g. a color attachment image) - prefer GPU memory |
| 3691 | if(deviceAccess) |
| 3692 | { |
| 3693 | // ...unless there is a clear preference from the user not to do so. |
| 3694 | if(preferHost) |
| 3695 | outNotPreferredFlags |= VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT; |
| 3696 | else |
| 3697 | outPreferredFlags |= VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT; |
| 3698 | } |
| 3699 | // No direct GPU access, no CPU access, just transfers. |
| 3700 | // It may be staging copy intended for e.g. preserving image for next frame (then better GPU memory) or |
| 3701 | // a "swap file" copy to free some GPU memory (then better CPU memory). |
| 3702 | // Up to the user to decide. If no preferece, assume the former and choose GPU memory. |
| 3703 | if(preferHost) |
| 3704 | outNotPreferredFlags |= VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT; |
| 3705 | else |
| 3706 | outPreferredFlags |= VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT; |
| 3707 | } |
| 3708 | break; |
| 3709 | } |
| 3710 | default: |
| 3711 | VMA_ASSERT(0); |
| 3712 | } |
| 3713 | |
| 3714 | // Avoid DEVICE_COHERENT unless explicitly requested. |
| 3715 | if(((allocCreateInfo.requiredFlags | allocCreateInfo.preferredFlags) & |
| 3716 | (VK_MEMORY_PROPERTY_DEVICE_COHERENT_BIT_AMD_COPY | VK_MEMORY_PROPERTY_DEVICE_UNCACHED_BIT_AMD_COPY)) == 0) |
| 3717 | { |
| 3718 | outNotPreferredFlags |= VK_MEMORY_PROPERTY_DEVICE_UNCACHED_BIT_AMD_COPY; |
| 3719 | } |
| 3720 | |
| 3721 | return true; |
| 3722 | } |
| 3723 | |
| 3724 | //////////////////////////////////////////////////////////////////////////////// |
| 3725 | // Memory allocation |
| 3726 | |
| 3727 | static void* VmaMalloc(const VkAllocationCallbacks* pAllocationCallbacks, size_t size, size_t alignment) |
| 3728 | { |
| 3729 | void* result = VMA_NULL; |
| 3730 | if ((pAllocationCallbacks != VMA_NULL) && |
| 3731 | (pAllocationCallbacks->pfnAllocation != VMA_NULL)) |
| 3732 | { |
| 3733 | result = (*pAllocationCallbacks->pfnAllocation)( |
| 3734 | pAllocationCallbacks->pUserData, |
| 3735 | size, |
| 3736 | alignment, |
| 3737 | VK_SYSTEM_ALLOCATION_SCOPE_OBJECT); |
| 3738 | } |
| 3739 | else |
| 3740 | { |
| 3741 | result = VMA_SYSTEM_ALIGNED_MALLOC(size, alignment); |
| 3742 | } |
| 3743 | VMA_ASSERT(result != VMA_NULL && "CPU memory allocation failed." ); |
| 3744 | return result; |
| 3745 | } |
| 3746 | |
| 3747 | static void VmaFree(const VkAllocationCallbacks* pAllocationCallbacks, void* ptr) |
| 3748 | { |
| 3749 | if ((pAllocationCallbacks != VMA_NULL) && |
| 3750 | (pAllocationCallbacks->pfnFree != VMA_NULL)) |
| 3751 | { |
| 3752 | (*pAllocationCallbacks->pfnFree)(pAllocationCallbacks->pUserData, ptr); |
| 3753 | } |
| 3754 | else |
| 3755 | { |
| 3756 | VMA_SYSTEM_ALIGNED_FREE(ptr); |
| 3757 | } |
| 3758 | } |
| 3759 | |
| 3760 | template<typename T> |
| 3761 | static T* VmaAllocate(const VkAllocationCallbacks* pAllocationCallbacks) |
| 3762 | { |
| 3763 | return (T*)VmaMalloc(pAllocationCallbacks, size: sizeof(T), VMA_ALIGN_OF(T)); |
| 3764 | } |
| 3765 | |
| 3766 | template<typename T> |
| 3767 | static T* VmaAllocateArray(const VkAllocationCallbacks* pAllocationCallbacks, size_t count) |
| 3768 | { |
| 3769 | return (T*)VmaMalloc(pAllocationCallbacks, size: sizeof(T) * count, VMA_ALIGN_OF(T)); |
| 3770 | } |
| 3771 | |
| 3772 | #define vma_new(allocator, type) new(VmaAllocate<type>(allocator))(type) |
| 3773 | |
| 3774 | #define vma_new_array(allocator, type, count) new(VmaAllocateArray<type>((allocator), (count)))(type) |
| 3775 | |
| 3776 | template<typename T> |
| 3777 | static void vma_delete(const VkAllocationCallbacks* pAllocationCallbacks, T* ptr) |
| 3778 | { |
| 3779 | ptr->~T(); |
| 3780 | VmaFree(pAllocationCallbacks, ptr); |
| 3781 | } |
| 3782 | |
| 3783 | template<typename T> |
| 3784 | static void vma_delete_array(const VkAllocationCallbacks* pAllocationCallbacks, T* ptr, size_t count) |
| 3785 | { |
| 3786 | if (ptr != VMA_NULL) |
| 3787 | { |
| 3788 | for (size_t i = count; i--; ) |
| 3789 | { |
| 3790 | ptr[i].~T(); |
| 3791 | } |
| 3792 | VmaFree(pAllocationCallbacks, ptr); |
| 3793 | } |
| 3794 | } |
| 3795 | |
| 3796 | static char* VmaCreateStringCopy(const VkAllocationCallbacks* allocs, const char* srcStr) |
| 3797 | { |
| 3798 | if (srcStr != VMA_NULL) |
| 3799 | { |
| 3800 | const size_t len = strlen(s: srcStr); |
| 3801 | char* const result = vma_new_array(allocs, char, len + 1); |
| 3802 | memcpy(dest: result, src: srcStr, n: len + 1); |
| 3803 | return result; |
| 3804 | } |
| 3805 | return VMA_NULL; |
| 3806 | } |
| 3807 | |
| 3808 | #if VMA_STATS_STRING_ENABLED |
| 3809 | static char* VmaCreateStringCopy(const VkAllocationCallbacks* allocs, const char* srcStr, size_t strLen) |
| 3810 | { |
| 3811 | if (srcStr != VMA_NULL) |
| 3812 | { |
| 3813 | char* const result = vma_new_array(allocs, char, strLen + 1); |
| 3814 | memcpy(dest: result, src: srcStr, n: strLen); |
| 3815 | result[strLen] = '\0'; |
| 3816 | return result; |
| 3817 | } |
| 3818 | return VMA_NULL; |
| 3819 | } |
| 3820 | #endif // VMA_STATS_STRING_ENABLED |
| 3821 | |
| 3822 | static void VmaFreeString(const VkAllocationCallbacks* allocs, char* str) |
| 3823 | { |
| 3824 | if (str != VMA_NULL) |
| 3825 | { |
| 3826 | const size_t len = strlen(s: str); |
| 3827 | vma_delete_array(pAllocationCallbacks: allocs, ptr: str, count: len + 1); |
| 3828 | } |
| 3829 | } |
| 3830 | |
| 3831 | template<typename CmpLess, typename VectorT> |
| 3832 | size_t VmaVectorInsertSorted(VectorT& vector, const typename VectorT::value_type& value) |
| 3833 | { |
| 3834 | const size_t indexToInsert = VmaBinaryFindFirstNotLess( |
| 3835 | vector.data(), |
| 3836 | vector.data() + vector.size(), |
| 3837 | value, |
| 3838 | CmpLess()) - vector.data(); |
| 3839 | VmaVectorInsert(vector, indexToInsert, value); |
| 3840 | return indexToInsert; |
| 3841 | } |
| 3842 | |
| 3843 | template<typename CmpLess, typename VectorT> |
| 3844 | bool VmaVectorRemoveSorted(VectorT& vector, const typename VectorT::value_type& value) |
| 3845 | { |
| 3846 | CmpLess comparator; |
| 3847 | typename VectorT::iterator it = VmaBinaryFindFirstNotLess( |
| 3848 | vector.begin(), |
| 3849 | vector.end(), |
| 3850 | value, |
| 3851 | comparator); |
| 3852 | if ((it != vector.end()) && !comparator(*it, value) && !comparator(value, *it)) |
| 3853 | { |
| 3854 | size_t indexToRemove = it - vector.begin(); |
| 3855 | VmaVectorRemove(vector, indexToRemove); |
| 3856 | return true; |
| 3857 | } |
| 3858 | return false; |
| 3859 | } |
| 3860 | #endif // _VMA_FUNCTIONS |
| 3861 | |
| 3862 | #ifndef _VMA_STATISTICS_FUNCTIONS |
| 3863 | |
| 3864 | static void VmaClearStatistics(VmaStatistics& outStats) |
| 3865 | { |
| 3866 | outStats.blockCount = 0; |
| 3867 | outStats.allocationCount = 0; |
| 3868 | outStats.blockBytes = 0; |
| 3869 | outStats.allocationBytes = 0; |
| 3870 | } |
| 3871 | |
| 3872 | static void VmaAddStatistics(VmaStatistics& inoutStats, const VmaStatistics& src) |
| 3873 | { |
| 3874 | inoutStats.blockCount += src.blockCount; |
| 3875 | inoutStats.allocationCount += src.allocationCount; |
| 3876 | inoutStats.blockBytes += src.blockBytes; |
| 3877 | inoutStats.allocationBytes += src.allocationBytes; |
| 3878 | } |
| 3879 | |
| 3880 | static void VmaClearDetailedStatistics(VmaDetailedStatistics& outStats) |
| 3881 | { |
| 3882 | VmaClearStatistics(outStats&: outStats.statistics); |
| 3883 | outStats.unusedRangeCount = 0; |
| 3884 | outStats.allocationSizeMin = VK_WHOLE_SIZE; |
| 3885 | outStats.allocationSizeMax = 0; |
| 3886 | outStats.unusedRangeSizeMin = VK_WHOLE_SIZE; |
| 3887 | outStats.unusedRangeSizeMax = 0; |
| 3888 | } |
| 3889 | |
| 3890 | static void VmaAddDetailedStatisticsAllocation(VmaDetailedStatistics& inoutStats, VkDeviceSize size) |
| 3891 | { |
| 3892 | inoutStats.statistics.allocationCount++; |
| 3893 | inoutStats.statistics.allocationBytes += size; |
| 3894 | inoutStats.allocationSizeMin = VMA_MIN(inoutStats.allocationSizeMin, size); |
| 3895 | inoutStats.allocationSizeMax = VMA_MAX(inoutStats.allocationSizeMax, size); |
| 3896 | } |
| 3897 | |
| 3898 | static void VmaAddDetailedStatisticsUnusedRange(VmaDetailedStatistics& inoutStats, VkDeviceSize size) |
| 3899 | { |
| 3900 | inoutStats.unusedRangeCount++; |
| 3901 | inoutStats.unusedRangeSizeMin = VMA_MIN(inoutStats.unusedRangeSizeMin, size); |
| 3902 | inoutStats.unusedRangeSizeMax = VMA_MAX(inoutStats.unusedRangeSizeMax, size); |
| 3903 | } |
| 3904 | |
| 3905 | static void VmaAddDetailedStatistics(VmaDetailedStatistics& inoutStats, const VmaDetailedStatistics& src) |
| 3906 | { |
| 3907 | VmaAddStatistics(inoutStats&: inoutStats.statistics, src: src.statistics); |
| 3908 | inoutStats.unusedRangeCount += src.unusedRangeCount; |
| 3909 | inoutStats.allocationSizeMin = VMA_MIN(inoutStats.allocationSizeMin, src.allocationSizeMin); |
| 3910 | inoutStats.allocationSizeMax = VMA_MAX(inoutStats.allocationSizeMax, src.allocationSizeMax); |
| 3911 | inoutStats.unusedRangeSizeMin = VMA_MIN(inoutStats.unusedRangeSizeMin, src.unusedRangeSizeMin); |
| 3912 | inoutStats.unusedRangeSizeMax = VMA_MAX(inoutStats.unusedRangeSizeMax, src.unusedRangeSizeMax); |
| 3913 | } |
| 3914 | |
| 3915 | #endif // _VMA_STATISTICS_FUNCTIONS |
| 3916 | |
| 3917 | #ifndef _VMA_MUTEX_LOCK |
| 3918 | // Helper RAII class to lock a mutex in constructor and unlock it in destructor (at the end of scope). |
| 3919 | struct VmaMutexLock |
| 3920 | { |
| 3921 | VMA_CLASS_NO_COPY(VmaMutexLock) |
| 3922 | public: |
| 3923 | VmaMutexLock(VMA_MUTEX& mutex, bool useMutex = true) : |
| 3924 | m_pMutex(useMutex ? &mutex : VMA_NULL) |
| 3925 | { |
| 3926 | if (m_pMutex) { m_pMutex->Lock(); } |
| 3927 | } |
| 3928 | ~VmaMutexLock() { if (m_pMutex) { m_pMutex->Unlock(); } } |
| 3929 | |
| 3930 | private: |
| 3931 | VMA_MUTEX* m_pMutex; |
| 3932 | }; |
| 3933 | |
| 3934 | // Helper RAII class to lock a RW mutex in constructor and unlock it in destructor (at the end of scope), for reading. |
| 3935 | struct VmaMutexLockRead |
| 3936 | { |
| 3937 | VMA_CLASS_NO_COPY(VmaMutexLockRead) |
| 3938 | public: |
| 3939 | VmaMutexLockRead(VMA_RW_MUTEX& mutex, bool useMutex) : |
| 3940 | m_pMutex(useMutex ? &mutex : VMA_NULL) |
| 3941 | { |
| 3942 | if (m_pMutex) { m_pMutex->LockRead(); } |
| 3943 | } |
| 3944 | ~VmaMutexLockRead() { if (m_pMutex) { m_pMutex->UnlockRead(); } } |
| 3945 | |
| 3946 | private: |
| 3947 | VMA_RW_MUTEX* m_pMutex; |
| 3948 | }; |
| 3949 | |
| 3950 | // Helper RAII class to lock a RW mutex in constructor and unlock it in destructor (at the end of scope), for writing. |
| 3951 | struct VmaMutexLockWrite |
| 3952 | { |
| 3953 | VMA_CLASS_NO_COPY(VmaMutexLockWrite) |
| 3954 | public: |
| 3955 | VmaMutexLockWrite(VMA_RW_MUTEX& mutex, bool useMutex) |
| 3956 | : m_pMutex(useMutex ? &mutex : VMA_NULL) |
| 3957 | { |
| 3958 | if (m_pMutex) { m_pMutex->LockWrite(); } |
| 3959 | } |
| 3960 | ~VmaMutexLockWrite() { if (m_pMutex) { m_pMutex->UnlockWrite(); } } |
| 3961 | |
| 3962 | private: |
| 3963 | VMA_RW_MUTEX* m_pMutex; |
| 3964 | }; |
| 3965 | |
| 3966 | #if VMA_DEBUG_GLOBAL_MUTEX |
| 3967 | static VMA_MUTEX gDebugGlobalMutex; |
| 3968 | #define VMA_DEBUG_GLOBAL_MUTEX_LOCK VmaMutexLock debugGlobalMutexLock(gDebugGlobalMutex, true); |
| 3969 | #else |
| 3970 | #define VMA_DEBUG_GLOBAL_MUTEX_LOCK |
| 3971 | #endif |
| 3972 | #endif // _VMA_MUTEX_LOCK |
| 3973 | |
| 3974 | #ifndef _VMA_ATOMIC_TRANSACTIONAL_INCREMENT |
| 3975 | // An object that increments given atomic but decrements it back in the destructor unless Commit() is called. |
| 3976 | template<typename T> |
| 3977 | struct AtomicTransactionalIncrement |
| 3978 | { |
| 3979 | public: |
| 3980 | typedef std::atomic<T> AtomicT; |
| 3981 | |
| 3982 | ~AtomicTransactionalIncrement() |
| 3983 | { |
| 3984 | if(m_Atomic) |
| 3985 | --(*m_Atomic); |
| 3986 | } |
| 3987 | |
| 3988 | void Commit() { m_Atomic = nullptr; } |
| 3989 | T Increment(AtomicT* atomic) |
| 3990 | { |
| 3991 | m_Atomic = atomic; |
| 3992 | return m_Atomic->fetch_add(1); |
| 3993 | } |
| 3994 | |
| 3995 | private: |
| 3996 | AtomicT* m_Atomic = nullptr; |
| 3997 | }; |
| 3998 | #endif // _VMA_ATOMIC_TRANSACTIONAL_INCREMENT |
| 3999 | |
| 4000 | #ifndef _VMA_STL_ALLOCATOR |
| 4001 | // STL-compatible allocator. |
| 4002 | template<typename T> |
| 4003 | struct VmaStlAllocator |
| 4004 | { |
| 4005 | const VkAllocationCallbacks* const m_pCallbacks; |
| 4006 | typedef T value_type; |
| 4007 | |
| 4008 | VmaStlAllocator(const VkAllocationCallbacks* pCallbacks) : m_pCallbacks(pCallbacks) {} |
| 4009 | template<typename U> |
| 4010 | VmaStlAllocator(const VmaStlAllocator<U>& src) : m_pCallbacks(src.m_pCallbacks) {} |
| 4011 | VmaStlAllocator(const VmaStlAllocator&) = default; |
| 4012 | VmaStlAllocator& operator=(const VmaStlAllocator&) = delete; |
| 4013 | |
| 4014 | T* allocate(size_t n) { return VmaAllocateArray<T>(m_pCallbacks, n); } |
| 4015 | void deallocate(T* p, size_t n) { VmaFree(m_pCallbacks, p); } |
| 4016 | |
| 4017 | template<typename U> |
| 4018 | bool operator==(const VmaStlAllocator<U>& rhs) const |
| 4019 | { |
| 4020 | return m_pCallbacks == rhs.m_pCallbacks; |
| 4021 | } |
| 4022 | template<typename U> |
| 4023 | bool operator!=(const VmaStlAllocator<U>& rhs) const |
| 4024 | { |
| 4025 | return m_pCallbacks != rhs.m_pCallbacks; |
| 4026 | } |
| 4027 | }; |
| 4028 | #endif // _VMA_STL_ALLOCATOR |
| 4029 | |
| 4030 | #ifndef _VMA_VECTOR |
| 4031 | /* Class with interface compatible with subset of std::vector. |
| 4032 | T must be POD because constructors and destructors are not called and memcpy is |
| 4033 | used for these objects. */ |
| 4034 | template<typename T, typename AllocatorT> |
| 4035 | class VmaVector |
| 4036 | { |
| 4037 | public: |
| 4038 | typedef T value_type; |
| 4039 | typedef T* iterator; |
| 4040 | typedef const T* const_iterator; |
| 4041 | |
| 4042 | VmaVector(const AllocatorT& allocator); |
| 4043 | VmaVector(size_t count, const AllocatorT& allocator); |
| 4044 | // This version of the constructor is here for compatibility with pre-C++14 std::vector. |
| 4045 | // value is unused. |
| 4046 | VmaVector(size_t count, const T& value, const AllocatorT& allocator) : VmaVector(count, allocator) {} |
| 4047 | VmaVector(const VmaVector<T, AllocatorT>& src); |
| 4048 | VmaVector& operator=(const VmaVector& rhs); |
| 4049 | ~VmaVector() { VmaFree(m_Allocator.m_pCallbacks, m_pArray); } |
| 4050 | |
| 4051 | bool empty() const { return m_Count == 0; } |
| 4052 | size_t size() const { return m_Count; } |
| 4053 | T* data() { return m_pArray; } |
| 4054 | T& front() { VMA_HEAVY_ASSERT(m_Count > 0); return m_pArray[0]; } |
| 4055 | T& back() { VMA_HEAVY_ASSERT(m_Count > 0); return m_pArray[m_Count - 1]; } |
| 4056 | const T* data() const { return m_pArray; } |
| 4057 | const T& front() const { VMA_HEAVY_ASSERT(m_Count > 0); return m_pArray[0]; } |
| 4058 | const T& back() const { VMA_HEAVY_ASSERT(m_Count > 0); return m_pArray[m_Count - 1]; } |
| 4059 | |
| 4060 | iterator begin() { return m_pArray; } |
| 4061 | iterator end() { return m_pArray + m_Count; } |
| 4062 | const_iterator cbegin() const { return m_pArray; } |
| 4063 | const_iterator cend() const { return m_pArray + m_Count; } |
| 4064 | const_iterator begin() const { return cbegin(); } |
| 4065 | const_iterator end() const { return cend(); } |
| 4066 | |
| 4067 | void pop_front() { VMA_HEAVY_ASSERT(m_Count > 0); remove(index: 0); } |
| 4068 | void pop_back() { VMA_HEAVY_ASSERT(m_Count > 0); resize(newCount: size() - 1); } |
| 4069 | void push_front(const T& src) { insert(index: 0, src); } |
| 4070 | |
| 4071 | void push_back(const T& src); |
| 4072 | void reserve(size_t newCapacity, bool freeMemory = false); |
| 4073 | void resize(size_t newCount); |
| 4074 | void clear() { resize(newCount: 0); } |
| 4075 | void shrink_to_fit(); |
| 4076 | void insert(size_t index, const T& src); |
| 4077 | void remove(size_t index); |
| 4078 | |
| 4079 | T& operator[](size_t index) { VMA_HEAVY_ASSERT(index < m_Count); return m_pArray[index]; } |
| 4080 | const T& operator[](size_t index) const { VMA_HEAVY_ASSERT(index < m_Count); return m_pArray[index]; } |
| 4081 | |
| 4082 | private: |
| 4083 | AllocatorT m_Allocator; |
| 4084 | T* m_pArray; |
| 4085 | size_t m_Count; |
| 4086 | size_t m_Capacity; |
| 4087 | }; |
| 4088 | |
| 4089 | #ifndef _VMA_VECTOR_FUNCTIONS |
| 4090 | template<typename T, typename AllocatorT> |
| 4091 | VmaVector<T, AllocatorT>::VmaVector(const AllocatorT& allocator) |
| 4092 | : m_Allocator(allocator), |
| 4093 | m_pArray(VMA_NULL), |
| 4094 | m_Count(0), |
| 4095 | m_Capacity(0) {} |
| 4096 | |
| 4097 | template<typename T, typename AllocatorT> |
| 4098 | VmaVector<T, AllocatorT>::VmaVector(size_t count, const AllocatorT& allocator) |
| 4099 | : m_Allocator(allocator), |
| 4100 | m_pArray(count ? (T*)VmaAllocateArray<T>(allocator.m_pCallbacks, count) : VMA_NULL), |
| 4101 | m_Count(count), |
| 4102 | m_Capacity(count) {} |
| 4103 | |
| 4104 | template<typename T, typename AllocatorT> |
| 4105 | VmaVector<T, AllocatorT>::VmaVector(const VmaVector& src) |
| 4106 | : m_Allocator(src.m_Allocator), |
| 4107 | m_pArray(src.m_Count ? (T*)VmaAllocateArray<T>(src.m_Allocator.m_pCallbacks, src.m_Count) : VMA_NULL), |
| 4108 | m_Count(src.m_Count), |
| 4109 | m_Capacity(src.m_Count) |
| 4110 | { |
| 4111 | if (m_Count != 0) |
| 4112 | { |
| 4113 | memcpy(m_pArray, src.m_pArray, m_Count * sizeof(T)); |
| 4114 | } |
| 4115 | } |
| 4116 | |
| 4117 | template<typename T, typename AllocatorT> |
| 4118 | VmaVector<T, AllocatorT>& VmaVector<T, AllocatorT>::operator=(const VmaVector& rhs) |
| 4119 | { |
| 4120 | if (&rhs != this) |
| 4121 | { |
| 4122 | resize(newCount: rhs.m_Count); |
| 4123 | if (m_Count != 0) |
| 4124 | { |
| 4125 | memcpy(m_pArray, rhs.m_pArray, m_Count * sizeof(T)); |
| 4126 | } |
| 4127 | } |
| 4128 | return *this; |
| 4129 | } |
| 4130 | |
| 4131 | template<typename T, typename AllocatorT> |
| 4132 | void VmaVector<T, AllocatorT>::push_back(const T& src) |
| 4133 | { |
| 4134 | const size_t newIndex = size(); |
| 4135 | resize(newCount: newIndex + 1); |
| 4136 | m_pArray[newIndex] = src; |
| 4137 | } |
| 4138 | |
| 4139 | template<typename T, typename AllocatorT> |
| 4140 | void VmaVector<T, AllocatorT>::reserve(size_t newCapacity, bool freeMemory) |
| 4141 | { |
| 4142 | newCapacity = VMA_MAX(newCapacity, m_Count); |
| 4143 | |
| 4144 | if ((newCapacity < m_Capacity) && !freeMemory) |
| 4145 | { |
| 4146 | newCapacity = m_Capacity; |
| 4147 | } |
| 4148 | |
| 4149 | if (newCapacity != m_Capacity) |
| 4150 | { |
| 4151 | T* const newArray = newCapacity ? VmaAllocateArray<T>(m_Allocator, newCapacity) : VMA_NULL; |
| 4152 | if (m_Count != 0) |
| 4153 | { |
| 4154 | memcpy(newArray, m_pArray, m_Count * sizeof(T)); |
| 4155 | } |
| 4156 | VmaFree(m_Allocator.m_pCallbacks, m_pArray); |
| 4157 | m_Capacity = newCapacity; |
| 4158 | m_pArray = newArray; |
| 4159 | } |
| 4160 | } |
| 4161 | |
| 4162 | template<typename T, typename AllocatorT> |
| 4163 | void VmaVector<T, AllocatorT>::resize(size_t newCount) |
| 4164 | { |
| 4165 | size_t newCapacity = m_Capacity; |
| 4166 | if (newCount > m_Capacity) |
| 4167 | { |
| 4168 | newCapacity = VMA_MAX(newCount, VMA_MAX(m_Capacity * 3 / 2, (size_t)8)); |
| 4169 | } |
| 4170 | |
| 4171 | if (newCapacity != m_Capacity) |
| 4172 | { |
| 4173 | T* const newArray = newCapacity ? VmaAllocateArray<T>(m_Allocator.m_pCallbacks, newCapacity) : VMA_NULL; |
| 4174 | const size_t elementsToCopy = VMA_MIN(m_Count, newCount); |
| 4175 | if (elementsToCopy != 0) |
| 4176 | { |
| 4177 | memcpy(newArray, m_pArray, elementsToCopy * sizeof(T)); |
| 4178 | } |
| 4179 | VmaFree(m_Allocator.m_pCallbacks, m_pArray); |
| 4180 | m_Capacity = newCapacity; |
| 4181 | m_pArray = newArray; |
| 4182 | } |
| 4183 | |
| 4184 | m_Count = newCount; |
| 4185 | } |
| 4186 | |
| 4187 | template<typename T, typename AllocatorT> |
| 4188 | void VmaVector<T, AllocatorT>::shrink_to_fit() |
| 4189 | { |
| 4190 | if (m_Capacity > m_Count) |
| 4191 | { |
| 4192 | T* newArray = VMA_NULL; |
| 4193 | if (m_Count > 0) |
| 4194 | { |
| 4195 | newArray = VmaAllocateArray<T>(m_Allocator.m_pCallbacks, m_Count); |
| 4196 | memcpy(newArray, m_pArray, m_Count * sizeof(T)); |
| 4197 | } |
| 4198 | VmaFree(m_Allocator.m_pCallbacks, m_pArray); |
| 4199 | m_Capacity = m_Count; |
| 4200 | m_pArray = newArray; |
| 4201 | } |
| 4202 | } |
| 4203 | |
| 4204 | template<typename T, typename AllocatorT> |
| 4205 | void VmaVector<T, AllocatorT>::insert(size_t index, const T& src) |
| 4206 | { |
| 4207 | VMA_HEAVY_ASSERT(index <= m_Count); |
| 4208 | const size_t oldCount = size(); |
| 4209 | resize(newCount: oldCount + 1); |
| 4210 | if (index < oldCount) |
| 4211 | { |
| 4212 | memmove(m_pArray + (index + 1), m_pArray + index, (oldCount - index) * sizeof(T)); |
| 4213 | } |
| 4214 | m_pArray[index] = src; |
| 4215 | } |
| 4216 | |
| 4217 | template<typename T, typename AllocatorT> |
| 4218 | void VmaVector<T, AllocatorT>::remove(size_t index) |
| 4219 | { |
| 4220 | VMA_HEAVY_ASSERT(index < m_Count); |
| 4221 | const size_t oldCount = size(); |
| 4222 | if (index < oldCount - 1) |
| 4223 | { |
| 4224 | memmove(m_pArray + index, m_pArray + (index + 1), (oldCount - index - 1) * sizeof(T)); |
| 4225 | } |
| 4226 | resize(newCount: oldCount - 1); |
| 4227 | } |
| 4228 | #endif // _VMA_VECTOR_FUNCTIONS |
| 4229 | |
| 4230 | template<typename T, typename allocatorT> |
| 4231 | static void VmaVectorInsert(VmaVector<T, allocatorT>& vec, size_t index, const T& item) |
| 4232 | { |
| 4233 | vec.insert(index, item); |
| 4234 | } |
| 4235 | |
| 4236 | template<typename T, typename allocatorT> |
| 4237 | static void VmaVectorRemove(VmaVector<T, allocatorT>& vec, size_t index) |
| 4238 | { |
| 4239 | vec.remove(index); |
| 4240 | } |
| 4241 | #endif // _VMA_VECTOR |
| 4242 | |
| 4243 | #ifndef _VMA_SMALL_VECTOR |
| 4244 | /* |
| 4245 | This is a vector (a variable-sized array), optimized for the case when the array is small. |
| 4246 | |
| 4247 | It contains some number of elements in-place, which allows it to avoid heap allocation |
| 4248 | when the actual number of elements is below that threshold. This allows normal "small" |
| 4249 | cases to be fast without losing generality for large inputs. |
| 4250 | */ |
| 4251 | template<typename T, typename AllocatorT, size_t N> |
| 4252 | class VmaSmallVector |
| 4253 | { |
| 4254 | public: |
| 4255 | typedef T value_type; |
| 4256 | typedef T* iterator; |
| 4257 | |
| 4258 | VmaSmallVector(const AllocatorT& allocator); |
| 4259 | VmaSmallVector(size_t count, const AllocatorT& allocator); |
| 4260 | template<typename SrcT, typename SrcAllocatorT, size_t SrcN> |
| 4261 | VmaSmallVector(const VmaSmallVector<SrcT, SrcAllocatorT, SrcN>&) = delete; |
| 4262 | template<typename SrcT, typename SrcAllocatorT, size_t SrcN> |
| 4263 | VmaSmallVector<T, AllocatorT, N>& operator=(const VmaSmallVector<SrcT, SrcAllocatorT, SrcN>&) = delete; |
| 4264 | ~VmaSmallVector() = default; |
| 4265 | |
| 4266 | bool empty() const { return m_Count == 0; } |
| 4267 | size_t size() const { return m_Count; } |
| 4268 | T* data() { return m_Count > N ? m_DynamicArray.data() : m_StaticArray; } |
| 4269 | T& front() { VMA_HEAVY_ASSERT(m_Count > 0); return data()[0]; } |
| 4270 | T& back() { VMA_HEAVY_ASSERT(m_Count > 0); return data()[m_Count - 1]; } |
| 4271 | const T* data() const { return m_Count > N ? m_DynamicArray.data() : m_StaticArray; } |
| 4272 | const T& front() const { VMA_HEAVY_ASSERT(m_Count > 0); return data()[0]; } |
| 4273 | const T& back() const { VMA_HEAVY_ASSERT(m_Count > 0); return data()[m_Count - 1]; } |
| 4274 | |
| 4275 | iterator begin() { return data(); } |
| 4276 | iterator end() { return data() + m_Count; } |
| 4277 | |
| 4278 | void pop_front() { VMA_HEAVY_ASSERT(m_Count > 0); remove(index: 0); } |
| 4279 | void pop_back() { VMA_HEAVY_ASSERT(m_Count > 0); resize(newCount: size() - 1); } |
| 4280 | void push_front(const T& src) { insert(index: 0, src); } |
| 4281 | |
| 4282 | void push_back(const T& src); |
| 4283 | void resize(size_t newCount, bool freeMemory = false); |
| 4284 | void clear(bool freeMemory = false); |
| 4285 | void insert(size_t index, const T& src); |
| 4286 | void remove(size_t index); |
| 4287 | |
| 4288 | T& operator[](size_t index) { VMA_HEAVY_ASSERT(index < m_Count); return data()[index]; } |
| 4289 | const T& operator[](size_t index) const { VMA_HEAVY_ASSERT(index < m_Count); return data()[index]; } |
| 4290 | |
| 4291 | private: |
| 4292 | size_t m_Count; |
| 4293 | T m_StaticArray[N]; // Used when m_Size <= N |
| 4294 | VmaVector<T, AllocatorT> m_DynamicArray; // Used when m_Size > N |
| 4295 | }; |
| 4296 | |
| 4297 | #ifndef _VMA_SMALL_VECTOR_FUNCTIONS |
| 4298 | template<typename T, typename AllocatorT, size_t N> |
| 4299 | VmaSmallVector<T, AllocatorT, N>::VmaSmallVector(const AllocatorT& allocator) |
| 4300 | : m_Count(0), |
| 4301 | m_DynamicArray(allocator) {} |
| 4302 | |
| 4303 | template<typename T, typename AllocatorT, size_t N> |
| 4304 | VmaSmallVector<T, AllocatorT, N>::VmaSmallVector(size_t count, const AllocatorT& allocator) |
| 4305 | : m_Count(count), |
| 4306 | m_DynamicArray(count > N ? count : 0, allocator) {} |
| 4307 | |
| 4308 | template<typename T, typename AllocatorT, size_t N> |
| 4309 | void VmaSmallVector<T, AllocatorT, N>::push_back(const T& src) |
| 4310 | { |
| 4311 | const size_t newIndex = size(); |
| 4312 | resize(newCount: newIndex + 1); |
| 4313 | data()[newIndex] = src; |
| 4314 | } |
| 4315 | |
| 4316 | template<typename T, typename AllocatorT, size_t N> |
| 4317 | void VmaSmallVector<T, AllocatorT, N>::resize(size_t newCount, bool freeMemory) |
| 4318 | { |
| 4319 | if (newCount > N && m_Count > N) |
| 4320 | { |
| 4321 | // Any direction, staying in m_DynamicArray |
| 4322 | m_DynamicArray.resize(newCount); |
| 4323 | if (freeMemory) |
| 4324 | { |
| 4325 | m_DynamicArray.shrink_to_fit(); |
| 4326 | } |
| 4327 | } |
| 4328 | else if (newCount > N && m_Count <= N) |
| 4329 | { |
| 4330 | // Growing, moving from m_StaticArray to m_DynamicArray |
| 4331 | m_DynamicArray.resize(newCount); |
| 4332 | if (m_Count > 0) |
| 4333 | { |
| 4334 | memcpy(m_DynamicArray.data(), m_StaticArray, m_Count * sizeof(T)); |
| 4335 | } |
| 4336 | } |
| 4337 | else if (newCount <= N && m_Count > N) |
| 4338 | { |
| 4339 | // Shrinking, moving from m_DynamicArray to m_StaticArray |
| 4340 | if (newCount > 0) |
| 4341 | { |
| 4342 | memcpy(m_StaticArray, m_DynamicArray.data(), newCount * sizeof(T)); |
| 4343 | } |
| 4344 | m_DynamicArray.resize(0); |
| 4345 | if (freeMemory) |
| 4346 | { |
| 4347 | m_DynamicArray.shrink_to_fit(); |
| 4348 | } |
| 4349 | } |
| 4350 | else |
| 4351 | { |
| 4352 | // Any direction, staying in m_StaticArray - nothing to do here |
| 4353 | } |
| 4354 | m_Count = newCount; |
| 4355 | } |
| 4356 | |
| 4357 | template<typename T, typename AllocatorT, size_t N> |
| 4358 | void VmaSmallVector<T, AllocatorT, N>::clear(bool freeMemory) |
| 4359 | { |
| 4360 | m_DynamicArray.clear(); |
| 4361 | if (freeMemory) |
| 4362 | { |
| 4363 | m_DynamicArray.shrink_to_fit(); |
| 4364 | } |
| 4365 | m_Count = 0; |
| 4366 | } |
| 4367 | |
| 4368 | template<typename T, typename AllocatorT, size_t N> |
| 4369 | void VmaSmallVector<T, AllocatorT, N>::insert(size_t index, const T& src) |
| 4370 | { |
| 4371 | VMA_HEAVY_ASSERT(index <= m_Count); |
| 4372 | const size_t oldCount = size(); |
| 4373 | resize(newCount: oldCount + 1); |
| 4374 | T* const dataPtr = data(); |
| 4375 | if (index < oldCount) |
| 4376 | { |
| 4377 | // I know, this could be more optimal for case where memmove can be memcpy directly from m_StaticArray to m_DynamicArray. |
| 4378 | memmove(dataPtr + (index + 1), dataPtr + index, (oldCount - index) * sizeof(T)); |
| 4379 | } |
| 4380 | dataPtr[index] = src; |
| 4381 | } |
| 4382 | |
| 4383 | template<typename T, typename AllocatorT, size_t N> |
| 4384 | void VmaSmallVector<T, AllocatorT, N>::remove(size_t index) |
| 4385 | { |
| 4386 | VMA_HEAVY_ASSERT(index < m_Count); |
| 4387 | const size_t oldCount = size(); |
| 4388 | if (index < oldCount - 1) |
| 4389 | { |
| 4390 | // I know, this could be more optimal for case where memmove can be memcpy directly from m_DynamicArray to m_StaticArray. |
| 4391 | T* const dataPtr = data(); |
| 4392 | memmove(dataPtr + index, dataPtr + (index + 1), (oldCount - index - 1) * sizeof(T)); |
| 4393 | } |
| 4394 | resize(newCount: oldCount - 1); |
| 4395 | } |
| 4396 | #endif // _VMA_SMALL_VECTOR_FUNCTIONS |
| 4397 | #endif // _VMA_SMALL_VECTOR |
| 4398 | |
| 4399 | #ifndef _VMA_POOL_ALLOCATOR |
| 4400 | /* |
| 4401 | Allocator for objects of type T using a list of arrays (pools) to speed up |
| 4402 | allocation. Number of elements that can be allocated is not bounded because |
| 4403 | allocator can create multiple blocks. |
| 4404 | */ |
| 4405 | template<typename T> |
| 4406 | class VmaPoolAllocator |
| 4407 | { |
| 4408 | VMA_CLASS_NO_COPY(VmaPoolAllocator) |
| 4409 | public: |
| 4410 | VmaPoolAllocator(const VkAllocationCallbacks* pAllocationCallbacks, uint32_t firstBlockCapacity); |
| 4411 | ~VmaPoolAllocator(); |
| 4412 | template<typename... Types> T* Alloc(Types&&... args); |
| 4413 | void Free(T* ptr); |
| 4414 | |
| 4415 | private: |
| 4416 | union Item |
| 4417 | { |
| 4418 | uint32_t NextFreeIndex; |
| 4419 | alignas(T) char Value[sizeof(T)]; |
| 4420 | }; |
| 4421 | struct ItemBlock |
| 4422 | { |
| 4423 | Item* pItems; |
| 4424 | uint32_t Capacity; |
| 4425 | uint32_t FirstFreeIndex; |
| 4426 | }; |
| 4427 | |
| 4428 | const VkAllocationCallbacks* m_pAllocationCallbacks; |
| 4429 | const uint32_t m_FirstBlockCapacity; |
| 4430 | VmaVector<ItemBlock, VmaStlAllocator<ItemBlock>> m_ItemBlocks; |
| 4431 | |
| 4432 | ItemBlock& CreateNewBlock(); |
| 4433 | }; |
| 4434 | |
| 4435 | #ifndef _VMA_POOL_ALLOCATOR_FUNCTIONS |
| 4436 | template<typename T> |
| 4437 | VmaPoolAllocator<T>::VmaPoolAllocator(const VkAllocationCallbacks* pAllocationCallbacks, uint32_t firstBlockCapacity) |
| 4438 | : m_pAllocationCallbacks(pAllocationCallbacks), |
| 4439 | m_FirstBlockCapacity(firstBlockCapacity), |
| 4440 | m_ItemBlocks(VmaStlAllocator<ItemBlock>(pAllocationCallbacks)) |
| 4441 | { |
| 4442 | VMA_ASSERT(m_FirstBlockCapacity > 1); |
| 4443 | } |
| 4444 | |
| 4445 | template<typename T> |
| 4446 | VmaPoolAllocator<T>::~VmaPoolAllocator() |
| 4447 | { |
| 4448 | for (size_t i = m_ItemBlocks.size(); i--;) |
| 4449 | vma_delete_array(m_pAllocationCallbacks, m_ItemBlocks[i].pItems, m_ItemBlocks[i].Capacity); |
| 4450 | m_ItemBlocks.clear(); |
| 4451 | } |
| 4452 | |
| 4453 | template<typename T> |
| 4454 | template<typename... Types> T* VmaPoolAllocator<T>::Alloc(Types&&... args) |
| 4455 | { |
| 4456 | for (size_t i = m_ItemBlocks.size(); i--; ) |
| 4457 | { |
| 4458 | ItemBlock& block = m_ItemBlocks[i]; |
| 4459 | // This block has some free items: Use first one. |
| 4460 | if (block.FirstFreeIndex != UINT32_MAX) |
| 4461 | { |
| 4462 | Item* const pItem = &block.pItems[block.FirstFreeIndex]; |
| 4463 | block.FirstFreeIndex = pItem->NextFreeIndex; |
| 4464 | T* result = (T*)&pItem->Value; |
| 4465 | new(result)T(std::forward<Types>(args)...); // Explicit constructor call. |
| 4466 | return result; |
| 4467 | } |
| 4468 | } |
| 4469 | |
| 4470 | // No block has free item: Create new one and use it. |
| 4471 | ItemBlock& newBlock = CreateNewBlock(); |
| 4472 | Item* const pItem = &newBlock.pItems[0]; |
| 4473 | newBlock.FirstFreeIndex = pItem->NextFreeIndex; |
| 4474 | T* result = (T*)&pItem->Value; |
| 4475 | new(result) T(std::forward<Types>(args)...); // Explicit constructor call. |
| 4476 | return result; |
| 4477 | } |
| 4478 | |
| 4479 | template<typename T> |
| 4480 | void VmaPoolAllocator<T>::Free(T* ptr) |
| 4481 | { |
| 4482 | // Search all memory blocks to find ptr. |
| 4483 | for (size_t i = m_ItemBlocks.size(); i--; ) |
| 4484 | { |
| 4485 | ItemBlock& block = m_ItemBlocks[i]; |
| 4486 | |
| 4487 | // Casting to union. |
| 4488 | Item* pItemPtr; |
| 4489 | memcpy(&pItemPtr, &ptr, sizeof(pItemPtr)); |
| 4490 | |
| 4491 | // Check if pItemPtr is in address range of this block. |
| 4492 | if ((pItemPtr >= block.pItems) && (pItemPtr < block.pItems + block.Capacity)) |
| 4493 | { |
| 4494 | ptr->~T(); // Explicit destructor call. |
| 4495 | const uint32_t index = static_cast<uint32_t>(pItemPtr - block.pItems); |
| 4496 | pItemPtr->NextFreeIndex = block.FirstFreeIndex; |
| 4497 | block.FirstFreeIndex = index; |
| 4498 | return; |
| 4499 | } |
| 4500 | } |
| 4501 | VMA_ASSERT(0 && "Pointer doesn't belong to this memory pool." ); |
| 4502 | } |
| 4503 | |
| 4504 | template<typename T> |
| 4505 | typename VmaPoolAllocator<T>::ItemBlock& VmaPoolAllocator<T>::CreateNewBlock() |
| 4506 | { |
| 4507 | const uint32_t newBlockCapacity = m_ItemBlocks.empty() ? |
| 4508 | m_FirstBlockCapacity : m_ItemBlocks.back().Capacity * 3 / 2; |
| 4509 | |
| 4510 | const ItemBlock newBlock = |
| 4511 | { |
| 4512 | vma_new_array(m_pAllocationCallbacks, Item, newBlockCapacity), |
| 4513 | newBlockCapacity, |
| 4514 | 0 |
| 4515 | }; |
| 4516 | |
| 4517 | m_ItemBlocks.push_back(newBlock); |
| 4518 | |
| 4519 | // Setup singly-linked list of all free items in this block. |
| 4520 | for (uint32_t i = 0; i < newBlockCapacity - 1; ++i) |
| 4521 | newBlock.pItems[i].NextFreeIndex = i + 1; |
| 4522 | newBlock.pItems[newBlockCapacity - 1].NextFreeIndex = UINT32_MAX; |
| 4523 | return m_ItemBlocks.back(); |
| 4524 | } |
| 4525 | #endif // _VMA_POOL_ALLOCATOR_FUNCTIONS |
| 4526 | #endif // _VMA_POOL_ALLOCATOR |
| 4527 | |
| 4528 | #ifndef _VMA_RAW_LIST |
| 4529 | template<typename T> |
| 4530 | struct VmaListItem |
| 4531 | { |
| 4532 | VmaListItem* pPrev; |
| 4533 | VmaListItem* pNext; |
| 4534 | T Value; |
| 4535 | }; |
| 4536 | |
| 4537 | // Doubly linked list. |
| 4538 | template<typename T> |
| 4539 | class VmaRawList |
| 4540 | { |
| 4541 | VMA_CLASS_NO_COPY(VmaRawList) |
| 4542 | public: |
| 4543 | typedef VmaListItem<T> ItemType; |
| 4544 | |
| 4545 | VmaRawList(const VkAllocationCallbacks* pAllocationCallbacks); |
| 4546 | // Intentionally not calling Clear, because that would be unnecessary |
| 4547 | // computations to return all items to m_ItemAllocator as free. |
| 4548 | ~VmaRawList() = default; |
| 4549 | |
| 4550 | size_t GetCount() const { return m_Count; } |
| 4551 | bool IsEmpty() const { return m_Count == 0; } |
| 4552 | |
| 4553 | ItemType* Front() { return m_pFront; } |
| 4554 | ItemType* Back() { return m_pBack; } |
| 4555 | const ItemType* Front() const { return m_pFront; } |
| 4556 | const ItemType* Back() const { return m_pBack; } |
| 4557 | |
| 4558 | ItemType* PushFront(); |
| 4559 | ItemType* PushBack(); |
| 4560 | ItemType* PushFront(const T& value); |
| 4561 | ItemType* PushBack(const T& value); |
| 4562 | void PopFront(); |
| 4563 | void PopBack(); |
| 4564 | |
| 4565 | // Item can be null - it means PushBack. |
| 4566 | ItemType* InsertBefore(ItemType* pItem); |
| 4567 | // Item can be null - it means PushFront. |
| 4568 | ItemType* InsertAfter(ItemType* pItem); |
| 4569 | ItemType* InsertBefore(ItemType* pItem, const T& value); |
| 4570 | ItemType* InsertAfter(ItemType* pItem, const T& value); |
| 4571 | |
| 4572 | void Clear(); |
| 4573 | void Remove(ItemType* pItem); |
| 4574 | |
| 4575 | private: |
| 4576 | const VkAllocationCallbacks* const m_pAllocationCallbacks; |
| 4577 | VmaPoolAllocator<ItemType> m_ItemAllocator; |
| 4578 | ItemType* m_pFront; |
| 4579 | ItemType* m_pBack; |
| 4580 | size_t m_Count; |
| 4581 | }; |
| 4582 | |
| 4583 | #ifndef _VMA_RAW_LIST_FUNCTIONS |
| 4584 | template<typename T> |
| 4585 | VmaRawList<T>::VmaRawList(const VkAllocationCallbacks* pAllocationCallbacks) |
| 4586 | : m_pAllocationCallbacks(pAllocationCallbacks), |
| 4587 | m_ItemAllocator(pAllocationCallbacks, 128), |
| 4588 | m_pFront(VMA_NULL), |
| 4589 | m_pBack(VMA_NULL), |
| 4590 | m_Count(0) {} |
| 4591 | |
| 4592 | template<typename T> |
| 4593 | VmaListItem<T>* VmaRawList<T>::PushFront() |
| 4594 | { |
| 4595 | ItemType* const pNewItem = m_ItemAllocator.Alloc(); |
| 4596 | pNewItem->pPrev = VMA_NULL; |
| 4597 | if (IsEmpty()) |
| 4598 | { |
| 4599 | pNewItem->pNext = VMA_NULL; |
| 4600 | m_pFront = pNewItem; |
| 4601 | m_pBack = pNewItem; |
| 4602 | m_Count = 1; |
| 4603 | } |
| 4604 | else |
| 4605 | { |
| 4606 | pNewItem->pNext = m_pFront; |
| 4607 | m_pFront->pPrev = pNewItem; |
| 4608 | m_pFront = pNewItem; |
| 4609 | ++m_Count; |
| 4610 | } |
| 4611 | return pNewItem; |
| 4612 | } |
| 4613 | |
| 4614 | template<typename T> |
| 4615 | VmaListItem<T>* VmaRawList<T>::PushBack() |
| 4616 | { |
| 4617 | ItemType* const pNewItem = m_ItemAllocator.Alloc(); |
| 4618 | pNewItem->pNext = VMA_NULL; |
| 4619 | if(IsEmpty()) |
| 4620 | { |
| 4621 | pNewItem->pPrev = VMA_NULL; |
| 4622 | m_pFront = pNewItem; |
| 4623 | m_pBack = pNewItem; |
| 4624 | m_Count = 1; |
| 4625 | } |
| 4626 | else |
| 4627 | { |
| 4628 | pNewItem->pPrev = m_pBack; |
| 4629 | m_pBack->pNext = pNewItem; |
| 4630 | m_pBack = pNewItem; |
| 4631 | ++m_Count; |
| 4632 | } |
| 4633 | return pNewItem; |
| 4634 | } |
| 4635 | |
| 4636 | template<typename T> |
| 4637 | VmaListItem<T>* VmaRawList<T>::PushFront(const T& value) |
| 4638 | { |
| 4639 | ItemType* const pNewItem = PushFront(); |
| 4640 | pNewItem->Value = value; |
| 4641 | return pNewItem; |
| 4642 | } |
| 4643 | |
| 4644 | template<typename T> |
| 4645 | VmaListItem<T>* VmaRawList<T>::PushBack(const T& value) |
| 4646 | { |
| 4647 | ItemType* const pNewItem = PushBack(); |
| 4648 | pNewItem->Value = value; |
| 4649 | return pNewItem; |
| 4650 | } |
| 4651 | |
| 4652 | template<typename T> |
| 4653 | void VmaRawList<T>::PopFront() |
| 4654 | { |
| 4655 | VMA_HEAVY_ASSERT(m_Count > 0); |
| 4656 | ItemType* const pFrontItem = m_pFront; |
| 4657 | ItemType* const pNextItem = pFrontItem->pNext; |
| 4658 | if (pNextItem != VMA_NULL) |
| 4659 | { |
| 4660 | pNextItem->pPrev = VMA_NULL; |
| 4661 | } |
| 4662 | m_pFront = pNextItem; |
| 4663 | m_ItemAllocator.Free(pFrontItem); |
| 4664 | --m_Count; |
| 4665 | } |
| 4666 | |
| 4667 | template<typename T> |
| 4668 | void VmaRawList<T>::PopBack() |
| 4669 | { |
| 4670 | VMA_HEAVY_ASSERT(m_Count > 0); |
| 4671 | ItemType* const pBackItem = m_pBack; |
| 4672 | ItemType* const pPrevItem = pBackItem->pPrev; |
| 4673 | if(pPrevItem != VMA_NULL) |
| 4674 | { |
| 4675 | pPrevItem->pNext = VMA_NULL; |
| 4676 | } |
| 4677 | m_pBack = pPrevItem; |
| 4678 | m_ItemAllocator.Free(pBackItem); |
| 4679 | --m_Count; |
| 4680 | } |
| 4681 | |
| 4682 | template<typename T> |
| 4683 | void VmaRawList<T>::Clear() |
| 4684 | { |
| 4685 | if (IsEmpty() == false) |
| 4686 | { |
| 4687 | ItemType* pItem = m_pBack; |
| 4688 | while (pItem != VMA_NULL) |
| 4689 | { |
| 4690 | ItemType* const pPrevItem = pItem->pPrev; |
| 4691 | m_ItemAllocator.Free(pItem); |
| 4692 | pItem = pPrevItem; |
| 4693 | } |
| 4694 | m_pFront = VMA_NULL; |
| 4695 | m_pBack = VMA_NULL; |
| 4696 | m_Count = 0; |
| 4697 | } |
| 4698 | } |
| 4699 | |
| 4700 | template<typename T> |
| 4701 | void VmaRawList<T>::Remove(ItemType* pItem) |
| 4702 | { |
| 4703 | VMA_HEAVY_ASSERT(pItem != VMA_NULL); |
| 4704 | VMA_HEAVY_ASSERT(m_Count > 0); |
| 4705 | |
| 4706 | if(pItem->pPrev != VMA_NULL) |
| 4707 | { |
| 4708 | pItem->pPrev->pNext = pItem->pNext; |
| 4709 | } |
| 4710 | else |
| 4711 | { |
| 4712 | VMA_HEAVY_ASSERT(m_pFront == pItem); |
| 4713 | m_pFront = pItem->pNext; |
| 4714 | } |
| 4715 | |
| 4716 | if(pItem->pNext != VMA_NULL) |
| 4717 | { |
| 4718 | pItem->pNext->pPrev = pItem->pPrev; |
| 4719 | } |
| 4720 | else |
| 4721 | { |
| 4722 | VMA_HEAVY_ASSERT(m_pBack == pItem); |
| 4723 | m_pBack = pItem->pPrev; |
| 4724 | } |
| 4725 | |
| 4726 | m_ItemAllocator.Free(pItem); |
| 4727 | --m_Count; |
| 4728 | } |
| 4729 | |
| 4730 | template<typename T> |
| 4731 | VmaListItem<T>* VmaRawList<T>::InsertBefore(ItemType* pItem) |
| 4732 | { |
| 4733 | if(pItem != VMA_NULL) |
| 4734 | { |
| 4735 | ItemType* const prevItem = pItem->pPrev; |
| 4736 | ItemType* const newItem = m_ItemAllocator.Alloc(); |
| 4737 | newItem->pPrev = prevItem; |
| 4738 | newItem->pNext = pItem; |
| 4739 | pItem->pPrev = newItem; |
| 4740 | if(prevItem != VMA_NULL) |
| 4741 | { |
| 4742 | prevItem->pNext = newItem; |
| 4743 | } |
| 4744 | else |
| 4745 | { |
| 4746 | VMA_HEAVY_ASSERT(m_pFront == pItem); |
| 4747 | m_pFront = newItem; |
| 4748 | } |
| 4749 | ++m_Count; |
| 4750 | return newItem; |
| 4751 | } |
| 4752 | else |
| 4753 | return PushBack(); |
| 4754 | } |
| 4755 | |
| 4756 | template<typename T> |
| 4757 | VmaListItem<T>* VmaRawList<T>::InsertAfter(ItemType* pItem) |
| 4758 | { |
| 4759 | if(pItem != VMA_NULL) |
| 4760 | { |
| 4761 | ItemType* const nextItem = pItem->pNext; |
| 4762 | ItemType* const newItem = m_ItemAllocator.Alloc(); |
| 4763 | newItem->pNext = nextItem; |
| 4764 | newItem->pPrev = pItem; |
| 4765 | pItem->pNext = newItem; |
| 4766 | if(nextItem != VMA_NULL) |
| 4767 | { |
| 4768 | nextItem->pPrev = newItem; |
| 4769 | } |
| 4770 | else |
| 4771 | { |
| 4772 | VMA_HEAVY_ASSERT(m_pBack == pItem); |
| 4773 | m_pBack = newItem; |
| 4774 | } |
| 4775 | ++m_Count; |
| 4776 | return newItem; |
| 4777 | } |
| 4778 | else |
| 4779 | return PushFront(); |
| 4780 | } |
| 4781 | |
| 4782 | template<typename T> |
| 4783 | VmaListItem<T>* VmaRawList<T>::InsertBefore(ItemType* pItem, const T& value) |
| 4784 | { |
| 4785 | ItemType* const newItem = InsertBefore(pItem); |
| 4786 | newItem->Value = value; |
| 4787 | return newItem; |
| 4788 | } |
| 4789 | |
| 4790 | template<typename T> |
| 4791 | VmaListItem<T>* VmaRawList<T>::InsertAfter(ItemType* pItem, const T& value) |
| 4792 | { |
| 4793 | ItemType* const newItem = InsertAfter(pItem); |
| 4794 | newItem->Value = value; |
| 4795 | return newItem; |
| 4796 | } |
| 4797 | #endif // _VMA_RAW_LIST_FUNCTIONS |
| 4798 | #endif // _VMA_RAW_LIST |
| 4799 | |
| 4800 | #ifndef _VMA_LIST |
| 4801 | template<typename T, typename AllocatorT> |
| 4802 | class VmaList |
| 4803 | { |
| 4804 | VMA_CLASS_NO_COPY(VmaList) |
| 4805 | public: |
| 4806 | class reverse_iterator; |
| 4807 | class const_iterator; |
| 4808 | class const_reverse_iterator; |
| 4809 | |
| 4810 | class iterator |
| 4811 | { |
| 4812 | friend class const_iterator; |
| 4813 | friend class VmaList<T, AllocatorT>; |
| 4814 | public: |
| 4815 | iterator() : m_pList(VMA_NULL), m_pItem(VMA_NULL) {} |
| 4816 | iterator(const reverse_iterator& src) : m_pList(src.m_pList), m_pItem(src.m_pItem) {} |
| 4817 | |
| 4818 | T& operator*() const { VMA_HEAVY_ASSERT(m_pItem != VMA_NULL); return m_pItem->Value; } |
| 4819 | T* operator->() const { VMA_HEAVY_ASSERT(m_pItem != VMA_NULL); return &m_pItem->Value; } |
| 4820 | |
| 4821 | bool operator==(const iterator& rhs) const { VMA_HEAVY_ASSERT(m_pList == rhs.m_pList); return m_pItem == rhs.m_pItem; } |
| 4822 | bool operator!=(const iterator& rhs) const { VMA_HEAVY_ASSERT(m_pList == rhs.m_pList); return m_pItem != rhs.m_pItem; } |
| 4823 | |
| 4824 | iterator operator++(int) { iterator result = *this; ++*this; return result; } |
| 4825 | iterator operator--(int) { iterator result = *this; --*this; return result; } |
| 4826 | |
| 4827 | iterator& operator++() { VMA_HEAVY_ASSERT(m_pItem != VMA_NULL); m_pItem = m_pItem->pNext; return *this; } |
| 4828 | iterator& operator--(); |
| 4829 | |
| 4830 | private: |
| 4831 | VmaRawList<T>* m_pList; |
| 4832 | VmaListItem<T>* m_pItem; |
| 4833 | |
| 4834 | iterator(VmaRawList<T>* pList, VmaListItem<T>* pItem) : m_pList(pList), m_pItem(pItem) {} |
| 4835 | }; |
| 4836 | class reverse_iterator |
| 4837 | { |
| 4838 | friend class const_reverse_iterator; |
| 4839 | friend class VmaList<T, AllocatorT>; |
| 4840 | public: |
| 4841 | reverse_iterator() : m_pList(VMA_NULL), m_pItem(VMA_NULL) {} |
| 4842 | reverse_iterator(const iterator& src) : m_pList(src.m_pList), m_pItem(src.m_pItem) {} |
| 4843 | |
| 4844 | T& operator*() const { VMA_HEAVY_ASSERT(m_pItem != VMA_NULL); return m_pItem->Value; } |
| 4845 | T* operator->() const { VMA_HEAVY_ASSERT(m_pItem != VMA_NULL); return &m_pItem->Value; } |
| 4846 | |
| 4847 | bool operator==(const reverse_iterator& rhs) const { VMA_HEAVY_ASSERT(m_pList == rhs.m_pList); return m_pItem == rhs.m_pItem; } |
| 4848 | bool operator!=(const reverse_iterator& rhs) const { VMA_HEAVY_ASSERT(m_pList == rhs.m_pList); return m_pItem != rhs.m_pItem; } |
| 4849 | |
| 4850 | reverse_iterator operator++(int) { reverse_iterator result = *this; ++* this; return result; } |
| 4851 | reverse_iterator operator--(int) { reverse_iterator result = *this; --* this; return result; } |
| 4852 | |
| 4853 | reverse_iterator& operator++() { VMA_HEAVY_ASSERT(m_pItem != VMA_NULL); m_pItem = m_pItem->pPrev; return *this; } |
| 4854 | reverse_iterator& operator--(); |
| 4855 | |
| 4856 | private: |
| 4857 | VmaRawList<T>* m_pList; |
| 4858 | VmaListItem<T>* m_pItem; |
| 4859 | |
| 4860 | reverse_iterator(VmaRawList<T>* pList, VmaListItem<T>* pItem) : m_pList(pList), m_pItem(pItem) {} |
| 4861 | }; |
| 4862 | class const_iterator |
| 4863 | { |
| 4864 | friend class VmaList<T, AllocatorT>; |
| 4865 | public: |
| 4866 | const_iterator() : m_pList(VMA_NULL), m_pItem(VMA_NULL) {} |
| 4867 | const_iterator(const iterator& src) : m_pList(src.m_pList), m_pItem(src.m_pItem) {} |
| 4868 | const_iterator(const reverse_iterator& src) : m_pList(src.m_pList), m_pItem(src.m_pItem) {} |
| 4869 | |
| 4870 | iterator drop_const() { return { const_cast<VmaRawList<T>*>(m_pList), const_cast<VmaListItem<T>*>(m_pItem) }; } |
| 4871 | |
| 4872 | const T& operator*() const { VMA_HEAVY_ASSERT(m_pItem != VMA_NULL); return m_pItem->Value; } |
| 4873 | const T* operator->() const { VMA_HEAVY_ASSERT(m_pItem != VMA_NULL); return &m_pItem->Value; } |
| 4874 | |
| 4875 | bool operator==(const const_iterator& rhs) const { VMA_HEAVY_ASSERT(m_pList == rhs.m_pList); return m_pItem == rhs.m_pItem; } |
| 4876 | bool operator!=(const const_iterator& rhs) const { VMA_HEAVY_ASSERT(m_pList == rhs.m_pList); return m_pItem != rhs.m_pItem; } |
| 4877 | |
| 4878 | const_iterator operator++(int) { const_iterator result = *this; ++* this; return result; } |
| 4879 | const_iterator operator--(int) { const_iterator result = *this; --* this; return result; } |
| 4880 | |
| 4881 | const_iterator& operator++() { VMA_HEAVY_ASSERT(m_pItem != VMA_NULL); m_pItem = m_pItem->pNext; return *this; } |
| 4882 | const_iterator& operator--(); |
| 4883 | |
| 4884 | private: |
| 4885 | const VmaRawList<T>* m_pList; |
| 4886 | const VmaListItem<T>* m_pItem; |
| 4887 | |
| 4888 | const_iterator(const VmaRawList<T>* pList, const VmaListItem<T>* pItem) : m_pList(pList), m_pItem(pItem) {} |
| 4889 | }; |
| 4890 | class const_reverse_iterator |
| 4891 | { |
| 4892 | friend class VmaList<T, AllocatorT>; |
| 4893 | public: |
| 4894 | const_reverse_iterator() : m_pList(VMA_NULL), m_pItem(VMA_NULL) {} |
| 4895 | const_reverse_iterator(const reverse_iterator& src) : m_pList(src.m_pList), m_pItem(src.m_pItem) {} |
| 4896 | const_reverse_iterator(const iterator& src) : m_pList(src.m_pList), m_pItem(src.m_pItem) {} |
| 4897 | |
| 4898 | reverse_iterator drop_const() { return { const_cast<VmaRawList<T>*>(m_pList), const_cast<VmaListItem<T>*>(m_pItem) }; } |
| 4899 | |
| 4900 | const T& operator*() const { VMA_HEAVY_ASSERT(m_pItem != VMA_NULL); return m_pItem->Value; } |
| 4901 | const T* operator->() const { VMA_HEAVY_ASSERT(m_pItem != VMA_NULL); return &m_pItem->Value; } |
| 4902 | |
| 4903 | bool operator==(const const_reverse_iterator& rhs) const { VMA_HEAVY_ASSERT(m_pList == rhs.m_pList); return m_pItem == rhs.m_pItem; } |
| 4904 | bool operator!=(const const_reverse_iterator& rhs) const { VMA_HEAVY_ASSERT(m_pList == rhs.m_pList); return m_pItem != rhs.m_pItem; } |
| 4905 | |
| 4906 | const_reverse_iterator operator++(int) { const_reverse_iterator result = *this; ++* this; return result; } |
| 4907 | const_reverse_iterator operator--(int) { const_reverse_iterator result = *this; --* this; return result; } |
| 4908 | |
| 4909 | const_reverse_iterator& operator++() { VMA_HEAVY_ASSERT(m_pItem != VMA_NULL); m_pItem = m_pItem->pPrev; return *this; } |
| 4910 | const_reverse_iterator& operator--(); |
| 4911 | |
| 4912 | private: |
| 4913 | const VmaRawList<T>* m_pList; |
| 4914 | const VmaListItem<T>* m_pItem; |
| 4915 | |
| 4916 | const_reverse_iterator(const VmaRawList<T>* pList, const VmaListItem<T>* pItem) : m_pList(pList), m_pItem(pItem) {} |
| 4917 | }; |
| 4918 | |
| 4919 | VmaList(const AllocatorT& allocator) : m_RawList(allocator.m_pCallbacks) {} |
| 4920 | |
| 4921 | bool empty() const { return m_RawList.IsEmpty(); } |
| 4922 | size_t size() const { return m_RawList.GetCount(); } |
| 4923 | |
| 4924 | iterator begin() { return iterator(&m_RawList, m_RawList.Front()); } |
| 4925 | iterator end() { return iterator(&m_RawList, VMA_NULL); } |
| 4926 | |
| 4927 | const_iterator cbegin() const { return const_iterator(&m_RawList, m_RawList.Front()); } |
| 4928 | const_iterator cend() const { return const_iterator(&m_RawList, VMA_NULL); } |
| 4929 | |
| 4930 | const_iterator begin() const { return cbegin(); } |
| 4931 | const_iterator end() const { return cend(); } |
| 4932 | |
| 4933 | reverse_iterator rbegin() { return reverse_iterator(&m_RawList, m_RawList.Back()); } |
| 4934 | reverse_iterator rend() { return reverse_iterator(&m_RawList, VMA_NULL); } |
| 4935 | |
| 4936 | const_reverse_iterator crbegin() const { return const_reverse_iterator(&m_RawList, m_RawList.Back()); } |
| 4937 | const_reverse_iterator crend() const { return const_reverse_iterator(&m_RawList, VMA_NULL); } |
| 4938 | |
| 4939 | const_reverse_iterator rbegin() const { return crbegin(); } |
| 4940 | const_reverse_iterator rend() const { return crend(); } |
| 4941 | |
| 4942 | void push_back(const T& value) { m_RawList.PushBack(value); } |
| 4943 | iterator insert(iterator it, const T& value) { return iterator(&m_RawList, m_RawList.InsertBefore(it.m_pItem, value)); } |
| 4944 | |
| 4945 | void clear() { m_RawList.Clear(); } |
| 4946 | void erase(iterator it) { m_RawList.Remove(it.m_pItem); } |
| 4947 | |
| 4948 | private: |
| 4949 | VmaRawList<T> m_RawList; |
| 4950 | }; |
| 4951 | |
| 4952 | #ifndef _VMA_LIST_FUNCTIONS |
| 4953 | template<typename T, typename AllocatorT> |
| 4954 | typename VmaList<T, AllocatorT>::iterator& VmaList<T, AllocatorT>::iterator::operator--() |
| 4955 | { |
| 4956 | if (m_pItem != VMA_NULL) |
| 4957 | { |
| 4958 | m_pItem = m_pItem->pPrev; |
| 4959 | } |
| 4960 | else |
| 4961 | { |
| 4962 | VMA_HEAVY_ASSERT(!m_pList->IsEmpty()); |
| 4963 | m_pItem = m_pList->Back(); |
| 4964 | } |
| 4965 | return *this; |
| 4966 | } |
| 4967 | |
| 4968 | template<typename T, typename AllocatorT> |
| 4969 | typename VmaList<T, AllocatorT>::reverse_iterator& VmaList<T, AllocatorT>::reverse_iterator::operator--() |
| 4970 | { |
| 4971 | if (m_pItem != VMA_NULL) |
| 4972 | { |
| 4973 | m_pItem = m_pItem->pNext; |
| 4974 | } |
| 4975 | else |
| 4976 | { |
| 4977 | VMA_HEAVY_ASSERT(!m_pList->IsEmpty()); |
| 4978 | m_pItem = m_pList->Front(); |
| 4979 | } |
| 4980 | return *this; |
| 4981 | } |
| 4982 | |
| 4983 | template<typename T, typename AllocatorT> |
| 4984 | typename VmaList<T, AllocatorT>::const_iterator& VmaList<T, AllocatorT>::const_iterator::operator--() |
| 4985 | { |
| 4986 | if (m_pItem != VMA_NULL) |
| 4987 | { |
| 4988 | m_pItem = m_pItem->pPrev; |
| 4989 | } |
| 4990 | else |
| 4991 | { |
| 4992 | VMA_HEAVY_ASSERT(!m_pList->IsEmpty()); |
| 4993 | m_pItem = m_pList->Back(); |
| 4994 | } |
| 4995 | return *this; |
| 4996 | } |
| 4997 | |
| 4998 | template<typename T, typename AllocatorT> |
| 4999 | typename VmaList<T, AllocatorT>::const_reverse_iterator& VmaList<T, AllocatorT>::const_reverse_iterator::operator--() |
| 5000 | { |
| 5001 | if (m_pItem != VMA_NULL) |
| 5002 | { |
| 5003 | m_pItem = m_pItem->pNext; |
| 5004 | } |
| 5005 | else |
| 5006 | { |
| 5007 | VMA_HEAVY_ASSERT(!m_pList->IsEmpty()); |
| 5008 | m_pItem = m_pList->Back(); |
| 5009 | } |
| 5010 | return *this; |
| 5011 | } |
| 5012 | #endif // _VMA_LIST_FUNCTIONS |
| 5013 | #endif // _VMA_LIST |
| 5014 | |
| 5015 | #ifndef _VMA_INTRUSIVE_LINKED_LIST |
| 5016 | /* |
| 5017 | Expected interface of ItemTypeTraits: |
| 5018 | struct MyItemTypeTraits |
| 5019 | { |
| 5020 | typedef MyItem ItemType; |
| 5021 | static ItemType* GetPrev(const ItemType* item) { return item->myPrevPtr; } |
| 5022 | static ItemType* GetNext(const ItemType* item) { return item->myNextPtr; } |
| 5023 | static ItemType*& AccessPrev(ItemType* item) { return item->myPrevPtr; } |
| 5024 | static ItemType*& AccessNext(ItemType* item) { return item->myNextPtr; } |
| 5025 | }; |
| 5026 | */ |
| 5027 | template<typename ItemTypeTraits> |
| 5028 | class VmaIntrusiveLinkedList |
| 5029 | { |
| 5030 | public: |
| 5031 | typedef typename ItemTypeTraits::ItemType ItemType; |
| 5032 | static ItemType* GetPrev(const ItemType* item) { return ItemTypeTraits::GetPrev(item); } |
| 5033 | static ItemType* GetNext(const ItemType* item) { return ItemTypeTraits::GetNext(item); } |
| 5034 | |
| 5035 | // Movable, not copyable. |
| 5036 | VmaIntrusiveLinkedList() = default; |
| 5037 | VmaIntrusiveLinkedList(VmaIntrusiveLinkedList && src); |
| 5038 | VmaIntrusiveLinkedList(const VmaIntrusiveLinkedList&) = delete; |
| 5039 | VmaIntrusiveLinkedList& operator=(VmaIntrusiveLinkedList&& src); |
| 5040 | VmaIntrusiveLinkedList& operator=(const VmaIntrusiveLinkedList&) = delete; |
| 5041 | ~VmaIntrusiveLinkedList() { VMA_HEAVY_ASSERT(IsEmpty()); } |
| 5042 | |
| 5043 | size_t GetCount() const { return m_Count; } |
| 5044 | bool IsEmpty() const { return m_Count == 0; } |
| 5045 | ItemType* Front() { return m_Front; } |
| 5046 | ItemType* Back() { return m_Back; } |
| 5047 | const ItemType* Front() const { return m_Front; } |
| 5048 | const ItemType* Back() const { return m_Back; } |
| 5049 | |
| 5050 | void PushBack(ItemType* item); |
| 5051 | void PushFront(ItemType* item); |
| 5052 | ItemType* PopBack(); |
| 5053 | ItemType* PopFront(); |
| 5054 | |
| 5055 | // MyItem can be null - it means PushBack. |
| 5056 | void InsertBefore(ItemType* existingItem, ItemType* newItem); |
| 5057 | // MyItem can be null - it means PushFront. |
| 5058 | void InsertAfter(ItemType* existingItem, ItemType* newItem); |
| 5059 | void Remove(ItemType* item); |
| 5060 | void RemoveAll(); |
| 5061 | |
| 5062 | private: |
| 5063 | ItemType* m_Front = VMA_NULL; |
| 5064 | ItemType* m_Back = VMA_NULL; |
| 5065 | size_t m_Count = 0; |
| 5066 | }; |
| 5067 | |
| 5068 | #ifndef _VMA_INTRUSIVE_LINKED_LIST_FUNCTIONS |
| 5069 | template<typename ItemTypeTraits> |
| 5070 | VmaIntrusiveLinkedList<ItemTypeTraits>::VmaIntrusiveLinkedList(VmaIntrusiveLinkedList&& src) |
| 5071 | : m_Front(src.m_Front), m_Back(src.m_Back), m_Count(src.m_Count) |
| 5072 | { |
| 5073 | src.m_Front = src.m_Back = VMA_NULL; |
| 5074 | src.m_Count = 0; |
| 5075 | } |
| 5076 | |
| 5077 | template<typename ItemTypeTraits> |
| 5078 | VmaIntrusiveLinkedList<ItemTypeTraits>& VmaIntrusiveLinkedList<ItemTypeTraits>::operator=(VmaIntrusiveLinkedList&& src) |
| 5079 | { |
| 5080 | if (&src != this) |
| 5081 | { |
| 5082 | VMA_HEAVY_ASSERT(IsEmpty()); |
| 5083 | m_Front = src.m_Front; |
| 5084 | m_Back = src.m_Back; |
| 5085 | m_Count = src.m_Count; |
| 5086 | src.m_Front = src.m_Back = VMA_NULL; |
| 5087 | src.m_Count = 0; |
| 5088 | } |
| 5089 | return *this; |
| 5090 | } |
| 5091 | |
| 5092 | template<typename ItemTypeTraits> |
| 5093 | void VmaIntrusiveLinkedList<ItemTypeTraits>::PushBack(ItemType* item) |
| 5094 | { |
| 5095 | VMA_HEAVY_ASSERT(ItemTypeTraits::GetPrev(item) == VMA_NULL && ItemTypeTraits::GetNext(item) == VMA_NULL); |
| 5096 | if (IsEmpty()) |
| 5097 | { |
| 5098 | m_Front = item; |
| 5099 | m_Back = item; |
| 5100 | m_Count = 1; |
| 5101 | } |
| 5102 | else |
| 5103 | { |
| 5104 | ItemTypeTraits::AccessPrev(item) = m_Back; |
| 5105 | ItemTypeTraits::AccessNext(m_Back) = item; |
| 5106 | m_Back = item; |
| 5107 | ++m_Count; |
| 5108 | } |
| 5109 | } |
| 5110 | |
| 5111 | template<typename ItemTypeTraits> |
| 5112 | void VmaIntrusiveLinkedList<ItemTypeTraits>::PushFront(ItemType* item) |
| 5113 | { |
| 5114 | VMA_HEAVY_ASSERT(ItemTypeTraits::GetPrev(item) == VMA_NULL && ItemTypeTraits::GetNext(item) == VMA_NULL); |
| 5115 | if (IsEmpty()) |
| 5116 | { |
| 5117 | m_Front = item; |
| 5118 | m_Back = item; |
| 5119 | m_Count = 1; |
| 5120 | } |
| 5121 | else |
| 5122 | { |
| 5123 | ItemTypeTraits::AccessNext(item) = m_Front; |
| 5124 | ItemTypeTraits::AccessPrev(m_Front) = item; |
| 5125 | m_Front = item; |
| 5126 | ++m_Count; |
| 5127 | } |
| 5128 | } |
| 5129 | |
| 5130 | template<typename ItemTypeTraits> |
| 5131 | typename VmaIntrusiveLinkedList<ItemTypeTraits>::ItemType* VmaIntrusiveLinkedList<ItemTypeTraits>::PopBack() |
| 5132 | { |
| 5133 | VMA_HEAVY_ASSERT(m_Count > 0); |
| 5134 | ItemType* const backItem = m_Back; |
| 5135 | ItemType* const prevItem = ItemTypeTraits::GetPrev(backItem); |
| 5136 | if (prevItem != VMA_NULL) |
| 5137 | { |
| 5138 | ItemTypeTraits::AccessNext(prevItem) = VMA_NULL; |
| 5139 | } |
| 5140 | m_Back = prevItem; |
| 5141 | --m_Count; |
| 5142 | ItemTypeTraits::AccessPrev(backItem) = VMA_NULL; |
| 5143 | ItemTypeTraits::AccessNext(backItem) = VMA_NULL; |
| 5144 | return backItem; |
| 5145 | } |
| 5146 | |
| 5147 | template<typename ItemTypeTraits> |
| 5148 | typename VmaIntrusiveLinkedList<ItemTypeTraits>::ItemType* VmaIntrusiveLinkedList<ItemTypeTraits>::PopFront() |
| 5149 | { |
| 5150 | VMA_HEAVY_ASSERT(m_Count > 0); |
| 5151 | ItemType* const frontItem = m_Front; |
| 5152 | ItemType* const nextItem = ItemTypeTraits::GetNext(frontItem); |
| 5153 | if (nextItem != VMA_NULL) |
| 5154 | { |
| 5155 | ItemTypeTraits::AccessPrev(nextItem) = VMA_NULL; |
| 5156 | } |
| 5157 | m_Front = nextItem; |
| 5158 | --m_Count; |
| 5159 | ItemTypeTraits::AccessPrev(frontItem) = VMA_NULL; |
| 5160 | ItemTypeTraits::AccessNext(frontItem) = VMA_NULL; |
| 5161 | return frontItem; |
| 5162 | } |
| 5163 | |
| 5164 | template<typename ItemTypeTraits> |
| 5165 | void VmaIntrusiveLinkedList<ItemTypeTraits>::InsertBefore(ItemType* existingItem, ItemType* newItem) |
| 5166 | { |
| 5167 | VMA_HEAVY_ASSERT(newItem != VMA_NULL && ItemTypeTraits::GetPrev(newItem) == VMA_NULL && ItemTypeTraits::GetNext(newItem) == VMA_NULL); |
| 5168 | if (existingItem != VMA_NULL) |
| 5169 | { |
| 5170 | ItemType* const prevItem = ItemTypeTraits::GetPrev(existingItem); |
| 5171 | ItemTypeTraits::AccessPrev(newItem) = prevItem; |
| 5172 | ItemTypeTraits::AccessNext(newItem) = existingItem; |
| 5173 | ItemTypeTraits::AccessPrev(existingItem) = newItem; |
| 5174 | if (prevItem != VMA_NULL) |
| 5175 | { |
| 5176 | ItemTypeTraits::AccessNext(prevItem) = newItem; |
| 5177 | } |
| 5178 | else |
| 5179 | { |
| 5180 | VMA_HEAVY_ASSERT(m_Front == existingItem); |
| 5181 | m_Front = newItem; |
| 5182 | } |
| 5183 | ++m_Count; |
| 5184 | } |
| 5185 | else |
| 5186 | PushBack(item: newItem); |
| 5187 | } |
| 5188 | |
| 5189 | template<typename ItemTypeTraits> |
| 5190 | void VmaIntrusiveLinkedList<ItemTypeTraits>::InsertAfter(ItemType* existingItem, ItemType* newItem) |
| 5191 | { |
| 5192 | VMA_HEAVY_ASSERT(newItem != VMA_NULL && ItemTypeTraits::GetPrev(newItem) == VMA_NULL && ItemTypeTraits::GetNext(newItem) == VMA_NULL); |
| 5193 | if (existingItem != VMA_NULL) |
| 5194 | { |
| 5195 | ItemType* const nextItem = ItemTypeTraits::GetNext(existingItem); |
| 5196 | ItemTypeTraits::AccessNext(newItem) = nextItem; |
| 5197 | ItemTypeTraits::AccessPrev(newItem) = existingItem; |
| 5198 | ItemTypeTraits::AccessNext(existingItem) = newItem; |
| 5199 | if (nextItem != VMA_NULL) |
| 5200 | { |
| 5201 | ItemTypeTraits::AccessPrev(nextItem) = newItem; |
| 5202 | } |
| 5203 | else |
| 5204 | { |
| 5205 | VMA_HEAVY_ASSERT(m_Back == existingItem); |
| 5206 | m_Back = newItem; |
| 5207 | } |
| 5208 | ++m_Count; |
| 5209 | } |
| 5210 | else |
| 5211 | return PushFront(item: newItem); |
| 5212 | } |
| 5213 | |
| 5214 | template<typename ItemTypeTraits> |
| 5215 | void VmaIntrusiveLinkedList<ItemTypeTraits>::Remove(ItemType* item) |
| 5216 | { |
| 5217 | VMA_HEAVY_ASSERT(item != VMA_NULL && m_Count > 0); |
| 5218 | if (ItemTypeTraits::GetPrev(item) != VMA_NULL) |
| 5219 | { |
| 5220 | ItemTypeTraits::AccessNext(ItemTypeTraits::AccessPrev(item)) = ItemTypeTraits::GetNext(item); |
| 5221 | } |
| 5222 | else |
| 5223 | { |
| 5224 | VMA_HEAVY_ASSERT(m_Front == item); |
| 5225 | m_Front = ItemTypeTraits::GetNext(item); |
| 5226 | } |
| 5227 | |
| 5228 | if (ItemTypeTraits::GetNext(item) != VMA_NULL) |
| 5229 | { |
| 5230 | ItemTypeTraits::AccessPrev(ItemTypeTraits::AccessNext(item)) = ItemTypeTraits::GetPrev(item); |
| 5231 | } |
| 5232 | else |
| 5233 | { |
| 5234 | VMA_HEAVY_ASSERT(m_Back == item); |
| 5235 | m_Back = ItemTypeTraits::GetPrev(item); |
| 5236 | } |
| 5237 | ItemTypeTraits::AccessPrev(item) = VMA_NULL; |
| 5238 | ItemTypeTraits::AccessNext(item) = VMA_NULL; |
| 5239 | --m_Count; |
| 5240 | } |
| 5241 | |
| 5242 | template<typename ItemTypeTraits> |
| 5243 | void VmaIntrusiveLinkedList<ItemTypeTraits>::RemoveAll() |
| 5244 | { |
| 5245 | if (!IsEmpty()) |
| 5246 | { |
| 5247 | ItemType* item = m_Back; |
| 5248 | while (item != VMA_NULL) |
| 5249 | { |
| 5250 | ItemType* const prevItem = ItemTypeTraits::AccessPrev(item); |
| 5251 | ItemTypeTraits::AccessPrev(item) = VMA_NULL; |
| 5252 | ItemTypeTraits::AccessNext(item) = VMA_NULL; |
| 5253 | item = prevItem; |
| 5254 | } |
| 5255 | m_Front = VMA_NULL; |
| 5256 | m_Back = VMA_NULL; |
| 5257 | m_Count = 0; |
| 5258 | } |
| 5259 | } |
| 5260 | #endif // _VMA_INTRUSIVE_LINKED_LIST_FUNCTIONS |
| 5261 | #endif // _VMA_INTRUSIVE_LINKED_LIST |
| 5262 | |
| 5263 | // Unused in this version. |
| 5264 | #if 0 |
| 5265 | |
| 5266 | #ifndef _VMA_PAIR |
| 5267 | template<typename T1, typename T2> |
| 5268 | struct VmaPair |
| 5269 | { |
| 5270 | T1 first; |
| 5271 | T2 second; |
| 5272 | |
| 5273 | VmaPair() : first(), second() {} |
| 5274 | VmaPair(const T1& firstSrc, const T2& secondSrc) : first(firstSrc), second(secondSrc) {} |
| 5275 | }; |
| 5276 | |
| 5277 | template<typename FirstT, typename SecondT> |
| 5278 | struct VmaPairFirstLess |
| 5279 | { |
| 5280 | bool operator()(const VmaPair<FirstT, SecondT>& lhs, const VmaPair<FirstT, SecondT>& rhs) const |
| 5281 | { |
| 5282 | return lhs.first < rhs.first; |
| 5283 | } |
| 5284 | bool operator()(const VmaPair<FirstT, SecondT>& lhs, const FirstT& rhsFirst) const |
| 5285 | { |
| 5286 | return lhs.first < rhsFirst; |
| 5287 | } |
| 5288 | }; |
| 5289 | #endif // _VMA_PAIR |
| 5290 | |
| 5291 | #ifndef _VMA_MAP |
| 5292 | /* Class compatible with subset of interface of std::unordered_map. |
| 5293 | KeyT, ValueT must be POD because they will be stored in VmaVector. |
| 5294 | */ |
| 5295 | template<typename KeyT, typename ValueT> |
| 5296 | class VmaMap |
| 5297 | { |
| 5298 | public: |
| 5299 | typedef VmaPair<KeyT, ValueT> PairType; |
| 5300 | typedef PairType* iterator; |
| 5301 | |
| 5302 | VmaMap(const VmaStlAllocator<PairType>& allocator) : m_Vector(allocator) {} |
| 5303 | |
| 5304 | iterator begin() { return m_Vector.begin(); } |
| 5305 | iterator end() { return m_Vector.end(); } |
| 5306 | size_t size() { return m_Vector.size(); } |
| 5307 | |
| 5308 | void insert(const PairType& pair); |
| 5309 | iterator find(const KeyT& key); |
| 5310 | void erase(iterator it); |
| 5311 | |
| 5312 | private: |
| 5313 | VmaVector< PairType, VmaStlAllocator<PairType>> m_Vector; |
| 5314 | }; |
| 5315 | |
| 5316 | #ifndef _VMA_MAP_FUNCTIONS |
| 5317 | template<typename KeyT, typename ValueT> |
| 5318 | void VmaMap<KeyT, ValueT>::insert(const PairType& pair) |
| 5319 | { |
| 5320 | const size_t indexToInsert = VmaBinaryFindFirstNotLess( |
| 5321 | m_Vector.data(), |
| 5322 | m_Vector.data() + m_Vector.size(), |
| 5323 | pair, |
| 5324 | VmaPairFirstLess<KeyT, ValueT>()) - m_Vector.data(); |
| 5325 | VmaVectorInsert(m_Vector, indexToInsert, pair); |
| 5326 | } |
| 5327 | |
| 5328 | template<typename KeyT, typename ValueT> |
| 5329 | VmaPair<KeyT, ValueT>* VmaMap<KeyT, ValueT>::find(const KeyT& key) |
| 5330 | { |
| 5331 | PairType* it = VmaBinaryFindFirstNotLess( |
| 5332 | m_Vector.data(), |
| 5333 | m_Vector.data() + m_Vector.size(), |
| 5334 | key, |
| 5335 | VmaPairFirstLess<KeyT, ValueT>()); |
| 5336 | if ((it != m_Vector.end()) && (it->first == key)) |
| 5337 | { |
| 5338 | return it; |
| 5339 | } |
| 5340 | else |
| 5341 | { |
| 5342 | return m_Vector.end(); |
| 5343 | } |
| 5344 | } |
| 5345 | |
| 5346 | template<typename KeyT, typename ValueT> |
| 5347 | void VmaMap<KeyT, ValueT>::erase(iterator it) |
| 5348 | { |
| 5349 | VmaVectorRemove(m_Vector, it - m_Vector.begin()); |
| 5350 | } |
| 5351 | #endif // _VMA_MAP_FUNCTIONS |
| 5352 | #endif // _VMA_MAP |
| 5353 | |
| 5354 | #endif // #if 0 |
| 5355 | |
| 5356 | #if !defined(_VMA_STRING_BUILDER) && VMA_STATS_STRING_ENABLED |
| 5357 | class VmaStringBuilder |
| 5358 | { |
| 5359 | public: |
| 5360 | VmaStringBuilder(const VkAllocationCallbacks* allocationCallbacks) : m_Data(VmaStlAllocator<char>(allocationCallbacks)) {} |
| 5361 | ~VmaStringBuilder() = default; |
| 5362 | |
| 5363 | size_t GetLength() const { return m_Data.size(); } |
| 5364 | const char* GetData() const { return m_Data.data(); } |
| 5365 | void AddNewLine() { Add(ch: '\n'); } |
| 5366 | void Add(char ch) { m_Data.push_back(src: ch); } |
| 5367 | |
| 5368 | void Add(const char* pStr); |
| 5369 | void AddNumber(uint32_t num); |
| 5370 | void AddNumber(uint64_t num); |
| 5371 | void AddPointer(const void* ptr); |
| 5372 | |
| 5373 | private: |
| 5374 | VmaVector<char, VmaStlAllocator<char>> m_Data; |
| 5375 | }; |
| 5376 | |
| 5377 | #ifndef _VMA_STRING_BUILDER_FUNCTIONS |
| 5378 | void VmaStringBuilder::Add(const char* pStr) |
| 5379 | { |
| 5380 | const size_t strLen = strlen(s: pStr); |
| 5381 | if (strLen > 0) |
| 5382 | { |
| 5383 | const size_t oldCount = m_Data.size(); |
| 5384 | m_Data.resize(newCount: oldCount + strLen); |
| 5385 | memcpy(dest: m_Data.data() + oldCount, src: pStr, n: strLen); |
| 5386 | } |
| 5387 | } |
| 5388 | |
| 5389 | void VmaStringBuilder::AddNumber(uint32_t num) |
| 5390 | { |
| 5391 | char buf[11]; |
| 5392 | buf[10] = '\0'; |
| 5393 | char* p = &buf[10]; |
| 5394 | do |
| 5395 | { |
| 5396 | *--p = '0' + (num % 10); |
| 5397 | num /= 10; |
| 5398 | } while (num); |
| 5399 | Add(pStr: p); |
| 5400 | } |
| 5401 | |
| 5402 | void VmaStringBuilder::AddNumber(uint64_t num) |
| 5403 | { |
| 5404 | char buf[21]; |
| 5405 | buf[20] = '\0'; |
| 5406 | char* p = &buf[20]; |
| 5407 | do |
| 5408 | { |
| 5409 | *--p = '0' + (num % 10); |
| 5410 | num /= 10; |
| 5411 | } while (num); |
| 5412 | Add(pStr: p); |
| 5413 | } |
| 5414 | |
| 5415 | void VmaStringBuilder::AddPointer(const void* ptr) |
| 5416 | { |
| 5417 | char buf[21]; |
| 5418 | VmaPtrToStr(outStr: buf, strLen: sizeof(buf), ptr); |
| 5419 | Add(pStr: buf); |
| 5420 | } |
| 5421 | #endif //_VMA_STRING_BUILDER_FUNCTIONS |
| 5422 | #endif // _VMA_STRING_BUILDER |
| 5423 | |
| 5424 | #if !defined(_VMA_JSON_WRITER) && VMA_STATS_STRING_ENABLED |
| 5425 | /* |
| 5426 | Allows to conveniently build a correct JSON document to be written to the |
| 5427 | VmaStringBuilder passed to the constructor. |
| 5428 | */ |
| 5429 | class VmaJsonWriter |
| 5430 | { |
| 5431 | VMA_CLASS_NO_COPY(VmaJsonWriter) |
| 5432 | public: |
| 5433 | // sb - string builder to write the document to. Must remain alive for the whole lifetime of this object. |
| 5434 | VmaJsonWriter(const VkAllocationCallbacks* pAllocationCallbacks, VmaStringBuilder& sb); |
| 5435 | ~VmaJsonWriter(); |
| 5436 | |
| 5437 | // Begins object by writing "{". |
| 5438 | // Inside an object, you must call pairs of WriteString and a value, e.g.: |
| 5439 | // j.BeginObject(true); j.WriteString("A"); j.WriteNumber(1); j.WriteString("B"); j.WriteNumber(2); j.EndObject(); |
| 5440 | // Will write: { "A": 1, "B": 2 } |
| 5441 | void BeginObject(bool singleLine = false); |
| 5442 | // Ends object by writing "}". |
| 5443 | void EndObject(); |
| 5444 | |
| 5445 | // Begins array by writing "[". |
| 5446 | // Inside an array, you can write a sequence of any values. |
| 5447 | void BeginArray(bool singleLine = false); |
| 5448 | // Ends array by writing "[". |
| 5449 | void EndArray(); |
| 5450 | |
| 5451 | // Writes a string value inside "". |
| 5452 | // pStr can contain any ANSI characters, including '"', new line etc. - they will be properly escaped. |
| 5453 | void WriteString(const char* pStr); |
| 5454 | |
| 5455 | // Begins writing a string value. |
| 5456 | // Call BeginString, ContinueString, ContinueString, ..., EndString instead of |
| 5457 | // WriteString to conveniently build the string content incrementally, made of |
| 5458 | // parts including numbers. |
| 5459 | void BeginString(const char* pStr = VMA_NULL); |
| 5460 | // Posts next part of an open string. |
| 5461 | void ContinueString(const char* pStr); |
| 5462 | // Posts next part of an open string. The number is converted to decimal characters. |
| 5463 | void ContinueString(uint32_t n); |
| 5464 | void ContinueString(uint64_t n); |
| 5465 | void ContinueString_Size(size_t n); |
| 5466 | // Posts next part of an open string. Pointer value is converted to characters |
| 5467 | // using "%p" formatting - shown as hexadecimal number, e.g.: 000000081276Ad00 |
| 5468 | void ContinueString_Pointer(const void* ptr); |
| 5469 | // Ends writing a string value by writing '"'. |
| 5470 | void EndString(const char* pStr = VMA_NULL); |
| 5471 | |
| 5472 | // Writes a number value. |
| 5473 | void WriteNumber(uint32_t n); |
| 5474 | void WriteNumber(uint64_t n); |
| 5475 | void WriteSize(size_t n); |
| 5476 | // Writes a boolean value - false or true. |
| 5477 | void WriteBool(bool b); |
| 5478 | // Writes a null value. |
| 5479 | void WriteNull(); |
| 5480 | |
| 5481 | private: |
| 5482 | enum COLLECTION_TYPE |
| 5483 | { |
| 5484 | COLLECTION_TYPE_OBJECT, |
| 5485 | COLLECTION_TYPE_ARRAY, |
| 5486 | }; |
| 5487 | struct StackItem |
| 5488 | { |
| 5489 | COLLECTION_TYPE type; |
| 5490 | uint32_t valueCount; |
| 5491 | bool singleLineMode; |
| 5492 | }; |
| 5493 | |
| 5494 | static const char* const INDENT; |
| 5495 | |
| 5496 | VmaStringBuilder& m_SB; |
| 5497 | VmaVector< StackItem, VmaStlAllocator<StackItem> > m_Stack; |
| 5498 | bool m_InsideString; |
| 5499 | |
| 5500 | // Write size_t for less than 64bits |
| 5501 | void WriteSize(size_t n, std::integral_constant<bool, false>) { m_SB.AddNumber(num: static_cast<uint32_t>(n)); } |
| 5502 | // Write size_t for 64bits |
| 5503 | void WriteSize(size_t n, std::integral_constant<bool, true>) { m_SB.AddNumber(num: static_cast<uint64_t>(n)); } |
| 5504 | |
| 5505 | void BeginValue(bool isString); |
| 5506 | void WriteIndent(bool oneLess = false); |
| 5507 | }; |
| 5508 | const char* const VmaJsonWriter::INDENT = " " ; |
| 5509 | |
| 5510 | #ifndef _VMA_JSON_WRITER_FUNCTIONS |
| 5511 | VmaJsonWriter::VmaJsonWriter(const VkAllocationCallbacks* pAllocationCallbacks, VmaStringBuilder& sb) |
| 5512 | : m_SB(sb), |
| 5513 | m_Stack(VmaStlAllocator<StackItem>(pAllocationCallbacks)), |
| 5514 | m_InsideString(false) {} |
| 5515 | |
| 5516 | VmaJsonWriter::~VmaJsonWriter() |
| 5517 | { |
| 5518 | VMA_ASSERT(!m_InsideString); |
| 5519 | VMA_ASSERT(m_Stack.empty()); |
| 5520 | } |
| 5521 | |
| 5522 | void VmaJsonWriter::BeginObject(bool singleLine) |
| 5523 | { |
| 5524 | VMA_ASSERT(!m_InsideString); |
| 5525 | |
| 5526 | BeginValue(isString: false); |
| 5527 | m_SB.Add(ch: '{'); |
| 5528 | |
| 5529 | StackItem item; |
| 5530 | item.type = COLLECTION_TYPE_OBJECT; |
| 5531 | item.valueCount = 0; |
| 5532 | item.singleLineMode = singleLine; |
| 5533 | m_Stack.push_back(src: item); |
| 5534 | } |
| 5535 | |
| 5536 | void VmaJsonWriter::EndObject() |
| 5537 | { |
| 5538 | VMA_ASSERT(!m_InsideString); |
| 5539 | |
| 5540 | WriteIndent(oneLess: true); |
| 5541 | m_SB.Add(ch: '}'); |
| 5542 | |
| 5543 | VMA_ASSERT(!m_Stack.empty() && m_Stack.back().type == COLLECTION_TYPE_OBJECT); |
| 5544 | m_Stack.pop_back(); |
| 5545 | } |
| 5546 | |
| 5547 | void VmaJsonWriter::BeginArray(bool singleLine) |
| 5548 | { |
| 5549 | VMA_ASSERT(!m_InsideString); |
| 5550 | |
| 5551 | BeginValue(isString: false); |
| 5552 | m_SB.Add(ch: '['); |
| 5553 | |
| 5554 | StackItem item; |
| 5555 | item.type = COLLECTION_TYPE_ARRAY; |
| 5556 | item.valueCount = 0; |
| 5557 | item.singleLineMode = singleLine; |
| 5558 | m_Stack.push_back(src: item); |
| 5559 | } |
| 5560 | |
| 5561 | void VmaJsonWriter::EndArray() |
| 5562 | { |
| 5563 | VMA_ASSERT(!m_InsideString); |
| 5564 | |
| 5565 | WriteIndent(oneLess: true); |
| 5566 | m_SB.Add(ch: ']'); |
| 5567 | |
| 5568 | VMA_ASSERT(!m_Stack.empty() && m_Stack.back().type == COLLECTION_TYPE_ARRAY); |
| 5569 | m_Stack.pop_back(); |
| 5570 | } |
| 5571 | |
| 5572 | void VmaJsonWriter::WriteString(const char* pStr) |
| 5573 | { |
| 5574 | BeginString(pStr); |
| 5575 | EndString(); |
| 5576 | } |
| 5577 | |
| 5578 | void VmaJsonWriter::BeginString(const char* pStr) |
| 5579 | { |
| 5580 | VMA_ASSERT(!m_InsideString); |
| 5581 | |
| 5582 | BeginValue(isString: true); |
| 5583 | m_SB.Add(ch: '"'); |
| 5584 | m_InsideString = true; |
| 5585 | if (pStr != VMA_NULL && pStr[0] != '\0') |
| 5586 | { |
| 5587 | ContinueString(pStr); |
| 5588 | } |
| 5589 | } |
| 5590 | |
| 5591 | void VmaJsonWriter::ContinueString(const char* pStr) |
| 5592 | { |
| 5593 | VMA_ASSERT(m_InsideString); |
| 5594 | |
| 5595 | const size_t strLen = strlen(s: pStr); |
| 5596 | for (size_t i = 0; i < strLen; ++i) |
| 5597 | { |
| 5598 | char ch = pStr[i]; |
| 5599 | if (ch == '\\') |
| 5600 | { |
| 5601 | m_SB.Add(pStr: "\\\\" ); |
| 5602 | } |
| 5603 | else if (ch == '"') |
| 5604 | { |
| 5605 | m_SB.Add(pStr: "\\\"" ); |
| 5606 | } |
| 5607 | else if (ch >= 32) |
| 5608 | { |
| 5609 | m_SB.Add(ch); |
| 5610 | } |
| 5611 | else switch (ch) |
| 5612 | { |
| 5613 | case '\b': |
| 5614 | m_SB.Add(pStr: "\\b" ); |
| 5615 | break; |
| 5616 | case '\f': |
| 5617 | m_SB.Add(pStr: "\\f" ); |
| 5618 | break; |
| 5619 | case '\n': |
| 5620 | m_SB.Add(pStr: "\\n" ); |
| 5621 | break; |
| 5622 | case '\r': |
| 5623 | m_SB.Add(pStr: "\\r" ); |
| 5624 | break; |
| 5625 | case '\t': |
| 5626 | m_SB.Add(pStr: "\\t" ); |
| 5627 | break; |
| 5628 | default: |
| 5629 | VMA_ASSERT(0 && "Character not currently supported." ); |
| 5630 | break; |
| 5631 | } |
| 5632 | } |
| 5633 | } |
| 5634 | |
| 5635 | void VmaJsonWriter::ContinueString(uint32_t n) |
| 5636 | { |
| 5637 | VMA_ASSERT(m_InsideString); |
| 5638 | m_SB.AddNumber(num: n); |
| 5639 | } |
| 5640 | |
| 5641 | void VmaJsonWriter::ContinueString(uint64_t n) |
| 5642 | { |
| 5643 | VMA_ASSERT(m_InsideString); |
| 5644 | m_SB.AddNumber(num: n); |
| 5645 | } |
| 5646 | |
| 5647 | void VmaJsonWriter::ContinueString_Size(size_t n) |
| 5648 | { |
| 5649 | VMA_ASSERT(m_InsideString); |
| 5650 | // Fix for AppleClang incorrect type casting |
| 5651 | // TODO: Change to if constexpr when C++17 used as minimal standard |
| 5652 | WriteSize(n, std::is_same<size_t, uint64_t>{}); |
| 5653 | } |
| 5654 | |
| 5655 | void VmaJsonWriter::ContinueString_Pointer(const void* ptr) |
| 5656 | { |
| 5657 | VMA_ASSERT(m_InsideString); |
| 5658 | m_SB.AddPointer(ptr); |
| 5659 | } |
| 5660 | |
| 5661 | void VmaJsonWriter::EndString(const char* pStr) |
| 5662 | { |
| 5663 | VMA_ASSERT(m_InsideString); |
| 5664 | if (pStr != VMA_NULL && pStr[0] != '\0') |
| 5665 | { |
| 5666 | ContinueString(pStr); |
| 5667 | } |
| 5668 | m_SB.Add(ch: '"'); |
| 5669 | m_InsideString = false; |
| 5670 | } |
| 5671 | |
| 5672 | void VmaJsonWriter::WriteNumber(uint32_t n) |
| 5673 | { |
| 5674 | VMA_ASSERT(!m_InsideString); |
| 5675 | BeginValue(isString: false); |
| 5676 | m_SB.AddNumber(num: n); |
| 5677 | } |
| 5678 | |
| 5679 | void VmaJsonWriter::WriteNumber(uint64_t n) |
| 5680 | { |
| 5681 | VMA_ASSERT(!m_InsideString); |
| 5682 | BeginValue(isString: false); |
| 5683 | m_SB.AddNumber(num: n); |
| 5684 | } |
| 5685 | |
| 5686 | void VmaJsonWriter::WriteSize(size_t n) |
| 5687 | { |
| 5688 | VMA_ASSERT(!m_InsideString); |
| 5689 | BeginValue(isString: false); |
| 5690 | // Fix for AppleClang incorrect type casting |
| 5691 | // TODO: Change to if constexpr when C++17 used as minimal standard |
| 5692 | WriteSize(n, std::is_same<size_t, uint64_t>{}); |
| 5693 | } |
| 5694 | |
| 5695 | void VmaJsonWriter::WriteBool(bool b) |
| 5696 | { |
| 5697 | VMA_ASSERT(!m_InsideString); |
| 5698 | BeginValue(isString: false); |
| 5699 | m_SB.Add(pStr: b ? "true" : "false" ); |
| 5700 | } |
| 5701 | |
| 5702 | void VmaJsonWriter::WriteNull() |
| 5703 | { |
| 5704 | VMA_ASSERT(!m_InsideString); |
| 5705 | BeginValue(isString: false); |
| 5706 | m_SB.Add(pStr: "null" ); |
| 5707 | } |
| 5708 | |
| 5709 | void VmaJsonWriter::BeginValue(bool isString) |
| 5710 | { |
| 5711 | if (!m_Stack.empty()) |
| 5712 | { |
| 5713 | StackItem& currItem = m_Stack.back(); |
| 5714 | if (currItem.type == COLLECTION_TYPE_OBJECT && |
| 5715 | currItem.valueCount % 2 == 0) |
| 5716 | { |
| 5717 | VMA_ASSERT(isString); |
| 5718 | } |
| 5719 | |
| 5720 | if (currItem.type == COLLECTION_TYPE_OBJECT && |
| 5721 | currItem.valueCount % 2 != 0) |
| 5722 | { |
| 5723 | m_SB.Add(pStr: ": " ); |
| 5724 | } |
| 5725 | else if (currItem.valueCount > 0) |
| 5726 | { |
| 5727 | m_SB.Add(pStr: ", " ); |
| 5728 | WriteIndent(); |
| 5729 | } |
| 5730 | else |
| 5731 | { |
| 5732 | WriteIndent(); |
| 5733 | } |
| 5734 | ++currItem.valueCount; |
| 5735 | } |
| 5736 | } |
| 5737 | |
| 5738 | void VmaJsonWriter::WriteIndent(bool oneLess) |
| 5739 | { |
| 5740 | if (!m_Stack.empty() && !m_Stack.back().singleLineMode) |
| 5741 | { |
| 5742 | m_SB.AddNewLine(); |
| 5743 | |
| 5744 | size_t count = m_Stack.size(); |
| 5745 | if (count > 0 && oneLess) |
| 5746 | { |
| 5747 | --count; |
| 5748 | } |
| 5749 | for (size_t i = 0; i < count; ++i) |
| 5750 | { |
| 5751 | m_SB.Add(pStr: INDENT); |
| 5752 | } |
| 5753 | } |
| 5754 | } |
| 5755 | #endif // _VMA_JSON_WRITER_FUNCTIONS |
| 5756 | |
| 5757 | static void VmaPrintDetailedStatistics(VmaJsonWriter& json, const VmaDetailedStatistics& stat) |
| 5758 | { |
| 5759 | json.BeginObject(); |
| 5760 | |
| 5761 | json.WriteString(pStr: "BlockCount" ); |
| 5762 | json.WriteNumber(n: stat.statistics.blockCount); |
| 5763 | json.WriteString(pStr: "BlockBytes" ); |
| 5764 | json.WriteNumber(n: stat.statistics.blockBytes); |
| 5765 | json.WriteString(pStr: "AllocationCount" ); |
| 5766 | json.WriteNumber(n: stat.statistics.allocationCount); |
| 5767 | json.WriteString(pStr: "AllocationBytes" ); |
| 5768 | json.WriteNumber(n: stat.statistics.allocationBytes); |
| 5769 | json.WriteString(pStr: "UnusedRangeCount" ); |
| 5770 | json.WriteNumber(n: stat.unusedRangeCount); |
| 5771 | |
| 5772 | if (stat.statistics.allocationCount > 1) |
| 5773 | { |
| 5774 | json.WriteString(pStr: "AllocationSizeMin" ); |
| 5775 | json.WriteNumber(n: stat.allocationSizeMin); |
| 5776 | json.WriteString(pStr: "AllocationSizeMax" ); |
| 5777 | json.WriteNumber(n: stat.allocationSizeMax); |
| 5778 | } |
| 5779 | if (stat.unusedRangeCount > 1) |
| 5780 | { |
| 5781 | json.WriteString(pStr: "UnusedRangeSizeMin" ); |
| 5782 | json.WriteNumber(n: stat.unusedRangeSizeMin); |
| 5783 | json.WriteString(pStr: "UnusedRangeSizeMax" ); |
| 5784 | json.WriteNumber(n: stat.unusedRangeSizeMax); |
| 5785 | } |
| 5786 | json.EndObject(); |
| 5787 | } |
| 5788 | #endif // _VMA_JSON_WRITER |
| 5789 | |
| 5790 | #ifndef _VMA_MAPPING_HYSTERESIS |
| 5791 | |
| 5792 | class VmaMappingHysteresis |
| 5793 | { |
| 5794 | VMA_CLASS_NO_COPY(VmaMappingHysteresis) |
| 5795 | public: |
| 5796 | VmaMappingHysteresis() = default; |
| 5797 | |
| 5798 | uint32_t () const { return m_ExtraMapping; } |
| 5799 | |
| 5800 | // Call when Map was called. |
| 5801 | // Returns true if switched to extra +1 mapping reference count. |
| 5802 | bool PostMap() |
| 5803 | { |
| 5804 | #if VMA_MAPPING_HYSTERESIS_ENABLED |
| 5805 | if(m_ExtraMapping == 0) |
| 5806 | { |
| 5807 | ++m_MajorCounter; |
| 5808 | if(m_MajorCounter >= COUNTER_MIN_EXTRA_MAPPING) |
| 5809 | { |
| 5810 | m_ExtraMapping = 1; |
| 5811 | m_MajorCounter = 0; |
| 5812 | m_MinorCounter = 0; |
| 5813 | return true; |
| 5814 | } |
| 5815 | } |
| 5816 | else // m_ExtraMapping == 1 |
| 5817 | PostMinorCounter(); |
| 5818 | #endif // #if VMA_MAPPING_HYSTERESIS_ENABLED |
| 5819 | return false; |
| 5820 | } |
| 5821 | |
| 5822 | // Call when Unmap was called. |
| 5823 | void PostUnmap() |
| 5824 | { |
| 5825 | #if VMA_MAPPING_HYSTERESIS_ENABLED |
| 5826 | if(m_ExtraMapping == 0) |
| 5827 | ++m_MajorCounter; |
| 5828 | else // m_ExtraMapping == 1 |
| 5829 | PostMinorCounter(); |
| 5830 | #endif // #if VMA_MAPPING_HYSTERESIS_ENABLED |
| 5831 | } |
| 5832 | |
| 5833 | // Call when allocation was made from the memory block. |
| 5834 | void PostAlloc() |
| 5835 | { |
| 5836 | #if VMA_MAPPING_HYSTERESIS_ENABLED |
| 5837 | if(m_ExtraMapping == 1) |
| 5838 | ++m_MajorCounter; |
| 5839 | else // m_ExtraMapping == 0 |
| 5840 | PostMinorCounter(); |
| 5841 | #endif // #if VMA_MAPPING_HYSTERESIS_ENABLED |
| 5842 | } |
| 5843 | |
| 5844 | // Call when allocation was freed from the memory block. |
| 5845 | // Returns true if switched to extra -1 mapping reference count. |
| 5846 | bool PostFree() |
| 5847 | { |
| 5848 | #if VMA_MAPPING_HYSTERESIS_ENABLED |
| 5849 | if(m_ExtraMapping == 1) |
| 5850 | { |
| 5851 | ++m_MajorCounter; |
| 5852 | if(m_MajorCounter >= COUNTER_MIN_EXTRA_MAPPING && |
| 5853 | m_MajorCounter > m_MinorCounter + 1) |
| 5854 | { |
| 5855 | m_ExtraMapping = 0; |
| 5856 | m_MajorCounter = 0; |
| 5857 | m_MinorCounter = 0; |
| 5858 | return true; |
| 5859 | } |
| 5860 | } |
| 5861 | else // m_ExtraMapping == 0 |
| 5862 | PostMinorCounter(); |
| 5863 | #endif // #if VMA_MAPPING_HYSTERESIS_ENABLED |
| 5864 | return false; |
| 5865 | } |
| 5866 | |
| 5867 | private: |
| 5868 | static const int32_t = 7; |
| 5869 | |
| 5870 | uint32_t m_MinorCounter = 0; |
| 5871 | uint32_t m_MajorCounter = 0; |
| 5872 | uint32_t = 0; // 0 or 1. |
| 5873 | |
| 5874 | void PostMinorCounter() |
| 5875 | { |
| 5876 | if(m_MinorCounter < m_MajorCounter) |
| 5877 | { |
| 5878 | ++m_MinorCounter; |
| 5879 | } |
| 5880 | else if(m_MajorCounter > 0) |
| 5881 | { |
| 5882 | --m_MajorCounter; |
| 5883 | --m_MinorCounter; |
| 5884 | } |
| 5885 | } |
| 5886 | }; |
| 5887 | |
| 5888 | #endif // _VMA_MAPPING_HYSTERESIS |
| 5889 | |
| 5890 | #ifndef _VMA_DEVICE_MEMORY_BLOCK |
| 5891 | /* |
| 5892 | Represents a single block of device memory (`VkDeviceMemory`) with all the |
| 5893 | data about its regions (aka suballocations, #VmaAllocation), assigned and free. |
| 5894 | |
| 5895 | Thread-safety: |
| 5896 | - Access to m_pMetadata must be externally synchronized. |
| 5897 | - Map, Unmap, Bind* are synchronized internally. |
| 5898 | */ |
| 5899 | class VmaDeviceMemoryBlock |
| 5900 | { |
| 5901 | VMA_CLASS_NO_COPY(VmaDeviceMemoryBlock) |
| 5902 | public: |
| 5903 | VmaBlockMetadata* m_pMetadata; |
| 5904 | |
| 5905 | VmaDeviceMemoryBlock(VmaAllocator hAllocator); |
| 5906 | ~VmaDeviceMemoryBlock(); |
| 5907 | |
| 5908 | // Always call after construction. |
| 5909 | void Init( |
| 5910 | VmaAllocator hAllocator, |
| 5911 | VmaPool hParentPool, |
| 5912 | uint32_t newMemoryTypeIndex, |
| 5913 | VkDeviceMemory newMemory, |
| 5914 | VkDeviceSize newSize, |
| 5915 | uint32_t id, |
| 5916 | uint32_t algorithm, |
| 5917 | VkDeviceSize bufferImageGranularity); |
| 5918 | // Always call before destruction. |
| 5919 | void Destroy(VmaAllocator allocator); |
| 5920 | |
| 5921 | VmaPool GetParentPool() const { return m_hParentPool; } |
| 5922 | VkDeviceMemory GetDeviceMemory() const { return m_hMemory; } |
| 5923 | uint32_t GetMemoryTypeIndex() const { return m_MemoryTypeIndex; } |
| 5924 | uint32_t GetId() const { return m_Id; } |
| 5925 | void* GetMappedData() const { return m_pMappedData; } |
| 5926 | uint32_t GetMapRefCount() const { return m_MapCount; } |
| 5927 | |
| 5928 | // Call when allocation/free was made from m_pMetadata. |
| 5929 | // Used for m_MappingHysteresis. |
| 5930 | void PostAlloc() { m_MappingHysteresis.PostAlloc(); } |
| 5931 | void PostFree(VmaAllocator hAllocator); |
| 5932 | |
| 5933 | // Validates all data structures inside this object. If not valid, returns false. |
| 5934 | bool Validate() const; |
| 5935 | VkResult CheckCorruption(VmaAllocator hAllocator); |
| 5936 | |
| 5937 | // ppData can be null. |
| 5938 | VkResult Map(VmaAllocator hAllocator, uint32_t count, void** ppData); |
| 5939 | void Unmap(VmaAllocator hAllocator, uint32_t count); |
| 5940 | |
| 5941 | VkResult WriteMagicValueAfterAllocation(VmaAllocator hAllocator, VkDeviceSize allocOffset, VkDeviceSize allocSize); |
| 5942 | VkResult ValidateMagicValueAfterAllocation(VmaAllocator hAllocator, VkDeviceSize allocOffset, VkDeviceSize allocSize); |
| 5943 | |
| 5944 | VkResult BindBufferMemory( |
| 5945 | const VmaAllocator hAllocator, |
| 5946 | const VmaAllocation hAllocation, |
| 5947 | VkDeviceSize allocationLocalOffset, |
| 5948 | VkBuffer hBuffer, |
| 5949 | const void* pNext); |
| 5950 | VkResult BindImageMemory( |
| 5951 | const VmaAllocator hAllocator, |
| 5952 | const VmaAllocation hAllocation, |
| 5953 | VkDeviceSize allocationLocalOffset, |
| 5954 | VkImage hImage, |
| 5955 | const void* pNext); |
| 5956 | |
| 5957 | private: |
| 5958 | VmaPool m_hParentPool; // VK_NULL_HANDLE if not belongs to custom pool. |
| 5959 | uint32_t m_MemoryTypeIndex; |
| 5960 | uint32_t m_Id; |
| 5961 | VkDeviceMemory m_hMemory; |
| 5962 | |
| 5963 | /* |
| 5964 | Protects access to m_hMemory so it is not used by multiple threads simultaneously, e.g. vkMapMemory, vkBindBufferMemory. |
| 5965 | Also protects m_MapCount, m_pMappedData. |
| 5966 | Allocations, deallocations, any change in m_pMetadata is protected by parent's VmaBlockVector::m_Mutex. |
| 5967 | */ |
| 5968 | VMA_MUTEX m_MapAndBindMutex; |
| 5969 | VmaMappingHysteresis m_MappingHysteresis; |
| 5970 | uint32_t m_MapCount; |
| 5971 | void* m_pMappedData; |
| 5972 | }; |
| 5973 | #endif // _VMA_DEVICE_MEMORY_BLOCK |
| 5974 | |
| 5975 | #ifndef _VMA_ALLOCATION_T |
| 5976 | struct VmaAllocation_T |
| 5977 | { |
| 5978 | friend struct VmaDedicatedAllocationListItemTraits; |
| 5979 | |
| 5980 | enum FLAGS |
| 5981 | { |
| 5982 | FLAG_PERSISTENT_MAP = 0x01, |
| 5983 | FLAG_MAPPING_ALLOWED = 0x02, |
| 5984 | }; |
| 5985 | |
| 5986 | public: |
| 5987 | enum ALLOCATION_TYPE |
| 5988 | { |
| 5989 | ALLOCATION_TYPE_NONE, |
| 5990 | ALLOCATION_TYPE_BLOCK, |
| 5991 | ALLOCATION_TYPE_DEDICATED, |
| 5992 | }; |
| 5993 | |
| 5994 | // This struct is allocated using VmaPoolAllocator. |
| 5995 | VmaAllocation_T(bool mappingAllowed); |
| 5996 | ~VmaAllocation_T(); |
| 5997 | |
| 5998 | void InitBlockAllocation( |
| 5999 | VmaDeviceMemoryBlock* block, |
| 6000 | VmaAllocHandle allocHandle, |
| 6001 | VkDeviceSize alignment, |
| 6002 | VkDeviceSize size, |
| 6003 | uint32_t memoryTypeIndex, |
| 6004 | VmaSuballocationType suballocationType, |
| 6005 | bool mapped); |
| 6006 | // pMappedData not null means allocation is created with MAPPED flag. |
| 6007 | void InitDedicatedAllocation( |
| 6008 | VmaPool hParentPool, |
| 6009 | uint32_t memoryTypeIndex, |
| 6010 | VkDeviceMemory hMemory, |
| 6011 | VmaSuballocationType suballocationType, |
| 6012 | void* pMappedData, |
| 6013 | VkDeviceSize size); |
| 6014 | |
| 6015 | ALLOCATION_TYPE GetType() const { return (ALLOCATION_TYPE)m_Type; } |
| 6016 | VkDeviceSize GetAlignment() const { return m_Alignment; } |
| 6017 | VkDeviceSize GetSize() const { return m_Size; } |
| 6018 | void* GetUserData() const { return m_pUserData; } |
| 6019 | const char* GetName() const { return m_pName; } |
| 6020 | VmaSuballocationType GetSuballocationType() const { return (VmaSuballocationType)m_SuballocationType; } |
| 6021 | |
| 6022 | VmaDeviceMemoryBlock* GetBlock() const { VMA_ASSERT(m_Type == ALLOCATION_TYPE_BLOCK); return m_BlockAllocation.m_Block; } |
| 6023 | uint32_t GetMemoryTypeIndex() const { return m_MemoryTypeIndex; } |
| 6024 | bool IsPersistentMap() const { return (m_Flags & FLAG_PERSISTENT_MAP) != 0; } |
| 6025 | bool IsMappingAllowed() const { return (m_Flags & FLAG_MAPPING_ALLOWED) != 0; } |
| 6026 | |
| 6027 | void SetUserData(VmaAllocator hAllocator, void* pUserData) { m_pUserData = pUserData; } |
| 6028 | void SetName(VmaAllocator hAllocator, const char* pName); |
| 6029 | void FreeName(VmaAllocator hAllocator); |
| 6030 | uint8_t SwapBlockAllocation(VmaAllocator hAllocator, VmaAllocation allocation); |
| 6031 | VmaAllocHandle GetAllocHandle() const; |
| 6032 | VkDeviceSize GetOffset() const; |
| 6033 | VmaPool GetParentPool() const; |
| 6034 | VkDeviceMemory GetMemory() const; |
| 6035 | void* GetMappedData() const; |
| 6036 | |
| 6037 | void BlockAllocMap(); |
| 6038 | void BlockAllocUnmap(); |
| 6039 | VkResult DedicatedAllocMap(VmaAllocator hAllocator, void** ppData); |
| 6040 | void DedicatedAllocUnmap(VmaAllocator hAllocator); |
| 6041 | |
| 6042 | #if VMA_STATS_STRING_ENABLED |
| 6043 | uint32_t GetBufferImageUsage() const { return m_BufferImageUsage; } |
| 6044 | |
| 6045 | void InitBufferImageUsage(uint32_t bufferImageUsage); |
| 6046 | void PrintParameters(class VmaJsonWriter& json) const; |
| 6047 | #endif |
| 6048 | |
| 6049 | private: |
| 6050 | // Allocation out of VmaDeviceMemoryBlock. |
| 6051 | struct BlockAllocation |
| 6052 | { |
| 6053 | VmaDeviceMemoryBlock* m_Block; |
| 6054 | VmaAllocHandle m_AllocHandle; |
| 6055 | }; |
| 6056 | // Allocation for an object that has its own private VkDeviceMemory. |
| 6057 | struct DedicatedAllocation |
| 6058 | { |
| 6059 | VmaPool m_hParentPool; // VK_NULL_HANDLE if not belongs to custom pool. |
| 6060 | VkDeviceMemory m_hMemory; |
| 6061 | void* m_pMappedData; // Not null means memory is mapped. |
| 6062 | VmaAllocation_T* m_Prev; |
| 6063 | VmaAllocation_T* m_Next; |
| 6064 | }; |
| 6065 | union |
| 6066 | { |
| 6067 | // Allocation out of VmaDeviceMemoryBlock. |
| 6068 | BlockAllocation m_BlockAllocation; |
| 6069 | // Allocation for an object that has its own private VkDeviceMemory. |
| 6070 | DedicatedAllocation m_DedicatedAllocation; |
| 6071 | }; |
| 6072 | |
| 6073 | VkDeviceSize m_Alignment; |
| 6074 | VkDeviceSize m_Size; |
| 6075 | void* m_pUserData; |
| 6076 | char* m_pName; |
| 6077 | uint32_t m_MemoryTypeIndex; |
| 6078 | uint8_t m_Type; // ALLOCATION_TYPE |
| 6079 | uint8_t m_SuballocationType; // VmaSuballocationType |
| 6080 | // Reference counter for vmaMapMemory()/vmaUnmapMemory(). |
| 6081 | uint8_t m_MapCount; |
| 6082 | uint8_t m_Flags; // enum FLAGS |
| 6083 | #if VMA_STATS_STRING_ENABLED |
| 6084 | uint32_t m_BufferImageUsage; // 0 if unknown. |
| 6085 | #endif |
| 6086 | }; |
| 6087 | #endif // _VMA_ALLOCATION_T |
| 6088 | |
| 6089 | #ifndef _VMA_DEDICATED_ALLOCATION_LIST_ITEM_TRAITS |
| 6090 | struct VmaDedicatedAllocationListItemTraits |
| 6091 | { |
| 6092 | typedef VmaAllocation_T ItemType; |
| 6093 | |
| 6094 | static ItemType* GetPrev(const ItemType* item) |
| 6095 | { |
| 6096 | VMA_HEAVY_ASSERT(item->GetType() == VmaAllocation_T::ALLOCATION_TYPE_DEDICATED); |
| 6097 | return item->m_DedicatedAllocation.m_Prev; |
| 6098 | } |
| 6099 | static ItemType* GetNext(const ItemType* item) |
| 6100 | { |
| 6101 | VMA_HEAVY_ASSERT(item->GetType() == VmaAllocation_T::ALLOCATION_TYPE_DEDICATED); |
| 6102 | return item->m_DedicatedAllocation.m_Next; |
| 6103 | } |
| 6104 | static ItemType*& AccessPrev(ItemType* item) |
| 6105 | { |
| 6106 | VMA_HEAVY_ASSERT(item->GetType() == VmaAllocation_T::ALLOCATION_TYPE_DEDICATED); |
| 6107 | return item->m_DedicatedAllocation.m_Prev; |
| 6108 | } |
| 6109 | static ItemType*& AccessNext(ItemType* item) |
| 6110 | { |
| 6111 | VMA_HEAVY_ASSERT(item->GetType() == VmaAllocation_T::ALLOCATION_TYPE_DEDICATED); |
| 6112 | return item->m_DedicatedAllocation.m_Next; |
| 6113 | } |
| 6114 | }; |
| 6115 | #endif // _VMA_DEDICATED_ALLOCATION_LIST_ITEM_TRAITS |
| 6116 | |
| 6117 | #ifndef _VMA_DEDICATED_ALLOCATION_LIST |
| 6118 | /* |
| 6119 | Stores linked list of VmaAllocation_T objects. |
| 6120 | Thread-safe, synchronized internally. |
| 6121 | */ |
| 6122 | class VmaDedicatedAllocationList |
| 6123 | { |
| 6124 | public: |
| 6125 | VmaDedicatedAllocationList() {} |
| 6126 | ~VmaDedicatedAllocationList(); |
| 6127 | |
| 6128 | void Init(bool useMutex) { m_UseMutex = useMutex; } |
| 6129 | bool Validate(); |
| 6130 | |
| 6131 | void AddDetailedStatistics(VmaDetailedStatistics& inoutStats); |
| 6132 | void AddStatistics(VmaStatistics& inoutStats); |
| 6133 | #if VMA_STATS_STRING_ENABLED |
| 6134 | // Writes JSON array with the list of allocations. |
| 6135 | void BuildStatsString(VmaJsonWriter& json); |
| 6136 | #endif |
| 6137 | |
| 6138 | bool IsEmpty(); |
| 6139 | void Register(VmaAllocation alloc); |
| 6140 | void Unregister(VmaAllocation alloc); |
| 6141 | |
| 6142 | private: |
| 6143 | typedef VmaIntrusiveLinkedList<VmaDedicatedAllocationListItemTraits> DedicatedAllocationLinkedList; |
| 6144 | |
| 6145 | bool m_UseMutex = true; |
| 6146 | VMA_RW_MUTEX m_Mutex; |
| 6147 | DedicatedAllocationLinkedList m_AllocationList; |
| 6148 | }; |
| 6149 | |
| 6150 | #ifndef _VMA_DEDICATED_ALLOCATION_LIST_FUNCTIONS |
| 6151 | |
| 6152 | VmaDedicatedAllocationList::~VmaDedicatedAllocationList() |
| 6153 | { |
| 6154 | VMA_HEAVY_ASSERT(Validate()); |
| 6155 | |
| 6156 | if (!m_AllocationList.IsEmpty()) |
| 6157 | { |
| 6158 | VMA_ASSERT(false && "Unfreed dedicated allocations found!" ); |
| 6159 | } |
| 6160 | } |
| 6161 | |
| 6162 | bool VmaDedicatedAllocationList::Validate() |
| 6163 | { |
| 6164 | const size_t declaredCount = m_AllocationList.GetCount(); |
| 6165 | size_t actualCount = 0; |
| 6166 | VmaMutexLockRead lock(m_Mutex, m_UseMutex); |
| 6167 | for (VmaAllocation alloc = m_AllocationList.Front(); |
| 6168 | alloc != VMA_NULL; alloc = m_AllocationList.GetNext(item: alloc)) |
| 6169 | { |
| 6170 | ++actualCount; |
| 6171 | } |
| 6172 | VMA_VALIDATE(actualCount == declaredCount); |
| 6173 | |
| 6174 | return true; |
| 6175 | } |
| 6176 | |
| 6177 | void VmaDedicatedAllocationList::AddDetailedStatistics(VmaDetailedStatistics& inoutStats) |
| 6178 | { |
| 6179 | for(auto* item = m_AllocationList.Front(); item != nullptr; item = DedicatedAllocationLinkedList::GetNext(item)) |
| 6180 | { |
| 6181 | const VkDeviceSize size = item->GetSize(); |
| 6182 | inoutStats.statistics.blockCount++; |
| 6183 | inoutStats.statistics.blockBytes += size; |
| 6184 | VmaAddDetailedStatisticsAllocation(inoutStats, size: item->GetSize()); |
| 6185 | } |
| 6186 | } |
| 6187 | |
| 6188 | void VmaDedicatedAllocationList::AddStatistics(VmaStatistics& inoutStats) |
| 6189 | { |
| 6190 | VmaMutexLockRead lock(m_Mutex, m_UseMutex); |
| 6191 | |
| 6192 | const uint32_t allocCount = (uint32_t)m_AllocationList.GetCount(); |
| 6193 | inoutStats.blockCount += allocCount; |
| 6194 | inoutStats.allocationCount += allocCount; |
| 6195 | |
| 6196 | for(auto* item = m_AllocationList.Front(); item != nullptr; item = DedicatedAllocationLinkedList::GetNext(item)) |
| 6197 | { |
| 6198 | const VkDeviceSize size = item->GetSize(); |
| 6199 | inoutStats.blockBytes += size; |
| 6200 | inoutStats.allocationBytes += size; |
| 6201 | } |
| 6202 | } |
| 6203 | |
| 6204 | #if VMA_STATS_STRING_ENABLED |
| 6205 | void VmaDedicatedAllocationList::BuildStatsString(VmaJsonWriter& json) |
| 6206 | { |
| 6207 | VmaMutexLockRead lock(m_Mutex, m_UseMutex); |
| 6208 | json.BeginArray(); |
| 6209 | for (VmaAllocation alloc = m_AllocationList.Front(); |
| 6210 | alloc != VMA_NULL; alloc = m_AllocationList.GetNext(item: alloc)) |
| 6211 | { |
| 6212 | json.BeginObject(singleLine: true); |
| 6213 | alloc->PrintParameters(json); |
| 6214 | json.EndObject(); |
| 6215 | } |
| 6216 | json.EndArray(); |
| 6217 | } |
| 6218 | #endif // VMA_STATS_STRING_ENABLED |
| 6219 | |
| 6220 | bool VmaDedicatedAllocationList::IsEmpty() |
| 6221 | { |
| 6222 | VmaMutexLockRead lock(m_Mutex, m_UseMutex); |
| 6223 | return m_AllocationList.IsEmpty(); |
| 6224 | } |
| 6225 | |
| 6226 | void VmaDedicatedAllocationList::Register(VmaAllocation alloc) |
| 6227 | { |
| 6228 | VmaMutexLockWrite lock(m_Mutex, m_UseMutex); |
| 6229 | m_AllocationList.PushBack(item: alloc); |
| 6230 | } |
| 6231 | |
| 6232 | void VmaDedicatedAllocationList::Unregister(VmaAllocation alloc) |
| 6233 | { |
| 6234 | VmaMutexLockWrite lock(m_Mutex, m_UseMutex); |
| 6235 | m_AllocationList.Remove(item: alloc); |
| 6236 | } |
| 6237 | #endif // _VMA_DEDICATED_ALLOCATION_LIST_FUNCTIONS |
| 6238 | #endif // _VMA_DEDICATED_ALLOCATION_LIST |
| 6239 | |
| 6240 | #ifndef _VMA_SUBALLOCATION |
| 6241 | /* |
| 6242 | Represents a region of VmaDeviceMemoryBlock that is either assigned and returned as |
| 6243 | allocated memory block or free. |
| 6244 | */ |
| 6245 | struct VmaSuballocation |
| 6246 | { |
| 6247 | VkDeviceSize offset; |
| 6248 | VkDeviceSize size; |
| 6249 | void* userData; |
| 6250 | VmaSuballocationType type; |
| 6251 | }; |
| 6252 | |
| 6253 | // Comparator for offsets. |
| 6254 | struct VmaSuballocationOffsetLess |
| 6255 | { |
| 6256 | bool operator()(const VmaSuballocation& lhs, const VmaSuballocation& rhs) const |
| 6257 | { |
| 6258 | return lhs.offset < rhs.offset; |
| 6259 | } |
| 6260 | }; |
| 6261 | |
| 6262 | struct VmaSuballocationOffsetGreater |
| 6263 | { |
| 6264 | bool operator()(const VmaSuballocation& lhs, const VmaSuballocation& rhs) const |
| 6265 | { |
| 6266 | return lhs.offset > rhs.offset; |
| 6267 | } |
| 6268 | }; |
| 6269 | |
| 6270 | struct VmaSuballocationItemSizeLess |
| 6271 | { |
| 6272 | bool operator()(const VmaSuballocationList::iterator lhs, |
| 6273 | const VmaSuballocationList::iterator rhs) const |
| 6274 | { |
| 6275 | return lhs->size < rhs->size; |
| 6276 | } |
| 6277 | |
| 6278 | bool operator()(const VmaSuballocationList::iterator lhs, |
| 6279 | VkDeviceSize rhsSize) const |
| 6280 | { |
| 6281 | return lhs->size < rhsSize; |
| 6282 | } |
| 6283 | }; |
| 6284 | #endif // _VMA_SUBALLOCATION |
| 6285 | |
| 6286 | #ifndef _VMA_ALLOCATION_REQUEST |
| 6287 | /* |
| 6288 | Parameters of planned allocation inside a VmaDeviceMemoryBlock. |
| 6289 | item points to a FREE suballocation. |
| 6290 | */ |
| 6291 | struct VmaAllocationRequest |
| 6292 | { |
| 6293 | VmaAllocHandle allocHandle; |
| 6294 | VkDeviceSize size; |
| 6295 | VmaSuballocationList::iterator item; |
| 6296 | void* customData; |
| 6297 | uint64_t algorithmData; |
| 6298 | VmaAllocationRequestType type; |
| 6299 | }; |
| 6300 | #endif // _VMA_ALLOCATION_REQUEST |
| 6301 | |
| 6302 | #ifndef _VMA_BLOCK_METADATA |
| 6303 | /* |
| 6304 | Data structure used for bookkeeping of allocations and unused ranges of memory |
| 6305 | in a single VkDeviceMemory block. |
| 6306 | */ |
| 6307 | class VmaBlockMetadata |
| 6308 | { |
| 6309 | public: |
| 6310 | // pAllocationCallbacks, if not null, must be owned externally - alive and unchanged for the whole lifetime of this object. |
| 6311 | VmaBlockMetadata(const VkAllocationCallbacks* pAllocationCallbacks, |
| 6312 | VkDeviceSize bufferImageGranularity, bool isVirtual); |
| 6313 | virtual ~VmaBlockMetadata() = default; |
| 6314 | |
| 6315 | virtual void Init(VkDeviceSize size) { m_Size = size; } |
| 6316 | bool IsVirtual() const { return m_IsVirtual; } |
| 6317 | VkDeviceSize GetSize() const { return m_Size; } |
| 6318 | |
| 6319 | // Validates all data structures inside this object. If not valid, returns false. |
| 6320 | virtual bool Validate() const = 0; |
| 6321 | virtual size_t GetAllocationCount() const = 0; |
| 6322 | virtual size_t GetFreeRegionsCount() const = 0; |
| 6323 | virtual VkDeviceSize GetSumFreeSize() const = 0; |
| 6324 | // Returns true if this block is empty - contains only single free suballocation. |
| 6325 | virtual bool IsEmpty() const = 0; |
| 6326 | virtual void GetAllocationInfo(VmaAllocHandle allocHandle, VmaVirtualAllocationInfo& outInfo) = 0; |
| 6327 | virtual VkDeviceSize GetAllocationOffset(VmaAllocHandle allocHandle) const = 0; |
| 6328 | virtual void* GetAllocationUserData(VmaAllocHandle allocHandle) const = 0; |
| 6329 | |
| 6330 | virtual VmaAllocHandle GetAllocationListBegin() const = 0; |
| 6331 | virtual VmaAllocHandle GetNextAllocation(VmaAllocHandle prevAlloc) const = 0; |
| 6332 | virtual VkDeviceSize GetNextFreeRegionSize(VmaAllocHandle alloc) const = 0; |
| 6333 | |
| 6334 | // Shouldn't modify blockCount. |
| 6335 | virtual void AddDetailedStatistics(VmaDetailedStatistics& inoutStats) const = 0; |
| 6336 | virtual void AddStatistics(VmaStatistics& inoutStats) const = 0; |
| 6337 | |
| 6338 | #if VMA_STATS_STRING_ENABLED |
| 6339 | virtual void PrintDetailedMap(class VmaJsonWriter& json) const = 0; |
| 6340 | #endif |
| 6341 | |
| 6342 | // Tries to find a place for suballocation with given parameters inside this block. |
| 6343 | // If succeeded, fills pAllocationRequest and returns true. |
| 6344 | // If failed, returns false. |
| 6345 | virtual bool CreateAllocationRequest( |
| 6346 | VkDeviceSize allocSize, |
| 6347 | VkDeviceSize allocAlignment, |
| 6348 | bool upperAddress, |
| 6349 | VmaSuballocationType allocType, |
| 6350 | // Always one of VMA_ALLOCATION_CREATE_STRATEGY_* or VMA_ALLOCATION_INTERNAL_STRATEGY_* flags. |
| 6351 | uint32_t strategy, |
| 6352 | VmaAllocationRequest* pAllocationRequest) = 0; |
| 6353 | |
| 6354 | virtual VkResult CheckCorruption(const void* pBlockData) = 0; |
| 6355 | |
| 6356 | // Makes actual allocation based on request. Request must already be checked and valid. |
| 6357 | virtual void Alloc( |
| 6358 | const VmaAllocationRequest& request, |
| 6359 | VmaSuballocationType type, |
| 6360 | void* userData) = 0; |
| 6361 | |
| 6362 | // Frees suballocation assigned to given memory region. |
| 6363 | virtual void Free(VmaAllocHandle allocHandle) = 0; |
| 6364 | |
| 6365 | // Frees all allocations. |
| 6366 | // Careful! Don't call it if there are VmaAllocation objects owned by userData of cleared allocations! |
| 6367 | virtual void Clear() = 0; |
| 6368 | |
| 6369 | virtual void SetAllocationUserData(VmaAllocHandle allocHandle, void* userData) = 0; |
| 6370 | virtual void DebugLogAllAllocations() const = 0; |
| 6371 | |
| 6372 | protected: |
| 6373 | const VkAllocationCallbacks* GetAllocationCallbacks() const { return m_pAllocationCallbacks; } |
| 6374 | VkDeviceSize GetBufferImageGranularity() const { return m_BufferImageGranularity; } |
| 6375 | VkDeviceSize GetDebugMargin() const { return IsVirtual() ? 0 : VMA_DEBUG_MARGIN; } |
| 6376 | |
| 6377 | void DebugLogAllocation(VkDeviceSize offset, VkDeviceSize size, void* userData) const; |
| 6378 | #if VMA_STATS_STRING_ENABLED |
| 6379 | // mapRefCount == UINT32_MAX means unspecified. |
| 6380 | void PrintDetailedMap_Begin(class VmaJsonWriter& json, |
| 6381 | VkDeviceSize unusedBytes, |
| 6382 | size_t allocationCount, |
| 6383 | size_t unusedRangeCount) const; |
| 6384 | void PrintDetailedMap_Allocation(class VmaJsonWriter& json, |
| 6385 | VkDeviceSize offset, VkDeviceSize size, void* userData) const; |
| 6386 | void PrintDetailedMap_UnusedRange(class VmaJsonWriter& json, |
| 6387 | VkDeviceSize offset, |
| 6388 | VkDeviceSize size) const; |
| 6389 | void PrintDetailedMap_End(class VmaJsonWriter& json) const; |
| 6390 | #endif |
| 6391 | |
| 6392 | private: |
| 6393 | VkDeviceSize m_Size; |
| 6394 | const VkAllocationCallbacks* m_pAllocationCallbacks; |
| 6395 | const VkDeviceSize m_BufferImageGranularity; |
| 6396 | const bool m_IsVirtual; |
| 6397 | }; |
| 6398 | |
| 6399 | #ifndef _VMA_BLOCK_METADATA_FUNCTIONS |
| 6400 | VmaBlockMetadata::VmaBlockMetadata(const VkAllocationCallbacks* pAllocationCallbacks, |
| 6401 | VkDeviceSize bufferImageGranularity, bool isVirtual) |
| 6402 | : m_Size(0), |
| 6403 | m_pAllocationCallbacks(pAllocationCallbacks), |
| 6404 | m_BufferImageGranularity(bufferImageGranularity), |
| 6405 | m_IsVirtual(isVirtual) {} |
| 6406 | |
| 6407 | void VmaBlockMetadata::DebugLogAllocation(VkDeviceSize offset, VkDeviceSize size, void* userData) const |
| 6408 | { |
| 6409 | if (IsVirtual()) |
| 6410 | { |
| 6411 | VMA_DEBUG_LOG("UNFREED VIRTUAL ALLOCATION; Offset: %llu; Size: %llu; UserData: %p" , offset, size, userData); |
| 6412 | } |
| 6413 | else |
| 6414 | { |
| 6415 | VMA_ASSERT(userData != VMA_NULL); |
| 6416 | VmaAllocation allocation = reinterpret_cast<VmaAllocation>(userData); |
| 6417 | |
| 6418 | userData = allocation->GetUserData(); |
| 6419 | const char* name = allocation->GetName(); |
| 6420 | |
| 6421 | #if VMA_STATS_STRING_ENABLED |
| 6422 | VMA_DEBUG_LOG("UNFREED ALLOCATION; Offset: %llu; Size: %llu; UserData: %p; Name: %s; Type: %s; Usage: %u" , |
| 6423 | offset, size, userData, name ? name : "vma_empty" , |
| 6424 | VMA_SUBALLOCATION_TYPE_NAMES[allocation->GetSuballocationType()], |
| 6425 | allocation->GetBufferImageUsage()); |
| 6426 | #else |
| 6427 | VMA_DEBUG_LOG("UNFREED ALLOCATION; Offset: %llu; Size: %llu; UserData: %p; Name: %s; Type: %u" , |
| 6428 | offset, size, userData, name ? name : "vma_empty" , |
| 6429 | (uint32_t)allocation->GetSuballocationType()); |
| 6430 | #endif // VMA_STATS_STRING_ENABLED |
| 6431 | } |
| 6432 | |
| 6433 | } |
| 6434 | |
| 6435 | #if VMA_STATS_STRING_ENABLED |
| 6436 | void VmaBlockMetadata::PrintDetailedMap_Begin(class VmaJsonWriter& json, |
| 6437 | VkDeviceSize unusedBytes, size_t allocationCount, size_t unusedRangeCount) const |
| 6438 | { |
| 6439 | json.WriteString(pStr: "TotalBytes" ); |
| 6440 | json.WriteNumber(n: GetSize()); |
| 6441 | |
| 6442 | json.WriteString(pStr: "UnusedBytes" ); |
| 6443 | json.WriteSize(n: unusedBytes); |
| 6444 | |
| 6445 | json.WriteString(pStr: "Allocations" ); |
| 6446 | json.WriteSize(n: allocationCount); |
| 6447 | |
| 6448 | json.WriteString(pStr: "UnusedRanges" ); |
| 6449 | json.WriteSize(n: unusedRangeCount); |
| 6450 | |
| 6451 | json.WriteString(pStr: "Suballocations" ); |
| 6452 | json.BeginArray(); |
| 6453 | } |
| 6454 | |
| 6455 | void VmaBlockMetadata::PrintDetailedMap_Allocation(class VmaJsonWriter& json, |
| 6456 | VkDeviceSize offset, VkDeviceSize size, void* userData) const |
| 6457 | { |
| 6458 | json.BeginObject(singleLine: true); |
| 6459 | |
| 6460 | json.WriteString(pStr: "Offset" ); |
| 6461 | json.WriteNumber(n: offset); |
| 6462 | |
| 6463 | if (IsVirtual()) |
| 6464 | { |
| 6465 | json.WriteString(pStr: "Size" ); |
| 6466 | json.WriteNumber(n: size); |
| 6467 | if (userData) |
| 6468 | { |
| 6469 | json.WriteString(pStr: "CustomData" ); |
| 6470 | json.BeginString(); |
| 6471 | json.ContinueString_Pointer(ptr: userData); |
| 6472 | json.EndString(); |
| 6473 | } |
| 6474 | } |
| 6475 | else |
| 6476 | { |
| 6477 | ((VmaAllocation)userData)->PrintParameters(json); |
| 6478 | } |
| 6479 | |
| 6480 | json.EndObject(); |
| 6481 | } |
| 6482 | |
| 6483 | void VmaBlockMetadata::PrintDetailedMap_UnusedRange(class VmaJsonWriter& json, |
| 6484 | VkDeviceSize offset, VkDeviceSize size) const |
| 6485 | { |
| 6486 | json.BeginObject(singleLine: true); |
| 6487 | |
| 6488 | json.WriteString(pStr: "Offset" ); |
| 6489 | json.WriteNumber(n: offset); |
| 6490 | |
| 6491 | json.WriteString(pStr: "Type" ); |
| 6492 | json.WriteString(pStr: VMA_SUBALLOCATION_TYPE_NAMES[VMA_SUBALLOCATION_TYPE_FREE]); |
| 6493 | |
| 6494 | json.WriteString(pStr: "Size" ); |
| 6495 | json.WriteNumber(n: size); |
| 6496 | |
| 6497 | json.EndObject(); |
| 6498 | } |
| 6499 | |
| 6500 | void VmaBlockMetadata::PrintDetailedMap_End(class VmaJsonWriter& json) const |
| 6501 | { |
| 6502 | json.EndArray(); |
| 6503 | } |
| 6504 | #endif // VMA_STATS_STRING_ENABLED |
| 6505 | #endif // _VMA_BLOCK_METADATA_FUNCTIONS |
| 6506 | #endif // _VMA_BLOCK_METADATA |
| 6507 | |
| 6508 | #ifndef _VMA_BLOCK_BUFFER_IMAGE_GRANULARITY |
| 6509 | // Before deleting object of this class remember to call 'Destroy()' |
| 6510 | class VmaBlockBufferImageGranularity final |
| 6511 | { |
| 6512 | public: |
| 6513 | struct ValidationContext |
| 6514 | { |
| 6515 | const VkAllocationCallbacks* allocCallbacks; |
| 6516 | uint16_t* pageAllocs; |
| 6517 | }; |
| 6518 | |
| 6519 | VmaBlockBufferImageGranularity(VkDeviceSize bufferImageGranularity); |
| 6520 | ~VmaBlockBufferImageGranularity(); |
| 6521 | |
| 6522 | bool IsEnabled() const { return m_BufferImageGranularity > MAX_LOW_BUFFER_IMAGE_GRANULARITY; } |
| 6523 | |
| 6524 | void Init(const VkAllocationCallbacks* pAllocationCallbacks, VkDeviceSize size); |
| 6525 | // Before destroying object you must call free it's memory |
| 6526 | void Destroy(const VkAllocationCallbacks* pAllocationCallbacks); |
| 6527 | |
| 6528 | void RoundupAllocRequest(VmaSuballocationType allocType, |
| 6529 | VkDeviceSize& inOutAllocSize, |
| 6530 | VkDeviceSize& inOutAllocAlignment) const; |
| 6531 | |
| 6532 | bool CheckConflictAndAlignUp(VkDeviceSize& inOutAllocOffset, |
| 6533 | VkDeviceSize allocSize, |
| 6534 | VkDeviceSize blockOffset, |
| 6535 | VkDeviceSize blockSize, |
| 6536 | VmaSuballocationType allocType) const; |
| 6537 | |
| 6538 | void AllocPages(uint8_t allocType, VkDeviceSize offset, VkDeviceSize size); |
| 6539 | void FreePages(VkDeviceSize offset, VkDeviceSize size); |
| 6540 | void Clear(); |
| 6541 | |
| 6542 | ValidationContext StartValidation(const VkAllocationCallbacks* pAllocationCallbacks, |
| 6543 | bool isVirutal) const; |
| 6544 | bool Validate(ValidationContext& ctx, VkDeviceSize offset, VkDeviceSize size) const; |
| 6545 | bool FinishValidation(ValidationContext& ctx) const; |
| 6546 | |
| 6547 | private: |
| 6548 | static const uint16_t MAX_LOW_BUFFER_IMAGE_GRANULARITY = 256; |
| 6549 | |
| 6550 | struct RegionInfo |
| 6551 | { |
| 6552 | uint8_t allocType; |
| 6553 | uint16_t allocCount; |
| 6554 | }; |
| 6555 | |
| 6556 | VkDeviceSize m_BufferImageGranularity; |
| 6557 | uint32_t m_RegionCount; |
| 6558 | RegionInfo* m_RegionInfo; |
| 6559 | |
| 6560 | uint32_t GetStartPage(VkDeviceSize offset) const { return OffsetToPageIndex(offset: offset & ~(m_BufferImageGranularity - 1)); } |
| 6561 | uint32_t GetEndPage(VkDeviceSize offset, VkDeviceSize size) const { return OffsetToPageIndex(offset: (offset + size - 1) & ~(m_BufferImageGranularity - 1)); } |
| 6562 | |
| 6563 | uint32_t OffsetToPageIndex(VkDeviceSize offset) const; |
| 6564 | void AllocPage(RegionInfo& page, uint8_t allocType); |
| 6565 | }; |
| 6566 | |
| 6567 | #ifndef _VMA_BLOCK_BUFFER_IMAGE_GRANULARITY_FUNCTIONS |
| 6568 | VmaBlockBufferImageGranularity::VmaBlockBufferImageGranularity(VkDeviceSize bufferImageGranularity) |
| 6569 | : m_BufferImageGranularity(bufferImageGranularity), |
| 6570 | m_RegionCount(0), |
| 6571 | m_RegionInfo(VMA_NULL) {} |
| 6572 | |
| 6573 | VmaBlockBufferImageGranularity::~VmaBlockBufferImageGranularity() |
| 6574 | { |
| 6575 | VMA_ASSERT(m_RegionInfo == VMA_NULL && "Free not called before destroying object!" ); |
| 6576 | } |
| 6577 | |
| 6578 | void VmaBlockBufferImageGranularity::Init(const VkAllocationCallbacks* pAllocationCallbacks, VkDeviceSize size) |
| 6579 | { |
| 6580 | if (IsEnabled()) |
| 6581 | { |
| 6582 | m_RegionCount = static_cast<uint32_t>(VmaDivideRoundingUp(x: size, y: m_BufferImageGranularity)); |
| 6583 | m_RegionInfo = vma_new_array(pAllocationCallbacks, RegionInfo, m_RegionCount); |
| 6584 | memset(s: m_RegionInfo, c: 0, n: m_RegionCount * sizeof(RegionInfo)); |
| 6585 | } |
| 6586 | } |
| 6587 | |
| 6588 | void VmaBlockBufferImageGranularity::Destroy(const VkAllocationCallbacks* pAllocationCallbacks) |
| 6589 | { |
| 6590 | if (m_RegionInfo) |
| 6591 | { |
| 6592 | vma_delete_array(pAllocationCallbacks, ptr: m_RegionInfo, count: m_RegionCount); |
| 6593 | m_RegionInfo = VMA_NULL; |
| 6594 | } |
| 6595 | } |
| 6596 | |
| 6597 | void VmaBlockBufferImageGranularity::RoundupAllocRequest(VmaSuballocationType allocType, |
| 6598 | VkDeviceSize& inOutAllocSize, |
| 6599 | VkDeviceSize& inOutAllocAlignment) const |
| 6600 | { |
| 6601 | if (m_BufferImageGranularity > 1 && |
| 6602 | m_BufferImageGranularity <= MAX_LOW_BUFFER_IMAGE_GRANULARITY) |
| 6603 | { |
| 6604 | if (allocType == VMA_SUBALLOCATION_TYPE_UNKNOWN || |
| 6605 | allocType == VMA_SUBALLOCATION_TYPE_IMAGE_UNKNOWN || |
| 6606 | allocType == VMA_SUBALLOCATION_TYPE_IMAGE_OPTIMAL) |
| 6607 | { |
| 6608 | inOutAllocAlignment = VMA_MAX(inOutAllocAlignment, m_BufferImageGranularity); |
| 6609 | inOutAllocSize = VmaAlignUp(val: inOutAllocSize, alignment: m_BufferImageGranularity); |
| 6610 | } |
| 6611 | } |
| 6612 | } |
| 6613 | |
| 6614 | bool VmaBlockBufferImageGranularity::CheckConflictAndAlignUp(VkDeviceSize& inOutAllocOffset, |
| 6615 | VkDeviceSize allocSize, |
| 6616 | VkDeviceSize blockOffset, |
| 6617 | VkDeviceSize blockSize, |
| 6618 | VmaSuballocationType allocType) const |
| 6619 | { |
| 6620 | if (IsEnabled()) |
| 6621 | { |
| 6622 | uint32_t startPage = GetStartPage(offset: inOutAllocOffset); |
| 6623 | if (m_RegionInfo[startPage].allocCount > 0 && |
| 6624 | VmaIsBufferImageGranularityConflict(suballocType1: static_cast<VmaSuballocationType>(m_RegionInfo[startPage].allocType), suballocType2: allocType)) |
| 6625 | { |
| 6626 | inOutAllocOffset = VmaAlignUp(val: inOutAllocOffset, alignment: m_BufferImageGranularity); |
| 6627 | if (blockSize < allocSize + inOutAllocOffset - blockOffset) |
| 6628 | return true; |
| 6629 | ++startPage; |
| 6630 | } |
| 6631 | uint32_t endPage = GetEndPage(offset: inOutAllocOffset, size: allocSize); |
| 6632 | if (endPage != startPage && |
| 6633 | m_RegionInfo[endPage].allocCount > 0 && |
| 6634 | VmaIsBufferImageGranularityConflict(suballocType1: static_cast<VmaSuballocationType>(m_RegionInfo[endPage].allocType), suballocType2: allocType)) |
| 6635 | { |
| 6636 | return true; |
| 6637 | } |
| 6638 | } |
| 6639 | return false; |
| 6640 | } |
| 6641 | |
| 6642 | void VmaBlockBufferImageGranularity::AllocPages(uint8_t allocType, VkDeviceSize offset, VkDeviceSize size) |
| 6643 | { |
| 6644 | if (IsEnabled()) |
| 6645 | { |
| 6646 | uint32_t startPage = GetStartPage(offset); |
| 6647 | AllocPage(page&: m_RegionInfo[startPage], allocType); |
| 6648 | |
| 6649 | uint32_t endPage = GetEndPage(offset, size); |
| 6650 | if (startPage != endPage) |
| 6651 | AllocPage(page&: m_RegionInfo[endPage], allocType); |
| 6652 | } |
| 6653 | } |
| 6654 | |
| 6655 | void VmaBlockBufferImageGranularity::FreePages(VkDeviceSize offset, VkDeviceSize size) |
| 6656 | { |
| 6657 | if (IsEnabled()) |
| 6658 | { |
| 6659 | uint32_t startPage = GetStartPage(offset); |
| 6660 | --m_RegionInfo[startPage].allocCount; |
| 6661 | if (m_RegionInfo[startPage].allocCount == 0) |
| 6662 | m_RegionInfo[startPage].allocType = VMA_SUBALLOCATION_TYPE_FREE; |
| 6663 | uint32_t endPage = GetEndPage(offset, size); |
| 6664 | if (startPage != endPage) |
| 6665 | { |
| 6666 | --m_RegionInfo[endPage].allocCount; |
| 6667 | if (m_RegionInfo[endPage].allocCount == 0) |
| 6668 | m_RegionInfo[endPage].allocType = VMA_SUBALLOCATION_TYPE_FREE; |
| 6669 | } |
| 6670 | } |
| 6671 | } |
| 6672 | |
| 6673 | void VmaBlockBufferImageGranularity::Clear() |
| 6674 | { |
| 6675 | if (m_RegionInfo) |
| 6676 | memset(s: m_RegionInfo, c: 0, n: m_RegionCount * sizeof(RegionInfo)); |
| 6677 | } |
| 6678 | |
| 6679 | VmaBlockBufferImageGranularity::ValidationContext VmaBlockBufferImageGranularity::StartValidation( |
| 6680 | const VkAllocationCallbacks* pAllocationCallbacks, bool isVirutal) const |
| 6681 | { |
| 6682 | ValidationContext ctx{ .allocCallbacks: pAllocationCallbacks, VMA_NULL }; |
| 6683 | if (!isVirutal && IsEnabled()) |
| 6684 | { |
| 6685 | ctx.pageAllocs = vma_new_array(pAllocationCallbacks, uint16_t, m_RegionCount); |
| 6686 | memset(s: ctx.pageAllocs, c: 0, n: m_RegionCount * sizeof(uint16_t)); |
| 6687 | } |
| 6688 | return ctx; |
| 6689 | } |
| 6690 | |
| 6691 | bool VmaBlockBufferImageGranularity::Validate(ValidationContext& ctx, |
| 6692 | VkDeviceSize offset, VkDeviceSize size) const |
| 6693 | { |
| 6694 | if (IsEnabled()) |
| 6695 | { |
| 6696 | uint32_t start = GetStartPage(offset); |
| 6697 | ++ctx.pageAllocs[start]; |
| 6698 | VMA_VALIDATE(m_RegionInfo[start].allocCount > 0); |
| 6699 | |
| 6700 | uint32_t end = GetEndPage(offset, size); |
| 6701 | if (start != end) |
| 6702 | { |
| 6703 | ++ctx.pageAllocs[end]; |
| 6704 | VMA_VALIDATE(m_RegionInfo[end].allocCount > 0); |
| 6705 | } |
| 6706 | } |
| 6707 | return true; |
| 6708 | } |
| 6709 | |
| 6710 | bool VmaBlockBufferImageGranularity::FinishValidation(ValidationContext& ctx) const |
| 6711 | { |
| 6712 | // Check proper page structure |
| 6713 | if (IsEnabled()) |
| 6714 | { |
| 6715 | VMA_ASSERT(ctx.pageAllocs != VMA_NULL && "Validation context not initialized!" ); |
| 6716 | |
| 6717 | for (uint32_t page = 0; page < m_RegionCount; ++page) |
| 6718 | { |
| 6719 | VMA_VALIDATE(ctx.pageAllocs[page] == m_RegionInfo[page].allocCount); |
| 6720 | } |
| 6721 | vma_delete_array(pAllocationCallbacks: ctx.allocCallbacks, ptr: ctx.pageAllocs, count: m_RegionCount); |
| 6722 | ctx.pageAllocs = VMA_NULL; |
| 6723 | } |
| 6724 | return true; |
| 6725 | } |
| 6726 | |
| 6727 | uint32_t VmaBlockBufferImageGranularity::OffsetToPageIndex(VkDeviceSize offset) const |
| 6728 | { |
| 6729 | return static_cast<uint32_t>(offset >> VMA_BITSCAN_MSB(m_BufferImageGranularity)); |
| 6730 | } |
| 6731 | |
| 6732 | void VmaBlockBufferImageGranularity::AllocPage(RegionInfo& page, uint8_t allocType) |
| 6733 | { |
| 6734 | // When current alloc type is free then it can be overriden by new type |
| 6735 | if (page.allocCount == 0 || (page.allocCount > 0 && page.allocType == VMA_SUBALLOCATION_TYPE_FREE)) |
| 6736 | page.allocType = allocType; |
| 6737 | |
| 6738 | ++page.allocCount; |
| 6739 | } |
| 6740 | #endif // _VMA_BLOCK_BUFFER_IMAGE_GRANULARITY_FUNCTIONS |
| 6741 | #endif // _VMA_BLOCK_BUFFER_IMAGE_GRANULARITY |
| 6742 | |
| 6743 | #if 0 |
| 6744 | #ifndef _VMA_BLOCK_METADATA_GENERIC |
| 6745 | class VmaBlockMetadata_Generic : public VmaBlockMetadata |
| 6746 | { |
| 6747 | friend class VmaDefragmentationAlgorithm_Generic; |
| 6748 | friend class VmaDefragmentationAlgorithm_Fast; |
| 6749 | VMA_CLASS_NO_COPY(VmaBlockMetadata_Generic) |
| 6750 | public: |
| 6751 | VmaBlockMetadata_Generic(const VkAllocationCallbacks* pAllocationCallbacks, |
| 6752 | VkDeviceSize bufferImageGranularity, bool isVirtual); |
| 6753 | virtual ~VmaBlockMetadata_Generic() = default; |
| 6754 | |
| 6755 | size_t GetAllocationCount() const override { return m_Suballocations.size() - m_FreeCount; } |
| 6756 | VkDeviceSize GetSumFreeSize() const override { return m_SumFreeSize; } |
| 6757 | bool IsEmpty() const override { return (m_Suballocations.size() == 1) && (m_FreeCount == 1); } |
| 6758 | void Free(VmaAllocHandle allocHandle) override { FreeSuballocation(FindAtOffset((VkDeviceSize)allocHandle - 1)); } |
| 6759 | VkDeviceSize GetAllocationOffset(VmaAllocHandle allocHandle) const override { return (VkDeviceSize)allocHandle - 1; }; |
| 6760 | |
| 6761 | void Init(VkDeviceSize size) override; |
| 6762 | bool Validate() const override; |
| 6763 | |
| 6764 | void AddDetailedStatistics(VmaDetailedStatistics& inoutStats) const override; |
| 6765 | void AddStatistics(VmaStatistics& inoutStats) const override; |
| 6766 | |
| 6767 | #if VMA_STATS_STRING_ENABLED |
| 6768 | void PrintDetailedMap(class VmaJsonWriter& json, uint32_t mapRefCount) const override; |
| 6769 | #endif |
| 6770 | |
| 6771 | bool CreateAllocationRequest( |
| 6772 | VkDeviceSize allocSize, |
| 6773 | VkDeviceSize allocAlignment, |
| 6774 | bool upperAddress, |
| 6775 | VmaSuballocationType allocType, |
| 6776 | uint32_t strategy, |
| 6777 | VmaAllocationRequest* pAllocationRequest) override; |
| 6778 | |
| 6779 | VkResult CheckCorruption(const void* pBlockData) override; |
| 6780 | |
| 6781 | void Alloc( |
| 6782 | const VmaAllocationRequest& request, |
| 6783 | VmaSuballocationType type, |
| 6784 | void* userData) override; |
| 6785 | |
| 6786 | void GetAllocationInfo(VmaAllocHandle allocHandle, VmaVirtualAllocationInfo& outInfo) override; |
| 6787 | void* GetAllocationUserData(VmaAllocHandle allocHandle) const override; |
| 6788 | VmaAllocHandle GetAllocationListBegin() const override; |
| 6789 | VmaAllocHandle GetNextAllocation(VmaAllocHandle prevAlloc) const override; |
| 6790 | void Clear() override; |
| 6791 | void SetAllocationUserData(VmaAllocHandle allocHandle, void* userData) override; |
| 6792 | void DebugLogAllAllocations() const override; |
| 6793 | |
| 6794 | private: |
| 6795 | uint32_t m_FreeCount; |
| 6796 | VkDeviceSize m_SumFreeSize; |
| 6797 | VmaSuballocationList m_Suballocations; |
| 6798 | // Suballocations that are free. Sorted by size, ascending. |
| 6799 | VmaVector<VmaSuballocationList::iterator, VmaStlAllocator<VmaSuballocationList::iterator>> m_FreeSuballocationsBySize; |
| 6800 | |
| 6801 | VkDeviceSize AlignAllocationSize(VkDeviceSize size) const { return IsVirtual() ? size : VmaAlignUp(size, (VkDeviceSize)16); } |
| 6802 | |
| 6803 | VmaSuballocationList::iterator FindAtOffset(VkDeviceSize offset) const; |
| 6804 | bool ValidateFreeSuballocationList() const; |
| 6805 | |
| 6806 | // Checks if requested suballocation with given parameters can be placed in given pFreeSuballocItem. |
| 6807 | // If yes, fills pOffset and returns true. If no, returns false. |
| 6808 | bool CheckAllocation( |
| 6809 | VkDeviceSize allocSize, |
| 6810 | VkDeviceSize allocAlignment, |
| 6811 | VmaSuballocationType allocType, |
| 6812 | VmaSuballocationList::const_iterator suballocItem, |
| 6813 | VmaAllocHandle* pAllocHandle) const; |
| 6814 | |
| 6815 | // Given free suballocation, it merges it with following one, which must also be free. |
| 6816 | void MergeFreeWithNext(VmaSuballocationList::iterator item); |
| 6817 | // Releases given suballocation, making it free. |
| 6818 | // Merges it with adjacent free suballocations if applicable. |
| 6819 | // Returns iterator to new free suballocation at this place. |
| 6820 | VmaSuballocationList::iterator FreeSuballocation(VmaSuballocationList::iterator suballocItem); |
| 6821 | // Given free suballocation, it inserts it into sorted list of |
| 6822 | // m_FreeSuballocationsBySize if it is suitable. |
| 6823 | void RegisterFreeSuballocation(VmaSuballocationList::iterator item); |
| 6824 | // Given free suballocation, it removes it from sorted list of |
| 6825 | // m_FreeSuballocationsBySize if it is suitable. |
| 6826 | void UnregisterFreeSuballocation(VmaSuballocationList::iterator item); |
| 6827 | }; |
| 6828 | |
| 6829 | #ifndef _VMA_BLOCK_METADATA_GENERIC_FUNCTIONS |
| 6830 | VmaBlockMetadata_Generic::VmaBlockMetadata_Generic(const VkAllocationCallbacks* pAllocationCallbacks, |
| 6831 | VkDeviceSize bufferImageGranularity, bool isVirtual) |
| 6832 | : VmaBlockMetadata(pAllocationCallbacks, bufferImageGranularity, isVirtual), |
| 6833 | m_FreeCount(0), |
| 6834 | m_SumFreeSize(0), |
| 6835 | m_Suballocations(VmaStlAllocator<VmaSuballocation>(pAllocationCallbacks)), |
| 6836 | m_FreeSuballocationsBySize(VmaStlAllocator<VmaSuballocationList::iterator>(pAllocationCallbacks)) {} |
| 6837 | |
| 6838 | void VmaBlockMetadata_Generic::Init(VkDeviceSize size) |
| 6839 | { |
| 6840 | VmaBlockMetadata::Init(size); |
| 6841 | |
| 6842 | m_FreeCount = 1; |
| 6843 | m_SumFreeSize = size; |
| 6844 | |
| 6845 | VmaSuballocation suballoc = {}; |
| 6846 | suballoc.offset = 0; |
| 6847 | suballoc.size = size; |
| 6848 | suballoc.type = VMA_SUBALLOCATION_TYPE_FREE; |
| 6849 | |
| 6850 | m_Suballocations.push_back(suballoc); |
| 6851 | m_FreeSuballocationsBySize.push_back(m_Suballocations.begin()); |
| 6852 | } |
| 6853 | |
| 6854 | bool VmaBlockMetadata_Generic::Validate() const |
| 6855 | { |
| 6856 | VMA_VALIDATE(!m_Suballocations.empty()); |
| 6857 | |
| 6858 | // Expected offset of new suballocation as calculated from previous ones. |
| 6859 | VkDeviceSize calculatedOffset = 0; |
| 6860 | // Expected number of free suballocations as calculated from traversing their list. |
| 6861 | uint32_t calculatedFreeCount = 0; |
| 6862 | // Expected sum size of free suballocations as calculated from traversing their list. |
| 6863 | VkDeviceSize calculatedSumFreeSize = 0; |
| 6864 | // Expected number of free suballocations that should be registered in |
| 6865 | // m_FreeSuballocationsBySize calculated from traversing their list. |
| 6866 | size_t freeSuballocationsToRegister = 0; |
| 6867 | // True if previous visited suballocation was free. |
| 6868 | bool prevFree = false; |
| 6869 | |
| 6870 | const VkDeviceSize debugMargin = GetDebugMargin(); |
| 6871 | |
| 6872 | for (const auto& subAlloc : m_Suballocations) |
| 6873 | { |
| 6874 | // Actual offset of this suballocation doesn't match expected one. |
| 6875 | VMA_VALIDATE(subAlloc.offset == calculatedOffset); |
| 6876 | |
| 6877 | const bool currFree = (subAlloc.type == VMA_SUBALLOCATION_TYPE_FREE); |
| 6878 | // Two adjacent free suballocations are invalid. They should be merged. |
| 6879 | VMA_VALIDATE(!prevFree || !currFree); |
| 6880 | |
| 6881 | VmaAllocation alloc = (VmaAllocation)subAlloc.userData; |
| 6882 | if (!IsVirtual()) |
| 6883 | { |
| 6884 | VMA_VALIDATE(currFree == (alloc == VK_NULL_HANDLE)); |
| 6885 | } |
| 6886 | |
| 6887 | if (currFree) |
| 6888 | { |
| 6889 | calculatedSumFreeSize += subAlloc.size; |
| 6890 | ++calculatedFreeCount; |
| 6891 | ++freeSuballocationsToRegister; |
| 6892 | |
| 6893 | // Margin required between allocations - every free space must be at least that large. |
| 6894 | VMA_VALIDATE(subAlloc.size >= debugMargin); |
| 6895 | } |
| 6896 | else |
| 6897 | { |
| 6898 | if (!IsVirtual()) |
| 6899 | { |
| 6900 | VMA_VALIDATE((VkDeviceSize)alloc->GetAllocHandle() == subAlloc.offset + 1); |
| 6901 | VMA_VALIDATE(alloc->GetSize() == subAlloc.size); |
| 6902 | } |
| 6903 | |
| 6904 | // Margin required between allocations - previous allocation must be free. |
| 6905 | VMA_VALIDATE(debugMargin == 0 || prevFree); |
| 6906 | } |
| 6907 | |
| 6908 | calculatedOffset += subAlloc.size; |
| 6909 | prevFree = currFree; |
| 6910 | } |
| 6911 | |
| 6912 | // Number of free suballocations registered in m_FreeSuballocationsBySize doesn't |
| 6913 | // match expected one. |
| 6914 | VMA_VALIDATE(m_FreeSuballocationsBySize.size() == freeSuballocationsToRegister); |
| 6915 | |
| 6916 | VkDeviceSize lastSize = 0; |
| 6917 | for (size_t i = 0; i < m_FreeSuballocationsBySize.size(); ++i) |
| 6918 | { |
| 6919 | VmaSuballocationList::iterator suballocItem = m_FreeSuballocationsBySize[i]; |
| 6920 | |
| 6921 | // Only free suballocations can be registered in m_FreeSuballocationsBySize. |
| 6922 | VMA_VALIDATE(suballocItem->type == VMA_SUBALLOCATION_TYPE_FREE); |
| 6923 | // They must be sorted by size ascending. |
| 6924 | VMA_VALIDATE(suballocItem->size >= lastSize); |
| 6925 | |
| 6926 | lastSize = suballocItem->size; |
| 6927 | } |
| 6928 | |
| 6929 | // Check if totals match calculated values. |
| 6930 | VMA_VALIDATE(ValidateFreeSuballocationList()); |
| 6931 | VMA_VALIDATE(calculatedOffset == GetSize()); |
| 6932 | VMA_VALIDATE(calculatedSumFreeSize == m_SumFreeSize); |
| 6933 | VMA_VALIDATE(calculatedFreeCount == m_FreeCount); |
| 6934 | |
| 6935 | return true; |
| 6936 | } |
| 6937 | |
| 6938 | void VmaBlockMetadata_Generic::AddDetailedStatistics(VmaDetailedStatistics& inoutStats) const |
| 6939 | { |
| 6940 | const uint32_t rangeCount = (uint32_t)m_Suballocations.size(); |
| 6941 | inoutStats.statistics.blockCount++; |
| 6942 | inoutStats.statistics.blockBytes += GetSize(); |
| 6943 | |
| 6944 | for (const auto& suballoc : m_Suballocations) |
| 6945 | { |
| 6946 | if (suballoc.type != VMA_SUBALLOCATION_TYPE_FREE) |
| 6947 | VmaAddDetailedStatisticsAllocation(inoutStats, suballoc.size); |
| 6948 | else |
| 6949 | VmaAddDetailedStatisticsUnusedRange(inoutStats, suballoc.size); |
| 6950 | } |
| 6951 | } |
| 6952 | |
| 6953 | void VmaBlockMetadata_Generic::AddStatistics(VmaStatistics& inoutStats) const |
| 6954 | { |
| 6955 | inoutStats.blockCount++; |
| 6956 | inoutStats.allocationCount += (uint32_t)m_Suballocations.size() - m_FreeCount; |
| 6957 | inoutStats.blockBytes += GetSize(); |
| 6958 | inoutStats.allocationBytes += GetSize() - m_SumFreeSize; |
| 6959 | } |
| 6960 | |
| 6961 | #if VMA_STATS_STRING_ENABLED |
| 6962 | void VmaBlockMetadata_Generic::PrintDetailedMap(class VmaJsonWriter& json, uint32_t mapRefCount) const |
| 6963 | { |
| 6964 | PrintDetailedMap_Begin(json, |
| 6965 | m_SumFreeSize, // unusedBytes |
| 6966 | m_Suballocations.size() - (size_t)m_FreeCount, // allocationCount |
| 6967 | m_FreeCount, // unusedRangeCount |
| 6968 | mapRefCount); |
| 6969 | |
| 6970 | for (const auto& suballoc : m_Suballocations) |
| 6971 | { |
| 6972 | if (suballoc.type == VMA_SUBALLOCATION_TYPE_FREE) |
| 6973 | { |
| 6974 | PrintDetailedMap_UnusedRange(json, suballoc.offset, suballoc.size); |
| 6975 | } |
| 6976 | else |
| 6977 | { |
| 6978 | PrintDetailedMap_Allocation(json, suballoc.offset, suballoc.size, suballoc.userData); |
| 6979 | } |
| 6980 | } |
| 6981 | |
| 6982 | PrintDetailedMap_End(json); |
| 6983 | } |
| 6984 | #endif // VMA_STATS_STRING_ENABLED |
| 6985 | |
| 6986 | bool VmaBlockMetadata_Generic::CreateAllocationRequest( |
| 6987 | VkDeviceSize allocSize, |
| 6988 | VkDeviceSize allocAlignment, |
| 6989 | bool upperAddress, |
| 6990 | VmaSuballocationType allocType, |
| 6991 | uint32_t strategy, |
| 6992 | VmaAllocationRequest* pAllocationRequest) |
| 6993 | { |
| 6994 | VMA_ASSERT(allocSize > 0); |
| 6995 | VMA_ASSERT(!upperAddress); |
| 6996 | VMA_ASSERT(allocType != VMA_SUBALLOCATION_TYPE_FREE); |
| 6997 | VMA_ASSERT(pAllocationRequest != VMA_NULL); |
| 6998 | VMA_HEAVY_ASSERT(Validate()); |
| 6999 | |
| 7000 | allocSize = AlignAllocationSize(allocSize); |
| 7001 | |
| 7002 | pAllocationRequest->type = VmaAllocationRequestType::Normal; |
| 7003 | pAllocationRequest->size = allocSize; |
| 7004 | |
| 7005 | const VkDeviceSize debugMargin = GetDebugMargin(); |
| 7006 | |
| 7007 | // There is not enough total free space in this block to fulfill the request: Early return. |
| 7008 | if (m_SumFreeSize < allocSize + debugMargin) |
| 7009 | { |
| 7010 | return false; |
| 7011 | } |
| 7012 | |
| 7013 | // New algorithm, efficiently searching freeSuballocationsBySize. |
| 7014 | const size_t freeSuballocCount = m_FreeSuballocationsBySize.size(); |
| 7015 | if (freeSuballocCount > 0) |
| 7016 | { |
| 7017 | if (strategy == 0 || |
| 7018 | strategy == VMA_ALLOCATION_CREATE_STRATEGY_MIN_MEMORY_BIT) |
| 7019 | { |
| 7020 | // Find first free suballocation with size not less than allocSize + debugMargin. |
| 7021 | VmaSuballocationList::iterator* const it = VmaBinaryFindFirstNotLess( |
| 7022 | m_FreeSuballocationsBySize.data(), |
| 7023 | m_FreeSuballocationsBySize.data() + freeSuballocCount, |
| 7024 | allocSize + debugMargin, |
| 7025 | VmaSuballocationItemSizeLess()); |
| 7026 | size_t index = it - m_FreeSuballocationsBySize.data(); |
| 7027 | for (; index < freeSuballocCount; ++index) |
| 7028 | { |
| 7029 | if (CheckAllocation( |
| 7030 | allocSize, |
| 7031 | allocAlignment, |
| 7032 | allocType, |
| 7033 | m_FreeSuballocationsBySize[index], |
| 7034 | &pAllocationRequest->allocHandle)) |
| 7035 | { |
| 7036 | pAllocationRequest->item = m_FreeSuballocationsBySize[index]; |
| 7037 | return true; |
| 7038 | } |
| 7039 | } |
| 7040 | } |
| 7041 | else if (strategy == VMA_ALLOCATION_INTERNAL_STRATEGY_MIN_OFFSET) |
| 7042 | { |
| 7043 | for (VmaSuballocationList::iterator it = m_Suballocations.begin(); |
| 7044 | it != m_Suballocations.end(); |
| 7045 | ++it) |
| 7046 | { |
| 7047 | if (it->type == VMA_SUBALLOCATION_TYPE_FREE && CheckAllocation( |
| 7048 | allocSize, |
| 7049 | allocAlignment, |
| 7050 | allocType, |
| 7051 | it, |
| 7052 | &pAllocationRequest->allocHandle)) |
| 7053 | { |
| 7054 | pAllocationRequest->item = it; |
| 7055 | return true; |
| 7056 | } |
| 7057 | } |
| 7058 | } |
| 7059 | else |
| 7060 | { |
| 7061 | VMA_ASSERT(strategy & (VMA_ALLOCATION_CREATE_STRATEGY_MIN_TIME_BIT | VMA_ALLOCATION_CREATE_STRATEGY_MIN_OFFSET_BIT )); |
| 7062 | // Search staring from biggest suballocations. |
| 7063 | for (size_t index = freeSuballocCount; index--; ) |
| 7064 | { |
| 7065 | if (CheckAllocation( |
| 7066 | allocSize, |
| 7067 | allocAlignment, |
| 7068 | allocType, |
| 7069 | m_FreeSuballocationsBySize[index], |
| 7070 | &pAllocationRequest->allocHandle)) |
| 7071 | { |
| 7072 | pAllocationRequest->item = m_FreeSuballocationsBySize[index]; |
| 7073 | return true; |
| 7074 | } |
| 7075 | } |
| 7076 | } |
| 7077 | } |
| 7078 | |
| 7079 | return false; |
| 7080 | } |
| 7081 | |
| 7082 | VkResult VmaBlockMetadata_Generic::CheckCorruption(const void* pBlockData) |
| 7083 | { |
| 7084 | for (auto& suballoc : m_Suballocations) |
| 7085 | { |
| 7086 | if (suballoc.type != VMA_SUBALLOCATION_TYPE_FREE) |
| 7087 | { |
| 7088 | if (!VmaValidateMagicValue(pBlockData, suballoc.offset + suballoc.size)) |
| 7089 | { |
| 7090 | VMA_ASSERT(0 && "MEMORY CORRUPTION DETECTED AFTER VALIDATED ALLOCATION!" ); |
| 7091 | return VK_ERROR_UNKNOWN_COPY; |
| 7092 | } |
| 7093 | } |
| 7094 | } |
| 7095 | |
| 7096 | return VK_SUCCESS; |
| 7097 | } |
| 7098 | |
| 7099 | void VmaBlockMetadata_Generic::Alloc( |
| 7100 | const VmaAllocationRequest& request, |
| 7101 | VmaSuballocationType type, |
| 7102 | void* userData) |
| 7103 | { |
| 7104 | VMA_ASSERT(request.type == VmaAllocationRequestType::Normal); |
| 7105 | VMA_ASSERT(request.item != m_Suballocations.end()); |
| 7106 | VmaSuballocation& suballoc = *request.item; |
| 7107 | // Given suballocation is a free block. |
| 7108 | VMA_ASSERT(suballoc.type == VMA_SUBALLOCATION_TYPE_FREE); |
| 7109 | |
| 7110 | // Given offset is inside this suballocation. |
| 7111 | VMA_ASSERT((VkDeviceSize)request.allocHandle - 1 >= suballoc.offset); |
| 7112 | const VkDeviceSize paddingBegin = (VkDeviceSize)request.allocHandle - suballoc.offset - 1; |
| 7113 | VMA_ASSERT(suballoc.size >= paddingBegin + request.size); |
| 7114 | const VkDeviceSize paddingEnd = suballoc.size - paddingBegin - request.size; |
| 7115 | |
| 7116 | // Unregister this free suballocation from m_FreeSuballocationsBySize and update |
| 7117 | // it to become used. |
| 7118 | UnregisterFreeSuballocation(request.item); |
| 7119 | |
| 7120 | suballoc.offset = (VkDeviceSize)request.allocHandle - 1; |
| 7121 | suballoc.size = request.size; |
| 7122 | suballoc.type = type; |
| 7123 | suballoc.userData = userData; |
| 7124 | |
| 7125 | // If there are any free bytes remaining at the end, insert new free suballocation after current one. |
| 7126 | if (paddingEnd) |
| 7127 | { |
| 7128 | VmaSuballocation paddingSuballoc = {}; |
| 7129 | paddingSuballoc.offset = suballoc.offset + suballoc.size; |
| 7130 | paddingSuballoc.size = paddingEnd; |
| 7131 | paddingSuballoc.type = VMA_SUBALLOCATION_TYPE_FREE; |
| 7132 | VmaSuballocationList::iterator next = request.item; |
| 7133 | ++next; |
| 7134 | const VmaSuballocationList::iterator paddingEndItem = |
| 7135 | m_Suballocations.insert(next, paddingSuballoc); |
| 7136 | RegisterFreeSuballocation(paddingEndItem); |
| 7137 | } |
| 7138 | |
| 7139 | // If there are any free bytes remaining at the beginning, insert new free suballocation before current one. |
| 7140 | if (paddingBegin) |
| 7141 | { |
| 7142 | VmaSuballocation paddingSuballoc = {}; |
| 7143 | paddingSuballoc.offset = suballoc.offset - paddingBegin; |
| 7144 | paddingSuballoc.size = paddingBegin; |
| 7145 | paddingSuballoc.type = VMA_SUBALLOCATION_TYPE_FREE; |
| 7146 | const VmaSuballocationList::iterator paddingBeginItem = |
| 7147 | m_Suballocations.insert(request.item, paddingSuballoc); |
| 7148 | RegisterFreeSuballocation(paddingBeginItem); |
| 7149 | } |
| 7150 | |
| 7151 | // Update totals. |
| 7152 | m_FreeCount = m_FreeCount - 1; |
| 7153 | if (paddingBegin > 0) |
| 7154 | { |
| 7155 | ++m_FreeCount; |
| 7156 | } |
| 7157 | if (paddingEnd > 0) |
| 7158 | { |
| 7159 | ++m_FreeCount; |
| 7160 | } |
| 7161 | m_SumFreeSize -= request.size; |
| 7162 | } |
| 7163 | |
| 7164 | void VmaBlockMetadata_Generic::GetAllocationInfo(VmaAllocHandle allocHandle, VmaVirtualAllocationInfo& outInfo) |
| 7165 | { |
| 7166 | outInfo.offset = (VkDeviceSize)allocHandle - 1; |
| 7167 | const VmaSuballocation& suballoc = *FindAtOffset(outInfo.offset); |
| 7168 | outInfo.size = suballoc.size; |
| 7169 | outInfo.pUserData = suballoc.userData; |
| 7170 | } |
| 7171 | |
| 7172 | void* VmaBlockMetadata_Generic::GetAllocationUserData(VmaAllocHandle allocHandle) const |
| 7173 | { |
| 7174 | return FindAtOffset((VkDeviceSize)allocHandle - 1)->userData; |
| 7175 | } |
| 7176 | |
| 7177 | VmaAllocHandle VmaBlockMetadata_Generic::GetAllocationListBegin() const |
| 7178 | { |
| 7179 | if (IsEmpty()) |
| 7180 | return VK_NULL_HANDLE; |
| 7181 | |
| 7182 | for (const auto& suballoc : m_Suballocations) |
| 7183 | { |
| 7184 | if (suballoc.type != VMA_SUBALLOCATION_TYPE_FREE) |
| 7185 | return (VmaAllocHandle)(suballoc.offset + 1); |
| 7186 | } |
| 7187 | VMA_ASSERT(false && "Should contain at least 1 allocation!" ); |
| 7188 | return VK_NULL_HANDLE; |
| 7189 | } |
| 7190 | |
| 7191 | VmaAllocHandle VmaBlockMetadata_Generic::GetNextAllocation(VmaAllocHandle prevAlloc) const |
| 7192 | { |
| 7193 | VmaSuballocationList::const_iterator prev = FindAtOffset((VkDeviceSize)prevAlloc - 1); |
| 7194 | |
| 7195 | for (VmaSuballocationList::const_iterator it = ++prev; it != m_Suballocations.end(); ++it) |
| 7196 | { |
| 7197 | if (it->type != VMA_SUBALLOCATION_TYPE_FREE) |
| 7198 | return (VmaAllocHandle)(it->offset + 1); |
| 7199 | } |
| 7200 | return VK_NULL_HANDLE; |
| 7201 | } |
| 7202 | |
| 7203 | void VmaBlockMetadata_Generic::Clear() |
| 7204 | { |
| 7205 | const VkDeviceSize size = GetSize(); |
| 7206 | |
| 7207 | VMA_ASSERT(IsVirtual()); |
| 7208 | m_FreeCount = 1; |
| 7209 | m_SumFreeSize = size; |
| 7210 | m_Suballocations.clear(); |
| 7211 | m_FreeSuballocationsBySize.clear(); |
| 7212 | |
| 7213 | VmaSuballocation suballoc = {}; |
| 7214 | suballoc.offset = 0; |
| 7215 | suballoc.size = size; |
| 7216 | suballoc.type = VMA_SUBALLOCATION_TYPE_FREE; |
| 7217 | m_Suballocations.push_back(suballoc); |
| 7218 | |
| 7219 | m_FreeSuballocationsBySize.push_back(m_Suballocations.begin()); |
| 7220 | } |
| 7221 | |
| 7222 | void VmaBlockMetadata_Generic::SetAllocationUserData(VmaAllocHandle allocHandle, void* userData) |
| 7223 | { |
| 7224 | VmaSuballocation& suballoc = *FindAtOffset((VkDeviceSize)allocHandle - 1); |
| 7225 | suballoc.userData = userData; |
| 7226 | } |
| 7227 | |
| 7228 | void VmaBlockMetadata_Generic::DebugLogAllAllocations() const |
| 7229 | { |
| 7230 | for (const auto& suballoc : m_Suballocations) |
| 7231 | { |
| 7232 | if (suballoc.type != VMA_SUBALLOCATION_TYPE_FREE) |
| 7233 | DebugLogAllocation(suballoc.offset, suballoc.size, suballoc.userData); |
| 7234 | } |
| 7235 | } |
| 7236 | |
| 7237 | VmaSuballocationList::iterator VmaBlockMetadata_Generic::FindAtOffset(VkDeviceSize offset) const |
| 7238 | { |
| 7239 | VMA_HEAVY_ASSERT(!m_Suballocations.empty()); |
| 7240 | const VkDeviceSize last = m_Suballocations.rbegin()->offset; |
| 7241 | if (last == offset) |
| 7242 | return m_Suballocations.rbegin().drop_const(); |
| 7243 | const VkDeviceSize first = m_Suballocations.begin()->offset; |
| 7244 | if (first == offset) |
| 7245 | return m_Suballocations.begin().drop_const(); |
| 7246 | |
| 7247 | const size_t suballocCount = m_Suballocations.size(); |
| 7248 | const VkDeviceSize step = (last - first + m_Suballocations.begin()->size) / suballocCount; |
| 7249 | auto findSuballocation = [&](auto begin, auto end) -> VmaSuballocationList::iterator |
| 7250 | { |
| 7251 | for (auto suballocItem = begin; |
| 7252 | suballocItem != end; |
| 7253 | ++suballocItem) |
| 7254 | { |
| 7255 | if (suballocItem->offset == offset) |
| 7256 | return suballocItem.drop_const(); |
| 7257 | } |
| 7258 | VMA_ASSERT(false && "Not found!" ); |
| 7259 | return m_Suballocations.end().drop_const(); |
| 7260 | }; |
| 7261 | // If requested offset is closer to the end of range, search from the end |
| 7262 | if (offset - first > suballocCount * step / 2) |
| 7263 | { |
| 7264 | return findSuballocation(m_Suballocations.rbegin(), m_Suballocations.rend()); |
| 7265 | } |
| 7266 | return findSuballocation(m_Suballocations.begin(), m_Suballocations.end()); |
| 7267 | } |
| 7268 | |
| 7269 | bool VmaBlockMetadata_Generic::ValidateFreeSuballocationList() const |
| 7270 | { |
| 7271 | VkDeviceSize lastSize = 0; |
| 7272 | for (size_t i = 0, count = m_FreeSuballocationsBySize.size(); i < count; ++i) |
| 7273 | { |
| 7274 | const VmaSuballocationList::iterator it = m_FreeSuballocationsBySize[i]; |
| 7275 | |
| 7276 | VMA_VALIDATE(it->type == VMA_SUBALLOCATION_TYPE_FREE); |
| 7277 | VMA_VALIDATE(it->size >= lastSize); |
| 7278 | lastSize = it->size; |
| 7279 | } |
| 7280 | return true; |
| 7281 | } |
| 7282 | |
| 7283 | bool VmaBlockMetadata_Generic::CheckAllocation( |
| 7284 | VkDeviceSize allocSize, |
| 7285 | VkDeviceSize allocAlignment, |
| 7286 | VmaSuballocationType allocType, |
| 7287 | VmaSuballocationList::const_iterator suballocItem, |
| 7288 | VmaAllocHandle* pAllocHandle) const |
| 7289 | { |
| 7290 | VMA_ASSERT(allocSize > 0); |
| 7291 | VMA_ASSERT(allocType != VMA_SUBALLOCATION_TYPE_FREE); |
| 7292 | VMA_ASSERT(suballocItem != m_Suballocations.cend()); |
| 7293 | VMA_ASSERT(pAllocHandle != VMA_NULL); |
| 7294 | |
| 7295 | const VkDeviceSize debugMargin = GetDebugMargin(); |
| 7296 | const VkDeviceSize bufferImageGranularity = GetBufferImageGranularity(); |
| 7297 | |
| 7298 | const VmaSuballocation& suballoc = *suballocItem; |
| 7299 | VMA_ASSERT(suballoc.type == VMA_SUBALLOCATION_TYPE_FREE); |
| 7300 | |
| 7301 | // Size of this suballocation is too small for this request: Early return. |
| 7302 | if (suballoc.size < allocSize) |
| 7303 | { |
| 7304 | return false; |
| 7305 | } |
| 7306 | |
| 7307 | // Start from offset equal to beginning of this suballocation. |
| 7308 | VkDeviceSize offset = suballoc.offset + (suballocItem == m_Suballocations.cbegin() ? 0 : GetDebugMargin()); |
| 7309 | |
| 7310 | // Apply debugMargin from the end of previous alloc. |
| 7311 | if (debugMargin > 0) |
| 7312 | { |
| 7313 | offset += debugMargin; |
| 7314 | } |
| 7315 | |
| 7316 | // Apply alignment. |
| 7317 | offset = VmaAlignUp(offset, allocAlignment); |
| 7318 | |
| 7319 | // Check previous suballocations for BufferImageGranularity conflicts. |
| 7320 | // Make bigger alignment if necessary. |
| 7321 | if (bufferImageGranularity > 1 && bufferImageGranularity != allocAlignment) |
| 7322 | { |
| 7323 | bool bufferImageGranularityConflict = false; |
| 7324 | VmaSuballocationList::const_iterator prevSuballocItem = suballocItem; |
| 7325 | while (prevSuballocItem != m_Suballocations.cbegin()) |
| 7326 | { |
| 7327 | --prevSuballocItem; |
| 7328 | const VmaSuballocation& prevSuballoc = *prevSuballocItem; |
| 7329 | if (VmaBlocksOnSamePage(prevSuballoc.offset, prevSuballoc.size, offset, bufferImageGranularity)) |
| 7330 | { |
| 7331 | if (VmaIsBufferImageGranularityConflict(prevSuballoc.type, allocType)) |
| 7332 | { |
| 7333 | bufferImageGranularityConflict = true; |
| 7334 | break; |
| 7335 | } |
| 7336 | } |
| 7337 | else |
| 7338 | // Already on previous page. |
| 7339 | break; |
| 7340 | } |
| 7341 | if (bufferImageGranularityConflict) |
| 7342 | { |
| 7343 | offset = VmaAlignUp(offset, bufferImageGranularity); |
| 7344 | } |
| 7345 | } |
| 7346 | |
| 7347 | // Calculate padding at the beginning based on current offset. |
| 7348 | const VkDeviceSize paddingBegin = offset - suballoc.offset; |
| 7349 | |
| 7350 | // Fail if requested size plus margin after is bigger than size of this suballocation. |
| 7351 | if (paddingBegin + allocSize + debugMargin > suballoc.size) |
| 7352 | { |
| 7353 | return false; |
| 7354 | } |
| 7355 | |
| 7356 | // Check next suballocations for BufferImageGranularity conflicts. |
| 7357 | // If conflict exists, allocation cannot be made here. |
| 7358 | if (allocSize % bufferImageGranularity || offset % bufferImageGranularity) |
| 7359 | { |
| 7360 | VmaSuballocationList::const_iterator nextSuballocItem = suballocItem; |
| 7361 | ++nextSuballocItem; |
| 7362 | while (nextSuballocItem != m_Suballocations.cend()) |
| 7363 | { |
| 7364 | const VmaSuballocation& nextSuballoc = *nextSuballocItem; |
| 7365 | if (VmaBlocksOnSamePage(offset, allocSize, nextSuballoc.offset, bufferImageGranularity)) |
| 7366 | { |
| 7367 | if (VmaIsBufferImageGranularityConflict(allocType, nextSuballoc.type)) |
| 7368 | { |
| 7369 | return false; |
| 7370 | } |
| 7371 | } |
| 7372 | else |
| 7373 | { |
| 7374 | // Already on next page. |
| 7375 | break; |
| 7376 | } |
| 7377 | ++nextSuballocItem; |
| 7378 | } |
| 7379 | } |
| 7380 | |
| 7381 | *pAllocHandle = (VmaAllocHandle)(offset + 1); |
| 7382 | // All tests passed: Success. pAllocHandle is already filled. |
| 7383 | return true; |
| 7384 | } |
| 7385 | |
| 7386 | void VmaBlockMetadata_Generic::MergeFreeWithNext(VmaSuballocationList::iterator item) |
| 7387 | { |
| 7388 | VMA_ASSERT(item != m_Suballocations.end()); |
| 7389 | VMA_ASSERT(item->type == VMA_SUBALLOCATION_TYPE_FREE); |
| 7390 | |
| 7391 | VmaSuballocationList::iterator nextItem = item; |
| 7392 | ++nextItem; |
| 7393 | VMA_ASSERT(nextItem != m_Suballocations.end()); |
| 7394 | VMA_ASSERT(nextItem->type == VMA_SUBALLOCATION_TYPE_FREE); |
| 7395 | |
| 7396 | item->size += nextItem->size; |
| 7397 | --m_FreeCount; |
| 7398 | m_Suballocations.erase(nextItem); |
| 7399 | } |
| 7400 | |
| 7401 | VmaSuballocationList::iterator VmaBlockMetadata_Generic::FreeSuballocation(VmaSuballocationList::iterator suballocItem) |
| 7402 | { |
| 7403 | // Change this suballocation to be marked as free. |
| 7404 | VmaSuballocation& suballoc = *suballocItem; |
| 7405 | suballoc.type = VMA_SUBALLOCATION_TYPE_FREE; |
| 7406 | suballoc.userData = VMA_NULL; |
| 7407 | |
| 7408 | // Update totals. |
| 7409 | ++m_FreeCount; |
| 7410 | m_SumFreeSize += suballoc.size; |
| 7411 | |
| 7412 | // Merge with previous and/or next suballocation if it's also free. |
| 7413 | bool mergeWithNext = false; |
| 7414 | bool mergeWithPrev = false; |
| 7415 | |
| 7416 | VmaSuballocationList::iterator nextItem = suballocItem; |
| 7417 | ++nextItem; |
| 7418 | if ((nextItem != m_Suballocations.end()) && (nextItem->type == VMA_SUBALLOCATION_TYPE_FREE)) |
| 7419 | { |
| 7420 | mergeWithNext = true; |
| 7421 | } |
| 7422 | |
| 7423 | VmaSuballocationList::iterator prevItem = suballocItem; |
| 7424 | if (suballocItem != m_Suballocations.begin()) |
| 7425 | { |
| 7426 | --prevItem; |
| 7427 | if (prevItem->type == VMA_SUBALLOCATION_TYPE_FREE) |
| 7428 | { |
| 7429 | mergeWithPrev = true; |
| 7430 | } |
| 7431 | } |
| 7432 | |
| 7433 | if (mergeWithNext) |
| 7434 | { |
| 7435 | UnregisterFreeSuballocation(nextItem); |
| 7436 | MergeFreeWithNext(suballocItem); |
| 7437 | } |
| 7438 | |
| 7439 | if (mergeWithPrev) |
| 7440 | { |
| 7441 | UnregisterFreeSuballocation(prevItem); |
| 7442 | MergeFreeWithNext(prevItem); |
| 7443 | RegisterFreeSuballocation(prevItem); |
| 7444 | return prevItem; |
| 7445 | } |
| 7446 | else |
| 7447 | { |
| 7448 | RegisterFreeSuballocation(suballocItem); |
| 7449 | return suballocItem; |
| 7450 | } |
| 7451 | } |
| 7452 | |
| 7453 | void VmaBlockMetadata_Generic::RegisterFreeSuballocation(VmaSuballocationList::iterator item) |
| 7454 | { |
| 7455 | VMA_ASSERT(item->type == VMA_SUBALLOCATION_TYPE_FREE); |
| 7456 | VMA_ASSERT(item->size > 0); |
| 7457 | |
| 7458 | // You may want to enable this validation at the beginning or at the end of |
| 7459 | // this function, depending on what do you want to check. |
| 7460 | VMA_HEAVY_ASSERT(ValidateFreeSuballocationList()); |
| 7461 | |
| 7462 | if (m_FreeSuballocationsBySize.empty()) |
| 7463 | { |
| 7464 | m_FreeSuballocationsBySize.push_back(item); |
| 7465 | } |
| 7466 | else |
| 7467 | { |
| 7468 | VmaVectorInsertSorted<VmaSuballocationItemSizeLess>(m_FreeSuballocationsBySize, item); |
| 7469 | } |
| 7470 | |
| 7471 | //VMA_HEAVY_ASSERT(ValidateFreeSuballocationList()); |
| 7472 | } |
| 7473 | |
| 7474 | void VmaBlockMetadata_Generic::UnregisterFreeSuballocation(VmaSuballocationList::iterator item) |
| 7475 | { |
| 7476 | VMA_ASSERT(item->type == VMA_SUBALLOCATION_TYPE_FREE); |
| 7477 | VMA_ASSERT(item->size > 0); |
| 7478 | |
| 7479 | // You may want to enable this validation at the beginning or at the end of |
| 7480 | // this function, depending on what do you want to check. |
| 7481 | VMA_HEAVY_ASSERT(ValidateFreeSuballocationList()); |
| 7482 | |
| 7483 | VmaSuballocationList::iterator* const it = VmaBinaryFindFirstNotLess( |
| 7484 | m_FreeSuballocationsBySize.data(), |
| 7485 | m_FreeSuballocationsBySize.data() + m_FreeSuballocationsBySize.size(), |
| 7486 | item, |
| 7487 | VmaSuballocationItemSizeLess()); |
| 7488 | for (size_t index = it - m_FreeSuballocationsBySize.data(); |
| 7489 | index < m_FreeSuballocationsBySize.size(); |
| 7490 | ++index) |
| 7491 | { |
| 7492 | if (m_FreeSuballocationsBySize[index] == item) |
| 7493 | { |
| 7494 | VmaVectorRemove(m_FreeSuballocationsBySize, index); |
| 7495 | return; |
| 7496 | } |
| 7497 | VMA_ASSERT((m_FreeSuballocationsBySize[index]->size == item->size) && "Not found." ); |
| 7498 | } |
| 7499 | VMA_ASSERT(0 && "Not found." ); |
| 7500 | |
| 7501 | //VMA_HEAVY_ASSERT(ValidateFreeSuballocationList()); |
| 7502 | } |
| 7503 | #endif // _VMA_BLOCK_METADATA_GENERIC_FUNCTIONS |
| 7504 | #endif // _VMA_BLOCK_METADATA_GENERIC |
| 7505 | #endif // #if 0 |
| 7506 | |
| 7507 | #ifndef _VMA_BLOCK_METADATA_LINEAR |
| 7508 | /* |
| 7509 | Allocations and their references in internal data structure look like this: |
| 7510 | |
| 7511 | if(m_2ndVectorMode == SECOND_VECTOR_EMPTY): |
| 7512 | |
| 7513 | 0 +-------+ |
| 7514 | | | |
| 7515 | | | |
| 7516 | | | |
| 7517 | +-------+ |
| 7518 | | Alloc | 1st[m_1stNullItemsBeginCount] |
| 7519 | +-------+ |
| 7520 | | Alloc | 1st[m_1stNullItemsBeginCount + 1] |
| 7521 | +-------+ |
| 7522 | | ... | |
| 7523 | +-------+ |
| 7524 | | Alloc | 1st[1st.size() - 1] |
| 7525 | +-------+ |
| 7526 | | | |
| 7527 | | | |
| 7528 | | | |
| 7529 | GetSize() +-------+ |
| 7530 | |
| 7531 | if(m_2ndVectorMode == SECOND_VECTOR_RING_BUFFER): |
| 7532 | |
| 7533 | 0 +-------+ |
| 7534 | | Alloc | 2nd[0] |
| 7535 | +-------+ |
| 7536 | | Alloc | 2nd[1] |
| 7537 | +-------+ |
| 7538 | | ... | |
| 7539 | +-------+ |
| 7540 | | Alloc | 2nd[2nd.size() - 1] |
| 7541 | +-------+ |
| 7542 | | | |
| 7543 | | | |
| 7544 | | | |
| 7545 | +-------+ |
| 7546 | | Alloc | 1st[m_1stNullItemsBeginCount] |
| 7547 | +-------+ |
| 7548 | | Alloc | 1st[m_1stNullItemsBeginCount + 1] |
| 7549 | +-------+ |
| 7550 | | ... | |
| 7551 | +-------+ |
| 7552 | | Alloc | 1st[1st.size() - 1] |
| 7553 | +-------+ |
| 7554 | | | |
| 7555 | GetSize() +-------+ |
| 7556 | |
| 7557 | if(m_2ndVectorMode == SECOND_VECTOR_DOUBLE_STACK): |
| 7558 | |
| 7559 | 0 +-------+ |
| 7560 | | | |
| 7561 | | | |
| 7562 | | | |
| 7563 | +-------+ |
| 7564 | | Alloc | 1st[m_1stNullItemsBeginCount] |
| 7565 | +-------+ |
| 7566 | | Alloc | 1st[m_1stNullItemsBeginCount + 1] |
| 7567 | +-------+ |
| 7568 | | ... | |
| 7569 | +-------+ |
| 7570 | | Alloc | 1st[1st.size() - 1] |
| 7571 | +-------+ |
| 7572 | | | |
| 7573 | | | |
| 7574 | | | |
| 7575 | +-------+ |
| 7576 | | Alloc | 2nd[2nd.size() - 1] |
| 7577 | +-------+ |
| 7578 | | ... | |
| 7579 | +-------+ |
| 7580 | | Alloc | 2nd[1] |
| 7581 | +-------+ |
| 7582 | | Alloc | 2nd[0] |
| 7583 | GetSize() +-------+ |
| 7584 | |
| 7585 | */ |
| 7586 | class VmaBlockMetadata_Linear : public VmaBlockMetadata |
| 7587 | { |
| 7588 | VMA_CLASS_NO_COPY(VmaBlockMetadata_Linear) |
| 7589 | public: |
| 7590 | VmaBlockMetadata_Linear(const VkAllocationCallbacks* pAllocationCallbacks, |
| 7591 | VkDeviceSize bufferImageGranularity, bool isVirtual); |
| 7592 | virtual ~VmaBlockMetadata_Linear() = default; |
| 7593 | |
| 7594 | VkDeviceSize GetSumFreeSize() const override { return m_SumFreeSize; } |
| 7595 | bool IsEmpty() const override { return GetAllocationCount() == 0; } |
| 7596 | VkDeviceSize GetAllocationOffset(VmaAllocHandle allocHandle) const override { return (VkDeviceSize)allocHandle - 1; }; |
| 7597 | |
| 7598 | void Init(VkDeviceSize size) override; |
| 7599 | bool Validate() const override; |
| 7600 | size_t GetAllocationCount() const override; |
| 7601 | size_t GetFreeRegionsCount() const override; |
| 7602 | |
| 7603 | void AddDetailedStatistics(VmaDetailedStatistics& inoutStats) const override; |
| 7604 | void AddStatistics(VmaStatistics& inoutStats) const override; |
| 7605 | |
| 7606 | #if VMA_STATS_STRING_ENABLED |
| 7607 | void PrintDetailedMap(class VmaJsonWriter& json) const override; |
| 7608 | #endif |
| 7609 | |
| 7610 | bool CreateAllocationRequest( |
| 7611 | VkDeviceSize allocSize, |
| 7612 | VkDeviceSize allocAlignment, |
| 7613 | bool upperAddress, |
| 7614 | VmaSuballocationType allocType, |
| 7615 | uint32_t strategy, |
| 7616 | VmaAllocationRequest* pAllocationRequest) override; |
| 7617 | |
| 7618 | VkResult CheckCorruption(const void* pBlockData) override; |
| 7619 | |
| 7620 | void Alloc( |
| 7621 | const VmaAllocationRequest& request, |
| 7622 | VmaSuballocationType type, |
| 7623 | void* userData) override; |
| 7624 | |
| 7625 | void Free(VmaAllocHandle allocHandle) override; |
| 7626 | void GetAllocationInfo(VmaAllocHandle allocHandle, VmaVirtualAllocationInfo& outInfo) override; |
| 7627 | void* GetAllocationUserData(VmaAllocHandle allocHandle) const override; |
| 7628 | VmaAllocHandle GetAllocationListBegin() const override; |
| 7629 | VmaAllocHandle GetNextAllocation(VmaAllocHandle prevAlloc) const override; |
| 7630 | VkDeviceSize GetNextFreeRegionSize(VmaAllocHandle alloc) const override; |
| 7631 | void Clear() override; |
| 7632 | void SetAllocationUserData(VmaAllocHandle allocHandle, void* userData) override; |
| 7633 | void DebugLogAllAllocations() const override; |
| 7634 | |
| 7635 | private: |
| 7636 | /* |
| 7637 | There are two suballocation vectors, used in ping-pong way. |
| 7638 | The one with index m_1stVectorIndex is called 1st. |
| 7639 | The one with index (m_1stVectorIndex ^ 1) is called 2nd. |
| 7640 | 2nd can be non-empty only when 1st is not empty. |
| 7641 | When 2nd is not empty, m_2ndVectorMode indicates its mode of operation. |
| 7642 | */ |
| 7643 | typedef VmaVector<VmaSuballocation, VmaStlAllocator<VmaSuballocation>> SuballocationVectorType; |
| 7644 | |
| 7645 | enum SECOND_VECTOR_MODE |
| 7646 | { |
| 7647 | SECOND_VECTOR_EMPTY, |
| 7648 | /* |
| 7649 | Suballocations in 2nd vector are created later than the ones in 1st, but they |
| 7650 | all have smaller offset. |
| 7651 | */ |
| 7652 | SECOND_VECTOR_RING_BUFFER, |
| 7653 | /* |
| 7654 | Suballocations in 2nd vector are upper side of double stack. |
| 7655 | They all have offsets higher than those in 1st vector. |
| 7656 | Top of this stack means smaller offsets, but higher indices in this vector. |
| 7657 | */ |
| 7658 | SECOND_VECTOR_DOUBLE_STACK, |
| 7659 | }; |
| 7660 | |
| 7661 | VkDeviceSize m_SumFreeSize; |
| 7662 | SuballocationVectorType m_Suballocations0, m_Suballocations1; |
| 7663 | uint32_t m_1stVectorIndex; |
| 7664 | SECOND_VECTOR_MODE m_2ndVectorMode; |
| 7665 | // Number of items in 1st vector with hAllocation = null at the beginning. |
| 7666 | size_t m_1stNullItemsBeginCount; |
| 7667 | // Number of other items in 1st vector with hAllocation = null somewhere in the middle. |
| 7668 | size_t m_1stNullItemsMiddleCount; |
| 7669 | // Number of items in 2nd vector with hAllocation = null. |
| 7670 | size_t m_2ndNullItemsCount; |
| 7671 | |
| 7672 | SuballocationVectorType& AccessSuballocations1st() { return m_1stVectorIndex ? m_Suballocations1 : m_Suballocations0; } |
| 7673 | SuballocationVectorType& AccessSuballocations2nd() { return m_1stVectorIndex ? m_Suballocations0 : m_Suballocations1; } |
| 7674 | const SuballocationVectorType& AccessSuballocations1st() const { return m_1stVectorIndex ? m_Suballocations1 : m_Suballocations0; } |
| 7675 | const SuballocationVectorType& AccessSuballocations2nd() const { return m_1stVectorIndex ? m_Suballocations0 : m_Suballocations1; } |
| 7676 | |
| 7677 | VmaSuballocation& FindSuballocation(VkDeviceSize offset) const; |
| 7678 | bool ShouldCompact1st() const; |
| 7679 | void CleanupAfterFree(); |
| 7680 | |
| 7681 | bool CreateAllocationRequest_LowerAddress( |
| 7682 | VkDeviceSize allocSize, |
| 7683 | VkDeviceSize allocAlignment, |
| 7684 | VmaSuballocationType allocType, |
| 7685 | uint32_t strategy, |
| 7686 | VmaAllocationRequest* pAllocationRequest); |
| 7687 | bool CreateAllocationRequest_UpperAddress( |
| 7688 | VkDeviceSize allocSize, |
| 7689 | VkDeviceSize allocAlignment, |
| 7690 | VmaSuballocationType allocType, |
| 7691 | uint32_t strategy, |
| 7692 | VmaAllocationRequest* pAllocationRequest); |
| 7693 | }; |
| 7694 | |
| 7695 | #ifndef _VMA_BLOCK_METADATA_LINEAR_FUNCTIONS |
| 7696 | VmaBlockMetadata_Linear::VmaBlockMetadata_Linear(const VkAllocationCallbacks* pAllocationCallbacks, |
| 7697 | VkDeviceSize bufferImageGranularity, bool isVirtual) |
| 7698 | : VmaBlockMetadata(pAllocationCallbacks, bufferImageGranularity, isVirtual), |
| 7699 | m_SumFreeSize(0), |
| 7700 | m_Suballocations0(VmaStlAllocator<VmaSuballocation>(pAllocationCallbacks)), |
| 7701 | m_Suballocations1(VmaStlAllocator<VmaSuballocation>(pAllocationCallbacks)), |
| 7702 | m_1stVectorIndex(0), |
| 7703 | m_2ndVectorMode(SECOND_VECTOR_EMPTY), |
| 7704 | m_1stNullItemsBeginCount(0), |
| 7705 | m_1stNullItemsMiddleCount(0), |
| 7706 | m_2ndNullItemsCount(0) {} |
| 7707 | |
| 7708 | void VmaBlockMetadata_Linear::Init(VkDeviceSize size) |
| 7709 | { |
| 7710 | VmaBlockMetadata::Init(size); |
| 7711 | m_SumFreeSize = size; |
| 7712 | } |
| 7713 | |
| 7714 | bool VmaBlockMetadata_Linear::Validate() const |
| 7715 | { |
| 7716 | const SuballocationVectorType& suballocations1st = AccessSuballocations1st(); |
| 7717 | const SuballocationVectorType& suballocations2nd = AccessSuballocations2nd(); |
| 7718 | |
| 7719 | VMA_VALIDATE(suballocations2nd.empty() == (m_2ndVectorMode == SECOND_VECTOR_EMPTY)); |
| 7720 | VMA_VALIDATE(!suballocations1st.empty() || |
| 7721 | suballocations2nd.empty() || |
| 7722 | m_2ndVectorMode != SECOND_VECTOR_RING_BUFFER); |
| 7723 | |
| 7724 | if (!suballocations1st.empty()) |
| 7725 | { |
| 7726 | // Null item at the beginning should be accounted into m_1stNullItemsBeginCount. |
| 7727 | VMA_VALIDATE(suballocations1st[m_1stNullItemsBeginCount].type != VMA_SUBALLOCATION_TYPE_FREE); |
| 7728 | // Null item at the end should be just pop_back(). |
| 7729 | VMA_VALIDATE(suballocations1st.back().type != VMA_SUBALLOCATION_TYPE_FREE); |
| 7730 | } |
| 7731 | if (!suballocations2nd.empty()) |
| 7732 | { |
| 7733 | // Null item at the end should be just pop_back(). |
| 7734 | VMA_VALIDATE(suballocations2nd.back().type != VMA_SUBALLOCATION_TYPE_FREE); |
| 7735 | } |
| 7736 | |
| 7737 | VMA_VALIDATE(m_1stNullItemsBeginCount + m_1stNullItemsMiddleCount <= suballocations1st.size()); |
| 7738 | VMA_VALIDATE(m_2ndNullItemsCount <= suballocations2nd.size()); |
| 7739 | |
| 7740 | VkDeviceSize sumUsedSize = 0; |
| 7741 | const size_t suballoc1stCount = suballocations1st.size(); |
| 7742 | const VkDeviceSize debugMargin = GetDebugMargin(); |
| 7743 | VkDeviceSize offset = 0; |
| 7744 | |
| 7745 | if (m_2ndVectorMode == SECOND_VECTOR_RING_BUFFER) |
| 7746 | { |
| 7747 | const size_t suballoc2ndCount = suballocations2nd.size(); |
| 7748 | size_t nullItem2ndCount = 0; |
| 7749 | for (size_t i = 0; i < suballoc2ndCount; ++i) |
| 7750 | { |
| 7751 | const VmaSuballocation& suballoc = suballocations2nd[i]; |
| 7752 | const bool currFree = (suballoc.type == VMA_SUBALLOCATION_TYPE_FREE); |
| 7753 | |
| 7754 | VmaAllocation const alloc = (VmaAllocation)suballoc.userData; |
| 7755 | if (!IsVirtual()) |
| 7756 | { |
| 7757 | VMA_VALIDATE(currFree == (alloc == VK_NULL_HANDLE)); |
| 7758 | } |
| 7759 | VMA_VALIDATE(suballoc.offset >= offset); |
| 7760 | |
| 7761 | if (!currFree) |
| 7762 | { |
| 7763 | if (!IsVirtual()) |
| 7764 | { |
| 7765 | VMA_VALIDATE((VkDeviceSize)alloc->GetAllocHandle() == suballoc.offset + 1); |
| 7766 | VMA_VALIDATE(alloc->GetSize() == suballoc.size); |
| 7767 | } |
| 7768 | sumUsedSize += suballoc.size; |
| 7769 | } |
| 7770 | else |
| 7771 | { |
| 7772 | ++nullItem2ndCount; |
| 7773 | } |
| 7774 | |
| 7775 | offset = suballoc.offset + suballoc.size + debugMargin; |
| 7776 | } |
| 7777 | |
| 7778 | VMA_VALIDATE(nullItem2ndCount == m_2ndNullItemsCount); |
| 7779 | } |
| 7780 | |
| 7781 | for (size_t i = 0; i < m_1stNullItemsBeginCount; ++i) |
| 7782 | { |
| 7783 | const VmaSuballocation& suballoc = suballocations1st[i]; |
| 7784 | VMA_VALIDATE(suballoc.type == VMA_SUBALLOCATION_TYPE_FREE && |
| 7785 | suballoc.userData == VMA_NULL); |
| 7786 | } |
| 7787 | |
| 7788 | size_t nullItem1stCount = m_1stNullItemsBeginCount; |
| 7789 | |
| 7790 | for (size_t i = m_1stNullItemsBeginCount; i < suballoc1stCount; ++i) |
| 7791 | { |
| 7792 | const VmaSuballocation& suballoc = suballocations1st[i]; |
| 7793 | const bool currFree = (suballoc.type == VMA_SUBALLOCATION_TYPE_FREE); |
| 7794 | |
| 7795 | VmaAllocation const alloc = (VmaAllocation)suballoc.userData; |
| 7796 | if (!IsVirtual()) |
| 7797 | { |
| 7798 | VMA_VALIDATE(currFree == (alloc == VK_NULL_HANDLE)); |
| 7799 | } |
| 7800 | VMA_VALIDATE(suballoc.offset >= offset); |
| 7801 | VMA_VALIDATE(i >= m_1stNullItemsBeginCount || currFree); |
| 7802 | |
| 7803 | if (!currFree) |
| 7804 | { |
| 7805 | if (!IsVirtual()) |
| 7806 | { |
| 7807 | VMA_VALIDATE((VkDeviceSize)alloc->GetAllocHandle() == suballoc.offset + 1); |
| 7808 | VMA_VALIDATE(alloc->GetSize() == suballoc.size); |
| 7809 | } |
| 7810 | sumUsedSize += suballoc.size; |
| 7811 | } |
| 7812 | else |
| 7813 | { |
| 7814 | ++nullItem1stCount; |
| 7815 | } |
| 7816 | |
| 7817 | offset = suballoc.offset + suballoc.size + debugMargin; |
| 7818 | } |
| 7819 | VMA_VALIDATE(nullItem1stCount == m_1stNullItemsBeginCount + m_1stNullItemsMiddleCount); |
| 7820 | |
| 7821 | if (m_2ndVectorMode == SECOND_VECTOR_DOUBLE_STACK) |
| 7822 | { |
| 7823 | const size_t suballoc2ndCount = suballocations2nd.size(); |
| 7824 | size_t nullItem2ndCount = 0; |
| 7825 | for (size_t i = suballoc2ndCount; i--; ) |
| 7826 | { |
| 7827 | const VmaSuballocation& suballoc = suballocations2nd[i]; |
| 7828 | const bool currFree = (suballoc.type == VMA_SUBALLOCATION_TYPE_FREE); |
| 7829 | |
| 7830 | VmaAllocation const alloc = (VmaAllocation)suballoc.userData; |
| 7831 | if (!IsVirtual()) |
| 7832 | { |
| 7833 | VMA_VALIDATE(currFree == (alloc == VK_NULL_HANDLE)); |
| 7834 | } |
| 7835 | VMA_VALIDATE(suballoc.offset >= offset); |
| 7836 | |
| 7837 | if (!currFree) |
| 7838 | { |
| 7839 | if (!IsVirtual()) |
| 7840 | { |
| 7841 | VMA_VALIDATE((VkDeviceSize)alloc->GetAllocHandle() == suballoc.offset + 1); |
| 7842 | VMA_VALIDATE(alloc->GetSize() == suballoc.size); |
| 7843 | } |
| 7844 | sumUsedSize += suballoc.size; |
| 7845 | } |
| 7846 | else |
| 7847 | { |
| 7848 | ++nullItem2ndCount; |
| 7849 | } |
| 7850 | |
| 7851 | offset = suballoc.offset + suballoc.size + debugMargin; |
| 7852 | } |
| 7853 | |
| 7854 | VMA_VALIDATE(nullItem2ndCount == m_2ndNullItemsCount); |
| 7855 | } |
| 7856 | |
| 7857 | VMA_VALIDATE(offset <= GetSize()); |
| 7858 | VMA_VALIDATE(m_SumFreeSize == GetSize() - sumUsedSize); |
| 7859 | |
| 7860 | return true; |
| 7861 | } |
| 7862 | |
| 7863 | size_t VmaBlockMetadata_Linear::GetAllocationCount() const |
| 7864 | { |
| 7865 | return AccessSuballocations1st().size() - m_1stNullItemsBeginCount - m_1stNullItemsMiddleCount + |
| 7866 | AccessSuballocations2nd().size() - m_2ndNullItemsCount; |
| 7867 | } |
| 7868 | |
| 7869 | size_t VmaBlockMetadata_Linear::GetFreeRegionsCount() const |
| 7870 | { |
| 7871 | // Function only used for defragmentation, which is disabled for this algorithm |
| 7872 | VMA_ASSERT(0); |
| 7873 | return SIZE_MAX; |
| 7874 | } |
| 7875 | |
| 7876 | void VmaBlockMetadata_Linear::AddDetailedStatistics(VmaDetailedStatistics& inoutStats) const |
| 7877 | { |
| 7878 | const VkDeviceSize size = GetSize(); |
| 7879 | const SuballocationVectorType& suballocations1st = AccessSuballocations1st(); |
| 7880 | const SuballocationVectorType& suballocations2nd = AccessSuballocations2nd(); |
| 7881 | const size_t suballoc1stCount = suballocations1st.size(); |
| 7882 | const size_t suballoc2ndCount = suballocations2nd.size(); |
| 7883 | |
| 7884 | inoutStats.statistics.blockCount++; |
| 7885 | inoutStats.statistics.blockBytes += size; |
| 7886 | |
| 7887 | VkDeviceSize lastOffset = 0; |
| 7888 | |
| 7889 | if (m_2ndVectorMode == SECOND_VECTOR_RING_BUFFER) |
| 7890 | { |
| 7891 | const VkDeviceSize freeSpace2ndTo1stEnd = suballocations1st[m_1stNullItemsBeginCount].offset; |
| 7892 | size_t nextAlloc2ndIndex = 0; |
| 7893 | while (lastOffset < freeSpace2ndTo1stEnd) |
| 7894 | { |
| 7895 | // Find next non-null allocation or move nextAllocIndex to the end. |
| 7896 | while (nextAlloc2ndIndex < suballoc2ndCount && |
| 7897 | suballocations2nd[nextAlloc2ndIndex].userData == VMA_NULL) |
| 7898 | { |
| 7899 | ++nextAlloc2ndIndex; |
| 7900 | } |
| 7901 | |
| 7902 | // Found non-null allocation. |
| 7903 | if (nextAlloc2ndIndex < suballoc2ndCount) |
| 7904 | { |
| 7905 | const VmaSuballocation& suballoc = suballocations2nd[nextAlloc2ndIndex]; |
| 7906 | |
| 7907 | // 1. Process free space before this allocation. |
| 7908 | if (lastOffset < suballoc.offset) |
| 7909 | { |
| 7910 | // There is free space from lastOffset to suballoc.offset. |
| 7911 | const VkDeviceSize unusedRangeSize = suballoc.offset - lastOffset; |
| 7912 | VmaAddDetailedStatisticsUnusedRange(inoutStats, size: unusedRangeSize); |
| 7913 | } |
| 7914 | |
| 7915 | // 2. Process this allocation. |
| 7916 | // There is allocation with suballoc.offset, suballoc.size. |
| 7917 | VmaAddDetailedStatisticsAllocation(inoutStats, size: suballoc.size); |
| 7918 | |
| 7919 | // 3. Prepare for next iteration. |
| 7920 | lastOffset = suballoc.offset + suballoc.size; |
| 7921 | ++nextAlloc2ndIndex; |
| 7922 | } |
| 7923 | // We are at the end. |
| 7924 | else |
| 7925 | { |
| 7926 | // There is free space from lastOffset to freeSpace2ndTo1stEnd. |
| 7927 | if (lastOffset < freeSpace2ndTo1stEnd) |
| 7928 | { |
| 7929 | const VkDeviceSize unusedRangeSize = freeSpace2ndTo1stEnd - lastOffset; |
| 7930 | VmaAddDetailedStatisticsUnusedRange(inoutStats, size: unusedRangeSize); |
| 7931 | } |
| 7932 | |
| 7933 | // End of loop. |
| 7934 | lastOffset = freeSpace2ndTo1stEnd; |
| 7935 | } |
| 7936 | } |
| 7937 | } |
| 7938 | |
| 7939 | size_t nextAlloc1stIndex = m_1stNullItemsBeginCount; |
| 7940 | const VkDeviceSize freeSpace1stTo2ndEnd = |
| 7941 | m_2ndVectorMode == SECOND_VECTOR_DOUBLE_STACK ? suballocations2nd.back().offset : size; |
| 7942 | while (lastOffset < freeSpace1stTo2ndEnd) |
| 7943 | { |
| 7944 | // Find next non-null allocation or move nextAllocIndex to the end. |
| 7945 | while (nextAlloc1stIndex < suballoc1stCount && |
| 7946 | suballocations1st[nextAlloc1stIndex].userData == VMA_NULL) |
| 7947 | { |
| 7948 | ++nextAlloc1stIndex; |
| 7949 | } |
| 7950 | |
| 7951 | // Found non-null allocation. |
| 7952 | if (nextAlloc1stIndex < suballoc1stCount) |
| 7953 | { |
| 7954 | const VmaSuballocation& suballoc = suballocations1st[nextAlloc1stIndex]; |
| 7955 | |
| 7956 | // 1. Process free space before this allocation. |
| 7957 | if (lastOffset < suballoc.offset) |
| 7958 | { |
| 7959 | // There is free space from lastOffset to suballoc.offset. |
| 7960 | const VkDeviceSize unusedRangeSize = suballoc.offset - lastOffset; |
| 7961 | VmaAddDetailedStatisticsUnusedRange(inoutStats, size: unusedRangeSize); |
| 7962 | } |
| 7963 | |
| 7964 | // 2. Process this allocation. |
| 7965 | // There is allocation with suballoc.offset, suballoc.size. |
| 7966 | VmaAddDetailedStatisticsAllocation(inoutStats, size: suballoc.size); |
| 7967 | |
| 7968 | // 3. Prepare for next iteration. |
| 7969 | lastOffset = suballoc.offset + suballoc.size; |
| 7970 | ++nextAlloc1stIndex; |
| 7971 | } |
| 7972 | // We are at the end. |
| 7973 | else |
| 7974 | { |
| 7975 | // There is free space from lastOffset to freeSpace1stTo2ndEnd. |
| 7976 | if (lastOffset < freeSpace1stTo2ndEnd) |
| 7977 | { |
| 7978 | const VkDeviceSize unusedRangeSize = freeSpace1stTo2ndEnd - lastOffset; |
| 7979 | VmaAddDetailedStatisticsUnusedRange(inoutStats, size: unusedRangeSize); |
| 7980 | } |
| 7981 | |
| 7982 | // End of loop. |
| 7983 | lastOffset = freeSpace1stTo2ndEnd; |
| 7984 | } |
| 7985 | } |
| 7986 | |
| 7987 | if (m_2ndVectorMode == SECOND_VECTOR_DOUBLE_STACK) |
| 7988 | { |
| 7989 | size_t nextAlloc2ndIndex = suballocations2nd.size() - 1; |
| 7990 | while (lastOffset < size) |
| 7991 | { |
| 7992 | // Find next non-null allocation or move nextAllocIndex to the end. |
| 7993 | while (nextAlloc2ndIndex != SIZE_MAX && |
| 7994 | suballocations2nd[nextAlloc2ndIndex].userData == VMA_NULL) |
| 7995 | { |
| 7996 | --nextAlloc2ndIndex; |
| 7997 | } |
| 7998 | |
| 7999 | // Found non-null allocation. |
| 8000 | if (nextAlloc2ndIndex != SIZE_MAX) |
| 8001 | { |
| 8002 | const VmaSuballocation& suballoc = suballocations2nd[nextAlloc2ndIndex]; |
| 8003 | |
| 8004 | // 1. Process free space before this allocation. |
| 8005 | if (lastOffset < suballoc.offset) |
| 8006 | { |
| 8007 | // There is free space from lastOffset to suballoc.offset. |
| 8008 | const VkDeviceSize unusedRangeSize = suballoc.offset - lastOffset; |
| 8009 | VmaAddDetailedStatisticsUnusedRange(inoutStats, size: unusedRangeSize); |
| 8010 | } |
| 8011 | |
| 8012 | // 2. Process this allocation. |
| 8013 | // There is allocation with suballoc.offset, suballoc.size. |
| 8014 | VmaAddDetailedStatisticsAllocation(inoutStats, size: suballoc.size); |
| 8015 | |
| 8016 | // 3. Prepare for next iteration. |
| 8017 | lastOffset = suballoc.offset + suballoc.size; |
| 8018 | --nextAlloc2ndIndex; |
| 8019 | } |
| 8020 | // We are at the end. |
| 8021 | else |
| 8022 | { |
| 8023 | // There is free space from lastOffset to size. |
| 8024 | if (lastOffset < size) |
| 8025 | { |
| 8026 | const VkDeviceSize unusedRangeSize = size - lastOffset; |
| 8027 | VmaAddDetailedStatisticsUnusedRange(inoutStats, size: unusedRangeSize); |
| 8028 | } |
| 8029 | |
| 8030 | // End of loop. |
| 8031 | lastOffset = size; |
| 8032 | } |
| 8033 | } |
| 8034 | } |
| 8035 | } |
| 8036 | |
| 8037 | void VmaBlockMetadata_Linear::AddStatistics(VmaStatistics& inoutStats) const |
| 8038 | { |
| 8039 | const SuballocationVectorType& suballocations1st = AccessSuballocations1st(); |
| 8040 | const SuballocationVectorType& suballocations2nd = AccessSuballocations2nd(); |
| 8041 | const VkDeviceSize size = GetSize(); |
| 8042 | const size_t suballoc1stCount = suballocations1st.size(); |
| 8043 | const size_t suballoc2ndCount = suballocations2nd.size(); |
| 8044 | |
| 8045 | inoutStats.blockCount++; |
| 8046 | inoutStats.blockBytes += size; |
| 8047 | inoutStats.allocationBytes += size - m_SumFreeSize; |
| 8048 | |
| 8049 | VkDeviceSize lastOffset = 0; |
| 8050 | |
| 8051 | if (m_2ndVectorMode == SECOND_VECTOR_RING_BUFFER) |
| 8052 | { |
| 8053 | const VkDeviceSize freeSpace2ndTo1stEnd = suballocations1st[m_1stNullItemsBeginCount].offset; |
| 8054 | size_t nextAlloc2ndIndex = m_1stNullItemsBeginCount; |
| 8055 | while (lastOffset < freeSpace2ndTo1stEnd) |
| 8056 | { |
| 8057 | // Find next non-null allocation or move nextAlloc2ndIndex to the end. |
| 8058 | while (nextAlloc2ndIndex < suballoc2ndCount && |
| 8059 | suballocations2nd[nextAlloc2ndIndex].userData == VMA_NULL) |
| 8060 | { |
| 8061 | ++nextAlloc2ndIndex; |
| 8062 | } |
| 8063 | |
| 8064 | // Found non-null allocation. |
| 8065 | if (nextAlloc2ndIndex < suballoc2ndCount) |
| 8066 | { |
| 8067 | const VmaSuballocation& suballoc = suballocations2nd[nextAlloc2ndIndex]; |
| 8068 | |
| 8069 | // 1. Process free space before this allocation. |
| 8070 | if (lastOffset < suballoc.offset) |
| 8071 | { |
| 8072 | // There is free space from lastOffset to suballoc.offset. |
| 8073 | const VkDeviceSize unusedRangeSize = suballoc.offset - lastOffset; |
| 8074 | } |
| 8075 | |
| 8076 | // 2. Process this allocation. |
| 8077 | // There is allocation with suballoc.offset, suballoc.size. |
| 8078 | ++inoutStats.allocationCount; |
| 8079 | |
| 8080 | // 3. Prepare for next iteration. |
| 8081 | lastOffset = suballoc.offset + suballoc.size; |
| 8082 | ++nextAlloc2ndIndex; |
| 8083 | } |
| 8084 | // We are at the end. |
| 8085 | else |
| 8086 | { |
| 8087 | if (lastOffset < freeSpace2ndTo1stEnd) |
| 8088 | { |
| 8089 | // There is free space from lastOffset to freeSpace2ndTo1stEnd. |
| 8090 | const VkDeviceSize unusedRangeSize = freeSpace2ndTo1stEnd - lastOffset; |
| 8091 | } |
| 8092 | |
| 8093 | // End of loop. |
| 8094 | lastOffset = freeSpace2ndTo1stEnd; |
| 8095 | } |
| 8096 | } |
| 8097 | } |
| 8098 | |
| 8099 | size_t nextAlloc1stIndex = m_1stNullItemsBeginCount; |
| 8100 | const VkDeviceSize freeSpace1stTo2ndEnd = |
| 8101 | m_2ndVectorMode == SECOND_VECTOR_DOUBLE_STACK ? suballocations2nd.back().offset : size; |
| 8102 | while (lastOffset < freeSpace1stTo2ndEnd) |
| 8103 | { |
| 8104 | // Find next non-null allocation or move nextAllocIndex to the end. |
| 8105 | while (nextAlloc1stIndex < suballoc1stCount && |
| 8106 | suballocations1st[nextAlloc1stIndex].userData == VMA_NULL) |
| 8107 | { |
| 8108 | ++nextAlloc1stIndex; |
| 8109 | } |
| 8110 | |
| 8111 | // Found non-null allocation. |
| 8112 | if (nextAlloc1stIndex < suballoc1stCount) |
| 8113 | { |
| 8114 | const VmaSuballocation& suballoc = suballocations1st[nextAlloc1stIndex]; |
| 8115 | |
| 8116 | // 1. Process free space before this allocation. |
| 8117 | if (lastOffset < suballoc.offset) |
| 8118 | { |
| 8119 | // There is free space from lastOffset to suballoc.offset. |
| 8120 | const VkDeviceSize unusedRangeSize = suballoc.offset - lastOffset; |
| 8121 | } |
| 8122 | |
| 8123 | // 2. Process this allocation. |
| 8124 | // There is allocation with suballoc.offset, suballoc.size. |
| 8125 | ++inoutStats.allocationCount; |
| 8126 | |
| 8127 | // 3. Prepare for next iteration. |
| 8128 | lastOffset = suballoc.offset + suballoc.size; |
| 8129 | ++nextAlloc1stIndex; |
| 8130 | } |
| 8131 | // We are at the end. |
| 8132 | else |
| 8133 | { |
| 8134 | if (lastOffset < freeSpace1stTo2ndEnd) |
| 8135 | { |
| 8136 | // There is free space from lastOffset to freeSpace1stTo2ndEnd. |
| 8137 | const VkDeviceSize unusedRangeSize = freeSpace1stTo2ndEnd - lastOffset; |
| 8138 | } |
| 8139 | |
| 8140 | // End of loop. |
| 8141 | lastOffset = freeSpace1stTo2ndEnd; |
| 8142 | } |
| 8143 | } |
| 8144 | |
| 8145 | if (m_2ndVectorMode == SECOND_VECTOR_DOUBLE_STACK) |
| 8146 | { |
| 8147 | size_t nextAlloc2ndIndex = suballocations2nd.size() - 1; |
| 8148 | while (lastOffset < size) |
| 8149 | { |
| 8150 | // Find next non-null allocation or move nextAlloc2ndIndex to the end. |
| 8151 | while (nextAlloc2ndIndex != SIZE_MAX && |
| 8152 | suballocations2nd[nextAlloc2ndIndex].userData == VMA_NULL) |
| 8153 | { |
| 8154 | --nextAlloc2ndIndex; |
| 8155 | } |
| 8156 | |
| 8157 | // Found non-null allocation. |
| 8158 | if (nextAlloc2ndIndex != SIZE_MAX) |
| 8159 | { |
| 8160 | const VmaSuballocation& suballoc = suballocations2nd[nextAlloc2ndIndex]; |
| 8161 | |
| 8162 | // 1. Process free space before this allocation. |
| 8163 | if (lastOffset < suballoc.offset) |
| 8164 | { |
| 8165 | // There is free space from lastOffset to suballoc.offset. |
| 8166 | const VkDeviceSize unusedRangeSize = suballoc.offset - lastOffset; |
| 8167 | } |
| 8168 | |
| 8169 | // 2. Process this allocation. |
| 8170 | // There is allocation with suballoc.offset, suballoc.size. |
| 8171 | ++inoutStats.allocationCount; |
| 8172 | |
| 8173 | // 3. Prepare for next iteration. |
| 8174 | lastOffset = suballoc.offset + suballoc.size; |
| 8175 | --nextAlloc2ndIndex; |
| 8176 | } |
| 8177 | // We are at the end. |
| 8178 | else |
| 8179 | { |
| 8180 | if (lastOffset < size) |
| 8181 | { |
| 8182 | // There is free space from lastOffset to size. |
| 8183 | const VkDeviceSize unusedRangeSize = size - lastOffset; |
| 8184 | } |
| 8185 | |
| 8186 | // End of loop. |
| 8187 | lastOffset = size; |
| 8188 | } |
| 8189 | } |
| 8190 | } |
| 8191 | } |
| 8192 | |
| 8193 | #if VMA_STATS_STRING_ENABLED |
| 8194 | void VmaBlockMetadata_Linear::PrintDetailedMap(class VmaJsonWriter& json) const |
| 8195 | { |
| 8196 | const VkDeviceSize size = GetSize(); |
| 8197 | const SuballocationVectorType& suballocations1st = AccessSuballocations1st(); |
| 8198 | const SuballocationVectorType& suballocations2nd = AccessSuballocations2nd(); |
| 8199 | const size_t suballoc1stCount = suballocations1st.size(); |
| 8200 | const size_t suballoc2ndCount = suballocations2nd.size(); |
| 8201 | |
| 8202 | // FIRST PASS |
| 8203 | |
| 8204 | size_t unusedRangeCount = 0; |
| 8205 | VkDeviceSize usedBytes = 0; |
| 8206 | |
| 8207 | VkDeviceSize lastOffset = 0; |
| 8208 | |
| 8209 | size_t alloc2ndCount = 0; |
| 8210 | if (m_2ndVectorMode == SECOND_VECTOR_RING_BUFFER) |
| 8211 | { |
| 8212 | const VkDeviceSize freeSpace2ndTo1stEnd = suballocations1st[m_1stNullItemsBeginCount].offset; |
| 8213 | size_t nextAlloc2ndIndex = 0; |
| 8214 | while (lastOffset < freeSpace2ndTo1stEnd) |
| 8215 | { |
| 8216 | // Find next non-null allocation or move nextAlloc2ndIndex to the end. |
| 8217 | while (nextAlloc2ndIndex < suballoc2ndCount && |
| 8218 | suballocations2nd[nextAlloc2ndIndex].userData == VMA_NULL) |
| 8219 | { |
| 8220 | ++nextAlloc2ndIndex; |
| 8221 | } |
| 8222 | |
| 8223 | // Found non-null allocation. |
| 8224 | if (nextAlloc2ndIndex < suballoc2ndCount) |
| 8225 | { |
| 8226 | const VmaSuballocation& suballoc = suballocations2nd[nextAlloc2ndIndex]; |
| 8227 | |
| 8228 | // 1. Process free space before this allocation. |
| 8229 | if (lastOffset < suballoc.offset) |
| 8230 | { |
| 8231 | // There is free space from lastOffset to suballoc.offset. |
| 8232 | ++unusedRangeCount; |
| 8233 | } |
| 8234 | |
| 8235 | // 2. Process this allocation. |
| 8236 | // There is allocation with suballoc.offset, suballoc.size. |
| 8237 | ++alloc2ndCount; |
| 8238 | usedBytes += suballoc.size; |
| 8239 | |
| 8240 | // 3. Prepare for next iteration. |
| 8241 | lastOffset = suballoc.offset + suballoc.size; |
| 8242 | ++nextAlloc2ndIndex; |
| 8243 | } |
| 8244 | // We are at the end. |
| 8245 | else |
| 8246 | { |
| 8247 | if (lastOffset < freeSpace2ndTo1stEnd) |
| 8248 | { |
| 8249 | // There is free space from lastOffset to freeSpace2ndTo1stEnd. |
| 8250 | ++unusedRangeCount; |
| 8251 | } |
| 8252 | |
| 8253 | // End of loop. |
| 8254 | lastOffset = freeSpace2ndTo1stEnd; |
| 8255 | } |
| 8256 | } |
| 8257 | } |
| 8258 | |
| 8259 | size_t nextAlloc1stIndex = m_1stNullItemsBeginCount; |
| 8260 | size_t alloc1stCount = 0; |
| 8261 | const VkDeviceSize freeSpace1stTo2ndEnd = |
| 8262 | m_2ndVectorMode == SECOND_VECTOR_DOUBLE_STACK ? suballocations2nd.back().offset : size; |
| 8263 | while (lastOffset < freeSpace1stTo2ndEnd) |
| 8264 | { |
| 8265 | // Find next non-null allocation or move nextAllocIndex to the end. |
| 8266 | while (nextAlloc1stIndex < suballoc1stCount && |
| 8267 | suballocations1st[nextAlloc1stIndex].userData == VMA_NULL) |
| 8268 | { |
| 8269 | ++nextAlloc1stIndex; |
| 8270 | } |
| 8271 | |
| 8272 | // Found non-null allocation. |
| 8273 | if (nextAlloc1stIndex < suballoc1stCount) |
| 8274 | { |
| 8275 | const VmaSuballocation& suballoc = suballocations1st[nextAlloc1stIndex]; |
| 8276 | |
| 8277 | // 1. Process free space before this allocation. |
| 8278 | if (lastOffset < suballoc.offset) |
| 8279 | { |
| 8280 | // There is free space from lastOffset to suballoc.offset. |
| 8281 | ++unusedRangeCount; |
| 8282 | } |
| 8283 | |
| 8284 | // 2. Process this allocation. |
| 8285 | // There is allocation with suballoc.offset, suballoc.size. |
| 8286 | ++alloc1stCount; |
| 8287 | usedBytes += suballoc.size; |
| 8288 | |
| 8289 | // 3. Prepare for next iteration. |
| 8290 | lastOffset = suballoc.offset + suballoc.size; |
| 8291 | ++nextAlloc1stIndex; |
| 8292 | } |
| 8293 | // We are at the end. |
| 8294 | else |
| 8295 | { |
| 8296 | if (lastOffset < size) |
| 8297 | { |
| 8298 | // There is free space from lastOffset to freeSpace1stTo2ndEnd. |
| 8299 | ++unusedRangeCount; |
| 8300 | } |
| 8301 | |
| 8302 | // End of loop. |
| 8303 | lastOffset = freeSpace1stTo2ndEnd; |
| 8304 | } |
| 8305 | } |
| 8306 | |
| 8307 | if (m_2ndVectorMode == SECOND_VECTOR_DOUBLE_STACK) |
| 8308 | { |
| 8309 | size_t nextAlloc2ndIndex = suballocations2nd.size() - 1; |
| 8310 | while (lastOffset < size) |
| 8311 | { |
| 8312 | // Find next non-null allocation or move nextAlloc2ndIndex to the end. |
| 8313 | while (nextAlloc2ndIndex != SIZE_MAX && |
| 8314 | suballocations2nd[nextAlloc2ndIndex].userData == VMA_NULL) |
| 8315 | { |
| 8316 | --nextAlloc2ndIndex; |
| 8317 | } |
| 8318 | |
| 8319 | // Found non-null allocation. |
| 8320 | if (nextAlloc2ndIndex != SIZE_MAX) |
| 8321 | { |
| 8322 | const VmaSuballocation& suballoc = suballocations2nd[nextAlloc2ndIndex]; |
| 8323 | |
| 8324 | // 1. Process free space before this allocation. |
| 8325 | if (lastOffset < suballoc.offset) |
| 8326 | { |
| 8327 | // There is free space from lastOffset to suballoc.offset. |
| 8328 | ++unusedRangeCount; |
| 8329 | } |
| 8330 | |
| 8331 | // 2. Process this allocation. |
| 8332 | // There is allocation with suballoc.offset, suballoc.size. |
| 8333 | ++alloc2ndCount; |
| 8334 | usedBytes += suballoc.size; |
| 8335 | |
| 8336 | // 3. Prepare for next iteration. |
| 8337 | lastOffset = suballoc.offset + suballoc.size; |
| 8338 | --nextAlloc2ndIndex; |
| 8339 | } |
| 8340 | // We are at the end. |
| 8341 | else |
| 8342 | { |
| 8343 | if (lastOffset < size) |
| 8344 | { |
| 8345 | // There is free space from lastOffset to size. |
| 8346 | ++unusedRangeCount; |
| 8347 | } |
| 8348 | |
| 8349 | // End of loop. |
| 8350 | lastOffset = size; |
| 8351 | } |
| 8352 | } |
| 8353 | } |
| 8354 | |
| 8355 | const VkDeviceSize unusedBytes = size - usedBytes; |
| 8356 | PrintDetailedMap_Begin(json, unusedBytes, allocationCount: alloc1stCount + alloc2ndCount, unusedRangeCount); |
| 8357 | |
| 8358 | // SECOND PASS |
| 8359 | lastOffset = 0; |
| 8360 | |
| 8361 | if (m_2ndVectorMode == SECOND_VECTOR_RING_BUFFER) |
| 8362 | { |
| 8363 | const VkDeviceSize freeSpace2ndTo1stEnd = suballocations1st[m_1stNullItemsBeginCount].offset; |
| 8364 | size_t nextAlloc2ndIndex = 0; |
| 8365 | while (lastOffset < freeSpace2ndTo1stEnd) |
| 8366 | { |
| 8367 | // Find next non-null allocation or move nextAlloc2ndIndex to the end. |
| 8368 | while (nextAlloc2ndIndex < suballoc2ndCount && |
| 8369 | suballocations2nd[nextAlloc2ndIndex].userData == VMA_NULL) |
| 8370 | { |
| 8371 | ++nextAlloc2ndIndex; |
| 8372 | } |
| 8373 | |
| 8374 | // Found non-null allocation. |
| 8375 | if (nextAlloc2ndIndex < suballoc2ndCount) |
| 8376 | { |
| 8377 | const VmaSuballocation& suballoc = suballocations2nd[nextAlloc2ndIndex]; |
| 8378 | |
| 8379 | // 1. Process free space before this allocation. |
| 8380 | if (lastOffset < suballoc.offset) |
| 8381 | { |
| 8382 | // There is free space from lastOffset to suballoc.offset. |
| 8383 | const VkDeviceSize unusedRangeSize = suballoc.offset - lastOffset; |
| 8384 | PrintDetailedMap_UnusedRange(json, offset: lastOffset, size: unusedRangeSize); |
| 8385 | } |
| 8386 | |
| 8387 | // 2. Process this allocation. |
| 8388 | // There is allocation with suballoc.offset, suballoc.size. |
| 8389 | PrintDetailedMap_Allocation(json, offset: suballoc.offset, size: suballoc.size, userData: suballoc.userData); |
| 8390 | |
| 8391 | // 3. Prepare for next iteration. |
| 8392 | lastOffset = suballoc.offset + suballoc.size; |
| 8393 | ++nextAlloc2ndIndex; |
| 8394 | } |
| 8395 | // We are at the end. |
| 8396 | else |
| 8397 | { |
| 8398 | if (lastOffset < freeSpace2ndTo1stEnd) |
| 8399 | { |
| 8400 | // There is free space from lastOffset to freeSpace2ndTo1stEnd. |
| 8401 | const VkDeviceSize unusedRangeSize = freeSpace2ndTo1stEnd - lastOffset; |
| 8402 | PrintDetailedMap_UnusedRange(json, offset: lastOffset, size: unusedRangeSize); |
| 8403 | } |
| 8404 | |
| 8405 | // End of loop. |
| 8406 | lastOffset = freeSpace2ndTo1stEnd; |
| 8407 | } |
| 8408 | } |
| 8409 | } |
| 8410 | |
| 8411 | nextAlloc1stIndex = m_1stNullItemsBeginCount; |
| 8412 | while (lastOffset < freeSpace1stTo2ndEnd) |
| 8413 | { |
| 8414 | // Find next non-null allocation or move nextAllocIndex to the end. |
| 8415 | while (nextAlloc1stIndex < suballoc1stCount && |
| 8416 | suballocations1st[nextAlloc1stIndex].userData == VMA_NULL) |
| 8417 | { |
| 8418 | ++nextAlloc1stIndex; |
| 8419 | } |
| 8420 | |
| 8421 | // Found non-null allocation. |
| 8422 | if (nextAlloc1stIndex < suballoc1stCount) |
| 8423 | { |
| 8424 | const VmaSuballocation& suballoc = suballocations1st[nextAlloc1stIndex]; |
| 8425 | |
| 8426 | // 1. Process free space before this allocation. |
| 8427 | if (lastOffset < suballoc.offset) |
| 8428 | { |
| 8429 | // There is free space from lastOffset to suballoc.offset. |
| 8430 | const VkDeviceSize unusedRangeSize = suballoc.offset - lastOffset; |
| 8431 | PrintDetailedMap_UnusedRange(json, offset: lastOffset, size: unusedRangeSize); |
| 8432 | } |
| 8433 | |
| 8434 | // 2. Process this allocation. |
| 8435 | // There is allocation with suballoc.offset, suballoc.size. |
| 8436 | PrintDetailedMap_Allocation(json, offset: suballoc.offset, size: suballoc.size, userData: suballoc.userData); |
| 8437 | |
| 8438 | // 3. Prepare for next iteration. |
| 8439 | lastOffset = suballoc.offset + suballoc.size; |
| 8440 | ++nextAlloc1stIndex; |
| 8441 | } |
| 8442 | // We are at the end. |
| 8443 | else |
| 8444 | { |
| 8445 | if (lastOffset < freeSpace1stTo2ndEnd) |
| 8446 | { |
| 8447 | // There is free space from lastOffset to freeSpace1stTo2ndEnd. |
| 8448 | const VkDeviceSize unusedRangeSize = freeSpace1stTo2ndEnd - lastOffset; |
| 8449 | PrintDetailedMap_UnusedRange(json, offset: lastOffset, size: unusedRangeSize); |
| 8450 | } |
| 8451 | |
| 8452 | // End of loop. |
| 8453 | lastOffset = freeSpace1stTo2ndEnd; |
| 8454 | } |
| 8455 | } |
| 8456 | |
| 8457 | if (m_2ndVectorMode == SECOND_VECTOR_DOUBLE_STACK) |
| 8458 | { |
| 8459 | size_t nextAlloc2ndIndex = suballocations2nd.size() - 1; |
| 8460 | while (lastOffset < size) |
| 8461 | { |
| 8462 | // Find next non-null allocation or move nextAlloc2ndIndex to the end. |
| 8463 | while (nextAlloc2ndIndex != SIZE_MAX && |
| 8464 | suballocations2nd[nextAlloc2ndIndex].userData == VMA_NULL) |
| 8465 | { |
| 8466 | --nextAlloc2ndIndex; |
| 8467 | } |
| 8468 | |
| 8469 | // Found non-null allocation. |
| 8470 | if (nextAlloc2ndIndex != SIZE_MAX) |
| 8471 | { |
| 8472 | const VmaSuballocation& suballoc = suballocations2nd[nextAlloc2ndIndex]; |
| 8473 | |
| 8474 | // 1. Process free space before this allocation. |
| 8475 | if (lastOffset < suballoc.offset) |
| 8476 | { |
| 8477 | // There is free space from lastOffset to suballoc.offset. |
| 8478 | const VkDeviceSize unusedRangeSize = suballoc.offset - lastOffset; |
| 8479 | PrintDetailedMap_UnusedRange(json, offset: lastOffset, size: unusedRangeSize); |
| 8480 | } |
| 8481 | |
| 8482 | // 2. Process this allocation. |
| 8483 | // There is allocation with suballoc.offset, suballoc.size. |
| 8484 | PrintDetailedMap_Allocation(json, offset: suballoc.offset, size: suballoc.size, userData: suballoc.userData); |
| 8485 | |
| 8486 | // 3. Prepare for next iteration. |
| 8487 | lastOffset = suballoc.offset + suballoc.size; |
| 8488 | --nextAlloc2ndIndex; |
| 8489 | } |
| 8490 | // We are at the end. |
| 8491 | else |
| 8492 | { |
| 8493 | if (lastOffset < size) |
| 8494 | { |
| 8495 | // There is free space from lastOffset to size. |
| 8496 | const VkDeviceSize unusedRangeSize = size - lastOffset; |
| 8497 | PrintDetailedMap_UnusedRange(json, offset: lastOffset, size: unusedRangeSize); |
| 8498 | } |
| 8499 | |
| 8500 | // End of loop. |
| 8501 | lastOffset = size; |
| 8502 | } |
| 8503 | } |
| 8504 | } |
| 8505 | |
| 8506 | PrintDetailedMap_End(json); |
| 8507 | } |
| 8508 | #endif // VMA_STATS_STRING_ENABLED |
| 8509 | |
| 8510 | bool VmaBlockMetadata_Linear::CreateAllocationRequest( |
| 8511 | VkDeviceSize allocSize, |
| 8512 | VkDeviceSize allocAlignment, |
| 8513 | bool upperAddress, |
| 8514 | VmaSuballocationType allocType, |
| 8515 | uint32_t strategy, |
| 8516 | VmaAllocationRequest* pAllocationRequest) |
| 8517 | { |
| 8518 | VMA_ASSERT(allocSize > 0); |
| 8519 | VMA_ASSERT(allocType != VMA_SUBALLOCATION_TYPE_FREE); |
| 8520 | VMA_ASSERT(pAllocationRequest != VMA_NULL); |
| 8521 | VMA_HEAVY_ASSERT(Validate()); |
| 8522 | pAllocationRequest->size = allocSize; |
| 8523 | return upperAddress ? |
| 8524 | CreateAllocationRequest_UpperAddress( |
| 8525 | allocSize, allocAlignment, allocType, strategy, pAllocationRequest) : |
| 8526 | CreateAllocationRequest_LowerAddress( |
| 8527 | allocSize, allocAlignment, allocType, strategy, pAllocationRequest); |
| 8528 | } |
| 8529 | |
| 8530 | VkResult VmaBlockMetadata_Linear::CheckCorruption(const void* pBlockData) |
| 8531 | { |
| 8532 | VMA_ASSERT(!IsVirtual()); |
| 8533 | SuballocationVectorType& suballocations1st = AccessSuballocations1st(); |
| 8534 | for (size_t i = m_1stNullItemsBeginCount, count = suballocations1st.size(); i < count; ++i) |
| 8535 | { |
| 8536 | const VmaSuballocation& suballoc = suballocations1st[i]; |
| 8537 | if (suballoc.type != VMA_SUBALLOCATION_TYPE_FREE) |
| 8538 | { |
| 8539 | if (!VmaValidateMagicValue(pData: pBlockData, offset: suballoc.offset + suballoc.size)) |
| 8540 | { |
| 8541 | VMA_ASSERT(0 && "MEMORY CORRUPTION DETECTED AFTER VALIDATED ALLOCATION!" ); |
| 8542 | return VK_ERROR_UNKNOWN_COPY; |
| 8543 | } |
| 8544 | } |
| 8545 | } |
| 8546 | |
| 8547 | SuballocationVectorType& suballocations2nd = AccessSuballocations2nd(); |
| 8548 | for (size_t i = 0, count = suballocations2nd.size(); i < count; ++i) |
| 8549 | { |
| 8550 | const VmaSuballocation& suballoc = suballocations2nd[i]; |
| 8551 | if (suballoc.type != VMA_SUBALLOCATION_TYPE_FREE) |
| 8552 | { |
| 8553 | if (!VmaValidateMagicValue(pData: pBlockData, offset: suballoc.offset + suballoc.size)) |
| 8554 | { |
| 8555 | VMA_ASSERT(0 && "MEMORY CORRUPTION DETECTED AFTER VALIDATED ALLOCATION!" ); |
| 8556 | return VK_ERROR_UNKNOWN_COPY; |
| 8557 | } |
| 8558 | } |
| 8559 | } |
| 8560 | |
| 8561 | return VK_SUCCESS; |
| 8562 | } |
| 8563 | |
| 8564 | void VmaBlockMetadata_Linear::Alloc( |
| 8565 | const VmaAllocationRequest& request, |
| 8566 | VmaSuballocationType type, |
| 8567 | void* userData) |
| 8568 | { |
| 8569 | const VkDeviceSize offset = (VkDeviceSize)request.allocHandle - 1; |
| 8570 | const VmaSuballocation newSuballoc = { .offset: offset, .size: request.size, .userData: userData, .type: type }; |
| 8571 | |
| 8572 | switch (request.type) |
| 8573 | { |
| 8574 | case VmaAllocationRequestType::UpperAddress: |
| 8575 | { |
| 8576 | VMA_ASSERT(m_2ndVectorMode != SECOND_VECTOR_RING_BUFFER && |
| 8577 | "CRITICAL ERROR: Trying to use linear allocator as double stack while it was already used as ring buffer." ); |
| 8578 | SuballocationVectorType& suballocations2nd = AccessSuballocations2nd(); |
| 8579 | suballocations2nd.push_back(src: newSuballoc); |
| 8580 | m_2ndVectorMode = SECOND_VECTOR_DOUBLE_STACK; |
| 8581 | } |
| 8582 | break; |
| 8583 | case VmaAllocationRequestType::EndOf1st: |
| 8584 | { |
| 8585 | SuballocationVectorType& suballocations1st = AccessSuballocations1st(); |
| 8586 | |
| 8587 | VMA_ASSERT(suballocations1st.empty() || |
| 8588 | offset >= suballocations1st.back().offset + suballocations1st.back().size); |
| 8589 | // Check if it fits before the end of the block. |
| 8590 | VMA_ASSERT(offset + request.size <= GetSize()); |
| 8591 | |
| 8592 | suballocations1st.push_back(src: newSuballoc); |
| 8593 | } |
| 8594 | break; |
| 8595 | case VmaAllocationRequestType::EndOf2nd: |
| 8596 | { |
| 8597 | SuballocationVectorType& suballocations1st = AccessSuballocations1st(); |
| 8598 | // New allocation at the end of 2-part ring buffer, so before first allocation from 1st vector. |
| 8599 | VMA_ASSERT(!suballocations1st.empty() && |
| 8600 | offset + request.size <= suballocations1st[m_1stNullItemsBeginCount].offset); |
| 8601 | SuballocationVectorType& suballocations2nd = AccessSuballocations2nd(); |
| 8602 | |
| 8603 | switch (m_2ndVectorMode) |
| 8604 | { |
| 8605 | case SECOND_VECTOR_EMPTY: |
| 8606 | // First allocation from second part ring buffer. |
| 8607 | VMA_ASSERT(suballocations2nd.empty()); |
| 8608 | m_2ndVectorMode = SECOND_VECTOR_RING_BUFFER; |
| 8609 | break; |
| 8610 | case SECOND_VECTOR_RING_BUFFER: |
| 8611 | // 2-part ring buffer is already started. |
| 8612 | VMA_ASSERT(!suballocations2nd.empty()); |
| 8613 | break; |
| 8614 | case SECOND_VECTOR_DOUBLE_STACK: |
| 8615 | VMA_ASSERT(0 && "CRITICAL ERROR: Trying to use linear allocator as ring buffer while it was already used as double stack." ); |
| 8616 | break; |
| 8617 | default: |
| 8618 | VMA_ASSERT(0); |
| 8619 | } |
| 8620 | |
| 8621 | suballocations2nd.push_back(src: newSuballoc); |
| 8622 | } |
| 8623 | break; |
| 8624 | default: |
| 8625 | VMA_ASSERT(0 && "CRITICAL INTERNAL ERROR." ); |
| 8626 | } |
| 8627 | |
| 8628 | m_SumFreeSize -= newSuballoc.size; |
| 8629 | } |
| 8630 | |
| 8631 | void VmaBlockMetadata_Linear::Free(VmaAllocHandle allocHandle) |
| 8632 | { |
| 8633 | SuballocationVectorType& suballocations1st = AccessSuballocations1st(); |
| 8634 | SuballocationVectorType& suballocations2nd = AccessSuballocations2nd(); |
| 8635 | VkDeviceSize offset = (VkDeviceSize)allocHandle - 1; |
| 8636 | |
| 8637 | if (!suballocations1st.empty()) |
| 8638 | { |
| 8639 | // First allocation: Mark it as next empty at the beginning. |
| 8640 | VmaSuballocation& firstSuballoc = suballocations1st[m_1stNullItemsBeginCount]; |
| 8641 | if (firstSuballoc.offset == offset) |
| 8642 | { |
| 8643 | firstSuballoc.type = VMA_SUBALLOCATION_TYPE_FREE; |
| 8644 | firstSuballoc.userData = VMA_NULL; |
| 8645 | m_SumFreeSize += firstSuballoc.size; |
| 8646 | ++m_1stNullItemsBeginCount; |
| 8647 | CleanupAfterFree(); |
| 8648 | return; |
| 8649 | } |
| 8650 | } |
| 8651 | |
| 8652 | // Last allocation in 2-part ring buffer or top of upper stack (same logic). |
| 8653 | if (m_2ndVectorMode == SECOND_VECTOR_RING_BUFFER || |
| 8654 | m_2ndVectorMode == SECOND_VECTOR_DOUBLE_STACK) |
| 8655 | { |
| 8656 | VmaSuballocation& lastSuballoc = suballocations2nd.back(); |
| 8657 | if (lastSuballoc.offset == offset) |
| 8658 | { |
| 8659 | m_SumFreeSize += lastSuballoc.size; |
| 8660 | suballocations2nd.pop_back(); |
| 8661 | CleanupAfterFree(); |
| 8662 | return; |
| 8663 | } |
| 8664 | } |
| 8665 | // Last allocation in 1st vector. |
| 8666 | else if (m_2ndVectorMode == SECOND_VECTOR_EMPTY) |
| 8667 | { |
| 8668 | VmaSuballocation& lastSuballoc = suballocations1st.back(); |
| 8669 | if (lastSuballoc.offset == offset) |
| 8670 | { |
| 8671 | m_SumFreeSize += lastSuballoc.size; |
| 8672 | suballocations1st.pop_back(); |
| 8673 | CleanupAfterFree(); |
| 8674 | return; |
| 8675 | } |
| 8676 | } |
| 8677 | |
| 8678 | VmaSuballocation refSuballoc; |
| 8679 | refSuballoc.offset = offset; |
| 8680 | // Rest of members stays uninitialized intentionally for better performance. |
| 8681 | |
| 8682 | // Item from the middle of 1st vector. |
| 8683 | { |
| 8684 | const SuballocationVectorType::iterator it = VmaBinaryFindSorted( |
| 8685 | beg: suballocations1st.begin() + m_1stNullItemsBeginCount, |
| 8686 | end: suballocations1st.end(), |
| 8687 | value: refSuballoc, |
| 8688 | cmp: VmaSuballocationOffsetLess()); |
| 8689 | if (it != suballocations1st.end()) |
| 8690 | { |
| 8691 | it->type = VMA_SUBALLOCATION_TYPE_FREE; |
| 8692 | it->userData = VMA_NULL; |
| 8693 | ++m_1stNullItemsMiddleCount; |
| 8694 | m_SumFreeSize += it->size; |
| 8695 | CleanupAfterFree(); |
| 8696 | return; |
| 8697 | } |
| 8698 | } |
| 8699 | |
| 8700 | if (m_2ndVectorMode != SECOND_VECTOR_EMPTY) |
| 8701 | { |
| 8702 | // Item from the middle of 2nd vector. |
| 8703 | const SuballocationVectorType::iterator it = m_2ndVectorMode == SECOND_VECTOR_RING_BUFFER ? |
| 8704 | VmaBinaryFindSorted(beg: suballocations2nd.begin(), end: suballocations2nd.end(), value: refSuballoc, cmp: VmaSuballocationOffsetLess()) : |
| 8705 | VmaBinaryFindSorted(beg: suballocations2nd.begin(), end: suballocations2nd.end(), value: refSuballoc, cmp: VmaSuballocationOffsetGreater()); |
| 8706 | if (it != suballocations2nd.end()) |
| 8707 | { |
| 8708 | it->type = VMA_SUBALLOCATION_TYPE_FREE; |
| 8709 | it->userData = VMA_NULL; |
| 8710 | ++m_2ndNullItemsCount; |
| 8711 | m_SumFreeSize += it->size; |
| 8712 | CleanupAfterFree(); |
| 8713 | return; |
| 8714 | } |
| 8715 | } |
| 8716 | |
| 8717 | VMA_ASSERT(0 && "Allocation to free not found in linear allocator!" ); |
| 8718 | } |
| 8719 | |
| 8720 | void VmaBlockMetadata_Linear::GetAllocationInfo(VmaAllocHandle allocHandle, VmaVirtualAllocationInfo& outInfo) |
| 8721 | { |
| 8722 | outInfo.offset = (VkDeviceSize)allocHandle - 1; |
| 8723 | VmaSuballocation& suballoc = FindSuballocation(offset: outInfo.offset); |
| 8724 | outInfo.size = suballoc.size; |
| 8725 | outInfo.pUserData = suballoc.userData; |
| 8726 | } |
| 8727 | |
| 8728 | void* VmaBlockMetadata_Linear::GetAllocationUserData(VmaAllocHandle allocHandle) const |
| 8729 | { |
| 8730 | return FindSuballocation(offset: (VkDeviceSize)allocHandle - 1).userData; |
| 8731 | } |
| 8732 | |
| 8733 | VmaAllocHandle VmaBlockMetadata_Linear::GetAllocationListBegin() const |
| 8734 | { |
| 8735 | // Function only used for defragmentation, which is disabled for this algorithm |
| 8736 | VMA_ASSERT(0); |
| 8737 | return VK_NULL_HANDLE; |
| 8738 | } |
| 8739 | |
| 8740 | VmaAllocHandle VmaBlockMetadata_Linear::GetNextAllocation(VmaAllocHandle prevAlloc) const |
| 8741 | { |
| 8742 | // Function only used for defragmentation, which is disabled for this algorithm |
| 8743 | VMA_ASSERT(0); |
| 8744 | return VK_NULL_HANDLE; |
| 8745 | } |
| 8746 | |
| 8747 | VkDeviceSize VmaBlockMetadata_Linear::GetNextFreeRegionSize(VmaAllocHandle alloc) const |
| 8748 | { |
| 8749 | // Function only used for defragmentation, which is disabled for this algorithm |
| 8750 | VMA_ASSERT(0); |
| 8751 | return 0; |
| 8752 | } |
| 8753 | |
| 8754 | void VmaBlockMetadata_Linear::Clear() |
| 8755 | { |
| 8756 | m_SumFreeSize = GetSize(); |
| 8757 | m_Suballocations0.clear(); |
| 8758 | m_Suballocations1.clear(); |
| 8759 | // Leaving m_1stVectorIndex unchanged - it doesn't matter. |
| 8760 | m_2ndVectorMode = SECOND_VECTOR_EMPTY; |
| 8761 | m_1stNullItemsBeginCount = 0; |
| 8762 | m_1stNullItemsMiddleCount = 0; |
| 8763 | m_2ndNullItemsCount = 0; |
| 8764 | } |
| 8765 | |
| 8766 | void VmaBlockMetadata_Linear::SetAllocationUserData(VmaAllocHandle allocHandle, void* userData) |
| 8767 | { |
| 8768 | VmaSuballocation& suballoc = FindSuballocation(offset: (VkDeviceSize)allocHandle - 1); |
| 8769 | suballoc.userData = userData; |
| 8770 | } |
| 8771 | |
| 8772 | void VmaBlockMetadata_Linear::DebugLogAllAllocations() const |
| 8773 | { |
| 8774 | const SuballocationVectorType& suballocations1st = AccessSuballocations1st(); |
| 8775 | for (auto it = suballocations1st.begin() + m_1stNullItemsBeginCount; it != suballocations1st.end(); ++it) |
| 8776 | if (it->type != VMA_SUBALLOCATION_TYPE_FREE) |
| 8777 | DebugLogAllocation(offset: it->offset, size: it->size, userData: it->userData); |
| 8778 | |
| 8779 | const SuballocationVectorType& suballocations2nd = AccessSuballocations2nd(); |
| 8780 | for (auto it = suballocations2nd.begin(); it != suballocations2nd.end(); ++it) |
| 8781 | if (it->type != VMA_SUBALLOCATION_TYPE_FREE) |
| 8782 | DebugLogAllocation(offset: it->offset, size: it->size, userData: it->userData); |
| 8783 | } |
| 8784 | |
| 8785 | VmaSuballocation& VmaBlockMetadata_Linear::FindSuballocation(VkDeviceSize offset) const |
| 8786 | { |
| 8787 | const SuballocationVectorType& suballocations1st = AccessSuballocations1st(); |
| 8788 | const SuballocationVectorType& suballocations2nd = AccessSuballocations2nd(); |
| 8789 | |
| 8790 | VmaSuballocation refSuballoc; |
| 8791 | refSuballoc.offset = offset; |
| 8792 | // Rest of members stays uninitialized intentionally for better performance. |
| 8793 | |
| 8794 | // Item from the 1st vector. |
| 8795 | { |
| 8796 | SuballocationVectorType::const_iterator it = VmaBinaryFindSorted( |
| 8797 | beg: suballocations1st.begin() + m_1stNullItemsBeginCount, |
| 8798 | end: suballocations1st.end(), |
| 8799 | value: refSuballoc, |
| 8800 | cmp: VmaSuballocationOffsetLess()); |
| 8801 | if (it != suballocations1st.end()) |
| 8802 | { |
| 8803 | return const_cast<VmaSuballocation&>(*it); |
| 8804 | } |
| 8805 | } |
| 8806 | |
| 8807 | if (m_2ndVectorMode != SECOND_VECTOR_EMPTY) |
| 8808 | { |
| 8809 | // Rest of members stays uninitialized intentionally for better performance. |
| 8810 | SuballocationVectorType::const_iterator it = m_2ndVectorMode == SECOND_VECTOR_RING_BUFFER ? |
| 8811 | VmaBinaryFindSorted(beg: suballocations2nd.begin(), end: suballocations2nd.end(), value: refSuballoc, cmp: VmaSuballocationOffsetLess()) : |
| 8812 | VmaBinaryFindSorted(beg: suballocations2nd.begin(), end: suballocations2nd.end(), value: refSuballoc, cmp: VmaSuballocationOffsetGreater()); |
| 8813 | if (it != suballocations2nd.end()) |
| 8814 | { |
| 8815 | return const_cast<VmaSuballocation&>(*it); |
| 8816 | } |
| 8817 | } |
| 8818 | |
| 8819 | VMA_ASSERT(0 && "Allocation not found in linear allocator!" ); |
| 8820 | return const_cast<VmaSuballocation&>(suballocations1st.back()); // Should never occur. |
| 8821 | } |
| 8822 | |
| 8823 | bool VmaBlockMetadata_Linear::ShouldCompact1st() const |
| 8824 | { |
| 8825 | const size_t nullItemCount = m_1stNullItemsBeginCount + m_1stNullItemsMiddleCount; |
| 8826 | const size_t suballocCount = AccessSuballocations1st().size(); |
| 8827 | return suballocCount > 32 && nullItemCount * 2 >= (suballocCount - nullItemCount) * 3; |
| 8828 | } |
| 8829 | |
| 8830 | void VmaBlockMetadata_Linear::CleanupAfterFree() |
| 8831 | { |
| 8832 | SuballocationVectorType& suballocations1st = AccessSuballocations1st(); |
| 8833 | SuballocationVectorType& suballocations2nd = AccessSuballocations2nd(); |
| 8834 | |
| 8835 | if (IsEmpty()) |
| 8836 | { |
| 8837 | suballocations1st.clear(); |
| 8838 | suballocations2nd.clear(); |
| 8839 | m_1stNullItemsBeginCount = 0; |
| 8840 | m_1stNullItemsMiddleCount = 0; |
| 8841 | m_2ndNullItemsCount = 0; |
| 8842 | m_2ndVectorMode = SECOND_VECTOR_EMPTY; |
| 8843 | } |
| 8844 | else |
| 8845 | { |
| 8846 | const size_t suballoc1stCount = suballocations1st.size(); |
| 8847 | const size_t nullItem1stCount = m_1stNullItemsBeginCount + m_1stNullItemsMiddleCount; |
| 8848 | VMA_ASSERT(nullItem1stCount <= suballoc1stCount); |
| 8849 | |
| 8850 | // Find more null items at the beginning of 1st vector. |
| 8851 | while (m_1stNullItemsBeginCount < suballoc1stCount && |
| 8852 | suballocations1st[m_1stNullItemsBeginCount].type == VMA_SUBALLOCATION_TYPE_FREE) |
| 8853 | { |
| 8854 | ++m_1stNullItemsBeginCount; |
| 8855 | --m_1stNullItemsMiddleCount; |
| 8856 | } |
| 8857 | |
| 8858 | // Find more null items at the end of 1st vector. |
| 8859 | while (m_1stNullItemsMiddleCount > 0 && |
| 8860 | suballocations1st.back().type == VMA_SUBALLOCATION_TYPE_FREE) |
| 8861 | { |
| 8862 | --m_1stNullItemsMiddleCount; |
| 8863 | suballocations1st.pop_back(); |
| 8864 | } |
| 8865 | |
| 8866 | // Find more null items at the end of 2nd vector. |
| 8867 | while (m_2ndNullItemsCount > 0 && |
| 8868 | suballocations2nd.back().type == VMA_SUBALLOCATION_TYPE_FREE) |
| 8869 | { |
| 8870 | --m_2ndNullItemsCount; |
| 8871 | suballocations2nd.pop_back(); |
| 8872 | } |
| 8873 | |
| 8874 | // Find more null items at the beginning of 2nd vector. |
| 8875 | while (m_2ndNullItemsCount > 0 && |
| 8876 | suballocations2nd[0].type == VMA_SUBALLOCATION_TYPE_FREE) |
| 8877 | { |
| 8878 | --m_2ndNullItemsCount; |
| 8879 | VmaVectorRemove(vec&: suballocations2nd, index: 0); |
| 8880 | } |
| 8881 | |
| 8882 | if (ShouldCompact1st()) |
| 8883 | { |
| 8884 | const size_t nonNullItemCount = suballoc1stCount - nullItem1stCount; |
| 8885 | size_t srcIndex = m_1stNullItemsBeginCount; |
| 8886 | for (size_t dstIndex = 0; dstIndex < nonNullItemCount; ++dstIndex) |
| 8887 | { |
| 8888 | while (suballocations1st[srcIndex].type == VMA_SUBALLOCATION_TYPE_FREE) |
| 8889 | { |
| 8890 | ++srcIndex; |
| 8891 | } |
| 8892 | if (dstIndex != srcIndex) |
| 8893 | { |
| 8894 | suballocations1st[dstIndex] = suballocations1st[srcIndex]; |
| 8895 | } |
| 8896 | ++srcIndex; |
| 8897 | } |
| 8898 | suballocations1st.resize(newCount: nonNullItemCount); |
| 8899 | m_1stNullItemsBeginCount = 0; |
| 8900 | m_1stNullItemsMiddleCount = 0; |
| 8901 | } |
| 8902 | |
| 8903 | // 2nd vector became empty. |
| 8904 | if (suballocations2nd.empty()) |
| 8905 | { |
| 8906 | m_2ndVectorMode = SECOND_VECTOR_EMPTY; |
| 8907 | } |
| 8908 | |
| 8909 | // 1st vector became empty. |
| 8910 | if (suballocations1st.size() - m_1stNullItemsBeginCount == 0) |
| 8911 | { |
| 8912 | suballocations1st.clear(); |
| 8913 | m_1stNullItemsBeginCount = 0; |
| 8914 | |
| 8915 | if (!suballocations2nd.empty() && m_2ndVectorMode == SECOND_VECTOR_RING_BUFFER) |
| 8916 | { |
| 8917 | // Swap 1st with 2nd. Now 2nd is empty. |
| 8918 | m_2ndVectorMode = SECOND_VECTOR_EMPTY; |
| 8919 | m_1stNullItemsMiddleCount = m_2ndNullItemsCount; |
| 8920 | while (m_1stNullItemsBeginCount < suballocations2nd.size() && |
| 8921 | suballocations2nd[m_1stNullItemsBeginCount].type == VMA_SUBALLOCATION_TYPE_FREE) |
| 8922 | { |
| 8923 | ++m_1stNullItemsBeginCount; |
| 8924 | --m_1stNullItemsMiddleCount; |
| 8925 | } |
| 8926 | m_2ndNullItemsCount = 0; |
| 8927 | m_1stVectorIndex ^= 1; |
| 8928 | } |
| 8929 | } |
| 8930 | } |
| 8931 | |
| 8932 | VMA_HEAVY_ASSERT(Validate()); |
| 8933 | } |
| 8934 | |
| 8935 | bool VmaBlockMetadata_Linear::CreateAllocationRequest_LowerAddress( |
| 8936 | VkDeviceSize allocSize, |
| 8937 | VkDeviceSize allocAlignment, |
| 8938 | VmaSuballocationType allocType, |
| 8939 | uint32_t strategy, |
| 8940 | VmaAllocationRequest* pAllocationRequest) |
| 8941 | { |
| 8942 | const VkDeviceSize blockSize = GetSize(); |
| 8943 | const VkDeviceSize debugMargin = GetDebugMargin(); |
| 8944 | const VkDeviceSize bufferImageGranularity = GetBufferImageGranularity(); |
| 8945 | SuballocationVectorType& suballocations1st = AccessSuballocations1st(); |
| 8946 | SuballocationVectorType& suballocations2nd = AccessSuballocations2nd(); |
| 8947 | |
| 8948 | if (m_2ndVectorMode == SECOND_VECTOR_EMPTY || m_2ndVectorMode == SECOND_VECTOR_DOUBLE_STACK) |
| 8949 | { |
| 8950 | // Try to allocate at the end of 1st vector. |
| 8951 | |
| 8952 | VkDeviceSize resultBaseOffset = 0; |
| 8953 | if (!suballocations1st.empty()) |
| 8954 | { |
| 8955 | const VmaSuballocation& lastSuballoc = suballocations1st.back(); |
| 8956 | resultBaseOffset = lastSuballoc.offset + lastSuballoc.size + debugMargin; |
| 8957 | } |
| 8958 | |
| 8959 | // Start from offset equal to beginning of free space. |
| 8960 | VkDeviceSize resultOffset = resultBaseOffset; |
| 8961 | |
| 8962 | // Apply alignment. |
| 8963 | resultOffset = VmaAlignUp(val: resultOffset, alignment: allocAlignment); |
| 8964 | |
| 8965 | // Check previous suballocations for BufferImageGranularity conflicts. |
| 8966 | // Make bigger alignment if necessary. |
| 8967 | if (bufferImageGranularity > 1 && bufferImageGranularity != allocAlignment && !suballocations1st.empty()) |
| 8968 | { |
| 8969 | bool bufferImageGranularityConflict = false; |
| 8970 | for (size_t prevSuballocIndex = suballocations1st.size(); prevSuballocIndex--; ) |
| 8971 | { |
| 8972 | const VmaSuballocation& prevSuballoc = suballocations1st[prevSuballocIndex]; |
| 8973 | if (VmaBlocksOnSamePage(resourceAOffset: prevSuballoc.offset, resourceASize: prevSuballoc.size, resourceBOffset: resultOffset, pageSize: bufferImageGranularity)) |
| 8974 | { |
| 8975 | if (VmaIsBufferImageGranularityConflict(suballocType1: prevSuballoc.type, suballocType2: allocType)) |
| 8976 | { |
| 8977 | bufferImageGranularityConflict = true; |
| 8978 | break; |
| 8979 | } |
| 8980 | } |
| 8981 | else |
| 8982 | // Already on previous page. |
| 8983 | break; |
| 8984 | } |
| 8985 | if (bufferImageGranularityConflict) |
| 8986 | { |
| 8987 | resultOffset = VmaAlignUp(val: resultOffset, alignment: bufferImageGranularity); |
| 8988 | } |
| 8989 | } |
| 8990 | |
| 8991 | const VkDeviceSize freeSpaceEnd = m_2ndVectorMode == SECOND_VECTOR_DOUBLE_STACK ? |
| 8992 | suballocations2nd.back().offset : blockSize; |
| 8993 | |
| 8994 | // There is enough free space at the end after alignment. |
| 8995 | if (resultOffset + allocSize + debugMargin <= freeSpaceEnd) |
| 8996 | { |
| 8997 | // Check next suballocations for BufferImageGranularity conflicts. |
| 8998 | // If conflict exists, allocation cannot be made here. |
| 8999 | if ((allocSize % bufferImageGranularity || resultOffset % bufferImageGranularity) && m_2ndVectorMode == SECOND_VECTOR_DOUBLE_STACK) |
| 9000 | { |
| 9001 | for (size_t nextSuballocIndex = suballocations2nd.size(); nextSuballocIndex--; ) |
| 9002 | { |
| 9003 | const VmaSuballocation& nextSuballoc = suballocations2nd[nextSuballocIndex]; |
| 9004 | if (VmaBlocksOnSamePage(resourceAOffset: resultOffset, resourceASize: allocSize, resourceBOffset: nextSuballoc.offset, pageSize: bufferImageGranularity)) |
| 9005 | { |
| 9006 | if (VmaIsBufferImageGranularityConflict(suballocType1: allocType, suballocType2: nextSuballoc.type)) |
| 9007 | { |
| 9008 | return false; |
| 9009 | } |
| 9010 | } |
| 9011 | else |
| 9012 | { |
| 9013 | // Already on previous page. |
| 9014 | break; |
| 9015 | } |
| 9016 | } |
| 9017 | } |
| 9018 | |
| 9019 | // All tests passed: Success. |
| 9020 | pAllocationRequest->allocHandle = (VmaAllocHandle)(resultOffset + 1); |
| 9021 | // pAllocationRequest->item, customData unused. |
| 9022 | pAllocationRequest->type = VmaAllocationRequestType::EndOf1st; |
| 9023 | return true; |
| 9024 | } |
| 9025 | } |
| 9026 | |
| 9027 | // Wrap-around to end of 2nd vector. Try to allocate there, watching for the |
| 9028 | // beginning of 1st vector as the end of free space. |
| 9029 | if (m_2ndVectorMode == SECOND_VECTOR_EMPTY || m_2ndVectorMode == SECOND_VECTOR_RING_BUFFER) |
| 9030 | { |
| 9031 | VMA_ASSERT(!suballocations1st.empty()); |
| 9032 | |
| 9033 | VkDeviceSize resultBaseOffset = 0; |
| 9034 | if (!suballocations2nd.empty()) |
| 9035 | { |
| 9036 | const VmaSuballocation& lastSuballoc = suballocations2nd.back(); |
| 9037 | resultBaseOffset = lastSuballoc.offset + lastSuballoc.size + debugMargin; |
| 9038 | } |
| 9039 | |
| 9040 | // Start from offset equal to beginning of free space. |
| 9041 | VkDeviceSize resultOffset = resultBaseOffset; |
| 9042 | |
| 9043 | // Apply alignment. |
| 9044 | resultOffset = VmaAlignUp(val: resultOffset, alignment: allocAlignment); |
| 9045 | |
| 9046 | // Check previous suballocations for BufferImageGranularity conflicts. |
| 9047 | // Make bigger alignment if necessary. |
| 9048 | if (bufferImageGranularity > 1 && bufferImageGranularity != allocAlignment && !suballocations2nd.empty()) |
| 9049 | { |
| 9050 | bool bufferImageGranularityConflict = false; |
| 9051 | for (size_t prevSuballocIndex = suballocations2nd.size(); prevSuballocIndex--; ) |
| 9052 | { |
| 9053 | const VmaSuballocation& prevSuballoc = suballocations2nd[prevSuballocIndex]; |
| 9054 | if (VmaBlocksOnSamePage(resourceAOffset: prevSuballoc.offset, resourceASize: prevSuballoc.size, resourceBOffset: resultOffset, pageSize: bufferImageGranularity)) |
| 9055 | { |
| 9056 | if (VmaIsBufferImageGranularityConflict(suballocType1: prevSuballoc.type, suballocType2: allocType)) |
| 9057 | { |
| 9058 | bufferImageGranularityConflict = true; |
| 9059 | break; |
| 9060 | } |
| 9061 | } |
| 9062 | else |
| 9063 | // Already on previous page. |
| 9064 | break; |
| 9065 | } |
| 9066 | if (bufferImageGranularityConflict) |
| 9067 | { |
| 9068 | resultOffset = VmaAlignUp(val: resultOffset, alignment: bufferImageGranularity); |
| 9069 | } |
| 9070 | } |
| 9071 | |
| 9072 | size_t index1st = m_1stNullItemsBeginCount; |
| 9073 | |
| 9074 | // There is enough free space at the end after alignment. |
| 9075 | if ((index1st == suballocations1st.size() && resultOffset + allocSize + debugMargin <= blockSize) || |
| 9076 | (index1st < suballocations1st.size() && resultOffset + allocSize + debugMargin <= suballocations1st[index1st].offset)) |
| 9077 | { |
| 9078 | // Check next suballocations for BufferImageGranularity conflicts. |
| 9079 | // If conflict exists, allocation cannot be made here. |
| 9080 | if (allocSize % bufferImageGranularity || resultOffset % bufferImageGranularity) |
| 9081 | { |
| 9082 | for (size_t nextSuballocIndex = index1st; |
| 9083 | nextSuballocIndex < suballocations1st.size(); |
| 9084 | nextSuballocIndex++) |
| 9085 | { |
| 9086 | const VmaSuballocation& nextSuballoc = suballocations1st[nextSuballocIndex]; |
| 9087 | if (VmaBlocksOnSamePage(resourceAOffset: resultOffset, resourceASize: allocSize, resourceBOffset: nextSuballoc.offset, pageSize: bufferImageGranularity)) |
| 9088 | { |
| 9089 | if (VmaIsBufferImageGranularityConflict(suballocType1: allocType, suballocType2: nextSuballoc.type)) |
| 9090 | { |
| 9091 | return false; |
| 9092 | } |
| 9093 | } |
| 9094 | else |
| 9095 | { |
| 9096 | // Already on next page. |
| 9097 | break; |
| 9098 | } |
| 9099 | } |
| 9100 | } |
| 9101 | |
| 9102 | // All tests passed: Success. |
| 9103 | pAllocationRequest->allocHandle = (VmaAllocHandle)(resultOffset + 1); |
| 9104 | pAllocationRequest->type = VmaAllocationRequestType::EndOf2nd; |
| 9105 | // pAllocationRequest->item, customData unused. |
| 9106 | return true; |
| 9107 | } |
| 9108 | } |
| 9109 | |
| 9110 | return false; |
| 9111 | } |
| 9112 | |
| 9113 | bool VmaBlockMetadata_Linear::CreateAllocationRequest_UpperAddress( |
| 9114 | VkDeviceSize allocSize, |
| 9115 | VkDeviceSize allocAlignment, |
| 9116 | VmaSuballocationType allocType, |
| 9117 | uint32_t strategy, |
| 9118 | VmaAllocationRequest* pAllocationRequest) |
| 9119 | { |
| 9120 | const VkDeviceSize blockSize = GetSize(); |
| 9121 | const VkDeviceSize bufferImageGranularity = GetBufferImageGranularity(); |
| 9122 | SuballocationVectorType& suballocations1st = AccessSuballocations1st(); |
| 9123 | SuballocationVectorType& suballocations2nd = AccessSuballocations2nd(); |
| 9124 | |
| 9125 | if (m_2ndVectorMode == SECOND_VECTOR_RING_BUFFER) |
| 9126 | { |
| 9127 | VMA_ASSERT(0 && "Trying to use pool with linear algorithm as double stack, while it is already being used as ring buffer." ); |
| 9128 | return false; |
| 9129 | } |
| 9130 | |
| 9131 | // Try to allocate before 2nd.back(), or end of block if 2nd.empty(). |
| 9132 | if (allocSize > blockSize) |
| 9133 | { |
| 9134 | return false; |
| 9135 | } |
| 9136 | VkDeviceSize resultBaseOffset = blockSize - allocSize; |
| 9137 | if (!suballocations2nd.empty()) |
| 9138 | { |
| 9139 | const VmaSuballocation& lastSuballoc = suballocations2nd.back(); |
| 9140 | resultBaseOffset = lastSuballoc.offset - allocSize; |
| 9141 | if (allocSize > lastSuballoc.offset) |
| 9142 | { |
| 9143 | return false; |
| 9144 | } |
| 9145 | } |
| 9146 | |
| 9147 | // Start from offset equal to end of free space. |
| 9148 | VkDeviceSize resultOffset = resultBaseOffset; |
| 9149 | |
| 9150 | const VkDeviceSize debugMargin = GetDebugMargin(); |
| 9151 | |
| 9152 | // Apply debugMargin at the end. |
| 9153 | if (debugMargin > 0) |
| 9154 | { |
| 9155 | if (resultOffset < debugMargin) |
| 9156 | { |
| 9157 | return false; |
| 9158 | } |
| 9159 | resultOffset -= debugMargin; |
| 9160 | } |
| 9161 | |
| 9162 | // Apply alignment. |
| 9163 | resultOffset = VmaAlignDown(val: resultOffset, alignment: allocAlignment); |
| 9164 | |
| 9165 | // Check next suballocations from 2nd for BufferImageGranularity conflicts. |
| 9166 | // Make bigger alignment if necessary. |
| 9167 | if (bufferImageGranularity > 1 && bufferImageGranularity != allocAlignment && !suballocations2nd.empty()) |
| 9168 | { |
| 9169 | bool bufferImageGranularityConflict = false; |
| 9170 | for (size_t nextSuballocIndex = suballocations2nd.size(); nextSuballocIndex--; ) |
| 9171 | { |
| 9172 | const VmaSuballocation& nextSuballoc = suballocations2nd[nextSuballocIndex]; |
| 9173 | if (VmaBlocksOnSamePage(resourceAOffset: resultOffset, resourceASize: allocSize, resourceBOffset: nextSuballoc.offset, pageSize: bufferImageGranularity)) |
| 9174 | { |
| 9175 | if (VmaIsBufferImageGranularityConflict(suballocType1: nextSuballoc.type, suballocType2: allocType)) |
| 9176 | { |
| 9177 | bufferImageGranularityConflict = true; |
| 9178 | break; |
| 9179 | } |
| 9180 | } |
| 9181 | else |
| 9182 | // Already on previous page. |
| 9183 | break; |
| 9184 | } |
| 9185 | if (bufferImageGranularityConflict) |
| 9186 | { |
| 9187 | resultOffset = VmaAlignDown(val: resultOffset, alignment: bufferImageGranularity); |
| 9188 | } |
| 9189 | } |
| 9190 | |
| 9191 | // There is enough free space. |
| 9192 | const VkDeviceSize endOf1st = !suballocations1st.empty() ? |
| 9193 | suballocations1st.back().offset + suballocations1st.back().size : |
| 9194 | 0; |
| 9195 | if (endOf1st + debugMargin <= resultOffset) |
| 9196 | { |
| 9197 | // Check previous suballocations for BufferImageGranularity conflicts. |
| 9198 | // If conflict exists, allocation cannot be made here. |
| 9199 | if (bufferImageGranularity > 1) |
| 9200 | { |
| 9201 | for (size_t prevSuballocIndex = suballocations1st.size(); prevSuballocIndex--; ) |
| 9202 | { |
| 9203 | const VmaSuballocation& prevSuballoc = suballocations1st[prevSuballocIndex]; |
| 9204 | if (VmaBlocksOnSamePage(resourceAOffset: prevSuballoc.offset, resourceASize: prevSuballoc.size, resourceBOffset: resultOffset, pageSize: bufferImageGranularity)) |
| 9205 | { |
| 9206 | if (VmaIsBufferImageGranularityConflict(suballocType1: allocType, suballocType2: prevSuballoc.type)) |
| 9207 | { |
| 9208 | return false; |
| 9209 | } |
| 9210 | } |
| 9211 | else |
| 9212 | { |
| 9213 | // Already on next page. |
| 9214 | break; |
| 9215 | } |
| 9216 | } |
| 9217 | } |
| 9218 | |
| 9219 | // All tests passed: Success. |
| 9220 | pAllocationRequest->allocHandle = (VmaAllocHandle)(resultOffset + 1); |
| 9221 | // pAllocationRequest->item unused. |
| 9222 | pAllocationRequest->type = VmaAllocationRequestType::UpperAddress; |
| 9223 | return true; |
| 9224 | } |
| 9225 | |
| 9226 | return false; |
| 9227 | } |
| 9228 | #endif // _VMA_BLOCK_METADATA_LINEAR_FUNCTIONS |
| 9229 | #endif // _VMA_BLOCK_METADATA_LINEAR |
| 9230 | |
| 9231 | #if 0 |
| 9232 | #ifndef _VMA_BLOCK_METADATA_BUDDY |
| 9233 | /* |
| 9234 | - GetSize() is the original size of allocated memory block. |
| 9235 | - m_UsableSize is this size aligned down to a power of two. |
| 9236 | All allocations and calculations happen relative to m_UsableSize. |
| 9237 | - GetUnusableSize() is the difference between them. |
| 9238 | It is reported as separate, unused range, not available for allocations. |
| 9239 | |
| 9240 | Node at level 0 has size = m_UsableSize. |
| 9241 | Each next level contains nodes with size 2 times smaller than current level. |
| 9242 | m_LevelCount is the maximum number of levels to use in the current object. |
| 9243 | */ |
| 9244 | class VmaBlockMetadata_Buddy : public VmaBlockMetadata |
| 9245 | { |
| 9246 | VMA_CLASS_NO_COPY(VmaBlockMetadata_Buddy) |
| 9247 | public: |
| 9248 | VmaBlockMetadata_Buddy(const VkAllocationCallbacks* pAllocationCallbacks, |
| 9249 | VkDeviceSize bufferImageGranularity, bool isVirtual); |
| 9250 | virtual ~VmaBlockMetadata_Buddy(); |
| 9251 | |
| 9252 | size_t GetAllocationCount() const override { return m_AllocationCount; } |
| 9253 | VkDeviceSize GetSumFreeSize() const override { return m_SumFreeSize + GetUnusableSize(); } |
| 9254 | bool IsEmpty() const override { return m_Root->type == Node::TYPE_FREE; } |
| 9255 | VkResult CheckCorruption(const void* pBlockData) override { return VK_ERROR_FEATURE_NOT_PRESENT; } |
| 9256 | VkDeviceSize GetAllocationOffset(VmaAllocHandle allocHandle) const override { return (VkDeviceSize)allocHandle - 1; }; |
| 9257 | void DebugLogAllAllocations() const override { DebugLogAllAllocationNode(m_Root, 0); } |
| 9258 | |
| 9259 | void Init(VkDeviceSize size) override; |
| 9260 | bool Validate() const override; |
| 9261 | |
| 9262 | void AddDetailedStatistics(VmaDetailedStatistics& inoutStats) const override; |
| 9263 | void AddStatistics(VmaStatistics& inoutStats) const override; |
| 9264 | |
| 9265 | #if VMA_STATS_STRING_ENABLED |
| 9266 | void PrintDetailedMap(class VmaJsonWriter& json, uint32_t mapRefCount) const override; |
| 9267 | #endif |
| 9268 | |
| 9269 | bool CreateAllocationRequest( |
| 9270 | VkDeviceSize allocSize, |
| 9271 | VkDeviceSize allocAlignment, |
| 9272 | bool upperAddress, |
| 9273 | VmaSuballocationType allocType, |
| 9274 | uint32_t strategy, |
| 9275 | VmaAllocationRequest* pAllocationRequest) override; |
| 9276 | |
| 9277 | void Alloc( |
| 9278 | const VmaAllocationRequest& request, |
| 9279 | VmaSuballocationType type, |
| 9280 | void* userData) override; |
| 9281 | |
| 9282 | void Free(VmaAllocHandle allocHandle) override; |
| 9283 | void GetAllocationInfo(VmaAllocHandle allocHandle, VmaVirtualAllocationInfo& outInfo) override; |
| 9284 | void* GetAllocationUserData(VmaAllocHandle allocHandle) const override; |
| 9285 | VmaAllocHandle GetAllocationListBegin() const override; |
| 9286 | VmaAllocHandle GetNextAllocation(VmaAllocHandle prevAlloc) const override; |
| 9287 | void Clear() override; |
| 9288 | void SetAllocationUserData(VmaAllocHandle allocHandle, void* userData) override; |
| 9289 | |
| 9290 | private: |
| 9291 | static const size_t MAX_LEVELS = 48; |
| 9292 | |
| 9293 | struct ValidationContext |
| 9294 | { |
| 9295 | size_t calculatedAllocationCount = 0; |
| 9296 | size_t calculatedFreeCount = 0; |
| 9297 | VkDeviceSize calculatedSumFreeSize = 0; |
| 9298 | }; |
| 9299 | struct Node |
| 9300 | { |
| 9301 | VkDeviceSize offset; |
| 9302 | enum TYPE |
| 9303 | { |
| 9304 | TYPE_FREE, |
| 9305 | TYPE_ALLOCATION, |
| 9306 | TYPE_SPLIT, |
| 9307 | TYPE_COUNT |
| 9308 | } type; |
| 9309 | Node* parent; |
| 9310 | Node* buddy; |
| 9311 | |
| 9312 | union |
| 9313 | { |
| 9314 | struct |
| 9315 | { |
| 9316 | Node* prev; |
| 9317 | Node* next; |
| 9318 | } free; |
| 9319 | struct |
| 9320 | { |
| 9321 | void* userData; |
| 9322 | } allocation; |
| 9323 | struct |
| 9324 | { |
| 9325 | Node* leftChild; |
| 9326 | } split; |
| 9327 | }; |
| 9328 | }; |
| 9329 | |
| 9330 | // Size of the memory block aligned down to a power of two. |
| 9331 | VkDeviceSize m_UsableSize; |
| 9332 | uint32_t m_LevelCount; |
| 9333 | VmaPoolAllocator<Node> m_NodeAllocator; |
| 9334 | Node* m_Root; |
| 9335 | struct |
| 9336 | { |
| 9337 | Node* front; |
| 9338 | Node* back; |
| 9339 | } m_FreeList[MAX_LEVELS]; |
| 9340 | |
| 9341 | // Number of nodes in the tree with type == TYPE_ALLOCATION. |
| 9342 | size_t m_AllocationCount; |
| 9343 | // Number of nodes in the tree with type == TYPE_FREE. |
| 9344 | size_t m_FreeCount; |
| 9345 | // Doesn't include space wasted due to internal fragmentation - allocation sizes are just aligned up to node sizes. |
| 9346 | // Doesn't include unusable size. |
| 9347 | VkDeviceSize m_SumFreeSize; |
| 9348 | |
| 9349 | VkDeviceSize GetUnusableSize() const { return GetSize() - m_UsableSize; } |
| 9350 | VkDeviceSize LevelToNodeSize(uint32_t level) const { return m_UsableSize >> level; } |
| 9351 | |
| 9352 | VkDeviceSize AlignAllocationSize(VkDeviceSize size) const |
| 9353 | { |
| 9354 | if (!IsVirtual()) |
| 9355 | { |
| 9356 | size = VmaAlignUp(size, (VkDeviceSize)16); |
| 9357 | } |
| 9358 | return VmaNextPow2(size); |
| 9359 | } |
| 9360 | Node* FindAllocationNode(VkDeviceSize offset, uint32_t& outLevel) const; |
| 9361 | void DeleteNodeChildren(Node* node); |
| 9362 | bool ValidateNode(ValidationContext& ctx, const Node* parent, const Node* curr, uint32_t level, VkDeviceSize levelNodeSize) const; |
| 9363 | uint32_t AllocSizeToLevel(VkDeviceSize allocSize) const; |
| 9364 | void AddNodeToDetailedStatistics(VmaDetailedStatistics& inoutStats, const Node* node, VkDeviceSize levelNodeSize) const; |
| 9365 | // Adds node to the front of FreeList at given level. |
| 9366 | // node->type must be FREE. |
| 9367 | // node->free.prev, next can be undefined. |
| 9368 | void AddToFreeListFront(uint32_t level, Node* node); |
| 9369 | // Removes node from FreeList at given level. |
| 9370 | // node->type must be FREE. |
| 9371 | // node->free.prev, next stay untouched. |
| 9372 | void RemoveFromFreeList(uint32_t level, Node* node); |
| 9373 | void DebugLogAllAllocationNode(Node* node, uint32_t level) const; |
| 9374 | |
| 9375 | #if VMA_STATS_STRING_ENABLED |
| 9376 | void PrintDetailedMapNode(class VmaJsonWriter& json, const Node* node, VkDeviceSize levelNodeSize) const; |
| 9377 | #endif |
| 9378 | }; |
| 9379 | |
| 9380 | #ifndef _VMA_BLOCK_METADATA_BUDDY_FUNCTIONS |
| 9381 | VmaBlockMetadata_Buddy::VmaBlockMetadata_Buddy(const VkAllocationCallbacks* pAllocationCallbacks, |
| 9382 | VkDeviceSize bufferImageGranularity, bool isVirtual) |
| 9383 | : VmaBlockMetadata(pAllocationCallbacks, bufferImageGranularity, isVirtual), |
| 9384 | m_NodeAllocator(pAllocationCallbacks, 32), // firstBlockCapacity |
| 9385 | m_Root(VMA_NULL), |
| 9386 | m_AllocationCount(0), |
| 9387 | m_FreeCount(1), |
| 9388 | m_SumFreeSize(0) |
| 9389 | { |
| 9390 | memset(m_FreeList, 0, sizeof(m_FreeList)); |
| 9391 | } |
| 9392 | |
| 9393 | VmaBlockMetadata_Buddy::~VmaBlockMetadata_Buddy() |
| 9394 | { |
| 9395 | DeleteNodeChildren(m_Root); |
| 9396 | m_NodeAllocator.Free(m_Root); |
| 9397 | } |
| 9398 | |
| 9399 | void VmaBlockMetadata_Buddy::Init(VkDeviceSize size) |
| 9400 | { |
| 9401 | VmaBlockMetadata::Init(size); |
| 9402 | |
| 9403 | m_UsableSize = VmaPrevPow2(size); |
| 9404 | m_SumFreeSize = m_UsableSize; |
| 9405 | |
| 9406 | // Calculate m_LevelCount. |
| 9407 | const VkDeviceSize minNodeSize = IsVirtual() ? 1 : 16; |
| 9408 | m_LevelCount = 1; |
| 9409 | while (m_LevelCount < MAX_LEVELS && |
| 9410 | LevelToNodeSize(m_LevelCount) >= minNodeSize) |
| 9411 | { |
| 9412 | ++m_LevelCount; |
| 9413 | } |
| 9414 | |
| 9415 | Node* rootNode = m_NodeAllocator.Alloc(); |
| 9416 | rootNode->offset = 0; |
| 9417 | rootNode->type = Node::TYPE_FREE; |
| 9418 | rootNode->parent = VMA_NULL; |
| 9419 | rootNode->buddy = VMA_NULL; |
| 9420 | |
| 9421 | m_Root = rootNode; |
| 9422 | AddToFreeListFront(0, rootNode); |
| 9423 | } |
| 9424 | |
| 9425 | bool VmaBlockMetadata_Buddy::Validate() const |
| 9426 | { |
| 9427 | // Validate tree. |
| 9428 | ValidationContext ctx; |
| 9429 | if (!ValidateNode(ctx, VMA_NULL, m_Root, 0, LevelToNodeSize(0))) |
| 9430 | { |
| 9431 | VMA_VALIDATE(false && "ValidateNode failed." ); |
| 9432 | } |
| 9433 | VMA_VALIDATE(m_AllocationCount == ctx.calculatedAllocationCount); |
| 9434 | VMA_VALIDATE(m_SumFreeSize == ctx.calculatedSumFreeSize); |
| 9435 | |
| 9436 | // Validate free node lists. |
| 9437 | for (uint32_t level = 0; level < m_LevelCount; ++level) |
| 9438 | { |
| 9439 | VMA_VALIDATE(m_FreeList[level].front == VMA_NULL || |
| 9440 | m_FreeList[level].front->free.prev == VMA_NULL); |
| 9441 | |
| 9442 | for (Node* node = m_FreeList[level].front; |
| 9443 | node != VMA_NULL; |
| 9444 | node = node->free.next) |
| 9445 | { |
| 9446 | VMA_VALIDATE(node->type == Node::TYPE_FREE); |
| 9447 | |
| 9448 | if (node->free.next == VMA_NULL) |
| 9449 | { |
| 9450 | VMA_VALIDATE(m_FreeList[level].back == node); |
| 9451 | } |
| 9452 | else |
| 9453 | { |
| 9454 | VMA_VALIDATE(node->free.next->free.prev == node); |
| 9455 | } |
| 9456 | } |
| 9457 | } |
| 9458 | |
| 9459 | // Validate that free lists ar higher levels are empty. |
| 9460 | for (uint32_t level = m_LevelCount; level < MAX_LEVELS; ++level) |
| 9461 | { |
| 9462 | VMA_VALIDATE(m_FreeList[level].front == VMA_NULL && m_FreeList[level].back == VMA_NULL); |
| 9463 | } |
| 9464 | |
| 9465 | return true; |
| 9466 | } |
| 9467 | |
| 9468 | void VmaBlockMetadata_Buddy::AddDetailedStatistics(VmaDetailedStatistics& inoutStats) const |
| 9469 | { |
| 9470 | inoutStats.statistics.blockCount++; |
| 9471 | inoutStats.statistics.blockBytes += GetSize(); |
| 9472 | |
| 9473 | AddNodeToDetailedStatistics(inoutStats, m_Root, LevelToNodeSize(0)); |
| 9474 | |
| 9475 | const VkDeviceSize unusableSize = GetUnusableSize(); |
| 9476 | if (unusableSize > 0) |
| 9477 | VmaAddDetailedStatisticsUnusedRange(inoutStats, unusableSize); |
| 9478 | } |
| 9479 | |
| 9480 | void VmaBlockMetadata_Buddy::AddStatistics(VmaStatistics& inoutStats) const |
| 9481 | { |
| 9482 | inoutStats.blockCount++; |
| 9483 | inoutStats.allocationCount += (uint32_t)m_AllocationCount; |
| 9484 | inoutStats.blockBytes += GetSize(); |
| 9485 | inoutStats.allocationBytes += GetSize() - m_SumFreeSize; |
| 9486 | } |
| 9487 | |
| 9488 | #if VMA_STATS_STRING_ENABLED |
| 9489 | void VmaBlockMetadata_Buddy::PrintDetailedMap(class VmaJsonWriter& json, uint32_t mapRefCount) const |
| 9490 | { |
| 9491 | VmaDetailedStatistics stats; |
| 9492 | VmaClearDetailedStatistics(stats); |
| 9493 | AddDetailedStatistics(stats); |
| 9494 | |
| 9495 | PrintDetailedMap_Begin( |
| 9496 | json, |
| 9497 | stats.statistics.blockBytes - stats.statistics.allocationBytes, |
| 9498 | stats.statistics.allocationCount, |
| 9499 | stats.unusedRangeCount, |
| 9500 | mapRefCount); |
| 9501 | |
| 9502 | PrintDetailedMapNode(json, m_Root, LevelToNodeSize(0)); |
| 9503 | |
| 9504 | const VkDeviceSize unusableSize = GetUnusableSize(); |
| 9505 | if (unusableSize > 0) |
| 9506 | { |
| 9507 | PrintDetailedMap_UnusedRange(json, |
| 9508 | m_UsableSize, // offset |
| 9509 | unusableSize); // size |
| 9510 | } |
| 9511 | |
| 9512 | PrintDetailedMap_End(json); |
| 9513 | } |
| 9514 | #endif // VMA_STATS_STRING_ENABLED |
| 9515 | |
| 9516 | bool VmaBlockMetadata_Buddy::CreateAllocationRequest( |
| 9517 | VkDeviceSize allocSize, |
| 9518 | VkDeviceSize allocAlignment, |
| 9519 | bool upperAddress, |
| 9520 | VmaSuballocationType allocType, |
| 9521 | uint32_t strategy, |
| 9522 | VmaAllocationRequest* pAllocationRequest) |
| 9523 | { |
| 9524 | VMA_ASSERT(!upperAddress && "VMA_ALLOCATION_CREATE_UPPER_ADDRESS_BIT can be used only with linear algorithm." ); |
| 9525 | |
| 9526 | allocSize = AlignAllocationSize(allocSize); |
| 9527 | |
| 9528 | // Simple way to respect bufferImageGranularity. May be optimized some day. |
| 9529 | // Whenever it might be an OPTIMAL image... |
| 9530 | if (allocType == VMA_SUBALLOCATION_TYPE_UNKNOWN || |
| 9531 | allocType == VMA_SUBALLOCATION_TYPE_IMAGE_UNKNOWN || |
| 9532 | allocType == VMA_SUBALLOCATION_TYPE_IMAGE_OPTIMAL) |
| 9533 | { |
| 9534 | allocAlignment = VMA_MAX(allocAlignment, GetBufferImageGranularity()); |
| 9535 | allocSize = VmaAlignUp(allocSize, GetBufferImageGranularity()); |
| 9536 | } |
| 9537 | |
| 9538 | if (allocSize > m_UsableSize) |
| 9539 | { |
| 9540 | return false; |
| 9541 | } |
| 9542 | |
| 9543 | const uint32_t targetLevel = AllocSizeToLevel(allocSize); |
| 9544 | for (uint32_t level = targetLevel; level--; ) |
| 9545 | { |
| 9546 | for (Node* freeNode = m_FreeList[level].front; |
| 9547 | freeNode != VMA_NULL; |
| 9548 | freeNode = freeNode->free.next) |
| 9549 | { |
| 9550 | if (freeNode->offset % allocAlignment == 0) |
| 9551 | { |
| 9552 | pAllocationRequest->type = VmaAllocationRequestType::Normal; |
| 9553 | pAllocationRequest->allocHandle = (VmaAllocHandle)(freeNode->offset + 1); |
| 9554 | pAllocationRequest->size = allocSize; |
| 9555 | pAllocationRequest->customData = (void*)(uintptr_t)level; |
| 9556 | return true; |
| 9557 | } |
| 9558 | } |
| 9559 | } |
| 9560 | |
| 9561 | return false; |
| 9562 | } |
| 9563 | |
| 9564 | void VmaBlockMetadata_Buddy::Alloc( |
| 9565 | const VmaAllocationRequest& request, |
| 9566 | VmaSuballocationType type, |
| 9567 | void* userData) |
| 9568 | { |
| 9569 | VMA_ASSERT(request.type == VmaAllocationRequestType::Normal); |
| 9570 | |
| 9571 | const uint32_t targetLevel = AllocSizeToLevel(request.size); |
| 9572 | uint32_t currLevel = (uint32_t)(uintptr_t)request.customData; |
| 9573 | |
| 9574 | Node* currNode = m_FreeList[currLevel].front; |
| 9575 | VMA_ASSERT(currNode != VMA_NULL && currNode->type == Node::TYPE_FREE); |
| 9576 | const VkDeviceSize offset = (VkDeviceSize)request.allocHandle - 1; |
| 9577 | while (currNode->offset != offset) |
| 9578 | { |
| 9579 | currNode = currNode->free.next; |
| 9580 | VMA_ASSERT(currNode != VMA_NULL && currNode->type == Node::TYPE_FREE); |
| 9581 | } |
| 9582 | |
| 9583 | // Go down, splitting free nodes. |
| 9584 | while (currLevel < targetLevel) |
| 9585 | { |
| 9586 | // currNode is already first free node at currLevel. |
| 9587 | // Remove it from list of free nodes at this currLevel. |
| 9588 | RemoveFromFreeList(currLevel, currNode); |
| 9589 | |
| 9590 | const uint32_t childrenLevel = currLevel + 1; |
| 9591 | |
| 9592 | // Create two free sub-nodes. |
| 9593 | Node* leftChild = m_NodeAllocator.Alloc(); |
| 9594 | Node* rightChild = m_NodeAllocator.Alloc(); |
| 9595 | |
| 9596 | leftChild->offset = currNode->offset; |
| 9597 | leftChild->type = Node::TYPE_FREE; |
| 9598 | leftChild->parent = currNode; |
| 9599 | leftChild->buddy = rightChild; |
| 9600 | |
| 9601 | rightChild->offset = currNode->offset + LevelToNodeSize(childrenLevel); |
| 9602 | rightChild->type = Node::TYPE_FREE; |
| 9603 | rightChild->parent = currNode; |
| 9604 | rightChild->buddy = leftChild; |
| 9605 | |
| 9606 | // Convert current currNode to split type. |
| 9607 | currNode->type = Node::TYPE_SPLIT; |
| 9608 | currNode->split.leftChild = leftChild; |
| 9609 | |
| 9610 | // Add child nodes to free list. Order is important! |
| 9611 | AddToFreeListFront(childrenLevel, rightChild); |
| 9612 | AddToFreeListFront(childrenLevel, leftChild); |
| 9613 | |
| 9614 | ++m_FreeCount; |
| 9615 | ++currLevel; |
| 9616 | currNode = m_FreeList[currLevel].front; |
| 9617 | |
| 9618 | /* |
| 9619 | We can be sure that currNode, as left child of node previously split, |
| 9620 | also fulfills the alignment requirement. |
| 9621 | */ |
| 9622 | } |
| 9623 | |
| 9624 | // Remove from free list. |
| 9625 | VMA_ASSERT(currLevel == targetLevel && |
| 9626 | currNode != VMA_NULL && |
| 9627 | currNode->type == Node::TYPE_FREE); |
| 9628 | RemoveFromFreeList(currLevel, currNode); |
| 9629 | |
| 9630 | // Convert to allocation node. |
| 9631 | currNode->type = Node::TYPE_ALLOCATION; |
| 9632 | currNode->allocation.userData = userData; |
| 9633 | |
| 9634 | ++m_AllocationCount; |
| 9635 | --m_FreeCount; |
| 9636 | m_SumFreeSize -= request.size; |
| 9637 | } |
| 9638 | |
| 9639 | void VmaBlockMetadata_Buddy::GetAllocationInfo(VmaAllocHandle allocHandle, VmaVirtualAllocationInfo& outInfo) |
| 9640 | { |
| 9641 | uint32_t level = 0; |
| 9642 | outInfo.offset = (VkDeviceSize)allocHandle - 1; |
| 9643 | const Node* const node = FindAllocationNode(outInfo.offset, level); |
| 9644 | outInfo.size = LevelToNodeSize(level); |
| 9645 | outInfo.pUserData = node->allocation.userData; |
| 9646 | } |
| 9647 | |
| 9648 | void* VmaBlockMetadata_Buddy::GetAllocationUserData(VmaAllocHandle allocHandle) const |
| 9649 | { |
| 9650 | uint32_t level = 0; |
| 9651 | const Node* const node = FindAllocationNode((VkDeviceSize)allocHandle - 1, level); |
| 9652 | return node->allocation.userData; |
| 9653 | } |
| 9654 | |
| 9655 | VmaAllocHandle VmaBlockMetadata_Buddy::GetAllocationListBegin() const |
| 9656 | { |
| 9657 | // Function only used for defragmentation, which is disabled for this algorithm |
| 9658 | return VK_NULL_HANDLE; |
| 9659 | } |
| 9660 | |
| 9661 | VmaAllocHandle VmaBlockMetadata_Buddy::GetNextAllocation(VmaAllocHandle prevAlloc) const |
| 9662 | { |
| 9663 | // Function only used for defragmentation, which is disabled for this algorithm |
| 9664 | return VK_NULL_HANDLE; |
| 9665 | } |
| 9666 | |
| 9667 | void VmaBlockMetadata_Buddy::DeleteNodeChildren(Node* node) |
| 9668 | { |
| 9669 | if (node->type == Node::TYPE_SPLIT) |
| 9670 | { |
| 9671 | DeleteNodeChildren(node->split.leftChild->buddy); |
| 9672 | DeleteNodeChildren(node->split.leftChild); |
| 9673 | const VkAllocationCallbacks* allocationCallbacks = GetAllocationCallbacks(); |
| 9674 | m_NodeAllocator.Free(node->split.leftChild->buddy); |
| 9675 | m_NodeAllocator.Free(node->split.leftChild); |
| 9676 | } |
| 9677 | } |
| 9678 | |
| 9679 | void VmaBlockMetadata_Buddy::Clear() |
| 9680 | { |
| 9681 | DeleteNodeChildren(m_Root); |
| 9682 | m_Root->type = Node::TYPE_FREE; |
| 9683 | m_AllocationCount = 0; |
| 9684 | m_FreeCount = 1; |
| 9685 | m_SumFreeSize = m_UsableSize; |
| 9686 | } |
| 9687 | |
| 9688 | void VmaBlockMetadata_Buddy::SetAllocationUserData(VmaAllocHandle allocHandle, void* userData) |
| 9689 | { |
| 9690 | uint32_t level = 0; |
| 9691 | Node* const node = FindAllocationNode((VkDeviceSize)allocHandle - 1, level); |
| 9692 | node->allocation.userData = userData; |
| 9693 | } |
| 9694 | |
| 9695 | VmaBlockMetadata_Buddy::Node* VmaBlockMetadata_Buddy::FindAllocationNode(VkDeviceSize offset, uint32_t& outLevel) const |
| 9696 | { |
| 9697 | Node* node = m_Root; |
| 9698 | VkDeviceSize nodeOffset = 0; |
| 9699 | outLevel = 0; |
| 9700 | VkDeviceSize levelNodeSize = LevelToNodeSize(0); |
| 9701 | while (node->type == Node::TYPE_SPLIT) |
| 9702 | { |
| 9703 | const VkDeviceSize nextLevelNodeSize = levelNodeSize >> 1; |
| 9704 | if (offset < nodeOffset + nextLevelNodeSize) |
| 9705 | { |
| 9706 | node = node->split.leftChild; |
| 9707 | } |
| 9708 | else |
| 9709 | { |
| 9710 | node = node->split.leftChild->buddy; |
| 9711 | nodeOffset += nextLevelNodeSize; |
| 9712 | } |
| 9713 | ++outLevel; |
| 9714 | levelNodeSize = nextLevelNodeSize; |
| 9715 | } |
| 9716 | |
| 9717 | VMA_ASSERT(node != VMA_NULL && node->type == Node::TYPE_ALLOCATION); |
| 9718 | return node; |
| 9719 | } |
| 9720 | |
| 9721 | bool VmaBlockMetadata_Buddy::ValidateNode(ValidationContext& ctx, const Node* parent, const Node* curr, uint32_t level, VkDeviceSize levelNodeSize) const |
| 9722 | { |
| 9723 | VMA_VALIDATE(level < m_LevelCount); |
| 9724 | VMA_VALIDATE(curr->parent == parent); |
| 9725 | VMA_VALIDATE((curr->buddy == VMA_NULL) == (parent == VMA_NULL)); |
| 9726 | VMA_VALIDATE(curr->buddy == VMA_NULL || curr->buddy->buddy == curr); |
| 9727 | switch (curr->type) |
| 9728 | { |
| 9729 | case Node::TYPE_FREE: |
| 9730 | // curr->free.prev, next are validated separately. |
| 9731 | ctx.calculatedSumFreeSize += levelNodeSize; |
| 9732 | ++ctx.calculatedFreeCount; |
| 9733 | break; |
| 9734 | case Node::TYPE_ALLOCATION: |
| 9735 | ++ctx.calculatedAllocationCount; |
| 9736 | if (!IsVirtual()) |
| 9737 | { |
| 9738 | VMA_VALIDATE(curr->allocation.userData != VMA_NULL); |
| 9739 | } |
| 9740 | break; |
| 9741 | case Node::TYPE_SPLIT: |
| 9742 | { |
| 9743 | const uint32_t childrenLevel = level + 1; |
| 9744 | const VkDeviceSize childrenLevelNodeSize = levelNodeSize >> 1; |
| 9745 | const Node* const leftChild = curr->split.leftChild; |
| 9746 | VMA_VALIDATE(leftChild != VMA_NULL); |
| 9747 | VMA_VALIDATE(leftChild->offset == curr->offset); |
| 9748 | if (!ValidateNode(ctx, curr, leftChild, childrenLevel, childrenLevelNodeSize)) |
| 9749 | { |
| 9750 | VMA_VALIDATE(false && "ValidateNode for left child failed." ); |
| 9751 | } |
| 9752 | const Node* const rightChild = leftChild->buddy; |
| 9753 | VMA_VALIDATE(rightChild->offset == curr->offset + childrenLevelNodeSize); |
| 9754 | if (!ValidateNode(ctx, curr, rightChild, childrenLevel, childrenLevelNodeSize)) |
| 9755 | { |
| 9756 | VMA_VALIDATE(false && "ValidateNode for right child failed." ); |
| 9757 | } |
| 9758 | } |
| 9759 | break; |
| 9760 | default: |
| 9761 | return false; |
| 9762 | } |
| 9763 | |
| 9764 | return true; |
| 9765 | } |
| 9766 | |
| 9767 | uint32_t VmaBlockMetadata_Buddy::AllocSizeToLevel(VkDeviceSize allocSize) const |
| 9768 | { |
| 9769 | // I know this could be optimized somehow e.g. by using std::log2p1 from C++20. |
| 9770 | uint32_t level = 0; |
| 9771 | VkDeviceSize currLevelNodeSize = m_UsableSize; |
| 9772 | VkDeviceSize nextLevelNodeSize = currLevelNodeSize >> 1; |
| 9773 | while (allocSize <= nextLevelNodeSize && level + 1 < m_LevelCount) |
| 9774 | { |
| 9775 | ++level; |
| 9776 | currLevelNodeSize >>= 1; |
| 9777 | nextLevelNodeSize >>= 1; |
| 9778 | } |
| 9779 | return level; |
| 9780 | } |
| 9781 | |
| 9782 | void VmaBlockMetadata_Buddy::Free(VmaAllocHandle allocHandle) |
| 9783 | { |
| 9784 | uint32_t level = 0; |
| 9785 | Node* node = FindAllocationNode((VkDeviceSize)allocHandle - 1, level); |
| 9786 | |
| 9787 | ++m_FreeCount; |
| 9788 | --m_AllocationCount; |
| 9789 | m_SumFreeSize += LevelToNodeSize(level); |
| 9790 | |
| 9791 | node->type = Node::TYPE_FREE; |
| 9792 | |
| 9793 | // Join free nodes if possible. |
| 9794 | while (level > 0 && node->buddy->type == Node::TYPE_FREE) |
| 9795 | { |
| 9796 | RemoveFromFreeList(level, node->buddy); |
| 9797 | Node* const parent = node->parent; |
| 9798 | |
| 9799 | m_NodeAllocator.Free(node->buddy); |
| 9800 | m_NodeAllocator.Free(node); |
| 9801 | parent->type = Node::TYPE_FREE; |
| 9802 | |
| 9803 | node = parent; |
| 9804 | --level; |
| 9805 | --m_FreeCount; |
| 9806 | } |
| 9807 | |
| 9808 | AddToFreeListFront(level, node); |
| 9809 | } |
| 9810 | |
| 9811 | void VmaBlockMetadata_Buddy::AddNodeToDetailedStatistics(VmaDetailedStatistics& inoutStats, const Node* node, VkDeviceSize levelNodeSize) const |
| 9812 | { |
| 9813 | switch (node->type) |
| 9814 | { |
| 9815 | case Node::TYPE_FREE: |
| 9816 | VmaAddDetailedStatisticsUnusedRange(inoutStats, levelNodeSize); |
| 9817 | break; |
| 9818 | case Node::TYPE_ALLOCATION: |
| 9819 | VmaAddDetailedStatisticsAllocation(inoutStats, levelNodeSize); |
| 9820 | break; |
| 9821 | case Node::TYPE_SPLIT: |
| 9822 | { |
| 9823 | const VkDeviceSize childrenNodeSize = levelNodeSize / 2; |
| 9824 | const Node* const leftChild = node->split.leftChild; |
| 9825 | AddNodeToDetailedStatistics(inoutStats, leftChild, childrenNodeSize); |
| 9826 | const Node* const rightChild = leftChild->buddy; |
| 9827 | AddNodeToDetailedStatistics(inoutStats, rightChild, childrenNodeSize); |
| 9828 | } |
| 9829 | break; |
| 9830 | default: |
| 9831 | VMA_ASSERT(0); |
| 9832 | } |
| 9833 | } |
| 9834 | |
| 9835 | void VmaBlockMetadata_Buddy::AddToFreeListFront(uint32_t level, Node* node) |
| 9836 | { |
| 9837 | VMA_ASSERT(node->type == Node::TYPE_FREE); |
| 9838 | |
| 9839 | // List is empty. |
| 9840 | Node* const frontNode = m_FreeList[level].front; |
| 9841 | if (frontNode == VMA_NULL) |
| 9842 | { |
| 9843 | VMA_ASSERT(m_FreeList[level].back == VMA_NULL); |
| 9844 | node->free.prev = node->free.next = VMA_NULL; |
| 9845 | m_FreeList[level].front = m_FreeList[level].back = node; |
| 9846 | } |
| 9847 | else |
| 9848 | { |
| 9849 | VMA_ASSERT(frontNode->free.prev == VMA_NULL); |
| 9850 | node->free.prev = VMA_NULL; |
| 9851 | node->free.next = frontNode; |
| 9852 | frontNode->free.prev = node; |
| 9853 | m_FreeList[level].front = node; |
| 9854 | } |
| 9855 | } |
| 9856 | |
| 9857 | void VmaBlockMetadata_Buddy::RemoveFromFreeList(uint32_t level, Node* node) |
| 9858 | { |
| 9859 | VMA_ASSERT(m_FreeList[level].front != VMA_NULL); |
| 9860 | |
| 9861 | // It is at the front. |
| 9862 | if (node->free.prev == VMA_NULL) |
| 9863 | { |
| 9864 | VMA_ASSERT(m_FreeList[level].front == node); |
| 9865 | m_FreeList[level].front = node->free.next; |
| 9866 | } |
| 9867 | else |
| 9868 | { |
| 9869 | Node* const prevFreeNode = node->free.prev; |
| 9870 | VMA_ASSERT(prevFreeNode->free.next == node); |
| 9871 | prevFreeNode->free.next = node->free.next; |
| 9872 | } |
| 9873 | |
| 9874 | // It is at the back. |
| 9875 | if (node->free.next == VMA_NULL) |
| 9876 | { |
| 9877 | VMA_ASSERT(m_FreeList[level].back == node); |
| 9878 | m_FreeList[level].back = node->free.prev; |
| 9879 | } |
| 9880 | else |
| 9881 | { |
| 9882 | Node* const nextFreeNode = node->free.next; |
| 9883 | VMA_ASSERT(nextFreeNode->free.prev == node); |
| 9884 | nextFreeNode->free.prev = node->free.prev; |
| 9885 | } |
| 9886 | } |
| 9887 | |
| 9888 | void VmaBlockMetadata_Buddy::DebugLogAllAllocationNode(Node* node, uint32_t level) const |
| 9889 | { |
| 9890 | switch (node->type) |
| 9891 | { |
| 9892 | case Node::TYPE_FREE: |
| 9893 | break; |
| 9894 | case Node::TYPE_ALLOCATION: |
| 9895 | DebugLogAllocation(node->offset, LevelToNodeSize(level), node->allocation.userData); |
| 9896 | break; |
| 9897 | case Node::TYPE_SPLIT: |
| 9898 | { |
| 9899 | ++level; |
| 9900 | DebugLogAllAllocationNode(node->split.leftChild, level); |
| 9901 | DebugLogAllAllocationNode(node->split.leftChild->buddy, level); |
| 9902 | } |
| 9903 | break; |
| 9904 | default: |
| 9905 | VMA_ASSERT(0); |
| 9906 | } |
| 9907 | } |
| 9908 | |
| 9909 | #if VMA_STATS_STRING_ENABLED |
| 9910 | void VmaBlockMetadata_Buddy::PrintDetailedMapNode(class VmaJsonWriter& json, const Node* node, VkDeviceSize levelNodeSize) const |
| 9911 | { |
| 9912 | switch (node->type) |
| 9913 | { |
| 9914 | case Node::TYPE_FREE: |
| 9915 | PrintDetailedMap_UnusedRange(json, node->offset, levelNodeSize); |
| 9916 | break; |
| 9917 | case Node::TYPE_ALLOCATION: |
| 9918 | PrintDetailedMap_Allocation(json, node->offset, levelNodeSize, node->allocation.userData); |
| 9919 | break; |
| 9920 | case Node::TYPE_SPLIT: |
| 9921 | { |
| 9922 | const VkDeviceSize childrenNodeSize = levelNodeSize / 2; |
| 9923 | const Node* const leftChild = node->split.leftChild; |
| 9924 | PrintDetailedMapNode(json, leftChild, childrenNodeSize); |
| 9925 | const Node* const rightChild = leftChild->buddy; |
| 9926 | PrintDetailedMapNode(json, rightChild, childrenNodeSize); |
| 9927 | } |
| 9928 | break; |
| 9929 | default: |
| 9930 | VMA_ASSERT(0); |
| 9931 | } |
| 9932 | } |
| 9933 | #endif // VMA_STATS_STRING_ENABLED |
| 9934 | #endif // _VMA_BLOCK_METADATA_BUDDY_FUNCTIONS |
| 9935 | #endif // _VMA_BLOCK_METADATA_BUDDY |
| 9936 | #endif // #if 0 |
| 9937 | |
| 9938 | #ifndef _VMA_BLOCK_METADATA_TLSF |
| 9939 | // To not search current larger region if first allocation won't succeed and skip to smaller range |
| 9940 | // use with VMA_ALLOCATION_CREATE_STRATEGY_MIN_MEMORY_BIT as strategy in CreateAllocationRequest(). |
| 9941 | // When fragmentation and reusal of previous blocks doesn't matter then use with |
| 9942 | // VMA_ALLOCATION_CREATE_STRATEGY_MIN_TIME_BIT for fastest alloc time possible. |
| 9943 | class VmaBlockMetadata_TLSF : public VmaBlockMetadata |
| 9944 | { |
| 9945 | VMA_CLASS_NO_COPY(VmaBlockMetadata_TLSF) |
| 9946 | public: |
| 9947 | VmaBlockMetadata_TLSF(const VkAllocationCallbacks* pAllocationCallbacks, |
| 9948 | VkDeviceSize bufferImageGranularity, bool isVirtual); |
| 9949 | virtual ~VmaBlockMetadata_TLSF(); |
| 9950 | |
| 9951 | size_t GetAllocationCount() const override { return m_AllocCount; } |
| 9952 | size_t GetFreeRegionsCount() const override { return m_BlocksFreeCount + 1; } |
| 9953 | VkDeviceSize GetSumFreeSize() const override { return m_BlocksFreeSize + m_NullBlock->size; } |
| 9954 | bool IsEmpty() const override { return m_NullBlock->offset == 0; } |
| 9955 | VkDeviceSize GetAllocationOffset(VmaAllocHandle allocHandle) const override { return ((Block*)allocHandle)->offset; }; |
| 9956 | |
| 9957 | void Init(VkDeviceSize size) override; |
| 9958 | bool Validate() const override; |
| 9959 | |
| 9960 | void AddDetailedStatistics(VmaDetailedStatistics& inoutStats) const override; |
| 9961 | void AddStatistics(VmaStatistics& inoutStats) const override; |
| 9962 | |
| 9963 | #if VMA_STATS_STRING_ENABLED |
| 9964 | void PrintDetailedMap(class VmaJsonWriter& json) const override; |
| 9965 | #endif |
| 9966 | |
| 9967 | bool CreateAllocationRequest( |
| 9968 | VkDeviceSize allocSize, |
| 9969 | VkDeviceSize allocAlignment, |
| 9970 | bool upperAddress, |
| 9971 | VmaSuballocationType allocType, |
| 9972 | uint32_t strategy, |
| 9973 | VmaAllocationRequest* pAllocationRequest) override; |
| 9974 | |
| 9975 | VkResult CheckCorruption(const void* pBlockData) override; |
| 9976 | void Alloc( |
| 9977 | const VmaAllocationRequest& request, |
| 9978 | VmaSuballocationType type, |
| 9979 | void* userData) override; |
| 9980 | |
| 9981 | void Free(VmaAllocHandle allocHandle) override; |
| 9982 | void GetAllocationInfo(VmaAllocHandle allocHandle, VmaVirtualAllocationInfo& outInfo) override; |
| 9983 | void* GetAllocationUserData(VmaAllocHandle allocHandle) const override; |
| 9984 | VmaAllocHandle GetAllocationListBegin() const override; |
| 9985 | VmaAllocHandle GetNextAllocation(VmaAllocHandle prevAlloc) const override; |
| 9986 | VkDeviceSize GetNextFreeRegionSize(VmaAllocHandle alloc) const override; |
| 9987 | void Clear() override; |
| 9988 | void SetAllocationUserData(VmaAllocHandle allocHandle, void* userData) override; |
| 9989 | void DebugLogAllAllocations() const override; |
| 9990 | |
| 9991 | private: |
| 9992 | // According to original paper it should be preferable 4 or 5: |
| 9993 | // M. Masmano, I. Ripoll, A. Crespo, and J. Real "TLSF: a New Dynamic Memory Allocator for Real-Time Systems" |
| 9994 | // http://www.gii.upv.es/tlsf/files/ecrts04_tlsf.pdf |
| 9995 | static const uint8_t SECOND_LEVEL_INDEX = 5; |
| 9996 | static const uint16_t SMALL_BUFFER_SIZE = 256; |
| 9997 | static const uint32_t INITIAL_BLOCK_ALLOC_COUNT = 16; |
| 9998 | static const uint8_t MEMORY_CLASS_SHIFT = 7; |
| 9999 | static const uint8_t MAX_MEMORY_CLASSES = 65 - MEMORY_CLASS_SHIFT; |
| 10000 | |
| 10001 | class Block |
| 10002 | { |
| 10003 | public: |
| 10004 | VkDeviceSize offset; |
| 10005 | VkDeviceSize size; |
| 10006 | Block* prevPhysical; |
| 10007 | Block* nextPhysical; |
| 10008 | |
| 10009 | void MarkFree() { prevFree = VMA_NULL; } |
| 10010 | void MarkTaken() { prevFree = this; } |
| 10011 | bool IsFree() const { return prevFree != this; } |
| 10012 | void*& UserData() { VMA_HEAVY_ASSERT(!IsFree()); return userData; } |
| 10013 | Block*& PrevFree() { return prevFree; } |
| 10014 | Block*& NextFree() { VMA_HEAVY_ASSERT(IsFree()); return nextFree; } |
| 10015 | |
| 10016 | private: |
| 10017 | Block* prevFree; // Address of the same block here indicates that block is taken |
| 10018 | union |
| 10019 | { |
| 10020 | Block* nextFree; |
| 10021 | void* userData; |
| 10022 | }; |
| 10023 | }; |
| 10024 | |
| 10025 | size_t m_AllocCount; |
| 10026 | // Total number of free blocks besides null block |
| 10027 | size_t m_BlocksFreeCount; |
| 10028 | // Total size of free blocks excluding null block |
| 10029 | VkDeviceSize m_BlocksFreeSize; |
| 10030 | uint32_t m_IsFreeBitmap; |
| 10031 | uint8_t m_MemoryClasses; |
| 10032 | uint32_t m_InnerIsFreeBitmap[MAX_MEMORY_CLASSES]; |
| 10033 | uint32_t m_ListsCount; |
| 10034 | /* |
| 10035 | * 0: 0-3 lists for small buffers |
| 10036 | * 1+: 0-(2^SLI-1) lists for normal buffers |
| 10037 | */ |
| 10038 | Block** m_FreeList; |
| 10039 | VmaPoolAllocator<Block> m_BlockAllocator; |
| 10040 | Block* m_NullBlock; |
| 10041 | VmaBlockBufferImageGranularity m_GranularityHandler; |
| 10042 | |
| 10043 | uint8_t SizeToMemoryClass(VkDeviceSize size) const; |
| 10044 | uint16_t SizeToSecondIndex(VkDeviceSize size, uint8_t memoryClass) const; |
| 10045 | uint32_t GetListIndex(uint8_t memoryClass, uint16_t secondIndex) const; |
| 10046 | uint32_t GetListIndex(VkDeviceSize size) const; |
| 10047 | |
| 10048 | void RemoveFreeBlock(Block* block); |
| 10049 | void InsertFreeBlock(Block* block); |
| 10050 | void MergeBlock(Block* block, Block* prev); |
| 10051 | |
| 10052 | Block* FindFreeBlock(VkDeviceSize size, uint32_t& listIndex) const; |
| 10053 | bool CheckBlock( |
| 10054 | Block& block, |
| 10055 | uint32_t listIndex, |
| 10056 | VkDeviceSize allocSize, |
| 10057 | VkDeviceSize allocAlignment, |
| 10058 | VmaSuballocationType allocType, |
| 10059 | VmaAllocationRequest* pAllocationRequest); |
| 10060 | }; |
| 10061 | |
| 10062 | #ifndef _VMA_BLOCK_METADATA_TLSF_FUNCTIONS |
| 10063 | VmaBlockMetadata_TLSF::VmaBlockMetadata_TLSF(const VkAllocationCallbacks* pAllocationCallbacks, |
| 10064 | VkDeviceSize bufferImageGranularity, bool isVirtual) |
| 10065 | : VmaBlockMetadata(pAllocationCallbacks, bufferImageGranularity, isVirtual), |
| 10066 | m_AllocCount(0), |
| 10067 | m_BlocksFreeCount(0), |
| 10068 | m_BlocksFreeSize(0), |
| 10069 | m_IsFreeBitmap(0), |
| 10070 | m_MemoryClasses(0), |
| 10071 | m_ListsCount(0), |
| 10072 | m_FreeList(VMA_NULL), |
| 10073 | m_BlockAllocator(pAllocationCallbacks, INITIAL_BLOCK_ALLOC_COUNT), |
| 10074 | m_NullBlock(VMA_NULL), |
| 10075 | m_GranularityHandler(bufferImageGranularity) {} |
| 10076 | |
| 10077 | VmaBlockMetadata_TLSF::~VmaBlockMetadata_TLSF() |
| 10078 | { |
| 10079 | if (m_FreeList) |
| 10080 | vma_delete_array(pAllocationCallbacks: GetAllocationCallbacks(), ptr: m_FreeList, count: m_ListsCount); |
| 10081 | m_GranularityHandler.Destroy(pAllocationCallbacks: GetAllocationCallbacks()); |
| 10082 | } |
| 10083 | |
| 10084 | void VmaBlockMetadata_TLSF::Init(VkDeviceSize size) |
| 10085 | { |
| 10086 | VmaBlockMetadata::Init(size); |
| 10087 | |
| 10088 | if (!IsVirtual()) |
| 10089 | m_GranularityHandler.Init(pAllocationCallbacks: GetAllocationCallbacks(), size); |
| 10090 | |
| 10091 | m_NullBlock = m_BlockAllocator.Alloc(); |
| 10092 | m_NullBlock->size = size; |
| 10093 | m_NullBlock->offset = 0; |
| 10094 | m_NullBlock->prevPhysical = VMA_NULL; |
| 10095 | m_NullBlock->nextPhysical = VMA_NULL; |
| 10096 | m_NullBlock->MarkFree(); |
| 10097 | m_NullBlock->NextFree() = VMA_NULL; |
| 10098 | m_NullBlock->PrevFree() = VMA_NULL; |
| 10099 | uint8_t memoryClass = SizeToMemoryClass(size); |
| 10100 | uint16_t sli = SizeToSecondIndex(size, memoryClass); |
| 10101 | m_ListsCount = (memoryClass == 0 ? 0 : (memoryClass - 1) * (1UL << SECOND_LEVEL_INDEX) + sli) + 1; |
| 10102 | if (IsVirtual()) |
| 10103 | m_ListsCount += 1UL << SECOND_LEVEL_INDEX; |
| 10104 | else |
| 10105 | m_ListsCount += 4; |
| 10106 | |
| 10107 | m_MemoryClasses = memoryClass + 2; |
| 10108 | memset(s: m_InnerIsFreeBitmap, c: 0, n: MAX_MEMORY_CLASSES * sizeof(uint32_t)); |
| 10109 | |
| 10110 | m_FreeList = vma_new_array(GetAllocationCallbacks(), Block*, m_ListsCount); |
| 10111 | memset(s: m_FreeList, c: 0, n: m_ListsCount * sizeof(Block*)); |
| 10112 | } |
| 10113 | |
| 10114 | bool VmaBlockMetadata_TLSF::Validate() const |
| 10115 | { |
| 10116 | VMA_VALIDATE(GetSumFreeSize() <= GetSize()); |
| 10117 | |
| 10118 | VkDeviceSize calculatedSize = m_NullBlock->size; |
| 10119 | VkDeviceSize calculatedFreeSize = m_NullBlock->size; |
| 10120 | size_t allocCount = 0; |
| 10121 | size_t freeCount = 0; |
| 10122 | |
| 10123 | // Check integrity of free lists |
| 10124 | for (uint32_t list = 0; list < m_ListsCount; ++list) |
| 10125 | { |
| 10126 | Block* block = m_FreeList[list]; |
| 10127 | if (block != VMA_NULL) |
| 10128 | { |
| 10129 | VMA_VALIDATE(block->IsFree()); |
| 10130 | VMA_VALIDATE(block->PrevFree() == VMA_NULL); |
| 10131 | while (block->NextFree()) |
| 10132 | { |
| 10133 | VMA_VALIDATE(block->NextFree()->IsFree()); |
| 10134 | VMA_VALIDATE(block->NextFree()->PrevFree() == block); |
| 10135 | block = block->NextFree(); |
| 10136 | } |
| 10137 | } |
| 10138 | } |
| 10139 | |
| 10140 | VkDeviceSize nextOffset = m_NullBlock->offset; |
| 10141 | auto validateCtx = m_GranularityHandler.StartValidation(pAllocationCallbacks: GetAllocationCallbacks(), isVirutal: IsVirtual()); |
| 10142 | |
| 10143 | VMA_VALIDATE(m_NullBlock->nextPhysical == VMA_NULL); |
| 10144 | if (m_NullBlock->prevPhysical) |
| 10145 | { |
| 10146 | VMA_VALIDATE(m_NullBlock->prevPhysical->nextPhysical == m_NullBlock); |
| 10147 | } |
| 10148 | // Check all blocks |
| 10149 | for (Block* prev = m_NullBlock->prevPhysical; prev != VMA_NULL; prev = prev->prevPhysical) |
| 10150 | { |
| 10151 | VMA_VALIDATE(prev->offset + prev->size == nextOffset); |
| 10152 | nextOffset = prev->offset; |
| 10153 | calculatedSize += prev->size; |
| 10154 | |
| 10155 | uint32_t listIndex = GetListIndex(size: prev->size); |
| 10156 | if (prev->IsFree()) |
| 10157 | { |
| 10158 | ++freeCount; |
| 10159 | // Check if free block belongs to free list |
| 10160 | Block* freeBlock = m_FreeList[listIndex]; |
| 10161 | VMA_VALIDATE(freeBlock != VMA_NULL); |
| 10162 | |
| 10163 | bool found = false; |
| 10164 | do |
| 10165 | { |
| 10166 | if (freeBlock == prev) |
| 10167 | found = true; |
| 10168 | |
| 10169 | freeBlock = freeBlock->NextFree(); |
| 10170 | } while (!found && freeBlock != VMA_NULL); |
| 10171 | |
| 10172 | VMA_VALIDATE(found); |
| 10173 | calculatedFreeSize += prev->size; |
| 10174 | } |
| 10175 | else |
| 10176 | { |
| 10177 | ++allocCount; |
| 10178 | // Check if taken block is not on a free list |
| 10179 | Block* freeBlock = m_FreeList[listIndex]; |
| 10180 | while (freeBlock) |
| 10181 | { |
| 10182 | VMA_VALIDATE(freeBlock != prev); |
| 10183 | freeBlock = freeBlock->NextFree(); |
| 10184 | } |
| 10185 | |
| 10186 | if (!IsVirtual()) |
| 10187 | { |
| 10188 | VMA_VALIDATE(m_GranularityHandler.Validate(validateCtx, prev->offset, prev->size)); |
| 10189 | } |
| 10190 | } |
| 10191 | |
| 10192 | if (prev->prevPhysical) |
| 10193 | { |
| 10194 | VMA_VALIDATE(prev->prevPhysical->nextPhysical == prev); |
| 10195 | } |
| 10196 | } |
| 10197 | |
| 10198 | if (!IsVirtual()) |
| 10199 | { |
| 10200 | VMA_VALIDATE(m_GranularityHandler.FinishValidation(validateCtx)); |
| 10201 | } |
| 10202 | |
| 10203 | VMA_VALIDATE(nextOffset == 0); |
| 10204 | VMA_VALIDATE(calculatedSize == GetSize()); |
| 10205 | VMA_VALIDATE(calculatedFreeSize == GetSumFreeSize()); |
| 10206 | VMA_VALIDATE(allocCount == m_AllocCount); |
| 10207 | VMA_VALIDATE(freeCount == m_BlocksFreeCount); |
| 10208 | |
| 10209 | return true; |
| 10210 | } |
| 10211 | |
| 10212 | void VmaBlockMetadata_TLSF::AddDetailedStatistics(VmaDetailedStatistics& inoutStats) const |
| 10213 | { |
| 10214 | inoutStats.statistics.blockCount++; |
| 10215 | inoutStats.statistics.blockBytes += GetSize(); |
| 10216 | if (m_NullBlock->size > 0) |
| 10217 | VmaAddDetailedStatisticsUnusedRange(inoutStats, size: m_NullBlock->size); |
| 10218 | |
| 10219 | for (Block* block = m_NullBlock->prevPhysical; block != VMA_NULL; block = block->prevPhysical) |
| 10220 | { |
| 10221 | if (block->IsFree()) |
| 10222 | VmaAddDetailedStatisticsUnusedRange(inoutStats, size: block->size); |
| 10223 | else |
| 10224 | VmaAddDetailedStatisticsAllocation(inoutStats, size: block->size); |
| 10225 | } |
| 10226 | } |
| 10227 | |
| 10228 | void VmaBlockMetadata_TLSF::AddStatistics(VmaStatistics& inoutStats) const |
| 10229 | { |
| 10230 | inoutStats.blockCount++; |
| 10231 | inoutStats.allocationCount += (uint32_t)m_AllocCount; |
| 10232 | inoutStats.blockBytes += GetSize(); |
| 10233 | inoutStats.allocationBytes += GetSize() - GetSumFreeSize(); |
| 10234 | } |
| 10235 | |
| 10236 | #if VMA_STATS_STRING_ENABLED |
| 10237 | void VmaBlockMetadata_TLSF::PrintDetailedMap(class VmaJsonWriter& json) const |
| 10238 | { |
| 10239 | size_t blockCount = m_AllocCount + m_BlocksFreeCount; |
| 10240 | VmaStlAllocator<Block*> allocator(GetAllocationCallbacks()); |
| 10241 | VmaVector<Block*, VmaStlAllocator<Block*>> blockList(blockCount, allocator); |
| 10242 | |
| 10243 | size_t i = blockCount; |
| 10244 | for (Block* block = m_NullBlock->prevPhysical; block != VMA_NULL; block = block->prevPhysical) |
| 10245 | { |
| 10246 | blockList[--i] = block; |
| 10247 | } |
| 10248 | VMA_ASSERT(i == 0); |
| 10249 | |
| 10250 | VmaDetailedStatistics stats; |
| 10251 | VmaClearDetailedStatistics(outStats&: stats); |
| 10252 | AddDetailedStatistics(inoutStats&: stats); |
| 10253 | |
| 10254 | PrintDetailedMap_Begin(json, |
| 10255 | unusedBytes: stats.statistics.blockBytes - stats.statistics.allocationBytes, |
| 10256 | allocationCount: stats.statistics.allocationCount, |
| 10257 | unusedRangeCount: stats.unusedRangeCount); |
| 10258 | |
| 10259 | for (; i < blockCount; ++i) |
| 10260 | { |
| 10261 | Block* block = blockList[i]; |
| 10262 | if (block->IsFree()) |
| 10263 | PrintDetailedMap_UnusedRange(json, offset: block->offset, size: block->size); |
| 10264 | else |
| 10265 | PrintDetailedMap_Allocation(json, offset: block->offset, size: block->size, userData: block->UserData()); |
| 10266 | } |
| 10267 | if (m_NullBlock->size > 0) |
| 10268 | PrintDetailedMap_UnusedRange(json, offset: m_NullBlock->offset, size: m_NullBlock->size); |
| 10269 | |
| 10270 | PrintDetailedMap_End(json); |
| 10271 | } |
| 10272 | #endif |
| 10273 | |
| 10274 | bool VmaBlockMetadata_TLSF::CreateAllocationRequest( |
| 10275 | VkDeviceSize allocSize, |
| 10276 | VkDeviceSize allocAlignment, |
| 10277 | bool upperAddress, |
| 10278 | VmaSuballocationType allocType, |
| 10279 | uint32_t strategy, |
| 10280 | VmaAllocationRequest* pAllocationRequest) |
| 10281 | { |
| 10282 | VMA_ASSERT(allocSize > 0 && "Cannot allocate empty block!" ); |
| 10283 | VMA_ASSERT(!upperAddress && "VMA_ALLOCATION_CREATE_UPPER_ADDRESS_BIT can be used only with linear algorithm." ); |
| 10284 | |
| 10285 | // For small granularity round up |
| 10286 | if (!IsVirtual()) |
| 10287 | m_GranularityHandler.RoundupAllocRequest(allocType, inOutAllocSize&: allocSize, inOutAllocAlignment&: allocAlignment); |
| 10288 | |
| 10289 | allocSize += GetDebugMargin(); |
| 10290 | // Quick check for too small pool |
| 10291 | if (allocSize > GetSumFreeSize()) |
| 10292 | return false; |
| 10293 | |
| 10294 | // If no free blocks in pool then check only null block |
| 10295 | if (m_BlocksFreeCount == 0) |
| 10296 | return CheckBlock(block&: *m_NullBlock, listIndex: m_ListsCount, allocSize, allocAlignment, allocType, pAllocationRequest); |
| 10297 | |
| 10298 | // Round up to the next block |
| 10299 | VkDeviceSize sizeForNextList = allocSize; |
| 10300 | VkDeviceSize smallSizeStep = SMALL_BUFFER_SIZE / (IsVirtual() ? 1 << SECOND_LEVEL_INDEX : 4); |
| 10301 | if (allocSize > SMALL_BUFFER_SIZE) |
| 10302 | { |
| 10303 | sizeForNextList += (1ULL << (VMA_BITSCAN_MSB(allocSize) - SECOND_LEVEL_INDEX)); |
| 10304 | } |
| 10305 | else if (allocSize > SMALL_BUFFER_SIZE - smallSizeStep) |
| 10306 | sizeForNextList = SMALL_BUFFER_SIZE + 1; |
| 10307 | else |
| 10308 | sizeForNextList += smallSizeStep; |
| 10309 | |
| 10310 | uint32_t nextListIndex = 0; |
| 10311 | uint32_t prevListIndex = 0; |
| 10312 | Block* nextListBlock = VMA_NULL; |
| 10313 | Block* prevListBlock = VMA_NULL; |
| 10314 | |
| 10315 | // Check blocks according to strategies |
| 10316 | if (strategy & VMA_ALLOCATION_CREATE_STRATEGY_MIN_TIME_BIT) |
| 10317 | { |
| 10318 | // Quick check for larger block first |
| 10319 | nextListBlock = FindFreeBlock(size: sizeForNextList, listIndex&: nextListIndex); |
| 10320 | if (nextListBlock != VMA_NULL && CheckBlock(block&: *nextListBlock, listIndex: nextListIndex, allocSize, allocAlignment, allocType, pAllocationRequest)) |
| 10321 | return true; |
| 10322 | |
| 10323 | // If not fitted then null block |
| 10324 | if (CheckBlock(block&: *m_NullBlock, listIndex: m_ListsCount, allocSize, allocAlignment, allocType, pAllocationRequest)) |
| 10325 | return true; |
| 10326 | |
| 10327 | // Null block failed, search larger bucket |
| 10328 | while (nextListBlock) |
| 10329 | { |
| 10330 | if (CheckBlock(block&: *nextListBlock, listIndex: nextListIndex, allocSize, allocAlignment, allocType, pAllocationRequest)) |
| 10331 | return true; |
| 10332 | nextListBlock = nextListBlock->NextFree(); |
| 10333 | } |
| 10334 | |
| 10335 | // Failed again, check best fit bucket |
| 10336 | prevListBlock = FindFreeBlock(size: allocSize, listIndex&: prevListIndex); |
| 10337 | while (prevListBlock) |
| 10338 | { |
| 10339 | if (CheckBlock(block&: *prevListBlock, listIndex: prevListIndex, allocSize, allocAlignment, allocType, pAllocationRequest)) |
| 10340 | return true; |
| 10341 | prevListBlock = prevListBlock->NextFree(); |
| 10342 | } |
| 10343 | } |
| 10344 | else if (strategy & VMA_ALLOCATION_CREATE_STRATEGY_MIN_MEMORY_BIT) |
| 10345 | { |
| 10346 | // Check best fit bucket |
| 10347 | prevListBlock = FindFreeBlock(size: allocSize, listIndex&: prevListIndex); |
| 10348 | while (prevListBlock) |
| 10349 | { |
| 10350 | if (CheckBlock(block&: *prevListBlock, listIndex: prevListIndex, allocSize, allocAlignment, allocType, pAllocationRequest)) |
| 10351 | return true; |
| 10352 | prevListBlock = prevListBlock->NextFree(); |
| 10353 | } |
| 10354 | |
| 10355 | // If failed check null block |
| 10356 | if (CheckBlock(block&: *m_NullBlock, listIndex: m_ListsCount, allocSize, allocAlignment, allocType, pAllocationRequest)) |
| 10357 | return true; |
| 10358 | |
| 10359 | // Check larger bucket |
| 10360 | nextListBlock = FindFreeBlock(size: sizeForNextList, listIndex&: nextListIndex); |
| 10361 | while (nextListBlock) |
| 10362 | { |
| 10363 | if (CheckBlock(block&: *nextListBlock, listIndex: nextListIndex, allocSize, allocAlignment, allocType, pAllocationRequest)) |
| 10364 | return true; |
| 10365 | nextListBlock = nextListBlock->NextFree(); |
| 10366 | } |
| 10367 | } |
| 10368 | else if (strategy & VMA_ALLOCATION_CREATE_STRATEGY_MIN_OFFSET_BIT ) |
| 10369 | { |
| 10370 | // Perform search from the start |
| 10371 | VmaStlAllocator<Block*> allocator(GetAllocationCallbacks()); |
| 10372 | VmaVector<Block*, VmaStlAllocator<Block*>> blockList(m_BlocksFreeCount, allocator); |
| 10373 | |
| 10374 | size_t i = m_BlocksFreeCount; |
| 10375 | for (Block* block = m_NullBlock->prevPhysical; block != VMA_NULL; block = block->prevPhysical) |
| 10376 | { |
| 10377 | if (block->IsFree() && block->size >= allocSize) |
| 10378 | blockList[--i] = block; |
| 10379 | } |
| 10380 | |
| 10381 | for (; i < m_BlocksFreeCount; ++i) |
| 10382 | { |
| 10383 | Block& block = *blockList[i]; |
| 10384 | if (CheckBlock(block, listIndex: GetListIndex(size: block.size), allocSize, allocAlignment, allocType, pAllocationRequest)) |
| 10385 | return true; |
| 10386 | } |
| 10387 | |
| 10388 | // If failed check null block |
| 10389 | if (CheckBlock(block&: *m_NullBlock, listIndex: m_ListsCount, allocSize, allocAlignment, allocType, pAllocationRequest)) |
| 10390 | return true; |
| 10391 | |
| 10392 | // Whole range searched, no more memory |
| 10393 | return false; |
| 10394 | } |
| 10395 | else |
| 10396 | { |
| 10397 | // Check larger bucket |
| 10398 | nextListBlock = FindFreeBlock(size: sizeForNextList, listIndex&: nextListIndex); |
| 10399 | while (nextListBlock) |
| 10400 | { |
| 10401 | if (CheckBlock(block&: *nextListBlock, listIndex: nextListIndex, allocSize, allocAlignment, allocType, pAllocationRequest)) |
| 10402 | return true; |
| 10403 | nextListBlock = nextListBlock->NextFree(); |
| 10404 | } |
| 10405 | |
| 10406 | // If failed check null block |
| 10407 | if (CheckBlock(block&: *m_NullBlock, listIndex: m_ListsCount, allocSize, allocAlignment, allocType, pAllocationRequest)) |
| 10408 | return true; |
| 10409 | |
| 10410 | // Check best fit bucket |
| 10411 | prevListBlock = FindFreeBlock(size: allocSize, listIndex&: prevListIndex); |
| 10412 | while (prevListBlock) |
| 10413 | { |
| 10414 | if (CheckBlock(block&: *prevListBlock, listIndex: prevListIndex, allocSize, allocAlignment, allocType, pAllocationRequest)) |
| 10415 | return true; |
| 10416 | prevListBlock = prevListBlock->NextFree(); |
| 10417 | } |
| 10418 | } |
| 10419 | |
| 10420 | // Worst case, full search has to be done |
| 10421 | while (++nextListIndex < m_ListsCount) |
| 10422 | { |
| 10423 | nextListBlock = m_FreeList[nextListIndex]; |
| 10424 | while (nextListBlock) |
| 10425 | { |
| 10426 | if (CheckBlock(block&: *nextListBlock, listIndex: nextListIndex, allocSize, allocAlignment, allocType, pAllocationRequest)) |
| 10427 | return true; |
| 10428 | nextListBlock = nextListBlock->NextFree(); |
| 10429 | } |
| 10430 | } |
| 10431 | |
| 10432 | // No more memory sadly |
| 10433 | return false; |
| 10434 | } |
| 10435 | |
| 10436 | VkResult VmaBlockMetadata_TLSF::CheckCorruption(const void* pBlockData) |
| 10437 | { |
| 10438 | for (Block* block = m_NullBlock->prevPhysical; block != VMA_NULL; block = block->prevPhysical) |
| 10439 | { |
| 10440 | if (!block->IsFree()) |
| 10441 | { |
| 10442 | if (!VmaValidateMagicValue(pData: pBlockData, offset: block->offset + block->size)) |
| 10443 | { |
| 10444 | VMA_ASSERT(0 && "MEMORY CORRUPTION DETECTED AFTER VALIDATED ALLOCATION!" ); |
| 10445 | return VK_ERROR_UNKNOWN_COPY; |
| 10446 | } |
| 10447 | } |
| 10448 | } |
| 10449 | |
| 10450 | return VK_SUCCESS; |
| 10451 | } |
| 10452 | |
| 10453 | void VmaBlockMetadata_TLSF::Alloc( |
| 10454 | const VmaAllocationRequest& request, |
| 10455 | VmaSuballocationType type, |
| 10456 | void* userData) |
| 10457 | { |
| 10458 | VMA_ASSERT(request.type == VmaAllocationRequestType::TLSF); |
| 10459 | |
| 10460 | // Get block and pop it from the free list |
| 10461 | Block* currentBlock = (Block*)request.allocHandle; |
| 10462 | VkDeviceSize offset = request.algorithmData; |
| 10463 | VMA_ASSERT(currentBlock != VMA_NULL); |
| 10464 | VMA_ASSERT(currentBlock->offset <= offset); |
| 10465 | |
| 10466 | if (currentBlock != m_NullBlock) |
| 10467 | RemoveFreeBlock(block: currentBlock); |
| 10468 | |
| 10469 | VkDeviceSize debugMargin = GetDebugMargin(); |
| 10470 | VkDeviceSize misssingAlignment = offset - currentBlock->offset; |
| 10471 | |
| 10472 | // Append missing alignment to prev block or create new one |
| 10473 | if (misssingAlignment) |
| 10474 | { |
| 10475 | Block* prevBlock = currentBlock->prevPhysical; |
| 10476 | VMA_ASSERT(prevBlock != VMA_NULL && "There should be no missing alignment at offset 0!" ); |
| 10477 | |
| 10478 | if (prevBlock->IsFree() && prevBlock->size != debugMargin) |
| 10479 | { |
| 10480 | uint32_t oldList = GetListIndex(size: prevBlock->size); |
| 10481 | prevBlock->size += misssingAlignment; |
| 10482 | // Check if new size crosses list bucket |
| 10483 | if (oldList != GetListIndex(size: prevBlock->size)) |
| 10484 | { |
| 10485 | prevBlock->size -= misssingAlignment; |
| 10486 | RemoveFreeBlock(block: prevBlock); |
| 10487 | prevBlock->size += misssingAlignment; |
| 10488 | InsertFreeBlock(block: prevBlock); |
| 10489 | } |
| 10490 | else |
| 10491 | m_BlocksFreeSize += misssingAlignment; |
| 10492 | } |
| 10493 | else |
| 10494 | { |
| 10495 | Block* newBlock = m_BlockAllocator.Alloc(); |
| 10496 | currentBlock->prevPhysical = newBlock; |
| 10497 | prevBlock->nextPhysical = newBlock; |
| 10498 | newBlock->prevPhysical = prevBlock; |
| 10499 | newBlock->nextPhysical = currentBlock; |
| 10500 | newBlock->size = misssingAlignment; |
| 10501 | newBlock->offset = currentBlock->offset; |
| 10502 | newBlock->MarkTaken(); |
| 10503 | |
| 10504 | InsertFreeBlock(block: newBlock); |
| 10505 | } |
| 10506 | |
| 10507 | currentBlock->size -= misssingAlignment; |
| 10508 | currentBlock->offset += misssingAlignment; |
| 10509 | } |
| 10510 | |
| 10511 | VkDeviceSize size = request.size + debugMargin; |
| 10512 | if (currentBlock->size == size) |
| 10513 | { |
| 10514 | if (currentBlock == m_NullBlock) |
| 10515 | { |
| 10516 | // Setup new null block |
| 10517 | m_NullBlock = m_BlockAllocator.Alloc(); |
| 10518 | m_NullBlock->size = 0; |
| 10519 | m_NullBlock->offset = currentBlock->offset + size; |
| 10520 | m_NullBlock->prevPhysical = currentBlock; |
| 10521 | m_NullBlock->nextPhysical = VMA_NULL; |
| 10522 | m_NullBlock->MarkFree(); |
| 10523 | m_NullBlock->PrevFree() = VMA_NULL; |
| 10524 | m_NullBlock->NextFree() = VMA_NULL; |
| 10525 | currentBlock->nextPhysical = m_NullBlock; |
| 10526 | currentBlock->MarkTaken(); |
| 10527 | } |
| 10528 | } |
| 10529 | else |
| 10530 | { |
| 10531 | VMA_ASSERT(currentBlock->size > size && "Proper block already found, shouldn't find smaller one!" ); |
| 10532 | |
| 10533 | // Create new free block |
| 10534 | Block* newBlock = m_BlockAllocator.Alloc(); |
| 10535 | newBlock->size = currentBlock->size - size; |
| 10536 | newBlock->offset = currentBlock->offset + size; |
| 10537 | newBlock->prevPhysical = currentBlock; |
| 10538 | newBlock->nextPhysical = currentBlock->nextPhysical; |
| 10539 | currentBlock->nextPhysical = newBlock; |
| 10540 | currentBlock->size = size; |
| 10541 | |
| 10542 | if (currentBlock == m_NullBlock) |
| 10543 | { |
| 10544 | m_NullBlock = newBlock; |
| 10545 | m_NullBlock->MarkFree(); |
| 10546 | m_NullBlock->NextFree() = VMA_NULL; |
| 10547 | m_NullBlock->PrevFree() = VMA_NULL; |
| 10548 | currentBlock->MarkTaken(); |
| 10549 | } |
| 10550 | else |
| 10551 | { |
| 10552 | newBlock->nextPhysical->prevPhysical = newBlock; |
| 10553 | newBlock->MarkTaken(); |
| 10554 | InsertFreeBlock(block: newBlock); |
| 10555 | } |
| 10556 | } |
| 10557 | currentBlock->UserData() = userData; |
| 10558 | |
| 10559 | if (debugMargin > 0) |
| 10560 | { |
| 10561 | currentBlock->size -= debugMargin; |
| 10562 | Block* newBlock = m_BlockAllocator.Alloc(); |
| 10563 | newBlock->size = debugMargin; |
| 10564 | newBlock->offset = currentBlock->offset + currentBlock->size; |
| 10565 | newBlock->prevPhysical = currentBlock; |
| 10566 | newBlock->nextPhysical = currentBlock->nextPhysical; |
| 10567 | newBlock->MarkTaken(); |
| 10568 | currentBlock->nextPhysical->prevPhysical = newBlock; |
| 10569 | currentBlock->nextPhysical = newBlock; |
| 10570 | InsertFreeBlock(block: newBlock); |
| 10571 | } |
| 10572 | |
| 10573 | if (!IsVirtual()) |
| 10574 | m_GranularityHandler.AllocPages(allocType: (uint8_t)(uintptr_t)request.customData, |
| 10575 | offset: currentBlock->offset, size: currentBlock->size); |
| 10576 | ++m_AllocCount; |
| 10577 | } |
| 10578 | |
| 10579 | void VmaBlockMetadata_TLSF::Free(VmaAllocHandle allocHandle) |
| 10580 | { |
| 10581 | Block* block = (Block*)allocHandle; |
| 10582 | Block* next = block->nextPhysical; |
| 10583 | VMA_ASSERT(!block->IsFree() && "Block is already free!" ); |
| 10584 | |
| 10585 | if (!IsVirtual()) |
| 10586 | m_GranularityHandler.FreePages(offset: block->offset, size: block->size); |
| 10587 | --m_AllocCount; |
| 10588 | |
| 10589 | VkDeviceSize debugMargin = GetDebugMargin(); |
| 10590 | if (debugMargin > 0) |
| 10591 | { |
| 10592 | RemoveFreeBlock(block: next); |
| 10593 | MergeBlock(block: next, prev: block); |
| 10594 | block = next; |
| 10595 | next = next->nextPhysical; |
| 10596 | } |
| 10597 | |
| 10598 | // Try merging |
| 10599 | Block* prev = block->prevPhysical; |
| 10600 | if (prev != VMA_NULL && prev->IsFree() && prev->size != debugMargin) |
| 10601 | { |
| 10602 | RemoveFreeBlock(block: prev); |
| 10603 | MergeBlock(block, prev); |
| 10604 | } |
| 10605 | |
| 10606 | if (!next->IsFree()) |
| 10607 | InsertFreeBlock(block); |
| 10608 | else if (next == m_NullBlock) |
| 10609 | MergeBlock(block: m_NullBlock, prev: block); |
| 10610 | else |
| 10611 | { |
| 10612 | RemoveFreeBlock(block: next); |
| 10613 | MergeBlock(block: next, prev: block); |
| 10614 | InsertFreeBlock(block: next); |
| 10615 | } |
| 10616 | } |
| 10617 | |
| 10618 | void VmaBlockMetadata_TLSF::GetAllocationInfo(VmaAllocHandle allocHandle, VmaVirtualAllocationInfo& outInfo) |
| 10619 | { |
| 10620 | Block* block = (Block*)allocHandle; |
| 10621 | VMA_ASSERT(!block->IsFree() && "Cannot get allocation info for free block!" ); |
| 10622 | outInfo.offset = block->offset; |
| 10623 | outInfo.size = block->size; |
| 10624 | outInfo.pUserData = block->UserData(); |
| 10625 | } |
| 10626 | |
| 10627 | void* VmaBlockMetadata_TLSF::GetAllocationUserData(VmaAllocHandle allocHandle) const |
| 10628 | { |
| 10629 | Block* block = (Block*)allocHandle; |
| 10630 | VMA_ASSERT(!block->IsFree() && "Cannot get user data for free block!" ); |
| 10631 | return block->UserData(); |
| 10632 | } |
| 10633 | |
| 10634 | VmaAllocHandle VmaBlockMetadata_TLSF::GetAllocationListBegin() const |
| 10635 | { |
| 10636 | if (m_AllocCount == 0) |
| 10637 | return VK_NULL_HANDLE; |
| 10638 | |
| 10639 | for (Block* block = m_NullBlock->prevPhysical; block; block = block->prevPhysical) |
| 10640 | { |
| 10641 | if (!block->IsFree()) |
| 10642 | return (VmaAllocHandle)block; |
| 10643 | } |
| 10644 | VMA_ASSERT(false && "If m_AllocCount > 0 then should find any allocation!" ); |
| 10645 | return VK_NULL_HANDLE; |
| 10646 | } |
| 10647 | |
| 10648 | VmaAllocHandle VmaBlockMetadata_TLSF::GetNextAllocation(VmaAllocHandle prevAlloc) const |
| 10649 | { |
| 10650 | Block* startBlock = (Block*)prevAlloc; |
| 10651 | VMA_ASSERT(!startBlock->IsFree() && "Incorrect block!" ); |
| 10652 | |
| 10653 | for (Block* block = startBlock->prevPhysical; block; block = block->prevPhysical) |
| 10654 | { |
| 10655 | if (!block->IsFree()) |
| 10656 | return (VmaAllocHandle)block; |
| 10657 | } |
| 10658 | return VK_NULL_HANDLE; |
| 10659 | } |
| 10660 | |
| 10661 | VkDeviceSize VmaBlockMetadata_TLSF::GetNextFreeRegionSize(VmaAllocHandle alloc) const |
| 10662 | { |
| 10663 | Block* block = (Block*)alloc; |
| 10664 | VMA_ASSERT(!block->IsFree() && "Incorrect block!" ); |
| 10665 | |
| 10666 | if (block->prevPhysical) |
| 10667 | return block->prevPhysical->IsFree() ? block->prevPhysical->size : 0; |
| 10668 | return 0; |
| 10669 | } |
| 10670 | |
| 10671 | void VmaBlockMetadata_TLSF::Clear() |
| 10672 | { |
| 10673 | m_AllocCount = 0; |
| 10674 | m_BlocksFreeCount = 0; |
| 10675 | m_BlocksFreeSize = 0; |
| 10676 | m_IsFreeBitmap = 0; |
| 10677 | m_NullBlock->offset = 0; |
| 10678 | m_NullBlock->size = GetSize(); |
| 10679 | Block* block = m_NullBlock->prevPhysical; |
| 10680 | m_NullBlock->prevPhysical = VMA_NULL; |
| 10681 | while (block) |
| 10682 | { |
| 10683 | Block* prev = block->prevPhysical; |
| 10684 | m_BlockAllocator.Free(ptr: block); |
| 10685 | block = prev; |
| 10686 | } |
| 10687 | memset(s: m_FreeList, c: 0, n: m_ListsCount * sizeof(Block*)); |
| 10688 | memset(s: m_InnerIsFreeBitmap, c: 0, n: m_MemoryClasses * sizeof(uint32_t)); |
| 10689 | m_GranularityHandler.Clear(); |
| 10690 | } |
| 10691 | |
| 10692 | void VmaBlockMetadata_TLSF::SetAllocationUserData(VmaAllocHandle allocHandle, void* userData) |
| 10693 | { |
| 10694 | Block* block = (Block*)allocHandle; |
| 10695 | VMA_ASSERT(!block->IsFree() && "Trying to set user data for not allocated block!" ); |
| 10696 | block->UserData() = userData; |
| 10697 | } |
| 10698 | |
| 10699 | void VmaBlockMetadata_TLSF::DebugLogAllAllocations() const |
| 10700 | { |
| 10701 | for (Block* block = m_NullBlock->prevPhysical; block != VMA_NULL; block = block->prevPhysical) |
| 10702 | if (!block->IsFree()) |
| 10703 | DebugLogAllocation(offset: block->offset, size: block->size, userData: block->UserData()); |
| 10704 | } |
| 10705 | |
| 10706 | uint8_t VmaBlockMetadata_TLSF::SizeToMemoryClass(VkDeviceSize size) const |
| 10707 | { |
| 10708 | if (size > SMALL_BUFFER_SIZE) |
| 10709 | return VMA_BITSCAN_MSB(size) - MEMORY_CLASS_SHIFT; |
| 10710 | return 0; |
| 10711 | } |
| 10712 | |
| 10713 | uint16_t VmaBlockMetadata_TLSF::SizeToSecondIndex(VkDeviceSize size, uint8_t memoryClass) const |
| 10714 | { |
| 10715 | if (memoryClass == 0) |
| 10716 | { |
| 10717 | if (IsVirtual()) |
| 10718 | return static_cast<uint16_t>((size - 1) / 8); |
| 10719 | else |
| 10720 | return static_cast<uint16_t>((size - 1) / 64); |
| 10721 | } |
| 10722 | return static_cast<uint16_t>((size >> (memoryClass + MEMORY_CLASS_SHIFT - SECOND_LEVEL_INDEX)) ^ (1U << SECOND_LEVEL_INDEX)); |
| 10723 | } |
| 10724 | |
| 10725 | uint32_t VmaBlockMetadata_TLSF::GetListIndex(uint8_t memoryClass, uint16_t secondIndex) const |
| 10726 | { |
| 10727 | if (memoryClass == 0) |
| 10728 | return secondIndex; |
| 10729 | |
| 10730 | const uint32_t index = static_cast<uint32_t>(memoryClass - 1) * (1 << SECOND_LEVEL_INDEX) + secondIndex; |
| 10731 | if (IsVirtual()) |
| 10732 | return index + (1 << SECOND_LEVEL_INDEX); |
| 10733 | else |
| 10734 | return index + 4; |
| 10735 | } |
| 10736 | |
| 10737 | uint32_t VmaBlockMetadata_TLSF::GetListIndex(VkDeviceSize size) const |
| 10738 | { |
| 10739 | uint8_t memoryClass = SizeToMemoryClass(size); |
| 10740 | return GetListIndex(memoryClass, secondIndex: SizeToSecondIndex(size, memoryClass)); |
| 10741 | } |
| 10742 | |
| 10743 | void VmaBlockMetadata_TLSF::RemoveFreeBlock(Block* block) |
| 10744 | { |
| 10745 | VMA_ASSERT(block != m_NullBlock); |
| 10746 | VMA_ASSERT(block->IsFree()); |
| 10747 | |
| 10748 | if (block->NextFree() != VMA_NULL) |
| 10749 | block->NextFree()->PrevFree() = block->PrevFree(); |
| 10750 | if (block->PrevFree() != VMA_NULL) |
| 10751 | block->PrevFree()->NextFree() = block->NextFree(); |
| 10752 | else |
| 10753 | { |
| 10754 | uint8_t memClass = SizeToMemoryClass(size: block->size); |
| 10755 | uint16_t secondIndex = SizeToSecondIndex(size: block->size, memoryClass: memClass); |
| 10756 | uint32_t index = GetListIndex(memoryClass: memClass, secondIndex); |
| 10757 | VMA_ASSERT(m_FreeList[index] == block); |
| 10758 | m_FreeList[index] = block->NextFree(); |
| 10759 | if (block->NextFree() == VMA_NULL) |
| 10760 | { |
| 10761 | m_InnerIsFreeBitmap[memClass] &= ~(1U << secondIndex); |
| 10762 | if (m_InnerIsFreeBitmap[memClass] == 0) |
| 10763 | m_IsFreeBitmap &= ~(1UL << memClass); |
| 10764 | } |
| 10765 | } |
| 10766 | block->MarkTaken(); |
| 10767 | block->UserData() = VMA_NULL; |
| 10768 | --m_BlocksFreeCount; |
| 10769 | m_BlocksFreeSize -= block->size; |
| 10770 | } |
| 10771 | |
| 10772 | void VmaBlockMetadata_TLSF::InsertFreeBlock(Block* block) |
| 10773 | { |
| 10774 | VMA_ASSERT(block != m_NullBlock); |
| 10775 | VMA_ASSERT(!block->IsFree() && "Cannot insert block twice!" ); |
| 10776 | |
| 10777 | uint8_t memClass = SizeToMemoryClass(size: block->size); |
| 10778 | uint16_t secondIndex = SizeToSecondIndex(size: block->size, memoryClass: memClass); |
| 10779 | uint32_t index = GetListIndex(memoryClass: memClass, secondIndex); |
| 10780 | VMA_ASSERT(index < m_ListsCount); |
| 10781 | block->PrevFree() = VMA_NULL; |
| 10782 | block->NextFree() = m_FreeList[index]; |
| 10783 | m_FreeList[index] = block; |
| 10784 | if (block->NextFree() != VMA_NULL) |
| 10785 | block->NextFree()->PrevFree() = block; |
| 10786 | else |
| 10787 | { |
| 10788 | m_InnerIsFreeBitmap[memClass] |= 1U << secondIndex; |
| 10789 | m_IsFreeBitmap |= 1UL << memClass; |
| 10790 | } |
| 10791 | ++m_BlocksFreeCount; |
| 10792 | m_BlocksFreeSize += block->size; |
| 10793 | } |
| 10794 | |
| 10795 | void VmaBlockMetadata_TLSF::MergeBlock(Block* block, Block* prev) |
| 10796 | { |
| 10797 | VMA_ASSERT(block->prevPhysical == prev && "Cannot merge seperate physical regions!" ); |
| 10798 | VMA_ASSERT(!prev->IsFree() && "Cannot merge block that belongs to free list!" ); |
| 10799 | |
| 10800 | block->offset = prev->offset; |
| 10801 | block->size += prev->size; |
| 10802 | block->prevPhysical = prev->prevPhysical; |
| 10803 | if (block->prevPhysical) |
| 10804 | block->prevPhysical->nextPhysical = block; |
| 10805 | m_BlockAllocator.Free(ptr: prev); |
| 10806 | } |
| 10807 | |
| 10808 | VmaBlockMetadata_TLSF::Block* VmaBlockMetadata_TLSF::FindFreeBlock(VkDeviceSize size, uint32_t& listIndex) const |
| 10809 | { |
| 10810 | uint8_t memoryClass = SizeToMemoryClass(size); |
| 10811 | uint32_t innerFreeMap = m_InnerIsFreeBitmap[memoryClass] & (~0U << SizeToSecondIndex(size, memoryClass)); |
| 10812 | if (!innerFreeMap) |
| 10813 | { |
| 10814 | // Check higher levels for avaiable blocks |
| 10815 | uint32_t freeMap = m_IsFreeBitmap & (~0UL << (memoryClass + 1)); |
| 10816 | if (!freeMap) |
| 10817 | return VMA_NULL; // No more memory avaible |
| 10818 | |
| 10819 | // Find lowest free region |
| 10820 | memoryClass = VMA_BITSCAN_LSB(freeMap); |
| 10821 | innerFreeMap = m_InnerIsFreeBitmap[memoryClass]; |
| 10822 | VMA_ASSERT(innerFreeMap != 0); |
| 10823 | } |
| 10824 | // Find lowest free subregion |
| 10825 | listIndex = GetListIndex(memoryClass, VMA_BITSCAN_LSB(innerFreeMap)); |
| 10826 | VMA_ASSERT(m_FreeList[listIndex]); |
| 10827 | return m_FreeList[listIndex]; |
| 10828 | } |
| 10829 | |
| 10830 | bool VmaBlockMetadata_TLSF::CheckBlock( |
| 10831 | Block& block, |
| 10832 | uint32_t listIndex, |
| 10833 | VkDeviceSize allocSize, |
| 10834 | VkDeviceSize allocAlignment, |
| 10835 | VmaSuballocationType allocType, |
| 10836 | VmaAllocationRequest* pAllocationRequest) |
| 10837 | { |
| 10838 | VMA_ASSERT(block.IsFree() && "Block is already taken!" ); |
| 10839 | |
| 10840 | VkDeviceSize alignedOffset = VmaAlignUp(val: block.offset, alignment: allocAlignment); |
| 10841 | if (block.size < allocSize + alignedOffset - block.offset) |
| 10842 | return false; |
| 10843 | |
| 10844 | // Check for granularity conflicts |
| 10845 | if (!IsVirtual() && |
| 10846 | m_GranularityHandler.CheckConflictAndAlignUp(inOutAllocOffset&: alignedOffset, allocSize, blockOffset: block.offset, blockSize: block.size, allocType)) |
| 10847 | return false; |
| 10848 | |
| 10849 | // Alloc successful |
| 10850 | pAllocationRequest->type = VmaAllocationRequestType::TLSF; |
| 10851 | pAllocationRequest->allocHandle = (VmaAllocHandle)█ |
| 10852 | pAllocationRequest->size = allocSize - GetDebugMargin(); |
| 10853 | pAllocationRequest->customData = (void*)allocType; |
| 10854 | pAllocationRequest->algorithmData = alignedOffset; |
| 10855 | |
| 10856 | // Place block at the start of list if it's normal block |
| 10857 | if (listIndex != m_ListsCount && block.PrevFree()) |
| 10858 | { |
| 10859 | block.PrevFree()->NextFree() = block.NextFree(); |
| 10860 | if (block.NextFree()) |
| 10861 | block.NextFree()->PrevFree() = block.PrevFree(); |
| 10862 | block.PrevFree() = VMA_NULL; |
| 10863 | block.NextFree() = m_FreeList[listIndex]; |
| 10864 | m_FreeList[listIndex] = █ |
| 10865 | if (block.NextFree()) |
| 10866 | block.NextFree()->PrevFree() = █ |
| 10867 | } |
| 10868 | |
| 10869 | return true; |
| 10870 | } |
| 10871 | #endif // _VMA_BLOCK_METADATA_TLSF_FUNCTIONS |
| 10872 | #endif // _VMA_BLOCK_METADATA_TLSF |
| 10873 | |
| 10874 | #ifndef _VMA_BLOCK_VECTOR |
| 10875 | /* |
| 10876 | Sequence of VmaDeviceMemoryBlock. Represents memory blocks allocated for a specific |
| 10877 | Vulkan memory type. |
| 10878 | |
| 10879 | Synchronized internally with a mutex. |
| 10880 | */ |
| 10881 | class VmaBlockVector |
| 10882 | { |
| 10883 | friend struct VmaDefragmentationContext_T; |
| 10884 | VMA_CLASS_NO_COPY(VmaBlockVector) |
| 10885 | public: |
| 10886 | VmaBlockVector( |
| 10887 | VmaAllocator hAllocator, |
| 10888 | VmaPool hParentPool, |
| 10889 | uint32_t memoryTypeIndex, |
| 10890 | VkDeviceSize preferredBlockSize, |
| 10891 | size_t minBlockCount, |
| 10892 | size_t maxBlockCount, |
| 10893 | VkDeviceSize bufferImageGranularity, |
| 10894 | bool explicitBlockSize, |
| 10895 | uint32_t algorithm, |
| 10896 | float priority, |
| 10897 | VkDeviceSize minAllocationAlignment, |
| 10898 | void* pMemoryAllocateNext); |
| 10899 | ~VmaBlockVector(); |
| 10900 | |
| 10901 | VmaAllocator GetAllocator() const { return m_hAllocator; } |
| 10902 | VmaPool GetParentPool() const { return m_hParentPool; } |
| 10903 | bool IsCustomPool() const { return m_hParentPool != VMA_NULL; } |
| 10904 | uint32_t GetMemoryTypeIndex() const { return m_MemoryTypeIndex; } |
| 10905 | VkDeviceSize GetPreferredBlockSize() const { return m_PreferredBlockSize; } |
| 10906 | VkDeviceSize GetBufferImageGranularity() const { return m_BufferImageGranularity; } |
| 10907 | uint32_t GetAlgorithm() const { return m_Algorithm; } |
| 10908 | bool HasExplicitBlockSize() const { return m_ExplicitBlockSize; } |
| 10909 | float GetPriority() const { return m_Priority; } |
| 10910 | const void* GetAllocationNextPtr() const { return m_pMemoryAllocateNext; } |
| 10911 | // To be used only while the m_Mutex is locked. Used during defragmentation. |
| 10912 | size_t GetBlockCount() const { return m_Blocks.size(); } |
| 10913 | // To be used only while the m_Mutex is locked. Used during defragmentation. |
| 10914 | VmaDeviceMemoryBlock* GetBlock(size_t index) const { return m_Blocks[index]; } |
| 10915 | VMA_RW_MUTEX &GetMutex() { return m_Mutex; } |
| 10916 | |
| 10917 | VkResult CreateMinBlocks(); |
| 10918 | void AddStatistics(VmaStatistics& inoutStats); |
| 10919 | void AddDetailedStatistics(VmaDetailedStatistics& inoutStats); |
| 10920 | bool IsEmpty(); |
| 10921 | bool IsCorruptionDetectionEnabled() const; |
| 10922 | |
| 10923 | VkResult Allocate( |
| 10924 | VkDeviceSize size, |
| 10925 | VkDeviceSize alignment, |
| 10926 | const VmaAllocationCreateInfo& createInfo, |
| 10927 | VmaSuballocationType suballocType, |
| 10928 | size_t allocationCount, |
| 10929 | VmaAllocation* pAllocations); |
| 10930 | |
| 10931 | void Free(const VmaAllocation hAllocation); |
| 10932 | |
| 10933 | #if VMA_STATS_STRING_ENABLED |
| 10934 | void PrintDetailedMap(class VmaJsonWriter& json); |
| 10935 | #endif |
| 10936 | |
| 10937 | VkResult CheckCorruption(); |
| 10938 | |
| 10939 | private: |
| 10940 | const VmaAllocator m_hAllocator; |
| 10941 | const VmaPool m_hParentPool; |
| 10942 | const uint32_t m_MemoryTypeIndex; |
| 10943 | const VkDeviceSize m_PreferredBlockSize; |
| 10944 | const size_t m_MinBlockCount; |
| 10945 | const size_t m_MaxBlockCount; |
| 10946 | const VkDeviceSize m_BufferImageGranularity; |
| 10947 | const bool m_ExplicitBlockSize; |
| 10948 | const uint32_t m_Algorithm; |
| 10949 | const float m_Priority; |
| 10950 | const VkDeviceSize m_MinAllocationAlignment; |
| 10951 | |
| 10952 | void* const m_pMemoryAllocateNext; |
| 10953 | VMA_RW_MUTEX m_Mutex; |
| 10954 | // Incrementally sorted by sumFreeSize, ascending. |
| 10955 | VmaVector<VmaDeviceMemoryBlock*, VmaStlAllocator<VmaDeviceMemoryBlock*>> m_Blocks; |
| 10956 | uint32_t m_NextBlockId; |
| 10957 | bool m_IncrementalSort = true; |
| 10958 | |
| 10959 | void SetIncrementalSort(bool val) { m_IncrementalSort = val; } |
| 10960 | |
| 10961 | VkDeviceSize CalcMaxBlockSize() const; |
| 10962 | // Finds and removes given block from vector. |
| 10963 | void Remove(VmaDeviceMemoryBlock* pBlock); |
| 10964 | // Performs single step in sorting m_Blocks. They may not be fully sorted |
| 10965 | // after this call. |
| 10966 | void IncrementallySortBlocks(); |
| 10967 | void SortByFreeSize(); |
| 10968 | |
| 10969 | VkResult AllocatePage( |
| 10970 | VkDeviceSize size, |
| 10971 | VkDeviceSize alignment, |
| 10972 | const VmaAllocationCreateInfo& createInfo, |
| 10973 | VmaSuballocationType suballocType, |
| 10974 | VmaAllocation* pAllocation); |
| 10975 | |
| 10976 | VkResult AllocateFromBlock( |
| 10977 | VmaDeviceMemoryBlock* pBlock, |
| 10978 | VkDeviceSize size, |
| 10979 | VkDeviceSize alignment, |
| 10980 | VmaAllocationCreateFlags allocFlags, |
| 10981 | void* pUserData, |
| 10982 | VmaSuballocationType suballocType, |
| 10983 | uint32_t strategy, |
| 10984 | VmaAllocation* pAllocation); |
| 10985 | |
| 10986 | VkResult CommitAllocationRequest( |
| 10987 | VmaAllocationRequest& allocRequest, |
| 10988 | VmaDeviceMemoryBlock* pBlock, |
| 10989 | VkDeviceSize alignment, |
| 10990 | VmaAllocationCreateFlags allocFlags, |
| 10991 | void* pUserData, |
| 10992 | VmaSuballocationType suballocType, |
| 10993 | VmaAllocation* pAllocation); |
| 10994 | |
| 10995 | VkResult CreateBlock(VkDeviceSize blockSize, size_t* pNewBlockIndex); |
| 10996 | bool HasEmptyBlock(); |
| 10997 | }; |
| 10998 | #endif // _VMA_BLOCK_VECTOR |
| 10999 | |
| 11000 | #ifndef _VMA_DEFRAGMENTATION_CONTEXT |
| 11001 | struct VmaDefragmentationContext_T |
| 11002 | { |
| 11003 | VMA_CLASS_NO_COPY(VmaDefragmentationContext_T) |
| 11004 | public: |
| 11005 | VmaDefragmentationContext_T( |
| 11006 | VmaAllocator hAllocator, |
| 11007 | const VmaDefragmentationInfo& info); |
| 11008 | ~VmaDefragmentationContext_T(); |
| 11009 | |
| 11010 | void GetStats(VmaDefragmentationStats& outStats) { outStats = m_GlobalStats; } |
| 11011 | |
| 11012 | VkResult DefragmentPassBegin(VmaDefragmentationPassMoveInfo& moveInfo); |
| 11013 | VkResult DefragmentPassEnd(VmaDefragmentationPassMoveInfo& moveInfo); |
| 11014 | |
| 11015 | private: |
| 11016 | // Max number of allocations to ignore due to size constraints before ending single pass |
| 11017 | static const uint8_t MAX_ALLOCS_TO_IGNORE = 16; |
| 11018 | enum class CounterStatus { Pass, Ignore, End }; |
| 11019 | |
| 11020 | struct FragmentedBlock |
| 11021 | { |
| 11022 | uint32_t data; |
| 11023 | VmaDeviceMemoryBlock* block; |
| 11024 | }; |
| 11025 | struct StateBalanced |
| 11026 | { |
| 11027 | VkDeviceSize avgFreeSize = 0; |
| 11028 | VkDeviceSize avgAllocSize = UINT64_MAX; |
| 11029 | }; |
| 11030 | struct StateExtensive |
| 11031 | { |
| 11032 | enum class Operation : uint8_t |
| 11033 | { |
| 11034 | FindFreeBlockBuffer, FindFreeBlockTexture, FindFreeBlockAll, |
| 11035 | MoveBuffers, MoveTextures, MoveAll, |
| 11036 | Cleanup, Done |
| 11037 | }; |
| 11038 | |
| 11039 | Operation operation = Operation::FindFreeBlockTexture; |
| 11040 | size_t firstFreeBlock = SIZE_MAX; |
| 11041 | }; |
| 11042 | struct MoveAllocationData |
| 11043 | { |
| 11044 | VkDeviceSize size; |
| 11045 | VkDeviceSize alignment; |
| 11046 | VmaSuballocationType type; |
| 11047 | VmaAllocationCreateFlags flags; |
| 11048 | VmaDefragmentationMove move = {}; |
| 11049 | }; |
| 11050 | |
| 11051 | const VkDeviceSize m_MaxPassBytes; |
| 11052 | const uint32_t m_MaxPassAllocations; |
| 11053 | |
| 11054 | VmaStlAllocator<VmaDefragmentationMove> m_MoveAllocator; |
| 11055 | VmaVector<VmaDefragmentationMove, VmaStlAllocator<VmaDefragmentationMove>> m_Moves; |
| 11056 | |
| 11057 | uint8_t m_IgnoredAllocs = 0; |
| 11058 | uint32_t m_Algorithm; |
| 11059 | uint32_t m_BlockVectorCount; |
| 11060 | VmaBlockVector* m_PoolBlockVector; |
| 11061 | VmaBlockVector** m_pBlockVectors; |
| 11062 | size_t m_ImmovableBlockCount = 0; |
| 11063 | VmaDefragmentationStats m_GlobalStats = { .bytesMoved: 0 }; |
| 11064 | VmaDefragmentationStats m_PassStats = { .bytesMoved: 0 }; |
| 11065 | void* m_AlgorithmState = VMA_NULL; |
| 11066 | |
| 11067 | static MoveAllocationData GetMoveData(VmaAllocHandle handle, VmaBlockMetadata* metadata); |
| 11068 | CounterStatus CheckCounters(VkDeviceSize bytes); |
| 11069 | bool IncrementCounters(VkDeviceSize bytes); |
| 11070 | bool ReallocWithinBlock(VmaBlockVector& vector, VmaDeviceMemoryBlock* block); |
| 11071 | bool AllocInOtherBlock(size_t start, size_t end, MoveAllocationData& data, VmaBlockVector& vector); |
| 11072 | |
| 11073 | bool ComputeDefragmentation(VmaBlockVector& vector, size_t index); |
| 11074 | bool ComputeDefragmentation_Fast(VmaBlockVector& vector); |
| 11075 | bool ComputeDefragmentation_Balanced(VmaBlockVector& vector, size_t index, bool update); |
| 11076 | bool ComputeDefragmentation_Full(VmaBlockVector& vector); |
| 11077 | bool ComputeDefragmentation_Extensive(VmaBlockVector& vector, size_t index); |
| 11078 | |
| 11079 | void UpdateVectorStatistics(VmaBlockVector& vector, StateBalanced& state); |
| 11080 | bool MoveDataToFreeBlocks(VmaSuballocationType currentType, |
| 11081 | VmaBlockVector& vector, size_t firstFreeBlock, |
| 11082 | bool& texturePresent, bool& bufferPresent, bool& otherPresent); |
| 11083 | }; |
| 11084 | #endif // _VMA_DEFRAGMENTATION_CONTEXT |
| 11085 | |
| 11086 | #ifndef _VMA_POOL_T |
| 11087 | struct VmaPool_T |
| 11088 | { |
| 11089 | friend struct VmaPoolListItemTraits; |
| 11090 | VMA_CLASS_NO_COPY(VmaPool_T) |
| 11091 | public: |
| 11092 | VmaBlockVector m_BlockVector; |
| 11093 | VmaDedicatedAllocationList m_DedicatedAllocations; |
| 11094 | |
| 11095 | VmaPool_T( |
| 11096 | VmaAllocator hAllocator, |
| 11097 | const VmaPoolCreateInfo& createInfo, |
| 11098 | VkDeviceSize preferredBlockSize); |
| 11099 | ~VmaPool_T(); |
| 11100 | |
| 11101 | uint32_t GetId() const { return m_Id; } |
| 11102 | void SetId(uint32_t id) { VMA_ASSERT(m_Id == 0); m_Id = id; } |
| 11103 | |
| 11104 | const char* GetName() const { return m_Name; } |
| 11105 | void SetName(const char* pName); |
| 11106 | |
| 11107 | #if VMA_STATS_STRING_ENABLED |
| 11108 | //void PrintDetailedMap(class VmaStringBuilder& sb); |
| 11109 | #endif |
| 11110 | |
| 11111 | private: |
| 11112 | uint32_t m_Id; |
| 11113 | char* m_Name; |
| 11114 | VmaPool_T* m_PrevPool = VMA_NULL; |
| 11115 | VmaPool_T* m_NextPool = VMA_NULL; |
| 11116 | }; |
| 11117 | |
| 11118 | struct VmaPoolListItemTraits |
| 11119 | { |
| 11120 | typedef VmaPool_T ItemType; |
| 11121 | |
| 11122 | static ItemType* GetPrev(const ItemType* item) { return item->m_PrevPool; } |
| 11123 | static ItemType* GetNext(const ItemType* item) { return item->m_NextPool; } |
| 11124 | static ItemType*& AccessPrev(ItemType* item) { return item->m_PrevPool; } |
| 11125 | static ItemType*& AccessNext(ItemType* item) { return item->m_NextPool; } |
| 11126 | }; |
| 11127 | #endif // _VMA_POOL_T |
| 11128 | |
| 11129 | #ifndef _VMA_CURRENT_BUDGET_DATA |
| 11130 | struct VmaCurrentBudgetData |
| 11131 | { |
| 11132 | VMA_ATOMIC_UINT32 m_BlockCount[VK_MAX_MEMORY_HEAPS]; |
| 11133 | VMA_ATOMIC_UINT32 m_AllocationCount[VK_MAX_MEMORY_HEAPS]; |
| 11134 | VMA_ATOMIC_UINT64 m_BlockBytes[VK_MAX_MEMORY_HEAPS]; |
| 11135 | VMA_ATOMIC_UINT64 m_AllocationBytes[VK_MAX_MEMORY_HEAPS]; |
| 11136 | |
| 11137 | #if VMA_MEMORY_BUDGET |
| 11138 | VMA_ATOMIC_UINT32 m_OperationsSinceBudgetFetch; |
| 11139 | VMA_RW_MUTEX m_BudgetMutex; |
| 11140 | uint64_t m_VulkanUsage[VK_MAX_MEMORY_HEAPS]; |
| 11141 | uint64_t m_VulkanBudget[VK_MAX_MEMORY_HEAPS]; |
| 11142 | uint64_t m_BlockBytesAtBudgetFetch[VK_MAX_MEMORY_HEAPS]; |
| 11143 | #endif // VMA_MEMORY_BUDGET |
| 11144 | |
| 11145 | VmaCurrentBudgetData(); |
| 11146 | |
| 11147 | void AddAllocation(uint32_t heapIndex, VkDeviceSize allocationSize); |
| 11148 | void RemoveAllocation(uint32_t heapIndex, VkDeviceSize allocationSize); |
| 11149 | }; |
| 11150 | |
| 11151 | #ifndef _VMA_CURRENT_BUDGET_DATA_FUNCTIONS |
| 11152 | VmaCurrentBudgetData::VmaCurrentBudgetData() |
| 11153 | { |
| 11154 | for (uint32_t heapIndex = 0; heapIndex < VK_MAX_MEMORY_HEAPS; ++heapIndex) |
| 11155 | { |
| 11156 | m_BlockCount[heapIndex] = 0; |
| 11157 | m_AllocationCount[heapIndex] = 0; |
| 11158 | m_BlockBytes[heapIndex] = 0; |
| 11159 | m_AllocationBytes[heapIndex] = 0; |
| 11160 | #if VMA_MEMORY_BUDGET |
| 11161 | m_VulkanUsage[heapIndex] = 0; |
| 11162 | m_VulkanBudget[heapIndex] = 0; |
| 11163 | m_BlockBytesAtBudgetFetch[heapIndex] = 0; |
| 11164 | #endif |
| 11165 | } |
| 11166 | |
| 11167 | #if VMA_MEMORY_BUDGET |
| 11168 | m_OperationsSinceBudgetFetch = 0; |
| 11169 | #endif |
| 11170 | } |
| 11171 | |
| 11172 | void VmaCurrentBudgetData::AddAllocation(uint32_t heapIndex, VkDeviceSize allocationSize) |
| 11173 | { |
| 11174 | m_AllocationBytes[heapIndex] += allocationSize; |
| 11175 | ++m_AllocationCount[heapIndex]; |
| 11176 | #if VMA_MEMORY_BUDGET |
| 11177 | ++m_OperationsSinceBudgetFetch; |
| 11178 | #endif |
| 11179 | } |
| 11180 | |
| 11181 | void VmaCurrentBudgetData::RemoveAllocation(uint32_t heapIndex, VkDeviceSize allocationSize) |
| 11182 | { |
| 11183 | VMA_ASSERT(m_AllocationBytes[heapIndex] >= allocationSize); |
| 11184 | m_AllocationBytes[heapIndex] -= allocationSize; |
| 11185 | VMA_ASSERT(m_AllocationCount[heapIndex] > 0); |
| 11186 | --m_AllocationCount[heapIndex]; |
| 11187 | #if VMA_MEMORY_BUDGET |
| 11188 | ++m_OperationsSinceBudgetFetch; |
| 11189 | #endif |
| 11190 | } |
| 11191 | #endif // _VMA_CURRENT_BUDGET_DATA_FUNCTIONS |
| 11192 | #endif // _VMA_CURRENT_BUDGET_DATA |
| 11193 | |
| 11194 | #ifndef _VMA_ALLOCATION_OBJECT_ALLOCATOR |
| 11195 | /* |
| 11196 | Thread-safe wrapper over VmaPoolAllocator free list, for allocation of VmaAllocation_T objects. |
| 11197 | */ |
| 11198 | class VmaAllocationObjectAllocator |
| 11199 | { |
| 11200 | VMA_CLASS_NO_COPY(VmaAllocationObjectAllocator) |
| 11201 | public: |
| 11202 | VmaAllocationObjectAllocator(const VkAllocationCallbacks* pAllocationCallbacks) |
| 11203 | : m_Allocator(pAllocationCallbacks, 1024) {} |
| 11204 | |
| 11205 | template<typename... Types> VmaAllocation Allocate(Types&&... args); |
| 11206 | void Free(VmaAllocation hAlloc); |
| 11207 | |
| 11208 | private: |
| 11209 | VMA_MUTEX m_Mutex; |
| 11210 | VmaPoolAllocator<VmaAllocation_T> m_Allocator; |
| 11211 | }; |
| 11212 | |
| 11213 | template<typename... Types> |
| 11214 | VmaAllocation VmaAllocationObjectAllocator::Allocate(Types&&... args) |
| 11215 | { |
| 11216 | VmaMutexLock mutexLock(m_Mutex); |
| 11217 | return m_Allocator.Alloc<Types...>(std::forward<Types>(args)...); |
| 11218 | } |
| 11219 | |
| 11220 | void VmaAllocationObjectAllocator::Free(VmaAllocation hAlloc) |
| 11221 | { |
| 11222 | VmaMutexLock mutexLock(m_Mutex); |
| 11223 | m_Allocator.Free(ptr: hAlloc); |
| 11224 | } |
| 11225 | #endif // _VMA_ALLOCATION_OBJECT_ALLOCATOR |
| 11226 | |
| 11227 | #ifndef _VMA_VIRTUAL_BLOCK_T |
| 11228 | struct VmaVirtualBlock_T |
| 11229 | { |
| 11230 | VMA_CLASS_NO_COPY(VmaVirtualBlock_T) |
| 11231 | public: |
| 11232 | const bool m_AllocationCallbacksSpecified; |
| 11233 | const VkAllocationCallbacks m_AllocationCallbacks; |
| 11234 | |
| 11235 | VmaVirtualBlock_T(const VmaVirtualBlockCreateInfo& createInfo); |
| 11236 | ~VmaVirtualBlock_T(); |
| 11237 | |
| 11238 | VkResult Init() { return VK_SUCCESS; } |
| 11239 | bool IsEmpty() const { return m_Metadata->IsEmpty(); } |
| 11240 | void Free(VmaVirtualAllocation allocation) { m_Metadata->Free(allocHandle: (VmaAllocHandle)allocation); } |
| 11241 | void SetAllocationUserData(VmaVirtualAllocation allocation, void* userData) { m_Metadata->SetAllocationUserData(allocHandle: (VmaAllocHandle)allocation, userData); } |
| 11242 | void Clear() { m_Metadata->Clear(); } |
| 11243 | |
| 11244 | const VkAllocationCallbacks* GetAllocationCallbacks() const; |
| 11245 | void GetAllocationInfo(VmaVirtualAllocation allocation, VmaVirtualAllocationInfo& outInfo); |
| 11246 | VkResult Allocate(const VmaVirtualAllocationCreateInfo& createInfo, VmaVirtualAllocation& outAllocation, |
| 11247 | VkDeviceSize* outOffset); |
| 11248 | void GetStatistics(VmaStatistics& outStats) const; |
| 11249 | void CalculateDetailedStatistics(VmaDetailedStatistics& outStats) const; |
| 11250 | #if VMA_STATS_STRING_ENABLED |
| 11251 | void BuildStatsString(bool detailedMap, VmaStringBuilder& sb) const; |
| 11252 | #endif |
| 11253 | |
| 11254 | private: |
| 11255 | VmaBlockMetadata* m_Metadata; |
| 11256 | }; |
| 11257 | |
| 11258 | #ifndef _VMA_VIRTUAL_BLOCK_T_FUNCTIONS |
| 11259 | VmaVirtualBlock_T::VmaVirtualBlock_T(const VmaVirtualBlockCreateInfo& createInfo) |
| 11260 | : m_AllocationCallbacksSpecified(createInfo.pAllocationCallbacks != VMA_NULL), |
| 11261 | m_AllocationCallbacks(createInfo.pAllocationCallbacks != VMA_NULL ? *createInfo.pAllocationCallbacks : VmaEmptyAllocationCallbacks) |
| 11262 | { |
| 11263 | const uint32_t algorithm = createInfo.flags & VMA_VIRTUAL_BLOCK_CREATE_ALGORITHM_MASK; |
| 11264 | switch (algorithm) |
| 11265 | { |
| 11266 | default: |
| 11267 | VMA_ASSERT(0); |
| 11268 | case 0: |
| 11269 | m_Metadata = vma_new(GetAllocationCallbacks(), VmaBlockMetadata_TLSF)(VK_NULL_HANDLE, 1, true); |
| 11270 | break; |
| 11271 | case VMA_VIRTUAL_BLOCK_CREATE_LINEAR_ALGORITHM_BIT: |
| 11272 | m_Metadata = vma_new(GetAllocationCallbacks(), VmaBlockMetadata_Linear)(VK_NULL_HANDLE, 1, true); |
| 11273 | break; |
| 11274 | } |
| 11275 | |
| 11276 | m_Metadata->Init(size: createInfo.size); |
| 11277 | } |
| 11278 | |
| 11279 | VmaVirtualBlock_T::~VmaVirtualBlock_T() |
| 11280 | { |
| 11281 | // Define macro VMA_DEBUG_LOG to receive the list of the unfreed allocations |
| 11282 | if (!m_Metadata->IsEmpty()) |
| 11283 | m_Metadata->DebugLogAllAllocations(); |
| 11284 | // This is the most important assert in the entire library. |
| 11285 | // Hitting it means you have some memory leak - unreleased virtual allocations. |
| 11286 | VMA_ASSERT(m_Metadata->IsEmpty() && "Some virtual allocations were not freed before destruction of this virtual block!" ); |
| 11287 | |
| 11288 | vma_delete(pAllocationCallbacks: GetAllocationCallbacks(), ptr: m_Metadata); |
| 11289 | } |
| 11290 | |
| 11291 | const VkAllocationCallbacks* VmaVirtualBlock_T::GetAllocationCallbacks() const |
| 11292 | { |
| 11293 | return m_AllocationCallbacksSpecified ? &m_AllocationCallbacks : VMA_NULL; |
| 11294 | } |
| 11295 | |
| 11296 | void VmaVirtualBlock_T::GetAllocationInfo(VmaVirtualAllocation allocation, VmaVirtualAllocationInfo& outInfo) |
| 11297 | { |
| 11298 | m_Metadata->GetAllocationInfo(allocHandle: (VmaAllocHandle)allocation, outInfo); |
| 11299 | } |
| 11300 | |
| 11301 | VkResult VmaVirtualBlock_T::Allocate(const VmaVirtualAllocationCreateInfo& createInfo, VmaVirtualAllocation& outAllocation, |
| 11302 | VkDeviceSize* outOffset) |
| 11303 | { |
| 11304 | VmaAllocationRequest request = {}; |
| 11305 | if (m_Metadata->CreateAllocationRequest( |
| 11306 | allocSize: createInfo.size, // allocSize |
| 11307 | VMA_MAX(createInfo.alignment, (VkDeviceSize)1), // allocAlignment |
| 11308 | upperAddress: (createInfo.flags & VMA_VIRTUAL_ALLOCATION_CREATE_UPPER_ADDRESS_BIT) != 0, // upperAddress |
| 11309 | allocType: VMA_SUBALLOCATION_TYPE_UNKNOWN, // allocType - unimportant |
| 11310 | strategy: createInfo.flags & VMA_VIRTUAL_ALLOCATION_CREATE_STRATEGY_MASK, // strategy |
| 11311 | pAllocationRequest: &request)) |
| 11312 | { |
| 11313 | m_Metadata->Alloc(request, |
| 11314 | type: VMA_SUBALLOCATION_TYPE_UNKNOWN, // type - unimportant |
| 11315 | userData: createInfo.pUserData); |
| 11316 | outAllocation = (VmaVirtualAllocation)request.allocHandle; |
| 11317 | if(outOffset) |
| 11318 | *outOffset = m_Metadata->GetAllocationOffset(allocHandle: request.allocHandle); |
| 11319 | return VK_SUCCESS; |
| 11320 | } |
| 11321 | outAllocation = (VmaVirtualAllocation)VK_NULL_HANDLE; |
| 11322 | if (outOffset) |
| 11323 | *outOffset = UINT64_MAX; |
| 11324 | return VK_ERROR_OUT_OF_DEVICE_MEMORY; |
| 11325 | } |
| 11326 | |
| 11327 | void VmaVirtualBlock_T::GetStatistics(VmaStatistics& outStats) const |
| 11328 | { |
| 11329 | VmaClearStatistics(outStats); |
| 11330 | m_Metadata->AddStatistics(inoutStats&: outStats); |
| 11331 | } |
| 11332 | |
| 11333 | void VmaVirtualBlock_T::CalculateDetailedStatistics(VmaDetailedStatistics& outStats) const |
| 11334 | { |
| 11335 | VmaClearDetailedStatistics(outStats); |
| 11336 | m_Metadata->AddDetailedStatistics(inoutStats&: outStats); |
| 11337 | } |
| 11338 | |
| 11339 | #if VMA_STATS_STRING_ENABLED |
| 11340 | void VmaVirtualBlock_T::BuildStatsString(bool detailedMap, VmaStringBuilder& sb) const |
| 11341 | { |
| 11342 | VmaJsonWriter json(GetAllocationCallbacks(), sb); |
| 11343 | json.BeginObject(); |
| 11344 | |
| 11345 | VmaDetailedStatistics stats; |
| 11346 | CalculateDetailedStatistics(outStats&: stats); |
| 11347 | |
| 11348 | json.WriteString(pStr: "Stats" ); |
| 11349 | VmaPrintDetailedStatistics(json, stat: stats); |
| 11350 | |
| 11351 | if (detailedMap) |
| 11352 | { |
| 11353 | json.WriteString(pStr: "Details" ); |
| 11354 | json.BeginObject(); |
| 11355 | m_Metadata->PrintDetailedMap(json); |
| 11356 | json.EndObject(); |
| 11357 | } |
| 11358 | |
| 11359 | json.EndObject(); |
| 11360 | } |
| 11361 | #endif // VMA_STATS_STRING_ENABLED |
| 11362 | #endif // _VMA_VIRTUAL_BLOCK_T_FUNCTIONS |
| 11363 | #endif // _VMA_VIRTUAL_BLOCK_T |
| 11364 | |
| 11365 | |
| 11366 | // Main allocator object. |
| 11367 | struct VmaAllocator_T |
| 11368 | { |
| 11369 | VMA_CLASS_NO_COPY(VmaAllocator_T) |
| 11370 | public: |
| 11371 | bool m_UseMutex; |
| 11372 | uint32_t m_VulkanApiVersion; |
| 11373 | bool m_UseKhrDedicatedAllocation; // Can be set only if m_VulkanApiVersion < VK_MAKE_VERSION(1, 1, 0). |
| 11374 | bool m_UseKhrBindMemory2; // Can be set only if m_VulkanApiVersion < VK_MAKE_VERSION(1, 1, 0). |
| 11375 | bool m_UseExtMemoryBudget; |
| 11376 | bool m_UseAmdDeviceCoherentMemory; |
| 11377 | bool m_UseKhrBufferDeviceAddress; |
| 11378 | bool m_UseExtMemoryPriority; |
| 11379 | VkDevice m_hDevice; |
| 11380 | VkInstance m_hInstance; |
| 11381 | bool m_AllocationCallbacksSpecified; |
| 11382 | VkAllocationCallbacks m_AllocationCallbacks; |
| 11383 | VmaDeviceMemoryCallbacks m_DeviceMemoryCallbacks; |
| 11384 | VmaAllocationObjectAllocator m_AllocationObjectAllocator; |
| 11385 | |
| 11386 | // Each bit (1 << i) is set if HeapSizeLimit is enabled for that heap, so cannot allocate more than the heap size. |
| 11387 | uint32_t m_HeapSizeLimitMask; |
| 11388 | |
| 11389 | VkPhysicalDeviceProperties m_PhysicalDeviceProperties; |
| 11390 | VkPhysicalDeviceMemoryProperties m_MemProps; |
| 11391 | |
| 11392 | // Default pools. |
| 11393 | VmaBlockVector* m_pBlockVectors[VK_MAX_MEMORY_TYPES]; |
| 11394 | VmaDedicatedAllocationList m_DedicatedAllocations[VK_MAX_MEMORY_TYPES]; |
| 11395 | |
| 11396 | VmaCurrentBudgetData m_Budget; |
| 11397 | VMA_ATOMIC_UINT32 m_DeviceMemoryCount; // Total number of VkDeviceMemory objects. |
| 11398 | |
| 11399 | VmaAllocator_T(const VmaAllocatorCreateInfo* pCreateInfo); |
| 11400 | VkResult Init(const VmaAllocatorCreateInfo* pCreateInfo); |
| 11401 | ~VmaAllocator_T(); |
| 11402 | |
| 11403 | const VkAllocationCallbacks* GetAllocationCallbacks() const |
| 11404 | { |
| 11405 | return m_AllocationCallbacksSpecified ? &m_AllocationCallbacks : VMA_NULL; |
| 11406 | } |
| 11407 | const VmaVulkanFunctions& GetVulkanFunctions() const |
| 11408 | { |
| 11409 | return m_VulkanFunctions; |
| 11410 | } |
| 11411 | |
| 11412 | VkPhysicalDevice GetPhysicalDevice() const { return m_PhysicalDevice; } |
| 11413 | |
| 11414 | VkDeviceSize GetBufferImageGranularity() const |
| 11415 | { |
| 11416 | return VMA_MAX( |
| 11417 | static_cast<VkDeviceSize>(VMA_DEBUG_MIN_BUFFER_IMAGE_GRANULARITY), |
| 11418 | m_PhysicalDeviceProperties.limits.bufferImageGranularity); |
| 11419 | } |
| 11420 | |
| 11421 | uint32_t GetMemoryHeapCount() const { return m_MemProps.memoryHeapCount; } |
| 11422 | uint32_t GetMemoryTypeCount() const { return m_MemProps.memoryTypeCount; } |
| 11423 | |
| 11424 | uint32_t MemoryTypeIndexToHeapIndex(uint32_t memTypeIndex) const |
| 11425 | { |
| 11426 | VMA_ASSERT(memTypeIndex < m_MemProps.memoryTypeCount); |
| 11427 | return m_MemProps.memoryTypes[memTypeIndex].heapIndex; |
| 11428 | } |
| 11429 | // True when specific memory type is HOST_VISIBLE but not HOST_COHERENT. |
| 11430 | bool IsMemoryTypeNonCoherent(uint32_t memTypeIndex) const |
| 11431 | { |
| 11432 | return (m_MemProps.memoryTypes[memTypeIndex].propertyFlags & (VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT | VK_MEMORY_PROPERTY_HOST_COHERENT_BIT)) == |
| 11433 | VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT; |
| 11434 | } |
| 11435 | // Minimum alignment for all allocations in specific memory type. |
| 11436 | VkDeviceSize GetMemoryTypeMinAlignment(uint32_t memTypeIndex) const |
| 11437 | { |
| 11438 | return IsMemoryTypeNonCoherent(memTypeIndex) ? |
| 11439 | VMA_MAX((VkDeviceSize)VMA_MIN_ALIGNMENT, m_PhysicalDeviceProperties.limits.nonCoherentAtomSize) : |
| 11440 | (VkDeviceSize)VMA_MIN_ALIGNMENT; |
| 11441 | } |
| 11442 | |
| 11443 | bool IsIntegratedGpu() const |
| 11444 | { |
| 11445 | return m_PhysicalDeviceProperties.deviceType == VK_PHYSICAL_DEVICE_TYPE_INTEGRATED_GPU; |
| 11446 | } |
| 11447 | |
| 11448 | uint32_t GetGlobalMemoryTypeBits() const { return m_GlobalMemoryTypeBits; } |
| 11449 | |
| 11450 | void GetBufferMemoryRequirements( |
| 11451 | VkBuffer hBuffer, |
| 11452 | VkMemoryRequirements& memReq, |
| 11453 | bool& requiresDedicatedAllocation, |
| 11454 | bool& prefersDedicatedAllocation) const; |
| 11455 | void GetImageMemoryRequirements( |
| 11456 | VkImage hImage, |
| 11457 | VkMemoryRequirements& memReq, |
| 11458 | bool& requiresDedicatedAllocation, |
| 11459 | bool& prefersDedicatedAllocation) const; |
| 11460 | VkResult FindMemoryTypeIndex( |
| 11461 | uint32_t memoryTypeBits, |
| 11462 | const VmaAllocationCreateInfo* pAllocationCreateInfo, |
| 11463 | VkFlags bufImgUsage, // VkBufferCreateInfo::usage or VkImageCreateInfo::usage. UINT32_MAX if unknown. |
| 11464 | uint32_t* pMemoryTypeIndex) const; |
| 11465 | |
| 11466 | // Main allocation function. |
| 11467 | VkResult AllocateMemory( |
| 11468 | const VkMemoryRequirements& vkMemReq, |
| 11469 | bool requiresDedicatedAllocation, |
| 11470 | bool prefersDedicatedAllocation, |
| 11471 | VkBuffer dedicatedBuffer, |
| 11472 | VkImage dedicatedImage, |
| 11473 | VkFlags dedicatedBufferImageUsage, // UINT32_MAX if unknown. |
| 11474 | const VmaAllocationCreateInfo& createInfo, |
| 11475 | VmaSuballocationType suballocType, |
| 11476 | size_t allocationCount, |
| 11477 | VmaAllocation* pAllocations); |
| 11478 | |
| 11479 | // Main deallocation function. |
| 11480 | void FreeMemory( |
| 11481 | size_t allocationCount, |
| 11482 | const VmaAllocation* pAllocations); |
| 11483 | |
| 11484 | void CalculateStatistics(VmaTotalStatistics* pStats); |
| 11485 | |
| 11486 | void GetHeapBudgets( |
| 11487 | VmaBudget* outBudgets, uint32_t firstHeap, uint32_t heapCount); |
| 11488 | |
| 11489 | #if VMA_STATS_STRING_ENABLED |
| 11490 | void PrintDetailedMap(class VmaJsonWriter& json); |
| 11491 | #endif |
| 11492 | |
| 11493 | void GetAllocationInfo(VmaAllocation hAllocation, VmaAllocationInfo* pAllocationInfo); |
| 11494 | |
| 11495 | VkResult CreatePool(const VmaPoolCreateInfo* pCreateInfo, VmaPool* pPool); |
| 11496 | void DestroyPool(VmaPool pool); |
| 11497 | void GetPoolStatistics(VmaPool pool, VmaStatistics* pPoolStats); |
| 11498 | void CalculatePoolStatistics(VmaPool pool, VmaDetailedStatistics* pPoolStats); |
| 11499 | |
| 11500 | void SetCurrentFrameIndex(uint32_t frameIndex); |
| 11501 | uint32_t GetCurrentFrameIndex() const { return m_CurrentFrameIndex.load(); } |
| 11502 | |
| 11503 | VkResult CheckPoolCorruption(VmaPool hPool); |
| 11504 | VkResult CheckCorruption(uint32_t memoryTypeBits); |
| 11505 | |
| 11506 | // Call to Vulkan function vkAllocateMemory with accompanying bookkeeping. |
| 11507 | VkResult AllocateVulkanMemory(const VkMemoryAllocateInfo* pAllocateInfo, VkDeviceMemory* pMemory); |
| 11508 | // Call to Vulkan function vkFreeMemory with accompanying bookkeeping. |
| 11509 | void FreeVulkanMemory(uint32_t memoryType, VkDeviceSize size, VkDeviceMemory hMemory); |
| 11510 | // Call to Vulkan function vkBindBufferMemory or vkBindBufferMemory2KHR. |
| 11511 | VkResult BindVulkanBuffer( |
| 11512 | VkDeviceMemory memory, |
| 11513 | VkDeviceSize memoryOffset, |
| 11514 | VkBuffer buffer, |
| 11515 | const void* pNext); |
| 11516 | // Call to Vulkan function vkBindImageMemory or vkBindImageMemory2KHR. |
| 11517 | VkResult BindVulkanImage( |
| 11518 | VkDeviceMemory memory, |
| 11519 | VkDeviceSize memoryOffset, |
| 11520 | VkImage image, |
| 11521 | const void* pNext); |
| 11522 | |
| 11523 | VkResult Map(VmaAllocation hAllocation, void** ppData); |
| 11524 | void Unmap(VmaAllocation hAllocation); |
| 11525 | |
| 11526 | VkResult BindBufferMemory( |
| 11527 | VmaAllocation hAllocation, |
| 11528 | VkDeviceSize allocationLocalOffset, |
| 11529 | VkBuffer hBuffer, |
| 11530 | const void* pNext); |
| 11531 | VkResult BindImageMemory( |
| 11532 | VmaAllocation hAllocation, |
| 11533 | VkDeviceSize allocationLocalOffset, |
| 11534 | VkImage hImage, |
| 11535 | const void* pNext); |
| 11536 | |
| 11537 | VkResult FlushOrInvalidateAllocation( |
| 11538 | VmaAllocation hAllocation, |
| 11539 | VkDeviceSize offset, VkDeviceSize size, |
| 11540 | VMA_CACHE_OPERATION op); |
| 11541 | VkResult FlushOrInvalidateAllocations( |
| 11542 | uint32_t allocationCount, |
| 11543 | const VmaAllocation* allocations, |
| 11544 | const VkDeviceSize* offsets, const VkDeviceSize* sizes, |
| 11545 | VMA_CACHE_OPERATION op); |
| 11546 | |
| 11547 | void FillAllocation(const VmaAllocation hAllocation, uint8_t pattern); |
| 11548 | |
| 11549 | /* |
| 11550 | Returns bit mask of memory types that can support defragmentation on GPU as |
| 11551 | they support creation of required buffer for copy operations. |
| 11552 | */ |
| 11553 | uint32_t GetGpuDefragmentationMemoryTypeBits(); |
| 11554 | |
| 11555 | #if VMA_EXTERNAL_MEMORY |
| 11556 | VkExternalMemoryHandleTypeFlagsKHR GetExternalMemoryHandleTypeFlags(uint32_t memTypeIndex) const |
| 11557 | { |
| 11558 | return m_TypeExternalMemoryHandleTypes[memTypeIndex]; |
| 11559 | } |
| 11560 | #endif // #if VMA_EXTERNAL_MEMORY |
| 11561 | |
| 11562 | private: |
| 11563 | VkDeviceSize m_PreferredLargeHeapBlockSize; |
| 11564 | |
| 11565 | VkPhysicalDevice m_PhysicalDevice; |
| 11566 | VMA_ATOMIC_UINT32 m_CurrentFrameIndex; |
| 11567 | VMA_ATOMIC_UINT32 m_GpuDefragmentationMemoryTypeBits; // UINT32_MAX means uninitialized. |
| 11568 | #if VMA_EXTERNAL_MEMORY |
| 11569 | VkExternalMemoryHandleTypeFlagsKHR m_TypeExternalMemoryHandleTypes[VK_MAX_MEMORY_TYPES]; |
| 11570 | #endif // #if VMA_EXTERNAL_MEMORY |
| 11571 | |
| 11572 | VMA_RW_MUTEX m_PoolsMutex; |
| 11573 | typedef VmaIntrusiveLinkedList<VmaPoolListItemTraits> PoolList; |
| 11574 | // Protected by m_PoolsMutex. |
| 11575 | PoolList m_Pools; |
| 11576 | uint32_t m_NextPoolId; |
| 11577 | |
| 11578 | VmaVulkanFunctions m_VulkanFunctions; |
| 11579 | |
| 11580 | // Global bit mask AND-ed with any memoryTypeBits to disallow certain memory types. |
| 11581 | uint32_t m_GlobalMemoryTypeBits; |
| 11582 | |
| 11583 | void ImportVulkanFunctions(const VmaVulkanFunctions* pVulkanFunctions); |
| 11584 | |
| 11585 | #if VMA_STATIC_VULKAN_FUNCTIONS == 1 |
| 11586 | void ImportVulkanFunctions_Static(); |
| 11587 | #endif |
| 11588 | |
| 11589 | void ImportVulkanFunctions_Custom(const VmaVulkanFunctions* pVulkanFunctions); |
| 11590 | |
| 11591 | #if VMA_DYNAMIC_VULKAN_FUNCTIONS == 1 |
| 11592 | void ImportVulkanFunctions_Dynamic(); |
| 11593 | #endif |
| 11594 | |
| 11595 | void ValidateVulkanFunctions(); |
| 11596 | |
| 11597 | VkDeviceSize CalcPreferredBlockSize(uint32_t memTypeIndex); |
| 11598 | |
| 11599 | VkResult AllocateMemoryOfType( |
| 11600 | VmaPool pool, |
| 11601 | VkDeviceSize size, |
| 11602 | VkDeviceSize alignment, |
| 11603 | bool dedicatedPreferred, |
| 11604 | VkBuffer dedicatedBuffer, |
| 11605 | VkImage dedicatedImage, |
| 11606 | VkFlags dedicatedBufferImageUsage, |
| 11607 | const VmaAllocationCreateInfo& createInfo, |
| 11608 | uint32_t memTypeIndex, |
| 11609 | VmaSuballocationType suballocType, |
| 11610 | VmaDedicatedAllocationList& dedicatedAllocations, |
| 11611 | VmaBlockVector& blockVector, |
| 11612 | size_t allocationCount, |
| 11613 | VmaAllocation* pAllocations); |
| 11614 | |
| 11615 | // Helper function only to be used inside AllocateDedicatedMemory. |
| 11616 | VkResult AllocateDedicatedMemoryPage( |
| 11617 | VmaPool pool, |
| 11618 | VkDeviceSize size, |
| 11619 | VmaSuballocationType suballocType, |
| 11620 | uint32_t memTypeIndex, |
| 11621 | const VkMemoryAllocateInfo& allocInfo, |
| 11622 | bool map, |
| 11623 | bool isUserDataString, |
| 11624 | bool isMappingAllowed, |
| 11625 | void* pUserData, |
| 11626 | VmaAllocation* pAllocation); |
| 11627 | |
| 11628 | // Allocates and registers new VkDeviceMemory specifically for dedicated allocations. |
| 11629 | VkResult AllocateDedicatedMemory( |
| 11630 | VmaPool pool, |
| 11631 | VkDeviceSize size, |
| 11632 | VmaSuballocationType suballocType, |
| 11633 | VmaDedicatedAllocationList& dedicatedAllocations, |
| 11634 | uint32_t memTypeIndex, |
| 11635 | bool map, |
| 11636 | bool isUserDataString, |
| 11637 | bool isMappingAllowed, |
| 11638 | bool canAliasMemory, |
| 11639 | void* pUserData, |
| 11640 | float priority, |
| 11641 | VkBuffer dedicatedBuffer, |
| 11642 | VkImage dedicatedImage, |
| 11643 | VkFlags dedicatedBufferImageUsage, |
| 11644 | size_t allocationCount, |
| 11645 | VmaAllocation* pAllocations, |
| 11646 | const void* pNextChain = nullptr); |
| 11647 | |
| 11648 | void FreeDedicatedMemory(const VmaAllocation allocation); |
| 11649 | |
| 11650 | VkResult CalcMemTypeParams( |
| 11651 | VmaAllocationCreateInfo& outCreateInfo, |
| 11652 | uint32_t memTypeIndex, |
| 11653 | VkDeviceSize size, |
| 11654 | size_t allocationCount); |
| 11655 | VkResult CalcAllocationParams( |
| 11656 | VmaAllocationCreateInfo& outCreateInfo, |
| 11657 | bool dedicatedRequired, |
| 11658 | bool dedicatedPreferred); |
| 11659 | |
| 11660 | /* |
| 11661 | Calculates and returns bit mask of memory types that can support defragmentation |
| 11662 | on GPU as they support creation of required buffer for copy operations. |
| 11663 | */ |
| 11664 | uint32_t CalculateGpuDefragmentationMemoryTypeBits() const; |
| 11665 | uint32_t CalculateGlobalMemoryTypeBits() const; |
| 11666 | |
| 11667 | bool GetFlushOrInvalidateRange( |
| 11668 | VmaAllocation allocation, |
| 11669 | VkDeviceSize offset, VkDeviceSize size, |
| 11670 | VkMappedMemoryRange& outRange) const; |
| 11671 | |
| 11672 | #if VMA_MEMORY_BUDGET |
| 11673 | void UpdateVulkanBudget(); |
| 11674 | #endif // #if VMA_MEMORY_BUDGET |
| 11675 | }; |
| 11676 | |
| 11677 | |
| 11678 | #ifndef _VMA_MEMORY_FUNCTIONS |
| 11679 | static void* VmaMalloc(VmaAllocator hAllocator, size_t size, size_t alignment) |
| 11680 | { |
| 11681 | return VmaMalloc(pAllocationCallbacks: &hAllocator->m_AllocationCallbacks, size, alignment); |
| 11682 | } |
| 11683 | |
| 11684 | static void VmaFree(VmaAllocator hAllocator, void* ptr) |
| 11685 | { |
| 11686 | VmaFree(pAllocationCallbacks: &hAllocator->m_AllocationCallbacks, ptr); |
| 11687 | } |
| 11688 | |
| 11689 | template<typename T> |
| 11690 | static T* VmaAllocate(VmaAllocator hAllocator) |
| 11691 | { |
| 11692 | return (T*)VmaMalloc(hAllocator, size: sizeof(T), VMA_ALIGN_OF(T)); |
| 11693 | } |
| 11694 | |
| 11695 | template<typename T> |
| 11696 | static T* VmaAllocateArray(VmaAllocator hAllocator, size_t count) |
| 11697 | { |
| 11698 | return (T*)VmaMalloc(hAllocator, size: sizeof(T) * count, VMA_ALIGN_OF(T)); |
| 11699 | } |
| 11700 | |
| 11701 | template<typename T> |
| 11702 | static void vma_delete(VmaAllocator hAllocator, T* ptr) |
| 11703 | { |
| 11704 | if(ptr != VMA_NULL) |
| 11705 | { |
| 11706 | ptr->~T(); |
| 11707 | VmaFree(hAllocator, ptr); |
| 11708 | } |
| 11709 | } |
| 11710 | |
| 11711 | template<typename T> |
| 11712 | static void vma_delete_array(VmaAllocator hAllocator, T* ptr, size_t count) |
| 11713 | { |
| 11714 | if(ptr != VMA_NULL) |
| 11715 | { |
| 11716 | for(size_t i = count; i--; ) |
| 11717 | ptr[i].~T(); |
| 11718 | VmaFree(hAllocator, ptr); |
| 11719 | } |
| 11720 | } |
| 11721 | #endif // _VMA_MEMORY_FUNCTIONS |
| 11722 | |
| 11723 | #ifndef _VMA_DEVICE_MEMORY_BLOCK_FUNCTIONS |
| 11724 | VmaDeviceMemoryBlock::VmaDeviceMemoryBlock(VmaAllocator hAllocator) |
| 11725 | : m_pMetadata(VMA_NULL), |
| 11726 | m_MemoryTypeIndex(UINT32_MAX), |
| 11727 | m_Id(0), |
| 11728 | m_hMemory(VK_NULL_HANDLE), |
| 11729 | m_MapCount(0), |
| 11730 | m_pMappedData(VMA_NULL) {} |
| 11731 | |
| 11732 | VmaDeviceMemoryBlock::~VmaDeviceMemoryBlock() |
| 11733 | { |
| 11734 | VMA_ASSERT(m_MapCount == 0 && "VkDeviceMemory block is being destroyed while it is still mapped." ); |
| 11735 | VMA_ASSERT(m_hMemory == VK_NULL_HANDLE); |
| 11736 | } |
| 11737 | |
| 11738 | void VmaDeviceMemoryBlock::Init( |
| 11739 | VmaAllocator hAllocator, |
| 11740 | VmaPool hParentPool, |
| 11741 | uint32_t newMemoryTypeIndex, |
| 11742 | VkDeviceMemory newMemory, |
| 11743 | VkDeviceSize newSize, |
| 11744 | uint32_t id, |
| 11745 | uint32_t algorithm, |
| 11746 | VkDeviceSize bufferImageGranularity) |
| 11747 | { |
| 11748 | VMA_ASSERT(m_hMemory == VK_NULL_HANDLE); |
| 11749 | |
| 11750 | m_hParentPool = hParentPool; |
| 11751 | m_MemoryTypeIndex = newMemoryTypeIndex; |
| 11752 | m_Id = id; |
| 11753 | m_hMemory = newMemory; |
| 11754 | |
| 11755 | switch (algorithm) |
| 11756 | { |
| 11757 | case VMA_POOL_CREATE_LINEAR_ALGORITHM_BIT: |
| 11758 | m_pMetadata = vma_new(hAllocator, VmaBlockMetadata_Linear)(hAllocator->GetAllocationCallbacks(), |
| 11759 | bufferImageGranularity, false); // isVirtual |
| 11760 | break; |
| 11761 | default: |
| 11762 | VMA_ASSERT(0); |
| 11763 | // Fall-through. |
| 11764 | case 0: |
| 11765 | m_pMetadata = vma_new(hAllocator, VmaBlockMetadata_TLSF)(hAllocator->GetAllocationCallbacks(), |
| 11766 | bufferImageGranularity, false); // isVirtual |
| 11767 | } |
| 11768 | m_pMetadata->Init(size: newSize); |
| 11769 | } |
| 11770 | |
| 11771 | void VmaDeviceMemoryBlock::Destroy(VmaAllocator allocator) |
| 11772 | { |
| 11773 | // Define macro VMA_DEBUG_LOG to receive the list of the unfreed allocations |
| 11774 | if (!m_pMetadata->IsEmpty()) |
| 11775 | m_pMetadata->DebugLogAllAllocations(); |
| 11776 | // This is the most important assert in the entire library. |
| 11777 | // Hitting it means you have some memory leak - unreleased VmaAllocation objects. |
| 11778 | VMA_ASSERT(m_pMetadata->IsEmpty() && "Some allocations were not freed before destruction of this memory block!" ); |
| 11779 | |
| 11780 | VMA_ASSERT(m_hMemory != VK_NULL_HANDLE); |
| 11781 | allocator->FreeVulkanMemory(memoryType: m_MemoryTypeIndex, size: m_pMetadata->GetSize(), hMemory: m_hMemory); |
| 11782 | m_hMemory = VK_NULL_HANDLE; |
| 11783 | |
| 11784 | vma_delete(hAllocator: allocator, ptr: m_pMetadata); |
| 11785 | m_pMetadata = VMA_NULL; |
| 11786 | } |
| 11787 | |
| 11788 | void VmaDeviceMemoryBlock::PostFree(VmaAllocator hAllocator) |
| 11789 | { |
| 11790 | if(m_MappingHysteresis.PostFree()) |
| 11791 | { |
| 11792 | VMA_ASSERT(m_MappingHysteresis.GetExtraMapping() == 0); |
| 11793 | if (m_MapCount == 0) |
| 11794 | { |
| 11795 | m_pMappedData = VMA_NULL; |
| 11796 | (*hAllocator->GetVulkanFunctions().vkUnmapMemory)(hAllocator->m_hDevice, m_hMemory); |
| 11797 | } |
| 11798 | } |
| 11799 | } |
| 11800 | |
| 11801 | bool VmaDeviceMemoryBlock::Validate() const |
| 11802 | { |
| 11803 | VMA_VALIDATE((m_hMemory != VK_NULL_HANDLE) && |
| 11804 | (m_pMetadata->GetSize() != 0)); |
| 11805 | |
| 11806 | return m_pMetadata->Validate(); |
| 11807 | } |
| 11808 | |
| 11809 | VkResult VmaDeviceMemoryBlock::CheckCorruption(VmaAllocator hAllocator) |
| 11810 | { |
| 11811 | void* pData = nullptr; |
| 11812 | VkResult res = Map(hAllocator, count: 1, ppData: &pData); |
| 11813 | if (res != VK_SUCCESS) |
| 11814 | { |
| 11815 | return res; |
| 11816 | } |
| 11817 | |
| 11818 | res = m_pMetadata->CheckCorruption(pBlockData: pData); |
| 11819 | |
| 11820 | Unmap(hAllocator, count: 1); |
| 11821 | |
| 11822 | return res; |
| 11823 | } |
| 11824 | |
| 11825 | VkResult VmaDeviceMemoryBlock::Map(VmaAllocator hAllocator, uint32_t count, void** ppData) |
| 11826 | { |
| 11827 | if (count == 0) |
| 11828 | { |
| 11829 | return VK_SUCCESS; |
| 11830 | } |
| 11831 | |
| 11832 | VmaMutexLock lock(m_MapAndBindMutex, hAllocator->m_UseMutex); |
| 11833 | const uint32_t oldTotalMapCount = m_MapCount + m_MappingHysteresis.GetExtraMapping(); |
| 11834 | m_MappingHysteresis.PostMap(); |
| 11835 | if (oldTotalMapCount != 0) |
| 11836 | { |
| 11837 | m_MapCount += count; |
| 11838 | VMA_ASSERT(m_pMappedData != VMA_NULL); |
| 11839 | if (ppData != VMA_NULL) |
| 11840 | { |
| 11841 | *ppData = m_pMappedData; |
| 11842 | } |
| 11843 | return VK_SUCCESS; |
| 11844 | } |
| 11845 | else |
| 11846 | { |
| 11847 | VkResult result = (*hAllocator->GetVulkanFunctions().vkMapMemory)( |
| 11848 | hAllocator->m_hDevice, |
| 11849 | m_hMemory, |
| 11850 | 0, // offset |
| 11851 | VK_WHOLE_SIZE, |
| 11852 | 0, // flags |
| 11853 | &m_pMappedData); |
| 11854 | if (result == VK_SUCCESS) |
| 11855 | { |
| 11856 | if (ppData != VMA_NULL) |
| 11857 | { |
| 11858 | *ppData = m_pMappedData; |
| 11859 | } |
| 11860 | m_MapCount = count; |
| 11861 | } |
| 11862 | return result; |
| 11863 | } |
| 11864 | } |
| 11865 | |
| 11866 | void VmaDeviceMemoryBlock::Unmap(VmaAllocator hAllocator, uint32_t count) |
| 11867 | { |
| 11868 | if (count == 0) |
| 11869 | { |
| 11870 | return; |
| 11871 | } |
| 11872 | |
| 11873 | VmaMutexLock lock(m_MapAndBindMutex, hAllocator->m_UseMutex); |
| 11874 | if (m_MapCount >= count) |
| 11875 | { |
| 11876 | m_MapCount -= count; |
| 11877 | const uint32_t totalMapCount = m_MapCount + m_MappingHysteresis.GetExtraMapping(); |
| 11878 | if (totalMapCount == 0) |
| 11879 | { |
| 11880 | m_pMappedData = VMA_NULL; |
| 11881 | (*hAllocator->GetVulkanFunctions().vkUnmapMemory)(hAllocator->m_hDevice, m_hMemory); |
| 11882 | } |
| 11883 | m_MappingHysteresis.PostUnmap(); |
| 11884 | } |
| 11885 | else |
| 11886 | { |
| 11887 | VMA_ASSERT(0 && "VkDeviceMemory block is being unmapped while it was not previously mapped." ); |
| 11888 | } |
| 11889 | } |
| 11890 | |
| 11891 | VkResult VmaDeviceMemoryBlock::WriteMagicValueAfterAllocation(VmaAllocator hAllocator, VkDeviceSize allocOffset, VkDeviceSize allocSize) |
| 11892 | { |
| 11893 | VMA_ASSERT(VMA_DEBUG_MARGIN > 0 && VMA_DEBUG_MARGIN % 4 == 0 && VMA_DEBUG_DETECT_CORRUPTION); |
| 11894 | |
| 11895 | void* pData; |
| 11896 | VkResult res = Map(hAllocator, count: 1, ppData: &pData); |
| 11897 | if (res != VK_SUCCESS) |
| 11898 | { |
| 11899 | return res; |
| 11900 | } |
| 11901 | |
| 11902 | VmaWriteMagicValue(pData, offset: allocOffset + allocSize); |
| 11903 | |
| 11904 | Unmap(hAllocator, count: 1); |
| 11905 | return VK_SUCCESS; |
| 11906 | } |
| 11907 | |
| 11908 | VkResult VmaDeviceMemoryBlock::ValidateMagicValueAfterAllocation(VmaAllocator hAllocator, VkDeviceSize allocOffset, VkDeviceSize allocSize) |
| 11909 | { |
| 11910 | VMA_ASSERT(VMA_DEBUG_MARGIN > 0 && VMA_DEBUG_MARGIN % 4 == 0 && VMA_DEBUG_DETECT_CORRUPTION); |
| 11911 | |
| 11912 | void* pData; |
| 11913 | VkResult res = Map(hAllocator, count: 1, ppData: &pData); |
| 11914 | if (res != VK_SUCCESS) |
| 11915 | { |
| 11916 | return res; |
| 11917 | } |
| 11918 | |
| 11919 | if (!VmaValidateMagicValue(pData, offset: allocOffset + allocSize)) |
| 11920 | { |
| 11921 | VMA_ASSERT(0 && "MEMORY CORRUPTION DETECTED AFTER FREED ALLOCATION!" ); |
| 11922 | } |
| 11923 | |
| 11924 | Unmap(hAllocator, count: 1); |
| 11925 | return VK_SUCCESS; |
| 11926 | } |
| 11927 | |
| 11928 | VkResult VmaDeviceMemoryBlock::BindBufferMemory( |
| 11929 | const VmaAllocator hAllocator, |
| 11930 | const VmaAllocation hAllocation, |
| 11931 | VkDeviceSize allocationLocalOffset, |
| 11932 | VkBuffer hBuffer, |
| 11933 | const void* pNext) |
| 11934 | { |
| 11935 | VMA_ASSERT(hAllocation->GetType() == VmaAllocation_T::ALLOCATION_TYPE_BLOCK && |
| 11936 | hAllocation->GetBlock() == this); |
| 11937 | VMA_ASSERT(allocationLocalOffset < hAllocation->GetSize() && |
| 11938 | "Invalid allocationLocalOffset. Did you forget that this offset is relative to the beginning of the allocation, not the whole memory block?" ); |
| 11939 | const VkDeviceSize memoryOffset = hAllocation->GetOffset() + allocationLocalOffset; |
| 11940 | // This lock is important so that we don't call vkBind... and/or vkMap... simultaneously on the same VkDeviceMemory from multiple threads. |
| 11941 | VmaMutexLock lock(m_MapAndBindMutex, hAllocator->m_UseMutex); |
| 11942 | return hAllocator->BindVulkanBuffer(memory: m_hMemory, memoryOffset, buffer: hBuffer, pNext); |
| 11943 | } |
| 11944 | |
| 11945 | VkResult VmaDeviceMemoryBlock::BindImageMemory( |
| 11946 | const VmaAllocator hAllocator, |
| 11947 | const VmaAllocation hAllocation, |
| 11948 | VkDeviceSize allocationLocalOffset, |
| 11949 | VkImage hImage, |
| 11950 | const void* pNext) |
| 11951 | { |
| 11952 | VMA_ASSERT(hAllocation->GetType() == VmaAllocation_T::ALLOCATION_TYPE_BLOCK && |
| 11953 | hAllocation->GetBlock() == this); |
| 11954 | VMA_ASSERT(allocationLocalOffset < hAllocation->GetSize() && |
| 11955 | "Invalid allocationLocalOffset. Did you forget that this offset is relative to the beginning of the allocation, not the whole memory block?" ); |
| 11956 | const VkDeviceSize memoryOffset = hAllocation->GetOffset() + allocationLocalOffset; |
| 11957 | // This lock is important so that we don't call vkBind... and/or vkMap... simultaneously on the same VkDeviceMemory from multiple threads. |
| 11958 | VmaMutexLock lock(m_MapAndBindMutex, hAllocator->m_UseMutex); |
| 11959 | return hAllocator->BindVulkanImage(memory: m_hMemory, memoryOffset, image: hImage, pNext); |
| 11960 | } |
| 11961 | #endif // _VMA_DEVICE_MEMORY_BLOCK_FUNCTIONS |
| 11962 | |
| 11963 | #ifndef _VMA_ALLOCATION_T_FUNCTIONS |
| 11964 | VmaAllocation_T::VmaAllocation_T(bool mappingAllowed) |
| 11965 | : m_Alignment{ 1 }, |
| 11966 | m_Size{ 0 }, |
| 11967 | m_pUserData{ VMA_NULL }, |
| 11968 | m_pName{ VMA_NULL }, |
| 11969 | m_MemoryTypeIndex{ 0 }, |
| 11970 | m_Type{ (uint8_t)ALLOCATION_TYPE_NONE }, |
| 11971 | m_SuballocationType{ (uint8_t)VMA_SUBALLOCATION_TYPE_UNKNOWN }, |
| 11972 | m_MapCount{ 0 }, |
| 11973 | m_Flags{ 0 } |
| 11974 | { |
| 11975 | if(mappingAllowed) |
| 11976 | m_Flags |= (uint8_t)FLAG_MAPPING_ALLOWED; |
| 11977 | |
| 11978 | #if VMA_STATS_STRING_ENABLED |
| 11979 | m_BufferImageUsage = 0; |
| 11980 | #endif |
| 11981 | } |
| 11982 | |
| 11983 | VmaAllocation_T::~VmaAllocation_T() |
| 11984 | { |
| 11985 | VMA_ASSERT(m_MapCount == 0 && "Allocation was not unmapped before destruction." ); |
| 11986 | |
| 11987 | // Check if owned string was freed. |
| 11988 | VMA_ASSERT(m_pName == VMA_NULL); |
| 11989 | } |
| 11990 | |
| 11991 | void VmaAllocation_T::InitBlockAllocation( |
| 11992 | VmaDeviceMemoryBlock* block, |
| 11993 | VmaAllocHandle allocHandle, |
| 11994 | VkDeviceSize alignment, |
| 11995 | VkDeviceSize size, |
| 11996 | uint32_t memoryTypeIndex, |
| 11997 | VmaSuballocationType suballocationType, |
| 11998 | bool mapped) |
| 11999 | { |
| 12000 | VMA_ASSERT(m_Type == ALLOCATION_TYPE_NONE); |
| 12001 | VMA_ASSERT(block != VMA_NULL); |
| 12002 | m_Type = (uint8_t)ALLOCATION_TYPE_BLOCK; |
| 12003 | m_Alignment = alignment; |
| 12004 | m_Size = size; |
| 12005 | m_MemoryTypeIndex = memoryTypeIndex; |
| 12006 | if(mapped) |
| 12007 | { |
| 12008 | VMA_ASSERT(IsMappingAllowed() && "Mapping is not allowed on this allocation! Please use one of the new VMA_ALLOCATION_CREATE_HOST_ACCESS_* flags when creating it." ); |
| 12009 | m_Flags |= (uint8_t)FLAG_PERSISTENT_MAP; |
| 12010 | } |
| 12011 | m_SuballocationType = (uint8_t)suballocationType; |
| 12012 | m_BlockAllocation.m_Block = block; |
| 12013 | m_BlockAllocation.m_AllocHandle = allocHandle; |
| 12014 | } |
| 12015 | |
| 12016 | void VmaAllocation_T::InitDedicatedAllocation( |
| 12017 | VmaPool hParentPool, |
| 12018 | uint32_t memoryTypeIndex, |
| 12019 | VkDeviceMemory hMemory, |
| 12020 | VmaSuballocationType suballocationType, |
| 12021 | void* pMappedData, |
| 12022 | VkDeviceSize size) |
| 12023 | { |
| 12024 | VMA_ASSERT(m_Type == ALLOCATION_TYPE_NONE); |
| 12025 | VMA_ASSERT(hMemory != VK_NULL_HANDLE); |
| 12026 | m_Type = (uint8_t)ALLOCATION_TYPE_DEDICATED; |
| 12027 | m_Alignment = 0; |
| 12028 | m_Size = size; |
| 12029 | m_MemoryTypeIndex = memoryTypeIndex; |
| 12030 | m_SuballocationType = (uint8_t)suballocationType; |
| 12031 | if(pMappedData != VMA_NULL) |
| 12032 | { |
| 12033 | VMA_ASSERT(IsMappingAllowed() && "Mapping is not allowed on this allocation! Please use one of the new VMA_ALLOCATION_CREATE_HOST_ACCESS_* flags when creating it." ); |
| 12034 | m_Flags |= (uint8_t)FLAG_PERSISTENT_MAP; |
| 12035 | } |
| 12036 | m_DedicatedAllocation.m_hParentPool = hParentPool; |
| 12037 | m_DedicatedAllocation.m_hMemory = hMemory; |
| 12038 | m_DedicatedAllocation.m_pMappedData = pMappedData; |
| 12039 | m_DedicatedAllocation.m_Prev = VMA_NULL; |
| 12040 | m_DedicatedAllocation.m_Next = VMA_NULL; |
| 12041 | } |
| 12042 | |
| 12043 | void VmaAllocation_T::SetName(VmaAllocator hAllocator, const char* pName) |
| 12044 | { |
| 12045 | VMA_ASSERT(pName == VMA_NULL || pName != m_pName); |
| 12046 | |
| 12047 | FreeName(hAllocator); |
| 12048 | |
| 12049 | if (pName != VMA_NULL) |
| 12050 | m_pName = VmaCreateStringCopy(allocs: hAllocator->GetAllocationCallbacks(), srcStr: pName); |
| 12051 | } |
| 12052 | |
| 12053 | uint8_t VmaAllocation_T::SwapBlockAllocation(VmaAllocator hAllocator, VmaAllocation allocation) |
| 12054 | { |
| 12055 | VMA_ASSERT(allocation != VMA_NULL); |
| 12056 | VMA_ASSERT(m_Type == ALLOCATION_TYPE_BLOCK); |
| 12057 | VMA_ASSERT(allocation->m_Type == ALLOCATION_TYPE_BLOCK); |
| 12058 | |
| 12059 | if (m_MapCount != 0) |
| 12060 | m_BlockAllocation.m_Block->Unmap(hAllocator, count: m_MapCount); |
| 12061 | |
| 12062 | m_BlockAllocation.m_Block->m_pMetadata->SetAllocationUserData(allocHandle: m_BlockAllocation.m_AllocHandle, userData: allocation); |
| 12063 | VMA_SWAP(m_BlockAllocation, allocation->m_BlockAllocation); |
| 12064 | m_BlockAllocation.m_Block->m_pMetadata->SetAllocationUserData(allocHandle: m_BlockAllocation.m_AllocHandle, userData: this); |
| 12065 | |
| 12066 | #if VMA_STATS_STRING_ENABLED |
| 12067 | VMA_SWAP(m_BufferImageUsage, allocation->m_BufferImageUsage); |
| 12068 | #endif |
| 12069 | return m_MapCount; |
| 12070 | } |
| 12071 | |
| 12072 | VmaAllocHandle VmaAllocation_T::GetAllocHandle() const |
| 12073 | { |
| 12074 | switch (m_Type) |
| 12075 | { |
| 12076 | case ALLOCATION_TYPE_BLOCK: |
| 12077 | return m_BlockAllocation.m_AllocHandle; |
| 12078 | case ALLOCATION_TYPE_DEDICATED: |
| 12079 | return VK_NULL_HANDLE; |
| 12080 | default: |
| 12081 | VMA_ASSERT(0); |
| 12082 | return VK_NULL_HANDLE; |
| 12083 | } |
| 12084 | } |
| 12085 | |
| 12086 | VkDeviceSize VmaAllocation_T::GetOffset() const |
| 12087 | { |
| 12088 | switch (m_Type) |
| 12089 | { |
| 12090 | case ALLOCATION_TYPE_BLOCK: |
| 12091 | return m_BlockAllocation.m_Block->m_pMetadata->GetAllocationOffset(allocHandle: m_BlockAllocation.m_AllocHandle); |
| 12092 | case ALLOCATION_TYPE_DEDICATED: |
| 12093 | return 0; |
| 12094 | default: |
| 12095 | VMA_ASSERT(0); |
| 12096 | return 0; |
| 12097 | } |
| 12098 | } |
| 12099 | |
| 12100 | VmaPool VmaAllocation_T::GetParentPool() const |
| 12101 | { |
| 12102 | switch (m_Type) |
| 12103 | { |
| 12104 | case ALLOCATION_TYPE_BLOCK: |
| 12105 | return m_BlockAllocation.m_Block->GetParentPool(); |
| 12106 | case ALLOCATION_TYPE_DEDICATED: |
| 12107 | return m_DedicatedAllocation.m_hParentPool; |
| 12108 | default: |
| 12109 | VMA_ASSERT(0); |
| 12110 | return VK_NULL_HANDLE; |
| 12111 | } |
| 12112 | } |
| 12113 | |
| 12114 | VkDeviceMemory VmaAllocation_T::GetMemory() const |
| 12115 | { |
| 12116 | switch (m_Type) |
| 12117 | { |
| 12118 | case ALLOCATION_TYPE_BLOCK: |
| 12119 | return m_BlockAllocation.m_Block->GetDeviceMemory(); |
| 12120 | case ALLOCATION_TYPE_DEDICATED: |
| 12121 | return m_DedicatedAllocation.m_hMemory; |
| 12122 | default: |
| 12123 | VMA_ASSERT(0); |
| 12124 | return VK_NULL_HANDLE; |
| 12125 | } |
| 12126 | } |
| 12127 | |
| 12128 | void* VmaAllocation_T::GetMappedData() const |
| 12129 | { |
| 12130 | switch (m_Type) |
| 12131 | { |
| 12132 | case ALLOCATION_TYPE_BLOCK: |
| 12133 | if (m_MapCount != 0 || IsPersistentMap()) |
| 12134 | { |
| 12135 | void* pBlockData = m_BlockAllocation.m_Block->GetMappedData(); |
| 12136 | VMA_ASSERT(pBlockData != VMA_NULL); |
| 12137 | return (char*)pBlockData + GetOffset(); |
| 12138 | } |
| 12139 | else |
| 12140 | { |
| 12141 | return VMA_NULL; |
| 12142 | } |
| 12143 | break; |
| 12144 | case ALLOCATION_TYPE_DEDICATED: |
| 12145 | VMA_ASSERT((m_DedicatedAllocation.m_pMappedData != VMA_NULL) == (m_MapCount != 0 || IsPersistentMap())); |
| 12146 | return m_DedicatedAllocation.m_pMappedData; |
| 12147 | default: |
| 12148 | VMA_ASSERT(0); |
| 12149 | return VMA_NULL; |
| 12150 | } |
| 12151 | } |
| 12152 | |
| 12153 | void VmaAllocation_T::BlockAllocMap() |
| 12154 | { |
| 12155 | VMA_ASSERT(GetType() == ALLOCATION_TYPE_BLOCK); |
| 12156 | VMA_ASSERT(IsMappingAllowed() && "Mapping is not allowed on this allocation! Please use one of the new VMA_ALLOCATION_CREATE_HOST_ACCESS_* flags when creating it." ); |
| 12157 | |
| 12158 | if (m_MapCount < 0xFF) |
| 12159 | { |
| 12160 | ++m_MapCount; |
| 12161 | } |
| 12162 | else |
| 12163 | { |
| 12164 | VMA_ASSERT(0 && "Allocation mapped too many times simultaneously." ); |
| 12165 | } |
| 12166 | } |
| 12167 | |
| 12168 | void VmaAllocation_T::BlockAllocUnmap() |
| 12169 | { |
| 12170 | VMA_ASSERT(GetType() == ALLOCATION_TYPE_BLOCK); |
| 12171 | |
| 12172 | if (m_MapCount > 0) |
| 12173 | { |
| 12174 | --m_MapCount; |
| 12175 | } |
| 12176 | else |
| 12177 | { |
| 12178 | VMA_ASSERT(0 && "Unmapping allocation not previously mapped." ); |
| 12179 | } |
| 12180 | } |
| 12181 | |
| 12182 | VkResult VmaAllocation_T::DedicatedAllocMap(VmaAllocator hAllocator, void** ppData) |
| 12183 | { |
| 12184 | VMA_ASSERT(GetType() == ALLOCATION_TYPE_DEDICATED); |
| 12185 | VMA_ASSERT(IsMappingAllowed() && "Mapping is not allowed on this allocation! Please use one of the new VMA_ALLOCATION_CREATE_HOST_ACCESS_* flags when creating it." ); |
| 12186 | |
| 12187 | if (m_MapCount != 0 || IsPersistentMap()) |
| 12188 | { |
| 12189 | if (m_MapCount < 0xFF) |
| 12190 | { |
| 12191 | VMA_ASSERT(m_DedicatedAllocation.m_pMappedData != VMA_NULL); |
| 12192 | *ppData = m_DedicatedAllocation.m_pMappedData; |
| 12193 | ++m_MapCount; |
| 12194 | return VK_SUCCESS; |
| 12195 | } |
| 12196 | else |
| 12197 | { |
| 12198 | VMA_ASSERT(0 && "Dedicated allocation mapped too many times simultaneously." ); |
| 12199 | return VK_ERROR_MEMORY_MAP_FAILED; |
| 12200 | } |
| 12201 | } |
| 12202 | else |
| 12203 | { |
| 12204 | VkResult result = (*hAllocator->GetVulkanFunctions().vkMapMemory)( |
| 12205 | hAllocator->m_hDevice, |
| 12206 | m_DedicatedAllocation.m_hMemory, |
| 12207 | 0, // offset |
| 12208 | VK_WHOLE_SIZE, |
| 12209 | 0, // flags |
| 12210 | ppData); |
| 12211 | if (result == VK_SUCCESS) |
| 12212 | { |
| 12213 | m_DedicatedAllocation.m_pMappedData = *ppData; |
| 12214 | m_MapCount = 1; |
| 12215 | } |
| 12216 | return result; |
| 12217 | } |
| 12218 | } |
| 12219 | |
| 12220 | void VmaAllocation_T::DedicatedAllocUnmap(VmaAllocator hAllocator) |
| 12221 | { |
| 12222 | VMA_ASSERT(GetType() == ALLOCATION_TYPE_DEDICATED); |
| 12223 | |
| 12224 | if (m_MapCount > 0) |
| 12225 | { |
| 12226 | --m_MapCount; |
| 12227 | if (m_MapCount == 0 && !IsPersistentMap()) |
| 12228 | { |
| 12229 | m_DedicatedAllocation.m_pMappedData = VMA_NULL; |
| 12230 | (*hAllocator->GetVulkanFunctions().vkUnmapMemory)( |
| 12231 | hAllocator->m_hDevice, |
| 12232 | m_DedicatedAllocation.m_hMemory); |
| 12233 | } |
| 12234 | } |
| 12235 | else |
| 12236 | { |
| 12237 | VMA_ASSERT(0 && "Unmapping dedicated allocation not previously mapped." ); |
| 12238 | } |
| 12239 | } |
| 12240 | |
| 12241 | #if VMA_STATS_STRING_ENABLED |
| 12242 | void VmaAllocation_T::InitBufferImageUsage(uint32_t bufferImageUsage) |
| 12243 | { |
| 12244 | VMA_ASSERT(m_BufferImageUsage == 0); |
| 12245 | m_BufferImageUsage = bufferImageUsage; |
| 12246 | } |
| 12247 | |
| 12248 | void VmaAllocation_T::PrintParameters(class VmaJsonWriter& json) const |
| 12249 | { |
| 12250 | json.WriteString(pStr: "Type" ); |
| 12251 | json.WriteString(pStr: VMA_SUBALLOCATION_TYPE_NAMES[m_SuballocationType]); |
| 12252 | |
| 12253 | json.WriteString(pStr: "Size" ); |
| 12254 | json.WriteNumber(n: m_Size); |
| 12255 | json.WriteString(pStr: "Usage" ); |
| 12256 | json.WriteNumber(n: m_BufferImageUsage); |
| 12257 | |
| 12258 | if (m_pUserData != VMA_NULL) |
| 12259 | { |
| 12260 | json.WriteString(pStr: "CustomData" ); |
| 12261 | json.BeginString(); |
| 12262 | json.ContinueString_Pointer(ptr: m_pUserData); |
| 12263 | json.EndString(); |
| 12264 | } |
| 12265 | if (m_pName != VMA_NULL) |
| 12266 | { |
| 12267 | json.WriteString(pStr: "Name" ); |
| 12268 | json.WriteString(pStr: m_pName); |
| 12269 | } |
| 12270 | } |
| 12271 | #endif // VMA_STATS_STRING_ENABLED |
| 12272 | |
| 12273 | void VmaAllocation_T::FreeName(VmaAllocator hAllocator) |
| 12274 | { |
| 12275 | if(m_pName) |
| 12276 | { |
| 12277 | VmaFreeString(allocs: hAllocator->GetAllocationCallbacks(), str: m_pName); |
| 12278 | m_pName = VMA_NULL; |
| 12279 | } |
| 12280 | } |
| 12281 | #endif // _VMA_ALLOCATION_T_FUNCTIONS |
| 12282 | |
| 12283 | #ifndef _VMA_BLOCK_VECTOR_FUNCTIONS |
| 12284 | VmaBlockVector::VmaBlockVector( |
| 12285 | VmaAllocator hAllocator, |
| 12286 | VmaPool hParentPool, |
| 12287 | uint32_t memoryTypeIndex, |
| 12288 | VkDeviceSize preferredBlockSize, |
| 12289 | size_t minBlockCount, |
| 12290 | size_t maxBlockCount, |
| 12291 | VkDeviceSize bufferImageGranularity, |
| 12292 | bool explicitBlockSize, |
| 12293 | uint32_t algorithm, |
| 12294 | float priority, |
| 12295 | VkDeviceSize minAllocationAlignment, |
| 12296 | void* pMemoryAllocateNext) |
| 12297 | : m_hAllocator(hAllocator), |
| 12298 | m_hParentPool(hParentPool), |
| 12299 | m_MemoryTypeIndex(memoryTypeIndex), |
| 12300 | m_PreferredBlockSize(preferredBlockSize), |
| 12301 | m_MinBlockCount(minBlockCount), |
| 12302 | m_MaxBlockCount(maxBlockCount), |
| 12303 | m_BufferImageGranularity(bufferImageGranularity), |
| 12304 | m_ExplicitBlockSize(explicitBlockSize), |
| 12305 | m_Algorithm(algorithm), |
| 12306 | m_Priority(priority), |
| 12307 | m_MinAllocationAlignment(minAllocationAlignment), |
| 12308 | m_pMemoryAllocateNext(pMemoryAllocateNext), |
| 12309 | m_Blocks(VmaStlAllocator<VmaDeviceMemoryBlock*>(hAllocator->GetAllocationCallbacks())), |
| 12310 | m_NextBlockId(0) {} |
| 12311 | |
| 12312 | VmaBlockVector::~VmaBlockVector() |
| 12313 | { |
| 12314 | for (size_t i = m_Blocks.size(); i--; ) |
| 12315 | { |
| 12316 | m_Blocks[i]->Destroy(allocator: m_hAllocator); |
| 12317 | vma_delete(hAllocator: m_hAllocator, ptr: m_Blocks[i]); |
| 12318 | } |
| 12319 | } |
| 12320 | |
| 12321 | VkResult VmaBlockVector::CreateMinBlocks() |
| 12322 | { |
| 12323 | for (size_t i = 0; i < m_MinBlockCount; ++i) |
| 12324 | { |
| 12325 | VkResult res = CreateBlock(blockSize: m_PreferredBlockSize, VMA_NULL); |
| 12326 | if (res != VK_SUCCESS) |
| 12327 | { |
| 12328 | return res; |
| 12329 | } |
| 12330 | } |
| 12331 | return VK_SUCCESS; |
| 12332 | } |
| 12333 | |
| 12334 | void VmaBlockVector::AddStatistics(VmaStatistics& inoutStats) |
| 12335 | { |
| 12336 | VmaMutexLockRead lock(m_Mutex, m_hAllocator->m_UseMutex); |
| 12337 | |
| 12338 | const size_t blockCount = m_Blocks.size(); |
| 12339 | for (uint32_t blockIndex = 0; blockIndex < blockCount; ++blockIndex) |
| 12340 | { |
| 12341 | const VmaDeviceMemoryBlock* const pBlock = m_Blocks[blockIndex]; |
| 12342 | VMA_ASSERT(pBlock); |
| 12343 | VMA_HEAVY_ASSERT(pBlock->Validate()); |
| 12344 | pBlock->m_pMetadata->AddStatistics(inoutStats); |
| 12345 | } |
| 12346 | } |
| 12347 | |
| 12348 | void VmaBlockVector::AddDetailedStatistics(VmaDetailedStatistics& inoutStats) |
| 12349 | { |
| 12350 | VmaMutexLockRead lock(m_Mutex, m_hAllocator->m_UseMutex); |
| 12351 | |
| 12352 | const size_t blockCount = m_Blocks.size(); |
| 12353 | for (uint32_t blockIndex = 0; blockIndex < blockCount; ++blockIndex) |
| 12354 | { |
| 12355 | const VmaDeviceMemoryBlock* const pBlock = m_Blocks[blockIndex]; |
| 12356 | VMA_ASSERT(pBlock); |
| 12357 | VMA_HEAVY_ASSERT(pBlock->Validate()); |
| 12358 | pBlock->m_pMetadata->AddDetailedStatistics(inoutStats); |
| 12359 | } |
| 12360 | } |
| 12361 | |
| 12362 | bool VmaBlockVector::IsEmpty() |
| 12363 | { |
| 12364 | VmaMutexLockRead lock(m_Mutex, m_hAllocator->m_UseMutex); |
| 12365 | return m_Blocks.empty(); |
| 12366 | } |
| 12367 | |
| 12368 | bool VmaBlockVector::IsCorruptionDetectionEnabled() const |
| 12369 | { |
| 12370 | const uint32_t requiredMemFlags = VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT | VK_MEMORY_PROPERTY_HOST_COHERENT_BIT; |
| 12371 | return (VMA_DEBUG_DETECT_CORRUPTION != 0) && |
| 12372 | (VMA_DEBUG_MARGIN > 0) && |
| 12373 | (m_Algorithm == 0 || m_Algorithm == VMA_POOL_CREATE_LINEAR_ALGORITHM_BIT) && |
| 12374 | (m_hAllocator->m_MemProps.memoryTypes[m_MemoryTypeIndex].propertyFlags & requiredMemFlags) == requiredMemFlags; |
| 12375 | } |
| 12376 | |
| 12377 | VkResult VmaBlockVector::Allocate( |
| 12378 | VkDeviceSize size, |
| 12379 | VkDeviceSize alignment, |
| 12380 | const VmaAllocationCreateInfo& createInfo, |
| 12381 | VmaSuballocationType suballocType, |
| 12382 | size_t allocationCount, |
| 12383 | VmaAllocation* pAllocations) |
| 12384 | { |
| 12385 | size_t allocIndex; |
| 12386 | VkResult res = VK_SUCCESS; |
| 12387 | |
| 12388 | alignment = VMA_MAX(alignment, m_MinAllocationAlignment); |
| 12389 | |
| 12390 | if (IsCorruptionDetectionEnabled()) |
| 12391 | { |
| 12392 | size = VmaAlignUp<VkDeviceSize>(val: size, alignment: sizeof(VMA_CORRUPTION_DETECTION_MAGIC_VALUE)); |
| 12393 | alignment = VmaAlignUp<VkDeviceSize>(val: alignment, alignment: sizeof(VMA_CORRUPTION_DETECTION_MAGIC_VALUE)); |
| 12394 | } |
| 12395 | |
| 12396 | { |
| 12397 | VmaMutexLockWrite lock(m_Mutex, m_hAllocator->m_UseMutex); |
| 12398 | for (allocIndex = 0; allocIndex < allocationCount; ++allocIndex) |
| 12399 | { |
| 12400 | res = AllocatePage( |
| 12401 | size, |
| 12402 | alignment, |
| 12403 | createInfo, |
| 12404 | suballocType, |
| 12405 | pAllocation: pAllocations + allocIndex); |
| 12406 | if (res != VK_SUCCESS) |
| 12407 | { |
| 12408 | break; |
| 12409 | } |
| 12410 | } |
| 12411 | } |
| 12412 | |
| 12413 | if (res != VK_SUCCESS) |
| 12414 | { |
| 12415 | // Free all already created allocations. |
| 12416 | while (allocIndex--) |
| 12417 | Free(hAllocation: pAllocations[allocIndex]); |
| 12418 | memset(s: pAllocations, c: 0, n: sizeof(VmaAllocation) * allocationCount); |
| 12419 | } |
| 12420 | |
| 12421 | return res; |
| 12422 | } |
| 12423 | |
| 12424 | VkResult VmaBlockVector::AllocatePage( |
| 12425 | VkDeviceSize size, |
| 12426 | VkDeviceSize alignment, |
| 12427 | const VmaAllocationCreateInfo& createInfo, |
| 12428 | VmaSuballocationType suballocType, |
| 12429 | VmaAllocation* pAllocation) |
| 12430 | { |
| 12431 | const bool isUpperAddress = (createInfo.flags & VMA_ALLOCATION_CREATE_UPPER_ADDRESS_BIT) != 0; |
| 12432 | |
| 12433 | VkDeviceSize freeMemory; |
| 12434 | { |
| 12435 | const uint32_t heapIndex = m_hAllocator->MemoryTypeIndexToHeapIndex(memTypeIndex: m_MemoryTypeIndex); |
| 12436 | VmaBudget heapBudget = {}; |
| 12437 | m_hAllocator->GetHeapBudgets(outBudgets: &heapBudget, firstHeap: heapIndex, heapCount: 1); |
| 12438 | freeMemory = (heapBudget.usage < heapBudget.budget) ? (heapBudget.budget - heapBudget.usage) : 0; |
| 12439 | } |
| 12440 | |
| 12441 | const bool canFallbackToDedicated = !HasExplicitBlockSize() && |
| 12442 | (createInfo.flags & VMA_ALLOCATION_CREATE_NEVER_ALLOCATE_BIT) == 0; |
| 12443 | const bool canCreateNewBlock = |
| 12444 | ((createInfo.flags & VMA_ALLOCATION_CREATE_NEVER_ALLOCATE_BIT) == 0) && |
| 12445 | (m_Blocks.size() < m_MaxBlockCount) && |
| 12446 | (freeMemory >= size || !canFallbackToDedicated); |
| 12447 | uint32_t strategy = createInfo.flags & VMA_ALLOCATION_CREATE_STRATEGY_MASK; |
| 12448 | |
| 12449 | // Upper address can only be used with linear allocator and within single memory block. |
| 12450 | if (isUpperAddress && |
| 12451 | (m_Algorithm != VMA_POOL_CREATE_LINEAR_ALGORITHM_BIT || m_MaxBlockCount > 1)) |
| 12452 | { |
| 12453 | return VK_ERROR_FEATURE_NOT_PRESENT; |
| 12454 | } |
| 12455 | |
| 12456 | // Early reject: requested allocation size is larger that maximum block size for this block vector. |
| 12457 | if (size + VMA_DEBUG_MARGIN > m_PreferredBlockSize) |
| 12458 | { |
| 12459 | return VK_ERROR_OUT_OF_DEVICE_MEMORY; |
| 12460 | } |
| 12461 | |
| 12462 | // 1. Search existing allocations. Try to allocate. |
| 12463 | if (m_Algorithm == VMA_POOL_CREATE_LINEAR_ALGORITHM_BIT) |
| 12464 | { |
| 12465 | // Use only last block. |
| 12466 | if (!m_Blocks.empty()) |
| 12467 | { |
| 12468 | VmaDeviceMemoryBlock* const pCurrBlock = m_Blocks.back(); |
| 12469 | VMA_ASSERT(pCurrBlock); |
| 12470 | VkResult res = AllocateFromBlock( |
| 12471 | pBlock: pCurrBlock, size, alignment, allocFlags: createInfo.flags, pUserData: createInfo.pUserData, suballocType, strategy, pAllocation); |
| 12472 | if (res == VK_SUCCESS) |
| 12473 | { |
| 12474 | VMA_DEBUG_LOG(" Returned from last block #%u" , pCurrBlock->GetId()); |
| 12475 | IncrementallySortBlocks(); |
| 12476 | return VK_SUCCESS; |
| 12477 | } |
| 12478 | } |
| 12479 | } |
| 12480 | else |
| 12481 | { |
| 12482 | if (strategy != VMA_ALLOCATION_CREATE_STRATEGY_MIN_TIME_BIT) // MIN_MEMORY or default |
| 12483 | { |
| 12484 | const bool isHostVisible = |
| 12485 | (m_hAllocator->m_MemProps.memoryTypes[m_MemoryTypeIndex].propertyFlags & VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT) != 0; |
| 12486 | if(isHostVisible) |
| 12487 | { |
| 12488 | const bool isMappingAllowed = (createInfo.flags & |
| 12489 | (VMA_ALLOCATION_CREATE_HOST_ACCESS_SEQUENTIAL_WRITE_BIT | VMA_ALLOCATION_CREATE_HOST_ACCESS_RANDOM_BIT)) != 0; |
| 12490 | /* |
| 12491 | For non-mappable allocations, check blocks that are not mapped first. |
| 12492 | For mappable allocations, check blocks that are already mapped first. |
| 12493 | This way, having many blocks, we will separate mappable and non-mappable allocations, |
| 12494 | hopefully limiting the number of blocks that are mapped, which will help tools like RenderDoc. |
| 12495 | */ |
| 12496 | for(size_t mappingI = 0; mappingI < 2; ++mappingI) |
| 12497 | { |
| 12498 | // Forward order in m_Blocks - prefer blocks with smallest amount of free space. |
| 12499 | for (size_t blockIndex = 0; blockIndex < m_Blocks.size(); ++blockIndex) |
| 12500 | { |
| 12501 | VmaDeviceMemoryBlock* const pCurrBlock = m_Blocks[blockIndex]; |
| 12502 | VMA_ASSERT(pCurrBlock); |
| 12503 | const bool isBlockMapped = pCurrBlock->GetMappedData() != VMA_NULL; |
| 12504 | if((mappingI == 0) == (isMappingAllowed == isBlockMapped)) |
| 12505 | { |
| 12506 | VkResult res = AllocateFromBlock( |
| 12507 | pBlock: pCurrBlock, size, alignment, allocFlags: createInfo.flags, pUserData: createInfo.pUserData, suballocType, strategy, pAllocation); |
| 12508 | if (res == VK_SUCCESS) |
| 12509 | { |
| 12510 | VMA_DEBUG_LOG(" Returned from existing block #%u" , pCurrBlock->GetId()); |
| 12511 | IncrementallySortBlocks(); |
| 12512 | return VK_SUCCESS; |
| 12513 | } |
| 12514 | } |
| 12515 | } |
| 12516 | } |
| 12517 | } |
| 12518 | else |
| 12519 | { |
| 12520 | // Forward order in m_Blocks - prefer blocks with smallest amount of free space. |
| 12521 | for (size_t blockIndex = 0; blockIndex < m_Blocks.size(); ++blockIndex) |
| 12522 | { |
| 12523 | VmaDeviceMemoryBlock* const pCurrBlock = m_Blocks[blockIndex]; |
| 12524 | VMA_ASSERT(pCurrBlock); |
| 12525 | VkResult res = AllocateFromBlock( |
| 12526 | pBlock: pCurrBlock, size, alignment, allocFlags: createInfo.flags, pUserData: createInfo.pUserData, suballocType, strategy, pAllocation); |
| 12527 | if (res == VK_SUCCESS) |
| 12528 | { |
| 12529 | VMA_DEBUG_LOG(" Returned from existing block #%u" , pCurrBlock->GetId()); |
| 12530 | IncrementallySortBlocks(); |
| 12531 | return VK_SUCCESS; |
| 12532 | } |
| 12533 | } |
| 12534 | } |
| 12535 | } |
| 12536 | else // VMA_ALLOCATION_CREATE_STRATEGY_MIN_TIME_BIT |
| 12537 | { |
| 12538 | // Backward order in m_Blocks - prefer blocks with largest amount of free space. |
| 12539 | for (size_t blockIndex = m_Blocks.size(); blockIndex--; ) |
| 12540 | { |
| 12541 | VmaDeviceMemoryBlock* const pCurrBlock = m_Blocks[blockIndex]; |
| 12542 | VMA_ASSERT(pCurrBlock); |
| 12543 | VkResult res = AllocateFromBlock(pBlock: pCurrBlock, size, alignment, allocFlags: createInfo.flags, pUserData: createInfo.pUserData, suballocType, strategy, pAllocation); |
| 12544 | if (res == VK_SUCCESS) |
| 12545 | { |
| 12546 | VMA_DEBUG_LOG(" Returned from existing block #%u" , pCurrBlock->GetId()); |
| 12547 | IncrementallySortBlocks(); |
| 12548 | return VK_SUCCESS; |
| 12549 | } |
| 12550 | } |
| 12551 | } |
| 12552 | } |
| 12553 | |
| 12554 | // 2. Try to create new block. |
| 12555 | if (canCreateNewBlock) |
| 12556 | { |
| 12557 | // Calculate optimal size for new block. |
| 12558 | VkDeviceSize newBlockSize = m_PreferredBlockSize; |
| 12559 | uint32_t newBlockSizeShift = 0; |
| 12560 | const uint32_t NEW_BLOCK_SIZE_SHIFT_MAX = 3; |
| 12561 | |
| 12562 | if (!m_ExplicitBlockSize) |
| 12563 | { |
| 12564 | // Allocate 1/8, 1/4, 1/2 as first blocks. |
| 12565 | const VkDeviceSize maxExistingBlockSize = CalcMaxBlockSize(); |
| 12566 | for (uint32_t i = 0; i < NEW_BLOCK_SIZE_SHIFT_MAX; ++i) |
| 12567 | { |
| 12568 | const VkDeviceSize smallerNewBlockSize = newBlockSize / 2; |
| 12569 | if (smallerNewBlockSize > maxExistingBlockSize && smallerNewBlockSize >= size * 2) |
| 12570 | { |
| 12571 | newBlockSize = smallerNewBlockSize; |
| 12572 | ++newBlockSizeShift; |
| 12573 | } |
| 12574 | else |
| 12575 | { |
| 12576 | break; |
| 12577 | } |
| 12578 | } |
| 12579 | } |
| 12580 | |
| 12581 | size_t newBlockIndex = 0; |
| 12582 | VkResult res = (newBlockSize <= freeMemory || !canFallbackToDedicated) ? |
| 12583 | CreateBlock(blockSize: newBlockSize, pNewBlockIndex: &newBlockIndex) : VK_ERROR_OUT_OF_DEVICE_MEMORY; |
| 12584 | // Allocation of this size failed? Try 1/2, 1/4, 1/8 of m_PreferredBlockSize. |
| 12585 | if (!m_ExplicitBlockSize) |
| 12586 | { |
| 12587 | while (res < 0 && newBlockSizeShift < NEW_BLOCK_SIZE_SHIFT_MAX) |
| 12588 | { |
| 12589 | const VkDeviceSize smallerNewBlockSize = newBlockSize / 2; |
| 12590 | if (smallerNewBlockSize >= size) |
| 12591 | { |
| 12592 | newBlockSize = smallerNewBlockSize; |
| 12593 | ++newBlockSizeShift; |
| 12594 | res = (newBlockSize <= freeMemory || !canFallbackToDedicated) ? |
| 12595 | CreateBlock(blockSize: newBlockSize, pNewBlockIndex: &newBlockIndex) : VK_ERROR_OUT_OF_DEVICE_MEMORY; |
| 12596 | } |
| 12597 | else |
| 12598 | { |
| 12599 | break; |
| 12600 | } |
| 12601 | } |
| 12602 | } |
| 12603 | |
| 12604 | if (res == VK_SUCCESS) |
| 12605 | { |
| 12606 | VmaDeviceMemoryBlock* const pBlock = m_Blocks[newBlockIndex]; |
| 12607 | VMA_ASSERT(pBlock->m_pMetadata->GetSize() >= size); |
| 12608 | |
| 12609 | res = AllocateFromBlock( |
| 12610 | pBlock, size, alignment, allocFlags: createInfo.flags, pUserData: createInfo.pUserData, suballocType, strategy, pAllocation); |
| 12611 | if (res == VK_SUCCESS) |
| 12612 | { |
| 12613 | VMA_DEBUG_LOG(" Created new block #%u Size=%llu" , pBlock->GetId(), newBlockSize); |
| 12614 | IncrementallySortBlocks(); |
| 12615 | return VK_SUCCESS; |
| 12616 | } |
| 12617 | else |
| 12618 | { |
| 12619 | // Allocation from new block failed, possibly due to VMA_DEBUG_MARGIN or alignment. |
| 12620 | return VK_ERROR_OUT_OF_DEVICE_MEMORY; |
| 12621 | } |
| 12622 | } |
| 12623 | } |
| 12624 | |
| 12625 | return VK_ERROR_OUT_OF_DEVICE_MEMORY; |
| 12626 | } |
| 12627 | |
| 12628 | void VmaBlockVector::Free(const VmaAllocation hAllocation) |
| 12629 | { |
| 12630 | VmaDeviceMemoryBlock* pBlockToDelete = VMA_NULL; |
| 12631 | |
| 12632 | bool budgetExceeded = false; |
| 12633 | { |
| 12634 | const uint32_t heapIndex = m_hAllocator->MemoryTypeIndexToHeapIndex(memTypeIndex: m_MemoryTypeIndex); |
| 12635 | VmaBudget heapBudget = {}; |
| 12636 | m_hAllocator->GetHeapBudgets(outBudgets: &heapBudget, firstHeap: heapIndex, heapCount: 1); |
| 12637 | budgetExceeded = heapBudget.usage >= heapBudget.budget; |
| 12638 | } |
| 12639 | |
| 12640 | // Scope for lock. |
| 12641 | { |
| 12642 | VmaMutexLockWrite lock(m_Mutex, m_hAllocator->m_UseMutex); |
| 12643 | |
| 12644 | VmaDeviceMemoryBlock* pBlock = hAllocation->GetBlock(); |
| 12645 | |
| 12646 | if (IsCorruptionDetectionEnabled()) |
| 12647 | { |
| 12648 | VkResult res = pBlock->ValidateMagicValueAfterAllocation(hAllocator: m_hAllocator, allocOffset: hAllocation->GetOffset(), allocSize: hAllocation->GetSize()); |
| 12649 | VMA_ASSERT(res == VK_SUCCESS && "Couldn't map block memory to validate magic value." ); |
| 12650 | } |
| 12651 | |
| 12652 | if (hAllocation->IsPersistentMap()) |
| 12653 | { |
| 12654 | pBlock->Unmap(hAllocator: m_hAllocator, count: 1); |
| 12655 | } |
| 12656 | |
| 12657 | const bool hadEmptyBlockBeforeFree = HasEmptyBlock(); |
| 12658 | pBlock->m_pMetadata->Free(allocHandle: hAllocation->GetAllocHandle()); |
| 12659 | pBlock->PostFree(hAllocator: m_hAllocator); |
| 12660 | VMA_HEAVY_ASSERT(pBlock->Validate()); |
| 12661 | |
| 12662 | VMA_DEBUG_LOG(" Freed from MemoryTypeIndex=%u" , m_MemoryTypeIndex); |
| 12663 | |
| 12664 | const bool canDeleteBlock = m_Blocks.size() > m_MinBlockCount; |
| 12665 | // pBlock became empty after this deallocation. |
| 12666 | if (pBlock->m_pMetadata->IsEmpty()) |
| 12667 | { |
| 12668 | // Already had empty block. We don't want to have two, so delete this one. |
| 12669 | if ((hadEmptyBlockBeforeFree || budgetExceeded) && canDeleteBlock) |
| 12670 | { |
| 12671 | pBlockToDelete = pBlock; |
| 12672 | Remove(pBlock); |
| 12673 | } |
| 12674 | // else: We now have one empty block - leave it. A hysteresis to avoid allocating whole block back and forth. |
| 12675 | } |
| 12676 | // pBlock didn't become empty, but we have another empty block - find and free that one. |
| 12677 | // (This is optional, heuristics.) |
| 12678 | else if (hadEmptyBlockBeforeFree && canDeleteBlock) |
| 12679 | { |
| 12680 | VmaDeviceMemoryBlock* pLastBlock = m_Blocks.back(); |
| 12681 | if (pLastBlock->m_pMetadata->IsEmpty()) |
| 12682 | { |
| 12683 | pBlockToDelete = pLastBlock; |
| 12684 | m_Blocks.pop_back(); |
| 12685 | } |
| 12686 | } |
| 12687 | |
| 12688 | IncrementallySortBlocks(); |
| 12689 | } |
| 12690 | |
| 12691 | // Destruction of a free block. Deferred until this point, outside of mutex |
| 12692 | // lock, for performance reason. |
| 12693 | if (pBlockToDelete != VMA_NULL) |
| 12694 | { |
| 12695 | VMA_DEBUG_LOG(" Deleted empty block #%u" , pBlockToDelete->GetId()); |
| 12696 | pBlockToDelete->Destroy(allocator: m_hAllocator); |
| 12697 | vma_delete(hAllocator: m_hAllocator, ptr: pBlockToDelete); |
| 12698 | } |
| 12699 | |
| 12700 | m_hAllocator->m_Budget.RemoveAllocation(heapIndex: m_hAllocator->MemoryTypeIndexToHeapIndex(memTypeIndex: m_MemoryTypeIndex), allocationSize: hAllocation->GetSize()); |
| 12701 | m_hAllocator->m_AllocationObjectAllocator.Free(hAlloc: hAllocation); |
| 12702 | } |
| 12703 | |
| 12704 | VkDeviceSize VmaBlockVector::CalcMaxBlockSize() const |
| 12705 | { |
| 12706 | VkDeviceSize result = 0; |
| 12707 | for (size_t i = m_Blocks.size(); i--; ) |
| 12708 | { |
| 12709 | result = VMA_MAX(result, m_Blocks[i]->m_pMetadata->GetSize()); |
| 12710 | if (result >= m_PreferredBlockSize) |
| 12711 | { |
| 12712 | break; |
| 12713 | } |
| 12714 | } |
| 12715 | return result; |
| 12716 | } |
| 12717 | |
| 12718 | void VmaBlockVector::Remove(VmaDeviceMemoryBlock* pBlock) |
| 12719 | { |
| 12720 | for (uint32_t blockIndex = 0; blockIndex < m_Blocks.size(); ++blockIndex) |
| 12721 | { |
| 12722 | if (m_Blocks[blockIndex] == pBlock) |
| 12723 | { |
| 12724 | VmaVectorRemove(vec&: m_Blocks, index: blockIndex); |
| 12725 | return; |
| 12726 | } |
| 12727 | } |
| 12728 | VMA_ASSERT(0); |
| 12729 | } |
| 12730 | |
| 12731 | void VmaBlockVector::IncrementallySortBlocks() |
| 12732 | { |
| 12733 | if (!m_IncrementalSort) |
| 12734 | return; |
| 12735 | if (m_Algorithm != VMA_POOL_CREATE_LINEAR_ALGORITHM_BIT) |
| 12736 | { |
| 12737 | // Bubble sort only until first swap. |
| 12738 | for (size_t i = 1; i < m_Blocks.size(); ++i) |
| 12739 | { |
| 12740 | if (m_Blocks[i - 1]->m_pMetadata->GetSumFreeSize() > m_Blocks[i]->m_pMetadata->GetSumFreeSize()) |
| 12741 | { |
| 12742 | VMA_SWAP(m_Blocks[i - 1], m_Blocks[i]); |
| 12743 | return; |
| 12744 | } |
| 12745 | } |
| 12746 | } |
| 12747 | } |
| 12748 | |
| 12749 | void VmaBlockVector::SortByFreeSize() |
| 12750 | { |
| 12751 | VMA_SORT(m_Blocks.begin(), m_Blocks.end(), |
| 12752 | [](VmaDeviceMemoryBlock* b1, VmaDeviceMemoryBlock* b2) -> bool |
| 12753 | { |
| 12754 | return b1->m_pMetadata->GetSumFreeSize() < b2->m_pMetadata->GetSumFreeSize(); |
| 12755 | }); |
| 12756 | } |
| 12757 | |
| 12758 | VkResult VmaBlockVector::AllocateFromBlock( |
| 12759 | VmaDeviceMemoryBlock* pBlock, |
| 12760 | VkDeviceSize size, |
| 12761 | VkDeviceSize alignment, |
| 12762 | VmaAllocationCreateFlags allocFlags, |
| 12763 | void* pUserData, |
| 12764 | VmaSuballocationType suballocType, |
| 12765 | uint32_t strategy, |
| 12766 | VmaAllocation* pAllocation) |
| 12767 | { |
| 12768 | const bool isUpperAddress = (allocFlags & VMA_ALLOCATION_CREATE_UPPER_ADDRESS_BIT) != 0; |
| 12769 | |
| 12770 | VmaAllocationRequest currRequest = {}; |
| 12771 | if (pBlock->m_pMetadata->CreateAllocationRequest( |
| 12772 | allocSize: size, |
| 12773 | allocAlignment: alignment, |
| 12774 | upperAddress: isUpperAddress, |
| 12775 | allocType: suballocType, |
| 12776 | strategy, |
| 12777 | pAllocationRequest: &currRequest)) |
| 12778 | { |
| 12779 | return CommitAllocationRequest(allocRequest&: currRequest, pBlock, alignment, allocFlags, pUserData, suballocType, pAllocation); |
| 12780 | } |
| 12781 | return VK_ERROR_OUT_OF_DEVICE_MEMORY; |
| 12782 | } |
| 12783 | |
| 12784 | VkResult VmaBlockVector::CommitAllocationRequest( |
| 12785 | VmaAllocationRequest& allocRequest, |
| 12786 | VmaDeviceMemoryBlock* pBlock, |
| 12787 | VkDeviceSize alignment, |
| 12788 | VmaAllocationCreateFlags allocFlags, |
| 12789 | void* pUserData, |
| 12790 | VmaSuballocationType suballocType, |
| 12791 | VmaAllocation* pAllocation) |
| 12792 | { |
| 12793 | const bool mapped = (allocFlags & VMA_ALLOCATION_CREATE_MAPPED_BIT) != 0; |
| 12794 | const bool isUserDataString = (allocFlags & VMA_ALLOCATION_CREATE_USER_DATA_COPY_STRING_BIT) != 0; |
| 12795 | const bool isMappingAllowed = (allocFlags & |
| 12796 | (VMA_ALLOCATION_CREATE_HOST_ACCESS_SEQUENTIAL_WRITE_BIT | VMA_ALLOCATION_CREATE_HOST_ACCESS_RANDOM_BIT)) != 0; |
| 12797 | |
| 12798 | pBlock->PostAlloc(); |
| 12799 | // Allocate from pCurrBlock. |
| 12800 | if (mapped) |
| 12801 | { |
| 12802 | VkResult res = pBlock->Map(hAllocator: m_hAllocator, count: 1, VMA_NULL); |
| 12803 | if (res != VK_SUCCESS) |
| 12804 | { |
| 12805 | return res; |
| 12806 | } |
| 12807 | } |
| 12808 | |
| 12809 | *pAllocation = m_hAllocator->m_AllocationObjectAllocator.Allocate(args: isMappingAllowed); |
| 12810 | pBlock->m_pMetadata->Alloc(request: allocRequest, type: suballocType, userData: *pAllocation); |
| 12811 | (*pAllocation)->InitBlockAllocation( |
| 12812 | block: pBlock, |
| 12813 | allocHandle: allocRequest.allocHandle, |
| 12814 | alignment, |
| 12815 | size: allocRequest.size, // Not size, as actual allocation size may be larger than requested! |
| 12816 | memoryTypeIndex: m_MemoryTypeIndex, |
| 12817 | suballocationType: suballocType, |
| 12818 | mapped); |
| 12819 | VMA_HEAVY_ASSERT(pBlock->Validate()); |
| 12820 | if (isUserDataString) |
| 12821 | (*pAllocation)->SetName(hAllocator: m_hAllocator, pName: (const char*)pUserData); |
| 12822 | else |
| 12823 | (*pAllocation)->SetUserData(hAllocator: m_hAllocator, pUserData); |
| 12824 | m_hAllocator->m_Budget.AddAllocation(heapIndex: m_hAllocator->MemoryTypeIndexToHeapIndex(memTypeIndex: m_MemoryTypeIndex), allocationSize: allocRequest.size); |
| 12825 | if (VMA_DEBUG_INITIALIZE_ALLOCATIONS) |
| 12826 | { |
| 12827 | m_hAllocator->FillAllocation(hAllocation: *pAllocation, pattern: VMA_ALLOCATION_FILL_PATTERN_CREATED); |
| 12828 | } |
| 12829 | if (IsCorruptionDetectionEnabled()) |
| 12830 | { |
| 12831 | VkResult res = pBlock->WriteMagicValueAfterAllocation(hAllocator: m_hAllocator, allocOffset: (*pAllocation)->GetOffset(), allocSize: allocRequest.size); |
| 12832 | VMA_ASSERT(res == VK_SUCCESS && "Couldn't map block memory to write magic value." ); |
| 12833 | } |
| 12834 | return VK_SUCCESS; |
| 12835 | } |
| 12836 | |
| 12837 | VkResult VmaBlockVector::CreateBlock(VkDeviceSize blockSize, size_t* pNewBlockIndex) |
| 12838 | { |
| 12839 | VkMemoryAllocateInfo allocInfo = { .sType: VK_STRUCTURE_TYPE_MEMORY_ALLOCATE_INFO }; |
| 12840 | allocInfo.pNext = m_pMemoryAllocateNext; |
| 12841 | allocInfo.memoryTypeIndex = m_MemoryTypeIndex; |
| 12842 | allocInfo.allocationSize = blockSize; |
| 12843 | |
| 12844 | #if VMA_BUFFER_DEVICE_ADDRESS |
| 12845 | // Every standalone block can potentially contain a buffer with VK_BUFFER_USAGE_SHADER_DEVICE_ADDRESS_BIT - always enable the feature. |
| 12846 | VkMemoryAllocateFlagsInfoKHR allocFlagsInfo = { .sType: VK_STRUCTURE_TYPE_MEMORY_ALLOCATE_FLAGS_INFO_KHR }; |
| 12847 | if (m_hAllocator->m_UseKhrBufferDeviceAddress) |
| 12848 | { |
| 12849 | allocFlagsInfo.flags = VK_MEMORY_ALLOCATE_DEVICE_ADDRESS_BIT_KHR; |
| 12850 | VmaPnextChainPushFront(mainStruct: &allocInfo, newStruct: &allocFlagsInfo); |
| 12851 | } |
| 12852 | #endif // VMA_BUFFER_DEVICE_ADDRESS |
| 12853 | |
| 12854 | #if VMA_MEMORY_PRIORITY |
| 12855 | VkMemoryPriorityAllocateInfoEXT priorityInfo = { .sType: VK_STRUCTURE_TYPE_MEMORY_PRIORITY_ALLOCATE_INFO_EXT }; |
| 12856 | if (m_hAllocator->m_UseExtMemoryPriority) |
| 12857 | { |
| 12858 | VMA_ASSERT(m_Priority >= 0.f && m_Priority <= 1.f); |
| 12859 | priorityInfo.priority = m_Priority; |
| 12860 | VmaPnextChainPushFront(mainStruct: &allocInfo, newStruct: &priorityInfo); |
| 12861 | } |
| 12862 | #endif // VMA_MEMORY_PRIORITY |
| 12863 | |
| 12864 | #if VMA_EXTERNAL_MEMORY |
| 12865 | // Attach VkExportMemoryAllocateInfoKHR if necessary. |
| 12866 | VkExportMemoryAllocateInfoKHR exportMemoryAllocInfo = { .sType: VK_STRUCTURE_TYPE_EXPORT_MEMORY_ALLOCATE_INFO_KHR }; |
| 12867 | exportMemoryAllocInfo.handleTypes = m_hAllocator->GetExternalMemoryHandleTypeFlags(memTypeIndex: m_MemoryTypeIndex); |
| 12868 | if (exportMemoryAllocInfo.handleTypes != 0) |
| 12869 | { |
| 12870 | VmaPnextChainPushFront(mainStruct: &allocInfo, newStruct: &exportMemoryAllocInfo); |
| 12871 | } |
| 12872 | #endif // VMA_EXTERNAL_MEMORY |
| 12873 | |
| 12874 | VkDeviceMemory mem = VK_NULL_HANDLE; |
| 12875 | VkResult res = m_hAllocator->AllocateVulkanMemory(pAllocateInfo: &allocInfo, pMemory: &mem); |
| 12876 | if (res < 0) |
| 12877 | { |
| 12878 | return res; |
| 12879 | } |
| 12880 | |
| 12881 | // New VkDeviceMemory successfully created. |
| 12882 | |
| 12883 | // Create new Allocation for it. |
| 12884 | VmaDeviceMemoryBlock* const pBlock = vma_new(m_hAllocator, VmaDeviceMemoryBlock)(m_hAllocator); |
| 12885 | pBlock->Init( |
| 12886 | hAllocator: m_hAllocator, |
| 12887 | hParentPool: m_hParentPool, |
| 12888 | newMemoryTypeIndex: m_MemoryTypeIndex, |
| 12889 | newMemory: mem, |
| 12890 | newSize: allocInfo.allocationSize, |
| 12891 | id: m_NextBlockId++, |
| 12892 | algorithm: m_Algorithm, |
| 12893 | bufferImageGranularity: m_BufferImageGranularity); |
| 12894 | |
| 12895 | m_Blocks.push_back(src: pBlock); |
| 12896 | if (pNewBlockIndex != VMA_NULL) |
| 12897 | { |
| 12898 | *pNewBlockIndex = m_Blocks.size() - 1; |
| 12899 | } |
| 12900 | |
| 12901 | return VK_SUCCESS; |
| 12902 | } |
| 12903 | |
| 12904 | bool VmaBlockVector::HasEmptyBlock() |
| 12905 | { |
| 12906 | for (size_t index = 0, count = m_Blocks.size(); index < count; ++index) |
| 12907 | { |
| 12908 | VmaDeviceMemoryBlock* const pBlock = m_Blocks[index]; |
| 12909 | if (pBlock->m_pMetadata->IsEmpty()) |
| 12910 | { |
| 12911 | return true; |
| 12912 | } |
| 12913 | } |
| 12914 | return false; |
| 12915 | } |
| 12916 | |
| 12917 | #if VMA_STATS_STRING_ENABLED |
| 12918 | void VmaBlockVector::PrintDetailedMap(class VmaJsonWriter& json) |
| 12919 | { |
| 12920 | VmaMutexLockRead lock(m_Mutex, m_hAllocator->m_UseMutex); |
| 12921 | |
| 12922 | |
| 12923 | json.BeginObject(); |
| 12924 | for (size_t i = 0; i < m_Blocks.size(); ++i) |
| 12925 | { |
| 12926 | json.BeginString(); |
| 12927 | json.ContinueString(n: m_Blocks[i]->GetId()); |
| 12928 | json.EndString(); |
| 12929 | |
| 12930 | json.BeginObject(); |
| 12931 | json.WriteString(pStr: "MapRefCount" ); |
| 12932 | json.WriteNumber(n: m_Blocks[i]->GetMapRefCount()); |
| 12933 | |
| 12934 | m_Blocks[i]->m_pMetadata->PrintDetailedMap(json); |
| 12935 | json.EndObject(); |
| 12936 | } |
| 12937 | json.EndObject(); |
| 12938 | } |
| 12939 | #endif // VMA_STATS_STRING_ENABLED |
| 12940 | |
| 12941 | VkResult VmaBlockVector::CheckCorruption() |
| 12942 | { |
| 12943 | if (!IsCorruptionDetectionEnabled()) |
| 12944 | { |
| 12945 | return VK_ERROR_FEATURE_NOT_PRESENT; |
| 12946 | } |
| 12947 | |
| 12948 | VmaMutexLockRead lock(m_Mutex, m_hAllocator->m_UseMutex); |
| 12949 | for (uint32_t blockIndex = 0; blockIndex < m_Blocks.size(); ++blockIndex) |
| 12950 | { |
| 12951 | VmaDeviceMemoryBlock* const pBlock = m_Blocks[blockIndex]; |
| 12952 | VMA_ASSERT(pBlock); |
| 12953 | VkResult res = pBlock->CheckCorruption(hAllocator: m_hAllocator); |
| 12954 | if (res != VK_SUCCESS) |
| 12955 | { |
| 12956 | return res; |
| 12957 | } |
| 12958 | } |
| 12959 | return VK_SUCCESS; |
| 12960 | } |
| 12961 | |
| 12962 | #endif // _VMA_BLOCK_VECTOR_FUNCTIONS |
| 12963 | |
| 12964 | #ifndef _VMA_DEFRAGMENTATION_CONTEXT_FUNCTIONS |
| 12965 | VmaDefragmentationContext_T::VmaDefragmentationContext_T( |
| 12966 | VmaAllocator hAllocator, |
| 12967 | const VmaDefragmentationInfo& info) |
| 12968 | : m_MaxPassBytes(info.maxBytesPerPass == 0 ? VK_WHOLE_SIZE : info.maxBytesPerPass), |
| 12969 | m_MaxPassAllocations(info.maxAllocationsPerPass == 0 ? UINT32_MAX : info.maxAllocationsPerPass), |
| 12970 | m_MoveAllocator(hAllocator->GetAllocationCallbacks()), |
| 12971 | m_Moves(m_MoveAllocator) |
| 12972 | { |
| 12973 | m_Algorithm = info.flags & VMA_DEFRAGMENTATION_FLAG_ALGORITHM_MASK; |
| 12974 | |
| 12975 | if (info.pool != VMA_NULL) |
| 12976 | { |
| 12977 | m_BlockVectorCount = 1; |
| 12978 | m_PoolBlockVector = &info.pool->m_BlockVector; |
| 12979 | m_pBlockVectors = &m_PoolBlockVector; |
| 12980 | m_PoolBlockVector->SetIncrementalSort(false); |
| 12981 | m_PoolBlockVector->SortByFreeSize(); |
| 12982 | } |
| 12983 | else |
| 12984 | { |
| 12985 | m_BlockVectorCount = hAllocator->GetMemoryTypeCount(); |
| 12986 | m_PoolBlockVector = VMA_NULL; |
| 12987 | m_pBlockVectors = hAllocator->m_pBlockVectors; |
| 12988 | for (uint32_t i = 0; i < m_BlockVectorCount; ++i) |
| 12989 | { |
| 12990 | VmaBlockVector* vector = m_pBlockVectors[i]; |
| 12991 | if (vector != VMA_NULL) |
| 12992 | { |
| 12993 | vector->SetIncrementalSort(false); |
| 12994 | vector->SortByFreeSize(); |
| 12995 | } |
| 12996 | } |
| 12997 | } |
| 12998 | |
| 12999 | switch (m_Algorithm) |
| 13000 | { |
| 13001 | case 0: // Default algorithm |
| 13002 | m_Algorithm = VMA_DEFRAGMENTATION_FLAG_ALGORITHM_BALANCED_BIT; |
| 13003 | case VMA_DEFRAGMENTATION_FLAG_ALGORITHM_BALANCED_BIT: |
| 13004 | { |
| 13005 | m_AlgorithmState = vma_new_array(hAllocator, StateBalanced, m_BlockVectorCount); |
| 13006 | break; |
| 13007 | } |
| 13008 | case VMA_DEFRAGMENTATION_FLAG_ALGORITHM_EXTENSIVE_BIT: |
| 13009 | { |
| 13010 | if (hAllocator->GetBufferImageGranularity() > 1) |
| 13011 | { |
| 13012 | m_AlgorithmState = vma_new_array(hAllocator, StateExtensive, m_BlockVectorCount); |
| 13013 | } |
| 13014 | break; |
| 13015 | } |
| 13016 | } |
| 13017 | } |
| 13018 | |
| 13019 | VmaDefragmentationContext_T::~VmaDefragmentationContext_T() |
| 13020 | { |
| 13021 | if (m_PoolBlockVector != VMA_NULL) |
| 13022 | { |
| 13023 | m_PoolBlockVector->SetIncrementalSort(true); |
| 13024 | } |
| 13025 | else |
| 13026 | { |
| 13027 | for (uint32_t i = 0; i < m_BlockVectorCount; ++i) |
| 13028 | { |
| 13029 | VmaBlockVector* vector = m_pBlockVectors[i]; |
| 13030 | if (vector != VMA_NULL) |
| 13031 | vector->SetIncrementalSort(true); |
| 13032 | } |
| 13033 | } |
| 13034 | |
| 13035 | if (m_AlgorithmState) |
| 13036 | { |
| 13037 | switch (m_Algorithm) |
| 13038 | { |
| 13039 | case VMA_DEFRAGMENTATION_FLAG_ALGORITHM_BALANCED_BIT: |
| 13040 | vma_delete_array(pAllocationCallbacks: m_MoveAllocator.m_pCallbacks, ptr: reinterpret_cast<StateBalanced*>(m_AlgorithmState), count: m_BlockVectorCount); |
| 13041 | break; |
| 13042 | case VMA_DEFRAGMENTATION_FLAG_ALGORITHM_EXTENSIVE_BIT: |
| 13043 | vma_delete_array(pAllocationCallbacks: m_MoveAllocator.m_pCallbacks, ptr: reinterpret_cast<StateExtensive*>(m_AlgorithmState), count: m_BlockVectorCount); |
| 13044 | break; |
| 13045 | default: |
| 13046 | VMA_ASSERT(0); |
| 13047 | } |
| 13048 | } |
| 13049 | } |
| 13050 | |
| 13051 | VkResult VmaDefragmentationContext_T::DefragmentPassBegin(VmaDefragmentationPassMoveInfo& moveInfo) |
| 13052 | { |
| 13053 | if (m_PoolBlockVector != VMA_NULL) |
| 13054 | { |
| 13055 | VmaMutexLockWrite lock(m_PoolBlockVector->GetMutex(), m_PoolBlockVector->GetAllocator()->m_UseMutex); |
| 13056 | |
| 13057 | if (m_PoolBlockVector->GetBlockCount() > 1) |
| 13058 | ComputeDefragmentation(vector&: *m_PoolBlockVector, index: 0); |
| 13059 | else if (m_PoolBlockVector->GetBlockCount() == 1) |
| 13060 | ReallocWithinBlock(vector&: *m_PoolBlockVector, block: m_PoolBlockVector->GetBlock(index: 0)); |
| 13061 | } |
| 13062 | else |
| 13063 | { |
| 13064 | for (uint32_t i = 0; i < m_BlockVectorCount; ++i) |
| 13065 | { |
| 13066 | if (m_pBlockVectors[i] != VMA_NULL) |
| 13067 | { |
| 13068 | VmaMutexLockWrite lock(m_pBlockVectors[i]->GetMutex(), m_pBlockVectors[i]->GetAllocator()->m_UseMutex); |
| 13069 | |
| 13070 | if (m_pBlockVectors[i]->GetBlockCount() > 1) |
| 13071 | { |
| 13072 | if (ComputeDefragmentation(vector&: *m_pBlockVectors[i], index: i)) |
| 13073 | break; |
| 13074 | } |
| 13075 | else if (m_pBlockVectors[i]->GetBlockCount() == 1) |
| 13076 | { |
| 13077 | if (ReallocWithinBlock(vector&: *m_pBlockVectors[i], block: m_pBlockVectors[i]->GetBlock(index: 0))) |
| 13078 | break; |
| 13079 | } |
| 13080 | } |
| 13081 | } |
| 13082 | } |
| 13083 | |
| 13084 | moveInfo.moveCount = static_cast<uint32_t>(m_Moves.size()); |
| 13085 | if (moveInfo.moveCount > 0) |
| 13086 | { |
| 13087 | moveInfo.pMoves = m_Moves.data(); |
| 13088 | return VK_INCOMPLETE; |
| 13089 | } |
| 13090 | |
| 13091 | moveInfo.pMoves = VMA_NULL; |
| 13092 | return VK_SUCCESS; |
| 13093 | } |
| 13094 | |
| 13095 | VkResult VmaDefragmentationContext_T::DefragmentPassEnd(VmaDefragmentationPassMoveInfo& moveInfo) |
| 13096 | { |
| 13097 | VMA_ASSERT(moveInfo.moveCount > 0 ? moveInfo.pMoves != VMA_NULL : true); |
| 13098 | |
| 13099 | VkResult result = VK_SUCCESS; |
| 13100 | VmaStlAllocator<FragmentedBlock> blockAllocator(m_MoveAllocator.m_pCallbacks); |
| 13101 | VmaVector<FragmentedBlock, VmaStlAllocator<FragmentedBlock>> immovableBlocks(blockAllocator); |
| 13102 | VmaVector<FragmentedBlock, VmaStlAllocator<FragmentedBlock>> mappedBlocks(blockAllocator); |
| 13103 | |
| 13104 | VmaAllocator allocator = VMA_NULL; |
| 13105 | for (uint32_t i = 0; i < moveInfo.moveCount; ++i) |
| 13106 | { |
| 13107 | VmaDefragmentationMove& move = moveInfo.pMoves[i]; |
| 13108 | size_t prevCount = 0, currentCount = 0; |
| 13109 | VkDeviceSize freedBlockSize = 0; |
| 13110 | |
| 13111 | uint32_t vectorIndex; |
| 13112 | VmaBlockVector* vector; |
| 13113 | if (m_PoolBlockVector != VMA_NULL) |
| 13114 | { |
| 13115 | vectorIndex = 0; |
| 13116 | vector = m_PoolBlockVector; |
| 13117 | } |
| 13118 | else |
| 13119 | { |
| 13120 | vectorIndex = move.srcAllocation->GetMemoryTypeIndex(); |
| 13121 | vector = m_pBlockVectors[vectorIndex]; |
| 13122 | VMA_ASSERT(vector != VMA_NULL); |
| 13123 | } |
| 13124 | |
| 13125 | switch (move.operation) |
| 13126 | { |
| 13127 | case VMA_DEFRAGMENTATION_MOVE_OPERATION_COPY: |
| 13128 | { |
| 13129 | uint8_t mapCount = move.srcAllocation->SwapBlockAllocation(hAllocator: vector->m_hAllocator, allocation: move.dstTmpAllocation); |
| 13130 | if (mapCount > 0) |
| 13131 | { |
| 13132 | allocator = vector->m_hAllocator; |
| 13133 | VmaDeviceMemoryBlock* newMapBlock = move.srcAllocation->GetBlock(); |
| 13134 | bool notPresent = true; |
| 13135 | for (FragmentedBlock& block : mappedBlocks) |
| 13136 | { |
| 13137 | if (block.block == newMapBlock) |
| 13138 | { |
| 13139 | notPresent = false; |
| 13140 | block.data += mapCount; |
| 13141 | break; |
| 13142 | } |
| 13143 | } |
| 13144 | if (notPresent) |
| 13145 | mappedBlocks.push_back(src: { .data: mapCount, .block: newMapBlock }); |
| 13146 | } |
| 13147 | |
| 13148 | // Scope for locks, Free have it's own lock |
| 13149 | { |
| 13150 | VmaMutexLockRead lock(vector->GetMutex(), vector->GetAllocator()->m_UseMutex); |
| 13151 | prevCount = vector->GetBlockCount(); |
| 13152 | freedBlockSize = move.dstTmpAllocation->GetBlock()->m_pMetadata->GetSize(); |
| 13153 | } |
| 13154 | vector->Free(hAllocation: move.dstTmpAllocation); |
| 13155 | { |
| 13156 | VmaMutexLockRead lock(vector->GetMutex(), vector->GetAllocator()->m_UseMutex); |
| 13157 | currentCount = vector->GetBlockCount(); |
| 13158 | } |
| 13159 | |
| 13160 | result = VK_INCOMPLETE; |
| 13161 | break; |
| 13162 | } |
| 13163 | case VMA_DEFRAGMENTATION_MOVE_OPERATION_IGNORE: |
| 13164 | { |
| 13165 | m_PassStats.bytesMoved -= move.srcAllocation->GetSize(); |
| 13166 | --m_PassStats.allocationsMoved; |
| 13167 | vector->Free(hAllocation: move.dstTmpAllocation); |
| 13168 | |
| 13169 | VmaDeviceMemoryBlock* newBlock = move.srcAllocation->GetBlock(); |
| 13170 | bool notPresent = true; |
| 13171 | for (const FragmentedBlock& block : immovableBlocks) |
| 13172 | { |
| 13173 | if (block.block == newBlock) |
| 13174 | { |
| 13175 | notPresent = false; |
| 13176 | break; |
| 13177 | } |
| 13178 | } |
| 13179 | if (notPresent) |
| 13180 | immovableBlocks.push_back(src: { .data: vectorIndex, .block: newBlock }); |
| 13181 | break; |
| 13182 | } |
| 13183 | case VMA_DEFRAGMENTATION_MOVE_OPERATION_DESTROY: |
| 13184 | { |
| 13185 | m_PassStats.bytesMoved -= move.srcAllocation->GetSize(); |
| 13186 | --m_PassStats.allocationsMoved; |
| 13187 | // Scope for locks, Free have it's own lock |
| 13188 | { |
| 13189 | VmaMutexLockRead lock(vector->GetMutex(), vector->GetAllocator()->m_UseMutex); |
| 13190 | prevCount = vector->GetBlockCount(); |
| 13191 | freedBlockSize = move.srcAllocation->GetBlock()->m_pMetadata->GetSize(); |
| 13192 | } |
| 13193 | vector->Free(hAllocation: move.srcAllocation); |
| 13194 | { |
| 13195 | VmaMutexLockRead lock(vector->GetMutex(), vector->GetAllocator()->m_UseMutex); |
| 13196 | currentCount = vector->GetBlockCount(); |
| 13197 | } |
| 13198 | freedBlockSize *= prevCount - currentCount; |
| 13199 | |
| 13200 | VkDeviceSize dstBlockSize; |
| 13201 | { |
| 13202 | VmaMutexLockRead lock(vector->GetMutex(), vector->GetAllocator()->m_UseMutex); |
| 13203 | dstBlockSize = move.dstTmpAllocation->GetBlock()->m_pMetadata->GetSize(); |
| 13204 | } |
| 13205 | vector->Free(hAllocation: move.dstTmpAllocation); |
| 13206 | { |
| 13207 | VmaMutexLockRead lock(vector->GetMutex(), vector->GetAllocator()->m_UseMutex); |
| 13208 | freedBlockSize += dstBlockSize * (currentCount - vector->GetBlockCount()); |
| 13209 | currentCount = vector->GetBlockCount(); |
| 13210 | } |
| 13211 | |
| 13212 | result = VK_INCOMPLETE; |
| 13213 | break; |
| 13214 | } |
| 13215 | default: |
| 13216 | VMA_ASSERT(0); |
| 13217 | } |
| 13218 | |
| 13219 | if (prevCount > currentCount) |
| 13220 | { |
| 13221 | size_t freedBlocks = prevCount - currentCount; |
| 13222 | m_PassStats.deviceMemoryBlocksFreed += static_cast<uint32_t>(freedBlocks); |
| 13223 | m_PassStats.bytesFreed += freedBlockSize; |
| 13224 | } |
| 13225 | |
| 13226 | switch (m_Algorithm) |
| 13227 | { |
| 13228 | case VMA_DEFRAGMENTATION_FLAG_ALGORITHM_EXTENSIVE_BIT: |
| 13229 | { |
| 13230 | if (m_AlgorithmState != VMA_NULL) |
| 13231 | { |
| 13232 | // Avoid unnecessary tries to allocate when new free block is avaiable |
| 13233 | StateExtensive& state = reinterpret_cast<StateExtensive*>(m_AlgorithmState)[vectorIndex]; |
| 13234 | if (state.firstFreeBlock != SIZE_MAX) |
| 13235 | { |
| 13236 | const size_t diff = prevCount - currentCount; |
| 13237 | if (state.firstFreeBlock >= diff) |
| 13238 | { |
| 13239 | state.firstFreeBlock -= diff; |
| 13240 | if (state.firstFreeBlock != 0) |
| 13241 | state.firstFreeBlock -= vector->GetBlock(index: state.firstFreeBlock - 1)->m_pMetadata->IsEmpty(); |
| 13242 | } |
| 13243 | else |
| 13244 | state.firstFreeBlock = 0; |
| 13245 | } |
| 13246 | } |
| 13247 | } |
| 13248 | } |
| 13249 | } |
| 13250 | moveInfo.moveCount = 0; |
| 13251 | moveInfo.pMoves = VMA_NULL; |
| 13252 | m_Moves.clear(); |
| 13253 | |
| 13254 | // Update stats |
| 13255 | m_GlobalStats.allocationsMoved += m_PassStats.allocationsMoved; |
| 13256 | m_GlobalStats.bytesFreed += m_PassStats.bytesFreed; |
| 13257 | m_GlobalStats.bytesMoved += m_PassStats.bytesMoved; |
| 13258 | m_GlobalStats.deviceMemoryBlocksFreed += m_PassStats.deviceMemoryBlocksFreed; |
| 13259 | m_PassStats = { .bytesMoved: 0 }; |
| 13260 | |
| 13261 | // Move blocks with immovable allocations according to algorithm |
| 13262 | if (immovableBlocks.size() > 0) |
| 13263 | { |
| 13264 | switch (m_Algorithm) |
| 13265 | { |
| 13266 | case VMA_DEFRAGMENTATION_FLAG_ALGORITHM_EXTENSIVE_BIT: |
| 13267 | { |
| 13268 | if (m_AlgorithmState != VMA_NULL) |
| 13269 | { |
| 13270 | bool swapped = false; |
| 13271 | // Move to the start of free blocks range |
| 13272 | for (const FragmentedBlock& block : immovableBlocks) |
| 13273 | { |
| 13274 | StateExtensive& state = reinterpret_cast<StateExtensive*>(m_AlgorithmState)[block.data]; |
| 13275 | if (state.operation != StateExtensive::Operation::Cleanup) |
| 13276 | { |
| 13277 | VmaBlockVector* vector = m_pBlockVectors[block.data]; |
| 13278 | VmaMutexLockWrite lock(vector->GetMutex(), vector->GetAllocator()->m_UseMutex); |
| 13279 | |
| 13280 | for (size_t i = 0, count = vector->GetBlockCount() - m_ImmovableBlockCount; i < count; ++i) |
| 13281 | { |
| 13282 | if (vector->GetBlock(index: i) == block.block) |
| 13283 | { |
| 13284 | VMA_SWAP(vector->m_Blocks[i], vector->m_Blocks[vector->GetBlockCount() - ++m_ImmovableBlockCount]); |
| 13285 | if (state.firstFreeBlock != SIZE_MAX) |
| 13286 | { |
| 13287 | if (i + 1 < state.firstFreeBlock) |
| 13288 | { |
| 13289 | if (state.firstFreeBlock > 1) |
| 13290 | VMA_SWAP(vector->m_Blocks[i], vector->m_Blocks[--state.firstFreeBlock]); |
| 13291 | else |
| 13292 | --state.firstFreeBlock; |
| 13293 | } |
| 13294 | } |
| 13295 | swapped = true; |
| 13296 | break; |
| 13297 | } |
| 13298 | } |
| 13299 | } |
| 13300 | } |
| 13301 | if (swapped) |
| 13302 | result = VK_INCOMPLETE; |
| 13303 | break; |
| 13304 | } |
| 13305 | } |
| 13306 | default: |
| 13307 | { |
| 13308 | // Move to the begining |
| 13309 | for (const FragmentedBlock& block : immovableBlocks) |
| 13310 | { |
| 13311 | VmaBlockVector* vector = m_pBlockVectors[block.data]; |
| 13312 | VmaMutexLockWrite lock(vector->GetMutex(), vector->GetAllocator()->m_UseMutex); |
| 13313 | |
| 13314 | for (size_t i = m_ImmovableBlockCount; i < vector->GetBlockCount(); ++i) |
| 13315 | { |
| 13316 | if (vector->GetBlock(index: i) == block.block) |
| 13317 | { |
| 13318 | VMA_SWAP(vector->m_Blocks[i], vector->m_Blocks[m_ImmovableBlockCount++]); |
| 13319 | break; |
| 13320 | } |
| 13321 | } |
| 13322 | } |
| 13323 | break; |
| 13324 | } |
| 13325 | } |
| 13326 | } |
| 13327 | |
| 13328 | // Bulk-map destination blocks |
| 13329 | for (const FragmentedBlock& block : mappedBlocks) |
| 13330 | { |
| 13331 | VkResult res = block.block->Map(hAllocator: allocator, count: block.data, VMA_NULL); |
| 13332 | VMA_ASSERT(res == VK_SUCCESS); |
| 13333 | } |
| 13334 | return result; |
| 13335 | } |
| 13336 | |
| 13337 | bool VmaDefragmentationContext_T::ComputeDefragmentation(VmaBlockVector& vector, size_t index) |
| 13338 | { |
| 13339 | switch (m_Algorithm) |
| 13340 | { |
| 13341 | case VMA_DEFRAGMENTATION_FLAG_ALGORITHM_FAST_BIT: |
| 13342 | return ComputeDefragmentation_Fast(vector); |
| 13343 | default: |
| 13344 | VMA_ASSERT(0); |
| 13345 | case VMA_DEFRAGMENTATION_FLAG_ALGORITHM_BALANCED_BIT: |
| 13346 | return ComputeDefragmentation_Balanced(vector, index, update: true); |
| 13347 | case VMA_DEFRAGMENTATION_FLAG_ALGORITHM_FULL_BIT: |
| 13348 | return ComputeDefragmentation_Full(vector); |
| 13349 | case VMA_DEFRAGMENTATION_FLAG_ALGORITHM_EXTENSIVE_BIT: |
| 13350 | return ComputeDefragmentation_Extensive(vector, index); |
| 13351 | } |
| 13352 | } |
| 13353 | |
| 13354 | VmaDefragmentationContext_T::MoveAllocationData VmaDefragmentationContext_T::GetMoveData( |
| 13355 | VmaAllocHandle handle, VmaBlockMetadata* metadata) |
| 13356 | { |
| 13357 | MoveAllocationData moveData; |
| 13358 | moveData.move.srcAllocation = (VmaAllocation)metadata->GetAllocationUserData(allocHandle: handle); |
| 13359 | moveData.size = moveData.move.srcAllocation->GetSize(); |
| 13360 | moveData.alignment = moveData.move.srcAllocation->GetAlignment(); |
| 13361 | moveData.type = moveData.move.srcAllocation->GetSuballocationType(); |
| 13362 | moveData.flags = 0; |
| 13363 | |
| 13364 | if (moveData.move.srcAllocation->IsPersistentMap()) |
| 13365 | moveData.flags |= VMA_ALLOCATION_CREATE_MAPPED_BIT; |
| 13366 | if (moveData.move.srcAllocation->IsMappingAllowed()) |
| 13367 | moveData.flags |= VMA_ALLOCATION_CREATE_HOST_ACCESS_SEQUENTIAL_WRITE_BIT | VMA_ALLOCATION_CREATE_HOST_ACCESS_RANDOM_BIT; |
| 13368 | |
| 13369 | return moveData; |
| 13370 | } |
| 13371 | |
| 13372 | VmaDefragmentationContext_T::CounterStatus VmaDefragmentationContext_T::CheckCounters(VkDeviceSize bytes) |
| 13373 | { |
| 13374 | // Ignore allocation if will exceed max size for copy |
| 13375 | if (m_PassStats.bytesMoved + bytes > m_MaxPassBytes) |
| 13376 | { |
| 13377 | if (++m_IgnoredAllocs < MAX_ALLOCS_TO_IGNORE) |
| 13378 | return CounterStatus::Ignore; |
| 13379 | else |
| 13380 | return CounterStatus::End; |
| 13381 | } |
| 13382 | return CounterStatus::Pass; |
| 13383 | } |
| 13384 | |
| 13385 | bool VmaDefragmentationContext_T::IncrementCounters(VkDeviceSize bytes) |
| 13386 | { |
| 13387 | m_PassStats.bytesMoved += bytes; |
| 13388 | // Early return when max found |
| 13389 | if (++m_PassStats.allocationsMoved >= m_MaxPassAllocations || m_PassStats.bytesMoved >= m_MaxPassBytes) |
| 13390 | { |
| 13391 | VMA_ASSERT(m_PassStats.allocationsMoved == m_MaxPassAllocations || |
| 13392 | m_PassStats.bytesMoved == m_MaxPassBytes && "Exceeded maximal pass threshold!" ); |
| 13393 | return true; |
| 13394 | } |
| 13395 | return false; |
| 13396 | } |
| 13397 | |
| 13398 | bool VmaDefragmentationContext_T::ReallocWithinBlock(VmaBlockVector& vector, VmaDeviceMemoryBlock* block) |
| 13399 | { |
| 13400 | VmaBlockMetadata* metadata = block->m_pMetadata; |
| 13401 | |
| 13402 | for (VmaAllocHandle handle = metadata->GetAllocationListBegin(); |
| 13403 | handle != VK_NULL_HANDLE; |
| 13404 | handle = metadata->GetNextAllocation(prevAlloc: handle)) |
| 13405 | { |
| 13406 | MoveAllocationData moveData = GetMoveData(handle, metadata); |
| 13407 | // Ignore newly created allocations by defragmentation algorithm |
| 13408 | if (moveData.move.srcAllocation->GetUserData() == this) |
| 13409 | continue; |
| 13410 | switch (CheckCounters(bytes: moveData.move.srcAllocation->GetSize())) |
| 13411 | { |
| 13412 | case CounterStatus::Ignore: |
| 13413 | continue; |
| 13414 | case CounterStatus::End: |
| 13415 | return true; |
| 13416 | default: |
| 13417 | VMA_ASSERT(0); |
| 13418 | case CounterStatus::Pass: |
| 13419 | break; |
| 13420 | } |
| 13421 | |
| 13422 | VkDeviceSize offset = moveData.move.srcAllocation->GetOffset(); |
| 13423 | if (offset != 0 && metadata->GetSumFreeSize() >= moveData.size) |
| 13424 | { |
| 13425 | VmaAllocationRequest request = {}; |
| 13426 | if (metadata->CreateAllocationRequest( |
| 13427 | allocSize: moveData.size, |
| 13428 | allocAlignment: moveData.alignment, |
| 13429 | upperAddress: false, |
| 13430 | allocType: moveData.type, |
| 13431 | strategy: VMA_ALLOCATION_CREATE_STRATEGY_MIN_OFFSET_BIT, |
| 13432 | pAllocationRequest: &request)) |
| 13433 | { |
| 13434 | if (metadata->GetAllocationOffset(allocHandle: request.allocHandle) < offset) |
| 13435 | { |
| 13436 | if (vector.CommitAllocationRequest( |
| 13437 | allocRequest&: request, |
| 13438 | pBlock: block, |
| 13439 | alignment: moveData.alignment, |
| 13440 | allocFlags: moveData.flags, |
| 13441 | pUserData: this, |
| 13442 | suballocType: moveData.type, |
| 13443 | pAllocation: &moveData.move.dstTmpAllocation) == VK_SUCCESS) |
| 13444 | { |
| 13445 | m_Moves.push_back(src: moveData.move); |
| 13446 | if (IncrementCounters(bytes: moveData.size)) |
| 13447 | return true; |
| 13448 | } |
| 13449 | } |
| 13450 | } |
| 13451 | } |
| 13452 | } |
| 13453 | return false; |
| 13454 | } |
| 13455 | |
| 13456 | bool VmaDefragmentationContext_T::AllocInOtherBlock(size_t start, size_t end, MoveAllocationData& data, VmaBlockVector& vector) |
| 13457 | { |
| 13458 | for (; start < end; ++start) |
| 13459 | { |
| 13460 | VmaDeviceMemoryBlock* dstBlock = vector.GetBlock(index: start); |
| 13461 | if (dstBlock->m_pMetadata->GetSumFreeSize() >= data.size) |
| 13462 | { |
| 13463 | if (vector.AllocateFromBlock(pBlock: dstBlock, |
| 13464 | size: data.size, |
| 13465 | alignment: data.alignment, |
| 13466 | allocFlags: data.flags, |
| 13467 | pUserData: this, |
| 13468 | suballocType: data.type, |
| 13469 | strategy: 0, |
| 13470 | pAllocation: &data.move.dstTmpAllocation) == VK_SUCCESS) |
| 13471 | { |
| 13472 | m_Moves.push_back(src: data.move); |
| 13473 | if (IncrementCounters(bytes: data.size)) |
| 13474 | return true; |
| 13475 | break; |
| 13476 | } |
| 13477 | } |
| 13478 | } |
| 13479 | return false; |
| 13480 | } |
| 13481 | |
| 13482 | bool VmaDefragmentationContext_T::ComputeDefragmentation_Fast(VmaBlockVector& vector) |
| 13483 | { |
| 13484 | // Move only between blocks |
| 13485 | |
| 13486 | // Go through allocations in last blocks and try to fit them inside first ones |
| 13487 | for (size_t i = vector.GetBlockCount() - 1; i > m_ImmovableBlockCount; --i) |
| 13488 | { |
| 13489 | VmaBlockMetadata* metadata = vector.GetBlock(index: i)->m_pMetadata; |
| 13490 | |
| 13491 | for (VmaAllocHandle handle = metadata->GetAllocationListBegin(); |
| 13492 | handle != VK_NULL_HANDLE; |
| 13493 | handle = metadata->GetNextAllocation(prevAlloc: handle)) |
| 13494 | { |
| 13495 | MoveAllocationData moveData = GetMoveData(handle, metadata); |
| 13496 | // Ignore newly created allocations by defragmentation algorithm |
| 13497 | if (moveData.move.srcAllocation->GetUserData() == this) |
| 13498 | continue; |
| 13499 | switch (CheckCounters(bytes: moveData.move.srcAllocation->GetSize())) |
| 13500 | { |
| 13501 | case CounterStatus::Ignore: |
| 13502 | continue; |
| 13503 | case CounterStatus::End: |
| 13504 | return true; |
| 13505 | default: |
| 13506 | VMA_ASSERT(0); |
| 13507 | case CounterStatus::Pass: |
| 13508 | break; |
| 13509 | } |
| 13510 | |
| 13511 | // Check all previous blocks for free space |
| 13512 | if (AllocInOtherBlock(start: 0, end: i, data&: moveData, vector)) |
| 13513 | return true; |
| 13514 | } |
| 13515 | } |
| 13516 | return false; |
| 13517 | } |
| 13518 | |
| 13519 | bool VmaDefragmentationContext_T::ComputeDefragmentation_Balanced(VmaBlockVector& vector, size_t index, bool update) |
| 13520 | { |
| 13521 | // Go over every allocation and try to fit it in previous blocks at lowest offsets, |
| 13522 | // if not possible: realloc within single block to minimize offset (exclude offset == 0), |
| 13523 | // but only if there are noticable gaps between them (some heuristic, ex. average size of allocation in block) |
| 13524 | VMA_ASSERT(m_AlgorithmState != VMA_NULL); |
| 13525 | |
| 13526 | StateBalanced& vectorState = reinterpret_cast<StateBalanced*>(m_AlgorithmState)[index]; |
| 13527 | if (update && vectorState.avgAllocSize == UINT64_MAX) |
| 13528 | UpdateVectorStatistics(vector, state&: vectorState); |
| 13529 | |
| 13530 | const size_t startMoveCount = m_Moves.size(); |
| 13531 | VkDeviceSize minimalFreeRegion = vectorState.avgFreeSize / 2; |
| 13532 | for (size_t i = vector.GetBlockCount() - 1; i > m_ImmovableBlockCount; --i) |
| 13533 | { |
| 13534 | VmaDeviceMemoryBlock* block = vector.GetBlock(index: i); |
| 13535 | VmaBlockMetadata* metadata = block->m_pMetadata; |
| 13536 | VkDeviceSize prevFreeRegionSize = 0; |
| 13537 | |
| 13538 | for (VmaAllocHandle handle = metadata->GetAllocationListBegin(); |
| 13539 | handle != VK_NULL_HANDLE; |
| 13540 | handle = metadata->GetNextAllocation(prevAlloc: handle)) |
| 13541 | { |
| 13542 | MoveAllocationData moveData = GetMoveData(handle, metadata); |
| 13543 | // Ignore newly created allocations by defragmentation algorithm |
| 13544 | if (moveData.move.srcAllocation->GetUserData() == this) |
| 13545 | continue; |
| 13546 | switch (CheckCounters(bytes: moveData.move.srcAllocation->GetSize())) |
| 13547 | { |
| 13548 | case CounterStatus::Ignore: |
| 13549 | continue; |
| 13550 | case CounterStatus::End: |
| 13551 | return true; |
| 13552 | default: |
| 13553 | VMA_ASSERT(0); |
| 13554 | case CounterStatus::Pass: |
| 13555 | break; |
| 13556 | } |
| 13557 | |
| 13558 | // Check all previous blocks for free space |
| 13559 | const size_t prevMoveCount = m_Moves.size(); |
| 13560 | if (AllocInOtherBlock(start: 0, end: i, data&: moveData, vector)) |
| 13561 | return true; |
| 13562 | |
| 13563 | VkDeviceSize nextFreeRegionSize = metadata->GetNextFreeRegionSize(alloc: handle); |
| 13564 | // If no room found then realloc within block for lower offset |
| 13565 | VkDeviceSize offset = moveData.move.srcAllocation->GetOffset(); |
| 13566 | if (prevMoveCount == m_Moves.size() && offset != 0 && metadata->GetSumFreeSize() >= moveData.size) |
| 13567 | { |
| 13568 | // Check if realloc will make sense |
| 13569 | if (prevFreeRegionSize >= minimalFreeRegion || |
| 13570 | nextFreeRegionSize >= minimalFreeRegion || |
| 13571 | moveData.size <= vectorState.avgFreeSize || |
| 13572 | moveData.size <= vectorState.avgAllocSize) |
| 13573 | { |
| 13574 | VmaAllocationRequest request = {}; |
| 13575 | if (metadata->CreateAllocationRequest( |
| 13576 | allocSize: moveData.size, |
| 13577 | allocAlignment: moveData.alignment, |
| 13578 | upperAddress: false, |
| 13579 | allocType: moveData.type, |
| 13580 | strategy: VMA_ALLOCATION_CREATE_STRATEGY_MIN_OFFSET_BIT, |
| 13581 | pAllocationRequest: &request)) |
| 13582 | { |
| 13583 | if (metadata->GetAllocationOffset(allocHandle: request.allocHandle) < offset) |
| 13584 | { |
| 13585 | if (vector.CommitAllocationRequest( |
| 13586 | allocRequest&: request, |
| 13587 | pBlock: block, |
| 13588 | alignment: moveData.alignment, |
| 13589 | allocFlags: moveData.flags, |
| 13590 | pUserData: this, |
| 13591 | suballocType: moveData.type, |
| 13592 | pAllocation: &moveData.move.dstTmpAllocation) == VK_SUCCESS) |
| 13593 | { |
| 13594 | m_Moves.push_back(src: moveData.move); |
| 13595 | if (IncrementCounters(bytes: moveData.size)) |
| 13596 | return true; |
| 13597 | } |
| 13598 | } |
| 13599 | } |
| 13600 | } |
| 13601 | } |
| 13602 | prevFreeRegionSize = nextFreeRegionSize; |
| 13603 | } |
| 13604 | } |
| 13605 | |
| 13606 | // No moves perfomed, update statistics to current vector state |
| 13607 | if (startMoveCount == m_Moves.size() && !update) |
| 13608 | { |
| 13609 | vectorState.avgAllocSize = UINT64_MAX; |
| 13610 | return ComputeDefragmentation_Balanced(vector, index, update: false); |
| 13611 | } |
| 13612 | return false; |
| 13613 | } |
| 13614 | |
| 13615 | bool VmaDefragmentationContext_T::ComputeDefragmentation_Full(VmaBlockVector& vector) |
| 13616 | { |
| 13617 | // Go over every allocation and try to fit it in previous blocks at lowest offsets, |
| 13618 | // if not possible: realloc within single block to minimize offset (exclude offset == 0) |
| 13619 | |
| 13620 | for (size_t i = vector.GetBlockCount() - 1; i > m_ImmovableBlockCount; --i) |
| 13621 | { |
| 13622 | VmaDeviceMemoryBlock* block = vector.GetBlock(index: i); |
| 13623 | VmaBlockMetadata* metadata = block->m_pMetadata; |
| 13624 | |
| 13625 | for (VmaAllocHandle handle = metadata->GetAllocationListBegin(); |
| 13626 | handle != VK_NULL_HANDLE; |
| 13627 | handle = metadata->GetNextAllocation(prevAlloc: handle)) |
| 13628 | { |
| 13629 | MoveAllocationData moveData = GetMoveData(handle, metadata); |
| 13630 | // Ignore newly created allocations by defragmentation algorithm |
| 13631 | if (moveData.move.srcAllocation->GetUserData() == this) |
| 13632 | continue; |
| 13633 | switch (CheckCounters(bytes: moveData.move.srcAllocation->GetSize())) |
| 13634 | { |
| 13635 | case CounterStatus::Ignore: |
| 13636 | continue; |
| 13637 | case CounterStatus::End: |
| 13638 | return true; |
| 13639 | default: |
| 13640 | VMA_ASSERT(0); |
| 13641 | case CounterStatus::Pass: |
| 13642 | break; |
| 13643 | } |
| 13644 | |
| 13645 | // Check all previous blocks for free space |
| 13646 | const size_t prevMoveCount = m_Moves.size(); |
| 13647 | if (AllocInOtherBlock(start: 0, end: i, data&: moveData, vector)) |
| 13648 | return true; |
| 13649 | |
| 13650 | // If no room found then realloc within block for lower offset |
| 13651 | VkDeviceSize offset = moveData.move.srcAllocation->GetOffset(); |
| 13652 | if (prevMoveCount == m_Moves.size() && offset != 0 && metadata->GetSumFreeSize() >= moveData.size) |
| 13653 | { |
| 13654 | VmaAllocationRequest request = {}; |
| 13655 | if (metadata->CreateAllocationRequest( |
| 13656 | allocSize: moveData.size, |
| 13657 | allocAlignment: moveData.alignment, |
| 13658 | upperAddress: false, |
| 13659 | allocType: moveData.type, |
| 13660 | strategy: VMA_ALLOCATION_CREATE_STRATEGY_MIN_OFFSET_BIT, |
| 13661 | pAllocationRequest: &request)) |
| 13662 | { |
| 13663 | if (metadata->GetAllocationOffset(allocHandle: request.allocHandle) < offset) |
| 13664 | { |
| 13665 | if (vector.CommitAllocationRequest( |
| 13666 | allocRequest&: request, |
| 13667 | pBlock: block, |
| 13668 | alignment: moveData.alignment, |
| 13669 | allocFlags: moveData.flags, |
| 13670 | pUserData: this, |
| 13671 | suballocType: moveData.type, |
| 13672 | pAllocation: &moveData.move.dstTmpAllocation) == VK_SUCCESS) |
| 13673 | { |
| 13674 | m_Moves.push_back(src: moveData.move); |
| 13675 | if (IncrementCounters(bytes: moveData.size)) |
| 13676 | return true; |
| 13677 | } |
| 13678 | } |
| 13679 | } |
| 13680 | } |
| 13681 | } |
| 13682 | } |
| 13683 | return false; |
| 13684 | } |
| 13685 | |
| 13686 | bool VmaDefragmentationContext_T::ComputeDefragmentation_Extensive(VmaBlockVector& vector, size_t index) |
| 13687 | { |
| 13688 | // First free single block, then populate it to the brim, then free another block, and so on |
| 13689 | |
| 13690 | // Fallback to previous algorithm since without granularity conflicts it can achieve max packing |
| 13691 | if (vector.m_BufferImageGranularity == 1) |
| 13692 | return ComputeDefragmentation_Full(vector); |
| 13693 | |
| 13694 | VMA_ASSERT(m_AlgorithmState != VMA_NULL); |
| 13695 | |
| 13696 | StateExtensive& vectorState = reinterpret_cast<StateExtensive*>(m_AlgorithmState)[index]; |
| 13697 | |
| 13698 | bool texturePresent = false, bufferPresent = false, otherPresent = false; |
| 13699 | switch (vectorState.operation) |
| 13700 | { |
| 13701 | case StateExtensive::Operation::Done: // Vector defragmented |
| 13702 | return false; |
| 13703 | case StateExtensive::Operation::FindFreeBlockBuffer: |
| 13704 | case StateExtensive::Operation::FindFreeBlockTexture: |
| 13705 | case StateExtensive::Operation::FindFreeBlockAll: |
| 13706 | { |
| 13707 | // No more blocks to free, just perform fast realloc and move to cleanup |
| 13708 | if (vectorState.firstFreeBlock == 0) |
| 13709 | { |
| 13710 | vectorState.operation = StateExtensive::Operation::Cleanup; |
| 13711 | return ComputeDefragmentation_Fast(vector); |
| 13712 | } |
| 13713 | |
| 13714 | // No free blocks, have to clear last one |
| 13715 | size_t last = (vectorState.firstFreeBlock == SIZE_MAX ? vector.GetBlockCount() : vectorState.firstFreeBlock) - 1; |
| 13716 | VmaBlockMetadata* freeMetadata = vector.GetBlock(index: last)->m_pMetadata; |
| 13717 | |
| 13718 | const size_t prevMoveCount = m_Moves.size(); |
| 13719 | for (VmaAllocHandle handle = freeMetadata->GetAllocationListBegin(); |
| 13720 | handle != VK_NULL_HANDLE; |
| 13721 | handle = freeMetadata->GetNextAllocation(prevAlloc: handle)) |
| 13722 | { |
| 13723 | MoveAllocationData moveData = GetMoveData(handle, metadata: freeMetadata); |
| 13724 | switch (CheckCounters(bytes: moveData.move.srcAllocation->GetSize())) |
| 13725 | { |
| 13726 | case CounterStatus::Ignore: |
| 13727 | continue; |
| 13728 | case CounterStatus::End: |
| 13729 | return true; |
| 13730 | default: |
| 13731 | VMA_ASSERT(0); |
| 13732 | case CounterStatus::Pass: |
| 13733 | break; |
| 13734 | } |
| 13735 | |
| 13736 | // Check all previous blocks for free space |
| 13737 | if (AllocInOtherBlock(start: 0, end: last, data&: moveData, vector)) |
| 13738 | { |
| 13739 | // Full clear performed already |
| 13740 | if (prevMoveCount != m_Moves.size() && freeMetadata->GetNextAllocation(prevAlloc: handle) == VK_NULL_HANDLE) |
| 13741 | reinterpret_cast<size_t*>(m_AlgorithmState)[index] = last; |
| 13742 | return true; |
| 13743 | } |
| 13744 | } |
| 13745 | |
| 13746 | if (prevMoveCount == m_Moves.size()) |
| 13747 | { |
| 13748 | // Cannot perform full clear, have to move data in other blocks around |
| 13749 | if (last != 0) |
| 13750 | { |
| 13751 | for (size_t i = last - 1; i; --i) |
| 13752 | { |
| 13753 | if (ReallocWithinBlock(vector, block: vector.GetBlock(index: i))) |
| 13754 | return true; |
| 13755 | } |
| 13756 | } |
| 13757 | |
| 13758 | if (prevMoveCount == m_Moves.size()) |
| 13759 | { |
| 13760 | // No possible reallocs within blocks, try to move them around fast |
| 13761 | return ComputeDefragmentation_Fast(vector); |
| 13762 | } |
| 13763 | } |
| 13764 | else |
| 13765 | { |
| 13766 | switch (vectorState.operation) |
| 13767 | { |
| 13768 | case StateExtensive::Operation::FindFreeBlockBuffer: |
| 13769 | vectorState.operation = StateExtensive::Operation::MoveBuffers; |
| 13770 | break; |
| 13771 | default: |
| 13772 | VMA_ASSERT(0); |
| 13773 | case StateExtensive::Operation::FindFreeBlockTexture: |
| 13774 | vectorState.operation = StateExtensive::Operation::MoveTextures; |
| 13775 | break; |
| 13776 | case StateExtensive::Operation::FindFreeBlockAll: |
| 13777 | vectorState.operation = StateExtensive::Operation::MoveAll; |
| 13778 | break; |
| 13779 | } |
| 13780 | vectorState.firstFreeBlock = last; |
| 13781 | // Nothing done, block found without reallocations, can perform another reallocs in same pass |
| 13782 | return ComputeDefragmentation_Extensive(vector, index); |
| 13783 | } |
| 13784 | break; |
| 13785 | } |
| 13786 | case StateExtensive::Operation::MoveTextures: |
| 13787 | { |
| 13788 | if (MoveDataToFreeBlocks(currentType: VMA_SUBALLOCATION_TYPE_IMAGE_OPTIMAL, vector, |
| 13789 | firstFreeBlock: vectorState.firstFreeBlock, texturePresent, bufferPresent, otherPresent)) |
| 13790 | { |
| 13791 | if (texturePresent) |
| 13792 | { |
| 13793 | vectorState.operation = StateExtensive::Operation::FindFreeBlockTexture; |
| 13794 | return ComputeDefragmentation_Extensive(vector, index); |
| 13795 | } |
| 13796 | |
| 13797 | if (!bufferPresent && !otherPresent) |
| 13798 | { |
| 13799 | vectorState.operation = StateExtensive::Operation::Cleanup; |
| 13800 | break; |
| 13801 | } |
| 13802 | |
| 13803 | // No more textures to move, check buffers |
| 13804 | vectorState.operation = StateExtensive::Operation::MoveBuffers; |
| 13805 | bufferPresent = false; |
| 13806 | otherPresent = false; |
| 13807 | } |
| 13808 | else |
| 13809 | break; |
| 13810 | } |
| 13811 | case StateExtensive::Operation::MoveBuffers: |
| 13812 | { |
| 13813 | if (MoveDataToFreeBlocks(currentType: VMA_SUBALLOCATION_TYPE_BUFFER, vector, |
| 13814 | firstFreeBlock: vectorState.firstFreeBlock, texturePresent, bufferPresent, otherPresent)) |
| 13815 | { |
| 13816 | if (bufferPresent) |
| 13817 | { |
| 13818 | vectorState.operation = StateExtensive::Operation::FindFreeBlockBuffer; |
| 13819 | return ComputeDefragmentation_Extensive(vector, index); |
| 13820 | } |
| 13821 | |
| 13822 | if (!otherPresent) |
| 13823 | { |
| 13824 | vectorState.operation = StateExtensive::Operation::Cleanup; |
| 13825 | break; |
| 13826 | } |
| 13827 | |
| 13828 | // No more buffers to move, check all others |
| 13829 | vectorState.operation = StateExtensive::Operation::MoveAll; |
| 13830 | otherPresent = false; |
| 13831 | } |
| 13832 | else |
| 13833 | break; |
| 13834 | } |
| 13835 | case StateExtensive::Operation::MoveAll: |
| 13836 | { |
| 13837 | if (MoveDataToFreeBlocks(currentType: VMA_SUBALLOCATION_TYPE_FREE, vector, |
| 13838 | firstFreeBlock: vectorState.firstFreeBlock, texturePresent, bufferPresent, otherPresent)) |
| 13839 | { |
| 13840 | if (otherPresent) |
| 13841 | { |
| 13842 | vectorState.operation = StateExtensive::Operation::FindFreeBlockBuffer; |
| 13843 | return ComputeDefragmentation_Extensive(vector, index); |
| 13844 | } |
| 13845 | // Everything moved |
| 13846 | vectorState.operation = StateExtensive::Operation::Cleanup; |
| 13847 | } |
| 13848 | break; |
| 13849 | } |
| 13850 | case StateExtensive::Operation::Cleanup: |
| 13851 | // Cleanup is handled below so that other operations may reuse the cleanup code. This case is here to prevent the unhandled enum value warning (C4062). |
| 13852 | break; |
| 13853 | } |
| 13854 | |
| 13855 | if (vectorState.operation == StateExtensive::Operation::Cleanup) |
| 13856 | { |
| 13857 | // All other work done, pack data in blocks even tighter if possible |
| 13858 | const size_t prevMoveCount = m_Moves.size(); |
| 13859 | for (size_t i = 0; i < vector.GetBlockCount(); ++i) |
| 13860 | { |
| 13861 | if (ReallocWithinBlock(vector, block: vector.GetBlock(index: i))) |
| 13862 | return true; |
| 13863 | } |
| 13864 | |
| 13865 | if (prevMoveCount == m_Moves.size()) |
| 13866 | vectorState.operation = StateExtensive::Operation::Done; |
| 13867 | } |
| 13868 | return false; |
| 13869 | } |
| 13870 | |
| 13871 | void VmaDefragmentationContext_T::UpdateVectorStatistics(VmaBlockVector& vector, StateBalanced& state) |
| 13872 | { |
| 13873 | size_t allocCount = 0; |
| 13874 | size_t freeCount = 0; |
| 13875 | state.avgFreeSize = 0; |
| 13876 | state.avgAllocSize = 0; |
| 13877 | |
| 13878 | for (size_t i = 0; i < vector.GetBlockCount(); ++i) |
| 13879 | { |
| 13880 | VmaBlockMetadata* metadata = vector.GetBlock(index: i)->m_pMetadata; |
| 13881 | |
| 13882 | allocCount += metadata->GetAllocationCount(); |
| 13883 | freeCount += metadata->GetFreeRegionsCount(); |
| 13884 | state.avgFreeSize += metadata->GetSumFreeSize(); |
| 13885 | state.avgAllocSize += metadata->GetSize(); |
| 13886 | } |
| 13887 | |
| 13888 | state.avgAllocSize = (state.avgAllocSize - state.avgFreeSize) / allocCount; |
| 13889 | state.avgFreeSize /= freeCount; |
| 13890 | } |
| 13891 | |
| 13892 | bool VmaDefragmentationContext_T::MoveDataToFreeBlocks(VmaSuballocationType currentType, |
| 13893 | VmaBlockVector& vector, size_t firstFreeBlock, |
| 13894 | bool& texturePresent, bool& bufferPresent, bool& otherPresent) |
| 13895 | { |
| 13896 | const size_t prevMoveCount = m_Moves.size(); |
| 13897 | for (size_t i = firstFreeBlock ; i;) |
| 13898 | { |
| 13899 | VmaDeviceMemoryBlock* block = vector.GetBlock(index: --i); |
| 13900 | VmaBlockMetadata* metadata = block->m_pMetadata; |
| 13901 | |
| 13902 | for (VmaAllocHandle handle = metadata->GetAllocationListBegin(); |
| 13903 | handle != VK_NULL_HANDLE; |
| 13904 | handle = metadata->GetNextAllocation(prevAlloc: handle)) |
| 13905 | { |
| 13906 | MoveAllocationData moveData = GetMoveData(handle, metadata); |
| 13907 | // Ignore newly created allocations by defragmentation algorithm |
| 13908 | if (moveData.move.srcAllocation->GetUserData() == this) |
| 13909 | continue; |
| 13910 | switch (CheckCounters(bytes: moveData.move.srcAllocation->GetSize())) |
| 13911 | { |
| 13912 | case CounterStatus::Ignore: |
| 13913 | continue; |
| 13914 | case CounterStatus::End: |
| 13915 | return true; |
| 13916 | default: |
| 13917 | VMA_ASSERT(0); |
| 13918 | case CounterStatus::Pass: |
| 13919 | break; |
| 13920 | } |
| 13921 | |
| 13922 | // Move only single type of resources at once |
| 13923 | if (!VmaIsBufferImageGranularityConflict(suballocType1: moveData.type, suballocType2: currentType)) |
| 13924 | { |
| 13925 | // Try to fit allocation into free blocks |
| 13926 | if (AllocInOtherBlock(start: firstFreeBlock, end: vector.GetBlockCount(), data&: moveData, vector)) |
| 13927 | return false; |
| 13928 | } |
| 13929 | |
| 13930 | if (!VmaIsBufferImageGranularityConflict(suballocType1: moveData.type, suballocType2: VMA_SUBALLOCATION_TYPE_IMAGE_OPTIMAL)) |
| 13931 | texturePresent = true; |
| 13932 | else if (!VmaIsBufferImageGranularityConflict(suballocType1: moveData.type, suballocType2: VMA_SUBALLOCATION_TYPE_BUFFER)) |
| 13933 | bufferPresent = true; |
| 13934 | else |
| 13935 | otherPresent = true; |
| 13936 | } |
| 13937 | } |
| 13938 | return prevMoveCount == m_Moves.size(); |
| 13939 | } |
| 13940 | #endif // _VMA_DEFRAGMENTATION_CONTEXT_FUNCTIONS |
| 13941 | |
| 13942 | #ifndef _VMA_POOL_T_FUNCTIONS |
| 13943 | VmaPool_T::VmaPool_T( |
| 13944 | VmaAllocator hAllocator, |
| 13945 | const VmaPoolCreateInfo& createInfo, |
| 13946 | VkDeviceSize preferredBlockSize) |
| 13947 | : m_BlockVector( |
| 13948 | hAllocator, |
| 13949 | this, // hParentPool |
| 13950 | createInfo.memoryTypeIndex, |
| 13951 | createInfo.blockSize != 0 ? createInfo.blockSize : preferredBlockSize, |
| 13952 | createInfo.minBlockCount, |
| 13953 | createInfo.maxBlockCount, |
| 13954 | (createInfo.flags& VMA_POOL_CREATE_IGNORE_BUFFER_IMAGE_GRANULARITY_BIT) != 0 ? 1 : hAllocator->GetBufferImageGranularity(), |
| 13955 | createInfo.blockSize != 0, // explicitBlockSize |
| 13956 | createInfo.flags & VMA_POOL_CREATE_ALGORITHM_MASK, // algorithm |
| 13957 | createInfo.priority, |
| 13958 | VMA_MAX(hAllocator->GetMemoryTypeMinAlignment(createInfo.memoryTypeIndex), createInfo.minAllocationAlignment), |
| 13959 | createInfo.pMemoryAllocateNext), |
| 13960 | m_Id(0), |
| 13961 | m_Name(VMA_NULL) {} |
| 13962 | |
| 13963 | VmaPool_T::~VmaPool_T() |
| 13964 | { |
| 13965 | VMA_ASSERT(m_PrevPool == VMA_NULL && m_NextPool == VMA_NULL); |
| 13966 | } |
| 13967 | |
| 13968 | void VmaPool_T::SetName(const char* pName) |
| 13969 | { |
| 13970 | const VkAllocationCallbacks* allocs = m_BlockVector.GetAllocator()->GetAllocationCallbacks(); |
| 13971 | VmaFreeString(allocs, str: m_Name); |
| 13972 | |
| 13973 | if (pName != VMA_NULL) |
| 13974 | { |
| 13975 | m_Name = VmaCreateStringCopy(allocs, srcStr: pName); |
| 13976 | } |
| 13977 | else |
| 13978 | { |
| 13979 | m_Name = VMA_NULL; |
| 13980 | } |
| 13981 | } |
| 13982 | #endif // _VMA_POOL_T_FUNCTIONS |
| 13983 | |
| 13984 | #ifndef _VMA_ALLOCATOR_T_FUNCTIONS |
| 13985 | VmaAllocator_T::VmaAllocator_T(const VmaAllocatorCreateInfo* pCreateInfo) : |
| 13986 | m_UseMutex((pCreateInfo->flags & VMA_ALLOCATOR_CREATE_EXTERNALLY_SYNCHRONIZED_BIT) == 0), |
| 13987 | m_VulkanApiVersion(pCreateInfo->vulkanApiVersion != 0 ? pCreateInfo->vulkanApiVersion : VK_API_VERSION_1_0), |
| 13988 | m_UseKhrDedicatedAllocation((pCreateInfo->flags & VMA_ALLOCATOR_CREATE_KHR_DEDICATED_ALLOCATION_BIT) != 0), |
| 13989 | m_UseKhrBindMemory2((pCreateInfo->flags & VMA_ALLOCATOR_CREATE_KHR_BIND_MEMORY2_BIT) != 0), |
| 13990 | m_UseExtMemoryBudget((pCreateInfo->flags & VMA_ALLOCATOR_CREATE_EXT_MEMORY_BUDGET_BIT) != 0), |
| 13991 | m_UseAmdDeviceCoherentMemory((pCreateInfo->flags & VMA_ALLOCATOR_CREATE_AMD_DEVICE_COHERENT_MEMORY_BIT) != 0), |
| 13992 | m_UseKhrBufferDeviceAddress((pCreateInfo->flags & VMA_ALLOCATOR_CREATE_BUFFER_DEVICE_ADDRESS_BIT) != 0), |
| 13993 | m_UseExtMemoryPriority((pCreateInfo->flags & VMA_ALLOCATOR_CREATE_EXT_MEMORY_PRIORITY_BIT) != 0), |
| 13994 | m_hDevice(pCreateInfo->device), |
| 13995 | m_hInstance(pCreateInfo->instance), |
| 13996 | m_AllocationCallbacksSpecified(pCreateInfo->pAllocationCallbacks != VMA_NULL), |
| 13997 | m_AllocationCallbacks(pCreateInfo->pAllocationCallbacks ? |
| 13998 | *pCreateInfo->pAllocationCallbacks : VmaEmptyAllocationCallbacks), |
| 13999 | m_AllocationObjectAllocator(&m_AllocationCallbacks), |
| 14000 | m_HeapSizeLimitMask(0), |
| 14001 | m_DeviceMemoryCount(0), |
| 14002 | m_PreferredLargeHeapBlockSize(0), |
| 14003 | m_PhysicalDevice(pCreateInfo->physicalDevice), |
| 14004 | m_GpuDefragmentationMemoryTypeBits(UINT32_MAX), |
| 14005 | m_NextPoolId(0), |
| 14006 | m_GlobalMemoryTypeBits(UINT32_MAX) |
| 14007 | { |
| 14008 | if(m_VulkanApiVersion >= VK_MAKE_VERSION(1, 1, 0)) |
| 14009 | { |
| 14010 | m_UseKhrDedicatedAllocation = false; |
| 14011 | m_UseKhrBindMemory2 = false; |
| 14012 | } |
| 14013 | |
| 14014 | if(VMA_DEBUG_DETECT_CORRUPTION) |
| 14015 | { |
| 14016 | // Needs to be multiply of uint32_t size because we are going to write VMA_CORRUPTION_DETECTION_MAGIC_VALUE to it. |
| 14017 | VMA_ASSERT(VMA_DEBUG_MARGIN % sizeof(uint32_t) == 0); |
| 14018 | } |
| 14019 | |
| 14020 | VMA_ASSERT(pCreateInfo->physicalDevice && pCreateInfo->device && pCreateInfo->instance); |
| 14021 | |
| 14022 | if(m_VulkanApiVersion < VK_MAKE_VERSION(1, 1, 0)) |
| 14023 | { |
| 14024 | #if !(VMA_DEDICATED_ALLOCATION) |
| 14025 | if((pCreateInfo->flags & VMA_ALLOCATOR_CREATE_KHR_DEDICATED_ALLOCATION_BIT) != 0) |
| 14026 | { |
| 14027 | VMA_ASSERT(0 && "VMA_ALLOCATOR_CREATE_KHR_DEDICATED_ALLOCATION_BIT set but required extensions are disabled by preprocessor macros." ); |
| 14028 | } |
| 14029 | #endif |
| 14030 | #if !(VMA_BIND_MEMORY2) |
| 14031 | if((pCreateInfo->flags & VMA_ALLOCATOR_CREATE_KHR_BIND_MEMORY2_BIT) != 0) |
| 14032 | { |
| 14033 | VMA_ASSERT(0 && "VMA_ALLOCATOR_CREATE_KHR_BIND_MEMORY2_BIT set but required extension is disabled by preprocessor macros." ); |
| 14034 | } |
| 14035 | #endif |
| 14036 | } |
| 14037 | #if !(VMA_MEMORY_BUDGET) |
| 14038 | if((pCreateInfo->flags & VMA_ALLOCATOR_CREATE_EXT_MEMORY_BUDGET_BIT) != 0) |
| 14039 | { |
| 14040 | VMA_ASSERT(0 && "VMA_ALLOCATOR_CREATE_EXT_MEMORY_BUDGET_BIT set but required extension is disabled by preprocessor macros." ); |
| 14041 | } |
| 14042 | #endif |
| 14043 | #if !(VMA_BUFFER_DEVICE_ADDRESS) |
| 14044 | if(m_UseKhrBufferDeviceAddress) |
| 14045 | { |
| 14046 | VMA_ASSERT(0 && "VMA_ALLOCATOR_CREATE_BUFFER_DEVICE_ADDRESS_BIT is set but required extension or Vulkan 1.2 is not available in your Vulkan header or its support in VMA has been disabled by a preprocessor macro." ); |
| 14047 | } |
| 14048 | #endif |
| 14049 | #if VMA_VULKAN_VERSION < 1002000 |
| 14050 | if(m_VulkanApiVersion >= VK_MAKE_VERSION(1, 2, 0)) |
| 14051 | { |
| 14052 | VMA_ASSERT(0 && "vulkanApiVersion >= VK_API_VERSION_1_2 but required Vulkan version is disabled by preprocessor macros." ); |
| 14053 | } |
| 14054 | #endif |
| 14055 | #if VMA_VULKAN_VERSION < 1001000 |
| 14056 | if(m_VulkanApiVersion >= VK_MAKE_VERSION(1, 1, 0)) |
| 14057 | { |
| 14058 | VMA_ASSERT(0 && "vulkanApiVersion >= VK_API_VERSION_1_1 but required Vulkan version is disabled by preprocessor macros." ); |
| 14059 | } |
| 14060 | #endif |
| 14061 | #if !(VMA_MEMORY_PRIORITY) |
| 14062 | if(m_UseExtMemoryPriority) |
| 14063 | { |
| 14064 | VMA_ASSERT(0 && "VMA_ALLOCATOR_CREATE_EXT_MEMORY_PRIORITY_BIT is set but required extension is not available in your Vulkan header or its support in VMA has been disabled by a preprocessor macro." ); |
| 14065 | } |
| 14066 | #endif |
| 14067 | |
| 14068 | memset(s: &m_DeviceMemoryCallbacks, c: 0 ,n: sizeof(m_DeviceMemoryCallbacks)); |
| 14069 | memset(s: &m_PhysicalDeviceProperties, c: 0, n: sizeof(m_PhysicalDeviceProperties)); |
| 14070 | memset(s: &m_MemProps, c: 0, n: sizeof(m_MemProps)); |
| 14071 | |
| 14072 | memset(s: &m_pBlockVectors, c: 0, n: sizeof(m_pBlockVectors)); |
| 14073 | memset(s: &m_VulkanFunctions, c: 0, n: sizeof(m_VulkanFunctions)); |
| 14074 | |
| 14075 | #if VMA_EXTERNAL_MEMORY |
| 14076 | memset(s: &m_TypeExternalMemoryHandleTypes, c: 0, n: sizeof(m_TypeExternalMemoryHandleTypes)); |
| 14077 | #endif // #if VMA_EXTERNAL_MEMORY |
| 14078 | |
| 14079 | if(pCreateInfo->pDeviceMemoryCallbacks != VMA_NULL) |
| 14080 | { |
| 14081 | m_DeviceMemoryCallbacks.pUserData = pCreateInfo->pDeviceMemoryCallbacks->pUserData; |
| 14082 | m_DeviceMemoryCallbacks.pfnAllocate = pCreateInfo->pDeviceMemoryCallbacks->pfnAllocate; |
| 14083 | m_DeviceMemoryCallbacks.pfnFree = pCreateInfo->pDeviceMemoryCallbacks->pfnFree; |
| 14084 | } |
| 14085 | |
| 14086 | ImportVulkanFunctions(pVulkanFunctions: pCreateInfo->pVulkanFunctions); |
| 14087 | |
| 14088 | (*m_VulkanFunctions.vkGetPhysicalDeviceProperties)(m_PhysicalDevice, &m_PhysicalDeviceProperties); |
| 14089 | (*m_VulkanFunctions.vkGetPhysicalDeviceMemoryProperties)(m_PhysicalDevice, &m_MemProps); |
| 14090 | |
| 14091 | VMA_ASSERT(VmaIsPow2(VMA_MIN_ALIGNMENT)); |
| 14092 | VMA_ASSERT(VmaIsPow2(VMA_DEBUG_MIN_BUFFER_IMAGE_GRANULARITY)); |
| 14093 | VMA_ASSERT(VmaIsPow2(m_PhysicalDeviceProperties.limits.bufferImageGranularity)); |
| 14094 | VMA_ASSERT(VmaIsPow2(m_PhysicalDeviceProperties.limits.nonCoherentAtomSize)); |
| 14095 | |
| 14096 | m_PreferredLargeHeapBlockSize = (pCreateInfo->preferredLargeHeapBlockSize != 0) ? |
| 14097 | pCreateInfo->preferredLargeHeapBlockSize : static_cast<VkDeviceSize>(VMA_DEFAULT_LARGE_HEAP_BLOCK_SIZE); |
| 14098 | |
| 14099 | m_GlobalMemoryTypeBits = CalculateGlobalMemoryTypeBits(); |
| 14100 | |
| 14101 | #if VMA_EXTERNAL_MEMORY |
| 14102 | if(pCreateInfo->pTypeExternalMemoryHandleTypes != VMA_NULL) |
| 14103 | { |
| 14104 | memcpy(dest: m_TypeExternalMemoryHandleTypes, src: pCreateInfo->pTypeExternalMemoryHandleTypes, |
| 14105 | n: sizeof(VkExternalMemoryHandleTypeFlagsKHR) * GetMemoryTypeCount()); |
| 14106 | } |
| 14107 | #endif // #if VMA_EXTERNAL_MEMORY |
| 14108 | |
| 14109 | if(pCreateInfo->pHeapSizeLimit != VMA_NULL) |
| 14110 | { |
| 14111 | for(uint32_t heapIndex = 0; heapIndex < GetMemoryHeapCount(); ++heapIndex) |
| 14112 | { |
| 14113 | const VkDeviceSize limit = pCreateInfo->pHeapSizeLimit[heapIndex]; |
| 14114 | if(limit != VK_WHOLE_SIZE) |
| 14115 | { |
| 14116 | m_HeapSizeLimitMask |= 1u << heapIndex; |
| 14117 | if(limit < m_MemProps.memoryHeaps[heapIndex].size) |
| 14118 | { |
| 14119 | m_MemProps.memoryHeaps[heapIndex].size = limit; |
| 14120 | } |
| 14121 | } |
| 14122 | } |
| 14123 | } |
| 14124 | |
| 14125 | for(uint32_t memTypeIndex = 0; memTypeIndex < GetMemoryTypeCount(); ++memTypeIndex) |
| 14126 | { |
| 14127 | // Create only supported types |
| 14128 | if((m_GlobalMemoryTypeBits & (1u << memTypeIndex)) != 0) |
| 14129 | { |
| 14130 | const VkDeviceSize preferredBlockSize = CalcPreferredBlockSize(memTypeIndex); |
| 14131 | m_pBlockVectors[memTypeIndex] = vma_new(this, VmaBlockVector)( |
| 14132 | this, |
| 14133 | VK_NULL_HANDLE, // hParentPool |
| 14134 | memTypeIndex, |
| 14135 | preferredBlockSize, |
| 14136 | 0, |
| 14137 | SIZE_MAX, |
| 14138 | GetBufferImageGranularity(), |
| 14139 | false, // explicitBlockSize |
| 14140 | 0, // algorithm |
| 14141 | 0.5f, // priority (0.5 is the default per Vulkan spec) |
| 14142 | GetMemoryTypeMinAlignment(memTypeIndex), // minAllocationAlignment |
| 14143 | VMA_NULL); // // pMemoryAllocateNext |
| 14144 | // No need to call m_pBlockVectors[memTypeIndex][blockVectorTypeIndex]->CreateMinBlocks here, |
| 14145 | // becase minBlockCount is 0. |
| 14146 | } |
| 14147 | } |
| 14148 | } |
| 14149 | |
| 14150 | VkResult VmaAllocator_T::Init(const VmaAllocatorCreateInfo* pCreateInfo) |
| 14151 | { |
| 14152 | VkResult res = VK_SUCCESS; |
| 14153 | |
| 14154 | #if VMA_MEMORY_BUDGET |
| 14155 | if(m_UseExtMemoryBudget) |
| 14156 | { |
| 14157 | UpdateVulkanBudget(); |
| 14158 | } |
| 14159 | #endif // #if VMA_MEMORY_BUDGET |
| 14160 | |
| 14161 | return res; |
| 14162 | } |
| 14163 | |
| 14164 | VmaAllocator_T::~VmaAllocator_T() |
| 14165 | { |
| 14166 | VMA_ASSERT(m_Pools.IsEmpty()); |
| 14167 | |
| 14168 | for(size_t memTypeIndex = GetMemoryTypeCount(); memTypeIndex--; ) |
| 14169 | { |
| 14170 | vma_delete(hAllocator: this, ptr: m_pBlockVectors[memTypeIndex]); |
| 14171 | } |
| 14172 | } |
| 14173 | |
| 14174 | void VmaAllocator_T::ImportVulkanFunctions(const VmaVulkanFunctions* pVulkanFunctions) |
| 14175 | { |
| 14176 | #if VMA_STATIC_VULKAN_FUNCTIONS == 1 |
| 14177 | ImportVulkanFunctions_Static(); |
| 14178 | #endif |
| 14179 | |
| 14180 | if(pVulkanFunctions != VMA_NULL) |
| 14181 | { |
| 14182 | ImportVulkanFunctions_Custom(pVulkanFunctions); |
| 14183 | } |
| 14184 | |
| 14185 | #if VMA_DYNAMIC_VULKAN_FUNCTIONS == 1 |
| 14186 | ImportVulkanFunctions_Dynamic(); |
| 14187 | #endif |
| 14188 | |
| 14189 | ValidateVulkanFunctions(); |
| 14190 | } |
| 14191 | |
| 14192 | #if VMA_STATIC_VULKAN_FUNCTIONS == 1 |
| 14193 | |
| 14194 | void VmaAllocator_T::ImportVulkanFunctions_Static() |
| 14195 | { |
| 14196 | // Vulkan 1.0 |
| 14197 | m_VulkanFunctions.vkGetInstanceProcAddr = (PFN_vkGetInstanceProcAddr)vkGetInstanceProcAddr; |
| 14198 | m_VulkanFunctions.vkGetDeviceProcAddr = (PFN_vkGetDeviceProcAddr)vkGetDeviceProcAddr; |
| 14199 | m_VulkanFunctions.vkGetPhysicalDeviceProperties = (PFN_vkGetPhysicalDeviceProperties)vkGetPhysicalDeviceProperties; |
| 14200 | m_VulkanFunctions.vkGetPhysicalDeviceMemoryProperties = (PFN_vkGetPhysicalDeviceMemoryProperties)vkGetPhysicalDeviceMemoryProperties; |
| 14201 | m_VulkanFunctions.vkAllocateMemory = (PFN_vkAllocateMemory)vkAllocateMemory; |
| 14202 | m_VulkanFunctions.vkFreeMemory = (PFN_vkFreeMemory)vkFreeMemory; |
| 14203 | m_VulkanFunctions.vkMapMemory = (PFN_vkMapMemory)vkMapMemory; |
| 14204 | m_VulkanFunctions.vkUnmapMemory = (PFN_vkUnmapMemory)vkUnmapMemory; |
| 14205 | m_VulkanFunctions.vkFlushMappedMemoryRanges = (PFN_vkFlushMappedMemoryRanges)vkFlushMappedMemoryRanges; |
| 14206 | m_VulkanFunctions.vkInvalidateMappedMemoryRanges = (PFN_vkInvalidateMappedMemoryRanges)vkInvalidateMappedMemoryRanges; |
| 14207 | m_VulkanFunctions.vkBindBufferMemory = (PFN_vkBindBufferMemory)vkBindBufferMemory; |
| 14208 | m_VulkanFunctions.vkBindImageMemory = (PFN_vkBindImageMemory)vkBindImageMemory; |
| 14209 | m_VulkanFunctions.vkGetBufferMemoryRequirements = (PFN_vkGetBufferMemoryRequirements)vkGetBufferMemoryRequirements; |
| 14210 | m_VulkanFunctions.vkGetImageMemoryRequirements = (PFN_vkGetImageMemoryRequirements)vkGetImageMemoryRequirements; |
| 14211 | m_VulkanFunctions.vkCreateBuffer = (PFN_vkCreateBuffer)vkCreateBuffer; |
| 14212 | m_VulkanFunctions.vkDestroyBuffer = (PFN_vkDestroyBuffer)vkDestroyBuffer; |
| 14213 | m_VulkanFunctions.vkCreateImage = (PFN_vkCreateImage)vkCreateImage; |
| 14214 | m_VulkanFunctions.vkDestroyImage = (PFN_vkDestroyImage)vkDestroyImage; |
| 14215 | m_VulkanFunctions.vkCmdCopyBuffer = (PFN_vkCmdCopyBuffer)vkCmdCopyBuffer; |
| 14216 | |
| 14217 | // Vulkan 1.1 |
| 14218 | #if VMA_VULKAN_VERSION >= 1001000 |
| 14219 | if(m_VulkanApiVersion >= VK_MAKE_VERSION(1, 1, 0)) |
| 14220 | { |
| 14221 | m_VulkanFunctions.vkGetBufferMemoryRequirements2KHR = (PFN_vkGetBufferMemoryRequirements2)vkGetBufferMemoryRequirements2; |
| 14222 | m_VulkanFunctions.vkGetImageMemoryRequirements2KHR = (PFN_vkGetImageMemoryRequirements2)vkGetImageMemoryRequirements2; |
| 14223 | m_VulkanFunctions.vkBindBufferMemory2KHR = (PFN_vkBindBufferMemory2)vkBindBufferMemory2; |
| 14224 | m_VulkanFunctions.vkBindImageMemory2KHR = (PFN_vkBindImageMemory2)vkBindImageMemory2; |
| 14225 | m_VulkanFunctions.vkGetPhysicalDeviceMemoryProperties2KHR = (PFN_vkGetPhysicalDeviceMemoryProperties2)vkGetPhysicalDeviceMemoryProperties2; |
| 14226 | } |
| 14227 | #endif |
| 14228 | |
| 14229 | #if VMA_VULKAN_VERSION >= 1003000 |
| 14230 | if(m_VulkanApiVersion >= VK_MAKE_VERSION(1, 3, 0)) |
| 14231 | { |
| 14232 | m_VulkanFunctions.vkGetDeviceBufferMemoryRequirements = (PFN_vkGetDeviceBufferMemoryRequirements)vkGetDeviceBufferMemoryRequirements; |
| 14233 | m_VulkanFunctions.vkGetDeviceImageMemoryRequirements = (PFN_vkGetDeviceImageMemoryRequirements)vkGetDeviceImageMemoryRequirements; |
| 14234 | } |
| 14235 | #endif |
| 14236 | } |
| 14237 | |
| 14238 | #endif // VMA_STATIC_VULKAN_FUNCTIONS == 1 |
| 14239 | |
| 14240 | void VmaAllocator_T::ImportVulkanFunctions_Custom(const VmaVulkanFunctions* pVulkanFunctions) |
| 14241 | { |
| 14242 | VMA_ASSERT(pVulkanFunctions != VMA_NULL); |
| 14243 | |
| 14244 | #define VMA_COPY_IF_NOT_NULL(funcName) \ |
| 14245 | if(pVulkanFunctions->funcName != VMA_NULL) m_VulkanFunctions.funcName = pVulkanFunctions->funcName; |
| 14246 | |
| 14247 | VMA_COPY_IF_NOT_NULL(vkGetInstanceProcAddr); |
| 14248 | VMA_COPY_IF_NOT_NULL(vkGetDeviceProcAddr); |
| 14249 | VMA_COPY_IF_NOT_NULL(vkGetPhysicalDeviceProperties); |
| 14250 | VMA_COPY_IF_NOT_NULL(vkGetPhysicalDeviceMemoryProperties); |
| 14251 | VMA_COPY_IF_NOT_NULL(vkAllocateMemory); |
| 14252 | VMA_COPY_IF_NOT_NULL(vkFreeMemory); |
| 14253 | VMA_COPY_IF_NOT_NULL(vkMapMemory); |
| 14254 | VMA_COPY_IF_NOT_NULL(vkUnmapMemory); |
| 14255 | VMA_COPY_IF_NOT_NULL(vkFlushMappedMemoryRanges); |
| 14256 | VMA_COPY_IF_NOT_NULL(vkInvalidateMappedMemoryRanges); |
| 14257 | VMA_COPY_IF_NOT_NULL(vkBindBufferMemory); |
| 14258 | VMA_COPY_IF_NOT_NULL(vkBindImageMemory); |
| 14259 | VMA_COPY_IF_NOT_NULL(vkGetBufferMemoryRequirements); |
| 14260 | VMA_COPY_IF_NOT_NULL(vkGetImageMemoryRequirements); |
| 14261 | VMA_COPY_IF_NOT_NULL(vkCreateBuffer); |
| 14262 | VMA_COPY_IF_NOT_NULL(vkDestroyBuffer); |
| 14263 | VMA_COPY_IF_NOT_NULL(vkCreateImage); |
| 14264 | VMA_COPY_IF_NOT_NULL(vkDestroyImage); |
| 14265 | VMA_COPY_IF_NOT_NULL(vkCmdCopyBuffer); |
| 14266 | |
| 14267 | #if VMA_DEDICATED_ALLOCATION || VMA_VULKAN_VERSION >= 1001000 |
| 14268 | VMA_COPY_IF_NOT_NULL(vkGetBufferMemoryRequirements2KHR); |
| 14269 | VMA_COPY_IF_NOT_NULL(vkGetImageMemoryRequirements2KHR); |
| 14270 | #endif |
| 14271 | |
| 14272 | #if VMA_BIND_MEMORY2 || VMA_VULKAN_VERSION >= 1001000 |
| 14273 | VMA_COPY_IF_NOT_NULL(vkBindBufferMemory2KHR); |
| 14274 | VMA_COPY_IF_NOT_NULL(vkBindImageMemory2KHR); |
| 14275 | #endif |
| 14276 | |
| 14277 | #if VMA_MEMORY_BUDGET |
| 14278 | VMA_COPY_IF_NOT_NULL(vkGetPhysicalDeviceMemoryProperties2KHR); |
| 14279 | #endif |
| 14280 | |
| 14281 | #if VMA_VULKAN_VERSION >= 1003000 |
| 14282 | VMA_COPY_IF_NOT_NULL(vkGetDeviceBufferMemoryRequirements); |
| 14283 | VMA_COPY_IF_NOT_NULL(vkGetDeviceImageMemoryRequirements); |
| 14284 | #endif |
| 14285 | |
| 14286 | #undef VMA_COPY_IF_NOT_NULL |
| 14287 | } |
| 14288 | |
| 14289 | #if VMA_DYNAMIC_VULKAN_FUNCTIONS == 1 |
| 14290 | |
| 14291 | void VmaAllocator_T::ImportVulkanFunctions_Dynamic() |
| 14292 | { |
| 14293 | VMA_ASSERT(m_VulkanFunctions.vkGetInstanceProcAddr && m_VulkanFunctions.vkGetDeviceProcAddr && |
| 14294 | "To use VMA_DYNAMIC_VULKAN_FUNCTIONS in new versions of VMA you now have to pass " |
| 14295 | "VmaVulkanFunctions::vkGetInstanceProcAddr and vkGetDeviceProcAddr as VmaAllocatorCreateInfo::pVulkanFunctions. " |
| 14296 | "Other members can be null." ); |
| 14297 | |
| 14298 | #define VMA_FETCH_INSTANCE_FUNC(memberName, functionPointerType, functionNameString) \ |
| 14299 | if(m_VulkanFunctions.memberName == VMA_NULL) \ |
| 14300 | m_VulkanFunctions.memberName = \ |
| 14301 | (functionPointerType)m_VulkanFunctions.vkGetInstanceProcAddr(m_hInstance, functionNameString); |
| 14302 | #define VMA_FETCH_DEVICE_FUNC(memberName, functionPointerType, functionNameString) \ |
| 14303 | if(m_VulkanFunctions.memberName == VMA_NULL) \ |
| 14304 | m_VulkanFunctions.memberName = \ |
| 14305 | (functionPointerType)m_VulkanFunctions.vkGetDeviceProcAddr(m_hDevice, functionNameString); |
| 14306 | |
| 14307 | VMA_FETCH_INSTANCE_FUNC(vkGetPhysicalDeviceProperties, PFN_vkGetPhysicalDeviceProperties, "vkGetPhysicalDeviceProperties" ); |
| 14308 | VMA_FETCH_INSTANCE_FUNC(vkGetPhysicalDeviceMemoryProperties, PFN_vkGetPhysicalDeviceMemoryProperties, "vkGetPhysicalDeviceMemoryProperties" ); |
| 14309 | VMA_FETCH_DEVICE_FUNC(vkAllocateMemory, PFN_vkAllocateMemory, "vkAllocateMemory" ); |
| 14310 | VMA_FETCH_DEVICE_FUNC(vkFreeMemory, PFN_vkFreeMemory, "vkFreeMemory" ); |
| 14311 | VMA_FETCH_DEVICE_FUNC(vkMapMemory, PFN_vkMapMemory, "vkMapMemory" ); |
| 14312 | VMA_FETCH_DEVICE_FUNC(vkUnmapMemory, PFN_vkUnmapMemory, "vkUnmapMemory" ); |
| 14313 | VMA_FETCH_DEVICE_FUNC(vkFlushMappedMemoryRanges, PFN_vkFlushMappedMemoryRanges, "vkFlushMappedMemoryRanges" ); |
| 14314 | VMA_FETCH_DEVICE_FUNC(vkInvalidateMappedMemoryRanges, PFN_vkInvalidateMappedMemoryRanges, "vkInvalidateMappedMemoryRanges" ); |
| 14315 | VMA_FETCH_DEVICE_FUNC(vkBindBufferMemory, PFN_vkBindBufferMemory, "vkBindBufferMemory" ); |
| 14316 | VMA_FETCH_DEVICE_FUNC(vkBindImageMemory, PFN_vkBindImageMemory, "vkBindImageMemory" ); |
| 14317 | VMA_FETCH_DEVICE_FUNC(vkGetBufferMemoryRequirements, PFN_vkGetBufferMemoryRequirements, "vkGetBufferMemoryRequirements" ); |
| 14318 | VMA_FETCH_DEVICE_FUNC(vkGetImageMemoryRequirements, PFN_vkGetImageMemoryRequirements, "vkGetImageMemoryRequirements" ); |
| 14319 | VMA_FETCH_DEVICE_FUNC(vkCreateBuffer, PFN_vkCreateBuffer, "vkCreateBuffer" ); |
| 14320 | VMA_FETCH_DEVICE_FUNC(vkDestroyBuffer, PFN_vkDestroyBuffer, "vkDestroyBuffer" ); |
| 14321 | VMA_FETCH_DEVICE_FUNC(vkCreateImage, PFN_vkCreateImage, "vkCreateImage" ); |
| 14322 | VMA_FETCH_DEVICE_FUNC(vkDestroyImage, PFN_vkDestroyImage, "vkDestroyImage" ); |
| 14323 | VMA_FETCH_DEVICE_FUNC(vkCmdCopyBuffer, PFN_vkCmdCopyBuffer, "vkCmdCopyBuffer" ); |
| 14324 | |
| 14325 | #if VMA_VULKAN_VERSION >= 1001000 |
| 14326 | if(m_VulkanApiVersion >= VK_MAKE_VERSION(1, 1, 0)) |
| 14327 | { |
| 14328 | VMA_FETCH_DEVICE_FUNC(vkGetBufferMemoryRequirements2KHR, PFN_vkGetBufferMemoryRequirements2, "vkGetBufferMemoryRequirements2" ); |
| 14329 | VMA_FETCH_DEVICE_FUNC(vkGetImageMemoryRequirements2KHR, PFN_vkGetImageMemoryRequirements2, "vkGetImageMemoryRequirements2" ); |
| 14330 | VMA_FETCH_DEVICE_FUNC(vkBindBufferMemory2KHR, PFN_vkBindBufferMemory2, "vkBindBufferMemory2" ); |
| 14331 | VMA_FETCH_DEVICE_FUNC(vkBindImageMemory2KHR, PFN_vkBindImageMemory2, "vkBindImageMemory2" ); |
| 14332 | VMA_FETCH_INSTANCE_FUNC(vkGetPhysicalDeviceMemoryProperties2KHR, PFN_vkGetPhysicalDeviceMemoryProperties2, "vkGetPhysicalDeviceMemoryProperties2" ); |
| 14333 | } |
| 14334 | #endif |
| 14335 | |
| 14336 | #if VMA_DEDICATED_ALLOCATION |
| 14337 | if(m_UseKhrDedicatedAllocation) |
| 14338 | { |
| 14339 | VMA_FETCH_DEVICE_FUNC(vkGetBufferMemoryRequirements2KHR, PFN_vkGetBufferMemoryRequirements2KHR, "vkGetBufferMemoryRequirements2KHR" ); |
| 14340 | VMA_FETCH_DEVICE_FUNC(vkGetImageMemoryRequirements2KHR, PFN_vkGetImageMemoryRequirements2KHR, "vkGetImageMemoryRequirements2KHR" ); |
| 14341 | } |
| 14342 | #endif |
| 14343 | |
| 14344 | #if VMA_BIND_MEMORY2 |
| 14345 | if(m_UseKhrBindMemory2) |
| 14346 | { |
| 14347 | VMA_FETCH_DEVICE_FUNC(vkBindBufferMemory2KHR, PFN_vkBindBufferMemory2KHR, "vkBindBufferMemory2KHR" ); |
| 14348 | VMA_FETCH_DEVICE_FUNC(vkBindImageMemory2KHR, PFN_vkBindImageMemory2KHR, "vkBindImageMemory2KHR" ); |
| 14349 | } |
| 14350 | #endif // #if VMA_BIND_MEMORY2 |
| 14351 | |
| 14352 | #if VMA_MEMORY_BUDGET |
| 14353 | if(m_UseExtMemoryBudget) |
| 14354 | { |
| 14355 | VMA_FETCH_INSTANCE_FUNC(vkGetPhysicalDeviceMemoryProperties2KHR, PFN_vkGetPhysicalDeviceMemoryProperties2KHR, "vkGetPhysicalDeviceMemoryProperties2KHR" ); |
| 14356 | } |
| 14357 | #endif // #if VMA_MEMORY_BUDGET |
| 14358 | |
| 14359 | #if VMA_VULKAN_VERSION >= 1003000 |
| 14360 | if(m_VulkanApiVersion >= VK_MAKE_VERSION(1, 3, 0)) |
| 14361 | { |
| 14362 | VMA_FETCH_DEVICE_FUNC(vkGetDeviceBufferMemoryRequirements, PFN_vkGetDeviceBufferMemoryRequirements, "vkGetDeviceBufferMemoryRequirements" ); |
| 14363 | VMA_FETCH_DEVICE_FUNC(vkGetDeviceImageMemoryRequirements, PFN_vkGetDeviceImageMemoryRequirements, "vkGetDeviceImageMemoryRequirements" ); |
| 14364 | } |
| 14365 | #endif |
| 14366 | |
| 14367 | #undef VMA_FETCH_DEVICE_FUNC |
| 14368 | #undef VMA_FETCH_INSTANCE_FUNC |
| 14369 | } |
| 14370 | |
| 14371 | #endif // VMA_DYNAMIC_VULKAN_FUNCTIONS == 1 |
| 14372 | |
| 14373 | void VmaAllocator_T::ValidateVulkanFunctions() |
| 14374 | { |
| 14375 | VMA_ASSERT(m_VulkanFunctions.vkGetPhysicalDeviceProperties != VMA_NULL); |
| 14376 | VMA_ASSERT(m_VulkanFunctions.vkGetPhysicalDeviceMemoryProperties != VMA_NULL); |
| 14377 | VMA_ASSERT(m_VulkanFunctions.vkAllocateMemory != VMA_NULL); |
| 14378 | VMA_ASSERT(m_VulkanFunctions.vkFreeMemory != VMA_NULL); |
| 14379 | VMA_ASSERT(m_VulkanFunctions.vkMapMemory != VMA_NULL); |
| 14380 | VMA_ASSERT(m_VulkanFunctions.vkUnmapMemory != VMA_NULL); |
| 14381 | VMA_ASSERT(m_VulkanFunctions.vkFlushMappedMemoryRanges != VMA_NULL); |
| 14382 | VMA_ASSERT(m_VulkanFunctions.vkInvalidateMappedMemoryRanges != VMA_NULL); |
| 14383 | VMA_ASSERT(m_VulkanFunctions.vkBindBufferMemory != VMA_NULL); |
| 14384 | VMA_ASSERT(m_VulkanFunctions.vkBindImageMemory != VMA_NULL); |
| 14385 | VMA_ASSERT(m_VulkanFunctions.vkGetBufferMemoryRequirements != VMA_NULL); |
| 14386 | VMA_ASSERT(m_VulkanFunctions.vkGetImageMemoryRequirements != VMA_NULL); |
| 14387 | VMA_ASSERT(m_VulkanFunctions.vkCreateBuffer != VMA_NULL); |
| 14388 | VMA_ASSERT(m_VulkanFunctions.vkDestroyBuffer != VMA_NULL); |
| 14389 | VMA_ASSERT(m_VulkanFunctions.vkCreateImage != VMA_NULL); |
| 14390 | VMA_ASSERT(m_VulkanFunctions.vkDestroyImage != VMA_NULL); |
| 14391 | VMA_ASSERT(m_VulkanFunctions.vkCmdCopyBuffer != VMA_NULL); |
| 14392 | |
| 14393 | #if VMA_DEDICATED_ALLOCATION || VMA_VULKAN_VERSION >= 1001000 |
| 14394 | if(m_VulkanApiVersion >= VK_MAKE_VERSION(1, 1, 0) || m_UseKhrDedicatedAllocation) |
| 14395 | { |
| 14396 | VMA_ASSERT(m_VulkanFunctions.vkGetBufferMemoryRequirements2KHR != VMA_NULL); |
| 14397 | VMA_ASSERT(m_VulkanFunctions.vkGetImageMemoryRequirements2KHR != VMA_NULL); |
| 14398 | } |
| 14399 | #endif |
| 14400 | |
| 14401 | #if VMA_BIND_MEMORY2 || VMA_VULKAN_VERSION >= 1001000 |
| 14402 | if(m_VulkanApiVersion >= VK_MAKE_VERSION(1, 1, 0) || m_UseKhrBindMemory2) |
| 14403 | { |
| 14404 | VMA_ASSERT(m_VulkanFunctions.vkBindBufferMemory2KHR != VMA_NULL); |
| 14405 | VMA_ASSERT(m_VulkanFunctions.vkBindImageMemory2KHR != VMA_NULL); |
| 14406 | } |
| 14407 | #endif |
| 14408 | |
| 14409 | #if VMA_MEMORY_BUDGET || VMA_VULKAN_VERSION >= 1001000 |
| 14410 | if(m_UseExtMemoryBudget || m_VulkanApiVersion >= VK_MAKE_VERSION(1, 1, 0)) |
| 14411 | { |
| 14412 | VMA_ASSERT(m_VulkanFunctions.vkGetPhysicalDeviceMemoryProperties2KHR != VMA_NULL); |
| 14413 | } |
| 14414 | #endif |
| 14415 | |
| 14416 | #if VMA_VULKAN_VERSION >= 1003000 |
| 14417 | if(m_VulkanApiVersion >= VK_MAKE_VERSION(1, 3, 0)) |
| 14418 | { |
| 14419 | VMA_ASSERT(m_VulkanFunctions.vkGetDeviceBufferMemoryRequirements != VMA_NULL); |
| 14420 | VMA_ASSERT(m_VulkanFunctions.vkGetDeviceImageMemoryRequirements != VMA_NULL); |
| 14421 | } |
| 14422 | #endif |
| 14423 | } |
| 14424 | |
| 14425 | VkDeviceSize VmaAllocator_T::CalcPreferredBlockSize(uint32_t memTypeIndex) |
| 14426 | { |
| 14427 | const uint32_t heapIndex = MemoryTypeIndexToHeapIndex(memTypeIndex); |
| 14428 | const VkDeviceSize heapSize = m_MemProps.memoryHeaps[heapIndex].size; |
| 14429 | const bool isSmallHeap = heapSize <= VMA_SMALL_HEAP_MAX_SIZE; |
| 14430 | return VmaAlignUp(val: isSmallHeap ? (heapSize / 8) : m_PreferredLargeHeapBlockSize, alignment: (VkDeviceSize)32); |
| 14431 | } |
| 14432 | |
| 14433 | VkResult VmaAllocator_T::AllocateMemoryOfType( |
| 14434 | VmaPool pool, |
| 14435 | VkDeviceSize size, |
| 14436 | VkDeviceSize alignment, |
| 14437 | bool dedicatedPreferred, |
| 14438 | VkBuffer dedicatedBuffer, |
| 14439 | VkImage dedicatedImage, |
| 14440 | VkFlags dedicatedBufferImageUsage, |
| 14441 | const VmaAllocationCreateInfo& createInfo, |
| 14442 | uint32_t memTypeIndex, |
| 14443 | VmaSuballocationType suballocType, |
| 14444 | VmaDedicatedAllocationList& dedicatedAllocations, |
| 14445 | VmaBlockVector& blockVector, |
| 14446 | size_t allocationCount, |
| 14447 | VmaAllocation* pAllocations) |
| 14448 | { |
| 14449 | VMA_ASSERT(pAllocations != VMA_NULL); |
| 14450 | VMA_DEBUG_LOG(" AllocateMemory: MemoryTypeIndex=%u, AllocationCount=%zu, Size=%llu" , memTypeIndex, allocationCount, size); |
| 14451 | |
| 14452 | VmaAllocationCreateInfo finalCreateInfo = createInfo; |
| 14453 | VkResult res = CalcMemTypeParams( |
| 14454 | outCreateInfo&: finalCreateInfo, |
| 14455 | memTypeIndex, |
| 14456 | size, |
| 14457 | allocationCount); |
| 14458 | if(res != VK_SUCCESS) |
| 14459 | return res; |
| 14460 | |
| 14461 | if((finalCreateInfo.flags & VMA_ALLOCATION_CREATE_DEDICATED_MEMORY_BIT) != 0) |
| 14462 | { |
| 14463 | return AllocateDedicatedMemory( |
| 14464 | pool, |
| 14465 | size, |
| 14466 | suballocType, |
| 14467 | dedicatedAllocations, |
| 14468 | memTypeIndex, |
| 14469 | map: (finalCreateInfo.flags & VMA_ALLOCATION_CREATE_MAPPED_BIT) != 0, |
| 14470 | isUserDataString: (finalCreateInfo.flags & VMA_ALLOCATION_CREATE_USER_DATA_COPY_STRING_BIT) != 0, |
| 14471 | isMappingAllowed: (finalCreateInfo.flags & |
| 14472 | (VMA_ALLOCATION_CREATE_HOST_ACCESS_SEQUENTIAL_WRITE_BIT | VMA_ALLOCATION_CREATE_HOST_ACCESS_RANDOM_BIT)) != 0, |
| 14473 | canAliasMemory: (finalCreateInfo.flags & VMA_ALLOCATION_CREATE_CAN_ALIAS_BIT) != 0, |
| 14474 | pUserData: finalCreateInfo.pUserData, |
| 14475 | priority: finalCreateInfo.priority, |
| 14476 | dedicatedBuffer, |
| 14477 | dedicatedImage, |
| 14478 | dedicatedBufferImageUsage, |
| 14479 | allocationCount, |
| 14480 | pAllocations, |
| 14481 | pNextChain: blockVector.GetAllocationNextPtr()); |
| 14482 | } |
| 14483 | else |
| 14484 | { |
| 14485 | const bool canAllocateDedicated = |
| 14486 | (finalCreateInfo.flags & VMA_ALLOCATION_CREATE_NEVER_ALLOCATE_BIT) == 0 && |
| 14487 | (pool == VK_NULL_HANDLE || !blockVector.HasExplicitBlockSize()); |
| 14488 | |
| 14489 | if(canAllocateDedicated) |
| 14490 | { |
| 14491 | // Heuristics: Allocate dedicated memory if requested size if greater than half of preferred block size. |
| 14492 | if(size > blockVector.GetPreferredBlockSize() / 2) |
| 14493 | { |
| 14494 | dedicatedPreferred = true; |
| 14495 | } |
| 14496 | // Protection against creating each allocation as dedicated when we reach or exceed heap size/budget, |
| 14497 | // which can quickly deplete maxMemoryAllocationCount: Don't prefer dedicated allocations when above |
| 14498 | // 3/4 of the maximum allocation count. |
| 14499 | if(m_DeviceMemoryCount.load() > m_PhysicalDeviceProperties.limits.maxMemoryAllocationCount * 3 / 4) |
| 14500 | { |
| 14501 | dedicatedPreferred = false; |
| 14502 | } |
| 14503 | |
| 14504 | if(dedicatedPreferred) |
| 14505 | { |
| 14506 | res = AllocateDedicatedMemory( |
| 14507 | pool, |
| 14508 | size, |
| 14509 | suballocType, |
| 14510 | dedicatedAllocations, |
| 14511 | memTypeIndex, |
| 14512 | map: (finalCreateInfo.flags & VMA_ALLOCATION_CREATE_MAPPED_BIT) != 0, |
| 14513 | isUserDataString: (finalCreateInfo.flags & VMA_ALLOCATION_CREATE_USER_DATA_COPY_STRING_BIT) != 0, |
| 14514 | isMappingAllowed: (finalCreateInfo.flags & |
| 14515 | (VMA_ALLOCATION_CREATE_HOST_ACCESS_SEQUENTIAL_WRITE_BIT | VMA_ALLOCATION_CREATE_HOST_ACCESS_RANDOM_BIT)) != 0, |
| 14516 | canAliasMemory: (finalCreateInfo.flags & VMA_ALLOCATION_CREATE_CAN_ALIAS_BIT) != 0, |
| 14517 | pUserData: finalCreateInfo.pUserData, |
| 14518 | priority: finalCreateInfo.priority, |
| 14519 | dedicatedBuffer, |
| 14520 | dedicatedImage, |
| 14521 | dedicatedBufferImageUsage, |
| 14522 | allocationCount, |
| 14523 | pAllocations, |
| 14524 | pNextChain: blockVector.GetAllocationNextPtr()); |
| 14525 | if(res == VK_SUCCESS) |
| 14526 | { |
| 14527 | // Succeeded: AllocateDedicatedMemory function already filld pMemory, nothing more to do here. |
| 14528 | VMA_DEBUG_LOG(" Allocated as DedicatedMemory" ); |
| 14529 | return VK_SUCCESS; |
| 14530 | } |
| 14531 | } |
| 14532 | } |
| 14533 | |
| 14534 | res = blockVector.Allocate( |
| 14535 | size, |
| 14536 | alignment, |
| 14537 | createInfo: finalCreateInfo, |
| 14538 | suballocType, |
| 14539 | allocationCount, |
| 14540 | pAllocations); |
| 14541 | if(res == VK_SUCCESS) |
| 14542 | return VK_SUCCESS; |
| 14543 | |
| 14544 | // Try dedicated memory. |
| 14545 | if(canAllocateDedicated && !dedicatedPreferred) |
| 14546 | { |
| 14547 | res = AllocateDedicatedMemory( |
| 14548 | pool, |
| 14549 | size, |
| 14550 | suballocType, |
| 14551 | dedicatedAllocations, |
| 14552 | memTypeIndex, |
| 14553 | map: (finalCreateInfo.flags & VMA_ALLOCATION_CREATE_MAPPED_BIT) != 0, |
| 14554 | isUserDataString: (finalCreateInfo.flags & VMA_ALLOCATION_CREATE_USER_DATA_COPY_STRING_BIT) != 0, |
| 14555 | isMappingAllowed: (finalCreateInfo.flags & |
| 14556 | (VMA_ALLOCATION_CREATE_HOST_ACCESS_SEQUENTIAL_WRITE_BIT | VMA_ALLOCATION_CREATE_HOST_ACCESS_RANDOM_BIT)) != 0, |
| 14557 | canAliasMemory: (finalCreateInfo.flags & VMA_ALLOCATION_CREATE_CAN_ALIAS_BIT) != 0, |
| 14558 | pUserData: finalCreateInfo.pUserData, |
| 14559 | priority: finalCreateInfo.priority, |
| 14560 | dedicatedBuffer, |
| 14561 | dedicatedImage, |
| 14562 | dedicatedBufferImageUsage, |
| 14563 | allocationCount, |
| 14564 | pAllocations, |
| 14565 | pNextChain: blockVector.GetAllocationNextPtr()); |
| 14566 | if(res == VK_SUCCESS) |
| 14567 | { |
| 14568 | // Succeeded: AllocateDedicatedMemory function already filld pMemory, nothing more to do here. |
| 14569 | VMA_DEBUG_LOG(" Allocated as DedicatedMemory" ); |
| 14570 | return VK_SUCCESS; |
| 14571 | } |
| 14572 | } |
| 14573 | // Everything failed: Return error code. |
| 14574 | VMA_DEBUG_LOG(" vkAllocateMemory FAILED" ); |
| 14575 | return res; |
| 14576 | } |
| 14577 | } |
| 14578 | |
| 14579 | VkResult VmaAllocator_T::AllocateDedicatedMemory( |
| 14580 | VmaPool pool, |
| 14581 | VkDeviceSize size, |
| 14582 | VmaSuballocationType suballocType, |
| 14583 | VmaDedicatedAllocationList& dedicatedAllocations, |
| 14584 | uint32_t memTypeIndex, |
| 14585 | bool map, |
| 14586 | bool isUserDataString, |
| 14587 | bool isMappingAllowed, |
| 14588 | bool canAliasMemory, |
| 14589 | void* pUserData, |
| 14590 | float priority, |
| 14591 | VkBuffer dedicatedBuffer, |
| 14592 | VkImage dedicatedImage, |
| 14593 | VkFlags dedicatedBufferImageUsage, |
| 14594 | size_t allocationCount, |
| 14595 | VmaAllocation* pAllocations, |
| 14596 | const void* pNextChain) |
| 14597 | { |
| 14598 | VMA_ASSERT(allocationCount > 0 && pAllocations); |
| 14599 | |
| 14600 | VkMemoryAllocateInfo allocInfo = { .sType: VK_STRUCTURE_TYPE_MEMORY_ALLOCATE_INFO }; |
| 14601 | allocInfo.memoryTypeIndex = memTypeIndex; |
| 14602 | allocInfo.allocationSize = size; |
| 14603 | allocInfo.pNext = pNextChain; |
| 14604 | |
| 14605 | #if VMA_DEDICATED_ALLOCATION || VMA_VULKAN_VERSION >= 1001000 |
| 14606 | VkMemoryDedicatedAllocateInfoKHR dedicatedAllocInfo = { .sType: VK_STRUCTURE_TYPE_MEMORY_DEDICATED_ALLOCATE_INFO_KHR }; |
| 14607 | if(!canAliasMemory) |
| 14608 | { |
| 14609 | if(m_UseKhrDedicatedAllocation || m_VulkanApiVersion >= VK_MAKE_VERSION(1, 1, 0)) |
| 14610 | { |
| 14611 | if(dedicatedBuffer != VK_NULL_HANDLE) |
| 14612 | { |
| 14613 | VMA_ASSERT(dedicatedImage == VK_NULL_HANDLE); |
| 14614 | dedicatedAllocInfo.buffer = dedicatedBuffer; |
| 14615 | VmaPnextChainPushFront(mainStruct: &allocInfo, newStruct: &dedicatedAllocInfo); |
| 14616 | } |
| 14617 | else if(dedicatedImage != VK_NULL_HANDLE) |
| 14618 | { |
| 14619 | dedicatedAllocInfo.image = dedicatedImage; |
| 14620 | VmaPnextChainPushFront(mainStruct: &allocInfo, newStruct: &dedicatedAllocInfo); |
| 14621 | } |
| 14622 | } |
| 14623 | } |
| 14624 | #endif // #if VMA_DEDICATED_ALLOCATION || VMA_VULKAN_VERSION >= 1001000 |
| 14625 | |
| 14626 | #if VMA_BUFFER_DEVICE_ADDRESS |
| 14627 | VkMemoryAllocateFlagsInfoKHR allocFlagsInfo = { .sType: VK_STRUCTURE_TYPE_MEMORY_ALLOCATE_FLAGS_INFO_KHR }; |
| 14628 | if(m_UseKhrBufferDeviceAddress) |
| 14629 | { |
| 14630 | bool canContainBufferWithDeviceAddress = true; |
| 14631 | if(dedicatedBuffer != VK_NULL_HANDLE) |
| 14632 | { |
| 14633 | canContainBufferWithDeviceAddress = dedicatedBufferImageUsage == UINT32_MAX || // Usage flags unknown |
| 14634 | (dedicatedBufferImageUsage & VK_BUFFER_USAGE_SHADER_DEVICE_ADDRESS_BIT_EXT) != 0; |
| 14635 | } |
| 14636 | else if(dedicatedImage != VK_NULL_HANDLE) |
| 14637 | { |
| 14638 | canContainBufferWithDeviceAddress = false; |
| 14639 | } |
| 14640 | if(canContainBufferWithDeviceAddress) |
| 14641 | { |
| 14642 | allocFlagsInfo.flags = VK_MEMORY_ALLOCATE_DEVICE_ADDRESS_BIT_KHR; |
| 14643 | VmaPnextChainPushFront(mainStruct: &allocInfo, newStruct: &allocFlagsInfo); |
| 14644 | } |
| 14645 | } |
| 14646 | #endif // #if VMA_BUFFER_DEVICE_ADDRESS |
| 14647 | |
| 14648 | #if VMA_MEMORY_PRIORITY |
| 14649 | VkMemoryPriorityAllocateInfoEXT priorityInfo = { .sType: VK_STRUCTURE_TYPE_MEMORY_PRIORITY_ALLOCATE_INFO_EXT }; |
| 14650 | if(m_UseExtMemoryPriority) |
| 14651 | { |
| 14652 | VMA_ASSERT(priority >= 0.f && priority <= 1.f); |
| 14653 | priorityInfo.priority = priority; |
| 14654 | VmaPnextChainPushFront(mainStruct: &allocInfo, newStruct: &priorityInfo); |
| 14655 | } |
| 14656 | #endif // #if VMA_MEMORY_PRIORITY |
| 14657 | |
| 14658 | #if VMA_EXTERNAL_MEMORY |
| 14659 | // Attach VkExportMemoryAllocateInfoKHR if necessary. |
| 14660 | VkExportMemoryAllocateInfoKHR exportMemoryAllocInfo = { .sType: VK_STRUCTURE_TYPE_EXPORT_MEMORY_ALLOCATE_INFO_KHR }; |
| 14661 | exportMemoryAllocInfo.handleTypes = GetExternalMemoryHandleTypeFlags(memTypeIndex); |
| 14662 | if(exportMemoryAllocInfo.handleTypes != 0) |
| 14663 | { |
| 14664 | VmaPnextChainPushFront(mainStruct: &allocInfo, newStruct: &exportMemoryAllocInfo); |
| 14665 | } |
| 14666 | #endif // #if VMA_EXTERNAL_MEMORY |
| 14667 | |
| 14668 | size_t allocIndex; |
| 14669 | VkResult res = VK_SUCCESS; |
| 14670 | for(allocIndex = 0; allocIndex < allocationCount; ++allocIndex) |
| 14671 | { |
| 14672 | res = AllocateDedicatedMemoryPage( |
| 14673 | pool, |
| 14674 | size, |
| 14675 | suballocType, |
| 14676 | memTypeIndex, |
| 14677 | allocInfo, |
| 14678 | map, |
| 14679 | isUserDataString, |
| 14680 | isMappingAllowed, |
| 14681 | pUserData, |
| 14682 | pAllocation: pAllocations + allocIndex); |
| 14683 | if(res != VK_SUCCESS) |
| 14684 | { |
| 14685 | break; |
| 14686 | } |
| 14687 | } |
| 14688 | |
| 14689 | if(res == VK_SUCCESS) |
| 14690 | { |
| 14691 | for (allocIndex = 0; allocIndex < allocationCount; ++allocIndex) |
| 14692 | { |
| 14693 | dedicatedAllocations.Register(alloc: pAllocations[allocIndex]); |
| 14694 | } |
| 14695 | VMA_DEBUG_LOG(" Allocated DedicatedMemory Count=%zu, MemoryTypeIndex=#%u" , allocationCount, memTypeIndex); |
| 14696 | } |
| 14697 | else |
| 14698 | { |
| 14699 | // Free all already created allocations. |
| 14700 | while(allocIndex--) |
| 14701 | { |
| 14702 | VmaAllocation currAlloc = pAllocations[allocIndex]; |
| 14703 | VkDeviceMemory hMemory = currAlloc->GetMemory(); |
| 14704 | |
| 14705 | /* |
| 14706 | There is no need to call this, because Vulkan spec allows to skip vkUnmapMemory |
| 14707 | before vkFreeMemory. |
| 14708 | |
| 14709 | if(currAlloc->GetMappedData() != VMA_NULL) |
| 14710 | { |
| 14711 | (*m_VulkanFunctions.vkUnmapMemory)(m_hDevice, hMemory); |
| 14712 | } |
| 14713 | */ |
| 14714 | |
| 14715 | FreeVulkanMemory(memoryType: memTypeIndex, size: currAlloc->GetSize(), hMemory); |
| 14716 | m_Budget.RemoveAllocation(heapIndex: MemoryTypeIndexToHeapIndex(memTypeIndex), allocationSize: currAlloc->GetSize()); |
| 14717 | m_AllocationObjectAllocator.Free(hAlloc: currAlloc); |
| 14718 | } |
| 14719 | |
| 14720 | memset(s: pAllocations, c: 0, n: sizeof(VmaAllocation) * allocationCount); |
| 14721 | } |
| 14722 | |
| 14723 | return res; |
| 14724 | } |
| 14725 | |
| 14726 | VkResult VmaAllocator_T::AllocateDedicatedMemoryPage( |
| 14727 | VmaPool pool, |
| 14728 | VkDeviceSize size, |
| 14729 | VmaSuballocationType suballocType, |
| 14730 | uint32_t memTypeIndex, |
| 14731 | const VkMemoryAllocateInfo& allocInfo, |
| 14732 | bool map, |
| 14733 | bool isUserDataString, |
| 14734 | bool isMappingAllowed, |
| 14735 | void* pUserData, |
| 14736 | VmaAllocation* pAllocation) |
| 14737 | { |
| 14738 | VkDeviceMemory hMemory = VK_NULL_HANDLE; |
| 14739 | VkResult res = AllocateVulkanMemory(pAllocateInfo: &allocInfo, pMemory: &hMemory); |
| 14740 | if(res < 0) |
| 14741 | { |
| 14742 | VMA_DEBUG_LOG(" vkAllocateMemory FAILED" ); |
| 14743 | return res; |
| 14744 | } |
| 14745 | |
| 14746 | void* pMappedData = VMA_NULL; |
| 14747 | if(map) |
| 14748 | { |
| 14749 | res = (*m_VulkanFunctions.vkMapMemory)( |
| 14750 | m_hDevice, |
| 14751 | hMemory, |
| 14752 | 0, |
| 14753 | VK_WHOLE_SIZE, |
| 14754 | 0, |
| 14755 | &pMappedData); |
| 14756 | if(res < 0) |
| 14757 | { |
| 14758 | VMA_DEBUG_LOG(" vkMapMemory FAILED" ); |
| 14759 | FreeVulkanMemory(memoryType: memTypeIndex, size, hMemory); |
| 14760 | return res; |
| 14761 | } |
| 14762 | } |
| 14763 | |
| 14764 | *pAllocation = m_AllocationObjectAllocator.Allocate(args&: isMappingAllowed); |
| 14765 | (*pAllocation)->InitDedicatedAllocation(hParentPool: pool, memoryTypeIndex: memTypeIndex, hMemory, suballocationType: suballocType, pMappedData, size); |
| 14766 | if (isUserDataString) |
| 14767 | (*pAllocation)->SetName(hAllocator: this, pName: (const char*)pUserData); |
| 14768 | else |
| 14769 | (*pAllocation)->SetUserData(hAllocator: this, pUserData); |
| 14770 | m_Budget.AddAllocation(heapIndex: MemoryTypeIndexToHeapIndex(memTypeIndex), allocationSize: size); |
| 14771 | if(VMA_DEBUG_INITIALIZE_ALLOCATIONS) |
| 14772 | { |
| 14773 | FillAllocation(hAllocation: *pAllocation, pattern: VMA_ALLOCATION_FILL_PATTERN_CREATED); |
| 14774 | } |
| 14775 | |
| 14776 | return VK_SUCCESS; |
| 14777 | } |
| 14778 | |
| 14779 | void VmaAllocator_T::GetBufferMemoryRequirements( |
| 14780 | VkBuffer hBuffer, |
| 14781 | VkMemoryRequirements& memReq, |
| 14782 | bool& requiresDedicatedAllocation, |
| 14783 | bool& prefersDedicatedAllocation) const |
| 14784 | { |
| 14785 | #if VMA_DEDICATED_ALLOCATION || VMA_VULKAN_VERSION >= 1001000 |
| 14786 | if(m_UseKhrDedicatedAllocation || m_VulkanApiVersion >= VK_MAKE_VERSION(1, 1, 0)) |
| 14787 | { |
| 14788 | VkBufferMemoryRequirementsInfo2KHR memReqInfo = { .sType: VK_STRUCTURE_TYPE_BUFFER_MEMORY_REQUIREMENTS_INFO_2_KHR }; |
| 14789 | memReqInfo.buffer = hBuffer; |
| 14790 | |
| 14791 | VkMemoryDedicatedRequirementsKHR memDedicatedReq = { .sType: VK_STRUCTURE_TYPE_MEMORY_DEDICATED_REQUIREMENTS_KHR }; |
| 14792 | |
| 14793 | VkMemoryRequirements2KHR memReq2 = { .sType: VK_STRUCTURE_TYPE_MEMORY_REQUIREMENTS_2_KHR }; |
| 14794 | VmaPnextChainPushFront(mainStruct: &memReq2, newStruct: &memDedicatedReq); |
| 14795 | |
| 14796 | (*m_VulkanFunctions.vkGetBufferMemoryRequirements2KHR)(m_hDevice, &memReqInfo, &memReq2); |
| 14797 | |
| 14798 | memReq = memReq2.memoryRequirements; |
| 14799 | requiresDedicatedAllocation = (memDedicatedReq.requiresDedicatedAllocation != VK_FALSE); |
| 14800 | prefersDedicatedAllocation = (memDedicatedReq.prefersDedicatedAllocation != VK_FALSE); |
| 14801 | } |
| 14802 | else |
| 14803 | #endif // #if VMA_DEDICATED_ALLOCATION || VMA_VULKAN_VERSION >= 1001000 |
| 14804 | { |
| 14805 | (*m_VulkanFunctions.vkGetBufferMemoryRequirements)(m_hDevice, hBuffer, &memReq); |
| 14806 | requiresDedicatedAllocation = false; |
| 14807 | prefersDedicatedAllocation = false; |
| 14808 | } |
| 14809 | } |
| 14810 | |
| 14811 | void VmaAllocator_T::GetImageMemoryRequirements( |
| 14812 | VkImage hImage, |
| 14813 | VkMemoryRequirements& memReq, |
| 14814 | bool& requiresDedicatedAllocation, |
| 14815 | bool& prefersDedicatedAllocation) const |
| 14816 | { |
| 14817 | #if VMA_DEDICATED_ALLOCATION || VMA_VULKAN_VERSION >= 1001000 |
| 14818 | if(m_UseKhrDedicatedAllocation || m_VulkanApiVersion >= VK_MAKE_VERSION(1, 1, 0)) |
| 14819 | { |
| 14820 | VkImageMemoryRequirementsInfo2KHR memReqInfo = { .sType: VK_STRUCTURE_TYPE_IMAGE_MEMORY_REQUIREMENTS_INFO_2_KHR }; |
| 14821 | memReqInfo.image = hImage; |
| 14822 | |
| 14823 | VkMemoryDedicatedRequirementsKHR memDedicatedReq = { .sType: VK_STRUCTURE_TYPE_MEMORY_DEDICATED_REQUIREMENTS_KHR }; |
| 14824 | |
| 14825 | VkMemoryRequirements2KHR memReq2 = { .sType: VK_STRUCTURE_TYPE_MEMORY_REQUIREMENTS_2_KHR }; |
| 14826 | VmaPnextChainPushFront(mainStruct: &memReq2, newStruct: &memDedicatedReq); |
| 14827 | |
| 14828 | (*m_VulkanFunctions.vkGetImageMemoryRequirements2KHR)(m_hDevice, &memReqInfo, &memReq2); |
| 14829 | |
| 14830 | memReq = memReq2.memoryRequirements; |
| 14831 | requiresDedicatedAllocation = (memDedicatedReq.requiresDedicatedAllocation != VK_FALSE); |
| 14832 | prefersDedicatedAllocation = (memDedicatedReq.prefersDedicatedAllocation != VK_FALSE); |
| 14833 | } |
| 14834 | else |
| 14835 | #endif // #if VMA_DEDICATED_ALLOCATION || VMA_VULKAN_VERSION >= 1001000 |
| 14836 | { |
| 14837 | (*m_VulkanFunctions.vkGetImageMemoryRequirements)(m_hDevice, hImage, &memReq); |
| 14838 | requiresDedicatedAllocation = false; |
| 14839 | prefersDedicatedAllocation = false; |
| 14840 | } |
| 14841 | } |
| 14842 | |
| 14843 | VkResult VmaAllocator_T::FindMemoryTypeIndex( |
| 14844 | uint32_t memoryTypeBits, |
| 14845 | const VmaAllocationCreateInfo* pAllocationCreateInfo, |
| 14846 | VkFlags bufImgUsage, |
| 14847 | uint32_t* pMemoryTypeIndex) const |
| 14848 | { |
| 14849 | memoryTypeBits &= GetGlobalMemoryTypeBits(); |
| 14850 | |
| 14851 | if(pAllocationCreateInfo->memoryTypeBits != 0) |
| 14852 | { |
| 14853 | memoryTypeBits &= pAllocationCreateInfo->memoryTypeBits; |
| 14854 | } |
| 14855 | |
| 14856 | VkMemoryPropertyFlags requiredFlags = 0, preferredFlags = 0, notPreferredFlags = 0; |
| 14857 | if(!FindMemoryPreferences( |
| 14858 | isIntegratedGPU: IsIntegratedGpu(), |
| 14859 | allocCreateInfo: *pAllocationCreateInfo, |
| 14860 | bufImgUsage, |
| 14861 | outRequiredFlags&: requiredFlags, outPreferredFlags&: preferredFlags, outNotPreferredFlags&: notPreferredFlags)) |
| 14862 | { |
| 14863 | return VK_ERROR_FEATURE_NOT_PRESENT; |
| 14864 | } |
| 14865 | |
| 14866 | *pMemoryTypeIndex = UINT32_MAX; |
| 14867 | uint32_t minCost = UINT32_MAX; |
| 14868 | for(uint32_t memTypeIndex = 0, memTypeBit = 1; |
| 14869 | memTypeIndex < GetMemoryTypeCount(); |
| 14870 | ++memTypeIndex, memTypeBit <<= 1) |
| 14871 | { |
| 14872 | // This memory type is acceptable according to memoryTypeBits bitmask. |
| 14873 | if((memTypeBit & memoryTypeBits) != 0) |
| 14874 | { |
| 14875 | const VkMemoryPropertyFlags currFlags = |
| 14876 | m_MemProps.memoryTypes[memTypeIndex].propertyFlags; |
| 14877 | // This memory type contains requiredFlags. |
| 14878 | if((requiredFlags & ~currFlags) == 0) |
| 14879 | { |
| 14880 | // Calculate cost as number of bits from preferredFlags not present in this memory type. |
| 14881 | uint32_t currCost = VMA_COUNT_BITS_SET(preferredFlags & ~currFlags) + |
| 14882 | VMA_COUNT_BITS_SET(currFlags & notPreferredFlags); |
| 14883 | // Remember memory type with lowest cost. |
| 14884 | if(currCost < minCost) |
| 14885 | { |
| 14886 | *pMemoryTypeIndex = memTypeIndex; |
| 14887 | if(currCost == 0) |
| 14888 | { |
| 14889 | return VK_SUCCESS; |
| 14890 | } |
| 14891 | minCost = currCost; |
| 14892 | } |
| 14893 | } |
| 14894 | } |
| 14895 | } |
| 14896 | return (*pMemoryTypeIndex != UINT32_MAX) ? VK_SUCCESS : VK_ERROR_FEATURE_NOT_PRESENT; |
| 14897 | } |
| 14898 | |
| 14899 | VkResult VmaAllocator_T::CalcMemTypeParams( |
| 14900 | VmaAllocationCreateInfo& inoutCreateInfo, |
| 14901 | uint32_t memTypeIndex, |
| 14902 | VkDeviceSize size, |
| 14903 | size_t allocationCount) |
| 14904 | { |
| 14905 | // If memory type is not HOST_VISIBLE, disable MAPPED. |
| 14906 | if((inoutCreateInfo.flags & VMA_ALLOCATION_CREATE_MAPPED_BIT) != 0 && |
| 14907 | (m_MemProps.memoryTypes[memTypeIndex].propertyFlags & VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT) == 0) |
| 14908 | { |
| 14909 | inoutCreateInfo.flags &= ~VMA_ALLOCATION_CREATE_MAPPED_BIT; |
| 14910 | } |
| 14911 | |
| 14912 | if((inoutCreateInfo.flags & VMA_ALLOCATION_CREATE_DEDICATED_MEMORY_BIT) != 0 && |
| 14913 | (inoutCreateInfo.flags & VMA_ALLOCATION_CREATE_WITHIN_BUDGET_BIT) != 0) |
| 14914 | { |
| 14915 | const uint32_t heapIndex = MemoryTypeIndexToHeapIndex(memTypeIndex); |
| 14916 | VmaBudget heapBudget = {}; |
| 14917 | GetHeapBudgets(outBudgets: &heapBudget, firstHeap: heapIndex, heapCount: 1); |
| 14918 | if(heapBudget.usage + size * allocationCount > heapBudget.budget) |
| 14919 | { |
| 14920 | return VK_ERROR_OUT_OF_DEVICE_MEMORY; |
| 14921 | } |
| 14922 | } |
| 14923 | return VK_SUCCESS; |
| 14924 | } |
| 14925 | |
| 14926 | VkResult VmaAllocator_T::CalcAllocationParams( |
| 14927 | VmaAllocationCreateInfo& inoutCreateInfo, |
| 14928 | bool dedicatedRequired, |
| 14929 | bool dedicatedPreferred) |
| 14930 | { |
| 14931 | VMA_ASSERT((inoutCreateInfo.flags & |
| 14932 | (VMA_ALLOCATION_CREATE_HOST_ACCESS_SEQUENTIAL_WRITE_BIT | VMA_ALLOCATION_CREATE_HOST_ACCESS_RANDOM_BIT)) != |
| 14933 | (VMA_ALLOCATION_CREATE_HOST_ACCESS_SEQUENTIAL_WRITE_BIT | VMA_ALLOCATION_CREATE_HOST_ACCESS_RANDOM_BIT) && |
| 14934 | "Specifying both flags VMA_ALLOCATION_CREATE_HOST_ACCESS_SEQUENTIAL_WRITE_BIT and VMA_ALLOCATION_CREATE_HOST_ACCESS_RANDOM_BIT is incorrect." ); |
| 14935 | VMA_ASSERT((((inoutCreateInfo.flags & VMA_ALLOCATION_CREATE_HOST_ACCESS_ALLOW_TRANSFER_INSTEAD_BIT) == 0 || |
| 14936 | (inoutCreateInfo.flags & (VMA_ALLOCATION_CREATE_HOST_ACCESS_SEQUENTIAL_WRITE_BIT | VMA_ALLOCATION_CREATE_HOST_ACCESS_RANDOM_BIT)) != 0)) && |
| 14937 | "Specifying VMA_ALLOCATION_CREATE_HOST_ACCESS_ALLOW_TRANSFER_INSTEAD_BIT requires also VMA_ALLOCATION_CREATE_HOST_ACCESS_SEQUENTIAL_WRITE_BIT or VMA_ALLOCATION_CREATE_HOST_ACCESS_RANDOM_BIT." ); |
| 14938 | if(inoutCreateInfo.usage == VMA_MEMORY_USAGE_AUTO || inoutCreateInfo.usage == VMA_MEMORY_USAGE_AUTO_PREFER_DEVICE || inoutCreateInfo.usage == VMA_MEMORY_USAGE_AUTO_PREFER_HOST) |
| 14939 | { |
| 14940 | if((inoutCreateInfo.flags & VMA_ALLOCATION_CREATE_MAPPED_BIT) != 0) |
| 14941 | { |
| 14942 | VMA_ASSERT((inoutCreateInfo.flags & (VMA_ALLOCATION_CREATE_HOST_ACCESS_SEQUENTIAL_WRITE_BIT | VMA_ALLOCATION_CREATE_HOST_ACCESS_RANDOM_BIT)) != 0 && |
| 14943 | "When using VMA_ALLOCATION_CREATE_MAPPED_BIT and usage = VMA_MEMORY_USAGE_AUTO*, you must also specify VMA_ALLOCATION_CREATE_HOST_ACCESS_SEQUENTIAL_WRITE_BIT or VMA_ALLOCATION_CREATE_HOST_ACCESS_RANDOM_BIT." ); |
| 14944 | } |
| 14945 | } |
| 14946 | |
| 14947 | // If memory is lazily allocated, it should be always dedicated. |
| 14948 | if(dedicatedRequired || |
| 14949 | inoutCreateInfo.usage == VMA_MEMORY_USAGE_GPU_LAZILY_ALLOCATED) |
| 14950 | { |
| 14951 | inoutCreateInfo.flags |= VMA_ALLOCATION_CREATE_DEDICATED_MEMORY_BIT; |
| 14952 | } |
| 14953 | |
| 14954 | if(inoutCreateInfo.pool != VK_NULL_HANDLE) |
| 14955 | { |
| 14956 | if(inoutCreateInfo.pool->m_BlockVector.HasExplicitBlockSize() && |
| 14957 | (inoutCreateInfo.flags & VMA_ALLOCATION_CREATE_DEDICATED_MEMORY_BIT) != 0) |
| 14958 | { |
| 14959 | VMA_ASSERT(0 && "Specifying VMA_ALLOCATION_CREATE_DEDICATED_MEMORY_BIT while current custom pool doesn't support dedicated allocations." ); |
| 14960 | return VK_ERROR_FEATURE_NOT_PRESENT; |
| 14961 | } |
| 14962 | inoutCreateInfo.priority = inoutCreateInfo.pool->m_BlockVector.GetPriority(); |
| 14963 | } |
| 14964 | |
| 14965 | if((inoutCreateInfo.flags & VMA_ALLOCATION_CREATE_DEDICATED_MEMORY_BIT) != 0 && |
| 14966 | (inoutCreateInfo.flags & VMA_ALLOCATION_CREATE_NEVER_ALLOCATE_BIT) != 0) |
| 14967 | { |
| 14968 | VMA_ASSERT(0 && "Specifying VMA_ALLOCATION_CREATE_DEDICATED_MEMORY_BIT together with VMA_ALLOCATION_CREATE_NEVER_ALLOCATE_BIT makes no sense." ); |
| 14969 | return VK_ERROR_FEATURE_NOT_PRESENT; |
| 14970 | } |
| 14971 | |
| 14972 | if(VMA_DEBUG_ALWAYS_DEDICATED_MEMORY && |
| 14973 | (inoutCreateInfo.flags & VMA_ALLOCATION_CREATE_NEVER_ALLOCATE_BIT) != 0) |
| 14974 | { |
| 14975 | inoutCreateInfo.flags |= VMA_ALLOCATION_CREATE_DEDICATED_MEMORY_BIT; |
| 14976 | } |
| 14977 | |
| 14978 | // Non-auto USAGE values imply HOST_ACCESS flags. |
| 14979 | // And so does VMA_MEMORY_USAGE_UNKNOWN because it is used with custom pools. |
| 14980 | // Which specific flag is used doesn't matter. They change things only when used with VMA_MEMORY_USAGE_AUTO*. |
| 14981 | // Otherwise they just protect from assert on mapping. |
| 14982 | if(inoutCreateInfo.usage != VMA_MEMORY_USAGE_AUTO && |
| 14983 | inoutCreateInfo.usage != VMA_MEMORY_USAGE_AUTO_PREFER_DEVICE && |
| 14984 | inoutCreateInfo.usage != VMA_MEMORY_USAGE_AUTO_PREFER_HOST) |
| 14985 | { |
| 14986 | if((inoutCreateInfo.flags & (VMA_ALLOCATION_CREATE_HOST_ACCESS_SEQUENTIAL_WRITE_BIT | VMA_ALLOCATION_CREATE_HOST_ACCESS_RANDOM_BIT)) == 0) |
| 14987 | { |
| 14988 | inoutCreateInfo.flags |= VMA_ALLOCATION_CREATE_HOST_ACCESS_RANDOM_BIT; |
| 14989 | } |
| 14990 | } |
| 14991 | |
| 14992 | return VK_SUCCESS; |
| 14993 | } |
| 14994 | |
| 14995 | VkResult VmaAllocator_T::AllocateMemory( |
| 14996 | const VkMemoryRequirements& vkMemReq, |
| 14997 | bool requiresDedicatedAllocation, |
| 14998 | bool prefersDedicatedAllocation, |
| 14999 | VkBuffer dedicatedBuffer, |
| 15000 | VkImage dedicatedImage, |
| 15001 | VkFlags dedicatedBufferImageUsage, |
| 15002 | const VmaAllocationCreateInfo& createInfo, |
| 15003 | VmaSuballocationType suballocType, |
| 15004 | size_t allocationCount, |
| 15005 | VmaAllocation* pAllocations) |
| 15006 | { |
| 15007 | memset(s: pAllocations, c: 0, n: sizeof(VmaAllocation) * allocationCount); |
| 15008 | |
| 15009 | VMA_ASSERT(VmaIsPow2(vkMemReq.alignment)); |
| 15010 | |
| 15011 | if(vkMemReq.size == 0) |
| 15012 | { |
| 15013 | return VK_ERROR_INITIALIZATION_FAILED; |
| 15014 | } |
| 15015 | |
| 15016 | VmaAllocationCreateInfo createInfoFinal = createInfo; |
| 15017 | VkResult res = CalcAllocationParams(inoutCreateInfo&: createInfoFinal, dedicatedRequired: requiresDedicatedAllocation, dedicatedPreferred: prefersDedicatedAllocation); |
| 15018 | if(res != VK_SUCCESS) |
| 15019 | return res; |
| 15020 | |
| 15021 | if(createInfoFinal.pool != VK_NULL_HANDLE) |
| 15022 | { |
| 15023 | VmaBlockVector& blockVector = createInfoFinal.pool->m_BlockVector; |
| 15024 | return AllocateMemoryOfType( |
| 15025 | pool: createInfoFinal.pool, |
| 15026 | size: vkMemReq.size, |
| 15027 | alignment: vkMemReq.alignment, |
| 15028 | dedicatedPreferred: prefersDedicatedAllocation, |
| 15029 | dedicatedBuffer, |
| 15030 | dedicatedImage, |
| 15031 | dedicatedBufferImageUsage, |
| 15032 | createInfo: createInfoFinal, |
| 15033 | memTypeIndex: blockVector.GetMemoryTypeIndex(), |
| 15034 | suballocType, |
| 15035 | dedicatedAllocations&: createInfoFinal.pool->m_DedicatedAllocations, |
| 15036 | blockVector, |
| 15037 | allocationCount, |
| 15038 | pAllocations); |
| 15039 | } |
| 15040 | else |
| 15041 | { |
| 15042 | // Bit mask of memory Vulkan types acceptable for this allocation. |
| 15043 | uint32_t memoryTypeBits = vkMemReq.memoryTypeBits; |
| 15044 | uint32_t memTypeIndex = UINT32_MAX; |
| 15045 | res = FindMemoryTypeIndex(memoryTypeBits, pAllocationCreateInfo: &createInfoFinal, bufImgUsage: dedicatedBufferImageUsage, pMemoryTypeIndex: &memTypeIndex); |
| 15046 | // Can't find any single memory type matching requirements. res is VK_ERROR_FEATURE_NOT_PRESENT. |
| 15047 | if(res != VK_SUCCESS) |
| 15048 | return res; |
| 15049 | do |
| 15050 | { |
| 15051 | VmaBlockVector* blockVector = m_pBlockVectors[memTypeIndex]; |
| 15052 | VMA_ASSERT(blockVector && "Trying to use unsupported memory type!" ); |
| 15053 | res = AllocateMemoryOfType( |
| 15054 | VK_NULL_HANDLE, |
| 15055 | size: vkMemReq.size, |
| 15056 | alignment: vkMemReq.alignment, |
| 15057 | dedicatedPreferred: requiresDedicatedAllocation || prefersDedicatedAllocation, |
| 15058 | dedicatedBuffer, |
| 15059 | dedicatedImage, |
| 15060 | dedicatedBufferImageUsage, |
| 15061 | createInfo: createInfoFinal, |
| 15062 | memTypeIndex, |
| 15063 | suballocType, |
| 15064 | dedicatedAllocations&: m_DedicatedAllocations[memTypeIndex], |
| 15065 | blockVector&: *blockVector, |
| 15066 | allocationCount, |
| 15067 | pAllocations); |
| 15068 | // Allocation succeeded |
| 15069 | if(res == VK_SUCCESS) |
| 15070 | return VK_SUCCESS; |
| 15071 | |
| 15072 | // Remove old memTypeIndex from list of possibilities. |
| 15073 | memoryTypeBits &= ~(1u << memTypeIndex); |
| 15074 | // Find alternative memTypeIndex. |
| 15075 | res = FindMemoryTypeIndex(memoryTypeBits, pAllocationCreateInfo: &createInfoFinal, bufImgUsage: dedicatedBufferImageUsage, pMemoryTypeIndex: &memTypeIndex); |
| 15076 | } while(res == VK_SUCCESS); |
| 15077 | |
| 15078 | // No other matching memory type index could be found. |
| 15079 | // Not returning res, which is VK_ERROR_FEATURE_NOT_PRESENT, because we already failed to allocate once. |
| 15080 | return VK_ERROR_OUT_OF_DEVICE_MEMORY; |
| 15081 | } |
| 15082 | } |
| 15083 | |
| 15084 | void VmaAllocator_T::FreeMemory( |
| 15085 | size_t allocationCount, |
| 15086 | const VmaAllocation* pAllocations) |
| 15087 | { |
| 15088 | VMA_ASSERT(pAllocations); |
| 15089 | |
| 15090 | for(size_t allocIndex = allocationCount; allocIndex--; ) |
| 15091 | { |
| 15092 | VmaAllocation allocation = pAllocations[allocIndex]; |
| 15093 | |
| 15094 | if(allocation != VK_NULL_HANDLE) |
| 15095 | { |
| 15096 | if(VMA_DEBUG_INITIALIZE_ALLOCATIONS) |
| 15097 | { |
| 15098 | FillAllocation(hAllocation: allocation, pattern: VMA_ALLOCATION_FILL_PATTERN_DESTROYED); |
| 15099 | } |
| 15100 | |
| 15101 | allocation->FreeName(hAllocator: this); |
| 15102 | |
| 15103 | switch(allocation->GetType()) |
| 15104 | { |
| 15105 | case VmaAllocation_T::ALLOCATION_TYPE_BLOCK: |
| 15106 | { |
| 15107 | VmaBlockVector* pBlockVector = VMA_NULL; |
| 15108 | VmaPool hPool = allocation->GetParentPool(); |
| 15109 | if(hPool != VK_NULL_HANDLE) |
| 15110 | { |
| 15111 | pBlockVector = &hPool->m_BlockVector; |
| 15112 | } |
| 15113 | else |
| 15114 | { |
| 15115 | const uint32_t memTypeIndex = allocation->GetMemoryTypeIndex(); |
| 15116 | pBlockVector = m_pBlockVectors[memTypeIndex]; |
| 15117 | VMA_ASSERT(pBlockVector && "Trying to free memory of unsupported type!" ); |
| 15118 | } |
| 15119 | pBlockVector->Free(hAllocation: allocation); |
| 15120 | } |
| 15121 | break; |
| 15122 | case VmaAllocation_T::ALLOCATION_TYPE_DEDICATED: |
| 15123 | FreeDedicatedMemory(allocation); |
| 15124 | break; |
| 15125 | default: |
| 15126 | VMA_ASSERT(0); |
| 15127 | } |
| 15128 | } |
| 15129 | } |
| 15130 | } |
| 15131 | |
| 15132 | void VmaAllocator_T::CalculateStatistics(VmaTotalStatistics* pStats) |
| 15133 | { |
| 15134 | // Initialize. |
| 15135 | VmaClearDetailedStatistics(outStats&: pStats->total); |
| 15136 | for(uint32_t i = 0; i < VK_MAX_MEMORY_TYPES; ++i) |
| 15137 | VmaClearDetailedStatistics(outStats&: pStats->memoryType[i]); |
| 15138 | for(uint32_t i = 0; i < VK_MAX_MEMORY_HEAPS; ++i) |
| 15139 | VmaClearDetailedStatistics(outStats&: pStats->memoryHeap[i]); |
| 15140 | |
| 15141 | // Process default pools. |
| 15142 | for(uint32_t memTypeIndex = 0; memTypeIndex < GetMemoryTypeCount(); ++memTypeIndex) |
| 15143 | { |
| 15144 | VmaBlockVector* const pBlockVector = m_pBlockVectors[memTypeIndex]; |
| 15145 | if (pBlockVector != VMA_NULL) |
| 15146 | pBlockVector->AddDetailedStatistics(inoutStats&: pStats->memoryType[memTypeIndex]); |
| 15147 | } |
| 15148 | |
| 15149 | // Process custom pools. |
| 15150 | { |
| 15151 | VmaMutexLockRead lock(m_PoolsMutex, m_UseMutex); |
| 15152 | for(VmaPool pool = m_Pools.Front(); pool != VMA_NULL; pool = m_Pools.GetNext(item: pool)) |
| 15153 | { |
| 15154 | VmaBlockVector& blockVector = pool->m_BlockVector; |
| 15155 | const uint32_t memTypeIndex = blockVector.GetMemoryTypeIndex(); |
| 15156 | blockVector.AddDetailedStatistics(inoutStats&: pStats->memoryType[memTypeIndex]); |
| 15157 | pool->m_DedicatedAllocations.AddDetailedStatistics(inoutStats&: pStats->memoryType[memTypeIndex]); |
| 15158 | } |
| 15159 | } |
| 15160 | |
| 15161 | // Process dedicated allocations. |
| 15162 | for(uint32_t memTypeIndex = 0; memTypeIndex < GetMemoryTypeCount(); ++memTypeIndex) |
| 15163 | { |
| 15164 | m_DedicatedAllocations[memTypeIndex].AddDetailedStatistics(inoutStats&: pStats->memoryType[memTypeIndex]); |
| 15165 | } |
| 15166 | |
| 15167 | // Sum from memory types to memory heaps. |
| 15168 | for(uint32_t memTypeIndex = 0; memTypeIndex < GetMemoryTypeCount(); ++memTypeIndex) |
| 15169 | { |
| 15170 | const uint32_t memHeapIndex = m_MemProps.memoryTypes[memTypeIndex].heapIndex; |
| 15171 | VmaAddDetailedStatistics(inoutStats&: pStats->memoryHeap[memHeapIndex], src: pStats->memoryType[memTypeIndex]); |
| 15172 | } |
| 15173 | |
| 15174 | // Sum from memory heaps to total. |
| 15175 | for(uint32_t memHeapIndex = 0; memHeapIndex < GetMemoryHeapCount(); ++memHeapIndex) |
| 15176 | VmaAddDetailedStatistics(inoutStats&: pStats->total, src: pStats->memoryHeap[memHeapIndex]); |
| 15177 | |
| 15178 | VMA_ASSERT(pStats->total.statistics.allocationCount == 0 || |
| 15179 | pStats->total.allocationSizeMax >= pStats->total.allocationSizeMin); |
| 15180 | VMA_ASSERT(pStats->total.unusedRangeCount == 0 || |
| 15181 | pStats->total.unusedRangeSizeMax >= pStats->total.unusedRangeSizeMin); |
| 15182 | } |
| 15183 | |
| 15184 | void VmaAllocator_T::GetHeapBudgets(VmaBudget* outBudgets, uint32_t firstHeap, uint32_t heapCount) |
| 15185 | { |
| 15186 | #if VMA_MEMORY_BUDGET |
| 15187 | if(m_UseExtMemoryBudget) |
| 15188 | { |
| 15189 | if(m_Budget.m_OperationsSinceBudgetFetch < 30) |
| 15190 | { |
| 15191 | VmaMutexLockRead lockRead(m_Budget.m_BudgetMutex, m_UseMutex); |
| 15192 | for(uint32_t i = 0; i < heapCount; ++i, ++outBudgets) |
| 15193 | { |
| 15194 | const uint32_t heapIndex = firstHeap + i; |
| 15195 | |
| 15196 | outBudgets->statistics.blockCount = m_Budget.m_BlockCount[heapIndex]; |
| 15197 | outBudgets->statistics.allocationCount = m_Budget.m_AllocationCount[heapIndex]; |
| 15198 | outBudgets->statistics.blockBytes = m_Budget.m_BlockBytes[heapIndex]; |
| 15199 | outBudgets->statistics.allocationBytes = m_Budget.m_AllocationBytes[heapIndex]; |
| 15200 | |
| 15201 | if(m_Budget.m_VulkanUsage[heapIndex] + outBudgets->statistics.blockBytes > m_Budget.m_BlockBytesAtBudgetFetch[heapIndex]) |
| 15202 | { |
| 15203 | outBudgets->usage = m_Budget.m_VulkanUsage[heapIndex] + |
| 15204 | outBudgets->statistics.blockBytes - m_Budget.m_BlockBytesAtBudgetFetch[heapIndex]; |
| 15205 | } |
| 15206 | else |
| 15207 | { |
| 15208 | outBudgets->usage = 0; |
| 15209 | } |
| 15210 | |
| 15211 | // Have to take MIN with heap size because explicit HeapSizeLimit is included in it. |
| 15212 | outBudgets->budget = VMA_MIN( |
| 15213 | m_Budget.m_VulkanBudget[heapIndex], m_MemProps.memoryHeaps[heapIndex].size); |
| 15214 | } |
| 15215 | } |
| 15216 | else |
| 15217 | { |
| 15218 | UpdateVulkanBudget(); // Outside of mutex lock |
| 15219 | GetHeapBudgets(outBudgets, firstHeap, heapCount); // Recursion |
| 15220 | } |
| 15221 | } |
| 15222 | else |
| 15223 | #endif |
| 15224 | { |
| 15225 | for(uint32_t i = 0; i < heapCount; ++i, ++outBudgets) |
| 15226 | { |
| 15227 | const uint32_t heapIndex = firstHeap + i; |
| 15228 | |
| 15229 | outBudgets->statistics.blockCount = m_Budget.m_BlockCount[heapIndex]; |
| 15230 | outBudgets->statistics.allocationCount = m_Budget.m_AllocationCount[heapIndex]; |
| 15231 | outBudgets->statistics.blockBytes = m_Budget.m_BlockBytes[heapIndex]; |
| 15232 | outBudgets->statistics.allocationBytes = m_Budget.m_AllocationBytes[heapIndex]; |
| 15233 | |
| 15234 | outBudgets->usage = outBudgets->statistics.blockBytes; |
| 15235 | outBudgets->budget = m_MemProps.memoryHeaps[heapIndex].size * 8 / 10; // 80% heuristics. |
| 15236 | } |
| 15237 | } |
| 15238 | } |
| 15239 | |
| 15240 | void VmaAllocator_T::GetAllocationInfo(VmaAllocation hAllocation, VmaAllocationInfo* pAllocationInfo) |
| 15241 | { |
| 15242 | pAllocationInfo->memoryType = hAllocation->GetMemoryTypeIndex(); |
| 15243 | pAllocationInfo->deviceMemory = hAllocation->GetMemory(); |
| 15244 | pAllocationInfo->offset = hAllocation->GetOffset(); |
| 15245 | pAllocationInfo->size = hAllocation->GetSize(); |
| 15246 | pAllocationInfo->pMappedData = hAllocation->GetMappedData(); |
| 15247 | pAllocationInfo->pUserData = hAllocation->GetUserData(); |
| 15248 | pAllocationInfo->pName = hAllocation->GetName(); |
| 15249 | } |
| 15250 | |
| 15251 | VkResult VmaAllocator_T::CreatePool(const VmaPoolCreateInfo* pCreateInfo, VmaPool* pPool) |
| 15252 | { |
| 15253 | VMA_DEBUG_LOG(" CreatePool: MemoryTypeIndex=%u, flags=%u" , pCreateInfo->memoryTypeIndex, pCreateInfo->flags); |
| 15254 | |
| 15255 | VmaPoolCreateInfo newCreateInfo = *pCreateInfo; |
| 15256 | |
| 15257 | // Protection against uninitialized new structure member. If garbage data are left there, this pointer dereference would crash. |
| 15258 | if(pCreateInfo->pMemoryAllocateNext) |
| 15259 | { |
| 15260 | VMA_ASSERT(((const VkBaseInStructure*)pCreateInfo->pMemoryAllocateNext)->sType != 0); |
| 15261 | } |
| 15262 | |
| 15263 | if(newCreateInfo.maxBlockCount == 0) |
| 15264 | { |
| 15265 | newCreateInfo.maxBlockCount = SIZE_MAX; |
| 15266 | } |
| 15267 | if(newCreateInfo.minBlockCount > newCreateInfo.maxBlockCount) |
| 15268 | { |
| 15269 | return VK_ERROR_INITIALIZATION_FAILED; |
| 15270 | } |
| 15271 | // Memory type index out of range or forbidden. |
| 15272 | if(pCreateInfo->memoryTypeIndex >= GetMemoryTypeCount() || |
| 15273 | ((1u << pCreateInfo->memoryTypeIndex) & m_GlobalMemoryTypeBits) == 0) |
| 15274 | { |
| 15275 | return VK_ERROR_FEATURE_NOT_PRESENT; |
| 15276 | } |
| 15277 | if(newCreateInfo.minAllocationAlignment > 0) |
| 15278 | { |
| 15279 | VMA_ASSERT(VmaIsPow2(newCreateInfo.minAllocationAlignment)); |
| 15280 | } |
| 15281 | |
| 15282 | const VkDeviceSize preferredBlockSize = CalcPreferredBlockSize(memTypeIndex: newCreateInfo.memoryTypeIndex); |
| 15283 | |
| 15284 | *pPool = vma_new(this, VmaPool_T)(this, newCreateInfo, preferredBlockSize); |
| 15285 | |
| 15286 | VkResult res = (*pPool)->m_BlockVector.CreateMinBlocks(); |
| 15287 | if(res != VK_SUCCESS) |
| 15288 | { |
| 15289 | vma_delete(hAllocator: this, ptr: *pPool); |
| 15290 | *pPool = VMA_NULL; |
| 15291 | return res; |
| 15292 | } |
| 15293 | |
| 15294 | // Add to m_Pools. |
| 15295 | { |
| 15296 | VmaMutexLockWrite lock(m_PoolsMutex, m_UseMutex); |
| 15297 | (*pPool)->SetId(m_NextPoolId++); |
| 15298 | m_Pools.PushBack(item: *pPool); |
| 15299 | } |
| 15300 | |
| 15301 | return VK_SUCCESS; |
| 15302 | } |
| 15303 | |
| 15304 | void VmaAllocator_T::DestroyPool(VmaPool pool) |
| 15305 | { |
| 15306 | // Remove from m_Pools. |
| 15307 | { |
| 15308 | VmaMutexLockWrite lock(m_PoolsMutex, m_UseMutex); |
| 15309 | m_Pools.Remove(item: pool); |
| 15310 | } |
| 15311 | |
| 15312 | vma_delete(hAllocator: this, ptr: pool); |
| 15313 | } |
| 15314 | |
| 15315 | void VmaAllocator_T::GetPoolStatistics(VmaPool pool, VmaStatistics* pPoolStats) |
| 15316 | { |
| 15317 | VmaClearStatistics(outStats&: *pPoolStats); |
| 15318 | pool->m_BlockVector.AddStatistics(inoutStats&: *pPoolStats); |
| 15319 | pool->m_DedicatedAllocations.AddStatistics(inoutStats&: *pPoolStats); |
| 15320 | } |
| 15321 | |
| 15322 | void VmaAllocator_T::CalculatePoolStatistics(VmaPool pool, VmaDetailedStatistics* pPoolStats) |
| 15323 | { |
| 15324 | VmaClearDetailedStatistics(outStats&: *pPoolStats); |
| 15325 | pool->m_BlockVector.AddDetailedStatistics(inoutStats&: *pPoolStats); |
| 15326 | pool->m_DedicatedAllocations.AddDetailedStatistics(inoutStats&: *pPoolStats); |
| 15327 | } |
| 15328 | |
| 15329 | void VmaAllocator_T::SetCurrentFrameIndex(uint32_t frameIndex) |
| 15330 | { |
| 15331 | m_CurrentFrameIndex.store(i: frameIndex); |
| 15332 | |
| 15333 | #if VMA_MEMORY_BUDGET |
| 15334 | if(m_UseExtMemoryBudget) |
| 15335 | { |
| 15336 | UpdateVulkanBudget(); |
| 15337 | } |
| 15338 | #endif // #if VMA_MEMORY_BUDGET |
| 15339 | } |
| 15340 | |
| 15341 | VkResult VmaAllocator_T::CheckPoolCorruption(VmaPool hPool) |
| 15342 | { |
| 15343 | return hPool->m_BlockVector.CheckCorruption(); |
| 15344 | } |
| 15345 | |
| 15346 | VkResult VmaAllocator_T::CheckCorruption(uint32_t memoryTypeBits) |
| 15347 | { |
| 15348 | VkResult finalRes = VK_ERROR_FEATURE_NOT_PRESENT; |
| 15349 | |
| 15350 | // Process default pools. |
| 15351 | for(uint32_t memTypeIndex = 0; memTypeIndex < GetMemoryTypeCount(); ++memTypeIndex) |
| 15352 | { |
| 15353 | VmaBlockVector* const pBlockVector = m_pBlockVectors[memTypeIndex]; |
| 15354 | if(pBlockVector != VMA_NULL) |
| 15355 | { |
| 15356 | VkResult localRes = pBlockVector->CheckCorruption(); |
| 15357 | switch(localRes) |
| 15358 | { |
| 15359 | case VK_ERROR_FEATURE_NOT_PRESENT: |
| 15360 | break; |
| 15361 | case VK_SUCCESS: |
| 15362 | finalRes = VK_SUCCESS; |
| 15363 | break; |
| 15364 | default: |
| 15365 | return localRes; |
| 15366 | } |
| 15367 | } |
| 15368 | } |
| 15369 | |
| 15370 | // Process custom pools. |
| 15371 | { |
| 15372 | VmaMutexLockRead lock(m_PoolsMutex, m_UseMutex); |
| 15373 | for(VmaPool pool = m_Pools.Front(); pool != VMA_NULL; pool = m_Pools.GetNext(item: pool)) |
| 15374 | { |
| 15375 | if(((1u << pool->m_BlockVector.GetMemoryTypeIndex()) & memoryTypeBits) != 0) |
| 15376 | { |
| 15377 | VkResult localRes = pool->m_BlockVector.CheckCorruption(); |
| 15378 | switch(localRes) |
| 15379 | { |
| 15380 | case VK_ERROR_FEATURE_NOT_PRESENT: |
| 15381 | break; |
| 15382 | case VK_SUCCESS: |
| 15383 | finalRes = VK_SUCCESS; |
| 15384 | break; |
| 15385 | default: |
| 15386 | return localRes; |
| 15387 | } |
| 15388 | } |
| 15389 | } |
| 15390 | } |
| 15391 | |
| 15392 | return finalRes; |
| 15393 | } |
| 15394 | |
| 15395 | VkResult VmaAllocator_T::AllocateVulkanMemory(const VkMemoryAllocateInfo* pAllocateInfo, VkDeviceMemory* pMemory) |
| 15396 | { |
| 15397 | AtomicTransactionalIncrement<uint32_t> deviceMemoryCountIncrement; |
| 15398 | const uint64_t prevDeviceMemoryCount = deviceMemoryCountIncrement.Increment(atomic: &m_DeviceMemoryCount); |
| 15399 | #if VMA_DEBUG_DONT_EXCEED_MAX_MEMORY_ALLOCATION_COUNT |
| 15400 | if(prevDeviceMemoryCount >= m_PhysicalDeviceProperties.limits.maxMemoryAllocationCount) |
| 15401 | { |
| 15402 | return VK_ERROR_TOO_MANY_OBJECTS; |
| 15403 | } |
| 15404 | #endif |
| 15405 | |
| 15406 | const uint32_t heapIndex = MemoryTypeIndexToHeapIndex(memTypeIndex: pAllocateInfo->memoryTypeIndex); |
| 15407 | |
| 15408 | // HeapSizeLimit is in effect for this heap. |
| 15409 | if((m_HeapSizeLimitMask & (1u << heapIndex)) != 0) |
| 15410 | { |
| 15411 | const VkDeviceSize heapSize = m_MemProps.memoryHeaps[heapIndex].size; |
| 15412 | VkDeviceSize blockBytes = m_Budget.m_BlockBytes[heapIndex]; |
| 15413 | for(;;) |
| 15414 | { |
| 15415 | const VkDeviceSize blockBytesAfterAllocation = blockBytes + pAllocateInfo->allocationSize; |
| 15416 | if(blockBytesAfterAllocation > heapSize) |
| 15417 | { |
| 15418 | return VK_ERROR_OUT_OF_DEVICE_MEMORY; |
| 15419 | } |
| 15420 | if(m_Budget.m_BlockBytes[heapIndex].compare_exchange_strong(i1&: blockBytes, i2: blockBytesAfterAllocation)) |
| 15421 | { |
| 15422 | break; |
| 15423 | } |
| 15424 | } |
| 15425 | } |
| 15426 | else |
| 15427 | { |
| 15428 | m_Budget.m_BlockBytes[heapIndex] += pAllocateInfo->allocationSize; |
| 15429 | } |
| 15430 | ++m_Budget.m_BlockCount[heapIndex]; |
| 15431 | |
| 15432 | // VULKAN CALL vkAllocateMemory. |
| 15433 | VkResult res = (*m_VulkanFunctions.vkAllocateMemory)(m_hDevice, pAllocateInfo, GetAllocationCallbacks(), pMemory); |
| 15434 | |
| 15435 | if(res == VK_SUCCESS) |
| 15436 | { |
| 15437 | #if VMA_MEMORY_BUDGET |
| 15438 | ++m_Budget.m_OperationsSinceBudgetFetch; |
| 15439 | #endif |
| 15440 | |
| 15441 | // Informative callback. |
| 15442 | if(m_DeviceMemoryCallbacks.pfnAllocate != VMA_NULL) |
| 15443 | { |
| 15444 | (*m_DeviceMemoryCallbacks.pfnAllocate)(this, pAllocateInfo->memoryTypeIndex, *pMemory, pAllocateInfo->allocationSize, m_DeviceMemoryCallbacks.pUserData); |
| 15445 | } |
| 15446 | |
| 15447 | deviceMemoryCountIncrement.Commit(); |
| 15448 | } |
| 15449 | else |
| 15450 | { |
| 15451 | --m_Budget.m_BlockCount[heapIndex]; |
| 15452 | m_Budget.m_BlockBytes[heapIndex] -= pAllocateInfo->allocationSize; |
| 15453 | } |
| 15454 | |
| 15455 | return res; |
| 15456 | } |
| 15457 | |
| 15458 | void VmaAllocator_T::FreeVulkanMemory(uint32_t memoryType, VkDeviceSize size, VkDeviceMemory hMemory) |
| 15459 | { |
| 15460 | // Informative callback. |
| 15461 | if(m_DeviceMemoryCallbacks.pfnFree != VMA_NULL) |
| 15462 | { |
| 15463 | (*m_DeviceMemoryCallbacks.pfnFree)(this, memoryType, hMemory, size, m_DeviceMemoryCallbacks.pUserData); |
| 15464 | } |
| 15465 | |
| 15466 | // VULKAN CALL vkFreeMemory. |
| 15467 | (*m_VulkanFunctions.vkFreeMemory)(m_hDevice, hMemory, GetAllocationCallbacks()); |
| 15468 | |
| 15469 | const uint32_t heapIndex = MemoryTypeIndexToHeapIndex(memTypeIndex: memoryType); |
| 15470 | --m_Budget.m_BlockCount[heapIndex]; |
| 15471 | m_Budget.m_BlockBytes[heapIndex] -= size; |
| 15472 | |
| 15473 | --m_DeviceMemoryCount; |
| 15474 | } |
| 15475 | |
| 15476 | VkResult VmaAllocator_T::BindVulkanBuffer( |
| 15477 | VkDeviceMemory memory, |
| 15478 | VkDeviceSize memoryOffset, |
| 15479 | VkBuffer buffer, |
| 15480 | const void* pNext) |
| 15481 | { |
| 15482 | if(pNext != VMA_NULL) |
| 15483 | { |
| 15484 | #if VMA_VULKAN_VERSION >= 1001000 || VMA_BIND_MEMORY2 |
| 15485 | if((m_UseKhrBindMemory2 || m_VulkanApiVersion >= VK_MAKE_VERSION(1, 1, 0)) && |
| 15486 | m_VulkanFunctions.vkBindBufferMemory2KHR != VMA_NULL) |
| 15487 | { |
| 15488 | VkBindBufferMemoryInfoKHR bindBufferMemoryInfo = { .sType: VK_STRUCTURE_TYPE_BIND_BUFFER_MEMORY_INFO_KHR }; |
| 15489 | bindBufferMemoryInfo.pNext = pNext; |
| 15490 | bindBufferMemoryInfo.buffer = buffer; |
| 15491 | bindBufferMemoryInfo.memory = memory; |
| 15492 | bindBufferMemoryInfo.memoryOffset = memoryOffset; |
| 15493 | return (*m_VulkanFunctions.vkBindBufferMemory2KHR)(m_hDevice, 1, &bindBufferMemoryInfo); |
| 15494 | } |
| 15495 | else |
| 15496 | #endif // #if VMA_VULKAN_VERSION >= 1001000 || VMA_BIND_MEMORY2 |
| 15497 | { |
| 15498 | return VK_ERROR_EXTENSION_NOT_PRESENT; |
| 15499 | } |
| 15500 | } |
| 15501 | else |
| 15502 | { |
| 15503 | return (*m_VulkanFunctions.vkBindBufferMemory)(m_hDevice, buffer, memory, memoryOffset); |
| 15504 | } |
| 15505 | } |
| 15506 | |
| 15507 | VkResult VmaAllocator_T::BindVulkanImage( |
| 15508 | VkDeviceMemory memory, |
| 15509 | VkDeviceSize memoryOffset, |
| 15510 | VkImage image, |
| 15511 | const void* pNext) |
| 15512 | { |
| 15513 | if(pNext != VMA_NULL) |
| 15514 | { |
| 15515 | #if VMA_VULKAN_VERSION >= 1001000 || VMA_BIND_MEMORY2 |
| 15516 | if((m_UseKhrBindMemory2 || m_VulkanApiVersion >= VK_MAKE_VERSION(1, 1, 0)) && |
| 15517 | m_VulkanFunctions.vkBindImageMemory2KHR != VMA_NULL) |
| 15518 | { |
| 15519 | VkBindImageMemoryInfoKHR bindBufferMemoryInfo = { .sType: VK_STRUCTURE_TYPE_BIND_IMAGE_MEMORY_INFO_KHR }; |
| 15520 | bindBufferMemoryInfo.pNext = pNext; |
| 15521 | bindBufferMemoryInfo.image = image; |
| 15522 | bindBufferMemoryInfo.memory = memory; |
| 15523 | bindBufferMemoryInfo.memoryOffset = memoryOffset; |
| 15524 | return (*m_VulkanFunctions.vkBindImageMemory2KHR)(m_hDevice, 1, &bindBufferMemoryInfo); |
| 15525 | } |
| 15526 | else |
| 15527 | #endif // #if VMA_BIND_MEMORY2 |
| 15528 | { |
| 15529 | return VK_ERROR_EXTENSION_NOT_PRESENT; |
| 15530 | } |
| 15531 | } |
| 15532 | else |
| 15533 | { |
| 15534 | return (*m_VulkanFunctions.vkBindImageMemory)(m_hDevice, image, memory, memoryOffset); |
| 15535 | } |
| 15536 | } |
| 15537 | |
| 15538 | VkResult VmaAllocator_T::Map(VmaAllocation hAllocation, void** ppData) |
| 15539 | { |
| 15540 | switch(hAllocation->GetType()) |
| 15541 | { |
| 15542 | case VmaAllocation_T::ALLOCATION_TYPE_BLOCK: |
| 15543 | { |
| 15544 | VmaDeviceMemoryBlock* const pBlock = hAllocation->GetBlock(); |
| 15545 | char *pBytes = VMA_NULL; |
| 15546 | VkResult res = pBlock->Map(hAllocator: this, count: 1, ppData: (void**)&pBytes); |
| 15547 | if(res == VK_SUCCESS) |
| 15548 | { |
| 15549 | *ppData = pBytes + (ptrdiff_t)hAllocation->GetOffset(); |
| 15550 | hAllocation->BlockAllocMap(); |
| 15551 | } |
| 15552 | return res; |
| 15553 | } |
| 15554 | case VmaAllocation_T::ALLOCATION_TYPE_DEDICATED: |
| 15555 | return hAllocation->DedicatedAllocMap(hAllocator: this, ppData); |
| 15556 | default: |
| 15557 | VMA_ASSERT(0); |
| 15558 | return VK_ERROR_MEMORY_MAP_FAILED; |
| 15559 | } |
| 15560 | } |
| 15561 | |
| 15562 | void VmaAllocator_T::Unmap(VmaAllocation hAllocation) |
| 15563 | { |
| 15564 | switch(hAllocation->GetType()) |
| 15565 | { |
| 15566 | case VmaAllocation_T::ALLOCATION_TYPE_BLOCK: |
| 15567 | { |
| 15568 | VmaDeviceMemoryBlock* const pBlock = hAllocation->GetBlock(); |
| 15569 | hAllocation->BlockAllocUnmap(); |
| 15570 | pBlock->Unmap(hAllocator: this, count: 1); |
| 15571 | } |
| 15572 | break; |
| 15573 | case VmaAllocation_T::ALLOCATION_TYPE_DEDICATED: |
| 15574 | hAllocation->DedicatedAllocUnmap(hAllocator: this); |
| 15575 | break; |
| 15576 | default: |
| 15577 | VMA_ASSERT(0); |
| 15578 | } |
| 15579 | } |
| 15580 | |
| 15581 | VkResult VmaAllocator_T::BindBufferMemory( |
| 15582 | VmaAllocation hAllocation, |
| 15583 | VkDeviceSize allocationLocalOffset, |
| 15584 | VkBuffer hBuffer, |
| 15585 | const void* pNext) |
| 15586 | { |
| 15587 | VkResult res = VK_SUCCESS; |
| 15588 | switch(hAllocation->GetType()) |
| 15589 | { |
| 15590 | case VmaAllocation_T::ALLOCATION_TYPE_DEDICATED: |
| 15591 | res = BindVulkanBuffer(memory: hAllocation->GetMemory(), memoryOffset: allocationLocalOffset, buffer: hBuffer, pNext); |
| 15592 | break; |
| 15593 | case VmaAllocation_T::ALLOCATION_TYPE_BLOCK: |
| 15594 | { |
| 15595 | VmaDeviceMemoryBlock* const pBlock = hAllocation->GetBlock(); |
| 15596 | VMA_ASSERT(pBlock && "Binding buffer to allocation that doesn't belong to any block." ); |
| 15597 | res = pBlock->BindBufferMemory(hAllocator: this, hAllocation, allocationLocalOffset, hBuffer, pNext); |
| 15598 | break; |
| 15599 | } |
| 15600 | default: |
| 15601 | VMA_ASSERT(0); |
| 15602 | } |
| 15603 | return res; |
| 15604 | } |
| 15605 | |
| 15606 | VkResult VmaAllocator_T::BindImageMemory( |
| 15607 | VmaAllocation hAllocation, |
| 15608 | VkDeviceSize allocationLocalOffset, |
| 15609 | VkImage hImage, |
| 15610 | const void* pNext) |
| 15611 | { |
| 15612 | VkResult res = VK_SUCCESS; |
| 15613 | switch(hAllocation->GetType()) |
| 15614 | { |
| 15615 | case VmaAllocation_T::ALLOCATION_TYPE_DEDICATED: |
| 15616 | res = BindVulkanImage(memory: hAllocation->GetMemory(), memoryOffset: allocationLocalOffset, image: hImage, pNext); |
| 15617 | break; |
| 15618 | case VmaAllocation_T::ALLOCATION_TYPE_BLOCK: |
| 15619 | { |
| 15620 | VmaDeviceMemoryBlock* pBlock = hAllocation->GetBlock(); |
| 15621 | VMA_ASSERT(pBlock && "Binding image to allocation that doesn't belong to any block." ); |
| 15622 | res = pBlock->BindImageMemory(hAllocator: this, hAllocation, allocationLocalOffset, hImage, pNext); |
| 15623 | break; |
| 15624 | } |
| 15625 | default: |
| 15626 | VMA_ASSERT(0); |
| 15627 | } |
| 15628 | return res; |
| 15629 | } |
| 15630 | |
| 15631 | VkResult VmaAllocator_T::FlushOrInvalidateAllocation( |
| 15632 | VmaAllocation hAllocation, |
| 15633 | VkDeviceSize offset, VkDeviceSize size, |
| 15634 | VMA_CACHE_OPERATION op) |
| 15635 | { |
| 15636 | VkResult res = VK_SUCCESS; |
| 15637 | |
| 15638 | VkMappedMemoryRange memRange = {}; |
| 15639 | if(GetFlushOrInvalidateRange(allocation: hAllocation, offset, size, outRange&: memRange)) |
| 15640 | { |
| 15641 | switch(op) |
| 15642 | { |
| 15643 | case VMA_CACHE_FLUSH: |
| 15644 | res = (*GetVulkanFunctions().vkFlushMappedMemoryRanges)(m_hDevice, 1, &memRange); |
| 15645 | break; |
| 15646 | case VMA_CACHE_INVALIDATE: |
| 15647 | res = (*GetVulkanFunctions().vkInvalidateMappedMemoryRanges)(m_hDevice, 1, &memRange); |
| 15648 | break; |
| 15649 | default: |
| 15650 | VMA_ASSERT(0); |
| 15651 | } |
| 15652 | } |
| 15653 | // else: Just ignore this call. |
| 15654 | return res; |
| 15655 | } |
| 15656 | |
| 15657 | VkResult VmaAllocator_T::FlushOrInvalidateAllocations( |
| 15658 | uint32_t allocationCount, |
| 15659 | const VmaAllocation* allocations, |
| 15660 | const VkDeviceSize* offsets, const VkDeviceSize* sizes, |
| 15661 | VMA_CACHE_OPERATION op) |
| 15662 | { |
| 15663 | typedef VmaStlAllocator<VkMappedMemoryRange> RangeAllocator; |
| 15664 | typedef VmaSmallVector<VkMappedMemoryRange, RangeAllocator, 16> RangeVector; |
| 15665 | RangeVector ranges = RangeVector(RangeAllocator(GetAllocationCallbacks())); |
| 15666 | |
| 15667 | for(uint32_t allocIndex = 0; allocIndex < allocationCount; ++allocIndex) |
| 15668 | { |
| 15669 | const VmaAllocation alloc = allocations[allocIndex]; |
| 15670 | const VkDeviceSize offset = offsets != VMA_NULL ? offsets[allocIndex] : 0; |
| 15671 | const VkDeviceSize size = sizes != VMA_NULL ? sizes[allocIndex] : VK_WHOLE_SIZE; |
| 15672 | VkMappedMemoryRange newRange; |
| 15673 | if(GetFlushOrInvalidateRange(allocation: alloc, offset, size, outRange&: newRange)) |
| 15674 | { |
| 15675 | ranges.push_back(src: newRange); |
| 15676 | } |
| 15677 | } |
| 15678 | |
| 15679 | VkResult res = VK_SUCCESS; |
| 15680 | if(!ranges.empty()) |
| 15681 | { |
| 15682 | switch(op) |
| 15683 | { |
| 15684 | case VMA_CACHE_FLUSH: |
| 15685 | res = (*GetVulkanFunctions().vkFlushMappedMemoryRanges)(m_hDevice, (uint32_t)ranges.size(), ranges.data()); |
| 15686 | break; |
| 15687 | case VMA_CACHE_INVALIDATE: |
| 15688 | res = (*GetVulkanFunctions().vkInvalidateMappedMemoryRanges)(m_hDevice, (uint32_t)ranges.size(), ranges.data()); |
| 15689 | break; |
| 15690 | default: |
| 15691 | VMA_ASSERT(0); |
| 15692 | } |
| 15693 | } |
| 15694 | // else: Just ignore this call. |
| 15695 | return res; |
| 15696 | } |
| 15697 | |
| 15698 | void VmaAllocator_T::FreeDedicatedMemory(const VmaAllocation allocation) |
| 15699 | { |
| 15700 | VMA_ASSERT(allocation && allocation->GetType() == VmaAllocation_T::ALLOCATION_TYPE_DEDICATED); |
| 15701 | |
| 15702 | const uint32_t memTypeIndex = allocation->GetMemoryTypeIndex(); |
| 15703 | VmaPool parentPool = allocation->GetParentPool(); |
| 15704 | if(parentPool == VK_NULL_HANDLE) |
| 15705 | { |
| 15706 | // Default pool |
| 15707 | m_DedicatedAllocations[memTypeIndex].Unregister(alloc: allocation); |
| 15708 | } |
| 15709 | else |
| 15710 | { |
| 15711 | // Custom pool |
| 15712 | parentPool->m_DedicatedAllocations.Unregister(alloc: allocation); |
| 15713 | } |
| 15714 | |
| 15715 | VkDeviceMemory hMemory = allocation->GetMemory(); |
| 15716 | |
| 15717 | /* |
| 15718 | There is no need to call this, because Vulkan spec allows to skip vkUnmapMemory |
| 15719 | before vkFreeMemory. |
| 15720 | |
| 15721 | if(allocation->GetMappedData() != VMA_NULL) |
| 15722 | { |
| 15723 | (*m_VulkanFunctions.vkUnmapMemory)(m_hDevice, hMemory); |
| 15724 | } |
| 15725 | */ |
| 15726 | |
| 15727 | FreeVulkanMemory(memoryType: memTypeIndex, size: allocation->GetSize(), hMemory); |
| 15728 | |
| 15729 | m_Budget.RemoveAllocation(heapIndex: MemoryTypeIndexToHeapIndex(memTypeIndex: allocation->GetMemoryTypeIndex()), allocationSize: allocation->GetSize()); |
| 15730 | m_AllocationObjectAllocator.Free(hAlloc: allocation); |
| 15731 | |
| 15732 | VMA_DEBUG_LOG(" Freed DedicatedMemory MemoryTypeIndex=%u" , memTypeIndex); |
| 15733 | } |
| 15734 | |
| 15735 | uint32_t VmaAllocator_T::CalculateGpuDefragmentationMemoryTypeBits() const |
| 15736 | { |
| 15737 | VkBufferCreateInfo dummyBufCreateInfo; |
| 15738 | VmaFillGpuDefragmentationBufferCreateInfo(outBufCreateInfo&: dummyBufCreateInfo); |
| 15739 | |
| 15740 | uint32_t memoryTypeBits = 0; |
| 15741 | |
| 15742 | // Create buffer. |
| 15743 | VkBuffer buf = VK_NULL_HANDLE; |
| 15744 | VkResult res = (*GetVulkanFunctions().vkCreateBuffer)( |
| 15745 | m_hDevice, &dummyBufCreateInfo, GetAllocationCallbacks(), &buf); |
| 15746 | if(res == VK_SUCCESS) |
| 15747 | { |
| 15748 | // Query for supported memory types. |
| 15749 | VkMemoryRequirements memReq; |
| 15750 | (*GetVulkanFunctions().vkGetBufferMemoryRequirements)(m_hDevice, buf, &memReq); |
| 15751 | memoryTypeBits = memReq.memoryTypeBits; |
| 15752 | |
| 15753 | // Destroy buffer. |
| 15754 | (*GetVulkanFunctions().vkDestroyBuffer)(m_hDevice, buf, GetAllocationCallbacks()); |
| 15755 | } |
| 15756 | |
| 15757 | return memoryTypeBits; |
| 15758 | } |
| 15759 | |
| 15760 | uint32_t VmaAllocator_T::CalculateGlobalMemoryTypeBits() const |
| 15761 | { |
| 15762 | // Make sure memory information is already fetched. |
| 15763 | VMA_ASSERT(GetMemoryTypeCount() > 0); |
| 15764 | |
| 15765 | uint32_t memoryTypeBits = UINT32_MAX; |
| 15766 | |
| 15767 | if(!m_UseAmdDeviceCoherentMemory) |
| 15768 | { |
| 15769 | // Exclude memory types that have VK_MEMORY_PROPERTY_DEVICE_COHERENT_BIT_AMD. |
| 15770 | for(uint32_t memTypeIndex = 0; memTypeIndex < GetMemoryTypeCount(); ++memTypeIndex) |
| 15771 | { |
| 15772 | if((m_MemProps.memoryTypes[memTypeIndex].propertyFlags & VK_MEMORY_PROPERTY_DEVICE_COHERENT_BIT_AMD_COPY) != 0) |
| 15773 | { |
| 15774 | memoryTypeBits &= ~(1u << memTypeIndex); |
| 15775 | } |
| 15776 | } |
| 15777 | } |
| 15778 | |
| 15779 | return memoryTypeBits; |
| 15780 | } |
| 15781 | |
| 15782 | bool VmaAllocator_T::GetFlushOrInvalidateRange( |
| 15783 | VmaAllocation allocation, |
| 15784 | VkDeviceSize offset, VkDeviceSize size, |
| 15785 | VkMappedMemoryRange& outRange) const |
| 15786 | { |
| 15787 | const uint32_t memTypeIndex = allocation->GetMemoryTypeIndex(); |
| 15788 | if(size > 0 && IsMemoryTypeNonCoherent(memTypeIndex)) |
| 15789 | { |
| 15790 | const VkDeviceSize nonCoherentAtomSize = m_PhysicalDeviceProperties.limits.nonCoherentAtomSize; |
| 15791 | const VkDeviceSize allocationSize = allocation->GetSize(); |
| 15792 | VMA_ASSERT(offset <= allocationSize); |
| 15793 | |
| 15794 | outRange.sType = VK_STRUCTURE_TYPE_MAPPED_MEMORY_RANGE; |
| 15795 | outRange.pNext = VMA_NULL; |
| 15796 | outRange.memory = allocation->GetMemory(); |
| 15797 | |
| 15798 | switch(allocation->GetType()) |
| 15799 | { |
| 15800 | case VmaAllocation_T::ALLOCATION_TYPE_DEDICATED: |
| 15801 | outRange.offset = VmaAlignDown(val: offset, alignment: nonCoherentAtomSize); |
| 15802 | if(size == VK_WHOLE_SIZE) |
| 15803 | { |
| 15804 | outRange.size = allocationSize - outRange.offset; |
| 15805 | } |
| 15806 | else |
| 15807 | { |
| 15808 | VMA_ASSERT(offset + size <= allocationSize); |
| 15809 | outRange.size = VMA_MIN( |
| 15810 | VmaAlignUp(size + (offset - outRange.offset), nonCoherentAtomSize), |
| 15811 | allocationSize - outRange.offset); |
| 15812 | } |
| 15813 | break; |
| 15814 | case VmaAllocation_T::ALLOCATION_TYPE_BLOCK: |
| 15815 | { |
| 15816 | // 1. Still within this allocation. |
| 15817 | outRange.offset = VmaAlignDown(val: offset, alignment: nonCoherentAtomSize); |
| 15818 | if(size == VK_WHOLE_SIZE) |
| 15819 | { |
| 15820 | size = allocationSize - offset; |
| 15821 | } |
| 15822 | else |
| 15823 | { |
| 15824 | VMA_ASSERT(offset + size <= allocationSize); |
| 15825 | } |
| 15826 | outRange.size = VmaAlignUp(val: size + (offset - outRange.offset), alignment: nonCoherentAtomSize); |
| 15827 | |
| 15828 | // 2. Adjust to whole block. |
| 15829 | const VkDeviceSize allocationOffset = allocation->GetOffset(); |
| 15830 | VMA_ASSERT(allocationOffset % nonCoherentAtomSize == 0); |
| 15831 | const VkDeviceSize blockSize = allocation->GetBlock()->m_pMetadata->GetSize(); |
| 15832 | outRange.offset += allocationOffset; |
| 15833 | outRange.size = VMA_MIN(outRange.size, blockSize - outRange.offset); |
| 15834 | |
| 15835 | break; |
| 15836 | } |
| 15837 | default: |
| 15838 | VMA_ASSERT(0); |
| 15839 | } |
| 15840 | return true; |
| 15841 | } |
| 15842 | return false; |
| 15843 | } |
| 15844 | |
| 15845 | #if VMA_MEMORY_BUDGET |
| 15846 | void VmaAllocator_T::UpdateVulkanBudget() |
| 15847 | { |
| 15848 | VMA_ASSERT(m_UseExtMemoryBudget); |
| 15849 | |
| 15850 | VkPhysicalDeviceMemoryProperties2KHR memProps = { .sType: VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_MEMORY_PROPERTIES_2_KHR }; |
| 15851 | |
| 15852 | VkPhysicalDeviceMemoryBudgetPropertiesEXT budgetProps = { .sType: VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_MEMORY_BUDGET_PROPERTIES_EXT }; |
| 15853 | VmaPnextChainPushFront(mainStruct: &memProps, newStruct: &budgetProps); |
| 15854 | |
| 15855 | GetVulkanFunctions().vkGetPhysicalDeviceMemoryProperties2KHR(m_PhysicalDevice, &memProps); |
| 15856 | |
| 15857 | { |
| 15858 | VmaMutexLockWrite lockWrite(m_Budget.m_BudgetMutex, m_UseMutex); |
| 15859 | |
| 15860 | for(uint32_t heapIndex = 0; heapIndex < GetMemoryHeapCount(); ++heapIndex) |
| 15861 | { |
| 15862 | m_Budget.m_VulkanUsage[heapIndex] = budgetProps.heapUsage[heapIndex]; |
| 15863 | m_Budget.m_VulkanBudget[heapIndex] = budgetProps.heapBudget[heapIndex]; |
| 15864 | m_Budget.m_BlockBytesAtBudgetFetch[heapIndex] = m_Budget.m_BlockBytes[heapIndex].load(); |
| 15865 | |
| 15866 | // Some bugged drivers return the budget incorrectly, e.g. 0 or much bigger than heap size. |
| 15867 | if(m_Budget.m_VulkanBudget[heapIndex] == 0) |
| 15868 | { |
| 15869 | m_Budget.m_VulkanBudget[heapIndex] = m_MemProps.memoryHeaps[heapIndex].size * 8 / 10; // 80% heuristics. |
| 15870 | } |
| 15871 | else if(m_Budget.m_VulkanBudget[heapIndex] > m_MemProps.memoryHeaps[heapIndex].size) |
| 15872 | { |
| 15873 | m_Budget.m_VulkanBudget[heapIndex] = m_MemProps.memoryHeaps[heapIndex].size; |
| 15874 | } |
| 15875 | if(m_Budget.m_VulkanUsage[heapIndex] == 0 && m_Budget.m_BlockBytesAtBudgetFetch[heapIndex] > 0) |
| 15876 | { |
| 15877 | m_Budget.m_VulkanUsage[heapIndex] = m_Budget.m_BlockBytesAtBudgetFetch[heapIndex]; |
| 15878 | } |
| 15879 | } |
| 15880 | m_Budget.m_OperationsSinceBudgetFetch = 0; |
| 15881 | } |
| 15882 | } |
| 15883 | #endif // VMA_MEMORY_BUDGET |
| 15884 | |
| 15885 | void VmaAllocator_T::FillAllocation(const VmaAllocation hAllocation, uint8_t pattern) |
| 15886 | { |
| 15887 | if(VMA_DEBUG_INITIALIZE_ALLOCATIONS && |
| 15888 | hAllocation->IsMappingAllowed() && |
| 15889 | (m_MemProps.memoryTypes[hAllocation->GetMemoryTypeIndex()].propertyFlags & VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT) != 0) |
| 15890 | { |
| 15891 | void* pData = VMA_NULL; |
| 15892 | VkResult res = Map(hAllocation, ppData: &pData); |
| 15893 | if(res == VK_SUCCESS) |
| 15894 | { |
| 15895 | memset(s: pData, c: (int)pattern, n: (size_t)hAllocation->GetSize()); |
| 15896 | FlushOrInvalidateAllocation(hAllocation, offset: 0, VK_WHOLE_SIZE, op: VMA_CACHE_FLUSH); |
| 15897 | Unmap(hAllocation); |
| 15898 | } |
| 15899 | else |
| 15900 | { |
| 15901 | VMA_ASSERT(0 && "VMA_DEBUG_INITIALIZE_ALLOCATIONS is enabled, but couldn't map memory to fill allocation." ); |
| 15902 | } |
| 15903 | } |
| 15904 | } |
| 15905 | |
| 15906 | uint32_t VmaAllocator_T::GetGpuDefragmentationMemoryTypeBits() |
| 15907 | { |
| 15908 | uint32_t memoryTypeBits = m_GpuDefragmentationMemoryTypeBits.load(); |
| 15909 | if(memoryTypeBits == UINT32_MAX) |
| 15910 | { |
| 15911 | memoryTypeBits = CalculateGpuDefragmentationMemoryTypeBits(); |
| 15912 | m_GpuDefragmentationMemoryTypeBits.store(i: memoryTypeBits); |
| 15913 | } |
| 15914 | return memoryTypeBits; |
| 15915 | } |
| 15916 | |
| 15917 | #if VMA_STATS_STRING_ENABLED |
| 15918 | void VmaAllocator_T::PrintDetailedMap(VmaJsonWriter& json) |
| 15919 | { |
| 15920 | json.WriteString(pStr: "DefaultPools" ); |
| 15921 | json.BeginObject(); |
| 15922 | { |
| 15923 | for (uint32_t memTypeIndex = 0; memTypeIndex < GetMemoryTypeCount(); ++memTypeIndex) |
| 15924 | { |
| 15925 | VmaBlockVector* pBlockVector = m_pBlockVectors[memTypeIndex]; |
| 15926 | VmaDedicatedAllocationList& dedicatedAllocList = m_DedicatedAllocations[memTypeIndex]; |
| 15927 | if (pBlockVector != VMA_NULL) |
| 15928 | { |
| 15929 | json.BeginString(pStr: "Type " ); |
| 15930 | json.ContinueString(n: memTypeIndex); |
| 15931 | json.EndString(); |
| 15932 | json.BeginObject(); |
| 15933 | { |
| 15934 | json.WriteString(pStr: "PreferredBlockSize" ); |
| 15935 | json.WriteNumber(n: pBlockVector->GetPreferredBlockSize()); |
| 15936 | |
| 15937 | json.WriteString(pStr: "Blocks" ); |
| 15938 | pBlockVector->PrintDetailedMap(json); |
| 15939 | |
| 15940 | json.WriteString(pStr: "DedicatedAllocations" ); |
| 15941 | dedicatedAllocList.BuildStatsString(json); |
| 15942 | } |
| 15943 | json.EndObject(); |
| 15944 | } |
| 15945 | } |
| 15946 | } |
| 15947 | json.EndObject(); |
| 15948 | |
| 15949 | json.WriteString(pStr: "CustomPools" ); |
| 15950 | json.BeginObject(); |
| 15951 | { |
| 15952 | VmaMutexLockRead lock(m_PoolsMutex, m_UseMutex); |
| 15953 | if (!m_Pools.IsEmpty()) |
| 15954 | { |
| 15955 | for (uint32_t memTypeIndex = 0; memTypeIndex < GetMemoryTypeCount(); ++memTypeIndex) |
| 15956 | { |
| 15957 | bool displayType = true; |
| 15958 | size_t index = 0; |
| 15959 | for (VmaPool pool = m_Pools.Front(); pool != VMA_NULL; pool = m_Pools.GetNext(item: pool)) |
| 15960 | { |
| 15961 | VmaBlockVector& blockVector = pool->m_BlockVector; |
| 15962 | if (blockVector.GetMemoryTypeIndex() == memTypeIndex) |
| 15963 | { |
| 15964 | if (displayType) |
| 15965 | { |
| 15966 | json.BeginString(pStr: "Type " ); |
| 15967 | json.ContinueString(n: memTypeIndex); |
| 15968 | json.EndString(); |
| 15969 | json.BeginArray(); |
| 15970 | displayType = false; |
| 15971 | } |
| 15972 | |
| 15973 | json.BeginObject(); |
| 15974 | { |
| 15975 | json.WriteString(pStr: "Name" ); |
| 15976 | json.BeginString(); |
| 15977 | json.ContinueString_Size(n: index++); |
| 15978 | if (pool->GetName()) |
| 15979 | { |
| 15980 | json.ContinueString(pStr: " - " ); |
| 15981 | json.ContinueString(pStr: pool->GetName()); |
| 15982 | } |
| 15983 | json.EndString(); |
| 15984 | |
| 15985 | json.WriteString(pStr: "PreferredBlockSize" ); |
| 15986 | json.WriteNumber(n: blockVector.GetPreferredBlockSize()); |
| 15987 | |
| 15988 | json.WriteString(pStr: "Blocks" ); |
| 15989 | blockVector.PrintDetailedMap(json); |
| 15990 | |
| 15991 | json.WriteString(pStr: "DedicatedAllocations" ); |
| 15992 | pool->m_DedicatedAllocations.BuildStatsString(json); |
| 15993 | } |
| 15994 | json.EndObject(); |
| 15995 | } |
| 15996 | } |
| 15997 | |
| 15998 | if (!displayType) |
| 15999 | json.EndArray(); |
| 16000 | } |
| 16001 | } |
| 16002 | } |
| 16003 | json.EndObject(); |
| 16004 | } |
| 16005 | #endif // VMA_STATS_STRING_ENABLED |
| 16006 | #endif // _VMA_ALLOCATOR_T_FUNCTIONS |
| 16007 | |
| 16008 | |
| 16009 | #ifndef _VMA_PUBLIC_INTERFACE |
| 16010 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaCreateAllocator( |
| 16011 | const VmaAllocatorCreateInfo* pCreateInfo, |
| 16012 | VmaAllocator* pAllocator) |
| 16013 | { |
| 16014 | VMA_ASSERT(pCreateInfo && pAllocator); |
| 16015 | VMA_ASSERT(pCreateInfo->vulkanApiVersion == 0 || |
| 16016 | (VK_VERSION_MAJOR(pCreateInfo->vulkanApiVersion) == 1 && VK_VERSION_MINOR(pCreateInfo->vulkanApiVersion) <= 3)); |
| 16017 | VMA_DEBUG_LOG("vmaCreateAllocator" ); |
| 16018 | *pAllocator = vma_new(pCreateInfo->pAllocationCallbacks, VmaAllocator_T)(pCreateInfo); |
| 16019 | VkResult result = (*pAllocator)->Init(pCreateInfo); |
| 16020 | if(result < 0) |
| 16021 | { |
| 16022 | vma_delete(pAllocationCallbacks: pCreateInfo->pAllocationCallbacks, ptr: *pAllocator); |
| 16023 | *pAllocator = VK_NULL_HANDLE; |
| 16024 | } |
| 16025 | return result; |
| 16026 | } |
| 16027 | |
| 16028 | VMA_CALL_PRE void VMA_CALL_POST vmaDestroyAllocator( |
| 16029 | VmaAllocator allocator) |
| 16030 | { |
| 16031 | if(allocator != VK_NULL_HANDLE) |
| 16032 | { |
| 16033 | VMA_DEBUG_LOG("vmaDestroyAllocator" ); |
| 16034 | VkAllocationCallbacks allocationCallbacks = allocator->m_AllocationCallbacks; // Have to copy the callbacks when destroying. |
| 16035 | vma_delete(pAllocationCallbacks: &allocationCallbacks, ptr: allocator); |
| 16036 | } |
| 16037 | } |
| 16038 | |
| 16039 | VMA_CALL_PRE void VMA_CALL_POST vmaGetAllocatorInfo(VmaAllocator allocator, VmaAllocatorInfo* pAllocatorInfo) |
| 16040 | { |
| 16041 | VMA_ASSERT(allocator && pAllocatorInfo); |
| 16042 | pAllocatorInfo->instance = allocator->m_hInstance; |
| 16043 | pAllocatorInfo->physicalDevice = allocator->GetPhysicalDevice(); |
| 16044 | pAllocatorInfo->device = allocator->m_hDevice; |
| 16045 | } |
| 16046 | |
| 16047 | VMA_CALL_PRE void VMA_CALL_POST vmaGetPhysicalDeviceProperties( |
| 16048 | VmaAllocator allocator, |
| 16049 | const VkPhysicalDeviceProperties **ppPhysicalDeviceProperties) |
| 16050 | { |
| 16051 | VMA_ASSERT(allocator && ppPhysicalDeviceProperties); |
| 16052 | *ppPhysicalDeviceProperties = &allocator->m_PhysicalDeviceProperties; |
| 16053 | } |
| 16054 | |
| 16055 | VMA_CALL_PRE void VMA_CALL_POST vmaGetMemoryProperties( |
| 16056 | VmaAllocator allocator, |
| 16057 | const VkPhysicalDeviceMemoryProperties** ppPhysicalDeviceMemoryProperties) |
| 16058 | { |
| 16059 | VMA_ASSERT(allocator && ppPhysicalDeviceMemoryProperties); |
| 16060 | *ppPhysicalDeviceMemoryProperties = &allocator->m_MemProps; |
| 16061 | } |
| 16062 | |
| 16063 | VMA_CALL_PRE void VMA_CALL_POST vmaGetMemoryTypeProperties( |
| 16064 | VmaAllocator allocator, |
| 16065 | uint32_t memoryTypeIndex, |
| 16066 | VkMemoryPropertyFlags* pFlags) |
| 16067 | { |
| 16068 | VMA_ASSERT(allocator && pFlags); |
| 16069 | VMA_ASSERT(memoryTypeIndex < allocator->GetMemoryTypeCount()); |
| 16070 | *pFlags = allocator->m_MemProps.memoryTypes[memoryTypeIndex].propertyFlags; |
| 16071 | } |
| 16072 | |
| 16073 | VMA_CALL_PRE void VMA_CALL_POST vmaSetCurrentFrameIndex( |
| 16074 | VmaAllocator allocator, |
| 16075 | uint32_t frameIndex) |
| 16076 | { |
| 16077 | VMA_ASSERT(allocator); |
| 16078 | |
| 16079 | VMA_DEBUG_GLOBAL_MUTEX_LOCK |
| 16080 | |
| 16081 | allocator->SetCurrentFrameIndex(frameIndex); |
| 16082 | } |
| 16083 | |
| 16084 | VMA_CALL_PRE void VMA_CALL_POST vmaCalculateStatistics( |
| 16085 | VmaAllocator allocator, |
| 16086 | VmaTotalStatistics* pStats) |
| 16087 | { |
| 16088 | VMA_ASSERT(allocator && pStats); |
| 16089 | VMA_DEBUG_GLOBAL_MUTEX_LOCK |
| 16090 | allocator->CalculateStatistics(pStats); |
| 16091 | } |
| 16092 | |
| 16093 | VMA_CALL_PRE void VMA_CALL_POST vmaGetHeapBudgets( |
| 16094 | VmaAllocator allocator, |
| 16095 | VmaBudget* pBudgets) |
| 16096 | { |
| 16097 | VMA_ASSERT(allocator && pBudgets); |
| 16098 | VMA_DEBUG_GLOBAL_MUTEX_LOCK |
| 16099 | allocator->GetHeapBudgets(outBudgets: pBudgets, firstHeap: 0, heapCount: allocator->GetMemoryHeapCount()); |
| 16100 | } |
| 16101 | |
| 16102 | #if VMA_STATS_STRING_ENABLED |
| 16103 | |
| 16104 | VMA_CALL_PRE void VMA_CALL_POST vmaBuildStatsString( |
| 16105 | VmaAllocator allocator, |
| 16106 | char** ppStatsString, |
| 16107 | VkBool32 detailedMap) |
| 16108 | { |
| 16109 | VMA_ASSERT(allocator && ppStatsString); |
| 16110 | VMA_DEBUG_GLOBAL_MUTEX_LOCK |
| 16111 | |
| 16112 | VmaStringBuilder sb(allocator->GetAllocationCallbacks()); |
| 16113 | { |
| 16114 | VmaBudget budgets[VK_MAX_MEMORY_HEAPS]; |
| 16115 | allocator->GetHeapBudgets(outBudgets: budgets, firstHeap: 0, heapCount: allocator->GetMemoryHeapCount()); |
| 16116 | |
| 16117 | VmaTotalStatistics stats; |
| 16118 | allocator->CalculateStatistics(pStats: &stats); |
| 16119 | |
| 16120 | VmaJsonWriter json(allocator->GetAllocationCallbacks(), sb); |
| 16121 | json.BeginObject(); |
| 16122 | { |
| 16123 | json.WriteString(pStr: "General" ); |
| 16124 | json.BeginObject(); |
| 16125 | { |
| 16126 | const VkPhysicalDeviceProperties& deviceProperties = allocator->m_PhysicalDeviceProperties; |
| 16127 | const VkPhysicalDeviceMemoryProperties& memoryProperties = allocator->m_MemProps; |
| 16128 | |
| 16129 | json.WriteString(pStr: "API" ); |
| 16130 | json.WriteString(pStr: "Vulkan" ); |
| 16131 | |
| 16132 | json.WriteString(pStr: "apiVersion" ); |
| 16133 | json.BeginString(); |
| 16134 | json.ContinueString(VK_VERSION_MAJOR(deviceProperties.apiVersion)); |
| 16135 | json.ContinueString(pStr: "." ); |
| 16136 | json.ContinueString(VK_VERSION_MINOR(deviceProperties.apiVersion)); |
| 16137 | json.ContinueString(pStr: "." ); |
| 16138 | json.ContinueString(VK_VERSION_PATCH(deviceProperties.apiVersion)); |
| 16139 | json.EndString(); |
| 16140 | |
| 16141 | json.WriteString(pStr: "GPU" ); |
| 16142 | json.WriteString(pStr: deviceProperties.deviceName); |
| 16143 | json.WriteString(pStr: "deviceType" ); |
| 16144 | json.WriteNumber(n: static_cast<uint32_t>(deviceProperties.deviceType)); |
| 16145 | |
| 16146 | json.WriteString(pStr: "maxMemoryAllocationCount" ); |
| 16147 | json.WriteNumber(n: deviceProperties.limits.maxMemoryAllocationCount); |
| 16148 | json.WriteString(pStr: "bufferImageGranularity" ); |
| 16149 | json.WriteNumber(n: deviceProperties.limits.bufferImageGranularity); |
| 16150 | json.WriteString(pStr: "nonCoherentAtomSize" ); |
| 16151 | json.WriteNumber(n: deviceProperties.limits.nonCoherentAtomSize); |
| 16152 | |
| 16153 | json.WriteString(pStr: "memoryHeapCount" ); |
| 16154 | json.WriteNumber(n: memoryProperties.memoryHeapCount); |
| 16155 | json.WriteString(pStr: "memoryTypeCount" ); |
| 16156 | json.WriteNumber(n: memoryProperties.memoryTypeCount); |
| 16157 | } |
| 16158 | json.EndObject(); |
| 16159 | } |
| 16160 | { |
| 16161 | json.WriteString(pStr: "Total" ); |
| 16162 | VmaPrintDetailedStatistics(json, stat: stats.total); |
| 16163 | } |
| 16164 | { |
| 16165 | json.WriteString(pStr: "MemoryInfo" ); |
| 16166 | json.BeginObject(); |
| 16167 | { |
| 16168 | for (uint32_t heapIndex = 0; heapIndex < allocator->GetMemoryHeapCount(); ++heapIndex) |
| 16169 | { |
| 16170 | json.BeginString(pStr: "Heap " ); |
| 16171 | json.ContinueString(n: heapIndex); |
| 16172 | json.EndString(); |
| 16173 | json.BeginObject(); |
| 16174 | { |
| 16175 | const VkMemoryHeap& heapInfo = allocator->m_MemProps.memoryHeaps[heapIndex]; |
| 16176 | json.WriteString(pStr: "Flags" ); |
| 16177 | json.BeginArray(singleLine: true); |
| 16178 | { |
| 16179 | if (heapInfo.flags & VK_MEMORY_HEAP_DEVICE_LOCAL_BIT) |
| 16180 | json.WriteString(pStr: "DEVICE_LOCAL" ); |
| 16181 | #if VMA_VULKAN_VERSION >= 1001000 |
| 16182 | if (heapInfo.flags & VK_MEMORY_HEAP_MULTI_INSTANCE_BIT) |
| 16183 | json.WriteString(pStr: "MULTI_INSTANCE" ); |
| 16184 | #endif |
| 16185 | |
| 16186 | VkMemoryHeapFlags flags = heapInfo.flags & |
| 16187 | ~(VK_MEMORY_HEAP_DEVICE_LOCAL_BIT |
| 16188 | #if VMA_VULKAN_VERSION >= 1001000 |
| 16189 | | VK_MEMORY_HEAP_MULTI_INSTANCE_BIT |
| 16190 | #endif |
| 16191 | ); |
| 16192 | if (flags != 0) |
| 16193 | json.WriteNumber(n: flags); |
| 16194 | } |
| 16195 | json.EndArray(); |
| 16196 | |
| 16197 | json.WriteString(pStr: "Size" ); |
| 16198 | json.WriteNumber(n: heapInfo.size); |
| 16199 | |
| 16200 | json.WriteString(pStr: "Budget" ); |
| 16201 | json.BeginObject(); |
| 16202 | { |
| 16203 | json.WriteString(pStr: "BudgetBytes" ); |
| 16204 | json.WriteNumber(n: budgets[heapIndex].budget); |
| 16205 | json.WriteString(pStr: "UsageBytes" ); |
| 16206 | json.WriteNumber(n: budgets[heapIndex].usage); |
| 16207 | } |
| 16208 | json.EndObject(); |
| 16209 | |
| 16210 | json.WriteString(pStr: "Stats" ); |
| 16211 | VmaPrintDetailedStatistics(json, stat: stats.memoryHeap[heapIndex]); |
| 16212 | |
| 16213 | json.WriteString(pStr: "MemoryPools" ); |
| 16214 | json.BeginObject(); |
| 16215 | { |
| 16216 | for (uint32_t typeIndex = 0; typeIndex < allocator->GetMemoryTypeCount(); ++typeIndex) |
| 16217 | { |
| 16218 | if (allocator->MemoryTypeIndexToHeapIndex(memTypeIndex: typeIndex) == heapIndex) |
| 16219 | { |
| 16220 | json.BeginString(pStr: "Type " ); |
| 16221 | json.ContinueString(n: typeIndex); |
| 16222 | json.EndString(); |
| 16223 | json.BeginObject(); |
| 16224 | { |
| 16225 | json.WriteString(pStr: "Flags" ); |
| 16226 | json.BeginArray(singleLine: true); |
| 16227 | { |
| 16228 | VkMemoryPropertyFlags flags = allocator->m_MemProps.memoryTypes[typeIndex].propertyFlags; |
| 16229 | if (flags & VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT) |
| 16230 | json.WriteString(pStr: "DEVICE_LOCAL" ); |
| 16231 | if (flags & VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT) |
| 16232 | json.WriteString(pStr: "HOST_VISIBLE" ); |
| 16233 | if (flags & VK_MEMORY_PROPERTY_HOST_COHERENT_BIT) |
| 16234 | json.WriteString(pStr: "HOST_COHERENT" ); |
| 16235 | if (flags & VK_MEMORY_PROPERTY_HOST_CACHED_BIT) |
| 16236 | json.WriteString(pStr: "HOST_CACHED" ); |
| 16237 | if (flags & VK_MEMORY_PROPERTY_LAZILY_ALLOCATED_BIT) |
| 16238 | json.WriteString(pStr: "LAZILY_ALLOCATED" ); |
| 16239 | #if VMA_VULKAN_VERSION >= 1001000 |
| 16240 | if (flags & VK_MEMORY_PROPERTY_PROTECTED_BIT) |
| 16241 | json.WriteString(pStr: "PROTECTED" ); |
| 16242 | #endif |
| 16243 | #if VK_AMD_device_coherent_memory |
| 16244 | if (flags & VK_MEMORY_PROPERTY_DEVICE_COHERENT_BIT_AMD_COPY) |
| 16245 | json.WriteString(pStr: "DEVICE_COHERENT_AMD" ); |
| 16246 | if (flags & VK_MEMORY_PROPERTY_DEVICE_UNCACHED_BIT_AMD_COPY) |
| 16247 | json.WriteString(pStr: "DEVICE_UNCACHED_AMD" ); |
| 16248 | #endif |
| 16249 | |
| 16250 | flags &= ~(VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT |
| 16251 | #if VMA_VULKAN_VERSION >= 1001000 |
| 16252 | | VK_MEMORY_PROPERTY_LAZILY_ALLOCATED_BIT |
| 16253 | #endif |
| 16254 | #if VK_AMD_device_coherent_memory |
| 16255 | | VK_MEMORY_PROPERTY_DEVICE_COHERENT_BIT_AMD_COPY |
| 16256 | | VK_MEMORY_PROPERTY_DEVICE_UNCACHED_BIT_AMD_COPY |
| 16257 | #endif |
| 16258 | | VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT |
| 16259 | | VK_MEMORY_PROPERTY_HOST_COHERENT_BIT |
| 16260 | | VK_MEMORY_PROPERTY_HOST_CACHED_BIT); |
| 16261 | if (flags != 0) |
| 16262 | json.WriteNumber(n: flags); |
| 16263 | } |
| 16264 | json.EndArray(); |
| 16265 | |
| 16266 | json.WriteString(pStr: "Stats" ); |
| 16267 | VmaPrintDetailedStatistics(json, stat: stats.memoryType[typeIndex]); |
| 16268 | } |
| 16269 | json.EndObject(); |
| 16270 | } |
| 16271 | } |
| 16272 | |
| 16273 | } |
| 16274 | json.EndObject(); |
| 16275 | } |
| 16276 | json.EndObject(); |
| 16277 | } |
| 16278 | } |
| 16279 | json.EndObject(); |
| 16280 | } |
| 16281 | |
| 16282 | if (detailedMap == VK_TRUE) |
| 16283 | allocator->PrintDetailedMap(json); |
| 16284 | |
| 16285 | json.EndObject(); |
| 16286 | } |
| 16287 | |
| 16288 | *ppStatsString = VmaCreateStringCopy(allocs: allocator->GetAllocationCallbacks(), srcStr: sb.GetData(), strLen: sb.GetLength()); |
| 16289 | } |
| 16290 | |
| 16291 | VMA_CALL_PRE void VMA_CALL_POST vmaFreeStatsString( |
| 16292 | VmaAllocator allocator, |
| 16293 | char* pStatsString) |
| 16294 | { |
| 16295 | if(pStatsString != VMA_NULL) |
| 16296 | { |
| 16297 | VMA_ASSERT(allocator); |
| 16298 | VmaFreeString(allocs: allocator->GetAllocationCallbacks(), str: pStatsString); |
| 16299 | } |
| 16300 | } |
| 16301 | |
| 16302 | #endif // VMA_STATS_STRING_ENABLED |
| 16303 | |
| 16304 | /* |
| 16305 | This function is not protected by any mutex because it just reads immutable data. |
| 16306 | */ |
| 16307 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaFindMemoryTypeIndex( |
| 16308 | VmaAllocator allocator, |
| 16309 | uint32_t memoryTypeBits, |
| 16310 | const VmaAllocationCreateInfo* pAllocationCreateInfo, |
| 16311 | uint32_t* pMemoryTypeIndex) |
| 16312 | { |
| 16313 | VMA_ASSERT(allocator != VK_NULL_HANDLE); |
| 16314 | VMA_ASSERT(pAllocationCreateInfo != VMA_NULL); |
| 16315 | VMA_ASSERT(pMemoryTypeIndex != VMA_NULL); |
| 16316 | |
| 16317 | return allocator->FindMemoryTypeIndex(memoryTypeBits, pAllocationCreateInfo, UINT32_MAX, pMemoryTypeIndex); |
| 16318 | } |
| 16319 | |
| 16320 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaFindMemoryTypeIndexForBufferInfo( |
| 16321 | VmaAllocator allocator, |
| 16322 | const VkBufferCreateInfo* pBufferCreateInfo, |
| 16323 | const VmaAllocationCreateInfo* pAllocationCreateInfo, |
| 16324 | uint32_t* pMemoryTypeIndex) |
| 16325 | { |
| 16326 | VMA_ASSERT(allocator != VK_NULL_HANDLE); |
| 16327 | VMA_ASSERT(pBufferCreateInfo != VMA_NULL); |
| 16328 | VMA_ASSERT(pAllocationCreateInfo != VMA_NULL); |
| 16329 | VMA_ASSERT(pMemoryTypeIndex != VMA_NULL); |
| 16330 | |
| 16331 | const VkDevice hDev = allocator->m_hDevice; |
| 16332 | const VmaVulkanFunctions* funcs = &allocator->GetVulkanFunctions(); |
| 16333 | VkResult res; |
| 16334 | |
| 16335 | #if VMA_VULKAN_VERSION >= 1003000 |
| 16336 | if(funcs->vkGetDeviceBufferMemoryRequirements) |
| 16337 | { |
| 16338 | // Can query straight from VkBufferCreateInfo :) |
| 16339 | VkDeviceBufferMemoryRequirements devBufMemReq = {.sType: VK_STRUCTURE_TYPE_DEVICE_BUFFER_MEMORY_REQUIREMENTS}; |
| 16340 | devBufMemReq.pCreateInfo = pBufferCreateInfo; |
| 16341 | |
| 16342 | VkMemoryRequirements2 memReq = {.sType: VK_STRUCTURE_TYPE_MEMORY_REQUIREMENTS_2}; |
| 16343 | (*funcs->vkGetDeviceBufferMemoryRequirements)(hDev, &devBufMemReq, &memReq); |
| 16344 | |
| 16345 | res = allocator->FindMemoryTypeIndex( |
| 16346 | memoryTypeBits: memReq.memoryRequirements.memoryTypeBits, pAllocationCreateInfo, bufImgUsage: pBufferCreateInfo->usage, pMemoryTypeIndex); |
| 16347 | } |
| 16348 | else |
| 16349 | #endif // #if VMA_VULKAN_VERSION >= 1003000 |
| 16350 | { |
| 16351 | // Must create a dummy buffer to query :( |
| 16352 | VkBuffer hBuffer = VK_NULL_HANDLE; |
| 16353 | res = funcs->vkCreateBuffer( |
| 16354 | hDev, pBufferCreateInfo, allocator->GetAllocationCallbacks(), &hBuffer); |
| 16355 | if(res == VK_SUCCESS) |
| 16356 | { |
| 16357 | VkMemoryRequirements memReq = {}; |
| 16358 | funcs->vkGetBufferMemoryRequirements(hDev, hBuffer, &memReq); |
| 16359 | |
| 16360 | res = allocator->FindMemoryTypeIndex( |
| 16361 | memoryTypeBits: memReq.memoryTypeBits, pAllocationCreateInfo, bufImgUsage: pBufferCreateInfo->usage, pMemoryTypeIndex); |
| 16362 | |
| 16363 | funcs->vkDestroyBuffer( |
| 16364 | hDev, hBuffer, allocator->GetAllocationCallbacks()); |
| 16365 | } |
| 16366 | } |
| 16367 | return res; |
| 16368 | } |
| 16369 | |
| 16370 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaFindMemoryTypeIndexForImageInfo( |
| 16371 | VmaAllocator allocator, |
| 16372 | const VkImageCreateInfo* pImageCreateInfo, |
| 16373 | const VmaAllocationCreateInfo* pAllocationCreateInfo, |
| 16374 | uint32_t* pMemoryTypeIndex) |
| 16375 | { |
| 16376 | VMA_ASSERT(allocator != VK_NULL_HANDLE); |
| 16377 | VMA_ASSERT(pImageCreateInfo != VMA_NULL); |
| 16378 | VMA_ASSERT(pAllocationCreateInfo != VMA_NULL); |
| 16379 | VMA_ASSERT(pMemoryTypeIndex != VMA_NULL); |
| 16380 | |
| 16381 | const VkDevice hDev = allocator->m_hDevice; |
| 16382 | const VmaVulkanFunctions* funcs = &allocator->GetVulkanFunctions(); |
| 16383 | VkResult res; |
| 16384 | |
| 16385 | #if VMA_VULKAN_VERSION >= 1003000 |
| 16386 | if(funcs->vkGetDeviceImageMemoryRequirements) |
| 16387 | { |
| 16388 | // Can query straight from VkImageCreateInfo :) |
| 16389 | VkDeviceImageMemoryRequirements devImgMemReq = {.sType: VK_STRUCTURE_TYPE_DEVICE_IMAGE_MEMORY_REQUIREMENTS}; |
| 16390 | devImgMemReq.pCreateInfo = pImageCreateInfo; |
| 16391 | VMA_ASSERT(pImageCreateInfo->tiling != VK_IMAGE_TILING_DRM_FORMAT_MODIFIER_EXT_COPY && (pImageCreateInfo->flags & VK_IMAGE_CREATE_DISJOINT_BIT_COPY) == 0 && |
| 16392 | "Cannot use this VkImageCreateInfo with vmaFindMemoryTypeIndexForImageInfo as I don't know what to pass as VkDeviceImageMemoryRequirements::planeAspect." ); |
| 16393 | |
| 16394 | VkMemoryRequirements2 memReq = {.sType: VK_STRUCTURE_TYPE_MEMORY_REQUIREMENTS_2}; |
| 16395 | (*funcs->vkGetDeviceImageMemoryRequirements)(hDev, &devImgMemReq, &memReq); |
| 16396 | |
| 16397 | res = allocator->FindMemoryTypeIndex( |
| 16398 | memoryTypeBits: memReq.memoryRequirements.memoryTypeBits, pAllocationCreateInfo, bufImgUsage: pImageCreateInfo->usage, pMemoryTypeIndex); |
| 16399 | } |
| 16400 | else |
| 16401 | #endif // #if VMA_VULKAN_VERSION >= 1003000 |
| 16402 | { |
| 16403 | // Must create a dummy image to query :( |
| 16404 | VkImage hImage = VK_NULL_HANDLE; |
| 16405 | res = funcs->vkCreateImage( |
| 16406 | hDev, pImageCreateInfo, allocator->GetAllocationCallbacks(), &hImage); |
| 16407 | if(res == VK_SUCCESS) |
| 16408 | { |
| 16409 | VkMemoryRequirements memReq = {}; |
| 16410 | funcs->vkGetImageMemoryRequirements(hDev, hImage, &memReq); |
| 16411 | |
| 16412 | res = allocator->FindMemoryTypeIndex( |
| 16413 | memoryTypeBits: memReq.memoryTypeBits, pAllocationCreateInfo, bufImgUsage: pImageCreateInfo->usage, pMemoryTypeIndex); |
| 16414 | |
| 16415 | funcs->vkDestroyImage( |
| 16416 | hDev, hImage, allocator->GetAllocationCallbacks()); |
| 16417 | } |
| 16418 | } |
| 16419 | return res; |
| 16420 | } |
| 16421 | |
| 16422 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaCreatePool( |
| 16423 | VmaAllocator allocator, |
| 16424 | const VmaPoolCreateInfo* pCreateInfo, |
| 16425 | VmaPool* pPool) |
| 16426 | { |
| 16427 | VMA_ASSERT(allocator && pCreateInfo && pPool); |
| 16428 | |
| 16429 | VMA_DEBUG_LOG("vmaCreatePool" ); |
| 16430 | |
| 16431 | VMA_DEBUG_GLOBAL_MUTEX_LOCK |
| 16432 | |
| 16433 | return allocator->CreatePool(pCreateInfo, pPool); |
| 16434 | } |
| 16435 | |
| 16436 | VMA_CALL_PRE void VMA_CALL_POST vmaDestroyPool( |
| 16437 | VmaAllocator allocator, |
| 16438 | VmaPool pool) |
| 16439 | { |
| 16440 | VMA_ASSERT(allocator); |
| 16441 | |
| 16442 | if(pool == VK_NULL_HANDLE) |
| 16443 | { |
| 16444 | return; |
| 16445 | } |
| 16446 | |
| 16447 | VMA_DEBUG_LOG("vmaDestroyPool" ); |
| 16448 | |
| 16449 | VMA_DEBUG_GLOBAL_MUTEX_LOCK |
| 16450 | |
| 16451 | allocator->DestroyPool(pool); |
| 16452 | } |
| 16453 | |
| 16454 | VMA_CALL_PRE void VMA_CALL_POST vmaGetPoolStatistics( |
| 16455 | VmaAllocator allocator, |
| 16456 | VmaPool pool, |
| 16457 | VmaStatistics* pPoolStats) |
| 16458 | { |
| 16459 | VMA_ASSERT(allocator && pool && pPoolStats); |
| 16460 | |
| 16461 | VMA_DEBUG_GLOBAL_MUTEX_LOCK |
| 16462 | |
| 16463 | allocator->GetPoolStatistics(pool, pPoolStats); |
| 16464 | } |
| 16465 | |
| 16466 | VMA_CALL_PRE void VMA_CALL_POST vmaCalculatePoolStatistics( |
| 16467 | VmaAllocator allocator, |
| 16468 | VmaPool pool, |
| 16469 | VmaDetailedStatistics* pPoolStats) |
| 16470 | { |
| 16471 | VMA_ASSERT(allocator && pool && pPoolStats); |
| 16472 | |
| 16473 | VMA_DEBUG_GLOBAL_MUTEX_LOCK |
| 16474 | |
| 16475 | allocator->CalculatePoolStatistics(pool, pPoolStats); |
| 16476 | } |
| 16477 | |
| 16478 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaCheckPoolCorruption(VmaAllocator allocator, VmaPool pool) |
| 16479 | { |
| 16480 | VMA_ASSERT(allocator && pool); |
| 16481 | |
| 16482 | VMA_DEBUG_GLOBAL_MUTEX_LOCK |
| 16483 | |
| 16484 | VMA_DEBUG_LOG("vmaCheckPoolCorruption" ); |
| 16485 | |
| 16486 | return allocator->CheckPoolCorruption(hPool: pool); |
| 16487 | } |
| 16488 | |
| 16489 | VMA_CALL_PRE void VMA_CALL_POST vmaGetPoolName( |
| 16490 | VmaAllocator allocator, |
| 16491 | VmaPool pool, |
| 16492 | const char** ppName) |
| 16493 | { |
| 16494 | VMA_ASSERT(allocator && pool && ppName); |
| 16495 | |
| 16496 | VMA_DEBUG_LOG("vmaGetPoolName" ); |
| 16497 | |
| 16498 | VMA_DEBUG_GLOBAL_MUTEX_LOCK |
| 16499 | |
| 16500 | *ppName = pool->GetName(); |
| 16501 | } |
| 16502 | |
| 16503 | VMA_CALL_PRE void VMA_CALL_POST vmaSetPoolName( |
| 16504 | VmaAllocator allocator, |
| 16505 | VmaPool pool, |
| 16506 | const char* pName) |
| 16507 | { |
| 16508 | VMA_ASSERT(allocator && pool); |
| 16509 | |
| 16510 | VMA_DEBUG_LOG("vmaSetPoolName" ); |
| 16511 | |
| 16512 | VMA_DEBUG_GLOBAL_MUTEX_LOCK |
| 16513 | |
| 16514 | pool->SetName(pName); |
| 16515 | } |
| 16516 | |
| 16517 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaAllocateMemory( |
| 16518 | VmaAllocator allocator, |
| 16519 | const VkMemoryRequirements* pVkMemoryRequirements, |
| 16520 | const VmaAllocationCreateInfo* pCreateInfo, |
| 16521 | VmaAllocation* pAllocation, |
| 16522 | VmaAllocationInfo* pAllocationInfo) |
| 16523 | { |
| 16524 | VMA_ASSERT(allocator && pVkMemoryRequirements && pCreateInfo && pAllocation); |
| 16525 | |
| 16526 | VMA_DEBUG_LOG("vmaAllocateMemory" ); |
| 16527 | |
| 16528 | VMA_DEBUG_GLOBAL_MUTEX_LOCK |
| 16529 | |
| 16530 | VkResult result = allocator->AllocateMemory( |
| 16531 | vkMemReq: *pVkMemoryRequirements, |
| 16532 | requiresDedicatedAllocation: false, // requiresDedicatedAllocation |
| 16533 | prefersDedicatedAllocation: false, // prefersDedicatedAllocation |
| 16534 | VK_NULL_HANDLE, // dedicatedBuffer |
| 16535 | VK_NULL_HANDLE, // dedicatedImage |
| 16536 | UINT32_MAX, // dedicatedBufferImageUsage |
| 16537 | createInfo: *pCreateInfo, |
| 16538 | suballocType: VMA_SUBALLOCATION_TYPE_UNKNOWN, |
| 16539 | allocationCount: 1, // allocationCount |
| 16540 | pAllocations: pAllocation); |
| 16541 | |
| 16542 | if(pAllocationInfo != VMA_NULL && result == VK_SUCCESS) |
| 16543 | { |
| 16544 | allocator->GetAllocationInfo(hAllocation: *pAllocation, pAllocationInfo); |
| 16545 | } |
| 16546 | |
| 16547 | return result; |
| 16548 | } |
| 16549 | |
| 16550 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaAllocateMemoryPages( |
| 16551 | VmaAllocator allocator, |
| 16552 | const VkMemoryRequirements* pVkMemoryRequirements, |
| 16553 | const VmaAllocationCreateInfo* pCreateInfo, |
| 16554 | size_t allocationCount, |
| 16555 | VmaAllocation* pAllocations, |
| 16556 | VmaAllocationInfo* pAllocationInfo) |
| 16557 | { |
| 16558 | if(allocationCount == 0) |
| 16559 | { |
| 16560 | return VK_SUCCESS; |
| 16561 | } |
| 16562 | |
| 16563 | VMA_ASSERT(allocator && pVkMemoryRequirements && pCreateInfo && pAllocations); |
| 16564 | |
| 16565 | VMA_DEBUG_LOG("vmaAllocateMemoryPages" ); |
| 16566 | |
| 16567 | VMA_DEBUG_GLOBAL_MUTEX_LOCK |
| 16568 | |
| 16569 | VkResult result = allocator->AllocateMemory( |
| 16570 | vkMemReq: *pVkMemoryRequirements, |
| 16571 | requiresDedicatedAllocation: false, // requiresDedicatedAllocation |
| 16572 | prefersDedicatedAllocation: false, // prefersDedicatedAllocation |
| 16573 | VK_NULL_HANDLE, // dedicatedBuffer |
| 16574 | VK_NULL_HANDLE, // dedicatedImage |
| 16575 | UINT32_MAX, // dedicatedBufferImageUsage |
| 16576 | createInfo: *pCreateInfo, |
| 16577 | suballocType: VMA_SUBALLOCATION_TYPE_UNKNOWN, |
| 16578 | allocationCount, |
| 16579 | pAllocations); |
| 16580 | |
| 16581 | if(pAllocationInfo != VMA_NULL && result == VK_SUCCESS) |
| 16582 | { |
| 16583 | for(size_t i = 0; i < allocationCount; ++i) |
| 16584 | { |
| 16585 | allocator->GetAllocationInfo(hAllocation: pAllocations[i], pAllocationInfo: pAllocationInfo + i); |
| 16586 | } |
| 16587 | } |
| 16588 | |
| 16589 | return result; |
| 16590 | } |
| 16591 | |
| 16592 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaAllocateMemoryForBuffer( |
| 16593 | VmaAllocator allocator, |
| 16594 | VkBuffer buffer, |
| 16595 | const VmaAllocationCreateInfo* pCreateInfo, |
| 16596 | VmaAllocation* pAllocation, |
| 16597 | VmaAllocationInfo* pAllocationInfo) |
| 16598 | { |
| 16599 | VMA_ASSERT(allocator && buffer != VK_NULL_HANDLE && pCreateInfo && pAllocation); |
| 16600 | |
| 16601 | VMA_DEBUG_LOG("vmaAllocateMemoryForBuffer" ); |
| 16602 | |
| 16603 | VMA_DEBUG_GLOBAL_MUTEX_LOCK |
| 16604 | |
| 16605 | VkMemoryRequirements vkMemReq = {}; |
| 16606 | bool requiresDedicatedAllocation = false; |
| 16607 | bool prefersDedicatedAllocation = false; |
| 16608 | allocator->GetBufferMemoryRequirements(hBuffer: buffer, memReq&: vkMemReq, |
| 16609 | requiresDedicatedAllocation, |
| 16610 | prefersDedicatedAllocation); |
| 16611 | |
| 16612 | VkResult result = allocator->AllocateMemory( |
| 16613 | vkMemReq, |
| 16614 | requiresDedicatedAllocation, |
| 16615 | prefersDedicatedAllocation, |
| 16616 | dedicatedBuffer: buffer, // dedicatedBuffer |
| 16617 | VK_NULL_HANDLE, // dedicatedImage |
| 16618 | UINT32_MAX, // dedicatedBufferImageUsage |
| 16619 | createInfo: *pCreateInfo, |
| 16620 | suballocType: VMA_SUBALLOCATION_TYPE_BUFFER, |
| 16621 | allocationCount: 1, // allocationCount |
| 16622 | pAllocations: pAllocation); |
| 16623 | |
| 16624 | if(pAllocationInfo && result == VK_SUCCESS) |
| 16625 | { |
| 16626 | allocator->GetAllocationInfo(hAllocation: *pAllocation, pAllocationInfo); |
| 16627 | } |
| 16628 | |
| 16629 | return result; |
| 16630 | } |
| 16631 | |
| 16632 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaAllocateMemoryForImage( |
| 16633 | VmaAllocator allocator, |
| 16634 | VkImage image, |
| 16635 | const VmaAllocationCreateInfo* pCreateInfo, |
| 16636 | VmaAllocation* pAllocation, |
| 16637 | VmaAllocationInfo* pAllocationInfo) |
| 16638 | { |
| 16639 | VMA_ASSERT(allocator && image != VK_NULL_HANDLE && pCreateInfo && pAllocation); |
| 16640 | |
| 16641 | VMA_DEBUG_LOG("vmaAllocateMemoryForImage" ); |
| 16642 | |
| 16643 | VMA_DEBUG_GLOBAL_MUTEX_LOCK |
| 16644 | |
| 16645 | VkMemoryRequirements vkMemReq = {}; |
| 16646 | bool requiresDedicatedAllocation = false; |
| 16647 | bool prefersDedicatedAllocation = false; |
| 16648 | allocator->GetImageMemoryRequirements(hImage: image, memReq&: vkMemReq, |
| 16649 | requiresDedicatedAllocation, prefersDedicatedAllocation); |
| 16650 | |
| 16651 | VkResult result = allocator->AllocateMemory( |
| 16652 | vkMemReq, |
| 16653 | requiresDedicatedAllocation, |
| 16654 | prefersDedicatedAllocation, |
| 16655 | VK_NULL_HANDLE, // dedicatedBuffer |
| 16656 | dedicatedImage: image, // dedicatedImage |
| 16657 | UINT32_MAX, // dedicatedBufferImageUsage |
| 16658 | createInfo: *pCreateInfo, |
| 16659 | suballocType: VMA_SUBALLOCATION_TYPE_IMAGE_UNKNOWN, |
| 16660 | allocationCount: 1, // allocationCount |
| 16661 | pAllocations: pAllocation); |
| 16662 | |
| 16663 | if(pAllocationInfo && result == VK_SUCCESS) |
| 16664 | { |
| 16665 | allocator->GetAllocationInfo(hAllocation: *pAllocation, pAllocationInfo); |
| 16666 | } |
| 16667 | |
| 16668 | return result; |
| 16669 | } |
| 16670 | |
| 16671 | VMA_CALL_PRE void VMA_CALL_POST vmaFreeMemory( |
| 16672 | VmaAllocator allocator, |
| 16673 | VmaAllocation allocation) |
| 16674 | { |
| 16675 | VMA_ASSERT(allocator); |
| 16676 | |
| 16677 | if(allocation == VK_NULL_HANDLE) |
| 16678 | { |
| 16679 | return; |
| 16680 | } |
| 16681 | |
| 16682 | VMA_DEBUG_LOG("vmaFreeMemory" ); |
| 16683 | |
| 16684 | VMA_DEBUG_GLOBAL_MUTEX_LOCK |
| 16685 | |
| 16686 | allocator->FreeMemory( |
| 16687 | allocationCount: 1, // allocationCount |
| 16688 | pAllocations: &allocation); |
| 16689 | } |
| 16690 | |
| 16691 | VMA_CALL_PRE void VMA_CALL_POST vmaFreeMemoryPages( |
| 16692 | VmaAllocator allocator, |
| 16693 | size_t allocationCount, |
| 16694 | const VmaAllocation* pAllocations) |
| 16695 | { |
| 16696 | if(allocationCount == 0) |
| 16697 | { |
| 16698 | return; |
| 16699 | } |
| 16700 | |
| 16701 | VMA_ASSERT(allocator); |
| 16702 | |
| 16703 | VMA_DEBUG_LOG("vmaFreeMemoryPages" ); |
| 16704 | |
| 16705 | VMA_DEBUG_GLOBAL_MUTEX_LOCK |
| 16706 | |
| 16707 | allocator->FreeMemory(allocationCount, pAllocations); |
| 16708 | } |
| 16709 | |
| 16710 | VMA_CALL_PRE void VMA_CALL_POST vmaGetAllocationInfo( |
| 16711 | VmaAllocator allocator, |
| 16712 | VmaAllocation allocation, |
| 16713 | VmaAllocationInfo* pAllocationInfo) |
| 16714 | { |
| 16715 | VMA_ASSERT(allocator && allocation && pAllocationInfo); |
| 16716 | |
| 16717 | VMA_DEBUG_GLOBAL_MUTEX_LOCK |
| 16718 | |
| 16719 | allocator->GetAllocationInfo(hAllocation: allocation, pAllocationInfo); |
| 16720 | } |
| 16721 | |
| 16722 | VMA_CALL_PRE void VMA_CALL_POST vmaSetAllocationUserData( |
| 16723 | VmaAllocator allocator, |
| 16724 | VmaAllocation allocation, |
| 16725 | void* pUserData) |
| 16726 | { |
| 16727 | VMA_ASSERT(allocator && allocation); |
| 16728 | |
| 16729 | VMA_DEBUG_GLOBAL_MUTEX_LOCK |
| 16730 | |
| 16731 | allocation->SetUserData(hAllocator: allocator, pUserData); |
| 16732 | } |
| 16733 | |
| 16734 | VMA_CALL_PRE void VMA_CALL_POST vmaSetAllocationName( |
| 16735 | VmaAllocator VMA_NOT_NULL allocator, |
| 16736 | VmaAllocation VMA_NOT_NULL allocation, |
| 16737 | const char* VMA_NULLABLE pName) |
| 16738 | { |
| 16739 | allocation->SetName(hAllocator: allocator, pName); |
| 16740 | } |
| 16741 | |
| 16742 | VMA_CALL_PRE void VMA_CALL_POST vmaGetAllocationMemoryProperties( |
| 16743 | VmaAllocator VMA_NOT_NULL allocator, |
| 16744 | VmaAllocation VMA_NOT_NULL allocation, |
| 16745 | VkMemoryPropertyFlags* VMA_NOT_NULL pFlags) |
| 16746 | { |
| 16747 | VMA_ASSERT(allocator && allocation && pFlags); |
| 16748 | const uint32_t memTypeIndex = allocation->GetMemoryTypeIndex(); |
| 16749 | *pFlags = allocator->m_MemProps.memoryTypes[memTypeIndex].propertyFlags; |
| 16750 | } |
| 16751 | |
| 16752 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaMapMemory( |
| 16753 | VmaAllocator allocator, |
| 16754 | VmaAllocation allocation, |
| 16755 | void** ppData) |
| 16756 | { |
| 16757 | VMA_ASSERT(allocator && allocation && ppData); |
| 16758 | |
| 16759 | VMA_DEBUG_GLOBAL_MUTEX_LOCK |
| 16760 | |
| 16761 | return allocator->Map(hAllocation: allocation, ppData); |
| 16762 | } |
| 16763 | |
| 16764 | VMA_CALL_PRE void VMA_CALL_POST vmaUnmapMemory( |
| 16765 | VmaAllocator allocator, |
| 16766 | VmaAllocation allocation) |
| 16767 | { |
| 16768 | VMA_ASSERT(allocator && allocation); |
| 16769 | |
| 16770 | VMA_DEBUG_GLOBAL_MUTEX_LOCK |
| 16771 | |
| 16772 | allocator->Unmap(hAllocation: allocation); |
| 16773 | } |
| 16774 | |
| 16775 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaFlushAllocation( |
| 16776 | VmaAllocator allocator, |
| 16777 | VmaAllocation allocation, |
| 16778 | VkDeviceSize offset, |
| 16779 | VkDeviceSize size) |
| 16780 | { |
| 16781 | VMA_ASSERT(allocator && allocation); |
| 16782 | |
| 16783 | VMA_DEBUG_LOG("vmaFlushAllocation" ); |
| 16784 | |
| 16785 | VMA_DEBUG_GLOBAL_MUTEX_LOCK |
| 16786 | |
| 16787 | const VkResult res = allocator->FlushOrInvalidateAllocation(hAllocation: allocation, offset, size, op: VMA_CACHE_FLUSH); |
| 16788 | |
| 16789 | return res; |
| 16790 | } |
| 16791 | |
| 16792 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaInvalidateAllocation( |
| 16793 | VmaAllocator allocator, |
| 16794 | VmaAllocation allocation, |
| 16795 | VkDeviceSize offset, |
| 16796 | VkDeviceSize size) |
| 16797 | { |
| 16798 | VMA_ASSERT(allocator && allocation); |
| 16799 | |
| 16800 | VMA_DEBUG_LOG("vmaInvalidateAllocation" ); |
| 16801 | |
| 16802 | VMA_DEBUG_GLOBAL_MUTEX_LOCK |
| 16803 | |
| 16804 | const VkResult res = allocator->FlushOrInvalidateAllocation(hAllocation: allocation, offset, size, op: VMA_CACHE_INVALIDATE); |
| 16805 | |
| 16806 | return res; |
| 16807 | } |
| 16808 | |
| 16809 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaFlushAllocations( |
| 16810 | VmaAllocator allocator, |
| 16811 | uint32_t allocationCount, |
| 16812 | const VmaAllocation* allocations, |
| 16813 | const VkDeviceSize* offsets, |
| 16814 | const VkDeviceSize* sizes) |
| 16815 | { |
| 16816 | VMA_ASSERT(allocator); |
| 16817 | |
| 16818 | if(allocationCount == 0) |
| 16819 | { |
| 16820 | return VK_SUCCESS; |
| 16821 | } |
| 16822 | |
| 16823 | VMA_ASSERT(allocations); |
| 16824 | |
| 16825 | VMA_DEBUG_LOG("vmaFlushAllocations" ); |
| 16826 | |
| 16827 | VMA_DEBUG_GLOBAL_MUTEX_LOCK |
| 16828 | |
| 16829 | const VkResult res = allocator->FlushOrInvalidateAllocations(allocationCount, allocations, offsets, sizes, op: VMA_CACHE_FLUSH); |
| 16830 | |
| 16831 | return res; |
| 16832 | } |
| 16833 | |
| 16834 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaInvalidateAllocations( |
| 16835 | VmaAllocator allocator, |
| 16836 | uint32_t allocationCount, |
| 16837 | const VmaAllocation* allocations, |
| 16838 | const VkDeviceSize* offsets, |
| 16839 | const VkDeviceSize* sizes) |
| 16840 | { |
| 16841 | VMA_ASSERT(allocator); |
| 16842 | |
| 16843 | if(allocationCount == 0) |
| 16844 | { |
| 16845 | return VK_SUCCESS; |
| 16846 | } |
| 16847 | |
| 16848 | VMA_ASSERT(allocations); |
| 16849 | |
| 16850 | VMA_DEBUG_LOG("vmaInvalidateAllocations" ); |
| 16851 | |
| 16852 | VMA_DEBUG_GLOBAL_MUTEX_LOCK |
| 16853 | |
| 16854 | const VkResult res = allocator->FlushOrInvalidateAllocations(allocationCount, allocations, offsets, sizes, op: VMA_CACHE_INVALIDATE); |
| 16855 | |
| 16856 | return res; |
| 16857 | } |
| 16858 | |
| 16859 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaCheckCorruption( |
| 16860 | VmaAllocator allocator, |
| 16861 | uint32_t memoryTypeBits) |
| 16862 | { |
| 16863 | VMA_ASSERT(allocator); |
| 16864 | |
| 16865 | VMA_DEBUG_LOG("vmaCheckCorruption" ); |
| 16866 | |
| 16867 | VMA_DEBUG_GLOBAL_MUTEX_LOCK |
| 16868 | |
| 16869 | return allocator->CheckCorruption(memoryTypeBits); |
| 16870 | } |
| 16871 | |
| 16872 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaBeginDefragmentation( |
| 16873 | VmaAllocator allocator, |
| 16874 | const VmaDefragmentationInfo* pInfo, |
| 16875 | VmaDefragmentationContext* pContext) |
| 16876 | { |
| 16877 | VMA_ASSERT(allocator && pInfo && pContext); |
| 16878 | |
| 16879 | VMA_DEBUG_LOG("vmaBeginDefragmentation" ); |
| 16880 | |
| 16881 | if (pInfo->pool != VMA_NULL) |
| 16882 | { |
| 16883 | // Check if run on supported algorithms |
| 16884 | if (pInfo->pool->m_BlockVector.GetAlgorithm() & VMA_POOL_CREATE_LINEAR_ALGORITHM_BIT) |
| 16885 | return VK_ERROR_FEATURE_NOT_PRESENT; |
| 16886 | } |
| 16887 | |
| 16888 | VMA_DEBUG_GLOBAL_MUTEX_LOCK |
| 16889 | |
| 16890 | *pContext = vma_new(allocator, VmaDefragmentationContext_T)(allocator, *pInfo); |
| 16891 | return VK_SUCCESS; |
| 16892 | } |
| 16893 | |
| 16894 | VMA_CALL_PRE void VMA_CALL_POST vmaEndDefragmentation( |
| 16895 | VmaAllocator allocator, |
| 16896 | VmaDefragmentationContext context, |
| 16897 | VmaDefragmentationStats* pStats) |
| 16898 | { |
| 16899 | VMA_ASSERT(allocator && context); |
| 16900 | |
| 16901 | VMA_DEBUG_LOG("vmaEndDefragmentation" ); |
| 16902 | |
| 16903 | VMA_DEBUG_GLOBAL_MUTEX_LOCK |
| 16904 | |
| 16905 | if (pStats) |
| 16906 | context->GetStats(outStats&: *pStats); |
| 16907 | vma_delete(hAllocator: allocator, ptr: context); |
| 16908 | } |
| 16909 | |
| 16910 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaBeginDefragmentationPass( |
| 16911 | VmaAllocator VMA_NOT_NULL allocator, |
| 16912 | VmaDefragmentationContext VMA_NOT_NULL context, |
| 16913 | VmaDefragmentationPassMoveInfo* VMA_NOT_NULL pPassInfo) |
| 16914 | { |
| 16915 | VMA_ASSERT(context && pPassInfo); |
| 16916 | |
| 16917 | VMA_DEBUG_LOG("vmaBeginDefragmentationPass" ); |
| 16918 | |
| 16919 | VMA_DEBUG_GLOBAL_MUTEX_LOCK |
| 16920 | |
| 16921 | return context->DefragmentPassBegin(moveInfo&: *pPassInfo); |
| 16922 | } |
| 16923 | |
| 16924 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaEndDefragmentationPass( |
| 16925 | VmaAllocator VMA_NOT_NULL allocator, |
| 16926 | VmaDefragmentationContext VMA_NOT_NULL context, |
| 16927 | VmaDefragmentationPassMoveInfo* VMA_NOT_NULL pPassInfo) |
| 16928 | { |
| 16929 | VMA_ASSERT(context && pPassInfo); |
| 16930 | |
| 16931 | VMA_DEBUG_LOG("vmaEndDefragmentationPass" ); |
| 16932 | |
| 16933 | VMA_DEBUG_GLOBAL_MUTEX_LOCK |
| 16934 | |
| 16935 | return context->DefragmentPassEnd(moveInfo&: *pPassInfo); |
| 16936 | } |
| 16937 | |
| 16938 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaBindBufferMemory( |
| 16939 | VmaAllocator allocator, |
| 16940 | VmaAllocation allocation, |
| 16941 | VkBuffer buffer) |
| 16942 | { |
| 16943 | VMA_ASSERT(allocator && allocation && buffer); |
| 16944 | |
| 16945 | VMA_DEBUG_LOG("vmaBindBufferMemory" ); |
| 16946 | |
| 16947 | VMA_DEBUG_GLOBAL_MUTEX_LOCK |
| 16948 | |
| 16949 | return allocator->BindBufferMemory(hAllocation: allocation, allocationLocalOffset: 0, hBuffer: buffer, VMA_NULL); |
| 16950 | } |
| 16951 | |
| 16952 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaBindBufferMemory2( |
| 16953 | VmaAllocator allocator, |
| 16954 | VmaAllocation allocation, |
| 16955 | VkDeviceSize allocationLocalOffset, |
| 16956 | VkBuffer buffer, |
| 16957 | const void* pNext) |
| 16958 | { |
| 16959 | VMA_ASSERT(allocator && allocation && buffer); |
| 16960 | |
| 16961 | VMA_DEBUG_LOG("vmaBindBufferMemory2" ); |
| 16962 | |
| 16963 | VMA_DEBUG_GLOBAL_MUTEX_LOCK |
| 16964 | |
| 16965 | return allocator->BindBufferMemory(hAllocation: allocation, allocationLocalOffset, hBuffer: buffer, pNext); |
| 16966 | } |
| 16967 | |
| 16968 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaBindImageMemory( |
| 16969 | VmaAllocator allocator, |
| 16970 | VmaAllocation allocation, |
| 16971 | VkImage image) |
| 16972 | { |
| 16973 | VMA_ASSERT(allocator && allocation && image); |
| 16974 | |
| 16975 | VMA_DEBUG_LOG("vmaBindImageMemory" ); |
| 16976 | |
| 16977 | VMA_DEBUG_GLOBAL_MUTEX_LOCK |
| 16978 | |
| 16979 | return allocator->BindImageMemory(hAllocation: allocation, allocationLocalOffset: 0, hImage: image, VMA_NULL); |
| 16980 | } |
| 16981 | |
| 16982 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaBindImageMemory2( |
| 16983 | VmaAllocator allocator, |
| 16984 | VmaAllocation allocation, |
| 16985 | VkDeviceSize allocationLocalOffset, |
| 16986 | VkImage image, |
| 16987 | const void* pNext) |
| 16988 | { |
| 16989 | VMA_ASSERT(allocator && allocation && image); |
| 16990 | |
| 16991 | VMA_DEBUG_LOG("vmaBindImageMemory2" ); |
| 16992 | |
| 16993 | VMA_DEBUG_GLOBAL_MUTEX_LOCK |
| 16994 | |
| 16995 | return allocator->BindImageMemory(hAllocation: allocation, allocationLocalOffset, hImage: image, pNext); |
| 16996 | } |
| 16997 | |
| 16998 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaCreateBuffer( |
| 16999 | VmaAllocator allocator, |
| 17000 | const VkBufferCreateInfo* pBufferCreateInfo, |
| 17001 | const VmaAllocationCreateInfo* pAllocationCreateInfo, |
| 17002 | VkBuffer* pBuffer, |
| 17003 | VmaAllocation* pAllocation, |
| 17004 | VmaAllocationInfo* pAllocationInfo) |
| 17005 | { |
| 17006 | VMA_ASSERT(allocator && pBufferCreateInfo && pAllocationCreateInfo && pBuffer && pAllocation); |
| 17007 | |
| 17008 | if(pBufferCreateInfo->size == 0) |
| 17009 | { |
| 17010 | return VK_ERROR_INITIALIZATION_FAILED; |
| 17011 | } |
| 17012 | if((pBufferCreateInfo->usage & VK_BUFFER_USAGE_SHADER_DEVICE_ADDRESS_BIT_COPY) != 0 && |
| 17013 | !allocator->m_UseKhrBufferDeviceAddress) |
| 17014 | { |
| 17015 | VMA_ASSERT(0 && "Creating a buffer with VK_BUFFER_USAGE_SHADER_DEVICE_ADDRESS_BIT is not valid if VMA_ALLOCATOR_CREATE_BUFFER_DEVICE_ADDRESS_BIT was not used." ); |
| 17016 | return VK_ERROR_INITIALIZATION_FAILED; |
| 17017 | } |
| 17018 | |
| 17019 | VMA_DEBUG_LOG("vmaCreateBuffer" ); |
| 17020 | |
| 17021 | VMA_DEBUG_GLOBAL_MUTEX_LOCK |
| 17022 | |
| 17023 | *pBuffer = VK_NULL_HANDLE; |
| 17024 | *pAllocation = VK_NULL_HANDLE; |
| 17025 | |
| 17026 | // 1. Create VkBuffer. |
| 17027 | VkResult res = (*allocator->GetVulkanFunctions().vkCreateBuffer)( |
| 17028 | allocator->m_hDevice, |
| 17029 | pBufferCreateInfo, |
| 17030 | allocator->GetAllocationCallbacks(), |
| 17031 | pBuffer); |
| 17032 | if(res >= 0) |
| 17033 | { |
| 17034 | // 2. vkGetBufferMemoryRequirements. |
| 17035 | VkMemoryRequirements vkMemReq = {}; |
| 17036 | bool requiresDedicatedAllocation = false; |
| 17037 | bool prefersDedicatedAllocation = false; |
| 17038 | allocator->GetBufferMemoryRequirements(hBuffer: *pBuffer, memReq&: vkMemReq, |
| 17039 | requiresDedicatedAllocation, prefersDedicatedAllocation); |
| 17040 | |
| 17041 | // 3. Allocate memory using allocator. |
| 17042 | res = allocator->AllocateMemory( |
| 17043 | vkMemReq, |
| 17044 | requiresDedicatedAllocation, |
| 17045 | prefersDedicatedAllocation, |
| 17046 | dedicatedBuffer: *pBuffer, // dedicatedBuffer |
| 17047 | VK_NULL_HANDLE, // dedicatedImage |
| 17048 | dedicatedBufferImageUsage: pBufferCreateInfo->usage, // dedicatedBufferImageUsage |
| 17049 | createInfo: *pAllocationCreateInfo, |
| 17050 | suballocType: VMA_SUBALLOCATION_TYPE_BUFFER, |
| 17051 | allocationCount: 1, // allocationCount |
| 17052 | pAllocations: pAllocation); |
| 17053 | |
| 17054 | if(res >= 0) |
| 17055 | { |
| 17056 | // 3. Bind buffer with memory. |
| 17057 | if((pAllocationCreateInfo->flags & VMA_ALLOCATION_CREATE_DONT_BIND_BIT) == 0) |
| 17058 | { |
| 17059 | res = allocator->BindBufferMemory(hAllocation: *pAllocation, allocationLocalOffset: 0, hBuffer: *pBuffer, VMA_NULL); |
| 17060 | } |
| 17061 | if(res >= 0) |
| 17062 | { |
| 17063 | // All steps succeeded. |
| 17064 | #if VMA_STATS_STRING_ENABLED |
| 17065 | (*pAllocation)->InitBufferImageUsage(bufferImageUsage: pBufferCreateInfo->usage); |
| 17066 | #endif |
| 17067 | if(pAllocationInfo != VMA_NULL) |
| 17068 | { |
| 17069 | allocator->GetAllocationInfo(hAllocation: *pAllocation, pAllocationInfo); |
| 17070 | } |
| 17071 | |
| 17072 | return VK_SUCCESS; |
| 17073 | } |
| 17074 | allocator->FreeMemory( |
| 17075 | allocationCount: 1, // allocationCount |
| 17076 | pAllocations: pAllocation); |
| 17077 | *pAllocation = VK_NULL_HANDLE; |
| 17078 | (*allocator->GetVulkanFunctions().vkDestroyBuffer)(allocator->m_hDevice, *pBuffer, allocator->GetAllocationCallbacks()); |
| 17079 | *pBuffer = VK_NULL_HANDLE; |
| 17080 | return res; |
| 17081 | } |
| 17082 | (*allocator->GetVulkanFunctions().vkDestroyBuffer)(allocator->m_hDevice, *pBuffer, allocator->GetAllocationCallbacks()); |
| 17083 | *pBuffer = VK_NULL_HANDLE; |
| 17084 | return res; |
| 17085 | } |
| 17086 | return res; |
| 17087 | } |
| 17088 | |
| 17089 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaCreateBufferWithAlignment( |
| 17090 | VmaAllocator allocator, |
| 17091 | const VkBufferCreateInfo* pBufferCreateInfo, |
| 17092 | const VmaAllocationCreateInfo* pAllocationCreateInfo, |
| 17093 | VkDeviceSize minAlignment, |
| 17094 | VkBuffer* pBuffer, |
| 17095 | VmaAllocation* pAllocation, |
| 17096 | VmaAllocationInfo* pAllocationInfo) |
| 17097 | { |
| 17098 | VMA_ASSERT(allocator && pBufferCreateInfo && pAllocationCreateInfo && VmaIsPow2(minAlignment) && pBuffer && pAllocation); |
| 17099 | |
| 17100 | if(pBufferCreateInfo->size == 0) |
| 17101 | { |
| 17102 | return VK_ERROR_INITIALIZATION_FAILED; |
| 17103 | } |
| 17104 | if((pBufferCreateInfo->usage & VK_BUFFER_USAGE_SHADER_DEVICE_ADDRESS_BIT_COPY) != 0 && |
| 17105 | !allocator->m_UseKhrBufferDeviceAddress) |
| 17106 | { |
| 17107 | VMA_ASSERT(0 && "Creating a buffer with VK_BUFFER_USAGE_SHADER_DEVICE_ADDRESS_BIT is not valid if VMA_ALLOCATOR_CREATE_BUFFER_DEVICE_ADDRESS_BIT was not used." ); |
| 17108 | return VK_ERROR_INITIALIZATION_FAILED; |
| 17109 | } |
| 17110 | |
| 17111 | VMA_DEBUG_LOG("vmaCreateBufferWithAlignment" ); |
| 17112 | |
| 17113 | VMA_DEBUG_GLOBAL_MUTEX_LOCK |
| 17114 | |
| 17115 | *pBuffer = VK_NULL_HANDLE; |
| 17116 | *pAllocation = VK_NULL_HANDLE; |
| 17117 | |
| 17118 | // 1. Create VkBuffer. |
| 17119 | VkResult res = (*allocator->GetVulkanFunctions().vkCreateBuffer)( |
| 17120 | allocator->m_hDevice, |
| 17121 | pBufferCreateInfo, |
| 17122 | allocator->GetAllocationCallbacks(), |
| 17123 | pBuffer); |
| 17124 | if(res >= 0) |
| 17125 | { |
| 17126 | // 2. vkGetBufferMemoryRequirements. |
| 17127 | VkMemoryRequirements vkMemReq = {}; |
| 17128 | bool requiresDedicatedAllocation = false; |
| 17129 | bool prefersDedicatedAllocation = false; |
| 17130 | allocator->GetBufferMemoryRequirements(hBuffer: *pBuffer, memReq&: vkMemReq, |
| 17131 | requiresDedicatedAllocation, prefersDedicatedAllocation); |
| 17132 | |
| 17133 | // 2a. Include minAlignment |
| 17134 | vkMemReq.alignment = VMA_MAX(vkMemReq.alignment, minAlignment); |
| 17135 | |
| 17136 | // 3. Allocate memory using allocator. |
| 17137 | res = allocator->AllocateMemory( |
| 17138 | vkMemReq, |
| 17139 | requiresDedicatedAllocation, |
| 17140 | prefersDedicatedAllocation, |
| 17141 | dedicatedBuffer: *pBuffer, // dedicatedBuffer |
| 17142 | VK_NULL_HANDLE, // dedicatedImage |
| 17143 | dedicatedBufferImageUsage: pBufferCreateInfo->usage, // dedicatedBufferImageUsage |
| 17144 | createInfo: *pAllocationCreateInfo, |
| 17145 | suballocType: VMA_SUBALLOCATION_TYPE_BUFFER, |
| 17146 | allocationCount: 1, // allocationCount |
| 17147 | pAllocations: pAllocation); |
| 17148 | |
| 17149 | if(res >= 0) |
| 17150 | { |
| 17151 | // 3. Bind buffer with memory. |
| 17152 | if((pAllocationCreateInfo->flags & VMA_ALLOCATION_CREATE_DONT_BIND_BIT) == 0) |
| 17153 | { |
| 17154 | res = allocator->BindBufferMemory(hAllocation: *pAllocation, allocationLocalOffset: 0, hBuffer: *pBuffer, VMA_NULL); |
| 17155 | } |
| 17156 | if(res >= 0) |
| 17157 | { |
| 17158 | // All steps succeeded. |
| 17159 | #if VMA_STATS_STRING_ENABLED |
| 17160 | (*pAllocation)->InitBufferImageUsage(bufferImageUsage: pBufferCreateInfo->usage); |
| 17161 | #endif |
| 17162 | if(pAllocationInfo != VMA_NULL) |
| 17163 | { |
| 17164 | allocator->GetAllocationInfo(hAllocation: *pAllocation, pAllocationInfo); |
| 17165 | } |
| 17166 | |
| 17167 | return VK_SUCCESS; |
| 17168 | } |
| 17169 | allocator->FreeMemory( |
| 17170 | allocationCount: 1, // allocationCount |
| 17171 | pAllocations: pAllocation); |
| 17172 | *pAllocation = VK_NULL_HANDLE; |
| 17173 | (*allocator->GetVulkanFunctions().vkDestroyBuffer)(allocator->m_hDevice, *pBuffer, allocator->GetAllocationCallbacks()); |
| 17174 | *pBuffer = VK_NULL_HANDLE; |
| 17175 | return res; |
| 17176 | } |
| 17177 | (*allocator->GetVulkanFunctions().vkDestroyBuffer)(allocator->m_hDevice, *pBuffer, allocator->GetAllocationCallbacks()); |
| 17178 | *pBuffer = VK_NULL_HANDLE; |
| 17179 | return res; |
| 17180 | } |
| 17181 | return res; |
| 17182 | } |
| 17183 | |
| 17184 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaCreateAliasingBuffer( |
| 17185 | VmaAllocator VMA_NOT_NULL allocator, |
| 17186 | VmaAllocation VMA_NOT_NULL allocation, |
| 17187 | const VkBufferCreateInfo* VMA_NOT_NULL pBufferCreateInfo, |
| 17188 | VkBuffer VMA_NULLABLE_NON_DISPATCHABLE* VMA_NOT_NULL pBuffer) |
| 17189 | { |
| 17190 | VMA_ASSERT(allocator && pBufferCreateInfo && pBuffer && allocation); |
| 17191 | |
| 17192 | VMA_DEBUG_LOG("vmaCreateAliasingBuffer" ); |
| 17193 | |
| 17194 | *pBuffer = VK_NULL_HANDLE; |
| 17195 | |
| 17196 | if (pBufferCreateInfo->size == 0) |
| 17197 | { |
| 17198 | return VK_ERROR_INITIALIZATION_FAILED; |
| 17199 | } |
| 17200 | if ((pBufferCreateInfo->usage & VK_BUFFER_USAGE_SHADER_DEVICE_ADDRESS_BIT_COPY) != 0 && |
| 17201 | !allocator->m_UseKhrBufferDeviceAddress) |
| 17202 | { |
| 17203 | VMA_ASSERT(0 && "Creating a buffer with VK_BUFFER_USAGE_SHADER_DEVICE_ADDRESS_BIT is not valid if VMA_ALLOCATOR_CREATE_BUFFER_DEVICE_ADDRESS_BIT was not used." ); |
| 17204 | return VK_ERROR_INITIALIZATION_FAILED; |
| 17205 | } |
| 17206 | |
| 17207 | VMA_DEBUG_GLOBAL_MUTEX_LOCK |
| 17208 | |
| 17209 | // 1. Create VkBuffer. |
| 17210 | VkResult res = (*allocator->GetVulkanFunctions().vkCreateBuffer)( |
| 17211 | allocator->m_hDevice, |
| 17212 | pBufferCreateInfo, |
| 17213 | allocator->GetAllocationCallbacks(), |
| 17214 | pBuffer); |
| 17215 | if (res >= 0) |
| 17216 | { |
| 17217 | // 2. Bind buffer with memory. |
| 17218 | res = allocator->BindBufferMemory(hAllocation: allocation, allocationLocalOffset: 0, hBuffer: *pBuffer, VMA_NULL); |
| 17219 | if (res >= 0) |
| 17220 | { |
| 17221 | return VK_SUCCESS; |
| 17222 | } |
| 17223 | (*allocator->GetVulkanFunctions().vkDestroyBuffer)(allocator->m_hDevice, *pBuffer, allocator->GetAllocationCallbacks()); |
| 17224 | } |
| 17225 | return res; |
| 17226 | } |
| 17227 | |
| 17228 | VMA_CALL_PRE void VMA_CALL_POST vmaDestroyBuffer( |
| 17229 | VmaAllocator allocator, |
| 17230 | VkBuffer buffer, |
| 17231 | VmaAllocation allocation) |
| 17232 | { |
| 17233 | VMA_ASSERT(allocator); |
| 17234 | |
| 17235 | if(buffer == VK_NULL_HANDLE && allocation == VK_NULL_HANDLE) |
| 17236 | { |
| 17237 | return; |
| 17238 | } |
| 17239 | |
| 17240 | VMA_DEBUG_LOG("vmaDestroyBuffer" ); |
| 17241 | |
| 17242 | VMA_DEBUG_GLOBAL_MUTEX_LOCK |
| 17243 | |
| 17244 | if(buffer != VK_NULL_HANDLE) |
| 17245 | { |
| 17246 | (*allocator->GetVulkanFunctions().vkDestroyBuffer)(allocator->m_hDevice, buffer, allocator->GetAllocationCallbacks()); |
| 17247 | } |
| 17248 | |
| 17249 | if(allocation != VK_NULL_HANDLE) |
| 17250 | { |
| 17251 | allocator->FreeMemory( |
| 17252 | allocationCount: 1, // allocationCount |
| 17253 | pAllocations: &allocation); |
| 17254 | } |
| 17255 | } |
| 17256 | |
| 17257 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaCreateImage( |
| 17258 | VmaAllocator allocator, |
| 17259 | const VkImageCreateInfo* pImageCreateInfo, |
| 17260 | const VmaAllocationCreateInfo* pAllocationCreateInfo, |
| 17261 | VkImage* pImage, |
| 17262 | VmaAllocation* pAllocation, |
| 17263 | VmaAllocationInfo* pAllocationInfo) |
| 17264 | { |
| 17265 | VMA_ASSERT(allocator && pImageCreateInfo && pAllocationCreateInfo && pImage && pAllocation); |
| 17266 | |
| 17267 | if(pImageCreateInfo->extent.width == 0 || |
| 17268 | pImageCreateInfo->extent.height == 0 || |
| 17269 | pImageCreateInfo->extent.depth == 0 || |
| 17270 | pImageCreateInfo->mipLevels == 0 || |
| 17271 | pImageCreateInfo->arrayLayers == 0) |
| 17272 | { |
| 17273 | return VK_ERROR_INITIALIZATION_FAILED; |
| 17274 | } |
| 17275 | |
| 17276 | VMA_DEBUG_LOG("vmaCreateImage" ); |
| 17277 | |
| 17278 | VMA_DEBUG_GLOBAL_MUTEX_LOCK |
| 17279 | |
| 17280 | *pImage = VK_NULL_HANDLE; |
| 17281 | *pAllocation = VK_NULL_HANDLE; |
| 17282 | |
| 17283 | // 1. Create VkImage. |
| 17284 | VkResult res = (*allocator->GetVulkanFunctions().vkCreateImage)( |
| 17285 | allocator->m_hDevice, |
| 17286 | pImageCreateInfo, |
| 17287 | allocator->GetAllocationCallbacks(), |
| 17288 | pImage); |
| 17289 | if(res >= 0) |
| 17290 | { |
| 17291 | VmaSuballocationType suballocType = pImageCreateInfo->tiling == VK_IMAGE_TILING_OPTIMAL ? |
| 17292 | VMA_SUBALLOCATION_TYPE_IMAGE_OPTIMAL : |
| 17293 | VMA_SUBALLOCATION_TYPE_IMAGE_LINEAR; |
| 17294 | |
| 17295 | // 2. Allocate memory using allocator. |
| 17296 | VkMemoryRequirements vkMemReq = {}; |
| 17297 | bool requiresDedicatedAllocation = false; |
| 17298 | bool prefersDedicatedAllocation = false; |
| 17299 | allocator->GetImageMemoryRequirements(hImage: *pImage, memReq&: vkMemReq, |
| 17300 | requiresDedicatedAllocation, prefersDedicatedAllocation); |
| 17301 | |
| 17302 | res = allocator->AllocateMemory( |
| 17303 | vkMemReq, |
| 17304 | requiresDedicatedAllocation, |
| 17305 | prefersDedicatedAllocation, |
| 17306 | VK_NULL_HANDLE, // dedicatedBuffer |
| 17307 | dedicatedImage: *pImage, // dedicatedImage |
| 17308 | dedicatedBufferImageUsage: pImageCreateInfo->usage, // dedicatedBufferImageUsage |
| 17309 | createInfo: *pAllocationCreateInfo, |
| 17310 | suballocType, |
| 17311 | allocationCount: 1, // allocationCount |
| 17312 | pAllocations: pAllocation); |
| 17313 | |
| 17314 | if(res >= 0) |
| 17315 | { |
| 17316 | // 3. Bind image with memory. |
| 17317 | if((pAllocationCreateInfo->flags & VMA_ALLOCATION_CREATE_DONT_BIND_BIT) == 0) |
| 17318 | { |
| 17319 | res = allocator->BindImageMemory(hAllocation: *pAllocation, allocationLocalOffset: 0, hImage: *pImage, VMA_NULL); |
| 17320 | } |
| 17321 | if(res >= 0) |
| 17322 | { |
| 17323 | // All steps succeeded. |
| 17324 | #if VMA_STATS_STRING_ENABLED |
| 17325 | (*pAllocation)->InitBufferImageUsage(bufferImageUsage: pImageCreateInfo->usage); |
| 17326 | #endif |
| 17327 | if(pAllocationInfo != VMA_NULL) |
| 17328 | { |
| 17329 | allocator->GetAllocationInfo(hAllocation: *pAllocation, pAllocationInfo); |
| 17330 | } |
| 17331 | |
| 17332 | return VK_SUCCESS; |
| 17333 | } |
| 17334 | allocator->FreeMemory( |
| 17335 | allocationCount: 1, // allocationCount |
| 17336 | pAllocations: pAllocation); |
| 17337 | *pAllocation = VK_NULL_HANDLE; |
| 17338 | (*allocator->GetVulkanFunctions().vkDestroyImage)(allocator->m_hDevice, *pImage, allocator->GetAllocationCallbacks()); |
| 17339 | *pImage = VK_NULL_HANDLE; |
| 17340 | return res; |
| 17341 | } |
| 17342 | (*allocator->GetVulkanFunctions().vkDestroyImage)(allocator->m_hDevice, *pImage, allocator->GetAllocationCallbacks()); |
| 17343 | *pImage = VK_NULL_HANDLE; |
| 17344 | return res; |
| 17345 | } |
| 17346 | return res; |
| 17347 | } |
| 17348 | |
| 17349 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaCreateAliasingImage( |
| 17350 | VmaAllocator VMA_NOT_NULL allocator, |
| 17351 | VmaAllocation VMA_NOT_NULL allocation, |
| 17352 | const VkImageCreateInfo* VMA_NOT_NULL pImageCreateInfo, |
| 17353 | VkImage VMA_NULLABLE_NON_DISPATCHABLE* VMA_NOT_NULL pImage) |
| 17354 | { |
| 17355 | VMA_ASSERT(allocator && pImageCreateInfo && pImage && allocation); |
| 17356 | |
| 17357 | *pImage = VK_NULL_HANDLE; |
| 17358 | |
| 17359 | VMA_DEBUG_LOG("vmaCreateImage" ); |
| 17360 | |
| 17361 | if (pImageCreateInfo->extent.width == 0 || |
| 17362 | pImageCreateInfo->extent.height == 0 || |
| 17363 | pImageCreateInfo->extent.depth == 0 || |
| 17364 | pImageCreateInfo->mipLevels == 0 || |
| 17365 | pImageCreateInfo->arrayLayers == 0) |
| 17366 | { |
| 17367 | return VK_ERROR_INITIALIZATION_FAILED; |
| 17368 | } |
| 17369 | |
| 17370 | VMA_DEBUG_GLOBAL_MUTEX_LOCK |
| 17371 | |
| 17372 | // 1. Create VkImage. |
| 17373 | VkResult res = (*allocator->GetVulkanFunctions().vkCreateImage)( |
| 17374 | allocator->m_hDevice, |
| 17375 | pImageCreateInfo, |
| 17376 | allocator->GetAllocationCallbacks(), |
| 17377 | pImage); |
| 17378 | if (res >= 0) |
| 17379 | { |
| 17380 | // 2. Bind image with memory. |
| 17381 | res = allocator->BindImageMemory(hAllocation: allocation, allocationLocalOffset: 0, hImage: *pImage, VMA_NULL); |
| 17382 | if (res >= 0) |
| 17383 | { |
| 17384 | return VK_SUCCESS; |
| 17385 | } |
| 17386 | (*allocator->GetVulkanFunctions().vkDestroyImage)(allocator->m_hDevice, *pImage, allocator->GetAllocationCallbacks()); |
| 17387 | } |
| 17388 | return res; |
| 17389 | } |
| 17390 | |
| 17391 | VMA_CALL_PRE void VMA_CALL_POST vmaDestroyImage( |
| 17392 | VmaAllocator VMA_NOT_NULL allocator, |
| 17393 | VkImage VMA_NULLABLE_NON_DISPATCHABLE image, |
| 17394 | VmaAllocation VMA_NULLABLE allocation) |
| 17395 | { |
| 17396 | VMA_ASSERT(allocator); |
| 17397 | |
| 17398 | if(image == VK_NULL_HANDLE && allocation == VK_NULL_HANDLE) |
| 17399 | { |
| 17400 | return; |
| 17401 | } |
| 17402 | |
| 17403 | VMA_DEBUG_LOG("vmaDestroyImage" ); |
| 17404 | |
| 17405 | VMA_DEBUG_GLOBAL_MUTEX_LOCK |
| 17406 | |
| 17407 | if(image != VK_NULL_HANDLE) |
| 17408 | { |
| 17409 | (*allocator->GetVulkanFunctions().vkDestroyImage)(allocator->m_hDevice, image, allocator->GetAllocationCallbacks()); |
| 17410 | } |
| 17411 | if(allocation != VK_NULL_HANDLE) |
| 17412 | { |
| 17413 | allocator->FreeMemory( |
| 17414 | allocationCount: 1, // allocationCount |
| 17415 | pAllocations: &allocation); |
| 17416 | } |
| 17417 | } |
| 17418 | |
| 17419 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaCreateVirtualBlock( |
| 17420 | const VmaVirtualBlockCreateInfo* VMA_NOT_NULL pCreateInfo, |
| 17421 | VmaVirtualBlock VMA_NULLABLE * VMA_NOT_NULL pVirtualBlock) |
| 17422 | { |
| 17423 | VMA_ASSERT(pCreateInfo && pVirtualBlock); |
| 17424 | VMA_ASSERT(pCreateInfo->size > 0); |
| 17425 | VMA_DEBUG_LOG("vmaCreateVirtualBlock" ); |
| 17426 | VMA_DEBUG_GLOBAL_MUTEX_LOCK; |
| 17427 | *pVirtualBlock = vma_new(pCreateInfo->pAllocationCallbacks, VmaVirtualBlock_T)(*pCreateInfo); |
| 17428 | VkResult res = (*pVirtualBlock)->Init(); |
| 17429 | if(res < 0) |
| 17430 | { |
| 17431 | vma_delete(pAllocationCallbacks: pCreateInfo->pAllocationCallbacks, ptr: *pVirtualBlock); |
| 17432 | *pVirtualBlock = VK_NULL_HANDLE; |
| 17433 | } |
| 17434 | return res; |
| 17435 | } |
| 17436 | |
| 17437 | VMA_CALL_PRE void VMA_CALL_POST vmaDestroyVirtualBlock(VmaVirtualBlock VMA_NULLABLE virtualBlock) |
| 17438 | { |
| 17439 | if(virtualBlock != VK_NULL_HANDLE) |
| 17440 | { |
| 17441 | VMA_DEBUG_LOG("vmaDestroyVirtualBlock" ); |
| 17442 | VMA_DEBUG_GLOBAL_MUTEX_LOCK; |
| 17443 | VkAllocationCallbacks allocationCallbacks = virtualBlock->m_AllocationCallbacks; // Have to copy the callbacks when destroying. |
| 17444 | vma_delete(pAllocationCallbacks: &allocationCallbacks, ptr: virtualBlock); |
| 17445 | } |
| 17446 | } |
| 17447 | |
| 17448 | VMA_CALL_PRE VkBool32 VMA_CALL_POST vmaIsVirtualBlockEmpty(VmaVirtualBlock VMA_NOT_NULL virtualBlock) |
| 17449 | { |
| 17450 | VMA_ASSERT(virtualBlock != VK_NULL_HANDLE); |
| 17451 | VMA_DEBUG_LOG("vmaIsVirtualBlockEmpty" ); |
| 17452 | VMA_DEBUG_GLOBAL_MUTEX_LOCK; |
| 17453 | return virtualBlock->IsEmpty() ? VK_TRUE : VK_FALSE; |
| 17454 | } |
| 17455 | |
| 17456 | VMA_CALL_PRE void VMA_CALL_POST vmaGetVirtualAllocationInfo(VmaVirtualBlock VMA_NOT_NULL virtualBlock, |
| 17457 | VmaVirtualAllocation VMA_NOT_NULL_NON_DISPATCHABLE allocation, VmaVirtualAllocationInfo* VMA_NOT_NULL pVirtualAllocInfo) |
| 17458 | { |
| 17459 | VMA_ASSERT(virtualBlock != VK_NULL_HANDLE && pVirtualAllocInfo != VMA_NULL); |
| 17460 | VMA_DEBUG_LOG("vmaGetVirtualAllocationInfo" ); |
| 17461 | VMA_DEBUG_GLOBAL_MUTEX_LOCK; |
| 17462 | virtualBlock->GetAllocationInfo(allocation, outInfo&: *pVirtualAllocInfo); |
| 17463 | } |
| 17464 | |
| 17465 | VMA_CALL_PRE VkResult VMA_CALL_POST vmaVirtualAllocate(VmaVirtualBlock VMA_NOT_NULL virtualBlock, |
| 17466 | const VmaVirtualAllocationCreateInfo* VMA_NOT_NULL pCreateInfo, VmaVirtualAllocation VMA_NULLABLE_NON_DISPATCHABLE* VMA_NOT_NULL pAllocation, |
| 17467 | VkDeviceSize* VMA_NULLABLE pOffset) |
| 17468 | { |
| 17469 | VMA_ASSERT(virtualBlock != VK_NULL_HANDLE && pCreateInfo != VMA_NULL && pAllocation != VMA_NULL); |
| 17470 | VMA_DEBUG_LOG("vmaVirtualAllocate" ); |
| 17471 | VMA_DEBUG_GLOBAL_MUTEX_LOCK; |
| 17472 | return virtualBlock->Allocate(createInfo: *pCreateInfo, outAllocation&: *pAllocation, outOffset: pOffset); |
| 17473 | } |
| 17474 | |
| 17475 | VMA_CALL_PRE void VMA_CALL_POST vmaVirtualFree(VmaVirtualBlock VMA_NOT_NULL virtualBlock, VmaVirtualAllocation VMA_NULLABLE_NON_DISPATCHABLE allocation) |
| 17476 | { |
| 17477 | if(allocation != VK_NULL_HANDLE) |
| 17478 | { |
| 17479 | VMA_ASSERT(virtualBlock != VK_NULL_HANDLE); |
| 17480 | VMA_DEBUG_LOG("vmaVirtualFree" ); |
| 17481 | VMA_DEBUG_GLOBAL_MUTEX_LOCK; |
| 17482 | virtualBlock->Free(allocation); |
| 17483 | } |
| 17484 | } |
| 17485 | |
| 17486 | VMA_CALL_PRE void VMA_CALL_POST vmaClearVirtualBlock(VmaVirtualBlock VMA_NOT_NULL virtualBlock) |
| 17487 | { |
| 17488 | VMA_ASSERT(virtualBlock != VK_NULL_HANDLE); |
| 17489 | VMA_DEBUG_LOG("vmaClearVirtualBlock" ); |
| 17490 | VMA_DEBUG_GLOBAL_MUTEX_LOCK; |
| 17491 | virtualBlock->Clear(); |
| 17492 | } |
| 17493 | |
| 17494 | VMA_CALL_PRE void VMA_CALL_POST vmaSetVirtualAllocationUserData(VmaVirtualBlock VMA_NOT_NULL virtualBlock, |
| 17495 | VmaVirtualAllocation VMA_NOT_NULL_NON_DISPATCHABLE allocation, void* VMA_NULLABLE pUserData) |
| 17496 | { |
| 17497 | VMA_ASSERT(virtualBlock != VK_NULL_HANDLE); |
| 17498 | VMA_DEBUG_LOG("vmaSetVirtualAllocationUserData" ); |
| 17499 | VMA_DEBUG_GLOBAL_MUTEX_LOCK; |
| 17500 | virtualBlock->SetAllocationUserData(allocation, userData: pUserData); |
| 17501 | } |
| 17502 | |
| 17503 | VMA_CALL_PRE void VMA_CALL_POST vmaGetVirtualBlockStatistics(VmaVirtualBlock VMA_NOT_NULL virtualBlock, |
| 17504 | VmaStatistics* VMA_NOT_NULL pStats) |
| 17505 | { |
| 17506 | VMA_ASSERT(virtualBlock != VK_NULL_HANDLE && pStats != VMA_NULL); |
| 17507 | VMA_DEBUG_LOG("vmaGetVirtualBlockStatistics" ); |
| 17508 | VMA_DEBUG_GLOBAL_MUTEX_LOCK; |
| 17509 | virtualBlock->GetStatistics(outStats&: *pStats); |
| 17510 | } |
| 17511 | |
| 17512 | VMA_CALL_PRE void VMA_CALL_POST vmaCalculateVirtualBlockStatistics(VmaVirtualBlock VMA_NOT_NULL virtualBlock, |
| 17513 | VmaDetailedStatistics* VMA_NOT_NULL pStats) |
| 17514 | { |
| 17515 | VMA_ASSERT(virtualBlock != VK_NULL_HANDLE && pStats != VMA_NULL); |
| 17516 | VMA_DEBUG_LOG("vmaCalculateVirtualBlockStatistics" ); |
| 17517 | VMA_DEBUG_GLOBAL_MUTEX_LOCK; |
| 17518 | virtualBlock->CalculateDetailedStatistics(outStats&: *pStats); |
| 17519 | } |
| 17520 | |
| 17521 | #if VMA_STATS_STRING_ENABLED |
| 17522 | |
| 17523 | VMA_CALL_PRE void VMA_CALL_POST vmaBuildVirtualBlockStatsString(VmaVirtualBlock VMA_NOT_NULL virtualBlock, |
| 17524 | char* VMA_NULLABLE * VMA_NOT_NULL ppStatsString, VkBool32 detailedMap) |
| 17525 | { |
| 17526 | VMA_ASSERT(virtualBlock != VK_NULL_HANDLE && ppStatsString != VMA_NULL); |
| 17527 | VMA_DEBUG_GLOBAL_MUTEX_LOCK; |
| 17528 | const VkAllocationCallbacks* allocationCallbacks = virtualBlock->GetAllocationCallbacks(); |
| 17529 | VmaStringBuilder sb(allocationCallbacks); |
| 17530 | virtualBlock->BuildStatsString(detailedMap: detailedMap != VK_FALSE, sb); |
| 17531 | *ppStatsString = VmaCreateStringCopy(allocs: allocationCallbacks, srcStr: sb.GetData(), strLen: sb.GetLength()); |
| 17532 | } |
| 17533 | |
| 17534 | VMA_CALL_PRE void VMA_CALL_POST vmaFreeVirtualBlockStatsString(VmaVirtualBlock VMA_NOT_NULL virtualBlock, |
| 17535 | char* VMA_NULLABLE pStatsString) |
| 17536 | { |
| 17537 | if(pStatsString != VMA_NULL) |
| 17538 | { |
| 17539 | VMA_ASSERT(virtualBlock != VK_NULL_HANDLE); |
| 17540 | VMA_DEBUG_GLOBAL_MUTEX_LOCK; |
| 17541 | VmaFreeString(allocs: virtualBlock->GetAllocationCallbacks(), str: pStatsString); |
| 17542 | } |
| 17543 | } |
| 17544 | #endif // VMA_STATS_STRING_ENABLED |
| 17545 | #endif // _VMA_PUBLIC_INTERFACE |
| 17546 | |
| 17547 | #if defined(__GNUC__) && !defined(__clang__) |
| 17548 | #pragma GCC diagnostic pop |
| 17549 | #elif defined(__clang__) |
| 17550 | #pragma clang diagnostic pop |
| 17551 | #endif |
| 17552 | |
| 17553 | #endif // VMA_IMPLEMENTATION |
| 17554 | |
| 17555 | /** |
| 17556 | \page quick_start Quick start |
| 17557 | |
| 17558 | \section quick_start_project_setup Project setup |
| 17559 | |
| 17560 | Vulkan Memory Allocator comes in form of a "stb-style" single header file. |
| 17561 | You don't need to build it as a separate library project. |
| 17562 | You can add this file directly to your project and submit it to code repository next to your other source files. |
| 17563 | |
| 17564 | "Single header" doesn't mean that everything is contained in C/C++ declarations, |
| 17565 | like it tends to be in case of inline functions or C++ templates. |
| 17566 | It means that implementation is bundled with interface in a single file and needs to be extracted using preprocessor macro. |
| 17567 | If you don't do it properly, you will get linker errors. |
| 17568 | |
| 17569 | To do it properly: |
| 17570 | |
| 17571 | -# Include "vk_mem_alloc.h" file in each CPP file where you want to use the library. |
| 17572 | This includes declarations of all members of the library. |
| 17573 | -# In exactly one CPP file define following macro before this include. |
| 17574 | It enables also internal definitions. |
| 17575 | |
| 17576 | \code |
| 17577 | #define VMA_IMPLEMENTATION |
| 17578 | #include "vk_mem_alloc.h" |
| 17579 | \endcode |
| 17580 | |
| 17581 | It may be a good idea to create dedicated CPP file just for this purpose. |
| 17582 | |
| 17583 | This library includes header `<vulkan/vulkan.h>`, which in turn |
| 17584 | includes `<windows.h>` on Windows. If you need some specific macros defined |
| 17585 | before including these headers (like `WIN32_LEAN_AND_MEAN` or |
| 17586 | `WINVER` for Windows, `VK_USE_PLATFORM_WIN32_KHR` for Vulkan), you must define |
| 17587 | them before every `#include` of this library. |
| 17588 | |
| 17589 | This library is written in C++, but has C-compatible interface. |
| 17590 | Thus you can include and use vk_mem_alloc.h in C or C++ code, but full |
| 17591 | implementation with `VMA_IMPLEMENTATION` macro must be compiled as C++, NOT as C. |
| 17592 | Some features of C++14 used. STL containers, RTTI, or C++ exceptions are not used. |
| 17593 | |
| 17594 | |
| 17595 | \section quick_start_initialization Initialization |
| 17596 | |
| 17597 | At program startup: |
| 17598 | |
| 17599 | -# Initialize Vulkan to have `VkPhysicalDevice`, `VkDevice` and `VkInstance` object. |
| 17600 | -# Fill VmaAllocatorCreateInfo structure and create #VmaAllocator object by |
| 17601 | calling vmaCreateAllocator(). |
| 17602 | |
| 17603 | Only members `physicalDevice`, `device`, `instance` are required. |
| 17604 | However, you should inform the library which Vulkan version do you use by setting |
| 17605 | VmaAllocatorCreateInfo::vulkanApiVersion and which extensions did you enable |
| 17606 | by setting VmaAllocatorCreateInfo::flags (like #VMA_ALLOCATOR_CREATE_BUFFER_DEVICE_ADDRESS_BIT for VK_KHR_buffer_device_address). |
| 17607 | Otherwise, VMA would use only features of Vulkan 1.0 core with no extensions. |
| 17608 | |
| 17609 | You may need to configure importing Vulkan functions. There are 3 ways to do this: |
| 17610 | |
| 17611 | -# **If you link with Vulkan static library** (e.g. "vulkan-1.lib" on Windows): |
| 17612 | - You don't need to do anything. |
| 17613 | - VMA will use these, as macro `VMA_STATIC_VULKAN_FUNCTIONS` is defined to 1 by default. |
| 17614 | -# **If you want VMA to fetch pointers to Vulkan functions dynamically** using `vkGetInstanceProcAddr`, |
| 17615 | `vkGetDeviceProcAddr` (this is the option presented in the example below): |
| 17616 | - Define `VMA_STATIC_VULKAN_FUNCTIONS` to 0, `VMA_DYNAMIC_VULKAN_FUNCTIONS` to 1. |
| 17617 | - Provide pointers to these two functions via VmaVulkanFunctions::vkGetInstanceProcAddr, |
| 17618 | VmaVulkanFunctions::vkGetDeviceProcAddr. |
| 17619 | - The library will fetch pointers to all other functions it needs internally. |
| 17620 | -# **If you fetch pointers to all Vulkan functions in a custom way**, e.g. using some loader like |
| 17621 | [Volk](https://github.com/zeux/volk): |
| 17622 | - Define `VMA_STATIC_VULKAN_FUNCTIONS` and `VMA_DYNAMIC_VULKAN_FUNCTIONS` to 0. |
| 17623 | - Pass these pointers via structure #VmaVulkanFunctions. |
| 17624 | |
| 17625 | \code |
| 17626 | VmaVulkanFunctions vulkanFunctions = {}; |
| 17627 | vulkanFunctions.vkGetInstanceProcAddr = &vkGetInstanceProcAddr; |
| 17628 | vulkanFunctions.vkGetDeviceProcAddr = &vkGetDeviceProcAddr; |
| 17629 | |
| 17630 | VmaAllocatorCreateInfo allocatorCreateInfo = {}; |
| 17631 | allocatorCreateInfo.vulkanApiVersion = VK_API_VERSION_1_2; |
| 17632 | allocatorCreateInfo.physicalDevice = physicalDevice; |
| 17633 | allocatorCreateInfo.device = device; |
| 17634 | allocatorCreateInfo.instance = instance; |
| 17635 | allocatorCreateInfo.pVulkanFunctions = &vulkanFunctions; |
| 17636 | |
| 17637 | VmaAllocator allocator; |
| 17638 | vmaCreateAllocator(&allocatorCreateInfo, &allocator); |
| 17639 | \endcode |
| 17640 | |
| 17641 | |
| 17642 | \section quick_start_resource_allocation Resource allocation |
| 17643 | |
| 17644 | When you want to create a buffer or image: |
| 17645 | |
| 17646 | -# Fill `VkBufferCreateInfo` / `VkImageCreateInfo` structure. |
| 17647 | -# Fill VmaAllocationCreateInfo structure. |
| 17648 | -# Call vmaCreateBuffer() / vmaCreateImage() to get `VkBuffer`/`VkImage` with memory |
| 17649 | already allocated and bound to it, plus #VmaAllocation objects that represents its underlying memory. |
| 17650 | |
| 17651 | \code |
| 17652 | VkBufferCreateInfo bufferInfo = { VK_STRUCTURE_TYPE_BUFFER_CREATE_INFO }; |
| 17653 | bufferInfo.size = 65536; |
| 17654 | bufferInfo.usage = VK_BUFFER_USAGE_VERTEX_BUFFER_BIT | VK_BUFFER_USAGE_TRANSFER_DST_BIT; |
| 17655 | |
| 17656 | VmaAllocationCreateInfo allocInfo = {}; |
| 17657 | allocInfo.usage = VMA_MEMORY_USAGE_AUTO; |
| 17658 | |
| 17659 | VkBuffer buffer; |
| 17660 | VmaAllocation allocation; |
| 17661 | vmaCreateBuffer(allocator, &bufferInfo, &allocInfo, &buffer, &allocation, nullptr); |
| 17662 | \endcode |
| 17663 | |
| 17664 | Don't forget to destroy your objects when no longer needed: |
| 17665 | |
| 17666 | \code |
| 17667 | vmaDestroyBuffer(allocator, buffer, allocation); |
| 17668 | vmaDestroyAllocator(allocator); |
| 17669 | \endcode |
| 17670 | |
| 17671 | |
| 17672 | \page choosing_memory_type Choosing memory type |
| 17673 | |
| 17674 | Physical devices in Vulkan support various combinations of memory heaps and |
| 17675 | types. Help with choosing correct and optimal memory type for your specific |
| 17676 | resource is one of the key features of this library. You can use it by filling |
| 17677 | appropriate members of VmaAllocationCreateInfo structure, as described below. |
| 17678 | You can also combine multiple methods. |
| 17679 | |
| 17680 | -# If you just want to find memory type index that meets your requirements, you |
| 17681 | can use function: vmaFindMemoryTypeIndexForBufferInfo(), |
| 17682 | vmaFindMemoryTypeIndexForImageInfo(), vmaFindMemoryTypeIndex(). |
| 17683 | -# If you want to allocate a region of device memory without association with any |
| 17684 | specific image or buffer, you can use function vmaAllocateMemory(). Usage of |
| 17685 | this function is not recommended and usually not needed. |
| 17686 | vmaAllocateMemoryPages() function is also provided for creating multiple allocations at once, |
| 17687 | which may be useful for sparse binding. |
| 17688 | -# If you already have a buffer or an image created, you want to allocate memory |
| 17689 | for it and then you will bind it yourself, you can use function |
| 17690 | vmaAllocateMemoryForBuffer(), vmaAllocateMemoryForImage(). |
| 17691 | For binding you should use functions: vmaBindBufferMemory(), vmaBindImageMemory() |
| 17692 | or their extended versions: vmaBindBufferMemory2(), vmaBindImageMemory2(). |
| 17693 | -# **This is the easiest and recommended way to use this library:** |
| 17694 | If you want to create a buffer or an image, allocate memory for it and bind |
| 17695 | them together, all in one call, you can use function vmaCreateBuffer(), |
| 17696 | vmaCreateImage(). |
| 17697 | |
| 17698 | When using 3. or 4., the library internally queries Vulkan for memory types |
| 17699 | supported for that buffer or image (function `vkGetBufferMemoryRequirements()`) |
| 17700 | and uses only one of these types. |
| 17701 | |
| 17702 | If no memory type can be found that meets all the requirements, these functions |
| 17703 | return `VK_ERROR_FEATURE_NOT_PRESENT`. |
| 17704 | |
| 17705 | You can leave VmaAllocationCreateInfo structure completely filled with zeros. |
| 17706 | It means no requirements are specified for memory type. |
| 17707 | It is valid, although not very useful. |
| 17708 | |
| 17709 | \section choosing_memory_type_usage Usage |
| 17710 | |
| 17711 | The easiest way to specify memory requirements is to fill member |
| 17712 | VmaAllocationCreateInfo::usage using one of the values of enum #VmaMemoryUsage. |
| 17713 | It defines high level, common usage types. |
| 17714 | Since version 3 of the library, it is recommended to use #VMA_MEMORY_USAGE_AUTO to let it select best memory type for your resource automatically. |
| 17715 | |
| 17716 | For example, if you want to create a uniform buffer that will be filled using |
| 17717 | transfer only once or infrequently and then used for rendering every frame as a uniform buffer, you can |
| 17718 | do it using following code. The buffer will most likely end up in a memory type with |
| 17719 | `VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT` to be fast to access by the GPU device. |
| 17720 | |
| 17721 | \code |
| 17722 | VkBufferCreateInfo bufferInfo = { VK_STRUCTURE_TYPE_BUFFER_CREATE_INFO }; |
| 17723 | bufferInfo.size = 65536; |
| 17724 | bufferInfo.usage = VK_BUFFER_USAGE_UNIFORM_BUFFER_BIT | VK_BUFFER_USAGE_TRANSFER_DST_BIT; |
| 17725 | |
| 17726 | VmaAllocationCreateInfo allocInfo = {}; |
| 17727 | allocInfo.usage = VMA_MEMORY_USAGE_AUTO; |
| 17728 | |
| 17729 | VkBuffer buffer; |
| 17730 | VmaAllocation allocation; |
| 17731 | vmaCreateBuffer(allocator, &bufferInfo, &allocInfo, &buffer, &allocation, nullptr); |
| 17732 | \endcode |
| 17733 | |
| 17734 | If you have a preference for putting the resource in GPU (device) memory or CPU (host) memory |
| 17735 | on systems with discrete graphics card that have the memories separate, you can use |
| 17736 | #VMA_MEMORY_USAGE_AUTO_PREFER_DEVICE or #VMA_MEMORY_USAGE_AUTO_PREFER_HOST. |
| 17737 | |
| 17738 | When using `VMA_MEMORY_USAGE_AUTO*` while you want to map the allocated memory, |
| 17739 | you also need to specify one of the host access flags: |
| 17740 | #VMA_ALLOCATION_CREATE_HOST_ACCESS_SEQUENTIAL_WRITE_BIT or #VMA_ALLOCATION_CREATE_HOST_ACCESS_RANDOM_BIT. |
| 17741 | This will help the library decide about preferred memory type to ensure it has `VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT` |
| 17742 | so you can map it. |
| 17743 | |
| 17744 | For example, a staging buffer that will be filled via mapped pointer and then |
| 17745 | used as a source of transfer to the buffer decribed previously can be created like this. |
| 17746 | It will likely and up in a memory type that is `HOST_VISIBLE` and `HOST_COHERENT` |
| 17747 | but not `HOST_CACHED` (meaning uncached, write-combined) and not `DEVICE_LOCAL` (meaning system RAM). |
| 17748 | |
| 17749 | \code |
| 17750 | VkBufferCreateInfo stagingBufferInfo = { VK_STRUCTURE_TYPE_BUFFER_CREATE_INFO }; |
| 17751 | stagingBufferInfo.size = 65536; |
| 17752 | stagingBufferInfo.usage = VK_BUFFER_USAGE_TRANSFER_SRC_BIT; |
| 17753 | |
| 17754 | VmaAllocationCreateInfo stagingAllocInfo = {}; |
| 17755 | stagingAllocInfo.usage = VMA_MEMORY_USAGE_AUTO; |
| 17756 | stagingAllocInfo.flags = VMA_ALLOCATION_CREATE_HOST_ACCESS_SEQUENTIAL_WRITE_BIT; |
| 17757 | |
| 17758 | VkBuffer stagingBuffer; |
| 17759 | VmaAllocation stagingAllocation; |
| 17760 | vmaCreateBuffer(allocator, &stagingBufferInfo, &stagingAllocInfo, &stagingBuffer, &stagingAllocation, nullptr); |
| 17761 | \endcode |
| 17762 | |
| 17763 | For more examples of creating different kinds of resources, see chapter \ref usage_patterns. |
| 17764 | |
| 17765 | Usage values `VMA_MEMORY_USAGE_AUTO*` are legal to use only when the library knows |
| 17766 | about the resource being created by having `VkBufferCreateInfo` / `VkImageCreateInfo` passed, |
| 17767 | so they work with functions like: vmaCreateBuffer(), vmaCreateImage(), vmaFindMemoryTypeIndexForBufferInfo() etc. |
| 17768 | If you allocate raw memory using function vmaAllocateMemory(), you have to use other means of selecting |
| 17769 | memory type, as decribed below. |
| 17770 | |
| 17771 | \note |
| 17772 | Old usage values (`VMA_MEMORY_USAGE_GPU_ONLY`, `VMA_MEMORY_USAGE_CPU_ONLY`, |
| 17773 | `VMA_MEMORY_USAGE_CPU_TO_GPU`, `VMA_MEMORY_USAGE_GPU_TO_CPU`, `VMA_MEMORY_USAGE_CPU_COPY`) |
| 17774 | are still available and work same way as in previous versions of the library |
| 17775 | for backward compatibility, but they are not recommended. |
| 17776 | |
| 17777 | \section choosing_memory_type_required_preferred_flags Required and preferred flags |
| 17778 | |
| 17779 | You can specify more detailed requirements by filling members |
| 17780 | VmaAllocationCreateInfo::requiredFlags and VmaAllocationCreateInfo::preferredFlags |
| 17781 | with a combination of bits from enum `VkMemoryPropertyFlags`. For example, |
| 17782 | if you want to create a buffer that will be persistently mapped on host (so it |
| 17783 | must be `HOST_VISIBLE`) and preferably will also be `HOST_COHERENT` and `HOST_CACHED`, |
| 17784 | use following code: |
| 17785 | |
| 17786 | \code |
| 17787 | VmaAllocationCreateInfo allocInfo = {}; |
| 17788 | allocInfo.requiredFlags = VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT; |
| 17789 | allocInfo.preferredFlags = VK_MEMORY_PROPERTY_HOST_COHERENT_BIT | VK_MEMORY_PROPERTY_HOST_CACHED_BIT; |
| 17790 | allocInfo.flags = VMA_ALLOCATION_CREATE_HOST_ACCESS_RANDOM_BIT | VMA_ALLOCATION_CREATE_MAPPED_BIT; |
| 17791 | |
| 17792 | VkBuffer buffer; |
| 17793 | VmaAllocation allocation; |
| 17794 | vmaCreateBuffer(allocator, &bufferInfo, &allocInfo, &buffer, &allocation, nullptr); |
| 17795 | \endcode |
| 17796 | |
| 17797 | A memory type is chosen that has all the required flags and as many preferred |
| 17798 | flags set as possible. |
| 17799 | |
| 17800 | Value passed in VmaAllocationCreateInfo::usage is internally converted to a set of required and preferred flags, |
| 17801 | plus some extra "magic" (heuristics). |
| 17802 | |
| 17803 | \section choosing_memory_type_explicit_memory_types Explicit memory types |
| 17804 | |
| 17805 | If you inspected memory types available on the physical device and you have |
| 17806 | a preference for memory types that you want to use, you can fill member |
| 17807 | VmaAllocationCreateInfo::memoryTypeBits. It is a bit mask, where each bit set |
| 17808 | means that a memory type with that index is allowed to be used for the |
| 17809 | allocation. Special value 0, just like `UINT32_MAX`, means there are no |
| 17810 | restrictions to memory type index. |
| 17811 | |
| 17812 | Please note that this member is NOT just a memory type index. |
| 17813 | Still you can use it to choose just one, specific memory type. |
| 17814 | For example, if you already determined that your buffer should be created in |
| 17815 | memory type 2, use following code: |
| 17816 | |
| 17817 | \code |
| 17818 | uint32_t memoryTypeIndex = 2; |
| 17819 | |
| 17820 | VmaAllocationCreateInfo allocInfo = {}; |
| 17821 | allocInfo.memoryTypeBits = 1u << memoryTypeIndex; |
| 17822 | |
| 17823 | VkBuffer buffer; |
| 17824 | VmaAllocation allocation; |
| 17825 | vmaCreateBuffer(allocator, &bufferInfo, &allocInfo, &buffer, &allocation, nullptr); |
| 17826 | \endcode |
| 17827 | |
| 17828 | |
| 17829 | \section choosing_memory_type_custom_memory_pools Custom memory pools |
| 17830 | |
| 17831 | If you allocate from custom memory pool, all the ways of specifying memory |
| 17832 | requirements described above are not applicable and the aforementioned members |
| 17833 | of VmaAllocationCreateInfo structure are ignored. Memory type is selected |
| 17834 | explicitly when creating the pool and then used to make all the allocations from |
| 17835 | that pool. For further details, see \ref custom_memory_pools. |
| 17836 | |
| 17837 | \section choosing_memory_type_dedicated_allocations Dedicated allocations |
| 17838 | |
| 17839 | Memory for allocations is reserved out of larger block of `VkDeviceMemory` |
| 17840 | allocated from Vulkan internally. That is the main feature of this whole library. |
| 17841 | You can still request a separate memory block to be created for an allocation, |
| 17842 | just like you would do in a trivial solution without using any allocator. |
| 17843 | In that case, a buffer or image is always bound to that memory at offset 0. |
| 17844 | This is called a "dedicated allocation". |
| 17845 | You can explicitly request it by using flag #VMA_ALLOCATION_CREATE_DEDICATED_MEMORY_BIT. |
| 17846 | The library can also internally decide to use dedicated allocation in some cases, e.g.: |
| 17847 | |
| 17848 | - When the size of the allocation is large. |
| 17849 | - When [VK_KHR_dedicated_allocation](@ref vk_khr_dedicated_allocation) extension is enabled |
| 17850 | and it reports that dedicated allocation is required or recommended for the resource. |
| 17851 | - When allocation of next big memory block fails due to not enough device memory, |
| 17852 | but allocation with the exact requested size succeeds. |
| 17853 | |
| 17854 | |
| 17855 | \page memory_mapping Memory mapping |
| 17856 | |
| 17857 | To "map memory" in Vulkan means to obtain a CPU pointer to `VkDeviceMemory`, |
| 17858 | to be able to read from it or write to it in CPU code. |
| 17859 | Mapping is possible only of memory allocated from a memory type that has |
| 17860 | `VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT` flag. |
| 17861 | Functions `vkMapMemory()`, `vkUnmapMemory()` are designed for this purpose. |
| 17862 | You can use them directly with memory allocated by this library, |
| 17863 | but it is not recommended because of following issue: |
| 17864 | Mapping the same `VkDeviceMemory` block multiple times is illegal - only one mapping at a time is allowed. |
| 17865 | This includes mapping disjoint regions. Mapping is not reference-counted internally by Vulkan. |
| 17866 | Because of this, Vulkan Memory Allocator provides following facilities: |
| 17867 | |
| 17868 | \note If you want to be able to map an allocation, you need to specify one of the flags |
| 17869 | #VMA_ALLOCATION_CREATE_HOST_ACCESS_SEQUENTIAL_WRITE_BIT or #VMA_ALLOCATION_CREATE_HOST_ACCESS_RANDOM_BIT |
| 17870 | in VmaAllocationCreateInfo::flags. These flags are required for an allocation to be mappable |
| 17871 | when using #VMA_MEMORY_USAGE_AUTO or other `VMA_MEMORY_USAGE_AUTO*` enum values. |
| 17872 | For other usage values they are ignored and every such allocation made in `HOST_VISIBLE` memory type is mappable, |
| 17873 | but they can still be used for consistency. |
| 17874 | |
| 17875 | \section memory_mapping_mapping_functions Mapping functions |
| 17876 | |
| 17877 | The library provides following functions for mapping of a specific #VmaAllocation: vmaMapMemory(), vmaUnmapMemory(). |
| 17878 | They are safer and more convenient to use than standard Vulkan functions. |
| 17879 | You can map an allocation multiple times simultaneously - mapping is reference-counted internally. |
| 17880 | You can also map different allocations simultaneously regardless of whether they use the same `VkDeviceMemory` block. |
| 17881 | The way it is implemented is that the library always maps entire memory block, not just region of the allocation. |
| 17882 | For further details, see description of vmaMapMemory() function. |
| 17883 | Example: |
| 17884 | |
| 17885 | \code |
| 17886 | // Having these objects initialized: |
| 17887 | struct ConstantBuffer |
| 17888 | { |
| 17889 | ... |
| 17890 | }; |
| 17891 | ConstantBuffer constantBufferData = ... |
| 17892 | |
| 17893 | VmaAllocator allocator = ... |
| 17894 | VkBuffer constantBuffer = ... |
| 17895 | VmaAllocation constantBufferAllocation = ... |
| 17896 | |
| 17897 | // You can map and fill your buffer using following code: |
| 17898 | |
| 17899 | void* mappedData; |
| 17900 | vmaMapMemory(allocator, constantBufferAllocation, &mappedData); |
| 17901 | memcpy(mappedData, &constantBufferData, sizeof(constantBufferData)); |
| 17902 | vmaUnmapMemory(allocator, constantBufferAllocation); |
| 17903 | \endcode |
| 17904 | |
| 17905 | When mapping, you may see a warning from Vulkan validation layer similar to this one: |
| 17906 | |
| 17907 | <i>Mapping an image with layout VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL can result in undefined behavior if this memory is used by the device. Only GENERAL or PREINITIALIZED should be used.</i> |
| 17908 | |
| 17909 | It happens because the library maps entire `VkDeviceMemory` block, where different |
| 17910 | types of images and buffers may end up together, especially on GPUs with unified memory like Intel. |
| 17911 | You can safely ignore it if you are sure you access only memory of the intended |
| 17912 | object that you wanted to map. |
| 17913 | |
| 17914 | |
| 17915 | \section memory_mapping_persistently_mapped_memory Persistently mapped memory |
| 17916 | |
| 17917 | Kepping your memory persistently mapped is generally OK in Vulkan. |
| 17918 | You don't need to unmap it before using its data on the GPU. |
| 17919 | The library provides a special feature designed for that: |
| 17920 | Allocations made with #VMA_ALLOCATION_CREATE_MAPPED_BIT flag set in |
| 17921 | VmaAllocationCreateInfo::flags stay mapped all the time, |
| 17922 | so you can just access CPU pointer to it any time |
| 17923 | without a need to call any "map" or "unmap" function. |
| 17924 | Example: |
| 17925 | |
| 17926 | \code |
| 17927 | VkBufferCreateInfo bufCreateInfo = { VK_STRUCTURE_TYPE_BUFFER_CREATE_INFO }; |
| 17928 | bufCreateInfo.size = sizeof(ConstantBuffer); |
| 17929 | bufCreateInfo.usage = VK_BUFFER_USAGE_TRANSFER_SRC_BIT; |
| 17930 | |
| 17931 | VmaAllocationCreateInfo allocCreateInfo = {}; |
| 17932 | allocCreateInfo.usage = VMA_MEMORY_USAGE_AUTO; |
| 17933 | allocCreateInfo.flags = VMA_ALLOCATION_CREATE_HOST_ACCESS_SEQUENTIAL_WRITE_BIT | |
| 17934 | VMA_ALLOCATION_CREATE_MAPPED_BIT; |
| 17935 | |
| 17936 | VkBuffer buf; |
| 17937 | VmaAllocation alloc; |
| 17938 | VmaAllocationInfo allocInfo; |
| 17939 | vmaCreateBuffer(allocator, &bufCreateInfo, &allocCreateInfo, &buf, &alloc, &allocInfo); |
| 17940 | |
| 17941 | // Buffer is already mapped. You can access its memory. |
| 17942 | memcpy(allocInfo.pMappedData, &constantBufferData, sizeof(constantBufferData)); |
| 17943 | \endcode |
| 17944 | |
| 17945 | \note #VMA_ALLOCATION_CREATE_MAPPED_BIT by itself doesn't guarantee that the allocation will end up |
| 17946 | in a mappable memory type. |
| 17947 | For this, you need to also specify #VMA_ALLOCATION_CREATE_HOST_ACCESS_SEQUENTIAL_WRITE_BIT or |
| 17948 | #VMA_ALLOCATION_CREATE_HOST_ACCESS_RANDOM_BIT. |
| 17949 | #VMA_ALLOCATION_CREATE_MAPPED_BIT only guarantees that if the memory is `HOST_VISIBLE`, the allocation will be mapped on creation. |
| 17950 | For an example of how to make use of this fact, see section \ref usage_patterns_advanced_data_uploading. |
| 17951 | |
| 17952 | \section memory_mapping_cache_control Cache flush and invalidate |
| 17953 | |
| 17954 | Memory in Vulkan doesn't need to be unmapped before using it on GPU, |
| 17955 | but unless a memory types has `VK_MEMORY_PROPERTY_HOST_COHERENT_BIT` flag set, |
| 17956 | you need to manually **invalidate** cache before reading of mapped pointer |
| 17957 | and **flush** cache after writing to mapped pointer. |
| 17958 | Map/unmap operations don't do that automatically. |
| 17959 | Vulkan provides following functions for this purpose `vkFlushMappedMemoryRanges()`, |
| 17960 | `vkInvalidateMappedMemoryRanges()`, but this library provides more convenient |
| 17961 | functions that refer to given allocation object: vmaFlushAllocation(), |
| 17962 | vmaInvalidateAllocation(), |
| 17963 | or multiple objects at once: vmaFlushAllocations(), vmaInvalidateAllocations(). |
| 17964 | |
| 17965 | Regions of memory specified for flush/invalidate must be aligned to |
| 17966 | `VkPhysicalDeviceLimits::nonCoherentAtomSize`. This is automatically ensured by the library. |
| 17967 | In any memory type that is `HOST_VISIBLE` but not `HOST_COHERENT`, all allocations |
| 17968 | within blocks are aligned to this value, so their offsets are always multiply of |
| 17969 | `nonCoherentAtomSize` and two different allocations never share same "line" of this size. |
| 17970 | |
| 17971 | Also, Windows drivers from all 3 PC GPU vendors (AMD, Intel, NVIDIA) |
| 17972 | currently provide `HOST_COHERENT` flag on all memory types that are |
| 17973 | `HOST_VISIBLE`, so on PC you may not need to bother. |
| 17974 | |
| 17975 | |
| 17976 | \page staying_within_budget Staying within budget |
| 17977 | |
| 17978 | When developing a graphics-intensive game or program, it is important to avoid allocating |
| 17979 | more GPU memory than it is physically available. When the memory is over-committed, |
| 17980 | various bad things can happen, depending on the specific GPU, graphics driver, and |
| 17981 | operating system: |
| 17982 | |
| 17983 | - It may just work without any problems. |
| 17984 | - The application may slow down because some memory blocks are moved to system RAM |
| 17985 | and the GPU has to access them through PCI Express bus. |
| 17986 | - A new allocation may take very long time to complete, even few seconds, and possibly |
| 17987 | freeze entire system. |
| 17988 | - The new allocation may fail with `VK_ERROR_OUT_OF_DEVICE_MEMORY`. |
| 17989 | - It may even result in GPU crash (TDR), observed as `VK_ERROR_DEVICE_LOST` |
| 17990 | returned somewhere later. |
| 17991 | |
| 17992 | \section staying_within_budget_querying_for_budget Querying for budget |
| 17993 | |
| 17994 | To query for current memory usage and available budget, use function vmaGetHeapBudgets(). |
| 17995 | Returned structure #VmaBudget contains quantities expressed in bytes, per Vulkan memory heap. |
| 17996 | |
| 17997 | Please note that this function returns different information and works faster than |
| 17998 | vmaCalculateStatistics(). vmaGetHeapBudgets() can be called every frame or even before every |
| 17999 | allocation, while vmaCalculateStatistics() is intended to be used rarely, |
| 18000 | only to obtain statistical information, e.g. for debugging purposes. |
| 18001 | |
| 18002 | It is recommended to use <b>VK_EXT_memory_budget</b> device extension to obtain information |
| 18003 | about the budget from Vulkan device. VMA is able to use this extension automatically. |
| 18004 | When not enabled, the allocator behaves same way, but then it estimates current usage |
| 18005 | and available budget based on its internal information and Vulkan memory heap sizes, |
| 18006 | which may be less precise. In order to use this extension: |
| 18007 | |
| 18008 | 1. Make sure extensions VK_EXT_memory_budget and VK_KHR_get_physical_device_properties2 |
| 18009 | required by it are available and enable them. Please note that the first is a device |
| 18010 | extension and the second is instance extension! |
| 18011 | 2. Use flag #VMA_ALLOCATOR_CREATE_EXT_MEMORY_BUDGET_BIT when creating #VmaAllocator object. |
| 18012 | 3. Make sure to call vmaSetCurrentFrameIndex() every frame. Budget is queried from |
| 18013 | Vulkan inside of it to avoid overhead of querying it with every allocation. |
| 18014 | |
| 18015 | \section staying_within_budget_controlling_memory_usage Controlling memory usage |
| 18016 | |
| 18017 | There are many ways in which you can try to stay within the budget. |
| 18018 | |
| 18019 | First, when making new allocation requires allocating a new memory block, the library |
| 18020 | tries not to exceed the budget automatically. If a block with default recommended size |
| 18021 | (e.g. 256 MB) would go over budget, a smaller block is allocated, possibly even |
| 18022 | dedicated memory for just this resource. |
| 18023 | |
| 18024 | If the size of the requested resource plus current memory usage is more than the |
| 18025 | budget, by default the library still tries to create it, leaving it to the Vulkan |
| 18026 | implementation whether the allocation succeeds or fails. You can change this behavior |
| 18027 | by using #VMA_ALLOCATION_CREATE_WITHIN_BUDGET_BIT flag. With it, the allocation is |
| 18028 | not made if it would exceed the budget or if the budget is already exceeded. |
| 18029 | VMA then tries to make the allocation from the next eligible Vulkan memory type. |
| 18030 | The all of them fail, the call then fails with `VK_ERROR_OUT_OF_DEVICE_MEMORY`. |
| 18031 | Example usage pattern may be to pass the #VMA_ALLOCATION_CREATE_WITHIN_BUDGET_BIT flag |
| 18032 | when creating resources that are not essential for the application (e.g. the texture |
| 18033 | of a specific object) and not to pass it when creating critically important resources |
| 18034 | (e.g. render targets). |
| 18035 | |
| 18036 | On AMD graphics cards there is a custom vendor extension available: <b>VK_AMD_memory_overallocation_behavior</b> |
| 18037 | that allows to control the behavior of the Vulkan implementation in out-of-memory cases - |
| 18038 | whether it should fail with an error code or still allow the allocation. |
| 18039 | Usage of this extension involves only passing extra structure on Vulkan device creation, |
| 18040 | so it is out of scope of this library. |
| 18041 | |
| 18042 | Finally, you can also use #VMA_ALLOCATION_CREATE_NEVER_ALLOCATE_BIT flag to make sure |
| 18043 | a new allocation is created only when it fits inside one of the existing memory blocks. |
| 18044 | If it would require to allocate a new block, if fails instead with `VK_ERROR_OUT_OF_DEVICE_MEMORY`. |
| 18045 | This also ensures that the function call is very fast because it never goes to Vulkan |
| 18046 | to obtain a new block. |
| 18047 | |
| 18048 | \note Creating \ref custom_memory_pools with VmaPoolCreateInfo::minBlockCount |
| 18049 | set to more than 0 will currently try to allocate memory blocks without checking whether they |
| 18050 | fit within budget. |
| 18051 | |
| 18052 | |
| 18053 | \page resource_aliasing Resource aliasing (overlap) |
| 18054 | |
| 18055 | New explicit graphics APIs (Vulkan and Direct3D 12), thanks to manual memory |
| 18056 | management, give an opportunity to alias (overlap) multiple resources in the |
| 18057 | same region of memory - a feature not available in the old APIs (Direct3D 11, OpenGL). |
| 18058 | It can be useful to save video memory, but it must be used with caution. |
| 18059 | |
| 18060 | For example, if you know the flow of your whole render frame in advance, you |
| 18061 | are going to use some intermediate textures or buffers only during a small range of render passes, |
| 18062 | and you know these ranges don't overlap in time, you can bind these resources to |
| 18063 | the same place in memory, even if they have completely different parameters (width, height, format etc.). |
| 18064 | |
| 18065 |  |
| 18066 | |
| 18067 | Such scenario is possible using VMA, but you need to create your images manually. |
| 18068 | Then you need to calculate parameters of an allocation to be made using formula: |
| 18069 | |
| 18070 | - allocation size = max(size of each image) |
| 18071 | - allocation alignment = max(alignment of each image) |
| 18072 | - allocation memoryTypeBits = bitwise AND(memoryTypeBits of each image) |
| 18073 | |
| 18074 | Following example shows two different images bound to the same place in memory, |
| 18075 | allocated to fit largest of them. |
| 18076 | |
| 18077 | \code |
| 18078 | // A 512x512 texture to be sampled. |
| 18079 | VkImageCreateInfo img1CreateInfo = { VK_STRUCTURE_TYPE_IMAGE_CREATE_INFO }; |
| 18080 | img1CreateInfo.imageType = VK_IMAGE_TYPE_2D; |
| 18081 | img1CreateInfo.extent.width = 512; |
| 18082 | img1CreateInfo.extent.height = 512; |
| 18083 | img1CreateInfo.extent.depth = 1; |
| 18084 | img1CreateInfo.mipLevels = 10; |
| 18085 | img1CreateInfo.arrayLayers = 1; |
| 18086 | img1CreateInfo.format = VK_FORMAT_R8G8B8A8_SRGB; |
| 18087 | img1CreateInfo.tiling = VK_IMAGE_TILING_OPTIMAL; |
| 18088 | img1CreateInfo.initialLayout = VK_IMAGE_LAYOUT_UNDEFINED; |
| 18089 | img1CreateInfo.usage = VK_IMAGE_USAGE_TRANSFER_DST_BIT | VK_IMAGE_USAGE_SAMPLED_BIT; |
| 18090 | img1CreateInfo.samples = VK_SAMPLE_COUNT_1_BIT; |
| 18091 | |
| 18092 | // A full screen texture to be used as color attachment. |
| 18093 | VkImageCreateInfo img2CreateInfo = { VK_STRUCTURE_TYPE_IMAGE_CREATE_INFO }; |
| 18094 | img2CreateInfo.imageType = VK_IMAGE_TYPE_2D; |
| 18095 | img2CreateInfo.extent.width = 1920; |
| 18096 | img2CreateInfo.extent.height = 1080; |
| 18097 | img2CreateInfo.extent.depth = 1; |
| 18098 | img2CreateInfo.mipLevels = 1; |
| 18099 | img2CreateInfo.arrayLayers = 1; |
| 18100 | img2CreateInfo.format = VK_FORMAT_R8G8B8A8_UNORM; |
| 18101 | img2CreateInfo.tiling = VK_IMAGE_TILING_OPTIMAL; |
| 18102 | img2CreateInfo.initialLayout = VK_IMAGE_LAYOUT_UNDEFINED; |
| 18103 | img2CreateInfo.usage = VK_IMAGE_USAGE_SAMPLED_BIT | VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT; |
| 18104 | img2CreateInfo.samples = VK_SAMPLE_COUNT_1_BIT; |
| 18105 | |
| 18106 | VkImage img1; |
| 18107 | res = vkCreateImage(device, &img1CreateInfo, nullptr, &img1); |
| 18108 | VkImage img2; |
| 18109 | res = vkCreateImage(device, &img2CreateInfo, nullptr, &img2); |
| 18110 | |
| 18111 | VkMemoryRequirements img1MemReq; |
| 18112 | vkGetImageMemoryRequirements(device, img1, &img1MemReq); |
| 18113 | VkMemoryRequirements img2MemReq; |
| 18114 | vkGetImageMemoryRequirements(device, img2, &img2MemReq); |
| 18115 | |
| 18116 | VkMemoryRequirements finalMemReq = {}; |
| 18117 | finalMemReq.size = std::max(img1MemReq.size, img2MemReq.size); |
| 18118 | finalMemReq.alignment = std::max(img1MemReq.alignment, img2MemReq.alignment); |
| 18119 | finalMemReq.memoryTypeBits = img1MemReq.memoryTypeBits & img2MemReq.memoryTypeBits; |
| 18120 | // Validate if(finalMemReq.memoryTypeBits != 0) |
| 18121 | |
| 18122 | VmaAllocationCreateInfo allocCreateInfo = {}; |
| 18123 | allocCreateInfo.preferredFlags = VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT; |
| 18124 | |
| 18125 | VmaAllocation alloc; |
| 18126 | res = vmaAllocateMemory(allocator, &finalMemReq, &allocCreateInfo, &alloc, nullptr); |
| 18127 | |
| 18128 | res = vmaBindImageMemory(allocator, alloc, img1); |
| 18129 | res = vmaBindImageMemory(allocator, alloc, img2); |
| 18130 | |
| 18131 | // You can use img1, img2 here, but not at the same time! |
| 18132 | |
| 18133 | vmaFreeMemory(allocator, alloc); |
| 18134 | vkDestroyImage(allocator, img2, nullptr); |
| 18135 | vkDestroyImage(allocator, img1, nullptr); |
| 18136 | \endcode |
| 18137 | |
| 18138 | Remember that using resources that alias in memory requires proper synchronization. |
| 18139 | You need to issue a memory barrier to make sure commands that use `img1` and `img2` |
| 18140 | don't overlap on GPU timeline. |
| 18141 | You also need to treat a resource after aliasing as uninitialized - containing garbage data. |
| 18142 | For example, if you use `img1` and then want to use `img2`, you need to issue |
| 18143 | an image memory barrier for `img2` with `oldLayout` = `VK_IMAGE_LAYOUT_UNDEFINED`. |
| 18144 | |
| 18145 | Additional considerations: |
| 18146 | |
| 18147 | - Vulkan also allows to interpret contents of memory between aliasing resources consistently in some cases. |
| 18148 | See chapter 11.8. "Memory Aliasing" of Vulkan specification or `VK_IMAGE_CREATE_ALIAS_BIT` flag. |
| 18149 | - You can create more complex layout where different images and buffers are bound |
| 18150 | at different offsets inside one large allocation. For example, one can imagine |
| 18151 | a big texture used in some render passes, aliasing with a set of many small buffers |
| 18152 | used between in some further passes. To bind a resource at non-zero offset in an allocation, |
| 18153 | use vmaBindBufferMemory2() / vmaBindImageMemory2(). |
| 18154 | - Before allocating memory for the resources you want to alias, check `memoryTypeBits` |
| 18155 | returned in memory requirements of each resource to make sure the bits overlap. |
| 18156 | Some GPUs may expose multiple memory types suitable e.g. only for buffers or |
| 18157 | images with `COLOR_ATTACHMENT` usage, so the sets of memory types supported by your |
| 18158 | resources may be disjoint. Aliasing them is not possible in that case. |
| 18159 | |
| 18160 | |
| 18161 | \page custom_memory_pools Custom memory pools |
| 18162 | |
| 18163 | A memory pool contains a number of `VkDeviceMemory` blocks. |
| 18164 | The library automatically creates and manages default pool for each memory type available on the device. |
| 18165 | Default memory pool automatically grows in size. |
| 18166 | Size of allocated blocks is also variable and managed automatically. |
| 18167 | |
| 18168 | You can create custom pool and allocate memory out of it. |
| 18169 | It can be useful if you want to: |
| 18170 | |
| 18171 | - Keep certain kind of allocations separate from others. |
| 18172 | - Enforce particular, fixed size of Vulkan memory blocks. |
| 18173 | - Limit maximum amount of Vulkan memory allocated for that pool. |
| 18174 | - Reserve minimum or fixed amount of Vulkan memory always preallocated for that pool. |
| 18175 | - Use extra parameters for a set of your allocations that are available in #VmaPoolCreateInfo but not in |
| 18176 | #VmaAllocationCreateInfo - e.g., custom minimum alignment, custom `pNext` chain. |
| 18177 | - Perform defragmentation on a specific subset of your allocations. |
| 18178 | |
| 18179 | To use custom memory pools: |
| 18180 | |
| 18181 | -# Fill VmaPoolCreateInfo structure. |
| 18182 | -# Call vmaCreatePool() to obtain #VmaPool handle. |
| 18183 | -# When making an allocation, set VmaAllocationCreateInfo::pool to this handle. |
| 18184 | You don't need to specify any other parameters of this structure, like `usage`. |
| 18185 | |
| 18186 | Example: |
| 18187 | |
| 18188 | \code |
| 18189 | // Find memoryTypeIndex for the pool. |
| 18190 | VkBufferCreateInfo sampleBufCreateInfo = { VK_STRUCTURE_TYPE_BUFFER_CREATE_INFO }; |
| 18191 | sampleBufCreateInfo.size = 0x10000; // Doesn't matter. |
| 18192 | sampleBufCreateInfo.usage = VK_BUFFER_USAGE_UNIFORM_BUFFER_BIT | VK_BUFFER_USAGE_TRANSFER_DST_BIT; |
| 18193 | |
| 18194 | VmaAllocationCreateInfo sampleAllocCreateInfo = {}; |
| 18195 | sampleAllocCreateInfo.usage = VMA_MEMORY_USAGE_AUTO; |
| 18196 | |
| 18197 | uint32_t memTypeIndex; |
| 18198 | VkResult res = vmaFindMemoryTypeIndexForBufferInfo(allocator, |
| 18199 | &sampleBufCreateInfo, &sampleAllocCreateInfo, &memTypeIndex); |
| 18200 | // Check res... |
| 18201 | |
| 18202 | // Create a pool that can have at most 2 blocks, 128 MiB each. |
| 18203 | VmaPoolCreateInfo poolCreateInfo = {}; |
| 18204 | poolCreateInfo.memoryTypeIndex = memTypeIndex; |
| 18205 | poolCreateInfo.blockSize = 128ull * 1024 * 1024; |
| 18206 | poolCreateInfo.maxBlockCount = 2; |
| 18207 | |
| 18208 | VmaPool pool; |
| 18209 | res = vmaCreatePool(allocator, &poolCreateInfo, &pool); |
| 18210 | // Check res... |
| 18211 | |
| 18212 | // Allocate a buffer out of it. |
| 18213 | VkBufferCreateInfo bufCreateInfo = { VK_STRUCTURE_TYPE_BUFFER_CREATE_INFO }; |
| 18214 | bufCreateInfo.size = 1024; |
| 18215 | bufCreateInfo.usage = VK_BUFFER_USAGE_UNIFORM_BUFFER_BIT | VK_BUFFER_USAGE_TRANSFER_DST_BIT; |
| 18216 | |
| 18217 | VmaAllocationCreateInfo allocCreateInfo = {}; |
| 18218 | allocCreateInfo.pool = pool; |
| 18219 | |
| 18220 | VkBuffer buf; |
| 18221 | VmaAllocation alloc; |
| 18222 | res = vmaCreateBuffer(allocator, &bufCreateInfo, &allocCreateInfo, &buf, &alloc, nullptr); |
| 18223 | // Check res... |
| 18224 | \endcode |
| 18225 | |
| 18226 | You have to free all allocations made from this pool before destroying it. |
| 18227 | |
| 18228 | \code |
| 18229 | vmaDestroyBuffer(allocator, buf, alloc); |
| 18230 | vmaDestroyPool(allocator, pool); |
| 18231 | \endcode |
| 18232 | |
| 18233 | New versions of this library support creating dedicated allocations in custom pools. |
| 18234 | It is supported only when VmaPoolCreateInfo::blockSize = 0. |
| 18235 | To use this feature, set VmaAllocationCreateInfo::pool to the pointer to your custom pool and |
| 18236 | VmaAllocationCreateInfo::flags to #VMA_ALLOCATION_CREATE_DEDICATED_MEMORY_BIT. |
| 18237 | |
| 18238 | \note Excessive use of custom pools is a common mistake when using this library. |
| 18239 | Custom pools may be useful for special purposes - when you want to |
| 18240 | keep certain type of resources separate e.g. to reserve minimum amount of memory |
| 18241 | for them or limit maximum amount of memory they can occupy. For most |
| 18242 | resources this is not needed and so it is not recommended to create #VmaPool |
| 18243 | objects and allocations out of them. Allocating from the default pool is sufficient. |
| 18244 | |
| 18245 | |
| 18246 | \section custom_memory_pools_MemTypeIndex Choosing memory type index |
| 18247 | |
| 18248 | When creating a pool, you must explicitly specify memory type index. |
| 18249 | To find the one suitable for your buffers or images, you can use helper functions |
| 18250 | vmaFindMemoryTypeIndexForBufferInfo(), vmaFindMemoryTypeIndexForImageInfo(). |
| 18251 | You need to provide structures with example parameters of buffers or images |
| 18252 | that you are going to create in that pool. |
| 18253 | |
| 18254 | \code |
| 18255 | VkBufferCreateInfo exampleBufCreateInfo = { VK_STRUCTURE_TYPE_BUFFER_CREATE_INFO }; |
| 18256 | exampleBufCreateInfo.size = 1024; // Doesn't matter |
| 18257 | exampleBufCreateInfo.usage = VK_BUFFER_USAGE_UNIFORM_BUFFER_BIT | VK_BUFFER_USAGE_TRANSFER_DST_BIT; |
| 18258 | |
| 18259 | VmaAllocationCreateInfo allocCreateInfo = {}; |
| 18260 | allocCreateInfo.usage = VMA_MEMORY_USAGE_AUTO; |
| 18261 | |
| 18262 | uint32_t memTypeIndex; |
| 18263 | vmaFindMemoryTypeIndexForBufferInfo(allocator, &exampleBufCreateInfo, &allocCreateInfo, &memTypeIndex); |
| 18264 | |
| 18265 | VmaPoolCreateInfo poolCreateInfo = {}; |
| 18266 | poolCreateInfo.memoryTypeIndex = memTypeIndex; |
| 18267 | // ... |
| 18268 | \endcode |
| 18269 | |
| 18270 | When creating buffers/images allocated in that pool, provide following parameters: |
| 18271 | |
| 18272 | - `VkBufferCreateInfo`: Prefer to pass same parameters as above. |
| 18273 | Otherwise you risk creating resources in a memory type that is not suitable for them, which may result in undefined behavior. |
| 18274 | Using different `VK_BUFFER_USAGE_` flags may work, but you shouldn't create images in a pool intended for buffers |
| 18275 | or the other way around. |
| 18276 | - VmaAllocationCreateInfo: You don't need to pass same parameters. Fill only `pool` member. |
| 18277 | Other members are ignored anyway. |
| 18278 | |
| 18279 | \section linear_algorithm Linear allocation algorithm |
| 18280 | |
| 18281 | Each Vulkan memory block managed by this library has accompanying metadata that |
| 18282 | keeps track of used and unused regions. By default, the metadata structure and |
| 18283 | algorithm tries to find best place for new allocations among free regions to |
| 18284 | optimize memory usage. This way you can allocate and free objects in any order. |
| 18285 | |
| 18286 |  |
| 18287 | |
| 18288 | Sometimes there is a need to use simpler, linear allocation algorithm. You can |
| 18289 | create custom pool that uses such algorithm by adding flag |
| 18290 | #VMA_POOL_CREATE_LINEAR_ALGORITHM_BIT to VmaPoolCreateInfo::flags while creating |
| 18291 | #VmaPool object. Then an alternative metadata management is used. It always |
| 18292 | creates new allocations after last one and doesn't reuse free regions after |
| 18293 | allocations freed in the middle. It results in better allocation performance and |
| 18294 | less memory consumed by metadata. |
| 18295 | |
| 18296 |  |
| 18297 | |
| 18298 | With this one flag, you can create a custom pool that can be used in many ways: |
| 18299 | free-at-once, stack, double stack, and ring buffer. See below for details. |
| 18300 | You don't need to specify explicitly which of these options you are going to use - it is detected automatically. |
| 18301 | |
| 18302 | \subsection linear_algorithm_free_at_once Free-at-once |
| 18303 | |
| 18304 | In a pool that uses linear algorithm, you still need to free all the allocations |
| 18305 | individually, e.g. by using vmaFreeMemory() or vmaDestroyBuffer(). You can free |
| 18306 | them in any order. New allocations are always made after last one - free space |
| 18307 | in the middle is not reused. However, when you release all the allocation and |
| 18308 | the pool becomes empty, allocation starts from the beginning again. This way you |
| 18309 | can use linear algorithm to speed up creation of allocations that you are going |
| 18310 | to release all at once. |
| 18311 | |
| 18312 |  |
| 18313 | |
| 18314 | This mode is also available for pools created with VmaPoolCreateInfo::maxBlockCount |
| 18315 | value that allows multiple memory blocks. |
| 18316 | |
| 18317 | \subsection linear_algorithm_stack Stack |
| 18318 | |
| 18319 | When you free an allocation that was created last, its space can be reused. |
| 18320 | Thanks to this, if you always release allocations in the order opposite to their |
| 18321 | creation (LIFO - Last In First Out), you can achieve behavior of a stack. |
| 18322 | |
| 18323 |  |
| 18324 | |
| 18325 | This mode is also available for pools created with VmaPoolCreateInfo::maxBlockCount |
| 18326 | value that allows multiple memory blocks. |
| 18327 | |
| 18328 | \subsection linear_algorithm_double_stack Double stack |
| 18329 | |
| 18330 | The space reserved by a custom pool with linear algorithm may be used by two |
| 18331 | stacks: |
| 18332 | |
| 18333 | - First, default one, growing up from offset 0. |
| 18334 | - Second, "upper" one, growing down from the end towards lower offsets. |
| 18335 | |
| 18336 | To make allocation from the upper stack, add flag #VMA_ALLOCATION_CREATE_UPPER_ADDRESS_BIT |
| 18337 | to VmaAllocationCreateInfo::flags. |
| 18338 | |
| 18339 |  |
| 18340 | |
| 18341 | Double stack is available only in pools with one memory block - |
| 18342 | VmaPoolCreateInfo::maxBlockCount must be 1. Otherwise behavior is undefined. |
| 18343 | |
| 18344 | When the two stacks' ends meet so there is not enough space between them for a |
| 18345 | new allocation, such allocation fails with usual |
| 18346 | `VK_ERROR_OUT_OF_DEVICE_MEMORY` error. |
| 18347 | |
| 18348 | \subsection linear_algorithm_ring_buffer Ring buffer |
| 18349 | |
| 18350 | When you free some allocations from the beginning and there is not enough free space |
| 18351 | for a new one at the end of a pool, allocator's "cursor" wraps around to the |
| 18352 | beginning and starts allocation there. Thanks to this, if you always release |
| 18353 | allocations in the same order as you created them (FIFO - First In First Out), |
| 18354 | you can achieve behavior of a ring buffer / queue. |
| 18355 | |
| 18356 |  |
| 18357 | |
| 18358 | Ring buffer is available only in pools with one memory block - |
| 18359 | VmaPoolCreateInfo::maxBlockCount must be 1. Otherwise behavior is undefined. |
| 18360 | |
| 18361 | \note \ref defragmentation is not supported in custom pools created with #VMA_POOL_CREATE_LINEAR_ALGORITHM_BIT. |
| 18362 | |
| 18363 | |
| 18364 | \page defragmentation Defragmentation |
| 18365 | |
| 18366 | Interleaved allocations and deallocations of many objects of varying size can |
| 18367 | cause fragmentation over time, which can lead to a situation where the library is unable |
| 18368 | to find a continuous range of free memory for a new allocation despite there is |
| 18369 | enough free space, just scattered across many small free ranges between existing |
| 18370 | allocations. |
| 18371 | |
| 18372 | To mitigate this problem, you can use defragmentation feature. |
| 18373 | It doesn't happen automatically though and needs your cooperation, |
| 18374 | because VMA is a low level library that only allocates memory. |
| 18375 | It cannot recreate buffers and images in a new place as it doesn't remember the contents of `VkBufferCreateInfo` / `VkImageCreateInfo` structures. |
| 18376 | It cannot copy their contents as it doesn't record any commands to a command buffer. |
| 18377 | |
| 18378 | Example: |
| 18379 | |
| 18380 | \code |
| 18381 | VmaDefragmentationInfo defragInfo = {}; |
| 18382 | defragInfo.pool = myPool; |
| 18383 | defragInfo.flags = VMA_DEFRAGMENTATION_FLAG_ALGORITHM_FAST_BIT; |
| 18384 | |
| 18385 | VmaDefragmentationContext defragCtx; |
| 18386 | VkResult res = vmaBeginDefragmentation(allocator, &defragInfo, &defragCtx); |
| 18387 | // Check res... |
| 18388 | |
| 18389 | for(;;) |
| 18390 | { |
| 18391 | VmaDefragmentationPassMoveInfo pass; |
| 18392 | res = vmaBeginDefragmentationPass(allocator, defragCtx, &pass); |
| 18393 | if(res == VK_SUCCESS) |
| 18394 | break; |
| 18395 | else if(res != VK_INCOMPLETE) |
| 18396 | // Handle error... |
| 18397 | |
| 18398 | for(uint32_t i = 0; i < pass.moveCount; ++i) |
| 18399 | { |
| 18400 | // Inspect pass.pMoves[i].srcAllocation, identify what buffer/image it represents. |
| 18401 | VmaAllocationInfo allocInfo; |
| 18402 | vmaGetAllocationInfo(allocator, pMoves[i].srcAllocation, &allocInfo); |
| 18403 | MyEngineResourceData* resData = (MyEngineResourceData*)allocInfo.pUserData; |
| 18404 | |
| 18405 | // Recreate and bind this buffer/image at: pass.pMoves[i].dstMemory, pass.pMoves[i].dstOffset. |
| 18406 | VkImageCreateInfo imgCreateInfo = ... |
| 18407 | VkImage newImg; |
| 18408 | res = vkCreateImage(device, &imgCreateInfo, nullptr, &newImg); |
| 18409 | // Check res... |
| 18410 | res = vmaBindImageMemory(allocator, pMoves[i].dstTmpAllocation, newImg); |
| 18411 | // Check res... |
| 18412 | |
| 18413 | // Issue a vkCmdCopyBuffer/vkCmdCopyImage to copy its content to the new place. |
| 18414 | vkCmdCopyImage(cmdBuf, resData->img, ..., newImg, ...); |
| 18415 | } |
| 18416 | |
| 18417 | // Make sure the copy commands finished executing. |
| 18418 | vkWaitForFences(...); |
| 18419 | |
| 18420 | // Destroy old buffers/images bound with pass.pMoves[i].srcAllocation. |
| 18421 | for(uint32_t i = 0; i < pass.moveCount; ++i) |
| 18422 | { |
| 18423 | // ... |
| 18424 | vkDestroyImage(device, resData->img, nullptr); |
| 18425 | } |
| 18426 | |
| 18427 | // Update appropriate descriptors to point to the new places... |
| 18428 | |
| 18429 | res = vmaEndDefragmentationPass(allocator, defragCtx, &pass); |
| 18430 | if(res == VK_SUCCESS) |
| 18431 | break; |
| 18432 | else if(res != VK_INCOMPLETE) |
| 18433 | // Handle error... |
| 18434 | } |
| 18435 | |
| 18436 | vmaEndDefragmentation(allocator, defragCtx, nullptr); |
| 18437 | \endcode |
| 18438 | |
| 18439 | Although functions like vmaCreateBuffer(), vmaCreateImage(), vmaDestroyBuffer(), vmaDestroyImage() |
| 18440 | create/destroy an allocation and a buffer/image at once, these are just a shortcut for |
| 18441 | creating the resource, allocating memory, and binding them together. |
| 18442 | Defragmentation works on memory allocations only. You must handle the rest manually. |
| 18443 | Defragmentation is an iterative process that should repreat "passes" as long as related functions |
| 18444 | return `VK_INCOMPLETE` not `VK_SUCCESS`. |
| 18445 | In each pass: |
| 18446 | |
| 18447 | 1. vmaBeginDefragmentationPass() function call: |
| 18448 | - Calculates and returns the list of allocations to be moved in this pass. |
| 18449 | Note this can be a time-consuming process. |
| 18450 | - Reserves destination memory for them by creating temporary destination allocations |
| 18451 | that you can query for their `VkDeviceMemory` + offset using vmaGetAllocationInfo(). |
| 18452 | 2. Inside the pass, **you should**: |
| 18453 | - Inspect the returned list of allocations to be moved. |
| 18454 | - Create new buffers/images and bind them at the returned destination temporary allocations. |
| 18455 | - Copy data from source to destination resources if necessary. |
| 18456 | - Destroy the source buffers/images, but NOT their allocations. |
| 18457 | 3. vmaEndDefragmentationPass() function call: |
| 18458 | - Frees the source memory reserved for the allocations that are moved. |
| 18459 | - Modifies source #VmaAllocation objects that are moved to point to the destination reserved memory. |
| 18460 | - Frees `VkDeviceMemory` blocks that became empty. |
| 18461 | |
| 18462 | Unlike in previous iterations of the defragmentation API, there is no list of "movable" allocations passed as a parameter. |
| 18463 | Defragmentation algorithm tries to move all suitable allocations. |
| 18464 | You can, however, refuse to move some of them inside a defragmentation pass, by setting |
| 18465 | `pass.pMoves[i].operation` to #VMA_DEFRAGMENTATION_MOVE_OPERATION_IGNORE. |
| 18466 | This is not recommended and may result in suboptimal packing of the allocations after defragmentation. |
| 18467 | If you cannot ensure any allocation can be moved, it is better to keep movable allocations separate in a custom pool. |
| 18468 | |
| 18469 | Inside a pass, for each allocation that should be moved: |
| 18470 | |
| 18471 | - You should copy its data from the source to the destination place by calling e.g. `vkCmdCopyBuffer()`, `vkCmdCopyImage()`. |
| 18472 | - You need to make sure these commands finished executing before destroying the source buffers/images and before calling vmaEndDefragmentationPass(). |
| 18473 | - If a resource doesn't contain any meaningful data, e.g. it is a transient color attachment image to be cleared, |
| 18474 | filled, and used temporarily in each rendering frame, you can just recreate this image |
| 18475 | without copying its data. |
| 18476 | - If the resource is in `HOST_VISIBLE` and `HOST_CACHED` memory, you can copy its data on the CPU |
| 18477 | using `memcpy()`. |
| 18478 | - If you cannot move the allocation, you can set `pass.pMoves[i].operation` to #VMA_DEFRAGMENTATION_MOVE_OPERATION_IGNORE. |
| 18479 | This will cancel the move. |
| 18480 | - vmaEndDefragmentationPass() will then free the destination memory |
| 18481 | not the source memory of the allocation, leaving it unchanged. |
| 18482 | - If you decide the allocation is unimportant and can be destroyed instead of moved (e.g. it wasn't used for long time), |
| 18483 | you can set `pass.pMoves[i].operation` to #VMA_DEFRAGMENTATION_MOVE_OPERATION_DESTROY. |
| 18484 | - vmaEndDefragmentationPass() will then free both source and destination memory, and will destroy the source #VmaAllocation object. |
| 18485 | |
| 18486 | You can defragment a specific custom pool by setting VmaDefragmentationInfo::pool |
| 18487 | (like in the example above) or all the default pools by setting this member to null. |
| 18488 | |
| 18489 | Defragmentation is always performed in each pool separately. |
| 18490 | Allocations are never moved between different Vulkan memory types. |
| 18491 | The size of the destination memory reserved for a moved allocation is the same as the original one. |
| 18492 | Alignment of an allocation as it was determined using `vkGetBufferMemoryRequirements()` etc. is also respected after defragmentation. |
| 18493 | Buffers/images should be recreated with the same `VkBufferCreateInfo` / `VkImageCreateInfo` parameters as the original ones. |
| 18494 | |
| 18495 | You can perform the defragmentation incrementally to limit the number of allocations and bytes to be moved |
| 18496 | in each pass, e.g. to call it in sync with render frames and not to experience too big hitches. |
| 18497 | See members: VmaDefragmentationInfo::maxBytesPerPass, VmaDefragmentationInfo::maxAllocationsPerPass. |
| 18498 | |
| 18499 | It is also safe to perform the defragmentation asynchronously to render frames and other Vulkan and VMA |
| 18500 | usage, possibly from multiple threads, with the exception that allocations |
| 18501 | returned in VmaDefragmentationPassMoveInfo::pMoves shouldn't be destroyed until the defragmentation pass is ended. |
| 18502 | |
| 18503 | <b>Mapping</b> is preserved on allocations that are moved during defragmentation. |
| 18504 | Whether through #VMA_ALLOCATION_CREATE_MAPPED_BIT or vmaMapMemory(), the allocations |
| 18505 | are mapped at their new place. Of course, pointer to the mapped data changes, so it needs to be queried |
| 18506 | using VmaAllocationInfo::pMappedData. |
| 18507 | |
| 18508 | \note Defragmentation is not supported in custom pools created with #VMA_POOL_CREATE_LINEAR_ALGORITHM_BIT. |
| 18509 | |
| 18510 | |
| 18511 | \page statistics Statistics |
| 18512 | |
| 18513 | This library contains several functions that return information about its internal state, |
| 18514 | especially the amount of memory allocated from Vulkan. |
| 18515 | |
| 18516 | \section statistics_numeric_statistics Numeric statistics |
| 18517 | |
| 18518 | If you need to obtain basic statistics about memory usage per heap, together with current budget, |
| 18519 | you can call function vmaGetHeapBudgets() and inspect structure #VmaBudget. |
| 18520 | This is useful to keep track of memory usage and stay withing budget |
| 18521 | (see also \ref staying_within_budget). |
| 18522 | Example: |
| 18523 | |
| 18524 | \code |
| 18525 | uint32_t heapIndex = ... |
| 18526 | |
| 18527 | VmaBudget budgets[VK_MAX_MEMORY_HEAPS]; |
| 18528 | vmaGetHeapBudgets(allocator, budgets); |
| 18529 | |
| 18530 | printf("My heap currently has %u allocations taking %llu B,\n", |
| 18531 | budgets[heapIndex].statistics.allocationCount, |
| 18532 | budgets[heapIndex].statistics.allocationBytes); |
| 18533 | printf("allocated out of %u Vulkan device memory blocks taking %llu B,\n", |
| 18534 | budgets[heapIndex].statistics.blockCount, |
| 18535 | budgets[heapIndex].statistics.blockBytes); |
| 18536 | printf("Vulkan reports total usage %llu B with budget %llu B.\n", |
| 18537 | budgets[heapIndex].usage, |
| 18538 | budgets[heapIndex].budget); |
| 18539 | \endcode |
| 18540 | |
| 18541 | You can query for more detailed statistics per memory heap, type, and totals, |
| 18542 | including minimum and maximum allocation size and unused range size, |
| 18543 | by calling function vmaCalculateStatistics() and inspecting structure #VmaTotalStatistics. |
| 18544 | This function is slower though, as it has to traverse all the internal data structures, |
| 18545 | so it should be used only for debugging purposes. |
| 18546 | |
| 18547 | You can query for statistics of a custom pool using function vmaGetPoolStatistics() |
| 18548 | or vmaCalculatePoolStatistics(). |
| 18549 | |
| 18550 | You can query for information about a specific allocation using function vmaGetAllocationInfo(). |
| 18551 | It fill structure #VmaAllocationInfo. |
| 18552 | |
| 18553 | \section statistics_json_dump JSON dump |
| 18554 | |
| 18555 | You can dump internal state of the allocator to a string in JSON format using function vmaBuildStatsString(). |
| 18556 | The result is guaranteed to be correct JSON. |
| 18557 | It uses ANSI encoding. |
| 18558 | Any strings provided by user (see [Allocation names](@ref allocation_names)) |
| 18559 | are copied as-is and properly escaped for JSON, so if they use UTF-8, ISO-8859-2 or any other encoding, |
| 18560 | this JSON string can be treated as using this encoding. |
| 18561 | It must be freed using function vmaFreeStatsString(). |
| 18562 | |
| 18563 | The format of this JSON string is not part of official documentation of the library, |
| 18564 | but it will not change in backward-incompatible way without increasing library major version number |
| 18565 | and appropriate mention in changelog. |
| 18566 | |
| 18567 | The JSON string contains all the data that can be obtained using vmaCalculateStatistics(). |
| 18568 | It can also contain detailed map of allocated memory blocks and their regions - |
| 18569 | free and occupied by allocations. |
| 18570 | This allows e.g. to visualize the memory or assess fragmentation. |
| 18571 | |
| 18572 | |
| 18573 | \page allocation_annotation Allocation names and user data |
| 18574 | |
| 18575 | \section allocation_user_data Allocation user data |
| 18576 | |
| 18577 | You can annotate allocations with your own information, e.g. for debugging purposes. |
| 18578 | To do that, fill VmaAllocationCreateInfo::pUserData field when creating |
| 18579 | an allocation. It is an opaque `void*` pointer. You can use it e.g. as a pointer, |
| 18580 | some handle, index, key, ordinal number or any other value that would associate |
| 18581 | the allocation with your custom metadata. |
| 18582 | It it useful to identify appropriate data structures in your engine given #VmaAllocation, |
| 18583 | e.g. when doing \ref defragmentation. |
| 18584 | |
| 18585 | \code |
| 18586 | VkBufferCreateInfo bufCreateInfo = ... |
| 18587 | |
| 18588 | MyBufferMetadata* pMetadata = CreateBufferMetadata(); |
| 18589 | |
| 18590 | VmaAllocationCreateInfo allocCreateInfo = {}; |
| 18591 | allocCreateInfo.usage = VMA_MEMORY_USAGE_AUTO; |
| 18592 | allocCreateInfo.pUserData = pMetadata; |
| 18593 | |
| 18594 | VkBuffer buffer; |
| 18595 | VmaAllocation allocation; |
| 18596 | vmaCreateBuffer(allocator, &bufCreateInfo, &allocCreateInfo, &buffer, &allocation, nullptr); |
| 18597 | \endcode |
| 18598 | |
| 18599 | The pointer may be later retrieved as VmaAllocationInfo::pUserData: |
| 18600 | |
| 18601 | \code |
| 18602 | VmaAllocationInfo allocInfo; |
| 18603 | vmaGetAllocationInfo(allocator, allocation, &allocInfo); |
| 18604 | MyBufferMetadata* pMetadata = (MyBufferMetadata*)allocInfo.pUserData; |
| 18605 | \endcode |
| 18606 | |
| 18607 | It can also be changed using function vmaSetAllocationUserData(). |
| 18608 | |
| 18609 | Values of (non-zero) allocations' `pUserData` are printed in JSON report created by |
| 18610 | vmaBuildStatsString() in hexadecimal form. |
| 18611 | |
| 18612 | \section allocation_names Allocation names |
| 18613 | |
| 18614 | An allocation can also carry a null-terminated string, giving a name to the allocation. |
| 18615 | To set it, call vmaSetAllocationName(). |
| 18616 | The library creates internal copy of the string, so the pointer you pass doesn't need |
| 18617 | to be valid for whole lifetime of the allocation. You can free it after the call. |
| 18618 | |
| 18619 | \code |
| 18620 | std::string imageName = "Texture: "; |
| 18621 | imageName += fileName; |
| 18622 | vmaSetAllocationName(allocator, allocation, imageName.c_str()); |
| 18623 | \endcode |
| 18624 | |
| 18625 | The string can be later retrieved by inspecting VmaAllocationInfo::pName. |
| 18626 | It is also printed in JSON report created by vmaBuildStatsString(). |
| 18627 | |
| 18628 | \note Setting string name to VMA allocation doesn't automatically set it to the Vulkan buffer or image created with it. |
| 18629 | You must do it manually using an extension like VK_EXT_debug_utils, which is independent of this library. |
| 18630 | |
| 18631 | |
| 18632 | \page virtual_allocator Virtual allocator |
| 18633 | |
| 18634 | As an extra feature, the core allocation algorithm of the library is exposed through a simple and convenient API of "virtual allocator". |
| 18635 | It doesn't allocate any real GPU memory. It just keeps track of used and free regions of a "virtual block". |
| 18636 | You can use it to allocate your own memory or other objects, even completely unrelated to Vulkan. |
| 18637 | A common use case is sub-allocation of pieces of one large GPU buffer. |
| 18638 | |
| 18639 | \section virtual_allocator_creating_virtual_block Creating virtual block |
| 18640 | |
| 18641 | To use this functionality, there is no main "allocator" object. |
| 18642 | You don't need to have #VmaAllocator object created. |
| 18643 | All you need to do is to create a separate #VmaVirtualBlock object for each block of memory you want to be managed by the allocator: |
| 18644 | |
| 18645 | -# Fill in #VmaVirtualBlockCreateInfo structure. |
| 18646 | -# Call vmaCreateVirtualBlock(). Get new #VmaVirtualBlock object. |
| 18647 | |
| 18648 | Example: |
| 18649 | |
| 18650 | \code |
| 18651 | VmaVirtualBlockCreateInfo blockCreateInfo = {}; |
| 18652 | blockCreateInfo.size = 1048576; // 1 MB |
| 18653 | |
| 18654 | VmaVirtualBlock block; |
| 18655 | VkResult res = vmaCreateVirtualBlock(&blockCreateInfo, &block); |
| 18656 | \endcode |
| 18657 | |
| 18658 | \section virtual_allocator_making_virtual_allocations Making virtual allocations |
| 18659 | |
| 18660 | #VmaVirtualBlock object contains internal data structure that keeps track of free and occupied regions |
| 18661 | using the same code as the main Vulkan memory allocator. |
| 18662 | Similarly to #VmaAllocation for standard GPU allocations, there is #VmaVirtualAllocation type |
| 18663 | that represents an opaque handle to an allocation withing the virtual block. |
| 18664 | |
| 18665 | In order to make such allocation: |
| 18666 | |
| 18667 | -# Fill in #VmaVirtualAllocationCreateInfo structure. |
| 18668 | -# Call vmaVirtualAllocate(). Get new #VmaVirtualAllocation object that represents the allocation. |
| 18669 | You can also receive `VkDeviceSize offset` that was assigned to the allocation. |
| 18670 | |
| 18671 | Example: |
| 18672 | |
| 18673 | \code |
| 18674 | VmaVirtualAllocationCreateInfo allocCreateInfo = {}; |
| 18675 | allocCreateInfo.size = 4096; // 4 KB |
| 18676 | |
| 18677 | VmaVirtualAllocation alloc; |
| 18678 | VkDeviceSize offset; |
| 18679 | res = vmaVirtualAllocate(block, &allocCreateInfo, &alloc, &offset); |
| 18680 | if(res == VK_SUCCESS) |
| 18681 | { |
| 18682 | // Use the 4 KB of your memory starting at offset. |
| 18683 | } |
| 18684 | else |
| 18685 | { |
| 18686 | // Allocation failed - no space for it could be found. Handle this error! |
| 18687 | } |
| 18688 | \endcode |
| 18689 | |
| 18690 | \section virtual_allocator_deallocation Deallocation |
| 18691 | |
| 18692 | When no longer needed, an allocation can be freed by calling vmaVirtualFree(). |
| 18693 | You can only pass to this function an allocation that was previously returned by vmaVirtualAllocate() |
| 18694 | called for the same #VmaVirtualBlock. |
| 18695 | |
| 18696 | When whole block is no longer needed, the block object can be released by calling vmaDestroyVirtualBlock(). |
| 18697 | All allocations must be freed before the block is destroyed, which is checked internally by an assert. |
| 18698 | However, if you don't want to call vmaVirtualFree() for each allocation, you can use vmaClearVirtualBlock() to free them all at once - |
| 18699 | a feature not available in normal Vulkan memory allocator. Example: |
| 18700 | |
| 18701 | \code |
| 18702 | vmaVirtualFree(block, alloc); |
| 18703 | vmaDestroyVirtualBlock(block); |
| 18704 | \endcode |
| 18705 | |
| 18706 | \section virtual_allocator_allocation_parameters Allocation parameters |
| 18707 | |
| 18708 | You can attach a custom pointer to each allocation by using vmaSetVirtualAllocationUserData(). |
| 18709 | Its default value is null. |
| 18710 | It can be used to store any data that needs to be associated with that allocation - e.g. an index, a handle, or a pointer to some |
| 18711 | larger data structure containing more information. Example: |
| 18712 | |
| 18713 | \code |
| 18714 | struct CustomAllocData |
| 18715 | { |
| 18716 | std::string m_AllocName; |
| 18717 | }; |
| 18718 | CustomAllocData* allocData = new CustomAllocData(); |
| 18719 | allocData->m_AllocName = "My allocation 1"; |
| 18720 | vmaSetVirtualAllocationUserData(block, alloc, allocData); |
| 18721 | \endcode |
| 18722 | |
| 18723 | The pointer can later be fetched, along with allocation offset and size, by passing the allocation handle to function |
| 18724 | vmaGetVirtualAllocationInfo() and inspecting returned structure #VmaVirtualAllocationInfo. |
| 18725 | If you allocated a new object to be used as the custom pointer, don't forget to delete that object before freeing the allocation! |
| 18726 | Example: |
| 18727 | |
| 18728 | \code |
| 18729 | VmaVirtualAllocationInfo allocInfo; |
| 18730 | vmaGetVirtualAllocationInfo(block, alloc, &allocInfo); |
| 18731 | delete (CustomAllocData*)allocInfo.pUserData; |
| 18732 | |
| 18733 | vmaVirtualFree(block, alloc); |
| 18734 | \endcode |
| 18735 | |
| 18736 | \section virtual_allocator_alignment_and_units Alignment and units |
| 18737 | |
| 18738 | It feels natural to express sizes and offsets in bytes. |
| 18739 | If an offset of an allocation needs to be aligned to a multiply of some number (e.g. 4 bytes), you can fill optional member |
| 18740 | VmaVirtualAllocationCreateInfo::alignment to request it. Example: |
| 18741 | |
| 18742 | \code |
| 18743 | VmaVirtualAllocationCreateInfo allocCreateInfo = {}; |
| 18744 | allocCreateInfo.size = 4096; // 4 KB |
| 18745 | allocCreateInfo.alignment = 4; // Returned offset must be a multiply of 4 B |
| 18746 | |
| 18747 | VmaVirtualAllocation alloc; |
| 18748 | res = vmaVirtualAllocate(block, &allocCreateInfo, &alloc, nullptr); |
| 18749 | \endcode |
| 18750 | |
| 18751 | Alignments of different allocations made from one block may vary. |
| 18752 | However, if all alignments and sizes are always multiply of some size e.g. 4 B or `sizeof(MyDataStruct)`, |
| 18753 | you can express all sizes, alignments, and offsets in multiples of that size instead of individual bytes. |
| 18754 | It might be more convenient, but you need to make sure to use this new unit consistently in all the places: |
| 18755 | |
| 18756 | - VmaVirtualBlockCreateInfo::size |
| 18757 | - VmaVirtualAllocationCreateInfo::size and VmaVirtualAllocationCreateInfo::alignment |
| 18758 | - Using offset returned by vmaVirtualAllocate() or in VmaVirtualAllocationInfo::offset |
| 18759 | |
| 18760 | \section virtual_allocator_statistics Statistics |
| 18761 | |
| 18762 | You can obtain statistics of a virtual block using vmaGetVirtualBlockStatistics() |
| 18763 | (to get brief statistics that are fast to calculate) |
| 18764 | or vmaCalculateVirtualBlockStatistics() (to get more detailed statistics, slower to calculate). |
| 18765 | The functions fill structures #VmaStatistics, #VmaDetailedStatistics respectively - same as used by the normal Vulkan memory allocator. |
| 18766 | Example: |
| 18767 | |
| 18768 | \code |
| 18769 | VmaStatistics stats; |
| 18770 | vmaGetVirtualBlockStatistics(block, &stats); |
| 18771 | printf("My virtual block has %llu bytes used by %u virtual allocations\n", |
| 18772 | stats.allocationBytes, stats.allocationCount); |
| 18773 | \endcode |
| 18774 | |
| 18775 | You can also request a full list of allocations and free regions as a string in JSON format by calling |
| 18776 | vmaBuildVirtualBlockStatsString(). |
| 18777 | Returned string must be later freed using vmaFreeVirtualBlockStatsString(). |
| 18778 | The format of this string differs from the one returned by the main Vulkan allocator, but it is similar. |
| 18779 | |
| 18780 | \section virtual_allocator_additional_considerations Additional considerations |
| 18781 | |
| 18782 | The "virtual allocator" functionality is implemented on a level of individual memory blocks. |
| 18783 | Keeping track of a whole collection of blocks, allocating new ones when out of free space, |
| 18784 | deleting empty ones, and deciding which one to try first for a new allocation must be implemented by the user. |
| 18785 | |
| 18786 | Alternative allocation algorithms are supported, just like in custom pools of the real GPU memory. |
| 18787 | See enum #VmaVirtualBlockCreateFlagBits to learn how to specify them (e.g. #VMA_VIRTUAL_BLOCK_CREATE_LINEAR_ALGORITHM_BIT). |
| 18788 | You can find their description in chapter \ref custom_memory_pools. |
| 18789 | Allocation strategies are also supported. |
| 18790 | See enum #VmaVirtualAllocationCreateFlagBits to learn how to specify them (e.g. #VMA_VIRTUAL_ALLOCATION_CREATE_STRATEGY_MIN_TIME_BIT). |
| 18791 | |
| 18792 | Following features are supported only by the allocator of the real GPU memory and not by virtual allocations: |
| 18793 | buffer-image granularity, `VMA_DEBUG_MARGIN`, `VMA_MIN_ALIGNMENT`. |
| 18794 | |
| 18795 | |
| 18796 | \page debugging_memory_usage Debugging incorrect memory usage |
| 18797 | |
| 18798 | If you suspect a bug with memory usage, like usage of uninitialized memory or |
| 18799 | memory being overwritten out of bounds of an allocation, |
| 18800 | you can use debug features of this library to verify this. |
| 18801 | |
| 18802 | \section debugging_memory_usage_initialization Memory initialization |
| 18803 | |
| 18804 | If you experience a bug with incorrect and nondeterministic data in your program and you suspect uninitialized memory to be used, |
| 18805 | you can enable automatic memory initialization to verify this. |
| 18806 | To do it, define macro `VMA_DEBUG_INITIALIZE_ALLOCATIONS` to 1. |
| 18807 | |
| 18808 | \code |
| 18809 | #define VMA_DEBUG_INITIALIZE_ALLOCATIONS 1 |
| 18810 | #include "vk_mem_alloc.h" |
| 18811 | \endcode |
| 18812 | |
| 18813 | It makes memory of new allocations initialized to bit pattern `0xDCDCDCDC`. |
| 18814 | Before an allocation is destroyed, its memory is filled with bit pattern `0xEFEFEFEF`. |
| 18815 | Memory is automatically mapped and unmapped if necessary. |
| 18816 | |
| 18817 | If you find these values while debugging your program, good chances are that you incorrectly |
| 18818 | read Vulkan memory that is allocated but not initialized, or already freed, respectively. |
| 18819 | |
| 18820 | Memory initialization works only with memory types that are `HOST_VISIBLE` and with allocations that can be mapped. |
| 18821 | It works also with dedicated allocations. |
| 18822 | |
| 18823 | \section debugging_memory_usage_margins Margins |
| 18824 | |
| 18825 | By default, allocations are laid out in memory blocks next to each other if possible |
| 18826 | (considering required alignment, `bufferImageGranularity`, and `nonCoherentAtomSize`). |
| 18827 | |
| 18828 |  |
| 18829 | |
| 18830 | Define macro `VMA_DEBUG_MARGIN` to some non-zero value (e.g. 16) to enforce specified |
| 18831 | number of bytes as a margin after every allocation. |
| 18832 | |
| 18833 | \code |
| 18834 | #define VMA_DEBUG_MARGIN 16 |
| 18835 | #include "vk_mem_alloc.h" |
| 18836 | \endcode |
| 18837 | |
| 18838 |  |
| 18839 | |
| 18840 | If your bug goes away after enabling margins, it means it may be caused by memory |
| 18841 | being overwritten outside of allocation boundaries. It is not 100% certain though. |
| 18842 | Change in application behavior may also be caused by different order and distribution |
| 18843 | of allocations across memory blocks after margins are applied. |
| 18844 | |
| 18845 | Margins work with all types of memory. |
| 18846 | |
| 18847 | Margin is applied only to allocations made out of memory blocks and not to dedicated |
| 18848 | allocations, which have their own memory block of specific size. |
| 18849 | It is thus not applied to allocations made using #VMA_ALLOCATION_CREATE_DEDICATED_MEMORY_BIT flag |
| 18850 | or those automatically decided to put into dedicated allocations, e.g. due to its |
| 18851 | large size or recommended by VK_KHR_dedicated_allocation extension. |
| 18852 | |
| 18853 | Margins appear in [JSON dump](@ref statistics_json_dump) as part of free space. |
| 18854 | |
| 18855 | Note that enabling margins increases memory usage and fragmentation. |
| 18856 | |
| 18857 | Margins do not apply to \ref virtual_allocator. |
| 18858 | |
| 18859 | \section debugging_memory_usage_corruption_detection Corruption detection |
| 18860 | |
| 18861 | You can additionally define macro `VMA_DEBUG_DETECT_CORRUPTION` to 1 to enable validation |
| 18862 | of contents of the margins. |
| 18863 | |
| 18864 | \code |
| 18865 | #define VMA_DEBUG_MARGIN 16 |
| 18866 | #define VMA_DEBUG_DETECT_CORRUPTION 1 |
| 18867 | #include "vk_mem_alloc.h" |
| 18868 | \endcode |
| 18869 | |
| 18870 | When this feature is enabled, number of bytes specified as `VMA_DEBUG_MARGIN` |
| 18871 | (it must be multiply of 4) after every allocation is filled with a magic number. |
| 18872 | This idea is also know as "canary". |
| 18873 | Memory is automatically mapped and unmapped if necessary. |
| 18874 | |
| 18875 | This number is validated automatically when the allocation is destroyed. |
| 18876 | If it is not equal to the expected value, `VMA_ASSERT()` is executed. |
| 18877 | It clearly means that either CPU or GPU overwritten the memory outside of boundaries of the allocation, |
| 18878 | which indicates a serious bug. |
| 18879 | |
| 18880 | You can also explicitly request checking margins of all allocations in all memory blocks |
| 18881 | that belong to specified memory types by using function vmaCheckCorruption(), |
| 18882 | or in memory blocks that belong to specified custom pool, by using function |
| 18883 | vmaCheckPoolCorruption(). |
| 18884 | |
| 18885 | Margin validation (corruption detection) works only for memory types that are |
| 18886 | `HOST_VISIBLE` and `HOST_COHERENT`. |
| 18887 | |
| 18888 | |
| 18889 | \page opengl_interop OpenGL Interop |
| 18890 | |
| 18891 | VMA provides some features that help with interoperability with OpenGL. |
| 18892 | |
| 18893 | \section opengl_interop_exporting_memory Exporting memory |
| 18894 | |
| 18895 | If you want to attach `VkExportMemoryAllocateInfoKHR` structure to `pNext` chain of memory allocations made by the library: |
| 18896 | |
| 18897 | It is recommended to create \ref custom_memory_pools for such allocations. |
| 18898 | Define and fill in your `VkExportMemoryAllocateInfoKHR` structure and attach it to VmaPoolCreateInfo::pMemoryAllocateNext |
| 18899 | while creating the custom pool. |
| 18900 | Please note that the structure must remain alive and unchanged for the whole lifetime of the #VmaPool, |
| 18901 | not only while creating it, as no copy of the structure is made, |
| 18902 | but its original pointer is used for each allocation instead. |
| 18903 | |
| 18904 | If you want to export all memory allocated by the library from certain memory types, |
| 18905 | also dedicated allocations or other allocations made from default pools, |
| 18906 | an alternative solution is to fill in VmaAllocatorCreateInfo::pTypeExternalMemoryHandleTypes. |
| 18907 | It should point to an array with `VkExternalMemoryHandleTypeFlagsKHR` to be automatically passed by the library |
| 18908 | through `VkExportMemoryAllocateInfoKHR` on each allocation made from a specific memory type. |
| 18909 | Please note that new versions of the library also support dedicated allocations created in custom pools. |
| 18910 | |
| 18911 | You should not mix these two methods in a way that allows to apply both to the same memory type. |
| 18912 | Otherwise, `VkExportMemoryAllocateInfoKHR` structure would be attached twice to the `pNext` chain of `VkMemoryAllocateInfo`. |
| 18913 | |
| 18914 | |
| 18915 | \section opengl_interop_custom_alignment Custom alignment |
| 18916 | |
| 18917 | Buffers or images exported to a different API like OpenGL may require a different alignment, |
| 18918 | higher than the one used by the library automatically, queried from functions like `vkGetBufferMemoryRequirements`. |
| 18919 | To impose such alignment: |
| 18920 | |
| 18921 | It is recommended to create \ref custom_memory_pools for such allocations. |
| 18922 | Set VmaPoolCreateInfo::minAllocationAlignment member to the minimum alignment required for each allocation |
| 18923 | to be made out of this pool. |
| 18924 | The alignment actually used will be the maximum of this member and the alignment returned for the specific buffer or image |
| 18925 | from a function like `vkGetBufferMemoryRequirements`, which is called by VMA automatically. |
| 18926 | |
| 18927 | If you want to create a buffer with a specific minimum alignment out of default pools, |
| 18928 | use special function vmaCreateBufferWithAlignment(), which takes additional parameter `minAlignment`. |
| 18929 | |
| 18930 | Note the problem of alignment affects only resources placed inside bigger `VkDeviceMemory` blocks and not dedicated |
| 18931 | allocations, as these, by definition, always have alignment = 0 because the resource is bound to the beginning of its dedicated block. |
| 18932 | Contrary to Direct3D 12, Vulkan doesn't have a concept of alignment of the entire memory block passed on its allocation. |
| 18933 | |
| 18934 | |
| 18935 | \page usage_patterns Recommended usage patterns |
| 18936 | |
| 18937 | Vulkan gives great flexibility in memory allocation. |
| 18938 | This chapter shows the most common patterns. |
| 18939 | |
| 18940 | See also slides from talk: |
| 18941 | [Sawicki, Adam. Advanced Graphics Techniques Tutorial: Memory management in Vulkan and DX12. Game Developers Conference, 2018](https://www.gdcvault.com/play/1025458/Advanced-Graphics-Techniques-Tutorial-New) |
| 18942 | |
| 18943 | |
| 18944 | \section usage_patterns_gpu_only GPU-only resource |
| 18945 | |
| 18946 | <b>When:</b> |
| 18947 | Any resources that you frequently write and read on GPU, |
| 18948 | e.g. images used as color attachments (aka "render targets"), depth-stencil attachments, |
| 18949 | images/buffers used as storage image/buffer (aka "Unordered Access View (UAV)"). |
| 18950 | |
| 18951 | <b>What to do:</b> |
| 18952 | Let the library select the optimal memory type, which will likely have `VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT`. |
| 18953 | |
| 18954 | \code |
| 18955 | VkImageCreateInfo imgCreateInfo = { VK_STRUCTURE_TYPE_IMAGE_CREATE_INFO }; |
| 18956 | imgCreateInfo.imageType = VK_IMAGE_TYPE_2D; |
| 18957 | imgCreateInfo.extent.width = 3840; |
| 18958 | imgCreateInfo.extent.height = 2160; |
| 18959 | imgCreateInfo.extent.depth = 1; |
| 18960 | imgCreateInfo.mipLevels = 1; |
| 18961 | imgCreateInfo.arrayLayers = 1; |
| 18962 | imgCreateInfo.format = VK_FORMAT_R8G8B8A8_UNORM; |
| 18963 | imgCreateInfo.tiling = VK_IMAGE_TILING_OPTIMAL; |
| 18964 | imgCreateInfo.initialLayout = VK_IMAGE_LAYOUT_UNDEFINED; |
| 18965 | imgCreateInfo.usage = VK_IMAGE_USAGE_SAMPLED_BIT | VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT; |
| 18966 | imgCreateInfo.samples = VK_SAMPLE_COUNT_1_BIT; |
| 18967 | |
| 18968 | VmaAllocationCreateInfo allocCreateInfo = {}; |
| 18969 | allocCreateInfo.usage = VMA_MEMORY_USAGE_AUTO; |
| 18970 | allocCreateInfo.flags = VMA_ALLOCATION_CREATE_DEDICATED_MEMORY_BIT; |
| 18971 | allocCreateInfo.priority = 1.0f; |
| 18972 | |
| 18973 | VkImage img; |
| 18974 | VmaAllocation alloc; |
| 18975 | vmaCreateImage(allocator, &imgCreateInfo, &allocCreateInfo, &img, &alloc, nullptr); |
| 18976 | \endcode |
| 18977 | |
| 18978 | <b>Also consider:</b> |
| 18979 | Consider creating them as dedicated allocations using #VMA_ALLOCATION_CREATE_DEDICATED_MEMORY_BIT, |
| 18980 | especially if they are large or if you plan to destroy and recreate them with different sizes |
| 18981 | e.g. when display resolution changes. |
| 18982 | Prefer to create such resources first and all other GPU resources (like textures and vertex buffers) later. |
| 18983 | When VK_EXT_memory_priority extension is enabled, it is also worth setting high priority to such allocation |
| 18984 | to decrease chances to be evicted to system memory by the operating system. |
| 18985 | |
| 18986 | \section usage_patterns_staging_copy_upload Staging copy for upload |
| 18987 | |
| 18988 | <b>When:</b> |
| 18989 | A "staging" buffer than you want to map and fill from CPU code, then use as a source od transfer |
| 18990 | to some GPU resource. |
| 18991 | |
| 18992 | <b>What to do:</b> |
| 18993 | Use flag #VMA_ALLOCATION_CREATE_HOST_ACCESS_SEQUENTIAL_WRITE_BIT. |
| 18994 | Let the library select the optimal memory type, which will always have `VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT`. |
| 18995 | |
| 18996 | \code |
| 18997 | VkBufferCreateInfo bufCreateInfo = { VK_STRUCTURE_TYPE_BUFFER_CREATE_INFO }; |
| 18998 | bufCreateInfo.size = 65536; |
| 18999 | bufCreateInfo.usage = VK_BUFFER_USAGE_TRANSFER_SRC_BIT; |
| 19000 | |
| 19001 | VmaAllocationCreateInfo allocCreateInfo = {}; |
| 19002 | allocCreateInfo.usage = VMA_MEMORY_USAGE_AUTO; |
| 19003 | allocCreateInfo.flags = VMA_ALLOCATION_CREATE_HOST_ACCESS_SEQUENTIAL_WRITE_BIT | |
| 19004 | VMA_ALLOCATION_CREATE_MAPPED_BIT; |
| 19005 | |
| 19006 | VkBuffer buf; |
| 19007 | VmaAllocation alloc; |
| 19008 | VmaAllocationInfo allocInfo; |
| 19009 | vmaCreateBuffer(allocator, &bufCreateInfo, &allocCreateInfo, &buf, &alloc, &allocInfo); |
| 19010 | |
| 19011 | ... |
| 19012 | |
| 19013 | memcpy(allocInfo.pMappedData, myData, myDataSize); |
| 19014 | \endcode |
| 19015 | |
| 19016 | <b>Also consider:</b> |
| 19017 | You can map the allocation using vmaMapMemory() or you can create it as persistenly mapped |
| 19018 | using #VMA_ALLOCATION_CREATE_MAPPED_BIT, as in the example above. |
| 19019 | |
| 19020 | |
| 19021 | \section usage_patterns_readback Readback |
| 19022 | |
| 19023 | <b>When:</b> |
| 19024 | Buffers for data written by or transferred from the GPU that you want to read back on the CPU, |
| 19025 | e.g. results of some computations. |
| 19026 | |
| 19027 | <b>What to do:</b> |
| 19028 | Use flag #VMA_ALLOCATION_CREATE_HOST_ACCESS_RANDOM_BIT. |
| 19029 | Let the library select the optimal memory type, which will always have `VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT` |
| 19030 | and `VK_MEMORY_PROPERTY_HOST_CACHED_BIT`. |
| 19031 | |
| 19032 | \code |
| 19033 | VkBufferCreateInfo bufCreateInfo = { VK_STRUCTURE_TYPE_BUFFER_CREATE_INFO }; |
| 19034 | bufCreateInfo.size = 65536; |
| 19035 | bufCreateInfo.usage = VK_BUFFER_USAGE_TRANSFER_DST_BIT; |
| 19036 | |
| 19037 | VmaAllocationCreateInfo allocCreateInfo = {}; |
| 19038 | allocCreateInfo.usage = VMA_MEMORY_USAGE_AUTO; |
| 19039 | allocCreateInfo.flags = VMA_ALLOCATION_CREATE_HOST_ACCESS_RANDOM_BIT | |
| 19040 | VMA_ALLOCATION_CREATE_MAPPED_BIT; |
| 19041 | |
| 19042 | VkBuffer buf; |
| 19043 | VmaAllocation alloc; |
| 19044 | VmaAllocationInfo allocInfo; |
| 19045 | vmaCreateBuffer(allocator, &bufCreateInfo, &allocCreateInfo, &buf, &alloc, &allocInfo); |
| 19046 | |
| 19047 | ... |
| 19048 | |
| 19049 | const float* downloadedData = (const float*)allocInfo.pMappedData; |
| 19050 | \endcode |
| 19051 | |
| 19052 | |
| 19053 | \section usage_patterns_advanced_data_uploading Advanced data uploading |
| 19054 | |
| 19055 | For resources that you frequently write on CPU via mapped pointer and |
| 19056 | freqnently read on GPU e.g. as a uniform buffer (also called "dynamic"), multiple options are possible: |
| 19057 | |
| 19058 | -# Easiest solution is to have one copy of the resource in `HOST_VISIBLE` memory, |
| 19059 | even if it means system RAM (not `DEVICE_LOCAL`) on systems with a discrete graphics card, |
| 19060 | and make the device reach out to that resource directly. |
| 19061 | - Reads performed by the device will then go through PCI Express bus. |
| 19062 | The performace of this access may be limited, but it may be fine depending on the size |
| 19063 | of this resource (whether it is small enough to quickly end up in GPU cache) and the sparsity |
| 19064 | of access. |
| 19065 | -# On systems with unified memory (e.g. AMD APU or Intel integrated graphics, mobile chips), |
| 19066 | a memory type may be available that is both `HOST_VISIBLE` (available for mapping) and `DEVICE_LOCAL` |
| 19067 | (fast to access from the GPU). Then, it is likely the best choice for such type of resource. |
| 19068 | -# Systems with a discrete graphics card and separate video memory may or may not expose |
| 19069 | a memory type that is both `HOST_VISIBLE` and `DEVICE_LOCAL`, also known as Base Address Register (BAR). |
| 19070 | If they do, it represents a piece of VRAM (or entire VRAM, if ReBAR is enabled in the motherboard BIOS) |
| 19071 | that is available to CPU for mapping. |
| 19072 | - Writes performed by the host to that memory go through PCI Express bus. |
| 19073 | The performance of these writes may be limited, but it may be fine, especially on PCIe 4.0, |
| 19074 | as long as rules of using uncached and write-combined memory are followed - only sequential writes and no reads. |
| 19075 | -# Finally, you may need or prefer to create a separate copy of the resource in `DEVICE_LOCAL` memory, |
| 19076 | a separate "staging" copy in `HOST_VISIBLE` memory and perform an explicit transfer command between them. |
| 19077 | |
| 19078 | Thankfully, VMA offers an aid to create and use such resources in the the way optimal |
| 19079 | for the current Vulkan device. To help the library make the best choice, |
| 19080 | use flag #VMA_ALLOCATION_CREATE_HOST_ACCESS_SEQUENTIAL_WRITE_BIT together with |
| 19081 | #VMA_ALLOCATION_CREATE_HOST_ACCESS_ALLOW_TRANSFER_INSTEAD_BIT. |
| 19082 | It will then prefer a memory type that is both `DEVICE_LOCAL` and `HOST_VISIBLE` (integrated memory or BAR), |
| 19083 | but if no such memory type is available or allocation from it fails |
| 19084 | (PC graphics cards have only 256 MB of BAR by default, unless ReBAR is supported and enabled in BIOS), |
| 19085 | it will fall back to `DEVICE_LOCAL` memory for fast GPU access. |
| 19086 | It is then up to you to detect that the allocation ended up in a memory type that is not `HOST_VISIBLE`, |
| 19087 | so you need to create another "staging" allocation and perform explicit transfers. |
| 19088 | |
| 19089 | \code |
| 19090 | VkBufferCreateInfo bufCreateInfo = { VK_STRUCTURE_TYPE_BUFFER_CREATE_INFO }; |
| 19091 | bufCreateInfo.size = 65536; |
| 19092 | bufCreateInfo.usage = VK_BUFFER_USAGE_UNIFORM_BUFFER_BIT | VK_BUFFER_USAGE_TRANSFER_DST_BIT; |
| 19093 | |
| 19094 | VmaAllocationCreateInfo allocCreateInfo = {}; |
| 19095 | allocCreateInfo.usage = VMA_MEMORY_USAGE_AUTO; |
| 19096 | allocCreateInfo.flags = VMA_ALLOCATION_CREATE_HOST_ACCESS_SEQUENTIAL_WRITE_BIT | |
| 19097 | VMA_ALLOCATION_CREATE_HOST_ACCESS_ALLOW_TRANSFER_INSTEAD_BIT | |
| 19098 | VMA_ALLOCATION_CREATE_MAPPED_BIT; |
| 19099 | |
| 19100 | VkBuffer buf; |
| 19101 | VmaAllocation alloc; |
| 19102 | VmaAllocationInfo allocInfo; |
| 19103 | vmaCreateBuffer(allocator, &bufCreateInfo, &allocCreateInfo, &buf, &alloc, &allocInfo); |
| 19104 | |
| 19105 | VkMemoryPropertyFlags memPropFlags; |
| 19106 | vmaGetAllocationMemoryProperties(allocator, alloc, &memPropFlags); |
| 19107 | |
| 19108 | if(memPropFlags & VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT) |
| 19109 | { |
| 19110 | // Allocation ended up in a mappable memory and is already mapped - write to it directly. |
| 19111 | |
| 19112 | // [Executed in runtime]: |
| 19113 | memcpy(allocInfo.pMappedData, myData, myDataSize); |
| 19114 | } |
| 19115 | else |
| 19116 | { |
| 19117 | // Allocation ended up in a non-mappable memory - need to transfer. |
| 19118 | VkBufferCreateInfo stagingBufCreateInfo = { VK_STRUCTURE_TYPE_BUFFER_CREATE_INFO }; |
| 19119 | stagingBufCreateInfo.size = 65536; |
| 19120 | stagingBufCreateInfo.usage = VK_BUFFER_USAGE_TRANSFER_SRC_BIT; |
| 19121 | |
| 19122 | VmaAllocationCreateInfo stagingAllocCreateInfo = {}; |
| 19123 | stagingAllocCreateInfo.usage = VMA_MEMORY_USAGE_AUTO; |
| 19124 | stagingAllocCreateInfo.flags = VMA_ALLOCATION_CREATE_HOST_ACCESS_SEQUENTIAL_WRITE_BIT | |
| 19125 | VMA_ALLOCATION_CREATE_MAPPED_BIT; |
| 19126 | |
| 19127 | VkBuffer stagingBuf; |
| 19128 | VmaAllocation stagingAlloc; |
| 19129 | VmaAllocationInfo stagingAllocInfo; |
| 19130 | vmaCreateBuffer(allocator, &stagingBufCreateInfo, &stagingAllocCreateInfo, |
| 19131 | &stagingBuf, &stagingAlloc, stagingAllocInfo); |
| 19132 | |
| 19133 | // [Executed in runtime]: |
| 19134 | memcpy(stagingAllocInfo.pMappedData, myData, myDataSize); |
| 19135 | //vkCmdPipelineBarrier: VK_ACCESS_HOST_WRITE_BIT --> VK_ACCESS_TRANSFER_READ_BIT |
| 19136 | VkBufferCopy bufCopy = { |
| 19137 | 0, // srcOffset |
| 19138 | 0, // dstOffset, |
| 19139 | myDataSize); // size |
| 19140 | vkCmdCopyBuffer(cmdBuf, stagingBuf, buf, 1, &bufCopy); |
| 19141 | } |
| 19142 | \endcode |
| 19143 | |
| 19144 | \section usage_patterns_other_use_cases Other use cases |
| 19145 | |
| 19146 | Here are some other, less obvious use cases and their recommended settings: |
| 19147 | |
| 19148 | - An image that is used only as transfer source and destination, but it should stay on the device, |
| 19149 | as it is used to temporarily store a copy of some texture, e.g. from the current to the next frame, |
| 19150 | for temporal antialiasing or other temporal effects. |
| 19151 | - Use `VkImageCreateInfo::usage = VK_IMAGE_USAGE_TRANSFER_SRC_BIT | VK_IMAGE_USAGE_TRANSFER_DST_BIT` |
| 19152 | - Use VmaAllocationCreateInfo::usage = #VMA_MEMORY_USAGE_AUTO |
| 19153 | - An image that is used only as transfer source and destination, but it should be placed |
| 19154 | in the system RAM despite it doesn't need to be mapped, because it serves as a "swap" copy to evict |
| 19155 | least recently used textures from VRAM. |
| 19156 | - Use `VkImageCreateInfo::usage = VK_IMAGE_USAGE_TRANSFER_SRC_BIT | VK_IMAGE_USAGE_TRANSFER_DST_BIT` |
| 19157 | - Use VmaAllocationCreateInfo::usage = #VMA_MEMORY_USAGE_AUTO_PREFER_HOST, |
| 19158 | as VMA needs a hint here to differentiate from the previous case. |
| 19159 | - A buffer that you want to map and write from the CPU, directly read from the GPU |
| 19160 | (e.g. as a uniform or vertex buffer), but you have a clear preference to place it in device or |
| 19161 | host memory due to its large size. |
| 19162 | - Use `VkBufferCreateInfo::usage = VK_BUFFER_USAGE_UNIFORM_BUFFER_BIT` |
| 19163 | - Use VmaAllocationCreateInfo::usage = #VMA_MEMORY_USAGE_AUTO_PREFER_DEVICE or #VMA_MEMORY_USAGE_AUTO_PREFER_HOST |
| 19164 | - Use VmaAllocationCreateInfo::flags = #VMA_ALLOCATION_CREATE_HOST_ACCESS_SEQUENTIAL_WRITE_BIT |
| 19165 | |
| 19166 | |
| 19167 | \page configuration Configuration |
| 19168 | |
| 19169 | Please check "CONFIGURATION SECTION" in the code to find macros that you can define |
| 19170 | before each include of this file or change directly in this file to provide |
| 19171 | your own implementation of basic facilities like assert, `min()` and `max()` functions, |
| 19172 | mutex, atomic etc. |
| 19173 | The library uses its own implementation of containers by default, but you can switch to using |
| 19174 | STL containers instead. |
| 19175 | |
| 19176 | For example, define `VMA_ASSERT(expr)` before including the library to provide |
| 19177 | custom implementation of the assertion, compatible with your project. |
| 19178 | By default it is defined to standard C `assert(expr)` in `_DEBUG` configuration |
| 19179 | and empty otherwise. |
| 19180 | |
| 19181 | \section config_Vulkan_functions Pointers to Vulkan functions |
| 19182 | |
| 19183 | There are multiple ways to import pointers to Vulkan functions in the library. |
| 19184 | In the simplest case you don't need to do anything. |
| 19185 | If the compilation or linking of your program or the initialization of the #VmaAllocator |
| 19186 | doesn't work for you, you can try to reconfigure it. |
| 19187 | |
| 19188 | First, the allocator tries to fetch pointers to Vulkan functions linked statically, |
| 19189 | like this: |
| 19190 | |
| 19191 | \code |
| 19192 | m_VulkanFunctions.vkAllocateMemory = (PFN_vkAllocateMemory)vkAllocateMemory; |
| 19193 | \endcode |
| 19194 | |
| 19195 | If you want to disable this feature, set configuration macro: `#define VMA_STATIC_VULKAN_FUNCTIONS 0`. |
| 19196 | |
| 19197 | Second, you can provide the pointers yourself by setting member VmaAllocatorCreateInfo::pVulkanFunctions. |
| 19198 | You can fetch them e.g. using functions `vkGetInstanceProcAddr` and `vkGetDeviceProcAddr` or |
| 19199 | by using a helper library like [volk](https://github.com/zeux/volk). |
| 19200 | |
| 19201 | Third, VMA tries to fetch remaining pointers that are still null by calling |
| 19202 | `vkGetInstanceProcAddr` and `vkGetDeviceProcAddr` on its own. |
| 19203 | You need to only fill in VmaVulkanFunctions::vkGetInstanceProcAddr and VmaVulkanFunctions::vkGetDeviceProcAddr. |
| 19204 | Other pointers will be fetched automatically. |
| 19205 | If you want to disable this feature, set configuration macro: `#define VMA_DYNAMIC_VULKAN_FUNCTIONS 0`. |
| 19206 | |
| 19207 | Finally, all the function pointers required by the library (considering selected |
| 19208 | Vulkan version and enabled extensions) are checked with `VMA_ASSERT` if they are not null. |
| 19209 | |
| 19210 | |
| 19211 | \section custom_memory_allocator Custom host memory allocator |
| 19212 | |
| 19213 | If you use custom allocator for CPU memory rather than default operator `new` |
| 19214 | and `delete` from C++, you can make this library using your allocator as well |
| 19215 | by filling optional member VmaAllocatorCreateInfo::pAllocationCallbacks. These |
| 19216 | functions will be passed to Vulkan, as well as used by the library itself to |
| 19217 | make any CPU-side allocations. |
| 19218 | |
| 19219 | \section allocation_callbacks Device memory allocation callbacks |
| 19220 | |
| 19221 | The library makes calls to `vkAllocateMemory()` and `vkFreeMemory()` internally. |
| 19222 | You can setup callbacks to be informed about these calls, e.g. for the purpose |
| 19223 | of gathering some statistics. To do it, fill optional member |
| 19224 | VmaAllocatorCreateInfo::pDeviceMemoryCallbacks. |
| 19225 | |
| 19226 | \section heap_memory_limit Device heap memory limit |
| 19227 | |
| 19228 | When device memory of certain heap runs out of free space, new allocations may |
| 19229 | fail (returning error code) or they may succeed, silently pushing some existing_ |
| 19230 | memory blocks from GPU VRAM to system RAM (which degrades performance). This |
| 19231 | behavior is implementation-dependent - it depends on GPU vendor and graphics |
| 19232 | driver. |
| 19233 | |
| 19234 | On AMD cards it can be controlled while creating Vulkan device object by using |
| 19235 | VK_AMD_memory_overallocation_behavior extension, if available. |
| 19236 | |
| 19237 | Alternatively, if you want to test how your program behaves with limited amount of Vulkan device |
| 19238 | memory available without switching your graphics card to one that really has |
| 19239 | smaller VRAM, you can use a feature of this library intended for this purpose. |
| 19240 | To do it, fill optional member VmaAllocatorCreateInfo::pHeapSizeLimit. |
| 19241 | |
| 19242 | |
| 19243 | |
| 19244 | \page vk_khr_dedicated_allocation VK_KHR_dedicated_allocation |
| 19245 | |
| 19246 | VK_KHR_dedicated_allocation is a Vulkan extension which can be used to improve |
| 19247 | performance on some GPUs. It augments Vulkan API with possibility to query |
| 19248 | driver whether it prefers particular buffer or image to have its own, dedicated |
| 19249 | allocation (separate `VkDeviceMemory` block) for better efficiency - to be able |
| 19250 | to do some internal optimizations. The extension is supported by this library. |
| 19251 | It will be used automatically when enabled. |
| 19252 | |
| 19253 | It has been promoted to core Vulkan 1.1, so if you use eligible Vulkan version |
| 19254 | and inform VMA about it by setting VmaAllocatorCreateInfo::vulkanApiVersion, |
| 19255 | you are all set. |
| 19256 | |
| 19257 | Otherwise, if you want to use it as an extension: |
| 19258 | |
| 19259 | 1 . When creating Vulkan device, check if following 2 device extensions are |
| 19260 | supported (call `vkEnumerateDeviceExtensionProperties()`). |
| 19261 | If yes, enable them (fill `VkDeviceCreateInfo::ppEnabledExtensionNames`). |
| 19262 | |
| 19263 | - VK_KHR_get_memory_requirements2 |
| 19264 | - VK_KHR_dedicated_allocation |
| 19265 | |
| 19266 | If you enabled these extensions: |
| 19267 | |
| 19268 | 2 . Use #VMA_ALLOCATOR_CREATE_KHR_DEDICATED_ALLOCATION_BIT flag when creating |
| 19269 | your #VmaAllocator to inform the library that you enabled required extensions |
| 19270 | and you want the library to use them. |
| 19271 | |
| 19272 | \code |
| 19273 | allocatorInfo.flags |= VMA_ALLOCATOR_CREATE_KHR_DEDICATED_ALLOCATION_BIT; |
| 19274 | |
| 19275 | vmaCreateAllocator(&allocatorInfo, &allocator); |
| 19276 | \endcode |
| 19277 | |
| 19278 | That is all. The extension will be automatically used whenever you create a |
| 19279 | buffer using vmaCreateBuffer() or image using vmaCreateImage(). |
| 19280 | |
| 19281 | When using the extension together with Vulkan Validation Layer, you will receive |
| 19282 | warnings like this: |
| 19283 | |
| 19284 | _vkBindBufferMemory(): Binding memory to buffer 0x33 but vkGetBufferMemoryRequirements() has not been called on that buffer._ |
| 19285 | |
| 19286 | It is OK, you should just ignore it. It happens because you use function |
| 19287 | `vkGetBufferMemoryRequirements2KHR()` instead of standard |
| 19288 | `vkGetBufferMemoryRequirements()`, while the validation layer seems to be |
| 19289 | unaware of it. |
| 19290 | |
| 19291 | To learn more about this extension, see: |
| 19292 | |
| 19293 | - [VK_KHR_dedicated_allocation in Vulkan specification](https://www.khronos.org/registry/vulkan/specs/1.2-extensions/html/chap50.html#VK_KHR_dedicated_allocation) |
| 19294 | - [VK_KHR_dedicated_allocation unofficial manual](http://asawicki.info/articles/VK_KHR_dedicated_allocation.php5) |
| 19295 | |
| 19296 | |
| 19297 | |
| 19298 | \page vk_ext_memory_priority VK_EXT_memory_priority |
| 19299 | |
| 19300 | VK_EXT_memory_priority is a device extension that allows to pass additional "priority" |
| 19301 | value to Vulkan memory allocations that the implementation may use prefer certain |
| 19302 | buffers and images that are critical for performance to stay in device-local memory |
| 19303 | in cases when the memory is over-subscribed, while some others may be moved to the system memory. |
| 19304 | |
| 19305 | VMA offers convenient usage of this extension. |
| 19306 | If you enable it, you can pass "priority" parameter when creating allocations or custom pools |
| 19307 | and the library automatically passes the value to Vulkan using this extension. |
| 19308 | |
| 19309 | If you want to use this extension in connection with VMA, follow these steps: |
| 19310 | |
| 19311 | \section vk_ext_memory_priority_initialization Initialization |
| 19312 | |
| 19313 | 1) Call `vkEnumerateDeviceExtensionProperties` for the physical device. |
| 19314 | Check if the extension is supported - if returned array of `VkExtensionProperties` contains "VK_EXT_memory_priority". |
| 19315 | |
| 19316 | 2) Call `vkGetPhysicalDeviceFeatures2` for the physical device instead of old `vkGetPhysicalDeviceFeatures`. |
| 19317 | Attach additional structure `VkPhysicalDeviceMemoryPriorityFeaturesEXT` to `VkPhysicalDeviceFeatures2::pNext` to be returned. |
| 19318 | Check if the device feature is really supported - check if `VkPhysicalDeviceMemoryPriorityFeaturesEXT::memoryPriority` is true. |
| 19319 | |
| 19320 | 3) While creating device with `vkCreateDevice`, enable this extension - add "VK_EXT_memory_priority" |
| 19321 | to the list passed as `VkDeviceCreateInfo::ppEnabledExtensionNames`. |
| 19322 | |
| 19323 | 4) While creating the device, also don't set `VkDeviceCreateInfo::pEnabledFeatures`. |
| 19324 | Fill in `VkPhysicalDeviceFeatures2` structure instead and pass it as `VkDeviceCreateInfo::pNext`. |
| 19325 | Enable this device feature - attach additional structure `VkPhysicalDeviceMemoryPriorityFeaturesEXT` to |
| 19326 | `VkPhysicalDeviceFeatures2::pNext` chain and set its member `memoryPriority` to `VK_TRUE`. |
| 19327 | |
| 19328 | 5) While creating #VmaAllocator with vmaCreateAllocator() inform VMA that you |
| 19329 | have enabled this extension and feature - add #VMA_ALLOCATOR_CREATE_EXT_MEMORY_PRIORITY_BIT |
| 19330 | to VmaAllocatorCreateInfo::flags. |
| 19331 | |
| 19332 | \section vk_ext_memory_priority_usage Usage |
| 19333 | |
| 19334 | When using this extension, you should initialize following member: |
| 19335 | |
| 19336 | - VmaAllocationCreateInfo::priority when creating a dedicated allocation with #VMA_ALLOCATION_CREATE_DEDICATED_MEMORY_BIT. |
| 19337 | - VmaPoolCreateInfo::priority when creating a custom pool. |
| 19338 | |
| 19339 | It should be a floating-point value between `0.0f` and `1.0f`, where recommended default is `0.5f`. |
| 19340 | Memory allocated with higher value can be treated by the Vulkan implementation as higher priority |
| 19341 | and so it can have lower chances of being pushed out to system memory, experiencing degraded performance. |
| 19342 | |
| 19343 | It might be a good idea to create performance-critical resources like color-attachment or depth-stencil images |
| 19344 | as dedicated and set high priority to them. For example: |
| 19345 | |
| 19346 | \code |
| 19347 | VkImageCreateInfo imgCreateInfo = { VK_STRUCTURE_TYPE_IMAGE_CREATE_INFO }; |
| 19348 | imgCreateInfo.imageType = VK_IMAGE_TYPE_2D; |
| 19349 | imgCreateInfo.extent.width = 3840; |
| 19350 | imgCreateInfo.extent.height = 2160; |
| 19351 | imgCreateInfo.extent.depth = 1; |
| 19352 | imgCreateInfo.mipLevels = 1; |
| 19353 | imgCreateInfo.arrayLayers = 1; |
| 19354 | imgCreateInfo.format = VK_FORMAT_R8G8B8A8_UNORM; |
| 19355 | imgCreateInfo.tiling = VK_IMAGE_TILING_OPTIMAL; |
| 19356 | imgCreateInfo.initialLayout = VK_IMAGE_LAYOUT_UNDEFINED; |
| 19357 | imgCreateInfo.usage = VK_IMAGE_USAGE_SAMPLED_BIT | VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT; |
| 19358 | imgCreateInfo.samples = VK_SAMPLE_COUNT_1_BIT; |
| 19359 | |
| 19360 | VmaAllocationCreateInfo allocCreateInfo = {}; |
| 19361 | allocCreateInfo.usage = VMA_MEMORY_USAGE_AUTO; |
| 19362 | allocCreateInfo.flags = VMA_ALLOCATION_CREATE_DEDICATED_MEMORY_BIT; |
| 19363 | allocCreateInfo.priority = 1.0f; |
| 19364 | |
| 19365 | VkImage img; |
| 19366 | VmaAllocation alloc; |
| 19367 | vmaCreateImage(allocator, &imgCreateInfo, &allocCreateInfo, &img, &alloc, nullptr); |
| 19368 | \endcode |
| 19369 | |
| 19370 | `priority` member is ignored in the following situations: |
| 19371 | |
| 19372 | - Allocations created in custom pools: They inherit the priority, along with all other allocation parameters |
| 19373 | from the parametrs passed in #VmaPoolCreateInfo when the pool was created. |
| 19374 | - Allocations created in default pools: They inherit the priority from the parameters |
| 19375 | VMA used when creating default pools, which means `priority == 0.5f`. |
| 19376 | |
| 19377 | |
| 19378 | \page vk_amd_device_coherent_memory VK_AMD_device_coherent_memory |
| 19379 | |
| 19380 | VK_AMD_device_coherent_memory is a device extension that enables access to |
| 19381 | additional memory types with `VK_MEMORY_PROPERTY_DEVICE_COHERENT_BIT_AMD` and |
| 19382 | `VK_MEMORY_PROPERTY_DEVICE_UNCACHED_BIT_AMD` flag. It is useful mostly for |
| 19383 | allocation of buffers intended for writing "breadcrumb markers" in between passes |
| 19384 | or draw calls, which in turn are useful for debugging GPU crash/hang/TDR cases. |
| 19385 | |
| 19386 | When the extension is available but has not been enabled, Vulkan physical device |
| 19387 | still exposes those memory types, but their usage is forbidden. VMA automatically |
| 19388 | takes care of that - it returns `VK_ERROR_FEATURE_NOT_PRESENT` when an attempt |
| 19389 | to allocate memory of such type is made. |
| 19390 | |
| 19391 | If you want to use this extension in connection with VMA, follow these steps: |
| 19392 | |
| 19393 | \section vk_amd_device_coherent_memory_initialization Initialization |
| 19394 | |
| 19395 | 1) Call `vkEnumerateDeviceExtensionProperties` for the physical device. |
| 19396 | Check if the extension is supported - if returned array of `VkExtensionProperties` contains "VK_AMD_device_coherent_memory". |
| 19397 | |
| 19398 | 2) Call `vkGetPhysicalDeviceFeatures2` for the physical device instead of old `vkGetPhysicalDeviceFeatures`. |
| 19399 | Attach additional structure `VkPhysicalDeviceCoherentMemoryFeaturesAMD` to `VkPhysicalDeviceFeatures2::pNext` to be returned. |
| 19400 | Check if the device feature is really supported - check if `VkPhysicalDeviceCoherentMemoryFeaturesAMD::deviceCoherentMemory` is true. |
| 19401 | |
| 19402 | 3) While creating device with `vkCreateDevice`, enable this extension - add "VK_AMD_device_coherent_memory" |
| 19403 | to the list passed as `VkDeviceCreateInfo::ppEnabledExtensionNames`. |
| 19404 | |
| 19405 | 4) While creating the device, also don't set `VkDeviceCreateInfo::pEnabledFeatures`. |
| 19406 | Fill in `VkPhysicalDeviceFeatures2` structure instead and pass it as `VkDeviceCreateInfo::pNext`. |
| 19407 | Enable this device feature - attach additional structure `VkPhysicalDeviceCoherentMemoryFeaturesAMD` to |
| 19408 | `VkPhysicalDeviceFeatures2::pNext` and set its member `deviceCoherentMemory` to `VK_TRUE`. |
| 19409 | |
| 19410 | 5) While creating #VmaAllocator with vmaCreateAllocator() inform VMA that you |
| 19411 | have enabled this extension and feature - add #VMA_ALLOCATOR_CREATE_AMD_DEVICE_COHERENT_MEMORY_BIT |
| 19412 | to VmaAllocatorCreateInfo::flags. |
| 19413 | |
| 19414 | \section vk_amd_device_coherent_memory_usage Usage |
| 19415 | |
| 19416 | After following steps described above, you can create VMA allocations and custom pools |
| 19417 | out of the special `DEVICE_COHERENT` and `DEVICE_UNCACHED` memory types on eligible |
| 19418 | devices. There are multiple ways to do it, for example: |
| 19419 | |
| 19420 | - You can request or prefer to allocate out of such memory types by adding |
| 19421 | `VK_MEMORY_PROPERTY_DEVICE_UNCACHED_BIT_AMD` to VmaAllocationCreateInfo::requiredFlags |
| 19422 | or VmaAllocationCreateInfo::preferredFlags. Those flags can be freely mixed with |
| 19423 | other ways of \ref choosing_memory_type, like setting VmaAllocationCreateInfo::usage. |
| 19424 | - If you manually found memory type index to use for this purpose, force allocation |
| 19425 | from this specific index by setting VmaAllocationCreateInfo::memoryTypeBits `= 1u << index`. |
| 19426 | |
| 19427 | \section vk_amd_device_coherent_memory_more_information More information |
| 19428 | |
| 19429 | To learn more about this extension, see [VK_AMD_device_coherent_memory in Vulkan specification](https://www.khronos.org/registry/vulkan/specs/1.2-extensions/man/html/VK_AMD_device_coherent_memory.html) |
| 19430 | |
| 19431 | Example use of this extension can be found in the code of the sample and test suite |
| 19432 | accompanying this library. |
| 19433 | |
| 19434 | |
| 19435 | \page enabling_buffer_device_address Enabling buffer device address |
| 19436 | |
| 19437 | Device extension VK_KHR_buffer_device_address |
| 19438 | allow to fetch raw GPU pointer to a buffer and pass it for usage in a shader code. |
| 19439 | It has been promoted to core Vulkan 1.2. |
| 19440 | |
| 19441 | If you want to use this feature in connection with VMA, follow these steps: |
| 19442 | |
| 19443 | \section enabling_buffer_device_address_initialization Initialization |
| 19444 | |
| 19445 | 1) (For Vulkan version < 1.2) Call `vkEnumerateDeviceExtensionProperties` for the physical device. |
| 19446 | Check if the extension is supported - if returned array of `VkExtensionProperties` contains |
| 19447 | "VK_KHR_buffer_device_address". |
| 19448 | |
| 19449 | 2) Call `vkGetPhysicalDeviceFeatures2` for the physical device instead of old `vkGetPhysicalDeviceFeatures`. |
| 19450 | Attach additional structure `VkPhysicalDeviceBufferDeviceAddressFeatures*` to `VkPhysicalDeviceFeatures2::pNext` to be returned. |
| 19451 | Check if the device feature is really supported - check if `VkPhysicalDeviceBufferDeviceAddressFeatures::bufferDeviceAddress` is true. |
| 19452 | |
| 19453 | 3) (For Vulkan version < 1.2) While creating device with `vkCreateDevice`, enable this extension - add |
| 19454 | "VK_KHR_buffer_device_address" to the list passed as `VkDeviceCreateInfo::ppEnabledExtensionNames`. |
| 19455 | |
| 19456 | 4) While creating the device, also don't set `VkDeviceCreateInfo::pEnabledFeatures`. |
| 19457 | Fill in `VkPhysicalDeviceFeatures2` structure instead and pass it as `VkDeviceCreateInfo::pNext`. |
| 19458 | Enable this device feature - attach additional structure `VkPhysicalDeviceBufferDeviceAddressFeatures*` to |
| 19459 | `VkPhysicalDeviceFeatures2::pNext` and set its member `bufferDeviceAddress` to `VK_TRUE`. |
| 19460 | |
| 19461 | 5) While creating #VmaAllocator with vmaCreateAllocator() inform VMA that you |
| 19462 | have enabled this feature - add #VMA_ALLOCATOR_CREATE_BUFFER_DEVICE_ADDRESS_BIT |
| 19463 | to VmaAllocatorCreateInfo::flags. |
| 19464 | |
| 19465 | \section enabling_buffer_device_address_usage Usage |
| 19466 | |
| 19467 | After following steps described above, you can create buffers with `VK_BUFFER_USAGE_SHADER_DEVICE_ADDRESS_BIT*` using VMA. |
| 19468 | The library automatically adds `VK_MEMORY_ALLOCATE_DEVICE_ADDRESS_BIT*` to |
| 19469 | allocated memory blocks wherever it might be needed. |
| 19470 | |
| 19471 | Please note that the library supports only `VK_BUFFER_USAGE_SHADER_DEVICE_ADDRESS_BIT*`. |
| 19472 | The second part of this functionality related to "capture and replay" is not supported, |
| 19473 | as it is intended for usage in debugging tools like RenderDoc, not in everyday Vulkan usage. |
| 19474 | |
| 19475 | \section enabling_buffer_device_address_more_information More information |
| 19476 | |
| 19477 | To learn more about this extension, see [VK_KHR_buffer_device_address in Vulkan specification](https://www.khronos.org/registry/vulkan/specs/1.2-extensions/html/chap46.html#VK_KHR_buffer_device_address) |
| 19478 | |
| 19479 | Example use of this extension can be found in the code of the sample and test suite |
| 19480 | accompanying this library. |
| 19481 | |
| 19482 | \page general_considerations General considerations |
| 19483 | |
| 19484 | \section general_considerations_thread_safety Thread safety |
| 19485 | |
| 19486 | - The library has no global state, so separate #VmaAllocator objects can be used |
| 19487 | independently. |
| 19488 | There should be no need to create multiple such objects though - one per `VkDevice` is enough. |
| 19489 | - By default, all calls to functions that take #VmaAllocator as first parameter |
| 19490 | are safe to call from multiple threads simultaneously because they are |
| 19491 | synchronized internally when needed. |
| 19492 | This includes allocation and deallocation from default memory pool, as well as custom #VmaPool. |
| 19493 | - When the allocator is created with #VMA_ALLOCATOR_CREATE_EXTERNALLY_SYNCHRONIZED_BIT |
| 19494 | flag, calls to functions that take such #VmaAllocator object must be |
| 19495 | synchronized externally. |
| 19496 | - Access to a #VmaAllocation object must be externally synchronized. For example, |
| 19497 | you must not call vmaGetAllocationInfo() and vmaMapMemory() from different |
| 19498 | threads at the same time if you pass the same #VmaAllocation object to these |
| 19499 | functions. |
| 19500 | - #VmaVirtualBlock is not safe to be used from multiple threads simultaneously. |
| 19501 | |
| 19502 | \section general_considerations_versioning_and_compatibility Versioning and compatibility |
| 19503 | |
| 19504 | The library uses [**Semantic Versioning**](https://semver.org/), |
| 19505 | which means version numbers follow convention: Major.Minor.Patch (e.g. 2.3.0), where: |
| 19506 | |
| 19507 | - Incremented Patch version means a release is backward- and forward-compatible, |
| 19508 | introducing only some internal improvements, bug fixes, optimizations etc. |
| 19509 | or changes that are out of scope of the official API described in this documentation. |
| 19510 | - Incremented Minor version means a release is backward-compatible, |
| 19511 | so existing code that uses the library should continue to work, while some new |
| 19512 | symbols could have been added: new structures, functions, new values in existing |
| 19513 | enums and bit flags, new structure members, but not new function parameters. |
| 19514 | - Incrementing Major version means a release could break some backward compatibility. |
| 19515 | |
| 19516 | All changes between official releases are documented in file "CHANGELOG.md". |
| 19517 | |
| 19518 | \warning Backward compatiblity is considered on the level of C++ source code, not binary linkage. |
| 19519 | Adding new members to existing structures is treated as backward compatible if initializing |
| 19520 | the new members to binary zero results in the old behavior. |
| 19521 | You should always fully initialize all library structures to zeros and not rely on their |
| 19522 | exact binary size. |
| 19523 | |
| 19524 | \section general_considerations_validation_layer_warnings Validation layer warnings |
| 19525 | |
| 19526 | When using this library, you can meet following types of warnings issued by |
| 19527 | Vulkan validation layer. They don't necessarily indicate a bug, so you may need |
| 19528 | to just ignore them. |
| 19529 | |
| 19530 | - *vkBindBufferMemory(): Binding memory to buffer 0xeb8e4 but vkGetBufferMemoryRequirements() has not been called on that buffer.* |
| 19531 | - It happens when VK_KHR_dedicated_allocation extension is enabled. |
| 19532 | `vkGetBufferMemoryRequirements2KHR` function is used instead, while validation layer seems to be unaware of it. |
| 19533 | - *Mapping an image with layout VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL can result in undefined behavior if this memory is used by the device. Only GENERAL or PREINITIALIZED should be used.* |
| 19534 | - It happens when you map a buffer or image, because the library maps entire |
| 19535 | `VkDeviceMemory` block, where different types of images and buffers may end |
| 19536 | up together, especially on GPUs with unified memory like Intel. |
| 19537 | - *Non-linear image 0xebc91 is aliased with linear buffer 0xeb8e4 which may indicate a bug.* |
| 19538 | - It may happen when you use [defragmentation](@ref defragmentation). |
| 19539 | |
| 19540 | \section general_considerations_allocation_algorithm Allocation algorithm |
| 19541 | |
| 19542 | The library uses following algorithm for allocation, in order: |
| 19543 | |
| 19544 | -# Try to find free range of memory in existing blocks. |
| 19545 | -# If failed, try to create a new block of `VkDeviceMemory`, with preferred block size. |
| 19546 | -# If failed, try to create such block with size / 2, size / 4, size / 8. |
| 19547 | -# If failed, try to allocate separate `VkDeviceMemory` for this allocation, |
| 19548 | just like when you use #VMA_ALLOCATION_CREATE_DEDICATED_MEMORY_BIT. |
| 19549 | -# If failed, choose other memory type that meets the requirements specified in |
| 19550 | VmaAllocationCreateInfo and go to point 1. |
| 19551 | -# If failed, return `VK_ERROR_OUT_OF_DEVICE_MEMORY`. |
| 19552 | |
| 19553 | \section general_considerations_features_not_supported Features not supported |
| 19554 | |
| 19555 | Features deliberately excluded from the scope of this library: |
| 19556 | |
| 19557 | -# **Data transfer.** Uploading (streaming) and downloading data of buffers and images |
| 19558 | between CPU and GPU memory and related synchronization is responsibility of the user. |
| 19559 | Defining some "texture" object that would automatically stream its data from a |
| 19560 | staging copy in CPU memory to GPU memory would rather be a feature of another, |
| 19561 | higher-level library implemented on top of VMA. |
| 19562 | VMA doesn't record any commands to a `VkCommandBuffer`. It just allocates memory. |
| 19563 | -# **Recreation of buffers and images.** Although the library has functions for |
| 19564 | buffer and image creation: vmaCreateBuffer(), vmaCreateImage(), you need to |
| 19565 | recreate these objects yourself after defragmentation. That is because the big |
| 19566 | structures `VkBufferCreateInfo`, `VkImageCreateInfo` are not stored in |
| 19567 | #VmaAllocation object. |
| 19568 | -# **Handling CPU memory allocation failures.** When dynamically creating small C++ |
| 19569 | objects in CPU memory (not Vulkan memory), allocation failures are not checked |
| 19570 | and handled gracefully, because that would complicate code significantly and |
| 19571 | is usually not needed in desktop PC applications anyway. |
| 19572 | Success of an allocation is just checked with an assert. |
| 19573 | -# **Code free of any compiler warnings.** Maintaining the library to compile and |
| 19574 | work correctly on so many different platforms is hard enough. Being free of |
| 19575 | any warnings, on any version of any compiler, is simply not feasible. |
| 19576 | There are many preprocessor macros that make some variables unused, function parameters unreferenced, |
| 19577 | or conditional expressions constant in some configurations. |
| 19578 | The code of this library should not be bigger or more complicated just to silence these warnings. |
| 19579 | It is recommended to disable such warnings instead. |
| 19580 | -# This is a C++ library with C interface. **Bindings or ports to any other programming languages** are welcome as external projects but |
| 19581 | are not going to be included into this repository. |
| 19582 | */ |
| 19583 | |