https://github.com/halide/Halide
Tip revision: b1629e0f52f821d836b8b80a0dc90217904a67cc authored by Alexander on 10 May 2024, 02:55:38 UTC
fix Reinterpret cmp
fix Reinterpret cmp
Tip revision: b1629e0
vulkan_memory.h
#ifndef HALIDE_RUNTIME_VULKAN_MEMORY_H
#define HALIDE_RUNTIME_VULKAN_MEMORY_H
#include "internal/block_allocator.h"
#include "internal/memory_resources.h"
#include "vulkan_internal.h"
// Uncomment to enable verbose memory allocation debugging
// #define HL_VK_DEBUG_MEM 1
namespace Halide {
namespace Runtime {
namespace Internal {
namespace Vulkan {
// --------------------------------------------------------------------------
// Enable external client to override Vulkan allocation callbacks (if they so desire)
WEAK ScopedSpinLock::AtomicFlag custom_allocation_callbacks_lock = 0;
WEAK const VkAllocationCallbacks *custom_allocation_callbacks = nullptr; // nullptr => use Vulkan runtime implementation
// --------------------------------------------------------------------------
// Runtime configuration parameters to adjust the behaviour of the block allocator
struct VulkanMemoryConfig {
size_t maximum_pool_size = 0; //< Maximum number of bytes to allocate for the entire pool (including all blocks). Specified in bytes. Zero means no constraint
size_t minimum_block_size = 32 * 1024 * 1024; //< Default block size is 32MB
size_t maximum_block_size = 0; //< Specified in bytes. Zero means no constraint
size_t maximum_block_count = 0; //< Maximum number of blocks to allocate. Zero means no constraint
size_t nearest_multiple = 32; //< Always round up the requested region sizes to the given integer value. Zero means no constraint
};
WEAK VulkanMemoryConfig memory_allocator_config;
// --------------------------------------------------------------------------
/** Vulkan Memory Allocator class interface for managing large
* memory requests stored as contiguous blocks of memory, which
* are then sub-allocated into smaller regions of
* memory to avoid the excessive cost of vkAllocate and the limited
* number of available allocation calls through the API.
*/
class VulkanMemoryAllocator {
public:
// disable copy constructors and assignment
VulkanMemoryAllocator(const VulkanMemoryAllocator &) = delete;
VulkanMemoryAllocator &operator=(const VulkanMemoryAllocator &) = delete;
// disable non-factory constrction
VulkanMemoryAllocator() = delete;
~VulkanMemoryAllocator() = delete;
// Factory methods for creation / destruction
static VulkanMemoryAllocator *create(void *user_context, const VulkanMemoryConfig &config,
VkDevice dev, VkPhysicalDevice phys_dev,
const SystemMemoryAllocatorFns &system_allocator,
const VkAllocationCallbacks *alloc_callbacks = nullptr);
static int destroy(void *user_context, VulkanMemoryAllocator *allocator);
// Public interface methods
MemoryRegion *reserve(void *user_context, const MemoryRequest &request);
int conform(void *user_context, MemoryRequest *request); //< conforms the given memory request into one that can be allocated
int release(void *user_context, MemoryRegion *region); //< unmark and cache the region for reuse
int reclaim(void *user_context, MemoryRegion *region); //< free the region and consolidate
int retain(void *user_context, MemoryRegion *region); //< retain the region and increase its use count
bool collect(void *user_context); //< returns true if any blocks were removed
int release(void *user_context);
int destroy(void *user_context);
void *map(void *user_context, MemoryRegion *region);
int unmap(void *user_context, MemoryRegion *region);
MemoryRegion *create_crop(void *user_context, MemoryRegion *region, uint64_t offset);
int destroy_crop(void *user_context, MemoryRegion *region);
MemoryRegion *owner_of(void *user_context, MemoryRegion *region);
VkDevice current_device() const {
return this->device;
}
VkPhysicalDevice current_physical_device() const {
return this->physical_device;
}
const VkAllocationCallbacks *callbacks() const {
return this->alloc_callbacks;
}
static const VulkanMemoryConfig &default_config();
static int allocate_block(void *instance_ptr, MemoryBlock *block);
static int deallocate_block(void *instance_ptr, MemoryBlock *block);
static int conform_block_request(void *instance_ptr, MemoryRequest *request);
static int allocate_region(void *instance_ptr, MemoryRegion *region);
static int deallocate_region(void *instance_ptr, MemoryRegion *region);
static int conform_region_request(void *instance_ptr, MemoryRequest *request);
size_t bytes_allocated_for_blocks() const;
size_t blocks_allocated() const;
size_t bytes_allocated_for_regions() const;
size_t regions_allocated() const;
private:
static constexpr uint32_t invalid_usage_flags = uint32_t(-1);
static constexpr uint32_t invalid_memory_type = uint32_t(VK_MAX_MEMORY_TYPES);
// Initializes a new instance
int initialize(void *user_context, const VulkanMemoryConfig &config,
VkDevice dev, VkPhysicalDevice phys_dev,
const SystemMemoryAllocatorFns &system_allocator,
const VkAllocationCallbacks *alloc_callbacks = nullptr);
uint32_t select_memory_usage(void *user_context, MemoryProperties properties) const;
uint32_t select_memory_type(void *user_context,
VkPhysicalDevice physical_device,
MemoryProperties properties,
uint32_t required_flags) const;
int lookup_requirements(void *user_context, size_t size, uint32_t usage_flags, VkMemoryRequirements *memory_requirements);
size_t block_byte_count = 0;
size_t block_count = 0;
size_t region_byte_count = 0;
size_t region_count = 0;
void *owner_context = nullptr;
VulkanMemoryConfig config;
VkDevice device = nullptr;
VkPhysicalDevice physical_device = nullptr;
VkPhysicalDeviceLimits physical_device_limits = {};
const VkAllocationCallbacks *alloc_callbacks = nullptr;
BlockAllocator *block_allocator = nullptr;
};
VulkanMemoryAllocator *VulkanMemoryAllocator::create(void *user_context,
const VulkanMemoryConfig &cfg, VkDevice dev, VkPhysicalDevice phys_dev,
const SystemMemoryAllocatorFns &system_allocator,
const VkAllocationCallbacks *alloc_callbacks) {
if (system_allocator.allocate == nullptr) {
error(user_context) << "VulkanBlockAllocator: Unable to create instance! Missing system allocator interface!\n";
return nullptr;
}
VulkanMemoryAllocator *result = reinterpret_cast<VulkanMemoryAllocator *>(
system_allocator.allocate(user_context, sizeof(VulkanMemoryAllocator)));
if (result == nullptr) {
error(user_context) << "VulkanMemoryAllocator: Failed to create instance! Out of memory!\n";
return nullptr; // caller must handle error case for out-of-memory
}
result->initialize(user_context, cfg, dev, phys_dev, system_allocator, alloc_callbacks);
return result;
}
int VulkanMemoryAllocator::destroy(void *user_context, VulkanMemoryAllocator *instance) {
if (instance == nullptr) {
error(user_context) << "VulkanBlockAllocator: Unable to destroy instance! Invalide instance pointer!\n";
return halide_error_code_internal_error;
}
const BlockAllocator::MemoryAllocators &allocators = instance->block_allocator->current_allocators();
instance->destroy(user_context);
BlockAllocator::destroy(user_context, instance->block_allocator);
if (allocators.system.deallocate == nullptr) {
error(user_context) << "VulkanBlockAllocator: Unable to destroy instance! Missing system allocator interface!\n";
return halide_error_code_internal_error;
}
allocators.system.deallocate(user_context, instance);
return halide_error_code_success;
}
int VulkanMemoryAllocator::initialize(void *user_context,
const VulkanMemoryConfig &cfg, VkDevice dev, VkPhysicalDevice phys_dev,
const SystemMemoryAllocatorFns &system_allocator,
const VkAllocationCallbacks *callbacks) {
owner_context = user_context;
config = cfg;
device = dev;
physical_device = phys_dev;
alloc_callbacks = callbacks;
region_count = 0;
region_byte_count = 0;
block_count = 0;
block_byte_count = 0;
BlockAllocator::MemoryAllocators allocators;
allocators.system = system_allocator;
allocators.block = {VulkanMemoryAllocator::allocate_block, VulkanMemoryAllocator::deallocate_block, VulkanMemoryAllocator::conform_block_request};
allocators.region = {VulkanMemoryAllocator::allocate_region, VulkanMemoryAllocator::deallocate_region, VulkanMemoryAllocator::conform_region_request};
BlockAllocator::Config block_allocator_config = {0};
block_allocator_config.maximum_pool_size = cfg.maximum_pool_size;
block_allocator_config.maximum_block_count = cfg.maximum_block_count;
block_allocator_config.maximum_block_size = cfg.maximum_block_size;
block_allocator_config.minimum_block_size = cfg.minimum_block_size;
block_allocator_config.nearest_multiple = cfg.nearest_multiple;
block_allocator = BlockAllocator::create(user_context, block_allocator_config, allocators);
if (block_allocator == nullptr) {
error(user_context) << "VulkanMemoryAllocator: Failed to create BlockAllocator! Out of memory?!\n";
return halide_error_code_out_of_memory;
}
// get the physical device properties to determine limits and allocation requirements
VkPhysicalDeviceProperties physical_device_properties = {0};
memset(&physical_device_limits, 0, sizeof(VkPhysicalDeviceLimits));
vkGetPhysicalDeviceProperties(physical_device, &physical_device_properties);
memcpy(&physical_device_limits, &(physical_device_properties.limits), sizeof(VkPhysicalDeviceLimits));
return halide_error_code_success;
}
MemoryRegion *VulkanMemoryAllocator::reserve(void *user_context, const MemoryRequest &request) {
#if defined(HL_VK_DEBUG_MEM)
debug(nullptr) << "VulkanMemoryAllocator: Reserving memory ("
<< "user_context=" << user_context << " "
<< "block_allocator=" << (void *)(block_allocator) << " "
<< "request_size=" << (uint32_t)(request.size) << " "
<< "device=" << (void *)(device) << " "
<< "physical_device=" << (void *)(physical_device) << ") ...\n";
#endif
if ((device == nullptr) || (physical_device == nullptr)) {
error(user_context) << "VulkanMemoryAllocator: Unable to reserve memory! Invalid device handle!\n";
return nullptr;
}
if (block_allocator == nullptr) {
error(user_context) << "VulkanMemoryAllocator: Unable to reserve memory! Invalid block allocator!\n";
return nullptr;
}
return block_allocator->reserve(this, request);
}
void *VulkanMemoryAllocator::map(void *user_context, MemoryRegion *region) {
#if defined(HL_VK_DEBUG_MEM)
debug(nullptr) << "VulkanMemoryAllocator: Mapping region ("
<< "user_context=" << user_context << " "
<< "device=" << (void *)(device) << " "
<< "physical_device=" << (void *)(physical_device) << " "
<< "region=" << (void *)(region) << " "
<< "region_size=" << (uint32_t)region->size << " "
<< "region_offset=" << (uint32_t)region->offset << " "
<< "crop_offset=" << (uint32_t)region->range.head_offset << ") ...\n";
#endif
if ((device == nullptr) || (physical_device == nullptr)) {
error(user_context) << "VulkanMemoryAllocator: Unable to map memory! Invalid device handle!\n";
return nullptr;
}
if (block_allocator == nullptr) {
error(user_context) << "VulkanMemoryAllocator: Unable to map memory! Invalid block allocator!\n";
return nullptr;
}
MemoryRegion *owner = owner_of(user_context, region);
RegionAllocator *region_allocator = RegionAllocator::find_allocator(user_context, owner);
if (region_allocator == nullptr) {
error(user_context) << "VulkanMemoryAllocator: Unable to map region! Invalid region allocator handle!\n";
return nullptr; // NOTE: caller must handle nullptr
}
BlockResource *block_resource = region_allocator->block_resource();
if (block_resource == nullptr) {
error(user_context) << "VulkanMemoryAllocator: Unable to map region! Invalid block resource handle!\n";
return nullptr; // NOTE: caller must handle nullptr
}
VkDeviceMemory *device_memory = reinterpret_cast<VkDeviceMemory *>(block_resource->memory.handle);
if (device_memory == nullptr) {
error(user_context) << "VulkanMemoryAllocator: Unable to map region! Invalid device memory handle!\n";
return nullptr; // NOTE: caller must handle nullptr
}
void *mapped_ptr = nullptr;
VkDeviceSize memory_offset = region->offset + region->range.head_offset;
VkDeviceSize memory_size = region->size - region->range.tail_offset - region->range.head_offset;
if (((double)region->size - (double)region->range.tail_offset - (double)region->range.head_offset) <= 0.0) {
error(user_context) << "VulkanMemoryAllocator: Unable to map region! Invalid memory range !\n";
return nullptr;
}
#if defined(HL_VK_DEBUG_MEM)
debug(nullptr) << "VulkanMemoryAllocator: MapMemory ("
<< "user_context=" << user_context << "\n"
<< " region_size=" << (uint32_t)region->size << "\n"
<< " region_offset=" << (uint32_t)region->offset << "\n"
<< " region_range.head_offset=" << (uint32_t)region->range.head_offset << "\n"
<< " region_range.tail_offset=" << (uint32_t)region->range.tail_offset << "\n"
<< " memory_offset=" << (uint32_t)memory_offset << "\n"
<< " memory_size=" << (uint32_t)memory_size << "\n)\n";
#endif
VkResult result = vkMapMemory(device, *device_memory, memory_offset, memory_size, 0, (void **)(&mapped_ptr));
if (result != VK_SUCCESS) {
error(user_context) << "VulkanMemoryAllocator: Mapping region failed! vkMapMemory returned error code: " << vk_get_error_name(result) << "\n";
return nullptr;
}
return mapped_ptr;
}
int VulkanMemoryAllocator::unmap(void *user_context, MemoryRegion *region) {
#if defined(HL_VK_DEBUG_MEM)
debug(nullptr) << "VulkanMemoryAllocator: Unmapping region ("
<< "user_context=" << user_context << " "
<< "device=" << (void *)(device) << " "
<< "physical_device=" << (void *)(physical_device) << " "
<< "region=" << (void *)(region) << " "
<< "region_size=" << (uint32_t)region->size << " "
<< "region_offset=" << (uint32_t)region->offset << " "
<< "crop_offset=" << (uint32_t)region->range.head_offset << ") ...\n";
#endif
if ((device == nullptr) || (physical_device == nullptr)) {
error(user_context) << "VulkanMemoryAllocator: Unable to unmap region! Invalid device handle!\n";
return halide_error_code_generic_error;
}
MemoryRegion *owner = owner_of(user_context, region);
RegionAllocator *region_allocator = RegionAllocator::find_allocator(user_context, owner);
if (region_allocator == nullptr) {
error(user_context) << "VulkanMemoryAllocator: Unable to unmap region! Invalid region allocator handle!\n";
return halide_error_code_internal_error;
}
BlockResource *block_resource = region_allocator->block_resource();
if (block_resource == nullptr) {
error(user_context) << "VulkanMemoryAllocator: Unable to unmap region! Invalid block resource handle!\n";
return halide_error_code_internal_error;
}
VkDeviceMemory *device_memory = reinterpret_cast<VkDeviceMemory *>(block_resource->memory.handle);
if (device_memory == nullptr) {
error(user_context) << "VulkanMemoryAllocator: Unable to unmap region! Invalid device memory handle!\n";
return halide_error_code_internal_error;
}
vkUnmapMemory(device, *device_memory);
return halide_error_code_success;
}
MemoryRegion *VulkanMemoryAllocator::create_crop(void *user_context, MemoryRegion *region, uint64_t offset) {
#if defined(HL_VK_DEBUG_MEM)
debug(nullptr) << "VulkanMemoryAllocator: Cropping region ("
<< "user_context=" << user_context << " "
<< "device=" << (void *)(device) << " "
<< "physical_device=" << (void *)(physical_device) << " "
<< "region=" << (void *)(region) << " "
<< "region_size=" << (uint32_t)region->size << " "
<< "region_offset=" << (uint32_t)region->offset << " "
<< "crop_offset=" << (int64_t)offset << ") ...\n";
#endif
if ((device == nullptr) || (physical_device == nullptr)) {
error(user_context) << "VulkanMemoryAllocator: Unable to crop region! Invalid device handle!\n";
return nullptr;
}
MemoryRegion *owner = owner_of(user_context, region);
RegionAllocator *region_allocator = RegionAllocator::find_allocator(user_context, owner);
if (region_allocator == nullptr) {
error(user_context) << "VulkanMemoryAllocator: Unable to unmap region! Invalid region allocator handle!\n";
return nullptr; // NOTE: caller must handle nullptr
}
// increment usage count
int error_code = region_allocator->retain(this, owner);
if (error_code != halide_error_code_success) {
error(user_context) << "VulkanMemoryAllocator: Unable to crop region! Failed to retain memory region!\n";
return nullptr; // NOTE: caller must handle nullptr
}
// create a new region to return, and copy all the other region's properties
const BlockAllocator::MemoryAllocators &allocators = block_allocator->current_allocators();
if (allocators.system.allocate == nullptr) {
error(user_context) << "VulkanMemoryAllocator: Unable to create crop! Missing system allocator interface!\n";
return nullptr;
}
MemoryRegion *memory_region = reinterpret_cast<MemoryRegion *>(
allocators.system.allocate(user_context, sizeof(MemoryRegion)));
if (memory_region == nullptr) {
error(user_context) << "VulkanMemoryAllocator: Failed to allocate memory region! Out of memory!\n";
return nullptr;
}
memcpy(memory_region, owner, sizeof(MemoryRegion));
// point the handle to the owner of the allocated region, and update the head offset
memory_region->is_owner = false;
memory_region->handle = (void *)owner;
memory_region->range.head_offset = owner->range.head_offset + offset;
return memory_region;
}
int VulkanMemoryAllocator::destroy_crop(void *user_context, MemoryRegion *region) {
if (region == nullptr) {
error(user_context) << "VulkanMemoryAllocator: Failed to destroy crop! Invalid memory region!\n";
return halide_error_code_generic_error;
}
MemoryRegion *owner = owner_of(user_context, region);
RegionAllocator *region_allocator = RegionAllocator::find_allocator(user_context, owner);
if (region_allocator == nullptr) {
error(user_context) << "VulkanMemoryAllocator: Unable to destroy crop region! Invalid region allocator handle!\n";
return halide_error_code_internal_error;
}
// decrement usage count
int error_code = region_allocator->release(this, owner);
if (error_code != halide_error_code_success) {
error(user_context) << "VulkanBlockAllocator: Unable to destroy crop region! Region allocator failed to release memory region!\n";
return error_code;
}
// discard the copied region struct
const BlockAllocator::MemoryAllocators &allocators = block_allocator->current_allocators();
if (allocators.system.deallocate == nullptr) {
error(user_context) << "VulkanBlockAllocator: Unable to destroy crop region! Missing system allocator interface!\n";
return halide_error_code_internal_error;
}
allocators.system.deallocate(user_context, region);
return halide_error_code_success;
}
MemoryRegion *VulkanMemoryAllocator::owner_of(void *user_context, MemoryRegion *region) {
if (region->is_owner) {
return region;
} else {
// If this is a cropped region, use the handle to retrieve the owner of the allocation
return reinterpret_cast<MemoryRegion *>(region->handle);
}
}
int VulkanMemoryAllocator::release(void *user_context, MemoryRegion *region) {
#if defined(HL_VK_DEBUG_MEM)
debug(nullptr) << "VulkanMemoryAllocator: Releasing region ("
<< "user_context=" << user_context << " "
<< "region=" << (void *)(region) << " "
<< "size=" << (uint32_t)region->size << " "
<< "offset=" << (uint32_t)region->offset << ") ...\n";
#endif
if ((device == nullptr) || (physical_device == nullptr)) {
error(user_context) << "VulkanMemoryAllocator: Unable to release region! Invalid device handle!\n";
return halide_error_code_generic_error;
}
if (block_allocator == nullptr) {
error(user_context) << "VulkanMemoryAllocator: Unable to release region! Invalid block allocator!\n";
return halide_error_code_generic_error;
}
return block_allocator->release(this, region);
}
int VulkanMemoryAllocator::reclaim(void *user_context, MemoryRegion *region) {
#if defined(HL_VK_DEBUG_MEM)
debug(nullptr) << "VulkanMemoryAllocator: Reclaiming region ("
<< "user_context=" << user_context << " "
<< "region=" << (void *)(region) << " "
<< "size=" << (uint32_t)region->size << " "
<< "offset=" << (uint32_t)region->offset << ") ...\n";
#endif
if ((device == nullptr) || (physical_device == nullptr)) {
error(user_context) << "VulkanMemoryAllocator: Unable to reclaim region! Invalid device handle!\n";
return halide_error_code_generic_error;
}
if (block_allocator == nullptr) {
error(user_context) << "VulkanMemoryAllocator: Unable to reclaim region! Invalid block allocator!\n";
return halide_error_code_generic_error;
}
return block_allocator->reclaim(this, region);
}
int VulkanMemoryAllocator::retain(void *user_context, MemoryRegion *region) {
#if defined(HL_VK_DEBUG_MEM)
debug(nullptr) << "VulkanMemoryAllocator: Retaining region ("
<< "user_context=" << user_context << " "
<< "region=" << (void *)(region) << " "
<< "size=" << (uint32_t)region->size << " "
<< "offset=" << (uint32_t)region->offset << ") ...\n";
#endif
if ((device == nullptr) || (physical_device == nullptr)) {
error(user_context) << "VulkanMemoryAllocator: Unable to retain region! Invalid device handle!\n";
return halide_error_code_generic_error;
}
if (block_allocator == nullptr) {
error(user_context) << "VulkanMemoryAllocator: Unable to retain region! Invalid block allocator!\n";
return halide_error_code_generic_error;
}
return block_allocator->retain(this, region);
}
bool VulkanMemoryAllocator::collect(void *user_context) {
#if defined(HL_VK_DEBUG_MEM)
debug(nullptr) << "VulkanMemoryAllocator: Collecting unused memory ("
<< "user_context=" << user_context << ") ... \n";
#endif
if ((device == nullptr) || (physical_device == nullptr) || (block_allocator == nullptr)) {
return false;
}
return block_allocator->collect(this);
}
int VulkanMemoryAllocator::release(void *user_context) {
#if defined(HL_VK_DEBUG_MEM)
debug(nullptr) << "VulkanMemoryAllocator: Releasing block allocator ("
<< "user_context=" << user_context << ") ... \n";
#endif
if ((device == nullptr) || (physical_device == nullptr)) {
error(user_context) << "VulkanMemoryAllocator: Unable to release allocator! Invalid device handle!\n";
return halide_error_code_generic_error;
}
if (block_allocator == nullptr) {
error(user_context) << "VulkanMemoryAllocator: Unable to release allocator! Invalid block allocator!\n";
return halide_error_code_generic_error;
}
return block_allocator->release(this);
}
int VulkanMemoryAllocator::destroy(void *user_context) {
#if defined(HL_VK_DEBUG_MEM)
debug(nullptr) << "VulkanMemoryAllocator: Destroying allocator ("
<< "user_context=" << user_context << ") ... \n";
#endif
if (block_allocator != nullptr) {
block_allocator->destroy(this);
}
region_count = 0;
region_byte_count = 0;
block_count = 0;
block_byte_count = 0;
return halide_error_code_success;
}
const VulkanMemoryConfig &
VulkanMemoryAllocator::default_config() {
static VulkanMemoryConfig result;
return result;
}
// --
int VulkanMemoryAllocator::lookup_requirements(void *user_context, size_t size, uint32_t usage_flags, VkMemoryRequirements *memory_requirements) {
#if defined(HL_VK_DEBUG_MEM)
debug(nullptr) << "VulkanMemoryAllocator: Looking up requirements ("
<< "user_context=" << user_context << " "
<< "size=" << (uint32_t)block->size << ", "
<< "usage_flags=" << usage_flags << ") ... \n";
#endif
VkBufferCreateInfo create_info = {
VK_STRUCTURE_TYPE_BUFFER_CREATE_INFO, // struct type
nullptr, // struct extending this
0, // create flags
size, // buffer size (in bytes)
usage_flags, // buffer usage flags
VK_SHARING_MODE_EXCLUSIVE, // sharing mode
0, nullptr};
// Create a buffer to determine alignment requirements
VkBuffer buffer = {0};
VkResult result = vkCreateBuffer(this->device, &create_info, this->alloc_callbacks, &buffer);
if (result != VK_SUCCESS) {
#if defined(HL_VK_DEBUG_MEM)
debug(nullptr) << "VulkanMemoryAllocator: Failed to create buffer to find requirements!\n\t"
<< "vkCreateBuffer returned: " << vk_get_error_name(result) << "\n";
#endif
return halide_error_code_device_malloc_failed;
}
vkGetBufferMemoryRequirements(this->device, buffer, memory_requirements);
vkDestroyBuffer(this->device, buffer, this->alloc_callbacks);
return halide_error_code_success;
}
int VulkanMemoryAllocator::conform_block_request(void *instance_ptr, MemoryRequest *request) {
VulkanMemoryAllocator *instance = reinterpret_cast<VulkanMemoryAllocator *>(instance_ptr);
if (instance == nullptr) {
return halide_error_code_internal_error;
}
void *user_context = instance->owner_context;
#if defined(HL_VK_DEBUG_MEM)
debug(nullptr) << "VulkanMemoryAllocator: Conforming block request ("
<< "user_context=" << user_context << " "
<< "request=" << (void *)(request) << ") ... \n";
#endif
if ((instance->device == nullptr) || (instance->physical_device == nullptr)) {
error(user_context) << "VulkanRegionAllocator: Unable to conform block request! Invalid device handle!\n";
return halide_error_code_internal_error;
}
VkMemoryRequirements memory_requirements = {0};
uint32_t usage_flags = instance->select_memory_usage(user_context, request->properties);
int error_code = instance->lookup_requirements(user_context, request->size, usage_flags, &memory_requirements);
if (error_code != halide_error_code_success) {
error(user_context) << "VulkanRegionAllocator: Failed to conform block request! Unable to lookup requirements!\n";
return error_code;
}
#if defined(HL_VK_DEBUG_MEM)
debug(nullptr) << "VulkanMemoryAllocator: Block allocated ("
<< "size=" << (uint32_t)request->size << ", "
<< "required_alignment=" << (uint32_t)memory_requirements.alignment << ", "
<< "required_size=" << (uint32_t)memory_requirements.size << ", "
<< "uniform_buffer_offset_alignment=" << (uint32_t)instance->physical_device_limits.minUniformBufferOffsetAlignment << ", "
<< "storage_buffer_offset_alignment=" << (uint32_t)instance->physical_device_limits.minStorageBufferOffsetAlignment << ", "
<< "dedicated=" << (request->dedicated ? "true" : "false") << ")\n";
#endif
request->size = memory_requirements.size;
request->properties.alignment = memory_requirements.alignment;
return halide_error_code_success;
}
int VulkanMemoryAllocator::allocate_block(void *instance_ptr, MemoryBlock *block) {
VulkanMemoryAllocator *instance = reinterpret_cast<VulkanMemoryAllocator *>(instance_ptr);
if (instance == nullptr) {
return halide_error_code_internal_error;
}
void *user_context = instance->owner_context;
if ((instance->device == nullptr) || (instance->physical_device == nullptr)) {
error(user_context) << "VulkanBlockAllocator: Unable to deallocate block! Invalid device handle!\n";
return halide_error_code_internal_error;
}
if (block == nullptr) {
error(user_context) << "VulkanBlockAllocator: Unable to deallocate block! Invalid pointer!\n";
return halide_error_code_internal_error;
}
#if defined(HL_VK_DEBUG_MEM)
debug(nullptr) << "VulkanMemoryAllocator: Allocating block ("
<< "user_context=" << user_context << " "
<< "block=" << (void *)(block) << " "
<< "size=" << (uint64_t)block->size << ", "
<< "dedicated=" << (block->dedicated ? "true" : "false") << " "
<< "usage=" << halide_memory_usage_name(block->properties.usage) << " "
<< "caching=" << halide_memory_caching_name(block->properties.caching) << " "
<< "visibility=" << halide_memory_visibility_name(block->properties.visibility) << ")\n";
#endif
// Find an appropriate memory type given the flags
uint32_t memory_type = instance->select_memory_type(user_context, instance->physical_device, block->properties, 0);
if (memory_type == invalid_memory_type) {
error(user_context) << "VulkanMemoryAllocator: Unable to find appropriate memory type for device!\n";
return halide_error_code_generic_error;
}
// Allocate memory
VkMemoryAllocateInfo alloc_info = {
VK_STRUCTURE_TYPE_MEMORY_ALLOCATE_INFO, // struct type
nullptr, // struct extending this
block->size, // size of allocation in bytes
memory_type // memory type index from physical device
};
VkDeviceMemory *device_memory = (VkDeviceMemory *)vk_host_malloc(nullptr, sizeof(VkDeviceMemory), 0, VK_SYSTEM_ALLOCATION_SCOPE_OBJECT, instance->alloc_callbacks);
if (device_memory == nullptr) {
debug(nullptr) << "VulkanBlockAllocator: Unable to allocate block! Failed to allocate device memory handle!\n";
return halide_error_code_out_of_memory;
}
VkResult result = vkAllocateMemory(instance->device, &alloc_info, instance->alloc_callbacks, device_memory);
if (result != VK_SUCCESS) {
debug(nullptr) << "VulkanMemoryAllocator: Allocation failed! vkAllocateMemory returned: " << vk_get_error_name(result) << "\n";
return halide_error_code_device_malloc_failed;
}
#ifdef DEBUG_RUNTIME
debug(nullptr) << "vkAllocateMemory: Allocated memory for device region (" << (uint64_t)block->size << " bytes) ...\n";
#endif
block->handle = (void *)device_memory;
instance->block_byte_count += block->size;
instance->block_count++;
return halide_error_code_success;
}
int VulkanMemoryAllocator::deallocate_block(void *instance_ptr, MemoryBlock *block) {
VulkanMemoryAllocator *instance = reinterpret_cast<VulkanMemoryAllocator *>(instance_ptr);
if (instance == nullptr) {
return halide_error_code_internal_error;
}
void *user_context = instance->owner_context;
#if defined(HL_VK_DEBUG_MEM)
debug(nullptr) << "VulkanMemoryAllocator: Deallocating block ("
<< "user_context=" << user_context << " "
<< "block=" << (void *)(block) << ") ... \n";
#endif
if ((instance->device == nullptr) || (instance->physical_device == nullptr)) {
error(user_context) << "VulkanBlockAllocator: Unable to deallocate block! Invalid device handle!\n";
return halide_error_code_internal_error;
}
if (block == nullptr) {
error(user_context) << "VulkanBlockAllocator: Unable to deallocate block! Invalid pointer!\n";
return halide_error_code_internal_error;
}
#if defined(HL_VK_DEBUG_MEM)
debug(nullptr) << "VulkanBlockAllocator: deallocating block ("
<< "size=" << (uint32_t)block->size << ", "
<< "dedicated=" << (block->dedicated ? "true" : "false") << " "
<< "usage=" << halide_memory_usage_name(block->properties.usage) << " "
<< "caching=" << halide_memory_caching_name(block->properties.caching) << " "
<< "visibility=" << halide_memory_visibility_name(block->properties.visibility) << ")\n";
#endif
if (block->handle == nullptr) {
error(user_context) << "VulkanBlockAllocator: Unable to deallocate block! Invalid handle!\n";
return halide_error_code_internal_error;
}
VkDeviceMemory *device_memory = reinterpret_cast<VkDeviceMemory *>(block->handle);
if (device_memory == nullptr) {
error(user_context) << "VulkanBlockAllocator: Unable to deallocate block! Invalid device memory handle!\n";
return halide_error_code_internal_error;
}
vkFreeMemory(instance->device, *device_memory, instance->alloc_callbacks);
#ifdef DEBUG_RUNTIME
debug(nullptr) << "vkFreeMemory: Deallocated memory for device region (" << (uint64_t)block->size << " bytes) ...\n";
#endif
if (instance->block_count > 0) {
instance->block_count--;
} else {
error(nullptr) << "VulkanRegionAllocator: Block counter invalid ... reseting to zero!\n";
instance->block_count = 0;
}
if (int64_t(instance->block_byte_count) - int64_t(block->size) >= 0) {
instance->block_byte_count -= block->size;
} else {
error(nullptr) << "VulkanRegionAllocator: Block byte counter invalid ... reseting to zero!\n";
instance->block_byte_count = 0;
}
block->handle = nullptr;
vk_host_free(nullptr, device_memory, instance->alloc_callbacks);
device_memory = nullptr;
return halide_error_code_success;
}
size_t VulkanMemoryAllocator::blocks_allocated() const {
return block_count;
}
size_t VulkanMemoryAllocator::bytes_allocated_for_blocks() const {
return block_byte_count;
}
uint32_t VulkanMemoryAllocator::select_memory_type(void *user_context,
VkPhysicalDevice physical_device,
MemoryProperties properties,
uint32_t required_flags) const {
uint32_t want_flags = 0; //< preferred memory flags for requested access type
uint32_t need_flags = 0; //< must have in order to enable requested access
switch (properties.visibility) {
case MemoryVisibility::HostOnly:
want_flags |= VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT;
break;
case MemoryVisibility::DeviceOnly:
need_flags |= VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT;
break;
case MemoryVisibility::DeviceToHost:
need_flags |= VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT;
want_flags |= VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT;
break;
case MemoryVisibility::HostToDevice:
need_flags |= VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT;
break;
case MemoryVisibility::DefaultVisibility:
case MemoryVisibility::InvalidVisibility:
default:
error(nullptr) << "VulkanMemoryAllocator: Unable to convert type! Invalid memory visibility request!\n\t"
<< "visibility=" << halide_memory_visibility_name(properties.visibility) << "\n";
return invalid_memory_type;
};
switch (properties.caching) {
case MemoryCaching::CachedCoherent:
if (need_flags & VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT) {
want_flags |= VK_MEMORY_PROPERTY_HOST_CACHED_BIT | VK_MEMORY_PROPERTY_HOST_COHERENT_BIT;
}
break;
case MemoryCaching::UncachedCoherent:
if (need_flags & VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT) {
want_flags |= VK_MEMORY_PROPERTY_HOST_COHERENT_BIT;
}
break;
case MemoryCaching::Cached:
if (need_flags & VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT) {
want_flags |= VK_MEMORY_PROPERTY_HOST_CACHED_BIT;
}
break;
case MemoryCaching::Uncached:
case MemoryCaching::DefaultCaching:
break;
case MemoryCaching::InvalidCaching:
default:
error(user_context) << "VulkanMemoryAllocator: Unable to convert type! Invalid memory caching request!\n\t"
<< "caching=" << halide_memory_caching_name(properties.caching) << "\n";
return invalid_memory_type;
};
VkPhysicalDeviceMemoryProperties device_memory_properties;
vkGetPhysicalDeviceMemoryProperties(physical_device, &device_memory_properties);
uint32_t result = invalid_memory_type;
for (uint32_t i = 0; i < device_memory_properties.memoryTypeCount; ++i) {
// if required flags are given, see if the memory type matches the requirement
if (required_flags) {
if (((required_flags >> i) & 1) == 0) {
continue;
}
}
const VkMemoryPropertyFlags properties = device_memory_properties.memoryTypes[i].propertyFlags;
if (need_flags) {
if ((properties & need_flags) != need_flags) {
continue;
}
}
if (want_flags) {
if ((properties & want_flags) != want_flags) {
continue;
}
}
result = i;
break;
}
if (result == invalid_memory_type) {
error(user_context) << "VulkanBlockAllocator: Failed to find appropriate memory type for given properties:\n\t"
<< "usage=" << halide_memory_usage_name(properties.usage) << " "
<< "caching=" << halide_memory_caching_name(properties.caching) << " "
<< "visibility=" << halide_memory_visibility_name(properties.visibility) << "\n";
return invalid_memory_type;
}
return result;
}
// --
int VulkanMemoryAllocator::conform(void *user_context, MemoryRequest *request) {
// NOTE: Vulkan will only allow us to bind device memory to a buffer if the memory requirements are met.
// So now we have to check those (on every allocation) and potentially recreate the buffer if the requirements
// don't match the requested VkBuffer's properties. Note that this is the internal storage for the driver,
// whose size may be required to larger than our requested size (even though we will only ever touch the
// size of the region we're managing as within our block)
VkMemoryRequirements memory_requirements = {0};
uint32_t usage_flags = select_memory_usage(user_context, request->properties);
int error_code = lookup_requirements(user_context, request->size, usage_flags, &memory_requirements);
if (error_code != halide_error_code_success) {
error(user_context) << "VulkanRegionAllocator: Failed to conform block request! Unable to lookup requirements!\n";
return error_code;
}
#if defined(HL_VK_DEBUG_MEM)
debug(nullptr) << "VulkanMemoryAllocator: Buffer requirements ("
<< "requested_size=" << (uint32_t)region->size << ", "
<< "required_alignment=" << (uint32_t)memory_requirements.alignment << ", "
<< "required_size=" << (uint32_t)memory_requirements.size << ")\n";
#endif
// Enforce any alignment constraints reported by the device limits for each usage type
if (usage_flags & VK_BUFFER_USAGE_STORAGE_BUFFER_BIT) {
if ((request->alignment % this->physical_device_limits.minStorageBufferOffsetAlignment) != 0) {
request->alignment = this->physical_device_limits.minStorageBufferOffsetAlignment;
}
} else if (usage_flags & VK_BUFFER_USAGE_UNIFORM_BUFFER_BIT) {
if ((request->alignment % this->physical_device_limits.minUniformBufferOffsetAlignment) != 0) {
request->alignment = this->physical_device_limits.minUniformBufferOffsetAlignment;
}
}
// Ensure the request ends on an aligned address
if (request->alignment > config.nearest_multiple) {
request->properties.nearest_multiple = request->alignment;
}
size_t actual_alignment = conform_alignment(request->alignment, memory_requirements.alignment);
size_t actual_offset = aligned_offset(request->offset, actual_alignment);
size_t actual_size = conform_size(actual_offset, memory_requirements.size, actual_alignment, request->properties.nearest_multiple);
#if defined(HL_VK_DEBUG_MEM)
if ((request->size != actual_size) || (request->alignment != actual_alignment) || (request->offset != actual_offset)) {
debug(nullptr) << "VulkanMemoryAllocator: Adjusting request to match requirements (\n"
<< " size = " << (uint64_t)request->size << " => " << (uint64_t)actual_size << ",\n"
<< " alignment = " << (uint64_t)request->alignment << " => " << (uint64_t)actual_alignment << ",\n"
<< " offset = " << (uint64_t)request->offset << " => " << (uint64_t)actual_offset << ",\n"
<< " required.size = " << (uint64_t)memory_requirements.size << ",\n"
<< " required.alignment = " << (uint64_t)memory_requirements.alignment << "\n)\n";
}
#endif
request->size = actual_size;
request->alignment = actual_alignment;
request->offset = actual_offset;
return halide_error_code_success;
}
int VulkanMemoryAllocator::conform_region_request(void *instance_ptr, MemoryRequest *request) {
VulkanMemoryAllocator *instance = reinterpret_cast<VulkanMemoryAllocator *>(instance_ptr);
if (instance == nullptr) {
return halide_error_code_internal_error;
}
void *user_context = instance->owner_context;
#if defined(HL_VK_DEBUG_MEM)
debug(nullptr) << "VulkanMemoryAllocator: Conforming region request ("
<< "user_context=" << user_context << " "
<< "request=" << (void *)(region) << ") ... \n";
#endif
if ((instance->device == nullptr) || (instance->physical_device == nullptr)) {
error(user_context) << "VulkanRegionAllocator: Unable to conform region request! Invalid device handle!\n";
return halide_error_code_internal_error;
}
#if defined(HL_VK_DEBUG_MEM)
debug(nullptr) << "VulkanRegionAllocator: Conforming region request ("
<< "size=" << (uint32_t)request->size << ", "
<< "offset=" << (uint32_t)request->offset << ", "
<< "dedicated=" << (request->dedicated ? "true" : "false") << " "
<< "usage=" << halide_memory_usage_name(request->properties.usage) << " "
<< "caching=" << halide_memory_caching_name(request->properties.caching) << " "
<< "visibility=" << halide_memory_visibility_name(request->properties.visibility) << ")\n";
#endif
return instance->conform(user_context, request);
}
int VulkanMemoryAllocator::allocate_region(void *instance_ptr, MemoryRegion *region) {
VulkanMemoryAllocator *instance = reinterpret_cast<VulkanMemoryAllocator *>(instance_ptr);
if (instance == nullptr) {
return halide_error_code_internal_error;
}
void *user_context = instance->owner_context;
#if defined(HL_VK_DEBUG_MEM)
debug(nullptr) << "VulkanMemoryAllocator: Allocating region ("
<< "user_context=" << user_context << " "
<< "region=" << (void *)(region) << ") ... \n";
#endif
if ((instance->device == nullptr) || (instance->physical_device == nullptr)) {
error(user_context) << "VulkanRegionAllocator: Unable to allocate region! Invalid device handle!\n";
return halide_error_code_internal_error;
}
if (region == nullptr) {
error(user_context) << "VulkanRegionAllocator: Unable to allocate region! Invalid pointer!\n";
return halide_error_code_internal_error;
}
#if defined(HL_VK_DEBUG_MEM)
debug(nullptr) << "VulkanRegionAllocator: Allocating region ("
<< "size=" << (uint32_t)region->size << ", "
<< "offset=" << (uint32_t)region->offset << ", "
<< "dedicated=" << (region->dedicated ? "true" : "false") << " "
<< "usage=" << halide_memory_usage_name(region->properties.usage) << " "
<< "caching=" << halide_memory_caching_name(region->properties.caching) << " "
<< "visibility=" << halide_memory_visibility_name(region->properties.visibility) << ")\n";
#endif
uint32_t usage_flags = instance->select_memory_usage(user_context, region->properties);
VkBufferCreateInfo create_info = {
VK_STRUCTURE_TYPE_BUFFER_CREATE_INFO, // struct type
nullptr, // struct extending this
0, // create flags
region->size, // buffer size (in bytes)
usage_flags, // buffer usage flags
VK_SHARING_MODE_EXCLUSIVE, // sharing mode
0, nullptr};
VkBuffer *buffer = (VkBuffer *)vk_host_malloc(nullptr, sizeof(VkBuffer), 0, VK_SYSTEM_ALLOCATION_SCOPE_OBJECT, instance->alloc_callbacks);
if (buffer == nullptr) {
error(user_context) << "VulkanRegionAllocator: Unable to allocate region! Failed to allocate buffer handle!\n";
return halide_error_code_out_of_memory;
}
VkResult result = vkCreateBuffer(instance->device, &create_info, instance->alloc_callbacks, buffer);
if (result != VK_SUCCESS) {
error(user_context) << "VulkanRegionAllocator: Failed to create buffer!\n\t"
<< "vkCreateBuffer returned: " << vk_get_error_name(result) << "\n";
return halide_error_code_device_malloc_failed;
}
// NOTE: Vulkan will only allow us to bind device memory to a buffer if the memory requirements are met.
// So now we have to check those (on every allocation) and potentially recreate the buffer if the requirements
// don't match the requested VkBuffer's properties. Note that this is the internal storage for the driver,
// whose size may be required to larger than our requested size (even though we will only ever touch the
// size of the region we're managing as within our block)
VkMemoryRequirements memory_requirements = {0};
vkGetBufferMemoryRequirements(instance->device, *buffer, &memory_requirements);
#if defined(HL_VK_DEBUG_MEM)
debug(nullptr) << "VulkanMemoryAllocator: Buffer requirements ("
<< "requested_size=" << (uint32_t)region->size << ", "
<< "required_alignment=" << (uint32_t)memory_requirements.alignment << ", "
<< "required_size=" << (uint32_t)memory_requirements.size << ")\n";
#endif
if (memory_requirements.size > region->size) {
vkDestroyBuffer(instance->device, *buffer, instance->alloc_callbacks);
#ifdef DEBUG_RUNTIME
debug(nullptr) << "VulkanMemoryAllocator: Reallocating buffer to match required size ("
<< (uint64_t)region->size << " => " << (uint64_t)memory_requirements.size << " bytes) ...\n";
#endif
create_info.size = memory_requirements.size;
VkResult result = vkCreateBuffer(instance->device, &create_info, instance->alloc_callbacks, buffer);
if (result != VK_SUCCESS) {
error(user_context) << "VulkanRegionAllocator: Failed to recreate buffer!\n\t"
<< "vkCreateBuffer returned: " << vk_get_error_name(result) << "\n";
return halide_error_code_device_malloc_failed;
}
}
#ifdef DEBUG_RUNTIME
debug(nullptr) << "vkCreateBuffer: Created buffer for device region (" << (uint64_t)region->size << " bytes) ...\n";
#endif
RegionAllocator *region_allocator = RegionAllocator::find_allocator(user_context, region);
if (region_allocator == nullptr) {
error(user_context) << "VulkanBlockAllocator: Unable to allocate region! Invalid region allocator!\n";
return halide_error_code_internal_error;
}
BlockResource *block_resource = region_allocator->block_resource();
if (block_resource == nullptr) {
error(user_context) << "VulkanBlockAllocator: Unable to allocate region! Invalid block resource handle!\n";
return halide_error_code_internal_error;
}
VkDeviceMemory *device_memory = reinterpret_cast<VkDeviceMemory *>(block_resource->memory.handle);
if (device_memory == nullptr) {
error(user_context) << "VulkanBlockAllocator: Unable to allocate region! Invalid device memory handle!\n";
return halide_error_code_internal_error;
}
// Finally, bind buffer to the device memory
result = vkBindBufferMemory(instance->device, *buffer, *device_memory, region->offset);
if (result != VK_SUCCESS) {
error(user_context) << "VulkanRegionAllocator: Failed to bind buffer!\n\t"
<< "vkBindBufferMemory returned: " << vk_get_error_name(result) << "\n";
return halide_error_code_generic_error;
}
region->handle = (void *)buffer;
region->is_owner = true;
instance->region_byte_count += region->size;
instance->region_count++;
return halide_error_code_success;
}
int VulkanMemoryAllocator::deallocate_region(void *instance_ptr, MemoryRegion *region) {
VulkanMemoryAllocator *instance = reinterpret_cast<VulkanMemoryAllocator *>(instance_ptr);
if (instance == nullptr) {
return halide_error_code_internal_error;
}
void *user_context = instance->owner_context;
#if defined(HL_VK_DEBUG_MEM)
debug(nullptr) << "VulkanMemoryAllocator: Deallocating region ("
<< "user_context=" << user_context << " "
<< "region=" << (void *)(region) << ") ... \n";
#endif
if ((instance->device == nullptr) || (instance->physical_device == nullptr)) {
error(user_context) << "VulkanRegionAllocator: Unable to deallocate region! Invalid device handle!\n";
return halide_error_code_internal_error;
}
if (region == nullptr) {
error(user_context) << "VulkanRegionAllocator: Unable to deallocate region! Invalid pointer!\n";
return halide_error_code_internal_error;
}
#if defined(HL_VK_DEBUG_MEM)
debug(nullptr) << "VulkanRegionAllocator: Deallocating region ("
<< "size=" << (uint32_t)region->size << ", "
<< "offset=" << (uint32_t)region->offset << ", "
<< "dedicated=" << (region->dedicated ? "true" : "false") << " "
<< "usage=" << halide_memory_usage_name(region->properties.usage) << " "
<< "caching=" << halide_memory_caching_name(region->properties.caching) << " "
<< "visibility=" << halide_memory_visibility_name(region->properties.visibility) << ")\n";
#endif
if (region->handle == nullptr) {
error(user_context) << "VulkanRegionAllocator: Unable to deallocate region! Invalid handle!\n";
return halide_error_code_internal_error;
}
VkBuffer *buffer = reinterpret_cast<VkBuffer *>(region->handle);
if (buffer == nullptr) {
error(user_context) << "VulkanRegionAllocator: Unable to deallocate region! Invalid buffer handle!\n";
return halide_error_code_internal_error;
}
vkDestroyBuffer(instance->device, *buffer, instance->alloc_callbacks);
#ifdef DEBUG_RUNTIME
debug(nullptr) << "vkDestroyBuffer: Destroyed buffer for device region (" << (uint64_t)region->size << " bytes) ...\n";
#endif
region->handle = nullptr;
if (instance->region_count > 0) {
instance->region_count--;
} else {
error(nullptr) << "VulkanRegionAllocator: Region counter invalid ... reseting to zero!\n";
instance->region_count = 0;
return halide_error_code_internal_error;
}
if (int64_t(instance->region_byte_count) - int64_t(region->size) >= 0) {
instance->region_byte_count -= region->size;
} else {
error(nullptr) << "VulkanRegionAllocator: Region byte counter invalid ... reseting to zero!\n";
instance->region_byte_count = 0;
return halide_error_code_internal_error;
}
vk_host_free(nullptr, buffer, instance->alloc_callbacks);
buffer = nullptr;
return halide_error_code_success;
}
size_t VulkanMemoryAllocator::regions_allocated() const {
return region_count;
}
size_t VulkanMemoryAllocator::bytes_allocated_for_regions() const {
return region_byte_count;
}
uint32_t VulkanMemoryAllocator::select_memory_usage(void *user_context, MemoryProperties properties) const {
uint32_t result = 0;
switch (properties.usage) {
case MemoryUsage::UniformStorage:
result |= VK_BUFFER_USAGE_UNIFORM_BUFFER_BIT;
break;
case MemoryUsage::DynamicStorage:
case MemoryUsage::StaticStorage:
result |= VK_BUFFER_USAGE_STORAGE_BUFFER_BIT;
break;
case MemoryUsage::TransferSrc:
result |= VK_BUFFER_USAGE_STORAGE_BUFFER_BIT | VK_BUFFER_USAGE_TRANSFER_SRC_BIT;
break;
case MemoryUsage::TransferDst:
result |= VK_BUFFER_USAGE_STORAGE_BUFFER_BIT | VK_BUFFER_USAGE_TRANSFER_DST_BIT;
break;
case MemoryUsage::TransferSrcDst:
result |= VK_BUFFER_USAGE_STORAGE_BUFFER_BIT | VK_BUFFER_USAGE_TRANSFER_SRC_BIT | VK_BUFFER_USAGE_TRANSFER_DST_BIT;
break;
case MemoryUsage::DefaultUsage:
case MemoryUsage::InvalidUsage:
default:
error(user_context) << "VulkanRegionAllocator: Unable to convert type! Invalid memory usage request!\n\t"
<< "usage=" << halide_memory_usage_name(properties.usage) << "\n";
return invalid_usage_flags;
};
if (result == invalid_usage_flags) {
error(user_context) << "VulkanRegionAllocator: Failed to find appropriate memory usage for given properties:\n\t"
<< "usage=" << halide_memory_usage_name(properties.usage) << " "
<< "caching=" << halide_memory_caching_name(properties.caching) << " "
<< "visibility=" << halide_memory_visibility_name(properties.visibility) << "\n";
return invalid_usage_flags;
}
return result;
}
// --------------------------------------------------------------------------
namespace {
// --------------------------------------------------------------------------
// Halide System allocator for host allocations
void *vk_system_malloc(void *user_context, size_t size) {
return malloc(size);
}
void vk_system_free(void *user_context, void *ptr) {
free(ptr);
}
// Vulkan host-side allocation
void *vk_host_malloc(void *user_context, size_t size, size_t alignment, VkSystemAllocationScope scope, const VkAllocationCallbacks *callbacks) {
if (callbacks) {
return callbacks->pfnAllocation(user_context, size, alignment, scope);
} else {
return vk_system_malloc(user_context, size);
}
}
void vk_host_free(void *user_context, void *ptr, const VkAllocationCallbacks *callbacks) {
if (callbacks) {
return callbacks->pfnFree(user_context, ptr);
} else {
return vk_system_free(user_context, ptr);
}
}
VulkanMemoryAllocator *vk_create_memory_allocator(void *user_context,
VkDevice device,
VkPhysicalDevice physical_device,
const VkAllocationCallbacks *alloc_callbacks) {
SystemMemoryAllocatorFns system_allocator = {vk_system_malloc, vk_system_free};
VulkanMemoryConfig config = memory_allocator_config;
// Parse the allocation config string (if specified).
//
// `HL_VK_ALLOC_CONFIG=N:N:N` will tell Halide to configure the Vulkan memory
// allocator use the given constraints specified as three integer values
// separated by a `:` or `;`. These values correspond to `minimum_block_size`,
// `maximum_block_size` and `maximum_block_count`.
//
const char *alloc_config = vk_get_alloc_config_internal(user_context);
if (!StringUtils::is_empty(alloc_config)) {
StringTable alloc_config_values;
alloc_config_values.parse(user_context, alloc_config, HL_VK_ENV_DELIM);
if (alloc_config_values.size() > 0) {
config.maximum_pool_size = atoi(alloc_config_values[0]) * 1024 * 1024;
print(user_context) << "Vulkan: Configuring allocator with " << (uint32_t)config.maximum_pool_size << " for maximum pool size (in bytes)\n";
}
if (alloc_config_values.size() > 1) {
config.minimum_block_size = atoi(alloc_config_values[1]) * 1024 * 1024;
print(user_context) << "Vulkan: Configuring allocator with " << (uint32_t)config.minimum_block_size << " for minimum block size (in bytes)\n";
}
if (alloc_config_values.size() > 2) {
config.maximum_block_size = atoi(alloc_config_values[2]) * 1024 * 1024;
print(user_context) << "Vulkan: Configuring allocator with " << (uint32_t)config.maximum_block_size << " for maximum block size (in bytes)\n";
}
if (alloc_config_values.size() > 3) {
config.maximum_block_count = atoi(alloc_config_values[3]);
print(user_context) << "Vulkan: Configuring allocator with " << (uint32_t)config.maximum_block_count << " for maximum block count\n";
}
if (alloc_config_values.size() > 4) {
config.nearest_multiple = atoi(alloc_config_values[4]);
print(user_context) << "Vulkan: Configuring allocator with " << (uint32_t)config.nearest_multiple << " for nearest multiple\n";
}
}
return VulkanMemoryAllocator::create(user_context,
config, device, physical_device,
system_allocator, alloc_callbacks);
}
int vk_destroy_memory_allocator(void *user_context, VulkanMemoryAllocator *allocator) {
if (allocator != nullptr) {
VulkanMemoryAllocator::destroy(user_context, allocator);
allocator = nullptr;
}
return halide_error_code_success;
}
// --------------------------------------------------------------------------
int vk_clear_device_buffer(void *user_context,
VulkanMemoryAllocator *allocator,
VkCommandPool command_pool,
VkQueue command_queue,
VkBuffer device_buffer) {
#ifdef DEBUG_RUNTIME
debug(user_context)
<< " vk_clear_device_buffer (user_context: " << user_context << ", "
<< "allocator: " << (void *)allocator << ", "
<< "command_pool: " << (void *)command_pool << ", "
<< "command_queue: " << (void *)command_queue << ", "
<< "device_buffer: " << (void *)device_buffer << ")\n";
#endif
// create a command buffer
VkCommandBuffer command_buffer;
int error_code = vk_create_command_buffer(user_context, allocator, command_pool, &command_buffer);
if (error_code != halide_error_code_success) {
error(user_context) << "Vulkan: Failed to create command buffer!\n";
return error_code;
}
// begin the command buffer
VkCommandBufferBeginInfo command_buffer_begin_info =
{
VK_STRUCTURE_TYPE_COMMAND_BUFFER_BEGIN_INFO, // struct type
nullptr, // pointer to struct extending this
VK_COMMAND_BUFFER_USAGE_ONE_TIME_SUBMIT_BIT, // flags
nullptr // pointer to parent command buffer
};
VkResult result = vkBeginCommandBuffer(command_buffer, &command_buffer_begin_info);
if (result != VK_SUCCESS) {
error(user_context) << "Vulkan: vkBeginCommandBuffer returned " << vk_get_error_name(result) << "\n";
return halide_error_code_generic_error;
}
// fill buffer with zero values up to the size of the buffer
vkCmdFillBuffer(command_buffer, device_buffer, 0, VK_WHOLE_SIZE, 0);
// end the command buffer
result = vkEndCommandBuffer(command_buffer);
if (result != VK_SUCCESS) {
error(user_context) << "Vulkan: vkEndCommandBuffer returned " << vk_get_error_name(result) << "\n";
return halide_error_code_generic_error;
}
// submit the command buffer
VkSubmitInfo submit_info =
{
VK_STRUCTURE_TYPE_SUBMIT_INFO, // struct type
nullptr, // pointer to struct extending this
0, // wait semaphore count
nullptr, // semaphores
nullptr, // pipeline stages where semaphore waits occur
1, // how many command buffers to execute
&command_buffer, // the command buffers
0, // number of semaphores to signal
nullptr // the semaphores to signal
};
result = vkQueueSubmit(command_queue, 1, &submit_info, 0);
if (result != VK_SUCCESS) {
error(user_context) << "Vulkan: vkQueueSubmit returned " << vk_get_error_name(result) << "\n";
return halide_error_code_generic_error;
}
// wait for memset to finish
result = vkQueueWaitIdle(command_queue);
if (result != VK_SUCCESS) {
error(user_context) << "Vulkan: vkQueueWaitIdle returned " << vk_get_error_name(result) << "\n";
return halide_error_code_generic_error;
}
error_code = vk_destroy_command_buffer(user_context, allocator, command_pool, command_buffer);
if (error_code != halide_error_code_success) {
error(user_context) << "Vulkan: Failed to destroy command buffer!\n";
return error_code;
}
return halide_error_code_success;
}
// --------------------------------------------------------------------------
} // namespace
} // namespace Vulkan
} // namespace Internal
} // namespace Runtime
} // namespace Halide
// --------------------------------------------------------------------------
extern "C" {
// --------------------------------------------------------------------------
WEAK void halide_vulkan_set_allocation_callbacks(const VkAllocationCallbacks *callbacks) {
using namespace Halide::Runtime::Internal::Vulkan;
ScopedSpinLock lock(&custom_allocation_callbacks_lock);
custom_allocation_callbacks = callbacks;
}
WEAK const VkAllocationCallbacks *halide_vulkan_get_allocation_callbacks(void *user_context) {
using namespace Halide::Runtime::Internal::Vulkan;
ScopedSpinLock lock(&custom_allocation_callbacks_lock);
return custom_allocation_callbacks;
}
// --------------------------------------------------------------------------
} // extern "C"
#endif // HALIDE_RUNTIME_VULKAN_MEMORY_H