/* * Copyright 2015 The Android Open Source Project * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #include <hardware/hwvulkan.h> #include <algorithm> #include <array> #include <inttypes.h> #include <stdlib.h> #include <string.h> #include <log/log.h> #include <utils/Errors.h> #include "null_driver_gen.h" using namespace null_driver; struct VkPhysicalDevice_T { hwvulkan_dispatch_t dispatch; }; struct VkInstance_T { hwvulkan_dispatch_t dispatch; VkAllocationCallbacks allocator; VkPhysicalDevice_T physical_device; uint64_t next_callback_handle; }; struct VkQueue_T { hwvulkan_dispatch_t dispatch; }; struct VkCommandBuffer_T { hwvulkan_dispatch_t dispatch; }; namespace { // Handles for non-dispatchable objects are either pointers, or arbitrary // 64-bit non-zero values. We only use pointers when we need to keep state for // the object even in a null driver. For the rest, we form a handle as: // [63:63] = 1 to distinguish from pointer handles* // [62:56] = non-zero handle type enum value // [55: 0] = per-handle-type incrementing counter // * This works because virtual addresses with the high bit set are reserved // for kernel data in all ABIs we run on. // // We never reclaim handles on vkDestroy*. It's not even necessary for us to // have distinct handles for live objects, and practically speaking we won't // ever create 2^56 objects of the same type from a single VkDevice in a null // driver. // // Using a namespace here instead of 'enum class' since we want scoped // constants but also want implicit conversions to integral types. namespace HandleType { enum Enum { kBufferView, kDebugReportCallbackEXT, kDescriptorPool, kDescriptorSet, kDescriptorSetLayout, kEvent, kFence, kFramebuffer, kImageView, kPipeline, kPipelineCache, kPipelineLayout, kQueryPool, kRenderPass, kSampler, kSemaphore, kShaderModule, kNumTypes }; } // namespace HandleType const VkDeviceSize kMaxDeviceMemory = 0x10000000; // 256 MiB, arbitrary } // anonymous namespace struct VkDevice_T { hwvulkan_dispatch_t dispatch; VkAllocationCallbacks allocator; VkInstance_T* instance; VkQueue_T queue; std::array<uint64_t, HandleType::kNumTypes> next_handle; }; // ----------------------------------------------------------------------------- // Declare HAL_MODULE_INFO_SYM early so it can be referenced by nulldrv_device // later. namespace { int OpenDevice(const hw_module_t* module, const char* id, hw_device_t** device); hw_module_methods_t nulldrv_module_methods = {.open = OpenDevice}; } // namespace #pragma clang diagnostic push #pragma clang diagnostic ignored "-Wmissing-variable-declarations" __attribute__((visibility("default"))) hwvulkan_module_t HAL_MODULE_INFO_SYM = { .common = { .tag = HARDWARE_MODULE_TAG, .module_api_version = HWVULKAN_MODULE_API_VERSION_0_1, .hal_api_version = HARDWARE_HAL_API_VERSION, .id = HWVULKAN_HARDWARE_MODULE_ID, .name = "Null Vulkan Driver", .author = "The Android Open Source Project", .methods = &nulldrv_module_methods, }, }; #pragma clang diagnostic pop // ----------------------------------------------------------------------------- namespace { int CloseDevice(struct hw_device_t* /*device*/) { // nothing to do - opening a device doesn't allocate any resources return 0; } hwvulkan_device_t nulldrv_device = { .common = { .tag = HARDWARE_DEVICE_TAG, .version = HWVULKAN_DEVICE_API_VERSION_0_1, .module = &HAL_MODULE_INFO_SYM.common, .close = CloseDevice, }, .EnumerateInstanceExtensionProperties = EnumerateInstanceExtensionProperties, .CreateInstance = CreateInstance, .GetInstanceProcAddr = GetInstanceProcAddr}; int OpenDevice(const hw_module_t* /*module*/, const char* id, hw_device_t** device) { if (strcmp(id, HWVULKAN_DEVICE_0) == 0) { *device = &nulldrv_device.common; return 0; } return -ENOENT; } VkInstance_T* GetInstanceFromPhysicalDevice( VkPhysicalDevice_T* physical_device) { return reinterpret_cast<VkInstance_T*>( reinterpret_cast<uintptr_t>(physical_device) - offsetof(VkInstance_T, physical_device)); } uint64_t AllocHandle(uint64_t type, uint64_t* next_handle) { const uint64_t kHandleMask = (UINT64_C(1) << 56) - 1; ALOGE_IF(*next_handle == kHandleMask, "non-dispatchable handles of type=%" PRIu64 " are about to overflow", type); return (UINT64_C(1) << 63) | ((type & 0x7) << 56) | ((*next_handle)++ & kHandleMask); } template <class Handle> Handle AllocHandle(VkInstance instance, HandleType::Enum type) { return reinterpret_cast<Handle>( AllocHandle(type, &instance->next_callback_handle)); } template <class Handle> Handle AllocHandle(VkDevice device, HandleType::Enum type) { return reinterpret_cast<Handle>( AllocHandle(type, &device->next_handle[type])); } VKAPI_ATTR void* DefaultAllocate(void*, size_t size, size_t alignment, VkSystemAllocationScope) { void* ptr = nullptr; // Vulkan requires 'alignment' to be a power of two, but posix_memalign // additionally requires that it be at least sizeof(void*). int ret = posix_memalign(&ptr, std::max(alignment, sizeof(void*)), size); return ret == 0 ? ptr : nullptr; } VKAPI_ATTR void* DefaultReallocate(void*, void* ptr, size_t size, size_t alignment, VkSystemAllocationScope) { if (size == 0) { free(ptr); return nullptr; } // TODO(jessehall): Right now we never shrink allocations; if the new // request is smaller than the existing chunk, we just continue using it. // The null driver never reallocs, so this doesn't matter. If that changes, // or if this code is copied into some other project, this should probably // have a heuristic to allocate-copy-free when doing so will save "enough" // space. size_t old_size = ptr ? malloc_usable_size(ptr) : 0; if (size <= old_size) return ptr; void* new_ptr = nullptr; if (posix_memalign(&new_ptr, std::max(alignment, sizeof(void*)), size) != 0) return nullptr; if (ptr) { memcpy(new_ptr, ptr, std::min(old_size, size)); free(ptr); } return new_ptr; } VKAPI_ATTR void DefaultFree(void*, void* ptr) { free(ptr); } const VkAllocationCallbacks kDefaultAllocCallbacks = { .pUserData = nullptr, .pfnAllocation = DefaultAllocate, .pfnReallocation = DefaultReallocate, .pfnFree = DefaultFree, }; } // namespace namespace null_driver { #define DEFINE_OBJECT_HANDLE_CONVERSION(T) \ T* Get##T##FromHandle(Vk##T h); \ T* Get##T##FromHandle(Vk##T h) { \ return reinterpret_cast<T*>(uintptr_t(h)); \ } \ Vk##T GetHandleTo##T(const T* obj); \ Vk##T GetHandleTo##T(const T* obj) { \ return Vk##T(reinterpret_cast<uintptr_t>(obj)); \ } // ----------------------------------------------------------------------------- // Global VKAPI_ATTR VkResult EnumerateInstanceExtensionProperties( const char* layer_name, uint32_t* count, VkExtensionProperties* properties) { if (layer_name) { ALOGW( "Driver vkEnumerateInstanceExtensionProperties shouldn't be called " "with a layer name ('%s')", layer_name); } // NOTE: Change this to zero to report and extension, which can be useful // for testing changes to the loader. #if 1 (void)properties; // unused *count = 0; return VK_SUCCESS; #else const VkExtensionProperties kExtensions[] = { {VK_EXT_DEBUG_REPORT_EXTENSION_NAME, VK_EXT_DEBUG_REPORT_SPEC_VERSION}}; const uint32_t kExtensionsCount = sizeof(kExtensions) / sizeof(kExtensions[0]); if (!properties || *count > kExtensionsCount) *count = kExtensionsCount; if (properties) std::copy(kExtensions, kExtensions + *count, properties); return *count < kExtensionsCount ? VK_INCOMPLETE : VK_SUCCESS; #endif } VKAPI_ATTR VkResult CreateInstance(const VkInstanceCreateInfo* create_info, const VkAllocationCallbacks* allocator, VkInstance* out_instance) { if (!allocator) allocator = &kDefaultAllocCallbacks; VkInstance_T* instance = static_cast<VkInstance_T*>(allocator->pfnAllocation( allocator->pUserData, sizeof(VkInstance_T), alignof(VkInstance_T), VK_SYSTEM_ALLOCATION_SCOPE_INSTANCE)); if (!instance) return VK_ERROR_OUT_OF_HOST_MEMORY; instance->dispatch.magic = HWVULKAN_DISPATCH_MAGIC; instance->allocator = *allocator; instance->physical_device.dispatch.magic = HWVULKAN_DISPATCH_MAGIC; instance->next_callback_handle = 0; for (uint32_t i = 0; i < create_info->enabledExtensionCount; i++) { if (strcmp(create_info->ppEnabledExtensionNames[i], VK_EXT_DEBUG_REPORT_EXTENSION_NAME) == 0) { ALOGV("instance extension '%s' requested", create_info->ppEnabledExtensionNames[i]); } else { ALOGW("unsupported extension '%s' requested", create_info->ppEnabledExtensionNames[i]); } } *out_instance = instance; return VK_SUCCESS; } VKAPI_ATTR PFN_vkVoidFunction GetInstanceProcAddr(VkInstance instance, const char* name) { return instance ? GetInstanceProcAddr(name) : GetGlobalProcAddr(name); } VKAPI_ATTR PFN_vkVoidFunction GetDeviceProcAddr(VkDevice, const char* name) { return GetInstanceProcAddr(name); } // ----------------------------------------------------------------------------- // Instance void DestroyInstance(VkInstance instance, const VkAllocationCallbacks* /*allocator*/) { instance->allocator.pfnFree(instance->allocator.pUserData, instance); } // ----------------------------------------------------------------------------- // PhysicalDevice VkResult EnumeratePhysicalDevices(VkInstance instance, uint32_t* physical_device_count, VkPhysicalDevice* physical_devices) { if (physical_devices && *physical_device_count >= 1) physical_devices[0] = &instance->physical_device; *physical_device_count = 1; return VK_SUCCESS; } VkResult EnumerateDeviceLayerProperties(VkPhysicalDevice /*gpu*/, uint32_t* count, VkLayerProperties* /*properties*/) { ALOGW("Driver vkEnumerateDeviceLayerProperties shouldn't be called"); *count = 0; return VK_SUCCESS; } VkResult EnumerateDeviceExtensionProperties(VkPhysicalDevice /*gpu*/, const char* layer_name, uint32_t* count, VkExtensionProperties* properties) { if (layer_name) { ALOGW( "Driver vkEnumerateDeviceExtensionProperties shouldn't be called " "with a layer name ('%s')", layer_name); *count = 0; return VK_SUCCESS; } const VkExtensionProperties kExtensions[] = { {VK_ANDROID_NATIVE_BUFFER_EXTENSION_NAME, VK_ANDROID_NATIVE_BUFFER_SPEC_VERSION}}; const uint32_t kExtensionsCount = sizeof(kExtensions) / sizeof(kExtensions[0]); if (!properties || *count > kExtensionsCount) *count = kExtensionsCount; if (properties) std::copy(kExtensions, kExtensions + *count, properties); return *count < kExtensionsCount ? VK_INCOMPLETE : VK_SUCCESS; } void GetPhysicalDeviceProperties(VkPhysicalDevice, VkPhysicalDeviceProperties* properties) { properties->apiVersion = VK_MAKE_VERSION(1, 0, VK_HEADER_VERSION); properties->driverVersion = VK_MAKE_VERSION(0, 0, 1); properties->vendorID = 0; properties->deviceID = 0; properties->deviceType = VK_PHYSICAL_DEVICE_TYPE_OTHER; strcpy(properties->deviceName, "Android Vulkan Null Driver"); memset(properties->pipelineCacheUUID, 0, sizeof(properties->pipelineCacheUUID)); properties->limits = VkPhysicalDeviceLimits{ 4096, // maxImageDimension1D 4096, // maxImageDimension2D 256, // maxImageDimension3D 4096, // maxImageDimensionCube 256, // maxImageArrayLayers 65536, // maxTexelBufferElements 16384, // maxUniformBufferRange 1 << 27, // maxStorageBufferRange 128, // maxPushConstantsSize 4096, // maxMemoryAllocationCount 4000, // maxSamplerAllocationCount 1, // bufferImageGranularity 0, // sparseAddressSpaceSize 4, // maxBoundDescriptorSets 16, // maxPerStageDescriptorSamplers 12, // maxPerStageDescriptorUniformBuffers 4, // maxPerStageDescriptorStorageBuffers 16, // maxPerStageDescriptorSampledImages 4, // maxPerStageDescriptorStorageImages 4, // maxPerStageDescriptorInputAttachments 128, // maxPerStageResources 96, // maxDescriptorSetSamplers 72, // maxDescriptorSetUniformBuffers 8, // maxDescriptorSetUniformBuffersDynamic 24, // maxDescriptorSetStorageBuffers 4, // maxDescriptorSetStorageBuffersDynamic 96, // maxDescriptorSetSampledImages 24, // maxDescriptorSetStorageImages 4, // maxDescriptorSetInputAttachments 16, // maxVertexInputAttributes 16, // maxVertexInputBindings 2047, // maxVertexInputAttributeOffset 2048, // maxVertexInputBindingStride 64, // maxVertexOutputComponents 0, // maxTessellationGenerationLevel 0, // maxTessellationPatchSize 0, // maxTessellationControlPerVertexInputComponents 0, // maxTessellationControlPerVertexOutputComponents 0, // maxTessellationControlPerPatchOutputComponents 0, // maxTessellationControlTotalOutputComponents 0, // maxTessellationEvaluationInputComponents 0, // maxTessellationEvaluationOutputComponents 0, // maxGeometryShaderInvocations 0, // maxGeometryInputComponents 0, // maxGeometryOutputComponents 0, // maxGeometryOutputVertices 0, // maxGeometryTotalOutputComponents 64, // maxFragmentInputComponents 4, // maxFragmentOutputAttachments 0, // maxFragmentDualSrcAttachments 4, // maxFragmentCombinedOutputResources 16384, // maxComputeSharedMemorySize {65536, 65536, 65536}, // maxComputeWorkGroupCount[3] 128, // maxComputeWorkGroupInvocations {128, 128, 64}, // maxComputeWorkGroupSize[3] 4, // subPixelPrecisionBits 4, // subTexelPrecisionBits 4, // mipmapPrecisionBits UINT32_MAX, // maxDrawIndexedIndexValue 1, // maxDrawIndirectCount 2, // maxSamplerLodBias 1, // maxSamplerAnisotropy 1, // maxViewports {4096, 4096}, // maxViewportDimensions[2] {-8192.0f, 8191.0f}, // viewportBoundsRange[2] 0, // viewportSubPixelBits 64, // minMemoryMapAlignment 256, // minTexelBufferOffsetAlignment 256, // minUniformBufferOffsetAlignment 256, // minStorageBufferOffsetAlignment -8, // minTexelOffset 7, // maxTexelOffset 0, // minTexelGatherOffset 0, // maxTexelGatherOffset 0.0f, // minInterpolationOffset 0.0f, // maxInterpolationOffset 0, // subPixelInterpolationOffsetBits 4096, // maxFramebufferWidth 4096, // maxFramebufferHeight 256, // maxFramebufferLayers VK_SAMPLE_COUNT_1_BIT | VK_SAMPLE_COUNT_4_BIT, // framebufferColorSampleCounts VK_SAMPLE_COUNT_1_BIT | VK_SAMPLE_COUNT_4_BIT, // framebufferDepthSampleCounts VK_SAMPLE_COUNT_1_BIT | VK_SAMPLE_COUNT_4_BIT, // framebufferStencilSampleCounts VK_SAMPLE_COUNT_1_BIT | VK_SAMPLE_COUNT_4_BIT, // framebufferNoAttachmentsSampleCounts 4, // maxColorAttachments VK_SAMPLE_COUNT_1_BIT | VK_SAMPLE_COUNT_4_BIT, // sampledImageColorSampleCounts VK_SAMPLE_COUNT_1_BIT, // sampledImageIntegerSampleCounts VK_SAMPLE_COUNT_1_BIT | VK_SAMPLE_COUNT_4_BIT, // sampledImageDepthSampleCounts VK_SAMPLE_COUNT_1_BIT | VK_SAMPLE_COUNT_4_BIT, // sampledImageStencilSampleCounts VK_SAMPLE_COUNT_1_BIT, // storageImageSampleCounts 1, // maxSampleMaskWords VK_TRUE, // timestampComputeAndGraphics 1, // timestampPeriod 0, // maxClipDistances 0, // maxCullDistances 0, // maxCombinedClipAndCullDistances 2, // discreteQueuePriorities {1.0f, 1.0f}, // pointSizeRange[2] {1.0f, 1.0f}, // lineWidthRange[2] 0.0f, // pointSizeGranularity 0.0f, // lineWidthGranularity VK_TRUE, // strictLines VK_TRUE, // standardSampleLocations 1, // optimalBufferCopyOffsetAlignment 1, // optimalBufferCopyRowPitchAlignment 64, // nonCoherentAtomSize }; } void GetPhysicalDeviceQueueFamilyProperties( VkPhysicalDevice, uint32_t* count, VkQueueFamilyProperties* properties) { if (!properties || *count > 1) *count = 1; if (properties && *count == 1) { properties->queueFlags = VK_QUEUE_GRAPHICS_BIT | VK_QUEUE_COMPUTE_BIT | VK_QUEUE_TRANSFER_BIT; properties->queueCount = 1; properties->timestampValidBits = 64; properties->minImageTransferGranularity = VkExtent3D{1, 1, 1}; } } void GetPhysicalDeviceMemoryProperties( VkPhysicalDevice, VkPhysicalDeviceMemoryProperties* properties) { properties->memoryTypeCount = 1; properties->memoryTypes[0].propertyFlags = VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT | VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT | VK_MEMORY_PROPERTY_HOST_COHERENT_BIT | VK_MEMORY_PROPERTY_HOST_CACHED_BIT; properties->memoryTypes[0].heapIndex = 0; properties->memoryHeapCount = 1; properties->memoryHeaps[0].size = kMaxDeviceMemory; properties->memoryHeaps[0].flags = VK_MEMORY_HEAP_DEVICE_LOCAL_BIT; } void GetPhysicalDeviceFeatures(VkPhysicalDevice /*gpu*/, VkPhysicalDeviceFeatures* features) { *features = VkPhysicalDeviceFeatures{ VK_TRUE, // robustBufferAccess VK_FALSE, // fullDrawIndexUint32 VK_FALSE, // imageCubeArray VK_FALSE, // independentBlend VK_FALSE, // geometryShader VK_FALSE, // tessellationShader VK_FALSE, // sampleRateShading VK_FALSE, // dualSrcBlend VK_FALSE, // logicOp VK_FALSE, // multiDrawIndirect VK_FALSE, // drawIndirectFirstInstance VK_FALSE, // depthClamp VK_FALSE, // depthBiasClamp VK_FALSE, // fillModeNonSolid VK_FALSE, // depthBounds VK_FALSE, // wideLines VK_FALSE, // largePoints VK_FALSE, // alphaToOne VK_FALSE, // multiViewport VK_FALSE, // samplerAnisotropy VK_FALSE, // textureCompressionETC2 VK_FALSE, // textureCompressionASTC_LDR VK_FALSE, // textureCompressionBC VK_FALSE, // occlusionQueryPrecise VK_FALSE, // pipelineStatisticsQuery VK_FALSE, // vertexPipelineStoresAndAtomics VK_FALSE, // fragmentStoresAndAtomics VK_FALSE, // shaderTessellationAndGeometryPointSize VK_FALSE, // shaderImageGatherExtended VK_FALSE, // shaderStorageImageExtendedFormats VK_FALSE, // shaderStorageImageMultisample VK_FALSE, // shaderStorageImageReadWithoutFormat VK_FALSE, // shaderStorageImageWriteWithoutFormat VK_FALSE, // shaderUniformBufferArrayDynamicIndexing VK_FALSE, // shaderSampledImageArrayDynamicIndexing VK_FALSE, // shaderStorageBufferArrayDynamicIndexing VK_FALSE, // shaderStorageImageArrayDynamicIndexing VK_FALSE, // shaderClipDistance VK_FALSE, // shaderCullDistance VK_FALSE, // shaderFloat64 VK_FALSE, // shaderInt64 VK_FALSE, // shaderInt16 VK_FALSE, // shaderResourceResidency VK_FALSE, // shaderResourceMinLod VK_FALSE, // sparseBinding VK_FALSE, // sparseResidencyBuffer VK_FALSE, // sparseResidencyImage2D VK_FALSE, // sparseResidencyImage3D VK_FALSE, // sparseResidency2Samples VK_FALSE, // sparseResidency4Samples VK_FALSE, // sparseResidency8Samples VK_FALSE, // sparseResidency16Samples VK_FALSE, // sparseResidencyAliased VK_FALSE, // variableMultisampleRate VK_FALSE, // inheritedQueries }; } // ----------------------------------------------------------------------------- // Device VkResult CreateDevice(VkPhysicalDevice physical_device, const VkDeviceCreateInfo* create_info, const VkAllocationCallbacks* allocator, VkDevice* out_device) { VkInstance_T* instance = GetInstanceFromPhysicalDevice(physical_device); if (!allocator) allocator = &instance->allocator; VkDevice_T* device = static_cast<VkDevice_T*>(allocator->pfnAllocation( allocator->pUserData, sizeof(VkDevice_T), alignof(VkDevice_T), VK_SYSTEM_ALLOCATION_SCOPE_DEVICE)); if (!device) return VK_ERROR_OUT_OF_HOST_MEMORY; device->dispatch.magic = HWVULKAN_DISPATCH_MAGIC; device->allocator = *allocator; device->instance = instance; device->queue.dispatch.magic = HWVULKAN_DISPATCH_MAGIC; std::fill(device->next_handle.begin(), device->next_handle.end(), UINT64_C(0)); for (uint32_t i = 0; i < create_info->enabledExtensionCount; i++) { if (strcmp(create_info->ppEnabledExtensionNames[i], VK_ANDROID_NATIVE_BUFFER_EXTENSION_NAME) == 0) { ALOGV("Enabling " VK_ANDROID_NATIVE_BUFFER_EXTENSION_NAME); } } *out_device = device; return VK_SUCCESS; } void DestroyDevice(VkDevice device, const VkAllocationCallbacks* /*allocator*/) { if (!device) return; device->allocator.pfnFree(device->allocator.pUserData, device); } void GetDeviceQueue(VkDevice device, uint32_t, uint32_t, VkQueue* queue) { *queue = &device->queue; } // ----------------------------------------------------------------------------- // CommandPool struct CommandPool { typedef VkCommandPool HandleType; VkAllocationCallbacks allocator; }; DEFINE_OBJECT_HANDLE_CONVERSION(CommandPool) VkResult CreateCommandPool(VkDevice device, const VkCommandPoolCreateInfo* /*create_info*/, const VkAllocationCallbacks* allocator, VkCommandPool* cmd_pool) { if (!allocator) allocator = &device->allocator; CommandPool* pool = static_cast<CommandPool*>(allocator->pfnAllocation( allocator->pUserData, sizeof(CommandPool), alignof(CommandPool), VK_SYSTEM_ALLOCATION_SCOPE_OBJECT)); if (!pool) return VK_ERROR_OUT_OF_HOST_MEMORY; pool->allocator = *allocator; *cmd_pool = GetHandleToCommandPool(pool); return VK_SUCCESS; } void DestroyCommandPool(VkDevice /*device*/, VkCommandPool cmd_pool, const VkAllocationCallbacks* /*allocator*/) { CommandPool* pool = GetCommandPoolFromHandle(cmd_pool); pool->allocator.pfnFree(pool->allocator.pUserData, pool); } // ----------------------------------------------------------------------------- // CmdBuffer VkResult AllocateCommandBuffers(VkDevice /*device*/, const VkCommandBufferAllocateInfo* alloc_info, VkCommandBuffer* cmdbufs) { VkResult result = VK_SUCCESS; CommandPool& pool = *GetCommandPoolFromHandle(alloc_info->commandPool); std::fill(cmdbufs, cmdbufs + alloc_info->commandBufferCount, nullptr); for (uint32_t i = 0; i < alloc_info->commandBufferCount; i++) { cmdbufs[i] = static_cast<VkCommandBuffer_T*>(pool.allocator.pfnAllocation( pool.allocator.pUserData, sizeof(VkCommandBuffer_T), alignof(VkCommandBuffer_T), VK_SYSTEM_ALLOCATION_SCOPE_OBJECT)); if (!cmdbufs[i]) { result = VK_ERROR_OUT_OF_HOST_MEMORY; break; } cmdbufs[i]->dispatch.magic = HWVULKAN_DISPATCH_MAGIC; } if (result != VK_SUCCESS) { for (uint32_t i = 0; i < alloc_info->commandBufferCount; i++) { if (!cmdbufs[i]) break; pool.allocator.pfnFree(pool.allocator.pUserData, cmdbufs[i]); } } return result; } void FreeCommandBuffers(VkDevice /*device*/, VkCommandPool cmd_pool, uint32_t count, const VkCommandBuffer* cmdbufs) { CommandPool& pool = *GetCommandPoolFromHandle(cmd_pool); for (uint32_t i = 0; i < count; i++) pool.allocator.pfnFree(pool.allocator.pUserData, cmdbufs[i]); } // ----------------------------------------------------------------------------- // DeviceMemory struct DeviceMemory { typedef VkDeviceMemory HandleType; VkDeviceSize size; alignas(16) uint8_t data[0]; }; DEFINE_OBJECT_HANDLE_CONVERSION(DeviceMemory) VkResult AllocateMemory(VkDevice device, const VkMemoryAllocateInfo* alloc_info, const VkAllocationCallbacks* allocator, VkDeviceMemory* mem_handle) { if (SIZE_MAX - sizeof(DeviceMemory) <= alloc_info->allocationSize) return VK_ERROR_OUT_OF_HOST_MEMORY; if (!allocator) allocator = &device->allocator; size_t size = sizeof(DeviceMemory) + size_t(alloc_info->allocationSize); DeviceMemory* mem = static_cast<DeviceMemory*>(allocator->pfnAllocation( allocator->pUserData, size, alignof(DeviceMemory), VK_SYSTEM_ALLOCATION_SCOPE_OBJECT)); if (!mem) return VK_ERROR_OUT_OF_HOST_MEMORY; mem->size = size; *mem_handle = GetHandleToDeviceMemory(mem); return VK_SUCCESS; } void FreeMemory(VkDevice device, VkDeviceMemory mem_handle, const VkAllocationCallbacks* allocator) { if (!allocator) allocator = &device->allocator; DeviceMemory* mem = GetDeviceMemoryFromHandle(mem_handle); allocator->pfnFree(allocator->pUserData, mem); } VkResult MapMemory(VkDevice, VkDeviceMemory mem_handle, VkDeviceSize offset, VkDeviceSize, VkMemoryMapFlags, void** out_ptr) { DeviceMemory* mem = GetDeviceMemoryFromHandle(mem_handle); *out_ptr = &mem->data[0] + offset; return VK_SUCCESS; } // ----------------------------------------------------------------------------- // Buffer struct Buffer { typedef VkBuffer HandleType; VkDeviceSize size; }; DEFINE_OBJECT_HANDLE_CONVERSION(Buffer) VkResult CreateBuffer(VkDevice device, const VkBufferCreateInfo* create_info, const VkAllocationCallbacks* allocator, VkBuffer* buffer_handle) { ALOGW_IF(create_info->size > kMaxDeviceMemory, "CreateBuffer: requested size 0x%" PRIx64 " exceeds max device memory size 0x%" PRIx64, create_info->size, kMaxDeviceMemory); if (!allocator) allocator = &device->allocator; Buffer* buffer = static_cast<Buffer*>(allocator->pfnAllocation( allocator->pUserData, sizeof(Buffer), alignof(Buffer), VK_SYSTEM_ALLOCATION_SCOPE_OBJECT)); if (!buffer) return VK_ERROR_OUT_OF_HOST_MEMORY; buffer->size = create_info->size; *buffer_handle = GetHandleToBuffer(buffer); return VK_SUCCESS; } void GetBufferMemoryRequirements(VkDevice, VkBuffer buffer_handle, VkMemoryRequirements* requirements) { Buffer* buffer = GetBufferFromHandle(buffer_handle); requirements->size = buffer->size; requirements->alignment = 16; // allow fast Neon/SSE memcpy requirements->memoryTypeBits = 0x1; } void DestroyBuffer(VkDevice device, VkBuffer buffer_handle, const VkAllocationCallbacks* allocator) { if (!allocator) allocator = &device->allocator; Buffer* buffer = GetBufferFromHandle(buffer_handle); allocator->pfnFree(allocator->pUserData, buffer); } // ----------------------------------------------------------------------------- // Image struct Image { typedef VkImage HandleType; VkDeviceSize size; }; DEFINE_OBJECT_HANDLE_CONVERSION(Image) VkResult CreateImage(VkDevice device, const VkImageCreateInfo* create_info, const VkAllocationCallbacks* allocator, VkImage* image_handle) { if (create_info->imageType != VK_IMAGE_TYPE_2D || create_info->format != VK_FORMAT_R8G8B8A8_UNORM || create_info->mipLevels != 1) { ALOGE("CreateImage: not yet implemented: type=%d format=%d mips=%u", create_info->imageType, create_info->format, create_info->mipLevels); return VK_ERROR_OUT_OF_HOST_MEMORY; } VkDeviceSize size = VkDeviceSize(create_info->extent.width * create_info->extent.height) * create_info->arrayLayers * create_info->samples * 4u; ALOGW_IF(size > kMaxDeviceMemory, "CreateImage: image size 0x%" PRIx64 " exceeds max device memory size 0x%" PRIx64, size, kMaxDeviceMemory); if (!allocator) allocator = &device->allocator; Image* image = static_cast<Image*>(allocator->pfnAllocation( allocator->pUserData, sizeof(Image), alignof(Image), VK_SYSTEM_ALLOCATION_SCOPE_OBJECT)); if (!image) return VK_ERROR_OUT_OF_HOST_MEMORY; image->size = size; *image_handle = GetHandleToImage(image); return VK_SUCCESS; } void GetImageMemoryRequirements(VkDevice, VkImage image_handle, VkMemoryRequirements* requirements) { Image* image = GetImageFromHandle(image_handle); requirements->size = image->size; requirements->alignment = 16; // allow fast Neon/SSE memcpy requirements->memoryTypeBits = 0x1; } void DestroyImage(VkDevice device, VkImage image_handle, const VkAllocationCallbacks* allocator) { if (!allocator) allocator = &device->allocator; Image* image = GetImageFromHandle(image_handle); allocator->pfnFree(allocator->pUserData, image); } VkResult GetSwapchainGrallocUsageANDROID(VkDevice, VkFormat, VkImageUsageFlags, int* grallocUsage) { // The null driver never reads or writes the gralloc buffer *grallocUsage = 0; return VK_SUCCESS; } VkResult AcquireImageANDROID(VkDevice, VkImage, int fence, VkSemaphore, VkFence) { close(fence); return VK_SUCCESS; } VkResult QueueSignalReleaseImageANDROID(VkQueue, uint32_t, const VkSemaphore*, VkImage, int* fence) { *fence = -1; return VK_SUCCESS; } // ----------------------------------------------------------------------------- // No-op types VkResult CreateBufferView(VkDevice device, const VkBufferViewCreateInfo*, const VkAllocationCallbacks* /*allocator*/, VkBufferView* view) { *view = AllocHandle<VkBufferView>(device, HandleType::kBufferView); return VK_SUCCESS; } VkResult CreateDescriptorPool(VkDevice device, const VkDescriptorPoolCreateInfo*, const VkAllocationCallbacks* /*allocator*/, VkDescriptorPool* pool) { *pool = AllocHandle<VkDescriptorPool>(device, HandleType::kDescriptorPool); return VK_SUCCESS; } VkResult AllocateDescriptorSets(VkDevice device, const VkDescriptorSetAllocateInfo* alloc_info, VkDescriptorSet* descriptor_sets) { for (uint32_t i = 0; i < alloc_info->descriptorSetCount; i++) descriptor_sets[i] = AllocHandle<VkDescriptorSet>(device, HandleType::kDescriptorSet); return VK_SUCCESS; } VkResult CreateDescriptorSetLayout(VkDevice device, const VkDescriptorSetLayoutCreateInfo*, const VkAllocationCallbacks* /*allocator*/, VkDescriptorSetLayout* layout) { *layout = AllocHandle<VkDescriptorSetLayout>( device, HandleType::kDescriptorSetLayout); return VK_SUCCESS; } VkResult CreateEvent(VkDevice device, const VkEventCreateInfo*, const VkAllocationCallbacks* /*allocator*/, VkEvent* event) { *event = AllocHandle<VkEvent>(device, HandleType::kEvent); return VK_SUCCESS; } VkResult CreateFence(VkDevice device, const VkFenceCreateInfo*, const VkAllocationCallbacks* /*allocator*/, VkFence* fence) { *fence = AllocHandle<VkFence>(device, HandleType::kFence); return VK_SUCCESS; } VkResult CreateFramebuffer(VkDevice device, const VkFramebufferCreateInfo*, const VkAllocationCallbacks* /*allocator*/, VkFramebuffer* framebuffer) { *framebuffer = AllocHandle<VkFramebuffer>(device, HandleType::kFramebuffer); return VK_SUCCESS; } VkResult CreateImageView(VkDevice device, const VkImageViewCreateInfo*, const VkAllocationCallbacks* /*allocator*/, VkImageView* view) { *view = AllocHandle<VkImageView>(device, HandleType::kImageView); return VK_SUCCESS; } VkResult CreateGraphicsPipelines(VkDevice device, VkPipelineCache, uint32_t count, const VkGraphicsPipelineCreateInfo*, const VkAllocationCallbacks* /*allocator*/, VkPipeline* pipelines) { for (uint32_t i = 0; i < count; i++) pipelines[i] = AllocHandle<VkPipeline>(device, HandleType::kPipeline); return VK_SUCCESS; } VkResult CreateComputePipelines(VkDevice device, VkPipelineCache, uint32_t count, const VkComputePipelineCreateInfo*, const VkAllocationCallbacks* /*allocator*/, VkPipeline* pipelines) { for (uint32_t i = 0; i < count; i++) pipelines[i] = AllocHandle<VkPipeline>(device, HandleType::kPipeline); return VK_SUCCESS; } VkResult CreatePipelineCache(VkDevice device, const VkPipelineCacheCreateInfo*, const VkAllocationCallbacks* /*allocator*/, VkPipelineCache* cache) { *cache = AllocHandle<VkPipelineCache>(device, HandleType::kPipelineCache); return VK_SUCCESS; } VkResult CreatePipelineLayout(VkDevice device, const VkPipelineLayoutCreateInfo*, const VkAllocationCallbacks* /*allocator*/, VkPipelineLayout* layout) { *layout = AllocHandle<VkPipelineLayout>(device, HandleType::kPipelineLayout); return VK_SUCCESS; } VkResult CreateQueryPool(VkDevice device, const VkQueryPoolCreateInfo*, const VkAllocationCallbacks* /*allocator*/, VkQueryPool* pool) { *pool = AllocHandle<VkQueryPool>(device, HandleType::kQueryPool); return VK_SUCCESS; } VkResult CreateRenderPass(VkDevice device, const VkRenderPassCreateInfo*, const VkAllocationCallbacks* /*allocator*/, VkRenderPass* renderpass) { *renderpass = AllocHandle<VkRenderPass>(device, HandleType::kRenderPass); return VK_SUCCESS; } VkResult CreateSampler(VkDevice device, const VkSamplerCreateInfo*, const VkAllocationCallbacks* /*allocator*/, VkSampler* sampler) { *sampler = AllocHandle<VkSampler>(device, HandleType::kSampler); return VK_SUCCESS; } VkResult CreateSemaphore(VkDevice device, const VkSemaphoreCreateInfo*, const VkAllocationCallbacks* /*allocator*/, VkSemaphore* semaphore) { *semaphore = AllocHandle<VkSemaphore>(device, HandleType::kSemaphore); return VK_SUCCESS; } VkResult CreateShaderModule(VkDevice device, const VkShaderModuleCreateInfo*, const VkAllocationCallbacks* /*allocator*/, VkShaderModule* module) { *module = AllocHandle<VkShaderModule>(device, HandleType::kShaderModule); return VK_SUCCESS; } VkResult CreateDebugReportCallbackEXT(VkInstance instance, const VkDebugReportCallbackCreateInfoEXT*, const VkAllocationCallbacks*, VkDebugReportCallbackEXT* callback) { *callback = AllocHandle<VkDebugReportCallbackEXT>( instance, HandleType::kDebugReportCallbackEXT); return VK_SUCCESS; } // ----------------------------------------------------------------------------- // No-op entrypoints // clang-format off #pragma clang diagnostic push #pragma clang diagnostic ignored "-Wunused-parameter" void GetPhysicalDeviceFormatProperties(VkPhysicalDevice physicalDevice, VkFormat format, VkFormatProperties* pFormatProperties) { ALOGV("TODO: vk%s", __FUNCTION__); } VkResult GetPhysicalDeviceImageFormatProperties(VkPhysicalDevice physicalDevice, VkFormat format, VkImageType type, VkImageTiling tiling, VkImageUsageFlags usage, VkImageCreateFlags flags, VkImageFormatProperties* pImageFormatProperties) { ALOGV("TODO: vk%s", __FUNCTION__); return VK_SUCCESS; } VkResult EnumerateInstanceLayerProperties(uint32_t* pCount, VkLayerProperties* pProperties) { ALOGV("TODO: vk%s", __FUNCTION__); return VK_SUCCESS; } VkResult QueueSubmit(VkQueue queue, uint32_t submitCount, const VkSubmitInfo* pSubmitInfo, VkFence fence) { return VK_SUCCESS; } VkResult QueueWaitIdle(VkQueue queue) { ALOGV("TODO: vk%s", __FUNCTION__); return VK_SUCCESS; } VkResult DeviceWaitIdle(VkDevice device) { ALOGV("TODO: vk%s", __FUNCTION__); return VK_SUCCESS; } void UnmapMemory(VkDevice device, VkDeviceMemory mem) { } VkResult FlushMappedMemoryRanges(VkDevice device, uint32_t memRangeCount, const VkMappedMemoryRange* pMemRanges) { ALOGV("TODO: vk%s", __FUNCTION__); return VK_SUCCESS; } VkResult InvalidateMappedMemoryRanges(VkDevice device, uint32_t memRangeCount, const VkMappedMemoryRange* pMemRanges) { ALOGV("TODO: vk%s", __FUNCTION__); return VK_SUCCESS; } void GetDeviceMemoryCommitment(VkDevice device, VkDeviceMemory memory, VkDeviceSize* pCommittedMemoryInBytes) { ALOGV("TODO: vk%s", __FUNCTION__); } VkResult BindBufferMemory(VkDevice device, VkBuffer buffer, VkDeviceMemory mem, VkDeviceSize memOffset) { return VK_SUCCESS; } VkResult BindImageMemory(VkDevice device, VkImage image, VkDeviceMemory mem, VkDeviceSize memOffset) { return VK_SUCCESS; } void GetImageSparseMemoryRequirements(VkDevice device, VkImage image, uint32_t* pNumRequirements, VkSparseImageMemoryRequirements* pSparseMemoryRequirements) { ALOGV("TODO: vk%s", __FUNCTION__); } void GetPhysicalDeviceSparseImageFormatProperties(VkPhysicalDevice physicalDevice, VkFormat format, VkImageType type, VkSampleCountFlagBits samples, VkImageUsageFlags usage, VkImageTiling tiling, uint32_t* pNumProperties, VkSparseImageFormatProperties* pProperties) { ALOGV("TODO: vk%s", __FUNCTION__); } VkResult QueueBindSparse(VkQueue queue, uint32_t bindInfoCount, const VkBindSparseInfo* pBindInfo, VkFence fence) { ALOGV("TODO: vk%s", __FUNCTION__); return VK_SUCCESS; } void DestroyFence(VkDevice device, VkFence fence, const VkAllocationCallbacks* allocator) { } VkResult ResetFences(VkDevice device, uint32_t fenceCount, const VkFence* pFences) { return VK_SUCCESS; } VkResult GetFenceStatus(VkDevice device, VkFence fence) { ALOGV("TODO: vk%s", __FUNCTION__); return VK_SUCCESS; } VkResult WaitForFences(VkDevice device, uint32_t fenceCount, const VkFence* pFences, VkBool32 waitAll, uint64_t timeout) { return VK_SUCCESS; } void DestroySemaphore(VkDevice device, VkSemaphore semaphore, const VkAllocationCallbacks* allocator) { } void DestroyEvent(VkDevice device, VkEvent event, const VkAllocationCallbacks* allocator) { } VkResult GetEventStatus(VkDevice device, VkEvent event) { ALOGV("TODO: vk%s", __FUNCTION__); return VK_SUCCESS; } VkResult SetEvent(VkDevice device, VkEvent event) { ALOGV("TODO: vk%s", __FUNCTION__); return VK_SUCCESS; } VkResult ResetEvent(VkDevice device, VkEvent event) { ALOGV("TODO: vk%s", __FUNCTION__); return VK_SUCCESS; } void DestroyQueryPool(VkDevice device, VkQueryPool queryPool, const VkAllocationCallbacks* allocator) { } VkResult GetQueryPoolResults(VkDevice device, VkQueryPool queryPool, uint32_t startQuery, uint32_t queryCount, size_t dataSize, void* pData, VkDeviceSize stride, VkQueryResultFlags flags) { ALOGV("TODO: vk%s", __FUNCTION__); return VK_SUCCESS; } void DestroyBufferView(VkDevice device, VkBufferView bufferView, const VkAllocationCallbacks* allocator) { } void GetImageSubresourceLayout(VkDevice device, VkImage image, const VkImageSubresource* pSubresource, VkSubresourceLayout* pLayout) { ALOGV("TODO: vk%s", __FUNCTION__); } void DestroyImageView(VkDevice device, VkImageView imageView, const VkAllocationCallbacks* allocator) { } void DestroyShaderModule(VkDevice device, VkShaderModule shaderModule, const VkAllocationCallbacks* allocator) { } void DestroyPipelineCache(VkDevice device, VkPipelineCache pipelineCache, const VkAllocationCallbacks* allocator) { } VkResult GetPipelineCacheData(VkDevice device, VkPipelineCache pipelineCache, size_t* pDataSize, void* pData) { ALOGV("TODO: vk%s", __FUNCTION__); return VK_SUCCESS; } VkResult MergePipelineCaches(VkDevice device, VkPipelineCache destCache, uint32_t srcCacheCount, const VkPipelineCache* pSrcCaches) { ALOGV("TODO: vk%s", __FUNCTION__); return VK_SUCCESS; } void DestroyPipeline(VkDevice device, VkPipeline pipeline, const VkAllocationCallbacks* allocator) { } void DestroyPipelineLayout(VkDevice device, VkPipelineLayout pipelineLayout, const VkAllocationCallbacks* allocator) { } void DestroySampler(VkDevice device, VkSampler sampler, const VkAllocationCallbacks* allocator) { } void DestroyDescriptorSetLayout(VkDevice device, VkDescriptorSetLayout descriptorSetLayout, const VkAllocationCallbacks* allocator) { } void DestroyDescriptorPool(VkDevice device, VkDescriptorPool descriptorPool, const VkAllocationCallbacks* allocator) { } VkResult ResetDescriptorPool(VkDevice device, VkDescriptorPool descriptorPool, VkDescriptorPoolResetFlags flags) { ALOGV("TODO: vk%s", __FUNCTION__); return VK_SUCCESS; } void UpdateDescriptorSets(VkDevice device, uint32_t writeCount, const VkWriteDescriptorSet* pDescriptorWrites, uint32_t copyCount, const VkCopyDescriptorSet* pDescriptorCopies) { ALOGV("TODO: vk%s", __FUNCTION__); } VkResult FreeDescriptorSets(VkDevice device, VkDescriptorPool descriptorPool, uint32_t count, const VkDescriptorSet* pDescriptorSets) { ALOGV("TODO: vk%s", __FUNCTION__); return VK_SUCCESS; } void DestroyFramebuffer(VkDevice device, VkFramebuffer framebuffer, const VkAllocationCallbacks* allocator) { } void DestroyRenderPass(VkDevice device, VkRenderPass renderPass, const VkAllocationCallbacks* allocator) { } void GetRenderAreaGranularity(VkDevice device, VkRenderPass renderPass, VkExtent2D* pGranularity) { ALOGV("TODO: vk%s", __FUNCTION__); } VkResult ResetCommandPool(VkDevice device, VkCommandPool cmdPool, VkCommandPoolResetFlags flags) { ALOGV("TODO: vk%s", __FUNCTION__); return VK_SUCCESS; } VkResult BeginCommandBuffer(VkCommandBuffer cmdBuffer, const VkCommandBufferBeginInfo* pBeginInfo) { return VK_SUCCESS; } VkResult EndCommandBuffer(VkCommandBuffer cmdBuffer) { return VK_SUCCESS; } VkResult ResetCommandBuffer(VkCommandBuffer cmdBuffer, VkCommandBufferResetFlags flags) { ALOGV("TODO: vk%s", __FUNCTION__); return VK_SUCCESS; } void CmdBindPipeline(VkCommandBuffer cmdBuffer, VkPipelineBindPoint pipelineBindPoint, VkPipeline pipeline) { } void CmdSetViewport(VkCommandBuffer cmdBuffer, uint32_t firstViewport, uint32_t viewportCount, const VkViewport* pViewports) { } void CmdSetScissor(VkCommandBuffer cmdBuffer, uint32_t firstScissor, uint32_t scissorCount, const VkRect2D* pScissors) { } void CmdSetLineWidth(VkCommandBuffer cmdBuffer, float lineWidth) { } void CmdSetDepthBias(VkCommandBuffer cmdBuffer, float depthBias, float depthBiasClamp, float slopeScaledDepthBias) { } void CmdSetBlendConstants(VkCommandBuffer cmdBuffer, const float blendConst[4]) { } void CmdSetDepthBounds(VkCommandBuffer cmdBuffer, float minDepthBounds, float maxDepthBounds) { } void CmdSetStencilCompareMask(VkCommandBuffer cmdBuffer, VkStencilFaceFlags faceMask, uint32_t stencilCompareMask) { } void CmdSetStencilWriteMask(VkCommandBuffer cmdBuffer, VkStencilFaceFlags faceMask, uint32_t stencilWriteMask) { } void CmdSetStencilReference(VkCommandBuffer cmdBuffer, VkStencilFaceFlags faceMask, uint32_t stencilReference) { } void CmdBindDescriptorSets(VkCommandBuffer cmdBuffer, VkPipelineBindPoint pipelineBindPoint, VkPipelineLayout layout, uint32_t firstSet, uint32_t setCount, const VkDescriptorSet* pDescriptorSets, uint32_t dynamicOffsetCount, const uint32_t* pDynamicOffsets) { } void CmdBindIndexBuffer(VkCommandBuffer cmdBuffer, VkBuffer buffer, VkDeviceSize offset, VkIndexType indexType) { } void CmdBindVertexBuffers(VkCommandBuffer cmdBuffer, uint32_t startBinding, uint32_t bindingCount, const VkBuffer* pBuffers, const VkDeviceSize* pOffsets) { } void CmdDraw(VkCommandBuffer cmdBuffer, uint32_t vertexCount, uint32_t instanceCount, uint32_t firstVertex, uint32_t firstInstance) { } void CmdDrawIndexed(VkCommandBuffer cmdBuffer, uint32_t indexCount, uint32_t instanceCount, uint32_t firstIndex, int32_t vertexOffset, uint32_t firstInstance) { } void CmdDrawIndirect(VkCommandBuffer cmdBuffer, VkBuffer buffer, VkDeviceSize offset, uint32_t count, uint32_t stride) { } void CmdDrawIndexedIndirect(VkCommandBuffer cmdBuffer, VkBuffer buffer, VkDeviceSize offset, uint32_t count, uint32_t stride) { } void CmdDispatch(VkCommandBuffer cmdBuffer, uint32_t x, uint32_t y, uint32_t z) { } void CmdDispatchIndirect(VkCommandBuffer cmdBuffer, VkBuffer buffer, VkDeviceSize offset) { } void CmdCopyBuffer(VkCommandBuffer cmdBuffer, VkBuffer srcBuffer, VkBuffer destBuffer, uint32_t regionCount, const VkBufferCopy* pRegions) { } void CmdCopyImage(VkCommandBuffer cmdBuffer, VkImage srcImage, VkImageLayout srcImageLayout, VkImage destImage, VkImageLayout destImageLayout, uint32_t regionCount, const VkImageCopy* pRegions) { } void CmdBlitImage(VkCommandBuffer cmdBuffer, VkImage srcImage, VkImageLayout srcImageLayout, VkImage destImage, VkImageLayout destImageLayout, uint32_t regionCount, const VkImageBlit* pRegions, VkFilter filter) { } void CmdCopyBufferToImage(VkCommandBuffer cmdBuffer, VkBuffer srcBuffer, VkImage destImage, VkImageLayout destImageLayout, uint32_t regionCount, const VkBufferImageCopy* pRegions) { } void CmdCopyImageToBuffer(VkCommandBuffer cmdBuffer, VkImage srcImage, VkImageLayout srcImageLayout, VkBuffer destBuffer, uint32_t regionCount, const VkBufferImageCopy* pRegions) { } void CmdUpdateBuffer(VkCommandBuffer cmdBuffer, VkBuffer destBuffer, VkDeviceSize destOffset, VkDeviceSize dataSize, const uint32_t* pData) { } void CmdFillBuffer(VkCommandBuffer cmdBuffer, VkBuffer destBuffer, VkDeviceSize destOffset, VkDeviceSize fillSize, uint32_t data) { } void CmdClearColorImage(VkCommandBuffer cmdBuffer, VkImage image, VkImageLayout imageLayout, const VkClearColorValue* pColor, uint32_t rangeCount, const VkImageSubresourceRange* pRanges) { } void CmdClearDepthStencilImage(VkCommandBuffer cmdBuffer, VkImage image, VkImageLayout imageLayout, const VkClearDepthStencilValue* pDepthStencil, uint32_t rangeCount, const VkImageSubresourceRange* pRanges) { } void CmdClearAttachments(VkCommandBuffer cmdBuffer, uint32_t attachmentCount, const VkClearAttachment* pAttachments, uint32_t rectCount, const VkClearRect* pRects) { } void CmdResolveImage(VkCommandBuffer cmdBuffer, VkImage srcImage, VkImageLayout srcImageLayout, VkImage destImage, VkImageLayout destImageLayout, uint32_t regionCount, const VkImageResolve* pRegions) { } void CmdSetEvent(VkCommandBuffer cmdBuffer, VkEvent event, VkPipelineStageFlags stageMask) { } void CmdResetEvent(VkCommandBuffer cmdBuffer, VkEvent event, VkPipelineStageFlags stageMask) { } void CmdWaitEvents(VkCommandBuffer commandBuffer, uint32_t eventCount, const VkEvent* pEvents, VkPipelineStageFlags srcStageMask, VkPipelineStageFlags dstStageMask, uint32_t memoryBarrierCount, const VkMemoryBarrier* pMemoryBarriers, uint32_t bufferMemoryBarrierCount, const VkBufferMemoryBarrier* pBufferMemoryBarriers, uint32_t imageMemoryBarrierCount, const VkImageMemoryBarrier* pImageMemoryBarriers) { } void CmdPipelineBarrier(VkCommandBuffer commandBuffer, VkPipelineStageFlags srcStageMask, VkPipelineStageFlags dstStageMask, VkDependencyFlags dependencyFlags, uint32_t memoryBarrierCount, const VkMemoryBarrier* pMemoryBarriers, uint32_t bufferMemoryBarrierCount, const VkBufferMemoryBarrier* pBufferMemoryBarriers, uint32_t imageMemoryBarrierCount, const VkImageMemoryBarrier* pImageMemoryBarriers) { } void CmdBeginQuery(VkCommandBuffer cmdBuffer, VkQueryPool queryPool, uint32_t slot, VkQueryControlFlags flags) { } void CmdEndQuery(VkCommandBuffer cmdBuffer, VkQueryPool queryPool, uint32_t slot) { } void CmdResetQueryPool(VkCommandBuffer cmdBuffer, VkQueryPool queryPool, uint32_t startQuery, uint32_t queryCount) { } void CmdWriteTimestamp(VkCommandBuffer cmdBuffer, VkPipelineStageFlagBits pipelineStage, VkQueryPool queryPool, uint32_t slot) { } void CmdCopyQueryPoolResults(VkCommandBuffer cmdBuffer, VkQueryPool queryPool, uint32_t startQuery, uint32_t queryCount, VkBuffer destBuffer, VkDeviceSize destOffset, VkDeviceSize destStride, VkQueryResultFlags flags) { } void CmdPushConstants(VkCommandBuffer cmdBuffer, VkPipelineLayout layout, VkShaderStageFlags stageFlags, uint32_t start, uint32_t length, const void* values) { } void CmdBeginRenderPass(VkCommandBuffer cmdBuffer, const VkRenderPassBeginInfo* pRenderPassBegin, VkSubpassContents contents) { } void CmdNextSubpass(VkCommandBuffer cmdBuffer, VkSubpassContents contents) { } void CmdEndRenderPass(VkCommandBuffer cmdBuffer) { } void CmdExecuteCommands(VkCommandBuffer cmdBuffer, uint32_t cmdBuffersCount, const VkCommandBuffer* pCmdBuffers) { } void DestroyDebugReportCallbackEXT(VkInstance instance, VkDebugReportCallbackEXT callback, const VkAllocationCallbacks* pAllocator) { } void DebugReportMessageEXT(VkInstance instance, VkDebugReportFlagsEXT flags, VkDebugReportObjectTypeEXT objectType, uint64_t object, size_t location, int32_t messageCode, const char* pLayerPrefix, const char* pMessage) { } #pragma clang diagnostic pop // clang-format on } // namespace null_driver