Refresh/src/vulkan/Refresh_vulkan.c

13079 lines
424 KiB
C

/* Refresh - a cross-platform hardware-accelerated graphics library with modern capabilities
*
* Copyright (c) 2020-2024 Evan Hemsley
*
* This software is provided 'as-is', without any express or implied warranty.
* In no event will the authors be held liable for any damages arising from
* the use of this software.
*
* Permission is granted to anyone to use this software for any purpose,
* including commercial applications, and to alter it and redistribute it
* freely, subject to the following restrictions:
*
* 1. The origin of this software must not be misrepresented; you must not
* claim that you wrote the original software. If you use this software in a
* product, an acknowledgment in the product documentation would be
* appreciated but is not required.
*
* 2. Altered source versions must be plainly marked as such, and must not be
* misrepresented as being the original software.
*
* 3. This notice may not be removed or altered from any source distribution.
*
* Evan "cosmonaut" Hemsley <evan@moonside.games>
*
*/
#if REFRESH_VULKAN
/* Needed for VK_KHR_portability_subset */
#define VK_ENABLE_BETA_EXTENSIONS
#define VK_NO_PROTOTYPES
#include <vulkan/vulkan.h>
#include <SDL.h>
#include <SDL_vulkan.h>
#include <SDL_video.h>
#include "Refresh_driver.h"
#define VULKAN_INTERNAL_clamp(val, min, max) SDL_max(min, SDL_min(val, max))
/* Global Vulkan Loader Entry Points */
static PFN_vkGetInstanceProcAddr vkGetInstanceProcAddr = NULL;
#define VULKAN_GLOBAL_FUNCTION(name) \
static PFN_##name name = NULL;
#include "Refresh_vulkan_vkfuncs.h"
/* vkInstance/vkDevice function typedefs */
#define VULKAN_INSTANCE_FUNCTION(ext, ret, func, params) \
typedef ret (VKAPI_CALL *vkfntype_##func) params;
#define VULKAN_DEVICE_FUNCTION(ext, ret, func, params) \
typedef ret (VKAPI_CALL *vkfntype_##func) params;
#include "Refresh_vulkan_vkfuncs.h"
typedef struct VulkanExtensions
{
/* Globally supported */
Uint8 KHR_swapchain;
/* Core since 1.1 */
Uint8 KHR_maintenance1;
Uint8 KHR_get_memory_requirements2;
/* Core since 1.2 */
Uint8 KHR_driver_properties;
/* EXT, probably not going to be Core */
Uint8 EXT_vertex_attribute_divisor;
/* Only required for special implementations (i.e. MoltenVK) */
Uint8 KHR_portability_subset;
} VulkanExtensions;
/* Defines */
#define SMALL_ALLOCATION_THRESHOLD 2097152 /* 2 MiB */
#define SMALL_ALLOCATION_SIZE 16777216 /* 16 MiB */
#define LARGE_ALLOCATION_INCREMENT 67108864 /* 64 MiB */
#define MAX_UBO_SECTION_SIZE 4096 /* 4 KiB */
#define UNIFORM_BUFFER_SIZE 1048576 /* 1 MiB */
#define DESCRIPTOR_POOL_STARTING_SIZE 128
#define MAX_QUERIES 16
#define WINDOW_PROPERTY_DATA "Refresh_VulkanWindowPropertyData"
#define IDENTITY_SWIZZLE \
{ \
VK_COMPONENT_SWIZZLE_IDENTITY, \
VK_COMPONENT_SWIZZLE_IDENTITY, \
VK_COMPONENT_SWIZZLE_IDENTITY, \
VK_COMPONENT_SWIZZLE_IDENTITY \
}
#define NULL_DESC_LAYOUT (VkDescriptorSetLayout) 0
#define NULL_PIPELINE_LAYOUT (VkPipelineLayout) 0
#define NULL_RENDER_PASS (Refresh_RenderPass*) 0
#define EXPAND_ELEMENTS_IF_NEEDED(arr, initialValue, type) \
if (arr->count == arr->capacity) \
{ \
if (arr->capacity == 0) \
{ \
arr->capacity = initialValue; \
} \
else \
{ \
arr->capacity *= 2; \
} \
arr->elements = (type*) SDL_realloc( \
arr->elements, \
arr->capacity * sizeof(type) \
); \
}
#define EXPAND_ARRAY_IF_NEEDED(arr, elementType, newCount, capacity, newCapacity) \
if (newCount >= capacity) \
{ \
capacity = newCapacity; \
arr = (elementType*) SDL_realloc( \
arr, \
sizeof(elementType) * capacity \
); \
}
#define MOVE_ARRAY_CONTENTS_AND_RESET(i, dstArr, dstCount, srcArr, srcCount) \
for (i = 0; i < srcCount; i += 1) \
{ \
dstArr[i] = srcArr[i]; \
} \
dstCount = srcCount; \
srcCount = 0;
/* Conversions */
static const Uint8 DEVICE_PRIORITY[] =
{
0, /* VK_PHYSICAL_DEVICE_TYPE_OTHER */
3, /* VK_PHYSICAL_DEVICE_TYPE_INTEGRATED_GPU */
4, /* VK_PHYSICAL_DEVICE_TYPE_DISCRETE_GPU */
2, /* VK_PHYSICAL_DEVICE_TYPE_VIRTUAL_GPU */
1 /* VK_PHYSICAL_DEVICE_TYPE_CPU */
};
static VkPresentModeKHR SDLToVK_PresentMode[] =
{
VK_PRESENT_MODE_FIFO_KHR,
VK_PRESENT_MODE_IMMEDIATE_KHR,
VK_PRESENT_MODE_MAILBOX_KHR
};
static VkFormat SDLToVK_SurfaceFormat[] =
{
VK_FORMAT_R8G8B8A8_UNORM, /* R8G8B8A8 */
VK_FORMAT_B8G8R8A8_UNORM, /* B8G8R8A8 */
VK_FORMAT_R5G6B5_UNORM_PACK16, /* R5G6B5 */
VK_FORMAT_A1R5G5B5_UNORM_PACK16, /* A1R5G5B5 */
VK_FORMAT_B4G4R4A4_UNORM_PACK16, /* B4G4R4A4 */
VK_FORMAT_A2R10G10B10_UNORM_PACK32, /* A2R10G10B10 */
VK_FORMAT_A2B10G10R10_UNORM_PACK32, /* A2R10G10B10 */
VK_FORMAT_R16G16_UNORM, /* R16G16 */
VK_FORMAT_R16G16B16A16_UNORM, /* R16G16B16A16 */
VK_FORMAT_R8_UNORM, /* R8 */
VK_FORMAT_R8_UNORM, /* A8 */
VK_FORMAT_BC1_RGBA_UNORM_BLOCK, /* BC1 */
VK_FORMAT_BC2_UNORM_BLOCK, /* BC2 */
VK_FORMAT_BC3_UNORM_BLOCK, /* BC3 */
VK_FORMAT_BC7_UNORM_BLOCK, /* BC7 */
VK_FORMAT_R8G8_SNORM, /* R8G8_SNORM */
VK_FORMAT_R8G8B8A8_SNORM, /* R8G8B8A8_SNORM */
VK_FORMAT_R16_SFLOAT, /* R16_SFLOAT */
VK_FORMAT_R16G16_SFLOAT, /* R16G16_SFLOAT */
VK_FORMAT_R16G16B16A16_SFLOAT, /* R16G16B16A16_SFLOAT */
VK_FORMAT_R32_SFLOAT, /* R32_SFLOAT */
VK_FORMAT_R32G32_SFLOAT, /* R32G32_SFLOAT */
VK_FORMAT_R32G32B32A32_SFLOAT, /* R32G32B32A32_SFLOAT */
VK_FORMAT_R8_UINT, /* R8_UINT */
VK_FORMAT_R8G8_UINT, /* R8G8_UINT */
VK_FORMAT_R8G8B8A8_UINT, /* R8G8B8A8_UINT */
VK_FORMAT_R16_UINT, /* R16_UINT */
VK_FORMAT_R16G16_UINT, /* R16G16_UINT */
VK_FORMAT_R16G16B16A16_UINT, /* R16G16B16A16_UINT */
VK_FORMAT_R8G8B8A8_SRGB, /* R8G8B8A8_SRGB */
VK_FORMAT_B8G8R8A8_SRGB, /* B8G8R8A8_SRGB */
VK_FORMAT_BC3_SRGB_BLOCK, /* BC3_SRGB */
VK_FORMAT_BC7_SRGB_BLOCK, /* BC7_SRGB */
VK_FORMAT_D16_UNORM, /* D16_UNORM */
VK_FORMAT_X8_D24_UNORM_PACK32, /* D24_UNORM */
VK_FORMAT_D32_SFLOAT, /* D32_SFLOAT */
VK_FORMAT_D24_UNORM_S8_UINT, /* D24_UNORM_S8_UINT */
VK_FORMAT_D32_SFLOAT_S8_UINT, /* D32_SFLOAT_S8_UINT */
};
/* from SWAPCHAINCOMPOSITION */
static VkFormat SwapchainCompositionToFormat[] =
{
VK_FORMAT_B8G8R8A8_UNORM, /* SDR */
VK_FORMAT_B8G8R8A8_SRGB, /* SDR_SRGB */
VK_FORMAT_R16G16B16A16_SFLOAT, /* HDR */
VK_FORMAT_A2B10G10R10_UNORM_PACK32 /* HDR_ADVANCED*/
};
/* from SWAPCHAINCOMPOSITION */
static VkColorSpaceKHR SwapchainCompositionToColorSpace[] =
{
VK_COLOR_SPACE_SRGB_NONLINEAR_KHR,
VK_COLOR_SPACE_SRGB_NONLINEAR_KHR,
VK_COLOR_SPACE_EXTENDED_SRGB_LINEAR_EXT,
VK_COLOR_SPACE_HDR10_ST2084_EXT
};
static VkFormat SDLToVK_VertexFormat[] =
{
VK_FORMAT_R32_UINT, /* UINT */
VK_FORMAT_R32_SFLOAT, /* FLOAT */
VK_FORMAT_R32G32_SFLOAT, /* VECTOR2 */
VK_FORMAT_R32G32B32_SFLOAT, /* VECTOR3 */
VK_FORMAT_R32G32B32A32_SFLOAT, /* VECTOR4 */
VK_FORMAT_R8G8B8A8_UNORM, /* COLOR */
VK_FORMAT_R8G8B8A8_USCALED, /* BYTE4 */
VK_FORMAT_R16G16_SSCALED, /* SHORT2 */
VK_FORMAT_R16G16B16A16_SSCALED, /* SHORT4 */
VK_FORMAT_R16G16_SNORM, /* NORMALIZEDSHORT2 */
VK_FORMAT_R16G16B16A16_SNORM, /* NORMALIZEDSHORT4 */
VK_FORMAT_R16G16_SFLOAT, /* HALFVECTOR2 */
VK_FORMAT_R16G16B16A16_SFLOAT /* HALFVECTOR4 */
};
static VkIndexType SDLToVK_IndexType[] =
{
VK_INDEX_TYPE_UINT16,
VK_INDEX_TYPE_UINT32
};
static VkPrimitiveTopology SDLToVK_PrimitiveType[] =
{
VK_PRIMITIVE_TOPOLOGY_POINT_LIST,
VK_PRIMITIVE_TOPOLOGY_LINE_LIST,
VK_PRIMITIVE_TOPOLOGY_LINE_STRIP,
VK_PRIMITIVE_TOPOLOGY_TRIANGLE_LIST,
VK_PRIMITIVE_TOPOLOGY_TRIANGLE_STRIP
};
static VkPolygonMode SDLToVK_PolygonMode[] =
{
VK_POLYGON_MODE_FILL,
VK_POLYGON_MODE_LINE,
VK_POLYGON_MODE_POINT
};
static VkCullModeFlags SDLToVK_CullMode[] =
{
VK_CULL_MODE_NONE,
VK_CULL_MODE_FRONT_BIT,
VK_CULL_MODE_BACK_BIT,
VK_CULL_MODE_FRONT_AND_BACK
};
static VkFrontFace SDLToVK_FrontFace[] =
{
VK_FRONT_FACE_COUNTER_CLOCKWISE,
VK_FRONT_FACE_CLOCKWISE
};
static VkBlendFactor SDLToVK_BlendFactor[] =
{
VK_BLEND_FACTOR_ZERO,
VK_BLEND_FACTOR_ONE,
VK_BLEND_FACTOR_SRC_COLOR,
VK_BLEND_FACTOR_ONE_MINUS_SRC_COLOR,
VK_BLEND_FACTOR_DST_COLOR,
VK_BLEND_FACTOR_ONE_MINUS_DST_COLOR,
VK_BLEND_FACTOR_SRC_ALPHA,
VK_BLEND_FACTOR_ONE_MINUS_SRC_ALPHA,
VK_BLEND_FACTOR_DST_ALPHA,
VK_BLEND_FACTOR_ONE_MINUS_DST_ALPHA,
VK_BLEND_FACTOR_CONSTANT_COLOR,
VK_BLEND_FACTOR_ONE_MINUS_CONSTANT_COLOR,
VK_BLEND_FACTOR_CONSTANT_ALPHA,
VK_BLEND_FACTOR_ONE_MINUS_CONSTANT_ALPHA,
VK_BLEND_FACTOR_SRC_ALPHA_SATURATE
};
static VkBlendOp SDLToVK_BlendOp[] =
{
VK_BLEND_OP_ADD,
VK_BLEND_OP_SUBTRACT,
VK_BLEND_OP_REVERSE_SUBTRACT,
VK_BLEND_OP_MIN,
VK_BLEND_OP_MAX
};
static VkCompareOp SDLToVK_CompareOp[] =
{
VK_COMPARE_OP_NEVER,
VK_COMPARE_OP_LESS,
VK_COMPARE_OP_EQUAL,
VK_COMPARE_OP_LESS_OR_EQUAL,
VK_COMPARE_OP_GREATER,
VK_COMPARE_OP_NOT_EQUAL,
VK_COMPARE_OP_GREATER_OR_EQUAL,
VK_COMPARE_OP_ALWAYS
};
static VkStencilOp SDLToVK_StencilOp[] =
{
VK_STENCIL_OP_KEEP,
VK_STENCIL_OP_ZERO,
VK_STENCIL_OP_REPLACE,
VK_STENCIL_OP_INCREMENT_AND_CLAMP,
VK_STENCIL_OP_DECREMENT_AND_CLAMP,
VK_STENCIL_OP_INVERT,
VK_STENCIL_OP_INCREMENT_AND_WRAP,
VK_STENCIL_OP_DECREMENT_AND_WRAP
};
static VkAttachmentLoadOp SDLToVK_LoadOp[] =
{
VK_ATTACHMENT_LOAD_OP_LOAD,
VK_ATTACHMENT_LOAD_OP_CLEAR,
VK_ATTACHMENT_LOAD_OP_DONT_CARE
};
static VkAttachmentStoreOp SDLToVK_StoreOp[] =
{
VK_ATTACHMENT_STORE_OP_STORE,
VK_ATTACHMENT_STORE_OP_DONT_CARE
};
static VkSampleCountFlagBits SDLToVK_SampleCount[] =
{
VK_SAMPLE_COUNT_1_BIT,
VK_SAMPLE_COUNT_2_BIT,
VK_SAMPLE_COUNT_4_BIT,
VK_SAMPLE_COUNT_8_BIT,
VK_SAMPLE_COUNT_16_BIT,
VK_SAMPLE_COUNT_32_BIT,
VK_SAMPLE_COUNT_64_BIT
};
static VkVertexInputRate SDLToVK_VertexInputRate[] =
{
VK_VERTEX_INPUT_RATE_VERTEX,
VK_VERTEX_INPUT_RATE_INSTANCE
};
static VkFilter SDLToVK_Filter[] =
{
VK_FILTER_NEAREST,
VK_FILTER_LINEAR
};
static VkSamplerMipmapMode SDLToVK_SamplerMipmapMode[] =
{
VK_SAMPLER_MIPMAP_MODE_NEAREST,
VK_SAMPLER_MIPMAP_MODE_LINEAR
};
static VkSamplerAddressMode SDLToVK_SamplerAddressMode[] =
{
VK_SAMPLER_ADDRESS_MODE_REPEAT,
VK_SAMPLER_ADDRESS_MODE_MIRRORED_REPEAT,
VK_SAMPLER_ADDRESS_MODE_CLAMP_TO_EDGE,
VK_SAMPLER_ADDRESS_MODE_CLAMP_TO_BORDER
};
static VkBorderColor SDLToVK_BorderColor[] =
{
VK_BORDER_COLOR_FLOAT_TRANSPARENT_BLACK,
VK_BORDER_COLOR_INT_TRANSPARENT_BLACK,
VK_BORDER_COLOR_FLOAT_OPAQUE_BLACK,
VK_BORDER_COLOR_INT_OPAQUE_BLACK,
VK_BORDER_COLOR_FLOAT_OPAQUE_WHITE,
VK_BORDER_COLOR_INT_OPAQUE_WHITE
};
/* Structures */
typedef struct VulkanMemoryAllocation VulkanMemoryAllocation;
typedef struct VulkanBuffer VulkanBuffer;
typedef struct VulkanBufferContainer VulkanBufferContainer;
typedef struct VulkanTexture VulkanTexture;
typedef struct VulkanTextureContainer VulkanTextureContainer;
typedef struct VulkanFenceHandle
{
VkFence fence;
SDL_atomic_t referenceCount;
} VulkanFenceHandle;
/* Memory Allocation */
typedef struct VulkanMemoryFreeRegion
{
VulkanMemoryAllocation *allocation;
VkDeviceSize offset;
VkDeviceSize size;
Uint32 allocationIndex;
Uint32 sortedIndex;
} VulkanMemoryFreeRegion;
typedef struct VulkanMemoryUsedRegion
{
VulkanMemoryAllocation *allocation;
VkDeviceSize offset;
VkDeviceSize size;
VkDeviceSize resourceOffset; /* differs from offset based on alignment*/
VkDeviceSize resourceSize; /* differs from size based on alignment */
VkDeviceSize alignment;
Uint8 isBuffer;
union
{
VulkanBuffer *vulkanBuffer;
VulkanTexture *vulkanTexture;
};
} VulkanMemoryUsedRegion;
typedef struct VulkanMemorySubAllocator
{
Uint32 memoryTypeIndex;
VulkanMemoryAllocation **allocations;
Uint32 allocationCount;
VulkanMemoryFreeRegion **sortedFreeRegions;
Uint32 sortedFreeRegionCount;
Uint32 sortedFreeRegionCapacity;
} VulkanMemorySubAllocator;
struct VulkanMemoryAllocation
{
VulkanMemorySubAllocator *allocator;
VkDeviceMemory memory;
VkDeviceSize size;
VulkanMemoryUsedRegion **usedRegions;
Uint32 usedRegionCount;
Uint32 usedRegionCapacity;
VulkanMemoryFreeRegion **freeRegions;
Uint32 freeRegionCount;
Uint32 freeRegionCapacity;
Uint8 availableForAllocation;
VkDeviceSize freeSpace;
VkDeviceSize usedSpace;
Uint8 *mapPointer;
SDL_mutex *memoryLock;
};
typedef struct VulkanMemoryAllocator
{
VulkanMemorySubAllocator subAllocators[VK_MAX_MEMORY_TYPES];
} VulkanMemoryAllocator;
/* Memory structures */
/* We use pointer indirection so that defrag can occur without objects
* needing to be aware of the backing buffers changing.
*/
typedef struct VulkanBufferHandle
{
VulkanBuffer *vulkanBuffer;
VulkanBufferContainer *container;
} VulkanBufferHandle;
typedef enum VulkanBufferType
{
VULKAN_BUFFER_TYPE_GPU,
VULKAN_BUFFER_TYPE_UNIFORM,
VULKAN_BUFFER_TYPE_TRANSFER
} VulkanBufferType;
struct VulkanBuffer
{
VkBuffer buffer;
VkDeviceSize size;
VulkanMemoryUsedRegion *usedRegion;
VulkanBufferType type;
Refresh_BufferUsageFlags usageFlags;
SDL_atomic_t referenceCount; /* Tracks command buffer usage */
VulkanBufferHandle *handle;
SDL_bool transitioned;
Uint8 markedForDestroy; /* so that defrag doesn't double-free */
};
/* Buffer resources consist of multiple backing buffer handles so that data transfers
* can occur without blocking or the client having to manage extra resources.
*
* Cast from Refresh_Buffer or Refresh_TransferBuffer.
*/
struct VulkanBufferContainer
{
VulkanBufferHandle *activeBufferHandle;
/* These are all the buffer handles that have been used by this container.
* If the resource is bound and then updated with a cycle parameter, a new resource
* will be added to this list.
* These can be reused after they are submitted and command processing is complete.
*/
Uint32 bufferCapacity;
Uint32 bufferCount;
VulkanBufferHandle **bufferHandles;
char *debugName;
};
/* Renderer Structure */
typedef struct QueueFamilyIndices
{
Uint32 graphicsFamily;
Uint32 presentFamily;
Uint32 computeFamily;
Uint32 transferFamily;
} QueueFamilyIndices;
typedef struct VulkanSampler
{
VkSampler sampler;
SDL_atomic_t referenceCount;
} VulkanSampler;
typedef struct VulkanShader
{
VkShaderModule shaderModule;
const char *entryPointName;
SDL_atomic_t referenceCount;
} VulkanShader;
typedef struct VulkanTextureHandle
{
VulkanTexture *vulkanTexture;
VulkanTextureContainer *container;
} VulkanTextureHandle;
/* Textures are made up of individual slices.
* This helps us barrier the resource efficiently.
*/
typedef struct VulkanTextureSlice
{
VulkanTexture *parent;
Uint32 layer;
Uint32 level;
SDL_atomic_t referenceCount;
VkImageView view;
VulkanTextureHandle *msaaTexHandle; /* NULL if parent sample count is 1 or is depth target */
SDL_bool transitioned; /* used for layout tracking */
} VulkanTextureSlice;
struct VulkanTexture
{
VulkanMemoryUsedRegion *usedRegion;
VkImage image;
VkImageView view;
VkExtent2D dimensions;
Uint8 is3D;
Uint8 isCube;
Uint8 isRenderTarget;
Uint32 depth;
Uint32 layerCount;
Uint32 levelCount;
VkSampleCountFlagBits sampleCount; /* NOTE: This refers to the sample count of a render target pass using this texture, not the actual sample count of the texture */
VkFormat format;
VkComponentMapping swizzle;
Refresh_TextureUsageFlags usageFlags;
VkImageAspectFlags aspectFlags;
Uint32 sliceCount;
VulkanTextureSlice *slices;
VulkanTextureHandle *handle;
Uint8 markedForDestroy; /* so that defrag doesn't double-free */
};
/* Texture resources consist of multiple backing texture handles so that data transfers
* can occur without blocking or the client having to manage extra resources.
*
* Cast from Refresh_Texture.
*/
struct VulkanTextureContainer
{
VulkanTextureHandle *activeTextureHandle;
/* These are all the texture handles that have been used by this container.
* If the resource is bound and then updated with CYCLE, a new resource
* will be added to this list.
* These can be reused after they are submitted and command processing is complete.
*/
Uint32 textureCapacity;
Uint32 textureCount;
VulkanTextureHandle **textureHandles;
/* Swapchain images cannot be cycled */
Uint8 canBeCycled;
char *debugName;
};
typedef enum VulkanBufferUsageMode
{
VULKAN_BUFFER_USAGE_MODE_COPY_SOURCE,
VULKAN_BUFFER_USAGE_MODE_COPY_DESTINATION,
VULKAN_BUFFER_USAGE_MODE_VERTEX_READ,
VULKAN_BUFFER_USAGE_MODE_INDEX_READ,
VULKAN_BUFFER_USAGE_MODE_INDIRECT,
VULKAN_BUFFER_USAGE_MODE_GRAPHICS_STORAGE_READ,
VULKAN_BUFFER_USAGE_MODE_COMPUTE_STORAGE_READ,
VULKAN_BUFFER_USAGE_MODE_COMPUTE_STORAGE_READ_WRITE,
} VulkanBufferUsageMode;
typedef enum VulkanTextureUsageMode
{
VULKAN_TEXTURE_USAGE_MODE_COPY_SOURCE,
VULKAN_TEXTURE_USAGE_MODE_COPY_DESTINATION,
VULKAN_TEXTURE_USAGE_MODE_SAMPLER,
VULKAN_TEXTURE_USAGE_MODE_GRAPHICS_STORAGE_READ,
VULKAN_TEXTURE_USAGE_MODE_COMPUTE_STORAGE_READ,
VULKAN_TEXTURE_USAGE_MODE_COMPUTE_STORAGE_READ_WRITE,
VULKAN_TEXTURE_USAGE_MODE_COLOR_ATTACHMENT,
VULKAN_TEXTURE_USAGE_MODE_DEPTH_STENCIL_ATTACHMENT,
VULKAN_TEXTURE_USAGE_MODE_PRESENT
} VulkanTextureUsageMode;
typedef struct VulkanFramebuffer
{
VkFramebuffer framebuffer;
SDL_atomic_t referenceCount;
} VulkanFramebuffer;
typedef struct VulkanSwapchainData
{
/* Window surface */
VkSurfaceKHR surface;
/* Swapchain for window surface */
VkSwapchainKHR swapchain;
VkFormat format;
VkColorSpaceKHR colorSpace;
VkComponentMapping swapchainSwizzle;
VkPresentModeKHR presentMode;
/* Swapchain images */
VkExtent2D extent;
VulkanTextureContainer *textureContainers; /* use containers so that swapchain textures can use the same API as other textures */
Uint32 imageCount;
/* Synchronization primitives */
VkSemaphore imageAvailableSemaphore[MAX_FRAMES_IN_FLIGHT];
VkSemaphore renderFinishedSemaphore[MAX_FRAMES_IN_FLIGHT];
VulkanFenceHandle *inFlightFences[MAX_FRAMES_IN_FLIGHT];
Uint32 frameCounter;
} VulkanSwapchainData;
typedef struct WindowData
{
SDL_Window *window;
Refresh_SwapchainComposition swapchainComposition;
Refresh_PresentMode presentMode;
VulkanSwapchainData *swapchainData;
} WindowData;
typedef struct SwapchainSupportDetails
{
VkSurfaceCapabilitiesKHR capabilities;
VkSurfaceFormatKHR *formats;
Uint32 formatsLength;
VkPresentModeKHR *presentModes;
Uint32 presentModesLength;
} SwapchainSupportDetails;
typedef struct VulkanPresentData
{
WindowData *windowData;
Uint32 swapchainImageIndex;
} VulkanPresentData;
typedef struct VulkanUniformBuffer
{
VulkanBufferContainer *bufferContainer;
Uint32 drawOffset;
Uint32 offset;
} VulkanUniformBuffer;
typedef struct VulkanDescriptorInfo
{
VkDescriptorType descriptorType;
VkShaderStageFlagBits stageFlag;
} VulkanDescriptorInfo;
typedef struct DescriptorSetPool
{
SDL_mutex *lock;
VkDescriptorSetLayout descriptorSetLayout;
VulkanDescriptorInfo *descriptorInfos;
Uint32 descriptorInfoCount;
/* This is actually a descriptor set and descriptor pool simultaneously */
VkDescriptorPool *descriptorPools;
Uint32 descriptorPoolCount;
Uint32 nextPoolSize;
/* We just manage a pool ourselves instead of freeing the sets */
VkDescriptorSet *inactiveDescriptorSets;
Uint32 inactiveDescriptorSetCount;
Uint32 inactiveDescriptorSetCapacity;
} DescriptorSetPool;
typedef struct VulkanGraphicsPipelineResourceLayout
{
VkPipelineLayout pipelineLayout;
/*
* Descriptor set layout is as follows:
* 0: vertex resources
* 1: vertex uniform buffers
* 2: fragment resources
* 3: fragment uniform buffers
*/
DescriptorSetPool descriptorSetPools[4];
Uint32 vertexSamplerCount;
Uint32 vertexStorageBufferCount;
Uint32 vertexStorageTextureCount;
Uint32 vertexUniformBufferCount;
Uint32 fragmentSamplerCount;
Uint32 fragmentStorageBufferCount;
Uint32 fragmentStorageTextureCount;
Uint32 fragmentUniformBufferCount;
} VulkanGraphicsPipelineResourceLayout;
typedef struct VulkanGraphicsPipeline
{
VkPipeline pipeline;
Refresh_PrimitiveType primitiveType;
VulkanGraphicsPipelineResourceLayout resourceLayout;
VulkanShader *vertexShader;
VulkanShader *fragmentShader;
SDL_atomic_t referenceCount;
} VulkanGraphicsPipeline;
typedef struct VulkanComputePipelineResourceLayout
{
VkPipelineLayout pipelineLayout;
/*
* Descriptor set layout is as follows:
* 0: read-only textures, then read-only buffers
* 1: read-write textures, then read-write buffers
* 2: uniform buffers
*/
DescriptorSetPool descriptorSetPools[3];
Uint32 readOnlyStorageTextureCount;
Uint32 readOnlyStorageBufferCount;
Uint32 readWriteStorageTextureCount;
Uint32 readWriteStorageBufferCount;
Uint32 uniformBufferCount;
} VulkanComputePipelineResourceLayout;
typedef struct VulkanComputePipeline
{
VkPipeline pipeline;
VulkanComputePipelineResourceLayout resourceLayout;
VulkanShader *computeShader;
SDL_atomic_t referenceCount;
} VulkanComputePipeline;
typedef struct VulkanOcclusionQuery
{
Uint32 index;
} VulkanOcclusionQuery;
typedef struct RenderPassColorTargetDescription
{
VkFormat format;
Refresh_Color clearColor;
Refresh_LoadOp loadOp;
Refresh_StoreOp storeOp;
} RenderPassColorTargetDescription;
typedef struct RenderPassDepthStencilTargetDescription
{
VkFormat format;
Refresh_LoadOp loadOp;
Refresh_StoreOp storeOp;
Refresh_LoadOp stencilLoadOp;
Refresh_StoreOp stencilStoreOp;
} RenderPassDepthStencilTargetDescription;
typedef struct RenderPassHash
{
RenderPassColorTargetDescription colorTargetDescriptions[MAX_COLOR_TARGET_BINDINGS];
Uint32 colorAttachmentCount;
RenderPassDepthStencilTargetDescription depthStencilTargetDescription;
VkSampleCountFlagBits colorAttachmentSampleCount;
} RenderPassHash;
typedef struct RenderPassHashMap
{
RenderPassHash key;
VkRenderPass value;
} RenderPassHashMap;
typedef struct RenderPassHashArray
{
RenderPassHashMap *elements;
Sint32 count;
Sint32 capacity;
} RenderPassHashArray;
static inline Uint8 RenderPassHash_Compare(
RenderPassHash *a,
RenderPassHash *b
) {
Uint32 i;
if (a->colorAttachmentCount != b->colorAttachmentCount)
{
return 0;
}
if (a->colorAttachmentSampleCount != b->colorAttachmentSampleCount)
{
return 0;
}
for (i = 0; i < a->colorAttachmentCount; i += 1)
{
if (a->colorTargetDescriptions[i].format != b->colorTargetDescriptions[i].format)
{
return 0;
}
if ( a->colorTargetDescriptions[i].clearColor.r != b->colorTargetDescriptions[i].clearColor.r ||
a->colorTargetDescriptions[i].clearColor.g != b->colorTargetDescriptions[i].clearColor.g ||
a->colorTargetDescriptions[i].clearColor.b != b->colorTargetDescriptions[i].clearColor.b ||
a->colorTargetDescriptions[i].clearColor.a != b->colorTargetDescriptions[i].clearColor.a )
{
return 0;
}
if (a->colorTargetDescriptions[i].loadOp != b->colorTargetDescriptions[i].loadOp)
{
return 0;
}
if (a->colorTargetDescriptions[i].storeOp != b->colorTargetDescriptions[i].storeOp)
{
return 0;
}
}
if (a->depthStencilTargetDescription.format != b->depthStencilTargetDescription.format)
{
return 0;
}
if (a->depthStencilTargetDescription.loadOp != b->depthStencilTargetDescription.loadOp)
{
return 0;
}
if (a->depthStencilTargetDescription.storeOp != b->depthStencilTargetDescription.storeOp)
{
return 0;
}
if (a->depthStencilTargetDescription.stencilLoadOp != b->depthStencilTargetDescription.stencilLoadOp)
{
return 0;
}
if (a->depthStencilTargetDescription.stencilStoreOp != b->depthStencilTargetDescription.stencilStoreOp)
{
return 0;
}
return 1;
}
static inline VkRenderPass RenderPassHashArray_Fetch(
RenderPassHashArray *arr,
RenderPassHash *key
) {
Sint32 i;
for (i = 0; i < arr->count; i += 1)
{
RenderPassHash *e = &arr->elements[i].key;
if (RenderPassHash_Compare(e, key))
{
return arr->elements[i].value;
}
}
return VK_NULL_HANDLE;
}
static inline void RenderPassHashArray_Insert(
RenderPassHashArray *arr,
RenderPassHash key,
VkRenderPass value
) {
RenderPassHashMap map;
map.key = key;
map.value = value;
EXPAND_ELEMENTS_IF_NEEDED(arr, 4, RenderPassHashMap)
arr->elements[arr->count] = map;
arr->count += 1;
}
typedef struct FramebufferHash
{
VkImageView colorAttachmentViews[MAX_COLOR_TARGET_BINDINGS];
VkImageView colorMultiSampleAttachmentViews[MAX_COLOR_TARGET_BINDINGS];
Uint32 colorAttachmentCount;
VkImageView depthStencilAttachmentView;
Uint32 width;
Uint32 height;
} FramebufferHash;
typedef struct FramebufferHashMap
{
FramebufferHash key;
VulkanFramebuffer *value;
} FramebufferHashMap;
typedef struct FramebufferHashArray
{
FramebufferHashMap *elements;
Sint32 count;
Sint32 capacity;
} FramebufferHashArray;
static inline Uint8 FramebufferHash_Compare(
FramebufferHash *a,
FramebufferHash *b
) {
Uint32 i;
if (a->colorAttachmentCount != b->colorAttachmentCount)
{
return 0;
}
for (i = 0; i < a->colorAttachmentCount; i += 1)
{
if (a->colorAttachmentViews[i] != b->colorAttachmentViews[i])
{
return 0;
}
if (a->colorMultiSampleAttachmentViews[i] != b->colorMultiSampleAttachmentViews[i])
{
return 0;
}
}
if (a->depthStencilAttachmentView != b->depthStencilAttachmentView)
{
return 0;
}
if (a->width != b->width)
{
return 0;
}
if (a->height != b->height)
{
return 0;
}
return 1;
}
static inline VulkanFramebuffer* FramebufferHashArray_Fetch(
FramebufferHashArray *arr,
FramebufferHash *key
) {
Sint32 i;
for (i = 0; i < arr->count; i += 1)
{
FramebufferHash *e = &arr->elements[i].key;
if (FramebufferHash_Compare(e, key))
{
return arr->elements[i].value;
}
}
return VK_NULL_HANDLE;
}
static inline void FramebufferHashArray_Insert(
FramebufferHashArray *arr,
FramebufferHash key,
VulkanFramebuffer *value
) {
FramebufferHashMap map;
map.key = key;
map.value = value;
EXPAND_ELEMENTS_IF_NEEDED(arr, 4, FramebufferHashMap)
arr->elements[arr->count] = map;
arr->count += 1;
}
static inline void FramebufferHashArray_Remove(
FramebufferHashArray *arr,
Uint32 index
) {
if (index != arr->count - 1)
{
arr->elements[index] = arr->elements[arr->count - 1];
}
arr->count -= 1;
}
/* Command structures */
typedef struct DescriptorSetData
{
DescriptorSetPool *descriptorSetPool;
VkDescriptorSet descriptorSet;
} DescriptorSetData;
typedef struct VulkanFencePool
{
SDL_mutex *lock;
VulkanFenceHandle **availableFences;
Uint32 availableFenceCount;
Uint32 availableFenceCapacity;
} VulkanFencePool;
typedef struct VulkanCommandPool VulkanCommandPool;
typedef struct VulkanRenderer VulkanRenderer;
typedef struct VulkanCommandBuffer
{
CommandBufferCommonHeader common;
VulkanRenderer *renderer;
VkCommandBuffer commandBuffer;
VulkanCommandPool *commandPool;
VulkanPresentData *presentDatas;
Uint32 presentDataCount;
Uint32 presentDataCapacity;
VkSemaphore *waitSemaphores;
Uint32 waitSemaphoreCount;
Uint32 waitSemaphoreCapacity;
VkSemaphore *signalSemaphores;
Uint32 signalSemaphoreCount;
Uint32 signalSemaphoreCapacity;
VulkanComputePipeline *currentComputePipeline;
VulkanGraphicsPipeline *currentGraphicsPipeline;
/* Keep track of resources transitioned away from their default state to barrier them on pass end */
VulkanTextureSlice *colorAttachmentSlices[MAX_COLOR_TARGET_BINDINGS];
Uint32 colorAttachmentSliceCount;
VulkanTextureSlice *depthStencilAttachmentSlice; /* may be NULL */
/* Viewport/scissor state */
VkViewport currentViewport;
VkRect2D currentScissor;
/* Resource bind state */
SDL_bool needNewVertexResourceDescriptorSet;
SDL_bool needNewVertexUniformDescriptorSet;
SDL_bool needNewVertexUniformOffsets;
SDL_bool needNewFragmentResourceDescriptorSet;
SDL_bool needNewFragmentUniformDescriptorSet;
SDL_bool needNewFragmentUniformOffsets;
SDL_bool needNewComputeReadOnlyDescriptorSet;
SDL_bool needNewComputeReadWriteDescriptorSet;
SDL_bool needNewComputeUniformDescriptorSet;
SDL_bool needNewComputeUniformOffsets;
VkDescriptorSet vertexResourceDescriptorSet;
VkDescriptorSet vertexUniformDescriptorSet;
VkDescriptorSet fragmentResourceDescriptorSet;
VkDescriptorSet fragmentUniformDescriptorSet;
VkDescriptorSet computeReadOnlyDescriptorSet;
VkDescriptorSet computeReadWriteDescriptorSet;
VkDescriptorSet computeUniformDescriptorSet;
DescriptorSetData *boundDescriptorSetDatas;
Uint32 boundDescriptorSetDataCount;
Uint32 boundDescriptorSetDataCapacity;
VulkanTexture *vertexSamplerTextures[MAX_TEXTURE_SAMPLERS_PER_STAGE];
VulkanSampler *vertexSamplers[MAX_TEXTURE_SAMPLERS_PER_STAGE];
VulkanTextureSlice *vertexStorageTextureSlices[MAX_STORAGE_TEXTURES_PER_STAGE];
VulkanBuffer *vertexStorageBuffers[MAX_STORAGE_BUFFERS_PER_STAGE];
VulkanTexture *fragmentSamplerTextures[MAX_TEXTURE_SAMPLERS_PER_STAGE];
VulkanSampler *fragmentSamplers[MAX_TEXTURE_SAMPLERS_PER_STAGE];
VulkanTextureSlice *fragmentStorageTextureSlices[MAX_STORAGE_TEXTURES_PER_STAGE];
VulkanBuffer *fragmentStorageBuffers[MAX_STORAGE_BUFFERS_PER_STAGE];
VulkanTextureSlice *readWriteComputeStorageTextureSlices[MAX_STORAGE_TEXTURES_PER_STAGE];
VulkanBuffer *readWriteComputeStorageBuffers[MAX_STORAGE_BUFFERS_PER_STAGE];
VulkanTextureSlice *readOnlyComputeStorageTextureSlices[MAX_STORAGE_TEXTURES_PER_STAGE];
VulkanBuffer *readOnlyComputeStorageBuffers[MAX_STORAGE_BUFFERS_PER_STAGE];
/* Uniform buffers */
VulkanUniformBuffer *vertexUniformBuffers[MAX_UNIFORM_BUFFERS_PER_STAGE];
VulkanUniformBuffer *fragmentUniformBuffers[MAX_UNIFORM_BUFFERS_PER_STAGE];
VulkanUniformBuffer *computeUniformBuffers[MAX_UNIFORM_BUFFERS_PER_STAGE];
Uint32 initializedVertexUniformBufferCount;
Uint32 initializedFragmentUniformBufferCount;
Uint32 initializedComputeUniformBufferCount;
/* Track used resources */
VulkanBuffer **usedBuffers;
Uint32 usedBufferCount;
Uint32 usedBufferCapacity;
VulkanTextureSlice **usedTextureSlices;
Uint32 usedTextureSliceCount;
Uint32 usedTextureSliceCapacity;
VulkanSampler **usedSamplers;
Uint32 usedSamplerCount;
Uint32 usedSamplerCapacity;
VulkanGraphicsPipeline **usedGraphicsPipelines;
Uint32 usedGraphicsPipelineCount;
Uint32 usedGraphicsPipelineCapacity;
VulkanComputePipeline **usedComputePipelines;
Uint32 usedComputePipelineCount;
Uint32 usedComputePipelineCapacity;
VulkanFramebuffer **usedFramebuffers;
Uint32 usedFramebufferCount;
Uint32 usedFramebufferCapacity;
VulkanFenceHandle *inFlightFence;
Uint8 autoReleaseFence;
Uint8 isDefrag; /* Whether this CB was created for defragging */
} VulkanCommandBuffer;
struct VulkanCommandPool
{
SDL_threadID threadID;
VkCommandPool commandPool;
VulkanCommandBuffer **inactiveCommandBuffers;
Uint32 inactiveCommandBufferCapacity;
Uint32 inactiveCommandBufferCount;
};
#define NUM_COMMAND_POOL_BUCKETS 1031
typedef struct CommandPoolHash
{
SDL_threadID threadID;
} CommandPoolHash;
typedef struct CommandPoolHashMap
{
CommandPoolHash key;
VulkanCommandPool *value;
} CommandPoolHashMap;
typedef struct CommandPoolHashArray
{
CommandPoolHashMap *elements;
Uint32 count;
Uint32 capacity;
} CommandPoolHashArray;
typedef struct CommandPoolHashTable
{
CommandPoolHashArray buckets[NUM_COMMAND_POOL_BUCKETS];
} CommandPoolHashTable;
static inline uint64_t CommandPoolHashTable_GetHashCode(CommandPoolHash key)
{
const uint64_t HASH_FACTOR = 97;
uint64_t result = 1;
result = result * HASH_FACTOR + (uint64_t) key.threadID;
return result;
}
static inline VulkanCommandPool* CommandPoolHashTable_Fetch(
CommandPoolHashTable *table,
CommandPoolHash key
) {
Uint32 i;
uint64_t hashcode = CommandPoolHashTable_GetHashCode(key);
CommandPoolHashArray *arr = &table->buckets[hashcode % NUM_COMMAND_POOL_BUCKETS];
for (i = 0; i < arr->count; i += 1)
{
const CommandPoolHash *e = &arr->elements[i].key;
if (key.threadID == e->threadID)
{
return arr->elements[i].value;
}
}
return NULL;
}
static inline void CommandPoolHashTable_Insert(
CommandPoolHashTable *table,
CommandPoolHash key,
VulkanCommandPool *value
) {
uint64_t hashcode = CommandPoolHashTable_GetHashCode(key);
CommandPoolHashArray *arr = &table->buckets[hashcode % NUM_COMMAND_POOL_BUCKETS];
CommandPoolHashMap map;
map.key = key;
map.value = value;
EXPAND_ELEMENTS_IF_NEEDED(arr, 4, CommandPoolHashMap)
arr->elements[arr->count] = map;
arr->count += 1;
}
/* Context */
struct VulkanRenderer
{
VkInstance instance;
VkPhysicalDevice physicalDevice;
VkPhysicalDeviceProperties2 physicalDeviceProperties;
VkPhysicalDeviceDriverPropertiesKHR physicalDeviceDriverProperties;
VkDevice logicalDevice;
Uint8 integratedMemoryNotification;
Uint8 outOfDeviceLocalMemoryWarning;
Uint8 outofBARMemoryWarning;
Uint8 supportsDebugUtils;
Uint8 debugMode;
VulkanExtensions supports;
VulkanMemoryAllocator *memoryAllocator;
VkPhysicalDeviceMemoryProperties memoryProperties;
WindowData **claimedWindows;
Uint32 claimedWindowCount;
Uint32 claimedWindowCapacity;
Uint32 queueFamilyIndex;
VkQueue unifiedQueue;
VulkanCommandBuffer **submittedCommandBuffers;
Uint32 submittedCommandBufferCount;
Uint32 submittedCommandBufferCapacity;
VulkanFencePool fencePool;
CommandPoolHashTable commandPoolHashTable;
RenderPassHashArray renderPassHashArray;
FramebufferHashArray framebufferHashArray;
Uint32 minUBOAlignment;
/* Some drivers don't support D16 for some reason. Fun! */
VkFormat D16Format;
VkFormat D16S8Format;
/* Queries */
VkQueryPool queryPool;
Sint8 freeQueryIndexStack[MAX_QUERIES];
Sint8 freeQueryIndexStackHead;
/* Deferred resource destruction */
VulkanTexture **texturesToDestroy;
Uint32 texturesToDestroyCount;
Uint32 texturesToDestroyCapacity;
VulkanBuffer **buffersToDestroy;
Uint32 buffersToDestroyCount;
Uint32 buffersToDestroyCapacity;
VulkanSampler **samplersToDestroy;
Uint32 samplersToDestroyCount;
Uint32 samplersToDestroyCapacity;
VulkanGraphicsPipeline **graphicsPipelinesToDestroy;
Uint32 graphicsPipelinesToDestroyCount;
Uint32 graphicsPipelinesToDestroyCapacity;
VulkanComputePipeline **computePipelinesToDestroy;
Uint32 computePipelinesToDestroyCount;
Uint32 computePipelinesToDestroyCapacity;
VulkanShader **shadersToDestroy;
Uint32 shadersToDestroyCount;
Uint32 shadersToDestroyCapacity;
VulkanFramebuffer **framebuffersToDestroy;
Uint32 framebuffersToDestroyCount;
Uint32 framebuffersToDestroyCapacity;
SDL_mutex *allocatorLock;
SDL_mutex *disposeLock;
SDL_mutex *submitLock;
SDL_mutex *acquireCommandBufferLock;
SDL_mutex *renderPassFetchLock;
SDL_mutex *framebufferFetchLock;
SDL_mutex *queryLock;
Uint8 defragInProgress;
VulkanMemoryAllocation **allocationsToDefrag;
Uint32 allocationsToDefragCount;
Uint32 allocationsToDefragCapacity;
/* Support checks */
SDL_bool supportsPreciseOcclusionQueries;
#define VULKAN_INSTANCE_FUNCTION(ext, ret, func, params) \
vkfntype_##func func;
#define VULKAN_DEVICE_FUNCTION(ext, ret, func, params) \
vkfntype_##func func;
#include "Refresh_vulkan_vkfuncs.h"
};
/* Forward declarations */
static Uint8 VULKAN_INTERNAL_DefragmentMemory(VulkanRenderer *renderer);
static void VULKAN_INTERNAL_BeginCommandBuffer(VulkanRenderer *renderer, VulkanCommandBuffer *commandBuffer);
static void VULKAN_UnclaimWindow(Refresh_Renderer *driverData, SDL_Window *window);
static void VULKAN_Wait(Refresh_Renderer *driverData);
static void VULKAN_WaitForFences(Refresh_Renderer *driverData, SDL_bool waitAll, Uint32 fenceCount, Refresh_Fence **pFences);
static void VULKAN_Submit(Refresh_CommandBuffer *commandBuffer);
static VulkanTextureSlice* VULKAN_INTERNAL_FetchTextureSlice(VulkanTexture* texture, Uint32 layer, Uint32 level);
static VulkanTexture* VULKAN_INTERNAL_CreateTexture(
VulkanRenderer *renderer,
Uint32 width,
Uint32 height,
Uint32 depth,
Uint32 isCube,
Uint32 layerCount,
Uint32 levelCount,
VkSampleCountFlagBits sampleCount,
VkFormat format,
VkComponentMapping swizzle,
VkImageAspectFlags aspectMask,
Refresh_TextureUsageFlags textureUsageFlags
);
/* Error Handling */
static inline const char* VkErrorMessages(VkResult code)
{
#define ERR_TO_STR(e) \
case e: return #e;
switch (code)
{
ERR_TO_STR(VK_ERROR_OUT_OF_HOST_MEMORY)
ERR_TO_STR(VK_ERROR_OUT_OF_DEVICE_MEMORY)
ERR_TO_STR(VK_ERROR_FRAGMENTED_POOL)
ERR_TO_STR(VK_ERROR_OUT_OF_POOL_MEMORY)
ERR_TO_STR(VK_ERROR_INITIALIZATION_FAILED)
ERR_TO_STR(VK_ERROR_LAYER_NOT_PRESENT)
ERR_TO_STR(VK_ERROR_EXTENSION_NOT_PRESENT)
ERR_TO_STR(VK_ERROR_FEATURE_NOT_PRESENT)
ERR_TO_STR(VK_ERROR_TOO_MANY_OBJECTS)
ERR_TO_STR(VK_ERROR_DEVICE_LOST)
ERR_TO_STR(VK_ERROR_INCOMPATIBLE_DRIVER)
ERR_TO_STR(VK_ERROR_OUT_OF_DATE_KHR)
ERR_TO_STR(VK_ERROR_SURFACE_LOST_KHR)
ERR_TO_STR(VK_ERROR_FULL_SCREEN_EXCLUSIVE_MODE_LOST_EXT)
ERR_TO_STR(VK_SUBOPTIMAL_KHR)
default: return "Unhandled VkResult!";
}
#undef ERR_TO_STR
}
static inline void LogVulkanResultAsError(
const char* vulkanFunctionName,
VkResult result
) {
if (result != VK_SUCCESS)
{
SDL_LogError(
SDL_LOG_CATEGORY_APPLICATION,
"%s: %s",
vulkanFunctionName,
VkErrorMessages(result)
);
}
}
#define VULKAN_ERROR_CHECK(res, fn, ret) \
if (res != VK_SUCCESS) \
{ \
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "%s %s", #fn, VkErrorMessages(res)); \
return ret; \
}
/* Utility */
static inline SDL_bool VULKAN_INTERNAL_IsVulkanDepthFormat(VkFormat format)
{
/* FIXME: Can we refactor and use the regular IsDepthFormat for this? */
return (
format == SDLToVK_SurfaceFormat[REFRESH_TEXTUREFORMAT_D16_UNORM] ||
format == SDLToVK_SurfaceFormat[REFRESH_TEXTUREFORMAT_D24_UNORM] ||
format == SDLToVK_SurfaceFormat[REFRESH_TEXTUREFORMAT_D24_UNORM_S8_UINT] ||
format == SDLToVK_SurfaceFormat[REFRESH_TEXTUREFORMAT_D32_SFLOAT] ||
format == SDLToVK_SurfaceFormat[REFRESH_TEXTUREFORMAT_D32_SFLOAT_S8_UINT]
);
}
static inline VkSampleCountFlagBits VULKAN_INTERNAL_GetMaxMultiSampleCount(
VulkanRenderer *renderer,
VkSampleCountFlagBits multiSampleCount
) {
VkSampleCountFlags flags = renderer->physicalDeviceProperties.properties.limits.framebufferColorSampleCounts;
VkSampleCountFlagBits maxSupported = VK_SAMPLE_COUNT_1_BIT;
if (flags & VK_SAMPLE_COUNT_8_BIT)
{
maxSupported = VK_SAMPLE_COUNT_8_BIT;
}
else if (flags & VK_SAMPLE_COUNT_4_BIT)
{
maxSupported = VK_SAMPLE_COUNT_4_BIT;
}
else if (flags & VK_SAMPLE_COUNT_2_BIT)
{
maxSupported = VK_SAMPLE_COUNT_2_BIT;
}
return SDL_min(multiSampleCount, maxSupported);
}
static inline SDL_bool BGRToRGBSwapchainFormat(
VkFormat format,
VkFormat *outputFormat
) {
switch (format)
{
case VK_FORMAT_B8G8R8A8_UNORM:
*outputFormat = VK_FORMAT_R8G8B8A8_UNORM;
return SDL_TRUE;
case VK_FORMAT_A2B10G10R10_UNORM_PACK32:
*outputFormat = VK_FORMAT_A2R10G10B10_UNORM_PACK32;
return SDL_TRUE;
case VK_FORMAT_B8G8R8A8_SRGB:
*outputFormat = VK_FORMAT_R8G8B8A8_SRGB;
return SDL_TRUE;
default:
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "No valid fallback equivalent for this format exists!");
*outputFormat = VK_FORMAT_R8G8B8A8_UNORM;
return SDL_FALSE;
}
}
/* Memory Management */
/* Vulkan: Memory Allocation */
static inline VkDeviceSize VULKAN_INTERNAL_NextHighestAlignment(
VkDeviceSize n,
VkDeviceSize align
) {
return align * ((n + align - 1) / align);
}
static inline Uint32 VULKAN_INTERNAL_NextHighestAlignment32(
Uint32 n,
Uint32 align
) {
return align * ((n + align - 1) / align);
}
static void VULKAN_INTERNAL_MakeMemoryUnavailable(
VulkanRenderer* renderer,
VulkanMemoryAllocation *allocation
) {
Uint32 i, j;
VulkanMemoryFreeRegion *freeRegion;
allocation->availableForAllocation = 0;
for (i = 0; i < allocation->freeRegionCount; i += 1)
{
freeRegion = allocation->freeRegions[i];
/* close the gap in the sorted list */
if (allocation->allocator->sortedFreeRegionCount > 1)
{
for (j = freeRegion->sortedIndex; j < allocation->allocator->sortedFreeRegionCount - 1; j += 1)
{
allocation->allocator->sortedFreeRegions[j] =
allocation->allocator->sortedFreeRegions[j + 1];
allocation->allocator->sortedFreeRegions[j]->sortedIndex = j;
}
}
allocation->allocator->sortedFreeRegionCount -= 1;
}
}
static void VULKAN_INTERNAL_MarkAllocationsForDefrag(
VulkanRenderer *renderer
) {
Uint32 memoryType, allocationIndex;
VulkanMemorySubAllocator *currentAllocator;
for (memoryType = 0; memoryType < VK_MAX_MEMORY_TYPES; memoryType += 1)
{
currentAllocator = &renderer->memoryAllocator->subAllocators[memoryType];
for (allocationIndex = 0; allocationIndex < currentAllocator->allocationCount; allocationIndex += 1)
{
if (currentAllocator->allocations[allocationIndex]->availableForAllocation == 1)
{
if (currentAllocator->allocations[allocationIndex]->freeRegionCount > 1)
{
EXPAND_ARRAY_IF_NEEDED(
renderer->allocationsToDefrag,
VulkanMemoryAllocation*,
renderer->allocationsToDefragCount + 1,
renderer->allocationsToDefragCapacity,
renderer->allocationsToDefragCapacity * 2
);
renderer->allocationsToDefrag[renderer->allocationsToDefragCount] =
currentAllocator->allocations[allocationIndex];
renderer->allocationsToDefragCount += 1;
VULKAN_INTERNAL_MakeMemoryUnavailable(
renderer,
currentAllocator->allocations[allocationIndex]
);
}
}
}
}
}
static void VULKAN_INTERNAL_RemoveMemoryFreeRegion(
VulkanRenderer *renderer,
VulkanMemoryFreeRegion *freeRegion
) {
Uint32 i;
SDL_LockMutex(renderer->allocatorLock);
if (freeRegion->allocation->availableForAllocation)
{
/* close the gap in the sorted list */
if (freeRegion->allocation->allocator->sortedFreeRegionCount > 1)
{
for (i = freeRegion->sortedIndex; i < freeRegion->allocation->allocator->sortedFreeRegionCount - 1; i += 1)
{
freeRegion->allocation->allocator->sortedFreeRegions[i] =
freeRegion->allocation->allocator->sortedFreeRegions[i + 1];
freeRegion->allocation->allocator->sortedFreeRegions[i]->sortedIndex = i;
}
}
freeRegion->allocation->allocator->sortedFreeRegionCount -= 1;
}
/* close the gap in the buffer list */
if (freeRegion->allocation->freeRegionCount > 1 && freeRegion->allocationIndex != freeRegion->allocation->freeRegionCount - 1)
{
freeRegion->allocation->freeRegions[freeRegion->allocationIndex] =
freeRegion->allocation->freeRegions[freeRegion->allocation->freeRegionCount - 1];
freeRegion->allocation->freeRegions[freeRegion->allocationIndex]->allocationIndex =
freeRegion->allocationIndex;
}
freeRegion->allocation->freeRegionCount -= 1;
freeRegion->allocation->freeSpace -= freeRegion->size;
SDL_free(freeRegion);
SDL_UnlockMutex(renderer->allocatorLock);
}
static void VULKAN_INTERNAL_NewMemoryFreeRegion(
VulkanRenderer *renderer,
VulkanMemoryAllocation *allocation,
VkDeviceSize offset,
VkDeviceSize size
) {
VulkanMemoryFreeRegion *newFreeRegion;
VkDeviceSize newOffset, newSize;
Sint32 insertionIndex = 0;
Sint32 i;
SDL_LockMutex(renderer->allocatorLock);
/* look for an adjacent region to merge */
for (i = allocation->freeRegionCount - 1; i >= 0; i -= 1)
{
/* check left side */
if (allocation->freeRegions[i]->offset + allocation->freeRegions[i]->size == offset)
{
newOffset = allocation->freeRegions[i]->offset;
newSize = allocation->freeRegions[i]->size + size;
VULKAN_INTERNAL_RemoveMemoryFreeRegion(renderer, allocation->freeRegions[i]);
VULKAN_INTERNAL_NewMemoryFreeRegion(renderer, allocation, newOffset, newSize);
SDL_UnlockMutex(renderer->allocatorLock);
return;
}
/* check right side */
if (allocation->freeRegions[i]->offset == offset + size)
{
newOffset = offset;
newSize = allocation->freeRegions[i]->size + size;
VULKAN_INTERNAL_RemoveMemoryFreeRegion(renderer, allocation->freeRegions[i]);
VULKAN_INTERNAL_NewMemoryFreeRegion(renderer, allocation, newOffset, newSize);
SDL_UnlockMutex(renderer->allocatorLock);
return;
}
}
/* region is not contiguous with another free region, make a new one */
allocation->freeRegionCount += 1;
if (allocation->freeRegionCount > allocation->freeRegionCapacity)
{
allocation->freeRegionCapacity *= 2;
allocation->freeRegions = SDL_realloc(
allocation->freeRegions,
sizeof(VulkanMemoryFreeRegion*) * allocation->freeRegionCapacity
);
}
newFreeRegion = SDL_malloc(sizeof(VulkanMemoryFreeRegion));
newFreeRegion->offset = offset;
newFreeRegion->size = size;
newFreeRegion->allocation = allocation;
allocation->freeSpace += size;
allocation->freeRegions[allocation->freeRegionCount - 1] = newFreeRegion;
newFreeRegion->allocationIndex = allocation->freeRegionCount - 1;
if (allocation->availableForAllocation)
{
for (i = 0; i < allocation->allocator->sortedFreeRegionCount; i += 1)
{
if (allocation->allocator->sortedFreeRegions[i]->size < size)
{
/* this is where the new region should go */
break;
}
insertionIndex += 1;
}
if (allocation->allocator->sortedFreeRegionCount + 1 > allocation->allocator->sortedFreeRegionCapacity)
{
allocation->allocator->sortedFreeRegionCapacity *= 2;
allocation->allocator->sortedFreeRegions = SDL_realloc(
allocation->allocator->sortedFreeRegions,
sizeof(VulkanMemoryFreeRegion*) * allocation->allocator->sortedFreeRegionCapacity
);
}
/* perform insertion sort */
if (allocation->allocator->sortedFreeRegionCount > 0 && insertionIndex != allocation->allocator->sortedFreeRegionCount)
{
for (i = allocation->allocator->sortedFreeRegionCount; i > insertionIndex && i > 0; i -= 1)
{
allocation->allocator->sortedFreeRegions[i] = allocation->allocator->sortedFreeRegions[i - 1];
allocation->allocator->sortedFreeRegions[i]->sortedIndex = i;
}
}
allocation->allocator->sortedFreeRegionCount += 1;
allocation->allocator->sortedFreeRegions[insertionIndex] = newFreeRegion;
newFreeRegion->sortedIndex = insertionIndex;
}
SDL_UnlockMutex(renderer->allocatorLock);
}
static VulkanMemoryUsedRegion* VULKAN_INTERNAL_NewMemoryUsedRegion(
VulkanRenderer *renderer,
VulkanMemoryAllocation *allocation,
VkDeviceSize offset,
VkDeviceSize size,
VkDeviceSize resourceOffset,
VkDeviceSize resourceSize,
VkDeviceSize alignment
) {
VulkanMemoryUsedRegion *memoryUsedRegion;
SDL_LockMutex(renderer->allocatorLock);
if (allocation->usedRegionCount == allocation->usedRegionCapacity)
{
allocation->usedRegionCapacity *= 2;
allocation->usedRegions = SDL_realloc(
allocation->usedRegions,
allocation->usedRegionCapacity * sizeof(VulkanMemoryUsedRegion*)
);
}
memoryUsedRegion = SDL_malloc(sizeof(VulkanMemoryUsedRegion));
memoryUsedRegion->allocation = allocation;
memoryUsedRegion->offset = offset;
memoryUsedRegion->size = size;
memoryUsedRegion->resourceOffset = resourceOffset;
memoryUsedRegion->resourceSize = resourceSize;
memoryUsedRegion->alignment = alignment;
allocation->usedSpace += size;
allocation->usedRegions[allocation->usedRegionCount] = memoryUsedRegion;
allocation->usedRegionCount += 1;
SDL_UnlockMutex(renderer->allocatorLock);
return memoryUsedRegion;
}
static void VULKAN_INTERNAL_RemoveMemoryUsedRegion(
VulkanRenderer *renderer,
VulkanMemoryUsedRegion *usedRegion
) {
Uint32 i;
SDL_LockMutex(renderer->allocatorLock);
for (i = 0; i < usedRegion->allocation->usedRegionCount; i += 1)
{
if (usedRegion->allocation->usedRegions[i] == usedRegion)
{
/* plug the hole */
if (i != usedRegion->allocation->usedRegionCount - 1)
{
usedRegion->allocation->usedRegions[i] = usedRegion->allocation->usedRegions[usedRegion->allocation->usedRegionCount - 1];
}
break;
}
}
usedRegion->allocation->usedSpace -= usedRegion->size;
usedRegion->allocation->usedRegionCount -= 1;
VULKAN_INTERNAL_NewMemoryFreeRegion(
renderer,
usedRegion->allocation,
usedRegion->offset,
usedRegion->size
);
SDL_free(usedRegion);
SDL_UnlockMutex(renderer->allocatorLock);
}
static Uint8 VULKAN_INTERNAL_FindMemoryType(
VulkanRenderer *renderer,
Uint32 typeFilter,
VkMemoryPropertyFlags requiredProperties,
VkMemoryPropertyFlags ignoredProperties,
Uint32 *memoryTypeIndex
) {
Uint32 i;
for (i = *memoryTypeIndex; i < renderer->memoryProperties.memoryTypeCount; i += 1)
{
if ( (typeFilter & (1 << i)) &&
(renderer->memoryProperties.memoryTypes[i].propertyFlags & requiredProperties) == requiredProperties &&
(renderer->memoryProperties.memoryTypes[i].propertyFlags & ignoredProperties) == 0 )
{
*memoryTypeIndex = i;
return 1;
}
}
return 0;
}
static Uint8 VULKAN_INTERNAL_FindBufferMemoryRequirements(
VulkanRenderer *renderer,
VkBuffer buffer,
VkMemoryPropertyFlags requiredMemoryProperties,
VkMemoryPropertyFlags ignoredMemoryProperties,
VkMemoryRequirements2KHR *pMemoryRequirements,
Uint32 *pMemoryTypeIndex
) {
VkBufferMemoryRequirementsInfo2KHR bufferRequirementsInfo;
bufferRequirementsInfo.sType =
VK_STRUCTURE_TYPE_BUFFER_MEMORY_REQUIREMENTS_INFO_2_KHR;
bufferRequirementsInfo.pNext = NULL;
bufferRequirementsInfo.buffer = buffer;
renderer->vkGetBufferMemoryRequirements2KHR(
renderer->logicalDevice,
&bufferRequirementsInfo,
pMemoryRequirements
);
return VULKAN_INTERNAL_FindMemoryType(
renderer,
pMemoryRequirements->memoryRequirements.memoryTypeBits,
requiredMemoryProperties,
ignoredMemoryProperties,
pMemoryTypeIndex
);
}
static Uint8 VULKAN_INTERNAL_FindImageMemoryRequirements(
VulkanRenderer *renderer,
VkImage image,
VkMemoryPropertyFlags requiredMemoryPropertyFlags,
VkMemoryRequirements2KHR *pMemoryRequirements,
Uint32 *pMemoryTypeIndex
) {
VkImageMemoryRequirementsInfo2KHR imageRequirementsInfo;
imageRequirementsInfo.sType =
VK_STRUCTURE_TYPE_IMAGE_MEMORY_REQUIREMENTS_INFO_2_KHR;
imageRequirementsInfo.pNext = NULL;
imageRequirementsInfo.image = image;
renderer->vkGetImageMemoryRequirements2KHR(
renderer->logicalDevice,
&imageRequirementsInfo,
pMemoryRequirements
);
return VULKAN_INTERNAL_FindMemoryType(
renderer,
pMemoryRequirements->memoryRequirements.memoryTypeBits,
requiredMemoryPropertyFlags,
0,
pMemoryTypeIndex
);
}
static void VULKAN_INTERNAL_DeallocateMemory(
VulkanRenderer *renderer,
VulkanMemorySubAllocator *allocator,
Uint32 allocationIndex
) {
Uint32 i;
VulkanMemoryAllocation *allocation = allocator->allocations[allocationIndex];
SDL_LockMutex(renderer->allocatorLock);
/* If this allocation was marked for defrag, cancel that */
for (i = 0; i < renderer->allocationsToDefragCount; i += 1)
{
if (allocation == renderer->allocationsToDefrag[i])
{
renderer->allocationsToDefrag[i] = renderer->allocationsToDefrag[renderer->allocationsToDefragCount - 1];
renderer->allocationsToDefragCount -= 1;
break;
}
}
for (i = 0; i < allocation->freeRegionCount; i += 1)
{
VULKAN_INTERNAL_RemoveMemoryFreeRegion(
renderer,
allocation->freeRegions[i]
);
}
SDL_free(allocation->freeRegions);
/* no need to iterate used regions because deallocate
* only happens when there are 0 used regions
*/
SDL_free(allocation->usedRegions);
renderer->vkFreeMemory(
renderer->logicalDevice,
allocation->memory,
NULL
);
SDL_DestroyMutex(allocation->memoryLock);
SDL_free(allocation);
if (allocationIndex != allocator->allocationCount - 1)
{
allocator->allocations[allocationIndex] = allocator->allocations[allocator->allocationCount - 1];
}
allocator->allocationCount -= 1;
SDL_UnlockMutex(renderer->allocatorLock);
}
static Uint8 VULKAN_INTERNAL_AllocateMemory(
VulkanRenderer *renderer,
VkBuffer buffer,
VkImage image,
Uint32 memoryTypeIndex,
VkDeviceSize allocationSize,
Uint8 isHostVisible,
VulkanMemoryAllocation **pMemoryAllocation)
{
VulkanMemoryAllocation *allocation;
VulkanMemorySubAllocator *allocator = &renderer->memoryAllocator->subAllocators[memoryTypeIndex];
VkMemoryAllocateInfo allocInfo;
VkResult result;
allocInfo.sType = VK_STRUCTURE_TYPE_MEMORY_ALLOCATE_INFO;
allocInfo.pNext = NULL;
allocInfo.memoryTypeIndex = memoryTypeIndex;
allocInfo.allocationSize = allocationSize;
allocation = SDL_malloc(sizeof(VulkanMemoryAllocation));
allocation->size = allocationSize;
allocation->freeSpace = 0; /* added by FreeRegions */
allocation->usedSpace = 0; /* added by UsedRegions */
allocation->memoryLock = SDL_CreateMutex();
allocator->allocationCount += 1;
allocator->allocations = SDL_realloc(
allocator->allocations,
sizeof(VulkanMemoryAllocation*) * allocator->allocationCount
);
allocator->allocations[
allocator->allocationCount - 1
] = allocation;
allocInfo.pNext = NULL;
allocation->availableForAllocation = 1;
allocation->usedRegions = SDL_malloc(sizeof(VulkanMemoryUsedRegion*));
allocation->usedRegionCount = 0;
allocation->usedRegionCapacity = 1;
allocation->freeRegions = SDL_malloc(sizeof(VulkanMemoryFreeRegion*));
allocation->freeRegionCount = 0;
allocation->freeRegionCapacity = 1;
allocation->allocator = allocator;
result = renderer->vkAllocateMemory(
renderer->logicalDevice,
&allocInfo,
NULL,
&allocation->memory
);
if (result != VK_SUCCESS)
{
/* Uh oh, we couldn't allocate, time to clean up */
SDL_free(allocation->freeRegions);
allocator->allocationCount -= 1;
allocator->allocations = SDL_realloc(
allocator->allocations,
sizeof(VulkanMemoryAllocation*) * allocator->allocationCount
);
SDL_free(allocation);
return 0;
}
/* Persistent mapping for host-visible memory */
if (isHostVisible)
{
result = renderer->vkMapMemory(
renderer->logicalDevice,
allocation->memory,
0,
VK_WHOLE_SIZE,
0,
(void**) &allocation->mapPointer
);
VULKAN_ERROR_CHECK(result, vkMapMemory, 0)
}
else
{
allocation->mapPointer = NULL;
}
VULKAN_INTERNAL_NewMemoryFreeRegion(
renderer,
allocation,
0,
allocation->size
);
*pMemoryAllocation = allocation;
return 1;
}
static Uint8 VULKAN_INTERNAL_BindBufferMemory(
VulkanRenderer *renderer,
VulkanMemoryUsedRegion *usedRegion,
VkDeviceSize alignedOffset,
VkBuffer buffer
) {
VkResult vulkanResult;
SDL_LockMutex(usedRegion->allocation->memoryLock);
vulkanResult = renderer->vkBindBufferMemory(
renderer->logicalDevice,
buffer,
usedRegion->allocation->memory,
alignedOffset
);
SDL_UnlockMutex(usedRegion->allocation->memoryLock);
VULKAN_ERROR_CHECK(vulkanResult, vkBindBufferMemory, 0)
return 1;
}
static Uint8 VULKAN_INTERNAL_BindImageMemory(
VulkanRenderer *renderer,
VulkanMemoryUsedRegion *usedRegion,
VkDeviceSize alignedOffset,
VkImage image
) {
VkResult vulkanResult;
SDL_LockMutex(usedRegion->allocation->memoryLock);
vulkanResult = renderer->vkBindImageMemory(
renderer->logicalDevice,
image,
usedRegion->allocation->memory,
alignedOffset
);
SDL_UnlockMutex(usedRegion->allocation->memoryLock);
VULKAN_ERROR_CHECK(vulkanResult, vkBindBufferMemory, 0)
return 1;
}
static Uint8 VULKAN_INTERNAL_BindResourceMemory(
VulkanRenderer* renderer,
Uint32 memoryTypeIndex,
VkMemoryRequirements2KHR* memoryRequirements,
VkDeviceSize resourceSize, /* may be different from requirements size! */
VkBuffer buffer, /* may be VK_NULL_HANDLE */
VkImage image, /* may be VK_NULL_HANDLE */
VulkanMemoryUsedRegion** pMemoryUsedRegion
) {
VulkanMemoryAllocation *allocation;
VulkanMemorySubAllocator *allocator;
VulkanMemoryFreeRegion *region;
VulkanMemoryFreeRegion *selectedRegion;
VulkanMemoryUsedRegion *usedRegion;
VkDeviceSize requiredSize, allocationSize;
VkDeviceSize alignedOffset;
Uint32 newRegionSize, newRegionOffset;
Uint8 isHostVisible, smallAllocation, allocationResult;
Sint32 i;
isHostVisible =
(renderer->memoryProperties.memoryTypes[memoryTypeIndex].propertyFlags &
VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT) != 0;
allocator = &renderer->memoryAllocator->subAllocators[memoryTypeIndex];
requiredSize = memoryRequirements->memoryRequirements.size;
smallAllocation = requiredSize <= SMALL_ALLOCATION_THRESHOLD;
if ( (buffer == VK_NULL_HANDLE && image == VK_NULL_HANDLE) ||
(buffer != VK_NULL_HANDLE && image != VK_NULL_HANDLE) )
{
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "BindResourceMemory must be given either a VulkanBuffer or a VulkanTexture");
return 0;
}
SDL_LockMutex(renderer->allocatorLock);
selectedRegion = NULL;
for (i = allocator->sortedFreeRegionCount - 1; i >= 0; i -= 1)
{
region = allocator->sortedFreeRegions[i];
if (smallAllocation && region->allocation->size != SMALL_ALLOCATION_SIZE)
{
/* region is not in a small allocation */
continue;
}
if (!smallAllocation && region->allocation->size == SMALL_ALLOCATION_SIZE)
{
/* allocation is not small and current region is in a small allocation */
continue;
}
alignedOffset = VULKAN_INTERNAL_NextHighestAlignment(
region->offset,
memoryRequirements->memoryRequirements.alignment
);
if (alignedOffset + requiredSize <= region->offset + region->size)
{
selectedRegion = region;
break;
}
}
if (selectedRegion != NULL)
{
region = selectedRegion;
allocation = region->allocation;
usedRegion = VULKAN_INTERNAL_NewMemoryUsedRegion(
renderer,
allocation,
region->offset,
requiredSize + (alignedOffset - region->offset),
alignedOffset,
resourceSize,
memoryRequirements->memoryRequirements.alignment
);
usedRegion->isBuffer = buffer != VK_NULL_HANDLE;
newRegionSize = region->size - ((alignedOffset - region->offset) + requiredSize);
newRegionOffset = alignedOffset + requiredSize;
/* remove and add modified region to re-sort */
VULKAN_INTERNAL_RemoveMemoryFreeRegion(renderer, region);
/* if size is 0, no need to re-insert */
if (newRegionSize != 0)
{
VULKAN_INTERNAL_NewMemoryFreeRegion(
renderer,
allocation,
newRegionOffset,
newRegionSize
);
}
SDL_UnlockMutex(renderer->allocatorLock);
if (buffer != VK_NULL_HANDLE)
{
if (!VULKAN_INTERNAL_BindBufferMemory(
renderer,
usedRegion,
alignedOffset,
buffer
)) {
VULKAN_INTERNAL_RemoveMemoryUsedRegion(
renderer,
usedRegion
);
return 0;
}
}
else if (image != VK_NULL_HANDLE)
{
if (!VULKAN_INTERNAL_BindImageMemory(
renderer,
usedRegion,
alignedOffset,
image
)) {
VULKAN_INTERNAL_RemoveMemoryUsedRegion(
renderer,
usedRegion
);
return 0;
}
}
*pMemoryUsedRegion = usedRegion;
return 1;
}
/* No suitable free regions exist, allocate a new memory region */
if (
renderer->allocationsToDefragCount == 0 &&
!renderer->defragInProgress
) {
/* Mark currently fragmented allocations for defrag */
VULKAN_INTERNAL_MarkAllocationsForDefrag(renderer);
}
if (requiredSize > SMALL_ALLOCATION_THRESHOLD)
{
/* allocate a page of required size aligned to LARGE_ALLOCATION_INCREMENT increments */
allocationSize =
VULKAN_INTERNAL_NextHighestAlignment(requiredSize, LARGE_ALLOCATION_INCREMENT);
}
else
{
allocationSize = SMALL_ALLOCATION_SIZE;
}
allocationResult = VULKAN_INTERNAL_AllocateMemory(
renderer,
buffer,
image,
memoryTypeIndex,
allocationSize,
isHostVisible,
&allocation
);
/* Uh oh, we're out of memory */
if (allocationResult == 0)
{
SDL_UnlockMutex(renderer->allocatorLock);
/* Responsibility of the caller to handle being out of memory */
return 2;
}
usedRegion = VULKAN_INTERNAL_NewMemoryUsedRegion(
renderer,
allocation,
0,
requiredSize,
0,
resourceSize,
memoryRequirements->memoryRequirements.alignment
);
usedRegion->isBuffer = buffer != VK_NULL_HANDLE;
region = allocation->freeRegions[0];
newRegionOffset = region->offset + requiredSize;
newRegionSize = region->size - requiredSize;
VULKAN_INTERNAL_RemoveMemoryFreeRegion(renderer, region);
if (newRegionSize != 0)
{
VULKAN_INTERNAL_NewMemoryFreeRegion(
renderer,
allocation,
newRegionOffset,
newRegionSize
);
}
SDL_UnlockMutex(renderer->allocatorLock);
if (buffer != VK_NULL_HANDLE)
{
if (!VULKAN_INTERNAL_BindBufferMemory(
renderer,
usedRegion,
0,
buffer
)) {
VULKAN_INTERNAL_RemoveMemoryUsedRegion(
renderer,
usedRegion
);
return 0;
}
}
else if (image != VK_NULL_HANDLE)
{
if (!VULKAN_INTERNAL_BindImageMemory(
renderer,
usedRegion,
0,
image
)) {
VULKAN_INTERNAL_RemoveMemoryUsedRegion(
renderer,
usedRegion
);
return 0;
}
}
*pMemoryUsedRegion = usedRegion;
return 1;
}
static Uint8 VULKAN_INTERNAL_BindMemoryForImage(
VulkanRenderer* renderer,
VkImage image,
VulkanMemoryUsedRegion** usedRegion
) {
Uint8 bindResult = 0;
Uint32 memoryTypeIndex = 0;
VkMemoryPropertyFlags requiredMemoryPropertyFlags;
VkMemoryRequirements2KHR memoryRequirements =
{
VK_STRUCTURE_TYPE_MEMORY_REQUIREMENTS_2_KHR,
NULL
};
/* Prefer GPU allocation for textures */
requiredMemoryPropertyFlags = VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT;
while (VULKAN_INTERNAL_FindImageMemoryRequirements(
renderer,
image,
requiredMemoryPropertyFlags,
&memoryRequirements,
&memoryTypeIndex
)) {
bindResult = VULKAN_INTERNAL_BindResourceMemory(
renderer,
memoryTypeIndex,
&memoryRequirements,
memoryRequirements.memoryRequirements.size,
VK_NULL_HANDLE,
image,
usedRegion
);
if (bindResult == 1)
{
break;
}
else /* Bind failed, try the next device-local heap */
{
memoryTypeIndex += 1;
}
}
/* Bind _still_ failed, try again without device local */
if (bindResult != 1)
{
memoryTypeIndex = 0;
requiredMemoryPropertyFlags = 0;
SDL_LogWarn(SDL_LOG_CATEGORY_APPLICATION, "Out of device-local memory, allocating textures on host-local memory!");
while (VULKAN_INTERNAL_FindImageMemoryRequirements(
renderer,
image,
requiredMemoryPropertyFlags,
&memoryRequirements,
&memoryTypeIndex
)) {
bindResult = VULKAN_INTERNAL_BindResourceMemory(
renderer,
memoryTypeIndex,
&memoryRequirements,
memoryRequirements.memoryRequirements.size,
VK_NULL_HANDLE,
image,
usedRegion
);
if (bindResult == 1)
{
break;
}
else /* Bind failed, try the next heap */
{
memoryTypeIndex += 1;
}
}
}
return bindResult;
}
static Uint8 VULKAN_INTERNAL_BindMemoryForBuffer(
VulkanRenderer* renderer,
VkBuffer buffer,
VkDeviceSize size,
VulkanBufferType type,
VulkanMemoryUsedRegion** usedRegion
) {
Uint8 bindResult = 0;
Uint32 memoryTypeIndex = 0;
VkMemoryPropertyFlags requiredMemoryPropertyFlags = 0;
VkMemoryPropertyFlags ignoredMemoryPropertyFlags = 0;
VkMemoryRequirements2KHR memoryRequirements =
{
VK_STRUCTURE_TYPE_MEMORY_REQUIREMENTS_2_KHR,
NULL
};
if (type == VULKAN_BUFFER_TYPE_GPU)
{
requiredMemoryPropertyFlags |=
VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT;
}
else if (type == VULKAN_BUFFER_TYPE_UNIFORM)
{
requiredMemoryPropertyFlags |=
VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT |
VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT |
VK_MEMORY_PROPERTY_HOST_COHERENT_BIT;
}
else if (type == VULKAN_BUFFER_TYPE_TRANSFER)
{
requiredMemoryPropertyFlags |=
VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT |
VK_MEMORY_PROPERTY_HOST_CACHED_BIT;
ignoredMemoryPropertyFlags |=
VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT;
}
else
{
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Unrecognized buffer type!");
return 0;
}
while (VULKAN_INTERNAL_FindBufferMemoryRequirements(
renderer,
buffer,
requiredMemoryPropertyFlags,
ignoredMemoryPropertyFlags,
&memoryRequirements,
&memoryTypeIndex
)) {
bindResult = VULKAN_INTERNAL_BindResourceMemory(
renderer,
memoryTypeIndex,
&memoryRequirements,
size,
buffer,
VK_NULL_HANDLE,
usedRegion
);
if (bindResult == 1)
{
break;
}
else /* Bind failed, try the next device-local heap */
{
memoryTypeIndex += 1;
}
}
/* Bind failed, try again without preferred flags */
if (bindResult != 1)
{
memoryTypeIndex = 0;
requiredMemoryPropertyFlags = 0;
ignoredMemoryPropertyFlags = 0;
if (type == VULKAN_BUFFER_TYPE_GPU)
{
if (!renderer->outOfDeviceLocalMemoryWarning)
{
SDL_LogInfo(SDL_LOG_CATEGORY_APPLICATION, "Out of device-local memory, allocating buffers on host-local memory, expect degraded performance!");
renderer->outOfDeviceLocalMemoryWarning = 1;
}
}
else if (type == VULKAN_BUFFER_TYPE_UNIFORM)
{
requiredMemoryPropertyFlags =
VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT |
VK_MEMORY_PROPERTY_HOST_COHERENT_BIT;
if (!renderer->outofBARMemoryWarning)
{
SDL_LogInfo(SDL_LOG_CATEGORY_APPLICATION, "Out of BAR memory, allocating uniform buffers on host-local memory, expect degraded performance!");
renderer->outofBARMemoryWarning = 1;
}
}
else if (type == VULKAN_BUFFER_TYPE_TRANSFER)
{
requiredMemoryPropertyFlags =
VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT |
VK_MEMORY_PROPERTY_HOST_CACHED_BIT;
if (!renderer->integratedMemoryNotification)
{
SDL_LogInfo(SDL_LOG_CATEGORY_APPLICATION, "Integrated memory detected, allocating TransferBuffers on device-local memory!");
renderer->integratedMemoryNotification = 1;
}
}
while (VULKAN_INTERNAL_FindBufferMemoryRequirements(
renderer,
buffer,
requiredMemoryPropertyFlags,
ignoredMemoryPropertyFlags,
&memoryRequirements,
&memoryTypeIndex
)) {
bindResult = VULKAN_INTERNAL_BindResourceMemory(
renderer,
memoryTypeIndex,
&memoryRequirements,
size,
buffer,
VK_NULL_HANDLE,
usedRegion
);
if (bindResult == 1)
{
break;
}
else /* Bind failed, try the next heap */
{
memoryTypeIndex += 1;
}
}
}
return bindResult;
}
/* Resource tracking */
#define ADD_TO_ARRAY_UNIQUE(resource, type, array, count, capacity) \
Uint32 i; \
\
for (i = 0; i < commandBuffer->count; i += 1) \
{ \
if (commandBuffer->array[i] == resource) \
{ \
return; \
} \
} \
\
if (commandBuffer->count == commandBuffer->capacity) \
{ \
commandBuffer->capacity += 1; \
commandBuffer->array = SDL_realloc( \
commandBuffer->array, \
commandBuffer->capacity * sizeof(type) \
); \
} \
commandBuffer->array[commandBuffer->count] = resource; \
commandBuffer->count += 1;
#define TRACK_RESOURCE(resource, type, array, count, capacity) \
Uint32 i; \
\
for (i = 0; i < commandBuffer->count; i += 1) \
{ \
if (commandBuffer->array[i] == resource) \
{ \
return; \
} \
} \
\
if (commandBuffer->count == commandBuffer->capacity) \
{ \
commandBuffer->capacity += 1; \
commandBuffer->array = SDL_realloc( \
commandBuffer->array, \
commandBuffer->capacity * sizeof(type) \
); \
} \
commandBuffer->array[commandBuffer->count] = resource; \
commandBuffer->count += 1; \
SDL_AtomicIncRef(&resource->referenceCount);
static void VULKAN_INTERNAL_TrackBuffer(
VulkanRenderer *renderer,
VulkanCommandBuffer *commandBuffer,
VulkanBuffer *buffer
) {
TRACK_RESOURCE(
buffer,
VulkanBuffer*,
usedBuffers,
usedBufferCount,
usedBufferCapacity
)
}
static void VULKAN_INTERNAL_TrackTextureSlice(
VulkanRenderer *renderer,
VulkanCommandBuffer *commandBuffer,
VulkanTextureSlice *textureSlice
) {
TRACK_RESOURCE(
textureSlice,
VulkanTextureSlice*,
usedTextureSlices,
usedTextureSliceCount,
usedTextureSliceCapacity
)
}
static void VULKAN_INTERNAL_TrackSampler(
VulkanRenderer *renderer,
VulkanCommandBuffer *commandBuffer,
VulkanSampler *sampler
) {
TRACK_RESOURCE(
sampler,
VulkanSampler*,
usedSamplers,
usedSamplerCount,
usedSamplerCapacity
)
}
static void VULKAN_INTERNAL_TrackGraphicsPipeline(
VulkanRenderer *renderer,
VulkanCommandBuffer *commandBuffer,
VulkanGraphicsPipeline *graphicsPipeline
) {
TRACK_RESOURCE(
graphicsPipeline,
VulkanGraphicsPipeline*,
usedGraphicsPipelines,
usedGraphicsPipelineCount,
usedGraphicsPipelineCapacity
)
}
static void VULKAN_INTERNAL_TrackComputePipeline(
VulkanRenderer *renderer,
VulkanCommandBuffer *commandBuffer,
VulkanComputePipeline *computePipeline
) {
TRACK_RESOURCE(
computePipeline,
VulkanComputePipeline*,
usedComputePipelines,
usedComputePipelineCount,
usedComputePipelineCapacity
)
}
static void VULKAN_INTERNAL_TrackFramebuffer(
VulkanRenderer *renderer,
VulkanCommandBuffer *commandBuffer,
VulkanFramebuffer *framebuffer
) {
TRACK_RESOURCE(
framebuffer,
VulkanFramebuffer*,
usedFramebuffers,
usedFramebufferCount,
usedFramebufferCapacity
);
}
#undef TRACK_RESOURCE
/* Memory Barriers */
/*
* In Vulkan, we must manually synchronize operations that write to resources on the GPU
* so that read-after-write, write-after-read, and write-after-write hazards do not occur.
* Additionally, textures are required to be in specific layouts for specific use cases.
* Both of these tasks are accomplished with vkCmdPipelineBarrier.
*
* To insert the correct barriers, we keep track of "usage modes" for buffers and textures.
* These indicate the current usage of that resource on the command buffer.
* The transition from one usage mode to another indicates how the barrier should be constructed.
*
* Pipeline barriers cannot be inserted during a render pass, but they can be inserted
* during a compute or copy pass.
*
* This means that the "default" usage mode of any given resource should be that it should be
* ready for a graphics-read operation, because we cannot barrier during a render pass.
* In the case where a resource is only used in compute, its default usage mode can be compute-read.
* This strategy allows us to avoid expensive record keeping of command buffer/resource usage mode pairs,
* and it fully covers synchronization between all combinations of stages.
*
* In Upload and Copy functions, we transition the resource immediately before and after the copy command.
*
* When binding a resource for compute, we transition when the Bind functions are called.
* If a bind slot containing a resource is overwritten, we transition the resource in that slot back to its default.
* When EndComputePass is called we transition all bound resources back to their default state.
*
* When binding a texture as a render pass attachment, we transition the resource on BeginRenderPass
* and transition it back to its default on EndRenderPass.
*
* This strategy imposes certain limitations on resource usage flags.
* For example, a texture cannot have both the SAMPLER and GRAPHICS_STORAGE usage flags,
* because then it is imposible for the backend to infer which default usage mode the texture should use.
*
* Sync hazards can be detected by setting VK_KHRONOS_VALIDATION_VALIDATE_SYNC=1 when using validation layers.
*/
static void VULKAN_INTERNAL_BufferMemoryBarrier(
VulkanRenderer *renderer,
VulkanCommandBuffer *commandBuffer,
VulkanBufferUsageMode sourceUsageMode,
VulkanBufferUsageMode destinationUsageMode,
VulkanBuffer *buffer
) {
VkPipelineStageFlags srcStages = 0;
VkPipelineStageFlags dstStages = 0;
VkBufferMemoryBarrier memoryBarrier;
memoryBarrier.sType = VK_STRUCTURE_TYPE_BUFFER_MEMORY_BARRIER;
memoryBarrier.pNext = NULL;
memoryBarrier.srcAccessMask = 0;
memoryBarrier.dstAccessMask = 0;
memoryBarrier.srcQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED;
memoryBarrier.dstQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED;
memoryBarrier.buffer = buffer->buffer;
memoryBarrier.offset = 0;
memoryBarrier.size = buffer->size;
if (sourceUsageMode == VULKAN_BUFFER_USAGE_MODE_COPY_SOURCE)
{
srcStages = VK_PIPELINE_STAGE_TRANSFER_BIT;
memoryBarrier.srcAccessMask = VK_ACCESS_TRANSFER_READ_BIT;
}
else if (sourceUsageMode == VULKAN_BUFFER_USAGE_MODE_COPY_DESTINATION)
{
srcStages = VK_PIPELINE_STAGE_TRANSFER_BIT;
memoryBarrier.srcAccessMask = VK_ACCESS_TRANSFER_WRITE_BIT;
}
else if (sourceUsageMode == VULKAN_BUFFER_USAGE_MODE_VERTEX_READ)
{
srcStages = VK_PIPELINE_STAGE_VERTEX_INPUT_BIT;
memoryBarrier.srcAccessMask = VK_ACCESS_VERTEX_ATTRIBUTE_READ_BIT;
}
else if (sourceUsageMode == VULKAN_BUFFER_USAGE_MODE_INDEX_READ)
{
srcStages = VK_PIPELINE_STAGE_VERTEX_INPUT_BIT;
memoryBarrier.srcAccessMask = VK_ACCESS_INDEX_READ_BIT;
}
else if (sourceUsageMode == VULKAN_BUFFER_USAGE_MODE_INDIRECT)
{
srcStages = VK_PIPELINE_STAGE_DRAW_INDIRECT_BIT;
memoryBarrier.srcAccessMask = VK_ACCESS_INDIRECT_COMMAND_READ_BIT;
}
else if (sourceUsageMode == VULKAN_BUFFER_USAGE_MODE_GRAPHICS_STORAGE_READ)
{
srcStages = VK_PIPELINE_STAGE_VERTEX_SHADER_BIT | VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT;
memoryBarrier.srcAccessMask = VK_ACCESS_SHADER_READ_BIT;
}
else if (sourceUsageMode == VULKAN_BUFFER_USAGE_MODE_COMPUTE_STORAGE_READ)
{
srcStages = VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT;
memoryBarrier.srcAccessMask = VK_ACCESS_SHADER_READ_BIT;
}
else if (sourceUsageMode == VULKAN_BUFFER_USAGE_MODE_COMPUTE_STORAGE_READ_WRITE)
{
srcStages = VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT;
memoryBarrier.srcAccessMask = VK_ACCESS_SHADER_READ_BIT | VK_ACCESS_SHADER_WRITE_BIT;
}
else
{
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Unrecognized buffer source barrier type!");
return;
}
if (destinationUsageMode == VULKAN_BUFFER_USAGE_MODE_COPY_SOURCE)
{
dstStages = VK_PIPELINE_STAGE_TRANSFER_BIT;
memoryBarrier.dstAccessMask = VK_ACCESS_TRANSFER_READ_BIT;
}
else if (destinationUsageMode == VULKAN_BUFFER_USAGE_MODE_COPY_DESTINATION)
{
dstStages = VK_PIPELINE_STAGE_TRANSFER_BIT;
memoryBarrier.dstAccessMask = VK_ACCESS_TRANSFER_WRITE_BIT;
}
else if (destinationUsageMode == VULKAN_BUFFER_USAGE_MODE_VERTEX_READ)
{
dstStages = VK_PIPELINE_STAGE_VERTEX_INPUT_BIT;
memoryBarrier.dstAccessMask = VK_ACCESS_VERTEX_ATTRIBUTE_READ_BIT;
}
else if (destinationUsageMode == VULKAN_BUFFER_USAGE_MODE_INDEX_READ)
{
dstStages = VK_PIPELINE_STAGE_VERTEX_INPUT_BIT;
memoryBarrier.dstAccessMask = VK_ACCESS_INDEX_READ_BIT;
}
else if (destinationUsageMode == VULKAN_BUFFER_USAGE_MODE_INDIRECT)
{
dstStages = VK_PIPELINE_STAGE_DRAW_INDIRECT_BIT;
memoryBarrier.dstAccessMask = VK_ACCESS_INDIRECT_COMMAND_READ_BIT;
}
else if (destinationUsageMode == VULKAN_BUFFER_USAGE_MODE_GRAPHICS_STORAGE_READ)
{
dstStages = VK_PIPELINE_STAGE_VERTEX_SHADER_BIT | VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT;
memoryBarrier.dstAccessMask = VK_ACCESS_SHADER_READ_BIT;
}
else if (destinationUsageMode == VULKAN_BUFFER_USAGE_MODE_COMPUTE_STORAGE_READ)
{
dstStages = VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT;
memoryBarrier.dstAccessMask = VK_ACCESS_SHADER_READ_BIT;
}
else if (destinationUsageMode == VULKAN_BUFFER_USAGE_MODE_COMPUTE_STORAGE_READ_WRITE)
{
dstStages = VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT;
memoryBarrier.dstAccessMask = VK_ACCESS_SHADER_READ_BIT | VK_ACCESS_SHADER_WRITE_BIT;
}
else
{
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Unrecognized buffer destination barrier type!");
return;
}
renderer->vkCmdPipelineBarrier(
commandBuffer->commandBuffer,
srcStages,
dstStages,
0,
0,
NULL,
1,
&memoryBarrier,
0,
NULL
);
buffer->transitioned = SDL_TRUE;
}
static void VULKAN_INTERNAL_ImageMemoryBarrier(
VulkanRenderer *renderer,
VulkanCommandBuffer *commandBuffer,
VulkanTextureUsageMode sourceUsageMode,
VulkanTextureUsageMode destinationUsageMode,
VulkanTextureSlice *textureSlice
) {
VkPipelineStageFlags srcStages = 0;
VkPipelineStageFlags dstStages = 0;
VkImageMemoryBarrier memoryBarrier;
memoryBarrier.sType = VK_STRUCTURE_TYPE_IMAGE_MEMORY_BARRIER;
memoryBarrier.pNext = NULL;
memoryBarrier.srcAccessMask = 0;
memoryBarrier.dstAccessMask = 0;
memoryBarrier.oldLayout = VK_IMAGE_LAYOUT_UNDEFINED;
memoryBarrier.newLayout = VK_IMAGE_LAYOUT_UNDEFINED;
memoryBarrier.srcQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED;
memoryBarrier.dstQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED;
memoryBarrier.image = textureSlice->parent->image;
memoryBarrier.subresourceRange.aspectMask = textureSlice->parent->aspectFlags;
memoryBarrier.subresourceRange.baseArrayLayer = textureSlice->layer;
memoryBarrier.subresourceRange.layerCount = 1;
memoryBarrier.subresourceRange.baseMipLevel = textureSlice->level;
memoryBarrier.subresourceRange.levelCount = 1;
if (sourceUsageMode == VULKAN_TEXTURE_USAGE_MODE_COPY_SOURCE)
{
srcStages = VK_PIPELINE_STAGE_TRANSFER_BIT;
memoryBarrier.srcAccessMask = VK_ACCESS_TRANSFER_READ_BIT;
memoryBarrier.oldLayout = VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL;
}
else if (sourceUsageMode == VULKAN_TEXTURE_USAGE_MODE_COPY_DESTINATION)
{
srcStages = VK_PIPELINE_STAGE_TRANSFER_BIT;
memoryBarrier.srcAccessMask = VK_ACCESS_TRANSFER_WRITE_BIT;
memoryBarrier.oldLayout = VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL;
}
else if (sourceUsageMode == VULKAN_TEXTURE_USAGE_MODE_SAMPLER)
{
srcStages = VK_PIPELINE_STAGE_VERTEX_SHADER_BIT | VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT;
memoryBarrier.srcAccessMask = VK_ACCESS_SHADER_READ_BIT;
memoryBarrier.oldLayout = VK_IMAGE_LAYOUT_SHADER_READ_ONLY_OPTIMAL;
}
else if (sourceUsageMode == VULKAN_TEXTURE_USAGE_MODE_GRAPHICS_STORAGE_READ)
{
srcStages = VK_PIPELINE_STAGE_VERTEX_SHADER_BIT | VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT;
memoryBarrier.srcAccessMask = VK_ACCESS_SHADER_READ_BIT;
memoryBarrier.oldLayout = VK_IMAGE_LAYOUT_GENERAL;
}
else if (sourceUsageMode == VULKAN_TEXTURE_USAGE_MODE_COMPUTE_STORAGE_READ)
{
srcStages = VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT;
memoryBarrier.srcAccessMask = VK_ACCESS_SHADER_READ_BIT;
memoryBarrier.oldLayout = VK_IMAGE_LAYOUT_GENERAL;
}
else if (sourceUsageMode == VULKAN_TEXTURE_USAGE_MODE_COMPUTE_STORAGE_READ_WRITE)
{
srcStages = VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT;
memoryBarrier.srcAccessMask = VK_ACCESS_SHADER_READ_BIT | VK_ACCESS_SHADER_WRITE_BIT;
memoryBarrier.oldLayout = VK_IMAGE_LAYOUT_GENERAL;
}
else if (sourceUsageMode == VULKAN_TEXTURE_USAGE_MODE_COLOR_ATTACHMENT)
{
srcStages = VK_PIPELINE_STAGE_COLOR_ATTACHMENT_OUTPUT_BIT;
memoryBarrier.srcAccessMask = VK_ACCESS_COLOR_ATTACHMENT_WRITE_BIT;
memoryBarrier.oldLayout = VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL;
}
else if (sourceUsageMode == VULKAN_TEXTURE_USAGE_MODE_DEPTH_STENCIL_ATTACHMENT)
{
srcStages = VK_PIPELINE_STAGE_EARLY_FRAGMENT_TESTS_BIT | VK_PIPELINE_STAGE_LATE_FRAGMENT_TESTS_BIT;
memoryBarrier.srcAccessMask = VK_ACCESS_DEPTH_STENCIL_ATTACHMENT_WRITE_BIT;
memoryBarrier.oldLayout = VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL;
}
else
{
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Unrecognized texture source barrier type!");
return;
}
if (!textureSlice->transitioned)
{
memoryBarrier.oldLayout = VK_IMAGE_LAYOUT_UNDEFINED;
}
if (destinationUsageMode == VULKAN_TEXTURE_USAGE_MODE_COPY_SOURCE)
{
dstStages = VK_PIPELINE_STAGE_TRANSFER_BIT;
memoryBarrier.dstAccessMask = VK_ACCESS_TRANSFER_READ_BIT;
memoryBarrier.newLayout = VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL;
}
else if (destinationUsageMode == VULKAN_TEXTURE_USAGE_MODE_COPY_DESTINATION)
{
dstStages = VK_PIPELINE_STAGE_TRANSFER_BIT;
memoryBarrier.dstAccessMask = VK_ACCESS_TRANSFER_WRITE_BIT;
memoryBarrier.newLayout = VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL;
}
else if (destinationUsageMode == VULKAN_TEXTURE_USAGE_MODE_SAMPLER)
{
dstStages = VK_PIPELINE_STAGE_VERTEX_SHADER_BIT | VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT;
memoryBarrier.dstAccessMask = VK_ACCESS_SHADER_READ_BIT;
memoryBarrier.newLayout = VK_IMAGE_LAYOUT_SHADER_READ_ONLY_OPTIMAL;
}
else if (destinationUsageMode == VULKAN_TEXTURE_USAGE_MODE_GRAPHICS_STORAGE_READ)
{
dstStages = VK_PIPELINE_STAGE_VERTEX_SHADER_BIT | VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT;
memoryBarrier.dstAccessMask = VK_ACCESS_SHADER_READ_BIT;
memoryBarrier.newLayout = VK_IMAGE_LAYOUT_GENERAL;
}
else if (destinationUsageMode == VULKAN_TEXTURE_USAGE_MODE_COMPUTE_STORAGE_READ)
{
dstStages = VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT;
memoryBarrier.dstAccessMask = VK_ACCESS_SHADER_READ_BIT;
memoryBarrier.newLayout = VK_IMAGE_LAYOUT_GENERAL;
}
else if (destinationUsageMode == VULKAN_TEXTURE_USAGE_MODE_COMPUTE_STORAGE_READ_WRITE)
{
dstStages = VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT;
memoryBarrier.dstAccessMask = VK_ACCESS_SHADER_READ_BIT | VK_ACCESS_SHADER_WRITE_BIT;
memoryBarrier.newLayout = VK_IMAGE_LAYOUT_GENERAL;
}
else if (destinationUsageMode == VULKAN_TEXTURE_USAGE_MODE_COLOR_ATTACHMENT)
{
dstStages = VK_PIPELINE_STAGE_COLOR_ATTACHMENT_OUTPUT_BIT;
memoryBarrier.dstAccessMask = VK_ACCESS_COLOR_ATTACHMENT_READ_BIT | VK_ACCESS_COLOR_ATTACHMENT_WRITE_BIT;
memoryBarrier.newLayout = VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL;
}
else if (destinationUsageMode == VULKAN_TEXTURE_USAGE_MODE_DEPTH_STENCIL_ATTACHMENT)
{
dstStages = VK_PIPELINE_STAGE_EARLY_FRAGMENT_TESTS_BIT | VK_PIPELINE_STAGE_LATE_FRAGMENT_TESTS_BIT;
memoryBarrier.dstAccessMask = VK_ACCESS_DEPTH_STENCIL_ATTACHMENT_READ_BIT | VK_ACCESS_DEPTH_STENCIL_ATTACHMENT_WRITE_BIT;
memoryBarrier.newLayout = VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL;
}
else if (destinationUsageMode == VULKAN_TEXTURE_USAGE_MODE_PRESENT)
{
dstStages = VK_PIPELINE_STAGE_BOTTOM_OF_PIPE_BIT;
memoryBarrier.dstAccessMask = 0;
memoryBarrier.newLayout = VK_IMAGE_LAYOUT_PRESENT_SRC_KHR;
}
else
{
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Unrecognized texture destination barrier type!");
return;
}
renderer->vkCmdPipelineBarrier(
commandBuffer->commandBuffer,
srcStages,
dstStages,
0,
0,
NULL,
0,
NULL,
1,
&memoryBarrier
);
textureSlice->transitioned = SDL_TRUE;
}
static VulkanBufferUsageMode VULKAN_INTERNAL_DefaultBufferUsageMode(
VulkanBuffer *buffer
) {
/* NOTE: order matters here! */
if (buffer->usageFlags & REFRESH_BUFFERUSAGE_VERTEX_BIT)
{
return VULKAN_BUFFER_USAGE_MODE_VERTEX_READ;
}
else if (buffer->usageFlags & REFRESH_BUFFERUSAGE_INDEX_BIT)
{
return VULKAN_BUFFER_USAGE_MODE_INDEX_READ;
}
else if (buffer->usageFlags & REFRESH_BUFFERUSAGE_INDIRECT_BIT)
{
return VULKAN_BUFFER_USAGE_MODE_INDIRECT;
}
else if (buffer->usageFlags & REFRESH_BUFFERUSAGE_GRAPHICS_STORAGE_READ_BIT)
{
return VULKAN_BUFFER_USAGE_MODE_GRAPHICS_STORAGE_READ;
}
else if (buffer->usageFlags & REFRESH_BUFFERUSAGE_COMPUTE_STORAGE_READ_BIT)
{
return VULKAN_BUFFER_USAGE_MODE_COMPUTE_STORAGE_READ;
}
else if (buffer->usageFlags & REFRESH_BUFFERUSAGE_COMPUTE_STORAGE_WRITE_BIT)
{
return VULKAN_BUFFER_USAGE_MODE_COMPUTE_STORAGE_READ_WRITE;
}
else
{
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Buffer has no default usage mode!");
return VULKAN_BUFFER_USAGE_MODE_VERTEX_READ;
}
}
static VulkanTextureUsageMode VULKAN_INTERNAL_DefaultTextureUsageMode(
VulkanTexture *texture
) {
/* NOTE: order matters here! */
/* NOTE: graphics storage bits and sampler bit are mutually exclusive! */
if (texture->usageFlags & REFRESH_TEXTUREUSAGE_SAMPLER_BIT)
{
return VULKAN_TEXTURE_USAGE_MODE_SAMPLER;
}
else if (texture->usageFlags & REFRESH_TEXTUREUSAGE_GRAPHICS_STORAGE_READ_BIT)
{
return VULKAN_TEXTURE_USAGE_MODE_GRAPHICS_STORAGE_READ;
}
else if (texture->usageFlags & REFRESH_TEXTUREUSAGE_COLOR_TARGET_BIT)
{
return VULKAN_TEXTURE_USAGE_MODE_COLOR_ATTACHMENT;
}
else if (texture->usageFlags & REFRESH_TEXTUREUSAGE_DEPTH_STENCIL_TARGET_BIT)
{
return VULKAN_TEXTURE_USAGE_MODE_DEPTH_STENCIL_ATTACHMENT;
}
else if (texture->usageFlags & REFRESH_TEXTUREUSAGE_COMPUTE_STORAGE_READ_BIT)
{
return VULKAN_TEXTURE_USAGE_MODE_COMPUTE_STORAGE_READ;
}
else if (texture->usageFlags & REFRESH_TEXTUREUSAGE_COMPUTE_STORAGE_WRITE_BIT)
{
return VULKAN_TEXTURE_USAGE_MODE_COMPUTE_STORAGE_READ_WRITE;
}
else
{
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Texture has no default usage mode!");
return VULKAN_TEXTURE_USAGE_MODE_SAMPLER;
}
}
static void VULKAN_INTERNAL_BufferTransitionFromDefaultUsage(
VulkanRenderer *renderer,
VulkanCommandBuffer *commandBuffer,
VulkanBufferUsageMode destinationUsageMode,
VulkanBuffer *buffer
) {
VULKAN_INTERNAL_BufferMemoryBarrier(
renderer,
commandBuffer,
VULKAN_INTERNAL_DefaultBufferUsageMode(buffer),
destinationUsageMode,
buffer
);
}
static void VULKAN_INTERNAL_BufferTransitionToDefaultUsage(
VulkanRenderer *renderer,
VulkanCommandBuffer *commandBuffer,
VulkanBufferUsageMode sourceUsageMode,
VulkanBuffer *buffer
) {
VULKAN_INTERNAL_BufferMemoryBarrier(
renderer,
commandBuffer,
sourceUsageMode,
VULKAN_INTERNAL_DefaultBufferUsageMode(buffer),
buffer
);
}
static void VULKAN_INTERNAL_TextureTransitionFromDefaultUsage(
VulkanRenderer *renderer,
VulkanCommandBuffer *commandBuffer,
VulkanTextureUsageMode destinationUsageMode,
VulkanTextureSlice *textureSlice
) {
VULKAN_INTERNAL_ImageMemoryBarrier(
renderer,
commandBuffer,
VULKAN_INTERNAL_DefaultTextureUsageMode(textureSlice->parent),
destinationUsageMode,
textureSlice
);
}
static void VULKAN_INTERNAL_TextureTransitionToDefaultUsage(
VulkanRenderer *renderer,
VulkanCommandBuffer *commandBuffer,
VulkanTextureUsageMode sourceUsageMode,
VulkanTextureSlice *textureSlice
) {
VULKAN_INTERNAL_ImageMemoryBarrier(
renderer,
commandBuffer,
sourceUsageMode,
VULKAN_INTERNAL_DefaultTextureUsageMode(textureSlice->parent),
textureSlice
);
}
/* Resource Disposal */
static void VULKAN_INTERNAL_ReleaseFramebuffer(
VulkanRenderer *renderer,
VulkanFramebuffer *framebuffer
) {
SDL_LockMutex(renderer->disposeLock);
EXPAND_ARRAY_IF_NEEDED(
renderer->framebuffersToDestroy,
VulkanFramebuffer*,
renderer->framebuffersToDestroyCount + 1,
renderer->framebuffersToDestroyCapacity,
renderer->framebuffersToDestroyCapacity * 2
)
renderer->framebuffersToDestroy[renderer->framebuffersToDestroyCount] = framebuffer;
renderer->framebuffersToDestroyCount += 1;
SDL_UnlockMutex(renderer->disposeLock);
}
static void VULKAN_INTERNAL_DestroyFramebuffer(
VulkanRenderer *renderer,
VulkanFramebuffer *framebuffer
) {
renderer->vkDestroyFramebuffer(
renderer->logicalDevice,
framebuffer->framebuffer,
NULL
);
SDL_free(framebuffer);
}
static void VULKAN_INTERNAL_RemoveFramebuffersContainingView(
VulkanRenderer *renderer,
VkImageView view
) {
FramebufferHash *hash;
Sint32 i, j;
SDL_LockMutex(renderer->framebufferFetchLock);
for (i = renderer->framebufferHashArray.count - 1; i >= 0; i -= 1)
{
hash = &renderer->framebufferHashArray.elements[i].key;
for (j = 0; j < hash->colorAttachmentCount; j += 1)
{
if (hash->colorAttachmentViews[j] == view)
{
/* FIXME: do we actually need to queue this?
* The framebuffer should not be in use once the associated texture is being destroyed
*/
VULKAN_INTERNAL_ReleaseFramebuffer(
renderer,
renderer->framebufferHashArray.elements[i].value
);
FramebufferHashArray_Remove(
&renderer->framebufferHashArray,
i
);
break;
}
}
}
SDL_UnlockMutex(renderer->framebufferFetchLock);
}
static void VULKAN_INTERNAL_DestroyTexture(
VulkanRenderer* renderer,
VulkanTexture* texture
) {
Uint32 sliceIndex;
/* Clean up slices */
for (sliceIndex = 0; sliceIndex < texture->sliceCount; sliceIndex += 1)
{
if (texture->isRenderTarget)
{
VULKAN_INTERNAL_RemoveFramebuffersContainingView(
renderer,
texture->slices[sliceIndex].view
);
if (texture->slices[sliceIndex].msaaTexHandle != NULL)
{
VULKAN_INTERNAL_DestroyTexture(
renderer,
texture->slices[sliceIndex].msaaTexHandle->vulkanTexture
);
SDL_free(texture->slices[sliceIndex].msaaTexHandle);
}
}
renderer->vkDestroyImageView(
renderer->logicalDevice,
texture->slices[sliceIndex].view,
NULL
);
}
SDL_free(texture->slices);
renderer->vkDestroyImageView(
renderer->logicalDevice,
texture->view,
NULL
);
renderer->vkDestroyImage(
renderer->logicalDevice,
texture->image,
NULL
);
VULKAN_INTERNAL_RemoveMemoryUsedRegion(
renderer,
texture->usedRegion
);
SDL_free(texture);
}
static void VULKAN_INTERNAL_DestroyBuffer(
VulkanRenderer* renderer,
VulkanBuffer* buffer
) {
renderer->vkDestroyBuffer(
renderer->logicalDevice,
buffer->buffer,
NULL
);
VULKAN_INTERNAL_RemoveMemoryUsedRegion(
renderer,
buffer->usedRegion
);
SDL_free(buffer);
}
static void VULKAN_INTERNAL_DestroyBufferContainer(
VulkanRenderer *renderer,
VulkanBufferContainer *bufferContainer
) {
Uint32 i;
SDL_LockMutex(renderer->disposeLock);
for (i = 0; i < bufferContainer->bufferCount; i += 1)
{
VULKAN_INTERNAL_DestroyBuffer(renderer, bufferContainer->bufferHandles[i]->vulkanBuffer);
SDL_free(bufferContainer->bufferHandles[i]);
}
/* Containers are just client handles, so we can free immediately */
if (bufferContainer->debugName != NULL)
{
SDL_free(bufferContainer->debugName);
}
SDL_free(bufferContainer->bufferHandles);
SDL_free(bufferContainer);
SDL_UnlockMutex(renderer->disposeLock);
}
static void VULKAN_INTERNAL_DestroyCommandPool(
VulkanRenderer *renderer,
VulkanCommandPool *commandPool
) {
Uint32 i, j;
VulkanCommandBuffer* commandBuffer;
renderer->vkDestroyCommandPool(
renderer->logicalDevice,
commandPool->commandPool,
NULL
);
for (i = 0; i < commandPool->inactiveCommandBufferCount; i += 1)
{
commandBuffer = commandPool->inactiveCommandBuffers[i];
for (j = 0; j < commandBuffer->initializedVertexUniformBufferCount; j += 1)
{
VULKAN_INTERNAL_DestroyBufferContainer(
renderer,
commandBuffer->vertexUniformBuffers[j]->bufferContainer
);
SDL_free(commandBuffer->vertexUniformBuffers[j]);
}
for (j = 0; j < commandBuffer->initializedFragmentUniformBufferCount; j += 1)
{
VULKAN_INTERNAL_DestroyBufferContainer(
renderer,
commandBuffer->fragmentUniformBuffers[j]->bufferContainer
);
SDL_free(commandBuffer->fragmentUniformBuffers[j]);
}
for (j = 0; j < commandBuffer->initializedComputeUniformBufferCount; j += 1)
{
VULKAN_INTERNAL_DestroyBufferContainer(
renderer,
commandBuffer->computeUniformBuffers[j]->bufferContainer
);
SDL_free(commandBuffer->computeUniformBuffers[j]);
}
SDL_free(commandBuffer->presentDatas);
SDL_free(commandBuffer->waitSemaphores);
SDL_free(commandBuffer->signalSemaphores);
SDL_free(commandBuffer->boundDescriptorSetDatas);
SDL_free(commandBuffer->usedBuffers);
SDL_free(commandBuffer->usedTextureSlices);
SDL_free(commandBuffer->usedSamplers);
SDL_free(commandBuffer->usedGraphicsPipelines);
SDL_free(commandBuffer->usedComputePipelines);
SDL_free(commandBuffer->usedFramebuffers);
SDL_free(commandBuffer);
}
SDL_free(commandPool->inactiveCommandBuffers);
SDL_free(commandPool);
}
static void VULKAN_INTERNAL_DestroyDescriptorSetPool(
VulkanRenderer *renderer,
DescriptorSetPool *pool
) {
Uint32 i;
if (pool == NULL)
{
return;
}
for (i = 0; i < pool->descriptorPoolCount; i += 1)
{
renderer->vkDestroyDescriptorPool(
renderer->logicalDevice,
pool->descriptorPools[i],
NULL
);
}
renderer->vkDestroyDescriptorSetLayout(
renderer->logicalDevice,
pool->descriptorSetLayout,
NULL
);
SDL_free(pool->descriptorInfos);
SDL_free(pool->descriptorPools);
SDL_free(pool->inactiveDescriptorSets);
SDL_DestroyMutex(pool->lock);
}
static void VULKAN_INTERNAL_DestroyGraphicsPipeline(
VulkanRenderer *renderer,
VulkanGraphicsPipeline *graphicsPipeline
) {
Uint32 i;
renderer->vkDestroyPipeline(
renderer->logicalDevice,
graphicsPipeline->pipeline,
NULL
);
renderer->vkDestroyPipelineLayout(
renderer->logicalDevice,
graphicsPipeline->resourceLayout.pipelineLayout,
NULL
);
for (i = 0; i < 4; i += 1)
{
VULKAN_INTERNAL_DestroyDescriptorSetPool(
renderer,
&graphicsPipeline->resourceLayout.descriptorSetPools[i]
);
}
(void)SDL_AtomicDecRef(&graphicsPipeline->vertexShader->referenceCount);
(void)SDL_AtomicDecRef(&graphicsPipeline->fragmentShader->referenceCount);
SDL_free(graphicsPipeline);
}
static void VULKAN_INTERNAL_DestroyComputePipeline(
VulkanRenderer *renderer,
VulkanComputePipeline *computePipeline
) {
Uint32 i;
renderer->vkDestroyPipeline(
renderer->logicalDevice,
computePipeline->pipeline,
NULL
);
renderer->vkDestroyPipelineLayout(
renderer->logicalDevice,
computePipeline->resourceLayout.pipelineLayout,
NULL
);
for (i = 0; i < 3; i += 1)
{
VULKAN_INTERNAL_DestroyDescriptorSetPool(
renderer,
&computePipeline->resourceLayout.descriptorSetPools[i]
);
}
(void)SDL_AtomicDecRef(&computePipeline->computeShader->referenceCount);
SDL_free(computePipeline);
}
static void VULKAN_INTERNAL_DestroyShaderModule(
VulkanRenderer *renderer,
VulkanShader *vulkanShaderModule
) {
renderer->vkDestroyShaderModule(
renderer->logicalDevice,
vulkanShaderModule->shaderModule,
NULL
);
SDL_free(vulkanShaderModule);
}
static void VULKAN_INTERNAL_DestroySampler(
VulkanRenderer *renderer,
VulkanSampler *vulkanSampler
) {
renderer->vkDestroySampler(
renderer->logicalDevice,
vulkanSampler->sampler,
NULL
);
SDL_free(vulkanSampler);
}
static void VULKAN_INTERNAL_DestroySwapchain(
VulkanRenderer* renderer,
WindowData *windowData
) {
Uint32 i;
VulkanSwapchainData *swapchainData;
if (windowData == NULL)
{
return;
}
swapchainData = windowData->swapchainData;
if (swapchainData == NULL)
{
return;
}
for (i = 0; i < swapchainData->imageCount; i += 1)
{
renderer->vkDestroyImageView(
renderer->logicalDevice,
swapchainData->textureContainers[i].activeTextureHandle->vulkanTexture->slices[0].view,
NULL
);
SDL_free(swapchainData->textureContainers[i].activeTextureHandle->vulkanTexture->slices);
renderer->vkDestroyImageView(
renderer->logicalDevice,
swapchainData->textureContainers[i].activeTextureHandle->vulkanTexture->view,
NULL
);
SDL_free(swapchainData->textureContainers[i].activeTextureHandle->vulkanTexture);
SDL_free(swapchainData->textureContainers[i].activeTextureHandle);
}
SDL_free(swapchainData->textureContainers);
renderer->vkDestroySwapchainKHR(
renderer->logicalDevice,
swapchainData->swapchain,
NULL
);
renderer->vkDestroySurfaceKHR(
renderer->instance,
swapchainData->surface,
NULL
);
for (i = 0; i < MAX_FRAMES_IN_FLIGHT; i += 1)
{
renderer->vkDestroySemaphore(
renderer->logicalDevice,
swapchainData->imageAvailableSemaphore[i],
NULL
);
renderer->vkDestroySemaphore(
renderer->logicalDevice,
swapchainData->renderFinishedSemaphore[i],
NULL
);
}
windowData->swapchainData = NULL;
SDL_free(swapchainData);
}
/* Descriptor pool stuff */
static SDL_bool VULKAN_INTERNAL_CreateDescriptorPool(
VulkanRenderer *renderer,
VulkanDescriptorInfo *descriptorInfos,
Uint32 descriptorInfoCount,
Uint32 descriptorSetPoolSize,
VkDescriptorPool *pDescriptorPool
) {
VkDescriptorPoolSize *descriptorPoolSizes;
VkDescriptorPoolCreateInfo descriptorPoolInfo;
VkResult vulkanResult;
Uint32 i;
descriptorPoolSizes = NULL;
if (descriptorInfoCount > 0)
{
descriptorPoolSizes = SDL_stack_alloc(VkDescriptorPoolSize, descriptorInfoCount);
for (i = 0; i < descriptorInfoCount; i += 1)
{
descriptorPoolSizes[i].type = descriptorInfos[i].descriptorType;
descriptorPoolSizes[i].descriptorCount = descriptorSetPoolSize;
}
}
descriptorPoolInfo.sType = VK_STRUCTURE_TYPE_DESCRIPTOR_POOL_CREATE_INFO;
descriptorPoolInfo.pNext = NULL;
descriptorPoolInfo.flags = 0;
descriptorPoolInfo.maxSets = descriptorSetPoolSize;
descriptorPoolInfo.poolSizeCount = descriptorInfoCount;
descriptorPoolInfo.pPoolSizes = descriptorPoolSizes;
vulkanResult = renderer->vkCreateDescriptorPool(
renderer->logicalDevice,
&descriptorPoolInfo,
NULL,
pDescriptorPool
);
SDL_stack_free(descriptorPoolSizes);
if (vulkanResult != VK_SUCCESS)
{
LogVulkanResultAsError("vkCreateDescriptorPool", vulkanResult);
return SDL_FALSE;
}
return SDL_TRUE;
}
static SDL_bool VULKAN_INTERNAL_AllocateDescriptorSets(
VulkanRenderer *renderer,
VkDescriptorPool descriptorPool,
VkDescriptorSetLayout descriptorSetLayout,
Uint32 descriptorSetCount,
VkDescriptorSet *descriptorSetArray
) {
VkDescriptorSetAllocateInfo descriptorSetAllocateInfo;
VkDescriptorSetLayout *descriptorSetLayouts = SDL_stack_alloc(VkDescriptorSetLayout, descriptorSetCount);
VkResult vulkanResult;
Uint32 i;
for (i = 0; i < descriptorSetCount; i += 1)
{
descriptorSetLayouts[i] = descriptorSetLayout;
}
descriptorSetAllocateInfo.sType = VK_STRUCTURE_TYPE_DESCRIPTOR_SET_ALLOCATE_INFO;
descriptorSetAllocateInfo.pNext = NULL;
descriptorSetAllocateInfo.descriptorPool = descriptorPool;
descriptorSetAllocateInfo.descriptorSetCount = descriptorSetCount;
descriptorSetAllocateInfo.pSetLayouts = descriptorSetLayouts;
vulkanResult = renderer->vkAllocateDescriptorSets(
renderer->logicalDevice,
&descriptorSetAllocateInfo,
descriptorSetArray
);
if (vulkanResult != VK_SUCCESS)
{
LogVulkanResultAsError("vkAllocateDescriptorSets", vulkanResult);
SDL_stack_free(descriptorSetLayouts);
return SDL_FALSE;
}
SDL_stack_free(descriptorSetLayouts);
return SDL_TRUE;
}
static void VULKAN_INTERNAL_InitializeDescriptorSetPool(
VulkanRenderer *renderer,
DescriptorSetPool *descriptorSetPool
) {
descriptorSetPool->lock = SDL_CreateMutex();
/* Descriptor set layout and descriptor infos are already set when this function is called */
descriptorSetPool->descriptorPoolCount = 1;
descriptorSetPool->descriptorPools = SDL_malloc(sizeof(VkDescriptorPool));
descriptorSetPool->nextPoolSize = DESCRIPTOR_POOL_STARTING_SIZE * 2;
VULKAN_INTERNAL_CreateDescriptorPool(
renderer,
descriptorSetPool->descriptorInfos,
descriptorSetPool->descriptorInfoCount,
DESCRIPTOR_POOL_STARTING_SIZE,
&descriptorSetPool->descriptorPools[0]
);
descriptorSetPool->inactiveDescriptorSetCapacity = DESCRIPTOR_POOL_STARTING_SIZE;
descriptorSetPool->inactiveDescriptorSetCount = DESCRIPTOR_POOL_STARTING_SIZE;
descriptorSetPool->inactiveDescriptorSets = SDL_malloc(
sizeof(VkDescriptorSet) * DESCRIPTOR_POOL_STARTING_SIZE
);
VULKAN_INTERNAL_AllocateDescriptorSets(
renderer,
descriptorSetPool->descriptorPools[0],
descriptorSetPool->descriptorSetLayout,
DESCRIPTOR_POOL_STARTING_SIZE,
descriptorSetPool->inactiveDescriptorSets
);
}
static SDL_bool VULKAN_INTERNAL_InitializeGraphicsPipelineResourceLayout(
VulkanRenderer *renderer,
Refresh_GraphicsPipelineResourceInfo *vertexResourceInfo,
Refresh_GraphicsPipelineResourceInfo *fragmentResourceInfo,
VulkanGraphicsPipelineResourceLayout *pipelineResourceLayout
) {
VkDescriptorSetLayoutBinding descriptorSetLayoutBindings[MAX_TEXTURE_SAMPLERS_PER_STAGE + MAX_STORAGE_TEXTURES_PER_STAGE + MAX_STORAGE_BUFFERS_PER_STAGE];
VkDescriptorSetLayoutCreateInfo descriptorSetLayoutCreateInfo;
VkDescriptorSetLayout descriptorSetLayouts[4];
VkPipelineLayoutCreateInfo pipelineLayoutCreateInfo;
DescriptorSetPool *descriptorSetPool;
VkResult vulkanResult;
Uint32 i;
pipelineResourceLayout->vertexSamplerCount = vertexResourceInfo->samplerCount;
pipelineResourceLayout->vertexStorageTextureCount = vertexResourceInfo->storageTextureCount;
pipelineResourceLayout->vertexStorageBufferCount = vertexResourceInfo->storageBufferCount;
pipelineResourceLayout->vertexUniformBufferCount = vertexResourceInfo->uniformBufferCount;
pipelineResourceLayout->fragmentSamplerCount = fragmentResourceInfo->samplerCount;
pipelineResourceLayout->fragmentStorageTextureCount = fragmentResourceInfo->storageTextureCount;
pipelineResourceLayout->fragmentStorageBufferCount = fragmentResourceInfo->storageBufferCount;
pipelineResourceLayout->fragmentUniformBufferCount = fragmentResourceInfo->uniformBufferCount;
/* Vertex Resources */
descriptorSetLayoutCreateInfo.sType = VK_STRUCTURE_TYPE_DESCRIPTOR_SET_LAYOUT_CREATE_INFO;
descriptorSetLayoutCreateInfo.pNext = NULL;
descriptorSetLayoutCreateInfo.flags = 0;
descriptorSetLayoutCreateInfo.pBindings = NULL;
descriptorSetLayoutCreateInfo.bindingCount =
vertexResourceInfo->samplerCount +
vertexResourceInfo->storageTextureCount +
vertexResourceInfo->storageBufferCount;
descriptorSetPool = &pipelineResourceLayout->descriptorSetPools[0];
descriptorSetPool->descriptorInfoCount = descriptorSetLayoutCreateInfo.bindingCount;
descriptorSetPool->descriptorInfos = NULL;
if (descriptorSetLayoutCreateInfo.bindingCount > 0)
{
descriptorSetPool->descriptorInfos = SDL_malloc(
descriptorSetPool->descriptorInfoCount * sizeof(VulkanDescriptorInfo)
);
for (i = 0; i < vertexResourceInfo->samplerCount; i += 1)
{
descriptorSetLayoutBindings[i].binding = i;
descriptorSetLayoutBindings[i].descriptorCount = 1;
descriptorSetLayoutBindings[i].descriptorType = VK_DESCRIPTOR_TYPE_COMBINED_IMAGE_SAMPLER;
descriptorSetLayoutBindings[i].stageFlags = VK_SHADER_STAGE_VERTEX_BIT;
descriptorSetLayoutBindings[i].pImmutableSamplers = NULL;
descriptorSetPool->descriptorInfos[i].descriptorType = VK_DESCRIPTOR_TYPE_COMBINED_IMAGE_SAMPLER;
descriptorSetPool->descriptorInfos[i].stageFlag = VK_SHADER_STAGE_VERTEX_BIT;
}
for (i = vertexResourceInfo->samplerCount; i < vertexResourceInfo->samplerCount + vertexResourceInfo->storageTextureCount; i += 1)
{
descriptorSetLayoutBindings[i].binding = i;
descriptorSetLayoutBindings[i].descriptorCount = 1;
descriptorSetLayoutBindings[i].descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_IMAGE;
descriptorSetLayoutBindings[i].stageFlags = VK_SHADER_STAGE_VERTEX_BIT;
descriptorSetLayoutBindings[i].pImmutableSamplers = NULL;
descriptorSetPool->descriptorInfos[i].descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_IMAGE;
descriptorSetPool->descriptorInfos[i].stageFlag = VK_SHADER_STAGE_VERTEX_BIT;
}
for (i = vertexResourceInfo->samplerCount + vertexResourceInfo->storageTextureCount; i < descriptorSetLayoutCreateInfo.bindingCount; i += 1)
{
descriptorSetLayoutBindings[i].binding = i;
descriptorSetLayoutBindings[i].descriptorCount = 1;
descriptorSetLayoutBindings[i].descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_BUFFER;
descriptorSetLayoutBindings[i].stageFlags = VK_SHADER_STAGE_VERTEX_BIT;
descriptorSetLayoutBindings[i].pImmutableSamplers = NULL;
descriptorSetPool->descriptorInfos[i].descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_BUFFER;
descriptorSetPool->descriptorInfos[i].stageFlag = VK_SHADER_STAGE_VERTEX_BIT;
}
descriptorSetLayoutCreateInfo.pBindings = descriptorSetLayoutBindings;
}
vulkanResult = renderer->vkCreateDescriptorSetLayout(
renderer->logicalDevice,
&descriptorSetLayoutCreateInfo,
NULL,
&descriptorSetPool->descriptorSetLayout
);
descriptorSetLayouts[0] = descriptorSetPool->descriptorSetLayout;
if (vulkanResult != VK_SUCCESS)
{
LogVulkanResultAsError("vkCreateDescriptorSetLayout", vulkanResult);
return SDL_FALSE;
}
/* Vertex UBOs */
descriptorSetPool = &pipelineResourceLayout->descriptorSetPools[1];
descriptorSetLayoutCreateInfo.bindingCount = pipelineResourceLayout->vertexUniformBufferCount;
descriptorSetLayoutCreateInfo.pBindings = NULL;
descriptorSetPool->descriptorInfoCount = descriptorSetLayoutCreateInfo.bindingCount;
descriptorSetPool->descriptorInfos = NULL;
if (descriptorSetLayoutCreateInfo.bindingCount > 0)
{
descriptorSetPool->descriptorInfos = SDL_malloc(
descriptorSetPool->descriptorInfoCount * sizeof(VulkanDescriptorInfo)
);
for (i = 0; i < vertexResourceInfo->uniformBufferCount; i += 1)
{
descriptorSetLayoutBindings[i].binding = i;
descriptorSetLayoutBindings[i].descriptorCount = 1;
descriptorSetLayoutBindings[i].descriptorType = VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER_DYNAMIC;
descriptorSetLayoutBindings[i].stageFlags = VK_SHADER_STAGE_VERTEX_BIT;
descriptorSetLayoutBindings[i].pImmutableSamplers = NULL;
descriptorSetPool->descriptorInfos[i].descriptorType = VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER_DYNAMIC;
descriptorSetPool->descriptorInfos[i].stageFlag = VK_SHADER_STAGE_VERTEX_BIT;
}
descriptorSetLayoutCreateInfo.pBindings = descriptorSetLayoutBindings;
}
vulkanResult = renderer->vkCreateDescriptorSetLayout(
renderer->logicalDevice,
&descriptorSetLayoutCreateInfo,
NULL,
&descriptorSetPool->descriptorSetLayout
);
descriptorSetLayouts[1] = descriptorSetPool->descriptorSetLayout;
if (vulkanResult != VK_SUCCESS)
{
LogVulkanResultAsError("vkCreateDescriptorSetLayout", vulkanResult);
return SDL_FALSE;
}
/* Fragment resources */
descriptorSetPool = &pipelineResourceLayout->descriptorSetPools[2];
descriptorSetLayoutCreateInfo.bindingCount =
fragmentResourceInfo->samplerCount +
fragmentResourceInfo->storageTextureCount +
fragmentResourceInfo->storageBufferCount;
descriptorSetLayoutCreateInfo.pBindings = NULL;
descriptorSetPool->descriptorInfoCount = descriptorSetLayoutCreateInfo.bindingCount;
descriptorSetPool->descriptorInfos = NULL;
if (descriptorSetLayoutCreateInfo.bindingCount > 0)
{
descriptorSetPool->descriptorInfos = SDL_malloc(
descriptorSetPool->descriptorInfoCount * sizeof(VulkanDescriptorInfo)
);
for (i = 0; i < fragmentResourceInfo->samplerCount; i += 1)
{
descriptorSetLayoutBindings[i].binding = i;
descriptorSetLayoutBindings[i].descriptorCount = 1;
descriptorSetLayoutBindings[i].descriptorType = VK_DESCRIPTOR_TYPE_COMBINED_IMAGE_SAMPLER;
descriptorSetLayoutBindings[i].stageFlags = VK_SHADER_STAGE_FRAGMENT_BIT;
descriptorSetLayoutBindings[i].pImmutableSamplers = NULL;
descriptorSetPool->descriptorInfos[i].descriptorType = VK_DESCRIPTOR_TYPE_COMBINED_IMAGE_SAMPLER;
descriptorSetPool->descriptorInfos[i].stageFlag = VK_SHADER_STAGE_FRAGMENT_BIT;
}
for (i = fragmentResourceInfo->samplerCount; i < fragmentResourceInfo->samplerCount + fragmentResourceInfo->storageTextureCount; i += 1)
{
descriptorSetLayoutBindings[i].binding = i;
descriptorSetLayoutBindings[i].descriptorCount = 1;
descriptorSetLayoutBindings[i].descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_IMAGE;
descriptorSetLayoutBindings[i].stageFlags = VK_SHADER_STAGE_FRAGMENT_BIT;
descriptorSetLayoutBindings[i].pImmutableSamplers = NULL;
descriptorSetPool->descriptorInfos[i].descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_IMAGE;
descriptorSetPool->descriptorInfos[i].stageFlag = VK_SHADER_STAGE_FRAGMENT_BIT;
}
for (i = fragmentResourceInfo->samplerCount + fragmentResourceInfo->storageTextureCount; i < descriptorSetLayoutCreateInfo.bindingCount; i += 1)
{
descriptorSetLayoutBindings[i].binding = i;
descriptorSetLayoutBindings[i].descriptorCount = 1;
descriptorSetLayoutBindings[i].descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_BUFFER;
descriptorSetLayoutBindings[i].stageFlags = VK_SHADER_STAGE_FRAGMENT_BIT;
descriptorSetLayoutBindings[i].pImmutableSamplers = NULL;
descriptorSetPool->descriptorInfos[i].descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_IMAGE;
descriptorSetPool->descriptorInfos[i].stageFlag = VK_SHADER_STAGE_FRAGMENT_BIT;
}
descriptorSetLayoutCreateInfo.pBindings = descriptorSetLayoutBindings;
}
vulkanResult = renderer->vkCreateDescriptorSetLayout(
renderer->logicalDevice,
&descriptorSetLayoutCreateInfo,
NULL,
&descriptorSetPool->descriptorSetLayout
);
descriptorSetLayouts[2] = descriptorSetPool->descriptorSetLayout;
if (vulkanResult != VK_SUCCESS)
{
LogVulkanResultAsError("vkCreateDescriptorSetLayout", vulkanResult);
return SDL_FALSE;
}
/* Fragment UBOs */
descriptorSetPool = &pipelineResourceLayout->descriptorSetPools[3];
descriptorSetLayoutCreateInfo.bindingCount =
pipelineResourceLayout->fragmentUniformBufferCount;
descriptorSetLayoutCreateInfo.pBindings = NULL;
descriptorSetPool->descriptorInfoCount = descriptorSetLayoutCreateInfo.bindingCount;
descriptorSetPool->descriptorInfos = NULL;
if (descriptorSetLayoutCreateInfo.bindingCount > 0)
{
descriptorSetPool->descriptorInfos = SDL_malloc(
descriptorSetPool->descriptorInfoCount * sizeof(VulkanDescriptorInfo)
);
for (i = 0; i < fragmentResourceInfo->uniformBufferCount; i += 1)
{
descriptorSetLayoutBindings[i].binding = i;
descriptorSetLayoutBindings[i].descriptorCount = 1;
descriptorSetLayoutBindings[i].descriptorType = VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER_DYNAMIC;
descriptorSetLayoutBindings[i].stageFlags = VK_SHADER_STAGE_FRAGMENT_BIT;
descriptorSetLayoutBindings[i].pImmutableSamplers = NULL;
descriptorSetPool->descriptorInfos[i].descriptorType = VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER_DYNAMIC;
descriptorSetPool->descriptorInfos[i].stageFlag = VK_SHADER_STAGE_FRAGMENT_BIT;
}
descriptorSetLayoutCreateInfo.pBindings = descriptorSetLayoutBindings;
}
vulkanResult = renderer->vkCreateDescriptorSetLayout(
renderer->logicalDevice,
&descriptorSetLayoutCreateInfo,
NULL,
&descriptorSetPool->descriptorSetLayout
);
descriptorSetLayouts[3] = descriptorSetPool->descriptorSetLayout;
if (vulkanResult != VK_SUCCESS)
{
LogVulkanResultAsError("vkCreateDescriptorSetLayout", vulkanResult);
return SDL_FALSE;
}
/* Create the pipeline layout */
pipelineLayoutCreateInfo.sType = VK_STRUCTURE_TYPE_PIPELINE_LAYOUT_CREATE_INFO;
pipelineLayoutCreateInfo.pNext = NULL;
pipelineLayoutCreateInfo.flags = 0;
pipelineLayoutCreateInfo.setLayoutCount = 4;
pipelineLayoutCreateInfo.pSetLayouts = descriptorSetLayouts;
pipelineLayoutCreateInfo.pushConstantRangeCount = 0;
pipelineLayoutCreateInfo.pPushConstantRanges = NULL;
vulkanResult = renderer->vkCreatePipelineLayout(
renderer->logicalDevice,
&pipelineLayoutCreateInfo,
NULL,
&pipelineResourceLayout->pipelineLayout
);
if (vulkanResult != VK_SUCCESS)
{
LogVulkanResultAsError("vkCreatePipelineLayout", vulkanResult);
return SDL_FALSE;
}
for (i = 0; i < 4; i += 1)
{
VULKAN_INTERNAL_InitializeDescriptorSetPool(
renderer,
&pipelineResourceLayout->descriptorSetPools[i]
);
}
return SDL_TRUE;
}
static SDL_bool VULKAN_INTERNAL_InitializeComputePipelineResourceLayout(
VulkanRenderer *renderer,
Refresh_ComputePipelineResourceInfo *resourceLayoutInfo,
VulkanComputePipelineResourceLayout *pipelineResourceLayout
) {
VkDescriptorSetLayoutBinding descriptorSetLayoutBindings[MAX_UNIFORM_BUFFERS_PER_STAGE];
VkDescriptorSetLayoutCreateInfo descriptorSetLayoutCreateInfo;
VkDescriptorSetLayout descriptorSetLayouts[3];
VkPipelineLayoutCreateInfo pipelineLayoutCreateInfo;
DescriptorSetPool *descriptorSetPool;
VkResult vulkanResult;
Uint32 i;
pipelineResourceLayout->readOnlyStorageTextureCount = resourceLayoutInfo->readOnlyStorageTextureCount;
pipelineResourceLayout->readOnlyStorageBufferCount = resourceLayoutInfo->readOnlyStorageBufferCount;
pipelineResourceLayout->readWriteStorageTextureCount = resourceLayoutInfo->readWriteStorageTextureCount;
pipelineResourceLayout->readWriteStorageBufferCount = resourceLayoutInfo->readWriteStorageBufferCount;
pipelineResourceLayout->uniformBufferCount = resourceLayoutInfo->uniformBufferCount;
/* Read-only resources */
descriptorSetLayoutCreateInfo.sType = VK_STRUCTURE_TYPE_DESCRIPTOR_SET_LAYOUT_CREATE_INFO;
descriptorSetLayoutCreateInfo.pNext = NULL;
descriptorSetLayoutCreateInfo.flags = 0;
descriptorSetLayoutCreateInfo.pBindings = NULL;
descriptorSetLayoutCreateInfo.bindingCount =
resourceLayoutInfo->readOnlyStorageTextureCount +
resourceLayoutInfo->readOnlyStorageBufferCount;
descriptorSetPool = &pipelineResourceLayout->descriptorSetPools[0];
descriptorSetPool->descriptorInfoCount = descriptorSetLayoutCreateInfo.bindingCount;
descriptorSetPool->descriptorInfos = NULL;
if (descriptorSetLayoutCreateInfo.bindingCount > 0)
{
descriptorSetPool->descriptorInfos = SDL_malloc(
descriptorSetPool->descriptorInfoCount * sizeof(VulkanDescriptorInfo)
);
for (i = 0; i < resourceLayoutInfo->readOnlyStorageTextureCount; i += 1)
{
descriptorSetLayoutBindings[i].binding = i;
descriptorSetLayoutBindings[i].descriptorCount = 1;
descriptorSetLayoutBindings[i].descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_IMAGE;
descriptorSetLayoutBindings[i].stageFlags = VK_SHADER_STAGE_COMPUTE_BIT;
descriptorSetLayoutBindings[i].pImmutableSamplers = NULL;
descriptorSetPool->descriptorInfos[i].descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_IMAGE;
descriptorSetPool->descriptorInfos[i].stageFlag = VK_SHADER_STAGE_COMPUTE_BIT;
}
for (i = resourceLayoutInfo->readOnlyStorageTextureCount; i < descriptorSetLayoutCreateInfo.bindingCount; i += 1)
{
descriptorSetLayoutBindings[i].binding = i;
descriptorSetLayoutBindings[i].descriptorCount = 1;
descriptorSetLayoutBindings[i].descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_BUFFER;
descriptorSetLayoutBindings[i].stageFlags = VK_SHADER_STAGE_COMPUTE_BIT;
descriptorSetLayoutBindings[i].pImmutableSamplers = NULL;
descriptorSetPool->descriptorInfos[i].descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_BUFFER;
descriptorSetPool->descriptorInfos[i].stageFlag = VK_SHADER_STAGE_COMPUTE_BIT;
}
descriptorSetLayoutCreateInfo.pBindings = descriptorSetLayoutBindings;
}
vulkanResult = renderer->vkCreateDescriptorSetLayout(
renderer->logicalDevice,
&descriptorSetLayoutCreateInfo,
NULL,
&descriptorSetPool->descriptorSetLayout
);
descriptorSetLayouts[0] = descriptorSetPool->descriptorSetLayout;
if (vulkanResult != VK_SUCCESS)
{
LogVulkanResultAsError("vkCreateDescriptorSetLayout", vulkanResult);
return SDL_FALSE;
}
/* Read-write resources */
descriptorSetLayoutCreateInfo.bindingCount =
resourceLayoutInfo->readWriteStorageTextureCount +
resourceLayoutInfo->readWriteStorageBufferCount;
descriptorSetLayoutCreateInfo.pBindings = NULL;
descriptorSetPool = &pipelineResourceLayout->descriptorSetPools[1];
descriptorSetPool->descriptorInfoCount = descriptorSetLayoutCreateInfo.bindingCount;
descriptorSetPool->descriptorInfos = NULL;
if (descriptorSetLayoutCreateInfo.bindingCount > 0)
{
descriptorSetPool->descriptorInfos = SDL_malloc(
descriptorSetPool->descriptorInfoCount * sizeof(VulkanDescriptorInfo)
);
for (i = 0; i < resourceLayoutInfo->readWriteStorageTextureCount; i += 1)
{
descriptorSetLayoutBindings[i].binding = i;
descriptorSetLayoutBindings[i].descriptorCount = 1;
descriptorSetLayoutBindings[i].descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_IMAGE;
descriptorSetLayoutBindings[i].stageFlags = VK_SHADER_STAGE_COMPUTE_BIT;
descriptorSetLayoutBindings[i].pImmutableSamplers = NULL;
descriptorSetPool->descriptorInfos[i].descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_IMAGE;
descriptorSetPool->descriptorInfos[i].stageFlag = VK_SHADER_STAGE_COMPUTE_BIT;
}
for (i = resourceLayoutInfo->readWriteStorageTextureCount; i < descriptorSetLayoutCreateInfo.bindingCount; i += 1)
{
descriptorSetLayoutBindings[i].binding = i;
descriptorSetLayoutBindings[i].descriptorCount = 1;
descriptorSetLayoutBindings[i].descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_BUFFER;
descriptorSetLayoutBindings[i].stageFlags = VK_SHADER_STAGE_COMPUTE_BIT;
descriptorSetLayoutBindings[i].pImmutableSamplers = NULL;
descriptorSetPool->descriptorInfos[i].descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_BUFFER;
descriptorSetPool->descriptorInfos[i].stageFlag = VK_SHADER_STAGE_COMPUTE_BIT;
}
descriptorSetLayoutCreateInfo.pBindings = descriptorSetLayoutBindings;
}
vulkanResult = renderer->vkCreateDescriptorSetLayout(
renderer->logicalDevice,
&descriptorSetLayoutCreateInfo,
NULL,
&descriptorSetPool->descriptorSetLayout
);
descriptorSetLayouts[1] = descriptorSetPool->descriptorSetLayout;
if (vulkanResult != VK_SUCCESS)
{
LogVulkanResultAsError("vkCreateDescriptorSetLayout", vulkanResult);
return SDL_FALSE;
}
/* Uniform buffers */
descriptorSetPool = &pipelineResourceLayout->descriptorSetPools[2];
descriptorSetLayoutCreateInfo.bindingCount = resourceLayoutInfo->uniformBufferCount;
descriptorSetLayoutCreateInfo.pBindings = NULL;
descriptorSetPool->descriptorInfoCount = descriptorSetLayoutCreateInfo.bindingCount;
descriptorSetPool->descriptorInfos = NULL;
if (descriptorSetLayoutCreateInfo.bindingCount > 0)
{
descriptorSetPool->descriptorInfos = SDL_malloc(
descriptorSetPool->descriptorInfoCount * sizeof(VulkanDescriptorInfo)
);
for (i = 0; i < resourceLayoutInfo->uniformBufferCount; i += 1)
{
descriptorSetLayoutBindings[i].binding = i;
descriptorSetLayoutBindings[i].descriptorCount = 1;
descriptorSetLayoutBindings[i].descriptorType = VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER_DYNAMIC;
descriptorSetLayoutBindings[i].stageFlags = VK_SHADER_STAGE_COMPUTE_BIT;
descriptorSetLayoutBindings[i].pImmutableSamplers = NULL;
descriptorSetPool->descriptorInfos[i].descriptorType = VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER_DYNAMIC;
descriptorSetPool->descriptorInfos[i].stageFlag = VK_SHADER_STAGE_COMPUTE_BIT;
}
descriptorSetLayoutCreateInfo.pBindings = descriptorSetLayoutBindings;
}
vulkanResult = renderer->vkCreateDescriptorSetLayout(
renderer->logicalDevice,
&descriptorSetLayoutCreateInfo,
NULL,
&descriptorSetPool->descriptorSetLayout
);
descriptorSetLayouts[2] = descriptorSetPool->descriptorSetLayout;
if (vulkanResult != VK_SUCCESS)
{
LogVulkanResultAsError("vkCreateDescriptorSetLayout", vulkanResult);
return SDL_FALSE;
}
/* Create the pipeline layout */
pipelineLayoutCreateInfo.sType = VK_STRUCTURE_TYPE_PIPELINE_LAYOUT_CREATE_INFO;
pipelineLayoutCreateInfo.pNext = NULL;
pipelineLayoutCreateInfo.flags = 0;
pipelineLayoutCreateInfo.setLayoutCount = 3;
pipelineLayoutCreateInfo.pSetLayouts = descriptorSetLayouts;
pipelineLayoutCreateInfo.pushConstantRangeCount = 0;
pipelineLayoutCreateInfo.pPushConstantRanges = NULL;
vulkanResult = renderer->vkCreatePipelineLayout(
renderer->logicalDevice,
&pipelineLayoutCreateInfo,
NULL,
&pipelineResourceLayout->pipelineLayout
);
if (vulkanResult != VK_SUCCESS)
{
LogVulkanResultAsError("vkCreatePipelineLayout", vulkanResult);
return SDL_FALSE;
}
for (i = 0; i < 3; i += 1)
{
VULKAN_INTERNAL_InitializeDescriptorSetPool(
renderer,
&pipelineResourceLayout->descriptorSetPools[i]
);
}
return SDL_TRUE;
}
/* Data Buffer */
static VulkanBuffer* VULKAN_INTERNAL_CreateBuffer(
VulkanRenderer *renderer,
VkDeviceSize size,
Refresh_BufferUsageFlags usageFlags,
VulkanBufferType type
) {
VulkanBuffer* buffer;
VkResult vulkanResult;
VkBufferCreateInfo bufferCreateInfo;
VkBufferUsageFlags vulkanUsageFlags = 0;
Uint8 bindResult;
if (usageFlags & REFRESH_BUFFERUSAGE_VERTEX_BIT)
{
vulkanUsageFlags |= VK_BUFFER_USAGE_VERTEX_BUFFER_BIT;
}
if (usageFlags & REFRESH_BUFFERUSAGE_INDEX_BIT)
{
vulkanUsageFlags |= VK_BUFFER_USAGE_INDEX_BUFFER_BIT;
}
if (usageFlags & (
REFRESH_BUFFERUSAGE_GRAPHICS_STORAGE_READ_BIT |
REFRESH_BUFFERUSAGE_COMPUTE_STORAGE_READ_BIT |
REFRESH_BUFFERUSAGE_COMPUTE_STORAGE_WRITE_BIT
)) {
vulkanUsageFlags |= VK_BUFFER_USAGE_STORAGE_BUFFER_BIT;
}
if (usageFlags & REFRESH_BUFFERUSAGE_INDIRECT_BIT)
{
vulkanUsageFlags |= VK_BUFFER_USAGE_INDIRECT_BUFFER_BIT;
}
if (type == VULKAN_BUFFER_TYPE_UNIFORM)
{
vulkanUsageFlags |= VK_BUFFER_USAGE_UNIFORM_BUFFER_BIT;
}
else
{
/* GPU buffers need transfer bits for defrag, transfer buffers need them for transfers */
vulkanUsageFlags |= VK_BUFFER_USAGE_TRANSFER_SRC_BIT | VK_BUFFER_USAGE_TRANSFER_DST_BIT;
}
buffer = SDL_malloc(sizeof(VulkanBuffer));
buffer->size = size;
buffer->usageFlags = usageFlags;
buffer->type = type;
buffer->markedForDestroy = 0;
buffer->transitioned = SDL_FALSE;
bufferCreateInfo.sType = VK_STRUCTURE_TYPE_BUFFER_CREATE_INFO;
bufferCreateInfo.pNext = NULL;
bufferCreateInfo.flags = 0;
bufferCreateInfo.size = size;
bufferCreateInfo.usage = vulkanUsageFlags;
bufferCreateInfo.sharingMode = VK_SHARING_MODE_EXCLUSIVE;
bufferCreateInfo.queueFamilyIndexCount = 1;
bufferCreateInfo.pQueueFamilyIndices = &renderer->queueFamilyIndex;
/* Set transfer bits so we can defrag */
bufferCreateInfo.usage |= VK_BUFFER_USAGE_TRANSFER_SRC_BIT | VK_BUFFER_USAGE_TRANSFER_DST_BIT;
vulkanResult = renderer->vkCreateBuffer(
renderer->logicalDevice,
&bufferCreateInfo,
NULL,
&buffer->buffer
);
VULKAN_ERROR_CHECK(vulkanResult, vkCreateBuffer, 0)
bindResult = VULKAN_INTERNAL_BindMemoryForBuffer(
renderer,
buffer->buffer,
buffer->size,
buffer->type,
&buffer->usedRegion
);
if (bindResult != 1)
{
renderer->vkDestroyBuffer(
renderer->logicalDevice,
buffer->buffer,
NULL);
return NULL;
}
buffer->usedRegion->vulkanBuffer = buffer; /* lol */
buffer->handle = NULL;
SDL_AtomicSet(&buffer->referenceCount, 0);
return buffer;
}
/* Indirection so we can cleanly defrag buffers */
static VulkanBufferHandle* VULKAN_INTERNAL_CreateBufferHandle(
VulkanRenderer *renderer,
Uint32 sizeInBytes,
Refresh_BufferUsageFlags usageFlags,
VulkanBufferType type
) {
VulkanBufferHandle* bufferHandle;
VulkanBuffer* buffer;
buffer = VULKAN_INTERNAL_CreateBuffer(
renderer,
sizeInBytes,
usageFlags,
type
);
if (buffer == NULL)
{
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Failed to create buffer!");
return NULL;
}
bufferHandle = SDL_malloc(sizeof(VulkanBufferHandle));
bufferHandle->vulkanBuffer = buffer;
bufferHandle->container = NULL;
buffer->handle = bufferHandle;
return bufferHandle;
}
static VulkanBufferContainer* VULKAN_INTERNAL_CreateBufferContainer(
VulkanRenderer *renderer,
Uint32 sizeInBytes,
Refresh_BufferUsageFlags usageFlags,
VulkanBufferType type
) {
VulkanBufferContainer *bufferContainer;
VulkanBufferHandle *bufferHandle;
bufferHandle = VULKAN_INTERNAL_CreateBufferHandle(
renderer,
sizeInBytes,
usageFlags,
type
);
if (bufferHandle == NULL)
{
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Failed to create buffer container!");
return NULL;
}
bufferContainer = SDL_malloc(sizeof(VulkanBufferContainer));
bufferContainer->activeBufferHandle = bufferHandle;
bufferHandle->container = bufferContainer;
bufferContainer->bufferCapacity = 1;
bufferContainer->bufferCount = 1;
bufferContainer->bufferHandles = SDL_malloc(
bufferContainer->bufferCapacity * sizeof(VulkanBufferHandle*)
);
bufferContainer->bufferHandles[0] = bufferContainer->activeBufferHandle;
bufferContainer->debugName = NULL;
return bufferContainer;
}
/* Texture Slice Utilities */
static Uint32 VULKAN_INTERNAL_GetTextureSliceIndex(
VulkanTexture *texture,
Uint32 layer,
Uint32 level
) {
return (layer * texture->levelCount) + level;
}
static VulkanTextureSlice* VULKAN_INTERNAL_FetchTextureSlice(
VulkanTexture *texture,
Uint32 layer,
Uint32 level
) {
return &texture->slices[
VULKAN_INTERNAL_GetTextureSliceIndex(
texture,
layer,
level
)
];
}
static VulkanTextureSlice* VULKAN_INTERNAL_SDLToVulkanTextureSlice(
Refresh_TextureSlice *textureSlice
) {
return VULKAN_INTERNAL_FetchTextureSlice(
((VulkanTextureContainer*) textureSlice->texture)->activeTextureHandle->vulkanTexture,
textureSlice->layer,
textureSlice->mipLevel
);
}
static void VULKAN_INTERNAL_CreateSliceView(
VulkanRenderer *renderer,
VulkanTexture *texture,
Uint32 layer,
Uint32 level,
VkComponentMapping swizzle,
VkImageView *pView
) {
VkResult vulkanResult;
VkImageViewCreateInfo imageViewCreateInfo;
/* create framebuffer compatible views for RenderTarget */
imageViewCreateInfo.sType = VK_STRUCTURE_TYPE_IMAGE_VIEW_CREATE_INFO;
imageViewCreateInfo.pNext = NULL;
imageViewCreateInfo.flags = 0;
imageViewCreateInfo.image = texture->image;
imageViewCreateInfo.format = texture->format;
imageViewCreateInfo.components = swizzle;
imageViewCreateInfo.subresourceRange.aspectMask = texture->aspectFlags;
imageViewCreateInfo.subresourceRange.baseMipLevel = level;
imageViewCreateInfo.subresourceRange.levelCount = 1;
imageViewCreateInfo.subresourceRange.baseArrayLayer = layer;
imageViewCreateInfo.subresourceRange.layerCount = 1;
imageViewCreateInfo.viewType = VK_IMAGE_VIEW_TYPE_2D;
vulkanResult = renderer->vkCreateImageView(
renderer->logicalDevice,
&imageViewCreateInfo,
NULL,
pView
);
if (vulkanResult != VK_SUCCESS)
{
LogVulkanResultAsError(
"vkCreateImageView",
vulkanResult
);
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Failed to create color attachment image view");
*pView = (VkImageView) VK_NULL_HANDLE;
return;
}
}
/* Swapchain */
static Uint8 VULKAN_INTERNAL_QuerySwapchainSupport(
VulkanRenderer *renderer,
VkPhysicalDevice physicalDevice,
VkSurfaceKHR surface,
SwapchainSupportDetails *outputDetails
) {
VkResult result;
VkBool32 supportsPresent;
renderer->vkGetPhysicalDeviceSurfaceSupportKHR(
physicalDevice,
renderer->queueFamilyIndex,
surface,
&supportsPresent
);
if (!supportsPresent)
{
SDL_LogWarn(SDL_LOG_CATEGORY_APPLICATION, "This surface does not support presenting!");
return 0;
}
/* Initialize these in case anything fails */
outputDetails->formatsLength = 0;
outputDetails->presentModesLength = 0;
/* Run the device surface queries */
result = renderer->vkGetPhysicalDeviceSurfaceCapabilitiesKHR(
physicalDevice,
surface,
&outputDetails->capabilities
);
VULKAN_ERROR_CHECK(result, vkGetPhysicalDeviceSurfaceCapabilitiesKHR, 0)
if (!(outputDetails->capabilities.supportedCompositeAlpha & VK_COMPOSITE_ALPHA_OPAQUE_BIT_KHR))
{
SDL_LogWarn(SDL_LOG_CATEGORY_APPLICATION, "Opaque presentation unsupported! Expect weird transparency bugs!");
}
result = renderer->vkGetPhysicalDeviceSurfaceFormatsKHR(
physicalDevice,
surface,
&outputDetails->formatsLength,
NULL
);
VULKAN_ERROR_CHECK(result, vkGetPhysicalDeviceSurfaceFormatsKHR, 0)
result = renderer->vkGetPhysicalDeviceSurfacePresentModesKHR(
physicalDevice,
surface,
&outputDetails->presentModesLength,
NULL
);
VULKAN_ERROR_CHECK(result, vkGetPhysicalDeviceSurfacePresentModesKHR, 0)
/* Generate the arrays, if applicable */
outputDetails->formats = NULL;
if (outputDetails->formatsLength != 0)
{
outputDetails->formats = (VkSurfaceFormatKHR*) SDL_malloc(
sizeof(VkSurfaceFormatKHR) * outputDetails->formatsLength
);
if (!outputDetails->formats)
{
SDL_OutOfMemory();
return 0;
}
result = renderer->vkGetPhysicalDeviceSurfaceFormatsKHR(
physicalDevice,
surface,
&outputDetails->formatsLength,
outputDetails->formats
);
if (result != VK_SUCCESS)
{
SDL_LogError(
SDL_LOG_CATEGORY_APPLICATION,
"vkGetPhysicalDeviceSurfaceFormatsKHR: %s",
VkErrorMessages(result)
);
SDL_free(outputDetails->formats);
return 0;
}
}
outputDetails->presentModes = NULL;
if (outputDetails->presentModesLength != 0)
{
outputDetails->presentModes = (VkPresentModeKHR*) SDL_malloc(
sizeof(VkPresentModeKHR) * outputDetails->presentModesLength
);
if (!outputDetails->presentModes)
{
SDL_OutOfMemory();
return 0;
}
result = renderer->vkGetPhysicalDeviceSurfacePresentModesKHR(
physicalDevice,
surface,
&outputDetails->presentModesLength,
outputDetails->presentModes
);
if (result != VK_SUCCESS)
{
SDL_LogError(
SDL_LOG_CATEGORY_APPLICATION,
"vkGetPhysicalDeviceSurfacePresentModesKHR: %s",
VkErrorMessages(result)
);
SDL_free(outputDetails->formats);
SDL_free(outputDetails->presentModes);
return 0;
}
}
/* If we made it here, all the queries were successful. This does NOT
* necessarily mean there are any supported formats or present modes!
*/
return 1;
}
static SDL_bool VULKAN_INTERNAL_VerifySwapSurfaceFormat(
VkFormat desiredFormat,
VkColorSpaceKHR desiredColorSpace,
VkSurfaceFormatKHR *availableFormats,
Uint32 availableFormatsLength
) {
Uint32 i;
for (i = 0; i < availableFormatsLength; i += 1)
{
if ( availableFormats[i].format == desiredFormat &&
availableFormats[i].colorSpace == desiredColorSpace )
{
return SDL_TRUE;
}
}
return SDL_FALSE;
}
static SDL_bool VULKAN_INTERNAL_VerifySwapPresentMode(
VkPresentModeKHR presentMode,
VkPresentModeKHR *availablePresentModes,
Uint32 availablePresentModesLength
) {
Uint32 i;
for (i = 0; i < availablePresentModesLength; i += 1)
{
if (availablePresentModes[i] == presentMode)
{
return SDL_TRUE;
}
}
return SDL_FALSE;
}
static Uint8 VULKAN_INTERNAL_CreateSwapchain(
VulkanRenderer *renderer,
WindowData *windowData
) {
VkResult vulkanResult;
VulkanSwapchainData *swapchainData;
VkSwapchainCreateInfoKHR swapchainCreateInfo;
VkImage *swapchainImages;
VkImageViewCreateInfo imageViewCreateInfo;
VkSemaphoreCreateInfo semaphoreCreateInfo;
SwapchainSupportDetails swapchainSupportDetails;
Sint32 drawableWidth, drawableHeight;
Uint32 i;
swapchainData = SDL_malloc(sizeof(VulkanSwapchainData));
swapchainData->frameCounter = 0;
/* Each swapchain must have its own surface. */
if (!SDL_Vulkan_CreateSurface(
windowData->window,
renderer->instance,
&swapchainData->surface
)) {
SDL_free(swapchainData);
SDL_LogError(
SDL_LOG_CATEGORY_APPLICATION,
"Vulkan_CreateSurface failed: %s",
SDL_GetError()
);
return 0;
}
if (!VULKAN_INTERNAL_QuerySwapchainSupport(
renderer,
renderer->physicalDevice,
swapchainData->surface,
&swapchainSupportDetails
)) {
renderer->vkDestroySurfaceKHR(
renderer->instance,
swapchainData->surface,
NULL
);
if (swapchainSupportDetails.formatsLength > 0)
{
SDL_free(swapchainSupportDetails.formats);
}
if (swapchainSupportDetails.presentModesLength > 0)
{
SDL_free(swapchainSupportDetails.presentModes);
}
SDL_free(swapchainData);
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Device does not support swap chain creation");
return 0;
}
if ( swapchainSupportDetails.capabilities.currentExtent.width == 0 ||
swapchainSupportDetails.capabilities.currentExtent.height == 0)
{
/* Not an error, just minimize behavior! */
renderer->vkDestroySurfaceKHR(
renderer->instance,
swapchainData->surface,
NULL
);
if (swapchainSupportDetails.formatsLength > 0)
{
SDL_free(swapchainSupportDetails.formats);
}
if (swapchainSupportDetails.presentModesLength > 0)
{
SDL_free(swapchainSupportDetails.presentModes);
}
SDL_free(swapchainData);
return 0;
}
swapchainData->format = SwapchainCompositionToFormat[windowData->swapchainComposition];
swapchainData->colorSpace = SwapchainCompositionToColorSpace[windowData->swapchainComposition];
swapchainData->swapchainSwizzle.r = VK_COMPONENT_SWIZZLE_IDENTITY;
swapchainData->swapchainSwizzle.g = VK_COMPONENT_SWIZZLE_IDENTITY;
swapchainData->swapchainSwizzle.b = VK_COMPONENT_SWIZZLE_IDENTITY;
swapchainData->swapchainSwizzle.a = VK_COMPONENT_SWIZZLE_IDENTITY;
if (!VULKAN_INTERNAL_VerifySwapSurfaceFormat(
swapchainData->format,
swapchainData->colorSpace,
swapchainSupportDetails.formats,
swapchainSupportDetails.formatsLength
)) {
if (!BGRToRGBSwapchainFormat(
swapchainData->format,
&swapchainData->format
)) {
renderer->vkDestroySurfaceKHR(
renderer->instance,
swapchainData->surface,
NULL
);
if (swapchainSupportDetails.formatsLength > 0)
{
SDL_free(swapchainSupportDetails.formats);
}
if (swapchainSupportDetails.presentModesLength > 0)
{
SDL_free(swapchainSupportDetails.presentModes);
}
SDL_free(swapchainData);
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Device does not support requested colorspace!");
return 0;
}
if (!VULKAN_INTERNAL_VerifySwapSurfaceFormat(
swapchainData->format,
swapchainData->colorSpace,
swapchainSupportDetails.formats,
swapchainSupportDetails.formatsLength
)) {
renderer->vkDestroySurfaceKHR(
renderer->instance,
swapchainData->surface,
NULL
);
if (swapchainSupportDetails.formatsLength > 0)
{
SDL_free(swapchainSupportDetails.formats);
}
if (swapchainSupportDetails.presentModesLength > 0)
{
SDL_free(swapchainSupportDetails.presentModes);
}
SDL_free(swapchainData);
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Device does not support requested colorspace!");
return 0;
}
}
if (!VULKAN_INTERNAL_VerifySwapPresentMode(
SDLToVK_PresentMode[windowData->presentMode],
swapchainSupportDetails.presentModes,
swapchainSupportDetails.presentModesLength
)) {
renderer->vkDestroySurfaceKHR(
renderer->instance,
swapchainData->surface,
NULL
);
if (swapchainSupportDetails.formatsLength > 0)
{
SDL_free(swapchainSupportDetails.formats);
}
if (swapchainSupportDetails.presentModesLength > 0)
{
SDL_free(swapchainSupportDetails.presentModes);
}
SDL_free(swapchainData);
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Device does not support requested present mode!");
return 0;
}
swapchainData->presentMode = SDLToVK_PresentMode[windowData->presentMode];
/* Should sync now to be sure that our swapchain size is correct */
/* TODO: how to backport SyncWindow? */
SDL_GetWindowSizeInPixels(
windowData->window,
&drawableWidth,
&drawableHeight
);
if ( drawableWidth < swapchainSupportDetails.capabilities.minImageExtent.width ||
drawableWidth > swapchainSupportDetails.capabilities.maxImageExtent.width ||
drawableHeight < swapchainSupportDetails.capabilities.minImageExtent.height ||
drawableHeight > swapchainSupportDetails.capabilities.maxImageExtent.height )
{
if (swapchainSupportDetails.capabilities.currentExtent.width != UINT32_MAX)
{
drawableWidth = VULKAN_INTERNAL_clamp(
drawableWidth,
swapchainSupportDetails.capabilities.minImageExtent.width,
swapchainSupportDetails.capabilities.maxImageExtent.width
);
drawableHeight = VULKAN_INTERNAL_clamp(
drawableHeight,
swapchainSupportDetails.capabilities.minImageExtent.height,
swapchainSupportDetails.capabilities.maxImageExtent.height
);
}
else
{
renderer->vkDestroySurfaceKHR(
renderer->instance,
swapchainData->surface,
NULL
);
if (swapchainSupportDetails.formatsLength > 0)
{
SDL_free(swapchainSupportDetails.formats);
}
if (swapchainSupportDetails.presentModesLength > 0)
{
SDL_free(swapchainSupportDetails.presentModes);
}
SDL_free(swapchainData);
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "No fallback swapchain size available!");
return 0;
}
}
swapchainData->extent.width = drawableWidth;
swapchainData->extent.height = drawableHeight;
swapchainData->imageCount = swapchainSupportDetails.capabilities.minImageCount + 1;
if ( swapchainSupportDetails.capabilities.maxImageCount > 0 &&
swapchainData->imageCount > swapchainSupportDetails.capabilities.maxImageCount )
{
swapchainData->imageCount = swapchainSupportDetails.capabilities.maxImageCount;
}
if (swapchainData->presentMode == VK_PRESENT_MODE_MAILBOX_KHR)
{
/* Required for proper triple-buffering.
* Note that this is below the above maxImageCount check!
* If the driver advertises MAILBOX but does not support 3 swap
* images, it's not real mailbox support, so let it fail hard.
* -flibit
*/
swapchainData->imageCount = SDL_max(swapchainData->imageCount, 3);
}
swapchainCreateInfo.sType = VK_STRUCTURE_TYPE_SWAPCHAIN_CREATE_INFO_KHR;
swapchainCreateInfo.pNext = NULL;
swapchainCreateInfo.flags = 0;
swapchainCreateInfo.surface = swapchainData->surface;
swapchainCreateInfo.minImageCount = swapchainData->imageCount;
swapchainCreateInfo.imageFormat = swapchainData->format;
swapchainCreateInfo.imageColorSpace = swapchainData->colorSpace;
swapchainCreateInfo.imageExtent = swapchainData->extent;
swapchainCreateInfo.imageArrayLayers = 1;
swapchainCreateInfo.imageUsage =
VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT |
VK_IMAGE_USAGE_TRANSFER_SRC_BIT |
VK_IMAGE_USAGE_TRANSFER_DST_BIT;
swapchainCreateInfo.imageSharingMode = VK_SHARING_MODE_EXCLUSIVE;
swapchainCreateInfo.queueFamilyIndexCount = 0;
swapchainCreateInfo.pQueueFamilyIndices = NULL;
swapchainCreateInfo.preTransform = swapchainSupportDetails.capabilities.currentTransform;
swapchainCreateInfo.compositeAlpha = VK_COMPOSITE_ALPHA_OPAQUE_BIT_KHR;
swapchainCreateInfo.presentMode = swapchainData->presentMode;
swapchainCreateInfo.clipped = VK_TRUE;
swapchainCreateInfo.oldSwapchain = VK_NULL_HANDLE;
vulkanResult = renderer->vkCreateSwapchainKHR(
renderer->logicalDevice,
&swapchainCreateInfo,
NULL,
&swapchainData->swapchain
);
if (swapchainSupportDetails.formatsLength > 0)
{
SDL_free(swapchainSupportDetails.formats);
}
if (swapchainSupportDetails.presentModesLength > 0)
{
SDL_free(swapchainSupportDetails.presentModes);
}
if (vulkanResult != VK_SUCCESS)
{
renderer->vkDestroySurfaceKHR(
renderer->instance,
swapchainData->surface,
NULL
);
SDL_free(swapchainData);
LogVulkanResultAsError("vkCreateSwapchainKHR", vulkanResult);
return 0;
}
renderer->vkGetSwapchainImagesKHR(
renderer->logicalDevice,
swapchainData->swapchain,
&swapchainData->imageCount,
NULL
);
swapchainData->textureContainers = SDL_malloc(
sizeof(VulkanTextureContainer) * swapchainData->imageCount
);
if (!swapchainData->textureContainers)
{
SDL_OutOfMemory();
renderer->vkDestroySurfaceKHR(
renderer->instance,
swapchainData->surface,
NULL
);
SDL_free(swapchainData);
return 0;
}
swapchainImages = SDL_stack_alloc(VkImage, swapchainData->imageCount);
renderer->vkGetSwapchainImagesKHR(
renderer->logicalDevice,
swapchainData->swapchain,
&swapchainData->imageCount,
swapchainImages
);
imageViewCreateInfo.sType = VK_STRUCTURE_TYPE_IMAGE_VIEW_CREATE_INFO;
imageViewCreateInfo.pNext = NULL;
imageViewCreateInfo.flags = 0;
imageViewCreateInfo.viewType = VK_IMAGE_VIEW_TYPE_2D;
imageViewCreateInfo.format = swapchainData->format;
imageViewCreateInfo.components = swapchainData->swapchainSwizzle;
imageViewCreateInfo.subresourceRange.aspectMask = VK_IMAGE_ASPECT_COLOR_BIT;
imageViewCreateInfo.subresourceRange.baseMipLevel = 0;
imageViewCreateInfo.subresourceRange.levelCount = 1;
imageViewCreateInfo.subresourceRange.baseArrayLayer = 0;
imageViewCreateInfo.subresourceRange.layerCount = 1;
for (i = 0; i < swapchainData->imageCount; i += 1)
{
swapchainData->textureContainers[i].canBeCycled = 0;
swapchainData->textureContainers[i].textureCapacity = 0;
swapchainData->textureContainers[i].textureCount = 0;
swapchainData->textureContainers[i].textureHandles = NULL;
swapchainData->textureContainers[i].debugName = NULL;
swapchainData->textureContainers[i].activeTextureHandle = SDL_malloc(sizeof(VulkanTextureHandle));
swapchainData->textureContainers[i].activeTextureHandle->vulkanTexture = SDL_malloc(sizeof(VulkanTexture));
swapchainData->textureContainers[i].activeTextureHandle->vulkanTexture->image = swapchainImages[i];
imageViewCreateInfo.image = swapchainImages[i];
vulkanResult = renderer->vkCreateImageView(
renderer->logicalDevice,
&imageViewCreateInfo,
NULL,
&swapchainData->textureContainers[i].activeTextureHandle->vulkanTexture->view
);
if (vulkanResult != VK_SUCCESS)
{
renderer->vkDestroySurfaceKHR(
renderer->instance,
swapchainData->surface,
NULL
);
SDL_stack_free(swapchainImages);
SDL_free(swapchainData->textureContainers);
SDL_free(swapchainData);
LogVulkanResultAsError("vkCreateImageView", vulkanResult);
return 0;
}
/* Swapchain memory is managed by the driver */
swapchainData->textureContainers[i].activeTextureHandle->vulkanTexture->usedRegion = NULL;
swapchainData->textureContainers[i].activeTextureHandle->vulkanTexture->dimensions = swapchainData->extent;
swapchainData->textureContainers[i].activeTextureHandle->vulkanTexture->format = swapchainData->format;
swapchainData->textureContainers[i].activeTextureHandle->vulkanTexture->swizzle = swapchainData->swapchainSwizzle;
swapchainData->textureContainers[i].activeTextureHandle->vulkanTexture->is3D = 0;
swapchainData->textureContainers[i].activeTextureHandle->vulkanTexture->isCube = 0;
swapchainData->textureContainers[i].activeTextureHandle->vulkanTexture->isRenderTarget = 1;
swapchainData->textureContainers[i].activeTextureHandle->vulkanTexture->depth = 1;
swapchainData->textureContainers[i].activeTextureHandle->vulkanTexture->layerCount = 1;
swapchainData->textureContainers[i].activeTextureHandle->vulkanTexture->levelCount = 1;
swapchainData->textureContainers[i].activeTextureHandle->vulkanTexture->sampleCount = VK_SAMPLE_COUNT_1_BIT;
swapchainData->textureContainers[i].activeTextureHandle->vulkanTexture->usageFlags =
REFRESH_TEXTUREUSAGE_COLOR_TARGET_BIT;
swapchainData->textureContainers[i].activeTextureHandle->vulkanTexture->aspectFlags = VK_IMAGE_ASPECT_COLOR_BIT;
swapchainData->textureContainers[i].activeTextureHandle->container = NULL;
/* Create slice */
swapchainData->textureContainers[i].activeTextureHandle->vulkanTexture->sliceCount = 1;
swapchainData->textureContainers[i].activeTextureHandle->vulkanTexture->slices = SDL_malloc(sizeof(VulkanTextureSlice));
swapchainData->textureContainers[i].activeTextureHandle->vulkanTexture->slices[0].parent = swapchainData->textureContainers[i].activeTextureHandle->vulkanTexture;
swapchainData->textureContainers[i].activeTextureHandle->vulkanTexture->slices[0].layer = 0;
swapchainData->textureContainers[i].activeTextureHandle->vulkanTexture->slices[0].level = 0;
swapchainData->textureContainers[i].activeTextureHandle->vulkanTexture->slices[0].transitioned = SDL_TRUE;
swapchainData->textureContainers[i].activeTextureHandle->vulkanTexture->slices[0].msaaTexHandle = NULL;
VULKAN_INTERNAL_CreateSliceView(
renderer,
swapchainData->textureContainers[i].activeTextureHandle->vulkanTexture,
0,
0,
swapchainData->swapchainSwizzle,
&swapchainData->textureContainers[i].activeTextureHandle->vulkanTexture->slices[0].view
);
SDL_AtomicSet(&swapchainData->textureContainers[i].activeTextureHandle->vulkanTexture->slices[0].referenceCount, 0);
}
SDL_stack_free(swapchainImages);
semaphoreCreateInfo.sType = VK_STRUCTURE_TYPE_SEMAPHORE_CREATE_INFO;
semaphoreCreateInfo.pNext = NULL;
semaphoreCreateInfo.flags = 0;
for (i = 0; i < MAX_FRAMES_IN_FLIGHT; i += 1)
{
renderer->vkCreateSemaphore(
renderer->logicalDevice,
&semaphoreCreateInfo,
NULL,
&swapchainData->imageAvailableSemaphore[i]
);
renderer->vkCreateSemaphore(
renderer->logicalDevice,
&semaphoreCreateInfo,
NULL,
&swapchainData->renderFinishedSemaphore[i]
);
swapchainData->inFlightFences[i] = NULL;
}
windowData->swapchainData = swapchainData;
return 1;
}
/* Command Buffers */
static void VULKAN_INTERNAL_BeginCommandBuffer(
VulkanRenderer *renderer,
VulkanCommandBuffer *commandBuffer
) {
VkCommandBufferBeginInfo beginInfo;
VkResult result;
beginInfo.sType = VK_STRUCTURE_TYPE_COMMAND_BUFFER_BEGIN_INFO;
beginInfo.pNext = NULL;
beginInfo.flags = 0;
beginInfo.pInheritanceInfo = NULL;
beginInfo.flags = VK_COMMAND_BUFFER_USAGE_ONE_TIME_SUBMIT_BIT;
result = renderer->vkBeginCommandBuffer(
commandBuffer->commandBuffer,
&beginInfo
);
if (result != VK_SUCCESS)
{
LogVulkanResultAsError("vkBeginCommandBuffer", result);
}
}
static void VULKAN_INTERNAL_EndCommandBuffer(
VulkanRenderer* renderer,
VulkanCommandBuffer *commandBuffer
) {
VkResult result;
result = renderer->vkEndCommandBuffer(
commandBuffer->commandBuffer
);
if (result != VK_SUCCESS)
{
LogVulkanResultAsError("vkEndCommandBuffer", result);
}
}
static void VULKAN_DestroyDevice(
Refresh_Device *device
) {
VulkanRenderer* renderer = (VulkanRenderer*) device->driverData;
CommandPoolHashArray commandPoolHashArray;
VulkanMemorySubAllocator *allocator;
Sint32 i, j, k;
VULKAN_Wait(device->driverData);
for (i = renderer->claimedWindowCount - 1; i >= 0; i -= 1)
{
VULKAN_UnclaimWindow(device->driverData, renderer->claimedWindows[i]->window);
}
SDL_free(renderer->claimedWindows);
VULKAN_Wait(device->driverData);
SDL_free(renderer->submittedCommandBuffers);
for (i = 0; i < renderer->fencePool.availableFenceCount; i += 1)
{
renderer->vkDestroyFence(
renderer->logicalDevice,
renderer->fencePool.availableFences[i]->fence,
NULL
);
SDL_free(renderer->fencePool.availableFences[i]);
}
SDL_free(renderer->fencePool.availableFences);
SDL_DestroyMutex(renderer->fencePool.lock);
for (i = 0; i < NUM_COMMAND_POOL_BUCKETS; i += 1)
{
commandPoolHashArray = renderer->commandPoolHashTable.buckets[i];
for (j = 0; j < commandPoolHashArray.count; j += 1)
{
VULKAN_INTERNAL_DestroyCommandPool(
renderer,
commandPoolHashArray.elements[j].value
);
}
if (commandPoolHashArray.elements != NULL)
{
SDL_free(commandPoolHashArray.elements);
}
}
renderer->vkDestroyQueryPool(
renderer->logicalDevice,
renderer->queryPool,
NULL
);
for (i = 0; i < renderer->framebufferHashArray.count; i += 1)
{
VULKAN_INTERNAL_DestroyFramebuffer(
renderer,
renderer->framebufferHashArray.elements[i].value
);
}
SDL_free(renderer->framebufferHashArray.elements);
for (i = 0; i < renderer->renderPassHashArray.count; i += 1)
{
renderer->vkDestroyRenderPass(
renderer->logicalDevice,
renderer->renderPassHashArray.elements[i].value,
NULL
);
}
SDL_free(renderer->renderPassHashArray.elements);
for (i = 0; i < VK_MAX_MEMORY_TYPES; i += 1)
{
allocator = &renderer->memoryAllocator->subAllocators[i];
for (j = allocator->allocationCount - 1; j >= 0; j -= 1)
{
for (k = allocator->allocations[j]->usedRegionCount - 1; k >= 0; k -= 1)
{
VULKAN_INTERNAL_RemoveMemoryUsedRegion(
renderer,
allocator->allocations[j]->usedRegions[k]
);
}
VULKAN_INTERNAL_DeallocateMemory(
renderer,
allocator,
j
);
}
if (renderer->memoryAllocator->subAllocators[i].allocations != NULL)
{
SDL_free(renderer->memoryAllocator->subAllocators[i].allocations);
}
SDL_free(renderer->memoryAllocator->subAllocators[i].sortedFreeRegions);
}
SDL_free(renderer->memoryAllocator);
SDL_free(renderer->texturesToDestroy);
SDL_free(renderer->buffersToDestroy);
SDL_free(renderer->graphicsPipelinesToDestroy);
SDL_free(renderer->computePipelinesToDestroy);
SDL_free(renderer->shadersToDestroy);
SDL_free(renderer->samplersToDestroy);
SDL_free(renderer->framebuffersToDestroy);
SDL_DestroyMutex(renderer->allocatorLock);
SDL_DestroyMutex(renderer->disposeLock);
SDL_DestroyMutex(renderer->submitLock);
SDL_DestroyMutex(renderer->acquireCommandBufferLock);
SDL_DestroyMutex(renderer->renderPassFetchLock);
SDL_DestroyMutex(renderer->framebufferFetchLock);
SDL_DestroyMutex(renderer->queryLock);
renderer->vkDestroyDevice(renderer->logicalDevice, NULL);
renderer->vkDestroyInstance(renderer->instance, NULL);
SDL_free(renderer);
SDL_free(device);
SDL_Vulkan_UnloadLibrary();
}
static VkDescriptorSet VULKAN_INTERNAL_FetchDescriptorSet(
VulkanRenderer *renderer,
VulkanCommandBuffer *vulkanCommandBuffer,
DescriptorSetPool *descriptorSetPool
) {
VkDescriptorSet descriptorSet;
SDL_LockMutex(descriptorSetPool->lock);
/* If no inactive descriptor sets remain, create a new pool and allocate new inactive sets */
if (descriptorSetPool->inactiveDescriptorSetCount == 0)
{
descriptorSetPool->descriptorPoolCount += 1;
descriptorSetPool->descriptorPools = SDL_realloc(
descriptorSetPool->descriptorPools,
sizeof(VkDescriptorPool) * descriptorSetPool->descriptorPoolCount
);
if (!VULKAN_INTERNAL_CreateDescriptorPool(
renderer,
descriptorSetPool->descriptorInfos,
descriptorSetPool->descriptorInfoCount,
descriptorSetPool->nextPoolSize,
&descriptorSetPool->descriptorPools[descriptorSetPool->descriptorPoolCount - 1]
)) {
SDL_UnlockMutex(descriptorSetPool->lock);
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Failed to create descriptor pool!");
return VK_NULL_HANDLE;
}
descriptorSetPool->inactiveDescriptorSetCapacity += descriptorSetPool->nextPoolSize;
descriptorSetPool->inactiveDescriptorSets = SDL_realloc(
descriptorSetPool->inactiveDescriptorSets,
sizeof(VkDescriptorSet) * descriptorSetPool->inactiveDescriptorSetCapacity
);
if (!VULKAN_INTERNAL_AllocateDescriptorSets(
renderer,
descriptorSetPool->descriptorPools[descriptorSetPool->descriptorPoolCount - 1],
descriptorSetPool->descriptorSetLayout,
descriptorSetPool->nextPoolSize,
descriptorSetPool->inactiveDescriptorSets
)) {
SDL_UnlockMutex(descriptorSetPool->lock);
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Failed to allocate descriptor sets!");
return VK_NULL_HANDLE;
}
descriptorSetPool->inactiveDescriptorSetCount = descriptorSetPool->nextPoolSize;
descriptorSetPool->nextPoolSize *= 2;
}
descriptorSet = descriptorSetPool->inactiveDescriptorSets[descriptorSetPool->inactiveDescriptorSetCount - 1];
descriptorSetPool->inactiveDescriptorSetCount -= 1;
SDL_UnlockMutex(descriptorSetPool->lock);
if (vulkanCommandBuffer->boundDescriptorSetDataCount == vulkanCommandBuffer->boundDescriptorSetDataCapacity)
{
vulkanCommandBuffer->boundDescriptorSetDataCapacity *= 2;
vulkanCommandBuffer->boundDescriptorSetDatas = SDL_realloc(
vulkanCommandBuffer->boundDescriptorSetDatas,
vulkanCommandBuffer->boundDescriptorSetDataCapacity * sizeof(DescriptorSetData)
);
}
vulkanCommandBuffer->boundDescriptorSetDatas[vulkanCommandBuffer->boundDescriptorSetDataCount].descriptorSet = descriptorSet;
vulkanCommandBuffer->boundDescriptorSetDatas[vulkanCommandBuffer->boundDescriptorSetDataCount].descriptorSetPool = descriptorSetPool;
vulkanCommandBuffer->boundDescriptorSetDataCount += 1;
return descriptorSet;
}
static void VULKAN_INTERNAL_BindGraphicsDescriptorSets(
VulkanRenderer *renderer,
VulkanCommandBuffer *commandBuffer
) {
VulkanGraphicsPipelineResourceLayout *resourceLayout;
VkWriteDescriptorSet *writeDescriptorSets;
VkWriteDescriptorSet *currentWriteDescriptorSet;
DescriptorSetPool *descriptorSetPool;
VkDescriptorBufferInfo bufferInfos[MAX_STORAGE_BUFFERS_PER_STAGE];
VkDescriptorImageInfo imageInfos[MAX_TEXTURE_SAMPLERS_PER_STAGE + MAX_STORAGE_TEXTURES_PER_STAGE];
Uint32 dynamicOffsets[MAX_UNIFORM_BUFFERS_PER_STAGE];
Uint32 bufferInfoCount = 0;
Uint32 imageInfoCount = 0;
Uint32 i;
resourceLayout = &commandBuffer->currentGraphicsPipeline->resourceLayout;
if (commandBuffer->needNewVertexResourceDescriptorSet)
{
descriptorSetPool = &resourceLayout->descriptorSetPools[0];
commandBuffer->vertexResourceDescriptorSet = VULKAN_INTERNAL_FetchDescriptorSet(
renderer,
commandBuffer,
descriptorSetPool
);
writeDescriptorSets = SDL_stack_alloc(
VkWriteDescriptorSet,
resourceLayout->vertexSamplerCount +
resourceLayout->vertexStorageTextureCount +
resourceLayout->vertexStorageBufferCount
);
for (i = 0; i < resourceLayout->vertexSamplerCount; i += 1)
{
currentWriteDescriptorSet = &writeDescriptorSets[i];
currentWriteDescriptorSet->sType = VK_STRUCTURE_TYPE_WRITE_DESCRIPTOR_SET;
currentWriteDescriptorSet->pNext = NULL;
currentWriteDescriptorSet->descriptorCount = 1;
currentWriteDescriptorSet->descriptorType = VK_DESCRIPTOR_TYPE_COMBINED_IMAGE_SAMPLER;
currentWriteDescriptorSet->dstArrayElement = 0;
currentWriteDescriptorSet->dstBinding = i;
currentWriteDescriptorSet->dstSet = commandBuffer->vertexResourceDescriptorSet;
currentWriteDescriptorSet->pTexelBufferView = NULL;
currentWriteDescriptorSet->pBufferInfo = NULL;
imageInfos[imageInfoCount].sampler = commandBuffer->vertexSamplers[i]->sampler;
imageInfos[imageInfoCount].imageView = commandBuffer->vertexSamplerTextures[i]->view;
imageInfos[imageInfoCount].imageLayout = VK_IMAGE_LAYOUT_SHADER_READ_ONLY_OPTIMAL;
currentWriteDescriptorSet->pImageInfo = &imageInfos[imageInfoCount];
imageInfoCount += 1;
}
for (i = 0; i < resourceLayout->vertexStorageTextureCount; i += 1)
{
currentWriteDescriptorSet = &writeDescriptorSets[resourceLayout->vertexSamplerCount + i];
currentWriteDescriptorSet->sType = VK_STRUCTURE_TYPE_WRITE_DESCRIPTOR_SET;
currentWriteDescriptorSet->pNext = NULL;
currentWriteDescriptorSet->descriptorCount = 1;
currentWriteDescriptorSet->descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_IMAGE;
currentWriteDescriptorSet->dstArrayElement = 0;
currentWriteDescriptorSet->dstBinding = resourceLayout->vertexSamplerCount + i;
currentWriteDescriptorSet->dstSet = commandBuffer->vertexResourceDescriptorSet;
currentWriteDescriptorSet->pTexelBufferView = NULL;
currentWriteDescriptorSet->pBufferInfo = NULL;
imageInfos[imageInfoCount].sampler = VK_NULL_HANDLE;
imageInfos[imageInfoCount].imageView = commandBuffer->vertexStorageTextureSlices[i]->view;
imageInfos[imageInfoCount].imageLayout = VK_IMAGE_LAYOUT_SHADER_READ_ONLY_OPTIMAL;
currentWriteDescriptorSet->pImageInfo = &imageInfos[imageInfoCount];
imageInfoCount += 1;
}
for (i = 0; i < resourceLayout->vertexStorageBufferCount; i += 1)
{
currentWriteDescriptorSet = &writeDescriptorSets[resourceLayout->vertexSamplerCount + resourceLayout->vertexStorageTextureCount + i];
currentWriteDescriptorSet->sType = VK_STRUCTURE_TYPE_WRITE_DESCRIPTOR_SET;
currentWriteDescriptorSet->pNext = NULL;
currentWriteDescriptorSet->descriptorCount = 1;
currentWriteDescriptorSet->descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_BUFFER;
currentWriteDescriptorSet->dstArrayElement = 0;
currentWriteDescriptorSet->dstBinding = resourceLayout->vertexSamplerCount + resourceLayout->vertexStorageTextureCount + i;
currentWriteDescriptorSet->dstSet = commandBuffer->vertexResourceDescriptorSet;
currentWriteDescriptorSet->pTexelBufferView = NULL;
currentWriteDescriptorSet->pImageInfo = NULL;
bufferInfos[bufferInfoCount].buffer = commandBuffer->vertexStorageBuffers[i]->buffer;
bufferInfos[bufferInfoCount].offset = 0;
bufferInfos[bufferInfoCount].range = VK_WHOLE_SIZE;
currentWriteDescriptorSet->pBufferInfo = &bufferInfos[bufferInfoCount];
bufferInfoCount += 1;
}
renderer->vkUpdateDescriptorSets(
renderer->logicalDevice,
resourceLayout->vertexSamplerCount + resourceLayout->vertexStorageTextureCount + resourceLayout->vertexStorageBufferCount,
writeDescriptorSets,
0,
NULL
);
renderer->vkCmdBindDescriptorSets(
commandBuffer->commandBuffer,
VK_PIPELINE_BIND_POINT_GRAPHICS,
resourceLayout->pipelineLayout,
0,
1,
&commandBuffer->vertexResourceDescriptorSet,
0,
NULL
);
SDL_stack_free(writeDescriptorSets);
bufferInfoCount = 0;
imageInfoCount = 0;
commandBuffer->needNewVertexResourceDescriptorSet = SDL_FALSE;
}
if (commandBuffer->needNewVertexUniformDescriptorSet)
{
descriptorSetPool = &resourceLayout->descriptorSetPools[1];
commandBuffer->vertexUniformDescriptorSet = VULKAN_INTERNAL_FetchDescriptorSet(
renderer,
commandBuffer,
descriptorSetPool
);
writeDescriptorSets = SDL_stack_alloc(
VkWriteDescriptorSet,
resourceLayout->vertexUniformBufferCount
);
for (i = 0; i < resourceLayout->vertexUniformBufferCount; i += 1)
{
currentWriteDescriptorSet = &writeDescriptorSets[i];
currentWriteDescriptorSet->sType = VK_STRUCTURE_TYPE_WRITE_DESCRIPTOR_SET;
currentWriteDescriptorSet->pNext = NULL;
currentWriteDescriptorSet->descriptorCount = 1;
currentWriteDescriptorSet->descriptorType = VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER_DYNAMIC;
currentWriteDescriptorSet->dstArrayElement = 0;
currentWriteDescriptorSet->dstBinding = i;
currentWriteDescriptorSet->dstSet = commandBuffer->vertexUniformDescriptorSet;
currentWriteDescriptorSet->pTexelBufferView = NULL;
currentWriteDescriptorSet->pImageInfo = NULL;
bufferInfos[bufferInfoCount].buffer = commandBuffer->vertexUniformBuffers[i]->bufferContainer->activeBufferHandle->vulkanBuffer->buffer;
bufferInfos[bufferInfoCount].offset = 0;
bufferInfos[bufferInfoCount].range = MAX_UBO_SECTION_SIZE;
currentWriteDescriptorSet->pBufferInfo = &bufferInfos[bufferInfoCount];
bufferInfoCount += 1;
}
renderer->vkUpdateDescriptorSets(
renderer->logicalDevice,
resourceLayout->vertexUniformBufferCount,
writeDescriptorSets,
0,
NULL
);
SDL_stack_free(writeDescriptorSets);
bufferInfoCount = 0;
imageInfoCount = 0;
commandBuffer->needNewVertexUniformDescriptorSet = SDL_FALSE;
commandBuffer->needNewVertexUniformOffsets = SDL_TRUE;
}
if (commandBuffer->needNewVertexUniformOffsets)
{
for (i = 0; i < resourceLayout->vertexUniformBufferCount; i += 1)
{
dynamicOffsets[i] = commandBuffer->vertexUniformBuffers[i]->drawOffset;
}
renderer->vkCmdBindDescriptorSets(
commandBuffer->commandBuffer,
VK_PIPELINE_BIND_POINT_GRAPHICS,
resourceLayout->pipelineLayout,
1,
1,
&commandBuffer->vertexUniformDescriptorSet,
resourceLayout->vertexUniformBufferCount,
dynamicOffsets
);
commandBuffer->needNewVertexUniformOffsets = SDL_FALSE;
}
if (commandBuffer->needNewFragmentResourceDescriptorSet)
{
descriptorSetPool = &resourceLayout->descriptorSetPools[2];
commandBuffer->fragmentResourceDescriptorSet = VULKAN_INTERNAL_FetchDescriptorSet(
renderer,
commandBuffer,
descriptorSetPool
);
writeDescriptorSets = SDL_stack_alloc(
VkWriteDescriptorSet,
resourceLayout->fragmentSamplerCount +
resourceLayout->fragmentStorageTextureCount +
resourceLayout->fragmentStorageBufferCount
);
for (i = 0; i < resourceLayout->fragmentSamplerCount; i += 1)
{
currentWriteDescriptorSet = &writeDescriptorSets[i];
currentWriteDescriptorSet->sType = VK_STRUCTURE_TYPE_WRITE_DESCRIPTOR_SET;
currentWriteDescriptorSet->pNext = NULL;
currentWriteDescriptorSet->descriptorCount = 1;
currentWriteDescriptorSet->descriptorType = VK_DESCRIPTOR_TYPE_COMBINED_IMAGE_SAMPLER;
currentWriteDescriptorSet->dstArrayElement = 0;
currentWriteDescriptorSet->dstBinding = i;
currentWriteDescriptorSet->dstSet = commandBuffer->fragmentResourceDescriptorSet;
currentWriteDescriptorSet->pTexelBufferView = NULL;
currentWriteDescriptorSet->pBufferInfo = NULL;
imageInfos[imageInfoCount].sampler = commandBuffer->fragmentSamplers[i]->sampler;
imageInfos[imageInfoCount].imageView = commandBuffer->fragmentSamplerTextures[i]->view;
imageInfos[imageInfoCount].imageLayout = VK_IMAGE_LAYOUT_SHADER_READ_ONLY_OPTIMAL;
currentWriteDescriptorSet->pImageInfo = &imageInfos[imageInfoCount];
imageInfoCount += 1;
}
for (i = 0; i < resourceLayout->fragmentStorageTextureCount; i += 1)
{
currentWriteDescriptorSet = &writeDescriptorSets[resourceLayout->fragmentSamplerCount + i];
currentWriteDescriptorSet->sType = VK_STRUCTURE_TYPE_WRITE_DESCRIPTOR_SET;
currentWriteDescriptorSet->pNext = NULL;
currentWriteDescriptorSet->descriptorCount = 1;
currentWriteDescriptorSet->descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_IMAGE;
currentWriteDescriptorSet->dstArrayElement = 0;
currentWriteDescriptorSet->dstBinding = resourceLayout->fragmentSamplerCount + i;
currentWriteDescriptorSet->dstSet = commandBuffer->fragmentResourceDescriptorSet;
currentWriteDescriptorSet->pTexelBufferView = NULL;
currentWriteDescriptorSet->pBufferInfo = NULL;
imageInfos[imageInfoCount].sampler = VK_NULL_HANDLE;
imageInfos[imageInfoCount].imageView = commandBuffer->fragmentStorageTextureSlices[i]->view;
imageInfos[imageInfoCount].imageLayout = VK_IMAGE_LAYOUT_SHADER_READ_ONLY_OPTIMAL;
currentWriteDescriptorSet->pImageInfo = &imageInfos[imageInfoCount];
imageInfoCount += 1;
}
for (i = 0; i < resourceLayout->fragmentStorageBufferCount; i += 1)
{
currentWriteDescriptorSet = &writeDescriptorSets[resourceLayout->fragmentSamplerCount + resourceLayout->fragmentStorageTextureCount + i];
currentWriteDescriptorSet->sType = VK_STRUCTURE_TYPE_WRITE_DESCRIPTOR_SET;
currentWriteDescriptorSet->pNext = NULL;
currentWriteDescriptorSet->descriptorCount = 1;
currentWriteDescriptorSet->descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_IMAGE;
currentWriteDescriptorSet->dstArrayElement = 0;
currentWriteDescriptorSet->dstBinding = resourceLayout->fragmentSamplerCount + resourceLayout->fragmentStorageTextureCount + i;
currentWriteDescriptorSet->dstSet = commandBuffer->fragmentResourceDescriptorSet;
currentWriteDescriptorSet->pTexelBufferView = NULL;
currentWriteDescriptorSet->pImageInfo = NULL;
bufferInfos[bufferInfoCount].buffer = commandBuffer->fragmentStorageBuffers[i]->buffer;
bufferInfos[bufferInfoCount].offset = 0;
bufferInfos[bufferInfoCount].range = VK_WHOLE_SIZE;
currentWriteDescriptorSet->pBufferInfo = &bufferInfos[bufferInfoCount];
bufferInfoCount += 1;
}
renderer->vkUpdateDescriptorSets(
renderer->logicalDevice,
resourceLayout->fragmentSamplerCount + resourceLayout->fragmentStorageTextureCount + resourceLayout->fragmentStorageBufferCount,
writeDescriptorSets,
0,
NULL
);
renderer->vkCmdBindDescriptorSets(
commandBuffer->commandBuffer,
VK_PIPELINE_BIND_POINT_GRAPHICS,
resourceLayout->pipelineLayout,
2,
1,
&commandBuffer->fragmentResourceDescriptorSet,
0,
NULL
);
SDL_stack_free(writeDescriptorSets);
bufferInfoCount = 0;
imageInfoCount = 0;
commandBuffer->needNewFragmentResourceDescriptorSet = SDL_TRUE;
}
if (commandBuffer->needNewFragmentUniformDescriptorSet)
{
descriptorSetPool = &resourceLayout->descriptorSetPools[3];
commandBuffer->fragmentUniformDescriptorSet = VULKAN_INTERNAL_FetchDescriptorSet(
renderer,
commandBuffer,
descriptorSetPool
);
writeDescriptorSets = SDL_stack_alloc(
VkWriteDescriptorSet,
resourceLayout->fragmentUniformBufferCount
);
for (i = 0; i < resourceLayout->fragmentUniformBufferCount; i += 1)
{
currentWriteDescriptorSet = &writeDescriptorSets[i];
currentWriteDescriptorSet->sType = VK_STRUCTURE_TYPE_WRITE_DESCRIPTOR_SET;
currentWriteDescriptorSet->pNext = NULL;
currentWriteDescriptorSet->descriptorCount = 1;
currentWriteDescriptorSet->descriptorType = VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER_DYNAMIC;
currentWriteDescriptorSet->dstArrayElement = 0;
currentWriteDescriptorSet->dstBinding = i;
currentWriteDescriptorSet->dstSet = commandBuffer->fragmentUniformDescriptorSet;
currentWriteDescriptorSet->pTexelBufferView = NULL;
currentWriteDescriptorSet->pImageInfo = NULL;
bufferInfos[bufferInfoCount].buffer = commandBuffer->fragmentUniformBuffers[i]->bufferContainer->activeBufferHandle->vulkanBuffer->buffer;
bufferInfos[bufferInfoCount].offset = 0;
bufferInfos[bufferInfoCount].range = MAX_UBO_SECTION_SIZE;
currentWriteDescriptorSet->pBufferInfo = &bufferInfos[bufferInfoCount];
bufferInfoCount += 1;
}
renderer->vkUpdateDescriptorSets(
renderer->logicalDevice,
resourceLayout->fragmentUniformBufferCount,
writeDescriptorSets,
0,
NULL
);
SDL_stack_free(writeDescriptorSets);
bufferInfoCount = 0;
imageInfoCount = 0;
commandBuffer->needNewFragmentUniformDescriptorSet = SDL_FALSE;
commandBuffer->needNewFragmentUniformOffsets = SDL_TRUE;
}
if (commandBuffer->needNewFragmentUniformOffsets)
{
for (i = 0; i < resourceLayout->fragmentUniformBufferCount; i += 1)
{
dynamicOffsets[i] = commandBuffer->fragmentUniformBuffers[i]->drawOffset;
}
renderer->vkCmdBindDescriptorSets(
commandBuffer->commandBuffer,
VK_PIPELINE_BIND_POINT_GRAPHICS,
resourceLayout->pipelineLayout,
3,
1,
&commandBuffer->fragmentUniformDescriptorSet,
resourceLayout->fragmentUniformBufferCount,
dynamicOffsets
);
commandBuffer->needNewFragmentUniformOffsets = SDL_FALSE;
}
}
static void VULKAN_DrawIndexedPrimitives(
Refresh_CommandBuffer *commandBuffer,
Uint32 baseVertex,
Uint32 startIndex,
Uint32 primitiveCount,
Uint32 instanceCount
) {
VulkanCommandBuffer *vulkanCommandBuffer = (VulkanCommandBuffer*) commandBuffer;
VulkanRenderer* renderer = (VulkanRenderer*) vulkanCommandBuffer->renderer;
VULKAN_INTERNAL_BindGraphicsDescriptorSets(renderer, vulkanCommandBuffer);
renderer->vkCmdDrawIndexed(
vulkanCommandBuffer->commandBuffer,
PrimitiveVerts(
vulkanCommandBuffer->currentGraphicsPipeline->primitiveType,
primitiveCount
),
instanceCount,
startIndex,
baseVertex,
0
);
}
static void VULKAN_DrawPrimitives(
Refresh_CommandBuffer *commandBuffer,
Uint32 vertexStart,
Uint32 primitiveCount
) {
VulkanCommandBuffer *vulkanCommandBuffer = (VulkanCommandBuffer*) commandBuffer;
VulkanRenderer* renderer = (VulkanRenderer*) vulkanCommandBuffer->renderer;
VULKAN_INTERNAL_BindGraphicsDescriptorSets(renderer, vulkanCommandBuffer);
renderer->vkCmdDraw(
vulkanCommandBuffer->commandBuffer,
PrimitiveVerts(
vulkanCommandBuffer->currentGraphicsPipeline->primitiveType,
primitiveCount
),
1,
vertexStart,
0
);
}
static void VULKAN_DrawPrimitivesIndirect(
Refresh_CommandBuffer *commandBuffer,
Refresh_Buffer *buffer,
Uint32 offsetInBytes,
Uint32 drawCount,
Uint32 stride
) {
VulkanCommandBuffer *vulkanCommandBuffer = (VulkanCommandBuffer*) commandBuffer;
VulkanRenderer* renderer = (VulkanRenderer*) vulkanCommandBuffer->renderer;
VulkanBuffer *vulkanBuffer = ((VulkanBufferContainer*) buffer)->activeBufferHandle->vulkanBuffer;
VULKAN_INTERNAL_BindGraphicsDescriptorSets(renderer, vulkanCommandBuffer);
renderer->vkCmdDrawIndirect(
vulkanCommandBuffer->commandBuffer,
vulkanBuffer->buffer,
offsetInBytes,
drawCount,
stride
);
VULKAN_INTERNAL_TrackBuffer(renderer, vulkanCommandBuffer, vulkanBuffer);
}
static void VULKAN_DrawIndexedPrimitivesIndirect(
Refresh_CommandBuffer *commandBuffer,
Refresh_Buffer *buffer,
Uint32 offsetInBytes,
Uint32 drawCount,
Uint32 stride
) {
VulkanCommandBuffer *vulkanCommandBuffer = (VulkanCommandBuffer*) commandBuffer;
VulkanRenderer* renderer = (VulkanRenderer*) vulkanCommandBuffer->renderer;
VulkanBuffer *vulkanBuffer = ((VulkanBufferContainer*) buffer)->activeBufferHandle->vulkanBuffer;
VULKAN_INTERNAL_BindGraphicsDescriptorSets(renderer, vulkanCommandBuffer);
renderer->vkCmdDrawIndexedIndirect(
vulkanCommandBuffer->commandBuffer,
vulkanBuffer->buffer,
offsetInBytes,
drawCount,
stride
);
VULKAN_INTERNAL_TrackBuffer(renderer, vulkanCommandBuffer, vulkanBuffer);
}
/* Debug Naming */
static void VULKAN_INTERNAL_SetBufferName(
VulkanRenderer *renderer,
VulkanBuffer *buffer,
const char *text
) {
VkDebugUtilsObjectNameInfoEXT nameInfo;
if (renderer->debugMode && renderer->supportsDebugUtils)
{
nameInfo.sType = VK_STRUCTURE_TYPE_DEBUG_UTILS_OBJECT_NAME_INFO_EXT;
nameInfo.pNext = NULL;
nameInfo.pObjectName = text;
nameInfo.objectType = VK_OBJECT_TYPE_BUFFER;
nameInfo.objectHandle = (uint64_t) buffer->buffer;
renderer->vkSetDebugUtilsObjectNameEXT(
renderer->logicalDevice,
&nameInfo
);
}
}
static void VULKAN_SetBufferName(
Refresh_Renderer *driverData,
Refresh_Buffer *buffer,
const char *text
) {
VulkanRenderer *renderer = (VulkanRenderer*) driverData;
VulkanBufferContainer *container = (VulkanBufferContainer*) buffer;
if (renderer->debugMode && renderer->supportsDebugUtils)
{
container->debugName = SDL_realloc(
container->debugName,
SDL_strlen(text) + 1
);
SDL_utf8strlcpy(
container->debugName,
text,
SDL_strlen(text) + 1
);
for (Uint32 i = 0; i < container->bufferCount; i += 1)
{
VULKAN_INTERNAL_SetBufferName(
renderer,
container->bufferHandles[i]->vulkanBuffer,
text
);
}
}
}
static void VULKAN_INTERNAL_SetTextureName(
VulkanRenderer *renderer,
VulkanTexture *texture,
const char *text
) {
VkDebugUtilsObjectNameInfoEXT nameInfo;
if (renderer->debugMode && renderer->supportsDebugUtils)
{
nameInfo.sType = VK_STRUCTURE_TYPE_DEBUG_UTILS_OBJECT_NAME_INFO_EXT;
nameInfo.pNext = NULL;
nameInfo.pObjectName = text;
nameInfo.objectType = VK_OBJECT_TYPE_IMAGE;
nameInfo.objectHandle = (uint64_t) texture->image;
renderer->vkSetDebugUtilsObjectNameEXT(
renderer->logicalDevice,
&nameInfo
);
}
}
static void VULKAN_SetTextureName(
Refresh_Renderer *driverData,
Refresh_Texture *texture,
const char *text
) {
VulkanRenderer *renderer = (VulkanRenderer*) driverData;
VulkanTextureContainer *container = (VulkanTextureContainer*) texture;
if (renderer->debugMode && renderer->supportsDebugUtils)
{
container->debugName = SDL_realloc(
container->debugName,
SDL_strlen(text) + 1
);
SDL_utf8strlcpy(
container->debugName,
text,
SDL_strlen(text) + 1
);
for (Uint32 i = 0; i < container->textureCount; i += 1)
{
VULKAN_INTERNAL_SetTextureName(
renderer,
container->textureHandles[i]->vulkanTexture,
text
);
}
}
}
static void VULKAN_SetStringMarker(
Refresh_CommandBuffer *commandBuffer,
const char *text
) {
VulkanCommandBuffer *vulkanCommandBuffer = (VulkanCommandBuffer*) commandBuffer;
VulkanRenderer* renderer = (VulkanRenderer*) vulkanCommandBuffer->renderer;
VkDebugUtilsLabelEXT labelInfo;
if (renderer->supportsDebugUtils)
{
labelInfo.sType = VK_STRUCTURE_TYPE_DEBUG_UTILS_LABEL_EXT;
labelInfo.pNext = NULL;
labelInfo.pLabelName = text;
renderer->vkCmdInsertDebugUtilsLabelEXT(
vulkanCommandBuffer->commandBuffer,
&labelInfo
);
}
}
static VulkanTextureHandle* VULKAN_INTERNAL_CreateTextureHandle(
VulkanRenderer *renderer,
Uint32 width,
Uint32 height,
Uint32 depth,
Uint32 isCube,
Uint32 layerCount,
Uint32 levelCount,
VkSampleCountFlagBits sampleCount,
VkFormat format,
VkComponentMapping swizzle,
VkImageAspectFlags aspectMask,
Refresh_TextureUsageFlags textureUsageFlags
) {
VulkanTextureHandle *textureHandle;
VulkanTexture *texture;
texture = VULKAN_INTERNAL_CreateTexture(
renderer,
width,
height,
depth,
isCube,
layerCount,
levelCount,
sampleCount,
format,
swizzle,
aspectMask,
textureUsageFlags
);
if (texture == NULL)
{
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Failed to create texture!");
return NULL;
}
textureHandle = SDL_malloc(sizeof(VulkanTextureHandle));
textureHandle->vulkanTexture = texture;
textureHandle->container = NULL;
texture->handle = textureHandle;
return textureHandle;
}
static VulkanTexture* VULKAN_INTERNAL_CreateTexture(
VulkanRenderer *renderer,
Uint32 width,
Uint32 height,
Uint32 depth,
Uint32 isCube,
Uint32 layerCount,
Uint32 levelCount,
VkSampleCountFlagBits sampleCount,
VkFormat format,
VkComponentMapping swizzle,
VkImageAspectFlags aspectMask,
Refresh_TextureUsageFlags textureUsageFlags
) {
VkResult vulkanResult;
VkImageCreateInfo imageCreateInfo;
VkImageCreateFlags imageCreateFlags = 0;
VkImageViewCreateInfo imageViewCreateInfo;
Uint8 bindResult;
Uint8 is3D = depth > 1 ? 1 : 0;
Uint8 isRenderTarget =
((textureUsageFlags & REFRESH_TEXTUREUSAGE_COLOR_TARGET_BIT) != 0) ||
((textureUsageFlags & REFRESH_TEXTUREUSAGE_DEPTH_STENCIL_TARGET_BIT) != 0);
Uint32 i, j, sliceIndex;
VkImageUsageFlags vkUsageFlags = VK_IMAGE_USAGE_TRANSFER_SRC_BIT | VK_IMAGE_USAGE_TRANSFER_DST_BIT;
VulkanTexture *texture = SDL_malloc(sizeof(VulkanTexture));
texture->isCube = 0;
texture->is3D = 0;
texture->isRenderTarget = isRenderTarget;
texture->markedForDestroy = 0;
if (isCube)
{
imageCreateFlags |= VK_IMAGE_CREATE_CUBE_COMPATIBLE_BIT;
texture->isCube = 1;
}
else if (is3D)
{
imageCreateFlags |= VK_IMAGE_CREATE_2D_ARRAY_COMPATIBLE_BIT;
texture->is3D = 1;
}
if (textureUsageFlags & REFRESH_TEXTUREUSAGE_SAMPLER_BIT)
{
vkUsageFlags |= VK_IMAGE_USAGE_SAMPLED_BIT;
}
if (textureUsageFlags & REFRESH_TEXTUREUSAGE_COLOR_TARGET_BIT)
{
vkUsageFlags |= VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT;
}
if (textureUsageFlags & REFRESH_TEXTUREUSAGE_DEPTH_STENCIL_TARGET_BIT)
{
vkUsageFlags |= VK_IMAGE_USAGE_DEPTH_STENCIL_ATTACHMENT_BIT;
}
if (textureUsageFlags & (
REFRESH_TEXTUREUSAGE_GRAPHICS_STORAGE_READ_BIT |
REFRESH_TEXTUREUSAGE_COMPUTE_STORAGE_READ_BIT |
REFRESH_TEXTUREUSAGE_COMPUTE_STORAGE_WRITE_BIT
)) {
vkUsageFlags |= VK_IMAGE_USAGE_STORAGE_BIT;
}
imageCreateInfo.sType = VK_STRUCTURE_TYPE_IMAGE_CREATE_INFO;
imageCreateInfo.pNext = NULL;
imageCreateInfo.flags = imageCreateFlags;
imageCreateInfo.imageType = is3D ? VK_IMAGE_TYPE_3D : VK_IMAGE_TYPE_2D;
imageCreateInfo.format = format;
imageCreateInfo.extent.width = width;
imageCreateInfo.extent.height = height;
imageCreateInfo.extent.depth = depth;
imageCreateInfo.mipLevels = levelCount;
imageCreateInfo.arrayLayers = layerCount;
imageCreateInfo.samples = VULKAN_INTERNAL_IsVulkanDepthFormat(format) ? sampleCount : VK_SAMPLE_COUNT_1_BIT;
imageCreateInfo.tiling = VK_IMAGE_TILING_OPTIMAL;
imageCreateInfo.usage = vkUsageFlags;
imageCreateInfo.sharingMode = VK_SHARING_MODE_EXCLUSIVE;
imageCreateInfo.queueFamilyIndexCount = 0;
imageCreateInfo.pQueueFamilyIndices = NULL;
imageCreateInfo.initialLayout = VK_IMAGE_LAYOUT_UNDEFINED;
vulkanResult = renderer->vkCreateImage(
renderer->logicalDevice,
&imageCreateInfo,
NULL,
&texture->image
);
VULKAN_ERROR_CHECK(vulkanResult, vkCreateImage, 0)
bindResult = VULKAN_INTERNAL_BindMemoryForImage(
renderer,
texture->image,
&texture->usedRegion
);
if (bindResult != 1)
{
renderer->vkDestroyImage(
renderer->logicalDevice,
texture->image,
NULL);
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Unable to bind memory for texture!");
return NULL;
}
texture->usedRegion->vulkanTexture = texture; /* lol */
imageViewCreateInfo.sType = VK_STRUCTURE_TYPE_IMAGE_VIEW_CREATE_INFO;
imageViewCreateInfo.pNext = NULL;
imageViewCreateInfo.flags = 0;
imageViewCreateInfo.image = texture->image;
imageViewCreateInfo.format = format;
imageViewCreateInfo.components = swizzle;
imageViewCreateInfo.subresourceRange.aspectMask = aspectMask;
imageViewCreateInfo.subresourceRange.baseMipLevel = 0;
imageViewCreateInfo.subresourceRange.levelCount = levelCount;
imageViewCreateInfo.subresourceRange.baseArrayLayer = 0;
imageViewCreateInfo.subresourceRange.layerCount = layerCount;
if (isCube)
{
imageViewCreateInfo.viewType = VK_IMAGE_VIEW_TYPE_CUBE;
}
else if (is3D)
{
imageViewCreateInfo.viewType = VK_IMAGE_VIEW_TYPE_3D;
}
else if (layerCount > 1)
{
imageViewCreateInfo.viewType = VK_IMAGE_VIEW_TYPE_2D_ARRAY;
}
else
{
imageViewCreateInfo.viewType = VK_IMAGE_VIEW_TYPE_2D;
}
vulkanResult = renderer->vkCreateImageView(
renderer->logicalDevice,
&imageViewCreateInfo,
NULL,
&texture->view
);
if (vulkanResult != VK_SUCCESS)
{
LogVulkanResultAsError("vkCreateImageView", vulkanResult);
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Failed to create texture image view");
return NULL;
}
texture->dimensions.width = width;
texture->dimensions.height = height;
texture->depth = depth;
texture->format = format;
texture->swizzle = swizzle;
texture->levelCount = levelCount;
texture->layerCount = layerCount;
texture->sampleCount = sampleCount;
texture->usageFlags = textureUsageFlags;
texture->aspectFlags = aspectMask;
/* Define slices */
texture->sliceCount =
texture->layerCount *
texture->levelCount;
texture->slices = SDL_malloc(
texture->sliceCount * sizeof(VulkanTextureSlice)
);
for (i = 0; i < texture->layerCount; i += 1)
{
for (j = 0; j < texture->levelCount; j += 1)
{
sliceIndex = VULKAN_INTERNAL_GetTextureSliceIndex(
texture,
i,
j
);
VULKAN_INTERNAL_CreateSliceView(
renderer,
texture,
i,
j,
swizzle,
&texture->slices[sliceIndex].view
);
texture->slices[sliceIndex].parent = texture;
texture->slices[sliceIndex].layer = i;
texture->slices[sliceIndex].level = j;
texture->slices[sliceIndex].msaaTexHandle = NULL;
texture->slices[sliceIndex].transitioned = SDL_FALSE;
SDL_AtomicSet(&texture->slices[sliceIndex].referenceCount, 0);
if (
sampleCount > VK_SAMPLE_COUNT_1_BIT &&
isRenderTarget &&
!VULKAN_INTERNAL_IsVulkanDepthFormat(texture->format)
) {
texture->slices[sliceIndex].msaaTexHandle = VULKAN_INTERNAL_CreateTextureHandle(
renderer,
texture->dimensions.width >> j,
texture->dimensions.height >> j,
1,
0,
1,
1,
VK_SAMPLE_COUNT_1_BIT,
texture->format,
texture->swizzle,
aspectMask,
textureUsageFlags
);
}
}
}
return texture;
}
static void VULKAN_INTERNAL_CycleActiveBuffer(
VulkanRenderer *renderer,
VulkanBufferContainer *bufferContainer
) {
VulkanBufferHandle *bufferHandle;
Uint32 i;
/* If a previously-cycled buffer is available, we can use that. */
for (i = 0; i < bufferContainer->bufferCount; i += 1)
{
bufferHandle = bufferContainer->bufferHandles[i];
if (SDL_AtomicGet(&bufferHandle->vulkanBuffer->referenceCount) == 0)
{
bufferContainer->activeBufferHandle = bufferHandle;
return;
}
}
/* No buffer handle is available, generate a new one. */
bufferContainer->activeBufferHandle = VULKAN_INTERNAL_CreateBufferHandle(
renderer,
bufferContainer->activeBufferHandle->vulkanBuffer->size,
bufferContainer->activeBufferHandle->vulkanBuffer->usageFlags,
bufferContainer->activeBufferHandle->vulkanBuffer->type
);
bufferContainer->activeBufferHandle->container = bufferContainer;
EXPAND_ARRAY_IF_NEEDED(
bufferContainer->bufferHandles,
VulkanBufferHandle*,
bufferContainer->bufferCount + 1,
bufferContainer->bufferCapacity,
bufferContainer->bufferCapacity * 2
);
bufferContainer->bufferHandles[
bufferContainer->bufferCount
] = bufferContainer->activeBufferHandle;
bufferContainer->bufferCount += 1;
if (
renderer->debugMode &&
renderer->supportsDebugUtils &&
bufferContainer->debugName != NULL
) {
VULKAN_INTERNAL_SetBufferName(
renderer,
bufferContainer->activeBufferHandle->vulkanBuffer,
bufferContainer->debugName
);
}
}
static void VULKAN_INTERNAL_CycleActiveTexture(
VulkanRenderer *renderer,
VulkanTextureContainer *textureContainer
) {
VulkanTextureHandle *textureHandle;
Uint32 i, j;
Sint32 refCountTotal;
/* If a previously-cycled texture is available, we can use that. */
for (i = 0; i < textureContainer->textureCount; i += 1)
{
textureHandle = textureContainer->textureHandles[i];
refCountTotal = 0;
for (j = 0; j < textureHandle->vulkanTexture->sliceCount; j += 1)
{
refCountTotal += SDL_AtomicGet(&textureHandle->vulkanTexture->slices[j].referenceCount);
}
if (refCountTotal == 0)
{
textureContainer->activeTextureHandle = textureHandle;
return;
}
}
/* No texture handle is available, generate a new one. */
textureContainer->activeTextureHandle = VULKAN_INTERNAL_CreateTextureHandle(
renderer,
textureContainer->activeTextureHandle->vulkanTexture->dimensions.width,
textureContainer->activeTextureHandle->vulkanTexture->dimensions.height,
textureContainer->activeTextureHandle->vulkanTexture->depth,
textureContainer->activeTextureHandle->vulkanTexture->isCube,
textureContainer->activeTextureHandle->vulkanTexture->layerCount,
textureContainer->activeTextureHandle->vulkanTexture->levelCount,
textureContainer->activeTextureHandle->vulkanTexture->sampleCount,
textureContainer->activeTextureHandle->vulkanTexture->format,
textureContainer->activeTextureHandle->vulkanTexture->swizzle,
textureContainer->activeTextureHandle->vulkanTexture->aspectFlags,
textureContainer->activeTextureHandle->vulkanTexture->usageFlags
);
textureContainer->activeTextureHandle->container = textureContainer;
EXPAND_ARRAY_IF_NEEDED(
textureContainer->textureHandles,
VulkanTextureHandle*,
textureContainer->textureCount + 1,
textureContainer->textureCapacity,
textureContainer->textureCapacity * 2
);
textureContainer->textureHandles[
textureContainer->textureCount
] = textureContainer->activeTextureHandle;
textureContainer->textureCount += 1;
if (
renderer->debugMode &&
renderer->supportsDebugUtils &&
textureContainer->debugName != NULL
) {
VULKAN_INTERNAL_SetTextureName(
renderer,
textureContainer->activeTextureHandle->vulkanTexture,
textureContainer->debugName
);
}
}
static VulkanBuffer* VULKAN_INTERNAL_PrepareBufferForWrite(
VulkanRenderer *renderer,
VulkanCommandBuffer *commandBuffer,
VulkanBufferContainer *bufferContainer,
SDL_bool cycle,
VulkanBufferUsageMode destinationUsageMode
) {
if (
cycle &&
SDL_AtomicGet(&bufferContainer->activeBufferHandle->vulkanBuffer->referenceCount) > 0
) {
VULKAN_INTERNAL_CycleActiveBuffer(
renderer,
bufferContainer
);
}
VULKAN_INTERNAL_BufferTransitionFromDefaultUsage(
renderer,
commandBuffer,
destinationUsageMode,
bufferContainer->activeBufferHandle->vulkanBuffer
);
return bufferContainer->activeBufferHandle->vulkanBuffer;
}
static VulkanTextureSlice* VULKAN_INTERNAL_PrepareTextureSliceForWrite(
VulkanRenderer *renderer,
VulkanCommandBuffer *commandBuffer,
VulkanTextureContainer *textureContainer,
Uint32 layer,
Uint32 level,
SDL_bool cycle,
VulkanTextureUsageMode destinationUsageMode
) {
VulkanTextureSlice *textureSlice = VULKAN_INTERNAL_FetchTextureSlice(
textureContainer->activeTextureHandle->vulkanTexture,
layer,
level
);
if (
cycle &&
textureContainer->canBeCycled &&
SDL_AtomicGet(&textureSlice->referenceCount) > 0
) {
VULKAN_INTERNAL_CycleActiveTexture(
renderer,
textureContainer
);
textureSlice = VULKAN_INTERNAL_FetchTextureSlice(
textureContainer->activeTextureHandle->vulkanTexture,
layer,
level
);
}
/* always do barrier because of layout transitions */
VULKAN_INTERNAL_TextureTransitionFromDefaultUsage(
renderer,
commandBuffer,
destinationUsageMode,
textureSlice
);
return textureSlice;
}
static VkRenderPass VULKAN_INTERNAL_CreateRenderPass(
VulkanRenderer *renderer,
VulkanCommandBuffer *commandBuffer,
Refresh_ColorAttachmentInfo *colorAttachmentInfos,
Uint32 colorAttachmentCount,
Refresh_DepthStencilAttachmentInfo *depthStencilAttachmentInfo
) {
VkResult vulkanResult;
VkAttachmentDescription attachmentDescriptions[2 * MAX_COLOR_TARGET_BINDINGS + 1];
VkAttachmentReference colorAttachmentReferences[MAX_COLOR_TARGET_BINDINGS];
VkAttachmentReference resolveReferences[MAX_COLOR_TARGET_BINDINGS + 1];
VkAttachmentReference depthStencilAttachmentReference;
VkRenderPassCreateInfo renderPassCreateInfo;
VkSubpassDescription subpass;
VkRenderPass renderPass;
Uint32 i;
Uint32 attachmentDescriptionCount = 0;
Uint32 colorAttachmentReferenceCount = 0;
Uint32 resolveReferenceCount = 0;
VulkanTexture *texture = NULL;
for (i = 0; i < colorAttachmentCount; i += 1)
{
texture = ((VulkanTextureContainer*) colorAttachmentInfos[i].textureSlice.texture)->activeTextureHandle->vulkanTexture;
if (texture->sampleCount > VK_SAMPLE_COUNT_1_BIT)
{
/* Resolve attachment and multisample attachment */
attachmentDescriptions[attachmentDescriptionCount].flags = 0;
attachmentDescriptions[attachmentDescriptionCount].format = texture->format;
attachmentDescriptions[attachmentDescriptionCount].samples =
VK_SAMPLE_COUNT_1_BIT;
attachmentDescriptions[attachmentDescriptionCount].loadOp = SDLToVK_LoadOp[
colorAttachmentInfos[i].loadOp
];
attachmentDescriptions[attachmentDescriptionCount].storeOp =
VK_ATTACHMENT_STORE_OP_STORE; /* Always store the resolve texture */
attachmentDescriptions[attachmentDescriptionCount].stencilLoadOp =
VK_ATTACHMENT_LOAD_OP_DONT_CARE;
attachmentDescriptions[attachmentDescriptionCount].stencilStoreOp =
VK_ATTACHMENT_STORE_OP_DONT_CARE;
attachmentDescriptions[attachmentDescriptionCount].initialLayout =
VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL;
attachmentDescriptions[attachmentDescriptionCount].finalLayout =
VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL;
resolveReferences[resolveReferenceCount].attachment =
attachmentDescriptionCount;
resolveReferences[resolveReferenceCount].layout =
VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL;
attachmentDescriptionCount += 1;
resolveReferenceCount += 1;
attachmentDescriptions[attachmentDescriptionCount].flags = 0;
attachmentDescriptions[attachmentDescriptionCount].format = texture->format;
attachmentDescriptions[attachmentDescriptionCount].samples = texture->sampleCount;
attachmentDescriptions[attachmentDescriptionCount].loadOp = SDLToVK_LoadOp[
colorAttachmentInfos[i].loadOp
];
attachmentDescriptions[attachmentDescriptionCount].storeOp = SDLToVK_StoreOp[
colorAttachmentInfos[i].storeOp
];
attachmentDescriptions[attachmentDescriptionCount].stencilLoadOp =
VK_ATTACHMENT_LOAD_OP_DONT_CARE;
attachmentDescriptions[attachmentDescriptionCount].stencilStoreOp =
VK_ATTACHMENT_STORE_OP_DONT_CARE;
attachmentDescriptions[attachmentDescriptionCount].initialLayout =
VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL;
attachmentDescriptions[attachmentDescriptionCount].finalLayout =
VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL;
colorAttachmentReferences[colorAttachmentReferenceCount].attachment =
attachmentDescriptionCount;
colorAttachmentReferences[colorAttachmentReferenceCount].layout =
VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL;
attachmentDescriptionCount += 1;
colorAttachmentReferenceCount += 1;
}
else
{
attachmentDescriptions[attachmentDescriptionCount].flags = 0;
attachmentDescriptions[attachmentDescriptionCount].format = texture->format;
attachmentDescriptions[attachmentDescriptionCount].samples =
VK_SAMPLE_COUNT_1_BIT;
attachmentDescriptions[attachmentDescriptionCount].loadOp = SDLToVK_LoadOp[
colorAttachmentInfos[i].loadOp
];
attachmentDescriptions[attachmentDescriptionCount].storeOp =
VK_ATTACHMENT_STORE_OP_STORE; /* Always store non-MSAA textures */
attachmentDescriptions[attachmentDescriptionCount].stencilLoadOp =
VK_ATTACHMENT_LOAD_OP_DONT_CARE;
attachmentDescriptions[attachmentDescriptionCount].stencilStoreOp =
VK_ATTACHMENT_STORE_OP_DONT_CARE;
attachmentDescriptions[attachmentDescriptionCount].initialLayout =
VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL;
attachmentDescriptions[attachmentDescriptionCount].finalLayout =
VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL;
colorAttachmentReferences[colorAttachmentReferenceCount].attachment = attachmentDescriptionCount;
colorAttachmentReferences[colorAttachmentReferenceCount].layout =
VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL;
attachmentDescriptionCount += 1;
colorAttachmentReferenceCount += 1;
}
}
subpass.pipelineBindPoint = VK_PIPELINE_BIND_POINT_GRAPHICS;
subpass.flags = 0;
subpass.inputAttachmentCount = 0;
subpass.pInputAttachments = NULL;
subpass.colorAttachmentCount = colorAttachmentCount;
subpass.pColorAttachments = colorAttachmentReferences;
subpass.preserveAttachmentCount = 0;
subpass.pPreserveAttachments = NULL;
if (depthStencilAttachmentInfo == NULL)
{
subpass.pDepthStencilAttachment = NULL;
}
else
{
texture = ((VulkanTextureContainer*) depthStencilAttachmentInfo->textureSlice.texture)->activeTextureHandle->vulkanTexture;
attachmentDescriptions[attachmentDescriptionCount].flags = 0;
attachmentDescriptions[attachmentDescriptionCount].format = texture->format;
attachmentDescriptions[attachmentDescriptionCount].samples = texture->sampleCount;
attachmentDescriptions[attachmentDescriptionCount].loadOp = SDLToVK_LoadOp[
depthStencilAttachmentInfo->loadOp
];
attachmentDescriptions[attachmentDescriptionCount].storeOp = SDLToVK_StoreOp[
depthStencilAttachmentInfo->storeOp
];
attachmentDescriptions[attachmentDescriptionCount].stencilLoadOp = SDLToVK_LoadOp[
depthStencilAttachmentInfo->stencilLoadOp
];
attachmentDescriptions[attachmentDescriptionCount].stencilStoreOp = SDLToVK_StoreOp[
depthStencilAttachmentInfo->stencilStoreOp
];
attachmentDescriptions[attachmentDescriptionCount].initialLayout =
VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL;
attachmentDescriptions[attachmentDescriptionCount].finalLayout =
VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL;
depthStencilAttachmentReference.attachment =
attachmentDescriptionCount;
depthStencilAttachmentReference.layout =
VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL;
subpass.pDepthStencilAttachment =
&depthStencilAttachmentReference;
attachmentDescriptionCount += 1;
}
if (texture != NULL && texture->sampleCount > VK_SAMPLE_COUNT_1_BIT)
{
subpass.pResolveAttachments = resolveReferences;
}
else
{
subpass.pResolveAttachments = NULL;
}
renderPassCreateInfo.sType = VK_STRUCTURE_TYPE_RENDER_PASS_CREATE_INFO;
renderPassCreateInfo.pNext = NULL;
renderPassCreateInfo.flags = 0;
renderPassCreateInfo.pAttachments = attachmentDescriptions;
renderPassCreateInfo.attachmentCount = attachmentDescriptionCount;
renderPassCreateInfo.subpassCount = 1;
renderPassCreateInfo.pSubpasses = &subpass;
renderPassCreateInfo.dependencyCount = 0;
renderPassCreateInfo.pDependencies = NULL;
vulkanResult = renderer->vkCreateRenderPass(
renderer->logicalDevice,
&renderPassCreateInfo,
NULL,
&renderPass
);
if (vulkanResult != VK_SUCCESS)
{
renderPass = VK_NULL_HANDLE;
LogVulkanResultAsError("vkCreateRenderPass", vulkanResult);
}
return renderPass;
}
static VkRenderPass VULKAN_INTERNAL_CreateTransientRenderPass(
VulkanRenderer *renderer,
Refresh_GraphicsPipelineAttachmentInfo attachmentInfo,
VkSampleCountFlagBits sampleCount
) {
VkAttachmentDescription attachmentDescriptions[2 * MAX_COLOR_TARGET_BINDINGS + 1];
VkAttachmentReference colorAttachmentReferences[MAX_COLOR_TARGET_BINDINGS];
VkAttachmentReference resolveReferences[MAX_COLOR_TARGET_BINDINGS + 1];
VkAttachmentReference depthStencilAttachmentReference;
Refresh_ColorAttachmentDescription attachmentDescription;
VkSubpassDescription subpass;
VkRenderPassCreateInfo renderPassCreateInfo;
VkRenderPass renderPass;
VkResult result;
Uint32 multisampling = 0;
Uint32 attachmentDescriptionCount = 0;
Uint32 colorAttachmentReferenceCount = 0;
Uint32 resolveReferenceCount = 0;
Uint32 i;
for (i = 0; i < attachmentInfo.colorAttachmentCount; i += 1)
{
attachmentDescription = attachmentInfo.colorAttachmentDescriptions[i];
if (sampleCount > VK_SAMPLE_COUNT_1_BIT)
{
multisampling = 1;
/* Resolve attachment and multisample attachment */
attachmentDescriptions[attachmentDescriptionCount].flags = 0;
attachmentDescriptions[attachmentDescriptionCount].format = SDLToVK_SurfaceFormat[
attachmentDescription.format
];
attachmentDescriptions[attachmentDescriptionCount].samples = VK_SAMPLE_COUNT_1_BIT;
attachmentDescriptions[attachmentDescriptionCount].loadOp = VK_ATTACHMENT_LOAD_OP_DONT_CARE;
attachmentDescriptions[attachmentDescriptionCount].storeOp = VK_ATTACHMENT_STORE_OP_DONT_CARE;
attachmentDescriptions[attachmentDescriptionCount].stencilLoadOp = VK_ATTACHMENT_LOAD_OP_DONT_CARE;
attachmentDescriptions[attachmentDescriptionCount].stencilStoreOp = VK_ATTACHMENT_STORE_OP_DONT_CARE;
attachmentDescriptions[attachmentDescriptionCount].initialLayout = VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL;
attachmentDescriptions[attachmentDescriptionCount].finalLayout = VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL;
resolveReferences[resolveReferenceCount].attachment = attachmentDescriptionCount;
resolveReferences[resolveReferenceCount].layout = VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL;
attachmentDescriptionCount += 1;
resolveReferenceCount += 1;
attachmentDescriptions[attachmentDescriptionCount].flags = 0;
attachmentDescriptions[attachmentDescriptionCount].format = SDLToVK_SurfaceFormat[
attachmentDescription.format
];
attachmentDescriptions[attachmentDescriptionCount].samples = sampleCount;
attachmentDescriptions[attachmentDescriptionCount].loadOp = VK_ATTACHMENT_LOAD_OP_DONT_CARE;
attachmentDescriptions[attachmentDescriptionCount].storeOp = VK_ATTACHMENT_STORE_OP_DONT_CARE;
attachmentDescriptions[attachmentDescriptionCount].stencilLoadOp = VK_ATTACHMENT_LOAD_OP_DONT_CARE;
attachmentDescriptions[attachmentDescriptionCount].stencilStoreOp = VK_ATTACHMENT_STORE_OP_DONT_CARE;
attachmentDescriptions[attachmentDescriptionCount].initialLayout = VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL;
attachmentDescriptions[attachmentDescriptionCount].finalLayout = VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL;
colorAttachmentReferences[colorAttachmentReferenceCount].attachment =
attachmentDescriptionCount;
colorAttachmentReferences[colorAttachmentReferenceCount].layout =
VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL;
attachmentDescriptionCount += 1;
colorAttachmentReferenceCount += 1;
}
else
{
attachmentDescriptions[attachmentDescriptionCount].flags = 0;
attachmentDescriptions[attachmentDescriptionCount].format = SDLToVK_SurfaceFormat[
attachmentDescription.format
];
attachmentDescriptions[attachmentDescriptionCount].samples =
VK_SAMPLE_COUNT_1_BIT;
attachmentDescriptions[attachmentDescriptionCount].loadOp =
VK_ATTACHMENT_LOAD_OP_DONT_CARE;
attachmentDescriptions[attachmentDescriptionCount].storeOp =
VK_ATTACHMENT_STORE_OP_DONT_CARE;
attachmentDescriptions[attachmentDescriptionCount].stencilLoadOp =
VK_ATTACHMENT_LOAD_OP_DONT_CARE;
attachmentDescriptions[attachmentDescriptionCount].stencilStoreOp =
VK_ATTACHMENT_STORE_OP_DONT_CARE;
attachmentDescriptions[attachmentDescriptionCount].initialLayout =
VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL;
attachmentDescriptions[attachmentDescriptionCount].finalLayout =
VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL;
colorAttachmentReferences[colorAttachmentReferenceCount].attachment = attachmentDescriptionCount;
colorAttachmentReferences[colorAttachmentReferenceCount].layout =
VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL;
attachmentDescriptionCount += 1;
colorAttachmentReferenceCount += 1;
}
}
subpass.pipelineBindPoint = VK_PIPELINE_BIND_POINT_GRAPHICS;
subpass.flags = 0;
subpass.inputAttachmentCount = 0;
subpass.pInputAttachments = NULL;
subpass.colorAttachmentCount = attachmentInfo.colorAttachmentCount;
subpass.pColorAttachments = colorAttachmentReferences;
subpass.preserveAttachmentCount = 0;
subpass.pPreserveAttachments = NULL;
if (attachmentInfo.hasDepthStencilAttachment)
{
attachmentDescriptions[attachmentDescriptionCount].flags = 0;
attachmentDescriptions[attachmentDescriptionCount].format =
SDLToVK_SurfaceFormat[attachmentInfo.depthStencilFormat];
attachmentDescriptions[attachmentDescriptionCount].samples = sampleCount;
attachmentDescriptions[attachmentDescriptionCount].loadOp = VK_ATTACHMENT_LOAD_OP_DONT_CARE;
attachmentDescriptions[attachmentDescriptionCount].storeOp = VK_ATTACHMENT_STORE_OP_DONT_CARE;
attachmentDescriptions[attachmentDescriptionCount].stencilLoadOp = VK_ATTACHMENT_LOAD_OP_DONT_CARE;
attachmentDescriptions[attachmentDescriptionCount].stencilStoreOp = VK_ATTACHMENT_STORE_OP_DONT_CARE;
attachmentDescriptions[attachmentDescriptionCount].initialLayout =
VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL;
attachmentDescriptions[attachmentDescriptionCount].finalLayout =
VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL;
depthStencilAttachmentReference.attachment =
attachmentDescriptionCount;
depthStencilAttachmentReference.layout =
VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL;
subpass.pDepthStencilAttachment =
&depthStencilAttachmentReference;
attachmentDescriptionCount += 1;
}
else
{
subpass.pDepthStencilAttachment = NULL;
}
if (multisampling)
{
subpass.pResolveAttachments = resolveReferences;
}
else
{
subpass.pResolveAttachments = NULL;
}
renderPassCreateInfo.sType = VK_STRUCTURE_TYPE_RENDER_PASS_CREATE_INFO;
renderPassCreateInfo.pNext = NULL;
renderPassCreateInfo.flags = 0;
renderPassCreateInfo.pAttachments = attachmentDescriptions;
renderPassCreateInfo.attachmentCount = attachmentDescriptionCount;
renderPassCreateInfo.subpassCount = 1;
renderPassCreateInfo.pSubpasses = &subpass;
renderPassCreateInfo.dependencyCount = 0;
renderPassCreateInfo.pDependencies = NULL;
result = renderer->vkCreateRenderPass(
renderer->logicalDevice,
&renderPassCreateInfo,
NULL,
&renderPass
);
if (result != VK_SUCCESS)
{
renderPass = VK_NULL_HANDLE;
LogVulkanResultAsError("vkCreateRenderPass", result);
}
return renderPass;
}
static Refresh_GraphicsPipeline* VULKAN_CreateGraphicsPipeline(
Refresh_Renderer *driverData,
Refresh_GraphicsPipelineCreateInfo *pipelineCreateInfo
) {
VkResult vulkanResult;
Uint32 i;
VkSampleCountFlagBits actualSampleCount;
VulkanGraphicsPipeline *graphicsPipeline = (VulkanGraphicsPipeline*) SDL_malloc(sizeof(VulkanGraphicsPipeline));
VkGraphicsPipelineCreateInfo vkPipelineCreateInfo;
VkPipelineShaderStageCreateInfo shaderStageCreateInfos[2];
VkPipelineVertexInputStateCreateInfo vertexInputStateCreateInfo;
VkPipelineVertexInputDivisorStateCreateInfoEXT divisorStateCreateInfo;
VkVertexInputBindingDescription *vertexInputBindingDescriptions = SDL_stack_alloc(VkVertexInputBindingDescription, pipelineCreateInfo->vertexInputState.vertexBindingCount);
VkVertexInputAttributeDescription *vertexInputAttributeDescriptions = SDL_stack_alloc(VkVertexInputAttributeDescription, pipelineCreateInfo->vertexInputState.vertexAttributeCount);
VkVertexInputBindingDivisorDescriptionEXT *divisorDescriptions = SDL_stack_alloc(VkVertexInputBindingDivisorDescriptionEXT, pipelineCreateInfo->vertexInputState.vertexBindingCount);
Uint32 divisorDescriptionCount = 0;
VkPipelineInputAssemblyStateCreateInfo inputAssemblyStateCreateInfo;
VkPipelineViewportStateCreateInfo viewportStateCreateInfo;
VkPipelineRasterizationStateCreateInfo rasterizationStateCreateInfo;
VkPipelineMultisampleStateCreateInfo multisampleStateCreateInfo;
VkPipelineDepthStencilStateCreateInfo depthStencilStateCreateInfo;
VkStencilOpState frontStencilState;
VkStencilOpState backStencilState;
VkPipelineColorBlendStateCreateInfo colorBlendStateCreateInfo;
VkPipelineColorBlendAttachmentState *colorBlendAttachmentStates = SDL_stack_alloc(
VkPipelineColorBlendAttachmentState,
pipelineCreateInfo->attachmentInfo.colorAttachmentCount
);
static const VkDynamicState dynamicStates[] =
{
VK_DYNAMIC_STATE_VIEWPORT,
VK_DYNAMIC_STATE_SCISSOR
};
VkPipelineDynamicStateCreateInfo dynamicStateCreateInfo;
VulkanRenderer *renderer = (VulkanRenderer*) driverData;
/* Find a compatible sample count to use */
actualSampleCount = VULKAN_INTERNAL_GetMaxMultiSampleCount(
renderer,
SDLToVK_SampleCount[pipelineCreateInfo->multisampleState.multisampleCount]
);
/* Create a "compatible" render pass */
VkRenderPass transientRenderPass = VULKAN_INTERNAL_CreateTransientRenderPass(
renderer,
pipelineCreateInfo->attachmentInfo,
actualSampleCount
);
/* Dynamic state */
dynamicStateCreateInfo.sType = VK_STRUCTURE_TYPE_PIPELINE_DYNAMIC_STATE_CREATE_INFO;
dynamicStateCreateInfo.pNext = NULL;
dynamicStateCreateInfo.flags = 0;
dynamicStateCreateInfo.dynamicStateCount = SDL_arraysize(dynamicStates);
dynamicStateCreateInfo.pDynamicStates = dynamicStates;
/* Shader stages */
graphicsPipeline->vertexShader = (VulkanShader*) pipelineCreateInfo->vertexShader;
SDL_AtomicIncRef(&graphicsPipeline->vertexShader->referenceCount);
shaderStageCreateInfos[0].sType = VK_STRUCTURE_TYPE_PIPELINE_SHADER_STAGE_CREATE_INFO;
shaderStageCreateInfos[0].pNext = NULL;
shaderStageCreateInfos[0].flags = 0;
shaderStageCreateInfos[0].stage = VK_SHADER_STAGE_VERTEX_BIT;
shaderStageCreateInfos[0].module = graphicsPipeline->vertexShader->shaderModule;
shaderStageCreateInfos[0].pName = graphicsPipeline->vertexShader->entryPointName;
shaderStageCreateInfos[0].pSpecializationInfo = NULL;
graphicsPipeline->fragmentShader = (VulkanShader*) pipelineCreateInfo->fragmentShader;
SDL_AtomicIncRef(&graphicsPipeline->fragmentShader->referenceCount);
shaderStageCreateInfos[1].sType = VK_STRUCTURE_TYPE_PIPELINE_SHADER_STAGE_CREATE_INFO;
shaderStageCreateInfos[1].pNext = NULL;
shaderStageCreateInfos[1].flags = 0;
shaderStageCreateInfos[1].stage = VK_SHADER_STAGE_FRAGMENT_BIT;
shaderStageCreateInfos[1].module = graphicsPipeline->fragmentShader->shaderModule;
shaderStageCreateInfos[1].pName = graphicsPipeline->fragmentShader->entryPointName;
shaderStageCreateInfos[1].pSpecializationInfo = NULL;
/* Vertex input */
for (i = 0; i < pipelineCreateInfo->vertexInputState.vertexBindingCount; i += 1)
{
vertexInputBindingDescriptions[i].binding = pipelineCreateInfo->vertexInputState.vertexBindings[i].binding;
vertexInputBindingDescriptions[i].inputRate = SDLToVK_VertexInputRate[
pipelineCreateInfo->vertexInputState.vertexBindings[i].inputRate
];
vertexInputBindingDescriptions[i].stride = pipelineCreateInfo->vertexInputState.vertexBindings[i].stride;
if (pipelineCreateInfo->vertexInputState.vertexBindings[i].inputRate == REFRESH_VERTEXINPUTRATE_INSTANCE)
{
divisorDescriptionCount += 1;
}
}
for (i = 0; i < pipelineCreateInfo->vertexInputState.vertexAttributeCount; i += 1)
{
vertexInputAttributeDescriptions[i].binding = pipelineCreateInfo->vertexInputState.vertexAttributes[i].binding;
vertexInputAttributeDescriptions[i].format = SDLToVK_VertexFormat[
pipelineCreateInfo->vertexInputState.vertexAttributes[i].format
];
vertexInputAttributeDescriptions[i].location = pipelineCreateInfo->vertexInputState.vertexAttributes[i].location;
vertexInputAttributeDescriptions[i].offset = pipelineCreateInfo->vertexInputState.vertexAttributes[i].offset;
}
vertexInputStateCreateInfo.sType = VK_STRUCTURE_TYPE_PIPELINE_VERTEX_INPUT_STATE_CREATE_INFO;
vertexInputStateCreateInfo.pNext = NULL;
vertexInputStateCreateInfo.flags = 0;
vertexInputStateCreateInfo.vertexBindingDescriptionCount = pipelineCreateInfo->vertexInputState.vertexBindingCount;
vertexInputStateCreateInfo.pVertexBindingDescriptions = vertexInputBindingDescriptions;
vertexInputStateCreateInfo.vertexAttributeDescriptionCount = pipelineCreateInfo->vertexInputState.vertexAttributeCount;
vertexInputStateCreateInfo.pVertexAttributeDescriptions = vertexInputAttributeDescriptions;
if (divisorDescriptionCount > 0)
{
divisorDescriptionCount = 0;
for (i = 0; i < pipelineCreateInfo->vertexInputState.vertexBindingCount; i += 1)
{
if (pipelineCreateInfo->vertexInputState.vertexBindings[i].inputRate == REFRESH_VERTEXINPUTRATE_INSTANCE)
{
divisorDescriptions[divisorDescriptionCount].binding = pipelineCreateInfo->vertexInputState.vertexBindings[i].binding;
divisorDescriptions[divisorDescriptionCount].divisor = pipelineCreateInfo->vertexInputState.vertexBindings[i].stepRate;
divisorDescriptionCount += 1;
}
}
divisorStateCreateInfo.sType = VK_STRUCTURE_TYPE_PIPELINE_VERTEX_INPUT_DIVISOR_STATE_CREATE_INFO_EXT;
divisorStateCreateInfo.pNext = NULL;
divisorStateCreateInfo.vertexBindingDivisorCount = divisorDescriptionCount;
divisorStateCreateInfo.pVertexBindingDivisors = divisorDescriptions;
vertexInputStateCreateInfo.pNext = &divisorStateCreateInfo;
}
/* Topology */
inputAssemblyStateCreateInfo.sType = VK_STRUCTURE_TYPE_PIPELINE_INPUT_ASSEMBLY_STATE_CREATE_INFO;
inputAssemblyStateCreateInfo.pNext = NULL;
inputAssemblyStateCreateInfo.flags = 0;
inputAssemblyStateCreateInfo.primitiveRestartEnable = VK_FALSE;
inputAssemblyStateCreateInfo.topology = SDLToVK_PrimitiveType[
pipelineCreateInfo->primitiveType
];
graphicsPipeline->primitiveType = pipelineCreateInfo->primitiveType;
/* Viewport */
/* NOTE: viewport and scissor are dynamic, and must be set using the command buffer */
viewportStateCreateInfo.sType = VK_STRUCTURE_TYPE_PIPELINE_VIEWPORT_STATE_CREATE_INFO;
viewportStateCreateInfo.pNext = NULL;
viewportStateCreateInfo.flags = 0;
viewportStateCreateInfo.viewportCount = 1;
viewportStateCreateInfo.pViewports = NULL;
viewportStateCreateInfo.scissorCount = 1;
viewportStateCreateInfo.pScissors = NULL;
/* Rasterization */
rasterizationStateCreateInfo.sType = VK_STRUCTURE_TYPE_PIPELINE_RASTERIZATION_STATE_CREATE_INFO;
rasterizationStateCreateInfo.pNext = NULL;
rasterizationStateCreateInfo.flags = 0;
rasterizationStateCreateInfo.depthClampEnable = VK_FALSE;
rasterizationStateCreateInfo.rasterizerDiscardEnable = VK_FALSE;
rasterizationStateCreateInfo.polygonMode = SDLToVK_PolygonMode[
pipelineCreateInfo->rasterizerState.fillMode
];
rasterizationStateCreateInfo.cullMode = SDLToVK_CullMode[
pipelineCreateInfo->rasterizerState.cullMode
];
rasterizationStateCreateInfo.frontFace = SDLToVK_FrontFace[
pipelineCreateInfo->rasterizerState.frontFace
];
rasterizationStateCreateInfo.depthBiasEnable =
pipelineCreateInfo->rasterizerState.depthBiasEnable;
rasterizationStateCreateInfo.depthBiasConstantFactor =
pipelineCreateInfo->rasterizerState.depthBiasConstantFactor;
rasterizationStateCreateInfo.depthBiasClamp =
pipelineCreateInfo->rasterizerState.depthBiasClamp;
rasterizationStateCreateInfo.depthBiasSlopeFactor =
pipelineCreateInfo->rasterizerState.depthBiasSlopeFactor;
rasterizationStateCreateInfo.lineWidth = 1.0f;
/* Multisample */
multisampleStateCreateInfo.sType = VK_STRUCTURE_TYPE_PIPELINE_MULTISAMPLE_STATE_CREATE_INFO;
multisampleStateCreateInfo.pNext = NULL;
multisampleStateCreateInfo.flags = 0;
multisampleStateCreateInfo.rasterizationSamples = actualSampleCount;
multisampleStateCreateInfo.sampleShadingEnable = VK_FALSE;
multisampleStateCreateInfo.minSampleShading = 1.0f;
multisampleStateCreateInfo.pSampleMask =
&pipelineCreateInfo->multisampleState.sampleMask;
multisampleStateCreateInfo.alphaToCoverageEnable = VK_FALSE;
multisampleStateCreateInfo.alphaToOneEnable = VK_FALSE;
/* Depth Stencil State */
frontStencilState.failOp = SDLToVK_StencilOp[
pipelineCreateInfo->depthStencilState.frontStencilState.failOp
];
frontStencilState.passOp = SDLToVK_StencilOp[
pipelineCreateInfo->depthStencilState.frontStencilState.passOp
];
frontStencilState.depthFailOp = SDLToVK_StencilOp[
pipelineCreateInfo->depthStencilState.frontStencilState.depthFailOp
];
frontStencilState.compareOp = SDLToVK_CompareOp[
pipelineCreateInfo->depthStencilState.frontStencilState.compareOp
];
frontStencilState.compareMask =
pipelineCreateInfo->depthStencilState.compareMask;
frontStencilState.writeMask =
pipelineCreateInfo->depthStencilState.writeMask;
frontStencilState.reference =
pipelineCreateInfo->depthStencilState.reference;
backStencilState.failOp = SDLToVK_StencilOp[
pipelineCreateInfo->depthStencilState.backStencilState.failOp
];
backStencilState.passOp = SDLToVK_StencilOp[
pipelineCreateInfo->depthStencilState.backStencilState.passOp
];
backStencilState.depthFailOp = SDLToVK_StencilOp[
pipelineCreateInfo->depthStencilState.backStencilState.depthFailOp
];
backStencilState.compareOp = SDLToVK_CompareOp[
pipelineCreateInfo->depthStencilState.backStencilState.compareOp
];
backStencilState.compareMask =
pipelineCreateInfo->depthStencilState.compareMask;
backStencilState.writeMask =
pipelineCreateInfo->depthStencilState.writeMask;
backStencilState.reference =
pipelineCreateInfo->depthStencilState.reference;
depthStencilStateCreateInfo.sType = VK_STRUCTURE_TYPE_PIPELINE_DEPTH_STENCIL_STATE_CREATE_INFO;
depthStencilStateCreateInfo.pNext = NULL;
depthStencilStateCreateInfo.flags = 0;
depthStencilStateCreateInfo.depthTestEnable =
pipelineCreateInfo->depthStencilState.depthTestEnable;
depthStencilStateCreateInfo.depthWriteEnable =
pipelineCreateInfo->depthStencilState.depthWriteEnable;
depthStencilStateCreateInfo.depthCompareOp = SDLToVK_CompareOp[
pipelineCreateInfo->depthStencilState.compareOp
];
depthStencilStateCreateInfo.depthBoundsTestEnable =
pipelineCreateInfo->depthStencilState.depthBoundsTestEnable;
depthStencilStateCreateInfo.stencilTestEnable =
pipelineCreateInfo->depthStencilState.stencilTestEnable;
depthStencilStateCreateInfo.front = frontStencilState;
depthStencilStateCreateInfo.back = backStencilState;
depthStencilStateCreateInfo.minDepthBounds =
pipelineCreateInfo->depthStencilState.minDepthBounds;
depthStencilStateCreateInfo.maxDepthBounds =
pipelineCreateInfo->depthStencilState.maxDepthBounds;
/* Color Blend */
for (i = 0; i < pipelineCreateInfo->attachmentInfo.colorAttachmentCount; i += 1)
{
Refresh_ColorAttachmentBlendState blendState = pipelineCreateInfo->attachmentInfo.colorAttachmentDescriptions[i].blendState;
colorBlendAttachmentStates[i].blendEnable =
blendState.blendEnable;
colorBlendAttachmentStates[i].srcColorBlendFactor = SDLToVK_BlendFactor[
blendState.srcColorBlendFactor
];
colorBlendAttachmentStates[i].dstColorBlendFactor = SDLToVK_BlendFactor[
blendState.dstColorBlendFactor
];
colorBlendAttachmentStates[i].colorBlendOp = SDLToVK_BlendOp[
blendState.colorBlendOp
];
colorBlendAttachmentStates[i].srcAlphaBlendFactor = SDLToVK_BlendFactor[
blendState.srcAlphaBlendFactor
];
colorBlendAttachmentStates[i].dstAlphaBlendFactor = SDLToVK_BlendFactor[
blendState.dstAlphaBlendFactor
];
colorBlendAttachmentStates[i].alphaBlendOp = SDLToVK_BlendOp[
blendState.alphaBlendOp
];
colorBlendAttachmentStates[i].colorWriteMask =
blendState.colorWriteMask;
}
colorBlendStateCreateInfo.sType = VK_STRUCTURE_TYPE_PIPELINE_COLOR_BLEND_STATE_CREATE_INFO;
colorBlendStateCreateInfo.pNext = NULL;
colorBlendStateCreateInfo.flags = 0;
colorBlendStateCreateInfo.attachmentCount =
pipelineCreateInfo->attachmentInfo.colorAttachmentCount;
colorBlendStateCreateInfo.pAttachments =
colorBlendAttachmentStates;
colorBlendStateCreateInfo.blendConstants[0] =
pipelineCreateInfo->blendConstants[0];
colorBlendStateCreateInfo.blendConstants[1] =
pipelineCreateInfo->blendConstants[1];
colorBlendStateCreateInfo.blendConstants[2] =
pipelineCreateInfo->blendConstants[2];
colorBlendStateCreateInfo.blendConstants[3] =
pipelineCreateInfo->blendConstants[3];
/* We don't support LogicOp, so this is easy. */
colorBlendStateCreateInfo.logicOpEnable = VK_FALSE;
colorBlendStateCreateInfo.logicOp = 0;
/* Pipeline Layout */
if (!VULKAN_INTERNAL_InitializeGraphicsPipelineResourceLayout(
renderer,
&pipelineCreateInfo->vertexResourceInfo,
&pipelineCreateInfo->fragmentResourceInfo,
&graphicsPipeline->resourceLayout
)) {
SDL_stack_free(vertexInputBindingDescriptions);
SDL_stack_free(vertexInputAttributeDescriptions);
SDL_stack_free(colorBlendAttachmentStates);
SDL_stack_free(divisorDescriptions);
SDL_free(graphicsPipeline);
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Failed to initialize pipeline resource layout!");
return NULL;
}
/* Pipeline */
vkPipelineCreateInfo.sType = VK_STRUCTURE_TYPE_GRAPHICS_PIPELINE_CREATE_INFO;
vkPipelineCreateInfo.pNext = NULL;
vkPipelineCreateInfo.flags = 0;
vkPipelineCreateInfo.stageCount = 2;
vkPipelineCreateInfo.pStages = shaderStageCreateInfos;
vkPipelineCreateInfo.pVertexInputState = &vertexInputStateCreateInfo;
vkPipelineCreateInfo.pInputAssemblyState = &inputAssemblyStateCreateInfo;
vkPipelineCreateInfo.pTessellationState = VK_NULL_HANDLE;
vkPipelineCreateInfo.pViewportState = &viewportStateCreateInfo;
vkPipelineCreateInfo.pRasterizationState = &rasterizationStateCreateInfo;
vkPipelineCreateInfo.pMultisampleState = &multisampleStateCreateInfo;
vkPipelineCreateInfo.pDepthStencilState = &depthStencilStateCreateInfo;
vkPipelineCreateInfo.pColorBlendState = &colorBlendStateCreateInfo;
vkPipelineCreateInfo.pDynamicState = &dynamicStateCreateInfo;
vkPipelineCreateInfo.layout = graphicsPipeline->resourceLayout.pipelineLayout;
vkPipelineCreateInfo.renderPass = transientRenderPass;
vkPipelineCreateInfo.subpass = 0;
vkPipelineCreateInfo.basePipelineHandle = VK_NULL_HANDLE;
vkPipelineCreateInfo.basePipelineIndex = 0;
/* TODO: enable pipeline caching */
vulkanResult = renderer->vkCreateGraphicsPipelines(
renderer->logicalDevice,
VK_NULL_HANDLE,
1,
&vkPipelineCreateInfo,
NULL,
&graphicsPipeline->pipeline
);
SDL_stack_free(vertexInputBindingDescriptions);
SDL_stack_free(vertexInputAttributeDescriptions);
SDL_stack_free(colorBlendAttachmentStates);
SDL_stack_free(divisorDescriptions);
renderer->vkDestroyRenderPass(
renderer->logicalDevice,
transientRenderPass,
NULL
);
if (vulkanResult != VK_SUCCESS)
{
SDL_free(graphicsPipeline);
LogVulkanResultAsError("vkCreateGraphicsPipelines", vulkanResult);
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Failed to create graphics pipeline!");
return NULL;
}
SDL_AtomicSet(&graphicsPipeline->referenceCount, 0);
return (Refresh_GraphicsPipeline*) graphicsPipeline;
}
static Refresh_ComputePipeline* VULKAN_CreateComputePipeline(
Refresh_Renderer *driverData,
Refresh_ComputePipelineCreateInfo *pipelineCreateInfo
) {
VkComputePipelineCreateInfo computePipelineCreateInfo;
VkPipelineShaderStageCreateInfo pipelineShaderStageCreateInfo;
VulkanRenderer *renderer = (VulkanRenderer*) driverData;
VulkanComputePipeline *vulkanComputePipeline = SDL_malloc(sizeof(VulkanComputePipeline));
vulkanComputePipeline->computeShader = (VulkanShader*) pipelineCreateInfo->computeShader;
SDL_AtomicIncRef(&vulkanComputePipeline->computeShader->referenceCount);
pipelineShaderStageCreateInfo.sType = VK_STRUCTURE_TYPE_PIPELINE_SHADER_STAGE_CREATE_INFO;
pipelineShaderStageCreateInfo.pNext = NULL;
pipelineShaderStageCreateInfo.flags = 0;
pipelineShaderStageCreateInfo.stage = VK_SHADER_STAGE_COMPUTE_BIT;
pipelineShaderStageCreateInfo.module = vulkanComputePipeline->computeShader->shaderModule;
pipelineShaderStageCreateInfo.pName = vulkanComputePipeline->computeShader->entryPointName;
pipelineShaderStageCreateInfo.pSpecializationInfo = NULL;
if (!VULKAN_INTERNAL_InitializeComputePipelineResourceLayout(
renderer,
&pipelineCreateInfo->pipelineResourceInfo,
&vulkanComputePipeline->resourceLayout
))
{
SDL_free(vulkanComputePipeline);
return NULL;
}
computePipelineCreateInfo.sType = VK_STRUCTURE_TYPE_COMPUTE_PIPELINE_CREATE_INFO;
computePipelineCreateInfo.pNext = NULL;
computePipelineCreateInfo.flags = 0;
computePipelineCreateInfo.stage = pipelineShaderStageCreateInfo;
computePipelineCreateInfo.layout =
vulkanComputePipeline->resourceLayout.pipelineLayout;
computePipelineCreateInfo.basePipelineHandle = (VkPipeline) VK_NULL_HANDLE;
computePipelineCreateInfo.basePipelineIndex = 0;
renderer->vkCreateComputePipelines(
renderer->logicalDevice,
(VkPipelineCache) VK_NULL_HANDLE,
1,
&computePipelineCreateInfo,
NULL,
&vulkanComputePipeline->pipeline
);
SDL_AtomicSet(&vulkanComputePipeline->referenceCount, 0);
return (Refresh_ComputePipeline*) vulkanComputePipeline;
}
static Refresh_Sampler* VULKAN_CreateSampler(
Refresh_Renderer *driverData,
Refresh_SamplerCreateInfo *samplerCreateInfo
) {
VulkanRenderer* renderer = (VulkanRenderer*)driverData;
VulkanSampler *vulkanSampler = SDL_malloc(sizeof(VulkanSampler));
VkResult vulkanResult;
VkSamplerCreateInfo vkSamplerCreateInfo;
vkSamplerCreateInfo.sType = VK_STRUCTURE_TYPE_SAMPLER_CREATE_INFO;
vkSamplerCreateInfo.pNext = NULL;
vkSamplerCreateInfo.flags = 0;
vkSamplerCreateInfo.magFilter = SDLToVK_Filter[
samplerCreateInfo->magFilter
];
vkSamplerCreateInfo.minFilter = SDLToVK_Filter[
samplerCreateInfo->minFilter
];
vkSamplerCreateInfo.mipmapMode = SDLToVK_SamplerMipmapMode[
samplerCreateInfo->mipmapMode
];
vkSamplerCreateInfo.addressModeU = SDLToVK_SamplerAddressMode[
samplerCreateInfo->addressModeU
];
vkSamplerCreateInfo.addressModeV = SDLToVK_SamplerAddressMode[
samplerCreateInfo->addressModeV
];
vkSamplerCreateInfo.addressModeW = SDLToVK_SamplerAddressMode[
samplerCreateInfo->addressModeW
];
vkSamplerCreateInfo.mipLodBias = samplerCreateInfo->mipLodBias;
vkSamplerCreateInfo.anisotropyEnable = samplerCreateInfo->anisotropyEnable;
vkSamplerCreateInfo.maxAnisotropy = samplerCreateInfo->maxAnisotropy;
vkSamplerCreateInfo.compareEnable = samplerCreateInfo->compareEnable;
vkSamplerCreateInfo.compareOp = SDLToVK_CompareOp[
samplerCreateInfo->compareOp
];
vkSamplerCreateInfo.minLod = samplerCreateInfo->minLod;
vkSamplerCreateInfo.maxLod = samplerCreateInfo->maxLod;
vkSamplerCreateInfo.borderColor = SDLToVK_BorderColor[
samplerCreateInfo->borderColor
];
vkSamplerCreateInfo.unnormalizedCoordinates = VK_FALSE;
vulkanResult = renderer->vkCreateSampler(
renderer->logicalDevice,
&vkSamplerCreateInfo,
NULL,
&vulkanSampler->sampler
);
if (vulkanResult != VK_SUCCESS)
{
SDL_free(vulkanSampler);
LogVulkanResultAsError("vkCreateSampler", vulkanResult);
return NULL;
}
SDL_AtomicSet(&vulkanSampler->referenceCount, 0);
return (Refresh_Sampler*) vulkanSampler;
}
static Refresh_Shader* VULKAN_CreateShader(
Refresh_Renderer *driverData,
Refresh_ShaderCreateInfo *shaderCreateInfo
) {
VulkanShader *vulkanShader = SDL_malloc(sizeof(VulkanShader));
VkResult vulkanResult;
VkShaderModuleCreateInfo vkShaderModuleCreateInfo;
VulkanRenderer *renderer = (VulkanRenderer*) driverData;
if (shaderCreateInfo->format != REFRESH_SHADERFORMAT_SPIRV)
{
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Incompatible shader format for Vulkan");
return NULL;
}
vkShaderModuleCreateInfo.sType = VK_STRUCTURE_TYPE_SHADER_MODULE_CREATE_INFO;
vkShaderModuleCreateInfo.pNext = NULL;
vkShaderModuleCreateInfo.flags = 0;
vkShaderModuleCreateInfo.codeSize = shaderCreateInfo->codeSize;
vkShaderModuleCreateInfo.pCode = (Uint32*) shaderCreateInfo->code;
vulkanResult = renderer->vkCreateShaderModule(
renderer->logicalDevice,
&vkShaderModuleCreateInfo,
NULL,
&vulkanShader->shaderModule
);
if (vulkanResult != VK_SUCCESS)
{
SDL_free(vulkanShader);
LogVulkanResultAsError("vkCreateShaderModule", vulkanResult);
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Failed to create shader module!");
return NULL;
}
vulkanShader->entryPointName = shaderCreateInfo->entryPointName;
SDL_AtomicSet(&vulkanShader->referenceCount, 0);
return (Refresh_Shader*) vulkanShader;
}
static Refresh_Texture* VULKAN_CreateTexture(
Refresh_Renderer *driverData,
Refresh_TextureCreateInfo *textureCreateInfo
) {
VulkanRenderer *renderer = (VulkanRenderer*) driverData;
VkImageAspectFlags imageAspectFlags;
Uint8 isDepthFormat = IsDepthFormat(textureCreateInfo->format);
VkFormat format;
VulkanTextureContainer *container;
VulkanTextureHandle *textureHandle;
VkComponentMapping swizzle = IDENTITY_SWIZZLE;
format = SDLToVK_SurfaceFormat[textureCreateInfo->format];
/* FIXME: We probably need a swizzle table like FNA3D Vulkan does */
if (textureCreateInfo->format == REFRESH_TEXTUREFORMAT_A8)
{
swizzle.r = VK_COMPONENT_SWIZZLE_ZERO;
swizzle.g = VK_COMPONENT_SWIZZLE_ZERO;
swizzle.b = VK_COMPONENT_SWIZZLE_ZERO;
swizzle.a = VK_COMPONENT_SWIZZLE_R;
}
if (isDepthFormat)
{
imageAspectFlags = VK_IMAGE_ASPECT_DEPTH_BIT;
if (IsStencilFormat(textureCreateInfo->format))
{
imageAspectFlags |= VK_IMAGE_ASPECT_STENCIL_BIT;
}
}
else
{
imageAspectFlags = VK_IMAGE_ASPECT_COLOR_BIT;
}
textureHandle = VULKAN_INTERNAL_CreateTextureHandle(
renderer,
textureCreateInfo->width,
textureCreateInfo->height,
textureCreateInfo->depth,
textureCreateInfo->isCube,
textureCreateInfo->layerCount,
textureCreateInfo->levelCount,
SDLToVK_SampleCount[textureCreateInfo->sampleCount],
format,
swizzle,
imageAspectFlags,
textureCreateInfo->usageFlags
);
if (textureHandle == NULL)
{
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Failed to create texture container!");
return NULL;
}
container = SDL_malloc(sizeof(VulkanTextureContainer));
container->canBeCycled = 1;
container->activeTextureHandle = textureHandle;
container->textureCapacity = 1;
container->textureCount = 1 ;
container->textureHandles = SDL_malloc(
container->textureCapacity * sizeof(VulkanTextureHandle*)
);
container->textureHandles[0] = container->activeTextureHandle;
container->debugName = NULL;
textureHandle->container = container;
return (Refresh_Texture*) container;
}
static Refresh_Buffer* VULKAN_CreateBuffer(
Refresh_Renderer *driverData,
Refresh_BufferUsageFlags usageFlags,
Uint32 sizeInBytes
) {
return (Refresh_Buffer*) VULKAN_INTERNAL_CreateBufferContainer(
(VulkanRenderer*) driverData,
sizeInBytes,
usageFlags,
VULKAN_BUFFER_TYPE_GPU
);
}
static VulkanUniformBuffer* VULKAN_INTERNAL_CreateUniformBuffer(
VulkanRenderer *renderer,
Uint32 sizeInBytes
) {
VulkanUniformBuffer *uniformBuffer = SDL_malloc(sizeof(VulkanUniformBuffer));
uniformBuffer->bufferContainer = VULKAN_INTERNAL_CreateBufferContainer(
renderer,
sizeInBytes,
0,
VULKAN_BUFFER_TYPE_UNIFORM
);
uniformBuffer->drawOffset = 0;
uniformBuffer->offset = 0;
return uniformBuffer;
}
static Refresh_TransferBuffer* VULKAN_CreateTransferBuffer(
Refresh_Renderer *driverData,
Refresh_TransferUsage usage, /* ignored on Vulkan */
Refresh_TransferBufferMapFlags mapFlags, /* ignored on Vulkan */
Uint32 sizeInBytes
) {
return (Refresh_TransferBuffer*) VULKAN_INTERNAL_CreateBufferContainer(
(VulkanRenderer*) driverData,
sizeInBytes,
0,
VULKAN_BUFFER_TYPE_TRANSFER
);
}
static void VULKAN_INTERNAL_ReleaseTexture(
VulkanRenderer *renderer,
VulkanTexture *vulkanTexture
) {
if (vulkanTexture->markedForDestroy) { return; }
SDL_LockMutex(renderer->disposeLock);
EXPAND_ARRAY_IF_NEEDED(
renderer->texturesToDestroy,
VulkanTexture*,
renderer->texturesToDestroyCount + 1,
renderer->texturesToDestroyCapacity,
renderer->texturesToDestroyCapacity * 2
)
renderer->texturesToDestroy[
renderer->texturesToDestroyCount
] = vulkanTexture;
renderer->texturesToDestroyCount += 1;
vulkanTexture->markedForDestroy = 1;
SDL_UnlockMutex(renderer->disposeLock);
}
static void VULKAN_ReleaseTexture(
Refresh_Renderer *driverData,
Refresh_Texture *texture
) {
VulkanRenderer *renderer = (VulkanRenderer*) driverData;
VulkanTextureContainer *vulkanTextureContainer = (VulkanTextureContainer*) texture;
Uint32 i;
SDL_LockMutex(renderer->disposeLock);
for (i = 0; i < vulkanTextureContainer->textureCount; i += 1)
{
VULKAN_INTERNAL_ReleaseTexture(renderer, vulkanTextureContainer->textureHandles[i]->vulkanTexture);
SDL_free(vulkanTextureContainer->textureHandles[i]);
}
/* Containers are just client handles, so we can destroy immediately */
if (vulkanTextureContainer->debugName != NULL)
{
SDL_free(vulkanTextureContainer->debugName);
}
SDL_free(vulkanTextureContainer->textureHandles);
SDL_free(vulkanTextureContainer);
SDL_UnlockMutex(renderer->disposeLock);
}
static void VULKAN_ReleaseSampler(
Refresh_Renderer *driverData,
Refresh_Sampler *sampler
) {
VulkanRenderer *renderer = (VulkanRenderer*) driverData;
VulkanSampler* vulkanSampler = (VulkanSampler*) sampler;
SDL_LockMutex(renderer->disposeLock);
EXPAND_ARRAY_IF_NEEDED(
renderer->samplersToDestroy,
VulkanSampler*,
renderer->samplersToDestroyCount + 1,
renderer->samplersToDestroyCapacity,
renderer->samplersToDestroyCapacity * 2
)
renderer->samplersToDestroy[renderer->samplersToDestroyCount] = vulkanSampler;
renderer->samplersToDestroyCount += 1;
SDL_UnlockMutex(renderer->disposeLock);
}
static void VULKAN_INTERNAL_ReleaseBuffer(
VulkanRenderer *renderer,
VulkanBuffer *vulkanBuffer
) {
if (vulkanBuffer->markedForDestroy) { return; }
SDL_LockMutex(renderer->disposeLock);
EXPAND_ARRAY_IF_NEEDED(
renderer->buffersToDestroy,
VulkanBuffer*,
renderer->buffersToDestroyCount + 1,
renderer->buffersToDestroyCapacity,
renderer->buffersToDestroyCapacity * 2
)
renderer->buffersToDestroy[
renderer->buffersToDestroyCount
] = vulkanBuffer;
renderer->buffersToDestroyCount += 1;
vulkanBuffer->markedForDestroy = 1;
SDL_UnlockMutex(renderer->disposeLock);
}
static void VULKAN_INTERNAL_ReleaseBufferContainer(
VulkanRenderer *renderer,
VulkanBufferContainer *bufferContainer
) {
Uint32 i;
SDL_LockMutex(renderer->disposeLock);
for (i = 0; i < bufferContainer->bufferCount; i += 1)
{
VULKAN_INTERNAL_ReleaseBuffer(renderer, bufferContainer->bufferHandles[i]->vulkanBuffer);
SDL_free(bufferContainer->bufferHandles[i]);
}
/* Containers are just client handles, so we can free immediately */
if (bufferContainer->debugName != NULL)
{
SDL_free(bufferContainer->debugName);
}
SDL_free(bufferContainer->bufferHandles);
SDL_free(bufferContainer);
SDL_UnlockMutex(renderer->disposeLock);
}
static void VULKAN_ReleaseBuffer(
Refresh_Renderer *driverData,
Refresh_Buffer *buffer
) {
VulkanRenderer *renderer = (VulkanRenderer*) driverData;
VulkanBufferContainer *vulkanBufferContainer = (VulkanBufferContainer*) buffer;
VULKAN_INTERNAL_ReleaseBufferContainer(
renderer,
vulkanBufferContainer
);
}
static void VULKAN_ReleaseTransferBuffer(
Refresh_Renderer *driverData,
Refresh_TransferBuffer *transferBuffer
) {
VulkanRenderer *renderer = (VulkanRenderer*) driverData;
VulkanBufferContainer *transferBufferContainer = (VulkanBufferContainer*) transferBuffer;
VULKAN_INTERNAL_ReleaseBufferContainer(
renderer,
transferBufferContainer
);
}
static void VULKAN_ReleaseShader(
Refresh_Renderer *driverData,
Refresh_Shader *shader
) {
VulkanRenderer *renderer = (VulkanRenderer*) driverData;
VulkanShader *vulkanShader = (VulkanShader*) shader;
SDL_LockMutex(renderer->disposeLock);
EXPAND_ARRAY_IF_NEEDED(
renderer->shadersToDestroy,
VulkanShader*,
renderer->shadersToDestroyCount + 1,
renderer->shadersToDestroyCapacity,
renderer->shadersToDestroyCapacity * 2
)
renderer->shadersToDestroy[renderer->shadersToDestroyCount] = vulkanShader;
renderer->shadersToDestroyCount += 1;
SDL_UnlockMutex(renderer->disposeLock);
}
static void VULKAN_ReleaseComputePipeline(
Refresh_Renderer *driverData,
Refresh_ComputePipeline *computePipeline
) {
VulkanRenderer *renderer = (VulkanRenderer*) driverData;
VulkanComputePipeline *vulkanComputePipeline = (VulkanComputePipeline*) computePipeline;
SDL_LockMutex(renderer->disposeLock);
EXPAND_ARRAY_IF_NEEDED(
renderer->computePipelinesToDestroy,
VulkanComputePipeline*,
renderer->computePipelinesToDestroyCount + 1,
renderer->computePipelinesToDestroyCapacity,
renderer->computePipelinesToDestroyCapacity * 2
)
renderer->computePipelinesToDestroy[renderer->computePipelinesToDestroyCount] = vulkanComputePipeline;
renderer->computePipelinesToDestroyCount += 1;
SDL_UnlockMutex(renderer->disposeLock);
}
static void VULKAN_ReleaseGraphicsPipeline(
Refresh_Renderer *driverData,
Refresh_GraphicsPipeline *graphicsPipeline
) {
VulkanRenderer *renderer = (VulkanRenderer*) driverData;
VulkanGraphicsPipeline *vulkanGraphicsPipeline = (VulkanGraphicsPipeline*) graphicsPipeline;
SDL_LockMutex(renderer->disposeLock);
EXPAND_ARRAY_IF_NEEDED(
renderer->graphicsPipelinesToDestroy,
VulkanGraphicsPipeline*,
renderer->graphicsPipelinesToDestroyCount + 1,
renderer->graphicsPipelinesToDestroyCapacity,
renderer->graphicsPipelinesToDestroyCapacity * 2
)
renderer->graphicsPipelinesToDestroy[renderer->graphicsPipelinesToDestroyCount] = vulkanGraphicsPipeline;
renderer->graphicsPipelinesToDestroyCount += 1;
SDL_UnlockMutex(renderer->disposeLock);
}
static void VULKAN_ReleaseOcclusionQuery(
Refresh_Renderer *driverData,
Refresh_OcclusionQuery *query
) {
VulkanRenderer *renderer = (VulkanRenderer*) driverData;
VulkanOcclusionQuery *vulkanQuery = (VulkanOcclusionQuery*) query;
SDL_LockMutex(renderer->queryLock);
/* Push the now-free index to the stack */
renderer->freeQueryIndexStack[vulkanQuery->index] =
renderer->freeQueryIndexStackHead;
renderer->freeQueryIndexStackHead = vulkanQuery->index;
SDL_UnlockMutex(renderer->queryLock);
}
/* Command Buffer render state */
static VkRenderPass VULKAN_INTERNAL_FetchRenderPass(
VulkanRenderer *renderer,
VulkanCommandBuffer *commandBuffer,
Refresh_ColorAttachmentInfo *colorAttachmentInfos,
Uint32 colorAttachmentCount,
Refresh_DepthStencilAttachmentInfo *depthStencilAttachmentInfo
) {
VkRenderPass renderPass;
RenderPassHash hash;
Uint32 i;
SDL_LockMutex(renderer->renderPassFetchLock);
for (i = 0; i < colorAttachmentCount; i += 1)
{
hash.colorTargetDescriptions[i].format = ((VulkanTextureContainer*) colorAttachmentInfos[i].textureSlice.texture)->activeTextureHandle->vulkanTexture->format;
hash.colorTargetDescriptions[i].clearColor = colorAttachmentInfos[i].clearColor;
hash.colorTargetDescriptions[i].loadOp = colorAttachmentInfos[i].loadOp;
hash.colorTargetDescriptions[i].storeOp = colorAttachmentInfos[i].storeOp;
}
hash.colorAttachmentSampleCount = VK_SAMPLE_COUNT_1_BIT;
if (colorAttachmentCount > 0)
{
hash.colorAttachmentSampleCount = ((VulkanTextureContainer*) colorAttachmentInfos[0].textureSlice.texture)->activeTextureHandle->vulkanTexture->sampleCount;
}
hash.colorAttachmentCount = colorAttachmentCount;
if (depthStencilAttachmentInfo == NULL)
{
hash.depthStencilTargetDescription.format = 0;
hash.depthStencilTargetDescription.loadOp = REFRESH_LOADOP_DONT_CARE;
hash.depthStencilTargetDescription.storeOp = REFRESH_STOREOP_DONT_CARE;
hash.depthStencilTargetDescription.stencilLoadOp = REFRESH_LOADOP_DONT_CARE;
hash.depthStencilTargetDescription.stencilStoreOp = REFRESH_STOREOP_DONT_CARE;
}
else
{
hash.depthStencilTargetDescription.format = ((VulkanTextureContainer*) depthStencilAttachmentInfo->textureSlice.texture)->activeTextureHandle->vulkanTexture->format;
hash.depthStencilTargetDescription.loadOp = depthStencilAttachmentInfo->loadOp;
hash.depthStencilTargetDescription.storeOp = depthStencilAttachmentInfo->storeOp;
hash.depthStencilTargetDescription.stencilLoadOp = depthStencilAttachmentInfo->stencilLoadOp;
hash.depthStencilTargetDescription.stencilStoreOp = depthStencilAttachmentInfo->stencilStoreOp;
}
renderPass = RenderPassHashArray_Fetch(
&renderer->renderPassHashArray,
&hash
);
if (renderPass != VK_NULL_HANDLE)
{
SDL_UnlockMutex(renderer->renderPassFetchLock);
return renderPass;
}
renderPass = VULKAN_INTERNAL_CreateRenderPass(
renderer,
commandBuffer,
colorAttachmentInfos,
colorAttachmentCount,
depthStencilAttachmentInfo
);
if (renderPass != VK_NULL_HANDLE)
{
RenderPassHashArray_Insert(
&renderer->renderPassHashArray,
hash,
renderPass
);
}
SDL_UnlockMutex(renderer->renderPassFetchLock);
return renderPass;
}
static VulkanFramebuffer* VULKAN_INTERNAL_FetchFramebuffer(
VulkanRenderer *renderer,
VkRenderPass renderPass,
Refresh_ColorAttachmentInfo *colorAttachmentInfos,
Uint32 colorAttachmentCount,
Refresh_DepthStencilAttachmentInfo *depthStencilAttachmentInfo,
Uint32 width,
Uint32 height
) {
VulkanFramebuffer *vulkanFramebuffer;
VkFramebufferCreateInfo framebufferInfo;
VkResult result;
VkImageView imageViewAttachments[2 * MAX_COLOR_TARGET_BINDINGS + 1];
FramebufferHash hash;
VulkanTextureSlice *textureSlice;
Uint32 attachmentCount = 0;
Uint32 i;
for (i = 0; i < MAX_COLOR_TARGET_BINDINGS; i += 1)
{
hash.colorAttachmentViews[i] = VK_NULL_HANDLE;
hash.colorMultiSampleAttachmentViews[i] = VK_NULL_HANDLE;
}
hash.colorAttachmentCount = colorAttachmentCount;
for (i = 0; i < colorAttachmentCount; i += 1)
{
textureSlice = VULKAN_INTERNAL_SDLToVulkanTextureSlice(&colorAttachmentInfos[i].textureSlice);
hash.colorAttachmentViews[i] = textureSlice->view;
if (textureSlice->msaaTexHandle != NULL)
{
hash.colorMultiSampleAttachmentViews[i] = textureSlice->msaaTexHandle->vulkanTexture->view;
}
}
if (depthStencilAttachmentInfo == NULL)
{
hash.depthStencilAttachmentView = VK_NULL_HANDLE;
}
else
{
textureSlice = VULKAN_INTERNAL_SDLToVulkanTextureSlice(&depthStencilAttachmentInfo->textureSlice);
hash.depthStencilAttachmentView = textureSlice->view;
}
hash.width = width;
hash.height = height;
SDL_LockMutex(renderer->framebufferFetchLock);
vulkanFramebuffer = FramebufferHashArray_Fetch(
&renderer->framebufferHashArray,
&hash
);
SDL_UnlockMutex(renderer->framebufferFetchLock);
if (vulkanFramebuffer != NULL)
{
return vulkanFramebuffer;
}
vulkanFramebuffer = SDL_malloc(sizeof(VulkanFramebuffer));
SDL_AtomicSet(&vulkanFramebuffer->referenceCount, 0);
/* Create a new framebuffer */
for (i = 0; i < colorAttachmentCount; i += 1)
{
textureSlice = VULKAN_INTERNAL_SDLToVulkanTextureSlice(&colorAttachmentInfos[i].textureSlice);
imageViewAttachments[attachmentCount] =
textureSlice->view;
attachmentCount += 1;
if (textureSlice->msaaTexHandle != NULL)
{
imageViewAttachments[attachmentCount] =
textureSlice->msaaTexHandle->vulkanTexture->view;
attachmentCount += 1;
}
}
if (depthStencilAttachmentInfo != NULL)
{
textureSlice = VULKAN_INTERNAL_SDLToVulkanTextureSlice(&depthStencilAttachmentInfo->textureSlice);
imageViewAttachments[attachmentCount] =
textureSlice->view;
attachmentCount += 1;
}
framebufferInfo.sType = VK_STRUCTURE_TYPE_FRAMEBUFFER_CREATE_INFO;
framebufferInfo.pNext = NULL;
framebufferInfo.flags = 0;
framebufferInfo.renderPass = renderPass;
framebufferInfo.attachmentCount = attachmentCount;
framebufferInfo.pAttachments = imageViewAttachments;
framebufferInfo.width = hash.width;
framebufferInfo.height = hash.height;
framebufferInfo.layers = 1;
result = renderer->vkCreateFramebuffer(
renderer->logicalDevice,
&framebufferInfo,
NULL,
&vulkanFramebuffer->framebuffer
);
if (result == VK_SUCCESS)
{
SDL_LockMutex(renderer->framebufferFetchLock);
FramebufferHashArray_Insert(
&renderer->framebufferHashArray,
hash,
vulkanFramebuffer
);
SDL_UnlockMutex(renderer->framebufferFetchLock);
}
else
{
LogVulkanResultAsError("vkCreateFramebuffer", result);
SDL_free(vulkanFramebuffer);
vulkanFramebuffer = NULL;
}
return vulkanFramebuffer;
}
static void VULKAN_INTERNAL_SetCurrentViewport(
VulkanCommandBuffer *commandBuffer,
Refresh_Viewport *viewport
) {
VulkanCommandBuffer *vulkanCommandBuffer = (VulkanCommandBuffer*) commandBuffer;
vulkanCommandBuffer->currentViewport.x = viewport->x;
vulkanCommandBuffer->currentViewport.width = viewport->w;
vulkanCommandBuffer->currentViewport.minDepth = viewport->minDepth;
vulkanCommandBuffer->currentViewport.maxDepth = viewport->maxDepth;
/* Viewport flip for consistency with other backends */
/* FIXME: need moltenVK hack */
vulkanCommandBuffer->currentViewport.y = viewport->y + viewport->h;
vulkanCommandBuffer->currentViewport.height = -viewport->h;
}
static void VULKAN_SetViewport(
Refresh_CommandBuffer *commandBuffer,
Refresh_Viewport *viewport
) {
VulkanCommandBuffer *vulkanCommandBuffer = (VulkanCommandBuffer*) commandBuffer;
VulkanRenderer *renderer = (VulkanRenderer*) vulkanCommandBuffer->renderer;
VULKAN_INTERNAL_SetCurrentViewport(
vulkanCommandBuffer,
viewport
);
renderer->vkCmdSetViewport(
vulkanCommandBuffer->commandBuffer,
0,
1,
&vulkanCommandBuffer->currentViewport
);
}
static void VULKAN_INTERNAL_SetCurrentScissor(
VulkanCommandBuffer *vulkanCommandBuffer,
Refresh_Rect *scissor
) {
vulkanCommandBuffer->currentScissor.offset.x = scissor->x;
vulkanCommandBuffer->currentScissor.offset.y = scissor->y;
vulkanCommandBuffer->currentScissor.extent.width = scissor->w;
vulkanCommandBuffer->currentScissor.extent.height = scissor->h;
}
static void VULKAN_SetScissor(
Refresh_CommandBuffer *commandBuffer,
Refresh_Rect *scissor
) {
VulkanCommandBuffer *vulkanCommandBuffer = (VulkanCommandBuffer*) commandBuffer;
VulkanRenderer *renderer = (VulkanRenderer*) vulkanCommandBuffer->renderer;
VULKAN_INTERNAL_SetCurrentScissor(
vulkanCommandBuffer,
scissor
);
renderer->vkCmdSetScissor(
vulkanCommandBuffer->commandBuffer,
0,
1,
&vulkanCommandBuffer->currentScissor
);
}
static void VULKAN_BindVertexSamplers(
Refresh_CommandBuffer *commandBuffer,
Uint32 firstSlot,
Refresh_TextureSamplerBinding *textureSamplerBindings,
Uint32 bindingCount
) {
VulkanCommandBuffer *vulkanCommandBuffer = (VulkanCommandBuffer*) commandBuffer;
VulkanRenderer *renderer = vulkanCommandBuffer->renderer;
VulkanTextureContainer *textureContainer;
Uint32 i, j;
for (i = 0; i < bindingCount; i += 1)
{
textureContainer = (VulkanTextureContainer*) textureSamplerBindings[i].texture;
vulkanCommandBuffer->vertexSamplerTextures[firstSlot + i] = textureContainer->activeTextureHandle->vulkanTexture;
vulkanCommandBuffer->vertexSamplers[firstSlot + i] = (VulkanSampler*) textureSamplerBindings[i].sampler;
VULKAN_INTERNAL_TrackSampler(
renderer,
vulkanCommandBuffer,
(VulkanSampler*) textureSamplerBindings[i].sampler
);
for (j = 0; j < textureContainer->activeTextureHandle->vulkanTexture->sliceCount; j += 1)
{
VULKAN_INTERNAL_TrackTextureSlice(
renderer,
vulkanCommandBuffer,
&textureContainer->activeTextureHandle->vulkanTexture->slices[j]
);
}
}
vulkanCommandBuffer->needNewVertexResourceDescriptorSet = SDL_TRUE;
}
static void VULKAN_BindVertexStorageTextures(
Refresh_CommandBuffer *commandBuffer,
Uint32 firstSlot,
Refresh_TextureSlice *storageTextureSlices,
Uint32 bindingCount
) {
VulkanCommandBuffer *vulkanCommandBuffer = (VulkanCommandBuffer*) commandBuffer;
VulkanRenderer *renderer = vulkanCommandBuffer->renderer;
VulkanTextureContainer *textureContainer;
VulkanTextureSlice *textureSlice;
Uint32 i;
for (i = 0; i < bindingCount; i += 1)
{
textureContainer = (VulkanTextureContainer*) storageTextureSlices[i].texture;
textureSlice = VULKAN_INTERNAL_FetchTextureSlice(
textureContainer->activeTextureHandle->vulkanTexture,
storageTextureSlices[i].layer,
storageTextureSlices[i].mipLevel
);
vulkanCommandBuffer->vertexStorageTextureSlices[firstSlot + i] = textureSlice;
VULKAN_INTERNAL_TrackTextureSlice(
renderer,
vulkanCommandBuffer,
textureSlice
);
}
vulkanCommandBuffer->needNewVertexResourceDescriptorSet = SDL_TRUE;
}
static void VULKAN_BindVertexStorageBuffers(
Refresh_CommandBuffer *commandBuffer,
Uint32 firstSlot,
Refresh_Buffer **storageBuffers,
Uint32 bindingCount
) {
VulkanCommandBuffer *vulkanCommandBuffer = (VulkanCommandBuffer*) commandBuffer;
VulkanRenderer *renderer = vulkanCommandBuffer->renderer;
VulkanBufferContainer *bufferContainer;
Uint32 i;
for (i = 0; i < bindingCount; i += 1)
{
bufferContainer = (VulkanBufferContainer*) storageBuffers[i];
vulkanCommandBuffer->vertexStorageBuffers[firstSlot + i] = bufferContainer->activeBufferHandle->vulkanBuffer;
VULKAN_INTERNAL_TrackBuffer(
renderer,
vulkanCommandBuffer,
bufferContainer->activeBufferHandle->vulkanBuffer
);
}
vulkanCommandBuffer->needNewVertexResourceDescriptorSet = SDL_TRUE;
}
static void VULKAN_BindFragmentSamplers(
Refresh_CommandBuffer *commandBuffer,
Uint32 firstSlot,
Refresh_TextureSamplerBinding *textureSamplerBindings,
Uint32 bindingCount
) {
VulkanCommandBuffer *vulkanCommandBuffer = (VulkanCommandBuffer*) commandBuffer;
VulkanRenderer *renderer = vulkanCommandBuffer->renderer;
VulkanTextureContainer *textureContainer;
Uint32 i, j;
for (i = 0; i < bindingCount; i += 1)
{
textureContainer = (VulkanTextureContainer*) textureSamplerBindings[i].texture;
vulkanCommandBuffer->fragmentSamplerTextures[firstSlot + i] = textureContainer->activeTextureHandle->vulkanTexture;
vulkanCommandBuffer->fragmentSamplers[firstSlot + i] = (VulkanSampler*) textureSamplerBindings[i].sampler;
VULKAN_INTERNAL_TrackSampler(
renderer,
vulkanCommandBuffer,
(VulkanSampler*) textureSamplerBindings[i].sampler
);
for (j = 0; j < textureContainer->activeTextureHandle->vulkanTexture->sliceCount; j += 1)
{
VULKAN_INTERNAL_TrackTextureSlice(
renderer,
vulkanCommandBuffer,
&textureContainer->activeTextureHandle->vulkanTexture->slices[j]
);
}
}
vulkanCommandBuffer->needNewFragmentResourceDescriptorSet = SDL_TRUE;
}
static void VULKAN_BindFragmentStorageTextures(
Refresh_CommandBuffer *commandBuffer,
Uint32 firstSlot,
Refresh_TextureSlice *storageTextureSlices,
Uint32 bindingCount
) {
VulkanCommandBuffer *vulkanCommandBuffer = (VulkanCommandBuffer*) commandBuffer;
VulkanRenderer *renderer = vulkanCommandBuffer->renderer;
VulkanTextureContainer *textureContainer;
VulkanTextureSlice *textureSlice;
Uint32 i;
for (i = 0; i < bindingCount; i += 1)
{
textureContainer = (VulkanTextureContainer*) storageTextureSlices[i].texture;
textureSlice = VULKAN_INTERNAL_FetchTextureSlice(
textureContainer->activeTextureHandle->vulkanTexture,
storageTextureSlices[i].layer,
storageTextureSlices[i].mipLevel
);
vulkanCommandBuffer->fragmentStorageTextureSlices[firstSlot + i] = textureSlice;
VULKAN_INTERNAL_TrackTextureSlice(
renderer,
vulkanCommandBuffer,
textureSlice
);
}
vulkanCommandBuffer->needNewFragmentResourceDescriptorSet = SDL_TRUE;
}
static void VULKAN_BindFragmentStorageBuffers(
Refresh_CommandBuffer *commandBuffer,
Uint32 firstSlot,
Refresh_Buffer **storageBuffers,
Uint32 bindingCount
) {
VulkanCommandBuffer *vulkanCommandBuffer = (VulkanCommandBuffer*) commandBuffer;
VulkanRenderer *renderer = vulkanCommandBuffer->renderer;
VulkanBufferContainer *bufferContainer;
Uint32 i;
for (i = 0; i < bindingCount; i += 1)
{
bufferContainer = (VulkanBufferContainer*) storageBuffers[i];
vulkanCommandBuffer->fragmentStorageBuffers[firstSlot + i] = bufferContainer->activeBufferHandle->vulkanBuffer;
VULKAN_INTERNAL_TrackBuffer(
renderer,
vulkanCommandBuffer,
bufferContainer->activeBufferHandle->vulkanBuffer
);
}
vulkanCommandBuffer->needNewFragmentResourceDescriptorSet = SDL_TRUE;
}
static void VULKAN_INTERNAL_PushUniformData(
VulkanRenderer *renderer,
VulkanCommandBuffer *commandBuffer,
VulkanUniformBuffer *uniformBuffer,
Refresh_ShaderStage shaderStage,
void *data,
Uint32 dataLengthInBytes
) {
Uint32 blockSize =
VULKAN_INTERNAL_NextHighestAlignment32(
dataLengthInBytes,
renderer->minUBOAlignment
);
/* If there is no more room, cycle the uniform buffer */
if (uniformBuffer->offset + blockSize + MAX_UBO_SECTION_SIZE >= uniformBuffer->bufferContainer->activeBufferHandle->vulkanBuffer->size)
{
VULKAN_INTERNAL_CycleActiveBuffer(
renderer,
uniformBuffer->bufferContainer
);
uniformBuffer->drawOffset = 0;
uniformBuffer->offset = 0;
VULKAN_INTERNAL_TrackBuffer(
renderer,
commandBuffer,
uniformBuffer->bufferContainer->activeBufferHandle->vulkanBuffer
);
if (shaderStage == REFRESH_SHADERSTAGE_VERTEX)
{
commandBuffer->needNewVertexUniformDescriptorSet = SDL_TRUE;
}
else if (shaderStage == REFRESH_SHADERSTAGE_FRAGMENT)
{
commandBuffer->needNewFragmentUniformDescriptorSet = SDL_TRUE;
}
else if (shaderStage == REFRESH_SHADERSTAGE_COMPUTE)
{
commandBuffer->needNewComputeUniformDescriptorSet = SDL_TRUE;
}
else
{
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Unrecognized shader stage!");
return;
}
}
uniformBuffer->drawOffset = uniformBuffer->offset;
Uint8 *dst =
uniformBuffer->bufferContainer->activeBufferHandle->vulkanBuffer->usedRegion->allocation->mapPointer +
uniformBuffer->bufferContainer->activeBufferHandle->vulkanBuffer->usedRegion->resourceOffset +
uniformBuffer->offset;
SDL_memcpy(
dst,
data,
dataLengthInBytes
);
uniformBuffer->offset += blockSize;
if (shaderStage == REFRESH_SHADERSTAGE_VERTEX)
{
commandBuffer->needNewVertexUniformOffsets = SDL_TRUE;
}
else if (shaderStage == REFRESH_SHADERSTAGE_FRAGMENT)
{
commandBuffer->needNewFragmentUniformOffsets = SDL_TRUE;
}
else if (shaderStage == REFRESH_SHADERSTAGE_COMPUTE)
{
commandBuffer->needNewComputeUniformOffsets = SDL_TRUE;
}
else
{
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Unrecognized shader stage!");
return;
}
}
static void VULKAN_BeginRenderPass(
Refresh_CommandBuffer *commandBuffer,
Refresh_ColorAttachmentInfo *colorAttachmentInfos,
Uint32 colorAttachmentCount,
Refresh_DepthStencilAttachmentInfo *depthStencilAttachmentInfo
) {
VulkanCommandBuffer *vulkanCommandBuffer = (VulkanCommandBuffer*) commandBuffer;
VulkanRenderer *renderer = (VulkanRenderer*) vulkanCommandBuffer->renderer;
VkRenderPass renderPass;
VulkanFramebuffer *framebuffer;
VulkanTextureContainer *textureContainer;
VulkanTextureSlice *textureSlice;
Uint32 w, h;
VkClearValue *clearValues;
Uint32 clearCount = colorAttachmentCount;
Uint32 multisampleAttachmentCount = 0;
Uint32 totalColorAttachmentCount = 0;
Uint32 i;
Refresh_Viewport defaultViewport;
Refresh_Rect defaultScissor;
Uint32 framebufferWidth = UINT32_MAX;
Uint32 framebufferHeight = UINT32_MAX;
for (i = 0; i < colorAttachmentCount; i += 1)
{
textureContainer = (VulkanTextureContainer*) colorAttachmentInfos[i].textureSlice.texture;
w = textureContainer->activeTextureHandle->vulkanTexture->dimensions.width >> colorAttachmentInfos[i].textureSlice.mipLevel;
h = textureContainer->activeTextureHandle->vulkanTexture->dimensions.height >> colorAttachmentInfos[i].textureSlice.mipLevel;
/* The framebuffer cannot be larger than the smallest attachment. */
if (w < framebufferWidth)
{
framebufferWidth = w;
}
if (h < framebufferHeight)
{
framebufferHeight = h;
}
if (!textureContainer->activeTextureHandle->vulkanTexture->isRenderTarget)
{
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Color attachment texture was not designated as a target!");
return;
}
}
if (depthStencilAttachmentInfo != NULL)
{
textureContainer = (VulkanTextureContainer*) depthStencilAttachmentInfo->textureSlice.texture;
w = textureContainer->activeTextureHandle->vulkanTexture->dimensions.width >> depthStencilAttachmentInfo->textureSlice.mipLevel;
h = textureContainer->activeTextureHandle->vulkanTexture->dimensions.height >> depthStencilAttachmentInfo->textureSlice.mipLevel;
/* The framebuffer cannot be larger than the smallest attachment. */
if (w < framebufferWidth)
{
framebufferWidth = w;
}
if (h < framebufferHeight)
{
framebufferHeight = h;
}
if (!textureContainer->activeTextureHandle->vulkanTexture->isRenderTarget)
{
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Depth stencil attachment texture was not designated as a target!");
return;
}
}
/* Layout transitions */
for (i = 0; i < colorAttachmentCount; i += 1)
{
SDL_bool cycle;
if (colorAttachmentInfos[i].loadOp)
{
cycle = SDL_FALSE;
}
else
{
cycle = colorAttachmentInfos[i].cycle;
}
textureContainer = (VulkanTextureContainer*) colorAttachmentInfos[i].textureSlice.texture;
textureSlice = VULKAN_INTERNAL_PrepareTextureSliceForWrite(
renderer,
vulkanCommandBuffer,
textureContainer,
colorAttachmentInfos[i].textureSlice.layer,
colorAttachmentInfos[i].textureSlice.mipLevel,
cycle,
VULKAN_TEXTURE_USAGE_MODE_COLOR_ATTACHMENT
);
if (textureSlice->msaaTexHandle != NULL)
{
/* Transition the multisample attachment */
VULKAN_INTERNAL_TextureTransitionFromDefaultUsage(
renderer,
vulkanCommandBuffer,
VULKAN_TEXTURE_USAGE_MODE_COLOR_ATTACHMENT,
&textureSlice->msaaTexHandle->vulkanTexture->slices[0]
);
clearCount += 1;
multisampleAttachmentCount += 1;
}
vulkanCommandBuffer->colorAttachmentSlices[i] = textureSlice;
VULKAN_INTERNAL_TrackTextureSlice(renderer, vulkanCommandBuffer, textureSlice);
/* TODO: do we need to track the msaa texture? or is it implicitly only used when the regular texture is used? */
}
vulkanCommandBuffer->colorAttachmentSliceCount = colorAttachmentCount;
if (depthStencilAttachmentInfo != NULL)
{
SDL_bool cycle;
if (
depthStencilAttachmentInfo->loadOp == REFRESH_LOADOP_LOAD ||
depthStencilAttachmentInfo->stencilLoadOp == REFRESH_LOADOP_LOAD
) {
cycle = SDL_FALSE;
}
else
{
cycle = depthStencilAttachmentInfo->cycle;
}
textureContainer = (VulkanTextureContainer*) depthStencilAttachmentInfo->textureSlice.texture;
textureSlice = VULKAN_INTERNAL_PrepareTextureSliceForWrite(
renderer,
vulkanCommandBuffer,
textureContainer,
depthStencilAttachmentInfo->textureSlice.layer,
depthStencilAttachmentInfo->textureSlice.mipLevel,
cycle,
VULKAN_TEXTURE_USAGE_MODE_DEPTH_STENCIL_ATTACHMENT
);
clearCount += 1;
vulkanCommandBuffer->depthStencilAttachmentSlice = textureSlice;
VULKAN_INTERNAL_TrackTextureSlice(renderer, vulkanCommandBuffer, textureSlice);
}
/* Fetch required render objects */
renderPass = VULKAN_INTERNAL_FetchRenderPass(
renderer,
vulkanCommandBuffer,
colorAttachmentInfos,
colorAttachmentCount,
depthStencilAttachmentInfo
);
framebuffer = VULKAN_INTERNAL_FetchFramebuffer(
renderer,
renderPass,
colorAttachmentInfos,
colorAttachmentCount,
depthStencilAttachmentInfo,
framebufferWidth,
framebufferHeight
);
VULKAN_INTERNAL_TrackFramebuffer(renderer, vulkanCommandBuffer, framebuffer);
/* Set clear values */
clearValues = SDL_stack_alloc(VkClearValue, clearCount);
totalColorAttachmentCount = colorAttachmentCount + multisampleAttachmentCount;
for (i = 0; i < totalColorAttachmentCount; i += 1)
{
clearValues[i].color.float32[0] = colorAttachmentInfos[i].clearColor.r;
clearValues[i].color.float32[1] = colorAttachmentInfos[i].clearColor.g;
clearValues[i].color.float32[2] = colorAttachmentInfos[i].clearColor.b;
clearValues[i].color.float32[3] = colorAttachmentInfos[i].clearColor.a;
textureSlice = VULKAN_INTERNAL_SDLToVulkanTextureSlice(&colorAttachmentInfos[i].textureSlice);
if (textureSlice->parent->sampleCount > VK_SAMPLE_COUNT_1_BIT)
{
clearValues[i+1].color.float32[0] = colorAttachmentInfos[i].clearColor.r;
clearValues[i+1].color.float32[1] = colorAttachmentInfos[i].clearColor.g;
clearValues[i+1].color.float32[2] = colorAttachmentInfos[i].clearColor.b;
clearValues[i+1].color.float32[3] = colorAttachmentInfos[i].clearColor.a;
i += 1;
}
}
if (depthStencilAttachmentInfo != NULL)
{
clearValues[totalColorAttachmentCount].depthStencil.depth =
depthStencilAttachmentInfo->depthStencilClearValue.depth;
clearValues[totalColorAttachmentCount].depthStencil.stencil =
depthStencilAttachmentInfo->depthStencilClearValue.stencil;
}
VkRenderPassBeginInfo renderPassBeginInfo;
renderPassBeginInfo.sType = VK_STRUCTURE_TYPE_RENDER_PASS_BEGIN_INFO;
renderPassBeginInfo.pNext = NULL;
renderPassBeginInfo.renderPass = renderPass;
renderPassBeginInfo.framebuffer = framebuffer->framebuffer;
renderPassBeginInfo.pClearValues = clearValues;
renderPassBeginInfo.clearValueCount = clearCount;
renderPassBeginInfo.renderArea.extent.width = framebufferWidth;
renderPassBeginInfo.renderArea.extent.height = framebufferHeight;
renderPassBeginInfo.renderArea.offset.x = 0;
renderPassBeginInfo.renderArea.offset.y = 0;
renderer->vkCmdBeginRenderPass(
vulkanCommandBuffer->commandBuffer,
&renderPassBeginInfo,
VK_SUBPASS_CONTENTS_INLINE
);
SDL_stack_free(clearValues);
/* Set sensible default viewport state */
defaultViewport.x = 0;
defaultViewport.y = 0;
defaultViewport.w = framebufferWidth;
defaultViewport.h = framebufferHeight;
defaultViewport.minDepth = 0;
defaultViewport.maxDepth = 1;
VULKAN_INTERNAL_SetCurrentViewport(
vulkanCommandBuffer,
&defaultViewport
);
defaultScissor.x = 0;
defaultScissor.y = 0;
defaultScissor.w = framebufferWidth;
defaultScissor.h = framebufferHeight;
VULKAN_INTERNAL_SetCurrentScissor(
vulkanCommandBuffer,
&defaultScissor
);
}
static void VULKAN_BindGraphicsPipeline(
Refresh_CommandBuffer *commandBuffer,
Refresh_GraphicsPipeline *graphicsPipeline
) {
VulkanCommandBuffer *vulkanCommandBuffer = (VulkanCommandBuffer*) commandBuffer;
VulkanRenderer *renderer = (VulkanRenderer*) vulkanCommandBuffer->renderer;
VulkanGraphicsPipeline* pipeline = (VulkanGraphicsPipeline*) graphicsPipeline;
Uint32 i;
renderer->vkCmdBindPipeline(
vulkanCommandBuffer->commandBuffer,
VK_PIPELINE_BIND_POINT_GRAPHICS,
pipeline->pipeline
);
vulkanCommandBuffer->currentGraphicsPipeline = pipeline;
VULKAN_INTERNAL_TrackGraphicsPipeline(renderer, vulkanCommandBuffer, pipeline);
renderer->vkCmdSetViewport(
vulkanCommandBuffer->commandBuffer,
0,
1,
&vulkanCommandBuffer->currentViewport
);
renderer->vkCmdSetScissor(
vulkanCommandBuffer->commandBuffer,
0,
1,
&vulkanCommandBuffer->currentScissor
);
for (i = vulkanCommandBuffer->initializedVertexUniformBufferCount; i < pipeline->resourceLayout.vertexUniformBufferCount; i += 1)
{
vulkanCommandBuffer->vertexUniformBuffers[i] = VULKAN_INTERNAL_CreateUniformBuffer(renderer, UNIFORM_BUFFER_SIZE);
vulkanCommandBuffer->initializedVertexUniformBufferCount += 1;
}
for (i = vulkanCommandBuffer->initializedFragmentUniformBufferCount; i < pipeline->resourceLayout.fragmentUniformBufferCount; i += 1)
{
vulkanCommandBuffer->fragmentUniformBuffers[i] = VULKAN_INTERNAL_CreateUniformBuffer(renderer, UNIFORM_BUFFER_SIZE);
vulkanCommandBuffer->initializedFragmentUniformBufferCount += 1;
}
for (i = 0; i < pipeline->resourceLayout.vertexUniformBufferCount; i += 1)
{
VULKAN_INTERNAL_TrackBuffer(
renderer,
vulkanCommandBuffer,
vulkanCommandBuffer->vertexUniformBuffers[i]->bufferContainer->activeBufferHandle->vulkanBuffer
);
}
for (i = 0; i < pipeline->resourceLayout.fragmentUniformBufferCount; i += 1)
{
VULKAN_INTERNAL_TrackBuffer(
renderer,
vulkanCommandBuffer,
vulkanCommandBuffer->fragmentUniformBuffers[i]->bufferContainer->activeBufferHandle->vulkanBuffer
);
}
vulkanCommandBuffer->needNewVertexResourceDescriptorSet = SDL_TRUE;
vulkanCommandBuffer->needNewFragmentResourceDescriptorSet = SDL_TRUE;
vulkanCommandBuffer->needNewVertexUniformDescriptorSet = SDL_TRUE;
vulkanCommandBuffer->needNewFragmentUniformDescriptorSet = SDL_TRUE;
vulkanCommandBuffer->needNewVertexUniformOffsets = SDL_TRUE;
vulkanCommandBuffer->needNewFragmentUniformOffsets = SDL_TRUE;
}
static void VULKAN_BindVertexBuffers(
Refresh_CommandBuffer *commandBuffer,
Uint32 firstBinding,
Refresh_BufferBinding *pBindings,
Uint32 bindingCount
) {
VulkanCommandBuffer *vulkanCommandBuffer = (VulkanCommandBuffer*) commandBuffer;
VulkanRenderer *renderer = (VulkanRenderer*) vulkanCommandBuffer->renderer;
VulkanBuffer *currentVulkanBuffer;
VkBuffer *buffers = SDL_stack_alloc(VkBuffer, bindingCount);
VkDeviceSize *offsets = SDL_stack_alloc(VkDeviceSize, bindingCount);
Uint32 i;
for (i = 0; i < bindingCount; i += 1)
{
currentVulkanBuffer = ((VulkanBufferContainer*) pBindings[i].buffer)->activeBufferHandle->vulkanBuffer;
buffers[i] = currentVulkanBuffer->buffer;
offsets[i] = (VkDeviceSize) pBindings[i].offset;
VULKAN_INTERNAL_TrackBuffer(renderer, vulkanCommandBuffer, currentVulkanBuffer);
}
renderer->vkCmdBindVertexBuffers(
vulkanCommandBuffer->commandBuffer,
firstBinding,
bindingCount,
buffers,
offsets
);
SDL_stack_free(buffers);
SDL_stack_free(offsets);
}
static void VULKAN_BindIndexBuffer(
Refresh_CommandBuffer *commandBuffer,
Refresh_BufferBinding *pBinding,
Refresh_IndexElementSize indexElementSize
) {
VulkanCommandBuffer *vulkanCommandBuffer = (VulkanCommandBuffer*) commandBuffer;
VulkanRenderer *renderer = (VulkanRenderer*) vulkanCommandBuffer->renderer;
VulkanBuffer* vulkanBuffer = ((VulkanBufferContainer*) pBinding->buffer)->activeBufferHandle->vulkanBuffer;
VULKAN_INTERNAL_TrackBuffer(renderer, vulkanCommandBuffer, vulkanBuffer);
renderer->vkCmdBindIndexBuffer(
vulkanCommandBuffer->commandBuffer,
vulkanBuffer->buffer,
(VkDeviceSize) pBinding->offset,
SDLToVK_IndexType[indexElementSize]
);
}
static void VULKAN_PushVertexUniformData(
Refresh_CommandBuffer *commandBuffer,
Uint32 slotIndex,
void *data,
Uint32 dataLengthInBytes
) {
VulkanCommandBuffer *vulkanCommandBuffer = (VulkanCommandBuffer*) commandBuffer;
if (slotIndex >= vulkanCommandBuffer->currentGraphicsPipeline->resourceLayout.vertexUniformBufferCount)
{
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "No vertex uniforms exist on slot %i for this pipeline", slotIndex);
return;
}
VULKAN_INTERNAL_PushUniformData(
vulkanCommandBuffer->renderer,
vulkanCommandBuffer,
vulkanCommandBuffer->vertexUniformBuffers[slotIndex],
REFRESH_SHADERSTAGE_VERTEX,
data,
dataLengthInBytes
);
}
static void VULKAN_PushFragmentUniformData(
Refresh_CommandBuffer *commandBuffer,
Uint32 slotIndex,
void *data,
Uint32 dataLengthInBytes
) {
VulkanCommandBuffer *vulkanCommandBuffer = (VulkanCommandBuffer*) commandBuffer;
if (slotIndex >= vulkanCommandBuffer->currentGraphicsPipeline->resourceLayout.fragmentUniformBufferCount)
{
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "No fragment uniforms exist on slot %i for this pipeline", slotIndex);
return;
}
VULKAN_INTERNAL_PushUniformData(
vulkanCommandBuffer->renderer,
vulkanCommandBuffer,
vulkanCommandBuffer->fragmentUniformBuffers[slotIndex],
REFRESH_SHADERSTAGE_FRAGMENT,
data,
dataLengthInBytes
);
}
static void VULKAN_EndRenderPass(
Refresh_CommandBuffer *commandBuffer
) {
VulkanCommandBuffer *vulkanCommandBuffer = (VulkanCommandBuffer*) commandBuffer;
VulkanRenderer *renderer = (VulkanRenderer*) vulkanCommandBuffer->renderer;
Uint32 i;
renderer->vkCmdEndRenderPass(
vulkanCommandBuffer->commandBuffer
);
for (i = 0; i < vulkanCommandBuffer->colorAttachmentSliceCount; i += 1)
{
VULKAN_INTERNAL_TextureTransitionToDefaultUsage(
renderer,
vulkanCommandBuffer,
VULKAN_TEXTURE_USAGE_MODE_COLOR_ATTACHMENT,
vulkanCommandBuffer->colorAttachmentSlices[i]
);
}
vulkanCommandBuffer->colorAttachmentSliceCount = 0;
if (vulkanCommandBuffer->depthStencilAttachmentSlice != NULL)
{
VULKAN_INTERNAL_TextureTransitionToDefaultUsage(
renderer,
vulkanCommandBuffer,
VULKAN_TEXTURE_USAGE_MODE_DEPTH_STENCIL_ATTACHMENT,
vulkanCommandBuffer->depthStencilAttachmentSlice
);
vulkanCommandBuffer->depthStencilAttachmentSlice = NULL;
}
vulkanCommandBuffer->currentGraphicsPipeline = NULL;
vulkanCommandBuffer->vertexResourceDescriptorSet = VK_NULL_HANDLE;
vulkanCommandBuffer->vertexUniformDescriptorSet = VK_NULL_HANDLE;
vulkanCommandBuffer->fragmentResourceDescriptorSet = VK_NULL_HANDLE;
vulkanCommandBuffer->fragmentUniformDescriptorSet = VK_NULL_HANDLE;
}
static void VULKAN_BeginComputePass(
Refresh_CommandBuffer *commandBuffer,
Refresh_StorageTextureReadWriteBinding *storageTextureBindings,
Uint32 storageTextureBindingCount,
Refresh_StorageBufferReadWriteBinding *storageBufferBindings,
Uint32 storageBufferBindingCount
) {
VulkanCommandBuffer *vulkanCommandBuffer = (VulkanCommandBuffer*) commandBuffer;
VulkanRenderer *renderer = vulkanCommandBuffer->renderer;
VulkanTextureContainer *textureContainer;
VulkanTextureSlice *textureSlice;
VulkanBufferContainer *bufferContainer;
VulkanBuffer *buffer;
Uint32 i;
for (i = 0; i < storageTextureBindingCount; i += 1)
{
textureContainer = (VulkanTextureContainer*) storageTextureBindings[i].textureSlice.texture;
textureSlice = VULKAN_INTERNAL_PrepareTextureSliceForWrite(
renderer,
vulkanCommandBuffer,
textureContainer,
storageTextureBindings[i].textureSlice.layer,
storageTextureBindings[i].textureSlice.mipLevel,
storageTextureBindings[i].cycle,
VULKAN_TEXTURE_USAGE_MODE_COMPUTE_STORAGE_READ_WRITE
);
vulkanCommandBuffer->readWriteComputeStorageTextureSlices[i] = textureSlice;
VULKAN_INTERNAL_TrackTextureSlice(
renderer,
vulkanCommandBuffer,
textureSlice
);
}
for (i = 0; i < storageBufferBindingCount; i += 1)
{
bufferContainer = (VulkanBufferContainer*) storageBufferBindings[i].buffer;
buffer = VULKAN_INTERNAL_PrepareBufferForWrite(
renderer,
vulkanCommandBuffer,
bufferContainer,
storageBufferBindings[i].cycle,
VULKAN_BUFFER_USAGE_MODE_COMPUTE_STORAGE_READ
);
vulkanCommandBuffer->readWriteComputeStorageBuffers[i] = buffer;
VULKAN_INTERNAL_TrackBuffer(
renderer,
vulkanCommandBuffer,
buffer
);
}
}
static void VULKAN_BindComputePipeline(
Refresh_CommandBuffer *commandBuffer,
Refresh_ComputePipeline *computePipeline
) {
VulkanCommandBuffer *vulkanCommandBuffer = (VulkanCommandBuffer*) commandBuffer;
VulkanRenderer *renderer = (VulkanRenderer*) vulkanCommandBuffer->renderer;
VulkanComputePipeline *vulkanComputePipeline = (VulkanComputePipeline*) computePipeline;
Uint32 i;
renderer->vkCmdBindPipeline(
vulkanCommandBuffer->commandBuffer,
VK_PIPELINE_BIND_POINT_COMPUTE,
vulkanComputePipeline->pipeline
);
vulkanCommandBuffer->currentComputePipeline = vulkanComputePipeline;
VULKAN_INTERNAL_TrackComputePipeline(renderer, vulkanCommandBuffer, vulkanComputePipeline);
for (i = vulkanCommandBuffer->initializedComputeUniformBufferCount; i < vulkanComputePipeline->resourceLayout.uniformBufferCount; i += 1)
{
vulkanCommandBuffer->computeUniformBuffers[i] = VULKAN_INTERNAL_CreateUniformBuffer(renderer, UNIFORM_BUFFER_SIZE);
vulkanCommandBuffer->initializedComputeUniformBufferCount += 1;
}
for (i = 0; i < vulkanComputePipeline->resourceLayout.uniformBufferCount; i += 1)
{
VULKAN_INTERNAL_TrackBuffer(
renderer,
vulkanCommandBuffer,
vulkanCommandBuffer->computeUniformBuffers[i]->bufferContainer->activeBufferHandle->vulkanBuffer
);
}
vulkanCommandBuffer->needNewComputeReadWriteDescriptorSet = SDL_TRUE;
vulkanCommandBuffer->needNewComputeReadOnlyDescriptorSet = SDL_TRUE;
vulkanCommandBuffer->needNewComputeUniformDescriptorSet = SDL_TRUE;
vulkanCommandBuffer->needNewComputeUniformOffsets = SDL_TRUE;
}
static void VULKAN_BindComputeStorageTextures(
Refresh_CommandBuffer *commandBuffer,
Uint32 firstSlot,
Refresh_TextureSlice *storageTextureSlices,
Uint32 bindingCount
) {
VulkanCommandBuffer *vulkanCommandBuffer = (VulkanCommandBuffer*) commandBuffer;
VulkanRenderer *renderer = vulkanCommandBuffer->renderer;
VulkanTextureContainer *textureContainer;
VulkanTextureSlice *textureSlice;
Uint32 i;
for (i = 0; i < bindingCount; i += 1)
{
if (vulkanCommandBuffer->readOnlyComputeStorageTextureSlices[firstSlot + i] != NULL)
{
VULKAN_INTERNAL_TextureTransitionToDefaultUsage(
renderer,
vulkanCommandBuffer,
VULKAN_TEXTURE_USAGE_MODE_COMPUTE_STORAGE_READ,
vulkanCommandBuffer->readOnlyComputeStorageTextureSlices[firstSlot + i]
);
}
textureContainer = (VulkanTextureContainer*) storageTextureSlices[i].texture;
textureSlice = VULKAN_INTERNAL_FetchTextureSlice(
textureContainer->activeTextureHandle->vulkanTexture,
storageTextureSlices[i].layer,
storageTextureSlices[i].mipLevel
);
vulkanCommandBuffer->readOnlyComputeStorageTextureSlices[firstSlot + i] = textureSlice;
VULKAN_INTERNAL_TextureTransitionFromDefaultUsage(
renderer,
vulkanCommandBuffer,
VULKAN_TEXTURE_USAGE_MODE_COMPUTE_STORAGE_READ,
textureSlice
);
VULKAN_INTERNAL_TrackTextureSlice(
renderer,
vulkanCommandBuffer,
textureSlice
);
}
vulkanCommandBuffer->needNewComputeReadOnlyDescriptorSet = SDL_TRUE;
}
static void VULKAN_BindComputeStorageBuffers(
Refresh_CommandBuffer *commandBuffer,
Uint32 firstSlot,
Refresh_Buffer **storageBuffers,
Uint32 bindingCount
) {
VulkanCommandBuffer *vulkanCommandBuffer = (VulkanCommandBuffer*) commandBuffer;
VulkanRenderer *renderer = vulkanCommandBuffer->renderer;
VulkanBufferContainer *bufferContainer;
Uint32 i;
for (i = 0; i < bindingCount; i += 1)
{
if (vulkanCommandBuffer->readOnlyComputeStorageBuffers[firstSlot + i] != NULL)
{
VULKAN_INTERNAL_BufferTransitionToDefaultUsage(
renderer,
vulkanCommandBuffer,
VULKAN_BUFFER_USAGE_MODE_COMPUTE_STORAGE_READ,
vulkanCommandBuffer->readOnlyComputeStorageBuffers[firstSlot + i]
);
}
bufferContainer = (VulkanBufferContainer*) storageBuffers[i];
vulkanCommandBuffer->readOnlyComputeStorageBuffers[firstSlot + i] = bufferContainer->activeBufferHandle->vulkanBuffer;
VULKAN_INTERNAL_BufferTransitionFromDefaultUsage(
renderer,
vulkanCommandBuffer,
VULKAN_BUFFER_USAGE_MODE_COMPUTE_STORAGE_READ,
bufferContainer->activeBufferHandle->vulkanBuffer
);
VULKAN_INTERNAL_TrackBuffer(
renderer,
vulkanCommandBuffer,
bufferContainer->activeBufferHandle->vulkanBuffer
);
}
vulkanCommandBuffer->needNewComputeReadOnlyDescriptorSet = SDL_TRUE;
}
static void VULKAN_PushComputeUniformData(
Refresh_CommandBuffer *commandBuffer,
Uint32 slotIndex,
void *data,
Uint32 dataLengthInBytes
) {
VulkanCommandBuffer *vulkanCommandBuffer = (VulkanCommandBuffer*) commandBuffer;
if (slotIndex >= vulkanCommandBuffer->currentComputePipeline->resourceLayout.uniformBufferCount)
{
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "No compute uniforms exist on slot %i for this pipeline", slotIndex);
return;
}
VULKAN_INTERNAL_PushUniformData(
vulkanCommandBuffer->renderer,
vulkanCommandBuffer,
vulkanCommandBuffer->computeUniformBuffers[slotIndex],
REFRESH_SHADERSTAGE_COMPUTE,
data,
dataLengthInBytes
);
}
static void VULKAN_INTERNAL_BindComputeDescriptorSets(
VulkanRenderer *renderer,
VulkanCommandBuffer *commandBuffer
) {
VulkanComputePipelineResourceLayout *resourceLayout;
VkWriteDescriptorSet *writeDescriptorSets;
VkWriteDescriptorSet *currentWriteDescriptorSet;
DescriptorSetPool *descriptorSetPool;
VkDescriptorBufferInfo bufferInfos[MAX_STORAGE_BUFFERS_PER_STAGE];
VkDescriptorImageInfo imageInfos[MAX_STORAGE_TEXTURES_PER_STAGE];
Uint32 dynamicOffsets[MAX_UNIFORM_BUFFERS_PER_STAGE];
Uint32 bufferInfoCount = 0;
Uint32 imageInfoCount = 0;
Uint32 i;
resourceLayout = &commandBuffer->currentComputePipeline->resourceLayout;
if (commandBuffer->needNewComputeReadOnlyDescriptorSet)
{
descriptorSetPool = &resourceLayout->descriptorSetPools[0];
commandBuffer->computeReadOnlyDescriptorSet = VULKAN_INTERNAL_FetchDescriptorSet(
renderer,
commandBuffer,
descriptorSetPool
);
writeDescriptorSets = SDL_stack_alloc(
VkWriteDescriptorSet,
resourceLayout->readOnlyStorageTextureCount +
resourceLayout->readOnlyStorageBufferCount
);
for (i = 0; i < resourceLayout->readOnlyStorageTextureCount; i += 1)
{
currentWriteDescriptorSet = &writeDescriptorSets[i];
currentWriteDescriptorSet->sType = VK_STRUCTURE_TYPE_WRITE_DESCRIPTOR_SET;
currentWriteDescriptorSet->pNext = NULL;
currentWriteDescriptorSet->descriptorCount = 1;
currentWriteDescriptorSet->descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_IMAGE;
currentWriteDescriptorSet->dstArrayElement = 0;
currentWriteDescriptorSet->dstBinding = i;
currentWriteDescriptorSet->dstSet = commandBuffer->computeReadOnlyDescriptorSet;
currentWriteDescriptorSet->pTexelBufferView = NULL;
currentWriteDescriptorSet->pBufferInfo = NULL;
imageInfos[imageInfoCount].sampler = VK_NULL_HANDLE;
imageInfos[imageInfoCount].imageView = commandBuffer->readOnlyComputeStorageTextureSlices[i]->view;
imageInfos[imageInfoCount].imageLayout = VK_IMAGE_LAYOUT_GENERAL;
currentWriteDescriptorSet->pImageInfo = &imageInfos[imageInfoCount];
imageInfoCount += 1;
}
for (i = 0; i < resourceLayout->readOnlyStorageBufferCount; i += 1)
{
currentWriteDescriptorSet = &writeDescriptorSets[resourceLayout->readOnlyStorageTextureCount + i];
currentWriteDescriptorSet->sType = VK_STRUCTURE_TYPE_WRITE_DESCRIPTOR_SET;
currentWriteDescriptorSet->pNext = NULL;
currentWriteDescriptorSet->descriptorCount = 1;
currentWriteDescriptorSet->descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_BUFFER;
currentWriteDescriptorSet->dstArrayElement = 0;
currentWriteDescriptorSet->dstBinding = resourceLayout->readOnlyStorageTextureCount + i;
currentWriteDescriptorSet->dstSet = commandBuffer->computeReadOnlyDescriptorSet;
currentWriteDescriptorSet->pTexelBufferView = NULL;
currentWriteDescriptorSet->pImageInfo = NULL;
bufferInfos[bufferInfoCount].buffer = commandBuffer->readOnlyComputeStorageBuffers[i]->buffer;
bufferInfos[bufferInfoCount].offset = 0;
bufferInfos[bufferInfoCount].range = VK_WHOLE_SIZE;
currentWriteDescriptorSet->pBufferInfo = &bufferInfos[bufferInfoCount];
bufferInfoCount += 1;
}
renderer->vkUpdateDescriptorSets(
renderer->logicalDevice,
resourceLayout->readOnlyStorageTextureCount + resourceLayout->readOnlyStorageBufferCount,
writeDescriptorSets,
0,
NULL
);
renderer->vkCmdBindDescriptorSets(
commandBuffer->commandBuffer,
VK_PIPELINE_BIND_POINT_COMPUTE,
resourceLayout->pipelineLayout,
0,
1,
&commandBuffer->computeReadOnlyDescriptorSet,
0,
NULL
);
SDL_stack_free(writeDescriptorSets);
bufferInfoCount = 0;
imageInfoCount = 0;
commandBuffer->needNewComputeReadOnlyDescriptorSet = SDL_FALSE;
}
if (commandBuffer->needNewComputeReadWriteDescriptorSet)
{
descriptorSetPool = &resourceLayout->descriptorSetPools[1];
commandBuffer->computeReadWriteDescriptorSet = VULKAN_INTERNAL_FetchDescriptorSet(
renderer,
commandBuffer,
descriptorSetPool
);
writeDescriptorSets = SDL_stack_alloc(
VkWriteDescriptorSet,
resourceLayout->readWriteStorageTextureCount +
resourceLayout->readWriteStorageBufferCount
);
for (i = 0; i < resourceLayout->readWriteStorageTextureCount; i += 1)
{
currentWriteDescriptorSet = &writeDescriptorSets[i];
currentWriteDescriptorSet->sType = VK_STRUCTURE_TYPE_WRITE_DESCRIPTOR_SET;
currentWriteDescriptorSet->pNext = NULL;
currentWriteDescriptorSet->descriptorCount = 1;
currentWriteDescriptorSet->descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_IMAGE;
currentWriteDescriptorSet->dstArrayElement = 0;
currentWriteDescriptorSet->dstBinding = i;
currentWriteDescriptorSet->dstSet = commandBuffer->computeReadWriteDescriptorSet;
currentWriteDescriptorSet->pTexelBufferView = NULL;
currentWriteDescriptorSet->pBufferInfo = NULL;
imageInfos[imageInfoCount].sampler = VK_NULL_HANDLE;
imageInfos[imageInfoCount].imageView = commandBuffer->readWriteComputeStorageTextureSlices[i]->view;
imageInfos[imageInfoCount].imageLayout = VK_IMAGE_LAYOUT_GENERAL;
currentWriteDescriptorSet->pImageInfo = &imageInfos[imageInfoCount];
imageInfoCount += 1;
}
for (i = 0; i < resourceLayout->readWriteStorageBufferCount; i += 1)
{
currentWriteDescriptorSet = &writeDescriptorSets[resourceLayout->readWriteStorageTextureCount + i];
currentWriteDescriptorSet->sType = VK_STRUCTURE_TYPE_WRITE_DESCRIPTOR_SET;
currentWriteDescriptorSet->pNext = NULL;
currentWriteDescriptorSet->descriptorCount = 1;
currentWriteDescriptorSet->descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_BUFFER;
currentWriteDescriptorSet->dstArrayElement = 0;
currentWriteDescriptorSet->dstBinding = resourceLayout->readWriteStorageTextureCount + i;
currentWriteDescriptorSet->dstSet = commandBuffer->computeReadWriteDescriptorSet;
currentWriteDescriptorSet->pTexelBufferView = NULL;
currentWriteDescriptorSet->pImageInfo = NULL;
bufferInfos[bufferInfoCount].buffer = commandBuffer->readWriteComputeStorageBuffers[i]->buffer;
bufferInfos[bufferInfoCount].offset = 0;
bufferInfos[bufferInfoCount].range = VK_WHOLE_SIZE;
currentWriteDescriptorSet->pBufferInfo = &bufferInfos[bufferInfoCount];
bufferInfoCount += 1;
}
renderer->vkUpdateDescriptorSets(
renderer->logicalDevice,
resourceLayout->readWriteStorageTextureCount + resourceLayout->readWriteStorageBufferCount,
writeDescriptorSets,
0,
NULL
);
renderer->vkCmdBindDescriptorSets(
commandBuffer->commandBuffer,
VK_PIPELINE_BIND_POINT_COMPUTE,
resourceLayout->pipelineLayout,
1,
1,
&commandBuffer->computeReadWriteDescriptorSet,
0,
NULL
);
SDL_stack_free(writeDescriptorSets);
bufferInfoCount = 0;
imageInfoCount = 0;
commandBuffer->needNewComputeReadWriteDescriptorSet = SDL_FALSE;
}
if (commandBuffer->needNewComputeUniformDescriptorSet)
{
descriptorSetPool = &resourceLayout->descriptorSetPools[2];
commandBuffer->computeUniformDescriptorSet = VULKAN_INTERNAL_FetchDescriptorSet(
renderer,
commandBuffer,
descriptorSetPool
);
writeDescriptorSets = SDL_stack_alloc(
VkWriteDescriptorSet,
resourceLayout->uniformBufferCount
);
for (i = 0; i < resourceLayout->uniformBufferCount; i += 1)
{
currentWriteDescriptorSet = &writeDescriptorSets[i];
currentWriteDescriptorSet->sType = VK_STRUCTURE_TYPE_WRITE_DESCRIPTOR_SET;
currentWriteDescriptorSet->pNext = NULL;
currentWriteDescriptorSet->descriptorCount = 1;
currentWriteDescriptorSet->descriptorType = VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER_DYNAMIC;
currentWriteDescriptorSet->dstArrayElement = 0;
currentWriteDescriptorSet->dstBinding = i;
currentWriteDescriptorSet->dstSet = commandBuffer->computeUniformDescriptorSet;
currentWriteDescriptorSet->pTexelBufferView = NULL;
currentWriteDescriptorSet->pImageInfo = NULL;
bufferInfos[bufferInfoCount].buffer = commandBuffer->computeUniformBuffers[i]->bufferContainer->activeBufferHandle->vulkanBuffer->buffer;
bufferInfos[bufferInfoCount].offset = 0;
bufferInfos[bufferInfoCount].range = MAX_UBO_SECTION_SIZE;
currentWriteDescriptorSet->pBufferInfo = &bufferInfos[bufferInfoCount];
bufferInfoCount += 1;
}
renderer->vkUpdateDescriptorSets(
renderer->logicalDevice,
resourceLayout->uniformBufferCount,
writeDescriptorSets,
0,
NULL
);
SDL_stack_free(writeDescriptorSets);
bufferInfoCount = 0;
imageInfoCount = 0;
commandBuffer->needNewComputeUniformDescriptorSet = SDL_FALSE;
commandBuffer->needNewComputeUniformOffsets = SDL_TRUE;
}
if (commandBuffer->needNewComputeUniformOffsets)
{
for (i = 0; i < resourceLayout->uniformBufferCount; i += 1)
{
dynamicOffsets[i] = commandBuffer->computeUniformBuffers[i]->drawOffset;
}
renderer->vkCmdBindDescriptorSets(
commandBuffer->commandBuffer,
VK_PIPELINE_BIND_POINT_COMPUTE,
resourceLayout->pipelineLayout,
2,
1,
&commandBuffer->computeUniformDescriptorSet,
resourceLayout->uniformBufferCount,
dynamicOffsets
);
commandBuffer->needNewComputeUniformOffsets = SDL_FALSE;
}
}
static void VULKAN_DispatchCompute(
Refresh_CommandBuffer *commandBuffer,
Uint32 groupCountX,
Uint32 groupCountY,
Uint32 groupCountZ
) {
VulkanCommandBuffer *vulkanCommandBuffer = (VulkanCommandBuffer*) commandBuffer;
VulkanRenderer *renderer = (VulkanRenderer*) vulkanCommandBuffer->renderer;
VULKAN_INTERNAL_BindComputeDescriptorSets(renderer, vulkanCommandBuffer);
renderer->vkCmdDispatch(
vulkanCommandBuffer->commandBuffer,
groupCountX,
groupCountY,
groupCountZ
);
}
static void VULKAN_EndComputePass(
Refresh_CommandBuffer *commandBuffer
) {
VulkanCommandBuffer *vulkanCommandBuffer = (VulkanCommandBuffer*) commandBuffer;
Uint32 i;
for (i = 0; i < MAX_STORAGE_TEXTURES_PER_STAGE; i += 1)
{
if (vulkanCommandBuffer->readWriteComputeStorageTextureSlices[i] != NULL)
{
VULKAN_INTERNAL_TextureTransitionToDefaultUsage(
vulkanCommandBuffer->renderer,
vulkanCommandBuffer,
VULKAN_TEXTURE_USAGE_MODE_COMPUTE_STORAGE_READ_WRITE,
vulkanCommandBuffer->readWriteComputeStorageTextureSlices[i]
);
vulkanCommandBuffer->readWriteComputeStorageTextureSlices[i] = NULL;
}
}
for (i = 0; i < MAX_STORAGE_BUFFERS_PER_STAGE; i += 1)
{
if (vulkanCommandBuffer->readWriteComputeStorageBuffers[i] != NULL)
{
VULKAN_INTERNAL_BufferTransitionToDefaultUsage(
vulkanCommandBuffer->renderer,
vulkanCommandBuffer,
VULKAN_BUFFER_USAGE_MODE_COMPUTE_STORAGE_READ_WRITE,
vulkanCommandBuffer->readWriteComputeStorageBuffers[i]
);
vulkanCommandBuffer->readWriteComputeStorageBuffers[i] = NULL;
}
}
for (i = 0; i < MAX_STORAGE_TEXTURES_PER_STAGE; i += 1)
{
if (vulkanCommandBuffer->readOnlyComputeStorageTextureSlices[i] != NULL)
{
VULKAN_INTERNAL_TextureTransitionToDefaultUsage(
vulkanCommandBuffer->renderer,
vulkanCommandBuffer,
VULKAN_TEXTURE_USAGE_MODE_COMPUTE_STORAGE_READ,
vulkanCommandBuffer->readOnlyComputeStorageTextureSlices[i]
);
vulkanCommandBuffer->readOnlyComputeStorageTextureSlices[i] = NULL;
}
}
for (i = 0; i < MAX_STORAGE_BUFFERS_PER_STAGE; i += 1)
{
if (vulkanCommandBuffer->readOnlyComputeStorageBuffers[i] != NULL)
{
VULKAN_INTERNAL_BufferTransitionToDefaultUsage(
vulkanCommandBuffer->renderer,
vulkanCommandBuffer,
VULKAN_BUFFER_USAGE_MODE_COMPUTE_STORAGE_READ,
vulkanCommandBuffer->readOnlyComputeStorageBuffers[i]
);
vulkanCommandBuffer->readOnlyComputeStorageBuffers[i] = NULL;
}
}
vulkanCommandBuffer->currentComputePipeline = NULL;
vulkanCommandBuffer->computeReadOnlyDescriptorSet = VK_NULL_HANDLE;
vulkanCommandBuffer->computeReadWriteDescriptorSet = VK_NULL_HANDLE;
vulkanCommandBuffer->computeUniformDescriptorSet = VK_NULL_HANDLE;
}
static void VULKAN_MapTransferBuffer(
Refresh_Renderer *driverData,
Refresh_TransferBuffer *transferBuffer,
SDL_bool cycle,
void **ppData
) {
VulkanRenderer *renderer = (VulkanRenderer*) driverData;
VulkanBufferContainer *transferBufferContainer = (VulkanBufferContainer*) transferBuffer;
if (
cycle &&
SDL_AtomicGet(&transferBufferContainer->activeBufferHandle->vulkanBuffer->referenceCount) > 0
) {
VULKAN_INTERNAL_CycleActiveBuffer(
renderer,
transferBufferContainer
);
}
Uint8 *bufferPointer =
transferBufferContainer->activeBufferHandle->vulkanBuffer->usedRegion->allocation->mapPointer +
transferBufferContainer->activeBufferHandle->vulkanBuffer->usedRegion->resourceOffset;
*ppData = bufferPointer;
}
static void VULKAN_UnmapTransferBuffer(
Refresh_Renderer *driverData,
Refresh_TransferBuffer *transferBuffer
) {
/* no-op because transfer buffers are persistently mapped */
(void)driverData;
(void)transferBuffer;
}
static void VULKAN_SetTransferData(
Refresh_Renderer *driverData,
void* data,
Refresh_TransferBuffer *transferBuffer,
Refresh_BufferCopy *copyParams,
SDL_bool cycle
) {
VulkanRenderer *renderer = (VulkanRenderer*) driverData;
VulkanBufferContainer *transferBufferContainer = (VulkanBufferContainer*) transferBuffer;
if (
cycle &&
SDL_AtomicGet(&transferBufferContainer->activeBufferHandle->vulkanBuffer->referenceCount) > 0
) {
VULKAN_INTERNAL_CycleActiveBuffer(
renderer,
transferBufferContainer
);
}
Uint8 *bufferPointer =
transferBufferContainer->activeBufferHandle->vulkanBuffer->usedRegion->allocation->mapPointer +
transferBufferContainer->activeBufferHandle->vulkanBuffer->usedRegion->resourceOffset +
copyParams->dstOffset;
SDL_memcpy(
bufferPointer,
((Uint8*) data) + copyParams->srcOffset,
copyParams->size
);
}
static void VULKAN_GetTransferData(
Refresh_Renderer *driverData,
Refresh_TransferBuffer *transferBuffer,
void* data,
Refresh_BufferCopy *copyParams
) {
(void) driverData; /* used by other backends */
VulkanBufferContainer *transferBufferContainer = (VulkanBufferContainer*) transferBuffer;
VulkanBuffer *vulkanBuffer = transferBufferContainer->activeBufferHandle->vulkanBuffer;
Uint8 *bufferPointer =
vulkanBuffer->usedRegion->allocation->mapPointer +
vulkanBuffer->usedRegion->resourceOffset +
copyParams->srcOffset;
SDL_memcpy(
((Uint8*) data) + copyParams->dstOffset,
bufferPointer,
copyParams->size
);
}
static void VULKAN_BeginCopyPass(
Refresh_CommandBuffer *commandBuffer
) {
/* no-op */
(void)commandBuffer;
}
static void VULKAN_UploadToTexture(
Refresh_CommandBuffer *commandBuffer,
Refresh_TransferBuffer *transferBuffer,
Refresh_TextureRegion *textureRegion,
Refresh_BufferImageCopy *copyParams,
SDL_bool cycle
) {
VulkanCommandBuffer *vulkanCommandBuffer = (VulkanCommandBuffer*) commandBuffer;
VulkanRenderer *renderer = (VulkanRenderer*) vulkanCommandBuffer->renderer;
VulkanBufferContainer *transferBufferContainer = (VulkanBufferContainer*) transferBuffer;
VulkanTextureContainer *vulkanTextureContainer = (VulkanTextureContainer*) textureRegion->textureSlice.texture;
VulkanTextureSlice *vulkanTextureSlice;
VkBufferImageCopy imageCopy;
/* Note that the transfer buffer does not need a barrier, as it is synced by the client */
vulkanTextureSlice = VULKAN_INTERNAL_PrepareTextureSliceForWrite(
renderer,
vulkanCommandBuffer,
vulkanTextureContainer,
textureRegion->textureSlice.layer,
textureRegion->textureSlice.mipLevel,
cycle,
VULKAN_TEXTURE_USAGE_MODE_COPY_DESTINATION
);
imageCopy.imageExtent.width = textureRegion->w;
imageCopy.imageExtent.height = textureRegion->h;
imageCopy.imageExtent.depth = textureRegion->d;
imageCopy.imageOffset.x = textureRegion->x;
imageCopy.imageOffset.y = textureRegion->y;
imageCopy.imageOffset.z = textureRegion->z;
imageCopy.imageSubresource.aspectMask = vulkanTextureSlice->parent->aspectFlags;
imageCopy.imageSubresource.baseArrayLayer = textureRegion->textureSlice.layer;
imageCopy.imageSubresource.layerCount = 1;
imageCopy.imageSubresource.mipLevel = textureRegion->textureSlice.mipLevel;
imageCopy.bufferOffset = copyParams->bufferOffset;
imageCopy.bufferRowLength = copyParams->bufferStride;
imageCopy.bufferImageHeight = copyParams->bufferImageHeight;
renderer->vkCmdCopyBufferToImage(
vulkanCommandBuffer->commandBuffer,
transferBufferContainer->activeBufferHandle->vulkanBuffer->buffer,
vulkanTextureSlice->parent->image,
VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL,
1,
&imageCopy
);
VULKAN_INTERNAL_TextureTransitionToDefaultUsage(
renderer,
vulkanCommandBuffer,
VULKAN_TEXTURE_USAGE_MODE_COPY_DESTINATION,
vulkanTextureSlice
);
VULKAN_INTERNAL_TrackBuffer(renderer, vulkanCommandBuffer, transferBufferContainer->activeBufferHandle->vulkanBuffer);
VULKAN_INTERNAL_TrackTextureSlice(renderer, vulkanCommandBuffer, vulkanTextureSlice);
}
static void VULKAN_UploadToBuffer(
Refresh_CommandBuffer *commandBuffer,
Refresh_TransferBuffer *transferBuffer,
Refresh_Buffer *buffer,
Refresh_BufferCopy *copyParams,
SDL_bool cycle
) {
VulkanCommandBuffer *vulkanCommandBuffer = (VulkanCommandBuffer*) commandBuffer;
VulkanRenderer *renderer = (VulkanRenderer*) vulkanCommandBuffer->renderer;
VulkanBufferContainer *transferBufferContainer = (VulkanBufferContainer*) transferBuffer;
VulkanBufferContainer *bufferContainer = (VulkanBufferContainer*) buffer;
VkBufferCopy bufferCopy;
/* Note that the transfer buffer does not need a barrier, as it is synced by the client */
VulkanBuffer *vulkanBuffer = VULKAN_INTERNAL_PrepareBufferForWrite(
renderer,
vulkanCommandBuffer,
bufferContainer,
cycle,
VULKAN_BUFFER_USAGE_MODE_COPY_DESTINATION
);
bufferCopy.srcOffset = copyParams->srcOffset;
bufferCopy.dstOffset = copyParams->dstOffset;
bufferCopy.size = copyParams->size;
renderer->vkCmdCopyBuffer(
vulkanCommandBuffer->commandBuffer,
transferBufferContainer->activeBufferHandle->vulkanBuffer->buffer,
vulkanBuffer->buffer,
1,
&bufferCopy
);
VULKAN_INTERNAL_BufferTransitionToDefaultUsage(
renderer,
vulkanCommandBuffer,
VULKAN_BUFFER_USAGE_MODE_COPY_DESTINATION,
vulkanBuffer
);
VULKAN_INTERNAL_TrackBuffer(renderer, vulkanCommandBuffer, transferBufferContainer->activeBufferHandle->vulkanBuffer);
VULKAN_INTERNAL_TrackBuffer(renderer, vulkanCommandBuffer, vulkanBuffer);
}
/* Readback */
static void VULKAN_DownloadFromTexture(
Refresh_CommandBuffer *commandBuffer,
Refresh_TextureRegion *textureRegion,
Refresh_TransferBuffer *transferBuffer,
Refresh_BufferImageCopy *copyParams
) {
VulkanCommandBuffer *vulkanCommandBuffer = (VulkanCommandBuffer*) commandBuffer;
VulkanRenderer *renderer = vulkanCommandBuffer->renderer;
VulkanTextureSlice *vulkanTextureSlice;
VulkanBufferContainer *transferBufferContainer = (VulkanBufferContainer*) transferBuffer;
VkBufferImageCopy imageCopy;
vulkanTextureSlice = VULKAN_INTERNAL_SDLToVulkanTextureSlice(&textureRegion->textureSlice);
/* Note that the transfer buffer does not need a barrier, as it is synced by the client */
VULKAN_INTERNAL_TextureTransitionFromDefaultUsage(
renderer,
vulkanCommandBuffer,
VULKAN_TEXTURE_USAGE_MODE_COPY_SOURCE,
vulkanTextureSlice
);
imageCopy.imageExtent.width = textureRegion->w;
imageCopy.imageExtent.height = textureRegion->h;
imageCopy.imageExtent.depth = textureRegion->d;
imageCopy.imageOffset.x = textureRegion->x;
imageCopy.imageOffset.y = textureRegion->y;
imageCopy.imageOffset.z = textureRegion->z;
imageCopy.imageSubresource.aspectMask = vulkanTextureSlice->parent->aspectFlags;
imageCopy.imageSubresource.baseArrayLayer = textureRegion->textureSlice.layer;
imageCopy.imageSubresource.layerCount = 1;
imageCopy.imageSubresource.mipLevel = textureRegion->textureSlice.mipLevel;
imageCopy.bufferOffset = copyParams->bufferOffset;
imageCopy.bufferRowLength = copyParams->bufferStride;
imageCopy.bufferImageHeight = copyParams->bufferImageHeight;
renderer->vkCmdCopyImageToBuffer(
vulkanCommandBuffer->commandBuffer,
vulkanTextureSlice->parent->image,
VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL,
transferBufferContainer->activeBufferHandle->vulkanBuffer->buffer,
1,
&imageCopy
);
VULKAN_INTERNAL_TextureTransitionToDefaultUsage(
renderer,
vulkanCommandBuffer,
VULKAN_TEXTURE_USAGE_MODE_COPY_SOURCE,
vulkanTextureSlice
);
VULKAN_INTERNAL_TrackBuffer(renderer, vulkanCommandBuffer, transferBufferContainer->activeBufferHandle->vulkanBuffer);
VULKAN_INTERNAL_TrackTextureSlice(renderer, vulkanCommandBuffer, vulkanTextureSlice);
}
static void VULKAN_DownloadFromBuffer(
Refresh_CommandBuffer *commandBuffer,
Refresh_Buffer *buffer,
Refresh_TransferBuffer *transferBuffer,
Refresh_BufferCopy *copyParams
) {
VulkanCommandBuffer *vulkanCommandBuffer = (VulkanCommandBuffer*) commandBuffer;
VulkanRenderer *renderer = vulkanCommandBuffer->renderer;
VulkanBufferContainer *bufferContainer = (VulkanBufferContainer*) buffer;
VulkanBufferContainer *transferBufferContainer = (VulkanBufferContainer*) transferBuffer;
VkBufferCopy bufferCopy;
/* Note that transfer buffer does not need a barrier, as it is synced by the client */
VULKAN_INTERNAL_BufferTransitionFromDefaultUsage(
renderer,
vulkanCommandBuffer,
VULKAN_BUFFER_USAGE_MODE_COPY_SOURCE,
bufferContainer->activeBufferHandle->vulkanBuffer
);
bufferCopy.srcOffset = copyParams->srcOffset;
bufferCopy.dstOffset = copyParams->dstOffset;
bufferCopy.size = copyParams->size;
renderer->vkCmdCopyBuffer(
vulkanCommandBuffer->commandBuffer,
bufferContainer->activeBufferHandle->vulkanBuffer->buffer,
transferBufferContainer->activeBufferHandle->vulkanBuffer->buffer,
1,
&bufferCopy
);
VULKAN_INTERNAL_BufferTransitionToDefaultUsage(
renderer,
vulkanCommandBuffer,
VULKAN_BUFFER_USAGE_MODE_COPY_SOURCE,
bufferContainer->activeBufferHandle->vulkanBuffer
);
VULKAN_INTERNAL_TrackBuffer(renderer, vulkanCommandBuffer, transferBufferContainer->activeBufferHandle->vulkanBuffer);
VULKAN_INTERNAL_TrackBuffer(renderer, vulkanCommandBuffer, bufferContainer->activeBufferHandle->vulkanBuffer);
}
static void VULKAN_CopyTextureToTexture(
Refresh_CommandBuffer *commandBuffer,
Refresh_TextureRegion *source,
Refresh_TextureRegion *destination,
SDL_bool cycle
) {
VulkanCommandBuffer *vulkanCommandBuffer = (VulkanCommandBuffer*) commandBuffer;
VulkanRenderer *renderer = (VulkanRenderer*) vulkanCommandBuffer->renderer;
VulkanTextureContainer *dstContainer = (VulkanTextureContainer*) destination->textureSlice.texture;
VulkanTextureSlice *srcSlice;
VulkanTextureSlice *dstSlice;
VkImageCopy imageCopy;
srcSlice = VULKAN_INTERNAL_SDLToVulkanTextureSlice(&source->textureSlice);
dstSlice = VULKAN_INTERNAL_PrepareTextureSliceForWrite(
renderer,
vulkanCommandBuffer,
dstContainer,
destination->textureSlice.layer,
destination->textureSlice.mipLevel,
cycle,
VULKAN_TEXTURE_USAGE_MODE_COPY_DESTINATION
);
VULKAN_INTERNAL_TextureTransitionFromDefaultUsage(
renderer,
vulkanCommandBuffer,
VULKAN_TEXTURE_USAGE_MODE_COPY_SOURCE,
srcSlice
);
imageCopy.srcOffset.x = source->x;
imageCopy.srcOffset.y = source->y;
imageCopy.srcOffset.z = source->z;
imageCopy.srcSubresource.aspectMask = srcSlice->parent->aspectFlags;
imageCopy.srcSubresource.baseArrayLayer = source->textureSlice.layer;
imageCopy.srcSubresource.layerCount = 1;
imageCopy.srcSubresource.mipLevel = source->textureSlice.mipLevel;
imageCopy.dstOffset.x = destination->x;
imageCopy.dstOffset.y = destination->y;
imageCopy.dstOffset.z = destination->z;
imageCopy.dstSubresource.aspectMask = dstSlice->parent->aspectFlags;
imageCopy.dstSubresource.baseArrayLayer = destination->textureSlice.layer;
imageCopy.dstSubresource.layerCount = 1;
imageCopy.dstSubresource.mipLevel = destination->textureSlice.mipLevel;
imageCopy.extent.width = source->w;
imageCopy.extent.height = source->h;
imageCopy.extent.depth = source->d;
renderer->vkCmdCopyImage(
vulkanCommandBuffer->commandBuffer,
srcSlice->parent->image,
VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL,
dstSlice->parent->image,
VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL,
1,
&imageCopy
);
VULKAN_INTERNAL_TextureTransitionToDefaultUsage(
renderer,
vulkanCommandBuffer,
VULKAN_TEXTURE_USAGE_MODE_COPY_SOURCE,
srcSlice
);
VULKAN_INTERNAL_TextureTransitionToDefaultUsage(
renderer,
vulkanCommandBuffer,
VULKAN_TEXTURE_USAGE_MODE_COPY_DESTINATION,
dstSlice
);
VULKAN_INTERNAL_TrackTextureSlice(renderer, vulkanCommandBuffer, srcSlice);
VULKAN_INTERNAL_TrackTextureSlice(renderer, vulkanCommandBuffer, dstSlice);
}
static void VULKAN_CopyBufferToBuffer(
Refresh_CommandBuffer *commandBuffer,
Refresh_Buffer *source,
Refresh_Buffer *destination,
Refresh_BufferCopy *copyParams,
SDL_bool cycle
) {
VulkanCommandBuffer *vulkanCommandBuffer = (VulkanCommandBuffer*) commandBuffer;
VulkanRenderer *renderer = (VulkanRenderer*) vulkanCommandBuffer->renderer;
VulkanBufferContainer *srcContainer = (VulkanBufferContainer*) source;
VulkanBufferContainer *dstContainer = (VulkanBufferContainer*) destination;
VkBufferCopy bufferCopy;
VulkanBuffer *dstBuffer = VULKAN_INTERNAL_PrepareBufferForWrite(
renderer,
vulkanCommandBuffer,
dstContainer,
cycle,
VULKAN_BUFFER_USAGE_MODE_COPY_DESTINATION
);
VULKAN_INTERNAL_BufferTransitionFromDefaultUsage(
renderer,
vulkanCommandBuffer,
VULKAN_BUFFER_USAGE_MODE_COPY_SOURCE,
srcContainer->activeBufferHandle->vulkanBuffer
);
bufferCopy.srcOffset = copyParams->srcOffset;
bufferCopy.dstOffset = copyParams->dstOffset;
bufferCopy.size = copyParams->size;
renderer->vkCmdCopyBuffer(
vulkanCommandBuffer->commandBuffer,
srcContainer->activeBufferHandle->vulkanBuffer->buffer,
dstBuffer->buffer,
1,
&bufferCopy
);
VULKAN_INTERNAL_BufferTransitionToDefaultUsage(
renderer,
vulkanCommandBuffer,
VULKAN_BUFFER_USAGE_MODE_COPY_SOURCE,
srcContainer->activeBufferHandle->vulkanBuffer
);
VULKAN_INTERNAL_BufferTransitionToDefaultUsage(
renderer,
vulkanCommandBuffer,
VULKAN_BUFFER_USAGE_MODE_COPY_DESTINATION,
dstBuffer
);
VULKAN_INTERNAL_TrackBuffer(renderer, vulkanCommandBuffer, srcContainer->activeBufferHandle->vulkanBuffer);
VULKAN_INTERNAL_TrackBuffer(renderer, vulkanCommandBuffer, dstBuffer);
}
static void VULKAN_GenerateMipmaps(
Refresh_CommandBuffer *commandBuffer,
Refresh_Texture *texture
) {
VulkanCommandBuffer *vulkanCommandBuffer = (VulkanCommandBuffer*) commandBuffer;
VulkanRenderer *renderer = (VulkanRenderer*) vulkanCommandBuffer->renderer;
VulkanTexture *vulkanTexture = ((VulkanTextureContainer*) texture)->activeTextureHandle->vulkanTexture;
VulkanTextureSlice *srcTextureSlice;
VulkanTextureSlice *dstTextureSlice;
VkImageBlit blit;
Uint32 layer, level;
if (vulkanTexture->levelCount <= 1) { return; }
/* Blit each slice sequentially. Barriers, barriers everywhere! */
for (layer = 0; layer < vulkanTexture->layerCount; layer += 1)
for (level = 1; level < vulkanTexture->levelCount; level += 1)
{
srcTextureSlice = VULKAN_INTERNAL_FetchTextureSlice(
vulkanTexture,
layer,
level - 1
);
dstTextureSlice = VULKAN_INTERNAL_FetchTextureSlice(
vulkanTexture,
layer,
level
);
VULKAN_INTERNAL_TextureTransitionFromDefaultUsage(
renderer,
vulkanCommandBuffer,
VULKAN_TEXTURE_USAGE_MODE_COPY_SOURCE,
srcTextureSlice
);
VULKAN_INTERNAL_TextureTransitionFromDefaultUsage(
renderer,
vulkanCommandBuffer,
VULKAN_TEXTURE_USAGE_MODE_COPY_DESTINATION,
dstTextureSlice
);
blit.srcOffsets[0].x = 0;
blit.srcOffsets[0].y = 0;
blit.srcOffsets[0].z = 0;
blit.srcOffsets[1].x = vulkanTexture->dimensions.width >> (level - 1);
blit.srcOffsets[1].y = vulkanTexture->dimensions.height >> (level - 1);
blit.srcOffsets[1].z = 1;
blit.dstOffsets[0].x = 0;
blit.dstOffsets[0].y = 0;
blit.dstOffsets[0].z = 0;
blit.dstOffsets[1].x = vulkanTexture->dimensions.width >> level;
blit.dstOffsets[1].y = vulkanTexture->dimensions.height >> level;
blit.dstOffsets[1].z = 1;
blit.srcSubresource.aspectMask = VK_IMAGE_ASPECT_COLOR_BIT;
blit.srcSubresource.baseArrayLayer = layer;
blit.srcSubresource.layerCount = 1;
blit.srcSubresource.mipLevel = level - 1;
blit.dstSubresource.aspectMask = VK_IMAGE_ASPECT_COLOR_BIT;
blit.dstSubresource.baseArrayLayer = layer;
blit.dstSubresource.layerCount = 1;
blit.dstSubresource.mipLevel = level;
renderer->vkCmdBlitImage(
vulkanCommandBuffer->commandBuffer,
vulkanTexture->image,
VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL,
vulkanTexture->image,
VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL,
1,
&blit,
VK_FILTER_LINEAR
);
VULKAN_INTERNAL_TextureTransitionToDefaultUsage(
renderer,
vulkanCommandBuffer,
VULKAN_TEXTURE_USAGE_MODE_COPY_SOURCE,
srcTextureSlice
);
VULKAN_INTERNAL_TextureTransitionToDefaultUsage(
renderer,
vulkanCommandBuffer,
VULKAN_TEXTURE_USAGE_MODE_COPY_DESTINATION,
dstTextureSlice
);
VULKAN_INTERNAL_TrackTextureSlice(renderer, vulkanCommandBuffer, srcTextureSlice);
VULKAN_INTERNAL_TrackTextureSlice(renderer, vulkanCommandBuffer, dstTextureSlice);
}
}
static void VULKAN_EndCopyPass(
Refresh_CommandBuffer *commandBuffer
) {
/* no-op */
(void) commandBuffer;
}
static void VULKAN_Blit(
Refresh_CommandBuffer *commandBuffer,
Refresh_TextureRegion *source,
Refresh_TextureRegion *destination,
Refresh_Filter filterMode,
SDL_bool cycle
) {
VulkanCommandBuffer *vulkanCommandBuffer = (VulkanCommandBuffer*) commandBuffer;
VulkanRenderer *renderer = (VulkanRenderer*) vulkanCommandBuffer->renderer;
VulkanTextureContainer *sourceTextureContainer = (VulkanTextureContainer*) source->textureSlice.texture;
VkImageBlit region;
VulkanTextureSlice *srcTextureSlice = VULKAN_INTERNAL_FetchTextureSlice(
sourceTextureContainer->activeTextureHandle->vulkanTexture,
source->textureSlice.layer,
source->textureSlice.mipLevel
);
VulkanTextureSlice *dstTextureSlice = VULKAN_INTERNAL_PrepareTextureSliceForWrite(
renderer,
vulkanCommandBuffer,
(VulkanTextureContainer*) destination->textureSlice.texture,
destination->textureSlice.layer,
destination->textureSlice.mipLevel,
cycle,
VULKAN_TEXTURE_USAGE_MODE_COPY_DESTINATION
);
VULKAN_INTERNAL_TextureTransitionFromDefaultUsage(
renderer,
vulkanCommandBuffer,
VULKAN_TEXTURE_USAGE_MODE_COPY_SOURCE,
srcTextureSlice
);
region.srcSubresource.aspectMask = srcTextureSlice->parent->aspectFlags;
region.srcSubresource.baseArrayLayer = srcTextureSlice->layer;
region.srcSubresource.layerCount = 1;
region.srcSubresource.mipLevel = srcTextureSlice->level;
region.srcOffsets[0].x = source->x;
region.srcOffsets[0].y = source->y;
region.srcOffsets[0].z = source->z;
region.srcOffsets[1].x = source->x + source->w;
region.srcOffsets[1].y = source->y + source->h;
region.srcOffsets[1].z = source->z + source->d;
region.dstSubresource.aspectMask = dstTextureSlice->parent->aspectFlags;
region.dstSubresource.baseArrayLayer = dstTextureSlice->layer;
region.dstSubresource.layerCount = 1;
region.dstSubresource.mipLevel = dstTextureSlice->level;
region.dstOffsets[0].x = destination->x;
region.dstOffsets[0].y = destination->y;
region.dstOffsets[0].z = destination->z;
region.dstOffsets[1].x = destination->x + destination->w;
region.dstOffsets[1].y = destination->y + destination->h;
region.dstOffsets[1].z = destination->z + destination->d;
renderer->vkCmdBlitImage(
vulkanCommandBuffer->commandBuffer,
srcTextureSlice->parent->image,
VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL,
dstTextureSlice->parent->image,
VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL,
1,
&region,
SDLToVK_Filter[filterMode]
);
VULKAN_INTERNAL_TextureTransitionToDefaultUsage(
renderer,
vulkanCommandBuffer,
VULKAN_TEXTURE_USAGE_MODE_COPY_SOURCE,
srcTextureSlice
);
VULKAN_INTERNAL_TextureTransitionToDefaultUsage(
renderer,
vulkanCommandBuffer,
VULKAN_TEXTURE_USAGE_MODE_COPY_DESTINATION,
dstTextureSlice
);
VULKAN_INTERNAL_TrackTextureSlice(renderer, vulkanCommandBuffer, srcTextureSlice);
VULKAN_INTERNAL_TrackTextureSlice(renderer, vulkanCommandBuffer, dstTextureSlice);
}
static void VULKAN_INTERNAL_AllocateCommandBuffers(
VulkanRenderer *renderer,
VulkanCommandPool *vulkanCommandPool,
Uint32 allocateCount
) {
VkCommandBufferAllocateInfo allocateInfo;
VkResult vulkanResult;
Uint32 i;
VkCommandBuffer *commandBuffers = SDL_stack_alloc(VkCommandBuffer, allocateCount);
VulkanCommandBuffer *commandBuffer;
vulkanCommandPool->inactiveCommandBufferCapacity += allocateCount;
vulkanCommandPool->inactiveCommandBuffers = SDL_realloc(
vulkanCommandPool->inactiveCommandBuffers,
sizeof(VulkanCommandBuffer*) *
vulkanCommandPool->inactiveCommandBufferCapacity
);
allocateInfo.sType = VK_STRUCTURE_TYPE_COMMAND_BUFFER_ALLOCATE_INFO;
allocateInfo.pNext = NULL;
allocateInfo.commandPool = vulkanCommandPool->commandPool;
allocateInfo.commandBufferCount = allocateCount;
allocateInfo.level = VK_COMMAND_BUFFER_LEVEL_PRIMARY;
vulkanResult = renderer->vkAllocateCommandBuffers(
renderer->logicalDevice,
&allocateInfo,
commandBuffers
);
if (vulkanResult != VK_SUCCESS)
{
LogVulkanResultAsError("vkAllocateCommandBuffers", vulkanResult);
SDL_stack_free(commandBuffers);
return;
}
for (i = 0; i < allocateCount; i += 1)
{
commandBuffer = SDL_malloc(sizeof(VulkanCommandBuffer));
commandBuffer->renderer = renderer;
commandBuffer->commandPool = vulkanCommandPool;
commandBuffer->commandBuffer = commandBuffers[i];
commandBuffer->inFlightFence = VK_NULL_HANDLE;
/* Presentation tracking */
commandBuffer->presentDataCapacity = 1;
commandBuffer->presentDataCount = 0;
commandBuffer->presentDatas = SDL_malloc(
commandBuffer->presentDataCapacity * sizeof(VkPresentInfoKHR)
);
commandBuffer->waitSemaphoreCapacity = 1;
commandBuffer->waitSemaphoreCount = 0;
commandBuffer->waitSemaphores = SDL_malloc(
commandBuffer->waitSemaphoreCapacity * sizeof(VkSemaphore)
);
commandBuffer->signalSemaphoreCapacity = 1;
commandBuffer->signalSemaphoreCount = 0;
commandBuffer->signalSemaphores = SDL_malloc(
commandBuffer->signalSemaphoreCapacity * sizeof(VkSemaphore)
);
/* Descriptor set tracking */
commandBuffer->boundDescriptorSetDataCapacity = 16;
commandBuffer->boundDescriptorSetDataCount = 0;
commandBuffer->boundDescriptorSetDatas = SDL_malloc(
commandBuffer->boundDescriptorSetDataCapacity * sizeof(DescriptorSetData)
);
/* Resource bind tracking */
commandBuffer->needNewVertexResourceDescriptorSet = SDL_TRUE;
commandBuffer->needNewVertexUniformDescriptorSet = SDL_TRUE;
commandBuffer->needNewVertexUniformOffsets = SDL_TRUE;
commandBuffer->needNewFragmentResourceDescriptorSet = SDL_TRUE;
commandBuffer->needNewFragmentUniformDescriptorSet = SDL_TRUE;
commandBuffer->needNewFragmentUniformOffsets = SDL_TRUE;
commandBuffer->needNewComputeReadWriteDescriptorSet = SDL_TRUE;
commandBuffer->needNewComputeReadOnlyDescriptorSet = SDL_TRUE;
commandBuffer->needNewComputeUniformDescriptorSet = SDL_TRUE;
commandBuffer->needNewComputeUniformOffsets = SDL_TRUE;
commandBuffer->vertexResourceDescriptorSet = VK_NULL_HANDLE;
commandBuffer->vertexUniformDescriptorSet = VK_NULL_HANDLE;
commandBuffer->fragmentResourceDescriptorSet = VK_NULL_HANDLE;
commandBuffer->fragmentUniformDescriptorSet = VK_NULL_HANDLE;
commandBuffer->computeReadOnlyDescriptorSet = VK_NULL_HANDLE;
commandBuffer->computeReadWriteDescriptorSet = VK_NULL_HANDLE;
commandBuffer->computeUniformDescriptorSet = VK_NULL_HANDLE;
/* Uniform buffers */
commandBuffer->initializedVertexUniformBufferCount = 0;
commandBuffer->initializedFragmentUniformBufferCount = 0;
commandBuffer->initializedComputeUniformBufferCount = 0;
/* Resource tracking */
commandBuffer->usedBufferCapacity = 4;
commandBuffer->usedBufferCount = 0;
commandBuffer->usedBuffers = SDL_malloc(
commandBuffer->usedBufferCapacity * sizeof(VulkanBuffer*)
);
commandBuffer->usedTextureSliceCapacity = 4;
commandBuffer->usedTextureSliceCount = 0;
commandBuffer->usedTextureSlices = SDL_malloc(
commandBuffer->usedTextureSliceCapacity * sizeof(VulkanTextureSlice*)
);
commandBuffer->usedSamplerCapacity = 4;
commandBuffer->usedSamplerCount = 0;
commandBuffer->usedSamplers = SDL_malloc(
commandBuffer->usedSamplerCapacity * sizeof(VulkanSampler*)
);
commandBuffer->usedGraphicsPipelineCapacity = 4;
commandBuffer->usedGraphicsPipelineCount = 0;
commandBuffer->usedGraphicsPipelines = SDL_malloc(
commandBuffer->usedGraphicsPipelineCapacity * sizeof(VulkanGraphicsPipeline*)
);
commandBuffer->usedComputePipelineCapacity = 4;
commandBuffer->usedComputePipelineCount = 0;
commandBuffer->usedComputePipelines = SDL_malloc(
commandBuffer->usedComputePipelineCapacity * sizeof(VulkanComputePipeline*)
);
commandBuffer->usedFramebufferCapacity = 4;
commandBuffer->usedFramebufferCount = 0;
commandBuffer->usedFramebuffers = SDL_malloc(
commandBuffer->usedFramebufferCapacity * sizeof(VulkanFramebuffer*)
);
/* Pool it! */
vulkanCommandPool->inactiveCommandBuffers[
vulkanCommandPool->inactiveCommandBufferCount
] = commandBuffer;
vulkanCommandPool->inactiveCommandBufferCount += 1;
}
SDL_stack_free(commandBuffers);
}
static VulkanCommandPool* VULKAN_INTERNAL_FetchCommandPool(
VulkanRenderer *renderer,
SDL_threadID threadID
) {
VulkanCommandPool *vulkanCommandPool;
VkCommandPoolCreateInfo commandPoolCreateInfo;
VkResult vulkanResult;
CommandPoolHash commandPoolHash;
commandPoolHash.threadID = threadID;
vulkanCommandPool = CommandPoolHashTable_Fetch(
&renderer->commandPoolHashTable,
commandPoolHash
);
if (vulkanCommandPool != NULL)
{
return vulkanCommandPool;
}
vulkanCommandPool = (VulkanCommandPool*) SDL_malloc(sizeof(VulkanCommandPool));
commandPoolCreateInfo.sType = VK_STRUCTURE_TYPE_COMMAND_POOL_CREATE_INFO;
commandPoolCreateInfo.pNext = NULL;
commandPoolCreateInfo.flags = VK_COMMAND_POOL_CREATE_RESET_COMMAND_BUFFER_BIT;
commandPoolCreateInfo.queueFamilyIndex = renderer->queueFamilyIndex;
vulkanResult = renderer->vkCreateCommandPool(
renderer->logicalDevice,
&commandPoolCreateInfo,
NULL,
&vulkanCommandPool->commandPool
);
if (vulkanResult != VK_SUCCESS)
{
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Failed to create command pool!");
LogVulkanResultAsError("vkCreateCommandPool", vulkanResult);
return NULL;
}
vulkanCommandPool->threadID = threadID;
vulkanCommandPool->inactiveCommandBufferCapacity = 0;
vulkanCommandPool->inactiveCommandBufferCount = 0;
vulkanCommandPool->inactiveCommandBuffers = NULL;
VULKAN_INTERNAL_AllocateCommandBuffers(
renderer,
vulkanCommandPool,
2
);
CommandPoolHashTable_Insert(
&renderer->commandPoolHashTable,
commandPoolHash,
vulkanCommandPool
);
return vulkanCommandPool;
}
static VulkanCommandBuffer* VULKAN_INTERNAL_GetInactiveCommandBufferFromPool(
VulkanRenderer *renderer,
SDL_threadID threadID
) {
VulkanCommandPool *commandPool =
VULKAN_INTERNAL_FetchCommandPool(renderer, threadID);
VulkanCommandBuffer *commandBuffer;
if (commandPool == NULL)
{
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Failed to fetch command pool!");
return NULL;
}
if (commandPool->inactiveCommandBufferCount == 0)
{
VULKAN_INTERNAL_AllocateCommandBuffers(
renderer,
commandPool,
commandPool->inactiveCommandBufferCapacity
);
}
commandBuffer = commandPool->inactiveCommandBuffers[commandPool->inactiveCommandBufferCount - 1];
commandPool->inactiveCommandBufferCount -= 1;
return commandBuffer;
}
static Refresh_CommandBuffer* VULKAN_AcquireCommandBuffer(
Refresh_Renderer *driverData
) {
VulkanRenderer *renderer = (VulkanRenderer*) driverData;
VkResult result;
Uint32 i;
SDL_threadID threadID = SDL_ThreadID();
SDL_LockMutex(renderer->acquireCommandBufferLock);
VulkanCommandBuffer *commandBuffer =
VULKAN_INTERNAL_GetInactiveCommandBufferFromPool(renderer, threadID);
SDL_UnlockMutex(renderer->acquireCommandBufferLock);
if (commandBuffer == NULL)
{
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Failed to acquire command buffer!");
return NULL;
}
/* Reset state */
commandBuffer->currentComputePipeline = NULL;
commandBuffer->currentGraphicsPipeline = NULL;
for (i = 0; i < MAX_COLOR_TARGET_BINDINGS; i += 1)
{
commandBuffer->colorAttachmentSlices[i] = NULL;
}
commandBuffer->depthStencilAttachmentSlice = NULL;
commandBuffer->needNewVertexResourceDescriptorSet = SDL_TRUE;
commandBuffer->needNewVertexUniformDescriptorSet = SDL_TRUE;
commandBuffer->needNewVertexUniformOffsets = SDL_TRUE;
commandBuffer->needNewFragmentResourceDescriptorSet = SDL_TRUE;
commandBuffer->needNewFragmentUniformDescriptorSet = SDL_TRUE;
commandBuffer->needNewFragmentUniformOffsets = SDL_TRUE;
commandBuffer->needNewComputeReadOnlyDescriptorSet = SDL_TRUE;
commandBuffer->needNewComputeUniformDescriptorSet = SDL_TRUE;
commandBuffer->needNewComputeUniformOffsets = SDL_TRUE;
commandBuffer->vertexResourceDescriptorSet = VK_NULL_HANDLE;
commandBuffer->vertexUniformDescriptorSet = VK_NULL_HANDLE;
commandBuffer->fragmentResourceDescriptorSet = VK_NULL_HANDLE;
commandBuffer->fragmentUniformDescriptorSet = VK_NULL_HANDLE;
commandBuffer->computeReadOnlyDescriptorSet = VK_NULL_HANDLE;
commandBuffer->computeReadWriteDescriptorSet = VK_NULL_HANDLE;
commandBuffer->computeUniformDescriptorSet = VK_NULL_HANDLE;
SDL_zeroa(commandBuffer->vertexSamplerTextures);
SDL_zeroa(commandBuffer->vertexSamplers);
SDL_zeroa(commandBuffer->vertexStorageTextureSlices);
SDL_zeroa(commandBuffer->vertexStorageBuffers);
SDL_zeroa(commandBuffer->fragmentSamplerTextures);
SDL_zeroa(commandBuffer->fragmentSamplers);
SDL_zeroa(commandBuffer->fragmentStorageTextureSlices);
SDL_zeroa(commandBuffer->fragmentStorageBuffers);
SDL_zeroa(commandBuffer->readWriteComputeStorageTextureSlices);
SDL_zeroa(commandBuffer->readWriteComputeStorageBuffers);
SDL_zeroa(commandBuffer->readOnlyComputeStorageTextureSlices);
SDL_zeroa(commandBuffer->readOnlyComputeStorageBuffers);
commandBuffer->autoReleaseFence = 1;
commandBuffer->isDefrag = 0;
/* Reset the command buffer here to avoid resets being called
* from a separate thread than where the command buffer was acquired
*/
result = renderer->vkResetCommandBuffer(
commandBuffer->commandBuffer,
VK_COMMAND_BUFFER_RESET_RELEASE_RESOURCES_BIT
);
if (result != VK_SUCCESS)
{
LogVulkanResultAsError("vkResetCommandBuffer", result);
}
VULKAN_INTERNAL_BeginCommandBuffer(renderer, commandBuffer);
return (Refresh_CommandBuffer*) commandBuffer;
}
static SDL_bool VULKAN_QueryFence(
Refresh_Renderer *driverData,
Refresh_Fence *fence
) {
VulkanRenderer* renderer = (VulkanRenderer*) driverData;
VkResult result;
result = renderer->vkGetFenceStatus(
renderer->logicalDevice,
((VulkanFenceHandle*) fence)->fence
);
if (result == VK_SUCCESS)
{
return 1;
}
else if (result == VK_NOT_READY)
{
return 0;
}
else
{
LogVulkanResultAsError("vkGetFenceStatus", result);
return 0;
}
}
static void VULKAN_INTERNAL_ReturnFenceToPool(
VulkanRenderer *renderer,
VulkanFenceHandle *fenceHandle
) {
SDL_LockMutex(renderer->fencePool.lock);
EXPAND_ARRAY_IF_NEEDED(
renderer->fencePool.availableFences,
VulkanFenceHandle*,
renderer->fencePool.availableFenceCount + 1,
renderer->fencePool.availableFenceCapacity,
renderer->fencePool.availableFenceCapacity * 2
);
renderer->fencePool.availableFences[renderer->fencePool.availableFenceCount] = fenceHandle;
renderer->fencePool.availableFenceCount += 1;
SDL_UnlockMutex(renderer->fencePool.lock);
}
static void VULKAN_ReleaseFence(
Refresh_Renderer *driverData,
Refresh_Fence *fence
) {
VulkanFenceHandle *handle = (VulkanFenceHandle*) fence;
if (SDL_AtomicDecRef(&handle->referenceCount))
{
VULKAN_INTERNAL_ReturnFenceToPool((VulkanRenderer*) driverData, handle);
}
}
static WindowData* VULKAN_INTERNAL_FetchWindowData(
SDL_Window *window
) {
return (WindowData*) SDL_GetWindowData(window, WINDOW_PROPERTY_DATA);
}
static SDL_bool VULKAN_SupportsSwapchainComposition(
Refresh_Renderer *driverData,
SDL_Window *window,
Refresh_SwapchainComposition swapchainComposition
) {
VulkanRenderer *renderer = (VulkanRenderer*) driverData;
WindowData *windowData = VULKAN_INTERNAL_FetchWindowData(window);
VkSurfaceKHR surface;
SwapchainSupportDetails supportDetails;
SDL_bool destroySurface = SDL_FALSE;
SDL_bool result = SDL_FALSE;
Uint32 i;
if (windowData == NULL || windowData->swapchainData == NULL)
{
/* Create a dummy surface is the window is not claimed */
destroySurface = SDL_TRUE;
if (!SDL_Vulkan_CreateSurface(
window,
renderer->instance,
&surface
)) {
return SDL_FALSE;
}
}
else
{
surface = windowData->swapchainData->surface;
}
VULKAN_INTERNAL_QuerySwapchainSupport(
renderer,
renderer->physicalDevice,
surface,
&supportDetails
);
for (i = 0; i < supportDetails.formatsLength; i += 1)
{
if ( supportDetails.formats[i].format == SwapchainCompositionToFormat[swapchainComposition] &&
supportDetails.formats[i].colorSpace == SwapchainCompositionToColorSpace[swapchainComposition] )
{
result = SDL_TRUE;
break;
}
}
SDL_free(supportDetails.formats);
SDL_free(supportDetails.presentModes);
if (destroySurface)
{
renderer->vkDestroySurfaceKHR(
renderer->instance,
surface,
NULL
);
}
return result;
}
static SDL_bool VULKAN_SupportsPresentMode(
Refresh_Renderer *driverData,
SDL_Window *window,
Refresh_PresentMode presentMode
) {
VulkanRenderer *renderer = (VulkanRenderer*) driverData;
WindowData *windowData = VULKAN_INTERNAL_FetchWindowData(window);
VkSurfaceKHR surface;
SwapchainSupportDetails supportDetails;
SDL_bool destroySurface = SDL_FALSE;
SDL_bool result = SDL_FALSE;
Uint32 i;
if (windowData == NULL || windowData->swapchainData == NULL)
{
/* Create a dummy surface is the window is not claimed */
destroySurface = SDL_TRUE;
if (!SDL_Vulkan_CreateSurface(
window,
renderer->instance,
&surface
)) {
return SDL_FALSE;
}
}
else
{
surface = windowData->swapchainData->surface;
}
VULKAN_INTERNAL_QuerySwapchainSupport(
renderer,
renderer->physicalDevice,
surface,
&supportDetails
);
for (i = 0; i < supportDetails.presentModesLength; i += 1)
{
if (supportDetails.presentModes[i] == SDLToVK_PresentMode[presentMode])
{
result = SDL_TRUE;
break;
}
}
SDL_free(supportDetails.formats);
SDL_free(supportDetails.presentModes);
if (destroySurface)
{
renderer->vkDestroySurfaceKHR(
renderer->instance,
surface,
NULL
);
}
return result;
}
static SDL_bool VULKAN_ClaimWindow(
Refresh_Renderer *driverData,
SDL_Window *window,
Refresh_SwapchainComposition swapchainComposition,
Refresh_PresentMode presentMode
) {
VulkanRenderer *renderer = (VulkanRenderer*) driverData;
WindowData *windowData = VULKAN_INTERNAL_FetchWindowData(window);
if (windowData == NULL)
{
windowData = SDL_malloc(sizeof(WindowData));
windowData->window = window;
windowData->presentMode = presentMode;
windowData->swapchainComposition = swapchainComposition;
if (VULKAN_INTERNAL_CreateSwapchain(renderer, windowData))
{
SDL_SetWindowData(window, WINDOW_PROPERTY_DATA, windowData);
if (renderer->claimedWindowCount >= renderer->claimedWindowCapacity)
{
renderer->claimedWindowCapacity *= 2;
renderer->claimedWindows = SDL_realloc(
renderer->claimedWindows,
renderer->claimedWindowCapacity * sizeof(WindowData*)
);
}
renderer->claimedWindows[renderer->claimedWindowCount] = windowData;
renderer->claimedWindowCount += 1;
return 1;
}
else
{
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Could not create swapchain, failed to claim window!");
SDL_free(windowData);
return 0;
}
}
else
{
SDL_LogWarn(SDL_LOG_CATEGORY_APPLICATION, "Window already claimed!");
return 0;
}
}
static void VULKAN_UnclaimWindow(
Refresh_Renderer *driverData,
SDL_Window *window
) {
VulkanRenderer *renderer = (VulkanRenderer*) driverData;
WindowData *windowData = VULKAN_INTERNAL_FetchWindowData(window);
Uint32 i;
if (windowData == NULL)
{
return;
}
if (windowData->swapchainData != NULL)
{
VULKAN_Wait(driverData);
for (i = 0; i < MAX_FRAMES_IN_FLIGHT; i += 1)
{
if (windowData->swapchainData->inFlightFences[i] != NULL)
{
VULKAN_ReleaseFence(
driverData,
(Refresh_Fence*) windowData->swapchainData->inFlightFences[i]
);
}
}
VULKAN_INTERNAL_DestroySwapchain(
(VulkanRenderer*) driverData,
windowData
);
}
for (i = 0; i < renderer->claimedWindowCount; i += 1)
{
if (renderer->claimedWindows[i]->window == window)
{
renderer->claimedWindows[i] = renderer->claimedWindows[renderer->claimedWindowCount - 1];
renderer->claimedWindowCount -= 1;
break;
}
}
SDL_free(windowData);
SDL_SetWindowData(window, WINDOW_PROPERTY_DATA, NULL);
}
static void VULKAN_INTERNAL_RecreateSwapchain(
VulkanRenderer* renderer,
WindowData *windowData
) {
Uint32 i;
if (windowData->swapchainData != NULL)
{
VULKAN_Wait((Refresh_Renderer *)renderer);
for (i = 0; i < MAX_FRAMES_IN_FLIGHT; i += 1)
{
if (windowData->swapchainData->inFlightFences[i] != NULL)
{
VULKAN_ReleaseFence(
(Refresh_Renderer*) renderer,
(Refresh_Fence*) windowData->swapchainData->inFlightFences[i]
);
}
}
}
VULKAN_INTERNAL_DestroySwapchain(renderer, windowData);
VULKAN_INTERNAL_CreateSwapchain(renderer, windowData);
}
static Refresh_Texture* VULKAN_AcquireSwapchainTexture(
Refresh_CommandBuffer *commandBuffer,
SDL_Window *window,
Uint32 *pWidth,
Uint32 *pHeight
) {
VulkanCommandBuffer *vulkanCommandBuffer = (VulkanCommandBuffer*) commandBuffer;
VulkanRenderer *renderer = (VulkanRenderer*) vulkanCommandBuffer->renderer;
Uint32 swapchainImageIndex;
WindowData *windowData;
VulkanSwapchainData *swapchainData;
VkResult acquireResult = VK_SUCCESS;
VulkanTextureContainer *swapchainTextureContainer = NULL;
VulkanPresentData *presentData;
windowData = VULKAN_INTERNAL_FetchWindowData(window);
if (windowData == NULL)
{
return NULL;
}
swapchainData = windowData->swapchainData;
/* Window is claimed but swapchain is invalid! */
if (swapchainData == NULL)
{
if (SDL_GetWindowFlags(window) & SDL_WINDOW_MINIMIZED)
{
/* Window is minimized, don't bother */
return NULL;
}
/* Let's try to recreate */
VULKAN_INTERNAL_RecreateSwapchain(renderer, windowData);
swapchainData = windowData->swapchainData;
if (swapchainData == NULL)
{
SDL_LogWarn(SDL_LOG_CATEGORY_APPLICATION, "Failed to recreate swapchain!");
return NULL;
}
}
if (swapchainData->inFlightFences[swapchainData->frameCounter] != NULL)
{
if (swapchainData->presentMode == VK_PRESENT_MODE_FIFO_KHR)
{
/* In VSYNC mode, block until the least recent presented frame is done */
VULKAN_WaitForFences(
(Refresh_Renderer*) renderer,
SDL_TRUE,
1,
(Refresh_Fence**) &swapchainData->inFlightFences[swapchainData->frameCounter]
);
}
else
{
if (!VULKAN_QueryFence(
(Refresh_Renderer*) renderer,
(Refresh_Fence*) swapchainData->inFlightFences[swapchainData->frameCounter]
)) {
/*
* In MAILBOX or IMMEDIATE mode, if the least recent fence is not signaled,
* return NULL to indicate that rendering should be skipped
*/
return NULL;
}
}
VULKAN_ReleaseFence(
(Refresh_Renderer*) renderer,
(Refresh_Fence*) swapchainData->inFlightFences[swapchainData->frameCounter]
);
swapchainData->inFlightFences[swapchainData->frameCounter] = NULL;
}
acquireResult = renderer->vkAcquireNextImageKHR(
renderer->logicalDevice,
swapchainData->swapchain,
UINT64_MAX,
swapchainData->imageAvailableSemaphore[swapchainData->frameCounter],
VK_NULL_HANDLE,
&swapchainImageIndex
);
/* Acquisition is invalid, let's try to recreate */
if (acquireResult != VK_SUCCESS && acquireResult != VK_SUBOPTIMAL_KHR)
{
VULKAN_INTERNAL_RecreateSwapchain(renderer, windowData);
swapchainData = windowData->swapchainData;
if (swapchainData == NULL)
{
SDL_LogWarn(SDL_LOG_CATEGORY_APPLICATION, "Failed to recreate swapchain!");
return NULL;
}
acquireResult = renderer->vkAcquireNextImageKHR(
renderer->logicalDevice,
swapchainData->swapchain,
UINT64_MAX,
swapchainData->imageAvailableSemaphore[swapchainData->frameCounter],
VK_NULL_HANDLE,
&swapchainImageIndex
);
if (acquireResult != VK_SUCCESS && acquireResult != VK_SUBOPTIMAL_KHR)
{
SDL_LogWarn(SDL_LOG_CATEGORY_APPLICATION, "Failed to acquire swapchain texture!");
return NULL;
}
}
swapchainTextureContainer = &swapchainData->textureContainers[swapchainImageIndex];
/* We need a special execution dependency with pWaitDstStageMask or image transition can start before acquire finishes */
VkImageMemoryBarrier imageBarrier;
imageBarrier.sType = VK_STRUCTURE_TYPE_IMAGE_MEMORY_BARRIER;
imageBarrier.pNext = NULL;
imageBarrier.srcAccessMask = 0;
imageBarrier.dstAccessMask = VK_ACCESS_COLOR_ATTACHMENT_WRITE_BIT;
imageBarrier.oldLayout = VK_IMAGE_LAYOUT_UNDEFINED;
imageBarrier.newLayout = VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL;
imageBarrier.srcQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED;
imageBarrier.dstQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED;
imageBarrier.image = swapchainTextureContainer->activeTextureHandle->vulkanTexture->image;
imageBarrier.subresourceRange.aspectMask = VK_IMAGE_ASPECT_COLOR_BIT;
imageBarrier.subresourceRange.baseMipLevel = 0;
imageBarrier.subresourceRange.levelCount = 1;
imageBarrier.subresourceRange.baseArrayLayer = 0;
imageBarrier.subresourceRange.layerCount = 1;
renderer->vkCmdPipelineBarrier(
vulkanCommandBuffer->commandBuffer,
VK_PIPELINE_STAGE_COLOR_ATTACHMENT_OUTPUT_BIT,
VK_PIPELINE_STAGE_COLOR_ATTACHMENT_OUTPUT_BIT,
0,
0,
NULL,
0,
NULL,
1,
&imageBarrier
);
/* Set up present struct */
if (vulkanCommandBuffer->presentDataCount == vulkanCommandBuffer->presentDataCapacity)
{
vulkanCommandBuffer->presentDataCapacity += 1;
vulkanCommandBuffer->presentDatas = SDL_realloc(
vulkanCommandBuffer->presentDatas,
vulkanCommandBuffer->presentDataCapacity * sizeof(VkPresentInfoKHR)
);
}
presentData = &vulkanCommandBuffer->presentDatas[vulkanCommandBuffer->presentDataCount];
vulkanCommandBuffer->presentDataCount += 1;
presentData->windowData = windowData;
presentData->swapchainImageIndex = swapchainImageIndex;
/* Set up present semaphores */
if (vulkanCommandBuffer->waitSemaphoreCount == vulkanCommandBuffer->waitSemaphoreCapacity)
{
vulkanCommandBuffer->waitSemaphoreCapacity += 1;
vulkanCommandBuffer->waitSemaphores = SDL_realloc(
vulkanCommandBuffer->waitSemaphores,
vulkanCommandBuffer->waitSemaphoreCapacity * sizeof(VkSemaphore)
);
}
vulkanCommandBuffer->waitSemaphores[vulkanCommandBuffer->waitSemaphoreCount] =
swapchainData->imageAvailableSemaphore[swapchainData->frameCounter];
vulkanCommandBuffer->waitSemaphoreCount += 1;
if (vulkanCommandBuffer->signalSemaphoreCount == vulkanCommandBuffer->signalSemaphoreCapacity)
{
vulkanCommandBuffer->signalSemaphoreCapacity += 1;
vulkanCommandBuffer->signalSemaphores = SDL_realloc(
vulkanCommandBuffer->signalSemaphores,
vulkanCommandBuffer->signalSemaphoreCapacity * sizeof(VkSemaphore)
);
}
vulkanCommandBuffer->signalSemaphores[vulkanCommandBuffer->signalSemaphoreCount] =
swapchainData->renderFinishedSemaphore[swapchainData->frameCounter];
vulkanCommandBuffer->signalSemaphoreCount += 1;
*pWidth = swapchainData->extent.width;
*pHeight = swapchainData->extent.height;
return (Refresh_Texture*) swapchainTextureContainer;
}
static Refresh_TextureFormat VULKAN_GetSwapchainTextureFormat(
Refresh_Renderer *driverData,
SDL_Window *window
) {
WindowData *windowData = VULKAN_INTERNAL_FetchWindowData(window);
if (windowData == NULL)
{
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Cannot get swapchain format, window has not been claimed!");
return 0;
}
if (windowData->swapchainData == NULL)
{
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Cannot get swapchain format, swapchain is currently invalid!");
return 0;
}
switch (windowData->swapchainData->format)
{
case VK_FORMAT_R8G8B8A8_UNORM:
return REFRESH_TEXTUREFORMAT_R8G8B8A8;
case VK_FORMAT_B8G8R8A8_UNORM:
return REFRESH_TEXTUREFORMAT_B8G8R8A8;
case VK_FORMAT_B8G8R8A8_SRGB:
return REFRESH_TEXTUREFORMAT_B8G8R8A8_SRGB;
case VK_FORMAT_R8G8B8A8_SRGB:
return REFRESH_TEXTUREFORMAT_R8G8B8A8_SRGB;
case VK_FORMAT_R16G16B16A16_SFLOAT:
return REFRESH_TEXTUREFORMAT_R16G16B16A16_SFLOAT;
case VK_FORMAT_A2R10G10B10_UNORM_PACK32:
return REFRESH_TEXTUREFORMAT_A2R10G10B10;
case VK_FORMAT_A2B10G10R10_UNORM_PACK32:
return REFRESH_TEXTUREFORMAT_A2B10G10R10;
default:
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Unrecognized swapchain format!");
return 0;
}
}
static void VULKAN_SetSwapchainParameters(
Refresh_Renderer *driverData,
SDL_Window *window,
Refresh_SwapchainComposition swapchainComposition,
Refresh_PresentMode presentMode
) {
WindowData *windowData = VULKAN_INTERNAL_FetchWindowData(window);
if (windowData == NULL)
{
SDL_LogWarn(SDL_LOG_CATEGORY_APPLICATION, "Cannot set present mode, window has not been claimed!");
return;
}
/* The window size may have changed, always update even if these params are the same */
windowData->presentMode = presentMode;
windowData->swapchainComposition = swapchainComposition;
VULKAN_INTERNAL_RecreateSwapchain(
(VulkanRenderer *)driverData,
windowData
);
}
/* Submission structure */
static VulkanFenceHandle* VULKAN_INTERNAL_AcquireFenceFromPool(
VulkanRenderer *renderer
) {
VulkanFenceHandle *handle;
VkFenceCreateInfo fenceCreateInfo;
VkFence fence;
VkResult vulkanResult;
if (renderer->fencePool.availableFenceCount == 0)
{
/* Create fence */
fenceCreateInfo.sType = VK_STRUCTURE_TYPE_FENCE_CREATE_INFO;
fenceCreateInfo.pNext = NULL;
fenceCreateInfo.flags = 0;
vulkanResult = renderer->vkCreateFence(
renderer->logicalDevice,
&fenceCreateInfo,
NULL,
&fence
);
if (vulkanResult != VK_SUCCESS)
{
LogVulkanResultAsError("vkCreateFence", vulkanResult);
return NULL;
}
handle = SDL_malloc(sizeof(VulkanFenceHandle));
handle->fence = fence;
SDL_AtomicSet(&handle->referenceCount, 0);
return handle;
}
SDL_LockMutex(renderer->fencePool.lock);
handle = renderer->fencePool.availableFences[renderer->fencePool.availableFenceCount - 1];
renderer->fencePool.availableFenceCount -= 1;
vulkanResult = renderer->vkResetFences(
renderer->logicalDevice,
1,
&handle->fence
);
if (vulkanResult != VK_SUCCESS)
{
LogVulkanResultAsError("vkResetFences", vulkanResult);
}
SDL_UnlockMutex(renderer->fencePool.lock);
return handle;
}
static void VULKAN_INTERNAL_PerformPendingDestroys(
VulkanRenderer *renderer
) {
Sint32 i, sliceIndex;
Sint32 refCountTotal;
SDL_LockMutex(renderer->disposeLock);
for (i = renderer->texturesToDestroyCount - 1; i >= 0; i -= 1)
{
refCountTotal = 0;
for (sliceIndex = 0; sliceIndex < renderer->texturesToDestroy[i]->sliceCount; sliceIndex += 1)
{
refCountTotal += SDL_AtomicGet(&renderer->texturesToDestroy[i]->slices[sliceIndex].referenceCount);
}
if (refCountTotal == 0)
{
VULKAN_INTERNAL_DestroyTexture(
renderer,
renderer->texturesToDestroy[i]
);
renderer->texturesToDestroy[i] = renderer->texturesToDestroy[renderer->texturesToDestroyCount - 1];
renderer->texturesToDestroyCount -= 1;
}
}
for (i = renderer->buffersToDestroyCount - 1; i >= 0; i -= 1)
{
if (SDL_AtomicGet(&renderer->buffersToDestroy[i]->referenceCount) == 0)
{
VULKAN_INTERNAL_DestroyBuffer(
renderer,
renderer->buffersToDestroy[i]);
renderer->buffersToDestroy[i] = renderer->buffersToDestroy[renderer->buffersToDestroyCount - 1];
renderer->buffersToDestroyCount -= 1;
}
}
for (i = renderer->graphicsPipelinesToDestroyCount - 1; i >= 0; i -= 1)
{
if (SDL_AtomicGet(&renderer->graphicsPipelinesToDestroy[i]->referenceCount) == 0)
{
VULKAN_INTERNAL_DestroyGraphicsPipeline(
renderer,
renderer->graphicsPipelinesToDestroy[i]
);
renderer->graphicsPipelinesToDestroy[i] = renderer->graphicsPipelinesToDestroy[renderer->graphicsPipelinesToDestroyCount - 1];
renderer->graphicsPipelinesToDestroyCount -= 1;
}
}
for (i = renderer->computePipelinesToDestroyCount - 1; i >= 0; i -= 1)
{
if (SDL_AtomicGet(&renderer->computePipelinesToDestroy[i]->referenceCount) == 0)
{
VULKAN_INTERNAL_DestroyComputePipeline(
renderer,
renderer->computePipelinesToDestroy[i]
);
renderer->computePipelinesToDestroy[i] = renderer->computePipelinesToDestroy[renderer->computePipelinesToDestroyCount - 1];
renderer->computePipelinesToDestroyCount -= 1 ;
}
}
for (i = renderer->shadersToDestroyCount - 1; i >= 0; i -= 1)
{
if (SDL_AtomicGet(&renderer->shadersToDestroy[i]->referenceCount) == 0)
{
VULKAN_INTERNAL_DestroyShaderModule(
renderer,
renderer->shadersToDestroy[i]
);
renderer->shadersToDestroy[i] = renderer->shadersToDestroy[renderer->shadersToDestroyCount - 1];
renderer->shadersToDestroyCount -= 1;
}
}
for (i = renderer->samplersToDestroyCount - 1; i >= 0; i -= 1)
{
if (SDL_AtomicGet(&renderer->samplersToDestroy[i]->referenceCount) == 0)
{
VULKAN_INTERNAL_DestroySampler(
renderer,
renderer->samplersToDestroy[i]
);
renderer->samplersToDestroy[i] = renderer->samplersToDestroy[renderer->samplersToDestroyCount - 1];
renderer->samplersToDestroyCount -= 1;
}
}
for (i = renderer->framebuffersToDestroyCount - 1; i >= 0; i -= 1)
{
if (SDL_AtomicGet(&renderer->framebuffersToDestroy[i]->referenceCount) == 0)
{
VULKAN_INTERNAL_DestroyFramebuffer(
renderer,
renderer->framebuffersToDestroy[i]
);
renderer->framebuffersToDestroy[i] = renderer->framebuffersToDestroy[renderer->framebuffersToDestroyCount - 1];
renderer->framebuffersToDestroyCount -= 1;
}
}
SDL_UnlockMutex(renderer->disposeLock);
}
static void VULKAN_INTERNAL_CleanCommandBuffer(
VulkanRenderer *renderer,
VulkanCommandBuffer *commandBuffer
) {
Uint32 i;
DescriptorSetData *descriptorSetData;
if (commandBuffer->autoReleaseFence)
{
VULKAN_ReleaseFence(
(Refresh_Renderer*) renderer,
(Refresh_Fence*) commandBuffer->inFlightFence
);
commandBuffer->inFlightFence = NULL;
}
/* Bound descriptor sets are now available */
for (i = 0; i < commandBuffer->boundDescriptorSetDataCount; i += 1)
{
descriptorSetData = &commandBuffer->boundDescriptorSetDatas[i];
SDL_LockMutex(descriptorSetData->descriptorSetPool->lock);
if (descriptorSetData->descriptorSetPool->inactiveDescriptorSetCount == descriptorSetData->descriptorSetPool->inactiveDescriptorSetCapacity)
{
descriptorSetData->descriptorSetPool->inactiveDescriptorSetCapacity *= 2;
descriptorSetData->descriptorSetPool->inactiveDescriptorSets = SDL_realloc(
descriptorSetData->descriptorSetPool->inactiveDescriptorSets,
descriptorSetData->descriptorSetPool->inactiveDescriptorSetCapacity * sizeof(VkDescriptorSet)
);
}
descriptorSetData->descriptorSetPool->inactiveDescriptorSets[descriptorSetData->descriptorSetPool->inactiveDescriptorSetCount] = descriptorSetData->descriptorSet;
descriptorSetData->descriptorSetPool->inactiveDescriptorSetCount += 1;
SDL_UnlockMutex(descriptorSetData->descriptorSetPool->lock);
}
commandBuffer->boundDescriptorSetDataCount = 0;
/* Decrement reference counts */
for (i = 0; i < commandBuffer->usedBufferCount; i += 1)
{
(void)SDL_AtomicDecRef(&commandBuffer->usedBuffers[i]->referenceCount);
}
commandBuffer->usedBufferCount = 0;
for (i = 0; i < commandBuffer->usedTextureSliceCount; i += 1)
{
(void)SDL_AtomicDecRef(&commandBuffer->usedTextureSlices[i]->referenceCount);
}
commandBuffer->usedTextureSliceCount = 0;
for (i = 0; i < commandBuffer->usedSamplerCount; i += 1)
{
(void)SDL_AtomicDecRef(&commandBuffer->usedSamplers[i]->referenceCount);
}
commandBuffer->usedSamplerCount = 0;
for (i = 0; i < commandBuffer->usedGraphicsPipelineCount; i += 1)
{
(void)SDL_AtomicDecRef(&commandBuffer->usedGraphicsPipelines[i]->referenceCount);
}
commandBuffer->usedGraphicsPipelineCount = 0;
for (i = 0; i < commandBuffer->usedComputePipelineCount; i += 1)
{
(void)SDL_AtomicDecRef(&commandBuffer->usedComputePipelines[i]->referenceCount);
}
commandBuffer->usedComputePipelineCount = 0;
for (i = 0; i < commandBuffer->usedFramebufferCount; i += 1)
{
(void)SDL_AtomicDecRef(&commandBuffer->usedFramebuffers[i]->referenceCount);
}
commandBuffer->usedFramebufferCount = 0;
/* Reset presentation data */
commandBuffer->presentDataCount = 0;
commandBuffer->waitSemaphoreCount = 0;
commandBuffer->signalSemaphoreCount = 0;
/* Reset defrag state */
if (commandBuffer->isDefrag)
{
renderer->defragInProgress = 0;
}
/* Return command buffer to pool */
SDL_LockMutex(renderer->acquireCommandBufferLock);
if (commandBuffer->commandPool->inactiveCommandBufferCount == commandBuffer->commandPool->inactiveCommandBufferCapacity)
{
commandBuffer->commandPool->inactiveCommandBufferCapacity += 1;
commandBuffer->commandPool->inactiveCommandBuffers = SDL_realloc(
commandBuffer->commandPool->inactiveCommandBuffers,
commandBuffer->commandPool->inactiveCommandBufferCapacity * sizeof(VulkanCommandBuffer*)
);
}
commandBuffer->commandPool->inactiveCommandBuffers[
commandBuffer->commandPool->inactiveCommandBufferCount
] = commandBuffer;
commandBuffer->commandPool->inactiveCommandBufferCount += 1;
SDL_UnlockMutex(renderer->acquireCommandBufferLock);
/* Remove this command buffer from the submitted list */
for (i = 0; i < renderer->submittedCommandBufferCount; i += 1)
{
if (renderer->submittedCommandBuffers[i] == commandBuffer)
{
renderer->submittedCommandBuffers[i] = renderer->submittedCommandBuffers[renderer->submittedCommandBufferCount - 1];
renderer->submittedCommandBufferCount -= 1;
}
}
}
static void VULKAN_WaitForFences(
Refresh_Renderer *driverData,
SDL_bool waitAll,
Uint32 fenceCount,
Refresh_Fence **pFences
) {
VulkanRenderer* renderer = (VulkanRenderer*) driverData;
VkFence* fences = SDL_stack_alloc(VkFence, fenceCount);
VkResult result;
Sint32 i;
for (i = 0; i < fenceCount; i += 1)
{
fences[i] = ((VulkanFenceHandle*) pFences[i])->fence;
}
result = renderer->vkWaitForFences(
renderer->logicalDevice,
fenceCount,
fences,
waitAll,
UINT64_MAX
);
if (result != VK_SUCCESS)
{
LogVulkanResultAsError("vkWaitForFences", result);
}
SDL_stack_free(fences);
SDL_LockMutex(renderer->submitLock);
for (i = renderer->submittedCommandBufferCount - 1; i >= 0; i -= 1)
{
result = renderer->vkGetFenceStatus(
renderer->logicalDevice,
renderer->submittedCommandBuffers[i]->inFlightFence->fence
);
if (result == VK_SUCCESS)
{
VULKAN_INTERNAL_CleanCommandBuffer(
renderer,
renderer->submittedCommandBuffers[i]
);
}
}
VULKAN_INTERNAL_PerformPendingDestroys(renderer);
SDL_UnlockMutex(renderer->submitLock);
}
static void VULKAN_Wait(
Refresh_Renderer *driverData
) {
VulkanRenderer *renderer = (VulkanRenderer*) driverData;
VulkanCommandBuffer *commandBuffer;
VkResult result;
Sint32 i;
result = renderer->vkDeviceWaitIdle(renderer->logicalDevice);
if (result != VK_SUCCESS)
{
LogVulkanResultAsError("vkDeviceWaitIdle", result);
return;
}
SDL_LockMutex(renderer->submitLock);
for (i = renderer->submittedCommandBufferCount - 1; i >= 0; i -= 1)
{
commandBuffer = renderer->submittedCommandBuffers[i];
VULKAN_INTERNAL_CleanCommandBuffer(renderer, commandBuffer);
}
VULKAN_INTERNAL_PerformPendingDestroys(renderer);
SDL_UnlockMutex(renderer->submitLock);
}
static Refresh_Fence* VULKAN_SubmitAndAcquireFence(
Refresh_CommandBuffer *commandBuffer
) {
VulkanCommandBuffer *vulkanCommandBuffer;
vulkanCommandBuffer = (VulkanCommandBuffer*) commandBuffer;
vulkanCommandBuffer->autoReleaseFence = 0;
VULKAN_Submit(commandBuffer);
return (Refresh_Fence*) vulkanCommandBuffer->inFlightFence;
}
static void VULKAN_Submit(
Refresh_CommandBuffer *commandBuffer
) {
VulkanCommandBuffer *vulkanCommandBuffer = (VulkanCommandBuffer*) commandBuffer;
VulkanRenderer *renderer = (VulkanRenderer*) vulkanCommandBuffer->renderer;
VkSubmitInfo submitInfo;
VkPresentInfoKHR presentInfo;
VulkanPresentData *presentData;
VkResult vulkanResult, presentResult = VK_SUCCESS;
VkPipelineStageFlags waitStages[MAX_PRESENT_COUNT];
Uint32 swapchainImageIndex;
VulkanTextureSlice *swapchainTextureSlice;
Uint8 commandBufferCleaned = 0;
VulkanMemorySubAllocator *allocator;
SDL_bool presenting = SDL_FALSE;
Sint32 i, j;
SDL_LockMutex(renderer->submitLock);
/* FIXME: Can this just be permanent? */
for (i = 0; i < MAX_PRESENT_COUNT; i += 1)
{
waitStages[i] = VK_PIPELINE_STAGE_COLOR_ATTACHMENT_OUTPUT_BIT;
}
for (j = 0; j < vulkanCommandBuffer->presentDataCount; j += 1)
{
swapchainImageIndex = vulkanCommandBuffer->presentDatas[j].swapchainImageIndex;
swapchainTextureSlice = VULKAN_INTERNAL_FetchTextureSlice(
vulkanCommandBuffer->presentDatas[j].windowData->swapchainData->textureContainers[swapchainImageIndex].activeTextureHandle->vulkanTexture,
0,
0
);
VULKAN_INTERNAL_TextureTransitionFromDefaultUsage(
renderer,
vulkanCommandBuffer,
VULKAN_TEXTURE_USAGE_MODE_PRESENT,
swapchainTextureSlice
);
}
VULKAN_INTERNAL_EndCommandBuffer(renderer, vulkanCommandBuffer);
vulkanCommandBuffer->inFlightFence = VULKAN_INTERNAL_AcquireFenceFromPool(renderer);
/* Command buffer has a reference to the in-flight fence */
(void)SDL_AtomicIncRef(&vulkanCommandBuffer->inFlightFence->referenceCount);
submitInfo.sType = VK_STRUCTURE_TYPE_SUBMIT_INFO;
submitInfo.pNext = NULL;
submitInfo.commandBufferCount = 1;
submitInfo.pCommandBuffers = &vulkanCommandBuffer->commandBuffer;
submitInfo.pWaitDstStageMask = waitStages;
submitInfo.pWaitSemaphores = vulkanCommandBuffer->waitSemaphores;
submitInfo.waitSemaphoreCount = vulkanCommandBuffer->waitSemaphoreCount;
submitInfo.pSignalSemaphores = vulkanCommandBuffer->signalSemaphores;
submitInfo.signalSemaphoreCount = vulkanCommandBuffer->signalSemaphoreCount;
vulkanResult = renderer->vkQueueSubmit(
renderer->unifiedQueue,
1,
&submitInfo,
vulkanCommandBuffer->inFlightFence->fence
);
if (vulkanResult != VK_SUCCESS)
{
LogVulkanResultAsError("vkQueueSubmit", vulkanResult);
}
/* Mark command buffers as submitted */
if (renderer->submittedCommandBufferCount + 1 >= renderer->submittedCommandBufferCapacity)
{
renderer->submittedCommandBufferCapacity = renderer->submittedCommandBufferCount + 1;
renderer->submittedCommandBuffers = SDL_realloc(
renderer->submittedCommandBuffers,
sizeof(VulkanCommandBuffer*) * renderer->submittedCommandBufferCapacity
);
}
renderer->submittedCommandBuffers[renderer->submittedCommandBufferCount] = vulkanCommandBuffer;
renderer->submittedCommandBufferCount += 1;
/* Present, if applicable */
for (j = 0; j < vulkanCommandBuffer->presentDataCount; j += 1)
{
presenting = SDL_TRUE;
presentData = &vulkanCommandBuffer->presentDatas[j];
presentInfo.sType = VK_STRUCTURE_TYPE_PRESENT_INFO_KHR;
presentInfo.pNext = NULL;
presentInfo.pWaitSemaphores =
&presentData->windowData->swapchainData->renderFinishedSemaphore[
presentData->windowData->swapchainData->frameCounter
];
presentInfo.waitSemaphoreCount = 1;
presentInfo.pSwapchains = &presentData->windowData->swapchainData->swapchain;
presentInfo.swapchainCount = 1;
presentInfo.pImageIndices = &presentData->swapchainImageIndex;
presentInfo.pResults = NULL;
presentResult = renderer->vkQueuePresentKHR(
renderer->unifiedQueue,
&presentInfo
);
presentData->windowData->swapchainData->frameCounter =
(presentData->windowData->swapchainData->frameCounter + 1) % MAX_FRAMES_IN_FLIGHT;
if (presentResult != VK_SUCCESS)
{
VULKAN_INTERNAL_RecreateSwapchain(
renderer,
presentData->windowData
);
}
else
{
/* If presenting, the swapchain is using the in-flight fence */
presentData->windowData->swapchainData->inFlightFences[
presentData->windowData->swapchainData->frameCounter
] = vulkanCommandBuffer->inFlightFence;
(void)SDL_AtomicIncRef(&vulkanCommandBuffer->inFlightFence->referenceCount);
}
}
/* Check if we can perform any cleanups */
for (i = renderer->submittedCommandBufferCount - 1; i >= 0; i -= 1)
{
vulkanResult = renderer->vkGetFenceStatus(
renderer->logicalDevice,
renderer->submittedCommandBuffers[i]->inFlightFence->fence
);
if (vulkanResult == VK_SUCCESS)
{
VULKAN_INTERNAL_CleanCommandBuffer(
renderer,
renderer->submittedCommandBuffers[i]
);
commandBufferCleaned = 1;
}
}
if (commandBufferCleaned)
{
SDL_LockMutex(renderer->allocatorLock);
for (i = 0; i < VK_MAX_MEMORY_TYPES; i += 1)
{
allocator = &renderer->memoryAllocator->subAllocators[i];
for (j = allocator->allocationCount - 1; j >= 0; j -= 1)
{
if (allocator->allocations[j]->usedRegionCount == 0)
{
VULKAN_INTERNAL_DeallocateMemory(
renderer,
allocator,
j
);
}
}
}
SDL_UnlockMutex(renderer->allocatorLock);
}
/* Check pending destroys */
VULKAN_INTERNAL_PerformPendingDestroys(renderer);
/* Defrag! */
if (
presenting &&
renderer->allocationsToDefragCount > 0 &&
!renderer->defragInProgress
) {
VULKAN_INTERNAL_DefragmentMemory(renderer);
}
SDL_UnlockMutex(renderer->submitLock);
}
static Uint8 VULKAN_INTERNAL_DefragmentMemory(
VulkanRenderer *renderer
) {
VulkanMemoryAllocation *allocation;
VulkanMemoryUsedRegion *currentRegion;
VulkanBuffer* newBuffer;
VulkanTexture* newTexture;
VkBufferCopy bufferCopy;
VkImageCopy imageCopy;
VulkanCommandBuffer *commandBuffer;
VulkanTextureSlice *srcSlice;
VulkanTextureSlice *dstSlice;
Uint32 i, sliceIndex;
SDL_LockMutex(renderer->allocatorLock);
renderer->defragInProgress = 1;
commandBuffer = (VulkanCommandBuffer*) VULKAN_AcquireCommandBuffer((Refresh_Renderer *) renderer);
commandBuffer->isDefrag = 1;
allocation = renderer->allocationsToDefrag[renderer->allocationsToDefragCount - 1];
renderer->allocationsToDefragCount -= 1;
/* For each used region in the allocation
* create a new resource, copy the data
* and re-point the resource containers
*/
for (i = 0; i < allocation->usedRegionCount; i += 1)
{
currentRegion = allocation->usedRegions[i];
if (currentRegion->isBuffer && !currentRegion->vulkanBuffer->markedForDestroy)
{
currentRegion->vulkanBuffer->usageFlags |= VK_BUFFER_USAGE_TRANSFER_DST_BIT;
newBuffer = VULKAN_INTERNAL_CreateBuffer(
renderer,
currentRegion->vulkanBuffer->size,
currentRegion->vulkanBuffer->usageFlags,
currentRegion->vulkanBuffer->type
);
if (newBuffer == NULL)
{
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Failed to create defrag buffer!");
return 0;
}
if (
renderer->debugMode &&
renderer->supportsDebugUtils &&
currentRegion->vulkanBuffer->handle != NULL &&
currentRegion->vulkanBuffer->handle->container != NULL &&
currentRegion->vulkanBuffer->handle->container->debugName != NULL
) {
VULKAN_INTERNAL_SetBufferName(
renderer,
newBuffer,
currentRegion->vulkanBuffer->handle->container->debugName
);
}
/* Copy buffer contents if necessary */
if (
currentRegion->vulkanBuffer->type == VULKAN_BUFFER_TYPE_GPU && currentRegion->vulkanBuffer->transitioned
) {
VULKAN_INTERNAL_BufferTransitionFromDefaultUsage(
renderer,
commandBuffer,
VULKAN_BUFFER_USAGE_MODE_COPY_SOURCE,
currentRegion->vulkanBuffer
);
VULKAN_INTERNAL_BufferTransitionFromDefaultUsage(
renderer,
commandBuffer,
VULKAN_BUFFER_USAGE_MODE_COPY_DESTINATION,
newBuffer
);
bufferCopy.srcOffset = 0;
bufferCopy.dstOffset = 0;
bufferCopy.size = currentRegion->resourceSize;
renderer->vkCmdCopyBuffer(
commandBuffer->commandBuffer,
currentRegion->vulkanBuffer->buffer,
newBuffer->buffer,
1,
&bufferCopy
);
VULKAN_INTERNAL_BufferTransitionToDefaultUsage(
renderer,
commandBuffer,
VULKAN_BUFFER_USAGE_MODE_COPY_DESTINATION,
newBuffer
);
VULKAN_INTERNAL_TrackBuffer(renderer, commandBuffer, currentRegion->vulkanBuffer);
VULKAN_INTERNAL_TrackBuffer(renderer, commandBuffer, newBuffer);
}
/* re-point original container to new buffer */
if (currentRegion->vulkanBuffer->handle != NULL)
{
newBuffer->handle = currentRegion->vulkanBuffer->handle;
newBuffer->handle->vulkanBuffer = newBuffer;
currentRegion->vulkanBuffer->handle = NULL;
}
VULKAN_INTERNAL_ReleaseBuffer(renderer, currentRegion->vulkanBuffer);
}
else if (!currentRegion->isBuffer && !currentRegion->vulkanTexture->markedForDestroy)
{
newTexture = VULKAN_INTERNAL_CreateTexture(
renderer,
currentRegion->vulkanTexture->dimensions.width,
currentRegion->vulkanTexture->dimensions.height,
currentRegion->vulkanTexture->depth,
currentRegion->vulkanTexture->isCube,
currentRegion->vulkanTexture->layerCount,
currentRegion->vulkanTexture->levelCount,
currentRegion->vulkanTexture->sampleCount,
currentRegion->vulkanTexture->format,
currentRegion->vulkanTexture->swizzle,
currentRegion->vulkanTexture->aspectFlags,
currentRegion->vulkanTexture->usageFlags
);
if (newTexture == NULL)
{
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Failed to create defrag texture!");
return 0;
}
for (sliceIndex = 0; sliceIndex < currentRegion->vulkanTexture->sliceCount; sliceIndex += 1)
{
/* copy slice if necessary */
srcSlice = &currentRegion->vulkanTexture->slices[sliceIndex];
dstSlice = &newTexture->slices[sliceIndex];
/* Set debug name if it exists */
if (
renderer->debugMode &&
renderer->supportsDebugUtils &&
srcSlice->parent->handle != NULL &&
srcSlice->parent->handle->container != NULL &&
srcSlice->parent->handle->container->debugName != NULL
) {
VULKAN_INTERNAL_SetTextureName(
renderer,
currentRegion->vulkanTexture,
srcSlice->parent->handle->container->debugName
);
}
if (srcSlice->transitioned)
{
VULKAN_INTERNAL_TextureTransitionFromDefaultUsage(
renderer,
commandBuffer,
VULKAN_TEXTURE_USAGE_MODE_COPY_SOURCE,
srcSlice
);
VULKAN_INTERNAL_TextureTransitionFromDefaultUsage(
renderer,
commandBuffer,
VULKAN_TEXTURE_USAGE_MODE_COPY_DESTINATION,
dstSlice
);
imageCopy.srcOffset.x = 0;
imageCopy.srcOffset.y = 0;
imageCopy.srcOffset.z = 0;
imageCopy.srcSubresource.aspectMask = srcSlice->parent->aspectFlags;
imageCopy.srcSubresource.baseArrayLayer = srcSlice->layer;
imageCopy.srcSubresource.layerCount = 1;
imageCopy.srcSubresource.mipLevel = srcSlice->level;
imageCopy.extent.width = SDL_max(1, srcSlice->parent->dimensions.width >> srcSlice->level);
imageCopy.extent.height = SDL_max(1, srcSlice->parent->dimensions.height >> srcSlice->level);
imageCopy.extent.depth = srcSlice->parent->depth;
imageCopy.dstOffset.x = 0;
imageCopy.dstOffset.y = 0;
imageCopy.dstOffset.z = 0;
imageCopy.dstSubresource.aspectMask = dstSlice->parent->aspectFlags;
imageCopy.dstSubresource.baseArrayLayer = dstSlice->layer;
imageCopy.dstSubresource.layerCount = 1;
imageCopy.dstSubresource.mipLevel = dstSlice->level;
renderer->vkCmdCopyImage(
commandBuffer->commandBuffer,
currentRegion->vulkanTexture->image,
VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL,
newTexture->image,
VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL,
1,
&imageCopy
);
VULKAN_INTERNAL_TextureTransitionToDefaultUsage(
renderer,
commandBuffer,
VULKAN_TEXTURE_USAGE_MODE_COPY_DESTINATION,
dstSlice
);
VULKAN_INTERNAL_TrackTextureSlice(renderer, commandBuffer, srcSlice);
VULKAN_INTERNAL_TrackTextureSlice(renderer, commandBuffer, dstSlice);
}
}
/* re-point original container to new texture */
newTexture->handle = currentRegion->vulkanTexture->handle;
newTexture->handle->vulkanTexture = newTexture;
currentRegion->vulkanTexture->handle = NULL;
VULKAN_INTERNAL_ReleaseTexture(renderer, currentRegion->vulkanTexture);
}
}
SDL_UnlockMutex(renderer->allocatorLock);
VULKAN_Submit(
(Refresh_CommandBuffer*) commandBuffer
);
return 1;
}
/* Queries */
static Refresh_OcclusionQuery* VULKAN_CreateOcclusionQuery(
Refresh_Renderer *driverData
) {
VulkanRenderer *renderer = (VulkanRenderer*) driverData;
VulkanOcclusionQuery *query = (VulkanOcclusionQuery*) SDL_malloc(sizeof(VulkanOcclusionQuery));
SDL_LockMutex(renderer->queryLock);
if (renderer->freeQueryIndexStackHead == -1)
{
SDL_LogError(
SDL_LOG_CATEGORY_APPLICATION,
"Query limit of %d has been exceeded!",
MAX_QUERIES
);
return NULL;
}
query->index = (Uint32) renderer->freeQueryIndexStackHead;
renderer->freeQueryIndexStackHead = renderer->freeQueryIndexStack[renderer->freeQueryIndexStackHead];
SDL_UnlockMutex(renderer->queryLock);
return (Refresh_OcclusionQuery*) query;
}
static void VULKAN_OcclusionQueryBegin(
Refresh_CommandBuffer *commandBuffer,
Refresh_OcclusionQuery *query
) {
VulkanCommandBuffer *vulkanCommandBuffer = (VulkanCommandBuffer*) commandBuffer;
VulkanRenderer *renderer = (VulkanRenderer*) vulkanCommandBuffer->renderer;
VulkanOcclusionQuery *vulkanQuery = (VulkanOcclusionQuery*) query;
renderer->vkCmdResetQueryPool(
vulkanCommandBuffer->commandBuffer,
renderer->queryPool,
vulkanQuery->index,
1
);
renderer->vkCmdBeginQuery(
vulkanCommandBuffer->commandBuffer,
renderer->queryPool,
vulkanQuery->index,
renderer->supportsPreciseOcclusionQueries ?
VK_QUERY_CONTROL_PRECISE_BIT :
0
);
}
static void VULKAN_OcclusionQueryEnd(
Refresh_CommandBuffer *commandBuffer,
Refresh_OcclusionQuery *query
) {
VulkanCommandBuffer *vulkanCommandBuffer = (VulkanCommandBuffer*) commandBuffer;
VulkanRenderer *renderer = (VulkanRenderer*) vulkanCommandBuffer->renderer;
VulkanOcclusionQuery *vulkanQuery = (VulkanOcclusionQuery*) query;
renderer->vkCmdEndQuery(
vulkanCommandBuffer->commandBuffer,
renderer->queryPool,
vulkanQuery->index
);
}
static SDL_bool VULKAN_OcclusionQueryPixelCount(
Refresh_Renderer *driverData,
Refresh_OcclusionQuery *query,
Uint32 *pixelCount
) {
VulkanRenderer *renderer = (VulkanRenderer*) driverData;
VulkanOcclusionQuery *vulkanQuery = (VulkanOcclusionQuery*) query;
VkResult vulkanResult;
Uint32 queryResult;
SDL_LockMutex(renderer->queryLock);
vulkanResult = renderer->vkGetQueryPoolResults(
renderer->logicalDevice,
renderer->queryPool,
vulkanQuery->index,
1,
sizeof(queryResult),
&queryResult,
0,
0
);
SDL_UnlockMutex(renderer->queryLock);
*pixelCount = queryResult;
return vulkanResult == VK_SUCCESS;
}
/* Format Info */
static SDL_bool VULKAN_IsTextureFormatSupported(
Refresh_Renderer *driverData,
Refresh_TextureFormat format,
Refresh_TextureType type,
Refresh_TextureUsageFlags usage
) {
VulkanRenderer *renderer = (VulkanRenderer*) driverData;
VkFormat vulkanFormat = SDLToVK_SurfaceFormat[format];
VkImageUsageFlags vulkanUsage = 0;
VkImageCreateFlags createFlags = 0;
VkImageFormatProperties properties;
VkResult vulkanResult;
if (usage & REFRESH_TEXTUREUSAGE_SAMPLER_BIT)
{
vulkanUsage |= VK_IMAGE_USAGE_SAMPLED_BIT;
}
if (usage & REFRESH_TEXTUREUSAGE_COLOR_TARGET_BIT)
{
vulkanUsage |= VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT;
}
if (usage & REFRESH_TEXTUREUSAGE_DEPTH_STENCIL_TARGET_BIT)
{
vulkanUsage |= VK_IMAGE_USAGE_DEPTH_STENCIL_ATTACHMENT_BIT;
}
if (usage & (
REFRESH_TEXTUREUSAGE_GRAPHICS_STORAGE_READ_BIT |
REFRESH_TEXTUREUSAGE_COMPUTE_STORAGE_READ_BIT |
REFRESH_TEXTUREUSAGE_COMPUTE_STORAGE_WRITE_BIT
)) {
vulkanUsage |= VK_IMAGE_USAGE_STORAGE_BIT;
}
if (type == REFRESH_TEXTURETYPE_CUBE)
{
createFlags = VK_IMAGE_CREATE_CUBE_COMPATIBLE_BIT;
}
vulkanResult = renderer->vkGetPhysicalDeviceImageFormatProperties(
renderer->physicalDevice,
vulkanFormat,
(type == REFRESH_TEXTURETYPE_3D) ? VK_IMAGE_TYPE_3D : VK_IMAGE_TYPE_2D,
VK_IMAGE_TILING_OPTIMAL,
vulkanUsage,
createFlags,
&properties
);
return vulkanResult == VK_SUCCESS;
}
static Refresh_SampleCount VULKAN_GetBestSampleCount(
Refresh_Renderer *driverData,
Refresh_TextureFormat format,
Refresh_SampleCount desiredSampleCount
) {
VulkanRenderer *renderer = (VulkanRenderer*) driverData;
Uint32 maxSupported;
VkSampleCountFlagBits bits = IsDepthFormat(format) ?
renderer->physicalDeviceProperties.properties.limits.framebufferDepthSampleCounts :
renderer->physicalDeviceProperties.properties.limits.framebufferColorSampleCounts;
if (bits & VK_SAMPLE_COUNT_8_BIT)
{
maxSupported = REFRESH_SAMPLECOUNT_8;
}
else if (bits & VK_SAMPLE_COUNT_4_BIT)
{
maxSupported = REFRESH_SAMPLECOUNT_4;
}
else if (bits & VK_SAMPLE_COUNT_2_BIT)
{
maxSupported = REFRESH_SAMPLECOUNT_2;
}
else
{
maxSupported = REFRESH_SAMPLECOUNT_1;
}
return (Refresh_SampleCount) SDL_min(maxSupported, desiredSampleCount);
}
/* SPIR-V Cross Interop */
static Refresh_Shader* VULKAN_CompileFromSPIRVCross(
Refresh_Renderer *driverData,
Refresh_ShaderStage shader_stage,
const char *entryPointName,
const char *source
) {
SDL_assert(!"You should not have gotten here!!!");
return NULL;
}
/* Device instantiation */
static inline Uint8 CheckDeviceExtensions(
VkExtensionProperties *extensions,
Uint32 numExtensions,
VulkanExtensions *supports
) {
Uint32 i;
SDL_memset(supports, '\0', sizeof(VulkanExtensions));
for (i = 0; i < numExtensions; i += 1)
{
const char *name = extensions[i].extensionName;
#define CHECK(ext) \
if (SDL_strcmp(name, "VK_" #ext) == 0) \
{ \
supports->ext = 1; \
}
CHECK(KHR_swapchain)
else CHECK(KHR_maintenance1)
else CHECK(KHR_get_memory_requirements2)
else CHECK(KHR_driver_properties)
else CHECK(EXT_vertex_attribute_divisor)
else CHECK(KHR_portability_subset)
#undef CHECK
}
return ( supports->KHR_swapchain &&
supports->KHR_maintenance1 &&
supports->KHR_get_memory_requirements2 );
}
static inline Uint32 GetDeviceExtensionCount(VulkanExtensions *supports)
{
return (
supports->KHR_swapchain +
supports->KHR_maintenance1 +
supports->KHR_get_memory_requirements2 +
supports->KHR_driver_properties +
supports->EXT_vertex_attribute_divisor +
supports->KHR_portability_subset
);
}
static inline void CreateDeviceExtensionArray(
VulkanExtensions *supports,
const char **extensions
) {
Uint8 cur = 0;
#define CHECK(ext) \
if (supports->ext) \
{ \
extensions[cur++] = "VK_" #ext; \
}
CHECK(KHR_swapchain)
CHECK(KHR_maintenance1)
CHECK(KHR_get_memory_requirements2)
CHECK(KHR_driver_properties)
CHECK(EXT_vertex_attribute_divisor)
CHECK(KHR_portability_subset)
#undef CHECK
}
static inline Uint8 SupportsInstanceExtension(
const char *ext,
VkExtensionProperties *availableExtensions,
Uint32 numAvailableExtensions
) {
Uint32 i;
for (i = 0; i < numAvailableExtensions; i += 1)
{
if (SDL_strcmp(ext, availableExtensions[i].extensionName) == 0)
{
return 1;
}
}
return 0;
}
static Uint8 VULKAN_INTERNAL_CheckInstanceExtensions(
const char **requiredExtensions,
Uint32 requiredExtensionsLength,
Uint8 *supportsDebugUtils
) {
Uint32 extensionCount, i;
VkExtensionProperties *availableExtensions;
Uint8 allExtensionsSupported = 1;
vkEnumerateInstanceExtensionProperties(
NULL,
&extensionCount,
NULL
);
availableExtensions = SDL_malloc(
extensionCount * sizeof(VkExtensionProperties)
);
vkEnumerateInstanceExtensionProperties(
NULL,
&extensionCount,
availableExtensions
);
for (i = 0; i < requiredExtensionsLength; i += 1)
{
if (!SupportsInstanceExtension(
requiredExtensions[i],
availableExtensions,
extensionCount
)) {
allExtensionsSupported = 0;
break;
}
}
/* This is optional, but nice to have! */
*supportsDebugUtils = SupportsInstanceExtension(
VK_EXT_DEBUG_UTILS_EXTENSION_NAME,
availableExtensions,
extensionCount
);
SDL_free(availableExtensions);
return allExtensionsSupported;
}
static Uint8 VULKAN_INTERNAL_CheckDeviceExtensions(
VulkanRenderer *renderer,
VkPhysicalDevice physicalDevice,
VulkanExtensions *physicalDeviceExtensions
) {
Uint32 extensionCount;
VkExtensionProperties *availableExtensions;
Uint8 allExtensionsSupported;
renderer->vkEnumerateDeviceExtensionProperties(
physicalDevice,
NULL,
&extensionCount,
NULL
);
availableExtensions = (VkExtensionProperties*) SDL_malloc(
extensionCount * sizeof(VkExtensionProperties)
);
renderer->vkEnumerateDeviceExtensionProperties(
physicalDevice,
NULL,
&extensionCount,
availableExtensions
);
allExtensionsSupported = CheckDeviceExtensions(
availableExtensions,
extensionCount,
physicalDeviceExtensions
);
SDL_free(availableExtensions);
return allExtensionsSupported;
}
static Uint8 VULKAN_INTERNAL_CheckValidationLayers(
const char** validationLayers,
Uint32 validationLayersLength
) {
Uint32 layerCount;
VkLayerProperties *availableLayers;
Uint32 i, j;
Uint8 layerFound = 0;
vkEnumerateInstanceLayerProperties(&layerCount, NULL);
availableLayers = (VkLayerProperties*) SDL_malloc(
layerCount * sizeof(VkLayerProperties)
);
vkEnumerateInstanceLayerProperties(&layerCount, availableLayers);
for (i = 0; i < validationLayersLength; i += 1)
{
layerFound = 0;
for (j = 0; j < layerCount; j += 1)
{
if (SDL_strcmp(validationLayers[i], availableLayers[j].layerName) == 0)
{
layerFound = 1;
break;
}
}
if (!layerFound)
{
break;
}
}
SDL_free(availableLayers);
return layerFound;
}
static Uint8 VULKAN_INTERNAL_CreateInstance(
VulkanRenderer *renderer,
SDL_Window *deviceWindowHandle
) {
VkResult vulkanResult;
VkApplicationInfo appInfo;
VkInstanceCreateFlags createFlags;
const char **instanceExtensionNames;
Uint32 instanceExtensionCount;
VkInstanceCreateInfo createInfo;
static const char *layerNames[] = { "VK_LAYER_KHRONOS_validation" };
appInfo.sType = VK_STRUCTURE_TYPE_APPLICATION_INFO;
appInfo.pNext = NULL;
appInfo.pApplicationName = NULL;
appInfo.applicationVersion = 0;
appInfo.pEngineName = "SDLGPU";
appInfo.engineVersion = ( (2 * 100 * 100) + (0 * 100) + (0) );
appInfo.apiVersion = VK_MAKE_VERSION(1, 0, 0);
createFlags = 0;
if (!SDL_Vulkan_GetInstanceExtensions(
(SDL_Window*) deviceWindowHandle,
&instanceExtensionCount,
NULL
)) {
SDL_LogError(
SDL_LOG_CATEGORY_APPLICATION,
"SDL_Vulkan_GetInstanceExtensions(): getExtensionCount: %s",
SDL_GetError()
);
return 0;
}
/* Extra space for the following extensions:
* VK_KHR_get_physical_device_properties2
* VK_EXT_swapchain_colorspace
* VK_EXT_debug_utils
* VK_KHR_portability_enumeration
*/
instanceExtensionNames = SDL_stack_alloc(
const char*,
instanceExtensionCount + 4
);
if (!SDL_Vulkan_GetInstanceExtensions(
deviceWindowHandle,
&instanceExtensionCount,
instanceExtensionNames
)) {
SDL_LogError(
SDL_LOG_CATEGORY_APPLICATION,
"SDL_Vulkan_GetInstanceExtensions(): %s",
SDL_GetError()
);
}
/* Core since 1.1 */
instanceExtensionNames[instanceExtensionCount++] =
VK_KHR_GET_PHYSICAL_DEVICE_PROPERTIES_2_EXTENSION_NAME;
instanceExtensionNames[instanceExtensionCount++] =
VK_EXT_SWAPCHAIN_COLOR_SPACE_EXTENSION_NAME;
#ifdef SDL_PLATFORM_APPLE
instanceExtensionNames[instanceExtensionCount++] =
VK_KHR_PORTABILITY_ENUMERATION_EXTENSION_NAME;
createFlags |= VK_INSTANCE_CREATE_ENUMERATE_PORTABILITY_BIT_KHR;
#endif
if (!VULKAN_INTERNAL_CheckInstanceExtensions(
instanceExtensionNames,
instanceExtensionCount,
&renderer->supportsDebugUtils
)) {
SDL_LogError(
SDL_LOG_CATEGORY_APPLICATION,
"Required Vulkan instance extensions not supported"
);
SDL_stack_free((char*) instanceExtensionNames);
return 0;
}
if (renderer->supportsDebugUtils)
{
/* Append the debug extension to the end */
instanceExtensionNames[instanceExtensionCount++] =
VK_EXT_DEBUG_UTILS_EXTENSION_NAME;
}
else
{
SDL_LogWarn(
SDL_LOG_CATEGORY_APPLICATION,
"%s is not supported!",
VK_EXT_DEBUG_UTILS_EXTENSION_NAME
);
}
createInfo.sType = VK_STRUCTURE_TYPE_INSTANCE_CREATE_INFO;
createInfo.pNext = NULL;
createInfo.flags = createFlags;
createInfo.pApplicationInfo = &appInfo;
createInfo.ppEnabledLayerNames = layerNames;
createInfo.enabledExtensionCount = instanceExtensionCount;
createInfo.ppEnabledExtensionNames = instanceExtensionNames;
if (renderer->debugMode)
{
createInfo.enabledLayerCount = SDL_arraysize(layerNames);
if (!VULKAN_INTERNAL_CheckValidationLayers(
layerNames,
createInfo.enabledLayerCount
)) {
SDL_LogWarn(SDL_LOG_CATEGORY_APPLICATION, "Validation layers not found, continuing without validation");
createInfo.enabledLayerCount = 0;
}
else
{
SDL_LogInfo(SDL_LOG_CATEGORY_APPLICATION, "Validation layers enabled, expect debug level performance!");
}
}
else
{
createInfo.enabledLayerCount = 0;
}
vulkanResult = vkCreateInstance(&createInfo, NULL, &renderer->instance);
if (vulkanResult != VK_SUCCESS)
{
SDL_LogError(
SDL_LOG_CATEGORY_APPLICATION,
"vkCreateInstance failed: %s",
VkErrorMessages(vulkanResult)
);
SDL_stack_free((char*) instanceExtensionNames);
return 0;
}
SDL_stack_free((char*) instanceExtensionNames);
return 1;
}
static Uint8 VULKAN_INTERNAL_IsDeviceSuitable(
VulkanRenderer *renderer,
VkPhysicalDevice physicalDevice,
VulkanExtensions *physicalDeviceExtensions,
VkSurfaceKHR surface,
Uint32 *queueFamilyIndex,
Uint8 *deviceRank
) {
Uint32 queueFamilyCount, queueFamilyRank, queueFamilyBest;
SwapchainSupportDetails swapchainSupportDetails;
VkQueueFamilyProperties *queueProps;
VkBool32 supportsPresent;
Uint8 querySuccess;
VkPhysicalDeviceProperties deviceProperties;
Uint32 i;
/* Get the device rank before doing any checks, in case one fails.
* Note: If no dedicated device exists, one that supports our features
* would be fine
*/
renderer->vkGetPhysicalDeviceProperties(
physicalDevice,
&deviceProperties
);
if (*deviceRank < DEVICE_PRIORITY[deviceProperties.deviceType])
{
/* This device outranks the best device we've found so far!
* This includes a dedicated GPU that has less features than an
* integrated GPU, because this is a freak case that is almost
* never intentionally desired by the end user
*/
*deviceRank = DEVICE_PRIORITY[deviceProperties.deviceType];
}
else if (*deviceRank > DEVICE_PRIORITY[deviceProperties.deviceType])
{
/* Device is outranked by a previous device, don't even try to
* run a query and reset the rank to avoid overwrites
*/
*deviceRank = 0;
return 0;
}
if (!VULKAN_INTERNAL_CheckDeviceExtensions(
renderer,
physicalDevice,
physicalDeviceExtensions
)) {
return 0;
}
renderer->vkGetPhysicalDeviceQueueFamilyProperties(
physicalDevice,
&queueFamilyCount,
NULL
);
queueProps = (VkQueueFamilyProperties*) SDL_stack_alloc(
VkQueueFamilyProperties,
queueFamilyCount
);
renderer->vkGetPhysicalDeviceQueueFamilyProperties(
physicalDevice,
&queueFamilyCount,
queueProps
);
queueFamilyBest = 0;
*queueFamilyIndex = UINT32_MAX;
for (i = 0; i < queueFamilyCount; i += 1)
{
renderer->vkGetPhysicalDeviceSurfaceSupportKHR(
physicalDevice,
i,
surface,
&supportsPresent
);
if ( !supportsPresent ||
!(queueProps[i].queueFlags & VK_QUEUE_GRAPHICS_BIT) )
{
/* Not a graphics family, ignore. */
continue;
}
/* The queue family bitflags are kind of annoying.
*
* We of course need a graphics family, but we ideally want the
* _primary_ graphics family. The spec states that at least one
* graphics family must also be a compute family, so generally
* drivers make that the first one. But hey, maybe something
* genuinely can't do compute or something, and FNA doesn't
* need it, so we'll be open to a non-compute queue family.
*
* Additionally, it's common to see the primary queue family
* have the transfer bit set, which is great! But this is
* actually optional; it's impossible to NOT have transfers in
* graphics/compute but it _is_ possible for a graphics/compute
* family, even the primary one, to just decide not to set the
* bitflag. Admittedly, a driver may want to isolate transfer
* queues to a dedicated family so that queues made solely for
* transfers can have an optimized DMA queue.
*
* That, or the driver author got lazy and decided not to set
* the bit. Looking at you, Android.
*
* -flibit
*/
if (queueProps[i].queueFlags & VK_QUEUE_COMPUTE_BIT)
{
if (queueProps[i].queueFlags & VK_QUEUE_TRANSFER_BIT)
{
/* Has all attribs! */
queueFamilyRank = 3;
}
else
{
/* Probably has a DMA transfer queue family */
queueFamilyRank = 2;
}
}
else
{
/* Just a graphics family, probably has something better */
queueFamilyRank = 1;
}
if (queueFamilyRank > queueFamilyBest)
{
*queueFamilyIndex = i;
queueFamilyBest = queueFamilyRank;
}
}
SDL_stack_free(queueProps);
if (*queueFamilyIndex == UINT32_MAX)
{
/* Somehow no graphics queues existed. Compute-only device? */
return 0;
}
/* FIXME: Need better structure for checking vs storing support details */
querySuccess = VULKAN_INTERNAL_QuerySwapchainSupport(
renderer,
physicalDevice,
surface,
&swapchainSupportDetails
);
if (swapchainSupportDetails.formatsLength > 0)
{
SDL_free(swapchainSupportDetails.formats);
}
if (swapchainSupportDetails.presentModesLength > 0)
{
SDL_free(swapchainSupportDetails.presentModes);
}
return ( querySuccess &&
swapchainSupportDetails.formatsLength > 0 &&
swapchainSupportDetails.presentModesLength > 0 );
}
static Uint8 VULKAN_INTERNAL_DeterminePhysicalDevice(
VulkanRenderer *renderer,
VkSurfaceKHR surface
) {
VkResult vulkanResult;
VkPhysicalDevice *physicalDevices;
VulkanExtensions *physicalDeviceExtensions;
Uint32 physicalDeviceCount, i, suitableIndex;
Uint32 queueFamilyIndex, suitableQueueFamilyIndex;
Uint8 deviceRank, highestRank;
vulkanResult = renderer->vkEnumeratePhysicalDevices(
renderer->instance,
&physicalDeviceCount,
NULL
);
VULKAN_ERROR_CHECK(vulkanResult, vkEnumeratePhysicalDevices, 0)
if (physicalDeviceCount == 0)
{
SDL_LogWarn(SDL_LOG_CATEGORY_APPLICATION, "Failed to find any GPUs with Vulkan support");
return 0;
}
physicalDevices = SDL_stack_alloc(VkPhysicalDevice, physicalDeviceCount);
physicalDeviceExtensions = SDL_stack_alloc(VulkanExtensions, physicalDeviceCount);
vulkanResult = renderer->vkEnumeratePhysicalDevices(
renderer->instance,
&physicalDeviceCount,
physicalDevices
);
/* This should be impossible to hit, but from what I can tell this can
* be triggered not because the array is too small, but because there
* were drivers that turned out to be bogus, so this is the loader's way
* of telling us that the list is now smaller than expected :shrug:
*/
if (vulkanResult == VK_INCOMPLETE)
{
SDL_LogWarn(SDL_LOG_CATEGORY_APPLICATION, "vkEnumeratePhysicalDevices returned VK_INCOMPLETE, will keep trying anyway...");
vulkanResult = VK_SUCCESS;
}
if (vulkanResult != VK_SUCCESS)
{
SDL_LogWarn(
SDL_LOG_CATEGORY_APPLICATION,
"vkEnumeratePhysicalDevices failed: %s",
VkErrorMessages(vulkanResult)
);
SDL_stack_free(physicalDevices);
SDL_stack_free(physicalDeviceExtensions);
return 0;
}
/* Any suitable device will do, but we'd like the best */
suitableIndex = -1;
suitableQueueFamilyIndex = 0;
highestRank = 0;
for (i = 0; i < physicalDeviceCount; i += 1)
{
deviceRank = highestRank;
if (VULKAN_INTERNAL_IsDeviceSuitable(
renderer,
physicalDevices[i],
&physicalDeviceExtensions[i],
surface,
&queueFamilyIndex,
&deviceRank
)) {
/* Use this for rendering.
* Note that this may override a previous device that
* supports rendering, but shares the same device rank.
*/
suitableIndex = i;
suitableQueueFamilyIndex = queueFamilyIndex;
highestRank = deviceRank;
}
else if (deviceRank > highestRank)
{
/* In this case, we found a... "realer?" GPU,
* but it doesn't actually support our Vulkan.
* We should disqualify all devices below as a
* result, because if we don't we end up
* ignoring real hardware and risk using
* something like LLVMpipe instead!
* -flibit
*/
suitableIndex = -1;
highestRank = deviceRank;
}
}
if (suitableIndex != -1)
{
renderer->supports = physicalDeviceExtensions[suitableIndex];
renderer->physicalDevice = physicalDevices[suitableIndex];
renderer->queueFamilyIndex = suitableQueueFamilyIndex;
}
else
{
SDL_stack_free(physicalDevices);
SDL_stack_free(physicalDeviceExtensions);
return 0;
}
renderer->physicalDeviceProperties.sType =
VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_PROPERTIES_2;
if (renderer->supports.KHR_driver_properties)
{
renderer->physicalDeviceDriverProperties.sType =
VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_DRIVER_PROPERTIES_KHR;
renderer->physicalDeviceDriverProperties.pNext = NULL;
renderer->physicalDeviceProperties.pNext =
&renderer->physicalDeviceDriverProperties;
}
else
{
renderer->physicalDeviceProperties.pNext = NULL;
}
renderer->vkGetPhysicalDeviceProperties2KHR(
renderer->physicalDevice,
&renderer->physicalDeviceProperties
);
renderer->vkGetPhysicalDeviceMemoryProperties(
renderer->physicalDevice,
&renderer->memoryProperties
);
SDL_stack_free(physicalDevices);
SDL_stack_free(physicalDeviceExtensions);
return 1;
}
static Uint8 VULKAN_INTERNAL_CreateLogicalDevice(
VulkanRenderer *renderer
) {
VkResult vulkanResult;
VkDeviceCreateInfo deviceCreateInfo;
VkPhysicalDeviceFeatures deviceFeatures;
VkPhysicalDevicePortabilitySubsetFeaturesKHR portabilityFeatures;
const char **deviceExtensions;
VkDeviceQueueCreateInfo queueCreateInfo;
float queuePriority = 1.0f;
queueCreateInfo.sType =
VK_STRUCTURE_TYPE_DEVICE_QUEUE_CREATE_INFO;
queueCreateInfo.pNext = NULL;
queueCreateInfo.flags = 0;
queueCreateInfo.queueFamilyIndex = renderer->queueFamilyIndex;
queueCreateInfo.queueCount = 1;
queueCreateInfo.pQueuePriorities = &queuePriority;
/* specifying used device features */
SDL_zero(deviceFeatures);
deviceFeatures.fillModeNonSolid = VK_TRUE;
deviceFeatures.samplerAnisotropy = VK_TRUE;
deviceFeatures.multiDrawIndirect = VK_TRUE;
deviceFeatures.independentBlend = VK_TRUE;
/* creating the logical device */
deviceCreateInfo.sType = VK_STRUCTURE_TYPE_DEVICE_CREATE_INFO;
if (renderer->supports.KHR_portability_subset)
{
portabilityFeatures.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_PORTABILITY_SUBSET_FEATURES_KHR;
portabilityFeatures.pNext = NULL;
portabilityFeatures.constantAlphaColorBlendFactors = VK_FALSE;
portabilityFeatures.events = VK_FALSE;
portabilityFeatures.imageViewFormatReinterpretation = VK_FALSE;
portabilityFeatures.imageViewFormatSwizzle = VK_TRUE;
portabilityFeatures.imageView2DOn3DImage = VK_FALSE;
portabilityFeatures.multisampleArrayImage = VK_FALSE;
portabilityFeatures.mutableComparisonSamplers = VK_FALSE;
portabilityFeatures.pointPolygons = VK_FALSE;
portabilityFeatures.samplerMipLodBias = VK_FALSE; /* Technically should be true, but eh */
portabilityFeatures.separateStencilMaskRef = VK_FALSE;
portabilityFeatures.shaderSampleRateInterpolationFunctions = VK_FALSE;
portabilityFeatures.tessellationIsolines = VK_FALSE;
portabilityFeatures.tessellationPointMode = VK_FALSE;
portabilityFeatures.triangleFans = VK_FALSE;
portabilityFeatures.vertexAttributeAccessBeyondStride = VK_FALSE;
deviceCreateInfo.pNext = &portabilityFeatures;
}
else
{
deviceCreateInfo.pNext = NULL;
}
deviceCreateInfo.flags = 0;
deviceCreateInfo.queueCreateInfoCount = 1;
deviceCreateInfo.pQueueCreateInfos = &queueCreateInfo;
deviceCreateInfo.enabledLayerCount = 0;
deviceCreateInfo.ppEnabledLayerNames = NULL;
deviceCreateInfo.enabledExtensionCount = GetDeviceExtensionCount(
&renderer->supports
);
deviceExtensions = SDL_stack_alloc(
const char*,
deviceCreateInfo.enabledExtensionCount
);
CreateDeviceExtensionArray(&renderer->supports, deviceExtensions);
deviceCreateInfo.ppEnabledExtensionNames = deviceExtensions;
deviceCreateInfo.pEnabledFeatures = &deviceFeatures;
vulkanResult = renderer->vkCreateDevice(
renderer->physicalDevice,
&deviceCreateInfo,
NULL,
&renderer->logicalDevice
);
SDL_stack_free(deviceExtensions);
VULKAN_ERROR_CHECK(vulkanResult, vkCreateDevice, 0)
/* Load vkDevice entry points */
#define VULKAN_DEVICE_FUNCTION(ext, ret, func, params) \
renderer->func = (vkfntype_##func) \
renderer->vkGetDeviceProcAddr( \
renderer->logicalDevice, \
#func \
);
#include "Refresh_vulkan_vkfuncs.h"
renderer->vkGetDeviceQueue(
renderer->logicalDevice,
renderer->queueFamilyIndex,
0,
&renderer->unifiedQueue
);
return 1;
}
static void VULKAN_INTERNAL_LoadEntryPoints(void)
{
/* Required for MoltenVK support */
SDL_setenv("MVK_CONFIG_FULL_IMAGE_VIEW_SWIZZLE", "1", 1);
/* Load Vulkan entry points */
if (SDL_Vulkan_LoadLibrary(NULL) < 0)
{
SDL_LogWarn(SDL_LOG_CATEGORY_APPLICATION, "Vulkan: SDL_Vulkan_LoadLibrary failed!");
return;
}
#ifdef HAVE_GCC_DIAGNOSTIC_PRAGMA
#pragma GCC diagnostic push
#pragma GCC diagnostic ignored "-Wpedantic"
#endif
vkGetInstanceProcAddr = (PFN_vkGetInstanceProcAddr) SDL_Vulkan_GetVkGetInstanceProcAddr();
#ifdef HAVE_GCC_DIAGNOSTIC_PRAGMA
#pragma GCC diagnostic pop
#endif
if (vkGetInstanceProcAddr == NULL)
{
SDL_LogWarn(
SDL_LOG_CATEGORY_APPLICATION,
"SDL_Vulkan_GetVkGetInstanceProcAddr(): %s",
SDL_GetError()
);
return;
}
#define VULKAN_GLOBAL_FUNCTION(name) \
name = (PFN_##name) vkGetInstanceProcAddr(VK_NULL_HANDLE, #name); \
if (name == NULL) \
{ \
SDL_LogWarn(SDL_LOG_CATEGORY_APPLICATION, "vkGetInstanceProcAddr(VK_NULL_HANDLE, \"" #name "\") failed"); \
return; \
}
#include "Refresh_vulkan_vkfuncs.h"
}
static Uint8 VULKAN_INTERNAL_PrepareVulkan(
VulkanRenderer *renderer
) {
SDL_Window *dummyWindowHandle;
VkSurfaceKHR surface;
SwapchainSupportDetails swapchainSupportDetails;
VULKAN_INTERNAL_LoadEntryPoints();
dummyWindowHandle = SDL_CreateWindow(
"Refresh_ Vulkan",
0, 0,
128, 128,
SDL_WINDOW_VULKAN | SDL_WINDOW_HIDDEN
);
if (dummyWindowHandle == NULL)
{
SDL_LogWarn(SDL_LOG_CATEGORY_APPLICATION, "Vulkan: Could not create dummy window");
return 0;
}
if (!VULKAN_INTERNAL_CreateInstance(renderer, dummyWindowHandle))
{
SDL_DestroyWindow(dummyWindowHandle);
SDL_LogWarn(SDL_LOG_CATEGORY_APPLICATION, "Vulkan: Could not create Vulkan instance");
return 0;
}
if (!SDL_Vulkan_CreateSurface(
dummyWindowHandle,
renderer->instance,
&surface
)) {
SDL_DestroyWindow(dummyWindowHandle);
SDL_LogWarn(
SDL_LOG_CATEGORY_APPLICATION,
"SDL_Vulkan_CreateSurface failed: %s",
SDL_GetError()
);
return 0;
}
#define VULKAN_INSTANCE_FUNCTION(ext, ret, func, params) \
renderer->func = (vkfntype_##func) vkGetInstanceProcAddr(renderer->instance, #func);
#include "Refresh_vulkan_vkfuncs.h"
if (!VULKAN_INTERNAL_DeterminePhysicalDevice(renderer, surface))
{
SDL_LogWarn(SDL_LOG_CATEGORY_APPLICATION, "Vulkan: Failed to determine a suitable physical device");
renderer->vkDestroySurfaceKHR(
renderer->instance,
surface,
NULL
);
SDL_DestroyWindow(dummyWindowHandle);
return 0;
}
if (!VULKAN_INTERNAL_QuerySwapchainSupport(
renderer,
renderer->physicalDevice,
surface,
&swapchainSupportDetails
))
{
renderer->vkDestroySurfaceKHR(
renderer->instance,
surface,
NULL
);
SDL_DestroyWindow(dummyWindowHandle);
return 0;
}
SDL_free(swapchainSupportDetails.formats);
SDL_free(swapchainSupportDetails.presentModes);
renderer->vkDestroySurfaceKHR(
renderer->instance,
surface,
NULL
);
SDL_DestroyWindow(dummyWindowHandle);
return 1;
}
static SDL_bool VULKAN_PrepareDriver()
{
/* Set up dummy VulkanRenderer */
VulkanRenderer *renderer;
SDL_bool result;
if (SDL_Vulkan_LoadLibrary(NULL) < 0)
{
return 0;
}
renderer = (VulkanRenderer*) SDL_malloc(sizeof(VulkanRenderer));
SDL_memset(renderer, '\0', sizeof(VulkanRenderer));
result = VULKAN_INTERNAL_PrepareVulkan(renderer);
if (result)
{
renderer->vkDestroyInstance(renderer->instance, NULL);
}
SDL_free(renderer);
SDL_Vulkan_UnloadLibrary();
return result;
}
static Refresh_Device* VULKAN_CreateDevice(SDL_bool debugMode)
{
VulkanRenderer *renderer;
Refresh_Device *result;
VkResult vulkanResult;
Uint32 i;
/* Variables: Image Format Detection */
VkImageFormatProperties imageFormatProperties;
/* Variables: Query Pool Creation */
VkQueryPoolCreateInfo queryPoolCreateInfo;
/* Variables: Device Feature Checks */
VkPhysicalDeviceFeatures physicalDeviceFeatures;
if (SDL_Vulkan_LoadLibrary(NULL) < 0)
{
SDL_assert(!"This should have failed in PrepareDevice first!");
return NULL;
}
renderer = (VulkanRenderer*) SDL_malloc(sizeof(VulkanRenderer));
SDL_memset(renderer, '\0', sizeof(VulkanRenderer));
renderer->debugMode = debugMode;
if (!VULKAN_INTERNAL_PrepareVulkan(renderer))
{
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Failed to initialize Vulkan!");
SDL_free(renderer);
SDL_Vulkan_UnloadLibrary();
return NULL;
}
SDL_LogInfo(SDL_LOG_CATEGORY_APPLICATION, "Refresh_ Driver: Vulkan");
SDL_LogInfo(
SDL_LOG_CATEGORY_APPLICATION,
"Vulkan Device: %s",
renderer->physicalDeviceProperties.properties.deviceName
);
SDL_LogInfo(
SDL_LOG_CATEGORY_APPLICATION,
"Vulkan Driver: %s %s",
renderer->physicalDeviceDriverProperties.driverName,
renderer->physicalDeviceDriverProperties.driverInfo
);
SDL_LogInfo(
SDL_LOG_CATEGORY_APPLICATION,
"Vulkan Conformance: %u.%u.%u",
renderer->physicalDeviceDriverProperties.conformanceVersion.major,
renderer->physicalDeviceDriverProperties.conformanceVersion.minor,
renderer->physicalDeviceDriverProperties.conformanceVersion.patch
);
if (!VULKAN_INTERNAL_CreateLogicalDevice(
renderer
)) {
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION, "Failed to create logical device");
SDL_free(renderer);
SDL_Vulkan_UnloadLibrary();
return NULL;
}
/* FIXME: just move this into this function */
result = (Refresh_Device*) SDL_malloc(sizeof(Refresh_Device));
ASSIGN_DRIVER(VULKAN)
result->driverData = (Refresh_Renderer*) renderer;
/*
* Create initial swapchain array
*/
renderer->claimedWindowCapacity = 1;
renderer->claimedWindowCount = 0;
renderer->claimedWindows = SDL_malloc(
renderer->claimedWindowCapacity * sizeof(WindowData*)
);
/* Threading */
renderer->allocatorLock = SDL_CreateMutex();
renderer->disposeLock = SDL_CreateMutex();
renderer->submitLock = SDL_CreateMutex();
renderer->acquireCommandBufferLock = SDL_CreateMutex();
renderer->renderPassFetchLock = SDL_CreateMutex();
renderer->framebufferFetchLock = SDL_CreateMutex();
renderer->queryLock = SDL_CreateMutex();
/*
* Create submitted command buffer list
*/
renderer->submittedCommandBufferCapacity = 16;
renderer->submittedCommandBufferCount = 0;
renderer->submittedCommandBuffers = SDL_malloc(sizeof(VulkanCommandBuffer*) * renderer->submittedCommandBufferCapacity);
/* Memory Allocator */
renderer->memoryAllocator = (VulkanMemoryAllocator*) SDL_malloc(
sizeof(VulkanMemoryAllocator)
);
for (i = 0; i < VK_MAX_MEMORY_TYPES; i += 1)
{
renderer->memoryAllocator->subAllocators[i].memoryTypeIndex = i;
renderer->memoryAllocator->subAllocators[i].allocations = NULL;
renderer->memoryAllocator->subAllocators[i].allocationCount = 0;
renderer->memoryAllocator->subAllocators[i].sortedFreeRegions = SDL_malloc(
sizeof(VulkanMemoryFreeRegion*) * 4
);
renderer->memoryAllocator->subAllocators[i].sortedFreeRegionCount = 0;
renderer->memoryAllocator->subAllocators[i].sortedFreeRegionCapacity = 4;
}
/* UBO alignment */
renderer->minUBOAlignment = (Uint32) renderer->physicalDeviceProperties.properties.limits.minUniformBufferOffsetAlignment;
/* Initialize query pool */
queryPoolCreateInfo.sType = VK_STRUCTURE_TYPE_QUERY_POOL_CREATE_INFO;
queryPoolCreateInfo.pNext = NULL;
queryPoolCreateInfo.flags = 0;
queryPoolCreateInfo.queryType = VK_QUERY_TYPE_OCCLUSION;
queryPoolCreateInfo.queryCount = MAX_QUERIES;
queryPoolCreateInfo.pipelineStatistics = 0;
vulkanResult = renderer->vkCreateQueryPool(
renderer->logicalDevice,
&queryPoolCreateInfo,
NULL,
&renderer->queryPool
);
VULKAN_ERROR_CHECK(vulkanResult, vkCreateQueryPool, NULL)
for (i = 0; i < MAX_QUERIES - 1; i += 1)
{
renderer->freeQueryIndexStack[i] = i + 1;
}
renderer->freeQueryIndexStack[MAX_QUERIES - 1] = -1;
/* Initialize caches */
for (i = 0; i < NUM_COMMAND_POOL_BUCKETS; i += 1)
{
renderer->commandPoolHashTable.buckets[i].elements = NULL;
renderer->commandPoolHashTable.buckets[i].count = 0;
renderer->commandPoolHashTable.buckets[i].capacity = 0;
}
renderer->renderPassHashArray.elements = NULL;
renderer->renderPassHashArray.count = 0;
renderer->renderPassHashArray.capacity = 0;
renderer->framebufferHashArray.elements = NULL;
renderer->framebufferHashArray.count = 0;
renderer->framebufferHashArray.capacity = 0;
/* Initialize fence pool */
renderer->fencePool.lock = SDL_CreateMutex();
renderer->fencePool.availableFenceCapacity = 4;
renderer->fencePool.availableFenceCount = 0;
renderer->fencePool.availableFences = SDL_malloc(
renderer->fencePool.availableFenceCapacity * sizeof(VulkanFenceHandle*)
);
/* Some drivers don't support D16, so we have to fall back to D32. */
vulkanResult = renderer->vkGetPhysicalDeviceImageFormatProperties(
renderer->physicalDevice,
VK_FORMAT_D16_UNORM,
VK_IMAGE_TYPE_2D,
VK_IMAGE_TILING_OPTIMAL,
VK_IMAGE_ASPECT_DEPTH_BIT,
0,
&imageFormatProperties
);
if (vulkanResult == VK_ERROR_FORMAT_NOT_SUPPORTED)
{
renderer->D16Format = VK_FORMAT_D32_SFLOAT;
}
else
{
renderer->D16Format = VK_FORMAT_D16_UNORM;
}
vulkanResult = renderer->vkGetPhysicalDeviceImageFormatProperties(
renderer->physicalDevice,
VK_FORMAT_D16_UNORM_S8_UINT,
VK_IMAGE_TYPE_2D,
VK_IMAGE_TILING_OPTIMAL,
VK_IMAGE_ASPECT_DEPTH_BIT | VK_IMAGE_ASPECT_STENCIL_BIT,
0,
&imageFormatProperties
);
if (vulkanResult == VK_ERROR_FORMAT_NOT_SUPPORTED)
{
renderer->D16S8Format = VK_FORMAT_D32_SFLOAT_S8_UINT;
}
else
{
renderer->D16S8Format = VK_FORMAT_D16_UNORM_S8_UINT;
}
/* Deferred destroy storage */
renderer->texturesToDestroyCapacity = 16;
renderer->texturesToDestroyCount = 0;
renderer->texturesToDestroy = (VulkanTexture**)SDL_malloc(
sizeof(VulkanTexture*) *
renderer->texturesToDestroyCapacity
);
renderer->buffersToDestroyCapacity = 16;
renderer->buffersToDestroyCount = 0;
renderer->buffersToDestroy = SDL_malloc(
sizeof(VulkanBuffer*) *
renderer->buffersToDestroyCapacity
);
renderer->samplersToDestroyCapacity = 16;
renderer->samplersToDestroyCount = 0;
renderer->samplersToDestroy = SDL_malloc(
sizeof(VulkanSampler*) *
renderer->samplersToDestroyCapacity
);
renderer->graphicsPipelinesToDestroyCapacity = 16;
renderer->graphicsPipelinesToDestroyCount = 0;
renderer->graphicsPipelinesToDestroy = SDL_malloc(
sizeof(VulkanGraphicsPipeline*) *
renderer->graphicsPipelinesToDestroyCapacity
);
renderer->computePipelinesToDestroyCapacity = 16;
renderer->computePipelinesToDestroyCount = 0;
renderer->computePipelinesToDestroy = SDL_malloc(
sizeof(VulkanComputePipeline*) *
renderer->computePipelinesToDestroyCapacity
);
renderer->shadersToDestroyCapacity = 16;
renderer->shadersToDestroyCount = 0;
renderer->shadersToDestroy = SDL_malloc(
sizeof(VulkanShader*) *
renderer->shadersToDestroyCapacity
);
renderer->framebuffersToDestroyCapacity = 16;
renderer->framebuffersToDestroyCount = 0;
renderer->framebuffersToDestroy = SDL_malloc(
sizeof(VulkanFramebuffer*) *
renderer->framebuffersToDestroyCapacity
);
/* Defrag state */
renderer->defragInProgress = 0;
renderer->allocationsToDefragCount = 0;
renderer->allocationsToDefragCapacity = 4;
renderer->allocationsToDefrag = SDL_malloc(
renderer->allocationsToDefragCapacity * sizeof(VulkanMemoryAllocation*)
);
/* Support checks */
renderer->vkGetPhysicalDeviceFeatures(
renderer->physicalDevice,
&physicalDeviceFeatures
);
renderer->supportsPreciseOcclusionQueries = physicalDeviceFeatures.occlusionQueryPrecise;
return result;
}
Refresh_Driver VulkanDriver = {
"Vulkan",
REFRESH_BACKEND_VULKAN,
VULKAN_PrepareDriver,
VULKAN_CreateDevice
};
#endif /* REFRESH_VULKAN */