/* * Copyright 2016 Józef Kucia for CodeWeavers * * This library is free software; you can redistribute it and/or * modify it under the terms of the GNU Lesser General Public * License as published by the Free Software Foundation; either * version 2.1 of the License, or (at your option) any later version. * * This library is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Lesser General Public License for more details. * * You should have received a copy of the GNU Lesser General Public * License along with this library; if not, write to the Free Software * Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301, USA */ #ifndef __VKD3D_PRIVATE_H #define __VKD3D_PRIVATE_H #define COBJMACROS #define VK_NO_PROTOTYPES #include "vkd3d_common.h" #include "vkd3d_memory.h" #include "vkd3d_utf8.h" #include "hashmap.h" #include "list.h" #include "rbtree.h" #include "vkd3d.h" #include "vkd3d_build.h" #include "vkd3d_version.h" #include "vkd3d_shader.h" #include "vkd3d_threads.h" #include "vkd3d_platform.h" #include "vkd3d_swapchain_factory.h" #include "vkd3d_command_list_vkd3d_ext.h" #include "vkd3d_device_vkd3d_ext.h" #include "vkd3d_string.h" #include #include #include #include #define VK_CALL(f) (vk_procs->f) #define MAKE_MAGIC(a,b,c,d) (((uint32_t)a) | (((uint32_t)b) << 8) | (((uint32_t)c) << 16) | (((uint32_t)d) << 24)) #define VKD3D_MAX_COMPATIBLE_FORMAT_COUNT 6u #define VKD3D_MAX_SHADER_EXTENSIONS 2u #define VKD3D_MAX_SHADER_STAGES 5u #define VKD3D_MAX_VK_SYNC_OBJECTS 4u #define VKD3D_MAX_DESCRIPTOR_SETS 10u #define VKD3D_MAX_BINDLESS_DESCRIPTOR_SETS 8u #define VKD3D_MAX_MUTABLE_DESCRIPTOR_TYPES 6u #define VKD3D_PIPELINE_BIND_POINT_COUNT 2u #define VKD3D_TILE_SIZE 65536 struct d3d12_command_list; struct d3d12_device; struct d3d12_resource; struct vkd3d_bindless_set_info; struct vkd3d_dynamic_state; struct vkd3d_vk_global_procs { PFN_vkCreateInstance vkCreateInstance; PFN_vkEnumerateInstanceVersion vkEnumerateInstanceVersion; PFN_vkEnumerateInstanceExtensionProperties vkEnumerateInstanceExtensionProperties; PFN_vkEnumerateInstanceLayerProperties vkEnumerateInstanceLayerProperties; PFN_vkGetInstanceProcAddr vkGetInstanceProcAddr; }; #define DECLARE_VK_PFN(name) PFN_##name name; struct vkd3d_vk_instance_procs { #define VK_INSTANCE_PFN DECLARE_VK_PFN #define VK_INSTANCE_EXT_PFN DECLARE_VK_PFN #include "vulkan_procs.h" }; struct vkd3d_vk_device_procs { #define VK_INSTANCE_PFN DECLARE_VK_PFN #define VK_INSTANCE_EXT_PFN DECLARE_VK_PFN #define VK_DEVICE_PFN DECLARE_VK_PFN #define VK_DEVICE_EXT_PFN DECLARE_VK_PFN #include "vulkan_procs.h" }; #undef DECLARE_VK_PFN HRESULT hresult_from_errno(int rc); HRESULT hresult_from_vk_result(VkResult vr); HRESULT hresult_from_vkd3d_result(int vkd3d_result); struct vkd3d_vulkan_info { /* EXT instance extensions */ bool EXT_debug_utils; /* KHR device extensions */ bool KHR_buffer_device_address; bool KHR_draw_indirect_count; bool KHR_image_format_list; bool KHR_push_descriptor; bool KHR_timeline_semaphore; bool KHR_shader_float16_int8; bool KHR_shader_subgroup_extended_types; bool KHR_ray_tracing_pipeline; bool KHR_acceleration_structure; bool KHR_deferred_host_operations; bool KHR_pipeline_library; bool KHR_spirv_1_4; bool KHR_shader_float_controls; bool KHR_fragment_shading_rate; bool KHR_create_renderpass2; bool KHR_sampler_mirror_clamp_to_edge; bool KHR_separate_depth_stencil_layouts; /* EXT device extensions */ bool EXT_calibrated_timestamps; bool EXT_conditional_rendering; bool EXT_conservative_rasterization; bool EXT_custom_border_color; bool EXT_depth_clip_enable; bool EXT_descriptor_indexing; bool EXT_inline_uniform_block; bool EXT_robustness2; bool EXT_sampler_filter_minmax; bool EXT_shader_demote_to_helper_invocation; bool EXT_shader_stencil_export; bool EXT_shader_viewport_index_layer; bool EXT_subgroup_size_control; bool EXT_texel_buffer_alignment; bool EXT_transform_feedback; bool EXT_vertex_attribute_divisor; bool EXT_extended_dynamic_state; bool EXT_external_memory_host; bool EXT_4444_formats; /* AMD device extensions */ bool AMD_buffer_marker; bool AMD_shader_core_properties; bool AMD_shader_core_properties2; /* NV device extensions */ bool NV_shader_sm_builtins; bool NVX_binary_import; bool NVX_image_view_handle; /* VALVE extensions */ bool VALVE_mutable_descriptor_type; bool rasterization_stream; bool transform_feedback_queries; bool vertex_attrib_zero_divisor; unsigned int max_vertex_attrib_divisor; VkPhysicalDeviceLimits device_limits; VkPhysicalDeviceSparseProperties sparse_properties; VkPhysicalDeviceTexelBufferAlignmentPropertiesEXT texel_buffer_alignment_properties; unsigned int shader_extension_count; enum vkd3d_shader_target_extension shader_extensions[VKD3D_MAX_SHADER_EXTENSIONS]; }; struct vkd3d_instance { VkInstance vk_instance; uint32_t instance_version; struct vkd3d_vk_instance_procs vk_procs; PFN_vkd3d_signal_event signal_event; PFN_vkd3d_create_thread create_thread; PFN_vkd3d_join_thread join_thread; struct vkd3d_vulkan_info vk_info; struct vkd3d_vk_global_procs vk_global_procs; void *libvulkan; VkDebugUtilsMessengerEXT vk_debug_callback; LONG refcount; }; extern uint64_t vkd3d_config_flags; extern const struct vkd3d_shader_quirk_info *vkd3d_shader_quirk_info; union vkd3d_thread_handle { pthread_t pthread; void *handle; }; HRESULT vkd3d_create_thread(struct vkd3d_instance *instance, PFN_vkd3d_thread thread_main, void *data, union vkd3d_thread_handle *thread); HRESULT vkd3d_join_thread(struct vkd3d_instance *instance, union vkd3d_thread_handle *thread); struct vkd3d_waiting_fence { struct d3d12_fence *fence; uint64_t value; }; struct vkd3d_fence_worker { union vkd3d_thread_handle thread; pthread_mutex_t mutex; pthread_cond_t cond; bool should_exit; uint32_t enqueued_fence_count; struct vkd3d_waiting_fence *enqueued_fences; size_t enqueued_fences_size; struct d3d12_device *device; }; HRESULT vkd3d_fence_worker_start(struct vkd3d_fence_worker *worker, struct d3d12_device *device); HRESULT vkd3d_fence_worker_stop(struct vkd3d_fence_worker *worker, struct d3d12_device *device); #define VKD3D_VA_BLOCK_SIZE_BITS (20) #define VKD3D_VA_BLOCK_SIZE (1ull << VKD3D_VA_BLOCK_SIZE_BITS) #define VKD3D_VA_LO_MASK (VKD3D_VA_BLOCK_SIZE - 1) #define VKD3D_VA_BLOCK_BITS (20) #define VKD3D_VA_BLOCK_COUNT (1ull << VKD3D_VA_BLOCK_BITS) #define VKD3D_VA_BLOCK_MASK (VKD3D_VA_BLOCK_COUNT - 1) #define VKD3D_VA_NEXT_BITS (12) #define VKD3D_VA_NEXT_COUNT (1ull << VKD3D_VA_NEXT_BITS) #define VKD3D_VA_NEXT_MASK (VKD3D_VA_NEXT_COUNT - 1) struct vkd3d_unique_resource; struct vkd3d_va_entry { DECLSPEC_ALIGN(8) VkDeviceAddress va; const struct vkd3d_unique_resource *resource; }; struct vkd3d_va_block { struct vkd3d_va_entry l; struct vkd3d_va_entry r; }; struct vkd3d_va_tree { struct vkd3d_va_block blocks[VKD3D_VA_BLOCK_COUNT]; struct vkd3d_va_tree *next[VKD3D_VA_NEXT_COUNT]; }; struct vkd3d_va_range { VkDeviceAddress base; VkDeviceSize size; }; struct vkd3d_va_allocator { pthread_mutex_t mutex; struct vkd3d_va_range *free_ranges; size_t free_ranges_size; size_t free_range_count; VkDeviceAddress next_va; }; struct vkd3d_va_map { struct vkd3d_va_tree va_tree; struct vkd3d_va_allocator va_allocator; pthread_mutex_t mutex; struct vkd3d_unique_resource **small_entries; size_t small_entries_size; size_t small_entries_count; }; void vkd3d_va_map_insert(struct vkd3d_va_map *va_map, struct vkd3d_unique_resource *resource); void vkd3d_va_map_remove(struct vkd3d_va_map *va_map, const struct vkd3d_unique_resource *resource); const struct vkd3d_unique_resource *vkd3d_va_map_deref(struct vkd3d_va_map *va_map, VkDeviceAddress va); VkAccelerationStructureKHR vkd3d_va_map_place_acceleration_structure(struct vkd3d_va_map *va_map, struct d3d12_device *device, VkDeviceAddress va); VkDeviceAddress vkd3d_va_map_alloc_fake_va(struct vkd3d_va_map *va_map, VkDeviceSize size); void vkd3d_va_map_free_fake_va(struct vkd3d_va_map *va_map, VkDeviceAddress va, VkDeviceSize size); void vkd3d_va_map_init(struct vkd3d_va_map *va_map); void vkd3d_va_map_cleanup(struct vkd3d_va_map *va_map); struct vkd3d_gpu_va_allocation { D3D12_GPU_VIRTUAL_ADDRESS base; size_t size; void *ptr; }; struct vkd3d_gpu_va_slab { size_t size; void *ptr; }; enum vkd3d_render_pass_key_flag { VKD3D_RENDER_PASS_KEY_DEPTH_ENABLE = (1u << 0), VKD3D_RENDER_PASS_KEY_STENCIL_ENABLE = (1u << 1), VKD3D_RENDER_PASS_KEY_DEPTH_WRITE = (1u << 2), VKD3D_RENDER_PASS_KEY_STENCIL_WRITE = (1u << 3), VKD3D_RENDER_PASS_KEY_VRS_ATTACHMENT = (1u << 4), VKD3D_RENDER_PASS_KEY_DEPTH_STENCIL_ENABLE = (VKD3D_RENDER_PASS_KEY_DEPTH_ENABLE | VKD3D_RENDER_PASS_KEY_STENCIL_ENABLE), VKD3D_RENDER_PASS_KEY_DEPTH_STENCIL_WRITE = (VKD3D_RENDER_PASS_KEY_DEPTH_WRITE | VKD3D_RENDER_PASS_KEY_STENCIL_WRITE), }; struct vkd3d_render_pass_key { uint32_t attachment_count; uint32_t flags; /* vkd3d_render_pass_key_flag */ uint32_t sample_count; VkFormat vk_formats[D3D12_SIMULTANEOUS_RENDER_TARGET_COUNT + 1]; }; struct vkd3d_render_pass_entry; struct vkd3d_render_pass_cache { struct vkd3d_render_pass_entry *render_passes; size_t render_pass_count; size_t render_passes_size; spinlock_t lock; }; void vkd3d_render_pass_cache_cleanup(struct vkd3d_render_pass_cache *cache, struct d3d12_device *device); HRESULT vkd3d_render_pass_cache_find(struct vkd3d_render_pass_cache *cache, struct d3d12_device *device, const struct vkd3d_render_pass_key *key, VkRenderPass *vk_render_pass); void vkd3d_render_pass_cache_init(struct vkd3d_render_pass_cache *cache); struct vkd3d_private_store { pthread_mutex_t mutex; struct list content; }; struct vkd3d_private_data { struct list entry; GUID tag; unsigned int size; bool is_object; union { BYTE data[1]; IUnknown *object; }; }; static inline void vkd3d_private_data_destroy(struct vkd3d_private_data *data) { if (data->is_object) IUnknown_Release(data->object); list_remove(&data->entry); vkd3d_free(data); } static inline HRESULT vkd3d_private_store_init(struct vkd3d_private_store *store) { int rc; list_init(&store->content); if ((rc = pthread_mutex_init(&store->mutex, NULL))) ERR("Failed to initialize mutex, error %d.\n", rc); return hresult_from_errno(rc); } static inline void vkd3d_private_store_destroy(struct vkd3d_private_store *store) { struct vkd3d_private_data *data, *cursor; LIST_FOR_EACH_ENTRY_SAFE(data, cursor, &store->content, struct vkd3d_private_data, entry) { vkd3d_private_data_destroy(data); } pthread_mutex_destroy(&store->mutex); } static inline HRESULT vkd3d_private_data_lock(struct vkd3d_private_store *store) { int rc; if ((rc = pthread_mutex_lock(&store->mutex))) { ERR("Failed to lock mutex, error %d.\n", rc); return hresult_from_errno(rc); } return S_OK; } static inline void vkd3d_private_data_unlock(struct vkd3d_private_store *store) { pthread_mutex_unlock(&store->mutex); } HRESULT vkd3d_get_private_data(struct vkd3d_private_store *store, const GUID *tag, unsigned int *out_size, void *out); HRESULT vkd3d_private_store_set_private_data(struct vkd3d_private_store *store, const GUID *tag, const void *data, unsigned int data_size, bool is_object); typedef void(*vkd3d_set_name_callback)(void *, const char *); static inline bool vkd3d_private_data_object_name_ptr(REFGUID guid, UINT data_size, const void *data, const char **out_name) { if (out_name) *out_name = NULL; /* This is also handled in the object_name implementation * but this avoids an additional, needless allocation * and some games may spam SetName. */ if (!(vkd3d_config_flags & VKD3D_CONFIG_FLAG_DEBUG_UTILS)) return false; if (IsEqualGUID(guid, &WKPDID_D3DDebugObjectName)) { const char *name = (const char *)data; if (!data || !data_size) return true; if (out_name) *out_name = name[data_size - 1] != '\0' ? vkd3d_strdup_n(name, data_size) : name; return true; } else if (IsEqualGUID(guid, &WKPDID_D3DDebugObjectNameW)) { const WCHAR *name = (const WCHAR *)data; if (!data || data_size < sizeof(WCHAR)) return true; if (out_name) *out_name = vkd3d_strdup_w_utf8(name, data_size / sizeof(WCHAR)); return true; } return false; } static inline HRESULT vkd3d_set_private_data(struct vkd3d_private_store *store, const GUID *tag, unsigned int data_size, const void *data, vkd3d_set_name_callback set_name_callback, void *calling_object) { const char *name; HRESULT hr; if (FAILED(hr = vkd3d_private_data_lock(store))) return hr; if (FAILED(hr = vkd3d_private_store_set_private_data(store, tag, data, data_size, false))) { vkd3d_private_data_unlock(store); return hr; } if (set_name_callback && vkd3d_private_data_object_name_ptr(tag, data_size, data, &name)) { set_name_callback(calling_object, name); if (name && name != data) vkd3d_free((void *)name); } vkd3d_private_data_unlock(store); return hr; } static inline HRESULT vkd3d_set_private_data_interface(struct vkd3d_private_store *store, const GUID *tag, const IUnknown *object, vkd3d_set_name_callback set_name_callback, void *calling_object) { const void *data = object ? object : (void *)&object; HRESULT hr; if (FAILED(hr = vkd3d_private_data_lock(store))) return hr; if (FAILED(hr = vkd3d_private_store_set_private_data(store, tag, data, sizeof(object), !!object))) { vkd3d_private_data_unlock(store); return hr; } if (set_name_callback && vkd3d_private_data_object_name_ptr(tag, 0, NULL, NULL)) set_name_callback(calling_object, NULL); vkd3d_private_data_unlock(store); return hr; } HRESULT STDMETHODCALLTYPE d3d12_object_SetName(ID3D12Object *iface, const WCHAR *name); /* ID3D12Fence */ typedef ID3D12Fence1 d3d12_fence_iface; struct d3d12_fence_value { uint64_t virtual_value; uint64_t physical_value; const struct vkd3d_queue *signalling_queue; }; enum vkd3d_waiting_event_type { VKD3D_WAITING_EVENT_TYPE_EVENT, VKD3D_WAITING_EVENT_TYPE_SEMAPHORE, }; struct d3d12_fence { d3d12_fence_iface ID3D12Fence_iface; LONG refcount_internal; LONG refcount; D3D12_FENCE_FLAGS d3d12_flags; VkSemaphore timeline_semaphore; uint64_t max_pending_virtual_timeline_value; uint64_t virtual_value; uint64_t physical_value; uint64_t counter; struct d3d12_fence_value *pending_updates; size_t pending_updates_count; size_t pending_updates_size; pthread_mutex_t mutex; pthread_cond_t cond; pthread_cond_t null_event_cond; struct vkd3d_waiting_event { uint64_t value; HANDLE event; enum vkd3d_waiting_event_type type; bool *latch; } *events; size_t events_size; size_t event_count; struct d3d12_device *device; struct vkd3d_private_store private_store; }; struct d3d12_fence *unsafe_impl_from_ID3D12Fence1(ID3D12Fence1 *iface); struct d3d12_fence *unsafe_impl_from_ID3D12Fence(ID3D12Fence *iface); HRESULT d3d12_fence_create(struct d3d12_device *device, uint64_t initial_value, D3D12_FENCE_FLAGS flags, struct d3d12_fence **fence); HRESULT d3d12_fence_set_event_on_completion(struct d3d12_fence *fence, UINT64 value, HANDLE event, enum vkd3d_waiting_event_type type); enum vkd3d_allocation_flag { VKD3D_ALLOCATION_FLAG_GLOBAL_BUFFER = (1u << 0), VKD3D_ALLOCATION_FLAG_GPU_ADDRESS = (1u << 1), VKD3D_ALLOCATION_FLAG_CPU_ACCESS = (1u << 2), VKD3D_ALLOCATION_FLAG_ALLOW_WRITE_WATCH = (1u << 3), }; #define VKD3D_MEMORY_CHUNK_SIZE (VKD3D_VA_BLOCK_SIZE * 16) struct vkd3d_memory_chunk; struct vkd3d_allocate_memory_info { VkMemoryRequirements memory_requirements; D3D12_HEAP_PROPERTIES heap_properties; D3D12_HEAP_FLAGS heap_flags; void *host_ptr; const void *pNext; uint32_t flags; }; struct vkd3d_allocate_heap_memory_info { D3D12_HEAP_DESC heap_desc; void *host_ptr; }; struct vkd3d_allocate_resource_memory_info { D3D12_HEAP_PROPERTIES heap_properties; D3D12_HEAP_FLAGS heap_flags; VkBuffer vk_buffer; VkImage vk_image; void *host_ptr; }; struct vkd3d_view_map; struct vkd3d_unique_resource { union { VkBuffer vk_buffer; VkImage vk_image; }; uint64_t cookie; VkDeviceAddress va; VkDeviceSize size; /* This is used to handle views when we cannot bind it to a * specific ID3D12Resource, i.e. RTAS. Only allocated as needed. */ struct vkd3d_view_map *view_map; }; struct vkd3d_memory_allocation { struct vkd3d_unique_resource resource; VkDeviceMemory vk_memory; VkDeviceSize offset; void *cpu_address; D3D12_HEAP_TYPE heap_type; D3D12_HEAP_FLAGS heap_flags; uint32_t vk_memory_type; uint32_t flags; uint64_t clear_semaphore_value; struct vkd3d_memory_chunk *chunk; }; static inline void vkd3d_memory_allocation_slice(struct vkd3d_memory_allocation *dst, const struct vkd3d_memory_allocation *src, VkDeviceSize offset, VkDeviceSize size) { *dst = *src; dst->offset += offset; dst->resource.size = size; dst->resource.va += offset; if (dst->cpu_address) dst->cpu_address = void_ptr_offset(dst->cpu_address, offset); } struct vkd3d_memory_free_range { VkDeviceSize offset; VkDeviceSize length; }; struct vkd3d_memory_chunk { struct vkd3d_memory_allocation allocation; struct vkd3d_memory_free_range *free_ranges; size_t free_ranges_size; size_t free_ranges_count; }; #define VKD3D_MEMORY_CLEAR_COMMAND_BUFFER_COUNT (16u) struct vkd3d_memory_clear_queue { pthread_mutex_t mutex; VkCommandBuffer vk_command_buffers[VKD3D_MEMORY_CLEAR_COMMAND_BUFFER_COUNT]; VkCommandPool vk_command_pool; VkSemaphore vk_semaphore; UINT64 last_known_value; UINT64 next_signal_value; VkDeviceSize num_bytes_pending; uint32_t command_buffer_index; struct vkd3d_memory_allocation **allocations; size_t allocations_size; size_t allocations_count; }; struct vkd3d_memory_allocator { pthread_mutex_t mutex; struct vkd3d_memory_chunk **chunks; size_t chunks_size; size_t chunks_count; struct vkd3d_va_map va_map; struct vkd3d_queue *vkd3d_queue; struct vkd3d_memory_clear_queue clear_queue; }; void vkd3d_free_memory(struct d3d12_device *device, struct vkd3d_memory_allocator *allocator, const struct vkd3d_memory_allocation *allocation); HRESULT vkd3d_allocate_memory(struct d3d12_device *device, struct vkd3d_memory_allocator *allocator, const struct vkd3d_allocate_memory_info *info, struct vkd3d_memory_allocation *allocation); HRESULT vkd3d_allocate_heap_memory(struct d3d12_device *device, struct vkd3d_memory_allocator *allocator, const struct vkd3d_allocate_heap_memory_info *info, struct vkd3d_memory_allocation *allocation); HRESULT vkd3d_memory_allocator_init(struct vkd3d_memory_allocator *allocator, struct d3d12_device *device); void vkd3d_memory_allocator_cleanup(struct vkd3d_memory_allocator *allocator, struct d3d12_device *device); HRESULT vkd3d_memory_allocator_flush_clears(struct vkd3d_memory_allocator *allocator, struct d3d12_device *device); /* ID3D12Heap */ typedef ID3D12Heap1 d3d12_heap_iface; struct d3d12_heap { d3d12_heap_iface ID3D12Heap_iface; LONG refcount; D3D12_HEAP_DESC desc; struct vkd3d_memory_allocation allocation; struct d3d12_device *device; struct vkd3d_private_store private_store; }; HRESULT d3d12_heap_create(struct d3d12_device *device, const D3D12_HEAP_DESC *desc, void *host_address, struct d3d12_heap **heap); struct d3d12_heap *unsafe_impl_from_ID3D12Heap(ID3D12Heap *iface); enum vkd3d_resource_flag { VKD3D_RESOURCE_COMMITTED = (1u << 0), VKD3D_RESOURCE_PLACED = (1u << 1), VKD3D_RESOURCE_RESERVED = (1u << 2), VKD3D_RESOURCE_ALLOCATION = (1u << 3), VKD3D_RESOURCE_LINEAR_TILING = (1u << 4), VKD3D_RESOURCE_EXTERNAL = (1u << 5), VKD3D_RESOURCE_ACCELERATION_STRUCTURE = (1u << 6), VKD3D_RESOURCE_SIMULTANEOUS_ACCESS = (1u << 7), }; struct d3d12_sparse_image_region { VkImageSubresource subresource; uint32_t subresource_index; VkOffset3D offset; VkExtent3D extent; }; struct d3d12_sparse_buffer_region { VkDeviceSize offset; VkDeviceSize length; }; struct d3d12_sparse_tile { union { struct d3d12_sparse_image_region image; struct d3d12_sparse_buffer_region buffer; }; VkDeviceMemory vk_memory; VkDeviceSize vk_offset; }; struct d3d12_sparse_info { uint32_t tile_count; uint32_t tiling_count; struct d3d12_sparse_tile *tiles; D3D12_TILE_SHAPE tile_shape; D3D12_PACKED_MIP_INFO packed_mips; D3D12_SUBRESOURCE_TILING *tilings; VkDeviceMemory vk_metadata_memory; }; struct vkd3d_view_map { spinlock_t spinlock; struct hash_map map; #ifdef VKD3D_ENABLE_DESCRIPTOR_QA uint64_t resource_cookie; #endif }; HRESULT vkd3d_view_map_init(struct vkd3d_view_map *view_map); void vkd3d_view_map_destroy(struct vkd3d_view_map *view_map, struct d3d12_device *device); /* ID3D12Resource */ typedef ID3D12Resource1 d3d12_resource_iface; struct d3d12_resource { d3d12_resource_iface ID3D12Resource_iface; LONG refcount; LONG internal_refcount; D3D12_RESOURCE_DESC desc; D3D12_HEAP_PROPERTIES heap_properties; D3D12_HEAP_FLAGS heap_flags; struct vkd3d_memory_allocation mem; struct vkd3d_unique_resource res; struct d3d12_heap *heap; uint32_t flags; /* To keep track of initial layout. */ VkImageLayout common_layout; D3D12_RESOURCE_STATES initial_state; uint32_t initial_layout_transition; struct d3d12_sparse_info sparse; struct vkd3d_view_map view_map; struct d3d12_device *device; const struct vkd3d_format *format; VkImageView vrs_view; struct vkd3d_private_store private_store; }; static inline bool d3d12_resource_is_buffer(const struct d3d12_resource *resource) { return resource->desc.Dimension == D3D12_RESOURCE_DIMENSION_BUFFER; } static inline bool d3d12_resource_is_acceleration_structure(const struct d3d12_resource *resource) { return !!(resource->flags & VKD3D_RESOURCE_ACCELERATION_STRUCTURE); } static inline bool d3d12_resource_is_texture(const struct d3d12_resource *resource) { return resource->desc.Dimension != D3D12_RESOURCE_DIMENSION_BUFFER; } static inline VkImageLayout d3d12_resource_pick_layout(const struct d3d12_resource *resource, VkImageLayout layout) { return resource->flags & (VKD3D_RESOURCE_LINEAR_TILING | VKD3D_RESOURCE_SIMULTANEOUS_ACCESS) ? resource->common_layout : layout; } LONG64 vkd3d_allocate_cookie(); bool d3d12_resource_is_cpu_accessible(const struct d3d12_resource *resource); HRESULT d3d12_resource_validate_desc(const D3D12_RESOURCE_DESC *desc, struct d3d12_device *device); VkImageSubresource d3d12_resource_get_vk_subresource(const struct d3d12_resource *resource, uint32_t subresource_idx, bool all_aspects); VkImageAspectFlags vk_image_aspect_flags_from_d3d12( const struct vkd3d_format *format, uint32_t plane_idx); VkImageSubresource vk_image_subresource_from_d3d12( const struct vkd3d_format *format, uint32_t subresource_idx, unsigned int miplevel_count, unsigned int layer_count, bool all_aspects); HRESULT d3d12_resource_create_committed(struct d3d12_device *device, const D3D12_RESOURCE_DESC *desc, const D3D12_HEAP_PROPERTIES *heap_properties, D3D12_HEAP_FLAGS heap_flags, D3D12_RESOURCE_STATES initial_state, const D3D12_CLEAR_VALUE *optimized_clear_value, struct d3d12_resource **resource); HRESULT d3d12_resource_create_placed(struct d3d12_device *device, const D3D12_RESOURCE_DESC *desc, struct d3d12_heap *heap, uint64_t heap_offset, D3D12_RESOURCE_STATES initial_state, const D3D12_CLEAR_VALUE *optimized_clear_value, struct d3d12_resource **resource); HRESULT d3d12_resource_create_reserved(struct d3d12_device *device, const D3D12_RESOURCE_DESC *desc, D3D12_RESOURCE_STATES initial_state, const D3D12_CLEAR_VALUE *optimized_clear_value, struct d3d12_resource **resource); struct d3d12_resource *unsafe_impl_from_ID3D12Resource(ID3D12Resource *iface); HRESULT vkd3d_allocate_device_memory(struct d3d12_device *device, VkDeviceSize size, VkMemoryPropertyFlags type_flags, uint32_t type_mask, void *pNext, VkDeviceMemory *vk_memory, uint32_t *vk_memory_type); HRESULT vkd3d_allocate_buffer_memory(struct d3d12_device *device, VkBuffer vk_buffer, VkMemoryPropertyFlags type_flags, VkDeviceMemory *vk_memory); HRESULT vkd3d_allocate_image_memory(struct d3d12_device *device, VkImage vk_image, VkMemoryPropertyFlags type_flags, VkDeviceMemory *vk_memory); HRESULT vkd3d_create_buffer(struct d3d12_device *device, const D3D12_HEAP_PROPERTIES *heap_properties, D3D12_HEAP_FLAGS heap_flags, const D3D12_RESOURCE_DESC *desc, VkBuffer *vk_buffer); HRESULT vkd3d_get_image_allocation_info(struct d3d12_device *device, const D3D12_RESOURCE_DESC *desc, D3D12_RESOURCE_ALLOCATION_INFO *allocation_info); enum vkd3d_view_type { VKD3D_VIEW_TYPE_BUFFER, VKD3D_VIEW_TYPE_IMAGE, VKD3D_VIEW_TYPE_SAMPLER, VKD3D_VIEW_TYPE_ACCELERATION_STRUCTURE }; struct vkd3d_view { LONG refcount; enum vkd3d_view_type type; uint64_t cookie; union { VkBufferView vk_buffer_view; VkImageView vk_image_view; VkSampler vk_sampler; VkAccelerationStructureKHR vk_acceleration_structure; }; const struct vkd3d_format *format; union { struct { VkDeviceSize offset; VkDeviceSize size; } buffer; struct { VkImageViewType vk_view_type; unsigned int miplevel_idx; unsigned int layer_idx; unsigned int layer_count; } texture; } info; }; void vkd3d_view_decref(struct vkd3d_view *view, struct d3d12_device *device); void vkd3d_view_incref(struct vkd3d_view *view); struct vkd3d_buffer_view_desc { VkBuffer buffer; const struct vkd3d_format *format; VkDeviceSize offset; VkDeviceSize size; }; struct vkd3d_texture_view_desc { VkImage image; VkImageViewType view_type; VkImageAspectFlags aspect_mask; const struct vkd3d_format *format; unsigned int miplevel_idx; unsigned int miplevel_count; unsigned int layer_idx; unsigned int layer_count; float miplevel_clamp; VkComponentMapping components; bool allowed_swizzle; }; bool vkd3d_create_buffer_view(struct d3d12_device *device, const struct vkd3d_buffer_view_desc *desc, struct vkd3d_view **view); bool vkd3d_create_raw_r32ui_vk_buffer_view(struct d3d12_device *device, VkBuffer vk_buffer, VkDeviceSize offset, VkDeviceSize range, VkBufferView *vk_view); bool vkd3d_create_acceleration_structure_view(struct d3d12_device *device, const struct vkd3d_buffer_view_desc *desc, struct vkd3d_view **view); bool vkd3d_create_texture_view(struct d3d12_device *device, const struct vkd3d_texture_view_desc *desc, struct vkd3d_view **view); enum vkd3d_descriptor_flag { VKD3D_DESCRIPTOR_FLAG_VIEW = (1 << 0), VKD3D_DESCRIPTOR_FLAG_RAW_VA_AUX_BUFFER = (1 << 1), VKD3D_DESCRIPTOR_FLAG_BUFFER_OFFSET = (1 << 2), VKD3D_DESCRIPTOR_FLAG_OFFSET_RANGE = (1 << 3), }; struct vkd3d_descriptor_binding { uint16_t set; uint16_t binding; }; struct vkd3d_descriptor_data { uint64_t cookie; uint32_t set_info_mask; uint32_t flags; }; #define D3D12_DESC_ALIGNMENT 64 struct d3d12_desc { /* Align d3d12_desc to 64 bytes for two reasons. * - We need a POT size when reporting GPU addresses. * In DXR, we will need to handle app-placed VAs in a local root signature, * and the fastest approach we can use is uint(VA) >> 6 to derive an index. * - Can avoid false sharing on cache lines if multiple threads * modify adjacent descriptors somehow. */ DECLSPEC_ALIGN(D3D12_DESC_ALIGNMENT) struct vkd3d_descriptor_data metadata; struct d3d12_descriptor_heap *heap; uint32_t heap_offset; union { VkDescriptorBufferInfo buffer; struct vkd3d_view *view; } info; }; STATIC_ASSERT(sizeof(struct d3d12_desc) == 64); static inline struct d3d12_desc *d3d12_desc_from_cpu_handle(D3D12_CPU_DESCRIPTOR_HANDLE cpu_handle) { return (struct d3d12_desc *)cpu_handle.ptr; } static inline struct d3d12_desc *d3d12_desc_from_gpu_handle(D3D12_GPU_DESCRIPTOR_HANDLE gpu_handle) { return (struct d3d12_desc *)(intptr_t)gpu_handle.ptr; } void d3d12_desc_copy(struct d3d12_desc *dst, struct d3d12_desc *src, unsigned int count, D3D12_DESCRIPTOR_HEAP_TYPE heap_type, struct d3d12_device *device); void d3d12_desc_create_cbv(struct d3d12_desc *descriptor, struct d3d12_device *device, const D3D12_CONSTANT_BUFFER_VIEW_DESC *desc); void d3d12_desc_create_srv(struct d3d12_desc *descriptor, struct d3d12_device *device, struct d3d12_resource *resource, const D3D12_SHADER_RESOURCE_VIEW_DESC *desc); void d3d12_desc_create_uav(struct d3d12_desc *descriptor, struct d3d12_device *device, struct d3d12_resource *resource, struct d3d12_resource *counter_resource, const D3D12_UNORDERED_ACCESS_VIEW_DESC *desc); void d3d12_desc_create_sampler(struct d3d12_desc *sampler, struct d3d12_device *device, const D3D12_SAMPLER_DESC *desc); bool vkd3d_create_raw_buffer_view(struct d3d12_device *device, D3D12_GPU_VIRTUAL_ADDRESS gpu_address, VkBufferView *vk_buffer_view); HRESULT d3d12_create_static_sampler(struct d3d12_device *device, const D3D12_STATIC_SAMPLER_DESC *desc, VkSampler *vk_sampler); struct d3d12_rtv_desc { DECLSPEC_ALIGN(D3D12_DESC_ALIGNMENT) VkSampleCountFlagBits sample_count; const struct vkd3d_format *format; unsigned int width; unsigned int height; unsigned int layer_count; struct vkd3d_view *view; struct d3d12_resource *resource; }; STATIC_ASSERT(sizeof(struct d3d12_rtv_desc) == 64); void d3d12_rtv_desc_copy(struct d3d12_rtv_desc *dst, struct d3d12_rtv_desc *src, unsigned int count); static inline struct d3d12_rtv_desc *d3d12_rtv_desc_from_cpu_handle(D3D12_CPU_DESCRIPTOR_HANDLE cpu_handle) { return (struct d3d12_rtv_desc *)cpu_handle.ptr; } void d3d12_rtv_desc_create_rtv(struct d3d12_rtv_desc *rtv_desc, struct d3d12_device *device, struct d3d12_resource *resource, const D3D12_RENDER_TARGET_VIEW_DESC *desc); void d3d12_rtv_desc_create_dsv(struct d3d12_rtv_desc *dsv_desc, struct d3d12_device *device, struct d3d12_resource *resource, const D3D12_DEPTH_STENCIL_VIEW_DESC *desc); struct vkd3d_bound_buffer_range { uint32_t byte_offset; uint32_t byte_count; uint32_t element_offset; uint32_t element_count; }; struct vkd3d_host_visible_buffer_range { VkDescriptorBufferInfo descriptor; void *host_ptr; }; union vkd3d_descriptor_info { VkBufferView buffer_view; VkDescriptorBufferInfo buffer; VkDescriptorImageInfo image; VkDeviceAddress va; }; /* ID3D12DescriptorHeap */ struct d3d12_null_descriptor_template { struct VkWriteDescriptorSet writes[VKD3D_MAX_BINDLESS_DESCRIPTOR_SETS]; VkDescriptorBufferInfo buffer; VkDescriptorImageInfo image; VkBufferView buffer_view; unsigned int num_writes; unsigned int set_info_mask; }; struct d3d12_descriptor_heap { ID3D12DescriptorHeap ID3D12DescriptorHeap_iface; LONG refcount; uint64_t gpu_va; D3D12_DESCRIPTOR_HEAP_DESC desc; VkDescriptorPool vk_descriptor_pool; VkDescriptorSet vk_descriptor_sets[VKD3D_MAX_BINDLESS_DESCRIPTOR_SETS]; VkDeviceMemory vk_memory; VkBuffer vk_buffer; void *host_memory; struct vkd3d_host_visible_buffer_range raw_va_aux_buffer; struct vkd3d_host_visible_buffer_range buffer_ranges; #ifdef VKD3D_ENABLE_DESCRIPTOR_QA struct vkd3d_host_visible_buffer_range descriptor_heap_info; uint64_t cookie; #endif struct d3d12_null_descriptor_template null_descriptor_template; struct d3d12_device *device; struct vkd3d_private_store private_store; DECLSPEC_ALIGN(D3D12_DESC_ALIGNMENT) BYTE descriptors[]; }; HRESULT d3d12_descriptor_heap_create(struct d3d12_device *device, const D3D12_DESCRIPTOR_HEAP_DESC *desc, struct d3d12_descriptor_heap **descriptor_heap); void d3d12_descriptor_heap_cleanup(struct d3d12_descriptor_heap *descriptor_heap); struct d3d12_descriptor_heap *unsafe_impl_from_ID3D12DescriptorHeap(ID3D12DescriptorHeap *iface); static inline uint32_t d3d12_desc_heap_offset(const struct d3d12_desc *dst) { return dst->heap_offset; } static inline uint32_t d3d12_desc_heap_offset_from_gpu_handle(D3D12_GPU_DESCRIPTOR_HANDLE handle) { return (uint32_t)handle.ptr / sizeof(struct d3d12_desc); } /* ID3D12QueryHeap */ struct d3d12_query_heap { ID3D12QueryHeap ID3D12QueryHeap_iface; LONG refcount; D3D12_QUERY_HEAP_DESC desc; VkQueryPool vk_query_pool; VkDeviceMemory vk_memory; VkBuffer vk_buffer; uint32_t initialized; struct d3d12_device *device; struct vkd3d_private_store private_store; }; HRESULT d3d12_query_heap_create(struct d3d12_device *device, const D3D12_QUERY_HEAP_DESC *desc, struct d3d12_query_heap **heap); struct d3d12_query_heap *unsafe_impl_from_ID3D12QueryHeap(ID3D12QueryHeap *iface); size_t d3d12_query_heap_type_get_data_size(D3D12_QUERY_HEAP_TYPE heap_type); static inline bool d3d12_query_heap_type_is_inline(D3D12_QUERY_HEAP_TYPE heap_type) { return heap_type == D3D12_QUERY_HEAP_TYPE_OCCLUSION || heap_type == D3D12_QUERY_HEAP_TYPE_SO_STATISTICS; } enum vkd3d_root_signature_flag { VKD3D_ROOT_SIGNATURE_USE_ROOT_DESCRIPTOR_SET = 0x00000001u, VKD3D_ROOT_SIGNATURE_USE_INLINE_UNIFORM_BLOCK = 0x00000002u, VKD3D_ROOT_SIGNATURE_USE_RAW_VA_AUX_BUFFER = 0x00000004u, VKD3D_ROOT_SIGNATURE_USE_SSBO_OFFSET_BUFFER = 0x00000008u, VKD3D_ROOT_SIGNATURE_USE_TYPED_OFFSET_BUFFER = 0x00000010u, }; /* ID3D12RootSignature */ struct d3d12_bind_point_layout { VkPipelineLayout vk_pipeline_layout; VkShaderStageFlags vk_push_stages; }; #define VKD3D_MAX_HOISTED_DESCRIPTORS 16 struct vkd3d_descriptor_hoist_desc { uint32_t table_index; uint32_t table_offset; uint32_t parameter_index; }; struct vkd3d_descriptor_hoist_info { struct vkd3d_descriptor_hoist_desc desc[VKD3D_MAX_HOISTED_DESCRIPTORS]; unsigned int num_desc; }; struct d3d12_root_signature { ID3D12RootSignature ID3D12RootSignature_iface; LONG refcount; struct d3d12_bind_point_layout graphics, compute, raygen; VkDescriptorSetLayout vk_sampler_descriptor_layout; VkDescriptorSetLayout vk_root_descriptor_layout; VkDescriptorPool vk_sampler_pool; VkDescriptorSet vk_sampler_set; struct vkd3d_shader_root_parameter *parameters; unsigned int parameter_count; uint32_t sampler_descriptor_set; uint32_t root_descriptor_set; uint64_t descriptor_table_mask; uint64_t root_constant_mask; uint64_t root_descriptor_raw_va_mask; uint64_t root_descriptor_push_mask; D3D12_ROOT_SIGNATURE_FLAGS d3d12_flags; unsigned int flags; /* vkd3d_root_signature_flag */ unsigned int binding_count; struct vkd3d_shader_resource_binding *bindings; unsigned int root_constant_count; struct vkd3d_shader_push_constant_buffer *root_constants; /* Use one global push constant range */ VkPushConstantRange push_constant_range; struct vkd3d_shader_descriptor_binding push_constant_ubo_binding; struct vkd3d_shader_descriptor_binding raw_va_aux_buffer_binding; struct vkd3d_shader_descriptor_binding offset_buffer_binding; #ifdef VKD3D_ENABLE_DESCRIPTOR_QA struct vkd3d_shader_descriptor_binding descriptor_qa_heap_binding; struct vkd3d_shader_descriptor_binding descriptor_qa_global_info; #endif uint32_t descriptor_table_offset; uint32_t descriptor_table_count; unsigned int static_sampler_count; VkSampler *static_samplers; struct vkd3d_descriptor_hoist_info hoist_info; struct d3d12_device *device; struct vkd3d_private_store private_store; }; HRESULT d3d12_root_signature_create(struct d3d12_device *device, const void *bytecode, size_t bytecode_length, struct d3d12_root_signature **root_signature); struct d3d12_root_signature *unsafe_impl_from_ID3D12RootSignature(ID3D12RootSignature *iface); unsigned int d3d12_root_signature_get_shader_interface_flags(const struct d3d12_root_signature *root_signature); int vkd3d_parse_root_signature_v_1_0(const struct vkd3d_shader_code *dxbc, struct vkd3d_versioned_root_signature_desc *desc); int vkd3d_parse_root_signature_v_1_1(const struct vkd3d_shader_code *dxbc, struct vkd3d_versioned_root_signature_desc *desc); #define VKD3D_MAX_DYNAMIC_STATE_COUNT (7) enum vkd3d_dynamic_state_flag { VKD3D_DYNAMIC_STATE_VIEWPORT = (1 << 0), VKD3D_DYNAMIC_STATE_SCISSOR = (1 << 1), VKD3D_DYNAMIC_STATE_BLEND_CONSTANTS = (1 << 2), VKD3D_DYNAMIC_STATE_STENCIL_REFERENCE = (1 << 3), VKD3D_DYNAMIC_STATE_DEPTH_BOUNDS = (1 << 4), VKD3D_DYNAMIC_STATE_TOPOLOGY = (1 << 5), VKD3D_DYNAMIC_STATE_VERTEX_BUFFER = (1 << 6), VKD3D_DYNAMIC_STATE_VIEWPORT_COUNT = (1 << 7), VKD3D_DYNAMIC_STATE_SCISSOR_COUNT = (1 << 8), VKD3D_DYNAMIC_STATE_VERTEX_BUFFER_STRIDE = (1 << 9), VKD3D_DYNAMIC_STATE_FRAGMENT_SHADING_RATE = (1 << 10), }; struct vkd3d_shader_debug_ring_spec_constants { uint64_t hash; uint64_t atomic_bda; uint64_t host_bda; uint32_t ring_words; }; struct vkd3d_shader_debug_ring_spec_info { struct vkd3d_shader_debug_ring_spec_constants constants; VkSpecializationMapEntry map_entries[4]; VkSpecializationInfo spec_info; }; enum vkd3d_graphics_pipeline_static_variant_flag { VKD3D_GRAPHICS_PIPELINE_STATIC_VARIANT_VRS_ATTACHMENT = (1u << 0), VKD3D_GRAPHICS_PIPELINE_STATIC_VARIANT_LAST_BIT = (1u << 1), }; /* One render pass for each plane optimal mask. */ #define VKD3D_RENDER_PASS_COMPATIBILITY_VARIANT_COUNT 4 struct vkd3d_render_pass_compatibility { VkRenderPass dsv_layouts[VKD3D_RENDER_PASS_COMPATIBILITY_VARIANT_COUNT]; }; enum vkd3d_plane_optimal_flag { VKD3D_DEPTH_PLANE_OPTIMAL = (1 << 0), VKD3D_STENCIL_PLANE_OPTIMAL = (1 << 1), }; #define VKD3D_GRAPHICS_PIPELINE_STATIC_VARIANT_COUNT ((uint32_t)VKD3D_GRAPHICS_PIPELINE_STATIC_VARIANT_LAST_BIT) struct d3d12_graphics_pipeline_state { struct vkd3d_shader_debug_ring_spec_info spec_info[VKD3D_MAX_SHADER_STAGES]; VkPipelineShaderStageCreateInfo stages[VKD3D_MAX_SHADER_STAGES]; struct vkd3d_shader_meta stage_meta[VKD3D_MAX_SHADER_STAGES]; size_t stage_count; VkVertexInputAttributeDescription attributes[D3D12_VS_INPUT_REGISTER_COUNT]; VkVertexInputRate input_rates[D3D12_IA_VERTEX_INPUT_RESOURCE_SLOT_COUNT]; VkVertexInputBindingDivisorDescriptionEXT instance_divisors[D3D12_IA_VERTEX_INPUT_RESOURCE_SLOT_COUNT]; VkVertexInputBindingDescription attribute_bindings[D3D12_IA_VERTEX_INPUT_RESOURCE_SLOT_COUNT]; uint32_t minimum_vertex_buffer_dynamic_stride[D3D12_IA_VERTEX_INPUT_RESOURCE_SLOT_COUNT]; uint32_t vertex_buffer_stride_align_mask[D3D12_IA_VERTEX_INPUT_RESOURCE_SLOT_COUNT]; size_t instance_divisor_count; size_t attribute_binding_count; size_t attribute_count; D3D12_PRIMITIVE_TOPOLOGY_TYPE primitive_topology_type; uint32_t vertex_buffer_mask; VkPipelineColorBlendAttachmentState blend_attachments[D3D12_SIMULTANEOUS_RENDER_TARGET_COUNT]; unsigned int rt_count; unsigned int null_attachment_mask; unsigned int patch_vertex_count; const struct vkd3d_format *dsv_format; VkFormat rtv_formats[D3D12_SIMULTANEOUS_RENDER_TARGET_COUNT]; uint32_t dsv_plane_optimal_mask; struct vkd3d_render_pass_compatibility render_pass[VKD3D_GRAPHICS_PIPELINE_STATIC_VARIANT_COUNT]; D3D12_INDEX_BUFFER_STRIP_CUT_VALUE index_buffer_strip_cut_value; VkPipelineRasterizationStateCreateInfo rs_desc; VkPipelineMultisampleStateCreateInfo ms_desc; VkPipelineDepthStencilStateCreateInfo ds_desc; VkPipelineColorBlendStateCreateInfo blend_desc; VkSampleMask sample_mask[2]; VkPipelineRasterizationConservativeStateCreateInfoEXT rs_conservative_info; VkPipelineRasterizationDepthClipStateCreateInfoEXT rs_depth_clip_info; VkPipelineRasterizationStateStreamCreateInfoEXT rs_stream_info; uint32_t dynamic_state_flags; /* vkd3d_dynamic_state_flag */ VkPipelineLayout pipeline_layout; VkPipeline pipeline[VKD3D_GRAPHICS_PIPELINE_STATIC_VARIANT_COUNT]; struct list compiled_fallback_pipelines; bool xfb_enabled; }; static inline unsigned int dsv_attachment_mask(const struct d3d12_graphics_pipeline_state *graphics) { return 1u << graphics->rt_count; } struct d3d12_compute_pipeline_state { VkPipeline vk_pipeline; struct vkd3d_shader_meta meta; }; /* ID3D12PipelineState */ struct d3d12_pipeline_state { ID3D12PipelineState ID3D12PipelineState_iface; LONG refcount; union { struct d3d12_graphics_pipeline_state graphics; struct d3d12_compute_pipeline_state compute; }; VkPipelineBindPoint vk_bind_point; VkPipelineCache vk_pso_cache; spinlock_t lock; ID3D12RootSignature *private_root_signature; struct d3d12_device *device; struct vkd3d_private_store private_store; }; static inline bool d3d12_pipeline_state_is_compute(const struct d3d12_pipeline_state *state) { return state && state->vk_bind_point == VK_PIPELINE_BIND_POINT_COMPUTE; } static inline bool d3d12_pipeline_state_is_graphics(const struct d3d12_pipeline_state *state) { return state && state->vk_bind_point == VK_PIPELINE_BIND_POINT_GRAPHICS; } static inline bool d3d12_pipeline_state_has_unknown_dsv_format(struct d3d12_pipeline_state *state) { if (d3d12_pipeline_state_is_graphics(state)) { struct d3d12_graphics_pipeline_state *graphics = &state->graphics; return graphics->null_attachment_mask & dsv_attachment_mask(graphics); } return false; } struct d3d12_pipeline_state_desc { ID3D12RootSignature *root_signature; D3D12_SHADER_BYTECODE vs; D3D12_SHADER_BYTECODE ps; D3D12_SHADER_BYTECODE ds; D3D12_SHADER_BYTECODE hs; D3D12_SHADER_BYTECODE gs; D3D12_SHADER_BYTECODE cs; D3D12_STREAM_OUTPUT_DESC stream_output; D3D12_BLEND_DESC blend_state; UINT sample_mask; D3D12_RASTERIZER_DESC rasterizer_state; D3D12_DEPTH_STENCIL_DESC1 depth_stencil_state; D3D12_INPUT_LAYOUT_DESC input_layout; D3D12_INDEX_BUFFER_STRIP_CUT_VALUE strip_cut_value; D3D12_PRIMITIVE_TOPOLOGY_TYPE primitive_topology_type; D3D12_RT_FORMAT_ARRAY rtv_formats; DXGI_FORMAT dsv_format; DXGI_SAMPLE_DESC sample_desc; D3D12_VIEW_INSTANCING_DESC view_instancing_desc; UINT node_mask; D3D12_CACHED_PIPELINE_STATE cached_pso; D3D12_PIPELINE_STATE_FLAGS flags; }; HRESULT vkd3d_pipeline_state_desc_from_d3d12_graphics_desc(struct d3d12_pipeline_state_desc *desc, const D3D12_GRAPHICS_PIPELINE_STATE_DESC *d3d12_desc); HRESULT vkd3d_pipeline_state_desc_from_d3d12_compute_desc(struct d3d12_pipeline_state_desc *desc, const D3D12_COMPUTE_PIPELINE_STATE_DESC *d3d12_desc); HRESULT vkd3d_pipeline_state_desc_from_d3d12_stream_desc(struct d3d12_pipeline_state_desc *desc, const D3D12_PIPELINE_STATE_STREAM_DESC *d3d12_desc, VkPipelineBindPoint *vk_bind_point); struct vkd3d_pipeline_key { D3D12_PRIMITIVE_TOPOLOGY topology; uint32_t viewport_count; uint32_t strides[D3D12_IA_VERTEX_INPUT_RESOURCE_SLOT_COUNT]; VkFormat dsv_format; bool dynamic_stride; bool dynamic_viewport; bool dynamic_topology; }; bool d3d12_pipeline_state_has_replaced_shaders(struct d3d12_pipeline_state *state); HRESULT d3d12_pipeline_state_create(struct d3d12_device *device, VkPipelineBindPoint bind_point, const struct d3d12_pipeline_state_desc *desc, struct d3d12_pipeline_state **state); VkPipeline d3d12_pipeline_state_get_or_create_pipeline(struct d3d12_pipeline_state *state, const struct vkd3d_dynamic_state *dyn_state, const struct vkd3d_format *dsv_format, const struct vkd3d_render_pass_compatibility **render_pass_compat, uint32_t *dynamic_state_flags, uint32_t variant_flags); VkPipeline d3d12_pipeline_state_get_pipeline(struct d3d12_pipeline_state *state, const struct vkd3d_dynamic_state *dyn_state, const struct vkd3d_format *dsv_format, const struct vkd3d_render_pass_compatibility **render_pass_compat, uint32_t *dynamic_state_flags, uint32_t variant_flags); VkPipeline d3d12_pipeline_state_create_pipeline_variant(struct d3d12_pipeline_state *state, const struct vkd3d_pipeline_key *key, const struct vkd3d_format *dsv_format, VkPipelineCache vk_cache, struct vkd3d_render_pass_compatibility *render_pass_compat, uint32_t *dynamic_state_flags, uint32_t variant_flags); struct d3d12_pipeline_state *unsafe_impl_from_ID3D12PipelineState(ID3D12PipelineState *iface); /* ID3D12PipelineLibrary */ typedef ID3D12PipelineLibrary1 d3d12_pipeline_library_iface; struct d3d12_pipeline_library { d3d12_pipeline_library_iface ID3D12PipelineLibrary_iface; LONG refcount; struct d3d12_device *device; pthread_mutex_t mutex; struct hash_map map; struct vkd3d_private_store private_store; }; HRESULT d3d12_pipeline_library_create(struct d3d12_device *device, const void *blob, size_t blob_length, struct d3d12_pipeline_library **pipeline_library); HRESULT vkd3d_create_pipeline_cache_from_d3d12_desc(struct d3d12_device *device, const D3D12_CACHED_PIPELINE_STATE *state, VkPipelineCache *cache); VkResult vkd3d_serialize_pipeline_state(const struct d3d12_pipeline_state *state, size_t *size, void *data); struct vkd3d_buffer { VkBuffer vk_buffer; VkDeviceMemory vk_memory; }; struct d3d12_descriptor_pool_cache { VkDescriptorPool vk_descriptor_pool; VkDescriptorPool *free_descriptor_pools; size_t free_descriptor_pools_size; size_t free_descriptor_pool_count; VkDescriptorPool *descriptor_pools; size_t descriptor_pools_size; size_t descriptor_pool_count; }; enum vkd3d_descriptor_pool_types { VKD3D_DESCRIPTOR_POOL_TYPE_STATIC = 0, VKD3D_DESCRIPTOR_POOL_TYPE_COUNT }; #define VKD3D_SCRATCH_BUFFER_SIZE (1ull << 20) #define VKD3D_SCRATCH_BUFFER_COUNT (32u) struct vkd3d_scratch_buffer { struct vkd3d_memory_allocation allocation; VkDeviceSize offset; }; #define VKD3D_QUERY_TYPE_INDEX_OCCLUSION (0u) #define VKD3D_QUERY_TYPE_INDEX_PIPELINE_STATISTICS (1u) #define VKD3D_QUERY_TYPE_INDEX_TRANSFORM_FEEDBACK (2u) #define VKD3D_QUERY_TYPE_INDEX_RT_COMPACTED_SIZE (3u) #define VKD3D_QUERY_TYPE_INDEX_RT_SERIALIZE_SIZE (4u) #define VKD3D_VIRTUAL_QUERY_TYPE_COUNT (5u) #define VKD3D_VIRTUAL_QUERY_POOL_COUNT (128u) struct vkd3d_query_pool { VkQueryPool vk_query_pool; uint32_t type_index; uint32_t query_count; uint32_t next_index; }; /* ID3D12CommandAllocator */ struct d3d12_command_allocator { ID3D12CommandAllocator ID3D12CommandAllocator_iface; LONG refcount; D3D12_COMMAND_LIST_TYPE type; VkQueueFlags vk_queue_flags; VkCommandPool vk_command_pool; struct d3d12_descriptor_pool_cache descriptor_pool_caches[VKD3D_DESCRIPTOR_POOL_TYPE_COUNT]; VkRenderPass *passes; size_t passes_size; size_t pass_count; VkFramebuffer *framebuffers; size_t framebuffers_size; size_t framebuffer_count; struct vkd3d_view **views; size_t views_size; size_t view_count; VkBufferView *buffer_views; size_t buffer_views_size; size_t buffer_view_count; VkCommandBuffer *command_buffers; size_t command_buffers_size; size_t command_buffer_count; struct vkd3d_scratch_buffer *scratch_buffers; size_t scratch_buffers_size; size_t scratch_buffer_count; struct vkd3d_query_pool *query_pools; size_t query_pools_size; size_t query_pool_count; struct vkd3d_query_pool active_query_pools[VKD3D_VIRTUAL_QUERY_TYPE_COUNT]; LONG outstanding_submissions_count; struct d3d12_command_list *current_command_list; struct d3d12_device *device; struct vkd3d_private_store private_store; }; HRESULT d3d12_command_allocator_create(struct d3d12_device *device, D3D12_COMMAND_LIST_TYPE type, struct d3d12_command_allocator **allocator); bool d3d12_command_allocator_allocate_query_from_type_index( struct d3d12_command_allocator *allocator, uint32_t type_index, VkQueryPool *query_pool, uint32_t *query_index); enum vkd3d_pipeline_dirty_flag { VKD3D_PIPELINE_DIRTY_STATIC_SAMPLER_SET = 0x00000001u, VKD3D_PIPELINE_DIRTY_DESCRIPTOR_TABLE_OFFSETS = 0x00000002u, VKD3D_PIPELINE_DIRTY_HOISTED_DESCRIPTORS = 0x00000004u, }; struct vkd3d_root_descriptor_info { VkDescriptorType vk_descriptor_type; union vkd3d_descriptor_info info; }; struct vkd3d_pipeline_bindings { const struct d3d12_root_signature *root_signature; /* RT in DXR happens in COMPUTE bind point. */ struct d3d12_bind_point_layout layout, rt_layout; VkDescriptorSet static_sampler_set; uint32_t dirty_flags; /* vkd3d_pipeline_dirty_flags */ uint32_t descriptor_tables[D3D12_MAX_ROOT_COST]; uint64_t descriptor_table_active_mask; uint64_t descriptor_heap_dirty_mask; /* Needed when VK_KHR_push_descriptor is not available. */ struct vkd3d_root_descriptor_info root_descriptors[D3D12_MAX_ROOT_COST]; uint64_t root_descriptor_dirty_mask; uint64_t root_descriptor_active_mask; uint32_t root_constants[D3D12_MAX_ROOT_COST]; uint64_t root_constant_dirty_mask; }; struct vkd3d_dynamic_state { uint32_t active_flags; /* vkd3d_dynamic_state_flags */ uint32_t dirty_flags; /* vkd3d_dynamic_state_flags */ uint32_t dirty_vbos; uint32_t dirty_vbo_strides; uint32_t viewport_count; VkViewport viewports[D3D12_VIEWPORT_AND_SCISSORRECT_OBJECT_COUNT_PER_PIPELINE]; VkRect2D scissors[D3D12_VIEWPORT_AND_SCISSORRECT_OBJECT_COUNT_PER_PIPELINE]; float blend_constants[4]; uint32_t stencil_reference; float min_depth_bounds; float max_depth_bounds; VkBuffer vertex_buffers[D3D12_IA_VERTEX_INPUT_RESOURCE_SLOT_COUNT]; VkDeviceSize vertex_offsets[D3D12_IA_VERTEX_INPUT_RESOURCE_SLOT_COUNT]; VkDeviceSize vertex_sizes[D3D12_IA_VERTEX_INPUT_RESOURCE_SLOT_COUNT]; VkDeviceSize vertex_strides[D3D12_IA_VERTEX_INPUT_RESOURCE_SLOT_COUNT]; D3D12_PRIMITIVE_TOPOLOGY primitive_topology; VkPrimitiveTopology vk_primitive_topology; struct { VkExtent2D fragment_size; VkFragmentShadingRateCombinerOpKHR combiner_ops[D3D12_RS_SET_SHADING_RATE_COMBINER_COUNT]; } fragment_shading_rate; uint32_t pipeline_stack_size; }; /* ID3D12CommandList */ typedef ID3D12GraphicsCommandList5 d3d12_command_list_iface; struct vkd3d_clear_attachment { VkImageAspectFlags aspect_mask; VkImageAspectFlags discard_mask; VkClearValue value; }; struct vkd3d_clear_state { uint64_t attachment_mask; struct vkd3d_clear_attachment attachments[D3D12_SIMULTANEOUS_RENDER_TARGET_COUNT + 1]; }; enum vkd3d_initial_transition_type { VKD3D_INITIAL_TRANSITION_TYPE_RESOURCE, VKD3D_INITIAL_TRANSITION_TYPE_QUERY_HEAP, }; struct vkd3d_initial_transition { enum vkd3d_initial_transition_type type; union { struct { struct d3d12_resource *resource; bool perform_initial_transition; } resource; struct d3d12_query_heap *query_heap; }; }; enum vkd3d_active_query_state { VKD3D_ACTIVE_QUERY_RESET, VKD3D_ACTIVE_QUERY_BEGUN, VKD3D_ACTIVE_QUERY_ENDED, }; struct vkd3d_active_query { struct d3d12_query_heap *heap; uint32_t index; D3D12_QUERY_TYPE type; VkQueryPool vk_pool; uint32_t vk_index; enum vkd3d_active_query_state state; uint32_t resolve_index; }; enum vkd3d_query_range_flag { VKD3D_QUERY_RANGE_RESET = 0x1, }; struct vkd3d_query_range { VkQueryPool vk_pool; uint32_t index; uint32_t count; uint32_t flags; }; /* ID3D12CommandListExt */ typedef ID3D12GraphicsCommandListExt d3d12_command_list_vkd3d_ext_iface; struct d3d12_state_object; struct d3d12_resource_tracking { const struct d3d12_resource *resource; uint32_t plane_optimal_mask; }; struct d3d12_command_list { d3d12_command_list_iface ID3D12GraphicsCommandList_iface; d3d12_command_list_vkd3d_ext_iface ID3D12GraphicsCommandListExt_iface; LONG refcount; D3D12_COMMAND_LIST_TYPE type; VkQueueFlags vk_queue_flags; bool is_recording; bool is_valid; bool debug_capture; bool has_replaced_shaders; bool has_valid_index_buffer; VkCommandBuffer vk_command_buffer; VkCommandBuffer vk_init_commands; DXGI_FORMAT index_buffer_format; struct d3d12_rtv_desc rtvs[D3D12_SIMULTANEOUS_RENDER_TARGET_COUNT]; struct d3d12_rtv_desc dsv; struct vkd3d_clear_state clear_state; uint32_t dsv_plane_optimal_mask; VkImageLayout dsv_layout; unsigned int fb_width; unsigned int fb_height; unsigned int fb_layer_count; bool xfb_enabled; bool render_pass_suspended; bool predicate_enabled; VkDeviceAddress predicate_va; VkFramebuffer current_framebuffer; /* This is VK_NULL_HANDLE when we are no longer sure which pipeline to bind, * if this is NULL, we might need to lookup a pipeline key in order to bind the correct pipeline. */ VkPipeline current_pipeline; /* This is the actual pipeline which is bound to the pipeline. This lets us elide * possible calls to vkCmdBindPipeline and avoids invalidating dynamic state. */ VkPipeline command_buffer_pipeline; VkRenderPass pso_render_pass; VkRenderPass current_render_pass; struct vkd3d_dynamic_state dynamic_state; struct vkd3d_pipeline_bindings pipeline_bindings[VKD3D_PIPELINE_BIND_POINT_COUNT]; VkPipelineBindPoint active_bind_point; VkDescriptorSet descriptor_heaps[VKD3D_MAX_BINDLESS_DESCRIPTOR_SETS]; struct d3d12_pipeline_state *state; struct d3d12_state_object *rt_state; struct d3d12_command_allocator *allocator; struct d3d12_device *device; VkBuffer so_counter_buffers[D3D12_SO_BUFFER_SLOT_COUNT]; VkDeviceSize so_counter_buffer_offsets[D3D12_SO_BUFFER_SLOT_COUNT]; struct vkd3d_initial_transition *init_transitions; size_t init_transitions_size; size_t init_transitions_count; struct vkd3d_query_range *query_ranges; size_t query_ranges_size; size_t query_ranges_count; struct vkd3d_active_query *active_queries; size_t active_queries_size; size_t active_queries_count; struct vkd3d_active_query *pending_queries; size_t pending_queries_size; size_t pending_queries_count; LONG *outstanding_submissions_count; const struct d3d12_desc *cbv_srv_uav_descriptors; struct d3d12_resource *vrs_image; struct d3d12_resource_tracking *dsv_resource_tracking; size_t dsv_resource_tracking_count; size_t dsv_resource_tracking_size; struct vkd3d_private_store private_store; }; HRESULT d3d12_command_list_create(struct d3d12_device *device, UINT node_mask, D3D12_COMMAND_LIST_TYPE type, struct d3d12_command_list **list); bool d3d12_command_list_reset_query(struct d3d12_command_list *list, VkQueryPool vk_pool, uint32_t index); #define VKD3D_BUNDLE_CHUNK_SIZE (256 << 10) #define VKD3D_BUNDLE_COMMAND_ALIGNMENT (sizeof(UINT64)) struct d3d12_bundle_allocator { ID3D12CommandAllocator ID3D12CommandAllocator_iface; LONG refcount; void **chunks; size_t chunks_size; size_t chunks_count; size_t chunk_offset; struct d3d12_bundle *current_bundle; struct d3d12_device *device; struct vkd3d_private_store private_store; }; HRESULT d3d12_bundle_allocator_create(struct d3d12_device *device, struct d3d12_bundle_allocator **allocator); typedef void (*pfn_d3d12_bundle_command)(d3d12_command_list_iface *command_list, const void *args); struct d3d12_bundle_command { pfn_d3d12_bundle_command proc; struct d3d12_bundle_command *next; }; struct d3d12_bundle { d3d12_command_list_iface ID3D12GraphicsCommandList_iface; LONG refcount; bool is_recording; struct d3d12_device *device; struct d3d12_bundle_allocator *allocator; struct d3d12_bundle_command *head; struct d3d12_bundle_command *tail; struct vkd3d_private_store private_store; }; HRESULT d3d12_bundle_create(struct d3d12_device *device, UINT node_mask, D3D12_COMMAND_LIST_TYPE type, struct d3d12_bundle **bundle); void d3d12_bundle_execute(struct d3d12_bundle *bundle, d3d12_command_list_iface *list); struct d3d12_bundle *d3d12_bundle_from_iface(ID3D12GraphicsCommandList *iface); struct vkd3d_queue { /* Access to VkQueue must be externally synchronized. */ pthread_mutex_t mutex; VkQueue vk_queue; VkCommandPool barrier_pool; VkCommandBuffer barrier_command_buffer; VkSemaphore serializing_binary_semaphore; uint32_t vk_family_index; VkQueueFlags vk_queue_flags; uint32_t timestamp_bits; uint32_t virtual_queue_count; VkSemaphore *wait_semaphores; size_t wait_semaphores_size; uint64_t *wait_values; size_t wait_values_size; VkPipelineStageFlags *wait_stages; size_t wait_stages_size; uint32_t wait_count; }; VkQueue vkd3d_queue_acquire(struct vkd3d_queue *queue); HRESULT vkd3d_queue_create(struct d3d12_device *device, uint32_t family_index, uint32_t queue_index, const VkQueueFamilyProperties *properties, struct vkd3d_queue **queue); void vkd3d_queue_destroy(struct vkd3d_queue *queue, struct d3d12_device *device); void vkd3d_queue_release(struct vkd3d_queue *queue); void vkd3d_queue_add_wait(struct vkd3d_queue *queue, VkSemaphore semaphore, uint64_t value); enum vkd3d_submission_type { VKD3D_SUBMISSION_WAIT, VKD3D_SUBMISSION_SIGNAL, VKD3D_SUBMISSION_EXECUTE, VKD3D_SUBMISSION_BIND_SPARSE, VKD3D_SUBMISSION_STOP, VKD3D_SUBMISSION_DRAIN }; enum vkd3d_sparse_memory_bind_mode { VKD3D_SPARSE_MEMORY_BIND_MODE_UPDATE, VKD3D_SPARSE_MEMORY_BIND_MODE_COPY, }; struct vkd3d_sparse_memory_bind { uint32_t dst_tile; uint32_t src_tile; VkDeviceMemory vk_memory; VkDeviceSize vk_offset; }; struct vkd3d_sparse_memory_bind_range { uint32_t tile_index; uint32_t tile_count; VkDeviceMemory vk_memory; VkDeviceSize vk_offset; }; struct d3d12_command_queue_submission_wait { struct d3d12_fence *fence; UINT64 value; }; struct d3d12_command_queue_submission_signal { struct d3d12_fence *fence; UINT64 value; }; struct d3d12_command_queue_submission_execute { VkCommandBuffer *cmd; LONG **outstanding_submissions_counters; UINT cmd_count; UINT outstanding_submissions_counter_count; struct vkd3d_initial_transition *transitions; size_t transition_count; bool debug_capture; }; struct d3d12_command_queue_submission_bind_sparse { enum vkd3d_sparse_memory_bind_mode mode; uint32_t bind_count; struct vkd3d_sparse_memory_bind *bind_infos; struct d3d12_resource *dst_resource; struct d3d12_resource *src_resource; }; struct d3d12_command_queue_submission { enum vkd3d_submission_type type; union { struct d3d12_command_queue_submission_wait wait; struct d3d12_command_queue_submission_signal signal; struct d3d12_command_queue_submission_execute execute; struct d3d12_command_queue_submission_bind_sparse bind_sparse; }; }; struct vkd3d_timeline_semaphore { VkSemaphore vk_semaphore; uint64_t last_signaled; }; /* IWineDXGISwapChainFactory */ struct d3d12_swapchain_factory { IWineDXGISwapChainFactory IWineDXGISwapChainFactory_iface; struct d3d12_command_queue *queue; }; HRESULT d3d12_swapchain_factory_init(struct d3d12_command_queue *queue, struct d3d12_swapchain_factory *factory); /* ID3D12CommandQueue */ struct d3d12_command_queue { ID3D12CommandQueue ID3D12CommandQueue_iface; LONG refcount; D3D12_COMMAND_QUEUE_DESC desc; struct vkd3d_queue *vkd3d_queue; struct d3d12_device *device; pthread_mutex_t queue_lock; pthread_cond_t queue_cond; pthread_t submission_thread; struct d3d12_command_queue_submission *submissions; size_t submissions_count; size_t submissions_size; uint64_t drain_count; uint64_t queue_drain_count; struct vkd3d_fence_worker fence_worker; struct vkd3d_private_store private_store; #ifdef VKD3D_BUILD_STANDALONE_D3D12 struct d3d12_swapchain_factory swapchain_factory; #endif }; HRESULT d3d12_command_queue_create(struct d3d12_device *device, const D3D12_COMMAND_QUEUE_DESC *desc, struct d3d12_command_queue **queue); void d3d12_command_queue_submit_stop(struct d3d12_command_queue *queue); /* ID3D12CommandSignature */ struct d3d12_command_signature { ID3D12CommandSignature ID3D12CommandSignature_iface; LONG refcount; D3D12_COMMAND_SIGNATURE_DESC desc; struct d3d12_device *device; struct vkd3d_private_store private_store; }; HRESULT d3d12_command_signature_create(struct d3d12_device *device, const D3D12_COMMAND_SIGNATURE_DESC *desc, struct d3d12_command_signature **signature); struct d3d12_command_signature *unsafe_impl_from_ID3D12CommandSignature(ID3D12CommandSignature *iface); /* Static samplers */ struct vkd3d_sampler_state { pthread_mutex_t mutex; struct hash_map map; VkDescriptorPool *vk_descriptor_pools; size_t vk_descriptor_pools_size; size_t vk_descriptor_pool_count; }; struct vkd3d_shader_debug_ring { VkBuffer host_buffer; VkBuffer device_atomic_buffer; VkDeviceMemory host_buffer_memory; VkDeviceMemory device_atomic_buffer_memory; void *mapped; VkDeviceAddress ring_device_address; VkDeviceAddress atomic_device_address; size_t ring_size; size_t ring_offset; pthread_t ring_thread; pthread_mutex_t ring_lock; pthread_cond_t ring_cond; bool active; }; HRESULT vkd3d_sampler_state_init(struct vkd3d_sampler_state *state, struct d3d12_device *device); void vkd3d_sampler_state_cleanup(struct vkd3d_sampler_state *state, struct d3d12_device *device); HRESULT vkd3d_sampler_state_create_static_sampler(struct vkd3d_sampler_state *state, struct d3d12_device *device, const D3D12_STATIC_SAMPLER_DESC *desc, VkSampler *vk_sampler); HRESULT vkd3d_sampler_state_allocate_descriptor_set(struct vkd3d_sampler_state *state, struct d3d12_device *device, VkDescriptorSetLayout vk_layout, VkDescriptorSet *vk_set, VkDescriptorPool *vk_pool); void vkd3d_sampler_state_free_descriptor_set(struct vkd3d_sampler_state *state, struct d3d12_device *device, VkDescriptorSet vk_set, VkDescriptorPool vk_pool); HRESULT vkd3d_shader_debug_ring_init(struct vkd3d_shader_debug_ring *state, struct d3d12_device *device); void vkd3d_shader_debug_ring_cleanup(struct vkd3d_shader_debug_ring *state, struct d3d12_device *device); void *vkd3d_shader_debug_ring_thread_main(void *arg); void vkd3d_shader_debug_ring_init_spec_constant(struct d3d12_device *device, struct vkd3d_shader_debug_ring_spec_info *info, vkd3d_shader_hash_t hash); void vkd3d_shader_debug_ring_end_command_buffer(struct d3d12_command_list *list); /* Bindless */ enum vkd3d_bindless_flags { VKD3D_BINDLESS_SAMPLER = (1u << 0), VKD3D_BINDLESS_CBV = (1u << 1), VKD3D_BINDLESS_SRV = (1u << 2), VKD3D_BINDLESS_UAV = (1u << 3), VKD3D_RAW_VA_AUX_BUFFER = (1u << 4), VKD3D_BINDLESS_CBV_AS_SSBO = (1u << 5), VKD3D_BINDLESS_RAW_SSBO = (1u << 6), VKD3D_SSBO_OFFSET_BUFFER = (1u << 7), VKD3D_TYPED_OFFSET_BUFFER = (1u << 8), VKD3D_RAW_VA_ROOT_DESCRIPTOR_CBV = (1u << 9), VKD3D_RAW_VA_ROOT_DESCRIPTOR_SRV_UAV = (1u << 10), VKD3D_BINDLESS_MUTABLE_TYPE = (1u << 11), VKD3D_HOIST_STATIC_TABLE_CBV = (1u << 12), }; #define VKD3D_BINDLESS_SET_MAX_EXTRA_BINDINGS 8 enum vkd3d_bindless_set_flag { VKD3D_BINDLESS_SET_SAMPLER = (1u << 0), VKD3D_BINDLESS_SET_CBV = (1u << 1), VKD3D_BINDLESS_SET_SRV = (1u << 2), VKD3D_BINDLESS_SET_UAV = (1u << 3), VKD3D_BINDLESS_SET_IMAGE = (1u << 4), VKD3D_BINDLESS_SET_BUFFER = (1u << 5), VKD3D_BINDLESS_SET_AUX_BUFFER = (1u << 6), VKD3D_BINDLESS_SET_RAW_SSBO = (1u << 7), VKD3D_BINDLESS_SET_MUTABLE = (1u << 8), VKD3D_BINDLESS_SET_EXTRA_RAW_VA_AUX_BUFFER = (1u << 24), VKD3D_BINDLESS_SET_EXTRA_OFFSET_BUFFER = (1u << 25), VKD3D_BINDLESS_SET_EXTRA_GLOBAL_HEAP_INFO_BUFFER = (1u << 26), VKD3D_BINDLESS_SET_EXTRA_DESCRIPTOR_HEAP_INFO_BUFFER = (1u << 27), VKD3D_BINDLESS_SET_EXTRA_MASK = 0xff000000u }; struct vkd3d_bindless_set_info { VkDescriptorType vk_descriptor_type; D3D12_DESCRIPTOR_HEAP_TYPE heap_type; uint32_t flags; /* vkd3d_bindless_set_flag */ uint32_t set_index; uint32_t binding_index; VkDescriptorSetLayout vk_set_layout; VkDescriptorSetLayout vk_host_set_layout; }; struct vkd3d_bindless_state { uint32_t flags; /* vkd3d_bindless_flags */ struct vkd3d_bindless_set_info set_info[VKD3D_MAX_BINDLESS_DESCRIPTOR_SETS]; unsigned int set_count; unsigned int cbv_srv_uav_count; }; HRESULT vkd3d_bindless_state_init(struct vkd3d_bindless_state *bindless_state, struct d3d12_device *device); void vkd3d_bindless_state_cleanup(struct vkd3d_bindless_state *bindless_state, struct d3d12_device *device); bool vkd3d_bindless_state_find_binding(const struct vkd3d_bindless_state *bindless_state, uint32_t flags, struct vkd3d_shader_descriptor_binding *binding); struct vkd3d_descriptor_binding vkd3d_bindless_state_find_set(const struct vkd3d_bindless_state *bindless_state, uint32_t flags); struct vkd3d_descriptor_binding vkd3d_bindless_state_binding_from_info_index( const struct vkd3d_bindless_state *bindless_state, uint32_t index); uint32_t vkd3d_bindless_state_find_set_info_index(const struct vkd3d_bindless_state *bindless_state, uint32_t flags); static inline VkDescriptorType vkd3d_bindless_state_get_cbv_descriptor_type(const struct vkd3d_bindless_state *bindless_state) { return bindless_state->flags & VKD3D_BINDLESS_CBV_AS_SSBO ? VK_DESCRIPTOR_TYPE_STORAGE_BUFFER : VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER; } struct vkd3d_format_compatibility_list { DXGI_FORMAT typeless_format; unsigned int format_count; VkFormat vk_formats[VKD3D_MAX_COMPATIBLE_FORMAT_COUNT]; }; struct vkd3d_memory_info { uint32_t global_mask; uint32_t buffer_type_mask; uint32_t sampled_type_mask; uint32_t rt_ds_type_mask; }; HRESULT vkd3d_memory_info_init(struct vkd3d_memory_info *info, struct d3d12_device *device); /* meta operations */ struct vkd3d_clear_uav_args { VkClearColorValue clear_color; VkOffset2D offset; VkExtent2D extent; }; struct vkd3d_clear_uav_pipelines { VkPipeline buffer; VkPipeline buffer_raw; VkPipeline image_1d; VkPipeline image_2d; VkPipeline image_3d; VkPipeline image_1d_array; VkPipeline image_2d_array; }; struct vkd3d_clear_uav_ops { VkDescriptorSetLayout vk_set_layout_buffer_raw; VkDescriptorSetLayout vk_set_layout_buffer; VkDescriptorSetLayout vk_set_layout_image; VkPipelineLayout vk_pipeline_layout_buffer_raw; VkPipelineLayout vk_pipeline_layout_buffer; VkPipelineLayout vk_pipeline_layout_image; struct vkd3d_clear_uav_pipelines clear_float; struct vkd3d_clear_uav_pipelines clear_uint; }; struct vkd3d_clear_uav_pipeline { VkDescriptorSetLayout vk_set_layout; VkPipelineLayout vk_pipeline_layout; VkPipeline vk_pipeline; }; HRESULT vkd3d_clear_uav_ops_init(struct vkd3d_clear_uav_ops *meta_clear_uav_ops, struct d3d12_device *device); void vkd3d_clear_uav_ops_cleanup(struct vkd3d_clear_uav_ops *meta_clear_uav_ops, struct d3d12_device *device); struct vkd3d_copy_image_args { VkOffset2D offset; }; struct vkd3d_copy_image_info { VkDescriptorSetLayout vk_set_layout; VkPipelineLayout vk_pipeline_layout; VkRenderPass vk_render_pass; VkPipeline vk_pipeline; }; struct vkd3d_copy_image_pipeline_key { const struct vkd3d_format *format; VkImageViewType view_type; VkSampleCountFlagBits sample_count; VkImageLayout layout; }; struct vkd3d_copy_image_pipeline { struct vkd3d_copy_image_pipeline_key key; VkRenderPass vk_render_pass; VkPipeline vk_pipeline; }; struct vkd3d_copy_image_ops { VkDescriptorSetLayout vk_set_layout; VkPipelineLayout vk_pipeline_layout; VkShaderModule vk_fs_float_module; VkShaderModule vk_fs_uint_module; pthread_mutex_t mutex; struct vkd3d_copy_image_pipeline *pipelines; size_t pipelines_size; size_t pipeline_count; }; HRESULT vkd3d_copy_image_ops_init(struct vkd3d_copy_image_ops *meta_copy_image_ops, struct d3d12_device *device); void vkd3d_copy_image_ops_cleanup(struct vkd3d_copy_image_ops *meta_copy_image_ops, struct d3d12_device *device); struct vkd3d_swapchain_pipeline_key { VkPipelineBindPoint bind_point; VkAttachmentLoadOp load_op; VkFormat format; VkFilter filter; }; struct vkd3d_swapchain_info { VkDescriptorSetLayout vk_set_layout; VkPipelineLayout vk_pipeline_layout; VkRenderPass vk_render_pass; VkPipeline vk_pipeline; }; struct vkd3d_swapchain_pipeline { VkRenderPass vk_render_pass; VkPipeline vk_pipeline; struct vkd3d_swapchain_pipeline_key key; }; struct vkd3d_swapchain_ops { VkDescriptorSetLayout vk_set_layouts[2]; VkPipelineLayout vk_pipeline_layouts[2]; VkShaderModule vk_vs_module; VkShaderModule vk_fs_module; VkSampler vk_samplers[2]; pthread_mutex_t mutex; struct vkd3d_swapchain_pipeline *pipelines; size_t pipelines_size; size_t pipeline_count; }; HRESULT vkd3d_swapchain_ops_init(struct vkd3d_swapchain_ops *meta_swapchain_ops, struct d3d12_device *device); void vkd3d_swapchain_ops_cleanup(struct vkd3d_swapchain_ops *meta_swapchain_ops, struct d3d12_device *device); #define VKD3D_QUERY_OP_WORKGROUP_SIZE (64) struct vkd3d_query_resolve_args { uint32_t dst_index; uint32_t src_index; uint32_t query_count; }; struct vkd3d_query_gather_args { uint32_t query_count; uint32_t entry_offset; }; struct vkd3d_query_gather_info { VkDescriptorSetLayout vk_set_layout; VkPipelineLayout vk_pipeline_layout; VkPipeline vk_pipeline; }; struct vkd3d_query_ops { VkDescriptorSetLayout vk_gather_set_layout; VkPipelineLayout vk_gather_pipeline_layout; VkPipeline vk_gather_occlusion_pipeline; VkPipeline vk_gather_so_statistics_pipeline; VkDescriptorSetLayout vk_resolve_set_layout; VkPipelineLayout vk_resolve_pipeline_layout; VkPipeline vk_resolve_binary_pipeline; }; HRESULT vkd3d_query_ops_init(struct vkd3d_query_ops *meta_query_ops, struct d3d12_device *device); void vkd3d_query_ops_cleanup(struct vkd3d_query_ops *meta_query_ops, struct d3d12_device *device); union vkd3d_predicate_command_direct_args { VkDispatchIndirectCommand dispatch; VkDrawIndirectCommand draw; VkDrawIndexedIndirectCommand draw_indexed; uint32_t draw_count; }; struct vkd3d_predicate_command_args { VkDeviceAddress predicate_va; VkDeviceAddress src_arg_va; VkDeviceAddress dst_arg_va; union vkd3d_predicate_command_direct_args args; }; enum vkd3d_predicate_command_type { VKD3D_PREDICATE_COMMAND_DRAW, VKD3D_PREDICATE_COMMAND_DRAW_INDEXED, VKD3D_PREDICATE_COMMAND_DRAW_INDIRECT, VKD3D_PREDICATE_COMMAND_DRAW_INDIRECT_COUNT, VKD3D_PREDICATE_COMMAND_DISPATCH, VKD3D_PREDICATE_COMMAND_DISPATCH_INDIRECT, VKD3D_PREDICATE_COMMAND_COUNT }; struct vkd3d_predicate_command_info { VkPipelineLayout vk_pipeline_layout; VkPipeline vk_pipeline; uint32_t data_size; }; struct vkd3d_predicate_resolve_args { VkDeviceAddress src_va; VkDeviceAddress dst_va; VkBool32 invert; }; struct vkd3d_predicate_ops { VkPipelineLayout vk_command_pipeline_layout; VkPipelineLayout vk_resolve_pipeline_layout; VkPipeline vk_command_pipelines[VKD3D_PREDICATE_COMMAND_COUNT]; VkPipeline vk_resolve_pipeline; uint32_t data_sizes[VKD3D_PREDICATE_COMMAND_COUNT]; }; HRESULT vkd3d_predicate_ops_init(struct vkd3d_predicate_ops *meta_predicate_ops, struct d3d12_device *device); void vkd3d_predicate_ops_cleanup(struct vkd3d_predicate_ops *meta_predicate_ops, struct d3d12_device *device); struct vkd3d_meta_ops_common { VkShaderModule vk_module_fullscreen_vs; VkShaderModule vk_module_fullscreen_gs; }; struct vkd3d_meta_ops { struct d3d12_device *device; struct vkd3d_meta_ops_common common; struct vkd3d_clear_uav_ops clear_uav; struct vkd3d_copy_image_ops copy_image; struct vkd3d_swapchain_ops swapchain; struct vkd3d_query_ops query; struct vkd3d_predicate_ops predicate; }; HRESULT vkd3d_meta_ops_init(struct vkd3d_meta_ops *meta_ops, struct d3d12_device *device); HRESULT vkd3d_meta_ops_cleanup(struct vkd3d_meta_ops *meta_ops, struct d3d12_device *device); struct vkd3d_clear_uav_pipeline vkd3d_meta_get_clear_buffer_uav_pipeline(struct vkd3d_meta_ops *meta_ops, bool as_uint, bool raw); struct vkd3d_clear_uav_pipeline vkd3d_meta_get_clear_image_uav_pipeline(struct vkd3d_meta_ops *meta_ops, VkImageViewType image_view_type, bool as_uint); VkExtent3D vkd3d_meta_get_clear_image_uav_workgroup_size(VkImageViewType view_type); static inline VkExtent3D vkd3d_meta_get_clear_buffer_uav_workgroup_size() { VkExtent3D result = { 128, 1, 1 }; return result; } HRESULT vkd3d_meta_get_copy_image_pipeline(struct vkd3d_meta_ops *meta_ops, const struct vkd3d_copy_image_pipeline_key *key, struct vkd3d_copy_image_info *info); VkImageViewType vkd3d_meta_get_copy_image_view_type(D3D12_RESOURCE_DIMENSION dim); const struct vkd3d_format *vkd3d_meta_get_copy_image_attachment_format(struct vkd3d_meta_ops *meta_ops, const struct vkd3d_format *dst_format, const struct vkd3d_format *src_format, VkImageAspectFlags dst_aspect, VkImageAspectFlags src_aspect); HRESULT vkd3d_meta_get_swapchain_pipeline(struct vkd3d_meta_ops *meta_ops, const struct vkd3d_swapchain_pipeline_key *key, struct vkd3d_swapchain_info *info); bool vkd3d_meta_get_query_gather_pipeline(struct vkd3d_meta_ops *meta_ops, D3D12_QUERY_HEAP_TYPE heap_type, struct vkd3d_query_gather_info *info); void vkd3d_meta_get_predicate_pipeline(struct vkd3d_meta_ops *meta_ops, enum vkd3d_predicate_command_type command_type, struct vkd3d_predicate_command_info *info); enum vkd3d_time_domain_flag { VKD3D_TIME_DOMAIN_DEVICE = 0x00000001u, VKD3D_TIME_DOMAIN_QPC = 0x00000002u, }; struct vkd3d_physical_device_info { /* properties */ VkPhysicalDeviceDescriptorIndexingPropertiesEXT descriptor_indexing_properties; VkPhysicalDeviceInlineUniformBlockPropertiesEXT inline_uniform_block_properties; VkPhysicalDevicePushDescriptorPropertiesKHR push_descriptor_properties; VkPhysicalDeviceMaintenance3Properties maintenance3_properties; VkPhysicalDeviceTexelBufferAlignmentPropertiesEXT texel_buffer_alignment_properties; VkPhysicalDeviceTransformFeedbackPropertiesEXT xfb_properties; VkPhysicalDeviceVertexAttributeDivisorPropertiesEXT vertex_divisor_properties; VkPhysicalDeviceSubgroupProperties subgroup_properties; VkPhysicalDeviceTimelineSemaphorePropertiesKHR timeline_semaphore_properties; VkPhysicalDeviceSubgroupSizeControlPropertiesEXT subgroup_size_control_properties; VkPhysicalDeviceCustomBorderColorPropertiesEXT custom_border_color_properties; VkPhysicalDeviceShaderCorePropertiesAMD shader_core_properties; VkPhysicalDeviceShaderCoreProperties2AMD shader_core_properties2; VkPhysicalDeviceShaderSMBuiltinsPropertiesNV shader_sm_builtins_properties; VkPhysicalDeviceSamplerFilterMinmaxPropertiesEXT sampler_filter_minmax_properties; VkPhysicalDeviceRobustness2PropertiesEXT robustness2_properties; VkPhysicalDeviceExternalMemoryHostPropertiesEXT external_memory_host_properties; VkPhysicalDeviceRayTracingPipelinePropertiesKHR ray_tracing_pipeline_properties; VkPhysicalDeviceAccelerationStructurePropertiesKHR acceleration_structure_properties; VkPhysicalDeviceFloatControlsPropertiesKHR float_control_properties; VkPhysicalDeviceFragmentShadingRatePropertiesKHR fragment_shading_rate_properties; VkPhysicalDeviceConservativeRasterizationPropertiesEXT conservative_rasterization_properties; VkPhysicalDeviceProperties2KHR properties2; /* features */ VkPhysicalDeviceBufferDeviceAddressFeaturesKHR buffer_device_address_features; VkPhysicalDeviceConditionalRenderingFeaturesEXT conditional_rendering_features; VkPhysicalDeviceDepthClipEnableFeaturesEXT depth_clip_features; VkPhysicalDeviceDescriptorIndexingFeaturesEXT descriptor_indexing_features; VkPhysicalDeviceShaderDemoteToHelperInvocationFeaturesEXT demote_features; VkPhysicalDeviceInlineUniformBlockFeaturesEXT inline_uniform_block_features; VkPhysicalDeviceTexelBufferAlignmentFeaturesEXT texel_buffer_alignment_features; VkPhysicalDeviceTransformFeedbackFeaturesEXT xfb_features; VkPhysicalDeviceVertexAttributeDivisorFeaturesEXT vertex_divisor_features; VkPhysicalDeviceCustomBorderColorFeaturesEXT custom_border_color_features; VkPhysicalDevice4444FormatsFeaturesEXT ext_4444_formats_features; VkPhysicalDeviceTimelineSemaphoreFeaturesKHR timeline_semaphore_features; VkPhysicalDeviceFloat16Int8FeaturesKHR float16_int8_features; VkPhysicalDevice16BitStorageFeatures storage_16bit_features; VkPhysicalDeviceShaderSubgroupExtendedTypesFeaturesKHR subgroup_extended_types_features; VkPhysicalDeviceRobustness2FeaturesEXT robustness2_features; VkPhysicalDeviceExtendedDynamicStateFeaturesEXT extended_dynamic_state_features; VkPhysicalDeviceMutableDescriptorTypeFeaturesVALVE mutable_descriptor_features; VkPhysicalDeviceRayTracingPipelineFeaturesKHR ray_tracing_pipeline_features; VkPhysicalDeviceAccelerationStructureFeaturesKHR acceleration_structure_features; VkPhysicalDeviceFragmentShadingRateFeaturesKHR fragment_shading_rate_features; VkPhysicalDeviceShaderDrawParametersFeatures shader_draw_parameters_features; VkPhysicalDeviceSubgroupSizeControlFeaturesEXT subgroup_size_control_features; VkPhysicalDeviceSeparateDepthStencilLayoutsFeaturesKHR separate_depth_stencil_layout_features; VkPhysicalDeviceFeatures2 features2; /* others, for extensions that have no feature bits */ uint32_t time_domains; /* vkd3d_time_domain_flag */ bool additional_shading_rates_supported; /* d3d12 additional fragment shading rates cap */ }; struct d3d12_caps { D3D12_FEATURE_DATA_D3D12_OPTIONS options; D3D12_FEATURE_DATA_D3D12_OPTIONS1 options1; D3D12_FEATURE_DATA_D3D12_OPTIONS2 options2; D3D12_FEATURE_DATA_D3D12_OPTIONS3 options3; D3D12_FEATURE_DATA_D3D12_OPTIONS4 options4; D3D12_FEATURE_DATA_D3D12_OPTIONS5 options5; D3D12_FEATURE_DATA_D3D12_OPTIONS6 options6; D3D12_FEATURE_DATA_D3D12_OPTIONS7 options7; D3D_FEATURE_LEVEL max_feature_level; D3D_SHADER_MODEL max_shader_model; }; enum vkd3d_queue_family { VKD3D_QUEUE_FAMILY_GRAPHICS, VKD3D_QUEUE_FAMILY_COMPUTE, VKD3D_QUEUE_FAMILY_TRANSFER, VKD3D_QUEUE_FAMILY_SPARSE_BINDING, /* Keep internal queues at the end */ VKD3D_QUEUE_FAMILY_INTERNAL_COMPUTE, VKD3D_QUEUE_FAMILY_COUNT }; struct vkd3d_queue_family_info { struct vkd3d_queue **queues; uint32_t queue_count; uint32_t vk_family_index; uint32_t timestamp_bits; VkQueueFlags vk_queue_flags; }; /* ID3D12Device */ typedef ID3D12Device6 d3d12_device_iface; struct vkd3d_descriptor_qa_global_info; struct vkd3d_descriptor_qa_heap_buffer_data; /* ID3D12DeviceExt */ typedef ID3D12DeviceExt d3d12_device_vkd3d_ext_iface; struct d3d12_device { d3d12_device_iface ID3D12Device_iface; d3d12_device_vkd3d_ext_iface ID3D12DeviceExt_iface; LONG refcount; VkDevice vk_device; uint32_t api_version; VkPhysicalDevice vk_physical_device; struct vkd3d_vk_device_procs vk_procs; PFN_vkd3d_signal_event signal_event; pthread_mutex_t mutex; struct vkd3d_render_pass_cache render_pass_cache; VkPhysicalDeviceMemoryProperties memory_properties; struct vkd3d_vulkan_info vk_info; struct vkd3d_physical_device_info device_info; struct vkd3d_queue_family_info *queue_families[VKD3D_QUEUE_FAMILY_COUNT]; uint32_t queue_family_indices[VKD3D_QUEUE_FAMILY_COUNT]; uint32_t queue_family_count; uint32_t unique_queue_mask; struct vkd3d_instance *vkd3d_instance; IUnknown *parent; LUID adapter_luid; struct vkd3d_private_store private_store; struct d3d12_caps d3d12_caps; struct vkd3d_memory_allocator memory_allocator; struct vkd3d_scratch_buffer scratch_buffers[VKD3D_SCRATCH_BUFFER_COUNT]; size_t scratch_buffer_count; struct vkd3d_query_pool query_pools[VKD3D_VIRTUAL_QUERY_POOL_COUNT]; size_t query_pool_count; uint32_t *descriptor_heap_gpu_vas; size_t descriptor_heap_gpu_va_count; size_t descriptor_heap_gpu_va_size; uint32_t descriptor_heap_gpu_next; HRESULT removed_reason; const struct vkd3d_format *formats; const struct vkd3d_format *depth_stencil_formats; unsigned int format_compatibility_list_count; const struct vkd3d_format_compatibility_list *format_compatibility_lists; struct vkd3d_bindless_state bindless_state; struct vkd3d_memory_info memory_info; struct vkd3d_meta_ops meta_ops; struct vkd3d_view_map sampler_map; struct vkd3d_sampler_state sampler_state; struct vkd3d_shader_debug_ring debug_ring; #ifdef VKD3D_ENABLE_DESCRIPTOR_QA struct vkd3d_descriptor_qa_global_info *descriptor_qa_global_info; #endif }; HRESULT d3d12_device_create(struct vkd3d_instance *instance, const struct vkd3d_device_create_info *create_info, struct d3d12_device **device); struct vkd3d_queue_family_info *d3d12_device_get_vkd3d_queue_family(struct d3d12_device *device, D3D12_COMMAND_LIST_TYPE type); struct vkd3d_queue *d3d12_device_allocate_vkd3d_queue(struct d3d12_device *device, struct vkd3d_queue_family_info *queue_family); void d3d12_device_unmap_vkd3d_queue(struct d3d12_device *device, struct vkd3d_queue *queue); bool d3d12_device_is_uma(struct d3d12_device *device, bool *coherent); void d3d12_device_mark_as_removed(struct d3d12_device *device, HRESULT reason, const char *message, ...) VKD3D_PRINTF_FUNC(3, 4); struct d3d12_device *unsafe_impl_from_ID3D12Device(d3d12_device_iface *iface); bool d3d12_device_validate_shader_meta(struct d3d12_device *device, const struct vkd3d_shader_meta *meta); HRESULT d3d12_device_get_scratch_buffer(struct d3d12_device *device, VkDeviceSize min_size, struct vkd3d_scratch_buffer *scratch); void d3d12_device_return_scratch_buffer(struct d3d12_device *device, const struct vkd3d_scratch_buffer *scratch); HRESULT d3d12_device_get_query_pool(struct d3d12_device *device, uint32_t type_index, struct vkd3d_query_pool *pool); void d3d12_device_return_query_pool(struct d3d12_device *device, const struct vkd3d_query_pool *pool); uint64_t d3d12_device_get_descriptor_heap_gpu_va(struct d3d12_device *device); void d3d12_device_return_descriptor_heap_gpu_va(struct d3d12_device *device, uint64_t va); static inline HRESULT d3d12_device_query_interface(struct d3d12_device *device, REFIID iid, void **object) { return ID3D12Device6_QueryInterface(&device->ID3D12Device_iface, iid, object); } static inline ULONG d3d12_device_add_ref(struct d3d12_device *device) { return ID3D12Device6_AddRef(&device->ID3D12Device_iface); } static inline ULONG d3d12_device_release(struct d3d12_device *device) { return ID3D12Device6_Release(&device->ID3D12Device_iface); } static inline unsigned int d3d12_device_get_descriptor_handle_increment_size(struct d3d12_device *device, D3D12_DESCRIPTOR_HEAP_TYPE descriptor_type) { return ID3D12Device6_GetDescriptorHandleIncrementSize(&device->ID3D12Device_iface, descriptor_type); } static inline bool d3d12_device_use_ssbo_raw_buffer(struct d3d12_device *device) { return (device->bindless_state.flags & VKD3D_BINDLESS_RAW_SSBO) != 0; } static inline VkDeviceSize d3d12_device_get_ssbo_alignment(struct d3d12_device *device) { return device->device_info.properties2.properties.limits.minStorageBufferOffsetAlignment; } static inline bool d3d12_device_use_ssbo_root_descriptors(struct d3d12_device *device) { /* We only know the VA of root SRV/UAVs, so we cannot * make any better assumptions about the alignment */ return d3d12_device_use_ssbo_raw_buffer(device) && d3d12_device_get_ssbo_alignment(device) <= 4; } bool d3d12_device_supports_variable_shading_rate_tier_1(struct d3d12_device *device); bool d3d12_device_supports_ray_tracing_tier_1_0(const struct d3d12_device *device); /* ID3DBlob */ struct d3d_blob { ID3D10Blob ID3DBlob_iface; LONG refcount; void *buffer; SIZE_T size; }; HRESULT d3d_blob_create(void *buffer, SIZE_T size, struct d3d_blob **blob); /* ID3D12StateObject */ typedef ID3D12StateObject d3d12_state_object_iface; typedef ID3D12StateObjectProperties d3d12_state_object_properties_iface; struct d3d12_state_object_identifier { WCHAR *mangled_export; WCHAR *plain_export; /* Must be a persistent pointer as long as the StateObject object is live. */ uint8_t identifier[D3D12_SHADER_IDENTIFIER_SIZE_IN_BYTES]; /* The index into pStages[]. */ uint32_t general_stage_index; uint32_t closest_stage_index; uint32_t anyhit_stage_index; uint32_t intersection_stage_index; VkShaderStageFlagBits general_stage; VkDeviceSize stack_size_general; VkDeviceSize stack_size_closest; VkDeviceSize stack_size_any; VkDeviceSize stack_size_intersection; /* The index into vkGetShaderStackSize and friends for pGroups[]. */ uint32_t group_index; }; struct d3d12_state_object_stack_info { uint32_t max_callable; uint32_t max_anyhit; uint32_t max_miss; uint32_t max_raygen; uint32_t max_intersect; uint32_t max_closest; }; struct d3d12_state_object { d3d12_state_object_iface ID3D12StateObject_iface; d3d12_state_object_properties_iface ID3D12StateObjectProperties_iface; LONG refcount; LONG internal_refcount; D3D12_STATE_OBJECT_TYPE type; D3D12_STATE_OBJECT_FLAGS flags; struct d3d12_device *device; /* Could potentially be a hashmap. */ struct d3d12_state_object_identifier *exports; size_t exports_size; size_t exports_count; struct vkd3d_shader_library_entry_point *entry_points; size_t entry_points_count; size_t stages_count; /* Normally stages_count == entry_points_count, but entry_points is the entry points we * export externally, and stages_count matches pStages[] size for purposes of index fixups. */ VkPipeline pipeline; UINT64 pipeline_stack_size; struct d3d12_state_object_stack_info stack; struct d3d12_state_object **collections; size_t collections_count; struct vkd3d_private_store private_store; }; HRESULT d3d12_state_object_create(struct d3d12_device *device, const D3D12_STATE_OBJECT_DESC *desc, struct d3d12_state_object **object); struct d3d12_state_object *impl_from_ID3D12StateObject(ID3D12StateObject *iface); /* utils */ enum vkd3d_format_type { VKD3D_FORMAT_TYPE_OTHER, VKD3D_FORMAT_TYPE_TYPELESS, VKD3D_FORMAT_TYPE_SINT, VKD3D_FORMAT_TYPE_UINT, }; struct vkd3d_format { DXGI_FORMAT dxgi_format; VkFormat vk_format; size_t byte_count; size_t block_width; size_t block_height; size_t block_byte_count; VkImageAspectFlags vk_aspect_mask; unsigned int plane_count; enum vkd3d_format_type type; bool is_emulated; }; static inline size_t vkd3d_format_get_data_offset(const struct vkd3d_format *format, unsigned int row_pitch, unsigned int slice_pitch, unsigned int x, unsigned int y, unsigned int z) { return z * slice_pitch + (y / format->block_height) * row_pitch + (x / format->block_width) * format->byte_count * format->block_byte_count; } static inline bool vkd3d_format_is_compressed(const struct vkd3d_format *format) { return format->block_byte_count != 1; } void vkd3d_format_copy_data(const struct vkd3d_format *format, const uint8_t *src, unsigned int src_row_pitch, unsigned int src_slice_pitch, uint8_t *dst, unsigned int dst_row_pitch, unsigned int dst_slice_pitch, unsigned int w, unsigned int h, unsigned int d); const struct vkd3d_format *vkd3d_get_format(const struct d3d12_device *device, DXGI_FORMAT dxgi_format, bool depth_stencil); DXGI_FORMAT vkd3d_get_typeless_format(const struct d3d12_device *device, DXGI_FORMAT dxgi_format); const struct vkd3d_format *vkd3d_find_uint_format(const struct d3d12_device *device, DXGI_FORMAT dxgi_format); VkFormat vkd3d_internal_get_vk_format(const struct d3d12_device *device, DXGI_FORMAT dxgi_format); HRESULT vkd3d_init_format_info(struct d3d12_device *device); void vkd3d_cleanup_format_info(struct d3d12_device *device); static inline const struct vkd3d_format *vkd3d_format_from_d3d12_resource_desc( const struct d3d12_device *device, const D3D12_RESOURCE_DESC *desc, DXGI_FORMAT view_format) { return vkd3d_get_format(device, view_format ? view_format : desc->Format, desc->Flags & D3D12_RESOURCE_FLAG_ALLOW_DEPTH_STENCIL); } static inline VkImageSubresourceRange vk_subresource_range_from_layers(const VkImageSubresourceLayers *layers) { VkImageSubresourceRange range; range.aspectMask = layers->aspectMask; range.baseMipLevel = layers->mipLevel; range.levelCount = 1; range.baseArrayLayer = layers->baseArrayLayer; range.layerCount = layers->layerCount; return range; } static inline VkImageSubresourceLayers vk_subresource_layers_from_subresource(const VkImageSubresource *subresource) { VkImageSubresourceLayers layers; layers.aspectMask = subresource->aspectMask; layers.mipLevel = subresource->mipLevel; layers.baseArrayLayer = subresource->arrayLayer; layers.layerCount = 1; return layers; } static inline VkImageSubresourceLayers vk_subresource_layers_from_view(const struct vkd3d_view *view) { VkImageSubresourceLayers layers; layers.aspectMask = view->format->vk_aspect_mask; layers.mipLevel = view->info.texture.miplevel_idx; layers.baseArrayLayer = view->info.texture.layer_idx; layers.layerCount = view->info.texture.layer_count; return layers; } static inline VkImageSubresourceRange vk_subresource_range_from_view(const struct vkd3d_view *view) { VkImageSubresourceLayers layers = vk_subresource_layers_from_view(view); return vk_subresource_range_from_layers(&layers); } static inline bool d3d12_box_is_empty(const D3D12_BOX *box) { return box->right <= box->left || box->bottom <= box->top || box->back <= box->front; } static inline unsigned int d3d12_resource_desc_get_width(const D3D12_RESOURCE_DESC *desc, unsigned int miplevel_idx) { return max(1, desc->Width >> miplevel_idx); } static inline unsigned int d3d12_resource_desc_get_height(const D3D12_RESOURCE_DESC *desc, unsigned int miplevel_idx) { return max(1, desc->Height >> miplevel_idx); } static inline unsigned int d3d12_resource_desc_get_depth(const D3D12_RESOURCE_DESC *desc, unsigned int miplevel_idx) { unsigned int d = desc->Dimension != D3D12_RESOURCE_DIMENSION_TEXTURE3D ? 1 : desc->DepthOrArraySize; return max(1, d >> miplevel_idx); } static inline unsigned int d3d12_resource_desc_get_layer_count(const D3D12_RESOURCE_DESC *desc) { return desc->Dimension != D3D12_RESOURCE_DIMENSION_TEXTURE3D ? desc->DepthOrArraySize : 1; } static inline unsigned int d3d12_resource_desc_get_sub_resource_count_per_plane(const D3D12_RESOURCE_DESC *desc) { return d3d12_resource_desc_get_layer_count(desc) * desc->MipLevels; } static inline unsigned int d3d12_resource_get_sub_resource_count(const struct d3d12_resource *resource) { return d3d12_resource_desc_get_sub_resource_count_per_plane(&resource->desc) * (resource->format ? vkd3d_popcount(resource->format->vk_aspect_mask) : 1); } VkDeviceAddress vkd3d_get_buffer_device_address(struct d3d12_device *device, VkBuffer vk_buffer); VkDeviceAddress vkd3d_get_acceleration_structure_device_address(struct d3d12_device *device, VkAccelerationStructureKHR vk_acceleration_structure); static inline VkDeviceAddress d3d12_resource_get_va(const struct d3d12_resource *resource, VkDeviceSize offset) { return vkd3d_get_buffer_device_address(resource->device, resource->res.vk_buffer) + resource->mem.offset + offset; } static inline unsigned int vkd3d_compute_workgroup_count(unsigned int thread_count, unsigned int workgroup_size) { return (thread_count + workgroup_size - 1) / workgroup_size; } VkCompareOp vk_compare_op_from_d3d12(D3D12_COMPARISON_FUNC op); VkSampleCountFlagBits vk_samples_from_dxgi_sample_desc(const DXGI_SAMPLE_DESC *desc); VkSampleCountFlagBits vk_samples_from_sample_count(unsigned int sample_count); bool is_valid_feature_level(D3D_FEATURE_LEVEL feature_level); bool is_valid_resource_state(D3D12_RESOURCE_STATES state); bool is_write_resource_state(D3D12_RESOURCE_STATES state); HRESULT return_interface(void *iface, REFIID iface_iid, REFIID requested_iid, void **object); const char *debug_dxgi_format(DXGI_FORMAT format); const char *debug_d3d12_box(const D3D12_BOX *box); const char *debug_d3d12_shader_component_mapping(unsigned int mapping); const char *debug_vk_extent_3d(VkExtent3D extent); #define VKD3D_DEBUG_FLAGS_BUFFER_SIZE 1024 const char *debug_vk_memory_heap_flags(VkMemoryHeapFlags flags, char buffer[VKD3D_DEBUG_FLAGS_BUFFER_SIZE]); const char *debug_vk_memory_property_flags(VkMemoryPropertyFlags flags, char buffer[VKD3D_DEBUG_FLAGS_BUFFER_SIZE]); const char *debug_vk_queue_flags(VkQueueFlags flags, char buffer[VKD3D_DEBUG_FLAGS_BUFFER_SIZE]); static inline void debug_ignored_node_mask(unsigned int mask) { if (mask && mask != 1) FIXME("Ignoring node mask 0x%08x.\n", mask); } HRESULT vkd3d_load_vk_global_procs(struct vkd3d_vk_global_procs *procs, PFN_vkGetInstanceProcAddr vkGetInstanceProcAddr); HRESULT vkd3d_load_vk_instance_procs(struct vkd3d_vk_instance_procs *procs, const struct vkd3d_vk_global_procs *global_procs, VkInstance instance); HRESULT vkd3d_load_vk_device_procs(struct vkd3d_vk_device_procs *procs, const struct vkd3d_vk_instance_procs *parent_procs, VkDevice device); HRESULT vkd3d_set_vk_object_name(struct d3d12_device *device, uint64_t vk_object, VkObjectType vk_object_type, const char *name); enum VkPrimitiveTopology vk_topology_from_d3d12_topology(D3D12_PRIMITIVE_TOPOLOGY topology); static inline void vk_prepend_struct(void *header, void *structure) { VkBaseOutStructure *vk_header = header, *vk_structure = structure; assert(!vk_structure->pNext); vk_structure->pNext = vk_header->pNext; vk_header->pNext = vk_structure; } #define VKD3D_NULL_BUFFER_SIZE 16 struct vkd3d_view_key { enum vkd3d_view_type view_type; union { struct vkd3d_buffer_view_desc buffer; struct vkd3d_texture_view_desc texture; D3D12_SAMPLER_DESC sampler; } u; }; struct vkd3d_view *vkd3d_view_map_create_view(struct vkd3d_view_map *view_map, struct d3d12_device *device, const struct vkd3d_view_key *key); /* Acceleration structure helpers. */ struct vkd3d_acceleration_structure_build_info { /* This is not a hard limit, just an arbitrary value which lets us avoid allocation for * the common case. */ #define VKD3D_BUILD_INFO_STACK_COUNT 16 const struct VkAccelerationStructureBuildRangeInfoKHR *build_range_ptr_stack[VKD3D_BUILD_INFO_STACK_COUNT]; VkAccelerationStructureBuildRangeInfoKHR build_range_stack[VKD3D_BUILD_INFO_STACK_COUNT]; VkAccelerationStructureGeometryKHR geometries_stack[VKD3D_BUILD_INFO_STACK_COUNT]; const VkAccelerationStructureBuildRangeInfoKHR **build_range_ptrs; uint32_t primitive_counts_stack[VKD3D_BUILD_INFO_STACK_COUNT]; VkAccelerationStructureBuildRangeInfoKHR *build_ranges; VkAccelerationStructureBuildGeometryInfoKHR build_info; VkAccelerationStructureGeometryKHR *geometries; uint32_t *primitive_counts; }; void vkd3d_acceleration_structure_build_info_cleanup( struct vkd3d_acceleration_structure_build_info *info); bool vkd3d_acceleration_structure_convert_inputs(const struct d3d12_device *device, struct vkd3d_acceleration_structure_build_info *info, const D3D12_BUILD_RAYTRACING_ACCELERATION_STRUCTURE_INPUTS *desc); void vkd3d_acceleration_structure_emit_postbuild_info( struct d3d12_command_list *list, const D3D12_RAYTRACING_ACCELERATION_STRUCTURE_POSTBUILD_INFO_DESC *desc, uint32_t count, const D3D12_GPU_VIRTUAL_ADDRESS *addresses); void vkd3d_acceleration_structure_emit_immediate_postbuild_info( struct d3d12_command_list *list, uint32_t count, const D3D12_RAYTRACING_ACCELERATION_STRUCTURE_POSTBUILD_INFO_DESC *desc, VkAccelerationStructureKHR vk_acceleration_structure); void vkd3d_acceleration_structure_copy( struct d3d12_command_list *list, D3D12_GPU_VIRTUAL_ADDRESS dst, D3D12_GPU_VIRTUAL_ADDRESS src, D3D12_RAYTRACING_ACCELERATION_STRUCTURE_COPY_MODE mode); #define VKD3D_VENDOR_ID_NVIDIA 0x10DE #define VKD3D_VENDOR_ID_AMD 0x1002 #define VKD3D_VENDOR_ID_INTEL 0x8086 #endif /* __VKD3D_PRIVATE_H */