mesa/src/microsoft/vulkan/dzn_cmd_buffer.cpp

4120 lines
154 KiB
C++
Raw Blame History

This file contains invisible Unicode characters

This file contains invisible Unicode characters that are indistinguishable to humans but may be processed differently by a computer. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.

/*
* Copyright © Microsoft Corporation
*
* Permission is hereby granted, free of charge, to any person obtaining a
* copy of this software and associated documentation files (the "Software"),
* to deal in the Software without restriction, including without limitation
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
* and/or sell copies of the Software, and to permit persons to whom the
* Software is furnished to do so, subject to the following conditions:
*
* The above copyright notice and this permission notice (including the next
* paragraph) shall be included in all copies or substantial portions of the
* Software.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
* IN THE SOFTWARE.
*/
#include "dzn_private.h"
#include "vk_alloc.h"
#include "vk_debug_report.h"
#include "vk_format.h"
#include "vk_util.h"
static void
dzn_cmd_buffer_destroy(struct vk_command_buffer *cbuf)
{
if (!cbuf)
return;
dzn_cmd_buffer *cmdbuf = container_of(cbuf, dzn_cmd_buffer, vk);
dzn_device *device = container_of(cbuf->base.device, dzn_device, vk);
if (cmdbuf->cmdlist)
cmdbuf->cmdlist->Release();
if (cmdbuf->cmdalloc)
cmdbuf->cmdalloc->Release();
list_for_each_entry_safe(dzn_internal_resource, res, &cmdbuf->internal_bufs, link) {
list_del(&res->link);
res->res->Release();
vk_free(&cbuf->pool->alloc, res);
}
dzn_descriptor_heap_pool_finish(&cmdbuf->cbv_srv_uav_pool);
dzn_descriptor_heap_pool_finish(&cmdbuf->sampler_pool);
dzn_descriptor_heap_pool_finish(&cmdbuf->rtvs.pool);
dzn_descriptor_heap_pool_finish(&cmdbuf->dsvs.pool);
util_dynarray_fini(&cmdbuf->events.wait);
util_dynarray_fini(&cmdbuf->events.signal);
util_dynarray_fini(&cmdbuf->queries.reset);
util_dynarray_fini(&cmdbuf->queries.wait);
util_dynarray_fini(&cmdbuf->queries.signal);
if (cmdbuf->rtvs.ht) {
hash_table_foreach(cmdbuf->rtvs.ht, he)
vk_free(&cbuf->pool->alloc, he->data);
_mesa_hash_table_destroy(cmdbuf->rtvs.ht, NULL);
}
if (cmdbuf->dsvs.ht) {
hash_table_foreach(cmdbuf->dsvs.ht, he)
vk_free(&cbuf->pool->alloc, he->data);
_mesa_hash_table_destroy(cmdbuf->dsvs.ht, NULL);
}
if (cmdbuf->events.ht)
_mesa_hash_table_destroy(cmdbuf->events.ht, NULL);
if (cmdbuf->queries.ht) {
hash_table_foreach(cmdbuf->queries.ht, he) {
dzn_cmd_buffer_query_pool_state *qpstate =
(dzn_cmd_buffer_query_pool_state *)he->data;
util_dynarray_fini(&qpstate->reset);
util_dynarray_fini(&qpstate->collect);
util_dynarray_fini(&qpstate->wait);
util_dynarray_fini(&qpstate->signal);
vk_free(&cbuf->pool->alloc, he->data);
}
_mesa_hash_table_destroy(cmdbuf->queries.ht, NULL);
}
vk_command_buffer_finish(&cmdbuf->vk);
vk_free(&cbuf->pool->alloc, cmdbuf);
}
static uint32_t
dzn_cmd_buffer_rtv_key_hash_function(const void *key)
{
return _mesa_hash_data(key, sizeof(dzn_cmd_buffer_rtv_key));
}
static bool
dzn_cmd_buffer_rtv_key_equals_function(const void *a, const void *b)
{
return memcmp(a, b, sizeof(dzn_cmd_buffer_rtv_key)) == 0;
}
static uint32_t
dzn_cmd_buffer_dsv_key_hash_function(const void *key)
{
return _mesa_hash_data(key, sizeof(dzn_cmd_buffer_dsv_key));
}
static bool
dzn_cmd_buffer_dsv_key_equals_function(const void *a, const void *b)
{
return memcmp(a, b, sizeof(dzn_cmd_buffer_dsv_key)) == 0;
}
static VkResult
dzn_cmd_buffer_create(const VkCommandBufferAllocateInfo *info,
VkCommandBuffer *out)
{
VK_FROM_HANDLE(vk_command_pool, pool, info->commandPool);
dzn_device *device = container_of(pool->base.device, dzn_device, vk);
dzn_physical_device *pdev =
container_of(device->vk.physical, dzn_physical_device, vk);
assert(pool->queue_family_index < pdev->queue_family_count);
D3D12_COMMAND_LIST_TYPE type =
pdev->queue_families[pool->queue_family_index].desc.Type;
dzn_cmd_buffer *cmdbuf = (dzn_cmd_buffer *)
vk_zalloc(&pool->alloc, sizeof(*cmdbuf), 8, VK_SYSTEM_ALLOCATION_SCOPE_OBJECT);
if (!cmdbuf)
return vk_error(pool->base.device, VK_ERROR_OUT_OF_HOST_MEMORY);
VkResult result =
vk_command_buffer_init(&cmdbuf->vk, pool, info->level);
if (result != VK_SUCCESS) {
vk_free(&pool->alloc, cmdbuf);
return result;
}
memset(&cmdbuf->state, 0, sizeof(cmdbuf->state));
list_inithead(&cmdbuf->internal_bufs);
util_dynarray_init(&cmdbuf->events.wait, NULL);
util_dynarray_init(&cmdbuf->events.signal, NULL);
util_dynarray_init(&cmdbuf->queries.reset, NULL);
util_dynarray_init(&cmdbuf->queries.wait, NULL);
util_dynarray_init(&cmdbuf->queries.signal, NULL);
dzn_descriptor_heap_pool_init(&cmdbuf->rtvs.pool, device,
D3D12_DESCRIPTOR_HEAP_TYPE_RTV,
false, &pool->alloc);
dzn_descriptor_heap_pool_init(&cmdbuf->dsvs.pool, device,
D3D12_DESCRIPTOR_HEAP_TYPE_DSV,
false, &pool->alloc);
dzn_descriptor_heap_pool_init(&cmdbuf->cbv_srv_uav_pool, device,
D3D12_DESCRIPTOR_HEAP_TYPE_CBV_SRV_UAV,
true, &pool->alloc);
dzn_descriptor_heap_pool_init(&cmdbuf->sampler_pool, device,
D3D12_DESCRIPTOR_HEAP_TYPE_SAMPLER,
true, &pool->alloc);
cmdbuf->events.ht =
_mesa_pointer_hash_table_create(NULL);
cmdbuf->queries.ht =
_mesa_pointer_hash_table_create(NULL);
cmdbuf->rtvs.ht =
_mesa_hash_table_create(NULL,
dzn_cmd_buffer_rtv_key_hash_function,
dzn_cmd_buffer_rtv_key_equals_function);
cmdbuf->dsvs.ht =
_mesa_hash_table_create(NULL,
dzn_cmd_buffer_dsv_key_hash_function,
dzn_cmd_buffer_dsv_key_equals_function);
if (!cmdbuf->events.ht || !cmdbuf->queries.ht ||
!cmdbuf->rtvs.ht || !cmdbuf->dsvs.ht) {
result = vk_error(device, VK_ERROR_OUT_OF_HOST_MEMORY);
goto out;
}
cmdbuf->vk.destroy = dzn_cmd_buffer_destroy;
if (FAILED(device->dev->CreateCommandAllocator(type,
IID_PPV_ARGS(&cmdbuf->cmdalloc)))) {
result = vk_error(device, VK_ERROR_OUT_OF_HOST_MEMORY);
goto out;
}
if (FAILED(device->dev->CreateCommandList(0, type,
cmdbuf->cmdalloc, NULL,
IID_PPV_ARGS(&cmdbuf->cmdlist)))) {
result = vk_error(device, VK_ERROR_OUT_OF_HOST_MEMORY);
goto out;
}
out:
if (result != VK_SUCCESS)
dzn_cmd_buffer_destroy(&cmdbuf->vk);
else
*out = dzn_cmd_buffer_to_handle(cmdbuf);
return result;
}
VkResult
dzn_cmd_buffer_reset(dzn_cmd_buffer *cmdbuf)
{
dzn_device *device = container_of(cmdbuf->vk.base.device, dzn_device, vk);
/* Reset the state */
memset(&cmdbuf->state, 0, sizeof(cmdbuf->state));
/* TODO: Return resources to the pool */
list_for_each_entry_safe(dzn_internal_resource, res, &cmdbuf->internal_bufs, link) {
list_del(&res->link);
res->res->Release();
vk_free(&cmdbuf->vk.pool->alloc, res);
}
cmdbuf->error = VK_SUCCESS;
util_dynarray_clear(&cmdbuf->events.wait);
util_dynarray_clear(&cmdbuf->events.signal);
util_dynarray_clear(&cmdbuf->queries.reset);
util_dynarray_clear(&cmdbuf->queries.wait);
util_dynarray_clear(&cmdbuf->queries.signal);
hash_table_foreach(cmdbuf->rtvs.ht, he)
vk_free(&cmdbuf->vk.pool->alloc, he->data);
_mesa_hash_table_clear(cmdbuf->rtvs.ht, NULL);
dzn_descriptor_heap_pool_reset(&cmdbuf->rtvs.pool);
hash_table_foreach(cmdbuf->dsvs.ht, he)
vk_free(&cmdbuf->vk.pool->alloc, he->data);
_mesa_hash_table_clear(cmdbuf->dsvs.ht, NULL);
hash_table_foreach(cmdbuf->queries.ht, he) {
dzn_cmd_buffer_query_pool_state *qpstate =
(dzn_cmd_buffer_query_pool_state *)he->data;
util_dynarray_fini(&qpstate->reset);
util_dynarray_fini(&qpstate->collect);
util_dynarray_fini(&qpstate->wait);
util_dynarray_fini(&qpstate->signal);
vk_free(&cmdbuf->vk.pool->alloc, he->data);
}
_mesa_hash_table_clear(cmdbuf->queries.ht, NULL);
_mesa_hash_table_clear(cmdbuf->events.ht, NULL);
dzn_descriptor_heap_pool_reset(&cmdbuf->dsvs.pool);
dzn_descriptor_heap_pool_reset(&cmdbuf->cbv_srv_uav_pool);
dzn_descriptor_heap_pool_reset(&cmdbuf->sampler_pool);
vk_command_buffer_reset(&cmdbuf->vk);
/* cmdlist->Reset() doesn't return the memory back the the command list
* allocator, and cmdalloc->Reset() can only be called if there's no live
* cmdlist allocated from the allocator, so we need to release and create
* a new command list.
*/
cmdbuf->cmdlist->Release();
cmdbuf->cmdlist = NULL;
cmdbuf->cmdalloc->Reset();
if (FAILED(device->dev->CreateCommandList(0, D3D12_COMMAND_LIST_TYPE_DIRECT,
cmdbuf->cmdalloc, NULL,
IID_PPV_ARGS(&cmdbuf->cmdlist)))) {
cmdbuf->error = vk_error(device, VK_ERROR_OUT_OF_HOST_MEMORY);
}
return cmdbuf->error;
}
VKAPI_ATTR VkResult VKAPI_CALL
dzn_AllocateCommandBuffers(VkDevice device,
const VkCommandBufferAllocateInfo *pAllocateInfo,
VkCommandBuffer *pCommandBuffers)
{
VK_FROM_HANDLE(vk_command_pool, pool, pAllocateInfo->commandPool);
VK_FROM_HANDLE(dzn_device, dev, device);
VkResult result = VK_SUCCESS;
uint32_t i;
for (i = 0; i < pAllocateInfo->commandBufferCount; i++) {
result = dzn_cmd_buffer_create(pAllocateInfo,
&pCommandBuffers[i]);
if (result != VK_SUCCESS)
break;
}
if (result != VK_SUCCESS) {
dev->vk.dispatch_table.FreeCommandBuffers(device, pAllocateInfo->commandPool,
i, pCommandBuffers);
for (i = 0; i < pAllocateInfo->commandBufferCount; i++)
pCommandBuffers[i] = VK_NULL_HANDLE;
}
return result;
}
VKAPI_ATTR VkResult VKAPI_CALL
dzn_ResetCommandBuffer(VkCommandBuffer commandBuffer,
VkCommandBufferResetFlags flags)
{
VK_FROM_HANDLE(dzn_cmd_buffer, cmdbuf, commandBuffer);
return dzn_cmd_buffer_reset(cmdbuf);
}
VkResult
dzn_BeginCommandBuffer(VkCommandBuffer commandBuffer,
const VkCommandBufferBeginInfo *info)
{
VK_FROM_HANDLE(dzn_cmd_buffer, cmdbuf, commandBuffer);
/* If this is the first vkBeginCommandBuffer, we must *initialize* the
* command buffer's state. Otherwise, we must *reset* its state. In both
* cases we reset it.
*
* From the Vulkan 1.0 spec:
*
* If a command buffer is in the executable state and the command buffer
* was allocated from a command pool with the
* VK_COMMAND_POOL_CREATE_RESET_COMMAND_BUFFER_BIT flag set, then
* vkBeginCommandBuffer implicitly resets the command buffer, behaving
* as if vkResetCommandBuffer had been called with
* VK_COMMAND_BUFFER_RESET_RELEASE_RESOURCES_BIT not set. It then puts
* the command buffer in the recording state.
*/
return dzn_cmd_buffer_reset(cmdbuf);
}
static void
dzn_cmd_buffer_gather_events(dzn_cmd_buffer *cmdbuf)
{
dzn_device *device = container_of(cmdbuf->vk.base.device, dzn_device, vk);
if (cmdbuf->error != VK_SUCCESS)
goto out;
hash_table_foreach(cmdbuf->events.ht, he) {
enum dzn_event_state state = (enum dzn_event_state)(uintptr_t)he->data;
if (state != DZN_EVENT_STATE_EXTERNAL_WAIT) {
dzn_cmd_event_signal signal = { (dzn_event *)he->key, state == DZN_EVENT_STATE_SET };
dzn_cmd_event_signal *entry = (dzn_cmd_event_signal *)
util_dynarray_grow(&cmdbuf->events.signal, dzn_cmd_event_signal, 1);
if (!entry) {
cmdbuf->error = vk_error(device, VK_ERROR_OUT_OF_HOST_MEMORY);
break;
}
*entry = signal;
}
}
out:
_mesa_hash_table_clear(cmdbuf->events.ht, NULL);
}
static VkResult
dzn_cmd_buffer_dynbitset_reserve(dzn_cmd_buffer *cmdbuf, struct util_dynarray *array, uint32_t bit)
{
dzn_device *device = container_of(cmdbuf->vk.base.device, dzn_device, vk);
if (bit < util_dynarray_num_elements(array, BITSET_WORD) * BITSET_WORDBITS)
return VK_SUCCESS;
unsigned old_sz = array->size;
void *ptr = util_dynarray_grow(array, BITSET_WORD, (bit + BITSET_WORDBITS) / BITSET_WORDBITS);
if (!ptr) {
cmdbuf->error = vk_error(device, VK_ERROR_OUT_OF_HOST_MEMORY);
return cmdbuf->error;
}
memset(ptr, 0, array->size - old_sz);
return VK_SUCCESS;
}
static bool
dzn_cmd_buffer_dynbitset_test(struct util_dynarray *array, uint32_t bit)
{
uint32_t nbits = util_dynarray_num_elements(array, BITSET_WORD) * BITSET_WORDBITS;
if (bit < nbits)
return BITSET_TEST(util_dynarray_element(array, BITSET_WORD, 0), bit);
return false;
}
static VkResult
dzn_cmd_buffer_dynbitset_set(dzn_cmd_buffer *cmdbuf, struct util_dynarray *array, uint32_t bit)
{
dzn_device *device = container_of(cmdbuf->vk.base.device, dzn_device, vk);
VkResult result = dzn_cmd_buffer_dynbitset_reserve(cmdbuf, array, bit);
if (result != VK_SUCCESS)
return result;
BITSET_SET(util_dynarray_element(array, BITSET_WORD, 0), bit);
return VK_SUCCESS;
}
static void
dzn_cmd_buffer_dynbitset_clear(dzn_cmd_buffer *cmdbuf, struct util_dynarray *array, uint32_t bit)
{
dzn_device *device = container_of(cmdbuf->vk.base.device, dzn_device, vk);
if (bit >= util_dynarray_num_elements(array, BITSET_WORD) * BITSET_WORDBITS)
return;
BITSET_CLEAR(util_dynarray_element(array, BITSET_WORD, 0), bit);
}
static VkResult
dzn_cmd_buffer_dynbitset_set_range(dzn_cmd_buffer *cmdbuf, struct util_dynarray *array,
uint32_t bit, uint32_t count)
{
dzn_device *device = container_of(cmdbuf->vk.base.device, dzn_device, vk);
VkResult result = dzn_cmd_buffer_dynbitset_reserve(cmdbuf, array, bit + count - 1);
if (result != VK_SUCCESS)
return result;
BITSET_SET_RANGE(util_dynarray_element(array, BITSET_WORD, 0), bit, bit + count - 1);
return VK_SUCCESS;
}
static void
dzn_cmd_buffer_dynbitset_clear_range(dzn_cmd_buffer *cmdbuf, struct util_dynarray *array,
uint32_t bit, uint32_t count)
{
dzn_device *device = container_of(cmdbuf->vk.base.device, dzn_device, vk);
uint32_t nbits = util_dynarray_num_elements(array, BITSET_WORD) * BITSET_WORDBITS;
if (!nbits)
return;
uint32_t end = MIN2(bit + count, nbits) - 1;
while (bit <= end) {
uint32_t subcount = MIN2(end + 1 - bit, 32 - (bit % 32));
BITSET_CLEAR_RANGE(util_dynarray_element(array, BITSET_WORD, 0), bit, bit + subcount - 1);
bit += subcount;
}
}
static dzn_cmd_buffer_query_pool_state *
dzn_cmd_buffer_create_query_pool_state(dzn_cmd_buffer *cmdbuf)
{
dzn_device *device = container_of(cmdbuf->vk.base.device, dzn_device, vk);
dzn_cmd_buffer_query_pool_state *state = (dzn_cmd_buffer_query_pool_state *)
vk_alloc(&cmdbuf->vk.pool->alloc, sizeof(*state),
8, VK_SYSTEM_ALLOCATION_SCOPE_OBJECT);
if (!state) {
cmdbuf->error = vk_error(device, VK_ERROR_OUT_OF_HOST_MEMORY);
return NULL;
}
util_dynarray_init(&state->reset, NULL);
util_dynarray_init(&state->collect, NULL);
util_dynarray_init(&state->wait, NULL);
util_dynarray_init(&state->signal, NULL);
return state;
}
static void
dzn_cmd_buffer_destroy_query_pool_state(dzn_cmd_buffer *cmdbuf,
dzn_cmd_buffer_query_pool_state *state)
{
util_dynarray_fini(&state->reset);
util_dynarray_fini(&state->collect);
util_dynarray_fini(&state->wait);
util_dynarray_fini(&state->signal);
vk_free(&cmdbuf->vk.pool->alloc, state);
}
static dzn_cmd_buffer_query_pool_state *
dzn_cmd_buffer_get_query_pool_state(dzn_cmd_buffer *cmdbuf,
dzn_query_pool *qpool)
{
dzn_device *device = container_of(cmdbuf->vk.base.device, dzn_device, vk);
dzn_cmd_buffer_query_pool_state *state = NULL;
struct hash_entry *he =
_mesa_hash_table_search(cmdbuf->queries.ht, qpool);
if (!he) {
state = dzn_cmd_buffer_create_query_pool_state(cmdbuf);
if (!state)
return NULL;
he = _mesa_hash_table_insert(cmdbuf->queries.ht, qpool, state);
if (!he) {
dzn_cmd_buffer_destroy_query_pool_state(cmdbuf, state);
cmdbuf->error = vk_error(device, VK_ERROR_OUT_OF_HOST_MEMORY);
return NULL;
}
} else {
state = (dzn_cmd_buffer_query_pool_state *)he->data;
}
return state;
}
static VkResult
dzn_cmd_buffer_collect_queries(dzn_cmd_buffer *cmdbuf,
const dzn_query_pool *qpool,
dzn_cmd_buffer_query_pool_state *state,
uint32_t first_query,
uint32_t query_count)
{
dzn_device *device = container_of(cmdbuf->vk.base.device, dzn_device, vk);
uint32_t nbits = util_dynarray_num_elements(&state->collect, BITSET_WORD) * BITSET_WORDBITS;
uint32_t start, end;
query_count = MIN2(query_count, nbits - first_query);
nbits = MIN2(first_query + query_count, nbits);
VkResult result =
dzn_cmd_buffer_dynbitset_reserve(cmdbuf, &state->signal, first_query + query_count - 1);
if (result != VK_SUCCESS)
return result;
BITSET_WORD *collect =
util_dynarray_element(&state->collect, BITSET_WORD, 0);
for (start = first_query, end = first_query,
__bitset_next_range(&start, &end, collect, nbits);
start < nbits;
__bitset_next_range(&start, &end, collect, nbits)) {
cmdbuf->cmdlist->ResolveQueryData(qpool->heap, qpool->queries[start].type,
start, end - start,
qpool->resolve_buffer,
qpool->query_size * start);
}
D3D12_RESOURCE_BARRIER barrier = {
.Type = D3D12_RESOURCE_BARRIER_TYPE_TRANSITION,
.Flags = D3D12_RESOURCE_BARRIER_FLAG_NONE,
.Transition = {
.pResource = qpool->resolve_buffer,
.StateBefore = D3D12_RESOURCE_STATE_COPY_DEST,
.StateAfter = D3D12_RESOURCE_STATE_COPY_SOURCE,
},
};
uint32_t offset = dzn_query_pool_get_result_offset(qpool, first_query);
uint32_t size = dzn_query_pool_get_result_size(qpool, query_count);
cmdbuf->cmdlist->ResourceBarrier(1, &barrier);
cmdbuf->cmdlist->CopyBufferRegion(qpool->collect_buffer, offset,
qpool->resolve_buffer, offset,
size);
for (start = first_query, end = first_query,
__bitset_next_range(&start, &end, collect, nbits);
start < nbits;
__bitset_next_range(&start, &end, collect, nbits)) {
uint32_t step = DZN_QUERY_REFS_SECTION_SIZE / sizeof(uint64_t);
uint32_t count = end - start;
for (unsigned i = 0; i < count; i+= step) {
uint32_t sub_count = MIN2(step, count - i);
cmdbuf->cmdlist->CopyBufferRegion(qpool->collect_buffer,
dzn_query_pool_get_availability_offset(qpool, start + i),
device->queries.refs,
DZN_QUERY_REFS_ALL_ONES_OFFSET,
sizeof(uint64_t) * sub_count);
}
dzn_cmd_buffer_dynbitset_set_range(cmdbuf, &state->signal, start, count);
dzn_cmd_buffer_dynbitset_clear_range(cmdbuf, &state->collect, start, count);
}
DZN_SWAP(barrier.Transition.StateBefore, barrier.Transition.StateAfter);
cmdbuf->cmdlist->ResourceBarrier(1, &barrier);
return VK_SUCCESS;
}
static VkResult
dzn_cmd_buffer_collect_query_ops(dzn_cmd_buffer *cmdbuf,
dzn_query_pool *qpool,
struct util_dynarray *bitset_array,
struct util_dynarray *ops_array)
{
dzn_device *device = container_of(cmdbuf->vk.base.device, dzn_device, vk);
BITSET_WORD *bitset = util_dynarray_element(bitset_array, BITSET_WORD, 0);
uint32_t nbits = util_dynarray_num_elements(bitset_array, BITSET_WORD) * BITSET_WORDBITS;
uint32_t start, end;
BITSET_FOREACH_RANGE(start, end, bitset, nbits) {
dzn_cmd_buffer_query_range range { qpool, start, end - start };
dzn_cmd_buffer_query_range *entry = (dzn_cmd_buffer_query_range *)
util_dynarray_grow(ops_array, dzn_cmd_buffer_query_range, 1);
if (!entry) {
cmdbuf->error = vk_error(device, VK_ERROR_OUT_OF_HOST_MEMORY);
return cmdbuf->error;
}
*entry = range;
}
return VK_SUCCESS;
}
static VkResult
dzn_cmd_buffer_gather_queries(dzn_cmd_buffer *cmdbuf)
{
hash_table_foreach(cmdbuf->queries.ht, he) {
dzn_query_pool *qpool = (dzn_query_pool *)he->key;
dzn_cmd_buffer_query_pool_state *state =
(dzn_cmd_buffer_query_pool_state *)he->data;
VkResult result =
dzn_cmd_buffer_collect_queries(cmdbuf, qpool, state, 0, qpool->query_count);
if (result != VK_SUCCESS)
return result;
result = dzn_cmd_buffer_collect_query_ops(cmdbuf, qpool, &state->reset, &cmdbuf->queries.reset);
if (result != VK_SUCCESS)
return result;
result = dzn_cmd_buffer_collect_query_ops(cmdbuf, qpool, &state->wait, &cmdbuf->queries.wait);
if (result != VK_SUCCESS)
return result;
result = dzn_cmd_buffer_collect_query_ops(cmdbuf, qpool, &state->signal, &cmdbuf->queries.signal);
if (result != VK_SUCCESS)
return result;
}
return VK_SUCCESS;
}
VKAPI_ATTR VkResult VKAPI_CALL
dzn_EndCommandBuffer(VkCommandBuffer commandBuffer)
{
VK_FROM_HANDLE(dzn_cmd_buffer, cmdbuf, commandBuffer);
if (cmdbuf->vk.level == VK_COMMAND_BUFFER_LEVEL_PRIMARY) {
dzn_cmd_buffer_gather_events(cmdbuf);
dzn_cmd_buffer_gather_queries(cmdbuf);
HRESULT hres = cmdbuf->cmdlist->Close();
if (FAILED(hres))
cmdbuf->error = vk_error(cmdbuf->vk.base.device, VK_ERROR_OUT_OF_HOST_MEMORY);
} else {
cmdbuf->error = cmdbuf->vk.cmd_queue.error;
}
assert(cmdbuf->error == VK_SUCCESS);
return cmdbuf->error;
}
VKAPI_ATTR void VKAPI_CALL
dzn_CmdPipelineBarrier2(VkCommandBuffer commandBuffer,
const VkDependencyInfo *info)
{
VK_FROM_HANDLE(dzn_cmd_buffer, cmdbuf, commandBuffer);
bool execution_barrier =
!info->memoryBarrierCount &&
!info->bufferMemoryBarrierCount &&
!info->imageMemoryBarrierCount;
if (execution_barrier) {
/* Execution barrier can be emulated with a NULL UAV barrier (AKA
* pipeline flush). That's the best we can do with the standard D3D12
* barrier API.
*/
D3D12_RESOURCE_BARRIER barrier = {
.Type = D3D12_RESOURCE_BARRIER_TYPE_UAV,
.Flags = D3D12_RESOURCE_BARRIER_FLAG_NONE,
.UAV = { .pResource = NULL },
};
cmdbuf->cmdlist->ResourceBarrier(1, &barrier);
}
/* Global memory barriers can be emulated with NULL UAV/Aliasing barriers.
* Scopes are not taken into account, but that's inherent to the current
* D3D12 barrier API.
*/
if (info->memoryBarrierCount) {
D3D12_RESOURCE_BARRIER barriers[2] = {};
barriers[0].Type = D3D12_RESOURCE_BARRIER_TYPE_UAV;
barriers[0].Flags = D3D12_RESOURCE_BARRIER_FLAG_NONE;
barriers[0].UAV.pResource = NULL;
barriers[1].Type = D3D12_RESOURCE_BARRIER_TYPE_ALIASING;
barriers[1].Flags = D3D12_RESOURCE_BARRIER_FLAG_NONE;
barriers[1].Aliasing.pResourceBefore = NULL;
barriers[1].Aliasing.pResourceAfter = NULL;
cmdbuf->cmdlist->ResourceBarrier(2, barriers);
}
for (uint32_t i = 0; i < info->bufferMemoryBarrierCount; i++) {
VK_FROM_HANDLE(dzn_buffer, buf, info->pBufferMemoryBarriers[i].buffer);
D3D12_RESOURCE_BARRIER barrier = {};
/* UAV are used only for storage buffers, skip all other buffers. */
if (!(buf->usage & VK_BUFFER_USAGE_STORAGE_BUFFER_BIT))
continue;
barrier.Type = D3D12_RESOURCE_BARRIER_TYPE_UAV;
barrier.Flags = D3D12_RESOURCE_BARRIER_FLAG_NONE;
barrier.UAV.pResource = buf->res;
cmdbuf->cmdlist->ResourceBarrier(1, &barrier);
}
for (uint32_t i = 0; i < info->imageMemoryBarrierCount; i++) {
const VkImageMemoryBarrier2 *ibarrier = &info->pImageMemoryBarriers[i];
const VkImageSubresourceRange *range = &ibarrier->subresourceRange;
VK_FROM_HANDLE(dzn_image, image, ibarrier->image);
/* We use placed resource's simple model, in which only one resource
* pointing to a given heap is active at a given time. To make the
* resource active we need to add an aliasing barrier.
*/
D3D12_RESOURCE_BARRIER aliasing_barrier = {
.Type = D3D12_RESOURCE_BARRIER_TYPE_ALIASING,
.Flags = D3D12_RESOURCE_BARRIER_FLAG_NONE,
.Aliasing = {
.pResourceBefore = NULL,
.pResourceAfter = image->res,
},
};
cmdbuf->cmdlist->ResourceBarrier(1, &aliasing_barrier);
D3D12_RESOURCE_BARRIER transition_barrier = {
.Type = D3D12_RESOURCE_BARRIER_TYPE_TRANSITION,
.Flags = D3D12_RESOURCE_BARRIER_FLAG_NONE,
.Transition = {
.pResource = image->res,
.StateAfter = dzn_image_layout_to_state(ibarrier->newLayout),
},
};
if (ibarrier->oldLayout == VK_IMAGE_LAYOUT_UNDEFINED ||
ibarrier->oldLayout == VK_IMAGE_LAYOUT_PREINITIALIZED)
transition_barrier.Transition.StateBefore = image->mem->initial_state;
else
transition_barrier.Transition.StateBefore = dzn_image_layout_to_state(ibarrier->oldLayout);
if (transition_barrier.Transition.StateBefore == transition_barrier.Transition.StateAfter)
continue;
/* some layouts map to the same states, and NOP-barriers are illegal */
uint32_t layer_count = dzn_get_layer_count(image, range);
uint32_t level_count = dzn_get_level_count(image, range);
for (uint32_t layer = 0; layer < layer_count; layer++) {
for (uint32_t lvl = 0; lvl < level_count; lvl++) {
dzn_foreach_aspect(aspect, range->aspectMask) {
transition_barrier.Transition.Subresource =
dzn_image_range_get_subresource_index(image, range, aspect, lvl, layer);
cmdbuf->cmdlist->ResourceBarrier(1, &transition_barrier);
}
}
}
}
}
D3D12_CPU_DESCRIPTOR_HANDLE
dzn_cmd_buffer_get_dsv(dzn_cmd_buffer *cmdbuf,
const dzn_image *image,
const D3D12_DEPTH_STENCIL_VIEW_DESC *desc)
{
dzn_device *device = container_of(cmdbuf->vk.base.device, dzn_device, vk);
dzn_cmd_buffer_dsv_key key { image, *desc };
struct hash_entry *he = _mesa_hash_table_search(cmdbuf->dsvs.ht, &key);
struct dzn_cmd_buffer_dsv_entry *dsve;
if (!he) {
dzn_descriptor_heap *heap;
uint32_t slot;
// TODO: error handling
dsve = (dzn_cmd_buffer_dsv_entry *)
vk_alloc(&cmdbuf->vk.pool->alloc, sizeof(*dsve), 8,
VK_SYSTEM_ALLOCATION_SCOPE_OBJECT);
dsve->key = key;
dzn_descriptor_heap_pool_alloc_slots(&cmdbuf->dsvs.pool, device, 1, &heap, &slot);
dsve->handle = dzn_descriptor_heap_get_cpu_handle(heap, slot);
device->dev->CreateDepthStencilView(image->res, desc, dsve->handle);
_mesa_hash_table_insert(cmdbuf->dsvs.ht, &dsve->key, dsve);
} else {
dsve = (dzn_cmd_buffer_dsv_entry *)he->data;
}
return dsve->handle;
}
D3D12_CPU_DESCRIPTOR_HANDLE
dzn_cmd_buffer_get_rtv(dzn_cmd_buffer *cmdbuf,
const dzn_image *image,
const D3D12_RENDER_TARGET_VIEW_DESC *desc)
{
dzn_device *device = container_of(cmdbuf->vk.base.device, dzn_device, vk);
dzn_cmd_buffer_rtv_key key { image, *desc };
struct hash_entry *he = _mesa_hash_table_search(cmdbuf->rtvs.ht, &key);
struct dzn_cmd_buffer_rtv_entry *rtve;
if (!he) {
struct dzn_descriptor_heap *heap;
uint32_t slot;
// TODO: error handling
rtve = (dzn_cmd_buffer_rtv_entry *)
vk_alloc(&cmdbuf->vk.pool->alloc, sizeof(*rtve), 8,
VK_SYSTEM_ALLOCATION_SCOPE_OBJECT);
rtve->key = key;
dzn_descriptor_heap_pool_alloc_slots(&cmdbuf->rtvs.pool, device, 1, &heap, &slot);
rtve->handle = dzn_descriptor_heap_get_cpu_handle(heap, slot);
device->dev->CreateRenderTargetView(image->res, desc, rtve->handle);
he = _mesa_hash_table_insert(cmdbuf->rtvs.ht, &rtve->key, rtve);
} else {
rtve = (dzn_cmd_buffer_rtv_entry *)he->data;
}
return rtve->handle;
}
static VkResult
dzn_cmd_buffer_alloc_internal_buf(dzn_cmd_buffer *cmdbuf,
uint32_t size,
D3D12_HEAP_TYPE heap_type,
D3D12_RESOURCE_STATES init_state,
ID3D12Resource **out)
{
dzn_device *device = container_of(cmdbuf->vk.base.device, dzn_device, vk);
ComPtr<ID3D12Resource> res;
*out = NULL;
/* Align size on 64k (the default alignment) */
size = ALIGN_POT(size, 64 * 1024);
D3D12_HEAP_PROPERTIES hprops =
device->dev->GetCustomHeapProperties(0, heap_type);
D3D12_RESOURCE_DESC rdesc = {
.Dimension = D3D12_RESOURCE_DIMENSION_BUFFER,
.Alignment = D3D12_DEFAULT_RESOURCE_PLACEMENT_ALIGNMENT,
.Width = size,
.Height = 1,
.DepthOrArraySize = 1,
.MipLevels = 1,
.Format = DXGI_FORMAT_UNKNOWN,
.SampleDesc = { .Count = 1, .Quality = 0 },
.Layout = D3D12_TEXTURE_LAYOUT_ROW_MAJOR,
.Flags = D3D12_RESOURCE_FLAG_ALLOW_UNORDERED_ACCESS,
};
HRESULT hres =
device->dev->CreateCommittedResource(&hprops, D3D12_HEAP_FLAG_NONE, &rdesc,
init_state,
NULL, IID_PPV_ARGS(&res));
if (FAILED(hres)) {
cmdbuf->error = vk_error(device, VK_ERROR_OUT_OF_DEVICE_MEMORY);
return cmdbuf->error;
}
dzn_internal_resource *entry = (dzn_internal_resource *)
vk_alloc(&cmdbuf->vk.pool->alloc, sizeof(*entry), 8,
VK_SYSTEM_ALLOCATION_SCOPE_OBJECT);
if (!entry) {
cmdbuf->error = vk_error(device, VK_ERROR_OUT_OF_DEVICE_MEMORY);
return cmdbuf->error;
}
entry->res = res.Detach();
list_addtail(&entry->link, &cmdbuf->internal_bufs);
*out = entry->res;
return VK_SUCCESS;
}
static void
dzn_cmd_buffer_clear_rects_with_copy(dzn_cmd_buffer *cmdbuf,
const dzn_image *image,
VkImageLayout layout,
const VkClearColorValue *color,
const VkImageSubresourceRange *range,
uint32_t rect_count, D3D12_RECT *rects)
{
enum pipe_format pfmt = vk_format_to_pipe_format(image->vk.format);
uint32_t blksize = util_format_get_blocksize(pfmt);
uint8_t buf[D3D12_TEXTURE_DATA_PITCH_ALIGNMENT * 3] = {};
uint32_t raw[4] = {};
assert(blksize <= sizeof(raw));
assert(!(sizeof(buf) % blksize));
util_format_write_4(pfmt, (void *)color, 0, (void *)raw, 0, 0, 0, 1, 1);
uint32_t fill_step = D3D12_TEXTURE_DATA_PITCH_ALIGNMENT;
while (fill_step % blksize)
fill_step += D3D12_TEXTURE_DATA_PITCH_ALIGNMENT;
uint32_t max_w = u_minify(image->vk.extent.width, range->baseMipLevel);
uint32_t max_h = u_minify(image->vk.extent.height, range->baseMipLevel);
uint32_t row_pitch = ALIGN_NPOT(max_w * blksize, fill_step);
uint32_t res_size = max_h * row_pitch;
assert(fill_step <= sizeof(buf));
for (uint32_t i = 0; i < fill_step; i += blksize)
memcpy(&buf[i], raw, blksize);
ID3D12Resource *src_res;
VkResult result =
dzn_cmd_buffer_alloc_internal_buf(cmdbuf, res_size,
D3D12_HEAP_TYPE_UPLOAD,
D3D12_RESOURCE_STATE_GENERIC_READ,
&src_res);
if (result != VK_SUCCESS)
return;
assert(!(res_size % fill_step));
uint8_t *cpu_ptr;
src_res->Map(0, NULL, (void **)&cpu_ptr);
for (uint32_t i = 0; i < res_size; i += fill_step)
memcpy(&cpu_ptr[i], buf, fill_step);
src_res->Unmap(0, NULL);
D3D12_TEXTURE_COPY_LOCATION src_loc = {
.pResource = src_res,
.Type = D3D12_TEXTURE_COPY_TYPE_PLACED_FOOTPRINT,
.PlacedFootprint = {
.Offset = 0,
.Footprint = {
.Width = max_w,
.Height = max_h,
.Depth = 1,
.RowPitch = (UINT)ALIGN_NPOT(max_w * blksize, fill_step),
},
},
};
D3D12_RESOURCE_STATES dst_state = dzn_image_layout_to_state(layout);
D3D12_RESOURCE_BARRIER barrier = {
.Type = D3D12_RESOURCE_BARRIER_TYPE_TRANSITION,
.Flags = D3D12_RESOURCE_BARRIER_FLAG_NONE,
.Transition = {
.pResource = src_res,
.StateBefore = D3D12_RESOURCE_STATE_GENERIC_READ,
.StateAfter = D3D12_RESOURCE_STATE_COPY_SOURCE,
},
};
cmdbuf->cmdlist->ResourceBarrier(1, &barrier);
barrier.Transition.pResource = image->res;
assert(dzn_get_level_count(image, range) == 1);
uint32_t layer_count = dzn_get_layer_count(image, range);
dzn_foreach_aspect(aspect, range->aspectMask) {
VkImageSubresourceLayers subres = {
.aspectMask = (VkImageAspectFlags)aspect,
.mipLevel = range->baseMipLevel,
.baseArrayLayer = range->baseArrayLayer,
.layerCount = layer_count,
};
for (uint32_t layer = 0; layer < layer_count; layer++) {
if (dst_state != D3D12_RESOURCE_STATE_COPY_DEST) {
barrier.Transition.Subresource =
dzn_image_range_get_subresource_index(image, range, aspect, 0, layer);
barrier.Transition.StateBefore = dst_state;
barrier.Transition.StateAfter = D3D12_RESOURCE_STATE_COPY_DEST;
cmdbuf->cmdlist->ResourceBarrier(1, &barrier);
}
D3D12_TEXTURE_COPY_LOCATION dst_loc =
dzn_image_get_copy_loc(image, &subres, aspect, layer);
src_loc.PlacedFootprint.Footprint.Format =
dst_loc.Type == D3D12_TEXTURE_COPY_TYPE_PLACED_FOOTPRINT ?
dst_loc.PlacedFootprint.Footprint.Format :
image->desc.Format;
for (uint32_t r = 0; r < rect_count; r++) {
D3D12_BOX src_box = {
.left = 0,
.top = 0,
.front = 0,
.right = (UINT)(rects[r].right - rects[r].left),
.bottom = (UINT)(rects[r].bottom - rects[r].top),
.back = 1,
};
cmdbuf->cmdlist->CopyTextureRegion(&dst_loc,
rects[r].left, rects[r].top, 0,
&src_loc, &src_box);
}
if (dst_state != D3D12_RESOURCE_STATE_COPY_DEST) {
barrier.Transition.StateAfter = dst_state;
barrier.Transition.StateBefore = D3D12_RESOURCE_STATE_COPY_DEST;
cmdbuf->cmdlist->ResourceBarrier(1, &barrier);
}
}
}
}
static VkClearColorValue
adjust_clear_color(VkFormat format, const VkClearColorValue &col)
{
VkClearColorValue out = col;
// D3D12 doesn't support bgra4, so we map it to rgba4 and swizzle things
// manually where it matters, like here, in the clear path.
if (format == VK_FORMAT_B4G4R4A4_UNORM_PACK16) {
DZN_SWAP(out.float32[0], out.float32[1]);
DZN_SWAP(out.float32[2], out.float32[3]);
}
return out;
}
static void
dzn_cmd_buffer_clear_ranges_with_copy(dzn_cmd_buffer *cmdbuf,
const dzn_image *image,
VkImageLayout layout,
const VkClearColorValue *color,
uint32_t range_count,
const VkImageSubresourceRange *ranges)
{
enum pipe_format pfmt = vk_format_to_pipe_format(image->vk.format);
uint32_t blksize = util_format_get_blocksize(pfmt);
uint8_t buf[D3D12_TEXTURE_DATA_PITCH_ALIGNMENT * 3] = {};
uint32_t raw[4] = {};
assert(blksize <= sizeof(raw));
assert(!(sizeof(buf) % blksize));
util_format_write_4(pfmt, (void *)color, 0, (void *)raw, 0, 0, 0, 1, 1);
uint32_t fill_step = D3D12_TEXTURE_DATA_PITCH_ALIGNMENT;
while (fill_step % blksize)
fill_step += D3D12_TEXTURE_DATA_PITCH_ALIGNMENT;
uint32_t res_size = 0;
for (uint32_t r = 0; r < range_count; r++) {
uint32_t w = u_minify(image->vk.extent.width, ranges[r].baseMipLevel);
uint32_t h = u_minify(image->vk.extent.height, ranges[r].baseMipLevel);
uint32_t d = u_minify(image->vk.extent.depth, ranges[r].baseMipLevel);
uint32_t row_pitch = ALIGN_NPOT(w * blksize, fill_step);
res_size = MAX2(res_size, h * d * row_pitch);
}
assert(fill_step <= sizeof(buf));
for (uint32_t i = 0; i < fill_step; i += blksize)
memcpy(&buf[i], raw, blksize);
ID3D12Resource *src_res;
VkResult result =
dzn_cmd_buffer_alloc_internal_buf(cmdbuf, res_size,
D3D12_HEAP_TYPE_UPLOAD,
D3D12_RESOURCE_STATE_GENERIC_READ,
&src_res);
if (result != VK_SUCCESS)
return;
assert(!(res_size % fill_step));
uint8_t *cpu_ptr;
src_res->Map(0, NULL, (void **)&cpu_ptr);
for (uint32_t i = 0; i < res_size; i += fill_step)
memcpy(&cpu_ptr[i], buf, fill_step);
src_res->Unmap(0, NULL);
D3D12_TEXTURE_COPY_LOCATION src_loc = {
.pResource = src_res,
.Type = D3D12_TEXTURE_COPY_TYPE_PLACED_FOOTPRINT,
.PlacedFootprint = {
.Offset = 0,
},
};
D3D12_RESOURCE_STATES dst_state = dzn_image_layout_to_state(layout);
D3D12_RESOURCE_BARRIER barrier = {
.Type = D3D12_RESOURCE_BARRIER_TYPE_TRANSITION,
.Flags = D3D12_RESOURCE_BARRIER_FLAG_NONE,
.Transition = {
.pResource = src_res,
.StateBefore = D3D12_RESOURCE_STATE_GENERIC_READ,
.StateAfter = D3D12_RESOURCE_STATE_COPY_SOURCE,
},
};
cmdbuf->cmdlist->ResourceBarrier(1, &barrier);
barrier.Transition.pResource = image->res;
for (uint32_t r = 0; r < range_count; r++) {
uint32_t level_count = dzn_get_level_count(image, &ranges[r]);
uint32_t layer_count = dzn_get_layer_count(image, &ranges[r]);
dzn_foreach_aspect(aspect, ranges[r].aspectMask) {
for (uint32_t lvl = 0; lvl < level_count; lvl++) {
uint32_t w = u_minify(image->vk.extent.width, ranges[r].baseMipLevel + lvl);
uint32_t h = u_minify(image->vk.extent.height, ranges[r].baseMipLevel + lvl);
uint32_t d = u_minify(image->vk.extent.depth, ranges[r].baseMipLevel + lvl);
VkImageSubresourceLayers subres = {
.aspectMask = (VkImageAspectFlags)aspect,
.mipLevel = ranges[r].baseMipLevel + lvl,
.baseArrayLayer = ranges[r].baseArrayLayer,
.layerCount = layer_count,
};
for (uint32_t layer = 0; layer < layer_count; layer++) {
if (dst_state != D3D12_RESOURCE_STATE_COPY_DEST) {
barrier.Transition.Subresource =
dzn_image_range_get_subresource_index(image, &ranges[r], aspect, lvl, layer);
barrier.Transition.StateBefore = dst_state;
barrier.Transition.StateAfter = D3D12_RESOURCE_STATE_COPY_DEST;
cmdbuf->cmdlist->ResourceBarrier(1, &barrier);
}
D3D12_TEXTURE_COPY_LOCATION dst_loc =
dzn_image_get_copy_loc(image, &subres, aspect, layer);
src_loc.PlacedFootprint.Footprint.Format =
dst_loc.Type == D3D12_TEXTURE_COPY_TYPE_PLACED_FOOTPRINT ?
dst_loc.PlacedFootprint.Footprint.Format :
image->desc.Format;
src_loc.PlacedFootprint.Footprint.Width = w;
src_loc.PlacedFootprint.Footprint.Height = h;
src_loc.PlacedFootprint.Footprint.Depth = d;
src_loc.PlacedFootprint.Footprint.RowPitch =
ALIGN_NPOT(w * blksize, fill_step);
D3D12_BOX src_box = {
.left = 0,
.top = 0,
.front = 0,
.right = w,
.bottom = h,
.back = d,
};
cmdbuf->cmdlist->CopyTextureRegion(&dst_loc, 0, 0, 0,
&src_loc, &src_box);
if (dst_state != D3D12_RESOURCE_STATE_COPY_DEST) {
barrier.Transition.StateAfter = dst_state;
barrier.Transition.StateBefore = D3D12_RESOURCE_STATE_COPY_DEST;
cmdbuf->cmdlist->ResourceBarrier(1, &barrier);
}
}
}
}
}
}
static void
dzn_cmd_buffer_clear_attachment(dzn_cmd_buffer *cmdbuf,
uint32_t idx,
const VkClearValue *value,
VkImageAspectFlags aspects,
uint32_t base_layer,
uint32_t layer_count,
uint32_t rect_count,
D3D12_RECT *rects)
{
if (idx == VK_ATTACHMENT_UNUSED)
return;
dzn_image_view *view = cmdbuf->state.framebuffer->attachments[idx];
dzn_image *image = container_of(view->vk.image, dzn_image, vk);
VkImageSubresourceRange range = {
.aspectMask = aspects,
.baseMipLevel = view->vk.base_mip_level,
.levelCount = 1,
.baseArrayLayer = view->vk.base_array_layer + base_layer,
.layerCount = layer_count,
};
bool all_layers =
base_layer == 0 &&
(layer_count == view->vk.layer_count ||
layer_count == VK_REMAINING_ARRAY_LAYERS);
if (vk_format_is_depth_or_stencil(view->vk.format)) {
D3D12_CLEAR_FLAGS flags = (D3D12_CLEAR_FLAGS)0;
if (aspects & VK_IMAGE_ASPECT_DEPTH_BIT)
flags |= D3D12_CLEAR_FLAG_DEPTH;
if (aspects & VK_IMAGE_ASPECT_STENCIL_BIT)
flags |= D3D12_CLEAR_FLAG_STENCIL;
if (flags != 0) {
auto desc = dzn_image_get_dsv_desc(image, &range, 0);
auto handle = dzn_cmd_buffer_get_dsv(cmdbuf, image, &desc);
cmdbuf->cmdlist->ClearDepthStencilView(handle, flags,
value->depthStencil.depth,
value->depthStencil.stencil,
rect_count, rects);
}
} else if (aspects & VK_IMAGE_ASPECT_COLOR_BIT) {
VkClearColorValue color = adjust_clear_color(view->vk.format, value->color);
bool clear_with_cpy = false;
float vals[4];
if (vk_format_is_sint(view->vk.format)) {
for (uint32_t i = 0; i < 4; i++) {
vals[i] = color.int32[i];
if (color.int32[i] != (int32_t)vals[i]) {
clear_with_cpy = true;
break;
}
}
} else if (vk_format_is_uint(view->vk.format)) {
for (uint32_t i = 0; i < 4; i++) {
vals[i] = color.uint32[i];
if (color.uint32[i] != (uint32_t)vals[i]) {
clear_with_cpy = true;
break;
}
}
} else {
for (uint32_t i = 0; i < 4; i++)
vals[i] = color.float32[i];
}
if (clear_with_cpy) {
dzn_cmd_buffer_clear_rects_with_copy(cmdbuf, image,
VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL,
&value->color,
&range, rect_count, rects);
} else {
auto desc = dzn_image_get_rtv_desc(image, &range, 0);
auto handle = dzn_cmd_buffer_get_rtv(cmdbuf, image, &desc);
cmdbuf->cmdlist->ClearRenderTargetView(handle, vals, rect_count, rects);
}
}
}
static void
dzn_cmd_buffer_clear_color(dzn_cmd_buffer *cmdbuf,
const dzn_image *image,
VkImageLayout layout,
const VkClearColorValue *col,
uint32_t range_count,
const VkImageSubresourceRange *ranges)
{
if (!(image->desc.Flags & D3D12_RESOURCE_FLAG_ALLOW_RENDER_TARGET)) {
dzn_cmd_buffer_clear_ranges_with_copy(cmdbuf, image, layout, col, range_count, ranges);
return;
}
VkClearColorValue color = adjust_clear_color(image->vk.format, *col);
float clear_vals[4];
enum pipe_format pfmt = vk_format_to_pipe_format(image->vk.format);
if (util_format_is_pure_sint(pfmt)) {
for (uint32_t c = 0; c < ARRAY_SIZE(clear_vals); c++) {
clear_vals[c] = color.int32[c];
if (color.int32[c] != (int32_t)clear_vals[c]) {
dzn_cmd_buffer_clear_ranges_with_copy(cmdbuf, image, layout, col, range_count, ranges);
return;
}
}
} else if (util_format_is_pure_uint(pfmt)) {
for (uint32_t c = 0; c < ARRAY_SIZE(clear_vals); c++) {
clear_vals[c] = color.uint32[c];
if (color.uint32[c] != (uint32_t)clear_vals[c]) {
dzn_cmd_buffer_clear_ranges_with_copy(cmdbuf, image, layout, col, range_count, ranges);
return;
}
}
} else {
memcpy(clear_vals, color.float32, sizeof(clear_vals));
}
for (uint32_t r = 0; r < range_count; r++) {
const VkImageSubresourceRange *range = &ranges[r];
uint32_t layer_count = dzn_get_layer_count(image, range);
uint32_t level_count = dzn_get_level_count(image, range);
for (uint32_t lvl = 0; lvl < level_count; lvl++) {
D3D12_RESOURCE_BARRIER barrier = {
.Type = D3D12_RESOURCE_BARRIER_TYPE_TRANSITION,
.Flags = D3D12_RESOURCE_BARRIER_FLAG_NONE,
.Transition = {
.pResource = image->res,
.StateBefore = dzn_image_layout_to_state(layout),
.StateAfter = D3D12_RESOURCE_STATE_RENDER_TARGET,
},
};
if (barrier.Transition.StateBefore != barrier.Transition.StateAfter) {
for (uint32_t layer = 0; layer < layer_count; layer++) {
barrier.Transition.Subresource =
dzn_image_range_get_subresource_index(image, range,
VK_IMAGE_ASPECT_COLOR_BIT,
lvl, layer);
cmdbuf->cmdlist->ResourceBarrier(1, &barrier);
}
}
VkImageSubresourceRange view_range = *range;
if (image->vk.image_type == VK_IMAGE_TYPE_3D) {
view_range.baseArrayLayer = 0;
view_range.layerCount = u_minify(image->vk.extent.depth, range->baseMipLevel + lvl);
}
auto desc = dzn_image_get_rtv_desc(image, &view_range, lvl);
auto handle = dzn_cmd_buffer_get_rtv(cmdbuf, image, &desc);
cmdbuf->cmdlist->ClearRenderTargetView(handle, clear_vals, 0, NULL);
if (barrier.Transition.StateBefore != barrier.Transition.StateAfter) {
DZN_SWAP(barrier.Transition.StateBefore, barrier.Transition.StateAfter);
for (uint32_t layer = 0; layer < layer_count; layer++) {
barrier.Transition.Subresource =
dzn_image_range_get_subresource_index(image, range, VK_IMAGE_ASPECT_COLOR_BIT, lvl, layer);
cmdbuf->cmdlist->ResourceBarrier(1, &barrier);
}
}
}
}
}
static void
dzn_cmd_buffer_clear_zs(dzn_cmd_buffer *cmdbuf,
const dzn_image *image,
VkImageLayout layout,
const VkClearDepthStencilValue *zs,
uint32_t range_count,
const VkImageSubresourceRange *ranges)
{
assert(image->desc.Flags & D3D12_RESOURCE_FLAG_ALLOW_DEPTH_STENCIL);
for (uint32_t r = 0; r < range_count; r++) {
const VkImageSubresourceRange *range = &ranges[r];
uint32_t layer_count = dzn_get_layer_count(image, range);
uint32_t level_count = dzn_get_level_count(image, range);
D3D12_CLEAR_FLAGS flags = (D3D12_CLEAR_FLAGS)0;
if (range->aspectMask & VK_IMAGE_ASPECT_DEPTH_BIT)
flags |= D3D12_CLEAR_FLAG_DEPTH;
if (range->aspectMask & VK_IMAGE_ASPECT_STENCIL_BIT)
flags |= D3D12_CLEAR_FLAG_STENCIL;
for (uint32_t lvl = 0; lvl < level_count; lvl++) {
D3D12_RESOURCE_BARRIER barrier = {
.Type = D3D12_RESOURCE_BARRIER_TYPE_TRANSITION,
.Flags = D3D12_RESOURCE_BARRIER_FLAG_NONE,
.Transition = {
.pResource = image->res,
.StateBefore = dzn_image_layout_to_state(layout),
.StateAfter = D3D12_RESOURCE_STATE_DEPTH_WRITE,
},
};
if (barrier.Transition.StateBefore != barrier.Transition.StateAfter) {
for (uint32_t layer = 0; layer < layer_count; layer++) {
dzn_foreach_aspect(aspect, range->aspectMask) {
barrier.Transition.Subresource =
dzn_image_range_get_subresource_index(image, range, aspect, lvl, layer);
cmdbuf->cmdlist->ResourceBarrier(1, &barrier);
}
}
}
auto desc = dzn_image_get_dsv_desc(image, range, lvl);
auto handle = dzn_cmd_buffer_get_dsv(cmdbuf, image, &desc);
cmdbuf->cmdlist->ClearDepthStencilView(handle, flags,
zs->depth, zs->stencil,
0, NULL);
if (barrier.Transition.StateBefore != barrier.Transition.StateAfter) {
DZN_SWAP(barrier.Transition.StateBefore, barrier.Transition.StateAfter);
for (uint32_t layer = 0; layer < layer_count; layer++) {
dzn_foreach_aspect(aspect, range->aspectMask) {
barrier.Transition.Subresource =
dzn_image_range_get_subresource_index(image, range, aspect, lvl, layer);
cmdbuf->cmdlist->ResourceBarrier(1, &barrier);
}
}
}
}
}
}
static void
dzn_cmd_buffer_copy_buf2img_region(dzn_cmd_buffer *cmdbuf,
const VkCopyBufferToImageInfo2 *info,
uint32_t r,
VkImageAspectFlagBits aspect,
uint32_t l)
{
dzn_device *device = container_of(cmdbuf->vk.base.device, dzn_device, vk);
VK_FROM_HANDLE(dzn_buffer, src_buffer, info->srcBuffer);
VK_FROM_HANDLE(dzn_image, dst_image, info->dstImage);
ID3D12Device *dev = device->dev;
ID3D12GraphicsCommandList *cmdlist = cmdbuf->cmdlist;
const VkBufferImageCopy2 *region = &info->pRegions[r];
enum pipe_format pfmt = vk_format_to_pipe_format(dst_image->vk.format);
uint32_t blkh = util_format_get_blockheight(pfmt);
uint32_t blkd = util_format_get_blockdepth(pfmt);
D3D12_TEXTURE_COPY_LOCATION dst_img_loc =
dzn_image_get_copy_loc(dst_image, &region->imageSubresource, aspect, l);
D3D12_TEXTURE_COPY_LOCATION src_buf_loc =
dzn_buffer_get_copy_loc(src_buffer, dst_image->vk.format, region, aspect, l);
if (dzn_buffer_supports_region_copy(&src_buf_loc)) {
/* RowPitch and Offset are properly aligned, we can copy
* the whole thing in one call.
*/
D3D12_BOX src_box = {
.left = 0,
.top = 0,
.front = 0,
.right = region->imageExtent.width,
.bottom = region->imageExtent.height,
.back = region->imageExtent.depth,
};
cmdlist->CopyTextureRegion(&dst_img_loc, region->imageOffset.x,
region->imageOffset.y, region->imageOffset.z,
&src_buf_loc, &src_box);
return;
}
/* Copy line-by-line if things are not properly aligned. */
D3D12_BOX src_box = {
.top = 0,
.front = 0,
.bottom = blkh,
.back = blkd,
};
for (uint32_t z = 0; z < region->imageExtent.depth; z += blkd) {
for (uint32_t y = 0; y < region->imageExtent.height; y += blkh) {
uint32_t src_x;
D3D12_TEXTURE_COPY_LOCATION src_buf_line_loc =
dzn_buffer_get_line_copy_loc(src_buffer, dst_image->vk.format,
region, &src_buf_loc,
y, z, &src_x);
src_box.left = src_x;
src_box.right = src_x + region->imageExtent.width;
cmdlist->CopyTextureRegion(&dst_img_loc,
region->imageOffset.x,
region->imageOffset.y + y,
region->imageOffset.z + z,
&src_buf_line_loc, &src_box);
}
}
}
static void
dzn_cmd_buffer_copy_img2buf_region(dzn_cmd_buffer *cmdbuf,
const VkCopyImageToBufferInfo2 *info,
uint32_t r,
VkImageAspectFlagBits aspect,
uint32_t l)
{
dzn_device *device = container_of(cmdbuf->vk.base.device, dzn_device, vk);
VK_FROM_HANDLE(dzn_image, src_image, info->srcImage);
VK_FROM_HANDLE(dzn_buffer, dst_buffer, info->dstBuffer);
ID3D12Device *dev = device->dev;
ID3D12GraphicsCommandList *cmdlist = cmdbuf->cmdlist;
const VkBufferImageCopy2 *region = &info->pRegions[r];
enum pipe_format pfmt = vk_format_to_pipe_format(src_image->vk.format);
uint32_t blkh = util_format_get_blockheight(pfmt);
uint32_t blkd = util_format_get_blockdepth(pfmt);
D3D12_TEXTURE_COPY_LOCATION src_img_loc =
dzn_image_get_copy_loc(src_image, &region->imageSubresource, aspect, l);
D3D12_TEXTURE_COPY_LOCATION dst_buf_loc =
dzn_buffer_get_copy_loc(dst_buffer, src_image->vk.format, region, aspect, l);
if (dzn_buffer_supports_region_copy(&dst_buf_loc)) {
/* RowPitch and Offset are properly aligned on 256 bytes, we can copy
* the whole thing in one call.
*/
D3D12_BOX src_box = {
.left = (UINT)region->imageOffset.x,
.top = (UINT)region->imageOffset.y,
.front = (UINT)region->imageOffset.z,
.right = (UINT)(region->imageOffset.x + region->imageExtent.width),
.bottom = (UINT)(region->imageOffset.y + region->imageExtent.height),
.back = (UINT)(region->imageOffset.z + region->imageExtent.depth),
};
cmdlist->CopyTextureRegion(&dst_buf_loc, 0, 0, 0,
&src_img_loc, &src_box);
return;
}
D3D12_BOX src_box = {
.left = (UINT)region->imageOffset.x,
.right = (UINT)(region->imageOffset.x + region->imageExtent.width),
};
/* Copy line-by-line if things are not properly aligned. */
for (uint32_t z = 0; z < region->imageExtent.depth; z += blkd) {
src_box.front = region->imageOffset.z + z;
src_box.back = src_box.front + blkd;
for (uint32_t y = 0; y < region->imageExtent.height; y += blkh) {
uint32_t dst_x;
D3D12_TEXTURE_COPY_LOCATION dst_buf_line_loc =
dzn_buffer_get_line_copy_loc(dst_buffer, src_image->vk.format,
region, &dst_buf_loc,
y, z, &dst_x);
src_box.top = region->imageOffset.y + y;
src_box.bottom = src_box.top + blkh;
cmdlist->CopyTextureRegion(&dst_buf_line_loc, dst_x, 0, 0,
&src_img_loc, &src_box);
}
}
}
static void
dzn_cmd_buffer_copy_img_chunk(dzn_cmd_buffer *cmdbuf,
const VkCopyImageInfo2 *info,
D3D12_RESOURCE_DESC &tmp_desc,
D3D12_TEXTURE_COPY_LOCATION &tmp_loc,
uint32_t r,
VkImageAspectFlagBits aspect,
uint32_t l)
{
dzn_device *device = container_of(cmdbuf->vk.base.device, dzn_device, vk);
VK_FROM_HANDLE(dzn_image, src, info->srcImage);
VK_FROM_HANDLE(dzn_image, dst, info->dstImage);
ID3D12Device *dev = device->dev;
ID3D12GraphicsCommandList *cmdlist = cmdbuf->cmdlist;
const VkImageCopy2 *region = &info->pRegions[r];
const VkImageSubresourceLayers *src_subres = &region->srcSubresource;
const VkImageSubresourceLayers *dst_subres = &region->dstSubresource;
VkFormat src_format =
dzn_image_get_plane_format(src->vk.format, aspect);
VkFormat dst_format =
dzn_image_get_plane_format(dst->vk.format, aspect);
enum pipe_format src_pfmt = vk_format_to_pipe_format(src_format);
uint32_t src_blkw = util_format_get_blockwidth(src_pfmt);
uint32_t src_blkh = util_format_get_blockheight(src_pfmt);
uint32_t src_blkd = util_format_get_blockdepth(src_pfmt);
enum pipe_format dst_pfmt = vk_format_to_pipe_format(dst_format);
uint32_t dst_blkw = util_format_get_blockwidth(dst_pfmt);
uint32_t dst_blkh = util_format_get_blockheight(dst_pfmt);
uint32_t dst_blkd = util_format_get_blockdepth(dst_pfmt);
assert(src_subres->layerCount == dst_subres->layerCount);
assert(src_subres->aspectMask == dst_subres->aspectMask);
auto dst_loc = dzn_image_get_copy_loc(dst, dst_subres, aspect, l);
auto src_loc = dzn_image_get_copy_loc(src, src_subres, aspect, l);
D3D12_BOX src_box = {
.left = (UINT)MAX2(region->srcOffset.x, 0),
.top = (UINT)MAX2(region->srcOffset.y, 0),
.front = (UINT)MAX2(region->srcOffset.z, 0),
.right = (UINT)region->srcOffset.x + region->extent.width,
.bottom = (UINT)region->srcOffset.y + region->extent.height,
.back = (UINT)region->srcOffset.z + region->extent.depth,
};
if (!tmp_loc.pResource) {
cmdlist->CopyTextureRegion(&dst_loc, region->dstOffset.x,
region->dstOffset.y, region->dstOffset.z,
&src_loc, &src_box);
return;
}
tmp_desc.Format =
dzn_image_get_placed_footprint_format(src->vk.format, aspect);
tmp_desc.Width = region->extent.width;
tmp_desc.Height = region->extent.height;
dev->GetCopyableFootprints(&tmp_desc,
0, 1, 0,
&tmp_loc.PlacedFootprint,
NULL, NULL, NULL);
tmp_loc.PlacedFootprint.Footprint.Depth = region->extent.depth;
D3D12_RESOURCE_BARRIER barrier = {
.Type = D3D12_RESOURCE_BARRIER_TYPE_TRANSITION,
.Flags = D3D12_RESOURCE_BARRIER_FLAG_NONE,
.Transition = {
.pResource = tmp_loc.pResource,
.Subresource = 0,
.StateBefore = D3D12_RESOURCE_STATE_COPY_SOURCE,
.StateAfter = D3D12_RESOURCE_STATE_COPY_DEST,
},
};
if (r > 0 || l > 0)
cmdlist->ResourceBarrier(1, &barrier);
cmdlist->CopyTextureRegion(&tmp_loc, 0, 0, 0, &src_loc, &src_box);
DZN_SWAP(barrier.Transition.StateBefore, barrier.Transition.StateAfter);
cmdlist->ResourceBarrier(1, &barrier);
tmp_desc.Format =
dzn_image_get_placed_footprint_format(dst->vk.format, aspect);
if (src_blkw != dst_blkw)
tmp_desc.Width = DIV_ROUND_UP(region->extent.width, src_blkw) * dst_blkw;
if (src_blkh != dst_blkh)
tmp_desc.Height = DIV_ROUND_UP(region->extent.height, src_blkh) * dst_blkh;
device->dev->GetCopyableFootprints(&tmp_desc,
0, 1, 0,
&tmp_loc.PlacedFootprint,
NULL, NULL, NULL);
if (src_blkd != dst_blkd) {
tmp_loc.PlacedFootprint.Footprint.Depth =
DIV_ROUND_UP(region->extent.depth, src_blkd) * dst_blkd;
} else {
tmp_loc.PlacedFootprint.Footprint.Depth = region->extent.depth;
}
D3D12_BOX tmp_box = {
.left = 0,
.top = 0,
.front = 0,
.right = tmp_loc.PlacedFootprint.Footprint.Width,
.bottom = tmp_loc.PlacedFootprint.Footprint.Height,
.back = tmp_loc.PlacedFootprint.Footprint.Depth,
};
cmdlist->CopyTextureRegion(&dst_loc,
region->dstOffset.x,
region->dstOffset.y,
region->dstOffset.z,
&tmp_loc, &tmp_box);
}
static void
dzn_cmd_buffer_blit_prepare_src_view(dzn_cmd_buffer *cmdbuf,
VkImage image,
VkImageAspectFlagBits aspect,
const VkImageSubresourceLayers *subres,
dzn_descriptor_heap *heap,
uint32_t heap_slot)
{
dzn_device *device = container_of(cmdbuf->vk.base.device, dzn_device, vk);
VK_FROM_HANDLE(dzn_image, img, image);
VkImageViewCreateInfo iview_info = {
.sType = VK_STRUCTURE_TYPE_IMAGE_VIEW_CREATE_INFO,
.image = image,
.format = img->vk.format,
.subresourceRange = {
.aspectMask = (VkImageAspectFlags)aspect,
.baseMipLevel = subres->mipLevel,
.levelCount = 1,
.baseArrayLayer = subres->baseArrayLayer,
.layerCount = subres->layerCount,
},
};
if (aspect == VK_IMAGE_ASPECT_STENCIL_BIT) {
iview_info.components.r = VK_COMPONENT_SWIZZLE_G;
iview_info.components.g = VK_COMPONENT_SWIZZLE_G;
iview_info.components.b = VK_COMPONENT_SWIZZLE_G;
iview_info.components.a = VK_COMPONENT_SWIZZLE_G;
} else if (aspect == VK_IMAGE_ASPECT_STENCIL_BIT) {
iview_info.components.r = VK_COMPONENT_SWIZZLE_R;
iview_info.components.g = VK_COMPONENT_SWIZZLE_R;
iview_info.components.b = VK_COMPONENT_SWIZZLE_R;
iview_info.components.a = VK_COMPONENT_SWIZZLE_R;
}
switch (img->vk.image_type) {
case VK_IMAGE_TYPE_1D:
iview_info.viewType = img->vk.array_layers > 1 ?
VK_IMAGE_VIEW_TYPE_1D_ARRAY : VK_IMAGE_VIEW_TYPE_1D;
break;
case VK_IMAGE_TYPE_2D:
iview_info.viewType = img->vk.array_layers > 1 ?
VK_IMAGE_VIEW_TYPE_2D_ARRAY : VK_IMAGE_VIEW_TYPE_2D;
break;
case VK_IMAGE_TYPE_3D:
iview_info.viewType = VK_IMAGE_VIEW_TYPE_3D;
break;
default:
unreachable("Invalid type");
}
dzn_image_view iview;
dzn_image_view_init(device, &iview, &iview_info);
dzn_descriptor_heap_write_image_view_desc(heap, heap_slot, false, false, &iview);
dzn_image_view_finish(&iview);
D3D12_GPU_DESCRIPTOR_HANDLE handle =
dzn_descriptor_heap_get_gpu_handle(heap, heap_slot);
cmdbuf->cmdlist->SetGraphicsRootDescriptorTable(0, handle);
}
static void
dzn_cmd_buffer_blit_prepare_dst_view(dzn_cmd_buffer *cmdbuf,
dzn_image *img,
VkImageAspectFlagBits aspect,
uint32_t level, uint32_t layer)
{
bool ds = aspect & (VK_IMAGE_ASPECT_DEPTH_BIT | VK_IMAGE_ASPECT_STENCIL_BIT);
VkImageSubresourceRange range = {
.aspectMask = (VkImageAspectFlags)aspect,
.baseMipLevel = level,
.levelCount = 1,
.baseArrayLayer = layer,
.layerCount = 1,
};
if (ds) {
auto desc = dzn_image_get_dsv_desc(img, &range, 0);
auto handle = dzn_cmd_buffer_get_dsv(cmdbuf, img, &desc);
cmdbuf->cmdlist->OMSetRenderTargets(0, NULL, TRUE, &handle);
} else {
auto desc = dzn_image_get_rtv_desc(img, &range, 0);
auto handle = dzn_cmd_buffer_get_rtv(cmdbuf, img, &desc);
cmdbuf->cmdlist->OMSetRenderTargets(1, &handle, FALSE, NULL);
}
}
static void
dzn_cmd_buffer_blit_set_pipeline(dzn_cmd_buffer *cmdbuf,
const dzn_image *src,
const dzn_image *dst,
VkImageAspectFlagBits aspect,
VkFilter filter, bool resolve)
{
dzn_device *device = container_of(cmdbuf->vk.base.device, dzn_device, vk);
enum pipe_format pfmt = vk_format_to_pipe_format(dst->vk.format);
VkImageUsageFlags usage =
vk_format_is_depth_or_stencil(dst->vk.format) ?
VK_IMAGE_USAGE_DEPTH_STENCIL_ATTACHMENT_BIT :
VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT;
struct dzn_meta_blit_key ctx_key = {
.out_format = dzn_image_get_dxgi_format(dst->vk.format, usage, aspect),
.samples = (uint32_t)src->vk.samples,
.loc = (uint32_t)(aspect == VK_IMAGE_ASPECT_DEPTH_BIT ?
FRAG_RESULT_DEPTH :
aspect == VK_IMAGE_ASPECT_STENCIL_BIT ?
FRAG_RESULT_STENCIL :
FRAG_RESULT_DATA0),
.out_type = (uint32_t)(util_format_is_pure_uint(pfmt) ? GLSL_TYPE_UINT :
util_format_is_pure_sint(pfmt) ? GLSL_TYPE_INT :
aspect == VK_IMAGE_ASPECT_STENCIL_BIT ? GLSL_TYPE_UINT :
GLSL_TYPE_FLOAT),
.sampler_dim = (uint32_t)(src->vk.image_type == VK_IMAGE_TYPE_1D ? GLSL_SAMPLER_DIM_1D :
src->vk.image_type == VK_IMAGE_TYPE_2D && src->vk.samples == 1 ? GLSL_SAMPLER_DIM_2D :
src->vk.image_type == VK_IMAGE_TYPE_2D && src->vk.samples > 1 ? GLSL_SAMPLER_DIM_MS :
GLSL_SAMPLER_DIM_3D),
.src_is_array = src->vk.array_layers > 1,
.resolve = resolve,
.linear_filter = filter == VK_FILTER_LINEAR,
.padding = 0,
};
const dzn_meta_blit *ctx =
dzn_meta_blits_get_context(device, &ctx_key);
assert(ctx);
cmdbuf->cmdlist->SetGraphicsRootSignature(ctx->root_sig);
cmdbuf->cmdlist->SetPipelineState(ctx->pipeline_state);
}
static void
dzn_cmd_buffer_blit_set_2d_region(dzn_cmd_buffer *cmdbuf,
const dzn_image *src,
const VkImageSubresourceLayers *src_subres,
const VkOffset3D *src_offsets,
const dzn_image *dst,
const VkImageSubresourceLayers *dst_subres,
const VkOffset3D *dst_offsets,
bool normalize_src_coords)
{
uint32_t dst_w = u_minify(dst->vk.extent.width, dst_subres->mipLevel);
uint32_t dst_h = u_minify(dst->vk.extent.height, dst_subres->mipLevel);
uint32_t src_w = u_minify(src->vk.extent.width, src_subres->mipLevel);
uint32_t src_h = u_minify(src->vk.extent.height, src_subres->mipLevel);
float dst_pos[4] = {
(2 * (float)dst_offsets[0].x / (float)dst_w) - 1.0f, -((2 * (float)dst_offsets[0].y / (float)dst_h) - 1.0f),
(2 * (float)dst_offsets[1].x / (float)dst_w) - 1.0f, -((2 * (float)dst_offsets[1].y / (float)dst_h) - 1.0f),
};
float src_pos[4] = {
(float)src_offsets[0].x, (float)src_offsets[0].y,
(float)src_offsets[1].x, (float)src_offsets[1].y,
};
if (normalize_src_coords) {
src_pos[0] /= src_w;
src_pos[1] /= src_h;
src_pos[2] /= src_w;
src_pos[3] /= src_h;
}
float coords[] = {
dst_pos[0], dst_pos[1], src_pos[0], src_pos[1],
dst_pos[2], dst_pos[1], src_pos[2], src_pos[1],
dst_pos[0], dst_pos[3], src_pos[0], src_pos[3],
dst_pos[2], dst_pos[3], src_pos[2], src_pos[3],
};
cmdbuf->cmdlist->SetGraphicsRoot32BitConstants(1, ARRAY_SIZE(coords), coords, 0);
D3D12_VIEWPORT vp = {
.TopLeftX = 0,
.TopLeftY = 0,
.Width = (float)dst_w,
.Height = (float)dst_h,
.MinDepth = 0,
.MaxDepth = 1,
};
cmdbuf->cmdlist->RSSetViewports(1, &vp);
D3D12_RECT scissor = {
.left = MIN2(dst_offsets[0].x, dst_offsets[1].x),
.top = MIN2(dst_offsets[0].y, dst_offsets[1].y),
.right = MAX2(dst_offsets[0].x, dst_offsets[1].x),
.bottom = MAX2(dst_offsets[0].y, dst_offsets[1].y),
};
cmdbuf->cmdlist->RSSetScissorRects(1, &scissor);
}
static void
dzn_cmd_buffer_blit_issue_barriers(dzn_cmd_buffer *cmdbuf,
dzn_image *src, VkImageLayout src_layout,
const VkImageSubresourceLayers *src_subres,
dzn_image *dst, VkImageLayout dst_layout,
const VkImageSubresourceLayers *dst_subres,
VkImageAspectFlagBits aspect,
bool post)
{
bool ds = aspect & (VK_IMAGE_ASPECT_DEPTH_BIT | VK_IMAGE_ASPECT_STENCIL_BIT);
D3D12_RESOURCE_BARRIER barriers[2] = {
{
.Type = D3D12_RESOURCE_BARRIER_TYPE_TRANSITION,
.Flags = D3D12_RESOURCE_BARRIER_FLAG_NONE,
.Transition = {
.pResource = src->res,
.StateBefore = dzn_image_layout_to_state(src_layout),
.StateAfter = D3D12_RESOURCE_STATE_PIXEL_SHADER_RESOURCE,
},
},
{
.Type = D3D12_RESOURCE_BARRIER_TYPE_TRANSITION,
.Flags = D3D12_RESOURCE_BARRIER_FLAG_NONE,
.Transition = {
.pResource = dst->res,
.StateBefore = dzn_image_layout_to_state(dst_layout),
.StateAfter = ds ?
D3D12_RESOURCE_STATE_DEPTH_WRITE :
D3D12_RESOURCE_STATE_RENDER_TARGET,
},
},
};
if (post) {
DZN_SWAP(barriers[0].Transition.StateBefore, barriers[0].Transition.StateAfter);
DZN_SWAP(barriers[1].Transition.StateBefore, barriers[1].Transition.StateAfter);
}
uint32_t layer_count = dzn_get_layer_count(src, src_subres);
uint32_t src_level = src_subres->mipLevel;
uint32_t dst_level = dst_subres->mipLevel;
assert(dzn_get_layer_count(dst, dst_subres) == layer_count);
assert(src_level < src->vk.mip_levels);
assert(dst_level < dst->vk.mip_levels);
for (uint32_t layer = 0; layer < layer_count; layer++) {
barriers[0].Transition.Subresource =
dzn_image_layers_get_subresource_index(src, src_subres, aspect, layer);
barriers[1].Transition.Subresource =
dzn_image_layers_get_subresource_index(dst, dst_subres, aspect, layer);
cmdbuf->cmdlist->ResourceBarrier(ARRAY_SIZE(barriers), barriers);
}
}
static void
dzn_cmd_buffer_blit_region(dzn_cmd_buffer *cmdbuf,
const VkBlitImageInfo2 *info,
dzn_descriptor_heap *heap,
uint32_t *heap_slot,
uint32_t r)
{
dzn_device *device = container_of(cmdbuf->vk.base.device, dzn_device, vk);
VK_FROM_HANDLE(dzn_image, src, info->srcImage);
VK_FROM_HANDLE(dzn_image, dst, info->dstImage);
ID3D12Device *dev = device->dev;
const VkImageBlit2 *region = &info->pRegions[r];
dzn_foreach_aspect(aspect, region->srcSubresource.aspectMask) {
dzn_cmd_buffer_blit_set_pipeline(cmdbuf, src, dst, aspect, info->filter, false);
dzn_cmd_buffer_blit_issue_barriers(cmdbuf,
src, info->srcImageLayout, &region->srcSubresource,
dst, info->dstImageLayout, &region->dstSubresource,
aspect, false);
dzn_cmd_buffer_blit_prepare_src_view(cmdbuf, info->srcImage,
aspect, &region->srcSubresource,
heap, (*heap_slot)++);
dzn_cmd_buffer_blit_set_2d_region(cmdbuf,
src, &region->srcSubresource, region->srcOffsets,
dst, &region->dstSubresource, region->dstOffsets,
src->vk.samples == 1);
uint32_t dst_depth =
region->dstOffsets[1].z > region->dstOffsets[0].z ?
region->dstOffsets[1].z - region->dstOffsets[0].z :
region->dstOffsets[0].z - region->dstOffsets[1].z;
uint32_t src_depth =
region->srcOffsets[1].z > region->srcOffsets[0].z ?
region->srcOffsets[1].z - region->srcOffsets[0].z :
region->srcOffsets[0].z - region->srcOffsets[1].z;
uint32_t layer_count = dzn_get_layer_count(src, &region->srcSubresource);
uint32_t dst_level = region->dstSubresource.mipLevel;
float src_slice_step = layer_count > 1 ? 1 : (float)src_depth / dst_depth;
if (region->srcOffsets[0].z > region->srcOffsets[1].z)
src_slice_step = -src_slice_step;
float src_z_coord = layer_count > 1 ?
0 : (float)region->srcOffsets[0].z + (src_slice_step * 0.5f);
uint32_t slice_count = layer_count > 1 ? layer_count : dst_depth;
uint32_t dst_z_coord = layer_count > 1 ?
region->dstSubresource.baseArrayLayer :
region->dstOffsets[0].z;
if (region->dstOffsets[0].z > region->dstOffsets[1].z)
dst_z_coord--;
uint32_t dst_slice_step = region->dstOffsets[0].z < region->dstOffsets[1].z ?
1 : -1;
/* Normalize the src coordinates/step */
if (layer_count == 1 && src->vk.samples == 1) {
src_z_coord /= src->vk.extent.depth;
src_slice_step /= src->vk.extent.depth;
}
for (uint32_t slice = 0; slice < slice_count; slice++) {
dzn_cmd_buffer_blit_prepare_dst_view(cmdbuf, dst, aspect, dst_level, dst_z_coord);
cmdbuf->cmdlist->SetGraphicsRoot32BitConstants(1, 1, &src_z_coord, 16);
cmdbuf->cmdlist->DrawInstanced(4, 1, 0, 0);
src_z_coord += src_slice_step;
dst_z_coord += dst_slice_step;
}
dzn_cmd_buffer_blit_issue_barriers(cmdbuf,
src, info->srcImageLayout, &region->srcSubresource,
dst, info->dstImageLayout, &region->dstSubresource,
aspect, true);
}
}
static void
dzn_cmd_buffer_resolve_region(dzn_cmd_buffer *cmdbuf,
const VkResolveImageInfo2 *info,
dzn_descriptor_heap *heap,
uint32_t *heap_slot,
uint32_t r)
{
dzn_device *device = container_of(cmdbuf->vk.base.device, dzn_device, vk);
VK_FROM_HANDLE(dzn_image, src, info->srcImage);
VK_FROM_HANDLE(dzn_image, dst, info->dstImage);
ID3D12Device *dev = device->dev;
const VkImageResolve2 *region = &info->pRegions[r];
dzn_foreach_aspect(aspect, region->srcSubresource.aspectMask) {
dzn_cmd_buffer_blit_set_pipeline(cmdbuf, src, dst, aspect, VK_FILTER_NEAREST, true);
dzn_cmd_buffer_blit_issue_barriers(cmdbuf,
src, info->srcImageLayout, &region->srcSubresource,
dst, info->dstImageLayout, &region->dstSubresource,
aspect, false);
dzn_cmd_buffer_blit_prepare_src_view(cmdbuf, info->srcImage, aspect,
&region->srcSubresource,
heap, (*heap_slot)++);
VkOffset3D src_offset[2] = {
{
.x = region->srcOffset.x,
.y = region->srcOffset.y,
},
{
.x = (int32_t)(region->srcOffset.x + region->extent.width),
.y = (int32_t)(region->srcOffset.y + region->extent.height),
},
};
VkOffset3D dst_offset[2] = {
{
.x = region->dstOffset.x,
.y = region->dstOffset.y,
},
{
.x = (int32_t)(region->dstOffset.x + region->extent.width),
.y = (int32_t)(region->dstOffset.y + region->extent.height),
},
};
dzn_cmd_buffer_blit_set_2d_region(cmdbuf,
src, &region->srcSubresource, src_offset,
dst, &region->dstSubresource, dst_offset,
false);
uint32_t layer_count = dzn_get_layer_count(src, &region->srcSubresource);
for (uint32_t layer = 0; layer < layer_count; layer++) {
float src_z_coord = layer;
dzn_cmd_buffer_blit_prepare_dst_view(cmdbuf,
dst, aspect, region->dstSubresource.mipLevel,
region->dstSubresource.baseArrayLayer + layer);
cmdbuf->cmdlist->SetGraphicsRoot32BitConstants(1, 1, &src_z_coord, 16);
cmdbuf->cmdlist->DrawInstanced(4, 1, 0, 0);
}
dzn_cmd_buffer_blit_issue_barriers(cmdbuf,
src, info->srcImageLayout, &region->srcSubresource,
dst, info->dstImageLayout, &region->dstSubresource,
aspect, true);
}
}
static void
dzn_cmd_buffer_clear_attachments(dzn_cmd_buffer *cmdbuf,
uint32_t attachment_count,
const VkClearAttachment *attachments,
uint32_t rect_count,
const VkClearRect *rects)
{
struct dzn_render_pass *pass = cmdbuf->state.pass;
const struct dzn_subpass *subpass = &pass->subpasses[cmdbuf->state.subpass];
for (unsigned i = 0; i < attachment_count; i++) {
uint32_t idx;
if (attachments[i].aspectMask & VK_IMAGE_ASPECT_COLOR_BIT)
idx = subpass->colors[attachments[i].colorAttachment].idx;
else
idx = subpass->zs.idx;
for (uint32_t j = 0; j < rect_count; j++) {
D3D12_RECT rect;
dzn_translate_rect(&rect, &rects[j].rect);
dzn_cmd_buffer_clear_attachment(cmdbuf,
idx, &attachments[i].clearValue,
attachments[i].aspectMask,
rects[j].baseArrayLayer,
rects[j].layerCount,
1, &rect);
}
}
}
static void
dzn_cmd_buffer_attachment_ref_transition(dzn_cmd_buffer *cmdbuf,
const dzn_attachment_ref *att)
{
const dzn_image_view *iview = cmdbuf->state.framebuffer->attachments[att->idx];
const dzn_image *image = container_of(iview->vk.image, dzn_image, vk);
if (att->before == att->during)
return;
D3D12_RESOURCE_BARRIER barrier = {
.Type = D3D12_RESOURCE_BARRIER_TYPE_TRANSITION,
.Flags = D3D12_RESOURCE_BARRIER_FLAG_NONE,
.Transition = {
.pResource = image->res,
.Subresource = 0, // YOLO
.StateBefore = att->before,
.StateAfter = att->during,
},
};
cmdbuf->cmdlist->ResourceBarrier(1, &barrier);
}
void
dzn_cmd_buffer_attachment_transition(dzn_cmd_buffer *cmdbuf,
const dzn_attachment *att)
{
const dzn_image_view *iview = cmdbuf->state.framebuffer->attachments[att->idx];
const dzn_image *image = container_of(iview->vk.image, dzn_image, vk);
if (att->last == att->after)
return;
D3D12_RESOURCE_BARRIER barrier = {
.Type = D3D12_RESOURCE_BARRIER_TYPE_TRANSITION,
.Flags = D3D12_RESOURCE_BARRIER_FLAG_NONE,
.Transition = {
.pResource = image->res,
.Subresource = 0, // YOLO
.StateBefore = att->last,
.StateAfter = att->after,
},
};
cmdbuf->cmdlist->ResourceBarrier(1, &barrier);
}
static void
dzn_cmd_buffer_resolve_attachment(dzn_cmd_buffer *cmdbuf, uint32_t i)
{
const struct dzn_subpass *subpass =
&cmdbuf->state.pass->subpasses[cmdbuf->state.subpass];
if (subpass->resolve[i].idx == VK_ATTACHMENT_UNUSED)
return;
const dzn_framebuffer *framebuffer = cmdbuf->state.framebuffer;
struct dzn_image_view *src = framebuffer->attachments[subpass->colors[i].idx];
struct dzn_image *src_img = container_of(src->vk.image, dzn_image, vk);
struct dzn_image_view *dst = framebuffer->attachments[subpass->resolve[i].idx];
struct dzn_image *dst_img = container_of(dst->vk.image, dzn_image, vk);
D3D12_RESOURCE_BARRIER barriers[2];
uint32_t barrier_count = 0;
/* TODO: 2DArrays/3D */
if (subpass->colors[i].during != D3D12_RESOURCE_STATE_RESOLVE_SOURCE) {
barriers[barrier_count++] = D3D12_RESOURCE_BARRIER {
.Type = D3D12_RESOURCE_BARRIER_TYPE_TRANSITION,
.Flags = D3D12_RESOURCE_BARRIER_FLAG_NONE,
.Transition = {
.pResource = src_img->res,
.Subresource = 0,
.StateBefore = subpass->colors[i].during,
.StateAfter = D3D12_RESOURCE_STATE_RESOLVE_SOURCE,
},
};
}
if (subpass->resolve[i].during != D3D12_RESOURCE_STATE_RESOLVE_DEST) {
barriers[barrier_count++] = D3D12_RESOURCE_BARRIER {
.Type = D3D12_RESOURCE_BARRIER_TYPE_TRANSITION,
.Flags = D3D12_RESOURCE_BARRIER_FLAG_NONE,
.Transition = {
.pResource = dst_img->res,
.Subresource = 0,
.StateBefore = subpass->resolve[i].during,
.StateAfter = D3D12_RESOURCE_STATE_RESOLVE_DEST,
},
};
}
if (barrier_count)
cmdbuf->cmdlist->ResourceBarrier(barrier_count, barriers);
cmdbuf->cmdlist->ResolveSubresource(dst_img->res, 0,
src_img->res, 0,
dst->srv_desc.Format);
for (uint32_t b = 0; b < barrier_count; b++)
DZN_SWAP(barriers[b].Transition.StateBefore, barriers[b].Transition.StateAfter);
if (barrier_count)
cmdbuf->cmdlist->ResourceBarrier(barrier_count, barriers);
}
static void
dzn_cmd_buffer_begin_subpass(dzn_cmd_buffer *cmdbuf)
{
struct dzn_framebuffer *framebuffer = cmdbuf->state.framebuffer;
struct dzn_render_pass *pass = cmdbuf->state.pass;
const struct dzn_subpass *subpass = &pass->subpasses[cmdbuf->state.subpass];
D3D12_CPU_DESCRIPTOR_HANDLE rt_handles[MAX_RTS] = { };
D3D12_CPU_DESCRIPTOR_HANDLE zs_handle = { 0 };
for (uint32_t i = 0; i < subpass->color_count; i++) {
if (subpass->colors[i].idx == VK_ATTACHMENT_UNUSED) continue;
dzn_image_view *iview = framebuffer->attachments[subpass->colors[i].idx];
dzn_image *img = container_of(iview->vk.image, dzn_image, vk);
rt_handles[i] = dzn_cmd_buffer_get_rtv(cmdbuf, img, &iview->rtv_desc);
}
if (subpass->zs.idx != VK_ATTACHMENT_UNUSED) {
dzn_image_view *iview = framebuffer->attachments[subpass->zs.idx];
dzn_image *img = container_of(iview->vk.image, dzn_image, vk);
zs_handle = dzn_cmd_buffer_get_dsv(cmdbuf, img, &iview->dsv_desc);
}
cmdbuf->cmdlist->OMSetRenderTargets(subpass->color_count,
subpass->color_count ? rt_handles : NULL,
FALSE, zs_handle.ptr ? &zs_handle : NULL);
for (uint32_t i = 0; i < subpass->color_count; i++)
dzn_cmd_buffer_attachment_ref_transition(cmdbuf, &subpass->colors[i]);
for (uint32_t i = 0; i < subpass->input_count; i++)
dzn_cmd_buffer_attachment_ref_transition(cmdbuf, &subpass->inputs[i]);
if (subpass->zs.idx != VK_ATTACHMENT_UNUSED)
dzn_cmd_buffer_attachment_ref_transition(cmdbuf, &subpass->zs);
}
static void
dzn_cmd_buffer_end_subpass(dzn_cmd_buffer *cmdbuf)
{
const dzn_subpass *subpass = &cmdbuf->state.pass->subpasses[cmdbuf->state.subpass];
for (uint32_t i = 0; i < subpass->color_count; i++)
dzn_cmd_buffer_resolve_attachment(cmdbuf, i);
}
static void
dzn_cmd_buffer_update_pipeline(dzn_cmd_buffer *cmdbuf, uint32_t bindpoint)
{
const dzn_pipeline *pipeline = cmdbuf->state.bindpoint[bindpoint].pipeline;
if (!pipeline)
return;
if (cmdbuf->state.bindpoint[bindpoint].dirty & DZN_CMD_BINDPOINT_DIRTY_PIPELINE) {
if (bindpoint == VK_PIPELINE_BIND_POINT_GRAPHICS) {
const dzn_graphics_pipeline *gfx =
reinterpret_cast<const dzn_graphics_pipeline *>(pipeline);
cmdbuf->cmdlist->SetGraphicsRootSignature(pipeline->root.sig);
cmdbuf->cmdlist->IASetPrimitiveTopology(gfx->ia.topology);
} else {
cmdbuf->cmdlist->SetComputeRootSignature(pipeline->root.sig);
}
}
if (cmdbuf->state.pipeline != pipeline) {
cmdbuf->cmdlist->SetPipelineState(pipeline->state);
cmdbuf->state.pipeline = pipeline;
}
}
static void
dzn_cmd_buffer_update_heaps(dzn_cmd_buffer *cmdbuf, uint32_t bindpoint)
{
dzn_device *device = container_of(cmdbuf->vk.base.device, dzn_device, vk);
struct dzn_descriptor_state *desc_state =
&cmdbuf->state.bindpoint[bindpoint].desc_state;
dzn_descriptor_heap *new_heaps[NUM_POOL_TYPES] = {
desc_state->heaps[D3D12_DESCRIPTOR_HEAP_TYPE_CBV_SRV_UAV],
desc_state->heaps[D3D12_DESCRIPTOR_HEAP_TYPE_SAMPLER]
};
uint32_t new_heap_offsets[NUM_POOL_TYPES] = {};
bool update_root_desc_table[NUM_POOL_TYPES] = {};
const struct dzn_pipeline *pipeline =
cmdbuf->state.bindpoint[bindpoint].pipeline;
if (!(cmdbuf->state.bindpoint[bindpoint].dirty & DZN_CMD_BINDPOINT_DIRTY_HEAPS))
goto set_heaps;
dzn_foreach_pool_type (type) {
uint32_t desc_count = pipeline->desc_count[type];
if (!desc_count)
continue;
dzn_descriptor_heap_pool *pool =
type == D3D12_DESCRIPTOR_HEAP_TYPE_CBV_SRV_UAV ?
&cmdbuf->cbv_srv_uav_pool : &cmdbuf->sampler_pool;
uint32_t dst_offset = 0;
dzn_descriptor_heap *dst_heap = NULL;
uint32_t dst_heap_offset = 0;
dzn_descriptor_heap_pool_alloc_slots(pool, device, desc_count,
&dst_heap, &dst_heap_offset);
new_heap_offsets[type] = dst_heap_offset;
update_root_desc_table[type] = true;
for (uint32_t s = 0; s < MAX_SETS; s++) {
const struct dzn_descriptor_set *set = desc_state->sets[s].set;
if (!set) continue;
uint32_t set_heap_offset = pipeline->sets[s].heap_offsets[type];
uint32_t set_desc_count = pipeline->sets[s].range_desc_count[type];
if (set_desc_count) {
mtx_lock(&set->pool->defragment_lock);
dzn_descriptor_heap_copy(dst_heap, dst_heap_offset + set_heap_offset,
&set->pool->heaps[type], set->heap_offsets[type],
set_desc_count);
mtx_unlock(&set->pool->defragment_lock);
}
if (type == D3D12_DESCRIPTOR_HEAP_TYPE_CBV_SRV_UAV) {
uint32_t dynamic_buffer_count = pipeline->sets[s].dynamic_buffer_count;
for (uint32_t o = 0; o < dynamic_buffer_count; o++) {
uint32_t desc_heap_offset =
pipeline->sets[s].dynamic_buffer_heap_offsets[o].srv;
dzn_buffer_desc bdesc = set->dynamic_buffers[o];
bdesc.offset += desc_state->sets[s].dynamic_offsets[o];
dzn_descriptor_heap_write_buffer_desc(dst_heap,
dst_heap_offset + set_heap_offset + desc_heap_offset,
false, &bdesc);
if (pipeline->sets[s].dynamic_buffer_heap_offsets[o].uav != ~0) {
desc_heap_offset = pipeline->sets[s].dynamic_buffer_heap_offsets[o].uav;
dzn_descriptor_heap_write_buffer_desc(dst_heap,
dst_heap_offset + set_heap_offset + desc_heap_offset,
true, &bdesc);
}
}
}
}
new_heaps[type] = dst_heap;
}
set_heaps:
if (new_heaps[D3D12_DESCRIPTOR_HEAP_TYPE_CBV_SRV_UAV] != cmdbuf->state.heaps[D3D12_DESCRIPTOR_HEAP_TYPE_CBV_SRV_UAV] ||
new_heaps[D3D12_DESCRIPTOR_HEAP_TYPE_SAMPLER] != cmdbuf->state.heaps[D3D12_DESCRIPTOR_HEAP_TYPE_SAMPLER]) {
ID3D12DescriptorHeap *desc_heaps[2];
uint32_t num_desc_heaps = 0;
if (new_heaps[D3D12_DESCRIPTOR_HEAP_TYPE_CBV_SRV_UAV])
desc_heaps[num_desc_heaps++] = new_heaps[D3D12_DESCRIPTOR_HEAP_TYPE_CBV_SRV_UAV]->heap;
if (new_heaps[D3D12_DESCRIPTOR_HEAP_TYPE_SAMPLER])
desc_heaps[num_desc_heaps++] = new_heaps[D3D12_DESCRIPTOR_HEAP_TYPE_SAMPLER]->heap;
cmdbuf->cmdlist->SetDescriptorHeaps(num_desc_heaps, desc_heaps);
for (unsigned h = 0; h < ARRAY_SIZE(cmdbuf->state.heaps); h++)
cmdbuf->state.heaps[h] = new_heaps[h];
}
for (uint32_t r = 0; r < pipeline->root.sets_param_count; r++) {
D3D12_DESCRIPTOR_HEAP_TYPE type = pipeline->root.type[r];
if (!update_root_desc_table[type])
continue;
D3D12_GPU_DESCRIPTOR_HANDLE handle =
dzn_descriptor_heap_get_gpu_handle(new_heaps[type], new_heap_offsets[type]);
if (bindpoint == VK_PIPELINE_BIND_POINT_GRAPHICS)
cmdbuf->cmdlist->SetGraphicsRootDescriptorTable(r, handle);
else
cmdbuf->cmdlist->SetComputeRootDescriptorTable(r, handle);
}
}
static void
dzn_cmd_buffer_update_sysvals(dzn_cmd_buffer *cmdbuf, uint32_t bindpoint)
{
if (!(cmdbuf->state.bindpoint[bindpoint].dirty & DZN_CMD_BINDPOINT_DIRTY_SYSVALS))
return;
const struct dzn_pipeline *pipeline = cmdbuf->state.bindpoint[bindpoint].pipeline;
uint32_t sysval_cbv_param_idx = pipeline->root.sysval_cbv_param_idx;
if (bindpoint == VK_PIPELINE_BIND_POINT_GRAPHICS) {
cmdbuf->cmdlist->SetGraphicsRoot32BitConstants(sysval_cbv_param_idx,
sizeof(cmdbuf->state.sysvals.gfx) / 4,
&cmdbuf->state.sysvals.gfx, 0);
} else {
cmdbuf->cmdlist->SetComputeRoot32BitConstants(sysval_cbv_param_idx,
sizeof(cmdbuf->state.sysvals.compute) / 4,
&cmdbuf->state.sysvals.compute, 0);
}
}
static void
dzn_cmd_buffer_update_viewports(dzn_cmd_buffer *cmdbuf)
{
const dzn_graphics_pipeline *pipeline =
reinterpret_cast<const dzn_graphics_pipeline *>(cmdbuf->state.pipeline);
if (!(cmdbuf->state.dirty & DZN_CMD_DIRTY_VIEWPORTS) ||
!pipeline->vp.count)
return;
cmdbuf->cmdlist->RSSetViewports(pipeline->vp.count, cmdbuf->state.viewports);
}
static void
dzn_cmd_buffer_update_scissors(dzn_cmd_buffer *cmdbuf)
{
const dzn_graphics_pipeline *pipeline =
reinterpret_cast<const dzn_graphics_pipeline *>(cmdbuf->state.pipeline);
if (!(cmdbuf->state.dirty & DZN_CMD_DIRTY_SCISSORS))
return;
if (!pipeline->scissor.count) {
/* Apply a scissor delimiting the render area. */
cmdbuf->cmdlist->RSSetScissorRects(1, &cmdbuf->state.render_area);
return;
}
D3D12_RECT scissors[MAX_SCISSOR];
uint32_t scissor_count = pipeline->scissor.count;
memcpy(scissors, cmdbuf->state.scissors, sizeof(D3D12_RECT) * pipeline->scissor.count);
for (uint32_t i = 0; i < pipeline->scissor.count; i++) {
scissors[i].left = MAX2(scissors[i].left, cmdbuf->state.render_area.left);
scissors[i].top = MAX2(scissors[i].top, cmdbuf->state.render_area.top);
scissors[i].right = MIN2(scissors[i].right, cmdbuf->state.render_area.right);
scissors[i].bottom = MIN2(scissors[i].bottom, cmdbuf->state.render_area.bottom);
}
cmdbuf->cmdlist->RSSetScissorRects(pipeline->scissor.count, scissors);
}
static void
dzn_cmd_buffer_update_vbviews(dzn_cmd_buffer *cmdbuf)
{
const dzn_graphics_pipeline *pipeline =
reinterpret_cast<const dzn_graphics_pipeline *>(cmdbuf->state.pipeline);
unsigned start, end;
BITSET_FOREACH_RANGE(start, end, cmdbuf->state.vb.dirty, MAX_VBS)
cmdbuf->cmdlist->IASetVertexBuffers(start, end - start, cmdbuf->state.vb.views);
BITSET_CLEAR_RANGE(cmdbuf->state.vb.dirty, 0, MAX_VBS);
}
static void
dzn_cmd_buffer_update_ibview(dzn_cmd_buffer *cmdbuf)
{
if (!(cmdbuf->state.dirty & DZN_CMD_DIRTY_IB))
return;
cmdbuf->cmdlist->IASetIndexBuffer(&cmdbuf->state.ib.view);
}
static void
dzn_cmd_buffer_update_push_constants(dzn_cmd_buffer *cmdbuf, uint32_t bindpoint)
{
struct dzn_cmd_buffer_push_constant_state *state =
bindpoint == VK_PIPELINE_BIND_POINT_GRAPHICS ?
&cmdbuf->state.push_constant.gfx : &cmdbuf->state.push_constant.compute;
uint32_t offset = state->offset / 4;
uint32_t end = ALIGN(state->end, 4) / 4;
uint32_t count = end - offset;
if (!count)
return;
uint32_t slot = cmdbuf->state.pipeline->root.push_constant_cbv_param_idx;
uint32_t *vals = state->values + offset;
if (bindpoint == VK_PIPELINE_BIND_POINT_GRAPHICS)
cmdbuf->cmdlist->SetGraphicsRoot32BitConstants(slot, count, vals, offset);
else
cmdbuf->cmdlist->SetComputeRoot32BitConstants(slot, count, vals, offset);
state->offset = 0;
state->end = 0;
}
void
dzn_cmd_buffer_update_zsa(dzn_cmd_buffer *cmdbuf)
{
if (cmdbuf->state.dirty & DZN_CMD_DIRTY_STENCIL_REF) {
const dzn_graphics_pipeline *gfx = (const dzn_graphics_pipeline *)
cmdbuf->state.bindpoint[VK_PIPELINE_BIND_POINT_GRAPHICS].pipeline;
uint32_t ref =
gfx->zsa.stencil_test.front.uses_ref ?
cmdbuf->state.zsa.stencil_test.front.ref :
cmdbuf->state.zsa.stencil_test.back.ref;
cmdbuf->cmdlist->OMSetStencilRef(ref);
}
}
static VkResult
dzn_cmd_buffer_triangle_fan_create_index(dzn_cmd_buffer *cmdbuf, uint32_t *vertex_count)
{
dzn_device *device = container_of(cmdbuf->vk.base.device, dzn_device, vk);
uint8_t index_size = *vertex_count <= 0xffff ? 2 : 4;
uint32_t triangle_count = MAX2(*vertex_count, 2) - 2;
*vertex_count = triangle_count * 3;
if (!*vertex_count)
return VK_SUCCESS;
ID3D12Resource *index_buf;
VkResult result =
dzn_cmd_buffer_alloc_internal_buf(cmdbuf, *vertex_count * index_size,
D3D12_HEAP_TYPE_UPLOAD,
D3D12_RESOURCE_STATE_GENERIC_READ,
&index_buf);
if (result != VK_SUCCESS)
return result;
void *cpu_ptr;
index_buf->Map(0, NULL, &cpu_ptr);
/* TODO: VK_PROVOKING_VERTEX_MODE_LAST_VERTEX_EXT */
if (index_size == 2) {
uint16_t *indices = (uint16_t *)cpu_ptr;
for (uint32_t t = 0; t < triangle_count; t++) {
indices[t * 3] = t + 1;
indices[(t * 3) + 1] = t + 2;
indices[(t * 3) + 2] = 0;
}
cmdbuf->state.ib.view.Format = DXGI_FORMAT_R16_UINT;
} else {
uint32_t *indices = (uint32_t *)cpu_ptr;
for (uint32_t t = 0; t < triangle_count; t++) {
indices[t * 3] = t + 1;
indices[(t * 3) + 1] = t + 2;
indices[(t * 3) + 2] = 0;
}
cmdbuf->state.ib.view.Format = DXGI_FORMAT_R32_UINT;
}
cmdbuf->state.ib.view.SizeInBytes = *vertex_count * index_size;
cmdbuf->state.ib.view.BufferLocation = index_buf->GetGPUVirtualAddress();
cmdbuf->state.dirty |= DZN_CMD_DIRTY_IB;
return VK_SUCCESS;
}
static VkResult
dzn_cmd_buffer_triangle_fan_rewrite_index(dzn_cmd_buffer *cmdbuf,
uint32_t *index_count,
uint32_t *first_index)
{
dzn_device *device = container_of(cmdbuf->vk.base.device, dzn_device, vk);
uint32_t triangle_count = MAX2(*index_count, 2) - 2;
*index_count = triangle_count * 3;
if (!*index_count)
return VK_SUCCESS;
/* New index is always 32bit to make the compute shader rewriting the
* index simpler */
ID3D12Resource *new_index_buf;
VkResult result =
dzn_cmd_buffer_alloc_internal_buf(cmdbuf, *index_count * 4,
D3D12_HEAP_TYPE_DEFAULT,
D3D12_RESOURCE_STATE_UNORDERED_ACCESS,
&new_index_buf);
if (result != VK_SUCCESS)
return result;
D3D12_GPU_VIRTUAL_ADDRESS old_index_buf_gpu =
cmdbuf->state.ib.view.BufferLocation;
enum dzn_index_type index_type =
dzn_index_type_from_dxgi_format(cmdbuf->state.ib.view.Format);
const dzn_meta_triangle_fan_rewrite_index *rewrite_index =
&device->triangle_fan[index_type];
const dzn_pipeline *compute_pipeline =
cmdbuf->state.bindpoint[VK_PIPELINE_BIND_POINT_COMPUTE].pipeline;
struct dzn_triangle_fan_rewrite_index_params params = {
.first_index = *first_index,
};
cmdbuf->cmdlist->SetComputeRootSignature(rewrite_index->root_sig);
cmdbuf->cmdlist->SetPipelineState(rewrite_index->pipeline_state);
cmdbuf->cmdlist->SetComputeRootUnorderedAccessView(0, new_index_buf->GetGPUVirtualAddress());
cmdbuf->cmdlist->SetComputeRoot32BitConstants(1, sizeof(params) / 4,
&params, 0);
cmdbuf->cmdlist->SetComputeRootShaderResourceView(2, old_index_buf_gpu);
cmdbuf->cmdlist->Dispatch(triangle_count, 1, 1);
D3D12_RESOURCE_BARRIER post_barriers[] = {
{
.Type = D3D12_RESOURCE_BARRIER_TYPE_TRANSITION,
.Flags = D3D12_RESOURCE_BARRIER_FLAG_NONE,
/* Transition the exec buffer to indirect arg so it can be
* pass to ExecuteIndirect() as an argument buffer.
*/
.Transition = {
.pResource = new_index_buf,
.Subresource = 0,
.StateBefore = D3D12_RESOURCE_STATE_UNORDERED_ACCESS,
.StateAfter = D3D12_RESOURCE_STATE_INDEX_BUFFER,
},
},
};
cmdbuf->cmdlist->ResourceBarrier(ARRAY_SIZE(post_barriers), post_barriers);
/* We don't mess up with the driver state when executing our internal
* compute shader, but we still change the D3D12 state, so let's mark
* things dirty if needed.
*/
cmdbuf->state.pipeline = NULL;
if (cmdbuf->state.bindpoint[VK_PIPELINE_BIND_POINT_COMPUTE].pipeline) {
cmdbuf->state.bindpoint[VK_PIPELINE_BIND_POINT_COMPUTE].dirty |=
DZN_CMD_BINDPOINT_DIRTY_PIPELINE;
}
cmdbuf->state.ib.view.SizeInBytes = *index_count * 4;
cmdbuf->state.ib.view.BufferLocation = new_index_buf->GetGPUVirtualAddress();
cmdbuf->state.ib.view.Format = DXGI_FORMAT_R32_UINT;
cmdbuf->state.dirty |= DZN_CMD_DIRTY_IB;
*first_index = 0;
return VK_SUCCESS;
}
static void
dzn_cmd_buffer_prepare_draw(dzn_cmd_buffer *cmdbuf, bool indexed)
{
dzn_cmd_buffer_update_pipeline(cmdbuf, VK_PIPELINE_BIND_POINT_GRAPHICS);
dzn_cmd_buffer_update_heaps(cmdbuf, VK_PIPELINE_BIND_POINT_GRAPHICS);
dzn_cmd_buffer_update_sysvals(cmdbuf, VK_PIPELINE_BIND_POINT_GRAPHICS);
dzn_cmd_buffer_update_viewports(cmdbuf);
dzn_cmd_buffer_update_scissors(cmdbuf);
dzn_cmd_buffer_update_vbviews(cmdbuf);
dzn_cmd_buffer_update_push_constants(cmdbuf, VK_PIPELINE_BIND_POINT_GRAPHICS);
dzn_cmd_buffer_update_zsa(cmdbuf);
if (indexed)
dzn_cmd_buffer_update_ibview(cmdbuf);
/* Reset the dirty states */
cmdbuf->state.bindpoint[VK_PIPELINE_BIND_POINT_GRAPHICS].dirty = 0;
cmdbuf->state.dirty = 0;
}
static uint32_t
dzn_cmd_buffer_triangle_fan_get_max_index_buf_size(dzn_cmd_buffer *cmdbuf, bool indexed)
{
dzn_graphics_pipeline *pipeline = (dzn_graphics_pipeline *)
cmdbuf->state.bindpoint[VK_PIPELINE_BIND_POINT_GRAPHICS].pipeline;
if (!pipeline->ia.triangle_fan)
return 0;
uint32_t max_triangles;
if (indexed) {
uint32_t index_size = cmdbuf->state.ib.view.Format == DXGI_FORMAT_R32_UINT ? 4 : 2;
uint32_t max_indices = cmdbuf->state.ib.view.SizeInBytes / index_size;
max_triangles = MAX2(max_indices, 2) - 2;
} else {
uint32_t max_vertex = 0;
for (uint32_t i = 0; i < pipeline->vb.count; i++) {
max_vertex =
MAX2(max_vertex,
cmdbuf->state.vb.views[i].SizeInBytes / cmdbuf->state.vb.views[i].StrideInBytes);
}
max_triangles = MAX2(max_vertex, 2) - 2;
}
return max_triangles * 3;
}
static void
dzn_cmd_buffer_indirect_draw(dzn_cmd_buffer *cmdbuf,
dzn_buffer *draw_buf,
size_t draw_buf_offset,
uint32_t draw_count,
uint32_t draw_buf_stride,
bool indexed)
{
dzn_device *device = container_of(cmdbuf->vk.base.device, dzn_device, vk);
dzn_graphics_pipeline *pipeline = (dzn_graphics_pipeline *)
cmdbuf->state.bindpoint[VK_PIPELINE_BIND_POINT_GRAPHICS].pipeline;
bool triangle_fan = pipeline->ia.triangle_fan;
uint32_t min_draw_buf_stride =
indexed ?
sizeof(struct dzn_indirect_indexed_draw_params) :
sizeof(struct dzn_indirect_draw_params);
draw_buf_stride = draw_buf_stride ? draw_buf_stride : min_draw_buf_stride;
assert(draw_buf_stride >= min_draw_buf_stride);
assert((draw_buf_stride & 3) == 0);
uint32_t sysvals_stride = ALIGN_POT(sizeof(cmdbuf->state.sysvals.gfx), 256);
uint32_t exec_buf_stride = 32;
uint32_t triangle_fan_index_buf_stride =
dzn_cmd_buffer_triangle_fan_get_max_index_buf_size(cmdbuf, indexed) *
sizeof(uint32_t);
uint32_t triangle_fan_exec_buf_stride =
sizeof(struct dzn_indirect_triangle_fan_rewrite_index_exec_params);
ID3D12Resource *exec_buf;
VkResult result =
dzn_cmd_buffer_alloc_internal_buf(cmdbuf, draw_count * exec_buf_stride,
D3D12_HEAP_TYPE_DEFAULT,
D3D12_RESOURCE_STATE_UNORDERED_ACCESS,
&exec_buf);
if (result != VK_SUCCESS)
return;
D3D12_GPU_VIRTUAL_ADDRESS draw_buf_gpu =
draw_buf->res->GetGPUVirtualAddress() + draw_buf_offset;
ID3D12Resource *triangle_fan_index_buf = NULL;
ID3D12Resource *triangle_fan_exec_buf = NULL;
if (triangle_fan_index_buf_stride) {
result =
dzn_cmd_buffer_alloc_internal_buf(cmdbuf,
draw_count * triangle_fan_index_buf_stride,
D3D12_HEAP_TYPE_DEFAULT,
D3D12_RESOURCE_STATE_UNORDERED_ACCESS,
&triangle_fan_index_buf);
if (result != VK_SUCCESS)
return;
result =
dzn_cmd_buffer_alloc_internal_buf(cmdbuf,
draw_count * triangle_fan_exec_buf_stride,
D3D12_HEAP_TYPE_DEFAULT,
D3D12_RESOURCE_STATE_UNORDERED_ACCESS,
&triangle_fan_exec_buf);
if (result != VK_SUCCESS)
return;
}
struct dzn_indirect_draw_triangle_fan_rewrite_params params = {
.draw_buf_stride = draw_buf_stride,
.triangle_fan_index_buf_stride = triangle_fan_index_buf_stride,
.triangle_fan_index_buf_start =
triangle_fan_index_buf ?
triangle_fan_index_buf->GetGPUVirtualAddress() : 0,
};
uint32_t params_size =
triangle_fan_index_buf_stride > 0 ?
sizeof(struct dzn_indirect_draw_triangle_fan_rewrite_params) :
sizeof(struct dzn_indirect_draw_rewrite_params);
enum dzn_indirect_draw_type draw_type;
if (indexed && triangle_fan_index_buf_stride > 0)
draw_type = DZN_INDIRECT_INDEXED_DRAW_TRIANGLE_FAN;
else if (!indexed && triangle_fan_index_buf_stride > 0)
draw_type = DZN_INDIRECT_DRAW_TRIANGLE_FAN;
else if (indexed)
draw_type = DZN_INDIRECT_INDEXED_DRAW;
else
draw_type = DZN_INDIRECT_DRAW;
dzn_meta_indirect_draw *indirect_draw = &device->indirect_draws[draw_type];
const dzn_pipeline *compute_pipeline =
cmdbuf->state.bindpoint[VK_PIPELINE_BIND_POINT_COMPUTE].pipeline;
cmdbuf->cmdlist->SetComputeRootSignature(indirect_draw->root_sig);
cmdbuf->cmdlist->SetPipelineState(indirect_draw->pipeline_state);
cmdbuf->cmdlist->SetComputeRoot32BitConstants(0, params_size / 4, (const void *)&params, 0);
cmdbuf->cmdlist->SetComputeRootShaderResourceView(1, draw_buf_gpu);
cmdbuf->cmdlist->SetComputeRootUnorderedAccessView(2, exec_buf->GetGPUVirtualAddress());
if (triangle_fan_exec_buf)
cmdbuf->cmdlist->SetComputeRootUnorderedAccessView(3, triangle_fan_exec_buf->GetGPUVirtualAddress());
cmdbuf->cmdlist->Dispatch(draw_count, 1, 1);
D3D12_RESOURCE_BARRIER post_barriers[] = {
{
.Type = D3D12_RESOURCE_BARRIER_TYPE_TRANSITION,
.Flags = D3D12_RESOURCE_BARRIER_FLAG_NONE,
/* Transition the exec buffer to indirect arg so it can be
* pass to ExecuteIndirect() as an argument buffer.
*/
.Transition = {
.pResource = exec_buf,
.Subresource = 0,
.StateBefore = D3D12_RESOURCE_STATE_UNORDERED_ACCESS,
.StateAfter = D3D12_RESOURCE_STATE_INDIRECT_ARGUMENT,
},
},
{
.Type = D3D12_RESOURCE_BARRIER_TYPE_TRANSITION,
.Flags = D3D12_RESOURCE_BARRIER_FLAG_NONE,
/* Transition the exec buffer to indirect arg so it can be
* pass to ExecuteIndirect() as an argument buffer.
*/
.Transition = {
.pResource = triangle_fan_exec_buf,
.Subresource = 0,
.StateBefore = D3D12_RESOURCE_STATE_UNORDERED_ACCESS,
.StateAfter = D3D12_RESOURCE_STATE_INDIRECT_ARGUMENT,
},
},
};
uint32_t post_barrier_count = triangle_fan_exec_buf ? 2 : 1;
cmdbuf->cmdlist->ResourceBarrier(post_barrier_count, post_barriers);
D3D12_INDEX_BUFFER_VIEW ib_view = {};
if (triangle_fan_exec_buf) {
auto index_type =
indexed ?
dzn_index_type_from_dxgi_format(cmdbuf->state.ib.view.Format) :
DZN_NO_INDEX;
dzn_meta_triangle_fan_rewrite_index *rewrite_index =
&device->triangle_fan[index_type];
struct dzn_triangle_fan_rewrite_index_params rewrite_index_params = {};
assert(rewrite_index->root_sig);
assert(rewrite_index->pipeline_state);
assert(rewrite_index->cmd_sig);
cmdbuf->cmdlist->SetComputeRootSignature(rewrite_index->root_sig);
cmdbuf->cmdlist->SetPipelineState(rewrite_index->pipeline_state);
cmdbuf->cmdlist->SetComputeRootUnorderedAccessView(0, triangle_fan_index_buf->GetGPUVirtualAddress());
cmdbuf->cmdlist->SetComputeRoot32BitConstants(1, sizeof(rewrite_index_params) / 4,
(const void *)&rewrite_index_params, 0);
if (indexed)
cmdbuf->cmdlist->SetComputeRootShaderResourceView(2, cmdbuf->state.ib.view.BufferLocation);
cmdbuf->cmdlist->ExecuteIndirect(rewrite_index->cmd_sig,
draw_count, triangle_fan_exec_buf,
0, NULL, 0);
D3D12_RESOURCE_BARRIER index_buf_barriers[] = {
{
.Type = D3D12_RESOURCE_BARRIER_TYPE_TRANSITION,
.Flags = D3D12_RESOURCE_BARRIER_FLAG_NONE,
.Transition = {
.pResource = triangle_fan_index_buf,
.Subresource = 0,
.StateBefore = D3D12_RESOURCE_STATE_UNORDERED_ACCESS,
.StateAfter = D3D12_RESOURCE_STATE_INDEX_BUFFER,
},
},
};
cmdbuf->cmdlist->ResourceBarrier(ARRAY_SIZE(index_buf_barriers), index_buf_barriers);
/* After our triangle-fan lowering the draw is indexed */
indexed = true;
ib_view = cmdbuf->state.ib.view;
cmdbuf->state.ib.view.BufferLocation = triangle_fan_index_buf->GetGPUVirtualAddress();
cmdbuf->state.ib.view.SizeInBytes = triangle_fan_index_buf_stride;
cmdbuf->state.ib.view.Format = DXGI_FORMAT_R32_UINT;
cmdbuf->state.dirty |= DZN_CMD_DIRTY_IB;
}
/* We don't mess up with the driver state when executing our internal
* compute shader, but we still change the D3D12 state, so let's mark
* things dirty if needed.
*/
cmdbuf->state.pipeline = NULL;
if (cmdbuf->state.bindpoint[VK_PIPELINE_BIND_POINT_COMPUTE].pipeline) {
cmdbuf->state.bindpoint[VK_PIPELINE_BIND_POINT_COMPUTE].dirty |=
DZN_CMD_BINDPOINT_DIRTY_PIPELINE;
}
cmdbuf->state.sysvals.gfx.first_vertex = 0;
cmdbuf->state.sysvals.gfx.base_instance = 0;
cmdbuf->state.sysvals.gfx.is_indexed_draw = indexed;
cmdbuf->state.bindpoint[VK_PIPELINE_BIND_POINT_GRAPHICS].dirty |=
DZN_CMD_BINDPOINT_DIRTY_SYSVALS;
dzn_cmd_buffer_prepare_draw(cmdbuf, indexed);
/* Restore the old IB view if we modified it during the triangle fan lowering */
if (ib_view.SizeInBytes) {
cmdbuf->state.ib.view = ib_view;
cmdbuf->state.dirty |= DZN_CMD_DIRTY_IB;
}
enum dzn_indirect_draw_cmd_sig_type cmd_sig_type =
triangle_fan_index_buf_stride > 0 ?
DZN_INDIRECT_DRAW_TRIANGLE_FAN_CMD_SIG :
indexed ?
DZN_INDIRECT_INDEXED_DRAW_CMD_SIG :
DZN_INDIRECT_DRAW_CMD_SIG;
ID3D12CommandSignature *cmdsig =
dzn_graphics_pipeline_get_indirect_cmd_sig(pipeline, cmd_sig_type);
if (!cmdsig) {
cmdbuf->error = vk_error(device, VK_ERROR_OUT_OF_DEVICE_MEMORY);
return;
}
cmdbuf->cmdlist->ExecuteIndirect(cmdsig,
draw_count, exec_buf, 0, NULL, 0);
}
static void
dzn_cmd_buffer_prepare_dispatch(dzn_cmd_buffer *cmdbuf)
{
dzn_cmd_buffer_update_pipeline(cmdbuf, VK_PIPELINE_BIND_POINT_COMPUTE);
dzn_cmd_buffer_update_heaps(cmdbuf, VK_PIPELINE_BIND_POINT_COMPUTE);
dzn_cmd_buffer_update_sysvals(cmdbuf, VK_PIPELINE_BIND_POINT_COMPUTE);
dzn_cmd_buffer_update_push_constants(cmdbuf, VK_PIPELINE_BIND_POINT_COMPUTE);
/* Reset the dirty states */
cmdbuf->state.bindpoint[VK_PIPELINE_BIND_POINT_COMPUTE].dirty = 0;
}
VKAPI_ATTR void VKAPI_CALL
dzn_CmdCopyBuffer2(VkCommandBuffer commandBuffer,
const VkCopyBufferInfo2 *info)
{
VK_FROM_HANDLE(dzn_cmd_buffer, cmdbuf, commandBuffer);
VK_FROM_HANDLE(dzn_buffer, src_buffer, info->srcBuffer);
VK_FROM_HANDLE(dzn_buffer, dst_buffer, info->dstBuffer);
for (int i = 0; i < info->regionCount; i++) {
auto &region = info->pRegions[i];
cmdbuf->cmdlist->CopyBufferRegion(dst_buffer->res, region.dstOffset,
src_buffer->res, region.srcOffset,
region.size);
}
}
VKAPI_ATTR void VKAPI_CALL
dzn_CmdCopyBufferToImage2(VkCommandBuffer commandBuffer,
const VkCopyBufferToImageInfo2 *info)
{
VK_FROM_HANDLE(dzn_cmd_buffer, cmdbuf, commandBuffer);
for (int i = 0; i < info->regionCount; i++) {
const VkBufferImageCopy2 &region = info->pRegions[i];
dzn_foreach_aspect(aspect, region.imageSubresource.aspectMask) {
for (uint32_t l = 0; l < region.imageSubresource.layerCount; l++)
dzn_cmd_buffer_copy_buf2img_region(cmdbuf, info, i, aspect, l);
}
}
}
VKAPI_ATTR void VKAPI_CALL
dzn_CmdCopyImageToBuffer2(VkCommandBuffer commandBuffer,
const VkCopyImageToBufferInfo2 *info)
{
VK_FROM_HANDLE(dzn_cmd_buffer, cmdbuf, commandBuffer);
for (int i = 0; i < info->regionCount; i++) {
const VkBufferImageCopy2 &region = info->pRegions[i];
dzn_foreach_aspect(aspect, region.imageSubresource.aspectMask) {
for (uint32_t l = 0; l < region.imageSubresource.layerCount; l++)
dzn_cmd_buffer_copy_img2buf_region(cmdbuf,info, i, aspect, l);
}
}
}
VKAPI_ATTR void VKAPI_CALL
dzn_CmdCopyImage2(VkCommandBuffer commandBuffer,
const VkCopyImageInfo2 *info)
{
VK_FROM_HANDLE(dzn_cmd_buffer, cmdbuf, commandBuffer);
dzn_device *device = container_of(cmdbuf->vk.base.device, dzn_device, vk);
VK_FROM_HANDLE(dzn_image, src, info->srcImage);
VK_FROM_HANDLE(dzn_image, dst, info->dstImage);
assert(src->vk.samples == dst->vk.samples);
bool requires_temp_res = src->vk.format != dst->vk.format &&
src->vk.tiling != VK_IMAGE_TILING_LINEAR &&
dst->vk.tiling != VK_IMAGE_TILING_LINEAR;
/* FIXME: multisample copies only work if we copy the entire subresource
* and if the the copy doesn't require a temporary linear resource. When
* these conditions are not met we should use a blit shader.
*/
if (src->vk.samples > 1) {
assert(requires_temp_res == false);
for (uint32_t i = 0; i < info->regionCount; i++) {
const VkImageCopy2 &region = info->pRegions[i];
uint32_t src_w = u_minify(src->vk.extent.width, region.srcSubresource.mipLevel);
uint32_t src_h = u_minify(src->vk.extent.width, region.srcSubresource.mipLevel);
assert(region.srcOffset.x == 0 && region.srcOffset.y == 0);
assert(region.extent.width == u_minify(src->vk.extent.width, region.srcSubresource.mipLevel));
assert(region.extent.height == u_minify(src->vk.extent.height, region.srcSubresource.mipLevel));
assert(region.dstOffset.x == 0 && region.dstOffset.y == 0);
assert(region.extent.width == u_minify(dst->vk.extent.width, region.dstSubresource.mipLevel));
assert(region.extent.height == u_minify(dst->vk.extent.height, region.dstSubresource.mipLevel));
}
}
D3D12_TEXTURE_COPY_LOCATION tmp_loc = {};
D3D12_RESOURCE_DESC tmp_desc = {
.Dimension = D3D12_RESOURCE_DIMENSION_TEXTURE2D,
.Alignment = D3D12_DEFAULT_RESOURCE_PLACEMENT_ALIGNMENT,
.DepthOrArraySize = 1,
.MipLevels = 1,
.Format = src->desc.Format,
.SampleDesc = { .Count = 1, .Quality = 0 },
.Layout = D3D12_TEXTURE_LAYOUT_UNKNOWN,
.Flags = D3D12_RESOURCE_FLAG_NONE,
};
if (requires_temp_res) {
ID3D12Device *dev = device->dev;
VkImageAspectFlags aspect = 0;
uint64_t max_size = 0;
if (vk_format_has_depth(src->vk.format))
aspect = VK_IMAGE_ASPECT_DEPTH_BIT;
else if (vk_format_has_stencil(src->vk.format))
aspect = VK_IMAGE_ASPECT_DEPTH_BIT;
else
aspect = VK_IMAGE_ASPECT_COLOR_BIT;
for (uint32_t i = 0; i < info->regionCount; i++) {
const VkImageCopy2 &region = info->pRegions[i];
uint64_t region_size = 0;
tmp_desc.Format =
dzn_image_get_dxgi_format(src->vk.format,
VK_IMAGE_USAGE_TRANSFER_DST_BIT,
aspect);
tmp_desc.Width = region.extent.width;
tmp_desc.Height = region.extent.height;
dev->GetCopyableFootprints(&src->desc,
0, 1, 0,
NULL, NULL, NULL,
&region_size);
max_size = MAX2(max_size, region_size * region.extent.depth);
}
VkResult result =
dzn_cmd_buffer_alloc_internal_buf(cmdbuf, max_size,
D3D12_HEAP_TYPE_DEFAULT,
D3D12_RESOURCE_STATE_COPY_DEST,
&tmp_loc.pResource);
if (result != VK_SUCCESS)
return;
tmp_loc.Type = D3D12_TEXTURE_COPY_TYPE_PLACED_FOOTPRINT;
}
for (int i = 0; i < info->regionCount; i++) {
const VkImageCopy2 &region = info->pRegions[i];
dzn_foreach_aspect(aspect, region.srcSubresource.aspectMask) {
for (uint32_t l = 0; l < region.srcSubresource.layerCount; l++)
dzn_cmd_buffer_copy_img_chunk(cmdbuf, info, tmp_desc, tmp_loc, i, aspect, l);
}
}
}
VKAPI_ATTR void VKAPI_CALL
dzn_CmdBlitImage2(VkCommandBuffer commandBuffer,
const VkBlitImageInfo2 *info)
{
VK_FROM_HANDLE(dzn_cmd_buffer, cmdbuf, commandBuffer);
dzn_device *device = container_of(cmdbuf->vk.base.device, dzn_device, vk);
if (info->regionCount == 0)
return;
uint32_t desc_count = 0;
for (uint32_t r = 0; r < info->regionCount; r++)
desc_count += util_bitcount(info->pRegions[r].srcSubresource.aspectMask);
dzn_descriptor_heap *heap;
uint32_t heap_slot;
VkResult result =
dzn_descriptor_heap_pool_alloc_slots(&cmdbuf->cbv_srv_uav_pool, device,
desc_count, &heap, &heap_slot);
if (result != VK_SUCCESS) {
cmdbuf->error = result;
return;
}
if (heap != cmdbuf->state.heaps[D3D12_DESCRIPTOR_HEAP_TYPE_CBV_SRV_UAV]) {
ID3D12DescriptorHeap * const heaps[] = { heap->heap };
cmdbuf->state.heaps[D3D12_DESCRIPTOR_HEAP_TYPE_CBV_SRV_UAV] = heap;
cmdbuf->cmdlist->SetDescriptorHeaps(ARRAY_SIZE(heaps), heaps);
}
cmdbuf->cmdlist->IASetPrimitiveTopology(D3D_PRIMITIVE_TOPOLOGY_TRIANGLESTRIP);
uint32_t heap_offset = 0;
for (uint32_t r = 0; r < info->regionCount; r++)
dzn_cmd_buffer_blit_region(cmdbuf, info, heap, &heap_slot, r);
cmdbuf->state.pipeline = NULL;
cmdbuf->state.dirty |= DZN_CMD_DIRTY_VIEWPORTS | DZN_CMD_DIRTY_SCISSORS;
if (cmdbuf->state.bindpoint[VK_PIPELINE_BIND_POINT_GRAPHICS].pipeline) {
cmdbuf->state.bindpoint[VK_PIPELINE_BIND_POINT_GRAPHICS].dirty |=
DZN_CMD_BINDPOINT_DIRTY_PIPELINE;
}
}
VKAPI_ATTR void VKAPI_CALL
dzn_CmdResolveImage2(VkCommandBuffer commandBuffer,
const VkResolveImageInfo2 *info)
{
VK_FROM_HANDLE(dzn_cmd_buffer, cmdbuf, commandBuffer);
dzn_device *device = container_of(cmdbuf->vk.base.device, dzn_device, vk);
if (info->regionCount == 0)
return;
uint32_t desc_count = 0;
for (uint32_t r = 0; r < info->regionCount; r++)
desc_count += util_bitcount(info->pRegions[r].srcSubresource.aspectMask);
dzn_descriptor_heap *heap;
uint32_t heap_slot;
VkResult result =
dzn_descriptor_heap_pool_alloc_slots(&cmdbuf->cbv_srv_uav_pool, device,
desc_count, &heap, &heap_slot);
if (result != VK_SUCCESS) {
cmdbuf->error = result;
return;
}
if (heap != cmdbuf->state.heaps[D3D12_DESCRIPTOR_HEAP_TYPE_CBV_SRV_UAV]) {
ID3D12DescriptorHeap * const heaps[] = { heap->heap };
cmdbuf->state.heaps[D3D12_DESCRIPTOR_HEAP_TYPE_CBV_SRV_UAV] = heap;
cmdbuf->cmdlist->SetDescriptorHeaps(ARRAY_SIZE(heaps), heaps);
}
cmdbuf->cmdlist->IASetPrimitiveTopology(D3D_PRIMITIVE_TOPOLOGY_TRIANGLESTRIP);
uint32_t heap_offset = 0;
for (uint32_t r = 0; r < info->regionCount; r++)
dzn_cmd_buffer_resolve_region(cmdbuf, info, heap, &heap_offset, r);
cmdbuf->state.pipeline = NULL;
cmdbuf->state.dirty |= DZN_CMD_DIRTY_VIEWPORTS | DZN_CMD_DIRTY_SCISSORS;
if (cmdbuf->state.bindpoint[VK_PIPELINE_BIND_POINT_GRAPHICS].pipeline) {
cmdbuf->state.bindpoint[VK_PIPELINE_BIND_POINT_GRAPHICS].dirty |=
DZN_CMD_BINDPOINT_DIRTY_PIPELINE;
}
}
VKAPI_ATTR void VKAPI_CALL
dzn_CmdClearColorImage(VkCommandBuffer commandBuffer,
VkImage image,
VkImageLayout imageLayout,
const VkClearColorValue *pColor,
uint32_t rangeCount,
const VkImageSubresourceRange *pRanges)
{
VK_FROM_HANDLE(dzn_cmd_buffer, cmdbuf, commandBuffer);
VK_FROM_HANDLE(dzn_image, img, image);
dzn_cmd_buffer_clear_color(cmdbuf, img, imageLayout, pColor, rangeCount, pRanges);
}
VKAPI_ATTR void VKAPI_CALL
dzn_CmdClearDepthStencilImage(VkCommandBuffer commandBuffer,
VkImage image,
VkImageLayout imageLayout,
const VkClearDepthStencilValue *pDepthStencil,
uint32_t rangeCount,
const VkImageSubresourceRange *pRanges)
{
VK_FROM_HANDLE(dzn_cmd_buffer, cmdbuf, commandBuffer);
VK_FROM_HANDLE(dzn_image, img, image);
dzn_cmd_buffer_clear_zs(cmdbuf, img, imageLayout, pDepthStencil, rangeCount, pRanges);
}
VKAPI_ATTR void VKAPI_CALL
dzn_CmdDispatch(VkCommandBuffer commandBuffer,
uint32_t groupCountX,
uint32_t groupCountY,
uint32_t groupCountZ)
{
VK_FROM_HANDLE(dzn_cmd_buffer, cmdbuf, commandBuffer);
cmdbuf->state.sysvals.compute.group_count_x = groupCountX;
cmdbuf->state.sysvals.compute.group_count_y = groupCountY;
cmdbuf->state.sysvals.compute.group_count_z = groupCountZ;
cmdbuf->state.bindpoint[VK_PIPELINE_BIND_POINT_COMPUTE].dirty |=
DZN_CMD_BINDPOINT_DIRTY_SYSVALS;
dzn_cmd_buffer_prepare_dispatch(cmdbuf);
cmdbuf->cmdlist->Dispatch(groupCountX, groupCountY, groupCountZ);
}
VKAPI_ATTR void VKAPI_CALL
dzn_CmdFillBuffer(VkCommandBuffer commandBuffer,
VkBuffer dstBuffer,
VkDeviceSize dstOffset,
VkDeviceSize size,
uint32_t data)
{
VK_FROM_HANDLE(dzn_cmd_buffer, cmdbuf, commandBuffer);
VK_FROM_HANDLE(dzn_buffer, buf, dstBuffer);
if (size == VK_WHOLE_SIZE)
size = buf->size - dstOffset;
size &= ~3ULL;
ID3D12Resource *src_res;
VkResult result =
dzn_cmd_buffer_alloc_internal_buf(cmdbuf, size,
D3D12_HEAP_TYPE_UPLOAD,
D3D12_RESOURCE_STATE_GENERIC_READ,
&src_res);
if (result != VK_SUCCESS)
return;
uint32_t *cpu_ptr;
src_res->Map(0, NULL, (void **)&cpu_ptr);
for (uint32_t i = 0; i < size / 4; i++)
cpu_ptr[i] = data;
src_res->Unmap(0, NULL);
cmdbuf->cmdlist->CopyBufferRegion(buf->res, dstOffset, src_res, 0, size);
}
VKAPI_ATTR void VKAPI_CALL
dzn_CmdUpdateBuffer(VkCommandBuffer commandBuffer,
VkBuffer dstBuffer,
VkDeviceSize dstOffset,
VkDeviceSize size,
const void *data)
{
VK_FROM_HANDLE(dzn_cmd_buffer, cmdbuf, commandBuffer);
VK_FROM_HANDLE(dzn_buffer, buf, dstBuffer);
if (size == VK_WHOLE_SIZE)
size = buf->size - dstOffset;
/*
* The spec says:
* "size is the number of bytes to fill, and must be either a multiple of
* 4, or VK_WHOLE_SIZE to fill the range from offset to the end of the
* buffer. If VK_WHOLE_SIZE is used and the remaining size of the buffer
* is not a multiple of 4, then the nearest smaller multiple is used."
*/
size &= ~3ULL;
ID3D12Resource *src_res;
VkResult result =
dzn_cmd_buffer_alloc_internal_buf(cmdbuf, size,
D3D12_HEAP_TYPE_UPLOAD,
D3D12_RESOURCE_STATE_GENERIC_READ,
&src_res);
if (result != VK_SUCCESS)
return;
void *cpu_ptr;
src_res->Map(0, NULL, &cpu_ptr);
memcpy(cpu_ptr, data, size),
src_res->Unmap(0, NULL);
cmdbuf->cmdlist->CopyBufferRegion(buf->res, dstOffset, src_res, 0, size);
}
VKAPI_ATTR void VKAPI_CALL
dzn_CmdClearAttachments(VkCommandBuffer commandBuffer,
uint32_t attachmentCount,
const VkClearAttachment *pAttachments,
uint32_t rectCount,
const VkClearRect *pRects)
{
VK_FROM_HANDLE(dzn_cmd_buffer, cmdbuf, commandBuffer);
dzn_cmd_buffer_clear_attachments(cmdbuf, attachmentCount, pAttachments, rectCount, pRects);
}
VKAPI_ATTR void VKAPI_CALL
dzn_CmdBeginRenderPass2(VkCommandBuffer commandBuffer,
const VkRenderPassBeginInfo *pRenderPassBeginInfo,
const VkSubpassBeginInfo *pSubpassBeginInfo)
{
VK_FROM_HANDLE(dzn_cmd_buffer, cmdbuf, commandBuffer);
VK_FROM_HANDLE(dzn_render_pass, pass, pRenderPassBeginInfo->renderPass);
VK_FROM_HANDLE(dzn_framebuffer, framebuffer, pRenderPassBeginInfo->framebuffer);
assert(pass->attachment_count == framebuffer->attachment_count);
cmdbuf->state.framebuffer = framebuffer;
cmdbuf->state.render_area = D3D12_RECT {
.left = pRenderPassBeginInfo->renderArea.offset.x,
.top = pRenderPassBeginInfo->renderArea.offset.y,
.right = (LONG)(pRenderPassBeginInfo->renderArea.offset.x + pRenderPassBeginInfo->renderArea.extent.width),
.bottom = (LONG)(pRenderPassBeginInfo->renderArea.offset.y + pRenderPassBeginInfo->renderArea.extent.height),
};
// The render area has an impact on the scissor state.
cmdbuf->state.dirty |= DZN_CMD_DIRTY_SCISSORS;
cmdbuf->state.pass = pass;
cmdbuf->state.subpass = 0;
dzn_cmd_buffer_begin_subpass(cmdbuf);
uint32_t clear_count =
MIN2(pRenderPassBeginInfo->clearValueCount, framebuffer->attachment_count);
for (int i = 0; i < clear_count; ++i) {
VkImageAspectFlags aspectMask = 0;
if (vk_format_is_depth_or_stencil(pass->attachments[i].format)) {
if (pass->attachments[i].clear.depth)
aspectMask |= VK_IMAGE_ASPECT_DEPTH_BIT;
if (pass->attachments[i].clear.stencil)
aspectMask |= VK_IMAGE_ASPECT_STENCIL_BIT;
} else if (pass->attachments[i].clear.color) {
aspectMask |= VK_IMAGE_ASPECT_COLOR_BIT;
}
dzn_cmd_buffer_clear_attachment(cmdbuf, i, &pRenderPassBeginInfo->pClearValues[i],
aspectMask, 0, ~0, 1, &cmdbuf->state.render_area);
}
}
VKAPI_ATTR void VKAPI_CALL
dzn_CmdEndRenderPass2(VkCommandBuffer commandBuffer,
const VkSubpassEndInfo *pSubpassEndInfo)
{
VK_FROM_HANDLE(dzn_cmd_buffer, cmdbuf, commandBuffer);
dzn_cmd_buffer_end_subpass(cmdbuf);
for (uint32_t i = 0; i < cmdbuf->state.pass->attachment_count; i++)
dzn_cmd_buffer_attachment_transition(cmdbuf, &cmdbuf->state.pass->attachments[i]);
cmdbuf->state.framebuffer = NULL;
cmdbuf->state.pass = NULL;
cmdbuf->state.subpass = 0;
}
VKAPI_ATTR void VKAPI_CALL
dzn_CmdNextSubpass2(VkCommandBuffer commandBuffer,
const VkSubpassBeginInfo *pSubpassBeginInfo,
const VkSubpassEndInfo *pSubpassEndInfo)
{
VK_FROM_HANDLE(dzn_cmd_buffer, cmdbuf, commandBuffer);
dzn_cmd_buffer_end_subpass(cmdbuf);
assert(cmdbuf->state.subpass + 1 < cmdbuf->state.pass->subpass_count);
cmdbuf->state.subpass++;
dzn_cmd_buffer_begin_subpass(cmdbuf);
}
VKAPI_ATTR void VKAPI_CALL
dzn_CmdBindPipeline(VkCommandBuffer commandBuffer,
VkPipelineBindPoint pipelineBindPoint,
VkPipeline pipe)
{
VK_FROM_HANDLE(dzn_cmd_buffer, cmdbuf, commandBuffer);
VK_FROM_HANDLE(dzn_pipeline, pipeline, pipe);
cmdbuf->state.bindpoint[pipelineBindPoint].pipeline = pipeline;
cmdbuf->state.bindpoint[pipelineBindPoint].dirty |= DZN_CMD_BINDPOINT_DIRTY_PIPELINE;
if (pipelineBindPoint == VK_PIPELINE_BIND_POINT_GRAPHICS) {
const dzn_graphics_pipeline *gfx = (const dzn_graphics_pipeline *)pipeline;
if (!gfx->vp.dynamic) {
memcpy(cmdbuf->state.viewports, gfx->vp.desc,
gfx->vp.count * sizeof(cmdbuf->state.viewports[0]));
cmdbuf->state.dirty |= DZN_CMD_DIRTY_VIEWPORTS;
}
if (!gfx->scissor.dynamic) {
memcpy(cmdbuf->state.scissors, gfx->scissor.desc,
gfx->scissor.count * sizeof(cmdbuf->state.scissors[0]));
cmdbuf->state.dirty |= DZN_CMD_DIRTY_SCISSORS;
}
if (gfx->zsa.stencil_test.enable && !gfx->zsa.stencil_test.dynamic_ref) {
cmdbuf->state.zsa.stencil_test.front.ref = gfx->zsa.stencil_test.front.ref;
cmdbuf->state.zsa.stencil_test.back.ref = gfx->zsa.stencil_test.back.ref;
cmdbuf->state.dirty |= DZN_CMD_DIRTY_STENCIL_REF;
}
for (uint32_t vb = 0; vb < gfx->vb.count; vb++)
cmdbuf->state.vb.views[vb].StrideInBytes = gfx->vb.strides[vb];
if (gfx->vb.count > 0)
BITSET_SET_RANGE(cmdbuf->state.vb.dirty, 0, gfx->vb.count - 1);
}
}
VKAPI_ATTR void VKAPI_CALL
dzn_CmdBindDescriptorSets(VkCommandBuffer commandBuffer,
VkPipelineBindPoint pipelineBindPoint,
VkPipelineLayout layout,
uint32_t firstSet,
uint32_t descriptorSetCount,
const VkDescriptorSet *pDescriptorSets,
uint32_t dynamicOffsetCount,
const uint32_t *pDynamicOffsets)
{
VK_FROM_HANDLE(dzn_cmd_buffer, cmdbuf, commandBuffer);
VK_FROM_HANDLE(dzn_pipeline_layout, playout, layout);
struct dzn_descriptor_state *desc_state =
&cmdbuf->state.bindpoint[pipelineBindPoint].desc_state;
uint32_t dirty = 0;
for (uint32_t i = 0; i < descriptorSetCount; i++) {
uint32_t idx = firstSet + i;
VK_FROM_HANDLE(dzn_descriptor_set, set, pDescriptorSets[i]);
if (desc_state->sets[idx].set != set) {
desc_state->sets[idx].set = set;
dirty |= DZN_CMD_BINDPOINT_DIRTY_HEAPS;
}
uint32_t dynamic_buffer_count = playout->sets[idx].dynamic_buffer_count;
if (dynamic_buffer_count) {
assert(dynamicOffsetCount >= dynamic_buffer_count);
for (uint32_t j = 0; j < dynamic_buffer_count; j++)
desc_state->sets[idx].dynamic_offsets[j] = pDynamicOffsets[j];
dynamicOffsetCount -= dynamic_buffer_count;
pDynamicOffsets += dynamic_buffer_count;
dirty |= DZN_CMD_BINDPOINT_DIRTY_HEAPS;
}
}
cmdbuf->state.bindpoint[pipelineBindPoint].dirty |= dirty;
}
VKAPI_ATTR void VKAPI_CALL
dzn_CmdSetViewport(VkCommandBuffer commandBuffer,
uint32_t firstViewport,
uint32_t viewportCount,
const VkViewport *pViewports)
{
VK_FROM_HANDLE(dzn_cmd_buffer, cmdbuf, commandBuffer);
STATIC_ASSERT(MAX_VP <= DXIL_SPIRV_MAX_VIEWPORT);
for (uint32_t i = 0; i < viewportCount; i++) {
uint32_t vp = i + firstViewport;
dzn_translate_viewport(&cmdbuf->state.viewports[vp], &pViewports[i]);
if (pViewports[i].minDepth > pViewports[i].maxDepth)
cmdbuf->state.sysvals.gfx.yz_flip_mask |= BITFIELD_BIT(vp + DXIL_SPIRV_Z_FLIP_SHIFT);
else
cmdbuf->state.sysvals.gfx.yz_flip_mask &= ~BITFIELD_BIT(vp + DXIL_SPIRV_Z_FLIP_SHIFT);
if (pViewports[i].height > 0)
cmdbuf->state.sysvals.gfx.yz_flip_mask |= BITFIELD_BIT(vp);
else
cmdbuf->state.sysvals.gfx.yz_flip_mask &= ~BITFIELD_BIT(vp);
}
if (viewportCount) {
cmdbuf->state.dirty |= DZN_CMD_DIRTY_VIEWPORTS;
cmdbuf->state.bindpoint[VK_PIPELINE_BIND_POINT_GRAPHICS].dirty |=
DZN_CMD_BINDPOINT_DIRTY_SYSVALS;
}
}
VKAPI_ATTR void VKAPI_CALL
dzn_CmdSetScissor(VkCommandBuffer commandBuffer,
uint32_t firstScissor,
uint32_t scissorCount,
const VkRect2D *pScissors)
{
VK_FROM_HANDLE(dzn_cmd_buffer, cmdbuf, commandBuffer);
for (uint32_t i = 0; i < scissorCount; i++)
dzn_translate_rect(&cmdbuf->state.scissors[i + firstScissor], &pScissors[i]);
if (scissorCount)
cmdbuf->state.dirty |= DZN_CMD_DIRTY_SCISSORS;
}
VKAPI_ATTR void VKAPI_CALL
dzn_CmdPushConstants(VkCommandBuffer commandBuffer, VkPipelineLayout layout,
VkShaderStageFlags stageFlags, uint32_t offset, uint32_t size,
const void *pValues)
{
VK_FROM_HANDLE(dzn_cmd_buffer, cmdbuf, commandBuffer);
struct dzn_cmd_buffer_push_constant_state *states[2];
uint32_t num_states = 0;
if (stageFlags & VK_SHADER_STAGE_ALL_GRAPHICS)
states[num_states++] = &cmdbuf->state.push_constant.gfx;
if (stageFlags & VK_SHADER_STAGE_COMPUTE_BIT)
states[num_states++] = &cmdbuf->state.push_constant.compute;
for (uint32_t i = 0; i < num_states; i++) {
memcpy(((char *)states[i]->values) + offset, pValues, size);
uint32_t current_offset = states[i]->offset;
uint32_t current_end = states[i]->end;
uint32_t end = offset + size;
if (current_end != 0) {
offset = MIN2(current_offset, offset);
end = MAX2(current_end, end);
}
states[i]->offset = offset;
states[i]->end = end;
}
}
VKAPI_ATTR void VKAPI_CALL
dzn_CmdDraw(VkCommandBuffer commandBuffer,
uint32_t vertexCount,
uint32_t instanceCount,
uint32_t firstVertex,
uint32_t firstInstance)
{
VK_FROM_HANDLE(dzn_cmd_buffer, cmdbuf, commandBuffer);
const dzn_graphics_pipeline *pipeline = (const dzn_graphics_pipeline *)
cmdbuf->state.bindpoint[VK_PIPELINE_BIND_POINT_GRAPHICS].pipeline;
cmdbuf->state.sysvals.gfx.first_vertex = firstVertex;
cmdbuf->state.sysvals.gfx.base_instance = firstInstance;
cmdbuf->state.bindpoint[VK_PIPELINE_BIND_POINT_GRAPHICS].dirty |=
DZN_CMD_BINDPOINT_DIRTY_SYSVALS;
if (pipeline->ia.triangle_fan) {
D3D12_INDEX_BUFFER_VIEW ib_view = cmdbuf->state.ib.view;
VkResult result =
dzn_cmd_buffer_triangle_fan_create_index(cmdbuf, &vertexCount);
if (result != VK_SUCCESS || !vertexCount)
return;
cmdbuf->state.sysvals.gfx.is_indexed_draw = true;
dzn_cmd_buffer_prepare_draw(cmdbuf, true);
cmdbuf->cmdlist->DrawIndexedInstanced(vertexCount, instanceCount, 0,
firstVertex, firstInstance);
/* Restore the IB view if we modified it when lowering triangle fans. */
if (ib_view.SizeInBytes > 0) {
cmdbuf->state.ib.view = ib_view;
cmdbuf->state.dirty |= DZN_CMD_DIRTY_IB;
}
} else {
cmdbuf->state.sysvals.gfx.is_indexed_draw = false;
dzn_cmd_buffer_prepare_draw(cmdbuf, false);
cmdbuf->cmdlist->DrawInstanced(vertexCount, instanceCount,
firstVertex, firstInstance);
}
}
VKAPI_ATTR void VKAPI_CALL
dzn_CmdDrawIndexed(VkCommandBuffer commandBuffer,
uint32_t indexCount,
uint32_t instanceCount,
uint32_t firstIndex,
int32_t vertexOffset,
uint32_t firstInstance)
{
VK_FROM_HANDLE(dzn_cmd_buffer, cmdbuf, commandBuffer);
const dzn_graphics_pipeline *pipeline = (const dzn_graphics_pipeline *)
cmdbuf->state.bindpoint[VK_PIPELINE_BIND_POINT_GRAPHICS].pipeline;
cmdbuf->state.sysvals.gfx.first_vertex = vertexOffset;
cmdbuf->state.sysvals.gfx.base_instance = firstInstance;
cmdbuf->state.sysvals.gfx.is_indexed_draw = true;
cmdbuf->state.bindpoint[VK_PIPELINE_BIND_POINT_GRAPHICS].dirty |=
DZN_CMD_BINDPOINT_DIRTY_SYSVALS;
D3D12_INDEX_BUFFER_VIEW ib_view = cmdbuf->state.ib.view;
if (pipeline->ia.triangle_fan) {
VkResult result =
dzn_cmd_buffer_triangle_fan_rewrite_index(cmdbuf, &indexCount, &firstIndex);
if (result != VK_SUCCESS || !indexCount)
return;
}
dzn_cmd_buffer_prepare_draw(cmdbuf, true);
cmdbuf->cmdlist->DrawIndexedInstanced(indexCount, instanceCount, firstIndex,
vertexOffset, firstInstance);
/* Restore the IB view if we modified it when lowering triangle fans. */
if (pipeline->ia.triangle_fan && ib_view.SizeInBytes) {
cmdbuf->state.ib.view = ib_view;
cmdbuf->state.dirty |= DZN_CMD_DIRTY_IB;
}
}
VKAPI_ATTR void VKAPI_CALL
dzn_CmdDrawIndirect(VkCommandBuffer commandBuffer,
VkBuffer buffer,
VkDeviceSize offset,
uint32_t drawCount,
uint32_t stride)
{
VK_FROM_HANDLE(dzn_cmd_buffer, cmdbuf, commandBuffer);
VK_FROM_HANDLE(dzn_buffer, buf, buffer);
dzn_cmd_buffer_indirect_draw(cmdbuf, buf, offset, drawCount, stride, false);
}
VKAPI_ATTR void VKAPI_CALL
dzn_CmdDrawIndexedIndirect(VkCommandBuffer commandBuffer,
VkBuffer buffer,
VkDeviceSize offset,
uint32_t drawCount,
uint32_t stride)
{
VK_FROM_HANDLE(dzn_cmd_buffer, cmdbuf, commandBuffer);
VK_FROM_HANDLE(dzn_buffer, buf, buffer);
dzn_cmd_buffer_indirect_draw(cmdbuf, buf, offset, drawCount, stride, true);
}
VKAPI_ATTR void VKAPI_CALL
dzn_CmdBindVertexBuffers(VkCommandBuffer commandBuffer,
uint32_t firstBinding,
uint32_t bindingCount,
const VkBuffer *pBuffers,
const VkDeviceSize *pOffsets)
{
VK_FROM_HANDLE(dzn_cmd_buffer, cmdbuf, commandBuffer);
if (!bindingCount)
return;
D3D12_VERTEX_BUFFER_VIEW *vbviews = cmdbuf->state.vb.views;
for (uint32_t i = 0; i < bindingCount; i++) {
VK_FROM_HANDLE(dzn_buffer, buf, pBuffers[i]);
vbviews[firstBinding + i].BufferLocation = buf->res->GetGPUVirtualAddress() + pOffsets[i];
vbviews[firstBinding + i].SizeInBytes = buf->size - pOffsets[i];
}
BITSET_SET_RANGE(cmdbuf->state.vb.dirty, firstBinding,
firstBinding + bindingCount - 1);
}
VKAPI_ATTR void VKAPI_CALL
dzn_CmdBindIndexBuffer(VkCommandBuffer commandBuffer,
VkBuffer buffer,
VkDeviceSize offset,
VkIndexType indexType)
{
VK_FROM_HANDLE(dzn_cmd_buffer, cmdbuf, commandBuffer);
VK_FROM_HANDLE(dzn_buffer, buf, buffer);
cmdbuf->state.ib.view.BufferLocation = buf->res->GetGPUVirtualAddress() + offset;
cmdbuf->state.ib.view.SizeInBytes = buf->size - offset;
switch (indexType) {
case VK_INDEX_TYPE_UINT16:
cmdbuf->state.ib.view.Format = DXGI_FORMAT_R16_UINT;
break;
case VK_INDEX_TYPE_UINT32:
cmdbuf->state.ib.view.Format = DXGI_FORMAT_R32_UINT;
break;
default: unreachable("Invalid index type");
}
cmdbuf->state.dirty |= DZN_CMD_DIRTY_IB;
}
VKAPI_ATTR void VKAPI_CALL
dzn_CmdResetEvent(VkCommandBuffer commandBuffer,
VkEvent event,
VkPipelineStageFlags stageMask)
{
VK_FROM_HANDLE(dzn_cmd_buffer, cmdbuf, commandBuffer);
dzn_device *device = container_of(cmdbuf->vk.base.device, dzn_device, vk);
VK_FROM_HANDLE(dzn_event, evt, event);
if (!_mesa_hash_table_insert(cmdbuf->events.ht, event, (void *)(uintptr_t)DZN_EVENT_STATE_RESET))
cmdbuf->error = vk_error(device, VK_ERROR_OUT_OF_HOST_MEMORY);
}
VKAPI_ATTR void VKAPI_CALL
dzn_CmdSetEvent(VkCommandBuffer commandBuffer,
VkEvent event,
VkPipelineStageFlags stageMask)
{
VK_FROM_HANDLE(dzn_cmd_buffer, cmdbuf, commandBuffer);
dzn_device *device = container_of(cmdbuf->vk.base.device, dzn_device, vk);
VK_FROM_HANDLE(dzn_event, evt, event);
if (!_mesa_hash_table_insert(cmdbuf->events.ht, event, (void *)(uintptr_t)DZN_EVENT_STATE_SET))
cmdbuf->error = vk_error(device, VK_ERROR_OUT_OF_HOST_MEMORY);
}
VKAPI_ATTR void VKAPI_CALL
dzn_CmdWaitEvents(VkCommandBuffer commandBuffer,
uint32_t eventCount,
const VkEvent *pEvents,
VkPipelineStageFlags srcStageMask,
VkPipelineStageFlags dstStageMask,
uint32_t memoryBarrierCount,
const VkMemoryBarrier *pMemoryBarriers,
uint32_t bufferMemoryBarrierCount,
const VkBufferMemoryBarrier *pBufferMemoryBarriers,
uint32_t imageMemoryBarrierCount,
const VkImageMemoryBarrier *pImageMemoryBarriers)
{
VK_FROM_HANDLE(dzn_cmd_buffer, cmdbuf, commandBuffer);
dzn_device *device = container_of(cmdbuf->vk.base.device, dzn_device, vk);
/* Intra-command list wait is handle by this pipeline flush, which is
* overkill, but that's the best we can do with the standard D3D12 barrier
* API.
*
* Inter-command list is taken care of by the serialization done at the
* ExecuteCommandList() level:
* "Calling ExecuteCommandLists twice in succession (from the same thread,
* or different threads) guarantees that the first workload (A) finishes
* before the second workload (B)"
*
* HOST -> DEVICE signaling is ignored and we assume events are always
* signaled when we reach the vkCmdWaitEvents() point.:
* "Command buffers in the submission can include vkCmdWaitEvents commands
* that wait on events that will not be signaled by earlier commands in the
* queue. Such events must be signaled by the application using vkSetEvent,
* and the vkCmdWaitEvents commands that wait upon them must not be inside
* a render pass instance.
* The event must be set before the vkCmdWaitEvents command is executed."
*/
bool flush_pipeline = false;
for (uint32_t i = 0; i < eventCount; i++) {
VK_FROM_HANDLE(dzn_event, event, pEvents[i]);
struct hash_entry *he =
_mesa_hash_table_search(cmdbuf->events.ht, event);
if (he) {
enum dzn_event_state state = (enum dzn_event_state)(uintptr_t)he->data;
assert(state != DZN_EVENT_STATE_RESET);
flush_pipeline = state == DZN_EVENT_STATE_SET;
} else {
if (!_mesa_hash_table_insert(cmdbuf->events.ht, event,
(void *)(uintptr_t)DZN_EVENT_STATE_EXTERNAL_WAIT)) {
cmdbuf->error = vk_error(device, VK_ERROR_OUT_OF_HOST_MEMORY);
return;
}
dzn_event **entry = (dzn_event **)
util_dynarray_grow(&cmdbuf->events.wait, dzn_event *, 1);
if (!entry) {
cmdbuf->error = vk_error(device, VK_ERROR_OUT_OF_HOST_MEMORY);
return;
}
*entry = event;
}
}
if (flush_pipeline) {
D3D12_RESOURCE_BARRIER barrier = {
.Type = D3D12_RESOURCE_BARRIER_TYPE_UAV,
.Flags = D3D12_RESOURCE_BARRIER_FLAG_NONE,
.UAV = { .pResource = NULL },
};
cmdbuf->cmdlist->ResourceBarrier(1, &barrier);
}
}
VKAPI_ATTR void VKAPI_CALL
dzn_CmdBeginQuery(VkCommandBuffer commandBuffer,
VkQueryPool queryPool,
uint32_t query,
VkQueryControlFlags flags)
{
VK_FROM_HANDLE(dzn_cmd_buffer, cmdbuf, commandBuffer);
dzn_device *device = container_of(cmdbuf->vk.base.device, dzn_device, vk);
VK_FROM_HANDLE(dzn_query_pool, qpool, queryPool);
dzn_cmd_buffer_query_pool_state *state =
dzn_cmd_buffer_get_query_pool_state(cmdbuf, qpool);
if (!state)
return;
qpool->queries[query].type = dzn_query_pool_get_query_type(qpool, flags);
dzn_cmd_buffer_dynbitset_clear(cmdbuf, &state->collect, query);
cmdbuf->cmdlist->BeginQuery(qpool->heap, qpool->queries[query].type, query);
}
VKAPI_ATTR void VKAPI_CALL
dzn_CmdEndQuery(VkCommandBuffer commandBuffer,
VkQueryPool queryPool,
uint32_t query)
{
VK_FROM_HANDLE(dzn_cmd_buffer, cmdbuf, commandBuffer);
VK_FROM_HANDLE(dzn_query_pool, qpool, queryPool);
dzn_cmd_buffer_query_pool_state *state =
dzn_cmd_buffer_get_query_pool_state(cmdbuf, qpool);
if (!state)
return;
dzn_cmd_buffer_dynbitset_set(cmdbuf, &state->collect, query);
cmdbuf->cmdlist->EndQuery(qpool->heap, qpool->queries[query].type, query);
}
VKAPI_ATTR void VKAPI_CALL
dzn_CmdWriteTimestamp2(VkCommandBuffer commandBuffer,
VkPipelineStageFlags2 stage,
VkQueryPool queryPool,
uint32_t query)
{
VK_FROM_HANDLE(dzn_cmd_buffer, cmdbuf, commandBuffer);
VK_FROM_HANDLE(dzn_query_pool, qpool, queryPool);
dzn_cmd_buffer_query_pool_state *state =
dzn_cmd_buffer_get_query_pool_state(cmdbuf, qpool);
if (!state)
return;
/* Execution barrier so the timestamp gets written after the pipeline flush. */
D3D12_RESOURCE_BARRIER barrier = {
.Type = D3D12_RESOURCE_BARRIER_TYPE_UAV,
.Flags = D3D12_RESOURCE_BARRIER_FLAG_NONE,
.UAV = { .pResource = NULL },
};
cmdbuf->cmdlist->ResourceBarrier(1, &barrier);
qpool->queries[query].type = D3D12_QUERY_TYPE_TIMESTAMP;
dzn_cmd_buffer_dynbitset_set(cmdbuf, &state->collect, query);
cmdbuf->cmdlist->EndQuery(qpool->heap, qpool->queries[query].type, query);
}
VKAPI_ATTR void VKAPI_CALL
dzn_CmdResetQueryPool(VkCommandBuffer commandBuffer,
VkQueryPool queryPool,
uint32_t firstQuery,
uint32_t queryCount)
{
VK_FROM_HANDLE(dzn_cmd_buffer, cmdbuf, commandBuffer);
dzn_device *device = container_of(cmdbuf->vk.base.device, dzn_device, vk);
VK_FROM_HANDLE(dzn_query_pool, qpool, queryPool);
dzn_cmd_buffer_query_pool_state *state =
dzn_cmd_buffer_get_query_pool_state(cmdbuf, qpool);
if (!state)
return;
uint32_t q_step = DZN_QUERY_REFS_SECTION_SIZE / sizeof(uint64_t);
for (uint32_t q = 0; q < queryCount; q += q_step) {
uint32_t q_count = MIN2(queryCount - q, q_step);
cmdbuf->cmdlist->CopyBufferRegion(qpool->collect_buffer,
dzn_query_pool_get_availability_offset(qpool, firstQuery + q),
device->queries.refs,
DZN_QUERY_REFS_ALL_ZEROS_OFFSET,
q_count * sizeof(uint64_t));
}
q_step = DZN_QUERY_REFS_SECTION_SIZE / qpool->query_size;
for (uint32_t q = 0; q < queryCount; q += q_step) {
cmdbuf->cmdlist->CopyBufferRegion(qpool->collect_buffer,
dzn_query_pool_get_result_offset(qpool, firstQuery + q),
device->queries.refs,
DZN_QUERY_REFS_ALL_ZEROS_OFFSET,
qpool->query_size);
}
dzn_cmd_buffer_dynbitset_set_range(cmdbuf, &state->reset, firstQuery, queryCount);
dzn_cmd_buffer_dynbitset_clear_range(cmdbuf, &state->collect, firstQuery, queryCount);
}
VKAPI_ATTR void VKAPI_CALL
dzn_CmdCopyQueryPoolResults(VkCommandBuffer commandBuffer,
VkQueryPool queryPool,
uint32_t firstQuery,
uint32_t queryCount,
VkBuffer dstBuffer,
VkDeviceSize dstOffset,
VkDeviceSize stride,
VkQueryResultFlags flags)
{
VK_FROM_HANDLE(dzn_cmd_buffer, cmdbuf, commandBuffer);
VK_FROM_HANDLE(dzn_query_pool, qpool, queryPool);
VK_FROM_HANDLE(dzn_buffer, buf, dstBuffer);
dzn_cmd_buffer_query_pool_state *qpstate =
dzn_cmd_buffer_get_query_pool_state(cmdbuf, qpool);
if (!qpstate)
return;
if (flags & VK_QUERY_RESULT_WAIT_BIT) {
for (uint32_t i = 0; i < queryCount; i++) {
if (!dzn_cmd_buffer_dynbitset_test(&qpstate->collect, firstQuery + i) &&
!dzn_cmd_buffer_dynbitset_test(&qpstate->signal, firstQuery + i))
dzn_cmd_buffer_dynbitset_set(cmdbuf, &qpstate->wait, firstQuery + i);
}
}
VkResult result =
dzn_cmd_buffer_collect_queries(cmdbuf, qpool, qpstate, firstQuery, queryCount);
if (result != VK_SUCCESS)
return;
bool raw_copy = (flags & VK_QUERY_RESULT_64_BIT) &&
stride == qpool->query_size &&
!(flags & VK_QUERY_RESULT_WITH_AVAILABILITY_BIT);
#define ALL_STATS \
(VK_QUERY_PIPELINE_STATISTIC_INPUT_ASSEMBLY_VERTICES_BIT | \
VK_QUERY_PIPELINE_STATISTIC_INPUT_ASSEMBLY_PRIMITIVES_BIT | \
VK_QUERY_PIPELINE_STATISTIC_VERTEX_SHADER_INVOCATIONS_BIT | \
VK_QUERY_PIPELINE_STATISTIC_GEOMETRY_SHADER_INVOCATIONS_BIT | \
VK_QUERY_PIPELINE_STATISTIC_GEOMETRY_SHADER_PRIMITIVES_BIT | \
VK_QUERY_PIPELINE_STATISTIC_CLIPPING_INVOCATIONS_BIT | \
VK_QUERY_PIPELINE_STATISTIC_CLIPPING_PRIMITIVES_BIT | \
VK_QUERY_PIPELINE_STATISTIC_FRAGMENT_SHADER_INVOCATIONS_BIT | \
VK_QUERY_PIPELINE_STATISTIC_TESSELLATION_CONTROL_SHADER_PATCHES_BIT | \
VK_QUERY_PIPELINE_STATISTIC_TESSELLATION_EVALUATION_SHADER_INVOCATIONS_BIT | \
VK_QUERY_PIPELINE_STATISTIC_COMPUTE_SHADER_INVOCATIONS_BIT)
if (qpool->heap_type == D3D12_QUERY_HEAP_TYPE_PIPELINE_STATISTICS &&
qpool->pipeline_statistics != ALL_STATS)
raw_copy = false;
#undef ALL_STATS
D3D12_RESOURCE_BARRIER barrier = {
.Type = D3D12_RESOURCE_BARRIER_TYPE_TRANSITION,
.Flags = D3D12_RESOURCE_BARRIER_FLAG_NONE,
.Transition = {
.pResource = qpool->collect_buffer,
.StateBefore = D3D12_RESOURCE_STATE_COPY_DEST,
.StateAfter = D3D12_RESOURCE_STATE_COPY_SOURCE,
},
};
cmdbuf->cmdlist->ResourceBarrier(1, &barrier);
if (raw_copy) {
cmdbuf->cmdlist->CopyBufferRegion(buf->res, dstOffset,
qpool->collect_buffer,
dzn_query_pool_get_result_offset(qpool, firstQuery),
dzn_query_pool_get_result_size(qpool, queryCount));
} else {
uint32_t step = flags & VK_QUERY_RESULT_64_BIT ? sizeof(uint64_t) : sizeof(uint32_t);
for (uint32_t q = 0; q < queryCount; q++) {
uint32_t res_offset = dzn_query_pool_get_result_offset(qpool, firstQuery + q);
uint32_t dst_counter_offset = 0;
if (qpool->heap_type == D3D12_QUERY_HEAP_TYPE_PIPELINE_STATISTICS) {
for (uint32_t c = 0; c < sizeof(D3D12_QUERY_DATA_PIPELINE_STATISTICS) / sizeof(uint64_t); c++) {
if (!(BITFIELD_BIT(c) & qpool->pipeline_statistics))
continue;
cmdbuf->cmdlist->CopyBufferRegion(buf->res, dstOffset + dst_counter_offset,
qpool->collect_buffer,
res_offset + (c * sizeof(uint64_t)),
step);
dst_counter_offset += step;
}
} else {
cmdbuf->cmdlist->CopyBufferRegion(buf->res, dstOffset,
qpool->collect_buffer,
res_offset, step);
dst_counter_offset += step;
}
if (flags & VK_QUERY_RESULT_WITH_AVAILABILITY_BIT) {
cmdbuf->cmdlist->CopyBufferRegion(buf->res, dstOffset + dst_counter_offset,
qpool->collect_buffer,
dzn_query_pool_get_availability_offset(qpool, firstQuery + q),
step);
}
dstOffset += stride;
}
}
DZN_SWAP(barrier.Transition.StateBefore, barrier.Transition.StateAfter);
cmdbuf->cmdlist->ResourceBarrier(1, &barrier);
}
VKAPI_ATTR void VKAPI_CALL
dzn_CmdDispatchIndirect(VkCommandBuffer commandBuffer,
VkBuffer buffer,
VkDeviceSize offset)
{
VK_FROM_HANDLE(dzn_cmd_buffer, cmdbuf, commandBuffer);
dzn_device *device = container_of(cmdbuf->vk.base.device, dzn_device, vk);
VK_FROM_HANDLE(dzn_buffer, buf, buffer);
cmdbuf->state.sysvals.compute.group_count_x = 0;
cmdbuf->state.sysvals.compute.group_count_y = 0;
cmdbuf->state.sysvals.compute.group_count_z = 0;
cmdbuf->state.bindpoint[VK_PIPELINE_BIND_POINT_COMPUTE].dirty |=
DZN_CMD_BINDPOINT_DIRTY_SYSVALS;
dzn_cmd_buffer_prepare_dispatch(cmdbuf);
dzn_compute_pipeline *pipeline = (dzn_compute_pipeline *)
cmdbuf->state.bindpoint[VK_PIPELINE_BIND_POINT_COMPUTE].pipeline;
ID3D12CommandSignature *cmdsig =
dzn_compute_pipeline_get_indirect_cmd_sig(pipeline);
if (!cmdsig) {
cmdbuf->error = vk_error(device, VK_ERROR_OUT_OF_HOST_MEMORY);
return;
}
ID3D12Resource *exec_buf;
VkResult result =
dzn_cmd_buffer_alloc_internal_buf(cmdbuf, sizeof(D3D12_DISPATCH_ARGUMENTS) * 2,
D3D12_HEAP_TYPE_DEFAULT,
D3D12_RESOURCE_STATE_COPY_DEST,
&exec_buf);
if (result != VK_SUCCESS)
return;
cmdbuf->cmdlist->CopyBufferRegion(exec_buf, 0,
buf->res,
offset,
sizeof(D3D12_DISPATCH_ARGUMENTS));
cmdbuf->cmdlist->CopyBufferRegion(exec_buf, sizeof(D3D12_DISPATCH_ARGUMENTS),
buf->res,
offset,
sizeof(D3D12_DISPATCH_ARGUMENTS));
D3D12_RESOURCE_BARRIER barriers[] = {
{
.Type = D3D12_RESOURCE_BARRIER_TYPE_TRANSITION,
.Flags = D3D12_RESOURCE_BARRIER_FLAG_NONE,
/* Transition the exec buffer to indirect arg so it can be
* passed to ExecuteIndirect() as an argument buffer.
*/
.Transition = {
.pResource = exec_buf,
.Subresource = 0,
.StateBefore = D3D12_RESOURCE_STATE_COPY_DEST,
.StateAfter = D3D12_RESOURCE_STATE_INDIRECT_ARGUMENT,
},
},
};
cmdbuf->cmdlist->ResourceBarrier(ARRAY_SIZE(barriers), barriers);
cmdbuf->cmdlist->ExecuteIndirect(cmdsig, 1, exec_buf, 0, NULL, 0);
}
VKAPI_ATTR void VKAPI_CALL
dzn_CmdSetLineWidth(VkCommandBuffer commandBuffer,
float lineWidth)
{
assert(lineWidth == 1.0f);
}
VKAPI_ATTR void VKAPI_CALL
dzn_CmdSetDepthBias(VkCommandBuffer commandBuffer,
float depthBiasConstantFactor,
float depthBiasClamp,
float depthBiasSlopeFactor)
{
dzn_stub();
}
VKAPI_ATTR void VKAPI_CALL
dzn_CmdSetBlendConstants(VkCommandBuffer commandBuffer,
const float blendConstants[4])
{
VK_FROM_HANDLE(dzn_cmd_buffer, cmdbuf, commandBuffer);
cmdbuf->cmdlist->OMSetBlendFactor(blendConstants);
}
VKAPI_ATTR void VKAPI_CALL
dzn_CmdSetDepthBounds(VkCommandBuffer commandBuffer,
float minDepthBounds,
float maxDepthBounds)
{
VK_FROM_HANDLE(dzn_cmd_buffer, cmdbuf, commandBuffer);
cmdbuf->cmdlist->OMSetDepthBounds(minDepthBounds, maxDepthBounds);
}
VKAPI_ATTR void VKAPI_CALL
dzn_CmdSetStencilCompareMask(VkCommandBuffer commandBuffer,
VkStencilFaceFlags faceMask,
uint32_t compareMask)
{
VK_FROM_HANDLE(dzn_cmd_buffer, cmdbuf, commandBuffer);
if (faceMask & VK_STENCIL_FACE_FRONT_BIT)
cmdbuf->state.zsa.stencil_test.front.compare_mask = compareMask;
if (faceMask & VK_STENCIL_FACE_BACK_BIT)
cmdbuf->state.zsa.stencil_test.back.compare_mask = compareMask;
cmdbuf->state.dirty |= DZN_CMD_DIRTY_STENCIL_COMPARE_MASK;
}
VKAPI_ATTR void VKAPI_CALL
dzn_CmdSetStencilWriteMask(VkCommandBuffer commandBuffer,
VkStencilFaceFlags faceMask,
uint32_t writeMask)
{
VK_FROM_HANDLE(dzn_cmd_buffer, cmdbuf, commandBuffer);
if (faceMask & VK_STENCIL_FACE_FRONT_BIT)
cmdbuf->state.zsa.stencil_test.front.write_mask = writeMask;
if (faceMask & VK_STENCIL_FACE_BACK_BIT)
cmdbuf->state.zsa.stencil_test.back.write_mask = writeMask;
cmdbuf->state.dirty |= DZN_CMD_DIRTY_STENCIL_WRITE_MASK;
}
VKAPI_ATTR void VKAPI_CALL
dzn_CmdSetStencilReference(VkCommandBuffer commandBuffer,
VkStencilFaceFlags faceMask,
uint32_t reference)
{
VK_FROM_HANDLE(dzn_cmd_buffer, cmdbuf, commandBuffer);
if (faceMask & VK_STENCIL_FACE_FRONT_BIT)
cmdbuf->state.zsa.stencil_test.front.ref = reference;
if (faceMask & VK_STENCIL_FACE_BACK_BIT)
cmdbuf->state.zsa.stencil_test.back.ref = reference;
cmdbuf->state.dirty |= DZN_CMD_DIRTY_STENCIL_REF;
}