vk: Implement VK_QUERY_RESULT_WITH_AVAILABILITY_BIT

This commit is contained in:
Kristian Høgsberg Kristensen 2016-01-06 21:57:24 -08:00
parent bbf3fc815b
commit a18b5e642c
2 changed files with 93 additions and 43 deletions

View File

@ -45,17 +45,15 @@ VkResult anv_CreateQueryPool(
switch (pCreateInfo->queryType) { switch (pCreateInfo->queryType) {
case VK_QUERY_TYPE_OCCLUSION: case VK_QUERY_TYPE_OCCLUSION:
slot_size = sizeof(struct anv_query_pool_slot); case VK_QUERY_TYPE_TIMESTAMP:
break; break;
case VK_QUERY_TYPE_PIPELINE_STATISTICS: case VK_QUERY_TYPE_PIPELINE_STATISTICS:
return VK_ERROR_INCOMPATIBLE_DRIVER; return VK_ERROR_INCOMPATIBLE_DRIVER;
case VK_QUERY_TYPE_TIMESTAMP:
slot_size = sizeof(uint64_t);
break;
default: default:
assert(!"Invalid query type"); assert(!"Invalid query type");
} }
slot_size = sizeof(struct anv_query_pool_slot);
pool = anv_alloc2(&device->alloc, pAllocator, sizeof(*pool), 8, pool = anv_alloc2(&device->alloc, pAllocator, sizeof(*pool), 8,
VK_SYSTEM_ALLOCATION_SCOPE_OBJECT); VK_SYSTEM_ALLOCATION_SCOPE_OBJECT);
if (pool == NULL) if (pool == NULL)
@ -110,12 +108,6 @@ VkResult anv_GetQueryPoolResults(
uint64_t result; uint64_t result;
int ret; int ret;
if (flags & VK_QUERY_RESULT_WITH_AVAILABILITY_BIT) {
/* Where is the availabilty info supposed to go? */
anv_finishme("VK_QUERY_RESULT_WITH_AVAILABILITY_BIT");
return VK_ERROR_INCOMPATIBLE_DRIVER;
}
assert(pool->type == VK_QUERY_TYPE_OCCLUSION || assert(pool->type == VK_QUERY_TYPE_OCCLUSION ||
pool->type == VK_QUERY_TYPE_TIMESTAMP); pool->type == VK_QUERY_TYPE_TIMESTAMP);
@ -132,11 +124,11 @@ VkResult anv_GetQueryPoolResults(
} }
void *data_end = pData + dataSize; void *data_end = pData + dataSize;
struct anv_query_pool_slot *slot = pool->bo.map;
for (uint32_t i = 0; i < queryCount; i++) { for (uint32_t i = 0; i < queryCount; i++) {
switch (pool->type) { switch (pool->type) {
case VK_QUERY_TYPE_OCCLUSION: { case VK_QUERY_TYPE_OCCLUSION: {
struct anv_query_pool_slot *slot = pool->bo.map;
result = slot[startQuery + i].end - slot[startQuery + i].begin; result = slot[startQuery + i].end - slot[startQuery + i].begin;
break; break;
} }
@ -144,8 +136,7 @@ VkResult anv_GetQueryPoolResults(
/* Not yet implemented */ /* Not yet implemented */
break; break;
case VK_QUERY_TYPE_TIMESTAMP: { case VK_QUERY_TYPE_TIMESTAMP: {
uint64_t *slot = pool->bo.map; result = slot[startQuery + i].begin;
result = slot[startQuery + i];
break; break;
} }
default: default:
@ -153,12 +144,19 @@ VkResult anv_GetQueryPoolResults(
} }
if (flags & VK_QUERY_RESULT_64_BIT) { if (flags & VK_QUERY_RESULT_64_BIT) {
*(uint64_t *)pData = result; uint64_t *dst = pData;
dst[0] = result;
if (flags & VK_QUERY_RESULT_WITH_AVAILABILITY_BIT)
dst[1] = slot[startQuery + i].available;
} else { } else {
uint32_t *dst = pData;
if (result > UINT32_MAX) if (result > UINT32_MAX)
result = UINT32_MAX; result = UINT32_MAX;
*(uint32_t *)pData = result; dst[0] = result;
if (flags & VK_QUERY_RESULT_WITH_AVAILABILITY_BIT)
dst[1] = slot[startQuery + i].available;
} }
pData += stride; pData += stride;
if (pData >= data_end) if (pData >= data_end)
break; break;
@ -173,5 +171,17 @@ void anv_CmdResetQueryPool(
uint32_t startQuery, uint32_t startQuery,
uint32_t queryCount) uint32_t queryCount)
{ {
stub(); ANV_FROM_HANDLE(anv_query_pool, pool, queryPool);
for (uint32_t i = 0; i < queryCount; i++) {
switch (pool->type) {
case VK_QUERY_TYPE_OCCLUSION: {
struct anv_query_pool_slot *slot = pool->bo.map;
slot[startQuery + i].available = 0;
break;
}
default:
assert(!"Invalid query type");
}
}
} }

View File

@ -875,6 +875,17 @@ emit_ps_depth_count(struct anv_batch *batch,
.Address = { bo, offset }); .Address = { bo, offset });
} }
static void
emit_query_availability(struct anv_batch *batch,
struct anv_bo *bo, uint32_t offset)
{
anv_batch_emit(batch, GENX(PIPE_CONTROL),
.DestinationAddressType = DAT_PPGTT,
.PostSyncOperation = WriteImmediateData,
.Address = { bo, offset },
.ImmediateData = 1);
}
void genX(CmdBeginQuery)( void genX(CmdBeginQuery)(
VkCommandBuffer commandBuffer, VkCommandBuffer commandBuffer,
VkQueryPool queryPool, VkQueryPool queryPool,
@ -908,6 +919,9 @@ void genX(CmdEndQuery)(
case VK_QUERY_TYPE_OCCLUSION: case VK_QUERY_TYPE_OCCLUSION:
emit_ps_depth_count(&cmd_buffer->batch, &pool->bo, emit_ps_depth_count(&cmd_buffer->batch, &pool->bo,
entry * sizeof(struct anv_query_pool_slot) + 8); entry * sizeof(struct anv_query_pool_slot) + 8);
emit_query_availability(&cmd_buffer->batch, &pool->bo,
entry * sizeof(struct anv_query_pool_slot) + 16);
break; break;
case VK_QUERY_TYPE_PIPELINE_STATISTICS: case VK_QUERY_TYPE_PIPELINE_STATISTICS:
@ -926,6 +940,7 @@ void genX(CmdWriteTimestamp)(
{ {
ANV_FROM_HANDLE(anv_cmd_buffer, cmd_buffer, commandBuffer); ANV_FROM_HANDLE(anv_cmd_buffer, cmd_buffer, commandBuffer);
ANV_FROM_HANDLE(anv_query_pool, pool, queryPool); ANV_FROM_HANDLE(anv_query_pool, pool, queryPool);
uint32_t offset = entry * sizeof(struct anv_query_pool_slot);
assert(pool->type == VK_QUERY_TYPE_TIMESTAMP); assert(pool->type == VK_QUERY_TYPE_TIMESTAMP);
@ -933,10 +948,10 @@ void genX(CmdWriteTimestamp)(
case VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT: case VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT:
anv_batch_emit(&cmd_buffer->batch, GENX(MI_STORE_REGISTER_MEM), anv_batch_emit(&cmd_buffer->batch, GENX(MI_STORE_REGISTER_MEM),
.RegisterAddress = TIMESTAMP, .RegisterAddress = TIMESTAMP,
.MemoryAddress = { &pool->bo, entry * 8 }); .MemoryAddress = { &pool->bo, offset });
anv_batch_emit(&cmd_buffer->batch, GENX(MI_STORE_REGISTER_MEM), anv_batch_emit(&cmd_buffer->batch, GENX(MI_STORE_REGISTER_MEM),
.RegisterAddress = TIMESTAMP + 4, .RegisterAddress = TIMESTAMP + 4,
.MemoryAddress = { &pool->bo, entry * 8 + 4 }); .MemoryAddress = { &pool->bo, offset + 4 });
break; break;
default: default:
@ -944,9 +959,11 @@ void genX(CmdWriteTimestamp)(
anv_batch_emit(&cmd_buffer->batch, GENX(PIPE_CONTROL), anv_batch_emit(&cmd_buffer->batch, GENX(PIPE_CONTROL),
.DestinationAddressType = DAT_PPGTT, .DestinationAddressType = DAT_PPGTT,
.PostSyncOperation = WriteTimestamp, .PostSyncOperation = WriteTimestamp,
.Address = { &pool->bo, entry * 8 }); .Address = { &pool->bo, offset });
break; break;
} }
emit_query_availability(&cmd_buffer->batch, &pool->bo, entry + 16);
} }
#define alu_opcode(v) __gen_field((v), 20, 31) #define alu_opcode(v) __gen_field((v), 20, 31)
@ -993,6 +1010,20 @@ emit_load_alu_reg_u64(struct anv_batch *batch, uint32_t reg,
.MemoryAddress = { bo, offset + 4 }); .MemoryAddress = { bo, offset + 4 });
} }
static void
store_query_result(struct anv_batch *batch, uint32_t reg,
struct anv_bo *bo, uint32_t offset, VkQueryResultFlags flags)
{
anv_batch_emit(batch, GENX(MI_STORE_REGISTER_MEM),
.RegisterAddress = reg,
.MemoryAddress = { bo, offset });
if (flags & VK_QUERY_RESULT_64_BIT)
anv_batch_emit(batch, GENX(MI_STORE_REGISTER_MEM),
.RegisterAddress = reg + 4,
.MemoryAddress = { bo, offset + 4 });
}
void genX(CmdCopyQueryPoolResults)( void genX(CmdCopyQueryPoolResults)(
VkCommandBuffer commandBuffer, VkCommandBuffer commandBuffer,
VkQueryPool queryPool, VkQueryPool queryPool,
@ -1008,15 +1039,6 @@ void genX(CmdCopyQueryPoolResults)(
ANV_FROM_HANDLE(anv_buffer, buffer, destBuffer); ANV_FROM_HANDLE(anv_buffer, buffer, destBuffer);
uint32_t slot_offset, dst_offset; uint32_t slot_offset, dst_offset;
if (flags & VK_QUERY_RESULT_WITH_AVAILABILITY_BIT) {
/* Where is the availabilty info supposed to go? */
anv_finishme("VK_QUERY_RESULT_WITH_AVAILABILITY_BIT");
return;
}
assert(pool->type == VK_QUERY_TYPE_OCCLUSION);
/* FIXME: If we're not waiting, should we just do this on the CPU? */
if (flags & VK_QUERY_RESULT_WAIT_BIT) if (flags & VK_QUERY_RESULT_WAIT_BIT)
anv_batch_emit(&cmd_buffer->batch, GENX(PIPE_CONTROL), anv_batch_emit(&cmd_buffer->batch, GENX(PIPE_CONTROL),
.CommandStreamerStallEnable = true, .CommandStreamerStallEnable = true,
@ -1026,26 +1048,44 @@ void genX(CmdCopyQueryPoolResults)(
for (uint32_t i = 0; i < queryCount; i++) { for (uint32_t i = 0; i < queryCount; i++) {
slot_offset = (startQuery + i) * sizeof(struct anv_query_pool_slot); slot_offset = (startQuery + i) * sizeof(struct anv_query_pool_slot);
switch (pool->type) {
case VK_QUERY_TYPE_OCCLUSION:
emit_load_alu_reg_u64(&cmd_buffer->batch,
CS_GPR(0), &pool->bo, slot_offset);
emit_load_alu_reg_u64(&cmd_buffer->batch,
CS_GPR(1), &pool->bo, slot_offset + 8);
emit_load_alu_reg_u64(&cmd_buffer->batch, CS_GPR(0), &pool->bo, slot_offset); /* FIXME: We need to clamp the result for 32 bit. */
emit_load_alu_reg_u64(&cmd_buffer->batch, CS_GPR(1), &pool->bo, slot_offset + 8);
/* FIXME: We need to clamp the result for 32 bit. */ uint32_t *dw = anv_batch_emitn(&cmd_buffer->batch, 5, GENX(MI_MATH));
dw[1] = alu(OPCODE_LOAD, OPERAND_SRCA, OPERAND_R1);
dw[2] = alu(OPCODE_LOAD, OPERAND_SRCB, OPERAND_R0);
dw[3] = alu(OPCODE_SUB, 0, 0);
dw[4] = alu(OPCODE_STORE, OPERAND_R2, OPERAND_ACCU);
break;
uint32_t *dw = anv_batch_emitn(&cmd_buffer->batch, 5, GENX(MI_MATH)); case VK_QUERY_TYPE_TIMESTAMP:
dw[1] = alu(OPCODE_LOAD, OPERAND_SRCA, OPERAND_R1); emit_load_alu_reg_u64(&cmd_buffer->batch,
dw[2] = alu(OPCODE_LOAD, OPERAND_SRCB, OPERAND_R0); CS_GPR(2), &pool->bo, slot_offset);
dw[3] = alu(OPCODE_SUB, 0, 0); break;
dw[4] = alu(OPCODE_STORE, OPERAND_R2, OPERAND_ACCU);
anv_batch_emit(&cmd_buffer->batch, GENX(MI_STORE_REGISTER_MEM), default:
.RegisterAddress = CS_GPR(2), unreachable("unhandled query type");
.MemoryAddress = { buffer->bo, dst_offset }); }
if (flags & VK_QUERY_RESULT_64_BIT) store_query_result(&cmd_buffer->batch,
anv_batch_emit(&cmd_buffer->batch, GENX(MI_STORE_REGISTER_MEM), CS_GPR(2), buffer->bo, dst_offset, flags);
.RegisterAddress = CS_GPR(2) + 4,
.MemoryAddress = { buffer->bo, dst_offset + 4 }); if (flags & VK_QUERY_RESULT_WITH_AVAILABILITY_BIT) {
emit_load_alu_reg_u64(&cmd_buffer->batch, CS_GPR(0),
&pool->bo, slot_offset + 16);
if (flags & VK_QUERY_RESULT_64_BIT)
store_query_result(&cmd_buffer->batch,
CS_GPR(0), buffer->bo, dst_offset + 8, flags);
else
store_query_result(&cmd_buffer->batch,
CS_GPR(0), buffer->bo, dst_offset + 4, flags);
}
dst_offset += destStride; dst_offset += destStride;
} }