radeonsi: rename si_gfx_* functions to si_cp_*
and write_event_eop -> release_mem
This commit is contained in:
parent
6e1cf6532d
commit
0d05581578
|
@ -146,6 +146,7 @@
|
||||||
#define WAIT_REG_MEM_EQUAL 3
|
#define WAIT_REG_MEM_EQUAL 3
|
||||||
#define WAIT_REG_MEM_NOT_EQUAL 4
|
#define WAIT_REG_MEM_NOT_EQUAL 4
|
||||||
#define WAIT_REG_MEM_MEM_SPACE(x) (((unsigned)(x) & 0x3) << 4)
|
#define WAIT_REG_MEM_MEM_SPACE(x) (((unsigned)(x) & 0x3) << 4)
|
||||||
|
#define WAIT_REG_MEM_PFP (1 << 8)
|
||||||
#define PKT3_MEM_WRITE 0x3D /* not on CIK */
|
#define PKT3_MEM_WRITE 0x3D /* not on CIK */
|
||||||
#define PKT3_INDIRECT_BUFFER_CIK 0x3F /* new on CIK */
|
#define PKT3_INDIRECT_BUFFER_CIK 0x3F /* new on CIK */
|
||||||
#define R_3F0_IB_BASE_LO 0x3F0
|
#define R_3F0_IB_BASE_LO 0x3F0
|
||||||
|
|
|
@ -66,11 +66,11 @@ struct si_multi_fence {
|
||||||
* \param old_value Previous fence value (for a bug workaround)
|
* \param old_value Previous fence value (for a bug workaround)
|
||||||
* \param new_value Fence value to write for this event.
|
* \param new_value Fence value to write for this event.
|
||||||
*/
|
*/
|
||||||
void si_gfx_write_event_eop(struct si_context *ctx,
|
void si_cp_release_mem(struct si_context *ctx,
|
||||||
unsigned event, unsigned event_flags,
|
unsigned event, unsigned event_flags,
|
||||||
unsigned dst_sel, unsigned int_sel, unsigned data_sel,
|
unsigned dst_sel, unsigned int_sel, unsigned data_sel,
|
||||||
struct r600_resource *buf, uint64_t va,
|
struct r600_resource *buf, uint64_t va,
|
||||||
uint32_t new_fence, unsigned query_type)
|
uint32_t new_fence, unsigned query_type)
|
||||||
{
|
{
|
||||||
struct radeon_cmdbuf *cs = ctx->gfx_cs;
|
struct radeon_cmdbuf *cs = ctx->gfx_cs;
|
||||||
unsigned op = EVENT_TYPE(event) |
|
unsigned op = EVENT_TYPE(event) |
|
||||||
|
@ -149,7 +149,7 @@ void si_gfx_write_event_eop(struct si_context *ctx,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
unsigned si_gfx_write_fence_dwords(struct si_screen *screen)
|
unsigned si_cp_write_fence_dwords(struct si_screen *screen)
|
||||||
{
|
{
|
||||||
unsigned dwords = 6;
|
unsigned dwords = 6;
|
||||||
|
|
||||||
|
@ -160,13 +160,13 @@ unsigned si_gfx_write_fence_dwords(struct si_screen *screen)
|
||||||
return dwords;
|
return dwords;
|
||||||
}
|
}
|
||||||
|
|
||||||
void si_gfx_wait_fence(struct si_context *ctx,
|
void si_cp_wait_mem(struct si_context *ctx,
|
||||||
uint64_t va, uint32_t ref, uint32_t mask)
|
uint64_t va, uint32_t ref, uint32_t mask, unsigned flags)
|
||||||
{
|
{
|
||||||
struct radeon_cmdbuf *cs = ctx->gfx_cs;
|
struct radeon_cmdbuf *cs = ctx->gfx_cs;
|
||||||
|
|
||||||
radeon_emit(cs, PKT3(PKT3_WAIT_REG_MEM, 5, 0));
|
radeon_emit(cs, PKT3(PKT3_WAIT_REG_MEM, 5, 0));
|
||||||
radeon_emit(cs, WAIT_REG_MEM_EQUAL | WAIT_REG_MEM_MEM_SPACE(1));
|
radeon_emit(cs, WAIT_REG_MEM_EQUAL | WAIT_REG_MEM_MEM_SPACE(1) | flags);
|
||||||
radeon_emit(cs, va);
|
radeon_emit(cs, va);
|
||||||
radeon_emit(cs, va >> 32);
|
radeon_emit(cs, va >> 32);
|
||||||
radeon_emit(cs, ref); /* reference value */
|
radeon_emit(cs, ref); /* reference value */
|
||||||
|
@ -275,13 +275,13 @@ static void si_fine_fence_set(struct si_context *ctx,
|
||||||
radeon_emit(cs, fence_va >> 32);
|
radeon_emit(cs, fence_va >> 32);
|
||||||
radeon_emit(cs, 0x80000000);
|
radeon_emit(cs, 0x80000000);
|
||||||
} else if (flags & PIPE_FLUSH_BOTTOM_OF_PIPE) {
|
} else if (flags & PIPE_FLUSH_BOTTOM_OF_PIPE) {
|
||||||
si_gfx_write_event_eop(ctx,
|
si_cp_release_mem(ctx,
|
||||||
V_028A90_BOTTOM_OF_PIPE_TS, 0,
|
V_028A90_BOTTOM_OF_PIPE_TS, 0,
|
||||||
EOP_DST_SEL_MEM,
|
EOP_DST_SEL_MEM,
|
||||||
EOP_INT_SEL_SEND_DATA_AFTER_WR_CONFIRM,
|
EOP_INT_SEL_SEND_DATA_AFTER_WR_CONFIRM,
|
||||||
EOP_DATA_SEL_VALUE_32BIT,
|
EOP_DATA_SEL_VALUE_32BIT,
|
||||||
NULL, fence_va, 0x80000000,
|
NULL, fence_va, 0x80000000,
|
||||||
PIPE_QUERY_GPU_FINISHED);
|
PIPE_QUERY_GPU_FINISHED);
|
||||||
} else {
|
} else {
|
||||||
assert(false);
|
assert(false);
|
||||||
}
|
}
|
||||||
|
|
|
@ -580,12 +580,12 @@ static void si_pc_emit_stop(struct si_context *sctx,
|
||||||
{
|
{
|
||||||
struct radeon_cmdbuf *cs = sctx->gfx_cs;
|
struct radeon_cmdbuf *cs = sctx->gfx_cs;
|
||||||
|
|
||||||
si_gfx_write_event_eop(sctx, V_028A90_BOTTOM_OF_PIPE_TS, 0,
|
si_cp_release_mem(sctx, V_028A90_BOTTOM_OF_PIPE_TS, 0,
|
||||||
EOP_DST_SEL_MEM,
|
EOP_DST_SEL_MEM,
|
||||||
EOP_INT_SEL_SEND_DATA_AFTER_WR_CONFIRM,
|
EOP_INT_SEL_SEND_DATA_AFTER_WR_CONFIRM,
|
||||||
EOP_DATA_SEL_VALUE_32BIT,
|
EOP_DATA_SEL_VALUE_32BIT,
|
||||||
buffer, va, 0, SI_NOT_QUERY);
|
buffer, va, 0, SI_NOT_QUERY);
|
||||||
si_gfx_wait_fence(sctx, va, 0, 0xffffffff);
|
si_cp_wait_mem(sctx, va, 0, 0xffffffff, 0);
|
||||||
|
|
||||||
radeon_emit(cs, PKT3(PKT3_EVENT_WRITE, 0, 0));
|
radeon_emit(cs, PKT3(PKT3_EVENT_WRITE, 0, 0));
|
||||||
radeon_emit(cs, EVENT_TYPE(V_028A90_PERFCOUNTER_SAMPLE) | EVENT_INDEX(0));
|
radeon_emit(cs, EVENT_TYPE(V_028A90_PERFCOUNTER_SAMPLE) | EVENT_INDEX(0));
|
||||||
|
@ -684,7 +684,7 @@ void si_init_perfcounters(struct si_screen *screen)
|
||||||
if (!pc)
|
if (!pc)
|
||||||
return;
|
return;
|
||||||
|
|
||||||
pc->num_stop_cs_dwords = 14 + si_gfx_write_fence_dwords(screen);
|
pc->num_stop_cs_dwords = 14 + si_cp_write_fence_dwords(screen);
|
||||||
pc->num_instance_cs_dwords = 3;
|
pc->num_instance_cs_dwords = 3;
|
||||||
|
|
||||||
pc->num_shader_types = ARRAY_SIZE(si_pc_shader_type_bits);
|
pc->num_shader_types = ARRAY_SIZE(si_pc_shader_type_bits);
|
||||||
|
|
|
@ -1170,14 +1170,14 @@ void si_screen_clear_buffer(struct si_screen *sscreen, struct pipe_resource *dst
|
||||||
uint64_t offset, uint64_t size, unsigned value);
|
uint64_t offset, uint64_t size, unsigned value);
|
||||||
|
|
||||||
/* si_fence.c */
|
/* si_fence.c */
|
||||||
void si_gfx_write_event_eop(struct si_context *ctx,
|
void si_cp_release_mem(struct si_context *ctx,
|
||||||
unsigned event, unsigned event_flags,
|
unsigned event, unsigned event_flags,
|
||||||
unsigned dst_sel, unsigned int_sel, unsigned data_sel,
|
unsigned dst_sel, unsigned int_sel, unsigned data_sel,
|
||||||
struct r600_resource *buf, uint64_t va,
|
struct r600_resource *buf, uint64_t va,
|
||||||
uint32_t new_fence, unsigned query_type);
|
uint32_t new_fence, unsigned query_type);
|
||||||
unsigned si_gfx_write_fence_dwords(struct si_screen *screen);
|
unsigned si_cp_write_fence_dwords(struct si_screen *screen);
|
||||||
void si_gfx_wait_fence(struct si_context *ctx,
|
void si_cp_wait_mem(struct si_context *ctx,
|
||||||
uint64_t va, uint32_t ref, uint32_t mask);
|
uint64_t va, uint32_t ref, uint32_t mask, unsigned flags);
|
||||||
void si_init_fence_functions(struct si_context *ctx);
|
void si_init_fence_functions(struct si_context *ctx);
|
||||||
void si_init_screen_fence_functions(struct si_screen *screen);
|
void si_init_screen_fence_functions(struct si_screen *screen);
|
||||||
struct pipe_fence_handle *si_create_fence(struct pipe_context *ctx,
|
struct pipe_fence_handle *si_create_fence(struct pipe_context *ctx,
|
||||||
|
|
|
@ -665,7 +665,7 @@ static struct pipe_query *si_query_hw_create(struct si_screen *sscreen,
|
||||||
case PIPE_QUERY_OCCLUSION_PREDICATE_CONSERVATIVE:
|
case PIPE_QUERY_OCCLUSION_PREDICATE_CONSERVATIVE:
|
||||||
query->result_size = 16 * sscreen->info.num_render_backends;
|
query->result_size = 16 * sscreen->info.num_render_backends;
|
||||||
query->result_size += 16; /* for the fence + alignment */
|
query->result_size += 16; /* for the fence + alignment */
|
||||||
query->num_cs_dw_end = 6 + si_gfx_write_fence_dwords(sscreen);
|
query->num_cs_dw_end = 6 + si_cp_write_fence_dwords(sscreen);
|
||||||
break;
|
break;
|
||||||
case SI_QUERY_TIME_ELAPSED_SDMA:
|
case SI_QUERY_TIME_ELAPSED_SDMA:
|
||||||
/* GET_GLOBAL_TIMESTAMP only works if the offset is a multiple of 32. */
|
/* GET_GLOBAL_TIMESTAMP only works if the offset is a multiple of 32. */
|
||||||
|
@ -674,11 +674,11 @@ static struct pipe_query *si_query_hw_create(struct si_screen *sscreen,
|
||||||
break;
|
break;
|
||||||
case PIPE_QUERY_TIME_ELAPSED:
|
case PIPE_QUERY_TIME_ELAPSED:
|
||||||
query->result_size = 24;
|
query->result_size = 24;
|
||||||
query->num_cs_dw_end = 8 + si_gfx_write_fence_dwords(sscreen);
|
query->num_cs_dw_end = 8 + si_cp_write_fence_dwords(sscreen);
|
||||||
break;
|
break;
|
||||||
case PIPE_QUERY_TIMESTAMP:
|
case PIPE_QUERY_TIMESTAMP:
|
||||||
query->result_size = 16;
|
query->result_size = 16;
|
||||||
query->num_cs_dw_end = 8 + si_gfx_write_fence_dwords(sscreen);
|
query->num_cs_dw_end = 8 + si_cp_write_fence_dwords(sscreen);
|
||||||
query->flags = SI_QUERY_HW_FLAG_NO_START;
|
query->flags = SI_QUERY_HW_FLAG_NO_START;
|
||||||
break;
|
break;
|
||||||
case PIPE_QUERY_PRIMITIVES_EMITTED:
|
case PIPE_QUERY_PRIMITIVES_EMITTED:
|
||||||
|
@ -699,7 +699,7 @@ static struct pipe_query *si_query_hw_create(struct si_screen *sscreen,
|
||||||
/* 11 values on GCN. */
|
/* 11 values on GCN. */
|
||||||
query->result_size = 11 * 16;
|
query->result_size = 11 * 16;
|
||||||
query->result_size += 8; /* for the fence + alignment */
|
query->result_size += 8; /* for the fence + alignment */
|
||||||
query->num_cs_dw_end = 6 + si_gfx_write_fence_dwords(sscreen);
|
query->num_cs_dw_end = 6 + si_cp_write_fence_dwords(sscreen);
|
||||||
break;
|
break;
|
||||||
default:
|
default:
|
||||||
assert(0);
|
assert(0);
|
||||||
|
@ -890,11 +890,11 @@ static void si_query_hw_do_emit_stop(struct si_context *sctx,
|
||||||
va += 8;
|
va += 8;
|
||||||
/* fall through */
|
/* fall through */
|
||||||
case PIPE_QUERY_TIMESTAMP:
|
case PIPE_QUERY_TIMESTAMP:
|
||||||
si_gfx_write_event_eop(sctx, V_028A90_BOTTOM_OF_PIPE_TS,
|
si_cp_release_mem(sctx, V_028A90_BOTTOM_OF_PIPE_TS,
|
||||||
0, EOP_DST_SEL_MEM,
|
0, EOP_DST_SEL_MEM,
|
||||||
EOP_INT_SEL_SEND_DATA_AFTER_WR_CONFIRM,
|
EOP_INT_SEL_SEND_DATA_AFTER_WR_CONFIRM,
|
||||||
EOP_DATA_SEL_TIMESTAMP, NULL, va,
|
EOP_DATA_SEL_TIMESTAMP, NULL, va,
|
||||||
0, query->b.type);
|
0, query->b.type);
|
||||||
fence_va = va + 8;
|
fence_va = va + 8;
|
||||||
break;
|
break;
|
||||||
case PIPE_QUERY_PIPELINE_STATISTICS: {
|
case PIPE_QUERY_PIPELINE_STATISTICS: {
|
||||||
|
@ -916,12 +916,12 @@ static void si_query_hw_do_emit_stop(struct si_context *sctx,
|
||||||
RADEON_PRIO_QUERY);
|
RADEON_PRIO_QUERY);
|
||||||
|
|
||||||
if (fence_va) {
|
if (fence_va) {
|
||||||
si_gfx_write_event_eop(sctx, V_028A90_BOTTOM_OF_PIPE_TS, 0,
|
si_cp_release_mem(sctx, V_028A90_BOTTOM_OF_PIPE_TS, 0,
|
||||||
EOP_DST_SEL_MEM,
|
EOP_DST_SEL_MEM,
|
||||||
EOP_INT_SEL_SEND_DATA_AFTER_WR_CONFIRM,
|
EOP_INT_SEL_SEND_DATA_AFTER_WR_CONFIRM,
|
||||||
EOP_DATA_SEL_VALUE_32BIT,
|
EOP_DATA_SEL_VALUE_32BIT,
|
||||||
query->buffer.buf, fence_va, 0x80000000,
|
query->buffer.buf, fence_va, 0x80000000,
|
||||||
query->b.type);
|
query->b.type);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1580,7 +1580,7 @@ static void si_query_hw_get_result_resource(struct si_context *sctx,
|
||||||
va = qbuf->buf->gpu_address + qbuf->results_end - query->result_size;
|
va = qbuf->buf->gpu_address + qbuf->results_end - query->result_size;
|
||||||
va += params.fence_offset;
|
va += params.fence_offset;
|
||||||
|
|
||||||
si_gfx_wait_fence(sctx, va, 0x80000000, 0x80000000);
|
si_cp_wait_mem(sctx, va, 0x80000000, 0x80000000, 0);
|
||||||
}
|
}
|
||||||
|
|
||||||
sctx->b.launch_grid(&sctx->b, &grid);
|
sctx->b.launch_grid(&sctx->b, &grid);
|
||||||
|
|
|
@ -918,11 +918,11 @@ void si_emit_cache_flush(struct si_context *sctx)
|
||||||
|
|
||||||
/* Necessary for DCC */
|
/* Necessary for DCC */
|
||||||
if (sctx->chip_class == VI)
|
if (sctx->chip_class == VI)
|
||||||
si_gfx_write_event_eop(sctx,
|
si_cp_release_mem(sctx,
|
||||||
V_028A90_FLUSH_AND_INV_CB_DATA_TS,
|
V_028A90_FLUSH_AND_INV_CB_DATA_TS,
|
||||||
0, EOP_DST_SEL_MEM, EOP_INT_SEL_NONE,
|
0, EOP_DST_SEL_MEM, EOP_INT_SEL_NONE,
|
||||||
EOP_DATA_SEL_DISCARD, NULL,
|
EOP_DATA_SEL_DISCARD, NULL,
|
||||||
0, 0, SI_NOT_QUERY);
|
0, 0, SI_NOT_QUERY);
|
||||||
}
|
}
|
||||||
if (flags & SI_CONTEXT_FLUSH_AND_INV_DB)
|
if (flags & SI_CONTEXT_FLUSH_AND_INV_DB)
|
||||||
cp_coher_cntl |= S_0085F0_DB_ACTION_ENA(1) |
|
cp_coher_cntl |= S_0085F0_DB_ACTION_ENA(1) |
|
||||||
|
@ -1035,13 +1035,13 @@ void si_emit_cache_flush(struct si_context *sctx)
|
||||||
va = sctx->wait_mem_scratch->gpu_address;
|
va = sctx->wait_mem_scratch->gpu_address;
|
||||||
sctx->wait_mem_number++;
|
sctx->wait_mem_number++;
|
||||||
|
|
||||||
si_gfx_write_event_eop(sctx, cb_db_event, tc_flags,
|
si_cp_release_mem(sctx, cb_db_event, tc_flags,
|
||||||
EOP_DST_SEL_MEM,
|
EOP_DST_SEL_MEM,
|
||||||
EOP_INT_SEL_SEND_DATA_AFTER_WR_CONFIRM,
|
EOP_INT_SEL_SEND_DATA_AFTER_WR_CONFIRM,
|
||||||
EOP_DATA_SEL_VALUE_32BIT,
|
EOP_DATA_SEL_VALUE_32BIT,
|
||||||
sctx->wait_mem_scratch, va,
|
sctx->wait_mem_scratch, va,
|
||||||
sctx->wait_mem_number, SI_NOT_QUERY);
|
sctx->wait_mem_number, SI_NOT_QUERY);
|
||||||
si_gfx_wait_fence(sctx, va, sctx->wait_mem_number, 0xffffffff);
|
si_cp_wait_mem(sctx, va, sctx->wait_mem_number, 0xffffffff, 0);
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Make sure ME is idle (it executes most packets) before continuing.
|
/* Make sure ME is idle (it executes most packets) before continuing.
|
||||||
|
|
Loading…
Reference in New Issue