radv: remove unused 'predicated' parameter from some functions

It's always false.

Signed-off-by: Samuel Pitoiset <samuel.pitoiset@gmail.com>
Reviewed-by: Dave Airlie <airlied@redhat.com>
This commit is contained in:
Samuel Pitoiset 2018-06-26 22:35:04 +02:00
parent a6b64d6dde
commit 9c09e7d66e
4 changed files with 15 additions and 25 deletions

View File

@ -4231,7 +4231,6 @@ static void write_event(struct radv_cmd_buffer *cmd_buffer,
* the stage mask. */ * the stage mask. */
si_cs_emit_write_event_eop(cs, si_cs_emit_write_event_eop(cs,
cmd_buffer->state.predicating,
cmd_buffer->device->physical_device->rad_info.chip_class, cmd_buffer->device->physical_device->rad_info.chip_class,
radv_cmd_buffer_uses_mec(cmd_buffer), radv_cmd_buffer_uses_mec(cmd_buffer),
V_028A90_BOTTOM_OF_PIPE_TS, 0, V_028A90_BOTTOM_OF_PIPE_TS, 0,
@ -4283,7 +4282,7 @@ void radv_CmdWaitEvents(VkCommandBuffer commandBuffer,
MAYBE_UNUSED unsigned cdw_max = radeon_check_space(cmd_buffer->device->ws, cs, 7); MAYBE_UNUSED unsigned cdw_max = radeon_check_space(cmd_buffer->device->ws, cs, 7);
si_emit_wait_fence(cs, false, va, 1, 0xffffffff); si_emit_wait_fence(cs, va, 1, 0xffffffff);
assert(cmd_buffer->cs->cdw <= cdw_max); assert(cmd_buffer->cs->cdw <= cdw_max);
} }

View File

@ -1065,7 +1065,6 @@ uint32_t si_get_ia_multi_vgt_param(struct radv_cmd_buffer *cmd_buffer,
bool instanced_draw, bool indirect_draw, bool instanced_draw, bool indirect_draw,
uint32_t draw_vertex_count); uint32_t draw_vertex_count);
void si_cs_emit_write_event_eop(struct radeon_cmdbuf *cs, void si_cs_emit_write_event_eop(struct radeon_cmdbuf *cs,
bool predicated,
enum chip_class chip_class, enum chip_class chip_class,
bool is_mec, bool is_mec,
unsigned event, unsigned event_flags, unsigned event, unsigned event_flags,
@ -1075,7 +1074,6 @@ void si_cs_emit_write_event_eop(struct radeon_cmdbuf *cs,
uint32_t new_fence); uint32_t new_fence);
void si_emit_wait_fence(struct radeon_cmdbuf *cs, void si_emit_wait_fence(struct radeon_cmdbuf *cs,
bool predicated,
uint64_t va, uint32_t ref, uint64_t va, uint32_t ref,
uint32_t mask); uint32_t mask);
void si_cs_emit_cache_flush(struct radeon_cmdbuf *cs, void si_cs_emit_cache_flush(struct radeon_cmdbuf *cs,

View File

@ -992,7 +992,7 @@ void radv_CmdCopyQueryPoolResults(
uint64_t avail_va = va + pool->availability_offset + 4 * query; uint64_t avail_va = va + pool->availability_offset + 4 * query;
/* This waits on the ME. All copies below are done on the ME */ /* This waits on the ME. All copies below are done on the ME */
si_emit_wait_fence(cs, false, avail_va, 1, 0xffffffff); si_emit_wait_fence(cs, avail_va, 1, 0xffffffff);
} }
} }
radv_query_shader(cmd_buffer, cmd_buffer->device->meta_state.query.pipeline_statistics_query_pipeline, radv_query_shader(cmd_buffer, cmd_buffer->device->meta_state.query.pipeline_statistics_query_pipeline,
@ -1015,7 +1015,7 @@ void radv_CmdCopyQueryPoolResults(
uint64_t avail_va = va + pool->availability_offset + 4 * query; uint64_t avail_va = va + pool->availability_offset + 4 * query;
/* This waits on the ME. All copies below are done on the ME */ /* This waits on the ME. All copies below are done on the ME */
si_emit_wait_fence(cs, false, avail_va, 1, 0xffffffff); si_emit_wait_fence(cs, avail_va, 1, 0xffffffff);
} }
if (flags & VK_QUERY_RESULT_WITH_AVAILABILITY_BIT) { if (flags & VK_QUERY_RESULT_WITH_AVAILABILITY_BIT) {
uint64_t avail_va = va + pool->availability_offset + 4 * query; uint64_t avail_va = va + pool->availability_offset + 4 * query;
@ -1176,7 +1176,6 @@ static void emit_end_query(struct radv_cmd_buffer *cmd_buffer,
radeon_emit(cs, va >> 32); radeon_emit(cs, va >> 32);
si_cs_emit_write_event_eop(cs, si_cs_emit_write_event_eop(cs,
false,
cmd_buffer->device->physical_device->rad_info.chip_class, cmd_buffer->device->physical_device->rad_info.chip_class,
radv_cmd_buffer_uses_mec(cmd_buffer), radv_cmd_buffer_uses_mec(cmd_buffer),
V_028A90_BOTTOM_OF_PIPE_TS, 0, V_028A90_BOTTOM_OF_PIPE_TS, 0,
@ -1300,14 +1299,12 @@ void radv_CmdWriteTimestamp(
break; break;
default: default:
si_cs_emit_write_event_eop(cs, si_cs_emit_write_event_eop(cs,
false,
cmd_buffer->device->physical_device->rad_info.chip_class, cmd_buffer->device->physical_device->rad_info.chip_class,
mec, mec,
V_028A90_BOTTOM_OF_PIPE_TS, 0, V_028A90_BOTTOM_OF_PIPE_TS, 0,
EOP_DATA_SEL_TIMESTAMP, EOP_DATA_SEL_TIMESTAMP,
query_va, 0, 0); query_va, 0, 0);
si_cs_emit_write_event_eop(cs, si_cs_emit_write_event_eop(cs,
false,
cmd_buffer->device->physical_device->rad_info.chip_class, cmd_buffer->device->physical_device->rad_info.chip_class,
mec, mec,
V_028A90_BOTTOM_OF_PIPE_TS, 0, V_028A90_BOTTOM_OF_PIPE_TS, 0,

View File

@ -673,7 +673,6 @@ si_get_ia_multi_vgt_param(struct radv_cmd_buffer *cmd_buffer,
} }
void si_cs_emit_write_event_eop(struct radeon_cmdbuf *cs, void si_cs_emit_write_event_eop(struct radeon_cmdbuf *cs,
bool predicated,
enum chip_class chip_class, enum chip_class chip_class,
bool is_mec, bool is_mec,
unsigned event, unsigned event_flags, unsigned event, unsigned event_flags,
@ -694,7 +693,7 @@ void si_cs_emit_write_event_eop(struct radeon_cmdbuf *cs,
sel |= EOP_INT_SEL(EOP_INT_SEL_SEND_DATA_AFTER_WR_CONFIRM); sel |= EOP_INT_SEL(EOP_INT_SEL_SEND_DATA_AFTER_WR_CONFIRM);
if (chip_class >= GFX9 || is_gfx8_mec) { if (chip_class >= GFX9 || is_gfx8_mec) {
radeon_emit(cs, PKT3(PKT3_RELEASE_MEM, is_gfx8_mec ? 5 : 6, predicated)); radeon_emit(cs, PKT3(PKT3_RELEASE_MEM, is_gfx8_mec ? 5 : 6, false));
radeon_emit(cs, op); radeon_emit(cs, op);
radeon_emit(cs, sel); radeon_emit(cs, sel);
radeon_emit(cs, va); /* address lo */ radeon_emit(cs, va); /* address lo */
@ -710,7 +709,7 @@ void si_cs_emit_write_event_eop(struct radeon_cmdbuf *cs,
* (and optional cache flushes executed) before the timestamp * (and optional cache flushes executed) before the timestamp
* is written. * is written.
*/ */
radeon_emit(cs, PKT3(PKT3_EVENT_WRITE_EOP, 4, predicated)); radeon_emit(cs, PKT3(PKT3_EVENT_WRITE_EOP, 4, false));
radeon_emit(cs, op); radeon_emit(cs, op);
radeon_emit(cs, va); radeon_emit(cs, va);
radeon_emit(cs, ((va >> 32) & 0xffff) | sel); radeon_emit(cs, ((va >> 32) & 0xffff) | sel);
@ -718,7 +717,7 @@ void si_cs_emit_write_event_eop(struct radeon_cmdbuf *cs,
radeon_emit(cs, 0); /* unused */ radeon_emit(cs, 0); /* unused */
} }
radeon_emit(cs, PKT3(PKT3_EVENT_WRITE_EOP, 4, predicated)); radeon_emit(cs, PKT3(PKT3_EVENT_WRITE_EOP, 4, false));
radeon_emit(cs, op); radeon_emit(cs, op);
radeon_emit(cs, va); radeon_emit(cs, va);
radeon_emit(cs, ((va >> 32) & 0xffff) | sel); radeon_emit(cs, ((va >> 32) & 0xffff) | sel);
@ -729,11 +728,10 @@ void si_cs_emit_write_event_eop(struct radeon_cmdbuf *cs,
void void
si_emit_wait_fence(struct radeon_cmdbuf *cs, si_emit_wait_fence(struct radeon_cmdbuf *cs,
bool predicated,
uint64_t va, uint32_t ref, uint64_t va, uint32_t ref,
uint32_t mask) uint32_t mask)
{ {
radeon_emit(cs, PKT3(PKT3_WAIT_REG_MEM, 5, predicated)); radeon_emit(cs, PKT3(PKT3_WAIT_REG_MEM, 5, false));
radeon_emit(cs, WAIT_REG_MEM_EQUAL | WAIT_REG_MEM_MEM_SPACE(1)); radeon_emit(cs, WAIT_REG_MEM_EQUAL | WAIT_REG_MEM_MEM_SPACE(1));
radeon_emit(cs, va); radeon_emit(cs, va);
radeon_emit(cs, va >> 32); radeon_emit(cs, va >> 32);
@ -745,13 +743,12 @@ si_emit_wait_fence(struct radeon_cmdbuf *cs,
static void static void
si_emit_acquire_mem(struct radeon_cmdbuf *cs, si_emit_acquire_mem(struct radeon_cmdbuf *cs,
bool is_mec, bool is_mec,
bool predicated,
bool is_gfx9, bool is_gfx9,
unsigned cp_coher_cntl) unsigned cp_coher_cntl)
{ {
if (is_mec || is_gfx9) { if (is_mec || is_gfx9) {
uint32_t hi_val = is_gfx9 ? 0xffffff : 0xff; uint32_t hi_val = is_gfx9 ? 0xffffff : 0xff;
radeon_emit(cs, PKT3(PKT3_ACQUIRE_MEM, 5, predicated) | radeon_emit(cs, PKT3(PKT3_ACQUIRE_MEM, 5, false) |
PKT3_SHADER_TYPE_S(is_mec)); PKT3_SHADER_TYPE_S(is_mec));
radeon_emit(cs, cp_coher_cntl); /* CP_COHER_CNTL */ radeon_emit(cs, cp_coher_cntl); /* CP_COHER_CNTL */
radeon_emit(cs, 0xffffffff); /* CP_COHER_SIZE */ radeon_emit(cs, 0xffffffff); /* CP_COHER_SIZE */
@ -761,7 +758,7 @@ si_emit_acquire_mem(struct radeon_cmdbuf *cs,
radeon_emit(cs, 0x0000000A); /* POLL_INTERVAL */ radeon_emit(cs, 0x0000000A); /* POLL_INTERVAL */
} else { } else {
/* ACQUIRE_MEM is only required on a compute ring. */ /* ACQUIRE_MEM is only required on a compute ring. */
radeon_emit(cs, PKT3(PKT3_SURFACE_SYNC, 3, predicated)); radeon_emit(cs, PKT3(PKT3_SURFACE_SYNC, 3, false));
radeon_emit(cs, cp_coher_cntl); /* CP_COHER_CNTL */ radeon_emit(cs, cp_coher_cntl); /* CP_COHER_CNTL */
radeon_emit(cs, 0xffffffff); /* CP_COHER_SIZE */ radeon_emit(cs, 0xffffffff); /* CP_COHER_SIZE */
radeon_emit(cs, 0); /* CP_COHER_BASE */ radeon_emit(cs, 0); /* CP_COHER_BASE */
@ -801,7 +798,6 @@ si_cs_emit_cache_flush(struct radeon_cmdbuf *cs,
/* Necessary for DCC */ /* Necessary for DCC */
if (chip_class >= VI) { if (chip_class >= VI) {
si_cs_emit_write_event_eop(cs, si_cs_emit_write_event_eop(cs,
false,
chip_class, chip_class,
is_mec, is_mec,
V_028A90_FLUSH_AND_INV_CB_DATA_TS, V_028A90_FLUSH_AND_INV_CB_DATA_TS,
@ -875,10 +871,10 @@ si_cs_emit_cache_flush(struct radeon_cmdbuf *cs,
assert(flush_cnt); assert(flush_cnt);
uint32_t old_fence = (*flush_cnt)++; uint32_t old_fence = (*flush_cnt)++;
si_cs_emit_write_event_eop(cs, false, chip_class, false, cb_db_event, tc_flags, si_cs_emit_write_event_eop(cs, chip_class, false, cb_db_event, tc_flags,
EOP_DATA_SEL_VALUE_32BIT, EOP_DATA_SEL_VALUE_32BIT,
flush_va, old_fence, *flush_cnt); flush_va, old_fence, *flush_cnt);
si_emit_wait_fence(cs, false, flush_va, *flush_cnt, 0xffffffff); si_emit_wait_fence(cs, flush_va, *flush_cnt, 0xffffffff);
} }
/* VGT state sync */ /* VGT state sync */
@ -902,7 +898,7 @@ si_cs_emit_cache_flush(struct radeon_cmdbuf *cs,
if ((flush_bits & RADV_CMD_FLAG_INV_GLOBAL_L2) || if ((flush_bits & RADV_CMD_FLAG_INV_GLOBAL_L2) ||
(chip_class <= CIK && (flush_bits & RADV_CMD_FLAG_WRITEBACK_GLOBAL_L2))) { (chip_class <= CIK && (flush_bits & RADV_CMD_FLAG_WRITEBACK_GLOBAL_L2))) {
si_emit_acquire_mem(cs, is_mec, false, chip_class >= GFX9, si_emit_acquire_mem(cs, is_mec, chip_class >= GFX9,
cp_coher_cntl | cp_coher_cntl |
S_0085F0_TC_ACTION_ENA(1) | S_0085F0_TC_ACTION_ENA(1) |
S_0085F0_TCL1_ACTION_ENA(1) | S_0085F0_TCL1_ACTION_ENA(1) |
@ -916,7 +912,7 @@ si_cs_emit_cache_flush(struct radeon_cmdbuf *cs,
* *
* WB doesn't work without NC. * WB doesn't work without NC.
*/ */
si_emit_acquire_mem(cs, is_mec, false, si_emit_acquire_mem(cs, is_mec,
chip_class >= GFX9, chip_class >= GFX9,
cp_coher_cntl | cp_coher_cntl |
S_0301F0_TC_WB_ACTION_ENA(1) | S_0301F0_TC_WB_ACTION_ENA(1) |
@ -925,7 +921,7 @@ si_cs_emit_cache_flush(struct radeon_cmdbuf *cs,
} }
if (flush_bits & RADV_CMD_FLAG_INV_VMEM_L1) { if (flush_bits & RADV_CMD_FLAG_INV_VMEM_L1) {
si_emit_acquire_mem(cs, is_mec, si_emit_acquire_mem(cs, is_mec,
false, chip_class >= GFX9, chip_class >= GFX9,
cp_coher_cntl | cp_coher_cntl |
S_0085F0_TCL1_ACTION_ENA(1)); S_0085F0_TCL1_ACTION_ENA(1));
cp_coher_cntl = 0; cp_coher_cntl = 0;
@ -936,7 +932,7 @@ si_cs_emit_cache_flush(struct radeon_cmdbuf *cs,
* Therefore, it should be last. Done in PFP. * Therefore, it should be last. Done in PFP.
*/ */
if (cp_coher_cntl) if (cp_coher_cntl)
si_emit_acquire_mem(cs, is_mec, false, chip_class >= GFX9, cp_coher_cntl); si_emit_acquire_mem(cs, is_mec, chip_class >= GFX9, cp_coher_cntl);
if (flush_bits & RADV_CMD_FLAG_START_PIPELINE_STATS) { if (flush_bits & RADV_CMD_FLAG_START_PIPELINE_STATS) {
radeon_emit(cs, PKT3(PKT3_EVENT_WRITE, 0, 0)); radeon_emit(cs, PKT3(PKT3_EVENT_WRITE, 0, 0));