mirror of https://gitlab.freedesktop.org/mesa/mesa
1465 lines
59 KiB
C
1465 lines
59 KiB
C
/*
|
|
* Copyright 2012 Advanced Micro Devices, Inc.
|
|
*
|
|
* SPDX-License-Identifier: MIT
|
|
*/
|
|
|
|
#include "ac_shader_util.h"
|
|
#include "ac_gpu_info.h"
|
|
|
|
#include "sid.h"
|
|
#include "util/u_math.h"
|
|
|
|
#include <assert.h>
|
|
#include <stdlib.h>
|
|
#include <string.h>
|
|
|
|
/* Set NIR options shared by ACO, LLVM, RADV, and radeonsi. */
|
|
void ac_set_nir_options(struct radeon_info *info, bool use_llvm,
|
|
nir_shader_compiler_options *options)
|
|
{
|
|
/* |---------------------------------- Performance & Availability --------------------------------|
|
|
* |MAD/MAC/MADAK/MADMK|MAD_LEGACY|MAC_LEGACY| FMA |FMAC/FMAAK/FMAMK|FMA_LEGACY|PK_FMA_F16,|Best choice
|
|
* Arch | F32,F16,F64 | F32,F16 | F32,F16 |F32,F16,F64 | F32,F16 | F32 |PK_FMAC_F16|F16,F32,F64
|
|
* ------------------------------------------------------------------------------------------------------------------
|
|
* gfx6,7 | 1 , - , - | 1 , - | 1 , - |1/4, - ,1/16| - , - | - | - , - | - ,MAD,FMA
|
|
* gfx8 | 1 , 1 , - | 1 , - | - , - |1/4, 1 ,1/16| - , - | - | - , - |MAD,MAD,FMA
|
|
* gfx9 | 1 ,1|0, - | 1 , - | - , - | 1 , 1 ,1/16| 0|1, - | - | 2 , - |FMA,MAD,FMA
|
|
* gfx10 | 1 , - , - | 1 , - | 1 , - | 1 , 1 ,1/16| 1 , 1 | - | 2 , 2 |FMA,MAD,FMA
|
|
* gfx10.3| - , - , - | - , - | - , - | 1 , 1 ,1/16| 1 , 1 | 1 | 2 , 2 | all FMA
|
|
* gfx11 | - , - , - | - , - | - , - | 2 , 2 ,1/16| 2 , 2 | 2 | 2 , 2 | all FMA
|
|
*
|
|
* Tahiti, Hawaii, Carrizo, Vega20: FMA_F32 is full rate, FMA_F64 is 1/4
|
|
* gfx9 supports MAD_F16 only on Vega10, Raven, Raven2, Renoir.
|
|
* gfx9 supports FMAC_F32 only on Vega20, but doesn't support FMAAK and FMAMK.
|
|
*
|
|
* gfx8 prefers MAD for F16 because of MAC/MADAK/MADMK.
|
|
* gfx9 and newer prefer FMA for F16 because of the packed instruction.
|
|
* gfx10 and older prefer MAD for F32 because of the legacy instruction.
|
|
*/
|
|
|
|
memset(options, 0, sizeof(*options));
|
|
options->vertex_id_zero_based = true;
|
|
options->lower_scmp = true;
|
|
options->lower_flrp16 = true;
|
|
options->lower_flrp32 = true;
|
|
options->lower_flrp64 = true;
|
|
options->lower_device_index_to_zero = true;
|
|
options->lower_fdiv = true;
|
|
options->lower_fmod = true;
|
|
options->lower_ineg = true;
|
|
options->lower_bitfield_insert = true;
|
|
options->lower_bitfield_extract = true;
|
|
options->lower_pack_snorm_4x8 = true;
|
|
options->lower_pack_unorm_4x8 = true;
|
|
options->lower_pack_half_2x16 = true;
|
|
options->lower_pack_64_2x32 = true;
|
|
options->lower_pack_64_4x16 = true;
|
|
options->lower_pack_32_2x16 = true;
|
|
options->lower_unpack_snorm_2x16 = true;
|
|
options->lower_unpack_snorm_4x8 = true;
|
|
options->lower_unpack_unorm_2x16 = true;
|
|
options->lower_unpack_unorm_4x8 = true;
|
|
options->lower_unpack_half_2x16 = true;
|
|
options->lower_fpow = true;
|
|
options->lower_mul_2x32_64 = true;
|
|
options->lower_iadd_sat = info->gfx_level <= GFX8;
|
|
options->lower_hadd = true;
|
|
options->lower_mul_32x16 = true;
|
|
options->has_bfe = true;
|
|
options->has_bfm = true;
|
|
options->has_bitfield_select = true;
|
|
options->has_fsub = true;
|
|
options->has_isub = true;
|
|
options->has_sdot_4x8 = info->has_accelerated_dot_product;
|
|
options->has_sudot_4x8 = info->has_accelerated_dot_product && info->gfx_level >= GFX11;
|
|
options->has_udot_4x8 = info->has_accelerated_dot_product;
|
|
options->has_sdot_4x8_sat = info->has_accelerated_dot_product;
|
|
options->has_sudot_4x8_sat = info->has_accelerated_dot_product && info->gfx_level >= GFX11;
|
|
options->has_udot_4x8_sat = info->has_accelerated_dot_product;
|
|
options->has_dot_2x16 = info->has_accelerated_dot_product && info->gfx_level < GFX11;
|
|
options->has_find_msb_rev = true;
|
|
options->has_pack_half_2x16_rtz = true;
|
|
options->has_bit_test = !use_llvm;
|
|
options->has_fmulz = true;
|
|
options->has_msad = true;
|
|
options->has_shfr32 = true;
|
|
options->use_interpolated_input_intrinsics = true;
|
|
options->lower_int64_options = nir_lower_imul64 | nir_lower_imul_high64 | nir_lower_imul_2x32_64 | nir_lower_divmod64 |
|
|
nir_lower_minmax64 | nir_lower_iabs64 | nir_lower_iadd_sat64 | nir_lower_conv64;
|
|
options->divergence_analysis_options = nir_divergence_view_index_uniform;
|
|
options->optimize_quad_vote_to_reduce = true;
|
|
options->lower_fisnormal = true;
|
|
options->support_16bit_alu = info->gfx_level >= GFX8;
|
|
options->vectorize_vec2_16bit = info->has_packed_math_16bit;
|
|
}
|
|
|
|
bool
|
|
ac_nir_mem_vectorize_callback(unsigned align_mul, unsigned align_offset, unsigned bit_size,
|
|
unsigned num_components, nir_intrinsic_instr *low,
|
|
nir_intrinsic_instr *high, void *data)
|
|
{
|
|
if (num_components > 4)
|
|
return false;
|
|
|
|
bool is_scratch = false;
|
|
switch (low->intrinsic) {
|
|
case nir_intrinsic_load_stack:
|
|
case nir_intrinsic_load_scratch:
|
|
case nir_intrinsic_store_stack:
|
|
case nir_intrinsic_store_scratch:
|
|
is_scratch = true;
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
/* >128 bit loads are split except with SMEM. On GFX6-8, >32 bit scratch loads are split. */
|
|
enum amd_gfx_level gfx_level = *(enum amd_gfx_level *)data;
|
|
if (bit_size * num_components > (is_scratch && gfx_level <= GFX8 ? 32 : 128))
|
|
return false;
|
|
|
|
uint32_t align;
|
|
if (align_offset)
|
|
align = 1 << (ffs(align_offset) - 1);
|
|
else
|
|
align = align_mul;
|
|
|
|
switch (low->intrinsic) {
|
|
case nir_intrinsic_load_global:
|
|
case nir_intrinsic_load_global_constant:
|
|
case nir_intrinsic_store_global:
|
|
case nir_intrinsic_store_ssbo:
|
|
case nir_intrinsic_load_ssbo:
|
|
case nir_intrinsic_load_ubo:
|
|
case nir_intrinsic_load_push_constant:
|
|
case nir_intrinsic_load_stack:
|
|
case nir_intrinsic_load_scratch:
|
|
case nir_intrinsic_store_stack:
|
|
case nir_intrinsic_store_scratch: {
|
|
unsigned max_components;
|
|
if (align % 4 == 0)
|
|
max_components = NIR_MAX_VEC_COMPONENTS;
|
|
else if (align % 2 == 0)
|
|
max_components = 16u / bit_size;
|
|
else
|
|
max_components = 8u / bit_size;
|
|
return (align % (bit_size / 8u)) == 0 && num_components <= max_components;
|
|
}
|
|
case nir_intrinsic_load_deref:
|
|
case nir_intrinsic_store_deref:
|
|
assert(nir_deref_mode_is(nir_src_as_deref(low->src[0]), nir_var_mem_shared));
|
|
FALLTHROUGH;
|
|
case nir_intrinsic_load_shared:
|
|
case nir_intrinsic_store_shared:
|
|
if (bit_size * num_components == 96) { /* 96 bit loads require 128 bit alignment and are split otherwise */
|
|
return align % 16 == 0;
|
|
} else if (bit_size == 16 && (align % 4)) {
|
|
/* AMD hardware can't do 2-byte aligned f16vec2 loads, but they are useful for ALU
|
|
* vectorization, because our vectorizer requires the scalar IR to already contain vectors.
|
|
*/
|
|
return (align % 2 == 0) && num_components <= 2;
|
|
} else {
|
|
if (num_components == 3) {
|
|
/* AMD hardware can't do 3-component loads except for 96-bit loads, handled above. */
|
|
return false;
|
|
}
|
|
unsigned req = bit_size * num_components;
|
|
if (req == 64 || req == 128) /* 64-bit and 128-bit loads can use ds_read2_b{32,64} */
|
|
req /= 2u;
|
|
return align % (req / 8u) == 0;
|
|
}
|
|
default:
|
|
return false;
|
|
}
|
|
return false;
|
|
}
|
|
|
|
unsigned ac_get_spi_shader_z_format(bool writes_z, bool writes_stencil, bool writes_samplemask,
|
|
bool writes_mrt0_alpha)
|
|
{
|
|
/* If writes_mrt0_alpha is true, one other flag must be true too. */
|
|
assert(!writes_mrt0_alpha || writes_z || writes_stencil || writes_samplemask);
|
|
|
|
if (writes_z || writes_mrt0_alpha) {
|
|
/* Z needs 32 bits. */
|
|
if (writes_samplemask || writes_mrt0_alpha)
|
|
return V_028710_SPI_SHADER_32_ABGR;
|
|
else if (writes_stencil)
|
|
return V_028710_SPI_SHADER_32_GR;
|
|
else
|
|
return V_028710_SPI_SHADER_32_R;
|
|
} else if (writes_stencil || writes_samplemask) {
|
|
/* Both stencil and sample mask need only 16 bits. */
|
|
return V_028710_SPI_SHADER_UINT16_ABGR;
|
|
} else {
|
|
return V_028710_SPI_SHADER_ZERO;
|
|
}
|
|
}
|
|
|
|
unsigned ac_get_cb_shader_mask(unsigned spi_shader_col_format)
|
|
{
|
|
unsigned i, cb_shader_mask = 0;
|
|
|
|
for (i = 0; i < 8; i++) {
|
|
switch ((spi_shader_col_format >> (i * 4)) & 0xf) {
|
|
case V_028714_SPI_SHADER_ZERO:
|
|
break;
|
|
case V_028714_SPI_SHADER_32_R:
|
|
cb_shader_mask |= 0x1 << (i * 4);
|
|
break;
|
|
case V_028714_SPI_SHADER_32_GR:
|
|
cb_shader_mask |= 0x3 << (i * 4);
|
|
break;
|
|
case V_028714_SPI_SHADER_32_AR:
|
|
cb_shader_mask |= 0x9u << (i * 4);
|
|
break;
|
|
case V_028714_SPI_SHADER_FP16_ABGR:
|
|
case V_028714_SPI_SHADER_UNORM16_ABGR:
|
|
case V_028714_SPI_SHADER_SNORM16_ABGR:
|
|
case V_028714_SPI_SHADER_UINT16_ABGR:
|
|
case V_028714_SPI_SHADER_SINT16_ABGR:
|
|
case V_028714_SPI_SHADER_32_ABGR:
|
|
cb_shader_mask |= 0xfu << (i * 4);
|
|
break;
|
|
default:
|
|
assert(0);
|
|
}
|
|
}
|
|
return cb_shader_mask;
|
|
}
|
|
|
|
/**
|
|
* Calculate the appropriate setting of VGT_GS_MODE when \p shader is a
|
|
* geometry shader.
|
|
*/
|
|
uint32_t ac_vgt_gs_mode(unsigned gs_max_vert_out, enum amd_gfx_level gfx_level)
|
|
{
|
|
unsigned cut_mode;
|
|
|
|
assert (gfx_level < GFX11);
|
|
|
|
if (gs_max_vert_out <= 128) {
|
|
cut_mode = V_028A40_GS_CUT_128;
|
|
} else if (gs_max_vert_out <= 256) {
|
|
cut_mode = V_028A40_GS_CUT_256;
|
|
} else if (gs_max_vert_out <= 512) {
|
|
cut_mode = V_028A40_GS_CUT_512;
|
|
} else {
|
|
assert(gs_max_vert_out <= 1024);
|
|
cut_mode = V_028A40_GS_CUT_1024;
|
|
}
|
|
|
|
return S_028A40_MODE(V_028A40_GS_SCENARIO_G) | S_028A40_CUT_MODE(cut_mode) |
|
|
S_028A40_ES_WRITE_OPTIMIZE(gfx_level <= GFX8) | S_028A40_GS_WRITE_OPTIMIZE(1) |
|
|
S_028A40_ONCHIP(gfx_level >= GFX9 ? 1 : 0);
|
|
}
|
|
|
|
/// Translate a (dfmt, nfmt) pair into a chip-appropriate combined format
|
|
/// value for LLVM8+ tbuffer intrinsics.
|
|
unsigned ac_get_tbuffer_format(enum amd_gfx_level gfx_level, unsigned dfmt, unsigned nfmt)
|
|
{
|
|
// Some games try to access vertex buffers without a valid format.
|
|
// This is a game bug, but we should still handle it gracefully.
|
|
if (dfmt == V_008F0C_GFX10_FORMAT_INVALID)
|
|
return V_008F0C_GFX10_FORMAT_INVALID;
|
|
|
|
if (gfx_level >= GFX11) {
|
|
switch (dfmt) {
|
|
default:
|
|
unreachable("bad dfmt");
|
|
case V_008F0C_BUF_DATA_FORMAT_INVALID:
|
|
return V_008F0C_GFX11_FORMAT_INVALID;
|
|
|
|
case V_008F0C_BUF_DATA_FORMAT_8:
|
|
switch (nfmt) {
|
|
case V_008F0C_BUF_NUM_FORMAT_UNORM:
|
|
return V_008F0C_GFX11_FORMAT_8_UNORM;
|
|
case V_008F0C_BUF_NUM_FORMAT_SNORM:
|
|
return V_008F0C_GFX11_FORMAT_8_SNORM;
|
|
case V_008F0C_BUF_NUM_FORMAT_USCALED:
|
|
return V_008F0C_GFX11_FORMAT_8_USCALED;
|
|
case V_008F0C_BUF_NUM_FORMAT_SSCALED:
|
|
return V_008F0C_GFX11_FORMAT_8_SSCALED;
|
|
default:
|
|
unreachable("bad nfmt");
|
|
case V_008F0C_BUF_NUM_FORMAT_UINT:
|
|
return V_008F0C_GFX11_FORMAT_8_UINT;
|
|
case V_008F0C_BUF_NUM_FORMAT_SINT:
|
|
return V_008F0C_GFX11_FORMAT_8_SINT;
|
|
}
|
|
|
|
case V_008F0C_BUF_DATA_FORMAT_8_8:
|
|
switch (nfmt) {
|
|
case V_008F0C_BUF_NUM_FORMAT_UNORM:
|
|
return V_008F0C_GFX11_FORMAT_8_8_UNORM;
|
|
case V_008F0C_BUF_NUM_FORMAT_SNORM:
|
|
return V_008F0C_GFX11_FORMAT_8_8_SNORM;
|
|
case V_008F0C_BUF_NUM_FORMAT_USCALED:
|
|
return V_008F0C_GFX11_FORMAT_8_8_USCALED;
|
|
case V_008F0C_BUF_NUM_FORMAT_SSCALED:
|
|
return V_008F0C_GFX11_FORMAT_8_8_SSCALED;
|
|
default:
|
|
unreachable("bad nfmt");
|
|
case V_008F0C_BUF_NUM_FORMAT_UINT:
|
|
return V_008F0C_GFX11_FORMAT_8_8_UINT;
|
|
case V_008F0C_BUF_NUM_FORMAT_SINT:
|
|
return V_008F0C_GFX11_FORMAT_8_8_SINT;
|
|
}
|
|
|
|
case V_008F0C_BUF_DATA_FORMAT_8_8_8_8:
|
|
switch (nfmt) {
|
|
case V_008F0C_BUF_NUM_FORMAT_UNORM:
|
|
return V_008F0C_GFX11_FORMAT_8_8_8_8_UNORM;
|
|
case V_008F0C_BUF_NUM_FORMAT_SNORM:
|
|
return V_008F0C_GFX11_FORMAT_8_8_8_8_SNORM;
|
|
case V_008F0C_BUF_NUM_FORMAT_USCALED:
|
|
return V_008F0C_GFX11_FORMAT_8_8_8_8_USCALED;
|
|
case V_008F0C_BUF_NUM_FORMAT_SSCALED:
|
|
return V_008F0C_GFX11_FORMAT_8_8_8_8_SSCALED;
|
|
default:
|
|
unreachable("bad nfmt");
|
|
case V_008F0C_BUF_NUM_FORMAT_UINT:
|
|
return V_008F0C_GFX11_FORMAT_8_8_8_8_UINT;
|
|
case V_008F0C_BUF_NUM_FORMAT_SINT:
|
|
return V_008F0C_GFX11_FORMAT_8_8_8_8_SINT;
|
|
}
|
|
|
|
case V_008F0C_BUF_DATA_FORMAT_16:
|
|
switch (nfmt) {
|
|
case V_008F0C_BUF_NUM_FORMAT_UNORM:
|
|
return V_008F0C_GFX11_FORMAT_16_UNORM;
|
|
case V_008F0C_BUF_NUM_FORMAT_SNORM:
|
|
return V_008F0C_GFX11_FORMAT_16_SNORM;
|
|
case V_008F0C_BUF_NUM_FORMAT_USCALED:
|
|
return V_008F0C_GFX11_FORMAT_16_USCALED;
|
|
case V_008F0C_BUF_NUM_FORMAT_SSCALED:
|
|
return V_008F0C_GFX11_FORMAT_16_SSCALED;
|
|
default:
|
|
unreachable("bad nfmt");
|
|
case V_008F0C_BUF_NUM_FORMAT_UINT:
|
|
return V_008F0C_GFX11_FORMAT_16_UINT;
|
|
case V_008F0C_BUF_NUM_FORMAT_SINT:
|
|
return V_008F0C_GFX11_FORMAT_16_SINT;
|
|
case V_008F0C_BUF_NUM_FORMAT_FLOAT:
|
|
return V_008F0C_GFX11_FORMAT_16_FLOAT;
|
|
}
|
|
|
|
case V_008F0C_BUF_DATA_FORMAT_16_16:
|
|
switch (nfmt) {
|
|
case V_008F0C_BUF_NUM_FORMAT_UNORM:
|
|
return V_008F0C_GFX11_FORMAT_16_16_UNORM;
|
|
case V_008F0C_BUF_NUM_FORMAT_SNORM:
|
|
return V_008F0C_GFX11_FORMAT_16_16_SNORM;
|
|
case V_008F0C_BUF_NUM_FORMAT_USCALED:
|
|
return V_008F0C_GFX11_FORMAT_16_16_USCALED;
|
|
case V_008F0C_BUF_NUM_FORMAT_SSCALED:
|
|
return V_008F0C_GFX11_FORMAT_16_16_SSCALED;
|
|
default:
|
|
unreachable("bad nfmt");
|
|
case V_008F0C_BUF_NUM_FORMAT_UINT:
|
|
return V_008F0C_GFX11_FORMAT_16_16_UINT;
|
|
case V_008F0C_BUF_NUM_FORMAT_SINT:
|
|
return V_008F0C_GFX11_FORMAT_16_16_SINT;
|
|
case V_008F0C_BUF_NUM_FORMAT_FLOAT:
|
|
return V_008F0C_GFX11_FORMAT_16_16_FLOAT;
|
|
}
|
|
|
|
case V_008F0C_BUF_DATA_FORMAT_16_16_16_16:
|
|
switch (nfmt) {
|
|
case V_008F0C_BUF_NUM_FORMAT_UNORM:
|
|
return V_008F0C_GFX11_FORMAT_16_16_16_16_UNORM;
|
|
case V_008F0C_BUF_NUM_FORMAT_SNORM:
|
|
return V_008F0C_GFX11_FORMAT_16_16_16_16_SNORM;
|
|
case V_008F0C_BUF_NUM_FORMAT_USCALED:
|
|
return V_008F0C_GFX11_FORMAT_16_16_16_16_USCALED;
|
|
case V_008F0C_BUF_NUM_FORMAT_SSCALED:
|
|
return V_008F0C_GFX11_FORMAT_16_16_16_16_SSCALED;
|
|
default:
|
|
unreachable("bad nfmt");
|
|
case V_008F0C_BUF_NUM_FORMAT_UINT:
|
|
return V_008F0C_GFX11_FORMAT_16_16_16_16_UINT;
|
|
case V_008F0C_BUF_NUM_FORMAT_SINT:
|
|
return V_008F0C_GFX11_FORMAT_16_16_16_16_SINT;
|
|
case V_008F0C_BUF_NUM_FORMAT_FLOAT:
|
|
return V_008F0C_GFX11_FORMAT_16_16_16_16_FLOAT;
|
|
}
|
|
|
|
case V_008F0C_BUF_DATA_FORMAT_32:
|
|
switch (nfmt) {
|
|
default:
|
|
unreachable("bad nfmt");
|
|
case V_008F0C_BUF_NUM_FORMAT_UINT:
|
|
return V_008F0C_GFX11_FORMAT_32_UINT;
|
|
case V_008F0C_BUF_NUM_FORMAT_SINT:
|
|
return V_008F0C_GFX11_FORMAT_32_SINT;
|
|
case V_008F0C_BUF_NUM_FORMAT_FLOAT:
|
|
return V_008F0C_GFX11_FORMAT_32_FLOAT;
|
|
}
|
|
|
|
case V_008F0C_BUF_DATA_FORMAT_32_32:
|
|
switch (nfmt) {
|
|
default:
|
|
unreachable("bad nfmt");
|
|
case V_008F0C_BUF_NUM_FORMAT_UINT:
|
|
return V_008F0C_GFX11_FORMAT_32_32_UINT;
|
|
case V_008F0C_BUF_NUM_FORMAT_SINT:
|
|
return V_008F0C_GFX11_FORMAT_32_32_SINT;
|
|
case V_008F0C_BUF_NUM_FORMAT_FLOAT:
|
|
return V_008F0C_GFX11_FORMAT_32_32_FLOAT;
|
|
}
|
|
|
|
case V_008F0C_BUF_DATA_FORMAT_32_32_32:
|
|
switch (nfmt) {
|
|
default:
|
|
unreachable("bad nfmt");
|
|
case V_008F0C_BUF_NUM_FORMAT_UINT:
|
|
return V_008F0C_GFX11_FORMAT_32_32_32_UINT;
|
|
case V_008F0C_BUF_NUM_FORMAT_SINT:
|
|
return V_008F0C_GFX11_FORMAT_32_32_32_SINT;
|
|
case V_008F0C_BUF_NUM_FORMAT_FLOAT:
|
|
return V_008F0C_GFX11_FORMAT_32_32_32_FLOAT;
|
|
}
|
|
|
|
case V_008F0C_BUF_DATA_FORMAT_32_32_32_32:
|
|
switch (nfmt) {
|
|
default:
|
|
unreachable("bad nfmt");
|
|
case V_008F0C_BUF_NUM_FORMAT_UINT:
|
|
return V_008F0C_GFX11_FORMAT_32_32_32_32_UINT;
|
|
case V_008F0C_BUF_NUM_FORMAT_SINT:
|
|
return V_008F0C_GFX11_FORMAT_32_32_32_32_SINT;
|
|
case V_008F0C_BUF_NUM_FORMAT_FLOAT:
|
|
return V_008F0C_GFX11_FORMAT_32_32_32_32_FLOAT;
|
|
}
|
|
|
|
case V_008F0C_BUF_DATA_FORMAT_2_10_10_10:
|
|
switch (nfmt) {
|
|
case V_008F0C_BUF_NUM_FORMAT_UNORM:
|
|
return V_008F0C_GFX11_FORMAT_2_10_10_10_UNORM;
|
|
case V_008F0C_BUF_NUM_FORMAT_SNORM:
|
|
return V_008F0C_GFX11_FORMAT_2_10_10_10_SNORM;
|
|
case V_008F0C_BUF_NUM_FORMAT_USCALED:
|
|
return V_008F0C_GFX11_FORMAT_2_10_10_10_USCALED;
|
|
case V_008F0C_BUF_NUM_FORMAT_SSCALED:
|
|
return V_008F0C_GFX11_FORMAT_2_10_10_10_SSCALED;
|
|
default:
|
|
unreachable("bad nfmt");
|
|
case V_008F0C_BUF_NUM_FORMAT_UINT:
|
|
return V_008F0C_GFX11_FORMAT_2_10_10_10_UINT;
|
|
case V_008F0C_BUF_NUM_FORMAT_SINT:
|
|
return V_008F0C_GFX11_FORMAT_2_10_10_10_SINT;
|
|
}
|
|
|
|
case V_008F0C_BUF_DATA_FORMAT_10_11_11:
|
|
switch (nfmt) {
|
|
default:
|
|
unreachable("bad nfmt");
|
|
case V_008F0C_BUF_NUM_FORMAT_FLOAT:
|
|
return V_008F0C_GFX11_FORMAT_10_11_11_FLOAT;
|
|
}
|
|
}
|
|
} else if (gfx_level >= GFX10) {
|
|
unsigned format;
|
|
switch (dfmt) {
|
|
default:
|
|
unreachable("bad dfmt");
|
|
case V_008F0C_BUF_DATA_FORMAT_INVALID:
|
|
format = V_008F0C_GFX10_FORMAT_INVALID;
|
|
break;
|
|
case V_008F0C_BUF_DATA_FORMAT_8:
|
|
format = V_008F0C_GFX10_FORMAT_8_UINT;
|
|
break;
|
|
case V_008F0C_BUF_DATA_FORMAT_8_8:
|
|
format = V_008F0C_GFX10_FORMAT_8_8_UINT;
|
|
break;
|
|
case V_008F0C_BUF_DATA_FORMAT_8_8_8_8:
|
|
format = V_008F0C_GFX10_FORMAT_8_8_8_8_UINT;
|
|
break;
|
|
case V_008F0C_BUF_DATA_FORMAT_16:
|
|
format = V_008F0C_GFX10_FORMAT_16_UINT;
|
|
break;
|
|
case V_008F0C_BUF_DATA_FORMAT_16_16:
|
|
format = V_008F0C_GFX10_FORMAT_16_16_UINT;
|
|
break;
|
|
case V_008F0C_BUF_DATA_FORMAT_16_16_16_16:
|
|
format = V_008F0C_GFX10_FORMAT_16_16_16_16_UINT;
|
|
break;
|
|
case V_008F0C_BUF_DATA_FORMAT_32:
|
|
format = V_008F0C_GFX10_FORMAT_32_UINT;
|
|
break;
|
|
case V_008F0C_BUF_DATA_FORMAT_32_32:
|
|
format = V_008F0C_GFX10_FORMAT_32_32_UINT;
|
|
break;
|
|
case V_008F0C_BUF_DATA_FORMAT_32_32_32:
|
|
format = V_008F0C_GFX10_FORMAT_32_32_32_UINT;
|
|
break;
|
|
case V_008F0C_BUF_DATA_FORMAT_32_32_32_32:
|
|
format = V_008F0C_GFX10_FORMAT_32_32_32_32_UINT;
|
|
break;
|
|
case V_008F0C_BUF_DATA_FORMAT_2_10_10_10:
|
|
format = V_008F0C_GFX10_FORMAT_2_10_10_10_UINT;
|
|
break;
|
|
case V_008F0C_BUF_DATA_FORMAT_10_11_11:
|
|
format = V_008F0C_GFX10_FORMAT_10_11_11_UINT;
|
|
break;
|
|
}
|
|
|
|
// Use the regularity properties of the combined format enum.
|
|
//
|
|
// Note: float is incompatible with 8-bit data formats,
|
|
// [us]{norm,scaled} are incompatible with 32-bit data formats.
|
|
// [us]scaled are not writable.
|
|
switch (nfmt) {
|
|
case V_008F0C_BUF_NUM_FORMAT_UNORM:
|
|
format -= 4;
|
|
break;
|
|
case V_008F0C_BUF_NUM_FORMAT_SNORM:
|
|
format -= 3;
|
|
break;
|
|
case V_008F0C_BUF_NUM_FORMAT_USCALED:
|
|
format -= 2;
|
|
break;
|
|
case V_008F0C_BUF_NUM_FORMAT_SSCALED:
|
|
format -= 1;
|
|
break;
|
|
default:
|
|
unreachable("bad nfmt");
|
|
case V_008F0C_BUF_NUM_FORMAT_UINT:
|
|
break;
|
|
case V_008F0C_BUF_NUM_FORMAT_SINT:
|
|
format += 1;
|
|
break;
|
|
case V_008F0C_BUF_NUM_FORMAT_FLOAT:
|
|
format += 2;
|
|
break;
|
|
}
|
|
|
|
return format;
|
|
} else {
|
|
return dfmt | (nfmt << 4);
|
|
}
|
|
}
|
|
|
|
static const struct ac_data_format_info data_format_table[] = {
|
|
[V_008F0C_BUF_DATA_FORMAT_INVALID] = {0, 4, 0, V_008F0C_BUF_DATA_FORMAT_INVALID},
|
|
[V_008F0C_BUF_DATA_FORMAT_8] = {1, 1, 1, V_008F0C_BUF_DATA_FORMAT_8},
|
|
[V_008F0C_BUF_DATA_FORMAT_16] = {2, 1, 2, V_008F0C_BUF_DATA_FORMAT_16},
|
|
[V_008F0C_BUF_DATA_FORMAT_8_8] = {2, 2, 1, V_008F0C_BUF_DATA_FORMAT_8},
|
|
[V_008F0C_BUF_DATA_FORMAT_32] = {4, 1, 4, V_008F0C_BUF_DATA_FORMAT_32},
|
|
[V_008F0C_BUF_DATA_FORMAT_16_16] = {4, 2, 2, V_008F0C_BUF_DATA_FORMAT_16},
|
|
[V_008F0C_BUF_DATA_FORMAT_10_11_11] = {4, 3, 0, V_008F0C_BUF_DATA_FORMAT_10_11_11},
|
|
[V_008F0C_BUF_DATA_FORMAT_11_11_10] = {4, 3, 0, V_008F0C_BUF_DATA_FORMAT_11_11_10},
|
|
[V_008F0C_BUF_DATA_FORMAT_10_10_10_2] = {4, 4, 0, V_008F0C_BUF_DATA_FORMAT_10_10_10_2},
|
|
[V_008F0C_BUF_DATA_FORMAT_2_10_10_10] = {4, 4, 0, V_008F0C_BUF_DATA_FORMAT_2_10_10_10},
|
|
[V_008F0C_BUF_DATA_FORMAT_8_8_8_8] = {4, 4, 1, V_008F0C_BUF_DATA_FORMAT_8},
|
|
[V_008F0C_BUF_DATA_FORMAT_32_32] = {8, 2, 4, V_008F0C_BUF_DATA_FORMAT_32},
|
|
[V_008F0C_BUF_DATA_FORMAT_16_16_16_16] = {8, 4, 2, V_008F0C_BUF_DATA_FORMAT_16},
|
|
[V_008F0C_BUF_DATA_FORMAT_32_32_32] = {12, 3, 4, V_008F0C_BUF_DATA_FORMAT_32},
|
|
[V_008F0C_BUF_DATA_FORMAT_32_32_32_32] = {16, 4, 4, V_008F0C_BUF_DATA_FORMAT_32},
|
|
};
|
|
|
|
const struct ac_data_format_info *ac_get_data_format_info(unsigned dfmt)
|
|
{
|
|
assert(dfmt < ARRAY_SIZE(data_format_table));
|
|
return &data_format_table[dfmt];
|
|
}
|
|
|
|
#define DUP2(v) v, v
|
|
#define DUP3(v) v, v, v
|
|
#define DUP4(v) v, v, v, v
|
|
|
|
#define FMT(dfmt, nfmt) 0xb, {HW_FMT(dfmt, nfmt), HW_FMT(dfmt##_##dfmt, nfmt), HW_FMT_INVALID, HW_FMT(dfmt##_##dfmt##_##dfmt##_##dfmt, nfmt)}
|
|
#define FMT_32(nfmt) 0xf, {HW_FMT(32, nfmt), HW_FMT(32_32, nfmt), HW_FMT(32_32_32, nfmt), HW_FMT(32_32_32_32, nfmt)}
|
|
#define FMT_64(nfmt) 0x3, {HW_FMT(32_32, nfmt), HW_FMT(32_32_32_32, nfmt), DUP2(HW_FMT_INVALID)}
|
|
#define FMTP(dfmt, nfmt) 0xf, {DUP4(HW_FMT(dfmt, nfmt))}
|
|
|
|
#define DST_SEL(x, y, z, w) \
|
|
(S_008F0C_DST_SEL_X(V_008F0C_SQ_SEL_##x) | S_008F0C_DST_SEL_Y(V_008F0C_SQ_SEL_##y) | \
|
|
S_008F0C_DST_SEL_Z(V_008F0C_SQ_SEL_##z) | S_008F0C_DST_SEL_W(V_008F0C_SQ_SEL_##w))
|
|
|
|
#define LIST_NFMT_8_16(nfmt) \
|
|
[(int)PIPE_FORMAT_R8_##nfmt] = {DST_SEL(X,0,0,1), 1, 1, 1, FMT(8, nfmt)}, \
|
|
[(int)PIPE_FORMAT_R8G8_##nfmt] = {DST_SEL(X,Y,0,1), 2, 2, 1, FMT(8, nfmt)}, \
|
|
[(int)PIPE_FORMAT_R8G8B8_##nfmt] = {DST_SEL(X,Y,Z,1), 3, 3, 1, FMT(8, nfmt)}, \
|
|
[(int)PIPE_FORMAT_B8G8R8_##nfmt] = {DST_SEL(Z,Y,X,1), 3, 3, 1, FMT(8, nfmt)}, \
|
|
[(int)PIPE_FORMAT_R8G8B8A8_##nfmt] = {DST_SEL(X,Y,Z,W), 4, 4, 1, FMT(8, nfmt)}, \
|
|
[(int)PIPE_FORMAT_B8G8R8A8_##nfmt] = {DST_SEL(Z,Y,X,W), 4, 4, 1, FMT(8, nfmt)}, \
|
|
[(int)PIPE_FORMAT_R16_##nfmt] = {DST_SEL(X,0,0,1), 2, 1, 2, FMT(16, nfmt)}, \
|
|
[(int)PIPE_FORMAT_R16G16_##nfmt] = {DST_SEL(X,Y,0,1), 4, 2, 2, FMT(16, nfmt)}, \
|
|
[(int)PIPE_FORMAT_R16G16B16_##nfmt] = {DST_SEL(X,Y,Z,1), 6, 3, 2, FMT(16, nfmt)}, \
|
|
[(int)PIPE_FORMAT_R16G16B16A16_##nfmt] = {DST_SEL(X,Y,Z,W), 8, 4, 2, FMT(16, nfmt)},
|
|
|
|
#define LIST_NFMT_32_64(nfmt) \
|
|
[(int)PIPE_FORMAT_R32_##nfmt] = {DST_SEL(X,0,0,1), 4, 1, 4, FMT_32(nfmt)}, \
|
|
[(int)PIPE_FORMAT_R32G32_##nfmt] = {DST_SEL(X,Y,0,1), 8, 2, 4, FMT_32(nfmt)}, \
|
|
[(int)PIPE_FORMAT_R32G32B32_##nfmt] = {DST_SEL(X,Y,Z,1), 12, 3, 4, FMT_32(nfmt)}, \
|
|
[(int)PIPE_FORMAT_R32G32B32A32_##nfmt] = {DST_SEL(X,Y,Z,W), 16, 4, 4, FMT_32(nfmt)}, \
|
|
[(int)PIPE_FORMAT_R64_##nfmt] = {DST_SEL(X,Y,0,0), 8, 1, 8, FMT_64(nfmt)}, \
|
|
[(int)PIPE_FORMAT_R64G64_##nfmt] = {DST_SEL(X,Y,Z,W), 16, 2, 8, FMT_64(nfmt)}, \
|
|
[(int)PIPE_FORMAT_R64G64B64_##nfmt] = {DST_SEL(X,Y,Z,W), 24, 3, 8, FMT_64(nfmt)}, \
|
|
[(int)PIPE_FORMAT_R64G64B64A64_##nfmt] = {DST_SEL(X,Y,Z,W), 32, 4, 8, FMT_64(nfmt)}, \
|
|
|
|
#define VB_FORMATS \
|
|
[(int)PIPE_FORMAT_NONE] = {DST_SEL(0,0,0,1), 0, 4, 0, 0xf, {DUP4(HW_FMT_INVALID)}}, \
|
|
LIST_NFMT_8_16(UNORM) \
|
|
LIST_NFMT_8_16(SNORM) \
|
|
LIST_NFMT_8_16(USCALED) \
|
|
LIST_NFMT_8_16(SSCALED) \
|
|
LIST_NFMT_8_16(UINT) \
|
|
LIST_NFMT_8_16(SINT) \
|
|
LIST_NFMT_32_64(UINT) \
|
|
LIST_NFMT_32_64(SINT) \
|
|
LIST_NFMT_32_64(FLOAT) \
|
|
[(int)PIPE_FORMAT_R16_FLOAT] = {DST_SEL(X,0,0,1), 2, 1, 2, FMT(16, FLOAT)}, \
|
|
[(int)PIPE_FORMAT_R16G16_FLOAT] = {DST_SEL(X,Y,0,1), 4, 2, 2, FMT(16, FLOAT)}, \
|
|
[(int)PIPE_FORMAT_R16G16B16_FLOAT] = {DST_SEL(X,Y,Z,1), 6, 3, 2, FMT(16, FLOAT)}, \
|
|
[(int)PIPE_FORMAT_R16G16B16A16_FLOAT] = {DST_SEL(X,Y,Z,W), 8, 4, 2, FMT(16, FLOAT)}, \
|
|
[(int)PIPE_FORMAT_B10G10R10A2_UNORM] = {DST_SEL(Z,Y,X,W), 4, 4, 0, FMTP(2_10_10_10, UNORM)}, \
|
|
[(int)PIPE_FORMAT_B10G10R10A2_SNORM] = {DST_SEL(Z,Y,X,W), 4, 4, 0, FMTP(2_10_10_10, SNORM), \
|
|
AA(AC_ALPHA_ADJUST_SNORM)}, \
|
|
[(int)PIPE_FORMAT_B10G10R10A2_USCALED] = {DST_SEL(Z,Y,X,W), 4, 4, 0, FMTP(2_10_10_10, USCALED)}, \
|
|
[(int)PIPE_FORMAT_B10G10R10A2_SSCALED] = {DST_SEL(Z,Y,X,W), 4, 4, 0, FMTP(2_10_10_10, SSCALED), \
|
|
AA(AC_ALPHA_ADJUST_SSCALED)}, \
|
|
[(int)PIPE_FORMAT_B10G10R10A2_UINT] = {DST_SEL(Z,Y,X,W), 4, 4, 0, FMTP(2_10_10_10, UINT)}, \
|
|
[(int)PIPE_FORMAT_B10G10R10A2_SINT] = {DST_SEL(Z,Y,X,W), 4, 4, 0, FMTP(2_10_10_10, SINT), \
|
|
AA(AC_ALPHA_ADJUST_SINT)}, \
|
|
[(int)PIPE_FORMAT_R10G10B10A2_UNORM] = {DST_SEL(X,Y,Z,W), 4, 4, 0, FMTP(2_10_10_10, UNORM)}, \
|
|
[(int)PIPE_FORMAT_R10G10B10A2_SNORM] = {DST_SEL(X,Y,Z,W), 4, 4, 0, FMTP(2_10_10_10, SNORM), \
|
|
AA(AC_ALPHA_ADJUST_SNORM)}, \
|
|
[(int)PIPE_FORMAT_R10G10B10A2_USCALED] = {DST_SEL(X,Y,Z,W), 4, 4, 0, FMTP(2_10_10_10, USCALED)}, \
|
|
[(int)PIPE_FORMAT_R10G10B10A2_SSCALED] = {DST_SEL(X,Y,Z,W), 4, 4, 0, FMTP(2_10_10_10, SSCALED), \
|
|
AA(AC_ALPHA_ADJUST_SSCALED)}, \
|
|
[(int)PIPE_FORMAT_R10G10B10A2_UINT] = {DST_SEL(X,Y,Z,W), 4, 4, 0, FMTP(2_10_10_10, UINT)}, \
|
|
[(int)PIPE_FORMAT_R10G10B10A2_SINT] = {DST_SEL(X,Y,Z,W), 4, 4, 0, FMTP(2_10_10_10, SINT), \
|
|
AA(AC_ALPHA_ADJUST_SINT)}, \
|
|
[(int)PIPE_FORMAT_R11G11B10_FLOAT] = {DST_SEL(X,Y,Z,1), 4, 3, 0, FMTP(10_11_11, FLOAT)}, \
|
|
|
|
#define HW_FMT(dfmt, nfmt) (V_008F0C_BUF_DATA_FORMAT_##dfmt | (V_008F0C_BUF_NUM_FORMAT_##nfmt << 4))
|
|
#define HW_FMT_INVALID (V_008F0C_BUF_DATA_FORMAT_INVALID | (V_008F0C_BUF_NUM_FORMAT_UNORM << 4))
|
|
#define AA(v) v
|
|
static const struct ac_vtx_format_info vb_formats_gfx6_alpha_adjust[] = {VB_FORMATS};
|
|
#undef AA
|
|
|
|
#define AA(v) AC_ALPHA_ADJUST_NONE
|
|
static const struct ac_vtx_format_info vb_formats_gfx6[] = {VB_FORMATS};
|
|
#undef HW_FMT_INVALID
|
|
#undef HW_FMT
|
|
|
|
#define HW_FMT(dfmt, nfmt) V_008F0C_GFX10_FORMAT_##dfmt##_##nfmt
|
|
#define HW_FMT_INVALID V_008F0C_GFX10_FORMAT_INVALID
|
|
static const struct ac_vtx_format_info vb_formats_gfx10[] = {VB_FORMATS};
|
|
#undef HW_FMT_INVALID
|
|
#undef HW_FMT
|
|
|
|
#define HW_FMT(dfmt, nfmt) V_008F0C_GFX11_FORMAT_##dfmt##_##nfmt
|
|
#define HW_FMT_INVALID V_008F0C_GFX11_FORMAT_INVALID
|
|
static const struct ac_vtx_format_info vb_formats_gfx11[] = {VB_FORMATS};
|
|
|
|
const struct ac_vtx_format_info *
|
|
ac_get_vtx_format_info_table(enum amd_gfx_level level, enum radeon_family family)
|
|
{
|
|
if (level >= GFX11)
|
|
return vb_formats_gfx11;
|
|
else if (level >= GFX10)
|
|
return vb_formats_gfx10;
|
|
bool alpha_adjust = level <= GFX8 && family != CHIP_STONEY;
|
|
return alpha_adjust ? vb_formats_gfx6_alpha_adjust : vb_formats_gfx6;
|
|
}
|
|
|
|
const struct ac_vtx_format_info *
|
|
ac_get_vtx_format_info(enum amd_gfx_level level, enum radeon_family family, enum pipe_format fmt)
|
|
{
|
|
return &ac_get_vtx_format_info_table(level, family)[fmt];
|
|
}
|
|
|
|
/**
|
|
* Check whether the specified fetch size is safe to use with MTBUF.
|
|
*
|
|
* Split typed vertex buffer loads when necessary to avoid any
|
|
* alignment issues that trigger memory violations and eventually a GPU
|
|
* hang. This can happen if the stride (static or dynamic) is unaligned and
|
|
* also if the VBO offset is aligned to a scalar (eg. stride is 8 and VBO
|
|
* offset is 2 for R16G16B16A16_SNORM).
|
|
*/
|
|
static bool
|
|
is_fetch_size_safe(const enum amd_gfx_level gfx_level, const struct ac_vtx_format_info* vtx_info,
|
|
const unsigned offset, const unsigned alignment, const unsigned channels)
|
|
{
|
|
if (!(vtx_info->has_hw_format & BITFIELD_BIT(channels - 1)))
|
|
return false;
|
|
|
|
unsigned vertex_byte_size = vtx_info->chan_byte_size * channels;
|
|
return (gfx_level >= GFX7 && gfx_level <= GFX9) ||
|
|
(offset % vertex_byte_size == 0 && MAX2(alignment, 1) % vertex_byte_size == 0);
|
|
}
|
|
|
|
/**
|
|
* Gets the number of channels that can be safely fetched by MTBUF (typed buffer load)
|
|
* instructions without triggering alignment-related issues.
|
|
*/
|
|
unsigned
|
|
ac_get_safe_fetch_size(const enum amd_gfx_level gfx_level, const struct ac_vtx_format_info* vtx_info,
|
|
const unsigned offset, const unsigned max_channels, const unsigned alignment,
|
|
const unsigned num_channels)
|
|
{
|
|
/* Packed formats can't be split. */
|
|
if (!vtx_info->chan_byte_size)
|
|
return vtx_info->num_channels;
|
|
|
|
/* Early exit if the specified number of channels is fine. */
|
|
if (is_fetch_size_safe(gfx_level, vtx_info, offset, alignment, num_channels))
|
|
return num_channels;
|
|
|
|
/* First, assume that more load instructions are worse and try using a larger data format. */
|
|
unsigned new_channels = num_channels + 1;
|
|
while (new_channels <= max_channels &&
|
|
!is_fetch_size_safe(gfx_level, vtx_info, offset, alignment, new_channels)) {
|
|
new_channels++;
|
|
}
|
|
|
|
/* Found a feasible load size. */
|
|
if (new_channels <= max_channels)
|
|
return new_channels;
|
|
|
|
/* Try decreasing load size (at the cost of more load instructions). */
|
|
new_channels = num_channels;
|
|
while (new_channels > 1 &&
|
|
!is_fetch_size_safe(gfx_level, vtx_info, offset, alignment, new_channels)) {
|
|
new_channels--;
|
|
}
|
|
|
|
return new_channels;
|
|
}
|
|
|
|
enum ac_image_dim ac_get_sampler_dim(enum amd_gfx_level gfx_level, enum glsl_sampler_dim dim,
|
|
bool is_array)
|
|
{
|
|
switch (dim) {
|
|
case GLSL_SAMPLER_DIM_1D:
|
|
if (gfx_level == GFX9)
|
|
return is_array ? ac_image_2darray : ac_image_2d;
|
|
return is_array ? ac_image_1darray : ac_image_1d;
|
|
case GLSL_SAMPLER_DIM_2D:
|
|
case GLSL_SAMPLER_DIM_RECT:
|
|
case GLSL_SAMPLER_DIM_EXTERNAL:
|
|
return is_array ? ac_image_2darray : ac_image_2d;
|
|
case GLSL_SAMPLER_DIM_3D:
|
|
return ac_image_3d;
|
|
case GLSL_SAMPLER_DIM_CUBE:
|
|
return ac_image_cube;
|
|
case GLSL_SAMPLER_DIM_MS:
|
|
return is_array ? ac_image_2darraymsaa : ac_image_2dmsaa;
|
|
case GLSL_SAMPLER_DIM_SUBPASS:
|
|
return ac_image_2darray;
|
|
case GLSL_SAMPLER_DIM_SUBPASS_MS:
|
|
return ac_image_2darraymsaa;
|
|
default:
|
|
unreachable("bad sampler dim");
|
|
}
|
|
}
|
|
|
|
enum ac_image_dim ac_get_image_dim(enum amd_gfx_level gfx_level, enum glsl_sampler_dim sdim,
|
|
bool is_array)
|
|
{
|
|
enum ac_image_dim dim = ac_get_sampler_dim(gfx_level, sdim, is_array);
|
|
|
|
/* Match the resource type set in the descriptor. */
|
|
if (dim == ac_image_cube || (gfx_level <= GFX8 && dim == ac_image_3d))
|
|
dim = ac_image_2darray;
|
|
else if (sdim == GLSL_SAMPLER_DIM_2D && !is_array && gfx_level == GFX9) {
|
|
/* When a single layer of a 3D texture is bound, the shader
|
|
* will refer to a 2D target, but the descriptor has a 3D type.
|
|
* Since the HW ignores BASE_ARRAY in this case, we need to
|
|
* send 3 coordinates. This doesn't hurt when the underlying
|
|
* texture is non-3D.
|
|
*/
|
|
dim = ac_image_3d;
|
|
}
|
|
|
|
return dim;
|
|
}
|
|
|
|
unsigned ac_get_fs_input_vgpr_cnt(const struct ac_shader_config *config,
|
|
uint8_t *num_fragcoord_components)
|
|
{
|
|
unsigned num_input_vgprs = 0;
|
|
unsigned fragcoord_components = 0;
|
|
|
|
if (G_0286CC_PERSP_SAMPLE_ENA(config->spi_ps_input_addr))
|
|
num_input_vgprs += 2;
|
|
if (G_0286CC_PERSP_CENTER_ENA(config->spi_ps_input_addr))
|
|
num_input_vgprs += 2;
|
|
if (G_0286CC_PERSP_CENTROID_ENA(config->spi_ps_input_addr))
|
|
num_input_vgprs += 2;
|
|
if (G_0286CC_PERSP_PULL_MODEL_ENA(config->spi_ps_input_addr))
|
|
num_input_vgprs += 3;
|
|
if (G_0286CC_LINEAR_SAMPLE_ENA(config->spi_ps_input_addr))
|
|
num_input_vgprs += 2;
|
|
if (G_0286CC_LINEAR_CENTER_ENA(config->spi_ps_input_addr))
|
|
num_input_vgprs += 2;
|
|
if (G_0286CC_LINEAR_CENTROID_ENA(config->spi_ps_input_addr))
|
|
num_input_vgprs += 2;
|
|
if (G_0286CC_LINE_STIPPLE_TEX_ENA(config->spi_ps_input_addr))
|
|
num_input_vgprs += 1;
|
|
if (G_0286CC_POS_X_FLOAT_ENA(config->spi_ps_input_addr)) {
|
|
num_input_vgprs += 1;
|
|
fragcoord_components++;
|
|
}
|
|
if (G_0286CC_POS_Y_FLOAT_ENA(config->spi_ps_input_addr)) {
|
|
num_input_vgprs += 1;
|
|
fragcoord_components++;
|
|
}
|
|
if (G_0286CC_POS_Z_FLOAT_ENA(config->spi_ps_input_addr)) {
|
|
num_input_vgprs += 1;
|
|
fragcoord_components++;
|
|
}
|
|
if (G_0286CC_POS_W_FLOAT_ENA(config->spi_ps_input_addr)) {
|
|
num_input_vgprs += 1;
|
|
fragcoord_components++;
|
|
}
|
|
if (G_0286CC_FRONT_FACE_ENA(config->spi_ps_input_addr))
|
|
num_input_vgprs += 1;
|
|
if (G_0286CC_ANCILLARY_ENA(config->spi_ps_input_addr))
|
|
num_input_vgprs += 1;
|
|
if (G_0286CC_SAMPLE_COVERAGE_ENA(config->spi_ps_input_addr))
|
|
num_input_vgprs += 1;
|
|
if (G_0286CC_POS_FIXED_PT_ENA(config->spi_ps_input_addr))
|
|
num_input_vgprs += 1;
|
|
|
|
if (num_fragcoord_components)
|
|
*num_fragcoord_components = fragcoord_components;
|
|
|
|
return num_input_vgprs;
|
|
}
|
|
|
|
uint16_t ac_get_ps_iter_mask(unsigned ps_iter_samples)
|
|
{
|
|
/* The bit pattern matches that used by fixed function fragment
|
|
* processing.
|
|
*/
|
|
switch (ps_iter_samples) {
|
|
case 1: return 0xffff;
|
|
case 2: return 0x5555;
|
|
case 4: return 0x1111;
|
|
case 8: return 0x0101;
|
|
case 16: return 0x0001;
|
|
default:
|
|
unreachable("invalid sample count");
|
|
}
|
|
}
|
|
|
|
void ac_choose_spi_color_formats(unsigned format, unsigned swap, unsigned ntype,
|
|
bool is_depth, bool use_rbplus,
|
|
struct ac_spi_color_formats *formats)
|
|
{
|
|
/* Alpha is needed for alpha-to-coverage.
|
|
* Blending may be with or without alpha.
|
|
*/
|
|
unsigned normal = 0; /* most optimal, may not support blending or export alpha */
|
|
unsigned alpha = 0; /* exports alpha, but may not support blending */
|
|
unsigned blend = 0; /* supports blending, but may not export alpha */
|
|
unsigned blend_alpha = 0; /* least optimal, supports blending and exports alpha */
|
|
|
|
/* Choose the SPI color formats. These are required values for RB+.
|
|
* Other chips have multiple choices, though they are not necessarily better.
|
|
*/
|
|
switch (format) {
|
|
case V_028C70_COLOR_5_6_5:
|
|
case V_028C70_COLOR_1_5_5_5:
|
|
case V_028C70_COLOR_5_5_5_1:
|
|
case V_028C70_COLOR_4_4_4_4:
|
|
case V_028C70_COLOR_10_11_11:
|
|
case V_028C70_COLOR_11_11_10:
|
|
case V_028C70_COLOR_5_9_9_9:
|
|
case V_028C70_COLOR_8:
|
|
case V_028C70_COLOR_8_8:
|
|
case V_028C70_COLOR_8_8_8_8:
|
|
case V_028C70_COLOR_10_10_10_2:
|
|
case V_028C70_COLOR_2_10_10_10:
|
|
if (ntype == V_028C70_NUMBER_UINT)
|
|
alpha = blend = blend_alpha = normal = V_028714_SPI_SHADER_UINT16_ABGR;
|
|
else if (ntype == V_028C70_NUMBER_SINT)
|
|
alpha = blend = blend_alpha = normal = V_028714_SPI_SHADER_SINT16_ABGR;
|
|
else
|
|
alpha = blend = blend_alpha = normal = V_028714_SPI_SHADER_FP16_ABGR;
|
|
|
|
if (!use_rbplus && format == V_028C70_COLOR_8 &&
|
|
ntype != V_028C70_NUMBER_SRGB && swap == V_028C70_SWAP_STD) /* R */ {
|
|
/* When RB+ is enabled, R8_UNORM should use FP16_ABGR for 2x
|
|
* exporting performance. Otherwise, use 32_R to remove useless
|
|
* instructions needed for 16-bit compressed exports.
|
|
*/
|
|
blend = normal = V_028714_SPI_SHADER_32_R;
|
|
}
|
|
break;
|
|
|
|
case V_028C70_COLOR_16:
|
|
case V_028C70_COLOR_16_16:
|
|
case V_028C70_COLOR_16_16_16_16:
|
|
if (ntype == V_028C70_NUMBER_UNORM || ntype == V_028C70_NUMBER_SNORM) {
|
|
/* UNORM16 and SNORM16 don't support blending */
|
|
if (ntype == V_028C70_NUMBER_UNORM)
|
|
normal = alpha = V_028714_SPI_SHADER_UNORM16_ABGR;
|
|
else
|
|
normal = alpha = V_028714_SPI_SHADER_SNORM16_ABGR;
|
|
|
|
/* Use 32 bits per channel for blending. */
|
|
if (format == V_028C70_COLOR_16) {
|
|
if (swap == V_028C70_SWAP_STD) { /* R */
|
|
blend = V_028714_SPI_SHADER_32_R;
|
|
blend_alpha = V_028714_SPI_SHADER_32_AR;
|
|
} else if (swap == V_028C70_SWAP_ALT_REV) /* A */
|
|
blend = blend_alpha = V_028714_SPI_SHADER_32_AR;
|
|
else
|
|
assert(0);
|
|
} else if (format == V_028C70_COLOR_16_16) {
|
|
if (swap == V_028C70_SWAP_STD || swap == V_028C70_SWAP_STD_REV) { /* RG or GR */
|
|
blend = V_028714_SPI_SHADER_32_GR;
|
|
blend_alpha = V_028714_SPI_SHADER_32_ABGR;
|
|
} else if (swap == V_028C70_SWAP_ALT) /* RA */
|
|
blend = blend_alpha = V_028714_SPI_SHADER_32_AR;
|
|
else
|
|
assert(0);
|
|
} else /* 16_16_16_16 */
|
|
blend = blend_alpha = V_028714_SPI_SHADER_32_ABGR;
|
|
} else if (ntype == V_028C70_NUMBER_UINT)
|
|
alpha = blend = blend_alpha = normal = V_028714_SPI_SHADER_UINT16_ABGR;
|
|
else if (ntype == V_028C70_NUMBER_SINT)
|
|
alpha = blend = blend_alpha = normal = V_028714_SPI_SHADER_SINT16_ABGR;
|
|
else if (ntype == V_028C70_NUMBER_FLOAT)
|
|
alpha = blend = blend_alpha = normal = V_028714_SPI_SHADER_FP16_ABGR;
|
|
else
|
|
assert(0);
|
|
break;
|
|
|
|
case V_028C70_COLOR_32:
|
|
if (swap == V_028C70_SWAP_STD) { /* R */
|
|
blend = normal = V_028714_SPI_SHADER_32_R;
|
|
alpha = blend_alpha = V_028714_SPI_SHADER_32_AR;
|
|
} else if (swap == V_028C70_SWAP_ALT_REV) /* A */
|
|
alpha = blend = blend_alpha = normal = V_028714_SPI_SHADER_32_AR;
|
|
else
|
|
assert(0);
|
|
break;
|
|
|
|
case V_028C70_COLOR_32_32:
|
|
if (swap == V_028C70_SWAP_STD || swap == V_028C70_SWAP_STD_REV) { /* RG or GR */
|
|
blend = normal = V_028714_SPI_SHADER_32_GR;
|
|
alpha = blend_alpha = V_028714_SPI_SHADER_32_ABGR;
|
|
} else if (swap == V_028C70_SWAP_ALT) /* RA */
|
|
alpha = blend = blend_alpha = normal = V_028714_SPI_SHADER_32_AR;
|
|
else
|
|
assert(0);
|
|
break;
|
|
|
|
case V_028C70_COLOR_32_32_32_32:
|
|
case V_028C70_COLOR_8_24:
|
|
case V_028C70_COLOR_24_8:
|
|
case V_028C70_COLOR_X24_8_32_FLOAT:
|
|
alpha = blend = blend_alpha = normal = V_028714_SPI_SHADER_32_ABGR;
|
|
break;
|
|
|
|
default:
|
|
assert(0);
|
|
return;
|
|
}
|
|
|
|
/* The DB->CB copy needs 32_ABGR. */
|
|
if (is_depth)
|
|
alpha = blend = blend_alpha = normal = V_028714_SPI_SHADER_32_ABGR;
|
|
|
|
formats->normal = normal;
|
|
formats->alpha = alpha;
|
|
formats->blend = blend;
|
|
formats->blend_alpha = blend_alpha;
|
|
}
|
|
|
|
void ac_compute_late_alloc(const struct radeon_info *info, bool ngg, bool ngg_culling,
|
|
bool uses_scratch, unsigned *late_alloc_wave64, unsigned *cu_mask)
|
|
{
|
|
*late_alloc_wave64 = 0; /* The limit is per SA. */
|
|
*cu_mask = 0xffff;
|
|
|
|
/* This should never be called on gfx12. Gfx12 doesn't need to mask CUs for late alloc. */
|
|
assert(info->gfx_level < GFX12);
|
|
|
|
/* CU masking can decrease performance and cause a hang with <= 2 CUs per SA. */
|
|
if (info->min_good_cu_per_sa <= 2)
|
|
return;
|
|
|
|
/* If scratch is used with late alloc, the GPU could deadlock if PS uses scratch too. A more
|
|
* complicated computation is needed to enable late alloc with scratch (see PAL).
|
|
*/
|
|
if (uses_scratch)
|
|
return;
|
|
|
|
/* Late alloc is not used for NGG on Navi14 due to a hw bug. */
|
|
if (ngg && info->family == CHIP_NAVI14)
|
|
return;
|
|
|
|
if (info->gfx_level >= GFX10) {
|
|
/* For Wave32, the hw will launch twice the number of late alloc waves, so 1 == 2x wave32.
|
|
* These limits are estimated because they are all safe but they vary in performance.
|
|
*/
|
|
if (ngg_culling)
|
|
*late_alloc_wave64 = info->min_good_cu_per_sa * 10;
|
|
else if (info->gfx_level >= GFX11)
|
|
*late_alloc_wave64 = 63;
|
|
else
|
|
*late_alloc_wave64 = info->min_good_cu_per_sa * 4;
|
|
|
|
/* Limit LATE_ALLOC_GS to prevent a hang (hw bug) on gfx10. */
|
|
if (info->gfx_level == GFX10 && ngg)
|
|
*late_alloc_wave64 = MIN2(*late_alloc_wave64, 64);
|
|
|
|
/* Gfx10: CU2 & CU3 must be disabled to prevent a hw deadlock.
|
|
* Others: CU1 must be disabled to prevent a hw deadlock.
|
|
*
|
|
* The deadlock is caused by late alloc, which usually increases performance.
|
|
*/
|
|
*cu_mask &= info->gfx_level == GFX10 ? ~BITFIELD_RANGE(2, 2) :
|
|
~BITFIELD_RANGE(1, 1);
|
|
} else {
|
|
if (info->min_good_cu_per_sa <= 4) {
|
|
/* Too few available compute units per SA. Disallowing VS to run on one CU could hurt us
|
|
* more than late VS allocation would help.
|
|
*
|
|
* 2 is the highest safe number that allows us to keep all CUs enabled.
|
|
*/
|
|
*late_alloc_wave64 = 2;
|
|
} else {
|
|
/* This is a good initial value, allowing 1 late_alloc wave per SIMD on num_cu - 2.
|
|
*/
|
|
*late_alloc_wave64 = (info->min_good_cu_per_sa - 2) * 4;
|
|
}
|
|
|
|
/* VS can't execute on one CU if the limit is > 2. */
|
|
if (*late_alloc_wave64 > 2)
|
|
*cu_mask = 0xfffe; /* 1 CU disabled */
|
|
}
|
|
|
|
/* Max number that fits into the register field. */
|
|
if (ngg) /* GS */
|
|
*late_alloc_wave64 = MIN2(*late_alloc_wave64, G_00B204_SPI_SHADER_LATE_ALLOC_GS_GFX10(~0u));
|
|
else /* VS */
|
|
*late_alloc_wave64 = MIN2(*late_alloc_wave64, G_00B11C_LIMIT(~0u));
|
|
}
|
|
|
|
unsigned ac_compute_cs_workgroup_size(const uint16_t sizes[3], bool variable, unsigned max)
|
|
{
|
|
if (variable)
|
|
return max;
|
|
|
|
return sizes[0] * sizes[1] * sizes[2];
|
|
}
|
|
|
|
unsigned ac_compute_lshs_workgroup_size(enum amd_gfx_level gfx_level, gl_shader_stage stage,
|
|
unsigned tess_num_patches,
|
|
unsigned tess_patch_in_vtx,
|
|
unsigned tess_patch_out_vtx)
|
|
{
|
|
/* When tessellation is used, API VS runs on HW LS, API TCS runs on HW HS.
|
|
* These two HW stages are merged on GFX9+.
|
|
*/
|
|
|
|
bool merged_shaders = gfx_level >= GFX9;
|
|
unsigned ls_workgroup_size = tess_num_patches * tess_patch_in_vtx;
|
|
unsigned hs_workgroup_size = tess_num_patches * tess_patch_out_vtx;
|
|
|
|
if (merged_shaders)
|
|
return MAX2(ls_workgroup_size, hs_workgroup_size);
|
|
else if (stage == MESA_SHADER_VERTEX)
|
|
return ls_workgroup_size;
|
|
else if (stage == MESA_SHADER_TESS_CTRL)
|
|
return hs_workgroup_size;
|
|
else
|
|
unreachable("invalid LSHS shader stage");
|
|
}
|
|
|
|
unsigned ac_compute_esgs_workgroup_size(enum amd_gfx_level gfx_level, unsigned wave_size,
|
|
unsigned es_verts, unsigned gs_inst_prims)
|
|
{
|
|
/* ESGS may operate in workgroups if on-chip GS (LDS rings) are enabled.
|
|
*
|
|
* GFX6: Not possible in the HW.
|
|
* GFX7-8 (unmerged): possible in the HW, but not implemented in Mesa.
|
|
* GFX9+ (merged): implemented in Mesa.
|
|
*/
|
|
|
|
if (gfx_level <= GFX8)
|
|
return wave_size;
|
|
|
|
unsigned workgroup_size = MAX2(es_verts, gs_inst_prims);
|
|
return CLAMP(workgroup_size, 1, 256);
|
|
}
|
|
|
|
unsigned ac_compute_ngg_workgroup_size(unsigned es_verts, unsigned gs_inst_prims,
|
|
unsigned max_vtx_out, unsigned prim_amp_factor)
|
|
{
|
|
/* NGG always operates in workgroups.
|
|
*
|
|
* For API VS/TES/GS:
|
|
* - 1 invocation per input vertex
|
|
* - 1 invocation per input primitive
|
|
*
|
|
* The same invocation can process both an input vertex and primitive,
|
|
* however 1 invocation can only output up to 1 vertex and 1 primitive.
|
|
*/
|
|
|
|
unsigned max_vtx_in = es_verts < 256 ? es_verts : 3 * gs_inst_prims;
|
|
unsigned max_prim_in = gs_inst_prims;
|
|
unsigned max_prim_out = gs_inst_prims * prim_amp_factor;
|
|
unsigned workgroup_size = MAX4(max_vtx_in, max_vtx_out, max_prim_in, max_prim_out);
|
|
|
|
return CLAMP(workgroup_size, 1, 256);
|
|
}
|
|
|
|
uint32_t ac_compute_num_tess_patches(const struct radeon_info *info, uint32_t num_tcs_input_cp,
|
|
uint32_t num_tcs_output_cp, uint32_t vram_per_patch,
|
|
uint32_t lds_per_patch, uint32_t wave_size,
|
|
bool tess_uses_primid)
|
|
{
|
|
/* The VGT HS block increments the patch ID unconditionally
|
|
* within a single threadgroup. This results in incorrect
|
|
* patch IDs when instanced draws are used.
|
|
*
|
|
* The intended solution is to restrict threadgroups to
|
|
* a single instance by setting SWITCH_ON_EOI, which
|
|
* should cause IA to split instances up. However, this
|
|
* doesn't work correctly on GFX6 when there is no other
|
|
* SE to switch to.
|
|
*/
|
|
const bool has_primid_instancing_bug = info->gfx_level == GFX6 && info->max_se == 1;
|
|
if (has_primid_instancing_bug && tess_uses_primid)
|
|
return 1;
|
|
|
|
/* Ensure that we only need 4 waves per CU, so that we don't need to check
|
|
* resource usage (such as whether we have enough VGPRs to fit the whole
|
|
* threadgroup into the CU). It also ensures that the number of tcs in and out
|
|
* vertices per threadgroup are at most 256, which is the hw limit.
|
|
*/
|
|
const unsigned max_verts_per_patch = MAX2(num_tcs_input_cp, num_tcs_output_cp);
|
|
unsigned num_patches = 256 / max_verts_per_patch;
|
|
|
|
/* Not necessary for correctness, but higher numbers are slower.
|
|
* The hardware can do more, but we prefer fully occupied waves.
|
|
* eg. 64 triangle patches means 3 fully occupied Wave64 waves.
|
|
*/
|
|
num_patches = MIN2(num_patches, 64);
|
|
|
|
/* When distributed tessellation is unsupported, switch between SEs
|
|
* at a higher frequency to manually balance the workload between SEs.
|
|
*/
|
|
if (!info->has_distributed_tess && info->max_se > 1)
|
|
num_patches = MIN2(num_patches, 16); /* recommended */
|
|
|
|
/* Make sure the output data fits in the offchip buffer */
|
|
if (vram_per_patch) {
|
|
const uint32_t tess_offchip_block_dw_size = info->family == CHIP_HAWAII ? 4096 : 8192;
|
|
num_patches =
|
|
MIN2(num_patches, (tess_offchip_block_dw_size * 4) / vram_per_patch);
|
|
}
|
|
|
|
/* Make sure that the data fits in LDS. This assumes the shaders only
|
|
* use LDS for the inputs and outputs.
|
|
*/
|
|
if (lds_per_patch) {
|
|
ASSERTED const unsigned max_lds_size = info->gfx_level >= GFX9 ? 64 * 1024 : 32 * 1024; /* hw limit */
|
|
const unsigned target_lds_size = max_lds_size / 2; /* target at least 2 workgroups per CU */
|
|
num_patches = MIN2(num_patches, target_lds_size / lds_per_patch);
|
|
assert(num_patches * lds_per_patch <= max_lds_size);
|
|
}
|
|
num_patches = MAX2(num_patches, 1);
|
|
|
|
/* Make sure that vector lanes are fully occupied by cutting off the last wave
|
|
* if it's only partially filled.
|
|
*/
|
|
const unsigned temp_verts_per_tg = num_patches * max_verts_per_patch;
|
|
|
|
if (temp_verts_per_tg > wave_size &&
|
|
(wave_size - temp_verts_per_tg % wave_size >= MAX2(max_verts_per_patch, 8)))
|
|
num_patches = (temp_verts_per_tg & ~(wave_size - 1)) / max_verts_per_patch;
|
|
|
|
if (info->gfx_level == GFX6) {
|
|
/* GFX6 bug workaround, related to power management. Limit LS-HS
|
|
* threadgroups to only one wave.
|
|
*/
|
|
const unsigned one_wave = wave_size / max_verts_per_patch;
|
|
num_patches = MIN2(num_patches, one_wave);
|
|
}
|
|
|
|
return num_patches;
|
|
}
|
|
|
|
uint32_t
|
|
ac_compute_tess_lds_size(const struct radeon_info *info, uint32_t lds_per_patch, uint32_t num_patches)
|
|
{
|
|
const unsigned lds_size = lds_per_patch * num_patches;
|
|
|
|
assert(lds_size <= (info->gfx_level >= GFX9 ? 65536 : 32768));
|
|
|
|
return align(lds_size, info->lds_encode_granularity) / info->lds_encode_granularity;
|
|
}
|
|
|
|
uint32_t ac_apply_cu_en(uint32_t value, uint32_t clear_mask, unsigned value_shift,
|
|
const struct radeon_info *info)
|
|
{
|
|
/* Register field position and mask. */
|
|
uint32_t cu_en_mask = ~clear_mask;
|
|
unsigned cu_en_shift = ffs(cu_en_mask) - 1;
|
|
/* The value being set. */
|
|
uint32_t cu_en = (value & cu_en_mask) >> cu_en_shift;
|
|
|
|
uint32_t set_cu_en = info->spi_cu_en;
|
|
|
|
if (info->gfx_level >= GFX12 && clear_mask == 0) {
|
|
/* The CU mask has 32 bits and is per SE, not per SA. This math doesn't work with
|
|
* asymmetric WGP harvesting because SA0 doesn't always end on the same bit.
|
|
*/
|
|
set_cu_en &= BITFIELD_MASK(info->max_good_cu_per_sa);
|
|
set_cu_en |= set_cu_en << info->max_good_cu_per_sa;
|
|
}
|
|
|
|
/* AND the field by spi_cu_en. */
|
|
uint32_t spi_cu_en = info->spi_cu_en >> value_shift;
|
|
return (value & ~cu_en_mask) |
|
|
(((cu_en & spi_cu_en) << cu_en_shift) & cu_en_mask);
|
|
}
|
|
|
|
/* Return the register value and tune bytes_per_wave to increase scratch performance. */
|
|
void ac_get_scratch_tmpring_size(const struct radeon_info *info,
|
|
unsigned bytes_per_wave, unsigned *max_seen_bytes_per_wave,
|
|
uint32_t *tmpring_size)
|
|
{
|
|
/* SPI_TMPRING_SIZE and COMPUTE_TMPRING_SIZE are essentially scratch buffer descriptors.
|
|
* WAVES means NUM_RECORDS. WAVESIZE is the size of each element, meaning STRIDE.
|
|
* Thus, WAVESIZE must be constant while the scratch buffer is being used by the GPU.
|
|
*
|
|
* If you want to increase WAVESIZE without waiting for idle, you need to allocate a new
|
|
* scratch buffer and use it instead. This will result in multiple scratch buffers being
|
|
* used at the same time, each with a different WAVESIZE.
|
|
*
|
|
* If you want to decrease WAVESIZE, you don't have to. There is no advantage in decreasing
|
|
* WAVESIZE after it's been increased.
|
|
*
|
|
* Shaders with SCRATCH_EN=0 don't allocate scratch space.
|
|
*/
|
|
const unsigned size_shift = info->gfx_level >= GFX11 ? 8 : 10;
|
|
const unsigned min_size_per_wave = BITFIELD_BIT(size_shift);
|
|
|
|
/* The LLVM shader backend should be reporting aligned scratch_sizes. */
|
|
assert((bytes_per_wave & BITFIELD_MASK(size_shift)) == 0 &&
|
|
"scratch size per wave should be aligned");
|
|
|
|
/* Add 1 scratch item to make the number of items odd. This should improve scratch
|
|
* performance by more randomly distributing scratch waves among memory channels.
|
|
*/
|
|
if (bytes_per_wave)
|
|
bytes_per_wave |= min_size_per_wave;
|
|
|
|
*max_seen_bytes_per_wave = MAX2(*max_seen_bytes_per_wave, bytes_per_wave);
|
|
|
|
unsigned max_scratch_waves = info->max_scratch_waves;
|
|
if (info->gfx_level >= GFX11)
|
|
max_scratch_waves /= info->max_se; /* WAVES is per SE */
|
|
|
|
/* TODO: We could decrease WAVES to make the whole buffer fit into the infinity cache. */
|
|
*tmpring_size = S_0286E8_WAVES(max_scratch_waves) |
|
|
S_0286E8_WAVESIZE(*max_seen_bytes_per_wave >> size_shift);
|
|
}
|
|
|
|
/* Get chip-agnostic memory instruction access flags (as opposed to chip-specific GLC/DLC/SLC)
|
|
* from a NIR memory intrinsic.
|
|
*/
|
|
enum gl_access_qualifier ac_get_mem_access_flags(const nir_intrinsic_instr *instr)
|
|
{
|
|
enum gl_access_qualifier access =
|
|
nir_intrinsic_has_access(instr) ? nir_intrinsic_access(instr) : 0;
|
|
|
|
/* Determine ACCESS_MAY_STORE_SUBDWORD. (for the GFX6 TC L1 bug workaround) */
|
|
if (!nir_intrinsic_infos[instr->intrinsic].has_dest) {
|
|
switch (instr->intrinsic) {
|
|
case nir_intrinsic_bindless_image_store:
|
|
access |= ACCESS_MAY_STORE_SUBDWORD;
|
|
break;
|
|
|
|
case nir_intrinsic_store_ssbo:
|
|
case nir_intrinsic_store_buffer_amd:
|
|
case nir_intrinsic_store_global:
|
|
case nir_intrinsic_store_global_amd:
|
|
if (access & ACCESS_USES_FORMAT_AMD ||
|
|
(nir_intrinsic_has_align_offset(instr) && nir_intrinsic_align(instr) % 4 != 0) ||
|
|
((instr->src[0].ssa->bit_size / 8) * instr->src[0].ssa->num_components) % 4 != 0)
|
|
access |= ACCESS_MAY_STORE_SUBDWORD;
|
|
break;
|
|
|
|
default:
|
|
unreachable("unexpected store instruction");
|
|
}
|
|
}
|
|
|
|
return access;
|
|
}
|
|
|
|
/* Convert chip-agnostic memory access flags into hw-specific cache flags.
|
|
*
|
|
* "access" must be a result of ac_get_mem_access_flags() with the appropriate ACCESS_TYPE_*
|
|
* flags set.
|
|
*/
|
|
union ac_hw_cache_flags ac_get_hw_cache_flags(const struct radeon_info *info,
|
|
enum gl_access_qualifier access)
|
|
{
|
|
union ac_hw_cache_flags result;
|
|
result.value = 0;
|
|
|
|
assert(util_bitcount(access & (ACCESS_TYPE_LOAD | ACCESS_TYPE_STORE |
|
|
ACCESS_TYPE_ATOMIC)) == 1);
|
|
assert(!(access & ACCESS_TYPE_SMEM) || access & ACCESS_TYPE_LOAD);
|
|
assert(!(access & ACCESS_IS_SWIZZLED_AMD) || !(access & ACCESS_TYPE_SMEM));
|
|
assert(!(access & ACCESS_MAY_STORE_SUBDWORD) || access & ACCESS_TYPE_STORE);
|
|
|
|
bool scope_is_device = access & (ACCESS_COHERENT | ACCESS_VOLATILE);
|
|
|
|
if (info->gfx_level >= GFX12) {
|
|
if (access & ACCESS_CP_GE_COHERENT_AMD) {
|
|
result.gfx12.scope = info->cp_sdma_ge_use_system_memory_scope ?
|
|
gfx12_scope_memory : gfx12_scope_device;
|
|
} else if (scope_is_device) {
|
|
result.gfx12.scope = gfx12_scope_device;
|
|
} else {
|
|
result.gfx12.scope = gfx12_scope_cu;
|
|
}
|
|
|
|
if (access & ACCESS_NON_TEMPORAL) {
|
|
if (access & ACCESS_TYPE_LOAD) {
|
|
/* Don't use non_temporal for SMEM because it can't set regular_temporal for MALL. */
|
|
if (!(access & ACCESS_TYPE_SMEM))
|
|
result.gfx12.temporal_hint = gfx12_load_near_non_temporal_far_regular_temporal;
|
|
} else if (access & ACCESS_TYPE_STORE) {
|
|
result.gfx12.temporal_hint = gfx12_store_near_non_temporal_far_regular_temporal;
|
|
} else {
|
|
result.gfx12.temporal_hint = gfx12_atomic_non_temporal;
|
|
}
|
|
}
|
|
} else if (info->gfx_level >= GFX11) {
|
|
/* GFX11 simplified it and exposes what is actually useful.
|
|
*
|
|
* GLC means device scope for loads only. (stores and atomics are always device scope)
|
|
* SLC means non-temporal for GL1 and GL2 caches. (GL1 = hit-evict, GL2 = stream, unavailable in SMEM)
|
|
* DLC means non-temporal for MALL. (noalloc, i.e. coherent bypass)
|
|
*
|
|
* GL0 doesn't have a non-temporal flag, so you always get LRU caching in CU scope.
|
|
*/
|
|
if (access & ACCESS_TYPE_LOAD && scope_is_device)
|
|
result.value |= ac_glc;
|
|
|
|
if (access & ACCESS_NON_TEMPORAL && !(access & ACCESS_TYPE_SMEM))
|
|
result.value |= ac_slc;
|
|
} else if (info->gfx_level >= GFX10) {
|
|
/* GFX10-10.3:
|
|
*
|
|
* VMEM and SMEM loads (SMEM only supports the first four):
|
|
* !GLC && !DLC && !SLC means CU scope <== use for normal loads with CU scope
|
|
* GLC && !DLC && !SLC means SA scope
|
|
* !GLC && DLC && !SLC means CU scope, GL1 bypass
|
|
* GLC && DLC && !SLC means device scope <== use for normal loads with device scope
|
|
* !GLC && !DLC && SLC means CU scope, non-temporal (GL0 = GL1 = hit-evict, GL2 = stream) <== use for non-temporal loads with CU scope
|
|
* GLC && !DLC && SLC means SA scope, non-temporal (GL1 = hit-evict, GL2 = stream)
|
|
* !GLC && DLC && SLC means CU scope, GL0 non-temporal, GL1-GL2 coherent bypass (GL0 = hit-evict, GL1 = bypass, GL2 = noalloc)
|
|
* GLC && DLC && SLC means device scope, GL2 coherent bypass (noalloc) <== use for non-temporal loads with device scope
|
|
*
|
|
* VMEM stores/atomics (stores are CU scope only if they overwrite the whole cache line,
|
|
* atomics are always device scope, GL1 is always bypassed):
|
|
* !GLC && !DLC && !SLC means CU scope <== use for normal stores with CU scope
|
|
* GLC && !DLC && !SLC means device scope <== use for normal stores with device scope
|
|
* !GLC && DLC && !SLC means CU scope, GL2 non-coherent bypass
|
|
* GLC && DLC && !SLC means device scope, GL2 non-coherent bypass
|
|
* !GLC && !DLC && SLC means CU scope, GL2 non-temporal (stream) <== use for non-temporal stores with CU scope
|
|
* GLC && !DLC && SLC means device scope, GL2 non-temporal (stream) <== use for non-temporal stores with device scope
|
|
* !GLC && DLC && SLC means CU scope, GL2 coherent bypass (noalloc)
|
|
* GLC && DLC && SLC means device scope, GL2 coherent bypass (noalloc)
|
|
*
|
|
* "stream" allows write combining in GL2. "coherent bypass" doesn't.
|
|
* "non-coherent bypass" doesn't guarantee ordering with any coherent stores.
|
|
*/
|
|
if (scope_is_device && !(access & ACCESS_TYPE_ATOMIC))
|
|
result.value |= ac_glc | (access & ACCESS_TYPE_LOAD ? ac_dlc : 0);
|
|
|
|
if (access & ACCESS_NON_TEMPORAL && !(access & ACCESS_TYPE_SMEM))
|
|
result.value |= ac_slc;
|
|
} else {
|
|
/* GFX6-GFX9:
|
|
*
|
|
* VMEM loads:
|
|
* !GLC && !SLC means CU scope
|
|
* GLC && !SLC means (GFX6: device scope, GFX7-9: device scope [*])
|
|
* !GLC && SLC means (GFX6: CU scope, GFX7: device scope, GFX8-9: CU scope), GL2 non-temporal (stream)
|
|
* GLC && SLC means device scope, GL2 non-temporal (stream)
|
|
*
|
|
* VMEM stores (atomics don't have [*]):
|
|
* !GLC && !SLC means (GFX6: CU scope, GFX7-9: device scope [*])
|
|
* GLC && !SLC means (GFX6-7: device scope, GFX8-9: device scope [*])
|
|
* !GLC && SLC means (GFX6: CU scope, GFX7-9: device scope [*]), GL2 non-temporal (stream)
|
|
* GLC && SLC means device scope, GL2 non-temporal (stream)
|
|
*
|
|
* [*] data can be cached in GL1 for future CU scope
|
|
*
|
|
* SMEM loads:
|
|
* GLC means device scope (available on GFX8+)
|
|
*/
|
|
if (scope_is_device && !(access & ACCESS_TYPE_ATOMIC)) {
|
|
/* SMEM doesn't support the device scope on GFX6-7. */
|
|
assert(info->gfx_level >= GFX8 || !(access & ACCESS_TYPE_SMEM));
|
|
result.value |= ac_glc;
|
|
}
|
|
|
|
if (access & ACCESS_NON_TEMPORAL && !(access & ACCESS_TYPE_SMEM))
|
|
result.value |= ac_slc;
|
|
|
|
/* GFX6 has a TC L1 bug causing corruption of 8bit/16bit stores. All store opcodes not
|
|
* aligned to a dword are affected.
|
|
*/
|
|
if (info->gfx_level == GFX6 && access & ACCESS_MAY_STORE_SUBDWORD)
|
|
result.value |= ac_glc;
|
|
}
|
|
|
|
if (access & ACCESS_IS_SWIZZLED_AMD) {
|
|
if (info->gfx_level >= GFX12)
|
|
result.gfx12.swizzled = true;
|
|
else
|
|
result.value |= ac_swizzled;
|
|
}
|
|
|
|
return result;
|
|
}
|
|
|
|
unsigned ac_get_all_edge_flag_bits(enum amd_gfx_level gfx_level)
|
|
{
|
|
return gfx_level >= GFX12 ?
|
|
((1u << 8) | (1u << 17) | (1u << 26)) :
|
|
((1u << 9) | (1u << 19) | (1u << 29));
|
|
}
|
|
|
|
/**
|
|
* Returns a unique index for a per-patch semantic name and index. The index
|
|
* must be less than 32, so that a 32-bit bitmask of used inputs or outputs
|
|
* can be calculated.
|
|
*/
|
|
unsigned
|
|
ac_shader_io_get_unique_index_patch(unsigned semantic)
|
|
{
|
|
switch (semantic) {
|
|
case VARYING_SLOT_TESS_LEVEL_OUTER:
|
|
return 0;
|
|
case VARYING_SLOT_TESS_LEVEL_INNER:
|
|
return 1;
|
|
default:
|
|
if (semantic >= VARYING_SLOT_PATCH0 && semantic < VARYING_SLOT_PATCH0 + 30)
|
|
return 2 + (semantic - VARYING_SLOT_PATCH0);
|
|
|
|
assert(!"invalid semantic");
|
|
return 0;
|
|
}
|
|
}
|