/* * Copyright (C) 2008 VMware, Inc. * Copyright (C) 2014 Broadcom * Copyright (C) 2018-2019 Alyssa Rosenzweig * Copyright (C) 2019-2020 Collabora, Ltd. * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the "Software"), * to deal in the Software without restriction, including without limitation * the rights to use, copy, modify, merge, publish, distribute, sublicense, * and/or sell copies of the Software, and to permit persons to whom the * Software is furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice (including the next * paragraph) shall be included in all copies or substantial portions of the * Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE * SOFTWARE. * */ #include "pan_texture.h" #include "util/macros.h" #include "util/u_math.h" #if PAN_ARCH >= 5 /* * Arm Scalable Texture Compression (ASTC) corresponds to just a few formats. * The block dimension is not part of the format. Instead, it is encoded as a * 6-bit tag on the payload pointer. Map the block size for a single dimension. */ static inline enum mali_astc_2d_dimension panfrost_astc_dim_2d(unsigned dim) { switch (dim) { case 4: return MALI_ASTC_2D_DIMENSION_4; case 5: return MALI_ASTC_2D_DIMENSION_5; case 6: return MALI_ASTC_2D_DIMENSION_6; case 8: return MALI_ASTC_2D_DIMENSION_8; case 10: return MALI_ASTC_2D_DIMENSION_10; case 12: return MALI_ASTC_2D_DIMENSION_12; default: unreachable("Invalid ASTC dimension"); } } static inline enum mali_astc_3d_dimension panfrost_astc_dim_3d(unsigned dim) { switch (dim) { case 3: return MALI_ASTC_3D_DIMENSION_3; case 4: return MALI_ASTC_3D_DIMENSION_4; case 5: return MALI_ASTC_3D_DIMENSION_5; case 6: return MALI_ASTC_3D_DIMENSION_6; default: unreachable("Invalid ASTC dimension"); } } #endif /* Texture addresses are tagged with information about compressed formats. * AFBC uses a bit for whether the colorspace transform is enabled (RGB and * RGBA only). * For ASTC, this is a "stretch factor" encoding the block size. */ static unsigned panfrost_compression_tag(const struct util_format_description *desc, enum mali_texture_dimension dim, uint64_t modifier) { #if PAN_ARCH >= 5 && PAN_ARCH <= 8 if (drm_is_afbc(modifier)) { unsigned flags = (modifier & AFBC_FORMAT_MOD_YTR) ? MALI_AFBC_SURFACE_FLAG_YTR : 0; #if PAN_ARCH >= 6 /* Prefetch enable */ flags |= MALI_AFBC_SURFACE_FLAG_PREFETCH; if (panfrost_afbc_is_wide(modifier)) flags |= MALI_AFBC_SURFACE_FLAG_WIDE_BLOCK; #endif #if PAN_ARCH >= 7 /* Tiled headers */ if (modifier & AFBC_FORMAT_MOD_TILED) flags |= MALI_AFBC_SURFACE_FLAG_TILED_HEADER; /* Used to make sure AFBC headers don't point outside the AFBC * body. HW is using the AFBC surface stride to do this check, * which doesn't work for 3D textures because the surface * stride does not cover the body. Only supported on v7+. */ if (dim != MALI_TEXTURE_DIMENSION_3D) flags |= MALI_AFBC_SURFACE_FLAG_CHECK_PAYLOAD_RANGE; #endif return flags; } else if (desc->layout == UTIL_FORMAT_LAYOUT_ASTC) { if (desc->block.depth > 1) { return (panfrost_astc_dim_3d(desc->block.depth) << 4) | (panfrost_astc_dim_3d(desc->block.height) << 2) | panfrost_astc_dim_3d(desc->block.width); } else { return (panfrost_astc_dim_2d(desc->block.height) << 3) | panfrost_astc_dim_2d(desc->block.width); } } #endif /* Tags are not otherwise used */ return 0; } /* Following the texture descriptor is a number of descriptors. How many? */ static unsigned panfrost_texture_num_elements(const struct pan_image_view *iview) { unsigned levels = 1 + iview->last_level - iview->first_level; unsigned layers = 1 + iview->last_layer - iview->first_layer; unsigned nr_samples = pan_image_view_get_nr_samples(iview); return levels * layers * MAX2(nr_samples, 1); } /* Conservative estimate of the size of the texture payload a priori. * Average case, size equal to the actual size. Worst case, off by 2x (if * a manual stride is not needed on a linear texture). Returned value * must be greater than or equal to the actual size, so it's safe to use * as an allocation amount */ unsigned GENX(panfrost_estimate_texture_payload_size)(const struct pan_image_view *iview) { size_t element_size; #if PAN_ARCH >= 9 element_size = pan_size(PLANE); /* 2-plane and 3-plane YUV use two plane descriptors. */ if (panfrost_format_is_yuv(iview->format) && iview->planes[1] != NULL) element_size *= 2; #elif PAN_ARCH == 7 if (panfrost_format_is_yuv(iview->format)) element_size = pan_size(MULTIPLANAR_SURFACE); else element_size = pan_size(SURFACE_WITH_STRIDE); #else /* Assume worst case. Overestimates on Midgard, but that's ok. */ element_size = pan_size(SURFACE_WITH_STRIDE); #endif unsigned elements = panfrost_texture_num_elements(iview); return element_size * elements; } static void panfrost_get_surface_strides(const struct pan_image_layout *layout, unsigned l, int32_t *row_stride, int32_t *surf_stride) { const struct pan_image_slice_layout *slice = &layout->slices[l]; if (drm_is_afbc(layout->modifier)) { /* Pre v7 don't have a row stride field. This field is * repurposed as a Y offset which we don't use */ *row_stride = PAN_ARCH < 7 ? 0 : slice->row_stride; *surf_stride = slice->afbc.surface_stride; } else { *row_stride = slice->row_stride; *surf_stride = slice->surface_stride; } } static mali_ptr panfrost_get_surface_pointer(const struct pan_image_layout *layout, enum mali_texture_dimension dim, mali_ptr base, unsigned l, unsigned i, unsigned s) { unsigned offset; if (layout->dim == MALI_TEXTURE_DIMENSION_3D) { assert(!s); offset = layout->slices[l].offset + i * panfrost_get_layer_stride(layout, l); } else { offset = panfrost_texture_offset(layout, l, i, s); } return base + offset; } #if PAN_ARCH <= 7 static void panfrost_emit_surface_with_stride(mali_ptr plane, int32_t row_stride, int32_t surface_stride, void **payload) { pan_pack(*payload, SURFACE_WITH_STRIDE, cfg) { cfg.pointer = plane; cfg.row_stride = row_stride; cfg.surface_stride = surface_stride; } *payload += pan_size(SURFACE_WITH_STRIDE); } #endif #if PAN_ARCH == 7 static void panfrost_emit_multiplanar_surface(mali_ptr planes[MAX_IMAGE_PLANES], int32_t row_strides[MAX_IMAGE_PLANES], void **payload) { assert(row_strides[2] == 0 || row_strides[1] == row_strides[2]); pan_pack(*payload, MULTIPLANAR_SURFACE, cfg) { cfg.plane_0_pointer = planes[0]; cfg.plane_0_row_stride = row_strides[0]; cfg.plane_1_2_row_stride = row_strides[1]; cfg.plane_1_pointer = planes[1]; cfg.plane_2_pointer = planes[2]; } *payload += pan_size(MULTIPLANAR_SURFACE); } #endif #if PAN_ARCH >= 9 /* clang-format off */ #define CLUMP_FMT(pipe, mali) [PIPE_FORMAT_ ## pipe] = MALI_CLUMP_FORMAT_ ## mali static enum mali_clump_format special_clump_formats[PIPE_FORMAT_COUNT] = { CLUMP_FMT(X32_S8X24_UINT, X32S8X24), CLUMP_FMT(X24S8_UINT, X24S8), CLUMP_FMT(S8X24_UINT, S8X24), CLUMP_FMT(S8_UINT, S8), CLUMP_FMT(L4A4_UNORM, L4A4), CLUMP_FMT(L8A8_UNORM, L8A8), CLUMP_FMT(L8A8_UINT, L8A8), CLUMP_FMT(L8A8_SINT, L8A8), CLUMP_FMT(A8_UNORM, A8), CLUMP_FMT(A8_UINT, A8), CLUMP_FMT(A8_SINT, A8), CLUMP_FMT(ETC1_RGB8, ETC2_RGB8), CLUMP_FMT(ETC2_RGB8, ETC2_RGB8), CLUMP_FMT(ETC2_SRGB8, ETC2_RGB8), CLUMP_FMT(ETC2_RGB8A1, ETC2_RGB8A1), CLUMP_FMT(ETC2_SRGB8A1, ETC2_RGB8A1), CLUMP_FMT(ETC2_RGBA8, ETC2_RGBA8), CLUMP_FMT(ETC2_SRGBA8, ETC2_RGBA8), CLUMP_FMT(ETC2_R11_UNORM, ETC2_R11_UNORM), CLUMP_FMT(ETC2_R11_SNORM, ETC2_R11_SNORM), CLUMP_FMT(ETC2_RG11_UNORM, ETC2_RG11_UNORM), CLUMP_FMT(ETC2_RG11_SNORM, ETC2_RG11_SNORM), CLUMP_FMT(DXT1_RGB, BC1_UNORM), CLUMP_FMT(DXT1_RGBA, BC1_UNORM), CLUMP_FMT(DXT1_SRGB, BC1_UNORM), CLUMP_FMT(DXT1_SRGBA, BC1_UNORM), CLUMP_FMT(DXT3_RGBA, BC2_UNORM), CLUMP_FMT(DXT3_SRGBA, BC2_UNORM), CLUMP_FMT(DXT5_RGBA, BC3_UNORM), CLUMP_FMT(DXT5_SRGBA, BC3_UNORM), CLUMP_FMT(RGTC1_UNORM, BC4_UNORM), CLUMP_FMT(RGTC1_SNORM, BC4_SNORM), CLUMP_FMT(RGTC2_UNORM, BC5_UNORM), CLUMP_FMT(RGTC2_SNORM, BC5_SNORM), CLUMP_FMT(BPTC_RGB_FLOAT, BC6H_SF16), CLUMP_FMT(BPTC_RGB_UFLOAT, BC6H_UF16), CLUMP_FMT(BPTC_RGBA_UNORM, BC7_UNORM), CLUMP_FMT(BPTC_SRGBA, BC7_UNORM), }; #undef CLUMP_FMT /* clang-format on */ static enum mali_clump_format panfrost_clump_format(enum pipe_format format) { /* First, try a special clump format. Note that the 0 encoding is for a * raw clump format, which will never be in the special table. */ if (special_clump_formats[format]) return special_clump_formats[format]; /* Else, it's a raw format. Raw formats must not be compressed. */ assert(!util_format_is_compressed(format)); /* YUV-sampling has special cases */ if (panfrost_format_is_yuv(format)) { switch (format) { case PIPE_FORMAT_R8G8_R8B8_UNORM: case PIPE_FORMAT_G8R8_B8R8_UNORM: case PIPE_FORMAT_R8B8_R8G8_UNORM: case PIPE_FORMAT_B8R8_G8R8_UNORM: return MALI_CLUMP_FORMAT_Y8_UV8_422; case PIPE_FORMAT_R8_G8B8_420_UNORM: case PIPE_FORMAT_R8_B8G8_420_UNORM: case PIPE_FORMAT_R8_G8_B8_420_UNORM: case PIPE_FORMAT_R8_B8_G8_420_UNORM: return MALI_CLUMP_FORMAT_Y8_UV8_420; default: unreachable("unhandled clump format"); } } /* Select the appropriate raw format. */ switch (util_format_get_blocksize(format)) { case 1: return MALI_CLUMP_FORMAT_RAW8; case 2: return MALI_CLUMP_FORMAT_RAW16; case 3: return MALI_CLUMP_FORMAT_RAW24; case 4: return MALI_CLUMP_FORMAT_RAW32; case 6: return MALI_CLUMP_FORMAT_RAW48; case 8: return MALI_CLUMP_FORMAT_RAW64; case 12: return MALI_CLUMP_FORMAT_RAW96; case 16: return MALI_CLUMP_FORMAT_RAW128; default: unreachable("Invalid bpp"); } } static enum mali_afbc_superblock_size translate_superblock_size(uint64_t modifier) { assert(drm_is_afbc(modifier)); switch (modifier & AFBC_FORMAT_MOD_BLOCK_SIZE_MASK) { case AFBC_FORMAT_MOD_BLOCK_SIZE_16x16: return MALI_AFBC_SUPERBLOCK_SIZE_16X16; case AFBC_FORMAT_MOD_BLOCK_SIZE_32x8: return MALI_AFBC_SUPERBLOCK_SIZE_32X8; case AFBC_FORMAT_MOD_BLOCK_SIZE_64x4: return MALI_AFBC_SUPERBLOCK_SIZE_64X4; default: unreachable("Invalid superblock size"); } } static void panfrost_emit_plane(const struct pan_image_layout *layout, enum pipe_format format, mali_ptr pointer, unsigned level, int32_t row_stride, int32_t surface_stride, mali_ptr plane2_ptr, void **payload) { const struct util_format_description *desc = util_format_description(layout->format); assert(row_stride >= 0 && surface_stride >= 0 && "negative stride"); bool afbc = drm_is_afbc(layout->modifier); // TODO: this isn't technically guaranteed to be YUV, but it is in practice. bool is_3_planar_yuv = desc->layout == UTIL_FORMAT_LAYOUT_PLANAR3; pan_pack(*payload, PLANE, cfg) { cfg.pointer = pointer; cfg.row_stride = row_stride; cfg.size = layout->data_size - layout->slices[level].offset; if (is_3_planar_yuv) { cfg.two_plane_yuv_chroma.secondary_pointer = plane2_ptr; } else if (!panfrost_format_is_yuv(layout->format)) { cfg.slice_stride = layout->nr_samples ? surface_stride : panfrost_get_layer_stride(layout, level); } if (desc->layout == UTIL_FORMAT_LAYOUT_ASTC) { assert(!afbc); if (desc->block.depth > 1) { cfg.plane_type = MALI_PLANE_TYPE_ASTC_3D; cfg.astc._3d.block_width = panfrost_astc_dim_3d(desc->block.width); cfg.astc._3d.block_height = panfrost_astc_dim_3d(desc->block.height); cfg.astc._3d.block_depth = panfrost_astc_dim_3d(desc->block.depth); } else { cfg.plane_type = MALI_PLANE_TYPE_ASTC_2D; cfg.astc._2d.block_width = panfrost_astc_dim_2d(desc->block.width); cfg.astc._2d.block_height = panfrost_astc_dim_2d(desc->block.height); } bool srgb = (desc->colorspace == UTIL_FORMAT_COLORSPACE_SRGB); /* Mesa does not advertise _HDR formats yet */ cfg.astc.decode_hdr = false; /* sRGB formats decode to RGBA8 sRGB, which is narrow. * * Non-sRGB formats decode to RGBA16F which is wide. * With a future extension, we could decode non-sRGB * formats narrowly too, but this isn't wired up in Mesa * yet. */ cfg.astc.decode_wide = !srgb; } else if (afbc) { cfg.plane_type = MALI_PLANE_TYPE_AFBC; cfg.afbc.superblock_size = translate_superblock_size(layout->modifier); cfg.afbc.ytr = (layout->modifier & AFBC_FORMAT_MOD_YTR); cfg.afbc.tiled_header = (layout->modifier & AFBC_FORMAT_MOD_TILED); cfg.afbc.prefetch = true; cfg.afbc.compression_mode = GENX(pan_afbc_compression_mode)(format); cfg.afbc.header_stride = layout->slices[level].afbc.header_size; } else { cfg.plane_type = is_3_planar_yuv ? MALI_PLANE_TYPE_CHROMA_2P : MALI_PLANE_TYPE_GENERIC; cfg.clump_format = panfrost_clump_format(format); } if (!afbc && layout->modifier == DRM_FORMAT_MOD_ARM_16X16_BLOCK_U_INTERLEAVED) cfg.clump_ordering = MALI_CLUMP_ORDERING_TILED_U_INTERLEAVED; else if (!afbc) cfg.clump_ordering = MALI_CLUMP_ORDERING_LINEAR; } *payload += pan_size(PLANE); } #endif static void panfrost_emit_surface(const struct pan_image_view *iview, unsigned level, unsigned index, unsigned sample, enum pipe_format format, void **payload) { ASSERTED const struct util_format_description *desc = util_format_description(format); const struct pan_image_layout *layouts[MAX_IMAGE_PLANES] = {0}; mali_ptr plane_ptrs[MAX_IMAGE_PLANES] = {0}; int32_t row_strides[MAX_IMAGE_PLANES] = {0}; int32_t surface_strides[MAX_IMAGE_PLANES] = {0}; for (int i = 0; i < MAX_IMAGE_PLANES; i++) { const struct pan_image *base_image = pan_image_view_get_plane(iview, i); if (!base_image) { /* Every texture should have at least one plane. */ assert(i > 0); break; } mali_ptr base = base_image->data.base + base_image->data.offset; if (iview->buf.size) { assert(iview->dim == MALI_TEXTURE_DIMENSION_1D); base += iview->buf.offset; } layouts[i] = &pan_image_view_get_plane(iview, i)->layout; /* v4 does not support compression */ assert(PAN_ARCH >= 5 || !drm_is_afbc(layouts[i]->modifier)); assert(PAN_ARCH >= 5 || desc->layout != UTIL_FORMAT_LAYOUT_ASTC); /* panfrost_compression_tag() wants the dimension of the resource, not the * one of the image view (those might differ). */ unsigned tag = panfrost_compression_tag(desc, layouts[i]->dim, layouts[i]->modifier); plane_ptrs[i] = panfrost_get_surface_pointer( layouts[i], iview->dim, base | tag, level, index, sample); panfrost_get_surface_strides(layouts[i], level, &row_strides[i], &surface_strides[i]); } #if PAN_ARCH >= 9 if (panfrost_format_is_yuv(format)) { for (int i = 0; i < MAX_IMAGE_PLANES; i++) { /* 3-plane YUV is submitted using two PLANE descriptors, where the * second one is of type CHROMA_2P */ if (i > 1) break; if (plane_ptrs[i] == 0) break; /* 3-plane YUV requires equal stride for both chroma planes */ assert(row_strides[2] == 0 || row_strides[1] == row_strides[2]); panfrost_emit_plane(layouts[i], format, plane_ptrs[i], level, row_strides[i], surface_strides[i], plane_ptrs[2], payload); } } else { panfrost_emit_plane(layouts[0], format, plane_ptrs[0], level, row_strides[0], surface_strides[0], 0, payload); } return; #endif #if PAN_ARCH <= 7 #if PAN_ARCH == 7 if (panfrost_format_is_yuv(format)) { panfrost_emit_multiplanar_surface(plane_ptrs, row_strides, payload); return; } #endif panfrost_emit_surface_with_stride(plane_ptrs[0], row_strides[0], surface_strides[0], payload); #endif } static void panfrost_emit_texture_payload(const struct pan_image_view *iview, enum pipe_format format, void *payload) { unsigned nr_samples = PAN_ARCH <= 7 ? pan_image_view_get_nr_samples(iview) : 1; /* Inject the addresses in, interleaving array indices, mip levels, * cube faces, and strides in that order. On Bifrost and older, each * sample had its own surface descriptor; on Valhall, they are fused * into a single plane descriptor. */ #if PAN_ARCH >= 7 /* V7 and later treats faces as extra layers */ for (int layer = iview->first_layer; layer <= iview->last_layer; ++layer) { for (int sample = 0; sample < nr_samples; ++sample) { for (int level = iview->first_level; level <= iview->last_level; ++level) { panfrost_emit_surface(iview, level, layer, sample, format, &payload); } } } #else unsigned first_layer = iview->first_layer, last_layer = iview->last_layer; unsigned face_count = 1; if (iview->dim == MALI_TEXTURE_DIMENSION_CUBE) { first_layer /= 6; last_layer /= 6; face_count = 6; } /* V6 and earlier has a different memory-layout */ for (int layer = first_layer; layer <= last_layer; ++layer) { for (int level = iview->first_level; level <= iview->last_level; ++level) { /* order of face and sample doesn't matter; we can only have multiple * of one or the other (no support for multisampled cubemaps) */ for (int face = 0; face < face_count; ++face) { for (int sample = 0; sample < nr_samples; ++sample) { panfrost_emit_surface(iview, level, (face_count * layer) + face, sample, format, &payload); } } } } #endif } #if PAN_ARCH <= 7 /* Map modifiers to mali_texture_layout for packing in a texture descriptor */ static enum mali_texture_layout panfrost_modifier_to_layout(uint64_t modifier) { if (drm_is_afbc(modifier)) return MALI_TEXTURE_LAYOUT_AFBC; else if (modifier == DRM_FORMAT_MOD_ARM_16X16_BLOCK_U_INTERLEAVED) return MALI_TEXTURE_LAYOUT_TILED; else if (modifier == DRM_FORMAT_MOD_LINEAR) return MALI_TEXTURE_LAYOUT_LINEAR; else unreachable("Invalid modifer"); } #endif /* * Generates a texture descriptor. Ideally, descriptors are immutable after the * texture is created, so we can keep these hanging around in GPU memory in a * dedicated BO and not have to worry. In practice there are some minor gotchas * with this (the driver sometimes will change the format of a texture on the * fly for compression) but it's fast enough to just regenerate the descriptor * in those cases, rather than monkeypatching at drawtime. A texture descriptor * consists of a 32-byte header followed by pointers. */ void GENX(panfrost_new_texture)(const struct pan_image_view *iview, void *out, const struct panfrost_ptr *payload) { const struct pan_image *base_image = pan_image_view_get_plane(iview, 0); const struct pan_image_layout *layout = &base_image->layout; enum pipe_format format = iview->format; uint32_t mali_format = GENX(panfrost_format_from_pipe_format)(format)->hw; unsigned char swizzle[4]; if (PAN_ARCH >= 7 && util_format_is_depth_or_stencil(format)) { /* v7+ doesn't have an _RRRR component order, combine the * user swizzle with a .XXXX swizzle to emulate that. */ static const unsigned char replicate_x[4] = { PIPE_SWIZZLE_X, PIPE_SWIZZLE_X, PIPE_SWIZZLE_X, PIPE_SWIZZLE_X, }; util_format_compose_swizzles(replicate_x, iview->swizzle, swizzle); } else if (PAN_ARCH == 7 && !panfrost_format_is_yuv(format)) { #if PAN_ARCH == 7 /* v7 (only) restricts component orders when AFBC is in use. * Rather than restrict AFBC, we use an allowed component order * with an invertible swizzle composed. */ enum mali_rgb_component_order orig = mali_format & BITFIELD_MASK(12); struct pan_decomposed_swizzle decomposed = GENX(pan_decompose_swizzle)(orig); /* Apply the new component order */ mali_format = (mali_format & ~orig) | decomposed.pre; /* Compose the new swizzle */ util_format_compose_swizzles(decomposed.post, iview->swizzle, swizzle); #endif } else { STATIC_ASSERT(sizeof(swizzle) == sizeof(iview->swizzle)); memcpy(swizzle, iview->swizzle, sizeof(swizzle)); } panfrost_emit_texture_payload(iview, format, payload->cpu); unsigned array_size = iview->last_layer - iview->first_layer + 1; /* If this is a cubemap, we expect the number of layers to be a multiple * of 6. */ if (iview->dim == MALI_TEXTURE_DIMENSION_CUBE) { assert(array_size % 6 == 0); array_size /= 6; } /* Multiplanar YUV textures require 2 surface descriptors. */ if (panfrost_format_is_yuv(iview->format) && PAN_ARCH >= 9 && pan_image_view_get_plane(iview, 1) != NULL) array_size *= 2; unsigned width; if (iview->buf.size) { assert(iview->dim == MALI_TEXTURE_DIMENSION_1D); assert(!iview->first_level && !iview->last_level); assert(!iview->first_layer && !iview->last_layer); assert(layout->nr_samples == 1); assert(layout->height == 1 && layout->depth == 1); assert(iview->buf.offset + iview->buf.size <= layout->width); width = iview->buf.size; } else { width = u_minify(layout->width, iview->first_level); } pan_pack(out, TEXTURE, cfg) { cfg.dimension = iview->dim; cfg.format = mali_format; cfg.width = width; cfg.height = u_minify(layout->height, iview->first_level); if (iview->dim == MALI_TEXTURE_DIMENSION_3D) cfg.depth = u_minify(layout->depth, iview->first_level); else cfg.sample_count = layout->nr_samples; cfg.swizzle = panfrost_translate_swizzle_4(swizzle); #if PAN_ARCH >= 9 cfg.texel_interleave = (layout->modifier != DRM_FORMAT_MOD_LINEAR) || util_format_is_compressed(format); #else cfg.texel_ordering = panfrost_modifier_to_layout(layout->modifier); #endif cfg.levels = iview->last_level - iview->first_level + 1; cfg.array_size = array_size; #if PAN_ARCH >= 6 cfg.surfaces = payload->gpu; /* We specify API-level LOD clamps in the sampler descriptor * and use these clamps simply for bounds checking. */ cfg.minimum_lod = 0; cfg.maximum_lod = cfg.levels - 1; #endif } } #if PAN_ARCH >= 9 enum mali_afbc_compression_mode GENX(pan_afbc_compression_mode)(enum pipe_format format) { /* There's a special case for texturing the stencil part from a combined * depth/stencil texture, handle it separately. */ if (format == PIPE_FORMAT_X24S8_UINT) return MALI_AFBC_COMPRESSION_MODE_X24S8; /* Otherwise, map canonical formats to the hardware enum. This only * needs to handle the subset of formats returned by * panfrost_afbc_format. */ /* clang-format off */ switch (panfrost_afbc_format(PAN_ARCH, format)) { case PAN_AFBC_MODE_R8: return MALI_AFBC_COMPRESSION_MODE_R8; case PAN_AFBC_MODE_R8G8: return MALI_AFBC_COMPRESSION_MODE_R8G8; case PAN_AFBC_MODE_R5G6B5: return MALI_AFBC_COMPRESSION_MODE_R5G6B5; case PAN_AFBC_MODE_R4G4B4A4: return MALI_AFBC_COMPRESSION_MODE_R4G4B4A4; case PAN_AFBC_MODE_R5G5B5A1: return MALI_AFBC_COMPRESSION_MODE_R5G5B5A1; case PAN_AFBC_MODE_R8G8B8: return MALI_AFBC_COMPRESSION_MODE_R8G8B8; case PAN_AFBC_MODE_R8G8B8A8: return MALI_AFBC_COMPRESSION_MODE_R8G8B8A8; case PAN_AFBC_MODE_R10G10B10A2: return MALI_AFBC_COMPRESSION_MODE_R10G10B10A2; case PAN_AFBC_MODE_R11G11B10: return MALI_AFBC_COMPRESSION_MODE_R11G11B10; case PAN_AFBC_MODE_S8: return MALI_AFBC_COMPRESSION_MODE_S8; case PAN_AFBC_MODE_INVALID: unreachable("Invalid AFBC format"); } /* clang-format on */ unreachable("all AFBC formats handled"); } #endif