/* * Copyright 2018 Collabora Ltd. * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the "Software"), * to deal in the Software without restriction, including without limitation * on the rights to use, copy, modify, merge, publish, distribute, sub * license, and/or sell copies of the Software, and to permit persons to whom * the Software is furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice (including the next * paragraph) shall be included in all copies or substantial portions of the * Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL * THE AUTHOR(S) AND/OR THEIR SUPPLIERS BE LIABLE FOR ANY CLAIM, * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE * USE OR OTHER DEALINGS IN THE SOFTWARE. */ #include "zink_screen.h" #include "zink_kopper.h" #include "zink_compiler.h" #include "zink_context.h" #include "zink_descriptors.h" #include "zink_fence.h" #include "vk_format.h" #include "zink_format.h" #include "zink_framebuffer.h" #include "zink_program.h" #include "zink_public.h" #include "zink_query.h" #include "zink_resource.h" #include "zink_state.h" #include "nir_to_spirv/nir_to_spirv.h" // for SPIRV_VERSION #include "util/u_debug.h" #include "util/u_dl.h" #include "util/os_file.h" #include "util/u_memory.h" #include "util/u_screen.h" #include "util/u_string.h" #include "util/perf/u_trace.h" #include "util/u_transfer_helper.h" #include "util/hex.h" #include "util/xmlconfig.h" #include "util/u_cpu_detect.h" #ifdef HAVE_LIBDRM #include #include #include #ifdef MAJOR_IN_MKDEV #include #endif #ifdef MAJOR_IN_SYSMACROS #include #endif #endif static int num_screens = 0; bool zink_tracing = false; #if DETECT_OS_WINDOWS #include #define VK_LIBNAME "vulkan-1.dll" #else #include #if DETECT_OS_APPLE #define VK_LIBNAME "libvulkan.1.dylib" #elif DETECT_OS_ANDROID #define VK_LIBNAME "libvulkan.so" #else #define VK_LIBNAME "libvulkan.so.1" #endif #endif #ifdef __APPLE__ #include "MoltenVK/mvk_vulkan.h" // Source of MVK_VERSION #include "MoltenVK/mvk_config.h" #define VK_NO_PROTOTYPES #include "MoltenVK/mvk_deprecated_api.h" #include "MoltenVK/mvk_private_api.h" #endif /* __APPLE__ */ #ifdef HAVE_LIBDRM #include "drm-uapi/dma-buf.h" #include #endif static const struct debug_named_value zink_debug_options[] = { { "nir", ZINK_DEBUG_NIR, "Dump NIR during program compile" }, { "spirv", ZINK_DEBUG_SPIRV, "Dump SPIR-V during program compile" }, { "tgsi", ZINK_DEBUG_TGSI, "Dump TGSI during program compile" }, { "validation", ZINK_DEBUG_VALIDATION, "Dump Validation layer output" }, { "vvl", ZINK_DEBUG_VALIDATION, "Dump Validation layer output" }, { "sync", ZINK_DEBUG_SYNC, "Force synchronization before draws/dispatches" }, { "compact", ZINK_DEBUG_COMPACT, "Use only 4 descriptor sets" }, { "noreorder", ZINK_DEBUG_NOREORDER, "Do not reorder command streams" }, { "gpl", ZINK_DEBUG_GPL, "Force using Graphics Pipeline Library for all shaders" }, { "shaderdb", ZINK_DEBUG_SHADERDB, "Do stuff to make shader-db work" }, { "rp", ZINK_DEBUG_RP, "Enable renderpass tracking/optimizations" }, { "norp", ZINK_DEBUG_NORP, "Disable renderpass tracking/optimizations" }, { "map", ZINK_DEBUG_MAP, "Track amount of mapped VRAM" }, { "flushsync", ZINK_DEBUG_FLUSHSYNC, "Force synchronous flushes/presents" }, { "noshobj", ZINK_DEBUG_NOSHOBJ, "Disable EXT_shader_object" }, { "optimal_keys", ZINK_DEBUG_OPTIMAL_KEYS, "Debug/use optimal_keys" }, { "noopt", ZINK_DEBUG_NOOPT, "Disable async optimized pipeline compiles" }, { "nobgc", ZINK_DEBUG_NOBGC, "Disable all async pipeline compiles" }, { "dgc", ZINK_DEBUG_DGC, "Use DGC (driver testing only)" }, { "mem", ZINK_DEBUG_MEM, "Debug memory allocations" }, { "quiet", ZINK_DEBUG_QUIET, "Suppress warnings" }, DEBUG_NAMED_VALUE_END }; DEBUG_GET_ONCE_FLAGS_OPTION(zink_debug, "ZINK_DEBUG", zink_debug_options, 0) uint32_t zink_debug; static const struct debug_named_value zink_descriptor_options[] = { { "auto", ZINK_DESCRIPTOR_MODE_AUTO, "Automatically detect best mode" }, { "lazy", ZINK_DESCRIPTOR_MODE_LAZY, "Don't cache, do least amount of updates" }, { "db", ZINK_DESCRIPTOR_MODE_DB, "Use descriptor buffers" }, DEBUG_NAMED_VALUE_END }; DEBUG_GET_ONCE_FLAGS_OPTION(zink_descriptor_mode, "ZINK_DESCRIPTORS", zink_descriptor_options, ZINK_DESCRIPTOR_MODE_AUTO) enum zink_descriptor_mode zink_descriptor_mode; static const char * zink_get_vendor(struct pipe_screen *pscreen) { return "Mesa"; } static const char * zink_get_device_vendor(struct pipe_screen *pscreen) { struct zink_screen *screen = zink_screen(pscreen); static char buf[1000]; snprintf(buf, sizeof(buf), "Unknown (vendor-id: 0x%04x)", screen->info.props.vendorID); return buf; } static const char * zink_get_name(struct pipe_screen *pscreen) { struct zink_screen *screen = zink_screen(pscreen); const char *driver_name = vk_DriverId_to_str(screen->info.driver_props.driverID) + strlen("VK_DRIVER_ID_"); static char buf[1000]; snprintf(buf, sizeof(buf), "zink Vulkan %d.%d(%s (%s))", VK_VERSION_MAJOR(screen->info.device_version), VK_VERSION_MINOR(screen->info.device_version), screen->info.props.deviceName, strstr(vk_DriverId_to_str(screen->info.driver_props.driverID), "VK_DRIVER_ID_") ? driver_name : "Driver Unknown" ); return buf; } static void zink_get_driver_uuid(struct pipe_screen *pscreen, char *uuid) { struct zink_screen *screen = zink_screen(pscreen); if (screen->vk_version >= VK_MAKE_VERSION(1,2,0)) { memcpy(uuid, screen->info.props11.driverUUID, VK_UUID_SIZE); } else { memcpy(uuid, screen->info.deviceid_props.driverUUID, VK_UUID_SIZE); } } static void zink_get_device_uuid(struct pipe_screen *pscreen, char *uuid) { struct zink_screen *screen = zink_screen(pscreen); if (screen->vk_version >= VK_MAKE_VERSION(1,2,0)) { memcpy(uuid, screen->info.props11.deviceUUID, VK_UUID_SIZE); } else { memcpy(uuid, screen->info.deviceid_props.deviceUUID, VK_UUID_SIZE); } } static void zink_get_device_luid(struct pipe_screen *pscreen, char *luid) { struct zink_screen *screen = zink_screen(pscreen); if (screen->info.have_vulkan12) { memcpy(luid, screen->info.props11.deviceLUID, VK_LUID_SIZE); } else { memcpy(luid, screen->info.deviceid_props.deviceLUID, VK_LUID_SIZE); } } static uint32_t zink_get_device_node_mask(struct pipe_screen *pscreen) { struct zink_screen *screen = zink_screen(pscreen); if (screen->info.have_vulkan12) { return screen->info.props11.deviceNodeMask; } else { return screen->info.deviceid_props.deviceNodeMask; } } static void zink_set_max_shader_compiler_threads(struct pipe_screen *pscreen, unsigned max_threads) { struct zink_screen *screen = zink_screen(pscreen); util_queue_adjust_num_threads(&screen->cache_get_thread, max_threads, false); } static bool zink_is_parallel_shader_compilation_finished(struct pipe_screen *screen, void *shader, enum pipe_shader_type shader_type) { if (shader_type == MESA_SHADER_COMPUTE) { struct zink_program *pg = shader; return !pg->can_precompile || util_queue_fence_is_signalled(&pg->cache_fence); } struct zink_shader *zs = shader; if (!util_queue_fence_is_signalled(&zs->precompile.fence)) return false; bool finished = true; set_foreach(zs->programs, entry) { struct zink_gfx_program *prog = (void*)entry->key; finished &= util_queue_fence_is_signalled(&prog->base.cache_fence); } return finished; } static VkDeviceSize get_video_mem(struct zink_screen *screen) { VkDeviceSize size = 0; for (uint32_t i = 0; i < screen->info.mem_props.memoryHeapCount; ++i) { if (screen->info.mem_props.memoryHeaps[i].flags & VK_MEMORY_HEAP_DEVICE_LOCAL_BIT) size += screen->info.mem_props.memoryHeaps[i].size; } return size; } /** * Creates the disk cache used by mesa/st frontend for caching the GLSL -> NIR * path. * * The output that gets stored in the frontend's cache is the result of * zink_shader_finalize(). So, our sha1 cache key here needs to include * everything that would change the NIR we generate from a given set of GLSL * source, including our driver build, the Vulkan device and driver (which could * affect the pipe caps we show the frontend), and any debug flags that change * codegen. * * This disk cache also gets used by zink itself for storing its output from NIR * -> SPIRV translation. */ static bool disk_cache_init(struct zink_screen *screen) { if (zink_debug & ZINK_DEBUG_SHADERDB) return true; #ifdef ENABLE_SHADER_CACHE struct mesa_sha1 ctx; _mesa_sha1_init(&ctx); #ifdef HAVE_DL_ITERATE_PHDR /* Hash in the zink driver build. */ const struct build_id_note *note = build_id_find_nhdr_for_addr(disk_cache_init); unsigned build_id_len = build_id_length(note); assert(note && build_id_len == 20); /* sha1 */ _mesa_sha1_update(&ctx, build_id_data(note), build_id_len); #endif /* Hash in the Vulkan pipeline cache UUID to identify the combination of * vulkan device and driver (or any inserted layer that would invalidate our * cached pipelines). * * "Although they have identical descriptions, VkPhysicalDeviceIDProperties * ::deviceUUID may differ from * VkPhysicalDeviceProperties2::pipelineCacheUUID. The former is intended to * identify and correlate devices across API and driver boundaries, while the * latter is used to identify a compatible device and driver combination to * use when serializing and de-serializing pipeline state." */ _mesa_sha1_update(&ctx, screen->info.props.pipelineCacheUUID, VK_UUID_SIZE); /* Hash in our debug flags that affect NIR generation as of finalize_nir */ unsigned shader_debug_flags = zink_debug & ZINK_DEBUG_COMPACT; _mesa_sha1_update(&ctx, &shader_debug_flags, sizeof(shader_debug_flags)); /* Some of the driconf options change shaders. Let's just hash the whole * thing to not forget any (especially as options get added). */ _mesa_sha1_update(&ctx, &screen->driconf, sizeof(screen->driconf)); /* EXT_shader_object causes different descriptor layouts for separate shaders */ _mesa_sha1_update(&ctx, &screen->info.have_EXT_shader_object, sizeof(screen->info.have_EXT_shader_object)); /* Finish the sha1 and format it as text. */ unsigned char sha1[20]; _mesa_sha1_final(&ctx, sha1); char cache_id[20 * 2 + 1]; mesa_bytes_to_hex(cache_id, sha1, 20); screen->disk_cache = disk_cache_create("zink", cache_id, 0); if (!screen->disk_cache) return true; if (!util_queue_init(&screen->cache_put_thread, "zcq", 8, 1, UTIL_QUEUE_INIT_RESIZE_IF_FULL, screen)) { mesa_loge("zink: Failed to create disk cache queue\n"); disk_cache_destroy(screen->disk_cache); screen->disk_cache = NULL; return false; } #endif return true; } static void cache_put_job(void *data, void *gdata, int thread_index) { struct zink_program *pg = data; struct zink_screen *screen = gdata; size_t size = 0; u_rwlock_rdlock(&pg->pipeline_cache_lock); VkResult result = VKSCR(GetPipelineCacheData)(screen->dev, pg->pipeline_cache, &size, NULL); if (result != VK_SUCCESS) { u_rwlock_rdunlock(&pg->pipeline_cache_lock); mesa_loge("ZINK: vkGetPipelineCacheData failed (%s)", vk_Result_to_str(result)); return; } if (pg->pipeline_cache_size == size) { u_rwlock_rdunlock(&pg->pipeline_cache_lock); return; } void *pipeline_data = malloc(size); if (!pipeline_data) { u_rwlock_rdunlock(&pg->pipeline_cache_lock); return; } result = VKSCR(GetPipelineCacheData)(screen->dev, pg->pipeline_cache, &size, pipeline_data); u_rwlock_rdunlock(&pg->pipeline_cache_lock); if (result == VK_SUCCESS) { pg->pipeline_cache_size = size; cache_key key; disk_cache_compute_key(screen->disk_cache, pg->sha1, sizeof(pg->sha1), key); disk_cache_put_nocopy(screen->disk_cache, key, pipeline_data, size, NULL); } else { mesa_loge("ZINK: vkGetPipelineCacheData failed (%s)", vk_Result_to_str(result)); } } void zink_screen_update_pipeline_cache(struct zink_screen *screen, struct zink_program *pg, bool in_thread) { if (!screen->disk_cache || !pg->pipeline_cache) return; if (in_thread) cache_put_job(pg, screen, 0); else if (util_queue_fence_is_signalled(&pg->cache_fence)) util_queue_add_job(&screen->cache_put_thread, pg, &pg->cache_fence, cache_put_job, NULL, 0); } static void cache_get_job(void *data, void *gdata, int thread_index) { struct zink_program *pg = data; struct zink_screen *screen = gdata; VkPipelineCacheCreateInfo pcci; pcci.sType = VK_STRUCTURE_TYPE_PIPELINE_CACHE_CREATE_INFO; pcci.pNext = NULL; pcci.flags = screen->info.have_EXT_pipeline_creation_cache_control ? VK_PIPELINE_CACHE_CREATE_EXTERNALLY_SYNCHRONIZED_BIT : 0; pcci.initialDataSize = 0; pcci.pInitialData = NULL; cache_key key; disk_cache_compute_key(screen->disk_cache, pg->sha1, sizeof(pg->sha1), key); pcci.pInitialData = disk_cache_get(screen->disk_cache, key, &pg->pipeline_cache_size); pcci.initialDataSize = pg->pipeline_cache_size; VkResult res = VKSCR(CreatePipelineCache)(screen->dev, &pcci, NULL, &pg->pipeline_cache); if (res != VK_SUCCESS) { mesa_loge("ZINK: vkCreatePipelineCache failed (%s)", vk_Result_to_str(res)); } free((void*)pcci.pInitialData); } void zink_screen_get_pipeline_cache(struct zink_screen *screen, struct zink_program *pg, bool in_thread) { if (!screen->disk_cache) return; if (in_thread) cache_get_job(pg, screen, 0); else util_queue_add_job(&screen->cache_get_thread, pg, &pg->cache_fence, cache_get_job, NULL, 0); } static int zink_get_compute_param(struct pipe_screen *pscreen, enum pipe_shader_ir ir_type, enum pipe_compute_cap param, void *ret) { struct zink_screen *screen = zink_screen(pscreen); #define RET(x) do { \ if (ret) \ memcpy(ret, x, sizeof(x)); \ return sizeof(x); \ } while (0) switch (param) { case PIPE_COMPUTE_CAP_ADDRESS_BITS: RET((uint32_t []){ 64 }); case PIPE_COMPUTE_CAP_IR_TARGET: if (ret) strcpy(ret, "nir"); return 4; case PIPE_COMPUTE_CAP_GRID_DIMENSION: RET((uint64_t []) { 3 }); case PIPE_COMPUTE_CAP_MAX_GRID_SIZE: RET(((uint64_t []) { screen->info.props.limits.maxComputeWorkGroupCount[0], screen->info.props.limits.maxComputeWorkGroupCount[1], screen->info.props.limits.maxComputeWorkGroupCount[2] })); case PIPE_COMPUTE_CAP_MAX_BLOCK_SIZE: /* MaxComputeWorkGroupSize[0..2] */ RET(((uint64_t []) {screen->info.props.limits.maxComputeWorkGroupSize[0], screen->info.props.limits.maxComputeWorkGroupSize[1], screen->info.props.limits.maxComputeWorkGroupSize[2]})); case PIPE_COMPUTE_CAP_MAX_THREADS_PER_BLOCK: case PIPE_COMPUTE_CAP_MAX_VARIABLE_THREADS_PER_BLOCK: RET((uint64_t []) { screen->info.props.limits.maxComputeWorkGroupInvocations }); case PIPE_COMPUTE_CAP_MAX_LOCAL_SIZE: RET((uint64_t []) { screen->info.props.limits.maxComputeSharedMemorySize }); case PIPE_COMPUTE_CAP_IMAGES_SUPPORTED: RET((uint32_t []) { 1 }); case PIPE_COMPUTE_CAP_SUBGROUP_SIZES: RET((uint32_t []) { screen->info.props11.subgroupSize }); case PIPE_COMPUTE_CAP_MAX_MEM_ALLOC_SIZE: RET((uint64_t []) { screen->clamp_video_mem }); case PIPE_COMPUTE_CAP_MAX_GLOBAL_SIZE: RET((uint64_t []) { screen->total_video_mem }); case PIPE_COMPUTE_CAP_MAX_COMPUTE_UNITS: // no way in vulkan to retrieve this information. RET((uint32_t []) { 1 }); case PIPE_COMPUTE_CAP_MAX_SUBGROUPS: case PIPE_COMPUTE_CAP_MAX_CLOCK_FREQUENCY: case PIPE_COMPUTE_CAP_MAX_PRIVATE_SIZE: case PIPE_COMPUTE_CAP_MAX_INPUT_SIZE: // XXX: I think these are for Clover... return 0; default: unreachable("unknown compute param"); } } static uint32_t get_smallest_buffer_heap(struct zink_screen *screen) { enum zink_heap heaps[] = { ZINK_HEAP_DEVICE_LOCAL, ZINK_HEAP_DEVICE_LOCAL_VISIBLE, ZINK_HEAP_HOST_VISIBLE_COHERENT, ZINK_HEAP_HOST_VISIBLE_COHERENT }; unsigned size = UINT32_MAX; for (unsigned i = 0; i < ARRAY_SIZE(heaps); i++) { for (unsigned j = 0; j < screen->heap_count[i]; j++) { unsigned heap_idx = screen->info.mem_props.memoryTypes[screen->heap_map[i][j]].heapIndex; size = MIN2(screen->info.mem_props.memoryHeaps[heap_idx].size, size); } } return size; } static inline bool have_fp32_filter_linear(struct zink_screen *screen) { const VkFormat fp32_formats[] = { VK_FORMAT_R32_SFLOAT, VK_FORMAT_R32G32_SFLOAT, VK_FORMAT_R32G32B32_SFLOAT, VK_FORMAT_R32G32B32A32_SFLOAT, VK_FORMAT_D32_SFLOAT, }; for (int i = 0; i < ARRAY_SIZE(fp32_formats); ++i) { VkFormatProperties props; VKSCR(GetPhysicalDeviceFormatProperties)(screen->pdev, fp32_formats[i], &props); if (((props.linearTilingFeatures | props.optimalTilingFeatures) & (VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT | VK_FORMAT_FEATURE_SAMPLED_IMAGE_FILTER_LINEAR_BIT)) == VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT) { return false; } } return true; } static int zink_get_param(struct pipe_screen *pscreen, enum pipe_cap param) { struct zink_screen *screen = zink_screen(pscreen); switch (param) { case PIPE_CAP_NULL_TEXTURES: return screen->info.rb_image_feats.robustImageAccess; case PIPE_CAP_TEXRECT: case PIPE_CAP_MULTI_DRAW_INDIRECT_PARTIAL_STRIDE: return 0; case PIPE_CAP_ANISOTROPIC_FILTER: return screen->info.feats.features.samplerAnisotropy; case PIPE_CAP_EMULATE_NONFIXED_PRIMITIVE_RESTART: return 1; case PIPE_CAP_SUPPORTED_PRIM_MODES_WITH_RESTART: { uint32_t modes = BITFIELD_BIT(MESA_PRIM_LINE_STRIP) | BITFIELD_BIT(MESA_PRIM_TRIANGLE_STRIP) | BITFIELD_BIT(MESA_PRIM_LINE_STRIP_ADJACENCY) | BITFIELD_BIT(MESA_PRIM_TRIANGLE_STRIP_ADJACENCY); if (screen->have_triangle_fans) modes |= BITFIELD_BIT(MESA_PRIM_TRIANGLE_FAN); if (screen->info.have_EXT_primitive_topology_list_restart) { modes |= BITFIELD_BIT(MESA_PRIM_POINTS) | BITFIELD_BIT(MESA_PRIM_LINES) | BITFIELD_BIT(MESA_PRIM_LINES_ADJACENCY) | BITFIELD_BIT(MESA_PRIM_TRIANGLES) | BITFIELD_BIT(MESA_PRIM_TRIANGLES_ADJACENCY); if (screen->info.list_restart_feats.primitiveTopologyPatchListRestart) modes |= BITFIELD_BIT(MESA_PRIM_PATCHES); } return modes; } case PIPE_CAP_SUPPORTED_PRIM_MODES: { uint32_t modes = BITFIELD_MASK(MESA_PRIM_COUNT); if (!screen->have_triangle_fans) modes &= ~BITFIELD_BIT(MESA_PRIM_QUADS); modes &= ~BITFIELD_BIT(MESA_PRIM_QUAD_STRIP); modes &= ~BITFIELD_BIT(MESA_PRIM_POLYGON); modes &= ~BITFIELD_BIT(MESA_PRIM_LINE_LOOP); if (!screen->have_triangle_fans) modes &= ~BITFIELD_BIT(MESA_PRIM_TRIANGLE_FAN); return modes; } case PIPE_CAP_FBFETCH: return 1; case PIPE_CAP_FBFETCH_COHERENT: return screen->info.have_EXT_rasterization_order_attachment_access; case PIPE_CAP_MEMOBJ: return screen->instance_info.have_KHR_external_memory_capabilities && (screen->info.have_KHR_external_memory_fd || screen->info.have_KHR_external_memory_win32); case PIPE_CAP_FENCE_SIGNAL: return screen->info.have_KHR_external_semaphore_fd || screen->info.have_KHR_external_semaphore_win32; case PIPE_CAP_NATIVE_FENCE_FD: return screen->instance_info.have_KHR_external_semaphore_capabilities && screen->info.have_KHR_external_semaphore_fd; case PIPE_CAP_RESOURCE_FROM_USER_MEMORY: return screen->info.have_EXT_external_memory_host; case PIPE_CAP_SURFACE_REINTERPRET_BLOCKS: return screen->info.have_vulkan11 || screen->info.have_KHR_maintenance2; case PIPE_CAP_VALIDATE_ALL_DIRTY_STATES: case PIPE_CAP_ALLOW_MAPPED_BUFFERS_DURING_EXECUTION: case PIPE_CAP_MAP_UNSYNCHRONIZED_THREAD_SAFE: case PIPE_CAP_SHAREABLE_SHADERS: case PIPE_CAP_DEVICE_RESET_STATUS_QUERY: case PIPE_CAP_QUERY_MEMORY_INFO: case PIPE_CAP_NPOT_TEXTURES: case PIPE_CAP_TGSI_TEXCOORD: case PIPE_CAP_DRAW_INDIRECT: case PIPE_CAP_TEXTURE_QUERY_LOD: case PIPE_CAP_GLSL_TESS_LEVELS_AS_INPUTS: case PIPE_CAP_COPY_BETWEEN_COMPRESSED_AND_PLAIN_FORMATS: case PIPE_CAP_FORCE_PERSAMPLE_INTERP: case PIPE_CAP_FRAMEBUFFER_NO_ATTACHMENT: case PIPE_CAP_SHADER_ARRAY_COMPONENTS: case PIPE_CAP_QUERY_BUFFER_OBJECT: case PIPE_CAP_CONDITIONAL_RENDER_INVERTED: case PIPE_CAP_CLIP_HALFZ: case PIPE_CAP_TEXTURE_QUERY_SAMPLES: case PIPE_CAP_TEXTURE_BARRIER: case PIPE_CAP_QUERY_SO_OVERFLOW: case PIPE_CAP_GL_SPIRV: case PIPE_CAP_CLEAR_SCISSORED: case PIPE_CAP_INVALIDATE_BUFFER: case PIPE_CAP_PREFER_REAL_BUFFER_IN_CONSTBUF0: case PIPE_CAP_PACKED_UNIFORMS: case PIPE_CAP_SHADER_PACK_HALF_FLOAT: case PIPE_CAP_CULL_DISTANCE_NOCOMBINE: case PIPE_CAP_SEAMLESS_CUBE_MAP_PER_TEXTURE: case PIPE_CAP_LOAD_CONSTBUF: case PIPE_CAP_MULTISAMPLE_Z_RESOLVE: case PIPE_CAP_ALLOW_GLTHREAD_BUFFER_SUBDATA_OPT: return 1; case PIPE_CAP_DRAW_VERTEX_STATE: return screen->info.have_EXT_vertex_input_dynamic_state; case PIPE_CAP_SURFACE_SAMPLE_COUNT: return screen->vk_version >= VK_MAKE_VERSION(1,2,0); case PIPE_CAP_SHADER_GROUP_VOTE: if (screen->info.have_vulkan11 && (screen->info.subgroup.supportedOperations & VK_SUBGROUP_FEATURE_VOTE_BIT) && (screen->info.subgroup.supportedStages & VK_SHADER_STAGE_COMPUTE_BIT)) return true; if (screen->info.have_EXT_shader_subgroup_vote) return true; return false; case PIPE_CAP_QUADS_FOLLOW_PROVOKING_VERTEX_CONVENTION: return 1; case PIPE_CAP_TEXTURE_MIRROR_CLAMP_TO_EDGE: return screen->info.have_KHR_sampler_mirror_clamp_to_edge || (screen->info.have_vulkan12 && screen->info.feats12.samplerMirrorClampToEdge); case PIPE_CAP_POLYGON_OFFSET_UNITS_UNSCALED: return 1; case PIPE_CAP_POLYGON_OFFSET_CLAMP: return screen->info.feats.features.depthBiasClamp; case PIPE_CAP_QUERY_PIPELINE_STATISTICS_SINGLE: return screen->info.feats.features.pipelineStatisticsQuery; case PIPE_CAP_ROBUST_BUFFER_ACCESS_BEHAVIOR: return screen->info.feats.features.robustBufferAccess && (screen->info.rb2_feats.robustImageAccess2 || screen->driver_workarounds.lower_robustImageAccess2); case PIPE_CAP_MULTI_DRAW_INDIRECT: return screen->info.feats.features.multiDrawIndirect; case PIPE_CAP_IMAGE_ATOMIC_FLOAT_ADD: return (screen->info.have_EXT_shader_atomic_float && screen->info.atomic_float_feats.shaderSharedFloat32AtomicAdd && screen->info.atomic_float_feats.shaderBufferFloat32AtomicAdd); case PIPE_CAP_SHADER_ATOMIC_INT64: return (screen->info.have_KHR_shader_atomic_int64 && screen->info.atomic_int_feats.shaderSharedInt64Atomics && screen->info.atomic_int_feats.shaderBufferInt64Atomics); case PIPE_CAP_MULTI_DRAW_INDIRECT_PARAMS: return screen->info.have_KHR_draw_indirect_count; case PIPE_CAP_START_INSTANCE: case PIPE_CAP_DRAW_PARAMETERS: return (screen->info.have_vulkan12 && screen->info.feats11.shaderDrawParameters) || screen->info.have_KHR_shader_draw_parameters; case PIPE_CAP_VERTEX_ELEMENT_INSTANCE_DIVISOR: return screen->info.have_EXT_vertex_attribute_divisor; case PIPE_CAP_MAX_VERTEX_STREAMS: return screen->info.tf_props.maxTransformFeedbackStreams; case PIPE_CAP_COMPUTE_SHADER_DERIVATIVES: return screen->info.have_NV_compute_shader_derivatives; case PIPE_CAP_INT64: case PIPE_CAP_DOUBLES: return 1; case PIPE_CAP_MAX_DUAL_SOURCE_RENDER_TARGETS: if (!screen->info.feats.features.dualSrcBlend) return 0; return screen->info.props.limits.maxFragmentDualSrcAttachments; case PIPE_CAP_MAX_RENDER_TARGETS: return screen->info.props.limits.maxColorAttachments; case PIPE_CAP_OCCLUSION_QUERY: return screen->info.feats.features.occlusionQueryPrecise; case PIPE_CAP_PROGRAMMABLE_SAMPLE_LOCATIONS: return screen->info.have_EXT_sample_locations && screen->info.have_EXT_extended_dynamic_state; case PIPE_CAP_QUERY_TIME_ELAPSED: return screen->timestamp_valid_bits > 0; case PIPE_CAP_TEXTURE_MULTISAMPLE: return 1; case PIPE_CAP_FRAGMENT_SHADER_INTERLOCK: return screen->info.have_EXT_fragment_shader_interlock; case PIPE_CAP_SHADER_CLOCK: return screen->info.have_KHR_shader_clock; case PIPE_CAP_SHADER_BALLOT: if (screen->info.props11.subgroupSize > 64) return false; if (screen->info.have_vulkan11 && screen->info.subgroup.supportedOperations & VK_SUBGROUP_FEATURE_BALLOT_BIT) return true; if (screen->info.have_EXT_shader_subgroup_ballot) return true; return false; case PIPE_CAP_DEMOTE_TO_HELPER_INVOCATION: return screen->spirv_version >= SPIRV_VERSION(1, 6) || screen->info.have_EXT_shader_demote_to_helper_invocation; case PIPE_CAP_SAMPLE_SHADING: return screen->info.feats.features.sampleRateShading; case PIPE_CAP_TEXTURE_SWIZZLE: return 1; case PIPE_CAP_VERTEX_ATTRIB_ELEMENT_ALIGNED_ONLY: return 1; case PIPE_CAP_GL_CLAMP: return 0; case PIPE_CAP_PREFER_IMM_ARRAYS_AS_CONSTBUF: return 0; /* Assume that the vk driver is capable of moving imm arrays to some sort of constant storage on its own. */ case PIPE_CAP_TEXTURE_BORDER_COLOR_QUIRK: { enum pipe_quirk_texture_border_color_swizzle quirk = PIPE_QUIRK_TEXTURE_BORDER_COLOR_SWIZZLE_ALPHA_NOT_W; if (!screen->info.border_color_feats.customBorderColorWithoutFormat) return quirk | PIPE_QUIRK_TEXTURE_BORDER_COLOR_SWIZZLE_FREEDRENO; /* assume that if drivers don't implement this extension they either: * - don't support custom border colors * - handle things correctly * - hate border color accuracy */ if (screen->info.have_EXT_border_color_swizzle && !screen->info.border_swizzle_feats.borderColorSwizzleFromImage) return quirk | PIPE_QUIRK_TEXTURE_BORDER_COLOR_SWIZZLE_NV50; return quirk; } case PIPE_CAP_MAX_TEXTURE_2D_SIZE: return MIN2(screen->info.props.limits.maxImageDimension1D, screen->info.props.limits.maxImageDimension2D); case PIPE_CAP_MAX_TEXTURE_3D_LEVELS: return 1 + util_logbase2(screen->info.props.limits.maxImageDimension3D); case PIPE_CAP_MAX_TEXTURE_CUBE_LEVELS: return 1 + util_logbase2(screen->info.props.limits.maxImageDimensionCube); case PIPE_CAP_FRAGMENT_SHADER_TEXTURE_LOD: case PIPE_CAP_FRAGMENT_SHADER_DERIVATIVES: return 1; case PIPE_CAP_BLEND_EQUATION_SEPARATE: case PIPE_CAP_INDEP_BLEND_ENABLE: case PIPE_CAP_INDEP_BLEND_FUNC: return screen->info.feats.features.independentBlend; case PIPE_CAP_DITHERING: return 0; case PIPE_CAP_MAX_STREAM_OUTPUT_BUFFERS: return screen->info.have_EXT_transform_feedback ? screen->info.tf_props.maxTransformFeedbackBuffers : 0; case PIPE_CAP_STREAM_OUTPUT_PAUSE_RESUME: case PIPE_CAP_STREAM_OUTPUT_INTERLEAVE_BUFFERS: return screen->info.have_EXT_transform_feedback; case PIPE_CAP_MAX_TEXTURE_ARRAY_LAYERS: return screen->info.props.limits.maxImageArrayLayers; case PIPE_CAP_DEPTH_CLIP_DISABLE: return screen->info.have_EXT_depth_clip_enable; case PIPE_CAP_SHADER_STENCIL_EXPORT: return screen->info.have_EXT_shader_stencil_export; case PIPE_CAP_VS_INSTANCEID: case PIPE_CAP_SEAMLESS_CUBE_MAP: return 1; case PIPE_CAP_MIN_TEXEL_OFFSET: return screen->info.props.limits.minTexelOffset; case PIPE_CAP_MAX_TEXEL_OFFSET: return screen->info.props.limits.maxTexelOffset; case PIPE_CAP_VERTEX_COLOR_UNCLAMPED: return 1; case PIPE_CAP_CONDITIONAL_RENDER: return 1; case PIPE_CAP_GLSL_FEATURE_LEVEL_COMPATIBILITY: case PIPE_CAP_GLSL_FEATURE_LEVEL: return 460; case PIPE_CAP_COMPUTE: return 1; case PIPE_CAP_CONSTANT_BUFFER_OFFSET_ALIGNMENT: return screen->info.props.limits.minUniformBufferOffsetAlignment; case PIPE_CAP_QUERY_TIMESTAMP: return screen->timestamp_valid_bits > 0; case PIPE_CAP_QUERY_TIMESTAMP_BITS: return screen->timestamp_valid_bits; case PIPE_CAP_TIMER_RESOLUTION: return ceil(screen->info.props.limits.timestampPeriod); case PIPE_CAP_MIN_MAP_BUFFER_ALIGNMENT: return 1 << MIN_SLAB_ORDER; case PIPE_CAP_CUBE_MAP_ARRAY: return screen->info.feats.features.imageCubeArray; case PIPE_CAP_TEXTURE_BUFFER_OBJECTS: case PIPE_CAP_PRIMITIVE_RESTART: return 1; case PIPE_CAP_BINDLESS_TEXTURE: if (zink_descriptor_mode == ZINK_DESCRIPTOR_MODE_DB && (screen->info.db_props.maxDescriptorBufferBindings < 2 || screen->info.db_props.maxSamplerDescriptorBufferBindings < 2)) return 0; return screen->info.have_EXT_descriptor_indexing; case PIPE_CAP_TEXTURE_BUFFER_OFFSET_ALIGNMENT: return screen->info.props.limits.minTexelBufferOffsetAlignment; case PIPE_CAP_TEXTURE_TRANSFER_MODES: { enum pipe_texture_transfer_mode mode = PIPE_TEXTURE_TRANSFER_BLIT; if (!screen->is_cpu && /* this needs substantial perf tuning */ screen->info.driver_props.driverID != VK_DRIVER_ID_MESA_TURNIP && screen->info.have_KHR_8bit_storage && screen->info.have_KHR_16bit_storage && screen->info.have_KHR_shader_float16_int8) mode |= PIPE_TEXTURE_TRANSFER_COMPUTE; return mode; } case PIPE_CAP_MAX_TEXEL_BUFFER_ELEMENTS_UINT: return MIN2(get_smallest_buffer_heap(screen), screen->info.props.limits.maxTexelBufferElements); case PIPE_CAP_ENDIANNESS: return PIPE_ENDIAN_NATIVE; /* unsure */ case PIPE_CAP_MAX_VIEWPORTS: return MIN2(screen->info.props.limits.maxViewports, PIPE_MAX_VIEWPORTS); case PIPE_CAP_IMAGE_LOAD_FORMATTED: return screen->info.feats.features.shaderStorageImageReadWithoutFormat; case PIPE_CAP_IMAGE_STORE_FORMATTED: return screen->info.feats.features.shaderStorageImageWriteWithoutFormat; case PIPE_CAP_MIXED_FRAMEBUFFER_SIZES: return 1; case PIPE_CAP_MAX_GEOMETRY_OUTPUT_VERTICES: return screen->info.props.limits.maxGeometryOutputVertices; case PIPE_CAP_MAX_GEOMETRY_TOTAL_OUTPUT_COMPONENTS: return screen->info.props.limits.maxGeometryTotalOutputComponents; case PIPE_CAP_MAX_TEXTURE_GATHER_COMPONENTS: return 4; case PIPE_CAP_MIN_TEXTURE_GATHER_OFFSET: return screen->info.props.limits.minTexelGatherOffset; case PIPE_CAP_MAX_TEXTURE_GATHER_OFFSET: return screen->info.props.limits.maxTexelGatherOffset; case PIPE_CAP_SAMPLER_REDUCTION_MINMAX_ARB: return screen->info.feats12.samplerFilterMinmax || screen->info.have_EXT_sampler_filter_minmax; case PIPE_CAP_OPENCL_INTEGER_FUNCTIONS: case PIPE_CAP_INTEGER_MULTIPLY_32X16: return screen->info.have_INTEL_shader_integer_functions2; case PIPE_CAP_FS_FINE_DERIVATIVE: return 1; case PIPE_CAP_VENDOR_ID: return screen->info.props.vendorID; case PIPE_CAP_DEVICE_ID: return screen->info.props.deviceID; case PIPE_CAP_ACCELERATED: return !screen->is_cpu; case PIPE_CAP_VIDEO_MEMORY: return get_video_mem(screen) >> 20; case PIPE_CAP_UMA: return screen->info.props.deviceType == VK_PHYSICAL_DEVICE_TYPE_INTEGRATED_GPU; case PIPE_CAP_MAX_VERTEX_ATTRIB_STRIDE: return screen->info.props.limits.maxVertexInputBindingStride; case PIPE_CAP_SAMPLER_VIEW_TARGET: return 1; case PIPE_CAP_VS_LAYER_VIEWPORT: case PIPE_CAP_TES_LAYER_VIEWPORT: return screen->info.have_EXT_shader_viewport_index_layer || (screen->spirv_version >= SPIRV_VERSION(1, 5) && screen->info.feats12.shaderOutputLayer && screen->info.feats12.shaderOutputViewportIndex); case PIPE_CAP_TEXTURE_FLOAT_LINEAR: return have_fp32_filter_linear(screen); case PIPE_CAP_TEXTURE_HALF_FLOAT_LINEAR: return 1; case PIPE_CAP_SHADER_BUFFER_OFFSET_ALIGNMENT: return screen->info.props.limits.minStorageBufferOffsetAlignment; case PIPE_CAP_PCI_GROUP: case PIPE_CAP_PCI_BUS: case PIPE_CAP_PCI_DEVICE: case PIPE_CAP_PCI_FUNCTION: return 0; /* TODO: figure these out */ case PIPE_CAP_CULL_DISTANCE: return screen->info.feats.features.shaderCullDistance; case PIPE_CAP_SPARSE_BUFFER_PAGE_SIZE: return screen->info.feats.features.sparseResidencyBuffer ? ZINK_SPARSE_BUFFER_PAGE_SIZE : 0; /* Sparse texture */ case PIPE_CAP_MAX_SPARSE_TEXTURE_SIZE: return screen->info.feats.features.sparseResidencyImage2D ? zink_get_param(pscreen, PIPE_CAP_MAX_TEXTURE_2D_SIZE) : 0; case PIPE_CAP_MAX_SPARSE_3D_TEXTURE_SIZE: return screen->info.feats.features.sparseResidencyImage3D ? (1 << (zink_get_param(pscreen, PIPE_CAP_MAX_TEXTURE_3D_LEVELS) - 1)) : 0; case PIPE_CAP_MAX_SPARSE_ARRAY_TEXTURE_LAYERS: return screen->info.feats.features.sparseResidencyImage2D ? zink_get_param(pscreen, PIPE_CAP_MAX_TEXTURE_ARRAY_LAYERS) : 0; case PIPE_CAP_SPARSE_TEXTURE_FULL_ARRAY_CUBE_MIPMAPS: return screen->info.feats.features.sparseResidencyImage2D ? 1 : 0; case PIPE_CAP_QUERY_SPARSE_TEXTURE_RESIDENCY: return screen->info.feats.features.sparseResidency2Samples && screen->info.feats.features.shaderResourceResidency ? 1 : 0; case PIPE_CAP_CLAMP_SPARSE_TEXTURE_LOD: return screen->info.feats.features.shaderResourceMinLod && screen->info.feats.features.sparseResidency2Samples && screen->info.feats.features.shaderResourceResidency ? 1 : 0; case PIPE_CAP_VIEWPORT_SUBPIXEL_BITS: return screen->info.props.limits.viewportSubPixelBits; case PIPE_CAP_MAX_GS_INVOCATIONS: return screen->info.props.limits.maxGeometryShaderInvocations; case PIPE_CAP_MAX_COMBINED_SHADER_BUFFERS: /* gallium handles this automatically */ return 0; case PIPE_CAP_MAX_SHADER_BUFFER_SIZE_UINT: /* 1<<27 is required by VK spec */ assert(screen->info.props.limits.maxStorageBufferRange >= 1 << 27); /* clamp to VK spec minimum */ return MIN2(get_smallest_buffer_heap(screen), screen->info.props.limits.maxStorageBufferRange); case PIPE_CAP_FS_COORD_ORIGIN_UPPER_LEFT: case PIPE_CAP_FS_COORD_PIXEL_CENTER_HALF_INTEGER: return 1; case PIPE_CAP_FS_COORD_ORIGIN_LOWER_LEFT: case PIPE_CAP_FS_COORD_PIXEL_CENTER_INTEGER: return 0; case PIPE_CAP_FS_FACE_IS_INTEGER_SYSVAL: case PIPE_CAP_FS_POINT_IS_SYSVAL: return 1; case PIPE_CAP_VIEWPORT_TRANSFORM_LOWERED: return 1; case PIPE_CAP_POINT_SIZE_FIXED: return screen->info.have_KHR_maintenance5 ? PIPE_POINT_SIZE_LOWER_USER_ONLY : PIPE_POINT_SIZE_LOWER_ALWAYS; case PIPE_CAP_FLATSHADE: case PIPE_CAP_ALPHA_TEST: case PIPE_CAP_CLIP_PLANES: case PIPE_CAP_TWO_SIDED_COLOR: return 0; case PIPE_CAP_MAX_SHADER_PATCH_VARYINGS: return screen->info.props.limits.maxTessellationControlPerPatchOutputComponents / 4; case PIPE_CAP_MAX_VARYINGS: /* need to reserve up to 60 of our varying components and 16 slots for streamout */ return MIN2(screen->info.props.limits.maxVertexOutputComponents / 4 / 2, 16); case PIPE_CAP_DMABUF: #if defined(HAVE_LIBDRM) && (DETECT_OS_LINUX || DETECT_OS_BSD) return screen->info.have_KHR_external_memory_fd && screen->info.have_EXT_external_memory_dma_buf && screen->info.have_EXT_queue_family_foreign ? DRM_PRIME_CAP_IMPORT | DRM_PRIME_CAP_EXPORT : 0; #else return 0; #endif case PIPE_CAP_DEPTH_BOUNDS_TEST: return screen->info.feats.features.depthBounds; case PIPE_CAP_POST_DEPTH_COVERAGE: return screen->info.have_EXT_post_depth_coverage; case PIPE_CAP_STRING_MARKER: return screen->instance_info.have_EXT_debug_utils; default: return u_pipe_screen_get_param_defaults(pscreen, param); } } static float zink_get_paramf(struct pipe_screen *pscreen, enum pipe_capf param) { struct zink_screen *screen = zink_screen(pscreen); switch (param) { case PIPE_CAPF_MIN_LINE_WIDTH: case PIPE_CAPF_MIN_LINE_WIDTH_AA: if (!screen->info.feats.features.wideLines) return 1.0f; return MAX2(screen->info.props.limits.lineWidthRange[0], 0.01); case PIPE_CAPF_MIN_POINT_SIZE: case PIPE_CAPF_MIN_POINT_SIZE_AA: if (!screen->info.feats.features.largePoints) return 1.0f; return MAX2(screen->info.props.limits.pointSizeRange[0], 0.01); case PIPE_CAPF_LINE_WIDTH_GRANULARITY: if (!screen->info.feats.features.wideLines) return 0.1f; return screen->info.props.limits.lineWidthGranularity; case PIPE_CAPF_POINT_SIZE_GRANULARITY: if (!screen->info.feats.features.largePoints) return 0.1f; return screen->info.props.limits.pointSizeGranularity; case PIPE_CAPF_MAX_LINE_WIDTH: case PIPE_CAPF_MAX_LINE_WIDTH_AA: if (!screen->info.feats.features.wideLines) return 1.0f; return screen->info.props.limits.lineWidthRange[1]; case PIPE_CAPF_MAX_POINT_SIZE: case PIPE_CAPF_MAX_POINT_SIZE_AA: if (!screen->info.feats.features.largePoints) return 1.0f; return screen->info.props.limits.pointSizeRange[1]; case PIPE_CAPF_MAX_TEXTURE_ANISOTROPY: if (!screen->info.feats.features.samplerAnisotropy) return 1.0f; return screen->info.props.limits.maxSamplerAnisotropy; case PIPE_CAPF_MAX_TEXTURE_LOD_BIAS: return screen->info.props.limits.maxSamplerLodBias; case PIPE_CAPF_MIN_CONSERVATIVE_RASTER_DILATE: case PIPE_CAPF_MAX_CONSERVATIVE_RASTER_DILATE: case PIPE_CAPF_CONSERVATIVE_RASTER_DILATE_GRANULARITY: return 0.0f; /* not implemented */ } /* should only get here on unhandled cases */ return 0.0f; } static int zink_get_shader_param(struct pipe_screen *pscreen, gl_shader_stage shader, enum pipe_shader_cap param) { struct zink_screen *screen = zink_screen(pscreen); switch (param) { case PIPE_SHADER_CAP_MAX_INSTRUCTIONS: switch (shader) { case MESA_SHADER_FRAGMENT: case MESA_SHADER_VERTEX: return INT_MAX; case MESA_SHADER_TESS_CTRL: case MESA_SHADER_TESS_EVAL: if (screen->info.feats.features.tessellationShader && screen->info.have_KHR_maintenance2) return INT_MAX; break; case MESA_SHADER_GEOMETRY: if (screen->info.feats.features.geometryShader) return INT_MAX; break; case MESA_SHADER_COMPUTE: return INT_MAX; default: break; } return 0; case PIPE_SHADER_CAP_MAX_ALU_INSTRUCTIONS: case PIPE_SHADER_CAP_MAX_TEX_INSTRUCTIONS: case PIPE_SHADER_CAP_MAX_TEX_INDIRECTIONS: case PIPE_SHADER_CAP_MAX_CONTROL_FLOW_DEPTH: return INT_MAX; case PIPE_SHADER_CAP_MAX_INPUTS: { uint32_t max = 0; switch (shader) { case MESA_SHADER_VERTEX: max = MIN2(screen->info.props.limits.maxVertexInputAttributes, PIPE_MAX_ATTRIBS); break; case MESA_SHADER_TESS_CTRL: max = screen->info.props.limits.maxTessellationControlPerVertexInputComponents / 4; break; case MESA_SHADER_TESS_EVAL: max = screen->info.props.limits.maxTessellationEvaluationInputComponents / 4; break; case MESA_SHADER_GEOMETRY: max = screen->info.props.limits.maxGeometryInputComponents / 4; break; case MESA_SHADER_FRAGMENT: /* intel drivers report fewer components, but it's a value that's compatible * with what we need for GL, so we can still force a conformant value here */ if (screen->info.driver_props.driverID == VK_DRIVER_ID_INTEL_OPEN_SOURCE_MESA || screen->info.driver_props.driverID == VK_DRIVER_ID_INTEL_PROPRIETARY_WINDOWS || (screen->info.driver_props.driverID == VK_DRIVER_ID_MESA_VENUS && screen->info.props.vendorID == 0x8086)) return 32; max = screen->info.props.limits.maxFragmentInputComponents / 4; break; default: return 0; /* unsupported stage */ } switch (shader) { case MESA_SHADER_VERTEX: case MESA_SHADER_TESS_EVAL: case MESA_SHADER_GEOMETRY: /* last vertex stage must support streamout, and this is capped in glsl compiler */ return MIN2(max, MAX_VARYING); default: break; } return MIN2(max, 64); // prevent overflowing struct shader_info::inputs_read } case PIPE_SHADER_CAP_MAX_OUTPUTS: { uint32_t max = 0; switch (shader) { case MESA_SHADER_VERTEX: max = screen->info.props.limits.maxVertexOutputComponents / 4; break; case MESA_SHADER_TESS_CTRL: max = screen->info.props.limits.maxTessellationControlPerVertexOutputComponents / 4; break; case MESA_SHADER_TESS_EVAL: max = screen->info.props.limits.maxTessellationEvaluationOutputComponents / 4; break; case MESA_SHADER_GEOMETRY: max = screen->info.props.limits.maxGeometryOutputComponents / 4; break; case MESA_SHADER_FRAGMENT: max = screen->info.props.limits.maxColorAttachments; break; default: return 0; /* unsupported stage */ } return MIN2(max, 64); // prevent overflowing struct shader_info::outputs_read/written } case PIPE_SHADER_CAP_MAX_CONST_BUFFER0_SIZE: /* At least 16384 is guaranteed by VK spec */ assert(screen->info.props.limits.maxUniformBufferRange >= 16384); /* but Gallium can't handle values that are too big */ return MIN3(get_smallest_buffer_heap(screen), screen->info.props.limits.maxUniformBufferRange, BITFIELD_BIT(31)); case PIPE_SHADER_CAP_MAX_CONST_BUFFERS: return MIN2(screen->info.props.limits.maxPerStageDescriptorUniformBuffers, PIPE_MAX_CONSTANT_BUFFERS); case PIPE_SHADER_CAP_MAX_TEMPS: return INT_MAX; case PIPE_SHADER_CAP_INTEGERS: return 1; case PIPE_SHADER_CAP_INDIRECT_CONST_ADDR: case PIPE_SHADER_CAP_INDIRECT_TEMP_ADDR: case PIPE_SHADER_CAP_INDIRECT_INPUT_ADDR: case PIPE_SHADER_CAP_INDIRECT_OUTPUT_ADDR: return 1; case PIPE_SHADER_CAP_SUBROUTINES: case PIPE_SHADER_CAP_INT64_ATOMICS: case PIPE_SHADER_CAP_GLSL_16BIT_CONSTS: return 0; /* not implemented */ case PIPE_SHADER_CAP_FP16_CONST_BUFFERS: //enabling this breaks GTF-GL46.gtf21.GL2Tests.glGetUniform.glGetUniform //return screen->info.feats11.uniformAndStorageBuffer16BitAccess || //(screen->info.have_KHR_16bit_storage && screen->info.storage_16bit_feats.uniformAndStorageBuffer16BitAccess); return 0; case PIPE_SHADER_CAP_FP16_DERIVATIVES: return 0; //spirv requires 32bit derivative srcs and dests case PIPE_SHADER_CAP_FP16: return screen->info.feats12.shaderFloat16 || (screen->info.have_KHR_shader_float16_int8 && screen->info.shader_float16_int8_feats.shaderFloat16); case PIPE_SHADER_CAP_INT16: return screen->info.feats.features.shaderInt16; case PIPE_SHADER_CAP_TGSI_SQRT_SUPPORTED: return 0; /* not implemented */ case PIPE_SHADER_CAP_MAX_TEXTURE_SAMPLERS: case PIPE_SHADER_CAP_MAX_SAMPLER_VIEWS: return MIN2(MIN2(screen->info.props.limits.maxPerStageDescriptorSamplers, screen->info.props.limits.maxPerStageDescriptorSampledImages), PIPE_MAX_SAMPLERS); case PIPE_SHADER_CAP_TGSI_ANY_INOUT_DECL_RANGE: return 0; /* no idea */ case PIPE_SHADER_CAP_MAX_SHADER_BUFFERS: switch (shader) { case MESA_SHADER_VERTEX: case MESA_SHADER_TESS_CTRL: case MESA_SHADER_TESS_EVAL: case MESA_SHADER_GEOMETRY: if (!screen->info.feats.features.vertexPipelineStoresAndAtomics) return 0; break; case MESA_SHADER_FRAGMENT: if (!screen->info.feats.features.fragmentStoresAndAtomics) return 0; break; default: break; } /* TODO: this limitation is dumb, and will need some fixes in mesa */ return MIN2(screen->info.props.limits.maxPerStageDescriptorStorageBuffers, PIPE_MAX_SHADER_BUFFERS); case PIPE_SHADER_CAP_SUPPORTED_IRS: return (1 << PIPE_SHADER_IR_NIR) | (1 << PIPE_SHADER_IR_TGSI); case PIPE_SHADER_CAP_MAX_SHADER_IMAGES: if (screen->info.feats.features.shaderStorageImageExtendedFormats && screen->info.feats.features.shaderStorageImageWriteWithoutFormat) return MIN2(screen->info.props.limits.maxPerStageDescriptorStorageImages, ZINK_MAX_SHADER_IMAGES); return 0; case PIPE_SHADER_CAP_MAX_HW_ATOMIC_COUNTERS: case PIPE_SHADER_CAP_MAX_HW_ATOMIC_COUNTER_BUFFERS: return 0; /* not implemented */ case PIPE_SHADER_CAP_CONT_SUPPORTED: return 1; } /* should only get here on unhandled cases */ return 0; } static VkSampleCountFlagBits vk_sample_count_flags(uint32_t sample_count) { switch (sample_count) { case 1: return VK_SAMPLE_COUNT_1_BIT; case 2: return VK_SAMPLE_COUNT_2_BIT; case 4: return VK_SAMPLE_COUNT_4_BIT; case 8: return VK_SAMPLE_COUNT_8_BIT; case 16: return VK_SAMPLE_COUNT_16_BIT; case 32: return VK_SAMPLE_COUNT_32_BIT; case 64: return VK_SAMPLE_COUNT_64_BIT; default: return 0; } } static bool zink_is_compute_copy_faster(struct pipe_screen *pscreen, enum pipe_format src_format, enum pipe_format dst_format, unsigned width, unsigned height, unsigned depth, bool cpu) { if (cpu) /* very basic for now, probably even worse for some cases, * but fixes lots of others */ return width * height * depth > 64 * 64; return false; } static bool zink_is_format_supported(struct pipe_screen *pscreen, enum pipe_format format, enum pipe_texture_target target, unsigned sample_count, unsigned storage_sample_count, unsigned bind) { struct zink_screen *screen = zink_screen(pscreen); if (storage_sample_count && !screen->info.feats.features.shaderStorageImageMultisample && bind & PIPE_BIND_SHADER_IMAGE) return false; if (format == PIPE_FORMAT_NONE) return screen->info.props.limits.framebufferNoAttachmentsSampleCounts & vk_sample_count_flags(sample_count); if (bind & PIPE_BIND_INDEX_BUFFER) { if (format == PIPE_FORMAT_R8_UINT && !screen->info.have_EXT_index_type_uint8) return false; if (format != PIPE_FORMAT_R8_UINT && format != PIPE_FORMAT_R16_UINT && format != PIPE_FORMAT_R32_UINT) return false; } /* always use superset to determine feature support */ VkFormat vkformat = zink_get_format(screen, PIPE_FORMAT_A8_UNORM ? zink_format_get_emulated_alpha(format) : format); if (vkformat == VK_FORMAT_UNDEFINED) return false; if (sample_count >= 1) { VkSampleCountFlagBits sample_mask = vk_sample_count_flags(sample_count); if (!sample_mask) return false; const struct util_format_description *desc = util_format_description(format); if (util_format_is_depth_or_stencil(format)) { if (util_format_has_depth(desc)) { if (bind & PIPE_BIND_DEPTH_STENCIL && (screen->info.props.limits.framebufferDepthSampleCounts & sample_mask) != sample_mask) return false; if (bind & PIPE_BIND_SAMPLER_VIEW && (screen->info.props.limits.sampledImageDepthSampleCounts & sample_mask) != sample_mask) return false; } if (util_format_has_stencil(desc)) { if (bind & PIPE_BIND_DEPTH_STENCIL && (screen->info.props.limits.framebufferStencilSampleCounts & sample_mask) != sample_mask) return false; if (bind & PIPE_BIND_SAMPLER_VIEW && (screen->info.props.limits.sampledImageStencilSampleCounts & sample_mask) != sample_mask) return false; } } else if (util_format_is_pure_integer(format)) { if (bind & PIPE_BIND_RENDER_TARGET && !(screen->info.props.limits.framebufferColorSampleCounts & sample_mask)) return false; if (bind & PIPE_BIND_SAMPLER_VIEW && !(screen->info.props.limits.sampledImageIntegerSampleCounts & sample_mask)) return false; } else { if (bind & PIPE_BIND_RENDER_TARGET && !(screen->info.props.limits.framebufferColorSampleCounts & sample_mask)) return false; if (bind & PIPE_BIND_SAMPLER_VIEW && !(screen->info.props.limits.sampledImageColorSampleCounts & sample_mask)) return false; } if (bind & PIPE_BIND_SHADER_IMAGE) { if (!(screen->info.props.limits.storageImageSampleCounts & sample_mask)) return false; } VkResult ret; VkImageFormatProperties image_props; VkImageFormatProperties2 props2; props2.sType = VK_STRUCTURE_TYPE_IMAGE_FORMAT_PROPERTIES_2; props2.pNext = NULL; VkPhysicalDeviceImageFormatInfo2 info; info.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_IMAGE_FORMAT_INFO_2; info.pNext = NULL; info.format = vkformat; info.flags = 0; info.usage = VK_IMAGE_USAGE_TRANSFER_SRC_BIT; info.tiling = VK_IMAGE_TILING_OPTIMAL; switch (target) { case PIPE_TEXTURE_1D: case PIPE_TEXTURE_1D_ARRAY: { bool need_2D = false; if (util_format_is_depth_or_stencil(format)) need_2D |= screen->need_2D_zs; info.type = need_2D ? VK_IMAGE_TYPE_2D : VK_IMAGE_TYPE_1D; break; } case PIPE_TEXTURE_CUBE: case PIPE_TEXTURE_CUBE_ARRAY: info.flags |= VK_IMAGE_CREATE_CUBE_COMPATIBLE_BIT; FALLTHROUGH; case PIPE_TEXTURE_2D: case PIPE_TEXTURE_2D_ARRAY: case PIPE_TEXTURE_RECT: info.type = VK_IMAGE_TYPE_2D; break; case PIPE_TEXTURE_3D: info.type = VK_IMAGE_TYPE_3D; if (bind & (PIPE_BIND_RENDER_TARGET | PIPE_BIND_DEPTH_STENCIL)) info.flags |= VK_IMAGE_CREATE_2D_ARRAY_COMPATIBLE_BIT; if (screen->info.have_EXT_image_2d_view_of_3d) info.flags |= VK_IMAGE_CREATE_2D_VIEW_COMPATIBLE_BIT_EXT; break; default: unreachable("unknown texture target"); } u_foreach_bit(b, bind) { switch (1<pdev, &info, &props2); /* this is using VK_IMAGE_CREATE_EXTENDED_USAGE_BIT and can't be validated */ if (vk_format_aspects(vkformat) & VK_IMAGE_ASPECT_PLANE_1_BIT) ret = VK_SUCCESS; image_props = props2.imageFormatProperties; } else { ret = VKSCR(GetPhysicalDeviceImageFormatProperties)(screen->pdev, vkformat, info.type, info.tiling, info.usage, info.flags, &image_props); } if (ret != VK_SUCCESS) return false; if (!(sample_count & image_props.sampleCounts)) return false; } struct zink_format_props props = screen->format_props[format]; if (target == PIPE_BUFFER) { if (bind & PIPE_BIND_VERTEX_BUFFER) { if (!(props.bufferFeatures & VK_FORMAT_FEATURE_VERTEX_BUFFER_BIT)) { enum pipe_format new_format = zink_decompose_vertex_format(format); if (!new_format) return false; if (!(screen->format_props[new_format].bufferFeatures & VK_FORMAT_FEATURE_VERTEX_BUFFER_BIT)) return false; } } if (bind & PIPE_BIND_SAMPLER_VIEW && !(props.bufferFeatures & VK_FORMAT_FEATURE_UNIFORM_TEXEL_BUFFER_BIT)) return false; if (bind & PIPE_BIND_SHADER_IMAGE && !(props.bufferFeatures & VK_FORMAT_FEATURE_STORAGE_TEXEL_BUFFER_BIT)) return false; } else { /* all other targets are texture-targets */ if (bind & PIPE_BIND_RENDER_TARGET && !(props.optimalTilingFeatures & VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BIT)) return false; if (bind & PIPE_BIND_BLENDABLE && !(props.optimalTilingFeatures & VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BLEND_BIT)) return false; if (bind & PIPE_BIND_SAMPLER_VIEW && !(props.optimalTilingFeatures & VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT)) return false; if (bind & PIPE_BIND_SAMPLER_REDUCTION_MINMAX && !(props.optimalTilingFeatures & VK_FORMAT_FEATURE_SAMPLED_IMAGE_FILTER_MINMAX_BIT)) return false; if ((bind & PIPE_BIND_SAMPLER_VIEW) || (bind & PIPE_BIND_RENDER_TARGET)) { /* if this is a 3-component texture, force gallium to give us 4 components by rejecting this one */ const struct util_format_description *desc = util_format_description(format); if (desc->nr_channels == 3 && (desc->block.bits == 24 || desc->block.bits == 48 || desc->block.bits == 96)) return false; } if (bind & PIPE_BIND_DEPTH_STENCIL && !(props.optimalTilingFeatures & VK_FORMAT_FEATURE_DEPTH_STENCIL_ATTACHMENT_BIT)) return false; if (bind & PIPE_BIND_SHADER_IMAGE && !(props.optimalTilingFeatures & VK_FORMAT_FEATURE_STORAGE_IMAGE_BIT)) return false; } return true; } static void zink_set_damage_region(struct pipe_screen *pscreen, struct pipe_resource *pres, unsigned int nrects, const struct pipe_box *rects) { struct zink_resource *res = zink_resource(pres); for (unsigned i = 0; i < nrects; i++) { int y = pres->height0 - rects[i].y - rects[i].height; res->damage.extent.width = MAX2(res->damage.extent.width, rects[i].x + rects[i].width); res->damage.extent.height = MAX2(res->damage.extent.height, y + rects[i].height); res->damage.offset.x = MIN2(res->damage.offset.x, rects[i].x); res->damage.offset.y = MIN2(res->damage.offset.y, y); } res->use_damage = nrects > 0; } static void zink_destroy_screen(struct pipe_screen *pscreen) { struct zink_screen *screen = zink_screen(pscreen); #ifdef HAVE_RENDERDOC_APP_H if (screen->renderdoc_capture_all && p_atomic_dec_zero(&num_screens)) screen->renderdoc_api->EndFrameCapture(RENDERDOC_DEVICEPOINTER_FROM_VKINSTANCE(screen->instance), NULL); #endif hash_table_foreach(&screen->dts, entry) zink_kopper_deinit_displaytarget(screen, entry->data); if (screen->copy_context) screen->copy_context->base.destroy(&screen->copy_context->base); struct zink_batch_state *bs = screen->free_batch_states; while (bs) { struct zink_batch_state *bs_next = bs->next; zink_batch_state_destroy(screen, bs); bs = bs_next; } if (VK_NULL_HANDLE != screen->debugUtilsCallbackHandle) { VKSCR(DestroyDebugUtilsMessengerEXT)(screen->instance, screen->debugUtilsCallbackHandle, NULL); } util_vertex_state_cache_deinit(&screen->vertex_state_cache); if (screen->gfx_push_constant_layout) VKSCR(DestroyPipelineLayout)(screen->dev, screen->gfx_push_constant_layout, NULL); u_transfer_helper_destroy(pscreen->transfer_helper); if (util_queue_is_initialized(&screen->cache_get_thread)) { util_queue_finish(&screen->cache_get_thread); util_queue_destroy(&screen->cache_get_thread); } #ifdef ENABLE_SHADER_CACHE if (screen->disk_cache && util_queue_is_initialized(&screen->cache_put_thread)) { util_queue_finish(&screen->cache_put_thread); disk_cache_wait_for_idle(screen->disk_cache); util_queue_destroy(&screen->cache_put_thread); } #endif disk_cache_destroy(screen->disk_cache); /* we don't have an API to check if a set is already initialized */ for (unsigned i = 0; i < ARRAY_SIZE(screen->pipeline_libs); i++) if (screen->pipeline_libs[i].table) _mesa_set_clear(&screen->pipeline_libs[i], NULL); zink_bo_deinit(screen); util_live_shader_cache_deinit(&screen->shaders); zink_descriptor_layouts_deinit(screen); if (screen->sem) VKSCR(DestroySemaphore)(screen->dev, screen->sem, NULL); if (screen->fence) VKSCR(DestroyFence)(screen->dev, screen->fence, NULL); if (util_queue_is_initialized(&screen->flush_queue)) util_queue_destroy(&screen->flush_queue); while (util_dynarray_contains(&screen->semaphores, VkSemaphore)) VKSCR(DestroySemaphore)(screen->dev, util_dynarray_pop(&screen->semaphores, VkSemaphore), NULL); while (util_dynarray_contains(&screen->fd_semaphores, VkSemaphore)) VKSCR(DestroySemaphore)(screen->dev, util_dynarray_pop(&screen->fd_semaphores, VkSemaphore), NULL); if (screen->bindless_layout) VKSCR(DestroyDescriptorSetLayout)(screen->dev, screen->bindless_layout, NULL); if (screen->dev) VKSCR(DestroyDevice)(screen->dev, NULL); if (screen->instance) VKSCR(DestroyInstance)(screen->instance, NULL); util_idalloc_mt_fini(&screen->buffer_ids); if (screen->loader_lib) util_dl_close(screen->loader_lib); if (screen->drm_fd != -1) close(screen->drm_fd); slab_destroy_parent(&screen->transfer_pool); slab_destroy(&screen->present_mempool); ralloc_free(screen); glsl_type_singleton_decref(); } static int zink_get_display_device(const struct zink_screen *screen, uint32_t pdev_count, const VkPhysicalDevice *pdevs, int64_t dev_major, int64_t dev_minor) { VkPhysicalDeviceDrmPropertiesEXT drm_props = { .sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_DRM_PROPERTIES_EXT, }; VkPhysicalDeviceProperties2 props = { .sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_PROPERTIES_2, .pNext = &drm_props, }; for (uint32_t i = 0; i < pdev_count; ++i) { VKSCR(GetPhysicalDeviceProperties2)(pdevs[i], &props); if (drm_props.renderMajor == dev_major && drm_props.renderMinor == dev_minor) return i; } return -1; } static int zink_get_cpu_device_type(const struct zink_screen *screen, uint32_t pdev_count, const VkPhysicalDevice *pdevs) { VkPhysicalDeviceProperties props; for (uint32_t i = 0; i < pdev_count; ++i) { VKSCR(GetPhysicalDeviceProperties)(pdevs[i], &props); /* if user wants cpu, only give them cpu */ if (props.deviceType == VK_PHYSICAL_DEVICE_TYPE_CPU) return i; } mesa_loge("ZINK: CPU device requested but none found!"); return -1; } static void choose_pdev(struct zink_screen *screen, int64_t dev_major, int64_t dev_minor) { bool cpu = debug_get_bool_option("LIBGL_ALWAYS_SOFTWARE", false) || debug_get_bool_option("D3D_ALWAYS_SOFTWARE", false); if (cpu || (dev_major > 0 && dev_major < 255)) { uint32_t pdev_count; int idx; VkPhysicalDevice *pdevs; VkResult result = VKSCR(EnumeratePhysicalDevices)(screen->instance, &pdev_count, NULL); if (result != VK_SUCCESS) { mesa_loge("ZINK: vkEnumeratePhysicalDevices failed (%s)", vk_Result_to_str(result)); return; } assert(pdev_count > 0); pdevs = malloc(sizeof(*pdevs) * pdev_count); if (!pdevs) { mesa_loge("ZINK: failed to allocate pdevs!"); return; } result = VKSCR(EnumeratePhysicalDevices)(screen->instance, &pdev_count, pdevs); assert(result == VK_SUCCESS); assert(pdev_count > 0); if (cpu) idx = zink_get_cpu_device_type(screen, pdev_count, pdevs); else idx = zink_get_display_device(screen, pdev_count, pdevs, dev_major, dev_minor); if (idx != -1) /* valid cpu device */ screen->pdev = pdevs[idx]; free(pdevs); if (idx == -1) return; } else { VkPhysicalDevice pdev; unsigned pdev_count = 1; VkResult result = VKSCR(EnumeratePhysicalDevices)(screen->instance, &pdev_count, &pdev); if (result != VK_SUCCESS && result != VK_INCOMPLETE) { mesa_loge("ZINK: vkEnumeratePhysicalDevices failed (%s)", vk_Result_to_str(result)); return; } screen->pdev = pdev; } VKSCR(GetPhysicalDeviceProperties)(screen->pdev, &screen->info.props); /* allow software rendering only if forced by the user */ if (!cpu && screen->info.props.deviceType == VK_PHYSICAL_DEVICE_TYPE_CPU) { screen->pdev = VK_NULL_HANDLE; return; } screen->info.device_version = screen->info.props.apiVersion; /* runtime version is the lesser of the instance version and device version */ screen->vk_version = MIN2(screen->info.device_version, screen->instance_info.loader_version); /* calculate SPIR-V version based on VK version */ if (screen->vk_version >= VK_MAKE_VERSION(1, 3, 0)) screen->spirv_version = SPIRV_VERSION(1, 6); else if (screen->vk_version >= VK_MAKE_VERSION(1, 2, 0)) screen->spirv_version = SPIRV_VERSION(1, 5); else if (screen->vk_version >= VK_MAKE_VERSION(1, 1, 0)) screen->spirv_version = SPIRV_VERSION(1, 3); else screen->spirv_version = SPIRV_VERSION(1, 0); } static void update_queue_props(struct zink_screen *screen) { uint32_t num_queues; VKSCR(GetPhysicalDeviceQueueFamilyProperties)(screen->pdev, &num_queues, NULL); assert(num_queues > 0); VkQueueFamilyProperties *props = malloc(sizeof(*props) * num_queues); if (!props) { mesa_loge("ZINK: failed to allocate props!"); return; } VKSCR(GetPhysicalDeviceQueueFamilyProperties)(screen->pdev, &num_queues, props); bool found_gfx = false; uint32_t sparse_only = UINT32_MAX; screen->sparse_queue = UINT32_MAX; for (uint32_t i = 0; i < num_queues; i++) { if (props[i].queueFlags & VK_QUEUE_GRAPHICS_BIT) { if (found_gfx) continue; screen->sparse_queue = screen->gfx_queue = i; screen->max_queues = props[i].queueCount; screen->timestamp_valid_bits = props[i].timestampValidBits; found_gfx = true; } else if (props[i].queueFlags & VK_QUEUE_SPARSE_BINDING_BIT) sparse_only = i; } if (sparse_only != UINT32_MAX) screen->sparse_queue = sparse_only; free(props); } static void init_queue(struct zink_screen *screen) { simple_mtx_init(&screen->queue_lock, mtx_plain); VKSCR(GetDeviceQueue)(screen->dev, screen->gfx_queue, 0, &screen->queue); if (screen->sparse_queue != screen->gfx_queue) VKSCR(GetDeviceQueue)(screen->dev, screen->sparse_queue, 0, &screen->queue_sparse); else screen->queue_sparse = screen->queue; } static void zink_flush_frontbuffer(struct pipe_screen *pscreen, struct pipe_context *pctx, struct pipe_resource *pres, unsigned level, unsigned layer, void *winsys_drawable_handle, unsigned nboxes, struct pipe_box *sub_box) { struct zink_screen *screen = zink_screen(pscreen); struct zink_resource *res = zink_resource(pres); struct zink_context *ctx = zink_context(pctx); /* if the surface is no longer a swapchain, this is a no-op */ if (!zink_is_swapchain(res)) return; ctx = zink_tc_context_unwrap(pctx, screen->threaded); if (!zink_kopper_acquired(res->obj->dt, res->obj->dt_idx)) { /* swapbuffers to an undefined surface: acquire and present garbage */ zink_kopper_acquire(ctx, res, UINT64_MAX); ctx->needs_present = res; /* set batch usage to submit acquire semaphore */ zink_batch_resource_usage_set(&ctx->batch, res, true, false); /* ensure the resource is set up to present garbage */ ctx->base.flush_resource(&ctx->base, pres); } /* handle any outstanding acquire submits (not just from above) */ if (ctx->batch.swapchain || ctx->needs_present) { ctx->batch.has_work = true; pctx->flush(pctx, NULL, PIPE_FLUSH_END_OF_FRAME); if (ctx->last_batch_state && screen->threaded_submit) { struct zink_batch_state *bs = ctx->last_batch_state; util_queue_fence_wait(&bs->flush_completed); } } res->use_damage = false; /* always verify that this was acquired */ assert(zink_kopper_acquired(res->obj->dt, res->obj->dt_idx)); zink_kopper_present_queue(screen, res, nboxes, sub_box); } bool zink_is_depth_format_supported(struct zink_screen *screen, VkFormat format) { VkFormatProperties props; VKSCR(GetPhysicalDeviceFormatProperties)(screen->pdev, format, &props); return (props.linearTilingFeatures | props.optimalTilingFeatures) & VK_FORMAT_FEATURE_DEPTH_STENCIL_ATTACHMENT_BIT; } static enum pipe_format emulate_x8(enum pipe_format format) { /* convert missing Xn variants to An */ switch (format) { case PIPE_FORMAT_B8G8R8X8_UNORM: return PIPE_FORMAT_B8G8R8A8_UNORM; case PIPE_FORMAT_B8G8R8X8_SRGB: return PIPE_FORMAT_B8G8R8A8_SRGB; case PIPE_FORMAT_R8G8B8X8_SRGB: return PIPE_FORMAT_R8G8B8A8_SRGB; case PIPE_FORMAT_R8G8B8X8_SINT: return PIPE_FORMAT_R8G8B8A8_SINT; case PIPE_FORMAT_R8G8B8X8_SNORM: return PIPE_FORMAT_R8G8B8A8_SNORM; case PIPE_FORMAT_R8G8B8X8_UNORM: return PIPE_FORMAT_R8G8B8A8_UNORM; case PIPE_FORMAT_R16G16B16X16_FLOAT: return PIPE_FORMAT_R16G16B16A16_FLOAT; case PIPE_FORMAT_R16G16B16X16_SINT: return PIPE_FORMAT_R16G16B16A16_SINT; case PIPE_FORMAT_R16G16B16X16_SNORM: return PIPE_FORMAT_R16G16B16A16_SNORM; case PIPE_FORMAT_R16G16B16X16_UNORM: return PIPE_FORMAT_R16G16B16A16_UNORM; case PIPE_FORMAT_R32G32B32X32_FLOAT: return PIPE_FORMAT_R32G32B32A32_FLOAT; case PIPE_FORMAT_R32G32B32X32_SINT: return PIPE_FORMAT_R32G32B32A32_SINT; default: return format; } } VkFormat zink_get_format(struct zink_screen *screen, enum pipe_format format) { if (format == PIPE_FORMAT_A8_UNORM && !screen->driver_workarounds.missing_a8_unorm) return VK_FORMAT_A8_UNORM_KHR; else if (!screen->driver_workarounds.broken_l4a4 || format != PIPE_FORMAT_L4A4_UNORM) format = zink_format_get_emulated_alpha(format); VkFormat ret = vk_format_from_pipe_format(emulate_x8(format)); if (format == PIPE_FORMAT_X32_S8X24_UINT && screen->have_D32_SFLOAT_S8_UINT) return VK_FORMAT_D32_SFLOAT_S8_UINT; if (format == PIPE_FORMAT_X24S8_UINT) /* valid when using aspects to extract stencil, * fails format test because it's emulated */ ret = VK_FORMAT_D24_UNORM_S8_UINT; if (ret == VK_FORMAT_X8_D24_UNORM_PACK32 && !screen->have_X8_D24_UNORM_PACK32) { assert(zink_is_depth_format_supported(screen, VK_FORMAT_D32_SFLOAT)); return VK_FORMAT_D32_SFLOAT; } if (ret == VK_FORMAT_D24_UNORM_S8_UINT && !screen->have_D24_UNORM_S8_UINT) { assert(screen->have_D32_SFLOAT_S8_UINT); return VK_FORMAT_D32_SFLOAT_S8_UINT; } if ((ret == VK_FORMAT_A4B4G4R4_UNORM_PACK16 && !screen->info.format_4444_feats.formatA4B4G4R4) || (ret == VK_FORMAT_A4R4G4B4_UNORM_PACK16 && !screen->info.format_4444_feats.formatA4R4G4B4)) return VK_FORMAT_UNDEFINED; if (format == PIPE_FORMAT_R4A4_UNORM) return VK_FORMAT_R4G4_UNORM_PACK8; return ret; } void zink_convert_color(const struct zink_screen *screen, enum pipe_format format, union pipe_color_union *dst, const union pipe_color_union *src) { const struct util_format_description *desc = util_format_description(format); union pipe_color_union tmp = *src; for (unsigned i = 0; i < 4; i++) zink_format_clamp_channel_color(desc, &tmp, src, i); if (zink_format_is_emulated_alpha(format) && /* Don't swizzle colors if the driver supports real A8_UNORM */ (format != PIPE_FORMAT_A8_UNORM || screen->driver_workarounds.missing_a8_unorm)) { if (util_format_is_alpha(format)) { tmp.ui[0] = tmp.ui[3]; tmp.ui[1] = 0; tmp.ui[2] = 0; tmp.ui[3] = 0; } else if (util_format_is_luminance(format)) { tmp.ui[1] = 0; tmp.ui[2] = 0; tmp.f[3] = 1.0; } else if (util_format_is_luminance_alpha(format)) { tmp.ui[1] = tmp.ui[3]; tmp.ui[2] = 0; tmp.f[3] = 1.0; } else /* zink_format_is_red_alpha */ { tmp.ui[1] = tmp.ui[3]; tmp.ui[2] = 0; tmp.ui[3] = 0; } } memcpy(dst, &tmp, sizeof(union pipe_color_union)); } static bool check_have_device_time(struct zink_screen *screen) { uint32_t num_domains = 0; VkTimeDomainEXT domains[8]; //current max is 4 VkResult result = VKSCR(GetPhysicalDeviceCalibrateableTimeDomainsEXT)(screen->pdev, &num_domains, NULL); if (result != VK_SUCCESS) { mesa_loge("ZINK: vkGetPhysicalDeviceCalibrateableTimeDomainsEXT failed (%s)", vk_Result_to_str(result)); } assert(num_domains > 0); assert(num_domains < ARRAY_SIZE(domains)); result = VKSCR(GetPhysicalDeviceCalibrateableTimeDomainsEXT)(screen->pdev, &num_domains, domains); if (result != VK_SUCCESS) { mesa_loge("ZINK: vkGetPhysicalDeviceCalibrateableTimeDomainsEXT failed (%s)", vk_Result_to_str(result)); } /* VK_TIME_DOMAIN_DEVICE_EXT is used for the ctx->get_timestamp hook and is the only one we really need */ for (unsigned i = 0; i < num_domains; i++) { if (domains[i] == VK_TIME_DOMAIN_DEVICE_EXT) { return true; } } return false; } static void zink_error(const char *msg) { } static void zink_warn(const char *msg) { } static void zink_info(const char *msg) { } static void zink_msg(const char *msg) { } static VKAPI_ATTR VkBool32 VKAPI_CALL zink_debug_util_callback( VkDebugUtilsMessageSeverityFlagBitsEXT messageSeverity, VkDebugUtilsMessageTypeFlagsEXT messageType, const VkDebugUtilsMessengerCallbackDataEXT *pCallbackData, void *pUserData) { // Pick message prefix and color to use. // Only MacOS and Linux have been tested for color support if (messageSeverity & VK_DEBUG_UTILS_MESSAGE_SEVERITY_ERROR_BIT_EXT) { zink_error(pCallbackData->pMessage); } else if (messageSeverity & VK_DEBUG_UTILS_MESSAGE_SEVERITY_WARNING_BIT_EXT) { zink_warn(pCallbackData->pMessage); } else if (messageSeverity & VK_DEBUG_UTILS_MESSAGE_SEVERITY_INFO_BIT_EXT) { zink_info(pCallbackData->pMessage); } else zink_msg(pCallbackData->pMessage); return VK_FALSE; } static bool create_debug(struct zink_screen *screen) { VkDebugUtilsMessengerCreateInfoEXT vkDebugUtilsMessengerCreateInfoEXT = { VK_STRUCTURE_TYPE_DEBUG_UTILS_MESSENGER_CREATE_INFO_EXT, NULL, 0, // flags VK_DEBUG_UTILS_MESSAGE_SEVERITY_VERBOSE_BIT_EXT | VK_DEBUG_UTILS_MESSAGE_SEVERITY_INFO_BIT_EXT | VK_DEBUG_UTILS_MESSAGE_SEVERITY_WARNING_BIT_EXT | VK_DEBUG_UTILS_MESSAGE_SEVERITY_ERROR_BIT_EXT, VK_DEBUG_UTILS_MESSAGE_TYPE_GENERAL_BIT_EXT | VK_DEBUG_UTILS_MESSAGE_TYPE_VALIDATION_BIT_EXT | VK_DEBUG_UTILS_MESSAGE_TYPE_PERFORMANCE_BIT_EXT, zink_debug_util_callback, NULL }; VkDebugUtilsMessengerEXT vkDebugUtilsCallbackEXT = VK_NULL_HANDLE; VkResult result = VKSCR(CreateDebugUtilsMessengerEXT)( screen->instance, &vkDebugUtilsMessengerCreateInfoEXT, NULL, &vkDebugUtilsCallbackEXT); if (result != VK_SUCCESS) { mesa_loge("ZINK: vkCreateDebugUtilsMessengerEXT failed (%s)", vk_Result_to_str(result)); } screen->debugUtilsCallbackHandle = vkDebugUtilsCallbackEXT; return true; } static bool zink_internal_setup_moltenvk(struct zink_screen *screen) { #if defined(MVK_VERSION) if (!screen->instance_info.have_MVK_moltenvk) return true; GET_PROC_ADDR_INSTANCE_LOCAL(screen, screen->instance, GetMoltenVKConfigurationMVK); GET_PROC_ADDR_INSTANCE_LOCAL(screen, screen->instance, SetMoltenVKConfigurationMVK); GET_PROC_ADDR_INSTANCE_LOCAL(screen, screen->instance, GetVersionStringsMVK); if (vk_GetVersionStringsMVK) { char molten_version[64] = {0}; char vulkan_version[64] = {0}; vk_GetVersionStringsMVK(molten_version, sizeof(molten_version) - 1, vulkan_version, sizeof(vulkan_version) - 1); printf("zink: MoltenVK %s Vulkan %s \n", molten_version, vulkan_version); } if (vk_GetMoltenVKConfigurationMVK && vk_SetMoltenVKConfigurationMVK) { MVKConfiguration molten_config = {0}; size_t molten_config_size = sizeof(molten_config); VkResult res = vk_GetMoltenVKConfigurationMVK(screen->instance, &molten_config, &molten_config_size); if (res == VK_SUCCESS || res == VK_INCOMPLETE) { // Needed to allow MoltenVK to accept VkImageView swizzles. // Encountered when using VK_FORMAT_R8G8_UNORM molten_config.fullImageViewSwizzle = VK_TRUE; vk_SetMoltenVKConfigurationMVK(screen->instance, &molten_config, &molten_config_size); } } #endif // MVK_VERSION return true; } static void check_vertex_formats(struct zink_screen *screen) { /* from vbuf */ enum pipe_format format_list[] = { /* not supported by vk PIPE_FORMAT_R32_FIXED, PIPE_FORMAT_R32G32_FIXED, PIPE_FORMAT_R32G32B32_FIXED, PIPE_FORMAT_R32G32B32A32_FIXED, */ PIPE_FORMAT_R16_FLOAT, PIPE_FORMAT_R16G16_FLOAT, PIPE_FORMAT_R16G16B16_FLOAT, PIPE_FORMAT_R16G16B16A16_FLOAT, /* not supported by vk PIPE_FORMAT_R64_FLOAT, PIPE_FORMAT_R64G64_FLOAT, PIPE_FORMAT_R64G64B64_FLOAT, PIPE_FORMAT_R64G64B64A64_FLOAT, PIPE_FORMAT_R32_UNORM, PIPE_FORMAT_R32G32_UNORM, PIPE_FORMAT_R32G32B32_UNORM, PIPE_FORMAT_R32G32B32A32_UNORM, PIPE_FORMAT_R32_SNORM, PIPE_FORMAT_R32G32_SNORM, PIPE_FORMAT_R32G32B32_SNORM, PIPE_FORMAT_R32G32B32A32_SNORM, PIPE_FORMAT_R32_USCALED, PIPE_FORMAT_R32G32_USCALED, PIPE_FORMAT_R32G32B32_USCALED, PIPE_FORMAT_R32G32B32A32_USCALED, PIPE_FORMAT_R32_SSCALED, PIPE_FORMAT_R32G32_SSCALED, PIPE_FORMAT_R32G32B32_SSCALED, PIPE_FORMAT_R32G32B32A32_SSCALED, */ PIPE_FORMAT_R16_UNORM, PIPE_FORMAT_R16G16_UNORM, PIPE_FORMAT_R16G16B16_UNORM, PIPE_FORMAT_R16G16B16A16_UNORM, PIPE_FORMAT_R16_SNORM, PIPE_FORMAT_R16G16_SNORM, PIPE_FORMAT_R16G16B16_SNORM, PIPE_FORMAT_R16G16B16_SINT, PIPE_FORMAT_R16G16B16_UINT, PIPE_FORMAT_R16G16B16A16_SNORM, PIPE_FORMAT_R16_USCALED, PIPE_FORMAT_R16G16_USCALED, PIPE_FORMAT_R16G16B16_USCALED, PIPE_FORMAT_R16G16B16A16_USCALED, PIPE_FORMAT_R16_SSCALED, PIPE_FORMAT_R16G16_SSCALED, PIPE_FORMAT_R16G16B16_SSCALED, PIPE_FORMAT_R16G16B16A16_SSCALED, PIPE_FORMAT_R8_UNORM, PIPE_FORMAT_R8G8_UNORM, PIPE_FORMAT_R8G8B8_UNORM, PIPE_FORMAT_R8G8B8A8_UNORM, PIPE_FORMAT_R8_SNORM, PIPE_FORMAT_R8G8_SNORM, PIPE_FORMAT_R8G8B8_SNORM, PIPE_FORMAT_R8G8B8A8_SNORM, PIPE_FORMAT_R8_USCALED, PIPE_FORMAT_R8G8_USCALED, PIPE_FORMAT_R8G8B8_USCALED, PIPE_FORMAT_R8G8B8A8_USCALED, PIPE_FORMAT_R8_SSCALED, PIPE_FORMAT_R8G8_SSCALED, PIPE_FORMAT_R8G8B8_SSCALED, PIPE_FORMAT_R8G8B8A8_SSCALED, }; for (unsigned i = 0; i < ARRAY_SIZE(format_list); i++) { if (zink_is_format_supported(&screen->base, format_list[i], PIPE_BUFFER, 0, 0, PIPE_BIND_VERTEX_BUFFER)) continue; if (util_format_get_nr_components(format_list[i]) == 1) continue; enum pipe_format decomposed = zink_decompose_vertex_format(format_list[i]); if (zink_is_format_supported(&screen->base, decomposed, PIPE_BUFFER, 0, 0, PIPE_BIND_VERTEX_BUFFER)) { screen->need_decompose_attrs = true; mesa_logw("zink: this application would be much faster if %s supported vertex format %s", screen->info.props.deviceName, util_format_name(format_list[i])); } } } static void populate_format_props(struct zink_screen *screen) { for (unsigned i = 0; i < PIPE_FORMAT_COUNT; i++) { VkFormat format; retry: format = zink_get_format(screen, i); if (!format) continue; if (VKSCR(GetPhysicalDeviceFormatProperties2)) { VkFormatProperties2 props = {0}; props.sType = VK_STRUCTURE_TYPE_FORMAT_PROPERTIES_2; VkDrmFormatModifierPropertiesListEXT mod_props; VkDrmFormatModifierPropertiesEXT mods[128]; if (screen->info.have_EXT_image_drm_format_modifier) { mod_props.sType = VK_STRUCTURE_TYPE_DRM_FORMAT_MODIFIER_PROPERTIES_LIST_EXT; mod_props.pNext = NULL; mod_props.drmFormatModifierCount = ARRAY_SIZE(mods); mod_props.pDrmFormatModifierProperties = mods; props.pNext = &mod_props; } VkFormatProperties3 props3 = {0}; if (screen->info.have_KHR_format_feature_flags2 || screen->info.have_vulkan13) { props3.sType = VK_STRUCTURE_TYPE_FORMAT_PROPERTIES_3; props3.pNext = props.pNext; props.pNext = &props3; } VKSCR(GetPhysicalDeviceFormatProperties2)(screen->pdev, format, &props); if (screen->info.have_KHR_format_feature_flags2 || screen->info.have_vulkan13) { screen->format_props[i].linearTilingFeatures = props3.linearTilingFeatures; screen->format_props[i].optimalTilingFeatures = props3.optimalTilingFeatures; screen->format_props[i].bufferFeatures = props3.bufferFeatures; if (props3.linearTilingFeatures & VK_FORMAT_FEATURE_2_LINEAR_COLOR_ATTACHMENT_BIT_NV) screen->format_props[i].linearTilingFeatures |= VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BIT; } else { // MoltenVk is 1.2 API screen->format_props[i].linearTilingFeatures = props.formatProperties.linearTilingFeatures; screen->format_props[i].optimalTilingFeatures = props.formatProperties.optimalTilingFeatures; screen->format_props[i].bufferFeatures = props.formatProperties.bufferFeatures; } if (screen->info.have_EXT_image_drm_format_modifier && mod_props.drmFormatModifierCount) { screen->modifier_props[i].drmFormatModifierCount = mod_props.drmFormatModifierCount; screen->modifier_props[i].pDrmFormatModifierProperties = ralloc_array(screen, VkDrmFormatModifierPropertiesEXT, mod_props.drmFormatModifierCount); if (mod_props.pDrmFormatModifierProperties) { for (unsigned j = 0; j < mod_props.drmFormatModifierCount; j++) screen->modifier_props[i].pDrmFormatModifierProperties[j] = mod_props.pDrmFormatModifierProperties[j]; } } } else { VkFormatProperties props = {0}; VKSCR(GetPhysicalDeviceFormatProperties)(screen->pdev, format, &props); screen->format_props[i].linearTilingFeatures = props.linearTilingFeatures; screen->format_props[i].optimalTilingFeatures = props.optimalTilingFeatures; screen->format_props[i].bufferFeatures = props.bufferFeatures; } if (i == PIPE_FORMAT_A8_UNORM && !screen->driver_workarounds.missing_a8_unorm) { if (!screen->format_props[i].linearTilingFeatures && !screen->format_props[i].optimalTilingFeatures && !screen->format_props[i].bufferFeatures) { screen->driver_workarounds.missing_a8_unorm = true; goto retry; } } if (zink_format_is_emulated_alpha(i)) { VkFormatFeatureFlags blocked = VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BIT | VK_FORMAT_FEATURE_STORAGE_IMAGE_BIT; screen->format_props[i].linearTilingFeatures &= ~blocked; screen->format_props[i].optimalTilingFeatures &= ~blocked; screen->format_props[i].bufferFeatures = 0; } } check_vertex_formats(screen); VkImageFormatProperties image_props; VkResult ret = VKSCR(GetPhysicalDeviceImageFormatProperties)(screen->pdev, VK_FORMAT_D32_SFLOAT, VK_IMAGE_TYPE_1D, VK_IMAGE_TILING_OPTIMAL, VK_IMAGE_USAGE_DEPTH_STENCIL_ATTACHMENT_BIT | VK_IMAGE_USAGE_SAMPLED_BIT, 0, &image_props); if (ret != VK_SUCCESS && ret != VK_ERROR_FORMAT_NOT_SUPPORTED) { mesa_loge("ZINK: vkGetPhysicalDeviceImageFormatProperties failed (%s)", vk_Result_to_str(ret)); } screen->need_2D_zs = ret != VK_SUCCESS; if (screen->info.feats.features.sparseResidencyImage2D) screen->need_2D_sparse = !screen->base.get_sparse_texture_virtual_page_size(&screen->base, PIPE_TEXTURE_1D, false, PIPE_FORMAT_R32_FLOAT, 0, 16, NULL, NULL, NULL); } static void setup_renderdoc(struct zink_screen *screen) { #ifdef HAVE_RENDERDOC_APP_H const char *capture_id = debug_get_option("ZINK_RENDERDOC", NULL); if (!capture_id) return; void *renderdoc = dlopen("librenderdoc.so", RTLD_NOW | RTLD_NOLOAD); /* not loaded */ if (!renderdoc) return; pRENDERDOC_GetAPI get_api = dlsym(renderdoc, "RENDERDOC_GetAPI"); if (!get_api) return; /* need synchronous dispatch for renderdoc coherency */ screen->threaded_submit = false; get_api(eRENDERDOC_API_Version_1_0_0, (void*)&screen->renderdoc_api); screen->renderdoc_api->SetActiveWindow(RENDERDOC_DEVICEPOINTER_FROM_VKINSTANCE(screen->instance), NULL); int count = sscanf(capture_id, "%u:%u", &screen->renderdoc_capture_start, &screen->renderdoc_capture_end); if (count != 2) { count = sscanf(capture_id, "%u", &screen->renderdoc_capture_start); if (!count) { if (!strcmp(capture_id, "all")) { screen->renderdoc_capture_all = true; } else { printf("`ZINK_RENDERDOC` usage: ZINK_RENDERDOC=all|frame_no[:end_frame_no]\n"); abort(); } } screen->renderdoc_capture_end = screen->renderdoc_capture_start; } p_atomic_set(&screen->renderdoc_frame, 1); #endif } bool zink_screen_init_semaphore(struct zink_screen *screen) { VkSemaphoreCreateInfo sci = {0}; VkSemaphoreTypeCreateInfo tci = {0}; sci.pNext = &tci; sci.sType = VK_STRUCTURE_TYPE_SEMAPHORE_CREATE_INFO; tci.sType = VK_STRUCTURE_TYPE_SEMAPHORE_TYPE_CREATE_INFO; tci.semaphoreType = VK_SEMAPHORE_TYPE_TIMELINE; return VKSCR(CreateSemaphore)(screen->dev, &sci, NULL, &screen->sem) == VK_SUCCESS; } VkSemaphore zink_create_exportable_semaphore(struct zink_screen *screen) { VkExportSemaphoreCreateInfo eci = { VK_STRUCTURE_TYPE_EXPORT_SEMAPHORE_CREATE_INFO, NULL, VK_EXTERNAL_SEMAPHORE_HANDLE_TYPE_SYNC_FD_BIT }; VkSemaphoreCreateInfo sci = { VK_STRUCTURE_TYPE_SEMAPHORE_CREATE_INFO, &eci, 0 }; VkSemaphore sem = VK_NULL_HANDLE; if (util_dynarray_contains(&screen->fd_semaphores, VkSemaphore)) { simple_mtx_lock(&screen->semaphores_lock); if (util_dynarray_contains(&screen->fd_semaphores, VkSemaphore)) sem = util_dynarray_pop(&screen->fd_semaphores, VkSemaphore); simple_mtx_unlock(&screen->semaphores_lock); } if (sem) return sem; VkResult ret = VKSCR(CreateSemaphore)(screen->dev, &sci, NULL, &sem); return ret == VK_SUCCESS ? sem : VK_NULL_HANDLE; } VkSemaphore zink_screen_export_dmabuf_semaphore(struct zink_screen *screen, struct zink_resource *res) { VkSemaphore sem = VK_NULL_HANDLE; #if defined(HAVE_LIBDRM) && (DETECT_OS_LINUX || DETECT_OS_BSD) struct dma_buf_export_sync_file export = { .flags = DMA_BUF_SYNC_RW, .fd = -1, }; int fd = -1; if (res->obj->is_aux) { fd = os_dupfd_cloexec(res->obj->handle); } else { VkMemoryGetFdInfoKHR fd_info = {0}; fd_info.sType = VK_STRUCTURE_TYPE_MEMORY_GET_FD_INFO_KHR; fd_info.memory = zink_bo_get_mem(res->obj->bo); fd_info.handleType = VK_EXTERNAL_MEMORY_HANDLE_TYPE_DMA_BUF_BIT_EXT; VKSCR(GetMemoryFdKHR)(screen->dev, &fd_info, &fd); } if (unlikely(fd < 0)) { mesa_loge("MESA: Unable to get a valid memory fd"); return VK_NULL_HANDLE; } int ret = drmIoctl(fd, DMA_BUF_IOCTL_EXPORT_SYNC_FILE, &export); if (ret) { if (errno == ENOTTY || errno == EBADF || errno == ENOSYS) { assert(!"how did this fail?"); return VK_NULL_HANDLE; } else { mesa_loge("MESA: failed to import sync file '%s'", strerror(errno)); return VK_NULL_HANDLE; } } sem = zink_create_exportable_semaphore(screen); const VkImportSemaphoreFdInfoKHR sdi = { .sType = VK_STRUCTURE_TYPE_IMPORT_SEMAPHORE_FD_INFO_KHR, .semaphore = sem, .flags = VK_SEMAPHORE_IMPORT_TEMPORARY_BIT, .handleType = VK_EXTERNAL_SEMAPHORE_HANDLE_TYPE_SYNC_FD_BIT, .fd = export.fd, }; bool success = VKSCR(ImportSemaphoreFdKHR)(screen->dev, &sdi) == VK_SUCCESS; close(fd); if (!success) { VKSCR(DestroySemaphore)(screen->dev, sem, NULL); return VK_NULL_HANDLE; } #endif return sem; } bool zink_screen_import_dmabuf_semaphore(struct zink_screen *screen, struct zink_resource *res, VkSemaphore sem) { #if defined(HAVE_LIBDRM) && (DETECT_OS_LINUX || DETECT_OS_BSD) const VkSemaphoreGetFdInfoKHR get_fd_info = { .sType = VK_STRUCTURE_TYPE_SEMAPHORE_GET_FD_INFO_KHR, .semaphore = sem, .handleType = VK_EXTERNAL_SEMAPHORE_HANDLE_TYPE_SYNC_FD_BIT, }; int sync_file_fd = -1; VkResult result = VKSCR(GetSemaphoreFdKHR)(screen->dev, &get_fd_info, &sync_file_fd); if (result != VK_SUCCESS) { return false; } bool ret = false; int fd; if (res->obj->is_aux) { fd = os_dupfd_cloexec(res->obj->handle); } else { VkMemoryGetFdInfoKHR fd_info = {0}; fd_info.sType = VK_STRUCTURE_TYPE_MEMORY_GET_FD_INFO_KHR; fd_info.memory = zink_bo_get_mem(res->obj->bo); fd_info.handleType = VK_EXTERNAL_MEMORY_HANDLE_TYPE_DMA_BUF_BIT_EXT; if (VKSCR(GetMemoryFdKHR)(screen->dev, &fd_info, &fd) != VK_SUCCESS) fd = -1; } if (fd != -1) { struct dma_buf_import_sync_file import = { .flags = DMA_BUF_SYNC_RW, .fd = sync_file_fd, }; int ioctl_ret = drmIoctl(fd, DMA_BUF_IOCTL_IMPORT_SYNC_FILE, &import); if (ioctl_ret) { if (errno == ENOTTY || errno == EBADF || errno == ENOSYS) { assert(!"how did this fail?"); } else { ret = true; } } close(fd); } close(sync_file_fd); return ret; #else return true; #endif } bool zink_screen_timeline_wait(struct zink_screen *screen, uint64_t batch_id, uint64_t timeout) { VkSemaphoreWaitInfo wi = {0}; if (zink_screen_check_last_finished(screen, batch_id)) return true; wi.sType = VK_STRUCTURE_TYPE_SEMAPHORE_WAIT_INFO; wi.semaphoreCount = 1; wi.pSemaphores = &screen->sem; wi.pValues = &batch_id; bool success = false; if (screen->device_lost) return true; VkResult ret = VKSCR(WaitSemaphores)(screen->dev, &wi, timeout); success = zink_screen_handle_vkresult(screen, ret); if (success) zink_screen_update_last_finished(screen, batch_id); return success; } static uint32_t zink_get_loader_version(struct zink_screen *screen) { uint32_t loader_version = VK_API_VERSION_1_0; // Get the Loader version GET_PROC_ADDR_INSTANCE_LOCAL(screen, NULL, EnumerateInstanceVersion); if (vk_EnumerateInstanceVersion) { uint32_t loader_version_temp = VK_API_VERSION_1_0; VkResult result = (*vk_EnumerateInstanceVersion)(&loader_version_temp); if (VK_SUCCESS == result) { loader_version = loader_version_temp; } else { mesa_loge("ZINK: vkEnumerateInstanceVersion failed (%s)", vk_Result_to_str(result)); } } return loader_version; } static void zink_query_memory_info(struct pipe_screen *pscreen, struct pipe_memory_info *info) { struct zink_screen *screen = zink_screen(pscreen); memset(info, 0, sizeof(struct pipe_memory_info)); if (screen->info.have_EXT_memory_budget && VKSCR(GetPhysicalDeviceMemoryProperties2)) { VkPhysicalDeviceMemoryProperties2 mem = {0}; mem.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_MEMORY_PROPERTIES_2; VkPhysicalDeviceMemoryBudgetPropertiesEXT budget = {0}; budget.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_MEMORY_BUDGET_PROPERTIES_EXT; mem.pNext = &budget; VKSCR(GetPhysicalDeviceMemoryProperties2)(screen->pdev, &mem); for (unsigned i = 0; i < mem.memoryProperties.memoryHeapCount; i++) { if (mem.memoryProperties.memoryHeaps[i].flags & VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT) { /* VRAM */ info->total_device_memory += mem.memoryProperties.memoryHeaps[i].size / 1024; info->avail_device_memory += (mem.memoryProperties.memoryHeaps[i].size - budget.heapUsage[i]) / 1024; } else { /* GART */ info->total_staging_memory += mem.memoryProperties.memoryHeaps[i].size / 1024; info->avail_staging_memory += (mem.memoryProperties.memoryHeaps[i].size - budget.heapUsage[i]) / 1024; } } /* evictions not yet supported in vulkan */ } else { for (unsigned i = 0; i < screen->info.mem_props.memoryHeapCount; i++) { if (screen->info.mem_props.memoryHeaps[i].flags & VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT) { /* VRAM */ info->total_device_memory += screen->info.mem_props.memoryHeaps[i].size / 1024; /* free real estate! */ info->avail_device_memory += info->total_device_memory; } else { /* GART */ info->total_staging_memory += screen->info.mem_props.memoryHeaps[i].size / 1024; /* free real estate! */ info->avail_staging_memory += info->total_staging_memory; } } } } static void zink_query_dmabuf_modifiers(struct pipe_screen *pscreen, enum pipe_format format, int max, uint64_t *modifiers, unsigned int *external_only, int *count) { struct zink_screen *screen = zink_screen(pscreen); *count = screen->modifier_props[format].drmFormatModifierCount; for (int i = 0; i < MIN2(max, *count); i++) { if (external_only) external_only[i] = 0; modifiers[i] = screen->modifier_props[format].pDrmFormatModifierProperties[i].drmFormatModifier; } } static bool zink_is_dmabuf_modifier_supported(struct pipe_screen *pscreen, uint64_t modifier, enum pipe_format format, bool *external_only) { struct zink_screen *screen = zink_screen(pscreen); for (unsigned i = 0; i < screen->modifier_props[format].drmFormatModifierCount; i++) if (screen->modifier_props[format].pDrmFormatModifierProperties[i].drmFormatModifier == modifier) return true; return false; } static unsigned zink_get_dmabuf_modifier_planes(struct pipe_screen *pscreen, uint64_t modifier, enum pipe_format format) { struct zink_screen *screen = zink_screen(pscreen); for (unsigned i = 0; i < screen->modifier_props[format].drmFormatModifierCount; i++) if (screen->modifier_props[format].pDrmFormatModifierProperties[i].drmFormatModifier == modifier) return screen->modifier_props[format].pDrmFormatModifierProperties[i].drmFormatModifierPlaneCount; return util_format_get_num_planes(format); } static int zink_get_sparse_texture_virtual_page_size(struct pipe_screen *pscreen, enum pipe_texture_target target, bool multi_sample, enum pipe_format pformat, unsigned offset, unsigned size, int *x, int *y, int *z) { struct zink_screen *screen = zink_screen(pscreen); static const int page_size_2d[][3] = { { 256, 256, 1 }, /* 8bpp */ { 256, 128, 1 }, /* 16bpp */ { 128, 128, 1 }, /* 32bpp */ { 128, 64, 1 }, /* 64bpp */ { 64, 64, 1 }, /* 128bpp */ }; static const int page_size_3d[][3] = { { 64, 32, 32 }, /* 8bpp */ { 32, 32, 32 }, /* 16bpp */ { 32, 32, 16 }, /* 32bpp */ { 32, 16, 16 }, /* 64bpp */ { 16, 16, 16 }, /* 128bpp */ }; /* Only support one type of page size. */ if (offset != 0) return 0; /* reject multisample if 2x isn't supported; assume none are */ if (multi_sample && !screen->info.feats.features.sparseResidency2Samples) return 0; VkFormat format = zink_get_format(screen, pformat); bool is_zs = util_format_is_depth_or_stencil(pformat); VkImageType type; switch (target) { case PIPE_TEXTURE_1D: case PIPE_TEXTURE_1D_ARRAY: type = (screen->need_2D_sparse || (screen->need_2D_zs && is_zs)) ? VK_IMAGE_TYPE_2D : VK_IMAGE_TYPE_1D; break; case PIPE_TEXTURE_2D: case PIPE_TEXTURE_CUBE: case PIPE_TEXTURE_RECT: case PIPE_TEXTURE_2D_ARRAY: case PIPE_TEXTURE_CUBE_ARRAY: type = VK_IMAGE_TYPE_2D; break; case PIPE_TEXTURE_3D: type = VK_IMAGE_TYPE_3D; break; case PIPE_BUFFER: goto hack_it_up; default: return 0; } VkImageUsageFlags use_flags = VK_IMAGE_USAGE_SAMPLED_BIT | VK_IMAGE_USAGE_TRANSFER_DST_BIT | VK_IMAGE_USAGE_TRANSFER_SRC_BIT | VK_IMAGE_USAGE_STORAGE_BIT; use_flags |= is_zs ? VK_IMAGE_USAGE_DEPTH_STENCIL_ATTACHMENT_BIT : VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT; VkImageUsageFlags flags = screen->format_props[pformat].optimalTilingFeatures & use_flags; VkSparseImageFormatProperties props[4]; //planar? unsigned prop_count = ARRAY_SIZE(props); VKSCR(GetPhysicalDeviceSparseImageFormatProperties)(screen->pdev, format, type, multi_sample ? VK_SAMPLE_COUNT_2_BIT : VK_SAMPLE_COUNT_1_BIT, flags, VK_IMAGE_TILING_OPTIMAL, &prop_count, props); if (!prop_count) { /* format may not support storage; try without */ flags &= ~VK_IMAGE_USAGE_STORAGE_BIT; prop_count = ARRAY_SIZE(props); VKSCR(GetPhysicalDeviceSparseImageFormatProperties)(screen->pdev, format, type, multi_sample ? VK_SAMPLE_COUNT_2_BIT : VK_SAMPLE_COUNT_1_BIT, flags, VK_IMAGE_TILING_OPTIMAL, &prop_count, props); if (!prop_count) return 0; } if (size) { if (x) *x = props[0].imageGranularity.width; if (y) *y = props[0].imageGranularity.height; if (z) *z = props[0].imageGranularity.depth; } return 1; hack_it_up: { const int (*page_sizes)[3] = target == PIPE_TEXTURE_3D ? page_size_3d : page_size_2d; int blk_size = util_format_get_blocksize(pformat); if (size) { unsigned index = util_logbase2(blk_size); if (x) *x = page_sizes[index][0]; if (y) *y = page_sizes[index][1]; if (z) *z = page_sizes[index][2]; } } return 1; } static VkDevice zink_create_logical_device(struct zink_screen *screen) { VkDevice dev = VK_NULL_HANDLE; VkDeviceQueueCreateInfo qci[2] = {0}; uint32_t queues[3] = { screen->gfx_queue, screen->sparse_queue, }; float dummy = 0.0f; for (unsigned i = 0; i < ARRAY_SIZE(qci); i++) { qci[i].sType = VK_STRUCTURE_TYPE_DEVICE_QUEUE_CREATE_INFO; qci[i].queueFamilyIndex = queues[i]; qci[i].queueCount = 1; qci[i].pQueuePriorities = &dummy; } unsigned num_queues = 1; if (screen->sparse_queue != screen->gfx_queue) num_queues++; VkDeviceCreateInfo dci = {0}; dci.sType = VK_STRUCTURE_TYPE_DEVICE_CREATE_INFO; dci.queueCreateInfoCount = num_queues; dci.pQueueCreateInfos = qci; /* extensions don't have bool members in pEnabledFeatures. * this requires us to pass the whole VkPhysicalDeviceFeatures2 struct */ if (screen->info.feats.sType == VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_FEATURES_2) { dci.pNext = &screen->info.feats; } else { dci.pEnabledFeatures = &screen->info.feats.features; } dci.ppEnabledExtensionNames = screen->info.extensions; dci.enabledExtensionCount = screen->info.num_extensions; VkResult result = VKSCR(CreateDevice)(screen->pdev, &dci, NULL, &dev); if (result != VK_SUCCESS) mesa_loge("ZINK: vkCreateDevice failed (%s)", vk_Result_to_str(result)); return dev; } static void check_base_requirements(struct zink_screen *screen) { if (zink_debug & ZINK_DEBUG_QUIET) return; if (screen->info.driver_props.driverID == VK_DRIVER_ID_MESA_V3DV) { /* v3dv doesn't support straddling i/o, but zink doesn't do that so this is effectively supported: * don't spam errors in this case */ screen->info.feats12.scalarBlockLayout = true; screen->info.have_EXT_scalar_block_layout = true; } if (!screen->info.feats.features.logicOp || !screen->info.feats.features.fillModeNonSolid || !screen->info.feats.features.shaderClipDistance || !(screen->info.feats12.scalarBlockLayout || screen->info.have_EXT_scalar_block_layout) || !screen->info.have_KHR_maintenance1 || !screen->info.have_EXT_custom_border_color || !screen->info.have_EXT_line_rasterization) { fprintf(stderr, "WARNING: Some incorrect rendering " "might occur because the selected Vulkan device (%s) doesn't support " "base Zink requirements: ", screen->info.props.deviceName); #define CHECK_OR_PRINT(X) \ if (!screen->info.X) \ fprintf(stderr, "%s ", #X) CHECK_OR_PRINT(feats.features.logicOp); CHECK_OR_PRINT(feats.features.fillModeNonSolid); CHECK_OR_PRINT(feats.features.shaderClipDistance); if (!screen->info.feats12.scalarBlockLayout && !screen->info.have_EXT_scalar_block_layout) fprintf(stderr, "scalarBlockLayout OR EXT_scalar_block_layout "); CHECK_OR_PRINT(have_KHR_maintenance1); CHECK_OR_PRINT(have_EXT_custom_border_color); CHECK_OR_PRINT(have_EXT_line_rasterization); fprintf(stderr, "\n"); } if (screen->info.driver_props.driverID == VK_DRIVER_ID_MESA_V3DV) { screen->info.feats12.scalarBlockLayout = false; screen->info.have_EXT_scalar_block_layout = false; } } static void zink_get_sample_pixel_grid(struct pipe_screen *pscreen, unsigned sample_count, unsigned *width, unsigned *height) { struct zink_screen *screen = zink_screen(pscreen); unsigned idx = util_logbase2_ceil(MAX2(sample_count, 1)); assert(idx < ARRAY_SIZE(screen->maxSampleLocationGridSize)); *width = screen->maxSampleLocationGridSize[idx].width; *height = screen->maxSampleLocationGridSize[idx].height; } static void init_driver_workarounds(struct zink_screen *screen) { /* enable implicit sync for all non-mesa drivers */ screen->driver_workarounds.implicit_sync = true; switch (screen->info.driver_props.driverID) { case VK_DRIVER_ID_MESA_RADV: case VK_DRIVER_ID_INTEL_OPEN_SOURCE_MESA: case VK_DRIVER_ID_MESA_LLVMPIPE: case VK_DRIVER_ID_MESA_TURNIP: case VK_DRIVER_ID_MESA_V3DV: case VK_DRIVER_ID_MESA_PANVK: case VK_DRIVER_ID_MESA_VENUS: screen->driver_workarounds.implicit_sync = false; break; default: break; } /* TODO: maybe compile multiple variants for different set counts for compact mode? */ if (screen->info.props.limits.maxBoundDescriptorSets < ZINK_DESCRIPTOR_ALL_TYPES || zink_debug & (ZINK_DEBUG_COMPACT | ZINK_DEBUG_NOSHOBJ)) screen->info.have_EXT_shader_object = false; /* EDS2 is only used with EDS1 */ if (!screen->info.have_EXT_extended_dynamic_state) { screen->info.have_EXT_extended_dynamic_state2 = false; /* CWE usage needs EDS1 */ screen->info.have_EXT_color_write_enable = false; } if (screen->info.driver_props.driverID == VK_DRIVER_ID_AMD_PROPRIETARY) /* this completely breaks xfb somehow */ screen->info.have_EXT_extended_dynamic_state2 = false; /* EDS3 is only used with EDS2 */ if (!screen->info.have_EXT_extended_dynamic_state2) screen->info.have_EXT_extended_dynamic_state3 = false; /* EXT_vertex_input_dynamic_state is only used with EDS2 and above */ if (!screen->info.have_EXT_extended_dynamic_state2) screen->info.have_EXT_vertex_input_dynamic_state = false; if (screen->info.line_rast_feats.stippledRectangularLines && screen->info.line_rast_feats.stippledBresenhamLines && screen->info.line_rast_feats.stippledSmoothLines && !screen->info.dynamic_state3_feats.extendedDynamicState3LineStippleEnable) screen->info.have_EXT_extended_dynamic_state3 = false; if (!screen->info.dynamic_state3_feats.extendedDynamicState3PolygonMode || !screen->info.dynamic_state3_feats.extendedDynamicState3DepthClampEnable || !screen->info.dynamic_state3_feats.extendedDynamicState3DepthClipNegativeOneToOne || !screen->info.dynamic_state3_feats.extendedDynamicState3DepthClipEnable || !screen->info.dynamic_state3_feats.extendedDynamicState3ProvokingVertexMode || !screen->info.dynamic_state3_feats.extendedDynamicState3LineRasterizationMode) screen->info.have_EXT_extended_dynamic_state3 = false; else if (screen->info.dynamic_state3_feats.extendedDynamicState3SampleMask && screen->info.dynamic_state3_feats.extendedDynamicState3AlphaToCoverageEnable && (!screen->info.feats.features.alphaToOne || screen->info.dynamic_state3_feats.extendedDynamicState3AlphaToOneEnable) && screen->info.dynamic_state3_feats.extendedDynamicState3ColorBlendEnable && screen->info.dynamic_state3_feats.extendedDynamicState3RasterizationSamples && screen->info.dynamic_state3_feats.extendedDynamicState3ColorWriteMask && screen->info.dynamic_state3_feats.extendedDynamicState3ColorBlendEquation && screen->info.dynamic_state3_feats.extendedDynamicState3LogicOpEnable && screen->info.dynamic_state2_feats.extendedDynamicState2LogicOp) screen->have_full_ds3 = true; if (screen->info.have_EXT_graphics_pipeline_library) screen->info.have_EXT_graphics_pipeline_library = screen->info.have_EXT_extended_dynamic_state && screen->info.have_EXT_extended_dynamic_state2 && ((zink_debug & ZINK_DEBUG_GPL) || screen->info.dynamic_state2_feats.extendedDynamicState2PatchControlPoints) && screen->info.have_EXT_extended_dynamic_state3 && screen->info.have_KHR_dynamic_rendering && screen->info.have_EXT_non_seamless_cube_map && (!(zink_debug & ZINK_DEBUG_GPL) || screen->info.gpl_props.graphicsPipelineLibraryFastLinking || screen->is_cpu); screen->driver_workarounds.broken_l4a4 = screen->info.driver_props.driverID == VK_DRIVER_ID_NVIDIA_PROPRIETARY; if (screen->info.driver_props.driverID == VK_DRIVER_ID_MESA_TURNIP) { /* performance */ screen->info.border_color_feats.customBorderColorWithoutFormat = VK_FALSE; } if (!screen->info.have_KHR_maintenance5) screen->driver_workarounds.missing_a8_unorm = true; if ((!screen->info.have_EXT_line_rasterization || !screen->info.line_rast_feats.stippledBresenhamLines) && screen->info.feats.features.geometryShader && screen->info.feats.features.sampleRateShading) { /* we're using stippledBresenhamLines as a proxy for all of these, to * avoid accidentally changing behavior on VK-drivers where we don't * want to add emulation. */ screen->driver_workarounds.no_linestipple = true; } if (screen->info.driver_props.driverID == VK_DRIVER_ID_IMAGINATION_PROPRIETARY) { assert(screen->info.feats.features.geometryShader); screen->driver_workarounds.no_linesmooth = true; } /* This is a workarround for the lack of * gl_PointSize + glPolygonMode(..., GL_LINE), in the imagination * proprietary driver. */ switch (screen->info.driver_props.driverID) { case VK_DRIVER_ID_IMAGINATION_PROPRIETARY: screen->driver_workarounds.no_hw_gl_point = true; break; default: screen->driver_workarounds.no_hw_gl_point = false; break; } if (screen->info.driver_props.driverID == VK_DRIVER_ID_AMD_OPEN_SOURCE || screen->info.driver_props.driverID == VK_DRIVER_ID_AMD_PROPRIETARY || screen->info.driver_props.driverID == VK_DRIVER_ID_NVIDIA_PROPRIETARY || screen->info.driver_props.driverID == VK_DRIVER_ID_MESA_RADV) screen->driver_workarounds.z24_unscaled_bias = 1<<23; else screen->driver_workarounds.z24_unscaled_bias = 1<<24; if (screen->info.driver_props.driverID == VK_DRIVER_ID_NVIDIA_PROPRIETARY) screen->driver_workarounds.z16_unscaled_bias = 1<<15; else screen->driver_workarounds.z16_unscaled_bias = 1<<16; /* these drivers don't use VK_PIPELINE_CREATE_COLOR_ATTACHMENT_FEEDBACK_LOOP_BIT_EXT, so it can always be set */ switch (screen->info.driver_props.driverID) { case VK_DRIVER_ID_MESA_RADV: case VK_DRIVER_ID_INTEL_OPEN_SOURCE_MESA: case VK_DRIVER_ID_MESA_LLVMPIPE: case VK_DRIVER_ID_MESA_VENUS: case VK_DRIVER_ID_NVIDIA_PROPRIETARY: case VK_DRIVER_ID_INTEL_PROPRIETARY_WINDOWS: case VK_DRIVER_ID_IMAGINATION_PROPRIETARY: screen->driver_workarounds.always_feedback_loop = screen->info.have_EXT_attachment_feedback_loop_layout; break; default: break; } /* these drivers don't use VK_PIPELINE_CREATE_DEPTH_STENCIL_ATTACHMENT_FEEDBACK_LOOP_BIT_EXT, so it can always be set */ switch (screen->info.driver_props.driverID) { case VK_DRIVER_ID_MESA_LLVMPIPE: case VK_DRIVER_ID_MESA_VENUS: case VK_DRIVER_ID_NVIDIA_PROPRIETARY: case VK_DRIVER_ID_IMAGINATION_PROPRIETARY: screen->driver_workarounds.always_feedback_loop_zs = screen->info.have_EXT_attachment_feedback_loop_layout; break; default: break; } /* use same mechanics if dynamic state is supported */ screen->driver_workarounds.always_feedback_loop |= screen->info.have_EXT_attachment_feedback_loop_dynamic_state; screen->driver_workarounds.always_feedback_loop_zs |= screen->info.have_EXT_attachment_feedback_loop_dynamic_state; /* these drivers cannot handle OOB gl_Layer values, and therefore need clamping in shader. * TODO: Vulkan extension that details whether vulkan driver can handle OOB layer values */ switch (screen->info.driver_props.driverID) { case VK_DRIVER_ID_IMAGINATION_PROPRIETARY: screen->driver_workarounds.needs_sanitised_layer = true; break; default: screen->driver_workarounds.needs_sanitised_layer = false; break; } /* these drivers will produce undefined results when using swizzle 1 with combined z/s textures * TODO: use a future device property when available */ switch (screen->info.driver_props.driverID) { case VK_DRIVER_ID_IMAGINATION_PROPRIETARY: case VK_DRIVER_ID_IMAGINATION_OPEN_SOURCE_MESA: screen->driver_workarounds.needs_zs_shader_swizzle = true; break; default: screen->driver_workarounds.needs_zs_shader_swizzle = false; break; } /* When robust contexts are advertised but robustImageAccess2 is not available */ screen->driver_workarounds.lower_robustImageAccess2 = !screen->info.rb2_feats.robustImageAccess2 && screen->info.feats.features.robustBufferAccess && screen->info.rb_image_feats.robustImageAccess; /* once more testing has been done, use the #if 0 block */ unsigned illegal = ZINK_DEBUG_RP | ZINK_DEBUG_NORP; if ((zink_debug & illegal) == illegal) { mesa_loge("Cannot specify ZINK_DEBUG=rp and ZINK_DEBUG=norp"); abort(); } /* these drivers benefit from renderpass optimization */ switch (screen->info.driver_props.driverID) { case VK_DRIVER_ID_MESA_LLVMPIPE: case VK_DRIVER_ID_MESA_TURNIP: case VK_DRIVER_ID_MESA_PANVK: case VK_DRIVER_ID_MESA_V3DV: case VK_DRIVER_ID_IMAGINATION_PROPRIETARY: case VK_DRIVER_ID_QUALCOMM_PROPRIETARY: case VK_DRIVER_ID_BROADCOM_PROPRIETARY: case VK_DRIVER_ID_ARM_PROPRIETARY: screen->driver_workarounds.track_renderpasses = true; //screen->info.primgen_feats.primitivesGeneratedQueryWithRasterizerDiscard break; default: break; } if (zink_debug & ZINK_DEBUG_RP) screen->driver_workarounds.track_renderpasses = true; else if (zink_debug & ZINK_DEBUG_NORP) screen->driver_workarounds.track_renderpasses = false; /* these drivers can't optimize non-overlapping copy ops */ switch (screen->info.driver_props.driverID) { case VK_DRIVER_ID_MESA_TURNIP: case VK_DRIVER_ID_QUALCOMM_PROPRIETARY: screen->driver_workarounds.broken_cache_semantics = true; break; default: break; } /* these drivers can successfully do INVALID <-> LINEAR dri3 modifier swap */ switch (screen->info.driver_props.driverID) { case VK_DRIVER_ID_MESA_TURNIP: case VK_DRIVER_ID_MESA_VENUS: screen->driver_workarounds.can_do_invalid_linear_modifier = true; break; default: break; } /* these drivers have no difference between unoptimized and optimized shader compilation */ switch (screen->info.driver_props.driverID) { case VK_DRIVER_ID_MESA_LLVMPIPE: screen->driver_workarounds.disable_optimized_compile = true; break; default: if (zink_debug & ZINK_DEBUG_NOOPT) screen->driver_workarounds.disable_optimized_compile = true; break; } switch (screen->info.driver_props.driverID) { case VK_DRIVER_ID_MESA_RADV: case VK_DRIVER_ID_AMD_OPEN_SOURCE: case VK_DRIVER_ID_AMD_PROPRIETARY: /* this has bad perf on AMD */ screen->info.have_KHR_push_descriptor = false; break; default: break; } if (!screen->resizable_bar) screen->info.have_EXT_host_image_copy = false; } static void fixup_driver_props(struct zink_screen *screen) { VkPhysicalDeviceProperties2 props = { VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_PROPERTIES_2 }; if (screen->info.have_EXT_host_image_copy) { /* fill in layouts */ screen->info.hic_props.pNext = props.pNext; props.pNext = &screen->info.hic_props; screen->info.hic_props.pCopySrcLayouts = ralloc_array(screen, VkImageLayout, screen->info.hic_props.copySrcLayoutCount); screen->info.hic_props.pCopyDstLayouts = ralloc_array(screen, VkImageLayout, screen->info.hic_props.copyDstLayoutCount); } if (props.pNext) screen->vk.GetPhysicalDeviceProperties2(screen->pdev, &props); if (screen->info.have_EXT_host_image_copy) { for (unsigned i = 0; i < screen->info.hic_props.copyDstLayoutCount; i++) { if (screen->info.hic_props.pCopyDstLayouts[i] == VK_IMAGE_LAYOUT_SHADER_READ_ONLY_OPTIMAL) { screen->can_hic_shader_read = true; break; } } } } static void init_optimal_keys(struct zink_screen *screen) { /* assume that anyone who knows enough to enable optimal_keys on turnip doesn't care about missing line stipple */ if (zink_debug & ZINK_DEBUG_OPTIMAL_KEYS && screen->info.driver_props.driverID == VK_DRIVER_ID_MESA_TURNIP) zink_debug |= ZINK_DEBUG_QUIET; screen->optimal_keys = !screen->need_decompose_attrs && screen->info.have_EXT_non_seamless_cube_map && screen->info.have_EXT_provoking_vertex && !screen->driconf.inline_uniforms && !screen->driver_workarounds.no_linestipple && !screen->driver_workarounds.no_linesmooth && !screen->driver_workarounds.no_hw_gl_point && !screen->driver_workarounds.lower_robustImageAccess2 && !screen->driconf.emulate_point_smooth && !screen->driver_workarounds.needs_zs_shader_swizzle; if (!screen->optimal_keys && zink_debug & ZINK_DEBUG_OPTIMAL_KEYS && !(zink_debug & ZINK_DEBUG_QUIET)) { fprintf(stderr, "The following criteria are preventing optimal_keys enablement:\n"); if (screen->need_decompose_attrs) fprintf(stderr, "missing vertex attribute formats\n"); if (screen->driconf.inline_uniforms) fprintf(stderr, "uniform inlining must be disabled (set ZINK_INLINE_UNIFORMS=0 in your env)\n"); if (screen->driconf.emulate_point_smooth) fprintf(stderr, "smooth point emulation is enabled\n"); if (screen->driver_workarounds.needs_zs_shader_swizzle) fprintf(stderr, "Z/S shader swizzle workaround is enabled\n"); CHECK_OR_PRINT(have_EXT_line_rasterization); CHECK_OR_PRINT(line_rast_feats.stippledBresenhamLines); CHECK_OR_PRINT(feats.features.geometryShader); CHECK_OR_PRINT(feats.features.sampleRateShading); CHECK_OR_PRINT(have_EXT_non_seamless_cube_map); CHECK_OR_PRINT(have_EXT_provoking_vertex); if (screen->driver_workarounds.no_linesmooth) fprintf(stderr, "driver does not support smooth lines\n"); if (screen->driver_workarounds.no_hw_gl_point) fprintf(stderr, "driver does not support hardware GL_POINT\n"); CHECK_OR_PRINT(rb2_feats.robustImageAccess2); CHECK_OR_PRINT(feats.features.robustBufferAccess); CHECK_OR_PRINT(rb_image_feats.robustImageAccess); printf("\n"); mesa_logw("zink: force-enabling optimal_keys despite missing features. Good luck!"); } if (zink_debug & ZINK_DEBUG_OPTIMAL_KEYS) screen->optimal_keys = true; if (!screen->optimal_keys) screen->info.have_EXT_graphics_pipeline_library = false; if (!screen->optimal_keys || !screen->info.have_KHR_maintenance5 || /* EXT_shader_object needs either dynamic feedback loop or per-app enablement */ (!screen->driconf.zink_shader_object_enable && !screen->info.have_EXT_attachment_feedback_loop_dynamic_state)) screen->info.have_EXT_shader_object = false; if (screen->info.have_EXT_shader_object) screen->have_full_ds3 = true; if (zink_debug & ZINK_DEBUG_DGC) { if (!screen->optimal_keys) { mesa_loge("zink: can't DGC without optimal_keys!"); zink_debug &= ~ZINK_DEBUG_DGC; } else { screen->info.have_EXT_multi_draw = false; screen->info.have_EXT_shader_object = false; screen->info.have_EXT_graphics_pipeline_library = false; screen->info.have_EXT_vertex_input_dynamic_state = false; } } } static struct disk_cache * zink_get_disk_shader_cache(struct pipe_screen *_screen) { struct zink_screen *screen = zink_screen(_screen); return screen->disk_cache; } VkSemaphore zink_create_semaphore(struct zink_screen *screen) { VkSemaphoreCreateInfo sci = { VK_STRUCTURE_TYPE_SEMAPHORE_CREATE_INFO, NULL, 0 }; VkSemaphore sem = VK_NULL_HANDLE; if (util_dynarray_contains(&screen->semaphores, VkSemaphore)) { simple_mtx_lock(&screen->semaphores_lock); if (util_dynarray_contains(&screen->semaphores, VkSemaphore)) sem = util_dynarray_pop(&screen->semaphores, VkSemaphore); simple_mtx_unlock(&screen->semaphores_lock); } if (sem) return sem; VkResult ret = VKSCR(CreateSemaphore)(screen->dev, &sci, NULL, &sem); return ret == VK_SUCCESS ? sem : VK_NULL_HANDLE; } void zink_screen_lock_context(struct zink_screen *screen) { simple_mtx_lock(&screen->copy_context_lock); if (!screen->copy_context) screen->copy_context = zink_context(screen->base.context_create(&screen->base, NULL, ZINK_CONTEXT_COPY_ONLY)); if (!screen->copy_context) { mesa_loge("zink: failed to create copy context"); /* realistically there's nothing that can be done here */ } } void zink_screen_unlock_context(struct zink_screen *screen) { simple_mtx_unlock(&screen->copy_context_lock); } static bool init_layouts(struct zink_screen *screen) { if (screen->info.have_EXT_descriptor_indexing) { VkDescriptorSetLayoutBinding bindings[4]; const unsigned num_bindings = 4; VkDescriptorSetLayoutCreateInfo dcslci = {0}; dcslci.sType = VK_STRUCTURE_TYPE_DESCRIPTOR_SET_LAYOUT_CREATE_INFO; dcslci.pNext = NULL; VkDescriptorSetLayoutBindingFlagsCreateInfo fci = {0}; VkDescriptorBindingFlags flags[4]; dcslci.pNext = &fci; if (zink_descriptor_mode == ZINK_DESCRIPTOR_MODE_DB) dcslci.flags = VK_DESCRIPTOR_SET_LAYOUT_CREATE_DESCRIPTOR_BUFFER_BIT_EXT; else dcslci.flags = VK_DESCRIPTOR_SET_LAYOUT_CREATE_UPDATE_AFTER_BIND_POOL_BIT; fci.sType = VK_STRUCTURE_TYPE_DESCRIPTOR_SET_LAYOUT_BINDING_FLAGS_CREATE_INFO; fci.bindingCount = num_bindings; fci.pBindingFlags = flags; for (unsigned i = 0; i < num_bindings; i++) { flags[i] = VK_DESCRIPTOR_BINDING_UPDATE_UNUSED_WHILE_PENDING_BIT | VK_DESCRIPTOR_BINDING_PARTIALLY_BOUND_BIT; if (zink_descriptor_mode != ZINK_DESCRIPTOR_MODE_DB) flags[i] |= VK_DESCRIPTOR_BINDING_UPDATE_AFTER_BIND_BIT; } /* there is exactly 1 bindless descriptor set per context, and it has 4 bindings, 1 for each descriptor type */ for (unsigned i = 0; i < num_bindings; i++) { bindings[i].binding = i; bindings[i].descriptorType = zink_descriptor_type_from_bindless_index(i); bindings[i].descriptorCount = ZINK_MAX_BINDLESS_HANDLES; bindings[i].stageFlags = VK_SHADER_STAGE_ALL_GRAPHICS | VK_SHADER_STAGE_COMPUTE_BIT; bindings[i].pImmutableSamplers = NULL; } dcslci.bindingCount = num_bindings; dcslci.pBindings = bindings; VkResult result = VKSCR(CreateDescriptorSetLayout)(screen->dev, &dcslci, 0, &screen->bindless_layout); if (result != VK_SUCCESS) { mesa_loge("ZINK: vkCreateDescriptorSetLayout failed (%s)", vk_Result_to_str(result)); return false; } } screen->gfx_push_constant_layout = zink_pipeline_layout_create(screen, NULL, 0, false, 0); return !!screen->gfx_push_constant_layout; } static int zink_screen_get_fd(struct pipe_screen *pscreen) { struct zink_screen *screen = zink_screen(pscreen); return screen->drm_fd; } static struct zink_screen * zink_internal_create_screen(const struct pipe_screen_config *config, int64_t dev_major, int64_t dev_minor) { if (getenv("ZINK_USE_LAVAPIPE")) { mesa_loge("ZINK_USE_LAVAPIPE is obsolete. Use LIBGL_ALWAYS_SOFTWARE\n"); return NULL; } struct zink_screen *screen = rzalloc(NULL, struct zink_screen); if (!screen) { mesa_loge("ZINK: failed to allocate screen"); return NULL; } screen->drm_fd = -1; glsl_type_singleton_init_or_ref(); zink_debug = debug_get_option_zink_debug(); if (zink_descriptor_mode == ZINK_DESCRIPTOR_MODE_AUTO) zink_descriptor_mode = debug_get_option_zink_descriptor_mode(); screen->threaded = util_get_cpu_caps()->nr_cpus > 1 && debug_get_bool_option("GALLIUM_THREAD", util_get_cpu_caps()->nr_cpus > 1); if (zink_debug & ZINK_DEBUG_FLUSHSYNC) screen->threaded_submit = false; else screen->threaded_submit = screen->threaded; screen->abort_on_hang = debug_get_bool_option("ZINK_HANG_ABORT", false); u_trace_state_init(); screen->loader_lib = util_dl_open(VK_LIBNAME); if (!screen->loader_lib) { mesa_loge("ZINK: failed to load "VK_LIBNAME); goto fail; } screen->vk_GetInstanceProcAddr = (PFN_vkGetInstanceProcAddr)util_dl_get_proc_address(screen->loader_lib, "vkGetInstanceProcAddr"); screen->vk_GetDeviceProcAddr = (PFN_vkGetDeviceProcAddr)util_dl_get_proc_address(screen->loader_lib, "vkGetDeviceProcAddr"); if (!screen->vk_GetInstanceProcAddr || !screen->vk_GetDeviceProcAddr) { mesa_loge("ZINK: failed to get proc address"); goto fail; } screen->instance_info.loader_version = zink_get_loader_version(screen); if (config) { driParseConfigFiles(config->options, config->options_info, 0, "zink", NULL, NULL, NULL, 0, NULL, 0); screen->driconf.dual_color_blend_by_location = driQueryOptionb(config->options, "dual_color_blend_by_location"); //screen->driconf.inline_uniforms = driQueryOptionb(config->options, "radeonsi_inline_uniforms"); screen->driconf.emulate_point_smooth = driQueryOptionb(config->options, "zink_emulate_point_smooth"); screen->driconf.zink_shader_object_enable = driQueryOptionb(config->options, "zink_shader_object_enable"); } if (!zink_create_instance(screen, dev_major > 0 && dev_major < 255)) goto fail; if (zink_debug & ZINK_DEBUG_VALIDATION) { if (!screen->instance_info.have_layer_KHRONOS_validation && !screen->instance_info.have_layer_LUNARG_standard_validation) { mesa_loge("Failed to load validation layer"); goto fail; } } vk_instance_uncompacted_dispatch_table_load(&screen->vk.instance, screen->vk_GetInstanceProcAddr, screen->instance); vk_physical_device_uncompacted_dispatch_table_load(&screen->vk.physical_device, screen->vk_GetInstanceProcAddr, screen->instance); zink_verify_instance_extensions(screen); if (screen->instance_info.have_EXT_debug_utils && (zink_debug & ZINK_DEBUG_VALIDATION) && !create_debug(screen)) debug_printf("ZINK: failed to setup debug utils\n"); choose_pdev(screen, dev_major, dev_minor); if (screen->pdev == VK_NULL_HANDLE) { mesa_loge("ZINK: failed to choose pdev"); goto fail; } screen->is_cpu = screen->info.props.deviceType == VK_PHYSICAL_DEVICE_TYPE_CPU; update_queue_props(screen); screen->have_X8_D24_UNORM_PACK32 = zink_is_depth_format_supported(screen, VK_FORMAT_X8_D24_UNORM_PACK32); screen->have_D24_UNORM_S8_UINT = zink_is_depth_format_supported(screen, VK_FORMAT_D24_UNORM_S8_UINT); screen->have_D32_SFLOAT_S8_UINT = zink_is_depth_format_supported(screen, VK_FORMAT_D32_SFLOAT_S8_UINT); if (!zink_get_physical_device_info(screen)) { debug_printf("ZINK: failed to detect features\n"); goto fail; } memset(&screen->heap_map, UINT8_MAX, sizeof(screen->heap_map)); for (enum zink_heap i = 0; i < ZINK_HEAP_MAX; i++) { for (unsigned j = 0; j < screen->info.mem_props.memoryTypeCount; j++) { VkMemoryPropertyFlags domains = vk_domain_from_heap(i); if ((screen->info.mem_props.memoryTypes[j].propertyFlags & domains) == domains) { screen->heap_map[i][screen->heap_count[i]++] = j; } } } /* iterate again to check for missing heaps */ for (enum zink_heap i = 0; i < ZINK_HEAP_MAX; i++) { /* not found: use compatible heap */ if (screen->heap_map[i][0] == UINT8_MAX) { /* only cached mem has a failure case for now */ assert(i == ZINK_HEAP_HOST_VISIBLE_COHERENT_CACHED || i == ZINK_HEAP_DEVICE_LOCAL_LAZY || i == ZINK_HEAP_DEVICE_LOCAL_VISIBLE); if (i == ZINK_HEAP_HOST_VISIBLE_COHERENT_CACHED) { memcpy(screen->heap_map[i], screen->heap_map[ZINK_HEAP_HOST_VISIBLE_COHERENT], screen->heap_count[ZINK_HEAP_HOST_VISIBLE_COHERENT]); screen->heap_count[i] = screen->heap_count[ZINK_HEAP_HOST_VISIBLE_COHERENT]; } else { memcpy(screen->heap_map[i], screen->heap_map[ZINK_HEAP_DEVICE_LOCAL], screen->heap_count[ZINK_HEAP_DEVICE_LOCAL]); screen->heap_count[i] = screen->heap_count[ZINK_HEAP_DEVICE_LOCAL]; } } } { uint64_t biggest_vis_vram = 0; for (unsigned i = 0; i < screen->heap_count[ZINK_HEAP_DEVICE_LOCAL_VISIBLE]; i++) biggest_vis_vram = MAX2(biggest_vis_vram, screen->info.mem_props.memoryHeaps[screen->info.mem_props.memoryTypes[screen->heap_map[ZINK_HEAP_DEVICE_LOCAL_VISIBLE][i]].heapIndex].size); uint64_t biggest_vram = 0; for (unsigned i = 0; i < screen->heap_count[ZINK_HEAP_DEVICE_LOCAL]; i++) biggest_vram = MAX2(biggest_vram, screen->info.mem_props.memoryHeaps[screen->info.mem_props.memoryTypes[screen->heap_map[ZINK_HEAP_DEVICE_LOCAL][i]].heapIndex].size); /* determine if vis vram is roughly equal to total vram */ if (biggest_vis_vram > biggest_vram * 0.9) screen->resizable_bar = true; } setup_renderdoc(screen); if (screen->threaded_submit && !util_queue_init(&screen->flush_queue, "zfq", 8, 1, UTIL_QUEUE_INIT_RESIZE_IF_FULL, screen)) { mesa_loge("zink: Failed to create flush queue.\n"); goto fail; } zink_internal_setup_moltenvk(screen); if (!screen->info.have_KHR_timeline_semaphore && !screen->info.feats12.timelineSemaphore) { mesa_loge("zink: KHR_timeline_semaphore is required"); goto fail; } if (zink_debug & ZINK_DEBUG_DGC) { if (!screen->info.have_NV_device_generated_commands) { mesa_loge("zink: can't use DGC without NV_device_generated_commands"); goto fail; } } if (zink_debug & ZINK_DEBUG_MEM) { simple_mtx_init(&screen->debug_mem_lock, mtx_plain); screen->debug_mem_sizes = _mesa_hash_table_create(screen, _mesa_hash_string, _mesa_key_string_equal); } fixup_driver_props(screen); init_driver_workarounds(screen); screen->dev = zink_create_logical_device(screen); if (!screen->dev) goto fail; vk_device_uncompacted_dispatch_table_load(&screen->vk.device, screen->vk_GetDeviceProcAddr, screen->dev); init_queue(screen); zink_verify_device_extensions(screen); /* descriptor set indexing is determined by 'compact' descriptor mode: * by default, 6 sets are used to provide more granular updating * in compact mode, a maximum of 4 sets are used, with like-types combined */ if ((zink_debug & ZINK_DEBUG_COMPACT) || screen->info.props.limits.maxBoundDescriptorSets < ZINK_MAX_DESCRIPTOR_SETS) { screen->desc_set_id[ZINK_DESCRIPTOR_TYPE_UNIFORMS] = 0; screen->desc_set_id[ZINK_DESCRIPTOR_TYPE_UBO] = 1; screen->desc_set_id[ZINK_DESCRIPTOR_TYPE_SSBO] = 1; screen->desc_set_id[ZINK_DESCRIPTOR_TYPE_SAMPLER_VIEW] = 2; screen->desc_set_id[ZINK_DESCRIPTOR_TYPE_IMAGE] = 2; screen->desc_set_id[ZINK_DESCRIPTOR_BINDLESS] = 3; screen->compact_descriptors = true; } else { screen->desc_set_id[ZINK_DESCRIPTOR_TYPE_UNIFORMS] = 0; screen->desc_set_id[ZINK_DESCRIPTOR_TYPE_UBO] = 1; screen->desc_set_id[ZINK_DESCRIPTOR_TYPE_SAMPLER_VIEW] = 2; screen->desc_set_id[ZINK_DESCRIPTOR_TYPE_SSBO] = 3; screen->desc_set_id[ZINK_DESCRIPTOR_TYPE_IMAGE] = 4; screen->desc_set_id[ZINK_DESCRIPTOR_BINDLESS] = 5; } if (screen->info.have_EXT_calibrated_timestamps && !check_have_device_time(screen)) goto fail; screen->have_triangle_fans = true; #if defined(VK_KHR_PORTABILITY_SUBSET_EXTENSION_NAME) if (screen->info.have_KHR_portability_subset) { screen->have_triangle_fans = (VK_TRUE == screen->info.portability_subset_feats.triangleFans); } #endif // VK_KHR_PORTABILITY_SUBSET_EXTENSION_NAME check_base_requirements(screen); util_live_shader_cache_init(&screen->shaders, zink_create_gfx_shader_state, zink_delete_shader_state); screen->base.get_name = zink_get_name; if (screen->instance_info.have_KHR_external_memory_capabilities) { screen->base.get_device_uuid = zink_get_device_uuid; screen->base.get_driver_uuid = zink_get_driver_uuid; } if (screen->info.have_KHR_external_memory_win32) { screen->base.get_device_luid = zink_get_device_luid; screen->base.get_device_node_mask = zink_get_device_node_mask; } screen->base.set_max_shader_compiler_threads = zink_set_max_shader_compiler_threads; screen->base.is_parallel_shader_compilation_finished = zink_is_parallel_shader_compilation_finished; screen->base.get_vendor = zink_get_vendor; screen->base.get_device_vendor = zink_get_device_vendor; screen->base.get_compute_param = zink_get_compute_param; screen->base.get_timestamp = zink_get_timestamp; screen->base.query_memory_info = zink_query_memory_info; screen->base.get_param = zink_get_param; screen->base.get_paramf = zink_get_paramf; screen->base.get_shader_param = zink_get_shader_param; screen->base.get_compiler_options = zink_get_compiler_options; screen->base.get_sample_pixel_grid = zink_get_sample_pixel_grid; screen->base.is_compute_copy_faster = zink_is_compute_copy_faster; screen->base.is_format_supported = zink_is_format_supported; screen->base.driver_thread_add_job = zink_driver_thread_add_job; if (screen->info.have_EXT_image_drm_format_modifier && screen->info.have_EXT_external_memory_dma_buf) { screen->base.query_dmabuf_modifiers = zink_query_dmabuf_modifiers; screen->base.is_dmabuf_modifier_supported = zink_is_dmabuf_modifier_supported; screen->base.get_dmabuf_modifier_planes = zink_get_dmabuf_modifier_planes; } #if defined(_WIN32) if (screen->info.have_KHR_external_memory_win32) screen->base.create_fence_win32 = zink_create_fence_win32; #endif screen->base.context_create = zink_context_create; screen->base.flush_frontbuffer = zink_flush_frontbuffer; screen->base.destroy = zink_destroy_screen; screen->base.finalize_nir = zink_shader_finalize; screen->base.get_disk_shader_cache = zink_get_disk_shader_cache; screen->base.get_sparse_texture_virtual_page_size = zink_get_sparse_texture_virtual_page_size; screen->base.get_driver_query_group_info = zink_get_driver_query_group_info; screen->base.get_driver_query_info = zink_get_driver_query_info; screen->base.set_damage_region = zink_set_damage_region; if (screen->info.have_EXT_sample_locations) { VkMultisamplePropertiesEXT prop; prop.sType = VK_STRUCTURE_TYPE_MULTISAMPLE_PROPERTIES_EXT; prop.pNext = NULL; for (unsigned i = 0; i < ARRAY_SIZE(screen->maxSampleLocationGridSize); i++) { if (screen->info.sample_locations_props.sampleLocationSampleCounts & (1 << i)) { VKSCR(GetPhysicalDeviceMultisamplePropertiesEXT)(screen->pdev, 1 << i, &prop); screen->maxSampleLocationGridSize[i] = prop.maxSampleLocationGridSize; } } } if (!zink_screen_resource_init(&screen->base)) goto fail; if (!zink_bo_init(screen)) { mesa_loge("ZINK: failed to initialize suballocator"); goto fail; } zink_screen_fence_init(&screen->base); zink_screen_init_compiler(screen); if (!disk_cache_init(screen)) { mesa_loge("ZINK: failed to initialize disk cache"); goto fail; } if (!util_queue_init(&screen->cache_get_thread, "zcfq", 8, 4, UTIL_QUEUE_INIT_RESIZE_IF_FULL, screen)) goto fail; populate_format_props(screen); slab_create_parent(&screen->transfer_pool, sizeof(struct zink_transfer), 16); slab_create(&screen->present_mempool, sizeof(struct zink_kopper_present_info), 16); screen->driconf.inline_uniforms = debug_get_bool_option("ZINK_INLINE_UNIFORMS", screen->is_cpu) && !(zink_debug & ZINK_DEBUG_DGC); screen->total_video_mem = get_video_mem(screen); screen->clamp_video_mem = screen->total_video_mem * 0.8; if (!os_get_total_physical_memory(&screen->total_mem)) { mesa_loge("ZINK: failed to get total physical memory"); goto fail; } if (!zink_screen_init_semaphore(screen)) { mesa_loge("zink: failed to create timeline semaphore"); goto fail; } bool can_db = true; { if (!screen->info.have_EXT_descriptor_buffer) { if (zink_descriptor_mode == ZINK_DESCRIPTOR_MODE_DB) { mesa_loge("Cannot use db descriptor mode without EXT_descriptor_buffer"); goto fail; } can_db = false; } if (!screen->resizable_bar) { if (zink_descriptor_mode == ZINK_DESCRIPTOR_MODE_DB) { mesa_loge("Cannot use db descriptor mode without resizable bar"); goto fail; } can_db = false; } if (!screen->info.have_EXT_non_seamless_cube_map) { if (zink_descriptor_mode == ZINK_DESCRIPTOR_MODE_DB) { mesa_loge("Cannot use db descriptor mode without EXT_non_seamless_cube_map"); goto fail; } can_db = false; } if (!screen->info.rb2_feats.nullDescriptor) { if (zink_descriptor_mode == ZINK_DESCRIPTOR_MODE_DB) { mesa_loge("Cannot use db descriptor mode without robustness2.nullDescriptor"); goto fail; } can_db = false; } if (ZINK_FBFETCH_DESCRIPTOR_SIZE < screen->info.db_props.inputAttachmentDescriptorSize) { if (zink_descriptor_mode == ZINK_DESCRIPTOR_MODE_DB) { mesa_loge("Cannot use db descriptor mode with inputAttachmentDescriptorSize(%u) > %u", (unsigned)screen->info.db_props.inputAttachmentDescriptorSize, ZINK_FBFETCH_DESCRIPTOR_SIZE); goto fail; } mesa_logw("zink: bug detected: inputAttachmentDescriptorSize(%u) > %u", (unsigned)screen->info.db_props.inputAttachmentDescriptorSize, ZINK_FBFETCH_DESCRIPTOR_SIZE); can_db = false; } if (screen->info.db_props.maxDescriptorBufferBindings < 2 || screen->info.db_props.maxSamplerDescriptorBufferBindings < 2) { if (zink_descriptor_mode == ZINK_DESCRIPTOR_MODE_DB) { /* allow for testing, but disable bindless */ mesa_logw("Cannot use bindless and db descriptor mode with (maxDescriptorBufferBindings||maxSamplerDescriptorBufferBindings) < 2"); } else { can_db = false; } } } if (zink_descriptor_mode == ZINK_DESCRIPTOR_MODE_AUTO) { /* descriptor buffer is not performant with virt yet */ if (screen->info.driver_props.driverID == VK_DRIVER_ID_MESA_VENUS) zink_descriptor_mode = ZINK_DESCRIPTOR_MODE_LAZY; else zink_descriptor_mode = can_db ? ZINK_DESCRIPTOR_MODE_DB : ZINK_DESCRIPTOR_MODE_LAZY; } if (zink_descriptor_mode == ZINK_DESCRIPTOR_MODE_DB) { const uint32_t sampler_size = MAX2(screen->info.db_props.combinedImageSamplerDescriptorSize, screen->info.db_props.robustUniformTexelBufferDescriptorSize); const uint32_t image_size = MAX2(screen->info.db_props.storageImageDescriptorSize, screen->info.db_props.robustStorageTexelBufferDescriptorSize); if (screen->compact_descriptors) { screen->db_size[ZINK_DESCRIPTOR_TYPE_UBO] = screen->info.db_props.robustUniformBufferDescriptorSize + screen->info.db_props.robustStorageBufferDescriptorSize; screen->db_size[ZINK_DESCRIPTOR_TYPE_SAMPLER_VIEW] = sampler_size + image_size; } else { screen->db_size[ZINK_DESCRIPTOR_TYPE_UBO] = screen->info.db_props.robustUniformBufferDescriptorSize; screen->db_size[ZINK_DESCRIPTOR_TYPE_SAMPLER_VIEW] = sampler_size; screen->db_size[ZINK_DESCRIPTOR_TYPE_SSBO] = screen->info.db_props.robustStorageBufferDescriptorSize; screen->db_size[ZINK_DESCRIPTOR_TYPE_IMAGE] = image_size; } screen->db_size[ZINK_DESCRIPTOR_TYPE_UNIFORMS] = screen->info.db_props.robustUniformBufferDescriptorSize; screen->info.have_KHR_push_descriptor = false; screen->base_descriptor_size = MAX4(screen->db_size[0], screen->db_size[1], screen->db_size[2], screen->db_size[3]); } simple_mtx_init(&screen->free_batch_states_lock, mtx_plain); simple_mtx_init(&screen->dt_lock, mtx_plain); util_idalloc_mt_init_tc(&screen->buffer_ids); simple_mtx_init(&screen->semaphores_lock, mtx_plain); util_dynarray_init(&screen->semaphores, screen); util_dynarray_init(&screen->fd_semaphores, screen); util_vertex_state_cache_init(&screen->vertex_state_cache, zink_create_vertex_state, zink_vertex_state_destroy); screen->base.create_vertex_state = zink_cache_create_vertex_state; screen->base.vertex_state_destroy = zink_cache_vertex_state_destroy; zink_synchronization_init(screen); zink_init_screen_pipeline_libs(screen); if (!init_layouts(screen)) { mesa_loge("ZINK: failed to initialize layouts"); goto fail; } if (!zink_descriptor_layouts_init(screen)) { mesa_loge("ZINK: failed to initialize descriptor layouts"); goto fail; } simple_mtx_init(&screen->copy_context_lock, mtx_plain); init_optimal_keys(screen); screen->screen_id = p_atomic_inc_return(&num_screens); zink_tracing = screen->instance_info.have_EXT_debug_utils && (u_trace_is_enabled(U_TRACE_TYPE_PERFETTO) || u_trace_is_enabled(U_TRACE_TYPE_MARKERS)); screen->frame_marker_emitted = zink_screen_debug_marker_begin(screen, "frame"); return screen; fail: zink_destroy_screen(&screen->base); return NULL; } struct pipe_screen * zink_create_screen(struct sw_winsys *winsys, const struct pipe_screen_config *config) { struct zink_screen *ret = zink_internal_create_screen(config, -1, -1); if (ret) { ret->drm_fd = -1; } return &ret->base; } static inline int zink_render_rdev(int fd, int64_t *dev_major, int64_t *dev_minor) { int ret = 0; *dev_major = *dev_minor = -1; #ifdef HAVE_LIBDRM struct stat stx; drmDevicePtr dev; if (fd == -1) return 0; if (drmGetDevice2(fd, 0, &dev)) return -1; if(!(dev->available_nodes & (1 << DRM_NODE_RENDER))) { ret = -1; goto free_device; } if(stat(dev->nodes[DRM_NODE_RENDER], &stx)) { ret = -1; goto free_device; } *dev_major = major(stx.st_rdev); *dev_minor = minor(stx.st_rdev); free_device: drmFreeDevice(&dev); #endif //HAVE_LIBDRM return ret; } struct pipe_screen * zink_drm_create_screen(int fd, const struct pipe_screen_config *config) { int64_t dev_major, dev_minor; struct zink_screen *ret; if (zink_render_rdev(fd, &dev_major, &dev_minor)) return NULL; ret = zink_internal_create_screen(config, dev_major, dev_minor); if (ret) ret->drm_fd = os_dupfd_cloexec(fd); if (ret && !ret->info.have_KHR_external_memory_fd) { debug_printf("ZINK: KHR_external_memory_fd required!\n"); zink_destroy_screen(&ret->base); return NULL; } return &ret->base; } void zink_stub_function_not_loaded() { /* this will be used by the zink_verify_*_extensions() functions on a * release build */ mesa_loge("ZINK: a Vulkan function was called without being loaded"); abort(); } bool zink_screen_debug_marker_begin(struct zink_screen *screen, const char *fmt, ...) { if (!zink_tracing) return false; char *name; va_list va; va_start(va, fmt); int ret = vasprintf(&name, fmt, va); va_end(va); if (ret == -1) return false; VkDebugUtilsLabelEXT info = { 0 }; info.sType = VK_STRUCTURE_TYPE_DEBUG_UTILS_LABEL_EXT; info.pLabelName = name; VKSCR(QueueBeginDebugUtilsLabelEXT)(screen->queue, &info); free(name); return true; } void zink_screen_debug_marker_end(struct zink_screen *screen, bool emitted) { if (emitted) VKSCR(QueueEndDebugUtilsLabelEXT)(screen->queue); }