panfrost: Make SLAB pool creation rely on BO helpers
There's no point duplicating the code, and it will help us simplify the bo_handles[] filling logic in panfrost_drm_submit_job(). Signed-off-by: Boris Brezillon <boris.brezillon@collabora.com>
This commit is contained in:
parent
c684a79669
commit
3557746e0d
|
@ -48,8 +48,8 @@ panfrost_allocate_chunk(struct panfrost_context *ctx, size_t size, unsigned heap
|
|||
struct panfrost_memory *backing = (struct panfrost_memory *) entry->slab;
|
||||
|
||||
struct panfrost_transfer transfer = {
|
||||
.cpu = backing->cpu + p_entry->offset,
|
||||
.gpu = backing->gpu + p_entry->offset
|
||||
.cpu = backing->bo->cpu + p_entry->offset,
|
||||
.gpu = backing->bo->gpu + p_entry->offset
|
||||
};
|
||||
|
||||
return transfer;
|
||||
|
@ -97,8 +97,8 @@ panfrost_allocate_transient(struct panfrost_context *ctx, size_t sz)
|
|||
struct panfrost_memory *backing = (struct panfrost_memory *) p_entry->base.slab;
|
||||
|
||||
struct panfrost_transfer ret = {
|
||||
.cpu = backing->cpu + p_entry->offset + pool->entry_offset,
|
||||
.gpu = backing->gpu + p_entry->offset + pool->entry_offset
|
||||
.cpu = backing->bo->cpu + p_entry->offset + pool->entry_offset,
|
||||
.gpu = backing->bo->gpu + p_entry->offset + pool->entry_offset
|
||||
};
|
||||
|
||||
/* Advance the pointer */
|
||||
|
@ -192,18 +192,19 @@ mali_ptr
|
|||
panfrost_upload(struct panfrost_memory *mem, const void *data, size_t sz, bool no_pad)
|
||||
{
|
||||
/* Bounds check */
|
||||
if ((mem->stack_bottom + sz) >= mem->size) {
|
||||
printf("Out of memory, tried to upload %zd but only %zd available\n", sz, mem->size - mem->stack_bottom);
|
||||
if ((mem->stack_bottom + sz) >= mem->bo->size) {
|
||||
printf("Out of memory, tried to upload %zd but only %zd available\n",
|
||||
sz, mem->bo->size - mem->stack_bottom);
|
||||
assert(0);
|
||||
}
|
||||
|
||||
return pandev_upload(-1, &mem->stack_bottom, mem->gpu, mem->cpu, data, sz, no_pad);
|
||||
return pandev_upload(-1, &mem->stack_bottom, mem->bo->gpu, mem->bo->cpu, data, sz, no_pad);
|
||||
}
|
||||
|
||||
mali_ptr
|
||||
panfrost_upload_sequential(struct panfrost_memory *mem, const void *data, size_t sz)
|
||||
{
|
||||
return pandev_upload(last_offset, &mem->stack_bottom, mem->gpu, mem->cpu, data, sz, true);
|
||||
return pandev_upload(last_offset, &mem->stack_bottom, mem->bo->gpu, mem->bo->cpu, data, sz, true);
|
||||
}
|
||||
|
||||
/* Simplified interface to allocate a chunk without any upload, to allow
|
||||
|
@ -215,6 +216,6 @@ panfrost_allocate_transfer(struct panfrost_memory *mem, size_t sz, mali_ptr *gpu
|
|||
{
|
||||
int offset = pandev_allocate_offset(&mem->stack_bottom, sz);
|
||||
|
||||
*gpu = mem->gpu + offset;
|
||||
return mem->cpu + offset;
|
||||
*gpu = mem->bo->gpu + offset;
|
||||
return mem->bo->cpu + offset;
|
||||
}
|
||||
|
|
|
@ -58,16 +58,28 @@ struct panfrost_transfer {
|
|||
mali_ptr gpu;
|
||||
};
|
||||
|
||||
struct panfrost_bo {
|
||||
struct pipe_reference reference;
|
||||
|
||||
/* Mapping for the entire object (all levels) */
|
||||
uint8_t *cpu;
|
||||
|
||||
/* GPU address for the object */
|
||||
mali_ptr gpu;
|
||||
|
||||
/* Size of all entire trees */
|
||||
size_t size;
|
||||
|
||||
int gem_handle;
|
||||
};
|
||||
|
||||
struct panfrost_memory {
|
||||
/* Subclassing slab object */
|
||||
struct pb_slab slab;
|
||||
|
||||
/* Backing for the slab in memory */
|
||||
uint8_t *cpu;
|
||||
mali_ptr gpu;
|
||||
struct panfrost_bo *bo;
|
||||
int stack_bottom;
|
||||
size_t size;
|
||||
int gem_handle;
|
||||
};
|
||||
|
||||
/* Slab entry sizes range from 2^min to 2^max. In this case, we range from 1k
|
||||
|
@ -109,7 +121,7 @@ static inline mali_ptr
|
|||
panfrost_reserve(struct panfrost_memory *mem, size_t sz)
|
||||
{
|
||||
mem->stack_bottom += sz;
|
||||
return mem->gpu + (mem->stack_bottom - sz);
|
||||
return mem->bo->gpu + (mem->stack_bottom - sz);
|
||||
}
|
||||
|
||||
struct panfrost_transfer
|
||||
|
|
|
@ -99,13 +99,13 @@ panfrost_emit_sfbd(struct panfrost_context *ctx, unsigned vertex_count)
|
|||
.unknown2 = 0x1f,
|
||||
.format = 0x30000000,
|
||||
.clear_flags = 0x1000,
|
||||
.unknown_address_0 = ctx->scratchpad.gpu,
|
||||
.tiler_polygon_list = ctx->tiler_polygon_list.gpu,
|
||||
.tiler_polygon_list_body = ctx->tiler_polygon_list.gpu + 40960,
|
||||
.unknown_address_0 = ctx->scratchpad.bo->gpu,
|
||||
.tiler_polygon_list = ctx->tiler_polygon_list.bo->gpu,
|
||||
.tiler_polygon_list_body = ctx->tiler_polygon_list.bo->gpu + 40960,
|
||||
.tiler_hierarchy_mask = 0xF0,
|
||||
.tiler_flags = 0x0,
|
||||
.tiler_heap_free = ctx->tiler_heap.gpu,
|
||||
.tiler_heap_end = ctx->tiler_heap.gpu + ctx->tiler_heap.size,
|
||||
.tiler_heap_free = ctx->tiler_heap.bo->gpu,
|
||||
.tiler_heap_end = ctx->tiler_heap.bo->gpu + ctx->tiler_heap.bo->size,
|
||||
};
|
||||
|
||||
panfrost_set_framebuffer_resolution(&framebuffer, ctx->pipe_framebuffer.width, ctx->pipe_framebuffer.height);
|
||||
|
@ -133,7 +133,7 @@ panfrost_emit_mfbd(struct panfrost_context *ctx, unsigned vertex_count)
|
|||
|
||||
.unknown2 = 0x1f,
|
||||
|
||||
.scratchpad = ctx->scratchpad.gpu,
|
||||
.scratchpad = ctx->scratchpad.bo->gpu,
|
||||
};
|
||||
|
||||
framebuffer.tiler_hierarchy_mask =
|
||||
|
@ -152,22 +152,22 @@ panfrost_emit_mfbd(struct panfrost_context *ctx, unsigned vertex_count)
|
|||
unsigned total_size = header_size + body_size;
|
||||
|
||||
if (framebuffer.tiler_hierarchy_mask) {
|
||||
assert(ctx->tiler_polygon_list.size >= total_size);
|
||||
assert(ctx->tiler_polygon_list.bo->size >= total_size);
|
||||
|
||||
/* Specify allocated tiler structures */
|
||||
framebuffer.tiler_polygon_list = ctx->tiler_polygon_list.gpu;
|
||||
framebuffer.tiler_polygon_list = ctx->tiler_polygon_list.bo->gpu;
|
||||
|
||||
/* Allow the entire tiler heap */
|
||||
framebuffer.tiler_heap_start = ctx->tiler_heap.gpu;
|
||||
framebuffer.tiler_heap_start = ctx->tiler_heap.bo->gpu;
|
||||
framebuffer.tiler_heap_end =
|
||||
ctx->tiler_heap.gpu + ctx->tiler_heap.size;
|
||||
ctx->tiler_heap.bo->gpu + ctx->tiler_heap.bo->size;
|
||||
} else {
|
||||
/* The tiler is disabled, so don't allow the tiler heap */
|
||||
framebuffer.tiler_heap_start = ctx->tiler_heap.gpu;
|
||||
framebuffer.tiler_heap_start = ctx->tiler_heap.bo->gpu;
|
||||
framebuffer.tiler_heap_end = framebuffer.tiler_heap_start;
|
||||
|
||||
/* Use a dummy polygon list */
|
||||
framebuffer.tiler_polygon_list = ctx->tiler_dummy.gpu;
|
||||
framebuffer.tiler_polygon_list = ctx->tiler_dummy.bo->gpu;
|
||||
|
||||
/* Also, set a "tiler disabled?" flag? */
|
||||
framebuffer.tiler_hierarchy_mask |= 0x1000;
|
||||
|
@ -529,7 +529,7 @@ panfrost_emit_varyings(
|
|||
unsigned stride,
|
||||
unsigned count)
|
||||
{
|
||||
mali_ptr varying_address = ctx->varying_mem.gpu + ctx->varying_height;
|
||||
mali_ptr varying_address = ctx->varying_mem.bo->gpu + ctx->varying_height;
|
||||
|
||||
/* Fill out the descriptor */
|
||||
slot->elements = varying_address | MALI_ATTR_LINEAR;
|
||||
|
@ -537,7 +537,7 @@ panfrost_emit_varyings(
|
|||
slot->size = stride * count;
|
||||
|
||||
ctx->varying_height += ALIGN(slot->size, 64);
|
||||
assert(ctx->varying_height < ctx->varying_mem.size);
|
||||
assert(ctx->varying_height < ctx->varying_mem.bo->size);
|
||||
|
||||
return varying_address;
|
||||
}
|
||||
|
|
|
@ -136,70 +136,18 @@ panfrost_drm_allocate_slab(struct panfrost_screen *screen,
|
|||
int commit_count,
|
||||
int extent)
|
||||
{
|
||||
struct drm_panfrost_create_bo create_bo = {
|
||||
.size = pages * 4096,
|
||||
.flags = 0, // TODO figure out proper flags..
|
||||
};
|
||||
struct drm_panfrost_mmap_bo mmap_bo = {0,};
|
||||
int ret;
|
||||
|
||||
// TODO cache allocations
|
||||
// TODO properly handle errors
|
||||
// TODO take into account extra_flags
|
||||
|
||||
ret = drmIoctl(screen->fd, DRM_IOCTL_PANFROST_CREATE_BO, &create_bo);
|
||||
if (ret) {
|
||||
fprintf(stderr, "DRM_IOCTL_PANFROST_CREATE_BO failed: %d\n", ret);
|
||||
assert(0);
|
||||
}
|
||||
|
||||
mem->gpu = create_bo.offset;
|
||||
mem->gem_handle = create_bo.handle;
|
||||
// TODO cache allocations
|
||||
// TODO properly handle errors
|
||||
// TODO take into account extra_flags
|
||||
mem->bo = panfrost_drm_create_bo(screen, pages * 4096, 0);
|
||||
mem->stack_bottom = 0;
|
||||
mem->size = create_bo.size;
|
||||
|
||||
// TODO map and unmap on demand?
|
||||
mmap_bo.handle = create_bo.handle;
|
||||
ret = drmIoctl(screen->fd, DRM_IOCTL_PANFROST_MMAP_BO, &mmap_bo);
|
||||
if (ret) {
|
||||
fprintf(stderr, "DRM_IOCTL_PANFROST_MMAP_BO failed: %d\n", ret);
|
||||
assert(0);
|
||||
}
|
||||
|
||||
mem->cpu = os_mmap(NULL, mem->size, PROT_READ | PROT_WRITE, MAP_SHARED,
|
||||
screen->fd, mmap_bo.offset);
|
||||
if (mem->cpu == MAP_FAILED) {
|
||||
fprintf(stderr, "mmap failed: %p\n", mem->cpu);
|
||||
assert(0);
|
||||
}
|
||||
|
||||
/* Record the mmap if we're tracing */
|
||||
if (pan_debug & PAN_DBG_TRACE)
|
||||
pandecode_inject_mmap(mem->gpu, mem->cpu, mem->size, NULL);
|
||||
}
|
||||
|
||||
void
|
||||
panfrost_drm_free_slab(struct panfrost_screen *screen, struct panfrost_memory *mem)
|
||||
{
|
||||
struct drm_gem_close gem_close = {
|
||||
.handle = mem->gem_handle,
|
||||
};
|
||||
int ret;
|
||||
|
||||
if (os_munmap((void *) (uintptr_t) mem->cpu, mem->size)) {
|
||||
perror("munmap");
|
||||
abort();
|
||||
}
|
||||
|
||||
mem->cpu = NULL;
|
||||
|
||||
ret = drmIoctl(screen->fd, DRM_IOCTL_GEM_CLOSE, &gem_close);
|
||||
if (ret) {
|
||||
fprintf(stderr, "DRM_IOCTL_GEM_CLOSE failed: %d\n", ret);
|
||||
assert(0);
|
||||
}
|
||||
|
||||
mem->gem_handle = -1;
|
||||
panfrost_bo_unreference(&screen->base, mem->bo);
|
||||
mem->bo = NULL;
|
||||
}
|
||||
|
||||
struct panfrost_bo *
|
||||
|
@ -267,11 +215,11 @@ panfrost_drm_submit_job(struct panfrost_context *ctx, u64 job_desc, int reqs, st
|
|||
|
||||
/* TODO: Add here the transient pools */
|
||||
/* TODO: Add here the BOs listed in the panfrost_job */
|
||||
bo_handles[submit.bo_handle_count++] = ctx->shaders.gem_handle;
|
||||
bo_handles[submit.bo_handle_count++] = ctx->scratchpad.gem_handle;
|
||||
bo_handles[submit.bo_handle_count++] = ctx->tiler_heap.gem_handle;
|
||||
bo_handles[submit.bo_handle_count++] = ctx->varying_mem.gem_handle;
|
||||
bo_handles[submit.bo_handle_count++] = ctx->tiler_polygon_list.gem_handle;
|
||||
bo_handles[submit.bo_handle_count++] = ctx->shaders.bo->gem_handle;
|
||||
bo_handles[submit.bo_handle_count++] = ctx->scratchpad.bo->gem_handle;
|
||||
bo_handles[submit.bo_handle_count++] = ctx->tiler_heap.bo->gem_handle;
|
||||
bo_handles[submit.bo_handle_count++] = ctx->varying_mem.bo->gem_handle;
|
||||
bo_handles[submit.bo_handle_count++] = ctx->tiler_polygon_list.bo->gem_handle;
|
||||
submit.bo_handles = (u64) (uintptr_t) bo_handles;
|
||||
|
||||
if (drmIoctl(screen->fd, DRM_IOCTL_PANFROST_SUBMIT, &submit)) {
|
||||
|
|
|
@ -57,21 +57,6 @@ struct panfrost_slice {
|
|||
bool initialized;
|
||||
};
|
||||
|
||||
struct panfrost_bo {
|
||||
struct pipe_reference reference;
|
||||
|
||||
/* Mapping for the entire object (all levels) */
|
||||
uint8_t *cpu;
|
||||
|
||||
/* GPU address for the object */
|
||||
mali_ptr gpu;
|
||||
|
||||
/* Size of all entire trees */
|
||||
size_t size;
|
||||
|
||||
int gem_handle;
|
||||
};
|
||||
|
||||
void
|
||||
panfrost_bo_reference(struct panfrost_bo *bo);
|
||||
|
||||
|
|
|
@ -306,7 +306,7 @@ panfrost_scoreboard_set_value(struct panfrost_job *batch)
|
|||
/* Okay, we do. Let's generate it */
|
||||
|
||||
struct panfrost_context *ctx = batch->ctx;
|
||||
mali_ptr polygon_list = ctx->tiler_polygon_list.gpu;
|
||||
mali_ptr polygon_list = ctx->tiler_polygon_list.bo->gpu;
|
||||
|
||||
struct panfrost_transfer job =
|
||||
panfrost_set_value_job(ctx, polygon_list);
|
||||
|
|
|
@ -55,14 +55,14 @@ panfrost_sfbd_clear(
|
|||
sfbd->clear_depth_3 = job->clear_depth;
|
||||
sfbd->clear_depth_4 = job->clear_depth;
|
||||
|
||||
sfbd->depth_buffer = ctx->depth_stencil_buffer.gpu;
|
||||
sfbd->depth_buffer = ctx->depth_stencil_buffer.bo->gpu;
|
||||
sfbd->depth_buffer_enable = MALI_DEPTH_STENCIL_ENABLE;
|
||||
}
|
||||
|
||||
if (job->clear & PIPE_CLEAR_STENCIL) {
|
||||
sfbd->clear_stencil = job->clear_stencil;
|
||||
|
||||
sfbd->stencil_buffer = ctx->depth_stencil_buffer.gpu;
|
||||
sfbd->stencil_buffer = ctx->depth_stencil_buffer.bo->gpu;
|
||||
sfbd->stencil_buffer_enable = MALI_DEPTH_STENCIL_ENABLE;
|
||||
}
|
||||
|
||||
|
|
Loading…
Reference in New Issue