Merge commit 'origin/gallium-0.1' into gallium-0.2
Conflicts: src/gallium/winsys/gdi/SConscript
This commit is contained in:
commit
8abc860bd4
|
@ -61,7 +61,7 @@ def AddOptions(opts):
|
|||
opts.Add(EnumOption('platform', 'target platform', default_platform,
|
||||
allowed_values=('linux', 'cell', 'windows', 'winddk', 'wince')))
|
||||
opts.Add(EnumOption('toolchain', 'compiler toolchain', 'default',
|
||||
allowed_values=('default', 'crossmingw')))
|
||||
allowed_values=('default', 'crossmingw', 'winddk')))
|
||||
opts.Add(BoolOption('llvm', 'use LLVM', 'no'))
|
||||
opts.Add(BoolOption('dri', 'build DRI drivers', default_dri))
|
||||
|
||||
|
|
|
@ -123,7 +123,7 @@ static void offset_first_tri( struct draw_stage *stage,
|
|||
{
|
||||
struct offset_stage *offset = offset_stage(stage);
|
||||
|
||||
offset->units = stage->draw->rasterizer->offset_units * stage->draw->mrd;
|
||||
offset->units = (float) (stage->draw->rasterizer->offset_units * stage->draw->mrd);
|
||||
offset->scale = stage->draw->rasterizer->offset_scale;
|
||||
|
||||
stage->tri = offset_tri;
|
||||
|
|
|
@ -11,7 +11,6 @@ C_SOURCES = \
|
|||
pb_bufmgr_debug.c \
|
||||
pb_bufmgr_fenced.c \
|
||||
pb_bufmgr_mm.c \
|
||||
pb_bufmgr_ondemand.c \
|
||||
pb_bufmgr_pool.c \
|
||||
pb_bufmgr_slab.c \
|
||||
pb_validate.c \
|
||||
|
|
|
@ -10,7 +10,6 @@ pipebuffer = env.ConvenienceLibrary(
|
|||
'pb_bufmgr_debug.c',
|
||||
'pb_bufmgr_fenced.c',
|
||||
'pb_bufmgr_mm.c',
|
||||
'pb_bufmgr_ondemand.c',
|
||||
'pb_bufmgr_pool.c',
|
||||
'pb_bufmgr_slab.c',
|
||||
'pb_validate.c',
|
||||
|
|
|
@ -37,7 +37,7 @@
|
|||
* There is no obligation of a winsys driver to use this library. And a pipe
|
||||
* driver should be completly agnostic about it.
|
||||
*
|
||||
* \author Jose Fonseca <jrfonseca@tungstengraphics.com>
|
||||
* \author Jos<EFBFBD> Fonseca <jrfonseca@tungstengraphics.com>
|
||||
*/
|
||||
|
||||
#ifndef PB_BUFFER_H_
|
||||
|
@ -46,7 +46,6 @@
|
|||
|
||||
#include "pipe/p_compiler.h"
|
||||
#include "pipe/p_debug.h"
|
||||
#include "pipe/p_error.h"
|
||||
#include "pipe/p_state.h"
|
||||
#include "pipe/p_inlines.h"
|
||||
|
||||
|
@ -57,8 +56,6 @@ extern "C" {
|
|||
|
||||
|
||||
struct pb_vtbl;
|
||||
struct pb_validate;
|
||||
|
||||
|
||||
/**
|
||||
* Buffer description.
|
||||
|
@ -107,13 +104,6 @@ struct pb_vtbl
|
|||
|
||||
void (*unmap)( struct pb_buffer *buf );
|
||||
|
||||
enum pipe_error (*validate)( struct pb_buffer *buf,
|
||||
struct pb_validate *vl,
|
||||
unsigned flags );
|
||||
|
||||
void (*fence)( struct pb_buffer *buf,
|
||||
struct pipe_fence_handle *fence );
|
||||
|
||||
/**
|
||||
* Get the base buffer and the offset.
|
||||
*
|
||||
|
@ -128,7 +118,6 @@ struct pb_vtbl
|
|||
void (*get_base_buffer)( struct pb_buffer *buf,
|
||||
struct pb_buffer **base_buf,
|
||||
unsigned *offset );
|
||||
|
||||
};
|
||||
|
||||
|
||||
|
@ -184,33 +173,10 @@ pb_get_base_buffer( struct pb_buffer *buf,
|
|||
offset = 0;
|
||||
return;
|
||||
}
|
||||
assert(buf->vtbl->get_base_buffer);
|
||||
buf->vtbl->get_base_buffer(buf, base_buf, offset);
|
||||
}
|
||||
|
||||
|
||||
static INLINE enum pipe_error
|
||||
pb_validate(struct pb_buffer *buf, struct pb_validate *vl, unsigned flags)
|
||||
{
|
||||
assert(buf);
|
||||
if(!buf)
|
||||
return PIPE_ERROR;
|
||||
assert(buf->vtbl->validate);
|
||||
return buf->vtbl->validate(buf, vl, flags);
|
||||
}
|
||||
|
||||
|
||||
static INLINE void
|
||||
pb_fence(struct pb_buffer *buf, struct pipe_fence_handle *fence)
|
||||
{
|
||||
assert(buf);
|
||||
if(!buf)
|
||||
return;
|
||||
assert(buf->vtbl->fence);
|
||||
buf->vtbl->fence(buf, fence);
|
||||
}
|
||||
|
||||
|
||||
static INLINE void
|
||||
pb_destroy(struct pb_buffer *buf)
|
||||
{
|
||||
|
|
|
@ -29,7 +29,7 @@
|
|||
* \file
|
||||
* Implementation of fenced buffers.
|
||||
*
|
||||
* \author Jose Fonseca <jrfonseca-at-tungstengraphics-dot-com>
|
||||
* \author José Fonseca <jrfonseca-at-tungstengraphics-dot-com>
|
||||
* \author Thomas Hellström <thomas-at-tungstengraphics-dot-com>
|
||||
*/
|
||||
|
||||
|
@ -59,6 +59,13 @@
|
|||
*/
|
||||
#define SUPER(__derived) (&(__derived)->base)
|
||||
|
||||
#define PIPE_BUFFER_USAGE_CPU_READ_WRITE \
|
||||
( PIPE_BUFFER_USAGE_CPU_READ | PIPE_BUFFER_USAGE_CPU_WRITE )
|
||||
#define PIPE_BUFFER_USAGE_GPU_READ_WRITE \
|
||||
( PIPE_BUFFER_USAGE_GPU_READ | PIPE_BUFFER_USAGE_GPU_WRITE )
|
||||
#define PIPE_BUFFER_USAGE_WRITE \
|
||||
( PIPE_BUFFER_USAGE_CPU_WRITE | PIPE_BUFFER_USAGE_GPU_WRITE )
|
||||
|
||||
|
||||
struct fenced_buffer_list
|
||||
{
|
||||
|
@ -90,8 +97,6 @@ struct fenced_buffer
|
|||
unsigned flags;
|
||||
|
||||
unsigned mapcount;
|
||||
struct pb_validate *vl;
|
||||
unsigned validation_flags;
|
||||
struct pipe_fence_handle *fence;
|
||||
|
||||
struct list_head head;
|
||||
|
@ -103,6 +108,7 @@ static INLINE struct fenced_buffer *
|
|||
fenced_buffer(struct pb_buffer *buf)
|
||||
{
|
||||
assert(buf);
|
||||
assert(buf->vtbl == &fenced_buffer_vtbl);
|
||||
return (struct fenced_buffer *)buf;
|
||||
}
|
||||
|
||||
|
@ -268,7 +274,6 @@ fenced_buffer_map(struct pb_buffer *buf,
|
|||
struct fenced_buffer *fenced_buf = fenced_buffer(buf);
|
||||
void *map;
|
||||
|
||||
assert(flags & PIPE_BUFFER_USAGE_CPU_READ_WRITE);
|
||||
assert(!(flags & ~PIPE_BUFFER_USAGE_CPU_READ_WRITE));
|
||||
flags &= PIPE_BUFFER_USAGE_CPU_READ_WRITE;
|
||||
|
||||
|
@ -310,101 +315,6 @@ fenced_buffer_unmap(struct pb_buffer *buf)
|
|||
}
|
||||
|
||||
|
||||
static enum pipe_error
|
||||
fenced_buffer_validate(struct pb_buffer *buf,
|
||||
struct pb_validate *vl,
|
||||
unsigned flags)
|
||||
{
|
||||
struct fenced_buffer *fenced_buf = fenced_buffer(buf);
|
||||
enum pipe_error ret;
|
||||
|
||||
if(!vl) {
|
||||
/* invalidate */
|
||||
fenced_buf->vl = NULL;
|
||||
fenced_buf->validation_flags = 0;
|
||||
return PIPE_OK;
|
||||
}
|
||||
|
||||
assert(flags & PIPE_BUFFER_USAGE_GPU_READ_WRITE);
|
||||
assert(!(flags & ~PIPE_BUFFER_USAGE_GPU_READ_WRITE));
|
||||
flags &= PIPE_BUFFER_USAGE_GPU_READ_WRITE;
|
||||
|
||||
/* Buffer cannot be validated in two different lists */
|
||||
if(fenced_buf->vl && fenced_buf->vl != vl)
|
||||
return PIPE_ERROR_RETRY;
|
||||
|
||||
/* Do not validate if buffer is still mapped */
|
||||
if(fenced_buf->flags & PIPE_BUFFER_USAGE_CPU_READ_WRITE) {
|
||||
/* TODO: wait for the thread that mapped the buffer to unmap it */
|
||||
return PIPE_ERROR_RETRY;
|
||||
}
|
||||
|
||||
/* Allow concurrent GPU reads, but serialize GPU writes */
|
||||
if(fenced_buf->flags & PIPE_BUFFER_USAGE_GPU_READ_WRITE) {
|
||||
if((fenced_buf->flags | flags) & PIPE_BUFFER_USAGE_GPU_WRITE) {
|
||||
_fenced_buffer_finish(fenced_buf);
|
||||
}
|
||||
}
|
||||
|
||||
if(fenced_buf->vl == vl &&
|
||||
(fenced_buf->validation_flags & flags) == flags) {
|
||||
/* Nothing to do -- buffer already validated */
|
||||
return PIPE_OK;
|
||||
}
|
||||
|
||||
/* Final sanity checking */
|
||||
assert(!(fenced_buf->flags & PIPE_BUFFER_USAGE_CPU_READ_WRITE));
|
||||
assert(!(fenced_buf->flags & PIPE_BUFFER_USAGE_GPU_WRITE));
|
||||
assert(!fenced_buf->mapcount);
|
||||
|
||||
ret = pb_validate(fenced_buf->buffer, vl, flags);
|
||||
if (ret != PIPE_OK)
|
||||
return ret;
|
||||
|
||||
fenced_buf->vl = vl;
|
||||
fenced_buf->validation_flags |= flags;
|
||||
|
||||
return PIPE_OK;
|
||||
}
|
||||
|
||||
|
||||
static void
|
||||
fenced_buffer_fence(struct pb_buffer *buf,
|
||||
struct pipe_fence_handle *fence)
|
||||
{
|
||||
struct fenced_buffer *fenced_buf;
|
||||
struct fenced_buffer_list *fenced_list;
|
||||
struct pipe_winsys *winsys;
|
||||
|
||||
fenced_buf = fenced_buffer(buf);
|
||||
fenced_list = fenced_buf->list;
|
||||
winsys = fenced_list->winsys;
|
||||
|
||||
if(fence == fenced_buf->fence) {
|
||||
/* Nothing to do */
|
||||
return;
|
||||
}
|
||||
|
||||
assert(fenced_buf->vl);
|
||||
assert(fenced_buf->validation_flags);
|
||||
|
||||
pipe_mutex_lock(fenced_list->mutex);
|
||||
if (fenced_buf->fence)
|
||||
_fenced_buffer_remove(fenced_list, fenced_buf);
|
||||
if (fence) {
|
||||
winsys->fence_reference(winsys, &fenced_buf->fence, fence);
|
||||
fenced_buf->flags |= fenced_buf->validation_flags;
|
||||
_fenced_buffer_add(fenced_buf);
|
||||
}
|
||||
pipe_mutex_unlock(fenced_list->mutex);
|
||||
|
||||
pb_fence(fenced_buf->buffer, fence);
|
||||
|
||||
fenced_buf->vl = NULL;
|
||||
fenced_buf->validation_flags = 0;
|
||||
}
|
||||
|
||||
|
||||
static void
|
||||
fenced_buffer_get_base_buffer(struct pb_buffer *buf,
|
||||
struct pb_buffer **base_buf,
|
||||
|
@ -415,13 +325,11 @@ fenced_buffer_get_base_buffer(struct pb_buffer *buf,
|
|||
}
|
||||
|
||||
|
||||
static const struct pb_vtbl
|
||||
const struct pb_vtbl
|
||||
fenced_buffer_vtbl = {
|
||||
fenced_buffer_destroy,
|
||||
fenced_buffer_map,
|
||||
fenced_buffer_unmap,
|
||||
fenced_buffer_validate,
|
||||
fenced_buffer_fence,
|
||||
fenced_buffer_get_base_buffer
|
||||
};
|
||||
|
||||
|
@ -454,6 +362,52 @@ fenced_buffer_create(struct fenced_buffer_list *fenced_list,
|
|||
}
|
||||
|
||||
|
||||
void
|
||||
buffer_fence(struct pb_buffer *buf,
|
||||
struct pipe_fence_handle *fence)
|
||||
{
|
||||
struct fenced_buffer *fenced_buf;
|
||||
struct fenced_buffer_list *fenced_list;
|
||||
struct pipe_winsys *winsys;
|
||||
/* FIXME: receive this as a parameter */
|
||||
unsigned flags = fence ? PIPE_BUFFER_USAGE_GPU_READ_WRITE : 0;
|
||||
|
||||
/* This is a public function, so be extra cautious with the buffer passed,
|
||||
* as happens frequently to receive null buffers, or pointer to buffers
|
||||
* other than fenced buffers. */
|
||||
assert(buf);
|
||||
if(!buf)
|
||||
return;
|
||||
assert(buf->vtbl == &fenced_buffer_vtbl);
|
||||
if(buf->vtbl != &fenced_buffer_vtbl)
|
||||
return;
|
||||
|
||||
fenced_buf = fenced_buffer(buf);
|
||||
fenced_list = fenced_buf->list;
|
||||
winsys = fenced_list->winsys;
|
||||
|
||||
if(!fence || fence == fenced_buf->fence) {
|
||||
/* Handle the same fence case specially, not only because it is a fast
|
||||
* path, but mostly to avoid serializing two writes with the same fence,
|
||||
* as that would bring the hardware down to synchronous operation without
|
||||
* any benefit.
|
||||
*/
|
||||
fenced_buf->flags |= flags & PIPE_BUFFER_USAGE_GPU_READ_WRITE;
|
||||
return;
|
||||
}
|
||||
|
||||
pipe_mutex_lock(fenced_list->mutex);
|
||||
if (fenced_buf->fence)
|
||||
_fenced_buffer_remove(fenced_list, fenced_buf);
|
||||
if (fence) {
|
||||
winsys->fence_reference(winsys, &fenced_buf->fence, fence);
|
||||
fenced_buf->flags |= flags & PIPE_BUFFER_USAGE_GPU_READ_WRITE;
|
||||
_fenced_buffer_add(fenced_buf);
|
||||
}
|
||||
pipe_mutex_unlock(fenced_list->mutex);
|
||||
}
|
||||
|
||||
|
||||
struct fenced_buffer_list *
|
||||
fenced_buffer_list_create(struct pipe_winsys *winsys)
|
||||
{
|
||||
|
|
|
@ -44,7 +44,7 @@
|
|||
* Between the handle's destruction, and the fence signalling, the buffer is
|
||||
* stored in a fenced buffer list.
|
||||
*
|
||||
* \author Jose Fonseca <jrfonseca@tungstengraphics.com>
|
||||
* \author José Fonseca <jrfonseca@tungstengraphics.com>
|
||||
*/
|
||||
|
||||
#ifndef PB_BUFFER_FENCED_H_
|
||||
|
@ -70,6 +70,14 @@ struct pipe_fence_handle;
|
|||
struct fenced_buffer_list;
|
||||
|
||||
|
||||
/**
|
||||
* The fenced buffer's virtual function table.
|
||||
*
|
||||
* NOTE: Made public for debugging purposes.
|
||||
*/
|
||||
extern const struct pb_vtbl fenced_buffer_vtbl;
|
||||
|
||||
|
||||
/**
|
||||
* Create a fenced buffer list.
|
||||
*
|
||||
|
@ -100,6 +108,17 @@ fenced_buffer_create(struct fenced_buffer_list *fenced,
|
|||
struct pb_buffer *buffer);
|
||||
|
||||
|
||||
/**
|
||||
* Set a buffer's fence.
|
||||
*
|
||||
* NOTE: Although it takes a generic pb_buffer argument, it will fail
|
||||
* on everything but buffers returned by fenced_buffer_create.
|
||||
*/
|
||||
void
|
||||
buffer_fence(struct pb_buffer *buf,
|
||||
struct pipe_fence_handle *fence);
|
||||
|
||||
|
||||
#ifdef __cplusplus
|
||||
}
|
||||
#endif
|
||||
|
|
|
@ -81,24 +81,6 @@ malloc_buffer_unmap(struct pb_buffer *buf)
|
|||
}
|
||||
|
||||
|
||||
static enum pipe_error
|
||||
malloc_buffer_validate(struct pb_buffer *buf,
|
||||
struct pb_validate *vl,
|
||||
unsigned flags)
|
||||
{
|
||||
assert(0);
|
||||
return PIPE_ERROR;
|
||||
}
|
||||
|
||||
|
||||
static void
|
||||
malloc_buffer_fence(struct pb_buffer *buf,
|
||||
struct pipe_fence_handle *fence)
|
||||
{
|
||||
assert(0);
|
||||
}
|
||||
|
||||
|
||||
static void
|
||||
malloc_buffer_get_base_buffer(struct pb_buffer *buf,
|
||||
struct pb_buffer **base_buf,
|
||||
|
@ -114,8 +96,6 @@ malloc_buffer_vtbl = {
|
|||
malloc_buffer_destroy,
|
||||
malloc_buffer_map,
|
||||
malloc_buffer_unmap,
|
||||
malloc_buffer_validate,
|
||||
malloc_buffer_fence,
|
||||
malloc_buffer_get_base_buffer
|
||||
};
|
||||
|
||||
|
|
|
@ -43,7 +43,7 @@
|
|||
* - the fenced buffer manager, which will delay buffer destruction until the
|
||||
* the moment the card finishing processing it.
|
||||
*
|
||||
* \author Jose Fonseca <jrfonseca@tungstengraphics.com>
|
||||
* \author José Fonseca <jrfonseca@tungstengraphics.com>
|
||||
*/
|
||||
|
||||
#ifndef PB_BUFMGR_H_
|
||||
|
@ -182,20 +182,6 @@ pb_alt_manager_create(struct pb_manager *provider1,
|
|||
struct pb_manager *provider2);
|
||||
|
||||
|
||||
/**
|
||||
* Ondemand buffer manager.
|
||||
*
|
||||
* Buffers are created in malloc'ed memory (fast and cached), and the constents
|
||||
* is transfered to a buffer from the provider (typically in slow uncached
|
||||
* memory) when there is an attempt to validate the buffer.
|
||||
*
|
||||
* Ideal for situations where one does not know before hand whether a given
|
||||
* buffer will effectively be used by the hardware or not.
|
||||
*/
|
||||
struct pb_manager *
|
||||
pb_ondemand_manager_create(struct pb_manager *provider);
|
||||
|
||||
|
||||
/**
|
||||
* Debug buffer manager to detect buffer under- and overflows.
|
||||
*
|
||||
|
|
|
@ -29,7 +29,7 @@
|
|||
* \file
|
||||
* Buffer cache.
|
||||
*
|
||||
* \author Jose Fonseca <jrfonseca-at-tungstengraphics-dot-com>
|
||||
* \author José Fonseca <jrfonseca-at-tungstengraphics-dot-com>
|
||||
* \author Thomas Hellström <thomas-at-tungstengraphics-dot-com>
|
||||
*/
|
||||
|
||||
|
@ -183,25 +183,6 @@ pb_cache_buffer_unmap(struct pb_buffer *_buf)
|
|||
}
|
||||
|
||||
|
||||
static enum pipe_error
|
||||
pb_cache_buffer_validate(struct pb_buffer *_buf,
|
||||
struct pb_validate *vl,
|
||||
unsigned flags)
|
||||
{
|
||||
struct pb_cache_buffer *buf = pb_cache_buffer(_buf);
|
||||
return pb_validate(buf->buffer, vl, flags);
|
||||
}
|
||||
|
||||
|
||||
static void
|
||||
pb_cache_buffer_fence(struct pb_buffer *_buf,
|
||||
struct pipe_fence_handle *fence)
|
||||
{
|
||||
struct pb_cache_buffer *buf = pb_cache_buffer(_buf);
|
||||
pb_fence(buf->buffer, fence);
|
||||
}
|
||||
|
||||
|
||||
static void
|
||||
pb_cache_buffer_get_base_buffer(struct pb_buffer *_buf,
|
||||
struct pb_buffer **base_buf,
|
||||
|
@ -217,8 +198,6 @@ pb_cache_buffer_vtbl = {
|
|||
pb_cache_buffer_destroy,
|
||||
pb_cache_buffer_map,
|
||||
pb_cache_buffer_unmap,
|
||||
pb_cache_buffer_validate,
|
||||
pb_cache_buffer_fence,
|
||||
pb_cache_buffer_get_base_buffer
|
||||
};
|
||||
|
||||
|
|
|
@ -29,7 +29,7 @@
|
|||
* \file
|
||||
* Debug buffer manager to detect buffer under- and overflows.
|
||||
*
|
||||
* \author Jose Fonseca <jrfonseca@tungstengraphics.com>
|
||||
* \author José Fonseca <jrfonseca@tungstengraphics.com>
|
||||
*/
|
||||
|
||||
|
||||
|
@ -255,35 +255,11 @@ pb_debug_buffer_get_base_buffer(struct pb_buffer *_buf,
|
|||
}
|
||||
|
||||
|
||||
static enum pipe_error
|
||||
pb_debug_buffer_validate(struct pb_buffer *_buf,
|
||||
struct pb_validate *vl,
|
||||
unsigned flags)
|
||||
{
|
||||
struct pb_debug_buffer *buf = pb_debug_buffer(_buf);
|
||||
|
||||
pb_debug_buffer_check(buf);
|
||||
|
||||
return pb_validate(buf->buffer, vl, flags);
|
||||
}
|
||||
|
||||
|
||||
static void
|
||||
pb_debug_buffer_fence(struct pb_buffer *_buf,
|
||||
struct pipe_fence_handle *fence)
|
||||
{
|
||||
struct pb_debug_buffer *buf = pb_debug_buffer(_buf);
|
||||
pb_fence(buf->buffer, fence);
|
||||
}
|
||||
|
||||
|
||||
const struct pb_vtbl
|
||||
pb_debug_buffer_vtbl = {
|
||||
pb_debug_buffer_destroy,
|
||||
pb_debug_buffer_map,
|
||||
pb_debug_buffer_unmap,
|
||||
pb_debug_buffer_validate,
|
||||
pb_debug_buffer_fence,
|
||||
pb_debug_buffer_get_base_buffer
|
||||
};
|
||||
|
||||
|
|
|
@ -30,7 +30,7 @@
|
|||
* \file
|
||||
* A buffer manager that wraps buffers in fenced buffers.
|
||||
*
|
||||
* \author Jose Fonseca <jrfonseca@tungstengraphics.dot.com>
|
||||
* \author José Fonseca <jrfonseca@tungstengraphics.dot.com>
|
||||
*/
|
||||
|
||||
|
||||
|
|
|
@ -29,7 +29,7 @@
|
|||
* \file
|
||||
* Buffer manager using the old texture memory manager.
|
||||
*
|
||||
* \author Jose Fonseca <jrfonseca@tungstengraphics.com>
|
||||
* \author José Fonseca <jrfonseca@tungstengraphics.com>
|
||||
*/
|
||||
|
||||
|
||||
|
@ -124,27 +124,6 @@ mm_buffer_unmap(struct pb_buffer *buf)
|
|||
}
|
||||
|
||||
|
||||
static enum pipe_error
|
||||
mm_buffer_validate(struct pb_buffer *buf,
|
||||
struct pb_validate *vl,
|
||||
unsigned flags)
|
||||
{
|
||||
struct mm_buffer *mm_buf = mm_buffer(buf);
|
||||
struct mm_pb_manager *mm = mm_buf->mgr;
|
||||
return pb_validate(mm->buffer, vl, flags);
|
||||
}
|
||||
|
||||
|
||||
static void
|
||||
mm_buffer_fence(struct pb_buffer *buf,
|
||||
struct pipe_fence_handle *fence)
|
||||
{
|
||||
struct mm_buffer *mm_buf = mm_buffer(buf);
|
||||
struct mm_pb_manager *mm = mm_buf->mgr;
|
||||
pb_fence(mm->buffer, fence);
|
||||
}
|
||||
|
||||
|
||||
static void
|
||||
mm_buffer_get_base_buffer(struct pb_buffer *buf,
|
||||
struct pb_buffer **base_buf,
|
||||
|
@ -162,8 +141,6 @@ mm_buffer_vtbl = {
|
|||
mm_buffer_destroy,
|
||||
mm_buffer_map,
|
||||
mm_buffer_unmap,
|
||||
mm_buffer_validate,
|
||||
mm_buffer_fence,
|
||||
mm_buffer_get_base_buffer
|
||||
};
|
||||
|
||||
|
|
|
@ -1,303 +0,0 @@
|
|||
/**************************************************************************
|
||||
*
|
||||
* Copyright 2008 Tungsten Graphics, Inc., Cedar Park, Texas.
|
||||
* All Rights Reserved.
|
||||
*
|
||||
* Permission is hereby granted, free of charge, to any person obtaining a
|
||||
* copy of this software and associated documentation files (the
|
||||
* "Software"), to deal in the Software without restriction, including
|
||||
* without limitation the rights to use, copy, modify, merge, publish,
|
||||
* distribute, sub license, and/or sell copies of the Software, and to
|
||||
* permit persons to whom the Software is furnished to do so, subject to
|
||||
* the following conditions:
|
||||
*
|
||||
* The above copyright notice and this permission notice (including the
|
||||
* next paragraph) shall be included in all copies or substantial portions
|
||||
* of the Software.
|
||||
*
|
||||
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
|
||||
* OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
|
||||
* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT.
|
||||
* IN NO EVENT SHALL TUNGSTEN GRAPHICS AND/OR ITS SUPPLIERS BE LIABLE FOR
|
||||
* ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT,
|
||||
* TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE
|
||||
* SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
|
||||
*
|
||||
**************************************************************************/
|
||||
|
||||
/**
|
||||
* @file
|
||||
* A variation of malloc buffers which get transferred to real graphics memory
|
||||
* when there is an attempt to validate them.
|
||||
*
|
||||
* @author Jose Fonseca <jrfonseca@tungstengraphics.com>
|
||||
*/
|
||||
|
||||
|
||||
#include "pipe/p_debug.h"
|
||||
#include "util/u_memory.h"
|
||||
#include "pb_buffer.h"
|
||||
#include "pb_bufmgr.h"
|
||||
|
||||
|
||||
struct pb_ondemand_manager;
|
||||
|
||||
|
||||
struct pb_ondemand_buffer
|
||||
{
|
||||
struct pb_buffer base;
|
||||
|
||||
struct pb_ondemand_manager *mgr;
|
||||
|
||||
/** Regular malloc'ed memory */
|
||||
void *data;
|
||||
unsigned mapcount;
|
||||
|
||||
/** Real buffer */
|
||||
struct pb_buffer *buffer;
|
||||
size_t size;
|
||||
struct pb_desc desc;
|
||||
};
|
||||
|
||||
|
||||
struct pb_ondemand_manager
|
||||
{
|
||||
struct pb_manager base;
|
||||
|
||||
struct pb_manager *provider;
|
||||
};
|
||||
|
||||
|
||||
extern const struct pb_vtbl pb_ondemand_buffer_vtbl;
|
||||
|
||||
static INLINE struct pb_ondemand_buffer *
|
||||
pb_ondemand_buffer(struct pb_buffer *buf)
|
||||
{
|
||||
assert(buf);
|
||||
assert(buf->vtbl == &pb_ondemand_buffer_vtbl);
|
||||
return (struct pb_ondemand_buffer *)buf;
|
||||
}
|
||||
|
||||
static INLINE struct pb_ondemand_manager *
|
||||
pb_ondemand_manager(struct pb_manager *mgr)
|
||||
{
|
||||
assert(mgr);
|
||||
return (struct pb_ondemand_manager *)mgr;
|
||||
}
|
||||
|
||||
|
||||
static void
|
||||
pb_ondemand_buffer_destroy(struct pb_buffer *_buf)
|
||||
{
|
||||
struct pb_ondemand_buffer *buf = pb_ondemand_buffer(_buf);
|
||||
|
||||
pb_reference(&buf->buffer, NULL);
|
||||
|
||||
align_free(buf->data);
|
||||
|
||||
FREE(buf);
|
||||
}
|
||||
|
||||
|
||||
static void *
|
||||
pb_ondemand_buffer_map(struct pb_buffer *_buf,
|
||||
unsigned flags)
|
||||
{
|
||||
struct pb_ondemand_buffer *buf = pb_ondemand_buffer(_buf);
|
||||
|
||||
if(buf->buffer) {
|
||||
assert(!buf->data);
|
||||
return pb_map(buf->buffer, flags);
|
||||
}
|
||||
else {
|
||||
assert(buf->data);
|
||||
++buf->mapcount;
|
||||
return buf->data;
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
static void
|
||||
pb_ondemand_buffer_unmap(struct pb_buffer *_buf)
|
||||
{
|
||||
struct pb_ondemand_buffer *buf = pb_ondemand_buffer(_buf);
|
||||
|
||||
if(buf->buffer) {
|
||||
assert(!buf->data);
|
||||
pb_unmap(buf->buffer);
|
||||
}
|
||||
else {
|
||||
assert(buf->data);
|
||||
assert(buf->mapcount);
|
||||
if(buf->mapcount)
|
||||
--buf->mapcount;
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
static enum pipe_error
|
||||
pb_ondemand_buffer_instantiate(struct pb_ondemand_buffer *buf)
|
||||
{
|
||||
if(!buf->buffer) {
|
||||
struct pb_manager *provider = buf->mgr->provider;
|
||||
uint8_t *map;
|
||||
|
||||
assert(!buf->mapcount);
|
||||
|
||||
buf->buffer = provider->create_buffer(provider, buf->size, &buf->desc);
|
||||
if(!buf->buffer)
|
||||
return PIPE_ERROR_OUT_OF_MEMORY;
|
||||
|
||||
map = pb_map(buf->buffer, PIPE_BUFFER_USAGE_CPU_READ);
|
||||
if(!map) {
|
||||
pb_reference(&buf->buffer, NULL);
|
||||
return PIPE_ERROR;
|
||||
}
|
||||
|
||||
memcpy(map, buf->data, buf->size);
|
||||
|
||||
pb_unmap(buf->buffer);
|
||||
|
||||
if(!buf->mapcount) {
|
||||
FREE(buf->data);
|
||||
buf->data = NULL;
|
||||
}
|
||||
}
|
||||
|
||||
return PIPE_OK;
|
||||
}
|
||||
|
||||
static enum pipe_error
|
||||
pb_ondemand_buffer_validate(struct pb_buffer *_buf,
|
||||
struct pb_validate *vl,
|
||||
unsigned flags)
|
||||
{
|
||||
struct pb_ondemand_buffer *buf = pb_ondemand_buffer(_buf);
|
||||
enum pipe_error ret;
|
||||
|
||||
assert(!buf->mapcount);
|
||||
if(buf->mapcount)
|
||||
return PIPE_ERROR;
|
||||
|
||||
ret = pb_ondemand_buffer_instantiate(buf);
|
||||
if(ret != PIPE_OK)
|
||||
return ret;
|
||||
|
||||
return pb_validate(buf->buffer, vl, flags);
|
||||
}
|
||||
|
||||
|
||||
static void
|
||||
pb_ondemand_buffer_fence(struct pb_buffer *_buf,
|
||||
struct pipe_fence_handle *fence)
|
||||
{
|
||||
struct pb_ondemand_buffer *buf = pb_ondemand_buffer(_buf);
|
||||
|
||||
assert(buf->buffer);
|
||||
if(!buf->buffer)
|
||||
return;
|
||||
|
||||
pb_fence(buf->buffer, fence);
|
||||
}
|
||||
|
||||
|
||||
static void
|
||||
pb_ondemand_buffer_get_base_buffer(struct pb_buffer *_buf,
|
||||
struct pb_buffer **base_buf,
|
||||
unsigned *offset)
|
||||
{
|
||||
struct pb_ondemand_buffer *buf = pb_ondemand_buffer(_buf);
|
||||
|
||||
if(pb_ondemand_buffer_instantiate(buf) != PIPE_OK) {
|
||||
assert(0);
|
||||
*base_buf = &buf->base;
|
||||
*offset = 0;
|
||||
return;
|
||||
}
|
||||
|
||||
pb_get_base_buffer(buf->buffer, base_buf, offset);
|
||||
}
|
||||
|
||||
|
||||
const struct pb_vtbl
|
||||
pb_ondemand_buffer_vtbl = {
|
||||
pb_ondemand_buffer_destroy,
|
||||
pb_ondemand_buffer_map,
|
||||
pb_ondemand_buffer_unmap,
|
||||
pb_ondemand_buffer_validate,
|
||||
pb_ondemand_buffer_fence,
|
||||
pb_ondemand_buffer_get_base_buffer
|
||||
};
|
||||
|
||||
|
||||
static struct pb_buffer *
|
||||
pb_ondemand_manager_create_buffer(struct pb_manager *_mgr,
|
||||
size_t size,
|
||||
const struct pb_desc *desc)
|
||||
{
|
||||
struct pb_ondemand_manager *mgr = pb_ondemand_manager(_mgr);
|
||||
struct pb_ondemand_buffer *buf;
|
||||
|
||||
buf = CALLOC_STRUCT(pb_ondemand_buffer);
|
||||
if(!buf)
|
||||
return NULL;
|
||||
|
||||
buf->base.base.refcount = 1;
|
||||
buf->base.base.alignment = desc->alignment;
|
||||
buf->base.base.usage = desc->usage;
|
||||
buf->base.base.size = size;
|
||||
buf->base.vtbl = &pb_ondemand_buffer_vtbl;
|
||||
|
||||
buf->mgr = mgr;
|
||||
|
||||
buf->data = align_malloc(size, desc->alignment < sizeof(void*) ? sizeof(void*) : desc->alignment);
|
||||
if(!buf->data) {
|
||||
FREE(buf);
|
||||
return NULL;
|
||||
}
|
||||
|
||||
buf->size = size;
|
||||
buf->desc = *desc;
|
||||
|
||||
return &buf->base;
|
||||
}
|
||||
|
||||
|
||||
static void
|
||||
pb_ondemand_manager_flush(struct pb_manager *_mgr)
|
||||
{
|
||||
struct pb_ondemand_manager *mgr = pb_ondemand_manager(_mgr);
|
||||
|
||||
mgr->provider->flush(mgr->provider);
|
||||
}
|
||||
|
||||
|
||||
static void
|
||||
pb_ondemand_manager_destroy(struct pb_manager *_mgr)
|
||||
{
|
||||
struct pb_ondemand_manager *mgr = pb_ondemand_manager(_mgr);
|
||||
|
||||
FREE(mgr);
|
||||
}
|
||||
|
||||
|
||||
struct pb_manager *
|
||||
pb_ondemand_manager_create(struct pb_manager *provider)
|
||||
{
|
||||
struct pb_ondemand_manager *mgr;
|
||||
|
||||
if(!provider)
|
||||
return NULL;
|
||||
|
||||
mgr = CALLOC_STRUCT(pb_ondemand_manager);
|
||||
if(!mgr)
|
||||
return NULL;
|
||||
|
||||
mgr->base.destroy = pb_ondemand_manager_destroy;
|
||||
mgr->base.create_buffer = pb_ondemand_manager_create_buffer;
|
||||
mgr->base.flush = pb_ondemand_manager_flush;
|
||||
|
||||
mgr->provider = provider;
|
||||
|
||||
return &mgr->base;
|
||||
}
|
|
@ -30,7 +30,7 @@
|
|||
* \file
|
||||
* Batch buffer pool management.
|
||||
*
|
||||
* \author Jose Fonseca <jrfonseca-at-tungstengraphics-dot-com>
|
||||
* \author José Fonseca <jrfonseca-at-tungstengraphics-dot-com>
|
||||
* \author Thomas Hellström <thomas-at-tungstengraphics-dot-com>
|
||||
*/
|
||||
|
||||
|
@ -138,27 +138,6 @@ pool_buffer_unmap(struct pb_buffer *buf)
|
|||
}
|
||||
|
||||
|
||||
static enum pipe_error
|
||||
pool_buffer_validate(struct pb_buffer *buf,
|
||||
struct pb_validate *vl,
|
||||
unsigned flags)
|
||||
{
|
||||
struct pool_buffer *pool_buf = pool_buffer(buf);
|
||||
struct pool_pb_manager *pool = pool_buf->mgr;
|
||||
return pb_validate(pool->buffer, vl, flags);
|
||||
}
|
||||
|
||||
|
||||
static void
|
||||
pool_buffer_fence(struct pb_buffer *buf,
|
||||
struct pipe_fence_handle *fence)
|
||||
{
|
||||
struct pool_buffer *pool_buf = pool_buffer(buf);
|
||||
struct pool_pb_manager *pool = pool_buf->mgr;
|
||||
pb_fence(pool->buffer, fence);
|
||||
}
|
||||
|
||||
|
||||
static void
|
||||
pool_buffer_get_base_buffer(struct pb_buffer *buf,
|
||||
struct pb_buffer **base_buf,
|
||||
|
@ -176,8 +155,6 @@ pool_buffer_vtbl = {
|
|||
pool_buffer_destroy,
|
||||
pool_buffer_map,
|
||||
pool_buffer_unmap,
|
||||
pool_buffer_validate,
|
||||
pool_buffer_fence,
|
||||
pool_buffer_get_base_buffer
|
||||
};
|
||||
|
||||
|
|
|
@ -248,25 +248,6 @@ pb_slab_buffer_unmap(struct pb_buffer *_buf)
|
|||
}
|
||||
|
||||
|
||||
static enum pipe_error
|
||||
pb_slab_buffer_validate(struct pb_buffer *_buf,
|
||||
struct pb_validate *vl,
|
||||
unsigned flags)
|
||||
{
|
||||
struct pb_slab_buffer *buf = pb_slab_buffer(_buf);
|
||||
return pb_validate(buf->slab->bo, vl, flags);
|
||||
}
|
||||
|
||||
|
||||
static void
|
||||
pb_slab_buffer_fence(struct pb_buffer *_buf,
|
||||
struct pipe_fence_handle *fence)
|
||||
{
|
||||
struct pb_slab_buffer *buf = pb_slab_buffer(_buf);
|
||||
pb_fence(buf->slab->bo, fence);
|
||||
}
|
||||
|
||||
|
||||
static void
|
||||
pb_slab_buffer_get_base_buffer(struct pb_buffer *_buf,
|
||||
struct pb_buffer **base_buf,
|
||||
|
@ -283,8 +264,6 @@ pb_slab_buffer_vtbl = {
|
|||
pb_slab_buffer_destroy,
|
||||
pb_slab_buffer_map,
|
||||
pb_slab_buffer_unmap,
|
||||
pb_slab_buffer_validate,
|
||||
pb_slab_buffer_fence,
|
||||
pb_slab_buffer_get_base_buffer
|
||||
};
|
||||
|
||||
|
|
|
@ -46,16 +46,9 @@
|
|||
#define PB_VALIDATE_INITIAL_SIZE 1 /* 512 */
|
||||
|
||||
|
||||
struct pb_validate_entry
|
||||
{
|
||||
struct pb_buffer *buf;
|
||||
unsigned flags;
|
||||
};
|
||||
|
||||
|
||||
struct pb_validate
|
||||
{
|
||||
struct pb_validate_entry *entries;
|
||||
struct pb_buffer **buffers;
|
||||
unsigned used;
|
||||
unsigned size;
|
||||
};
|
||||
|
@ -63,87 +56,54 @@ struct pb_validate
|
|||
|
||||
enum pipe_error
|
||||
pb_validate_add_buffer(struct pb_validate *vl,
|
||||
struct pb_buffer *buf,
|
||||
unsigned flags)
|
||||
struct pb_buffer *buf)
|
||||
{
|
||||
assert(buf);
|
||||
if(!buf)
|
||||
return PIPE_ERROR;
|
||||
|
||||
assert(flags & PIPE_BUFFER_USAGE_GPU_READ_WRITE);
|
||||
assert(!(flags & ~PIPE_BUFFER_USAGE_GPU_READ_WRITE));
|
||||
flags &= PIPE_BUFFER_USAGE_GPU_READ_WRITE;
|
||||
|
||||
/* We only need to store one reference for each buffer, so avoid storing
|
||||
* consecutive references for the same buffer. It might not be the most
|
||||
* common pattern, but it is easy to implement.
|
||||
* consecutive references for the same buffer. It might not be the more
|
||||
* common pasttern, but it is easy to implement.
|
||||
*/
|
||||
if(vl->used && vl->entries[vl->used - 1].buf == buf) {
|
||||
vl->entries[vl->used - 1].flags |= flags;
|
||||
if(vl->used && vl->buffers[vl->used - 1] == buf) {
|
||||
return PIPE_OK;
|
||||
}
|
||||
|
||||
/* Grow the table */
|
||||
if(vl->used == vl->size) {
|
||||
unsigned new_size;
|
||||
struct pb_validate_entry *new_entries;
|
||||
struct pb_buffer **new_buffers;
|
||||
|
||||
new_size = vl->size * 2;
|
||||
if(!new_size)
|
||||
return PIPE_ERROR_OUT_OF_MEMORY;
|
||||
|
||||
new_entries = (struct pb_validate_entry *)REALLOC(vl->entries,
|
||||
vl->size*sizeof(struct pb_validate_entry),
|
||||
new_size*sizeof(struct pb_validate_entry));
|
||||
if(!new_entries)
|
||||
new_buffers = (struct pb_buffer **)REALLOC(vl->buffers,
|
||||
vl->size*sizeof(struct pb_buffer *),
|
||||
new_size*sizeof(struct pb_buffer *));
|
||||
if(!new_buffers)
|
||||
return PIPE_ERROR_OUT_OF_MEMORY;
|
||||
|
||||
memset(new_entries + vl->size, 0, (new_size - vl->size)*sizeof(struct pb_validate_entry));
|
||||
memset(new_buffers + vl->size, 0, (new_size - vl->size)*sizeof(struct pb_buffer *));
|
||||
|
||||
vl->size = new_size;
|
||||
vl->entries = new_entries;
|
||||
vl->buffers = new_buffers;
|
||||
}
|
||||
|
||||
assert(!vl->entries[vl->used].buf);
|
||||
pb_reference(&vl->entries[vl->used].buf, buf);
|
||||
vl->entries[vl->used].flags = flags;
|
||||
assert(!vl->buffers[vl->used]);
|
||||
pb_reference(&vl->buffers[vl->used], buf);
|
||||
++vl->used;
|
||||
|
||||
return PIPE_OK;
|
||||
}
|
||||
|
||||
|
||||
enum pipe_error
|
||||
pb_validate_foreach(struct pb_validate *vl,
|
||||
enum pipe_error (*callback)(struct pb_buffer *buf, void *data),
|
||||
void *data)
|
||||
{
|
||||
unsigned i;
|
||||
for(i = 0; i < vl->used; ++i) {
|
||||
enum pipe_error ret;
|
||||
ret = callback(vl->entries[i].buf, data);
|
||||
if(ret != PIPE_OK)
|
||||
return ret;
|
||||
}
|
||||
return PIPE_OK;
|
||||
}
|
||||
|
||||
|
||||
enum pipe_error
|
||||
pb_validate_validate(struct pb_validate *vl)
|
||||
{
|
||||
unsigned i;
|
||||
|
||||
for(i = 0; i < vl->used; ++i) {
|
||||
enum pipe_error ret;
|
||||
ret = pb_validate(vl->entries[i].buf, vl, vl->entries[i].flags);
|
||||
if(ret != PIPE_OK) {
|
||||
while(i--)
|
||||
pb_validate(vl->entries[i].buf, NULL, 0);
|
||||
return ret;
|
||||
}
|
||||
}
|
||||
|
||||
/* FIXME: go through each buffer, ensure its not mapped, its address is
|
||||
* available -- requires a new pb_buffer interface */
|
||||
return PIPE_OK;
|
||||
}
|
||||
|
||||
|
@ -154,8 +114,8 @@ pb_validate_fence(struct pb_validate *vl,
|
|||
{
|
||||
unsigned i;
|
||||
for(i = 0; i < vl->used; ++i) {
|
||||
pb_fence(vl->entries[i].buf, fence);
|
||||
pb_reference(&vl->entries[i].buf, NULL);
|
||||
buffer_fence(vl->buffers[i], fence);
|
||||
pb_reference(&vl->buffers[i], NULL);
|
||||
}
|
||||
vl->used = 0;
|
||||
}
|
||||
|
@ -166,8 +126,8 @@ pb_validate_destroy(struct pb_validate *vl)
|
|||
{
|
||||
unsigned i;
|
||||
for(i = 0; i < vl->used; ++i)
|
||||
pb_reference(&vl->entries[i].buf, NULL);
|
||||
FREE(vl->entries);
|
||||
pb_reference(&vl->buffers[i], NULL);
|
||||
FREE(vl->buffers);
|
||||
FREE(vl);
|
||||
}
|
||||
|
||||
|
@ -182,8 +142,8 @@ pb_validate_create()
|
|||
return NULL;
|
||||
|
||||
vl->size = PB_VALIDATE_INITIAL_SIZE;
|
||||
vl->entries = (struct pb_validate_entry *)CALLOC(vl->size, sizeof(struct pb_validate_entry));
|
||||
if(!vl->entries) {
|
||||
vl->buffers = (struct pb_buffer **)CALLOC(vl->size, sizeof(struct pb_buffer *));
|
||||
if(!vl->buffers) {
|
||||
FREE(vl);
|
||||
return NULL;
|
||||
}
|
||||
|
|
|
@ -58,13 +58,7 @@ struct pb_validate;
|
|||
|
||||
enum pipe_error
|
||||
pb_validate_add_buffer(struct pb_validate *vl,
|
||||
struct pb_buffer *buf,
|
||||
unsigned flags);
|
||||
|
||||
enum pipe_error
|
||||
pb_validate_foreach(struct pb_validate *vl,
|
||||
enum pipe_error (*callback)(struct pb_buffer *buf, void *data),
|
||||
void *data);
|
||||
struct pb_buffer *buf);
|
||||
|
||||
/**
|
||||
* Validate all buffers for hardware access.
|
||||
|
@ -77,7 +71,7 @@ pb_validate_validate(struct pb_validate *vl);
|
|||
/**
|
||||
* Fence all buffers and clear the list.
|
||||
*
|
||||
* Should be called right after issuing commands to the hardware.
|
||||
* Should be called right before issuing commands to the hardware.
|
||||
*/
|
||||
void
|
||||
pb_validate_fence(struct pb_validate *vl,
|
||||
|
|
|
@ -30,7 +30,7 @@
|
|||
* Implementation of client buffer (also designated as "user buffers"), which
|
||||
* are just state-tracker owned data masqueraded as buffers.
|
||||
*
|
||||
* \author Jose Fonseca <jrfonseca@tungstengraphics.com>
|
||||
* \author José Fonseca <jrfonseca@tungstengraphics.com>
|
||||
*/
|
||||
|
||||
|
||||
|
@ -91,24 +91,6 @@ pb_user_buffer_unmap(struct pb_buffer *buf)
|
|||
}
|
||||
|
||||
|
||||
static enum pipe_error
|
||||
pb_user_buffer_validate(struct pb_buffer *buf,
|
||||
struct pb_validate *vl,
|
||||
unsigned flags)
|
||||
{
|
||||
assert(0);
|
||||
return PIPE_ERROR;
|
||||
}
|
||||
|
||||
|
||||
static void
|
||||
pb_user_buffer_fence(struct pb_buffer *buf,
|
||||
struct pipe_fence_handle *fence)
|
||||
{
|
||||
assert(0);
|
||||
}
|
||||
|
||||
|
||||
static void
|
||||
pb_user_buffer_get_base_buffer(struct pb_buffer *buf,
|
||||
struct pb_buffer **base_buf,
|
||||
|
@ -124,8 +106,6 @@ pb_user_buffer_vtbl = {
|
|||
pb_user_buffer_destroy,
|
||||
pb_user_buffer_map,
|
||||
pb_user_buffer_unmap,
|
||||
pb_user_buffer_validate,
|
||||
pb_user_buffer_fence,
|
||||
pb_user_buffer_get_base_buffer
|
||||
};
|
||||
|
||||
|
|
|
@ -66,6 +66,8 @@ struct blit_state
|
|||
void *fs;
|
||||
|
||||
struct pipe_buffer *vbuf; /**< quad vertices */
|
||||
unsigned vbuf_slot;
|
||||
|
||||
float vertices[4][2][4]; /**< vertex/texcoords for quad */
|
||||
};
|
||||
|
||||
|
@ -138,17 +140,7 @@ util_create_blit(struct pipe_context *pipe, struct cso_context *cso)
|
|||
|
||||
/* fragment shader */
|
||||
ctx->fs = util_make_fragment_tex_shader(pipe, &ctx->frag_shader);
|
||||
|
||||
ctx->vbuf = pipe_buffer_create(pipe->screen,
|
||||
32,
|
||||
PIPE_BUFFER_USAGE_VERTEX,
|
||||
sizeof(ctx->vertices));
|
||||
if (!ctx->vbuf) {
|
||||
FREE(ctx);
|
||||
ctx->pipe->delete_fs_state(ctx->pipe, ctx->fs);
|
||||
ctx->pipe->delete_vs_state(ctx->pipe, ctx->vs);
|
||||
return NULL;
|
||||
}
|
||||
ctx->vbuf = NULL;
|
||||
|
||||
/* init vertex data that doesn't change */
|
||||
for (i = 0; i < 4; i++) {
|
||||
|
@ -181,15 +173,35 @@ util_destroy_blit(struct blit_state *ctx)
|
|||
}
|
||||
|
||||
|
||||
static unsigned get_next_slot( struct blit_state *ctx )
|
||||
{
|
||||
const unsigned max_slots = 4096 / sizeof ctx->vertices;
|
||||
|
||||
if (ctx->vbuf_slot >= max_slots)
|
||||
util_blit_flush( ctx );
|
||||
|
||||
if (!ctx->vbuf) {
|
||||
ctx->vbuf = pipe_buffer_create(ctx->pipe->screen,
|
||||
32,
|
||||
PIPE_BUFFER_USAGE_VERTEX,
|
||||
max_slots * sizeof ctx->vertices);
|
||||
}
|
||||
|
||||
return ctx->vbuf_slot++ * sizeof ctx->vertices;
|
||||
}
|
||||
|
||||
|
||||
|
||||
/**
|
||||
* Setup vertex data for the textured quad we'll draw.
|
||||
* Note: y=0=top
|
||||
*/
|
||||
static void
|
||||
static unsigned
|
||||
setup_vertex_data(struct blit_state *ctx,
|
||||
float x0, float y0, float x1, float y1, float z)
|
||||
{
|
||||
void *buf;
|
||||
unsigned offset;
|
||||
|
||||
ctx->vertices[0][0][0] = x0;
|
||||
ctx->vertices[0][0][1] = y0;
|
||||
|
@ -215,12 +227,16 @@ setup_vertex_data(struct blit_state *ctx,
|
|||
ctx->vertices[3][1][0] = 0.0f;
|
||||
ctx->vertices[3][1][1] = 1.0f;
|
||||
|
||||
offset = get_next_slot( ctx );
|
||||
|
||||
buf = pipe_buffer_map(ctx->pipe->screen, ctx->vbuf,
|
||||
PIPE_BUFFER_USAGE_CPU_WRITE);
|
||||
|
||||
memcpy(buf, ctx->vertices, sizeof(ctx->vertices));
|
||||
memcpy((char *)buf + offset, ctx->vertices, sizeof(ctx->vertices));
|
||||
|
||||
pipe_buffer_unmap(ctx->pipe->screen, ctx->vbuf);
|
||||
|
||||
return offset;
|
||||
}
|
||||
|
||||
|
||||
|
@ -228,13 +244,14 @@ setup_vertex_data(struct blit_state *ctx,
|
|||
* Setup vertex data for the textured quad we'll draw.
|
||||
* Note: y=0=top
|
||||
*/
|
||||
static void
|
||||
static unsigned
|
||||
setup_vertex_data_tex(struct blit_state *ctx,
|
||||
float x0, float y0, float x1, float y1,
|
||||
float s0, float t0, float s1, float t1,
|
||||
float z)
|
||||
{
|
||||
void *buf;
|
||||
unsigned offset;
|
||||
|
||||
ctx->vertices[0][0][0] = x0;
|
||||
ctx->vertices[0][0][1] = y0;
|
||||
|
@ -260,12 +277,16 @@ setup_vertex_data_tex(struct blit_state *ctx,
|
|||
ctx->vertices[3][1][0] = s0;
|
||||
ctx->vertices[3][1][1] = t1;
|
||||
|
||||
offset = get_next_slot( ctx );
|
||||
|
||||
buf = pipe_buffer_map(ctx->pipe->screen, ctx->vbuf,
|
||||
PIPE_BUFFER_USAGE_CPU_WRITE);
|
||||
|
||||
memcpy(buf, ctx->vertices, sizeof(ctx->vertices));
|
||||
memcpy((char *)buf + offset, ctx->vertices, sizeof(ctx->vertices));
|
||||
|
||||
pipe_buffer_unmap(ctx->pipe->screen, ctx->vbuf);
|
||||
|
||||
return offset;
|
||||
}
|
||||
/**
|
||||
* Copy pixel block from src surface to dst surface.
|
||||
|
@ -291,6 +312,7 @@ util_blit_pixels(struct blit_state *ctx,
|
|||
const int srcH = abs(srcY1 - srcY0);
|
||||
const int srcLeft = MIN2(srcX0, srcX1);
|
||||
const int srcTop = MIN2(srcY0, srcY1);
|
||||
unsigned offset;
|
||||
|
||||
assert(filter == PIPE_TEX_MIPFILTER_NEAREST ||
|
||||
filter == PIPE_TEX_MIPFILTER_LINEAR);
|
||||
|
@ -398,11 +420,11 @@ util_blit_pixels(struct blit_state *ctx,
|
|||
cso_set_framebuffer(ctx->cso, &fb);
|
||||
|
||||
/* draw quad */
|
||||
setup_vertex_data(ctx,
|
||||
(float) dstX0, (float) dstY0,
|
||||
(float) dstX1, (float) dstY1, z);
|
||||
offset = setup_vertex_data(ctx,
|
||||
(float) dstX0, (float) dstY0,
|
||||
(float) dstX1, (float) dstY1, z);
|
||||
|
||||
util_draw_vertex_buffer(ctx->pipe, ctx->vbuf,
|
||||
util_draw_vertex_buffer(ctx->pipe, ctx->vbuf, offset,
|
||||
PIPE_PRIM_TRIANGLE_FAN,
|
||||
4, /* verts */
|
||||
2); /* attribs/vert */
|
||||
|
@ -421,6 +443,18 @@ util_blit_pixels(struct blit_state *ctx,
|
|||
screen->texture_release(screen, &tex);
|
||||
}
|
||||
|
||||
|
||||
/* Release vertex buffer at end of frame to avoid synchronous
|
||||
* rendering.
|
||||
*/
|
||||
void util_blit_flush( struct blit_state *ctx )
|
||||
{
|
||||
pipe_buffer_reference(ctx->pipe->screen, &ctx->vbuf, NULL);
|
||||
ctx->vbuf_slot = 0;
|
||||
}
|
||||
|
||||
|
||||
|
||||
/**
|
||||
* Copy pixel block from src texture to dst surface.
|
||||
* Overlapping regions are acceptable.
|
||||
|
@ -442,6 +476,7 @@ util_blit_pixels_tex(struct blit_state *ctx,
|
|||
struct pipe_screen *screen = pipe->screen;
|
||||
struct pipe_framebuffer_state fb;
|
||||
float s0, t0, s1, t1;
|
||||
unsigned offset;
|
||||
|
||||
assert(filter == PIPE_TEX_MIPFILTER_NEAREST ||
|
||||
filter == PIPE_TEX_MIPFILTER_LINEAR);
|
||||
|
@ -496,13 +531,14 @@ util_blit_pixels_tex(struct blit_state *ctx,
|
|||
cso_set_framebuffer(ctx->cso, &fb);
|
||||
|
||||
/* draw quad */
|
||||
setup_vertex_data_tex(ctx,
|
||||
(float) dstX0, (float) dstY0,
|
||||
(float) dstX1, (float) dstY1,
|
||||
s0, t0, s1, t1,
|
||||
z);
|
||||
offset = setup_vertex_data_tex(ctx,
|
||||
(float) dstX0, (float) dstY0,
|
||||
(float) dstX1, (float) dstY1,
|
||||
s0, t0, s1, t1,
|
||||
z);
|
||||
|
||||
util_draw_vertex_buffer(ctx->pipe, ctx->vbuf,
|
||||
util_draw_vertex_buffer(ctx->pipe,
|
||||
ctx->vbuf, offset,
|
||||
PIPE_PRIM_TRIANGLE_FAN,
|
||||
4, /* verts */
|
||||
2); /* attribs/vert */
|
||||
|
|
|
@ -70,6 +70,10 @@ util_blit_pixels_tex(struct blit_state *ctx,
|
|||
int dstX1, int dstY1,
|
||||
float z, uint filter);
|
||||
|
||||
/* Call at end of frame to avoid synchronous rendering.
|
||||
*/
|
||||
extern void
|
||||
util_blit_flush( struct blit_state *ctx );
|
||||
|
||||
#ifdef __cplusplus
|
||||
}
|
||||
|
|
|
@ -40,6 +40,7 @@
|
|||
void
|
||||
util_draw_vertex_buffer(struct pipe_context *pipe,
|
||||
struct pipe_buffer *vbuf,
|
||||
uint offset,
|
||||
uint prim_type,
|
||||
uint num_verts,
|
||||
uint num_attribs)
|
||||
|
@ -53,7 +54,7 @@ util_draw_vertex_buffer(struct pipe_context *pipe,
|
|||
/* tell pipe about the vertex buffer */
|
||||
vbuffer.buffer = vbuf;
|
||||
vbuffer.pitch = num_attribs * 4 * sizeof(float); /* vertex size */
|
||||
vbuffer.buffer_offset = 0;
|
||||
vbuffer.buffer_offset = offset;
|
||||
pipe->set_vertex_buffers(pipe, 1, &vbuffer);
|
||||
|
||||
/* tell pipe about the vertex attributes */
|
||||
|
@ -124,7 +125,7 @@ util_draw_texquad(struct pipe_context *pipe,
|
|||
v[29] = 1.0;
|
||||
|
||||
pipe_buffer_unmap(pipe->screen, vbuf);
|
||||
util_draw_vertex_buffer(pipe, vbuf, PIPE_PRIM_TRIANGLE_FAN, 4, 2);
|
||||
util_draw_vertex_buffer(pipe, vbuf, 0, PIPE_PRIM_TRIANGLE_FAN, 4, 2);
|
||||
}
|
||||
|
||||
pipe_buffer_reference(pipe->screen, &vbuf, NULL);
|
||||
|
|
|
@ -37,7 +37,7 @@ struct pipe_buffer;
|
|||
|
||||
extern void
|
||||
util_draw_vertex_buffer(struct pipe_context *pipe,
|
||||
struct pipe_buffer *vbuf,
|
||||
struct pipe_buffer *vbuf, uint offset,
|
||||
uint num_attribs, uint num_verts, uint prim_type);
|
||||
|
||||
|
||||
|
|
|
@ -69,6 +69,8 @@ struct gen_mipmap_state
|
|||
void *fs;
|
||||
|
||||
struct pipe_buffer *vbuf; /**< quad vertices */
|
||||
unsigned vbuf_slot;
|
||||
|
||||
float vertices[4][2][4]; /**< vertex/texcoords for quad */
|
||||
};
|
||||
|
||||
|
@ -779,10 +781,28 @@ util_create_gen_mipmap(struct pipe_context *pipe,
|
|||
}
|
||||
|
||||
|
||||
static void
|
||||
static unsigned get_next_slot( struct gen_mipmap_state *ctx )
|
||||
{
|
||||
const unsigned max_slots = 4096 / sizeof ctx->vertices;
|
||||
|
||||
if (ctx->vbuf_slot >= max_slots)
|
||||
util_gen_mipmap_flush( ctx );
|
||||
|
||||
if (!ctx->vbuf) {
|
||||
ctx->vbuf = pipe_buffer_create(ctx->pipe->screen,
|
||||
32,
|
||||
PIPE_BUFFER_USAGE_VERTEX,
|
||||
max_slots * sizeof ctx->vertices);
|
||||
}
|
||||
|
||||
return ctx->vbuf_slot++ * sizeof ctx->vertices;
|
||||
}
|
||||
|
||||
static unsigned
|
||||
set_vertex_data(struct gen_mipmap_state *ctx, float width, float height)
|
||||
{
|
||||
void *buf;
|
||||
unsigned offset;
|
||||
|
||||
ctx->vertices[0][0][0] = 0.0f; /*x*/
|
||||
ctx->vertices[0][0][1] = 0.0f; /*y*/
|
||||
|
@ -804,12 +824,16 @@ set_vertex_data(struct gen_mipmap_state *ctx, float width, float height)
|
|||
ctx->vertices[3][1][0] = 0.0f;
|
||||
ctx->vertices[3][1][1] = 1.0f;
|
||||
|
||||
offset = get_next_slot( ctx );
|
||||
|
||||
buf = pipe_buffer_map(ctx->pipe->screen, ctx->vbuf,
|
||||
PIPE_BUFFER_USAGE_CPU_WRITE);
|
||||
|
||||
memcpy(buf, ctx->vertices, sizeof(ctx->vertices));
|
||||
memcpy((char *)buf + offset, ctx->vertices, sizeof(ctx->vertices));
|
||||
|
||||
pipe_buffer_unmap(ctx->pipe->screen, ctx->vbuf);
|
||||
|
||||
return offset;
|
||||
}
|
||||
|
||||
|
||||
|
@ -834,6 +858,17 @@ util_destroy_gen_mipmap(struct gen_mipmap_state *ctx)
|
|||
}
|
||||
|
||||
|
||||
|
||||
/* Release vertex buffer at end of frame to avoid synchronous
|
||||
* rendering.
|
||||
*/
|
||||
void util_gen_mipmap_flush( struct gen_mipmap_state *ctx )
|
||||
{
|
||||
pipe_buffer_reference(ctx->pipe->screen, &ctx->vbuf, NULL);
|
||||
ctx->vbuf_slot = 0;
|
||||
}
|
||||
|
||||
|
||||
/**
|
||||
* Generate mipmap images. It's assumed all needed texture memory is
|
||||
* already allocated.
|
||||
|
@ -855,6 +890,7 @@ util_gen_mipmap(struct gen_mipmap_state *ctx,
|
|||
struct pipe_framebuffer_state fb;
|
||||
uint dstLevel;
|
||||
uint zslice = 0;
|
||||
uint offset;
|
||||
|
||||
/* check if we can render in the texture's format */
|
||||
if (!screen->is_format_supported(screen, pt->format, PIPE_TEXTURE_2D,
|
||||
|
@ -925,10 +961,13 @@ util_gen_mipmap(struct gen_mipmap_state *ctx,
|
|||
cso_set_sampler_textures(ctx->cso, 1, &pt);
|
||||
|
||||
/* quad coords in window coords (bypassing clipping, viewport mapping) */
|
||||
set_vertex_data(ctx,
|
||||
(float) pt->width[dstLevel],
|
||||
(float) pt->height[dstLevel]);
|
||||
util_draw_vertex_buffer(ctx->pipe, ctx->vbuf,
|
||||
offset = set_vertex_data(ctx,
|
||||
(float) pt->width[dstLevel],
|
||||
(float) pt->height[dstLevel]);
|
||||
|
||||
util_draw_vertex_buffer(ctx->pipe,
|
||||
ctx->vbuf,
|
||||
offset,
|
||||
PIPE_PRIM_TRIANGLE_FAN,
|
||||
4, /* verts */
|
||||
2); /* attribs/vert */
|
||||
|
|
|
@ -50,6 +50,11 @@ util_create_gen_mipmap(struct pipe_context *pipe, struct cso_context *cso);
|
|||
extern void
|
||||
util_destroy_gen_mipmap(struct gen_mipmap_state *ctx);
|
||||
|
||||
/* Release vertex buffer at end of frame to avoid synchronous
|
||||
* rendering.
|
||||
*/
|
||||
extern void
|
||||
util_gen_mipmap_flush( struct gen_mipmap_state *ctx );
|
||||
|
||||
|
||||
extern void
|
||||
|
|
|
@ -34,6 +34,8 @@
|
|||
#include "sp_surface.h"
|
||||
#include "sp_tile_cache.h"
|
||||
|
||||
#include "draw/draw_context.h"
|
||||
|
||||
|
||||
/**
|
||||
* XXX this might get moved someday
|
||||
|
|
|
@ -14,12 +14,6 @@ if env['platform'] == 'windows':
|
|||
env.Append(CPPDEFINES = [
|
||||
])
|
||||
|
||||
env.Append(CPPDEFINES = [
|
||||
'__GL_EXPORTS',
|
||||
'BUILD_GL32',
|
||||
'_GNU_H_WINDOWS32_DEFINES',
|
||||
])
|
||||
|
||||
sources = [
|
||||
'#src/mesa/state_tracker/wgl/opengl32.def',
|
||||
'gdi_softpipe_winsys.c',
|
||||
|
|
|
@ -201,6 +201,7 @@ STATETRACKER_SOURCES = \
|
|||
state_tracker/st_context.c \
|
||||
state_tracker/st_debug.c \
|
||||
state_tracker/st_draw.c \
|
||||
state_tracker/st_draw_feedback.c \
|
||||
state_tracker/st_extensions.c \
|
||||
state_tracker/st_format.c \
|
||||
state_tracker/st_framebuffer.c \
|
||||
|
|
|
@ -38,6 +38,7 @@
|
|||
#include "st_cb_accum.h"
|
||||
#include "st_cb_fbo.h"
|
||||
#include "st_draw.h"
|
||||
#include "st_public.h"
|
||||
#include "st_format.h"
|
||||
#include "pipe/p_context.h"
|
||||
#include "pipe/p_defines.h"
|
||||
|
@ -324,7 +325,7 @@ st_Accum(GLcontext *ctx, GLenum op, GLfloat value)
|
|||
const GLint height = ctx->DrawBuffer->_Ymax - ypos;
|
||||
|
||||
/* make sure color bufs aren't cached */
|
||||
pipe->flush(pipe, PIPE_FLUSH_RENDER_CACHE, NULL);
|
||||
st_flush( st, PIPE_FLUSH_RENDER_CACHE, NULL );
|
||||
|
||||
switch (op) {
|
||||
case GL_ADD:
|
||||
|
|
|
@ -349,8 +349,7 @@ make_bitmap_texture(GLcontext *ctx, GLsizei width, GLsizei height,
|
|||
return pt;
|
||||
}
|
||||
|
||||
|
||||
static void
|
||||
static GLuint
|
||||
setup_bitmap_vertex_data(struct st_context *st,
|
||||
int x, int y, int width, int height,
|
||||
float z, const float color[4])
|
||||
|
@ -369,12 +368,18 @@ setup_bitmap_vertex_data(struct st_context *st,
|
|||
const GLfloat clip_y0 = (GLfloat)(y0 / fb_height * 2.0 - 1.0);
|
||||
const GLfloat clip_x1 = (GLfloat)(x1 / fb_width * 2.0 - 1.0);
|
||||
const GLfloat clip_y1 = (GLfloat)(y1 / fb_height * 2.0 - 1.0);
|
||||
const GLuint max_slots = 4096 / sizeof(st->bitmap.vertices);
|
||||
GLuint i;
|
||||
void *buf;
|
||||
|
||||
if (st->bitmap.vbuf_slot >= max_slots) {
|
||||
pipe_buffer_reference(pipe->screen, &st->bitmap.vbuf, NULL);
|
||||
st->bitmap.vbuf_slot = 0;
|
||||
}
|
||||
|
||||
if (!st->bitmap.vbuf) {
|
||||
st->bitmap.vbuf = pipe_buffer_create(pipe->screen, 32, PIPE_BUFFER_USAGE_VERTEX,
|
||||
sizeof(st->bitmap.vertices));
|
||||
st->bitmap.vbuf = pipe_buffer_create(pipe->screen, 32,
|
||||
PIPE_BUFFER_USAGE_VERTEX,
|
||||
max_slots * sizeof(st->bitmap.vertices));
|
||||
}
|
||||
|
||||
/* Positions are in clip coords since we need to do clipping in case
|
||||
|
@ -413,9 +418,19 @@ setup_bitmap_vertex_data(struct st_context *st,
|
|||
}
|
||||
|
||||
/* put vertex data into vbuf */
|
||||
buf = pipe_buffer_map(pipe->screen, st->bitmap.vbuf, PIPE_BUFFER_USAGE_CPU_WRITE);
|
||||
memcpy(buf, st->bitmap.vertices, sizeof(st->bitmap.vertices));
|
||||
pipe_buffer_unmap(pipe->screen, st->bitmap.vbuf);
|
||||
{
|
||||
char *buf = pipe_buffer_map(pipe->screen,
|
||||
st->bitmap.vbuf,
|
||||
PIPE_BUFFER_USAGE_CPU_WRITE);
|
||||
|
||||
memcpy(buf + st->bitmap.vbuf_slot * sizeof st->bitmap.vertices,
|
||||
st->bitmap.vertices,
|
||||
sizeof st->bitmap.vertices);
|
||||
|
||||
pipe_buffer_unmap(pipe->screen, st->bitmap.vbuf);
|
||||
}
|
||||
|
||||
return st->bitmap.vbuf_slot++ * sizeof st->bitmap.vertices;
|
||||
}
|
||||
|
||||
|
||||
|
@ -434,6 +449,7 @@ draw_bitmap_quad(GLcontext *ctx, GLint x, GLint y, GLfloat z,
|
|||
struct cso_context *cso = ctx->st->cso_context;
|
||||
struct st_fragment_program *stfp;
|
||||
GLuint maxSize;
|
||||
GLuint offset;
|
||||
|
||||
stfp = combined_bitmap_fragment_program(ctx);
|
||||
|
||||
|
@ -518,11 +534,11 @@ draw_bitmap_quad(GLcontext *ctx, GLint x, GLint y, GLfloat z,
|
|||
}
|
||||
|
||||
/* draw textured quad */
|
||||
setup_bitmap_vertex_data(st, x, y, width, height,
|
||||
ctx->Current.RasterPos[2],
|
||||
color);
|
||||
offset = setup_bitmap_vertex_data(st, x, y, width, height,
|
||||
ctx->Current.RasterPos[2],
|
||||
color);
|
||||
|
||||
util_draw_vertex_buffer(pipe, st->bitmap.vbuf,
|
||||
util_draw_vertex_buffer(pipe, st->bitmap.vbuf, offset,
|
||||
PIPE_PRIM_TRIANGLE_FAN,
|
||||
4, /* verts */
|
||||
3); /* attribs/vert */
|
||||
|
@ -592,12 +608,12 @@ st_flush_bitmap_cache(struct st_context *st)
|
|||
struct pipe_screen *screen = pipe->screen;
|
||||
|
||||
assert(cache->xmin <= cache->xmax);
|
||||
/*
|
||||
printf("flush size %d x %d at %d, %d\n",
|
||||
|
||||
/* printf("flush size %d x %d at %d, %d\n",
|
||||
cache->xmax - cache->xmin,
|
||||
cache->ymax - cache->ymin,
|
||||
cache->xpos, cache->ypos);
|
||||
*/
|
||||
*/
|
||||
|
||||
/* The texture surface has been mapped until now.
|
||||
* So unmap and release the texture surface before drawing.
|
||||
|
@ -623,6 +639,20 @@ st_flush_bitmap_cache(struct st_context *st)
|
|||
}
|
||||
}
|
||||
|
||||
/* Flush bitmap cache and release vertex buffer.
|
||||
*/
|
||||
void
|
||||
st_flush_bitmap( struct st_context *st )
|
||||
{
|
||||
st_flush_bitmap_cache(st);
|
||||
|
||||
/* Release vertex buffer to avoid synchronous rendering if we were
|
||||
* to map it in the next frame.
|
||||
*/
|
||||
pipe_buffer_reference(st->pipe->screen, &st->bitmap.vbuf, NULL);
|
||||
st->bitmap.vbuf_slot = 0;
|
||||
}
|
||||
|
||||
|
||||
/**
|
||||
* Try to accumulate this glBitmap call in the bitmap cache.
|
||||
|
|
|
@ -42,5 +42,11 @@ st_destroy_bitmap(struct st_context *st);
|
|||
extern void
|
||||
st_flush_bitmap_cache(struct st_context *st);
|
||||
|
||||
/* Flush bitmap cache and release vertex buffer. Needed at end of
|
||||
* frame to avoid synchronous rendering.
|
||||
*/
|
||||
extern void
|
||||
st_flush_bitmap(struct st_context *st);
|
||||
|
||||
|
||||
#endif /* ST_CB_BITMAP_H */
|
||||
|
|
|
@ -148,12 +148,18 @@ draw_quad(GLcontext *ctx,
|
|||
{
|
||||
struct st_context *st = ctx->st;
|
||||
struct pipe_context *pipe = st->pipe;
|
||||
const GLuint max_slots = 1024 / sizeof(st->clear.vertices);
|
||||
GLuint i;
|
||||
void *buf;
|
||||
|
||||
if (st->clear.vbuf_slot >= max_slots) {
|
||||
pipe_buffer_reference(pipe->screen, &st->clear.vbuf, NULL);
|
||||
st->clear.vbuf_slot = 0;
|
||||
}
|
||||
|
||||
if (!st->clear.vbuf) {
|
||||
st->clear.vbuf = pipe_buffer_create(pipe->screen, 32, PIPE_BUFFER_USAGE_VERTEX,
|
||||
sizeof(st->clear.vertices));
|
||||
max_slots * sizeof(st->clear.vertices));
|
||||
}
|
||||
|
||||
/* positions */
|
||||
|
@ -181,14 +187,23 @@ draw_quad(GLcontext *ctx,
|
|||
|
||||
/* put vertex data into vbuf */
|
||||
buf = pipe_buffer_map(pipe->screen, st->clear.vbuf, PIPE_BUFFER_USAGE_CPU_WRITE);
|
||||
memcpy(buf, st->clear.vertices, sizeof(st->clear.vertices));
|
||||
|
||||
memcpy((char *)buf + st->clear.vbuf_slot * sizeof(st->clear.vertices),
|
||||
st->clear.vertices,
|
||||
sizeof(st->clear.vertices));
|
||||
|
||||
pipe_buffer_unmap(pipe->screen, st->clear.vbuf);
|
||||
|
||||
/* draw */
|
||||
util_draw_vertex_buffer(pipe, st->clear.vbuf,
|
||||
util_draw_vertex_buffer(pipe,
|
||||
st->clear.vbuf,
|
||||
st->clear.vbuf_slot * sizeof(st->clear.vertices),
|
||||
PIPE_PRIM_TRIANGLE_FAN,
|
||||
4, /* verts */
|
||||
2); /* attribs/vert */
|
||||
|
||||
/* Increment slot */
|
||||
st->clear.vbuf_slot++;
|
||||
}
|
||||
|
||||
|
||||
|
@ -508,6 +523,16 @@ clear_depth_stencil_buffer(GLcontext *ctx, struct gl_renderbuffer *rb)
|
|||
}
|
||||
|
||||
|
||||
void st_flush_clear( struct st_context *st )
|
||||
{
|
||||
/* Release vertex buffer to avoid synchronous rendering if we were
|
||||
* to map it in the next frame.
|
||||
*/
|
||||
pipe_buffer_reference(st->pipe->screen, &st->clear.vbuf, NULL);
|
||||
st->clear.vbuf_slot = 0;
|
||||
}
|
||||
|
||||
|
||||
|
||||
/**
|
||||
* Called via ctx->Driver.Clear()
|
||||
|
|
|
@ -37,6 +37,9 @@ st_init_clear(struct st_context *st);
|
|||
extern void
|
||||
st_destroy_clear(struct st_context *st);
|
||||
|
||||
extern void
|
||||
st_flush_clear(struct st_context *st);
|
||||
|
||||
|
||||
extern void
|
||||
st_init_clear_functions(struct dd_function_table *functions);
|
||||
|
|
|
@ -494,7 +494,7 @@ draw_quad(GLcontext *ctx, GLfloat x0, GLfloat y0, GLfloat z,
|
|||
memcpy(map, verts, sizeof(verts));
|
||||
pipe_buffer_unmap(pipe->screen, buf);
|
||||
|
||||
util_draw_vertex_buffer(pipe, buf,
|
||||
util_draw_vertex_buffer(pipe, buf, 0,
|
||||
PIPE_PRIM_QUADS,
|
||||
4, /* verts */
|
||||
3); /* attribs/vert */
|
||||
|
|
|
@ -426,7 +426,7 @@ st_finish_render_texture(GLcontext *ctx,
|
|||
if (!strb)
|
||||
return;
|
||||
|
||||
ctx->st->pipe->flush(ctx->st->pipe, PIPE_FLUSH_RENDER_CACHE, NULL);
|
||||
st_flush( ctx->st, PIPE_FLUSH_RENDER_CACHE, NULL );
|
||||
|
||||
if (strb->surface)
|
||||
screen->tex_surface_release( screen, &strb->surface );
|
||||
|
|
|
@ -37,11 +37,14 @@
|
|||
#include "st_context.h"
|
||||
#include "st_cb_bitmap.h"
|
||||
#include "st_cb_flush.h"
|
||||
#include "st_cb_clear.h"
|
||||
#include "st_cb_fbo.h"
|
||||
#include "st_public.h"
|
||||
#include "pipe/p_context.h"
|
||||
#include "pipe/p_defines.h"
|
||||
#include "pipe/p_winsys.h"
|
||||
#include "util/u_gen_mipmap.h"
|
||||
#include "util/u_blit.h"
|
||||
|
||||
|
||||
static INLINE GLboolean
|
||||
|
@ -78,7 +81,13 @@ void st_flush( struct st_context *st, uint pipeFlushFlags,
|
|||
{
|
||||
FLUSH_VERTICES(st->ctx, 0);
|
||||
|
||||
st_flush_bitmap_cache(st);
|
||||
/* Release any vertex buffers that might potentially be accessed in
|
||||
* successive frames:
|
||||
*/
|
||||
st_flush_bitmap(st);
|
||||
st_flush_clear(st);
|
||||
util_blit_flush(st->blit);
|
||||
util_gen_mipmap_flush(st->gen_mipmap);
|
||||
|
||||
st->pipe->flush( st->pipe, pipeFlushFlags, fence );
|
||||
}
|
||||
|
|
|
@ -317,10 +317,8 @@ st_readpixels(GLcontext *ctx, GLint x, GLint y, GLsizei width, GLsizei height,
|
|||
if (!dest)
|
||||
return;
|
||||
|
||||
st_flush_bitmap_cache(ctx->st);
|
||||
|
||||
/* make sure rendering has completed */
|
||||
pipe->flush(pipe, PIPE_FLUSH_RENDER_CACHE, NULL);
|
||||
st_flush(ctx->st, PIPE_FLUSH_RENDER_CACHE, NULL);
|
||||
|
||||
if (format == GL_STENCIL_INDEX) {
|
||||
st_read_stencil_pixels(ctx, x, y, width, height, type, pack, dest);
|
||||
|
|
|
@ -154,6 +154,7 @@ struct st_context
|
|||
void *vs;
|
||||
float vertices[4][3][4]; /**< vertex pos + color + texcoord */
|
||||
struct pipe_buffer *vbuf;
|
||||
unsigned vbuf_slot; /* next free slot in vbuf */
|
||||
struct bitmap_cache *cache;
|
||||
} bitmap;
|
||||
|
||||
|
@ -173,6 +174,7 @@ struct st_context
|
|||
void *fs;
|
||||
float vertices[4][2][4]; /**< vertex pos + color */
|
||||
struct pipe_buffer *vbuf;
|
||||
unsigned vbuf_slot;
|
||||
} clear;
|
||||
|
||||
void *passthrough_fs; /**< simple pass-through frag shader */
|
||||
|
|
|
@ -47,9 +47,6 @@
|
|||
#include "pipe/p_defines.h"
|
||||
#include "pipe/p_inlines.h"
|
||||
|
||||
#include "draw/draw_private.h"
|
||||
#include "draw/draw_context.h"
|
||||
|
||||
|
||||
static GLuint double_types[4] = {
|
||||
PIPE_FORMAT_R64_FLOAT,
|
||||
|
@ -161,8 +158,8 @@ static GLuint fixed_types[4] = {
|
|||
/**
|
||||
* Return a PIPE_FORMAT_x for the given GL datatype and size.
|
||||
*/
|
||||
static GLuint
|
||||
pipe_vertex_format(GLenum type, GLuint size, GLboolean normalized)
|
||||
GLuint
|
||||
st_pipe_vertex_format(GLenum type, GLuint size, GLboolean normalized)
|
||||
{
|
||||
assert((type >= GL_BYTE && type <= GL_DOUBLE) ||
|
||||
type == GL_FIXED);
|
||||
|
@ -393,9 +390,9 @@ setup_interleaved_attribs(GLcontext *ctx,
|
|||
velements[attr].vertex_buffer_index = 0;
|
||||
velements[attr].nr_components = arrays[mesaAttr]->Size;
|
||||
velements[attr].src_format =
|
||||
pipe_vertex_format(arrays[mesaAttr]->Type,
|
||||
arrays[mesaAttr]->Size,
|
||||
arrays[mesaAttr]->Normalized);
|
||||
st_pipe_vertex_format(arrays[mesaAttr]->Type,
|
||||
arrays[mesaAttr]->Size,
|
||||
arrays[mesaAttr]->Normalized);
|
||||
assert(velements[attr].src_format);
|
||||
}
|
||||
}
|
||||
|
@ -412,6 +409,7 @@ setup_non_interleaved_attribs(GLcontext *ctx,
|
|||
const struct st_vertex_program *vp,
|
||||
const struct gl_client_array **arrays,
|
||||
GLuint max_index,
|
||||
GLboolean *userSpace,
|
||||
struct pipe_vertex_buffer vbuffer[],
|
||||
struct pipe_vertex_element velements[])
|
||||
{
|
||||
|
@ -423,6 +421,8 @@ setup_non_interleaved_attribs(GLcontext *ctx,
|
|||
struct gl_buffer_object *bufobj = arrays[mesaAttr]->BufferObj;
|
||||
GLsizei stride = arrays[mesaAttr]->StrideB;
|
||||
|
||||
*userSpace = GL_FALSE;
|
||||
|
||||
if (bufobj && bufobj->Name) {
|
||||
/* Attribute data is in a VBO.
|
||||
* Recall that for VBOs, the gl_client_array->Ptr field is
|
||||
|
@ -442,6 +442,8 @@ setup_non_interleaved_attribs(GLcontext *ctx,
|
|||
uint bytes;
|
||||
/*printf("user-space array %d stride %d\n", attr, stride);*/
|
||||
|
||||
*userSpace = GL_TRUE;
|
||||
|
||||
/* wrap user data */
|
||||
if (arrays[mesaAttr]->Ptr) {
|
||||
/* user's vertex array */
|
||||
|
@ -475,9 +477,9 @@ setup_non_interleaved_attribs(GLcontext *ctx,
|
|||
velements[attr].vertex_buffer_index = attr;
|
||||
velements[attr].nr_components = arrays[mesaAttr]->Size;
|
||||
velements[attr].src_format
|
||||
= pipe_vertex_format(arrays[mesaAttr]->Type,
|
||||
arrays[mesaAttr]->Size,
|
||||
arrays[mesaAttr]->Normalized);
|
||||
= st_pipe_vertex_format(arrays[mesaAttr]->Type,
|
||||
arrays[mesaAttr]->Size,
|
||||
arrays[mesaAttr]->Normalized);
|
||||
assert(velements[attr].src_format);
|
||||
}
|
||||
}
|
||||
|
@ -558,7 +560,7 @@ st_draw_vbo(GLcontext *ctx,
|
|||
else {
|
||||
/*printf("Draw non-interleaved\n");*/
|
||||
setup_non_interleaved_attribs(ctx, vp, arrays, max_index,
|
||||
vbuffer, velements);
|
||||
&userSpace, vbuffer, velements);
|
||||
num_vbuffers = vp->num_inputs;
|
||||
num_velements = vp->num_inputs;
|
||||
}
|
||||
|
@ -672,223 +674,13 @@ st_draw_vbo(GLcontext *ctx,
|
|||
pipe_buffer_reference(pipe->screen, &vbuffer[attr].buffer, NULL);
|
||||
assert(!vbuffer[attr].buffer);
|
||||
}
|
||||
pipe->set_vertex_buffers(pipe, num_vbuffers, vbuffer);
|
||||
}
|
||||
|
||||
|
||||
#if FEATURE_feedback || FEATURE_drawpix
|
||||
|
||||
/**
|
||||
* Set the (private) draw module's post-transformed vertex format when in
|
||||
* GL_SELECT or GL_FEEDBACK mode or for glRasterPos.
|
||||
*/
|
||||
static void
|
||||
set_feedback_vertex_format(GLcontext *ctx)
|
||||
{
|
||||
#if 0
|
||||
struct st_context *st = ctx->st;
|
||||
struct vertex_info vinfo;
|
||||
GLuint i;
|
||||
|
||||
memset(&vinfo, 0, sizeof(vinfo));
|
||||
|
||||
if (ctx->RenderMode == GL_SELECT) {
|
||||
assert(ctx->RenderMode == GL_SELECT);
|
||||
vinfo.num_attribs = 1;
|
||||
vinfo.format[0] = FORMAT_4F;
|
||||
vinfo.interp_mode[0] = INTERP_LINEAR;
|
||||
}
|
||||
else {
|
||||
/* GL_FEEDBACK, or glRasterPos */
|
||||
/* emit all attribs (pos, color, texcoord) as GLfloat[4] */
|
||||
vinfo.num_attribs = st->state.vs->cso->state.num_outputs;
|
||||
for (i = 0; i < vinfo.num_attribs; i++) {
|
||||
vinfo.format[i] = FORMAT_4F;
|
||||
vinfo.interp_mode[i] = INTERP_LINEAR;
|
||||
}
|
||||
}
|
||||
|
||||
draw_set_vertex_info(st->draw, &vinfo);
|
||||
#endif
|
||||
}
|
||||
|
||||
|
||||
/**
|
||||
* Called by VBO to draw arrays when in selection or feedback mode and
|
||||
* to implement glRasterPos.
|
||||
* This is very much like the normal draw_vbo() function above.
|
||||
* Look at code refactoring some day.
|
||||
* Might move this into the failover module some day.
|
||||
*/
|
||||
void
|
||||
st_feedback_draw_vbo(GLcontext *ctx,
|
||||
const struct gl_client_array **arrays,
|
||||
const struct _mesa_prim *prims,
|
||||
GLuint nr_prims,
|
||||
const struct _mesa_index_buffer *ib,
|
||||
GLuint min_index,
|
||||
GLuint max_index)
|
||||
{
|
||||
struct st_context *st = ctx->st;
|
||||
struct pipe_context *pipe = st->pipe;
|
||||
struct draw_context *draw = st->draw;
|
||||
const struct st_vertex_program *vp;
|
||||
const struct pipe_shader_state *vs;
|
||||
struct pipe_buffer *index_buffer_handle = 0;
|
||||
struct pipe_vertex_buffer vbuffers[PIPE_MAX_SHADER_INPUTS];
|
||||
struct pipe_vertex_element velements[PIPE_MAX_ATTRIBS];
|
||||
GLuint attr, i;
|
||||
ubyte *mapped_constants;
|
||||
|
||||
assert(draw);
|
||||
|
||||
st_validate_state(ctx->st);
|
||||
|
||||
/* must get these after state validation! */
|
||||
vp = ctx->st->vp;
|
||||
vs = &st->vp->state;
|
||||
|
||||
if (!st->vp->draw_shader) {
|
||||
st->vp->draw_shader = draw_create_vertex_shader(draw, vs);
|
||||
}
|
||||
|
||||
/*
|
||||
* Set up the draw module's state.
|
||||
*
|
||||
* We'd like to do this less frequently, but the normal state-update
|
||||
* code sends state updates to the pipe, not to our private draw module.
|
||||
*/
|
||||
assert(draw);
|
||||
draw_set_viewport_state(draw, &st->state.viewport);
|
||||
draw_set_clip_state(draw, &st->state.clip);
|
||||
draw_set_rasterizer_state(draw, &st->state.rasterizer);
|
||||
draw_bind_vertex_shader(draw, st->vp->draw_shader);
|
||||
set_feedback_vertex_format(ctx);
|
||||
|
||||
/* loop over TGSI shader inputs to determine vertex buffer
|
||||
* and attribute info
|
||||
*/
|
||||
for (attr = 0; attr < vp->num_inputs; attr++) {
|
||||
const GLuint mesaAttr = vp->index_to_input[attr];
|
||||
struct gl_buffer_object *bufobj = arrays[mesaAttr]->BufferObj;
|
||||
void *map;
|
||||
|
||||
if (bufobj && bufobj->Name) {
|
||||
/* Attribute data is in a VBO.
|
||||
* Recall that for VBOs, the gl_client_array->Ptr field is
|
||||
* really an offset from the start of the VBO, not a pointer.
|
||||
*/
|
||||
struct st_buffer_object *stobj = st_buffer_object(bufobj);
|
||||
assert(stobj->buffer);
|
||||
|
||||
vbuffers[attr].buffer = NULL;
|
||||
pipe_buffer_reference(pipe->screen, &vbuffers[attr].buffer, stobj->buffer);
|
||||
vbuffers[attr].buffer_offset = (unsigned) arrays[0]->Ptr;/* in bytes */
|
||||
velements[attr].src_offset = arrays[mesaAttr]->Ptr - arrays[0]->Ptr;
|
||||
}
|
||||
else {
|
||||
/* attribute data is in user-space memory, not a VBO */
|
||||
uint bytes = (arrays[mesaAttr]->Size
|
||||
* _mesa_sizeof_type(arrays[mesaAttr]->Type)
|
||||
* (max_index + 1));
|
||||
|
||||
/* wrap user data */
|
||||
vbuffers[attr].buffer
|
||||
= pipe_user_buffer_create(pipe->screen, (void *) arrays[mesaAttr]->Ptr,
|
||||
bytes);
|
||||
vbuffers[attr].buffer_offset = 0;
|
||||
velements[attr].src_offset = 0;
|
||||
}
|
||||
|
||||
/* common-case setup */
|
||||
vbuffers[attr].pitch = arrays[mesaAttr]->StrideB; /* in bytes */
|
||||
vbuffers[attr].max_index = max_index;
|
||||
velements[attr].vertex_buffer_index = attr;
|
||||
velements[attr].nr_components = arrays[mesaAttr]->Size;
|
||||
velements[attr].src_format = pipe_vertex_format(arrays[mesaAttr]->Type,
|
||||
arrays[mesaAttr]->Size,
|
||||
arrays[mesaAttr]->Normalized);
|
||||
assert(velements[attr].src_format);
|
||||
|
||||
/* tell draw about this attribute */
|
||||
#if 0
|
||||
draw_set_vertex_buffer(draw, attr, &vbuffer[attr]);
|
||||
#endif
|
||||
|
||||
/* map the attrib buffer */
|
||||
map = pipe_buffer_map(pipe->screen, vbuffers[attr].buffer,
|
||||
PIPE_BUFFER_USAGE_CPU_READ);
|
||||
draw_set_mapped_vertex_buffer(draw, attr, map);
|
||||
}
|
||||
|
||||
draw_set_vertex_buffers(draw, vp->num_inputs, vbuffers);
|
||||
draw_set_vertex_elements(draw, vp->num_inputs, velements);
|
||||
|
||||
if (ib) {
|
||||
unsigned indexSize;
|
||||
struct gl_buffer_object *bufobj = ib->obj;
|
||||
struct st_buffer_object *stobj = st_buffer_object(bufobj);
|
||||
void *map;
|
||||
|
||||
index_buffer_handle = stobj->buffer;
|
||||
|
||||
switch (ib->type) {
|
||||
case GL_UNSIGNED_INT:
|
||||
indexSize = 4;
|
||||
break;
|
||||
case GL_UNSIGNED_SHORT:
|
||||
indexSize = 2;
|
||||
break;
|
||||
default:
|
||||
assert(0);
|
||||
return;
|
||||
}
|
||||
|
||||
map = pipe_buffer_map(pipe->screen, index_buffer_handle,
|
||||
PIPE_BUFFER_USAGE_CPU_READ);
|
||||
draw_set_mapped_element_buffer(draw, indexSize, map);
|
||||
}
|
||||
else {
|
||||
/* no index/element buffer */
|
||||
draw_set_mapped_element_buffer(draw, 0, NULL);
|
||||
}
|
||||
|
||||
|
||||
/* map constant buffers */
|
||||
mapped_constants = pipe_buffer_map(pipe->screen,
|
||||
st->state.constants[PIPE_SHADER_VERTEX].buffer,
|
||||
PIPE_BUFFER_USAGE_CPU_READ);
|
||||
draw_set_mapped_constant_buffer(st->draw, mapped_constants,
|
||||
st->state.constants[PIPE_SHADER_VERTEX].buffer->size);
|
||||
|
||||
|
||||
/* draw here */
|
||||
for (i = 0; i < nr_prims; i++) {
|
||||
draw_arrays(draw, prims[i].mode, prims[i].start, prims[i].count);
|
||||
}
|
||||
|
||||
|
||||
/* unmap constant buffers */
|
||||
pipe_buffer_unmap(pipe->screen, st->state.constants[PIPE_SHADER_VERTEX].buffer);
|
||||
|
||||
/*
|
||||
* unmap vertex/index buffers
|
||||
*/
|
||||
for (i = 0; i < PIPE_MAX_ATTRIBS; i++) {
|
||||
if (draw->pt.vertex_buffer[i].buffer) {
|
||||
pipe_buffer_unmap(pipe->screen, draw->pt.vertex_buffer[i].buffer);
|
||||
pipe_buffer_reference(pipe->screen, &draw->pt.vertex_buffer[i].buffer, NULL);
|
||||
draw_set_mapped_vertex_buffer(draw, i, NULL);
|
||||
}
|
||||
}
|
||||
if (ib) {
|
||||
pipe_buffer_unmap(pipe->screen, index_buffer_handle);
|
||||
draw_set_mapped_element_buffer(draw, 0, NULL);
|
||||
if (userSpace)
|
||||
{
|
||||
pipe->set_vertex_buffers(pipe, 0, NULL);
|
||||
}
|
||||
}
|
||||
|
||||
#endif /* FEATURE_feedback || FEATURE_drawpix */
|
||||
|
||||
|
||||
void st_init_draw( struct st_context *st )
|
||||
{
|
||||
|
|
|
@ -59,4 +59,9 @@ st_feedback_draw_vbo(GLcontext *ctx,
|
|||
GLuint min_index,
|
||||
GLuint max_index);
|
||||
|
||||
/* Internal function:
|
||||
*/
|
||||
extern GLuint
|
||||
st_pipe_vertex_format(GLenum type, GLuint size, GLboolean normalized);
|
||||
|
||||
#endif
|
||||
|
|
|
@ -0,0 +1,262 @@
|
|||
/**************************************************************************
|
||||
*
|
||||
* Copyright 2007 Tungsten Graphics, Inc., Cedar Park, Texas.
|
||||
* All Rights Reserved.
|
||||
*
|
||||
* Permission is hereby granted, free of charge, to any person obtaining a
|
||||
* copy of this software and associated documentation files (the
|
||||
* "Software"), to deal in the Software without restriction, including
|
||||
* without limitation the rights to use, copy, modify, merge, publish,
|
||||
* distribute, sub license, and/or sell copies of the Software, and to
|
||||
* permit persons to whom the Software is furnished to do so, subject to
|
||||
* the following conditions:
|
||||
*
|
||||
* The above copyright notice and this permission notice (including the
|
||||
* next paragraph) shall be included in all copies or substantial portions
|
||||
* of the Software.
|
||||
*
|
||||
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
|
||||
* OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
|
||||
* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT.
|
||||
* IN NO EVENT SHALL TUNGSTEN GRAPHICS AND/OR ITS SUPPLIERS BE LIABLE FOR
|
||||
* ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT,
|
||||
* TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE
|
||||
* SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
|
||||
*
|
||||
**************************************************************************/
|
||||
|
||||
#include "main/imports.h"
|
||||
#include "main/image.h"
|
||||
#include "main/macros.h"
|
||||
#include "shader/prog_uniform.h"
|
||||
|
||||
#include "vbo/vbo.h"
|
||||
|
||||
#include "st_context.h"
|
||||
#include "st_atom.h"
|
||||
#include "st_cb_bufferobjects.h"
|
||||
#include "st_draw.h"
|
||||
#include "st_program.h"
|
||||
|
||||
#include "pipe/p_context.h"
|
||||
#include "pipe/p_defines.h"
|
||||
#include "pipe/p_inlines.h"
|
||||
|
||||
#include "draw/draw_private.h"
|
||||
#include "draw/draw_context.h"
|
||||
|
||||
|
||||
#if FEATURE_feedback || FEATURE_drawpix
|
||||
|
||||
/**
|
||||
* Set the (private) draw module's post-transformed vertex format when in
|
||||
* GL_SELECT or GL_FEEDBACK mode or for glRasterPos.
|
||||
*/
|
||||
static void
|
||||
set_feedback_vertex_format(GLcontext *ctx)
|
||||
{
|
||||
#if 0
|
||||
struct st_context *st = ctx->st;
|
||||
struct vertex_info vinfo;
|
||||
GLuint i;
|
||||
|
||||
memset(&vinfo, 0, sizeof(vinfo));
|
||||
|
||||
if (ctx->RenderMode == GL_SELECT) {
|
||||
assert(ctx->RenderMode == GL_SELECT);
|
||||
vinfo.num_attribs = 1;
|
||||
vinfo.format[0] = FORMAT_4F;
|
||||
vinfo.interp_mode[0] = INTERP_LINEAR;
|
||||
}
|
||||
else {
|
||||
/* GL_FEEDBACK, or glRasterPos */
|
||||
/* emit all attribs (pos, color, texcoord) as GLfloat[4] */
|
||||
vinfo.num_attribs = st->state.vs->cso->state.num_outputs;
|
||||
for (i = 0; i < vinfo.num_attribs; i++) {
|
||||
vinfo.format[i] = FORMAT_4F;
|
||||
vinfo.interp_mode[i] = INTERP_LINEAR;
|
||||
}
|
||||
}
|
||||
|
||||
draw_set_vertex_info(st->draw, &vinfo);
|
||||
#endif
|
||||
}
|
||||
|
||||
|
||||
/**
|
||||
* Called by VBO to draw arrays when in selection or feedback mode and
|
||||
* to implement glRasterPos.
|
||||
* This is very much like the normal draw_vbo() function above.
|
||||
* Look at code refactoring some day.
|
||||
* Might move this into the failover module some day.
|
||||
*/
|
||||
void
|
||||
st_feedback_draw_vbo(GLcontext *ctx,
|
||||
const struct gl_client_array **arrays,
|
||||
const struct _mesa_prim *prims,
|
||||
GLuint nr_prims,
|
||||
const struct _mesa_index_buffer *ib,
|
||||
GLuint min_index,
|
||||
GLuint max_index)
|
||||
{
|
||||
struct st_context *st = ctx->st;
|
||||
struct pipe_context *pipe = st->pipe;
|
||||
struct draw_context *draw = st->draw;
|
||||
const struct st_vertex_program *vp;
|
||||
const struct pipe_shader_state *vs;
|
||||
struct pipe_buffer *index_buffer_handle = 0;
|
||||
struct pipe_vertex_buffer vbuffers[PIPE_MAX_SHADER_INPUTS];
|
||||
struct pipe_vertex_element velements[PIPE_MAX_ATTRIBS];
|
||||
GLuint attr, i;
|
||||
ubyte *mapped_constants;
|
||||
|
||||
assert(draw);
|
||||
|
||||
st_validate_state(ctx->st);
|
||||
|
||||
/* must get these after state validation! */
|
||||
vp = ctx->st->vp;
|
||||
vs = &st->vp->state;
|
||||
|
||||
if (!st->vp->draw_shader) {
|
||||
st->vp->draw_shader = draw_create_vertex_shader(draw, vs);
|
||||
}
|
||||
|
||||
/*
|
||||
* Set up the draw module's state.
|
||||
*
|
||||
* We'd like to do this less frequently, but the normal state-update
|
||||
* code sends state updates to the pipe, not to our private draw module.
|
||||
*/
|
||||
assert(draw);
|
||||
draw_set_viewport_state(draw, &st->state.viewport);
|
||||
draw_set_clip_state(draw, &st->state.clip);
|
||||
draw_set_rasterizer_state(draw, &st->state.rasterizer);
|
||||
draw_bind_vertex_shader(draw, st->vp->draw_shader);
|
||||
set_feedback_vertex_format(ctx);
|
||||
|
||||
/* loop over TGSI shader inputs to determine vertex buffer
|
||||
* and attribute info
|
||||
*/
|
||||
for (attr = 0; attr < vp->num_inputs; attr++) {
|
||||
const GLuint mesaAttr = vp->index_to_input[attr];
|
||||
struct gl_buffer_object *bufobj = arrays[mesaAttr]->BufferObj;
|
||||
void *map;
|
||||
|
||||
if (bufobj && bufobj->Name) {
|
||||
/* Attribute data is in a VBO.
|
||||
* Recall that for VBOs, the gl_client_array->Ptr field is
|
||||
* really an offset from the start of the VBO, not a pointer.
|
||||
*/
|
||||
struct st_buffer_object *stobj = st_buffer_object(bufobj);
|
||||
assert(stobj->buffer);
|
||||
|
||||
vbuffers[attr].buffer = NULL;
|
||||
pipe_buffer_reference(pipe->screen, &vbuffers[attr].buffer, stobj->buffer);
|
||||
vbuffers[attr].buffer_offset = (unsigned) arrays[0]->Ptr;/* in bytes */
|
||||
velements[attr].src_offset = arrays[mesaAttr]->Ptr - arrays[0]->Ptr;
|
||||
}
|
||||
else {
|
||||
/* attribute data is in user-space memory, not a VBO */
|
||||
uint bytes = (arrays[mesaAttr]->Size
|
||||
* _mesa_sizeof_type(arrays[mesaAttr]->Type)
|
||||
* (max_index + 1));
|
||||
|
||||
/* wrap user data */
|
||||
vbuffers[attr].buffer
|
||||
= pipe_user_buffer_create(pipe->screen, (void *) arrays[mesaAttr]->Ptr,
|
||||
bytes);
|
||||
vbuffers[attr].buffer_offset = 0;
|
||||
velements[attr].src_offset = 0;
|
||||
}
|
||||
|
||||
/* common-case setup */
|
||||
vbuffers[attr].pitch = arrays[mesaAttr]->StrideB; /* in bytes */
|
||||
vbuffers[attr].max_index = max_index;
|
||||
velements[attr].vertex_buffer_index = attr;
|
||||
velements[attr].nr_components = arrays[mesaAttr]->Size;
|
||||
velements[attr].src_format =
|
||||
st_pipe_vertex_format(arrays[mesaAttr]->Type,
|
||||
arrays[mesaAttr]->Size,
|
||||
arrays[mesaAttr]->Normalized);
|
||||
assert(velements[attr].src_format);
|
||||
|
||||
/* tell draw about this attribute */
|
||||
#if 0
|
||||
draw_set_vertex_buffer(draw, attr, &vbuffer[attr]);
|
||||
#endif
|
||||
|
||||
/* map the attrib buffer */
|
||||
map = pipe_buffer_map(pipe->screen, vbuffers[attr].buffer,
|
||||
PIPE_BUFFER_USAGE_CPU_READ);
|
||||
draw_set_mapped_vertex_buffer(draw, attr, map);
|
||||
}
|
||||
|
||||
draw_set_vertex_buffers(draw, vp->num_inputs, vbuffers);
|
||||
draw_set_vertex_elements(draw, vp->num_inputs, velements);
|
||||
|
||||
if (ib) {
|
||||
unsigned indexSize;
|
||||
struct gl_buffer_object *bufobj = ib->obj;
|
||||
struct st_buffer_object *stobj = st_buffer_object(bufobj);
|
||||
void *map;
|
||||
|
||||
index_buffer_handle = stobj->buffer;
|
||||
|
||||
switch (ib->type) {
|
||||
case GL_UNSIGNED_INT:
|
||||
indexSize = 4;
|
||||
break;
|
||||
case GL_UNSIGNED_SHORT:
|
||||
indexSize = 2;
|
||||
break;
|
||||
default:
|
||||
assert(0);
|
||||
return;
|
||||
}
|
||||
|
||||
map = pipe_buffer_map(pipe->screen, index_buffer_handle,
|
||||
PIPE_BUFFER_USAGE_CPU_READ);
|
||||
draw_set_mapped_element_buffer(draw, indexSize, map);
|
||||
}
|
||||
else {
|
||||
/* no index/element buffer */
|
||||
draw_set_mapped_element_buffer(draw, 0, NULL);
|
||||
}
|
||||
|
||||
|
||||
/* map constant buffers */
|
||||
mapped_constants = pipe_buffer_map(pipe->screen,
|
||||
st->state.constants[PIPE_SHADER_VERTEX].buffer,
|
||||
PIPE_BUFFER_USAGE_CPU_READ);
|
||||
draw_set_mapped_constant_buffer(st->draw, mapped_constants,
|
||||
st->state.constants[PIPE_SHADER_VERTEX].buffer->size);
|
||||
|
||||
|
||||
/* draw here */
|
||||
for (i = 0; i < nr_prims; i++) {
|
||||
draw_arrays(draw, prims[i].mode, prims[i].start, prims[i].count);
|
||||
}
|
||||
|
||||
|
||||
/* unmap constant buffers */
|
||||
pipe_buffer_unmap(pipe->screen, st->state.constants[PIPE_SHADER_VERTEX].buffer);
|
||||
|
||||
/*
|
||||
* unmap vertex/index buffers
|
||||
*/
|
||||
for (i = 0; i < PIPE_MAX_ATTRIBS; i++) {
|
||||
if (draw->pt.vertex_buffer[i].buffer) {
|
||||
pipe_buffer_unmap(pipe->screen, draw->pt.vertex_buffer[i].buffer);
|
||||
pipe_buffer_reference(pipe->screen, &draw->pt.vertex_buffer[i].buffer, NULL);
|
||||
draw_set_mapped_vertex_buffer(draw, i, NULL);
|
||||
}
|
||||
}
|
||||
if (ib) {
|
||||
pipe_buffer_unmap(pipe->screen, index_buffer_handle);
|
||||
draw_set_mapped_element_buffer(draw, 0, NULL);
|
||||
}
|
||||
}
|
||||
|
||||
#endif /* FEATURE_feedback || FEATURE_drawpix */
|
||||
|
|
@ -475,9 +475,15 @@ compile_instruction(
|
|||
break;
|
||||
case OPCODE_RSQ:
|
||||
fullinst->Instruction.Opcode = TGSI_OPCODE_RSQ;
|
||||
tgsi_util_set_full_src_register_sign_mode(
|
||||
&fullinst->FullSrcRegisters[0],
|
||||
TGSI_UTIL_SIGN_CLEAR );
|
||||
|
||||
/* KW: Don't do this here. If particular hardware needs to do
|
||||
* this, can do so in the driver..
|
||||
*/
|
||||
#if 0
|
||||
tgsi_util_set_full_src_register_sign_mode(
|
||||
&fullinst->FullSrcRegisters[0],
|
||||
TGSI_UTIL_SIGN_CLEAR );
|
||||
#endif
|
||||
break;
|
||||
case OPCODE_SCS:
|
||||
fullinst->Instruction.Opcode = TGSI_OPCODE_SCS;
|
||||
|
|
Loading…
Reference in New Issue