499 lines
12 KiB
C
499 lines
12 KiB
C
/**********************************************************
|
|
* Copyright 2009-2015 VMware, Inc. All rights reserved.
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person
|
|
* obtaining a copy of this software and associated documentation
|
|
* files (the "Software"), to deal in the Software without
|
|
* restriction, including without limitation the rights to use, copy,
|
|
* modify, merge, publish, distribute, sublicense, and/or sell copies
|
|
* of the Software, and to permit persons to whom the Software is
|
|
* furnished to do so, subject to the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice shall be
|
|
* included in all copies or substantial portions of the Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
|
|
* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
|
|
* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
|
|
* NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
|
|
* BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
|
|
* ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
|
|
* CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
|
|
* SOFTWARE.
|
|
*
|
|
**********************************************************/
|
|
#include <libsync.h>
|
|
|
|
#include "util/u_memory.h"
|
|
#include "util/u_atomic.h"
|
|
#include "util/list.h"
|
|
#include "os/os_thread.h"
|
|
|
|
#include "pipebuffer/pb_buffer_fenced.h"
|
|
|
|
#include "vmw_screen.h"
|
|
#include "vmw_fence.h"
|
|
|
|
struct vmw_fence_ops
|
|
{
|
|
/*
|
|
* Immutable members.
|
|
*/
|
|
struct pb_fence_ops base;
|
|
struct vmw_winsys_screen *vws;
|
|
|
|
mtx_t mutex;
|
|
|
|
/*
|
|
* Protected by mutex;
|
|
*/
|
|
struct list_head not_signaled;
|
|
uint32_t last_signaled;
|
|
uint32_t last_emitted;
|
|
};
|
|
|
|
struct vmw_fence
|
|
{
|
|
struct list_head ops_list;
|
|
int32_t refcount;
|
|
uint32_t handle;
|
|
uint32_t mask;
|
|
int32_t signalled;
|
|
uint32_t seqno;
|
|
int32_t fence_fd;
|
|
boolean imported; /* TRUE if imported from another process */
|
|
};
|
|
|
|
/**
|
|
* vmw_fence_seq_is_signaled - Check whether a fence seqno is
|
|
* signaled.
|
|
*
|
|
* @ops: Pointer to a struct pb_fence_ops.
|
|
*
|
|
*/
|
|
static inline boolean
|
|
vmw_fence_seq_is_signaled(uint32_t seq, uint32_t last, uint32_t cur)
|
|
{
|
|
return (cur - last <= cur - seq);
|
|
}
|
|
|
|
|
|
/**
|
|
* vmw_fence_ops - Return the vmw_fence_ops structure backing a
|
|
* struct pb_fence_ops pointer.
|
|
*
|
|
* @ops: Pointer to a struct pb_fence_ops.
|
|
*
|
|
*/
|
|
static inline struct vmw_fence_ops *
|
|
vmw_fence_ops(struct pb_fence_ops *ops)
|
|
{
|
|
assert(ops);
|
|
return (struct vmw_fence_ops *)ops;
|
|
}
|
|
|
|
|
|
/**
|
|
* vmw_fences_release - Release all fences from the not_signaled
|
|
* list.
|
|
*
|
|
* @ops: Pointer to a struct vmw_fence_ops.
|
|
*
|
|
*/
|
|
static void
|
|
vmw_fences_release(struct vmw_fence_ops *ops)
|
|
{
|
|
struct vmw_fence *fence, *n;
|
|
|
|
mtx_lock(&ops->mutex);
|
|
LIST_FOR_EACH_ENTRY_SAFE(fence, n, &ops->not_signaled, ops_list)
|
|
list_delinit(&fence->ops_list);
|
|
mtx_unlock(&ops->mutex);
|
|
}
|
|
|
|
/**
|
|
* vmw_fences_signal - Traverse the not_signaled list and try to
|
|
* signal unsignaled fences.
|
|
*
|
|
* @ops: Pointer to a struct pb_fence_ops.
|
|
* @signaled: Seqno that has signaled.
|
|
* @emitted: Last seqno emitted by the kernel.
|
|
* @has_emitted: Whether we provide the emitted value.
|
|
*
|
|
*/
|
|
void
|
|
vmw_fences_signal(struct pb_fence_ops *fence_ops,
|
|
uint32_t signaled,
|
|
uint32_t emitted,
|
|
boolean has_emitted)
|
|
{
|
|
struct vmw_fence_ops *ops = NULL;
|
|
struct vmw_fence *fence, *n;
|
|
|
|
if (fence_ops == NULL)
|
|
return;
|
|
|
|
ops = vmw_fence_ops(fence_ops);
|
|
mtx_lock(&ops->mutex);
|
|
|
|
if (!has_emitted) {
|
|
emitted = ops->last_emitted;
|
|
if (emitted - signaled > (1 << 30))
|
|
emitted = signaled;
|
|
}
|
|
|
|
if (signaled == ops->last_signaled && emitted == ops->last_emitted)
|
|
goto out_unlock;
|
|
|
|
LIST_FOR_EACH_ENTRY_SAFE(fence, n, &ops->not_signaled, ops_list) {
|
|
if (!vmw_fence_seq_is_signaled(fence->seqno, signaled, emitted))
|
|
break;
|
|
|
|
p_atomic_set(&fence->signalled, 1);
|
|
list_delinit(&fence->ops_list);
|
|
}
|
|
ops->last_signaled = signaled;
|
|
ops->last_emitted = emitted;
|
|
|
|
out_unlock:
|
|
mtx_unlock(&ops->mutex);
|
|
}
|
|
|
|
|
|
/**
|
|
* vmw_fence - return the vmw_fence object identified by a
|
|
* struct pipe_fence_handle *
|
|
*
|
|
* @fence: The opaque pipe fence handle.
|
|
*/
|
|
static inline struct vmw_fence *
|
|
vmw_fence(struct pipe_fence_handle *fence)
|
|
{
|
|
return (struct vmw_fence *) fence;
|
|
}
|
|
|
|
|
|
/**
|
|
* vmw_fence_create - Create a user-space fence object.
|
|
*
|
|
* @fence_ops: The fence_ops manager to register with.
|
|
* @handle: Handle identifying the kernel fence object.
|
|
* @mask: Mask of flags that this fence object may signal.
|
|
* @fd: File descriptor to associate with the fence
|
|
*
|
|
* Returns NULL on failure.
|
|
*/
|
|
struct pipe_fence_handle *
|
|
vmw_fence_create(struct pb_fence_ops *fence_ops, uint32_t handle,
|
|
uint32_t seqno, uint32_t mask, int32_t fd)
|
|
{
|
|
struct vmw_fence *fence = CALLOC_STRUCT(vmw_fence);
|
|
struct vmw_fence_ops *ops = NULL;
|
|
|
|
if (!fence)
|
|
return NULL;
|
|
|
|
p_atomic_set(&fence->refcount, 1);
|
|
fence->handle = handle;
|
|
fence->mask = mask;
|
|
fence->seqno = seqno;
|
|
fence->fence_fd = fd;
|
|
p_atomic_set(&fence->signalled, 0);
|
|
|
|
/*
|
|
* If the fence was not created by our device, then we won't
|
|
* manage it with our ops
|
|
*/
|
|
if (!fence_ops) {
|
|
fence->imported = true;
|
|
return (struct pipe_fence_handle *) fence;
|
|
}
|
|
|
|
ops = vmw_fence_ops(fence_ops);
|
|
|
|
mtx_lock(&ops->mutex);
|
|
|
|
if (vmw_fence_seq_is_signaled(seqno, ops->last_signaled, seqno)) {
|
|
p_atomic_set(&fence->signalled, 1);
|
|
list_inithead(&fence->ops_list);
|
|
} else {
|
|
p_atomic_set(&fence->signalled, 0);
|
|
list_addtail(&fence->ops_list, &ops->not_signaled);
|
|
}
|
|
|
|
mtx_unlock(&ops->mutex);
|
|
|
|
return (struct pipe_fence_handle *) fence;
|
|
}
|
|
|
|
|
|
/**
|
|
* vmw_fence_destroy - Frees a vmw fence object.
|
|
*
|
|
* Also closes the file handle associated with the object, if any
|
|
*/
|
|
static
|
|
void vmw_fence_destroy(struct vmw_fence *vfence)
|
|
{
|
|
if (vfence->fence_fd != -1)
|
|
close(vfence->fence_fd);
|
|
|
|
FREE(vfence);
|
|
}
|
|
|
|
|
|
/**
|
|
* vmw_fence_reference - Reference / unreference a vmw fence object.
|
|
*
|
|
* @vws: Pointer to the winsys screen.
|
|
* @ptr: Pointer to reference transfer destination.
|
|
* @fence: Pointer to object to reference. May be NULL.
|
|
*/
|
|
void
|
|
vmw_fence_reference(struct vmw_winsys_screen *vws,
|
|
struct pipe_fence_handle **ptr,
|
|
struct pipe_fence_handle *fence)
|
|
{
|
|
if (*ptr) {
|
|
struct vmw_fence *vfence = vmw_fence(*ptr);
|
|
|
|
if (p_atomic_dec_zero(&vfence->refcount)) {
|
|
struct vmw_fence_ops *ops = vmw_fence_ops(vws->fence_ops);
|
|
|
|
if (!vfence->imported) {
|
|
vmw_ioctl_fence_unref(vws, vfence->handle);
|
|
|
|
mtx_lock(&ops->mutex);
|
|
list_delinit(&vfence->ops_list);
|
|
mtx_unlock(&ops->mutex);
|
|
}
|
|
|
|
vmw_fence_destroy(vfence);
|
|
}
|
|
}
|
|
|
|
if (fence) {
|
|
struct vmw_fence *vfence = vmw_fence(fence);
|
|
|
|
p_atomic_inc(&vfence->refcount);
|
|
}
|
|
|
|
*ptr = fence;
|
|
}
|
|
|
|
|
|
/**
|
|
* vmw_fence_signalled - Check whether a fence object is signalled.
|
|
*
|
|
* @vws: Pointer to the winsys screen.
|
|
* @fence: Handle to the fence object.
|
|
* @flag: Fence flags to check. If the fence object can't signal
|
|
* a flag, it is assumed to be already signaled.
|
|
*
|
|
* Returns 0 if the fence object was signaled, nonzero otherwise.
|
|
*/
|
|
int
|
|
vmw_fence_signalled(struct vmw_winsys_screen *vws,
|
|
struct pipe_fence_handle *fence,
|
|
unsigned flag)
|
|
{
|
|
struct vmw_fence *vfence;
|
|
int32_t vflags = SVGA_FENCE_FLAG_EXEC;
|
|
int ret;
|
|
uint32_t old;
|
|
|
|
if (!fence)
|
|
return 0;
|
|
|
|
vfence = vmw_fence(fence);
|
|
old = p_atomic_read(&vfence->signalled);
|
|
|
|
vflags &= ~vfence->mask;
|
|
|
|
if ((old & vflags) == vflags)
|
|
return 0;
|
|
|
|
/*
|
|
* Currently we update signaled fences on each execbuf call.
|
|
* That should really be sufficient, and we can avoid
|
|
* a lot of kernel calls this way.
|
|
*/
|
|
#if 1
|
|
ret = vmw_ioctl_fence_signalled(vws, vfence->handle, vflags);
|
|
|
|
if (ret == 0)
|
|
p_atomic_set(&vfence->signalled, 1);
|
|
return ret;
|
|
#else
|
|
(void) ret;
|
|
return -1;
|
|
#endif
|
|
}
|
|
|
|
/**
|
|
* vmw_fence_finish - Wait for a fence object to signal.
|
|
*
|
|
* @vws: Pointer to the winsys screen.
|
|
* @fence: Handle to the fence object.
|
|
* @timeout: How long to wait before timing out.
|
|
* @flag: Fence flags to wait for. If the fence object can't signal
|
|
* a flag, it is assumed to be already signaled.
|
|
*
|
|
* Returns 0 if the wait succeeded. Nonzero otherwise.
|
|
*/
|
|
int
|
|
vmw_fence_finish(struct vmw_winsys_screen *vws,
|
|
struct pipe_fence_handle *fence,
|
|
uint64_t timeout,
|
|
unsigned flag)
|
|
{
|
|
struct vmw_fence *vfence;
|
|
int32_t vflags = SVGA_FENCE_FLAG_EXEC;
|
|
int ret;
|
|
uint32_t old;
|
|
|
|
if (!fence)
|
|
return 0;
|
|
|
|
vfence = vmw_fence(fence);
|
|
|
|
if (vfence->imported) {
|
|
ret = sync_wait(vfence->fence_fd, timeout / 1000000);
|
|
|
|
if (!ret)
|
|
p_atomic_set(&vfence->signalled, 1);
|
|
|
|
return !!ret;
|
|
}
|
|
|
|
old = p_atomic_read(&vfence->signalled);
|
|
vflags &= ~vfence->mask;
|
|
|
|
if ((old & vflags) == vflags)
|
|
return 0;
|
|
|
|
ret = vmw_ioctl_fence_finish(vws, vfence->handle, vflags);
|
|
|
|
if (ret == 0) {
|
|
int32_t prev = old;
|
|
|
|
do {
|
|
old = prev;
|
|
prev = p_atomic_cmpxchg(&vfence->signalled, old, old | vflags);
|
|
} while (prev != old);
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
/**
|
|
* vmw_fence_get_fd
|
|
*
|
|
* Returns the file descriptor associated with the fence
|
|
*/
|
|
int
|
|
vmw_fence_get_fd(struct pipe_fence_handle *fence)
|
|
{
|
|
struct vmw_fence *vfence;
|
|
|
|
if (!fence)
|
|
return -1;
|
|
|
|
vfence = vmw_fence(fence);
|
|
return vfence->fence_fd;
|
|
}
|
|
|
|
|
|
/**
|
|
* vmw_fence_ops_fence_reference - wrapper for the pb_fence_ops api.
|
|
*
|
|
* wrapper around vmw_fence_reference.
|
|
*/
|
|
static void
|
|
vmw_fence_ops_fence_reference(struct pb_fence_ops *ops,
|
|
struct pipe_fence_handle **ptr,
|
|
struct pipe_fence_handle *fence)
|
|
{
|
|
struct vmw_winsys_screen *vws = vmw_fence_ops(ops)->vws;
|
|
|
|
vmw_fence_reference(vws, ptr, fence);
|
|
}
|
|
|
|
/**
|
|
* vmw_fence_ops_fence_signalled - wrapper for the pb_fence_ops api.
|
|
*
|
|
* wrapper around vmw_fence_signalled.
|
|
*/
|
|
static int
|
|
vmw_fence_ops_fence_signalled(struct pb_fence_ops *ops,
|
|
struct pipe_fence_handle *fence,
|
|
unsigned flag)
|
|
{
|
|
struct vmw_winsys_screen *vws = vmw_fence_ops(ops)->vws;
|
|
|
|
return vmw_fence_signalled(vws, fence, flag);
|
|
}
|
|
|
|
|
|
/**
|
|
* vmw_fence_ops_fence_finish - wrapper for the pb_fence_ops api.
|
|
*
|
|
* wrapper around vmw_fence_finish.
|
|
*/
|
|
static int
|
|
vmw_fence_ops_fence_finish(struct pb_fence_ops *ops,
|
|
struct pipe_fence_handle *fence,
|
|
unsigned flag)
|
|
{
|
|
struct vmw_winsys_screen *vws = vmw_fence_ops(ops)->vws;
|
|
|
|
return vmw_fence_finish(vws, fence, PIPE_TIMEOUT_INFINITE, flag);
|
|
}
|
|
|
|
|
|
/**
|
|
* vmw_fence_ops_destroy - Destroy a pb_fence_ops function table.
|
|
*
|
|
* @ops: The function table to destroy.
|
|
*
|
|
* Part of the pb_fence_ops api.
|
|
*/
|
|
static void
|
|
vmw_fence_ops_destroy(struct pb_fence_ops *ops)
|
|
{
|
|
vmw_fences_release(vmw_fence_ops(ops));
|
|
FREE(ops);
|
|
}
|
|
|
|
|
|
/**
|
|
* vmw_fence_ops_create - Create a pb_fence_ops function table.
|
|
*
|
|
* @vws: Pointer to a struct vmw_winsys_screen.
|
|
*
|
|
* Returns a pointer to a pb_fence_ops function table to interface
|
|
* with pipe_buffer. This function is typically called on driver setup.
|
|
*
|
|
* Returns NULL on failure.
|
|
*/
|
|
struct pb_fence_ops *
|
|
vmw_fence_ops_create(struct vmw_winsys_screen *vws)
|
|
{
|
|
struct vmw_fence_ops *ops;
|
|
|
|
ops = CALLOC_STRUCT(vmw_fence_ops);
|
|
if(!ops)
|
|
return NULL;
|
|
|
|
(void) mtx_init(&ops->mutex, mtx_plain);
|
|
list_inithead(&ops->not_signaled);
|
|
ops->base.destroy = &vmw_fence_ops_destroy;
|
|
ops->base.fence_reference = &vmw_fence_ops_fence_reference;
|
|
ops->base.fence_signalled = &vmw_fence_ops_fence_signalled;
|
|
ops->base.fence_finish = &vmw_fence_ops_fence_finish;
|
|
|
|
ops->vws = vws;
|
|
|
|
return &ops->base;
|
|
}
|