290 lines
6.6 KiB
C
290 lines
6.6 KiB
C
// SPDX-License-Identifier: GPL-2.0-only OR MIT
|
|
/* Copyright (c) 2023 Imagination Technologies Ltd. */
|
|
|
|
#include <uapi/drm/pvr_drm.h>
|
|
|
|
#include <drm/drm_syncobj.h>
|
|
#include <drm/gpu_scheduler.h>
|
|
#include <linux/xarray.h>
|
|
#include <linux/dma-fence-unwrap.h>
|
|
|
|
#include "pvr_device.h"
|
|
#include "pvr_queue.h"
|
|
#include "pvr_sync.h"
|
|
|
|
static int
|
|
pvr_check_sync_op(const struct drm_pvr_sync_op *sync_op)
|
|
{
|
|
u8 handle_type;
|
|
|
|
if (sync_op->flags & ~DRM_PVR_SYNC_OP_FLAGS_MASK)
|
|
return -EINVAL;
|
|
|
|
handle_type = sync_op->flags & DRM_PVR_SYNC_OP_FLAG_HANDLE_TYPE_MASK;
|
|
if (handle_type != DRM_PVR_SYNC_OP_FLAG_HANDLE_TYPE_SYNCOBJ &&
|
|
handle_type != DRM_PVR_SYNC_OP_FLAG_HANDLE_TYPE_TIMELINE_SYNCOBJ)
|
|
return -EINVAL;
|
|
|
|
if (handle_type == DRM_PVR_SYNC_OP_FLAG_HANDLE_TYPE_SYNCOBJ &&
|
|
sync_op->value != 0)
|
|
return -EINVAL;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void
|
|
pvr_sync_signal_free(struct pvr_sync_signal *sig_sync)
|
|
{
|
|
if (!sig_sync)
|
|
return;
|
|
|
|
drm_syncobj_put(sig_sync->syncobj);
|
|
dma_fence_chain_free(sig_sync->chain);
|
|
dma_fence_put(sig_sync->fence);
|
|
kfree(sig_sync);
|
|
}
|
|
|
|
void
|
|
pvr_sync_signal_array_cleanup(struct xarray *array)
|
|
{
|
|
struct pvr_sync_signal *sig_sync;
|
|
unsigned long i;
|
|
|
|
xa_for_each(array, i, sig_sync)
|
|
pvr_sync_signal_free(sig_sync);
|
|
|
|
xa_destroy(array);
|
|
}
|
|
|
|
static struct pvr_sync_signal *
|
|
pvr_sync_signal_array_add(struct xarray *array, struct drm_file *file, u32 handle, u64 point)
|
|
{
|
|
struct pvr_sync_signal *sig_sync;
|
|
struct dma_fence *cur_fence;
|
|
int err;
|
|
u32 id;
|
|
|
|
sig_sync = kzalloc(sizeof(*sig_sync), GFP_KERNEL);
|
|
if (!sig_sync)
|
|
return ERR_PTR(-ENOMEM);
|
|
|
|
sig_sync->handle = handle;
|
|
sig_sync->point = point;
|
|
|
|
if (point > 0) {
|
|
sig_sync->chain = dma_fence_chain_alloc();
|
|
if (!sig_sync->chain) {
|
|
err = -ENOMEM;
|
|
goto err_free_sig_sync;
|
|
}
|
|
}
|
|
|
|
sig_sync->syncobj = drm_syncobj_find(file, handle);
|
|
if (!sig_sync->syncobj) {
|
|
err = -EINVAL;
|
|
goto err_free_sig_sync;
|
|
}
|
|
|
|
/* Retrieve the current fence attached to that point. It's
|
|
* perfectly fine to get a NULL fence here, it just means there's
|
|
* no fence attached to that point yet.
|
|
*/
|
|
if (!drm_syncobj_find_fence(file, handle, point, 0, &cur_fence))
|
|
sig_sync->fence = cur_fence;
|
|
|
|
err = xa_alloc(array, &id, sig_sync, xa_limit_32b, GFP_KERNEL);
|
|
if (err)
|
|
goto err_free_sig_sync;
|
|
|
|
return sig_sync;
|
|
|
|
err_free_sig_sync:
|
|
pvr_sync_signal_free(sig_sync);
|
|
return ERR_PTR(err);
|
|
}
|
|
|
|
static struct pvr_sync_signal *
|
|
pvr_sync_signal_array_search(struct xarray *array, u32 handle, u64 point)
|
|
{
|
|
struct pvr_sync_signal *sig_sync;
|
|
unsigned long i;
|
|
|
|
xa_for_each(array, i, sig_sync) {
|
|
if (handle == sig_sync->handle && point == sig_sync->point)
|
|
return sig_sync;
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
static struct pvr_sync_signal *
|
|
pvr_sync_signal_array_get(struct xarray *array, struct drm_file *file, u32 handle, u64 point)
|
|
{
|
|
struct pvr_sync_signal *sig_sync;
|
|
|
|
sig_sync = pvr_sync_signal_array_search(array, handle, point);
|
|
if (sig_sync)
|
|
return sig_sync;
|
|
|
|
return pvr_sync_signal_array_add(array, file, handle, point);
|
|
}
|
|
|
|
int
|
|
pvr_sync_signal_array_collect_ops(struct xarray *array,
|
|
struct drm_file *file,
|
|
u32 sync_op_count,
|
|
const struct drm_pvr_sync_op *sync_ops)
|
|
{
|
|
for (u32 i = 0; i < sync_op_count; i++) {
|
|
struct pvr_sync_signal *sig_sync;
|
|
int ret;
|
|
|
|
if (!(sync_ops[i].flags & DRM_PVR_SYNC_OP_FLAG_SIGNAL))
|
|
continue;
|
|
|
|
ret = pvr_check_sync_op(&sync_ops[i]);
|
|
if (ret)
|
|
return ret;
|
|
|
|
sig_sync = pvr_sync_signal_array_get(array, file,
|
|
sync_ops[i].handle,
|
|
sync_ops[i].value);
|
|
if (IS_ERR(sig_sync))
|
|
return PTR_ERR(sig_sync);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
pvr_sync_signal_array_update_fences(struct xarray *array,
|
|
u32 sync_op_count,
|
|
const struct drm_pvr_sync_op *sync_ops,
|
|
struct dma_fence *done_fence)
|
|
{
|
|
for (u32 i = 0; i < sync_op_count; i++) {
|
|
struct dma_fence *old_fence;
|
|
struct pvr_sync_signal *sig_sync;
|
|
|
|
if (!(sync_ops[i].flags & DRM_PVR_SYNC_OP_FLAG_SIGNAL))
|
|
continue;
|
|
|
|
sig_sync = pvr_sync_signal_array_search(array, sync_ops[i].handle,
|
|
sync_ops[i].value);
|
|
if (WARN_ON(!sig_sync))
|
|
return -EINVAL;
|
|
|
|
old_fence = sig_sync->fence;
|
|
sig_sync->fence = dma_fence_get(done_fence);
|
|
dma_fence_put(old_fence);
|
|
|
|
if (WARN_ON(!sig_sync->fence))
|
|
return -EINVAL;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
void
|
|
pvr_sync_signal_array_push_fences(struct xarray *array)
|
|
{
|
|
struct pvr_sync_signal *sig_sync;
|
|
unsigned long i;
|
|
|
|
xa_for_each(array, i, sig_sync) {
|
|
if (sig_sync->chain) {
|
|
drm_syncobj_add_point(sig_sync->syncobj, sig_sync->chain,
|
|
sig_sync->fence, sig_sync->point);
|
|
sig_sync->chain = NULL;
|
|
} else {
|
|
drm_syncobj_replace_fence(sig_sync->syncobj, sig_sync->fence);
|
|
}
|
|
}
|
|
}
|
|
|
|
static int
|
|
pvr_sync_add_dep_to_job(struct drm_sched_job *job, struct dma_fence *f)
|
|
{
|
|
struct dma_fence_unwrap iter;
|
|
u32 native_fence_count = 0;
|
|
struct dma_fence *uf;
|
|
int err = 0;
|
|
|
|
dma_fence_unwrap_for_each(uf, &iter, f) {
|
|
if (pvr_queue_fence_is_ufo_backed(uf))
|
|
native_fence_count++;
|
|
}
|
|
|
|
/* No need to unwrap the fence if it's fully non-native. */
|
|
if (!native_fence_count)
|
|
return drm_sched_job_add_dependency(job, f);
|
|
|
|
dma_fence_unwrap_for_each(uf, &iter, f) {
|
|
/* There's no dma_fence_unwrap_stop() helper cleaning up the refs
|
|
* owned by dma_fence_unwrap(), so let's just iterate over all
|
|
* entries without doing anything when something failed.
|
|
*/
|
|
if (err)
|
|
continue;
|
|
|
|
if (pvr_queue_fence_is_ufo_backed(uf)) {
|
|
struct drm_sched_fence *s_fence = to_drm_sched_fence(uf);
|
|
|
|
/* If this is a native dependency, we wait for the scheduled fence,
|
|
* and we will let pvr_queue_run_job() issue FW waits.
|
|
*/
|
|
err = drm_sched_job_add_dependency(job,
|
|
dma_fence_get(&s_fence->scheduled));
|
|
} else {
|
|
err = drm_sched_job_add_dependency(job, dma_fence_get(uf));
|
|
}
|
|
}
|
|
|
|
dma_fence_put(f);
|
|
return err;
|
|
}
|
|
|
|
int
|
|
pvr_sync_add_deps_to_job(struct pvr_file *pvr_file, struct drm_sched_job *job,
|
|
u32 sync_op_count,
|
|
const struct drm_pvr_sync_op *sync_ops,
|
|
struct xarray *signal_array)
|
|
{
|
|
int err = 0;
|
|
|
|
if (!sync_op_count)
|
|
return 0;
|
|
|
|
for (u32 i = 0; i < sync_op_count; i++) {
|
|
struct pvr_sync_signal *sig_sync;
|
|
struct dma_fence *fence;
|
|
|
|
if (sync_ops[i].flags & DRM_PVR_SYNC_OP_FLAG_SIGNAL)
|
|
continue;
|
|
|
|
err = pvr_check_sync_op(&sync_ops[i]);
|
|
if (err)
|
|
return err;
|
|
|
|
sig_sync = pvr_sync_signal_array_search(signal_array, sync_ops[i].handle,
|
|
sync_ops[i].value);
|
|
if (sig_sync) {
|
|
if (WARN_ON(!sig_sync->fence))
|
|
return -EINVAL;
|
|
|
|
fence = dma_fence_get(sig_sync->fence);
|
|
} else {
|
|
err = drm_syncobj_find_fence(from_pvr_file(pvr_file), sync_ops[i].handle,
|
|
sync_ops[i].value, 0, &fence);
|
|
if (err)
|
|
return err;
|
|
}
|
|
|
|
err = pvr_sync_add_dep_to_job(job, fence);
|
|
if (err)
|
|
return err;
|
|
}
|
|
|
|
return 0;
|
|
}
|