On 31/10/2025 13:16, Christian König wrote:
Just as proof of concept and minor cleanup.

Signed-off-by: Christian König <[email protected]>
---
  drivers/gpu/drm/scheduler/sched_fence.c | 11 +++++------
  include/drm/gpu_scheduler.h             |  4 ----
  2 files changed, 5 insertions(+), 10 deletions(-)

diff --git a/drivers/gpu/drm/scheduler/sched_fence.c 
b/drivers/gpu/drm/scheduler/sched_fence.c
index 9391d6f0dc01..7a94e03341cb 100644
--- a/drivers/gpu/drm/scheduler/sched_fence.c
+++ b/drivers/gpu/drm/scheduler/sched_fence.c
@@ -156,19 +156,19 @@ static void drm_sched_fence_set_deadline_finished(struct 
dma_fence *f,
        struct dma_fence *parent;
        unsigned long flags;
- spin_lock_irqsave(&fence->lock, flags);
+       dma_fence_lock(f, flags);

Moving to dma_fence_lock should either be a separate patch or squashed into the one which converts many other drivers. Even a separate patch before that previous patch would be better.

Naming wise, I however still think dma_fence_lock_irqsave would probably be better to stick with the same pattern everyone is so used too.

Regards,

Tvrtko

/* If we already have an earlier deadline, keep it: */
        if (test_bit(DRM_SCHED_FENCE_FLAG_HAS_DEADLINE_BIT, &f->flags) &&
            ktime_before(fence->deadline, deadline)) {
-               spin_unlock_irqrestore(&fence->lock, flags);
+               dma_fence_unlock(f, flags);
                return;
        }
fence->deadline = deadline;
        set_bit(DRM_SCHED_FENCE_FLAG_HAS_DEADLINE_BIT, &f->flags);
- spin_unlock_irqrestore(&fence->lock, flags);
+       dma_fence_unlock(f, flags);
/*
         * smp_load_aquire() to ensure that if we are racing another
@@ -217,7 +217,6 @@ struct drm_sched_fence *drm_sched_fence_alloc(struct 
drm_sched_entity *entity,
fence->owner = owner;
        fence->drm_client_id = drm_client_id;
-       spin_lock_init(&fence->lock);
return fence;
  }
@@ -230,9 +229,9 @@ void drm_sched_fence_init(struct drm_sched_fence *fence,
        fence->sched = entity->rq->sched;
        seq = atomic_inc_return(&entity->fence_seq);
        dma_fence_init(&fence->scheduled, &drm_sched_fence_ops_scheduled,
-                      &fence->lock, entity->fence_context, seq);
+                      NULL, entity->fence_context, seq);
        dma_fence_init(&fence->finished, &drm_sched_fence_ops_finished,
-                      &fence->lock, entity->fence_context + 1, seq);
+                      NULL, entity->fence_context + 1, seq);
  }
module_init(drm_sched_fence_slab_init);
diff --git a/include/drm/gpu_scheduler.h b/include/drm/gpu_scheduler.h
index fb88301b3c45..b77f24a783e3 100644
--- a/include/drm/gpu_scheduler.h
+++ b/include/drm/gpu_scheduler.h
@@ -297,10 +297,6 @@ struct drm_sched_fence {
           * belongs to.
           */
        struct drm_gpu_scheduler        *sched;
-        /**
-         * @lock: the lock used by the scheduled and the finished fences.
-         */
-       spinlock_t                      lock;
          /**
           * @owner: job owner for debugging
           */

Reply via email to