On Thu, 2025-04-24 at 11:55 +0200, Philipp Stanner wrote:
> The unit tests so far took care manually of avoiding memory leaks
> that
> might have occurred when calling drm_sched_fini().
> 
> The scheduler now takes care by itself of avoiding memory leaks if
> the
> driver provides the callback
> drm_sched_backend_ops.kill_fence_context().
> 
> Implement that callback for the unit tests. Remove the manual cleanup
> code.

@Tvrtko: On a scale from 1-10, how much do you love this patch? :)

P.

> 
> Signed-off-by: Philipp Stanner <pha...@kernel.org>
> ---
>  .../gpu/drm/scheduler/tests/mock_scheduler.c  | 34 ++++++++++++-----
> --
>  1 file changed, 21 insertions(+), 13 deletions(-)
> 
> diff --git a/drivers/gpu/drm/scheduler/tests/mock_scheduler.c
> b/drivers/gpu/drm/scheduler/tests/mock_scheduler.c
> index f999c8859cf7..a72d26ca8262 100644
> --- a/drivers/gpu/drm/scheduler/tests/mock_scheduler.c
> +++ b/drivers/gpu/drm/scheduler/tests/mock_scheduler.c
> @@ -228,10 +228,30 @@ static void mock_sched_free_job(struct
> drm_sched_job *sched_job)
>       /* Mock job itself is freed by the kunit framework. */
>  }
>  
> +static void mock_sched_fence_context_kill(struct drm_gpu_scheduler
> *gpu_sched)
> +{
> +     struct drm_mock_scheduler *sched =
> drm_sched_to_mock_sched(gpu_sched);
> +     struct drm_mock_sched_job *job;
> +     unsigned long flags;
> +
> +     spin_lock_irqsave(&sched->lock, flags);
> +     list_for_each_entry(job, &sched->job_list, link) {
> +             spin_lock(&job->lock);
> +             if (!dma_fence_is_signaled_locked(&job->hw_fence)) {
> +                     dma_fence_set_error(&job->hw_fence, -
> ECANCELED);
> +                     dma_fence_signal_locked(&job->hw_fence);
> +             }
> +             complete(&job->done);
> +             spin_unlock(&job->lock);
> +     }
> +     spin_unlock_irqrestore(&sched->lock, flags);
> +}
> +
>  static const struct drm_sched_backend_ops drm_mock_scheduler_ops = {
>       .run_job = mock_sched_run_job,
>       .timedout_job = mock_sched_timedout_job,
> -     .free_job = mock_sched_free_job
> +     .free_job = mock_sched_free_job,
> +     .kill_fence_context = mock_sched_fence_context_kill,
>  };
>  
>  /**
> @@ -300,18 +320,6 @@ void drm_mock_sched_fini(struct
> drm_mock_scheduler *sched)
>               drm_mock_sched_job_complete(job);
>       spin_unlock_irqrestore(&sched->lock, flags);
>  
> -     /*
> -      * Free completed jobs and jobs not yet processed by the DRM
> scheduler
> -      * free worker.
> -      */
> -     spin_lock_irqsave(&sched->lock, flags);
> -     list_for_each_entry_safe(job, next, &sched->done_list, link)
> -             list_move_tail(&job->link, &list);
> -     spin_unlock_irqrestore(&sched->lock, flags);
> -
> -     list_for_each_entry_safe(job, next, &list, link)
> -             mock_sched_free_job(&job->base);
> -
>       drm_sched_fini(&sched->base);
>  }
>  

Reply via email to