On Tue, Nov 12, 2024 at 10:06:21AM +0100, Philipp Stanner wrote:
> Hi Matt,
> 
> On Sat, 2024-11-09 at 09:29 -0800, Matthew Brost wrote:
> > Follow the semantics of DMA_RESV_USAGE_PREEMPT in the DRM scheduler
> > by
> > storing preemptive fences in a dedicated xarray, which is waited on
> > after all other fences are signaled. In addition to following these
> > semantics, pipeline preemptive fences by enabling signaling on all
> > preemptive fences before waiting on any of them.
> 
> the commit message lacks the *motivation*. Why is the current state a
> problem, why is that feature needed etc.
> 

Yes, I do this in the cover letter but this is missing here. Will add in
next rev.

> > 
> > Cc: Philipp Stanner <pstan...@redhat.com>
> > Cc: Danilo Krummrich <d...@kernel.org>
> > Cc: Luben Tuikov <ltuiko...@gmail.com>
> > Cc: Dave Airlie <airl...@redhat.com>
> > Cc: Simona Vetter <simona.vet...@ffwll.ch>
> > Cc: Christian Koenig <christian.koe...@amd.com>
> > Signed-off-by: Matthew Brost <matthew.br...@intel.com>
> > ---
> >  drivers/gpu/drm/scheduler/sched_entity.c | 29 ++++++++++++--
> >  drivers/gpu/drm/scheduler/sched_main.c   | 48 ++++++++++++++++------
> > --
> >  include/drm/gpu_scheduler.h              | 15 ++++++++
> >  3 files changed, 73 insertions(+), 19 deletions(-)
> > 
> > diff --git a/drivers/gpu/drm/scheduler/sched_entity.c
> > b/drivers/gpu/drm/scheduler/sched_entity.c
> > index 69bcf0e99d57..c6c4978aa65a 100644
> > --- a/drivers/gpu/drm/scheduler/sched_entity.c
> > +++ b/drivers/gpu/drm/scheduler/sched_entity.c
> > @@ -201,11 +201,13 @@ static void
> > drm_sched_entity_kill_jobs_cb(struct dma_fence *f,
> >     struct drm_sched_job *job = container_of(cb, struct
> > drm_sched_job,
> >                                              finish_cb);
> >     unsigned long index;
> > +   struct xarray *dependencies = &job->dependencies;
> >  
> >     dma_fence_put(f);
> >  
> > +again:
> >     /* Wait for all dependencies to avoid data corruptions */
> > -   xa_for_each(&job->dependencies, index, f) {
> > +   xa_for_each(dependencies, index, f) {
> >             struct drm_sched_fence *s_fence =
> > to_drm_sched_fence(f);
> >  
> >             if (s_fence && f == &s_fence->scheduled) {
> > @@ -223,7 +225,7 @@ static void drm_sched_entity_kill_jobs_cb(struct
> > dma_fence *f,
> >                     dma_fence_put(&s_fence->scheduled);
> >             }
> >  
> > -           xa_erase(&job->dependencies, index);
> > +           xa_erase(dependencies, index);
> >             if (f && !dma_fence_add_callback(f, &job->finish_cb,
> >                                             
> > drm_sched_entity_kill_jobs_cb))
> >                     return;
> > @@ -231,6 +233,11 @@ static void drm_sched_entity_kill_jobs_cb(struct
> > dma_fence *f,
> >             dma_fence_put(f);
> >     }
> >  
> > +   if (dependencies != &job->preempt_dependencies) {
> > +           dependencies = &job->preempt_dependencies;
> > +           goto again;
> > +   }
> > +
> 
> I think this should have a comment. It can only trigger once, right? So
> I guess that's why it doesn't make sense considering making it a loop
> instead of goto upwards?
> 

Yes, can only trigger once. I personally don't mind goto while others
find them offensive.

> >     INIT_WORK(&job->work, drm_sched_entity_kill_jobs_work);
> >     schedule_work(&job->work);
> >  }
> > @@ -456,17 +463,33 @@ drm_sched_job_dependency(struct drm_sched_job
> > *job,
> >                      struct drm_sched_entity *entity)
> >  {
> >     struct dma_fence *f;
> > +   struct xarray *dependencies;
> > +
> > +again:
> > +   dependencies = job->resolve_preempt_dependencies ?
> > +           &job->preempt_dependencies : &job->dependencies;
> 
> I don't think it's good to use the ternary operator for such long
> statements.
> 
> if-else is more readable.
>

Sure.
 
> >  
> >     /* We keep the fence around, so we can iterate over all
> > dependencies
> >      * in drm_sched_entity_kill_jobs_cb() to ensure all deps are
> > signaled
> >      * before killing the job.
> >      */
> > -   f = xa_load(&job->dependencies, job->last_dependency);
> > +   f = xa_load(dependencies, job->last_dependency);
> >     if (f) {
> >             job->last_dependency++;
> >             return dma_fence_get(f);
> >     }
> >  
> > +   /* Switch resolving preempt dependencies pipelining
> > signaling */
> 
> I don't understand this comment. I guess you want to say that this section 
> resolves preemption dependencies for the (fence) pipeline signaling?
> 

'Switch to resolving preempt dependencies. Enabling signaling on all
preempt dependencies to pipeline the hardware preemption'

Is that better more / clear?

> > +   if (!job->resolve_preempt_dependencies) {
> > +           unsigned long index;
> > +
> > +           xa_for_each(&job->preempt_dependencies, index, f)
> > +                   dma_fence_enable_sw_signaling(f);
> > +
> > +           job->resolve_preempt_dependencies = true;
> 
> Hm, is this set to false ever again? It seems it doesn't need to? So
> the goto again is only ever triggered once?
> 

resolve_preempt_dependencies can only go from 0 - > 1 exactly one time.

> > +           goto again;
> > +   }
> > +
> >     if (job->sched->ops->prepare_job)
> >             return job->sched->ops->prepare_job(job, entity);
> >  
> > diff --git a/drivers/gpu/drm/scheduler/sched_main.c
> > b/drivers/gpu/drm/scheduler/sched_main.c
> > index 7ce25281c74c..eceb9b8c6f5f 100644
> > --- a/drivers/gpu/drm/scheduler/sched_main.c
> > +++ b/drivers/gpu/drm/scheduler/sched_main.c
> > @@ -829,6 +829,7 @@ int drm_sched_job_init(struct drm_sched_job *job,
> >     INIT_LIST_HEAD(&job->list);
> >  
> >     xa_init_flags(&job->dependencies, XA_FLAGS_ALLOC);
> > +   xa_init_flags(&job->preempt_dependencies, XA_FLAGS_ALLOC);
> >  
> >     return 0;
> >  }
> > @@ -864,21 +865,14 @@ void drm_sched_job_arm(struct drm_sched_job
> > *job)
> >  }
> >  EXPORT_SYMBOL(drm_sched_job_arm);
> >  
> > -/**
> > - * drm_sched_job_add_dependency - adds the fence as a job dependency
> > - * @job: scheduler job to add the dependencies to
> > - * @fence: the dma_fence to add to the list of dependencies.
> > - *
> > - * Note that @fence is consumed in both the success and error cases.
> > - *
> > - * Returns:
> > - * 0 on success, or an error on failing to expand the array.
> > - */
> > -int drm_sched_job_add_dependency(struct drm_sched_job *job,
> > -                            struct dma_fence *fence)
> > +static int __drm_sched_job_add_dependency(struct drm_sched_job *job,
> > +                                     struct dma_fence *fence,
> > +                                     bool is_preempt)
> >  {
> >     struct dma_fence *entry;
> >     unsigned long index;
> > +   struct xarray *dependencies = is_preempt ? &job-
> > >preempt_dependencies :
> > +           &job->dependencies;
> 
> Same – is better as an if-else below
> 

Sure.

> >     u32 id = 0;
> >     int ret;
> >  
> > @@ -889,25 +883,41 @@ int drm_sched_job_add_dependency(struct
> > drm_sched_job *job,
> >      * This lets the size of the array of deps scale with the
> > number of
> >      * engines involved, rather than the number of BOs.
> >      */
> > -   xa_for_each(&job->dependencies, index, entry) {
> > +   xa_for_each(dependencies, index, entry) {
> >             if (entry->context != fence->context)
> >                     continue;
> >  
> >             if (dma_fence_is_later(fence, entry)) {
> >                     dma_fence_put(entry);
> > -                   xa_store(&job->dependencies, index, fence,
> > GFP_KERNEL);
> > +                   xa_store(dependencies, index, fence,
> > GFP_KERNEL);
> >             } else {
> >                     dma_fence_put(fence);
> >             }
> >             return 0;
> >     }
> >  
> > -   ret = xa_alloc(&job->dependencies, &id, fence, xa_limit_32b,
> > GFP_KERNEL);
> > +   ret = xa_alloc(dependencies, &id, fence, xa_limit_32b,
> > GFP_KERNEL);
> >     if (ret != 0)
> >             dma_fence_put(fence);
> >  
> >     return ret;
> >  }
> > +
> > +/**
> > + * drm_sched_job_add_dependency - adds the fence as a job dependency
> > + * @job: scheduler job to add the dependencies to
> > + * @fence: the dma_fence to add to the list of dependencies.
> > + *
> > + * Note that @fence is consumed in both the success and error cases.
> > + *
> > + * Returns:
> > + * 0 on success, or an error on failing to expand the array.
> > + */
> > +int drm_sched_job_add_dependency(struct drm_sched_job *job,
> > +                            struct dma_fence *fence)
> > +{
> > +   return __drm_sched_job_add_dependency(job, fence, false);
> > +}
> >  EXPORT_SYMBOL(drm_sched_job_add_dependency);
> >  
> >  /**
> > @@ -963,7 +973,9 @@ int drm_sched_job_add_resv_dependencies(struct
> > drm_sched_job *job,
> >     dma_resv_for_each_fence(&cursor, resv, usage, fence) {
> >             /* Make sure to grab an additional ref on the added
> > fence */
> >             dma_fence_get(fence);
> > -           ret = drm_sched_job_add_dependency(job, fence);
> > +           ret = __drm_sched_job_add_dependency(job, fence,
> > +                                               
> > cursor.fence_usage ==
> > +                                               
> > DMA_RESV_USAGE_PREEMPT);
> >             if (ret) {
> >                     dma_fence_put(fence);
> >                     return ret;
> > @@ -1030,6 +1042,10 @@ void drm_sched_job_cleanup(struct
> > drm_sched_job *job)
> >     }
> >     xa_destroy(&job->dependencies);
> >  
> > +   xa_for_each(&job->preempt_dependencies, index, fence) {
> > +           dma_fence_put(fence);
> > +   }
> > +   xa_destroy(&job->preempt_dependencies);
> >  }
> >  EXPORT_SYMBOL(drm_sched_job_cleanup);
> >  
> > diff --git a/include/drm/gpu_scheduler.h
> > b/include/drm/gpu_scheduler.h
> > index 95e17504e46a..de16cf6b1869 100644
> > --- a/include/drm/gpu_scheduler.h
> > +++ b/include/drm/gpu_scheduler.h
> > @@ -353,6 +353,13 @@ struct drm_sched_job {
> >  
> >     u32                             credits;
> >  
> > +   /**
> > +    * @resolve_preempt_dependencies:
> > +    *
> > +    * Job is currently resolving preempt dependencies.
> > +    */
> > +   bool                            resolve_preempt_dependencies
> > ;
> 
> I think this should be called "resolving_preempt_dependencies". Just 2
> letters more and it emphasizes that this is happening "currently".
> 

That is more clear. Will rename.

Matt

> 
> P.
> 
> > +
> >     /*
> >      * work is used only after finish_cb has been used and will
> > not be
> >      * accessed anymore.
> > @@ -376,6 +383,14 @@ struct drm_sched_job {
> >      */
> >     struct xarray                   dependencies;
> >  
> > +   /**
> > +    * @preempt_dependencies:
> > +    *
> > +    * Contains the dependencies as struct dma_fence for this
> > job which are
> > +    * preempt fences.
> > +    */
> > +   struct xarray                   preempt_dependencies;
> > +
> >     /** @last_dependency: tracks @dependencies as they signal */
> >     unsigned long                   last_dependency;
> >  
> 

Reply via email to