On 06/11/2025 14:46, Boris Brezillon wrote:
> If the group is already assigned a slot but was idle before this job
> submission, we need to make sure the priority rotation happens in the
> future. Extract the existing logic living in group_schedule_locked()
> and call this new sched_resume_tick() helper from the "group is
> assigned a slot" path.
> 
> Fixes: de8548813824 ("drm/panthor: Add the scheduler logical block")
> Signed-off-by: Boris Brezillon <[email protected]>

Reviewed-by: Steven Price <[email protected]>

> ---
>  drivers/gpu/drm/panthor/panthor_sched.c | 43 +++++++++++++++++++------
>  1 file changed, 34 insertions(+), 9 deletions(-)
> 
> diff --git a/drivers/gpu/drm/panthor/panthor_sched.c 
> b/drivers/gpu/drm/panthor/panthor_sched.c
> index 7b164228af7b..923816397751 100644
> --- a/drivers/gpu/drm/panthor/panthor_sched.c
> +++ b/drivers/gpu/drm/panthor/panthor_sched.c
> @@ -2541,14 +2541,33 @@ static void sync_upd_work(struct work_struct *work)
>               sched_queue_delayed_work(sched, tick, 0);
>  }
>  
> +static void sched_resume_tick(struct panthor_device *ptdev)
> +{
> +     struct panthor_scheduler *sched = ptdev->scheduler;
> +     u64 delay_jiffies, now;
> +
> +     drm_WARN_ON(&ptdev->base, sched->resched_target != U64_MAX);
> +
> +     /* Scheduler tick was off, recalculate the resched_target based on the
> +      * last tick event, and queue the scheduler work.
> +      */
> +     now = get_jiffies_64();
> +     sched->resched_target = sched->last_tick + sched->tick_period;
> +     if (sched->used_csg_slot_count == sched->csg_slot_count &&
> +         time_before64(now, sched->resched_target))
> +             delay_jiffies = min_t(unsigned long, sched->resched_target - 
> now, ULONG_MAX);
> +     else
> +             delay_jiffies = 0;
> +
> +     sched_queue_delayed_work(sched, tick, delay_jiffies);
> +}
> +
>  static void group_schedule_locked(struct panthor_group *group, u32 
> queue_mask)
>  {
>       struct panthor_device *ptdev = group->ptdev;
>       struct panthor_scheduler *sched = ptdev->scheduler;
>       struct list_head *queue = &sched->groups.runnable[group->priority];
> -     u64 delay_jiffies = 0;
>       bool was_idle;
> -     u64 now;
>  
>       if (!group_can_run(group))
>               return;
> @@ -2593,13 +2612,7 @@ static void group_schedule_locked(struct panthor_group 
> *group, u32 queue_mask)
>       /* Scheduler tick was off, recalculate the resched_target based on the
>        * last tick event, and queue the scheduler work.
>        */
> -     now = get_jiffies_64();
> -     sched->resched_target = sched->last_tick + sched->tick_period;
> -     if (sched->used_csg_slot_count == sched->csg_slot_count &&
> -         time_before64(now, sched->resched_target))
> -             delay_jiffies = min_t(unsigned long, sched->resched_target - 
> now, ULONG_MAX);
> -
> -     sched_queue_delayed_work(sched, tick, delay_jiffies);
> +     sched_resume_tick(ptdev);
>  }
>  
>  static void queue_stop(struct panthor_queue *queue,
> @@ -3200,6 +3213,18 @@ queue_run_job(struct drm_sched_job *sched_job)
>  
>               group_schedule_locked(group, BIT(job->queue_idx));
>       } else {
> +             u32 queue_mask = BIT(job->queue_idx);
> +             bool resume_tick = group_is_idle(group) &&
> +                                (group->idle_queues & queue_mask) &&
> +                                !(group->blocked_queues & queue_mask) &&
> +                                sched->resched_target == U64_MAX;
> +
> +             /* We just added something to the queue, so it's no longer 
> idle. */
> +             group->idle_queues &= ~BIT(job->queue_idx);
> +
> +             if (resume_tick)
> +                     sched_resume_tick(ptdev);
> +
>               gpu_write(ptdev, CSF_DOORBELL(queue->doorbell_id), 1);
>               if (!sched->pm.has_ref &&
>                   !(group->blocked_queues & BIT(job->queue_idx))) {

Reply via email to