On Wed, Nov 12, 2025 at 3:52 AM Boris Brezillon
<[email protected]> wrote:
>
> When rotating group priorities, we want the group with the
> highest priority to go back to the end of the queue, and all
> other active groups to get their priority bumped, otherwise
> some groups will never get a chance to run with the highest
> priority. This implies moving the rotation itself to
> tick_work(), and only dealing with old group ordering in
> tick_ctx_insert_old_group().
>
> v2:
> - Add R-b
> - Fix the commit message
>
> Fixes: de8548813824 ("drm/panthor: Add the scheduler logical block")
> Signed-off-by: Boris Brezillon <[email protected]>
> Reviewed-by: Steven Price <[email protected]>
> ---
>  drivers/gpu/drm/panthor/panthor_sched.c | 47 +++++++++++++++----------
>  1 file changed, 29 insertions(+), 18 deletions(-)
>
> diff --git a/drivers/gpu/drm/panthor/panthor_sched.c 
> b/drivers/gpu/drm/panthor/panthor_sched.c
> index 94514d464e64..f08a05d36fc0 100644
> --- a/drivers/gpu/drm/panthor/panthor_sched.c
> +++ b/drivers/gpu/drm/panthor/panthor_sched.c
> @@ -1960,31 +1960,22 @@ tick_ctx_pick_groups_from_list(const struct 
> panthor_scheduler *sched,
>  static void
>  tick_ctx_insert_old_group(struct panthor_scheduler *sched,
>                           struct panthor_sched_tick_ctx *ctx,
> -                         struct panthor_group *group,
> -                         bool full_tick)
> +                         struct panthor_group *group)
>  {
>         struct panthor_csg_slot *csg_slot = &sched->csg_slots[group->csg_id];
>         struct panthor_group *other_group;
>
> -       if (!full_tick) {
> -               list_add_tail(&group->run_node, 
> &ctx->old_groups[group->priority]);
> -               return;
> -       }
> -
> -       /* Rotate to make sure groups with lower CSG slot
> -        * priorities have a chance to get a higher CSG slot
> -        * priority next time they get picked. This priority
> -        * has an impact on resource request ordering, so it's
> -        * important to make sure we don't let one group starve
> -        * all other groups with the same group priority.
> -        */
> +       /* Class groups in descending priority order so we can easily rotate. 
> */
>         list_for_each_entry(other_group,
>                             &ctx->old_groups[csg_slot->group->priority],
>                             run_node) {
>                 struct panthor_csg_slot *other_csg_slot = 
> &sched->csg_slots[other_group->csg_id];
>
> -               if (other_csg_slot->priority > csg_slot->priority) {
> -                       list_add_tail(&csg_slot->group->run_node, 
> &other_group->run_node);
> +               /* Our group has a higher prio than the one we're testing 
> against,
> +                * place it just before.
> +                */
> +               if (csg_slot->priority > other_csg_slot->priority) {
> +                       list_add_tail(&group->run_node, 
> &other_group->run_node);
>                         return;
>                 }
>         }
> @@ -2033,7 +2024,7 @@ tick_ctx_init(struct panthor_scheduler *sched,
>                                 group->fatal_queues |= 
> GENMASK(group->queue_count - 1, 0);
>                 }
>
> -               tick_ctx_insert_old_group(sched, ctx, group, full_tick);
> +               tick_ctx_insert_old_group(sched, ctx, group);
This drops the only user of full_tick. We can omit the parameter from
tick_ctx_init.
>                 csgs_upd_ctx_queue_reqs(ptdev, &upd_ctx, i,
>                                         csg_iface->output->ack ^ 
> CSG_STATUS_UPDATE,
>                                         CSG_STATUS_UPDATE);
> @@ -2399,9 +2390,29 @@ static void tick_work(struct work_struct *work)
>         for (prio = PANTHOR_CSG_PRIORITY_COUNT - 1;
>              prio >= 0 && !tick_ctx_is_full(sched, &ctx);
>              prio--) {
> +               struct panthor_group *old_highest_prio_group =
> +                       list_first_entry_or_null(&ctx.old_groups[prio],
> +                                                struct panthor_group, 
> run_node);
> +
> +               /* Pull out the group with the highest prio for rotation. */
> +               if (old_highest_prio_group)
> +                       list_del(&old_highest_prio_group->run_node);
> +
> +               /* Re-insert old active groups so they get a chance to run 
> with higher prio. */
> +               tick_ctx_pick_groups_from_list(sched, &ctx, 
> &ctx.old_groups[prio], true, true);
> +
> +               /* Fill the remaining slots with runnable groups. */
>                 tick_ctx_pick_groups_from_list(sched, &ctx, 
> &sched->groups.runnable[prio],
>                                                true, false);
> -               tick_ctx_pick_groups_from_list(sched, &ctx, 
> &ctx.old_groups[prio], true, true);
> +
> +               /* Re-insert the old group with the highest prio, and give it 
> a chance to be
> +                * scheduled again (but with a lower prio) if there's room 
> left.
> +                */
> +               if (old_highest_prio_group) {
> +                       list_add_tail(&old_highest_prio_group->run_node, 
> &ctx.old_groups[prio]);
> +                       tick_ctx_pick_groups_from_list(sched, &ctx, 
> &ctx.old_groups[prio],
> +                                                      true, true);
> +               }
>         }
>
>         /* If we have free CSG slots left, pick idle groups */
> --
> 2.51.1
>

Reply via email to