On 06/11/2025 14:46, Boris Brezillon wrote:
> When we have multiple active groups with the same priority, we need to
> keep ticking for the priority rotation to take place. If we don't do
> that, we might starve slots with lower priorities.
>
> It's annoying to deal with that in tick_ctx_update_resched_target(),
> so let's add a ::stop_tick field to the tick context which is
> initialized to true, and downgraded to false as soon as we detect
> something that requires to tick to happen. This way we can complement
> the current logic with extra conditions if needed.
>
> Fixes: de8548813824 ("drm/panthor: Add the scheduler logical block")
> Signed-off-by: Boris Brezillon <[email protected]>
Reviewed-by: Steven Price <[email protected]>
> ---
> drivers/gpu/drm/panthor/panthor_sched.c | 36 ++++++++++++-------------
> 1 file changed, 17 insertions(+), 19 deletions(-)
>
> diff --git a/drivers/gpu/drm/panthor/panthor_sched.c
> b/drivers/gpu/drm/panthor/panthor_sched.c
> index 1eba56e7360d..7b164228af7b 100644
> --- a/drivers/gpu/drm/panthor/panthor_sched.c
> +++ b/drivers/gpu/drm/panthor/panthor_sched.c
> @@ -1882,6 +1882,7 @@ struct panthor_sched_tick_ctx {
> struct panthor_vm *vms[MAX_CS_PER_CSG];
> u32 as_count;
> bool immediate_tick;
> + bool stop_tick;
> u32 csg_upd_failed_mask;
> };
>
> @@ -1941,10 +1942,17 @@ tick_ctx_pick_groups_from_list(const struct
> panthor_scheduler *sched,
> if (!owned_by_tick_ctx)
> group_get(group);
>
> - list_move_tail(&group->run_node, &ctx->groups[group->priority]);
> ctx->group_count++;
> +
> + /* If we have more than one active group with the same priority,
> + * we need to keep ticking to rotate the CSG priority.
> + */
> if (group_is_idle(group))
> ctx->idle_group_count++;
> + else if (!list_empty(&ctx->groups[group->priority]))
> + ctx->stop_tick = false;
> +
> + list_move_tail(&group->run_node, &ctx->groups[group->priority]);
>
> if (i == ctx->as_count)
> ctx->vms[ctx->as_count++] = group->vm;
> @@ -1996,6 +2004,7 @@ tick_ctx_init(struct panthor_scheduler *sched,
> memset(ctx, 0, sizeof(*ctx));
> csgs_upd_ctx_init(&upd_ctx);
>
> + ctx->stop_tick = true;
> ctx->min_priority = PANTHOR_CSG_PRIORITY_COUNT;
> for (i = 0; i < ARRAY_SIZE(ctx->groups); i++) {
> INIT_LIST_HEAD(&ctx->groups[i]);
> @@ -2308,32 +2317,21 @@ static u64
> tick_ctx_update_resched_target(struct panthor_scheduler *sched,
> const struct panthor_sched_tick_ctx *ctx)
> {
> - /* We had space left, no need to reschedule until some external event
> happens. */
> - if (!tick_ctx_is_full(sched, ctx))
> - goto no_tick;
> + u64 resched_target;
>
> - /* If idle groups were scheduled, no need to wake up until some external
> - * event happens (group unblocked, new job submitted, ...).
> - */
> - if (ctx->idle_group_count)
> + if (ctx->stop_tick)
> goto no_tick;
>
> if (drm_WARN_ON(&sched->ptdev->base, ctx->min_priority >=
> PANTHOR_CSG_PRIORITY_COUNT))
> goto no_tick;
>
> - /* If there are groups of the same priority waiting, we need to
> - * keep the scheduler ticking, otherwise, we'll just wait for
> - * new groups with higher priority to be queued.
> - */
> - if (!list_empty(&sched->groups.runnable[ctx->min_priority])) {
> - u64 resched_target = sched->last_tick + sched->tick_period;
> + resched_target = sched->last_tick + sched->tick_period;
>
> - if (time_before64(sched->resched_target, sched->last_tick) ||
> - time_before64(resched_target, sched->resched_target))
> - sched->resched_target = resched_target;
> + if (time_before64(sched->resched_target, sched->last_tick) ||
> + time_before64(resched_target, sched->resched_target))
> + sched->resched_target = resched_target;
>
> - return sched->resched_target - sched->last_tick;
> - }
> + return sched->resched_target - sched->last_tick;
>
> no_tick:
> sched->resched_target = U64_MAX;