Since 83a0a96a5f26 (sched/fair: Leverage the idle state info when choosing the "idlest" cpu) find_idlest_cpu no longer returns -1.
Signed-off-by: Brendan Jackman <brendan.jack...@arm.com> Cc: Dietmar Eggemann <dietmar.eggem...@arm.com> Cc: Vincent Guittot <vincent.guit...@linaro.org> Cc: Josef Bacik <jo...@toxicpanda.com> Cc: Ingo Molnar <mi...@redhat.com> Cc: Morten Rasmussen <morten.rasmus...@arm.com> Cc: Peter Zijlstra <pet...@infradead.org> --- kernel/sched/fair.c | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c index c95880e216f6..64618d768546 100644 --- a/kernel/sched/fair.c +++ b/kernel/sched/fair.c @@ -5934,7 +5934,7 @@ select_task_rq_fair(struct task_struct *p, int prev_cpu, int sd_flag, int wake_f } new_cpu = find_idlest_cpu(group, p, cpu); - if (new_cpu == -1 || new_cpu == cpu) { + if (new_cpu == cpu) { /* Now try balancing at a lower domain level of cpu */ sd = sd->child; continue; -- 2.14.1