We want migrate_disable() tasks to get PULLs in order for them to PUSH
away the higher priority task.

Signed-off-by: Peter Zijlstra (Intel) <[email protected]>
---
 kernel/sched/cpudeadline.c |    4 ++--
 kernel/sched/cpupri.c      |    4 ++--
 kernel/sched/deadline.c    |    4 ++--
 kernel/sched/rt.c          |    4 ++--
 4 files changed, 8 insertions(+), 8 deletions(-)

--- a/kernel/sched/cpudeadline.c
+++ b/kernel/sched/cpudeadline.c
@@ -120,7 +120,7 @@ int cpudl_find(struct cpudl *cp, struct
        const struct sched_dl_entity *dl_se = &p->dl;
 
        if (later_mask &&
-           cpumask_and(later_mask, cp->free_cpus, p->cpus_ptr)) {
+           cpumask_and(later_mask, cp->free_cpus, &p->cpus_mask)) {
                unsigned long cap, max_cap = 0;
                int cpu, max_cpu = -1;
 
@@ -151,7 +151,7 @@ int cpudl_find(struct cpudl *cp, struct
 
                WARN_ON(best_cpu != -1 && !cpu_present(best_cpu));
 
-               if (cpumask_test_cpu(best_cpu, p->cpus_ptr) &&
+               if (cpumask_test_cpu(best_cpu, &p->cpus_mask) &&
                    dl_time_before(dl_se->deadline, cp->elements[0].dl)) {
                        if (later_mask)
                                cpumask_set_cpu(best_cpu, later_mask);
--- a/kernel/sched/cpupri.c
+++ b/kernel/sched/cpupri.c
@@ -73,11 +73,11 @@ static inline int __cpupri_find(struct c
        if (skip)
                return 0;
 
-       if (cpumask_any_and(p->cpus_ptr, vec->mask) >= nr_cpu_ids)
+       if (cpumask_any_and(&p->cpus_mask, vec->mask) >= nr_cpu_ids)
                return 0;
 
        if (lowest_mask) {
-               cpumask_and(lowest_mask, p->cpus_ptr, vec->mask);
+               cpumask_and(lowest_mask, &p->cpus_mask, vec->mask);
 
                /*
                 * We have to ensure that we have at least one bit
--- a/kernel/sched/deadline.c
+++ b/kernel/sched/deadline.c
@@ -1888,7 +1888,7 @@ static void task_fork_dl(struct task_str
 static int pick_dl_task(struct rq *rq, struct task_struct *p, int cpu)
 {
        if (!task_running(rq, p) &&
-           cpumask_test_cpu(cpu, p->cpus_ptr))
+           cpumask_test_cpu(cpu, &p->cpus_mask))
                return 1;
        return 0;
 }
@@ -2038,7 +2038,7 @@ static struct rq *find_lock_later_rq(str
                /* Retry if something changed. */
                if (double_lock_balance(rq, later_rq)) {
                        if (unlikely(task_rq(task) != rq ||
-                                    !cpumask_test_cpu(later_rq->cpu, 
task->cpus_ptr) ||
+                                    !cpumask_test_cpu(later_rq->cpu, 
&task->cpus_mask) ||
                                     task_running(rq, task) ||
                                     !dl_task(task) ||
                                     !task_on_rq_queued(task))) {
--- a/kernel/sched/rt.c
+++ b/kernel/sched/rt.c
@@ -1658,7 +1658,7 @@ static void put_prev_task_rt(struct rq *
 static int pick_rt_task(struct rq *rq, struct task_struct *p, int cpu)
 {
        if (!task_running(rq, p) &&
-           cpumask_test_cpu(cpu, p->cpus_ptr))
+           cpumask_test_cpu(cpu, &p->cpus_mask))
                return 1;
 
        return 0;
@@ -1811,7 +1811,7 @@ static struct rq *find_lock_lowest_rq(st
                         * Also make sure that it wasn't scheduled on its rq.
                         */
                        if (unlikely(task_rq(task) != rq ||
-                                    !cpumask_test_cpu(lowest_rq->cpu, 
task->cpus_ptr) ||
+                                    !cpumask_test_cpu(lowest_rq->cpu, 
&task->cpus_mask) ||
                                     task_running(rq, task) ||
                                     !rt_task(task) ||
                                     !task_on_rq_queued(task))) {


Reply via email to