On Tue, Jul 29, 2014 at 01:53:02PM +0400, Kirill Tkhai wrote:
> From: Kirill Tkhai <[email protected]>
> 
>     sched: Teach scheduler to understand ONRQ_MIGRATING state
>     
>     This is new on_rq state for the cases when task is migrating
>     from one src_rq to another dst_rq, and there is no necessity
>     to have both RQs locked at the same time.
>     
>     We will use the state this way:
>     
>       raw_spin_lock(&src_rq->lock);
>       dequeue_task(src_rq, p, 0);
>       p->on_rq = ONRQ_MIGRATING;
>       set_task_cpu(p, dst_cpu);
>       raw_spin_unlock(&src_rq->lock);
>     
>       raw_spin_lock(&dst_rq->lock);
>       p->on_rq = ONRQ_QUEUED;
>       enqueue_task(dst_rq, p, 0);
>       raw_spin_unlock(&dst_rq->lock);
>     
>     The profit is that double_rq_lock() is not needed now,
>     and this may reduce the latencies in some situations.

You forgot to explain how the spinning on task_migrated() is bounded and
thus doesn't make your beginning and end contradict itself.

>     Signed-off-by: Kirill Tkhai <[email protected]>
> 
> diff --git a/kernel/sched/core.c b/kernel/sched/core.c
> index 26aa7bc..00d7bcc 100644
> --- a/kernel/sched/core.c
> +++ b/kernel/sched/core.c
> @@ -333,7 +333,8 @@ static inline struct rq *__task_rq_lock(struct 
> task_struct *p)
>       for (;;) {
>               rq = task_rq(p);
>               raw_spin_lock(&rq->lock);
> -             if (likely(rq == task_rq(p)))
> +             if (likely(rq == task_rq(p) &&
> +                        !task_migrating(p)))
>                       return rq;
>               raw_spin_unlock(&rq->lock);
>       }

I would prefer an extra spin-loop like so, that avoids us spinning on
the rq-lock, which serves no purpose.

diff --git a/kernel/sched/core.c b/kernel/sched/core.c
index 2676866b4394..1e65a0bdbbc3 100644
--- a/kernel/sched/core.c
+++ b/kernel/sched/core.c
@@ -331,9 +331,12 @@ static inline struct rq *__task_rq_lock(struct task_struct 
*p)
        lockdep_assert_held(&p->pi_lock);
 
        for (;;) {
+               while (task_migrating(p))
+                       cpu_relax();
+
                rq = task_rq(p);
                raw_spin_lock(&rq->lock);
-               if (likely(rq == task_rq(p)))
+               if (likely(rq == task_rq(p) && !task_migrating(p)))
                        return rq;
                raw_spin_unlock(&rq->lock);
        }

> diff --git a/kernel/sched/sched.h b/kernel/sched/sched.h
> index e5a9b6d..f6773d7 100644
> --- a/kernel/sched/sched.h
> +++ b/kernel/sched/sched.h
> @@ -17,6 +17,7 @@ struct rq;
>  
>  /* .on_rq states of struct task_struct: */

The 'normal' way to write that is: task_struct::on_rq

>  #define ONRQ_QUEUED  1
> +#define ONRQ_MIGRATING       2
>  
>  extern __read_mostly int scheduler_running;
>  
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [email protected]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Reply via email to