On Tue, Oct 20, 2015 at 04:00:31PM +0200, Peter Zijlstra wrote:
> On Tue, Oct 20, 2015 at 09:28:08AM +0200, Daniel Wagner wrote:
> > diff --git a/arch/powerpc/kvm/book3s_hv.c b/arch/powerpc/kvm/book3s_hv.c
> > index 2280497..f534e15 100644
> > --- a/arch/powerpc/kvm/book3s_hv.c
> > +++ b/arch/powerpc/kvm/book3s_hv.c
> > @@ -2560,10 +2560,9 @@ static void kvmppc_vcore_blocked(struct kvmppc_vcore 
> > *vc)
> >  {
> >     struct kvm_vcpu *vcpu;
> >     int do_sleep = 1;
> > +   DECLARE_SWAITQUEUE(wait);
> >  
> > -   DEFINE_WAIT(wait);
> > -
> > -   prepare_to_wait(&vc->wq, &wait, TASK_INTERRUPTIBLE);
> > +   prepare_to_swait(&vc->wq, &wait, TASK_INTERRUPTIBLE);
> >  
> >     /*
> >      * Check one last time for pending exceptions and ceded state after
> > @@ -2577,7 +2576,7 @@ static void kvmppc_vcore_blocked(struct kvmppc_vcore 
> > *vc)
> >     }
> >  
> >     if (!do_sleep) {
> > -           finish_wait(&vc->wq, &wait);
> > +           finish_swait(&vc->wq, &wait);
> >             return;
> >     }
> >  
> > @@ -2585,7 +2584,7 @@ static void kvmppc_vcore_blocked(struct kvmppc_vcore 
> > *vc)
> >     trace_kvmppc_vcore_blocked(vc, 0);
> >     spin_unlock(&vc->lock);
> >     schedule();
> > -   finish_wait(&vc->wq, &wait);
> > +   finish_swait(&vc->wq, &wait);
> >     spin_lock(&vc->lock);
> >     vc->vcore_state = VCORE_INACTIVE;
> >     trace_kvmppc_vcore_blocked(vc, 1);
> 
> This one looks buggy, one should _NOT_ assume that your blocking
> condition is true after schedule().

Do you mean it's buggy in calling finish_swait there, or it's buggy in
not immediately re-checking the condition?  If the latter, then it's
OK because the sole caller of this function calls it in a loop and
checks the condition (all runnable vcpus in this vcore are idle) each
time around the loop.

> > diff --git a/virt/kvm/kvm_main.c b/virt/kvm/kvm_main.c
> > index 8db1d93..45ab55f 100644
> > --- a/virt/kvm/kvm_main.c
> > +++ b/virt/kvm/kvm_main.c
> > @@ -2019,7 +2018,7 @@ void kvm_vcpu_block(struct kvm_vcpu *vcpu)
> >     }
> >  
> >     for (;;) {
> > -           prepare_to_wait(&vcpu->wq, &wait, TASK_INTERRUPTIBLE);
> > +           prepare_to_swait(&vcpu->wq, &wait, TASK_INTERRUPTIBLE);
> >  
> >             if (kvm_vcpu_check_block(vcpu) < 0)
> >                     break;
> > @@ -2028,7 +2027,7 @@ void kvm_vcpu_block(struct kvm_vcpu *vcpu)
> >             schedule();
> >     }
> >  
> > -   finish_wait(&vcpu->wq, &wait);
> > +   finish_swait(&vcpu->wq, &wait);
> >     cur = ktime_get();
> >  
> >  out:
> 
> Should we not take this opportunity to get rid of these open-coded wait
> loops?
> 
> 
> Does this work?
> 
> ---
>  arch/powerpc/kvm/book3s_hv.c | 33 +++++++++++++++++----------------
>  virt/kvm/kvm_main.c          | 13 ++-----------
>  2 files changed, 19 insertions(+), 27 deletions(-)
> 
> diff --git a/arch/powerpc/kvm/book3s_hv.c b/arch/powerpc/kvm/book3s_hv.c
> index 228049786888..b5b8bcad5105 100644
> --- a/arch/powerpc/kvm/book3s_hv.c
> +++ b/arch/powerpc/kvm/book3s_hv.c
> @@ -2552,18 +2552,10 @@ static void kvmppc_wait_for_exec(struct kvmppc_vcore 
> *vc,
>       finish_wait(&vcpu->arch.cpu_run, &wait);
>  }
>  
> -/*
> - * All the vcpus in this vcore are idle, so wait for a decrementer
> - * or external interrupt to one of the vcpus.  vc->lock is held.
> - */
> -static void kvmppc_vcore_blocked(struct kvmppc_vcore *vc)
> +static inline bool kvmppc_vcore_should_sleep(struct kvmppc_vcore *vc)

This function could also be used in kvmppc_run_vcpu().

>  {
>       struct kvm_vcpu *vcpu;
> -     int do_sleep = 1;
> -
> -     DEFINE_WAIT(wait);
> -
> -     prepare_to_wait(&vc->wq, &wait, TASK_INTERRUPTIBLE);
> +     bool sleep = true;
>  
>       /*
>        * Check one last time for pending exceptions and ceded state after
> @@ -2571,26 +2563,35 @@ static void kvmppc_vcore_blocked(struct kvmppc_vcore 
> *vc)
>        */
>       list_for_each_entry(vcpu, &vc->runnable_threads, arch.run_list) {
>               if (vcpu->arch.pending_exceptions || !vcpu->arch.ceded) {
> -                     do_sleep = 0;
> +                     sleep = false;
>                       break;
>               }
>       }
>  
> -     if (!do_sleep) {
> -             finish_wait(&vc->wq, &wait);
> -             return;
> -     }
> +     return sleep;
> +}
>  
> +static inline void kvmppc_vcore_schedule(struct kvmppc_vcore *vc)
> +{
>       vc->vcore_state = VCORE_SLEEPING;
>       trace_kvmppc_vcore_blocked(vc, 0);
>       spin_unlock(&vc->lock);
>       schedule();
> -     finish_wait(&vc->wq, &wait);
>       spin_lock(&vc->lock);
>       vc->vcore_state = VCORE_INACTIVE;
>       trace_kvmppc_vcore_blocked(vc, 1);
>  }
>  
> +/*
> + * All the vcpus in this vcore are idle, so wait for a decrementer
> + * or external interrupt to one of the vcpus.  vc->lock is held.
> + */
> +static void kvmppc_vcore_blocked(struct kvmppc_vcore *vc)
> +{
> +     ___wait_event(vc->wq, !kvmppc_vcore_should_sleep(vc), TASK_IDLE, 0, 0,
> +                   kvmppc_vcore_schedule(vc));

Wow, triple underscores, that must be an ultra-trendy function. :)

> +}
> +
>  static int kvmppc_run_vcpu(struct kvm_run *kvm_run, struct kvm_vcpu *vcpu)
>  {
>       int n_ceded;

That all looks OK at a first glance, I'll give it a whirl.

Paul.
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [email protected]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Reply via email to