On 09/10/21 04:12, Sean Christopherson wrote:
diff --git a/arch/arm64/kvm/arch_timer.c b/arch/arm64/kvm/arch_timer.c
index 7e8396f74010..addd53b6eba6 100644
--- a/arch/arm64/kvm/arch_timer.c
+++ b/arch/arm64/kvm/arch_timer.c
@@ -649,7 +649,6 @@ void kvm_timer_vcpu_put(struct kvm_vcpu *vcpu)
  {
        struct arch_timer_cpu *timer = vcpu_timer(vcpu);
        struct timer_map map;
-       struct rcuwait *wait = kvm_arch_vcpu_get_wait(vcpu);
if (unlikely(!timer->enabled))
                return;
@@ -672,7 +671,7 @@ void kvm_timer_vcpu_put(struct kvm_vcpu *vcpu)
        if (map.emul_ptimer)
                soft_timer_cancel(&map.emul_ptimer->hrtimer);
- if (rcuwait_active(wait))
+       if (kvm_vcpu_is_blocking(vcpu))
                kvm_timer_blocking(vcpu);
/*

So this trick is what you're applying to x86 too instead of using vmx_pre_block, I see.

Paolo

_______________________________________________
kvmarm mailing list
[email protected]
https://lists.cs.columbia.edu/mailman/listinfo/kvmarm

Reply via email to