Gleb Natapov wrote:
> On Fri, Apr 17, 2009 at 03:12:57PM +0000, Dmitry Eremin-Solenikov wrote:
>> This patch does expose some problems on real HW. The first NMI completes w/o
>> problems. However If I try to boot the kernel w/ nmi_watchdog=1 or to trigger
>> two NMIs from the monitor, kernel is stuck somewhere.
>>
> Can you try this patch instead patch13:
> 
> 
> diff --git a/arch/x86/include/asm/kvm_host.h b/arch/x86/include/asm/kvm_host.h
> index 8b6f6e9..057a612 100644
> --- a/arch/x86/include/asm/kvm_host.h
> +++ b/arch/x86/include/asm/kvm_host.h
> @@ -766,6 +766,7 @@ enum {
>  #define HF_GIF_MASK          (1 << 0)
>  #define HF_HIF_MASK          (1 << 1)
>  #define HF_VINTR_MASK                (1 << 2)
> +#define HF_NMI_MASK          (1 << 3)
>  
>  /*
>   * Hardware virtualization extension instructions may fault if a
> diff --git a/arch/x86/kvm/svm.c b/arch/x86/kvm/svm.c
> index c605477..0a2b3f1 100644
> --- a/arch/x86/kvm/svm.c
> +++ b/arch/x86/kvm/svm.c
> @@ -1834,6 +1834,13 @@ static int cpuid_interception(struct vcpu_svm *svm, 
> struct kvm_run *kvm_run)
>       return 1;
>  }
>  
> +static int iret_interception(struct vcpu_svm *svm, struct kvm_run *kvm_run)
> +{
> +     svm->vmcb->control.intercept &= ~(1UL << INTERCEPT_IRET);
> +     svm->vcpu.arch.hflags &= ~HF_NMI_MASK;

Two minor issues:

++vcpu->stat.nmi_window_exits;

> +     return 1;
> +}
> +
>  static int invlpg_interception(struct vcpu_svm *svm, struct kvm_run *kvm_run)
>  {
>       if (emulate_instruction(&svm->vcpu, kvm_run, 0, 0, 0) != EMULATE_DONE)
> @@ -2111,6 +2118,7 @@ static int (*svm_exit_handlers[])(struct vcpu_svm *svm,
>       [SVM_EXIT_VINTR]                        = interrupt_window_interception,
>       /* [SVM_EXIT_CR0_SEL_WRITE]             = emulate_on_interception, */
>       [SVM_EXIT_CPUID]                        = cpuid_interception,
> +     [SVM_EXIT_IRET]                         = iret_interception,
>       [SVM_EXIT_INVD]                         = emulate_on_interception,
>       [SVM_EXIT_HLT]                          = halt_interception,
>       [SVM_EXIT_INVLPG]                       = invlpg_interception,
> @@ -2218,6 +2226,12 @@ static void pre_svm_run(struct vcpu_svm *svm)
>               new_asid(svm, svm_data);
>  }
>  
> +static void svm_inject_nmi(struct vcpu_svm *svm)
> +{
> +     svm->vmcb->control.event_inj = SVM_EVTINJ_VALID | SVM_EVTINJ_TYPE_NMI;
> +     svm->vcpu.arch.hflags |= HF_NMI_MASK;
> +     svm->vmcb->control.intercept |= (1UL << INTERCEPT_IRET);

and:

++svm->vcpu.stat.nmi_injections;

> +}
>  
>  static inline void svm_inject_irq(struct vcpu_svm *svm, int irq)
>  {
> @@ -2269,6 +2283,14 @@ static void update_cr8_intercept(struct kvm_vcpu *vcpu)
>               vmcb->control.intercept_cr_write |= INTERCEPT_CR8_MASK;
>  }
>  
> +static int svm_nmi_allowed(struct kvm_vcpu *vcpu)
> +{
> +     struct vcpu_svm *svm = to_svm(vcpu);
> +     struct vmcb *vmcb = svm->vmcb;
> +     return !(vmcb->control.int_state & SVM_INTERRUPT_SHADOW_MASK) &&
> +             !(svm->vcpu.arch.hflags & HF_NMI_MASK);
> +}
> +
>  static int svm_interrupt_allowed(struct kvm_vcpu *vcpu)
>  {
>       struct vcpu_svm *svm = to_svm(vcpu);
> @@ -2284,16 +2306,35 @@ static void enable_irq_window(struct kvm_vcpu *vcpu)
>       svm_inject_irq(to_svm(vcpu), 0x0);
>  }
>  
> +static void enable_nmi_window(struct kvm_vcpu *vcpu)
> +{
> +     struct vcpu_svm *svm = to_svm(vcpu);
> +
> +     if (svm->vmcb->control.int_state & SVM_INTERRUPT_SHADOW_MASK)
> +             enable_irq_window(vcpu);
> +}
> +
>  static void svm_intr_inject(struct kvm_vcpu *vcpu)
>  {
>       /* try to reinject previous events if any */
> +     if (vcpu->arch.nmi_injected) {
> +             svm_inject_nmi(to_svm(vcpu));
> +             return;
> +     }
> +
>       if (vcpu->arch.interrupt.pending) {
>               svm_queue_irq(to_svm(vcpu), vcpu->arch.interrupt.nr);
>               return;
>       }
>  
>       /* try to inject new event if pending */
> -     if (kvm_cpu_has_interrupt(vcpu)) {
> +     if (vcpu->arch.nmi_pending) {
> +             if (svm_nmi_allowed(vcpu)) {
> +                     vcpu->arch.nmi_pending = false;
> +                     vcpu->arch.nmi_injected = true;
> +                     svm_inject_nmi(vcpu);
> +             }
> +     } else if (kvm_cpu_has_interrupt(vcpu)) {
>               if (svm_interrupt_allowed(vcpu)) {
>                       kvm_queue_interrupt(vcpu, kvm_cpu_get_interrupt(vcpu));
>                       svm_queue_irq(to_svm(vcpu), vcpu->arch.interrupt.nr);
> @@ -2312,7 +2353,10 @@ static void svm_intr_assist(struct kvm_vcpu *vcpu, 
> struct kvm_run *kvm_run)
>  
>       svm_intr_inject(vcpu);
>  
> -     if (kvm_cpu_has_interrupt(vcpu) || req_int_win)
> +     /* enable NMI/IRQ window open exits if needed */
> +     if (vcpu->arch.nmi_pending)
> +             enable_nmi_window(vcpu);
> +     else if (kvm_cpu_has_interrupt(vcpu) || req_int_win)
>               enable_irq_window(vcpu);
>  
>  out:
> --
>                       Gleb.
> --
> To unsubscribe from this list: send the line "unsubscribe kvm" in
> the body of a message to [email protected]
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> 

Jan

-- 
Siemens AG, Corporate Technology, CT SE 2
Corporate Competence Center Embedded Linux
--
To unsubscribe from this list: send the line "unsubscribe kvm" in
the body of a message to [email protected]
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Reply via email to