Gleb Natapov wrote:
> The patch fixes two problems with task switching.
> 1. Back link is written to a wrong TSS.
> 2. Instruction emulation is not needed if the reason for task switch
>    is a task gate in IDT and access to it is caused by an external even.
> 
> 2 is currently solved only for VMX since there is not reliable way to
> skip an instruction in SVM. We should emulate it instead.

Does this series fix all issues Bernhard, Thomas and Julian stumbled over?

Jan

> 
> Signed-off-by: Gleb Natapov <[email protected]>
> ---
> 
>  arch/x86/include/asm/svm.h |    1 +
>  arch/x86/kvm/svm.c         |   25 ++++++++++++++++++-------
>  arch/x86/kvm/vmx.c         |   40 +++++++++++++++++++++++++++++-----------
>  arch/x86/kvm/x86.c         |   40 +++++++++++++++++++++++++++++++---------
>  4 files changed, 79 insertions(+), 27 deletions(-)
> 
> diff --git a/arch/x86/include/asm/svm.h b/arch/x86/include/asm/svm.h
> index 82ada75..85574b7 100644
> --- a/arch/x86/include/asm/svm.h
> +++ b/arch/x86/include/asm/svm.h
> @@ -225,6 +225,7 @@ struct __attribute__ ((__packed__)) vmcb {
>  #define SVM_EVTINJ_VALID_ERR (1 << 11)
>  
>  #define SVM_EXITINTINFO_VEC_MASK SVM_EVTINJ_VEC_MASK
> +#define SVM_EXITINTINFO_TYPE_MASK SVM_EVTINJ_TYPE_MASK
>  
>  #define      SVM_EXITINTINFO_TYPE_INTR SVM_EVTINJ_TYPE_INTR
>  #define      SVM_EXITINTINFO_TYPE_NMI SVM_EVTINJ_TYPE_NMI
> diff --git a/arch/x86/kvm/svm.c b/arch/x86/kvm/svm.c
> index 1fcbc17..3ffb695 100644
> --- a/arch/x86/kvm/svm.c
> +++ b/arch/x86/kvm/svm.c
> @@ -1823,17 +1823,28 @@ static int task_switch_interception(struct vcpu_svm 
> *svm,
>                                   struct kvm_run *kvm_run)
>  {
>       u16 tss_selector;
> +     int reason;
> +     int int_type = svm->vmcb->control.exit_int_info &
> +             SVM_EXITINTINFO_TYPE_MASK;
>  
>       tss_selector = (u16)svm->vmcb->control.exit_info_1;
> +
>       if (svm->vmcb->control.exit_info_2 &
>           (1ULL << SVM_EXITINFOSHIFT_TS_REASON_IRET))
> -             return kvm_task_switch(&svm->vcpu, tss_selector,
> -                                    TASK_SWITCH_IRET);
> -     if (svm->vmcb->control.exit_info_2 &
> -         (1ULL << SVM_EXITINFOSHIFT_TS_REASON_JMP))
> -             return kvm_task_switch(&svm->vcpu, tss_selector,
> -                                    TASK_SWITCH_JMP);
> -     return kvm_task_switch(&svm->vcpu, tss_selector, TASK_SWITCH_CALL);
> +             reason = TASK_SWITCH_IRET;
> +     else if (svm->vmcb->control.exit_info_2 &
> +              (1ULL << SVM_EXITINFOSHIFT_TS_REASON_JMP))
> +             reason = TASK_SWITCH_JMP;
> +     else if (svm->vmcb->control.exit_int_info & SVM_EXITINTINFO_VALID)
> +             reason = TASK_SWITCH_GATE;
> +     else
> +             reason = TASK_SWITCH_CALL;
> +
> +
> +     if (reason != TASK_SWITCH_GATE || int_type == SVM_EXITINTINFO_TYPE_SOFT)
> +             skip_emulated_instruction(&svm->vcpu);
> +
> +     return kvm_task_switch(&svm->vcpu, tss_selector, reason);
>  }
>  
>  static int cpuid_interception(struct vcpu_svm *svm, struct kvm_run *kvm_run)
> diff --git a/arch/x86/kvm/vmx.c b/arch/x86/kvm/vmx.c
> index 0da7a9e..01db958 100644
> --- a/arch/x86/kvm/vmx.c
> +++ b/arch/x86/kvm/vmx.c
> @@ -3025,22 +3025,40 @@ static int handle_task_switch(struct kvm_vcpu *vcpu, 
> struct kvm_run *kvm_run)
>       struct vcpu_vmx *vmx = to_vmx(vcpu);
>       unsigned long exit_qualification;
>       u16 tss_selector;
> -     int reason;
> +     int reason, type, idt_v;
> +
> +     idt_v = (vmx->idt_vectoring_info & VECTORING_INFO_VALID_MASK);
> +     type = (vmx->idt_vectoring_info & VECTORING_INFO_TYPE_MASK);
>  
>       exit_qualification = vmcs_readl(EXIT_QUALIFICATION);
>  
>       reason = (u32)exit_qualification >> 30;
> -     if (reason == TASK_SWITCH_GATE && vmx->vcpu.arch.nmi_injected &&
> -         (vmx->idt_vectoring_info & VECTORING_INFO_VALID_MASK) &&
> -         (vmx->idt_vectoring_info & VECTORING_INFO_TYPE_MASK)
> -         == INTR_TYPE_NMI_INTR) {
> -             vcpu->arch.nmi_injected = false;
> -             if (cpu_has_virtual_nmis())
> -                     vmcs_set_bits(GUEST_INTERRUPTIBILITY_INFO,
> -                                   GUEST_INTR_STATE_NMI);
> +     if (reason == TASK_SWITCH_GATE && idt_v) {
> +             switch (type) {
> +             case INTR_TYPE_NMI_INTR:
> +                     vcpu->arch.nmi_injected = false;
> +                     if (cpu_has_virtual_nmis())
> +                             vmcs_set_bits(GUEST_INTERRUPTIBILITY_INFO,
> +                                           GUEST_INTR_STATE_NMI);
> +                     break;
> +             case INTR_TYPE_EXT_INTR:
> +                     kvm_clear_interrupt_queue(vcpu);
> +                     break;
> +             case INTR_TYPE_HARD_EXCEPTION:
> +             case INTR_TYPE_SOFT_EXCEPTION:
> +                     kvm_clear_exception_queue(vcpu);
> +                     break;
> +             default:
> +                     break;
> +             }
>       }
>       tss_selector = exit_qualification;
>  
> +     if (!idt_v || (type != INTR_TYPE_HARD_EXCEPTION &&
> +                    type != INTR_TYPE_EXT_INTR &&
> +                    type != INTR_TYPE_NMI_INTR))
> +             skip_emulated_instruction(vcpu);
> +
>       if (!kvm_task_switch(vcpu, tss_selector, reason))
>               return 0;
>  
> @@ -3292,8 +3310,8 @@ static void vmx_complete_interrupts(struct vcpu_vmx 
> *vmx)
>  
>       vector = idt_vectoring_info & VECTORING_INFO_VECTOR_MASK;
>       type = idt_vectoring_info & VECTORING_INFO_TYPE_MASK;
> -     
> -     switch(type) {
> +
> +     switch (type) {
>       case INTR_TYPE_NMI_INTR:
>               vmx->vcpu.arch.nmi_injected = true;
>               /*
> diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c
> index ae4918c..573bb3f 100644
> --- a/arch/x86/kvm/x86.c
> +++ b/arch/x86/kvm/x86.c
> @@ -3697,7 +3697,6 @@ static void save_state_to_tss32(struct kvm_vcpu *vcpu,
>       tss->fs = get_segment_selector(vcpu, VCPU_SREG_FS);
>       tss->gs = get_segment_selector(vcpu, VCPU_SREG_GS);
>       tss->ldt_selector = get_segment_selector(vcpu, VCPU_SREG_LDTR);
> -     tss->prev_task_link = get_segment_selector(vcpu, VCPU_SREG_TR);
>  }
>  
>  static int load_state_from_tss32(struct kvm_vcpu *vcpu,
> @@ -3794,8 +3793,8 @@ static int load_state_from_tss16(struct kvm_vcpu *vcpu,
>  }
>  
>  static int kvm_task_switch_16(struct kvm_vcpu *vcpu, u16 tss_selector,
> -                    u32 old_tss_base,
> -                    struct desc_struct *nseg_desc)
> +                           u16 old_tss_sel, u32 old_tss_base,
> +                           struct desc_struct *nseg_desc)
>  {
>       struct tss_segment_16 tss_segment_16;
>       int ret = 0;
> @@ -3814,6 +3813,16 @@ static int kvm_task_switch_16(struct kvm_vcpu *vcpu, 
> u16 tss_selector,
>                          &tss_segment_16, sizeof tss_segment_16))
>               goto out;
>  
> +     if (old_tss_sel != 0xffff) {
> +             tss_segment_16.prev_task_link = old_tss_sel;
> +
> +             if (kvm_write_guest(vcpu->kvm,
> +                                 get_tss_base_addr(vcpu, nseg_desc),
> +                                 &tss_segment_16.prev_task_link,
> +                                 sizeof tss_segment_16.prev_task_link))
> +                     goto out;
> +     }
> +
>       if (load_state_from_tss16(vcpu, &tss_segment_16))
>               goto out;
>  
> @@ -3823,7 +3832,7 @@ out:
>  }
>  
>  static int kvm_task_switch_32(struct kvm_vcpu *vcpu, u16 tss_selector,
> -                    u32 old_tss_base,
> +                    u16 old_tss_sel, u32 old_tss_base,
>                      struct desc_struct *nseg_desc)
>  {
>       struct tss_segment_32 tss_segment_32;
> @@ -3843,6 +3852,16 @@ static int kvm_task_switch_32(struct kvm_vcpu *vcpu, 
> u16 tss_selector,
>                          &tss_segment_32, sizeof tss_segment_32))
>               goto out;
>  
> +     if (old_tss_sel != 0xffff) {
> +             tss_segment_32.prev_task_link = old_tss_sel;
> +
> +             if (kvm_write_guest(vcpu->kvm,
> +                                 get_tss_base_addr(vcpu, nseg_desc),
> +                                 &tss_segment_32.prev_task_link,
> +                                 sizeof tss_segment_32.prev_task_link))
> +                     goto out;
> +     }
> +
>       if (load_state_from_tss32(vcpu, &tss_segment_32))
>               goto out;
>  
> @@ -3896,14 +3915,17 @@ int kvm_task_switch(struct kvm_vcpu *vcpu, u16 
> tss_selector, int reason)
>               kvm_x86_ops->set_rflags(vcpu, eflags & ~X86_EFLAGS_NT);
>       }
>  
> -     kvm_x86_ops->skip_emulated_instruction(vcpu);
> +     /* set back link to prev task only if NT bit is set in eflags
> +        note that old_tss_sel is not used afetr this point */
> +     if (reason != TASK_SWITCH_CALL && reason != TASK_SWITCH_GATE)
> +             old_tss_sel = 0xffff;
>  
>       if (nseg_desc.type & 8)
> -             ret = kvm_task_switch_32(vcpu, tss_selector, old_tss_base,
> -                                      &nseg_desc);
> +             ret = kvm_task_switch_32(vcpu, tss_selector, old_tss_sel,
> +                                      old_tss_base, &nseg_desc);
>       else
> -             ret = kvm_task_switch_16(vcpu, tss_selector, old_tss_base,
> -                                      &nseg_desc);
> +             ret = kvm_task_switch_16(vcpu, tss_selector, old_tss_sel,
> +                                      old_tss_base, &nseg_desc);
>  
>       if (reason == TASK_SWITCH_CALL || reason == TASK_SWITCH_GATE) {
>               u32 eflags = kvm_x86_ops->get_rflags(vcpu);
> 
> --
> To unsubscribe from this list: send the line "unsubscribe kvm" in
> the body of a message to [email protected]
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> 


Attachment: signature.asc
Description: OpenPGP digital signature

Reply via email to