Hi Sean,

Thank you for reviewing my patches.

On 4/23/2025 9:14 PM, Sean Christopherson wrote:
> On Mon, Mar 24, 2025, Manali Shukla wrote:
>> +    if (vmcb02->save.rip && (svm->nested.ctl.bus_lock_rip == 
>> vmcb02->save.rip)) {
>> +            vmcb02->control.bus_lock_counter = 1;
>> +            svm->bus_lock_rip = svm->nested.ctl.bus_lock_rip;
>> +    } else {
>> +            vmcb02->control.bus_lock_counter = 0;
>> +    }
>> +    svm->nested.ctl.bus_lock_rip = INVALID_GPA;
>> +
>>      /* Done at vmrun: asid.  */
>>  
>>      /* Also overwritten later if necessary.  */
>> @@ -1039,6 +1069,18 @@ int nested_svm_vmexit(struct vcpu_svm *svm)
>>  
>>      }
>>  
>> +    /*
>> +     * If bus_lock_counter is nonzero and the guest has not moved past the
>> +     * guilty instruction, save bus_lock_rip in svm_nested_state. This will
>> +     * help determine at nested VMRUN whether to stash vmcb02's counter or
>> +     * reset it to '0'.
>> +     */
>> +    if (vmcb02->control.bus_lock_counter &&
>> +        svm->bus_lock_rip == vmcb02->save.rip)
>> +            svm->nested.ctl.bus_lock_rip = svm->bus_lock_rip;
>> +    else
>> +            svm->nested.ctl.bus_lock_rip = INVALID_GPA;
>> +
>>      nested_svm_copy_common_state(svm->nested.vmcb02.ptr, svm->vmcb01.ptr);
>>  
>>      svm_switch_vmcb(svm, &svm->vmcb01);
> 
> ...
> 
>> +static int bus_lock_exit(struct kvm_vcpu *vcpu)
>> +{
>> +    struct vcpu_svm *svm = to_svm(vcpu);
>> +
>> +    vcpu->run->exit_reason = KVM_EXIT_X86_BUS_LOCK;
>> +    vcpu->run->flags |= KVM_RUN_X86_BUS_LOCK;
>> +
>> +    vcpu->arch.cui_linear_rip = kvm_get_linear_rip(vcpu);
>> +    svm->bus_lock_rip = vcpu->arch.cui_linear_rip;
>> +    vcpu->arch.complete_userspace_io = complete_userspace_buslock;
>> +
>> +    return 0;
>> +}
> 
>> @@ -327,6 +328,7 @@ struct vcpu_svm {
>>  
>>      /* Guest GIF value, used when vGIF is not enabled */
>>      bool guest_gif;
>> +    u64 bus_lock_rip;
> 
> I don't think this field is necessary.  Rather than unconditionally invalidate
> on nested VMRUN and then conditionally restore on nested #VMEXIT, just leave
> svm->nested.ctl.bus_lock_rip set on VMRUN and conditionally invalidate on 
> #VMEXIT.
> And then in bus_lock_exit(), update the field if the exit occurred while L2 is
> active.
> 
> Completely untested:
> 
> diff --git a/arch/x86/kvm/svm/nested.c b/arch/x86/kvm/svm/nested.c
> index a42ef7dd9143..98e065a93516 100644
> --- a/arch/x86/kvm/svm/nested.c
> +++ b/arch/x86/kvm/svm/nested.c
> @@ -700,13 +700,10 @@ static void nested_vmcb02_prepare_control(struct 
> vcpu_svm *svm,
>          * L1 re-enters L2, the same instruction will trigger a VM-Exit and 
> the
>          * entire cycle start over.
>          */
> -       if (vmcb02->save.rip && (svm->nested.ctl.bus_lock_rip == 
> vmcb02->save.rip)) {
> +       if (vmcb02->save.rip && (svm->nested.ctl.bus_lock_rip == 
> vmcb02->save.rip))
>                 vmcb02->control.bus_lock_counter = 1;
> -               svm->bus_lock_rip = svm->nested.ctl.bus_lock_rip;
> -       } else {
> +       else
>                 vmcb02->control.bus_lock_counter = 0;
> -       }
> -       svm->nested.ctl.bus_lock_rip = INVALID_GPA;
>  
>         /* Done at vmrun: asid.  */
>  
> @@ -1070,15 +1067,10 @@ int nested_svm_vmexit(struct vcpu_svm *svm)
>         }
>  
>         /*
> -        * If bus_lock_counter is nonzero and the guest has not moved past the
> -        * guilty instruction, save bus_lock_rip in svm_nested_state. This 
> will
> -        * help determine at nested VMRUN whether to stash vmcb02's counter or
> -        * reset it to '0'.
> +        * Invalidate bus_lock_rip unless kVM is still waiting for the guest
> +        * to make forward progress before re-enabling bus lock detection.
>          */
> -       if (vmcb02->control.bus_lock_counter &&
> -           svm->bus_lock_rip == vmcb02->save.rip)
> -               svm->nested.ctl.bus_lock_rip = svm->bus_lock_rip;
> -       else
> +       if (!vmcb02->control.bus_lock_counter)
>                 svm->nested.ctl.bus_lock_rip = INVALID_GPA;
>  
>         nested_svm_copy_common_state(svm->nested.vmcb02.ptr, svm->vmcb01.ptr);
> diff --git a/arch/x86/kvm/svm/svm.c b/arch/x86/kvm/svm/svm.c
> index ea12e93ae983..11ce031323fd 100644
> --- a/arch/x86/kvm/svm/svm.c
> +++ b/arch/x86/kvm/svm/svm.c
> @@ -3333,9 +3333,10 @@ static int bus_lock_exit(struct kvm_vcpu *vcpu)
>         vcpu->run->flags |= KVM_RUN_X86_BUS_LOCK;
>  
>         vcpu->arch.cui_linear_rip = kvm_get_linear_rip(vcpu);
> -       svm->bus_lock_rip = vcpu->arch.cui_linear_rip;
>         vcpu->arch.complete_userspace_io = complete_userspace_buslock;
>  
> +       if (is_guest_mode(vcpu))
> +               svm->nested.ctl.bus_lock_rip = vcpu->arch.cui_linear_rip;
>         return 0;
>  }
>  
> diff --git a/arch/x86/kvm/svm/svm.h b/arch/x86/kvm/svm/svm.h
> index 7a4c5848c952..8667faccaedc 100644
> --- a/arch/x86/kvm/svm/svm.h
> +++ b/arch/x86/kvm/svm/svm.h
> @@ -328,7 +328,6 @@ struct vcpu_svm {
>  
>         /* Guest GIF value, used when vGIF is not enabled */
>         bool guest_gif;
> -       u64 bus_lock_rip;
>  };
>  
>  struct svm_cpu_data {
> 

I have added these changes and tested them. Everything looks good to me. I will 
include them in V5
and send it soon.

-Manali

Reply via email to