A single guest with nested virtualization needs two VMIDs: one for
the guest hypervisor (L1) and another for the nested guest (L2).

To support this, divide the VMID space into two equal parts when
nested virtualization is enabled.

Signed-off-by: Anup Patel <[email protected]>
---
 arch/riscv/include/asm/kvm_vmid.h |  1 +
 arch/riscv/kvm/main.c             |  4 ++--
 arch/riscv/kvm/tlb.c              | 11 +++++++++--
 arch/riscv/kvm/vmid.c             | 33 ++++++++++++++++++++++++++++---
 4 files changed, 42 insertions(+), 7 deletions(-)

diff --git a/arch/riscv/include/asm/kvm_vmid.h 
b/arch/riscv/include/asm/kvm_vmid.h
index db61b0525a8d..3048e12a639c 100644
--- a/arch/riscv/include/asm/kvm_vmid.h
+++ b/arch/riscv/include/asm/kvm_vmid.h
@@ -19,6 +19,7 @@ struct kvm_vmid {
 
 void __init kvm_riscv_gstage_vmid_detect(void);
 unsigned long kvm_riscv_gstage_vmid_bits(void);
+unsigned long kvm_riscv_gstage_nested_vmid(unsigned long vmid);
 int kvm_riscv_gstage_vmid_init(struct kvm *kvm);
 bool kvm_riscv_gstage_vmid_ver_changed(struct kvm_vmid *vmid);
 void kvm_riscv_gstage_vmid_update(struct kvm_vcpu *vcpu);
diff --git a/arch/riscv/kvm/main.c b/arch/riscv/kvm/main.c
index 5b4bf972d242..28044eefda47 100644
--- a/arch/riscv/kvm/main.c
+++ b/arch/riscv/kvm/main.c
@@ -123,8 +123,6 @@ static int __init riscv_kvm_init(void)
                return -ENODEV;
        }
 
-       kvm_riscv_gstage_vmid_detect();
-
        rc = kvm_riscv_aia_init();
        if (rc && rc != -ENODEV) {
                kvm_riscv_nacl_exit();
@@ -133,6 +131,8 @@ static int __init riscv_kvm_init(void)
 
        kvm_riscv_nested_init();
 
+       kvm_riscv_gstage_vmid_detect();
+
        kvm_info("hypervisor extension available\n");
 
        if (kvm_riscv_nacl_available()) {
diff --git a/arch/riscv/kvm/tlb.c b/arch/riscv/kvm/tlb.c
index ff1aeac4eb8e..a95aa5336560 100644
--- a/arch/riscv/kvm/tlb.c
+++ b/arch/riscv/kvm/tlb.c
@@ -160,7 +160,7 @@ void kvm_riscv_local_hfence_vvma_all(unsigned long vmid)
 
 void kvm_riscv_local_tlb_sanitize(struct kvm_vcpu *vcpu)
 {
-       unsigned long vmid;
+       unsigned long vmid, nvmid;
 
        if (!kvm_riscv_gstage_vmid_bits() ||
            vcpu->arch.last_exit_cpu == vcpu->cpu)
@@ -180,12 +180,19 @@ void kvm_riscv_local_tlb_sanitize(struct kvm_vcpu *vcpu)
        vmid = READ_ONCE(vcpu->kvm->arch.vmid.vmid);
        kvm_riscv_local_hfence_gvma_vmid_all(vmid);
 
+       nvmid = kvm_riscv_gstage_nested_vmid(vmid);
+       if (vmid != nvmid)
+               kvm_riscv_local_hfence_gvma_vmid_all(nvmid);
+
        /*
         * Flush VS-stage TLB entries for implementation where VS-stage
         * TLB does not cahce guest physical address and VMID.
         */
-       if (static_branch_unlikely(&kvm_riscv_vsstage_tlb_no_gpa))
+       if (static_branch_unlikely(&kvm_riscv_vsstage_tlb_no_gpa)) {
                kvm_riscv_local_hfence_vvma_all(vmid);
+               if (vmid != nvmid)
+                       kvm_riscv_local_hfence_vvma_all(nvmid);
+       }
 }
 
 void kvm_riscv_fence_i_process(struct kvm_vcpu *vcpu)
diff --git a/arch/riscv/kvm/vmid.c b/arch/riscv/kvm/vmid.c
index cf34d448289d..2ddd95fe2d9c 100644
--- a/arch/riscv/kvm/vmid.c
+++ b/arch/riscv/kvm/vmid.c
@@ -25,6 +25,8 @@ static DEFINE_SPINLOCK(vmid_lock);
 
 void __init kvm_riscv_gstage_vmid_detect(void)
 {
+       unsigned long min_vmids;
+
        /* Figure-out number of VMID bits in HW */
        csr_write(CSR_HGATP, (kvm_riscv_gstage_mode << HGATP_MODE_SHIFT) | 
HGATP_VMID);
        vmid_bits = csr_read(CSR_HGATP);
@@ -35,8 +37,23 @@ void __init kvm_riscv_gstage_vmid_detect(void)
        /* We polluted local TLB so flush all guest TLB */
        kvm_riscv_local_hfence_gvma_all();
 
-       /* We don't use VMID bits if they are not sufficient */
-       if ((1UL << vmid_bits) < num_possible_cpus())
+       /*
+        * A single guest with nested virtualization needs two
+        * VMIDs: one for the guest hypervisor (L1) and another
+        * for the nested guest (L2).
+        *
+        * Potentially, we can have a separate guest running on
+        * each host CPU so the number of VMIDs should not be:
+        *
+        * 1. less than the number of host CPUs for
+        *    nested virtualization disabled
+        * 2. less than twice the number of host CPUs for
+        *    nested virtualization enabled
+        */
+       min_vmids = num_possible_cpus();
+       if (kvm_riscv_nested_available())
+               min_vmids = min_vmids * 2;
+       if (BIT(vmid_bits) < min_vmids)
                vmid_bits = 0;
 }
 
@@ -45,6 +62,13 @@ unsigned long kvm_riscv_gstage_vmid_bits(void)
        return vmid_bits;
 }
 
+unsigned long kvm_riscv_gstage_nested_vmid(unsigned long vmid)
+{
+       if (kvm_riscv_nested_available())
+               return vmid | BIT(vmid_bits - 1);
+       return vmid;
+}
+
 int kvm_riscv_gstage_vmid_init(struct kvm *kvm)
 {
        /* Mark the initial VMID and VMID version invalid */
@@ -112,7 +136,10 @@ void kvm_riscv_gstage_vmid_update(struct kvm_vcpu *vcpu)
 
        vmid->vmid = vmid_next;
        vmid_next++;
-       vmid_next &= (1 << vmid_bits) - 1;
+       if (kvm_riscv_nested_available())
+               vmid_next &= BIT(vmid_bits - 1) - 1;
+       else
+               vmid_next &= BIT(vmid_bits) - 1;
 
        WRITE_ONCE(vmid->vmid_version, READ_ONCE(vmid_version));
 
-- 
2.43.0


Reply via email to