vgic v2 interface functions are passed vcpu, when the state
that they need is the vgic distributor, as well as the
kvm_cpu_context and the recently created vcpu_hyp_state. Reduce
the scope of its interface functions to these structs.

Pass the vgic distributor to fixup_guest_exit so that it's not
dependent on struct kvm for the vgic state. NOTE: this change to
fixup_guest_exit is temporary, and will be tidied up in in a
subsequent patch in this series.

Signed-off-by: Fuad Tabba <[email protected]>
---
 arch/arm64/include/asm/kvm_hyp.h         |  2 +-
 arch/arm64/kvm/hyp/include/hyp/switch.h  |  4 ++--
 arch/arm64/kvm/hyp/nvhe/switch.c         |  4 +++-
 arch/arm64/kvm/hyp/vgic-v2-cpuif-proxy.c | 16 ++++++----------
 arch/arm64/kvm/hyp/vhe/switch.c          |  3 ++-
 5 files changed, 14 insertions(+), 15 deletions(-)

diff --git a/arch/arm64/include/asm/kvm_hyp.h b/arch/arm64/include/asm/kvm_hyp.h
index 2737e05a16b2..d9a8872a7efb 100644
--- a/arch/arm64/include/asm/kvm_hyp.h
+++ b/arch/arm64/include/asm/kvm_hyp.h
@@ -55,7 +55,7 @@ DECLARE_PER_CPU(struct kvm_nvhe_init_params, kvm_init_params);
  */
 #define __kvm_swab32(x)        ___constant_swab32(x)
 
-int __vgic_v2_perform_cpuif_access(struct kvm_vcpu *vcpu);
+int __vgic_v2_perform_cpuif_access(struct vgic_dist *vgic, struct 
kvm_cpu_context *ctxt, struct vcpu_hyp_state *hyps);
 
 void __vgic_v3_save_state(struct vgic_v3_cpu_if *cpu_if);
 void __vgic_v3_restore_state(struct vgic_v3_cpu_if *cpu_if);
diff --git a/arch/arm64/kvm/hyp/include/hyp/switch.h 
b/arch/arm64/kvm/hyp/include/hyp/switch.h
index 075719c07009..30fcfe84f609 100644
--- a/arch/arm64/kvm/hyp/include/hyp/switch.h
+++ b/arch/arm64/kvm/hyp/include/hyp/switch.h
@@ -424,7 +424,7 @@ static inline bool __hyp_handle_ptrauth(struct kvm_vcpu 
*vcpu)
  * the guest, false when we should restore the host state and return to the
  * main run loop.
  */
-static inline bool fixup_guest_exit(struct kvm_vcpu *vcpu, u64 *exit_code)
+static inline bool fixup_guest_exit(struct kvm_vcpu *vcpu, struct vgic_dist 
*vgic, u64 *exit_code)
 {
        struct vcpu_hyp_state *vcpu_hyps = &hyp_state(vcpu);
        struct kvm_cpu_context *vcpu_ctxt = &vcpu_ctxt(vcpu);
@@ -486,7 +486,7 @@ static inline bool fixup_guest_exit(struct kvm_vcpu *vcpu, 
u64 *exit_code)
                        !kvm_vcpu_abt_iss1tw(vcpu);
 
                if (valid) {
-                       int ret = __vgic_v2_perform_cpuif_access(vcpu);
+                       int ret = __vgic_v2_perform_cpuif_access(vgic, 
vcpu_ctxt, vcpu_hyps);
 
                        if (ret == 1)
                                goto guest;
diff --git a/arch/arm64/kvm/hyp/nvhe/switch.c b/arch/arm64/kvm/hyp/nvhe/switch.c
index eadbf2ccaf68..164b0f899f7b 100644
--- a/arch/arm64/kvm/hyp/nvhe/switch.c
+++ b/arch/arm64/kvm/hyp/nvhe/switch.c
@@ -172,6 +172,8 @@ int __kvm_vcpu_run(struct kvm_vcpu *vcpu)
 {
        struct vcpu_hyp_state *vcpu_hyps = &hyp_state(vcpu);
        struct kvm_cpu_context *vcpu_ctxt = &vcpu_ctxt(vcpu);
+       struct kvm *kvm = kern_hyp_va(vcpu->kvm);
+       struct vgic_dist *vgic = &kvm->arch.vgic;
        struct kvm_cpu_context *host_ctxt;
        struct kvm_cpu_context *guest_ctxt;
        bool pmu_switch_needed;
@@ -230,7 +232,7 @@ int __kvm_vcpu_run(struct kvm_vcpu *vcpu)
                exit_code = __guest_enter(vcpu);
 
                /* And we're baaack! */
-       } while (fixup_guest_exit(vcpu, &exit_code));
+       } while (fixup_guest_exit(vcpu, vgic, &exit_code));
 
        __sysreg_save_state_nvhe(guest_ctxt);
        __sysreg32_save_state(vcpu);
diff --git a/arch/arm64/kvm/hyp/vgic-v2-cpuif-proxy.c 
b/arch/arm64/kvm/hyp/vgic-v2-cpuif-proxy.c
index acd0d21394e3..787f973af43a 100644
--- a/arch/arm64/kvm/hyp/vgic-v2-cpuif-proxy.c
+++ b/arch/arm64/kvm/hyp/vgic-v2-cpuif-proxy.c
@@ -34,19 +34,15 @@ static bool __is_be(struct kvm_cpu_context *vcpu_ctxt)
  *  0: Not a GICV access
  * -1: Illegal GICV access successfully performed
  */
-int __vgic_v2_perform_cpuif_access(struct kvm_vcpu *vcpu)
+int __vgic_v2_perform_cpuif_access(struct vgic_dist *vgic, struct 
kvm_cpu_context *vcpu_ctxt, struct vcpu_hyp_state *vcpu_hyps)
 {
-       struct vcpu_hyp_state *vcpu_hyps = &hyp_state(vcpu);
-       struct kvm_cpu_context *vcpu_ctxt = &vcpu_ctxt(vcpu);
-       struct kvm *kvm = kern_hyp_va(vcpu->kvm);
-       struct vgic_dist *vgic = &kvm->arch.vgic;
        phys_addr_t fault_ipa;
        void __iomem *addr;
        int rd;
 
        /* Build the full address */
-       fault_ipa  = kvm_vcpu_get_fault_ipa(vcpu);
-       fault_ipa |= kvm_vcpu_get_hfar(vcpu) & GENMASK(11, 0);
+       fault_ipa  = kvm_hyp_state_get_fault_ipa(vcpu_hyps);
+       fault_ipa |= kvm_hyp_state_get_hfar(vcpu_hyps) & GENMASK(11, 0);
 
        /* If not for GICV, move on */
        if (fault_ipa <  vgic->vgic_cpu_base ||
@@ -54,7 +50,7 @@ int __vgic_v2_perform_cpuif_access(struct kvm_vcpu *vcpu)
                return 0;
 
        /* Reject anything but a 32bit access */
-       if (kvm_vcpu_dabt_get_as(vcpu) != sizeof(u32)) {
+       if (kvm_hyp_state_dabt_get_as(vcpu_hyps) != sizeof(u32)) {
                __kvm_skip_instr(vcpu_ctxt, vcpu_hyps);
                return -1;
        }
@@ -65,11 +61,11 @@ int __vgic_v2_perform_cpuif_access(struct kvm_vcpu *vcpu)
                return -1;
        }
 
-       rd = kvm_vcpu_dabt_get_rd(vcpu);
+       rd = kvm_hyp_state_dabt_get_rd(vcpu_hyps);
        addr  = kvm_vgic_global_state.vcpu_hyp_va;
        addr += fault_ipa - vgic->vgic_cpu_base;
 
-       if (kvm_vcpu_dabt_iswrite(vcpu)) {
+       if (kvm_hyp_state_dabt_iswrite(vcpu_hyps)) {
                u32 data = ctxt_get_reg(vcpu_ctxt, rd);
                if (__is_be(vcpu_ctxt)) {
                        /* guest pre-swabbed data, undo this for writel() */
diff --git a/arch/arm64/kvm/hyp/vhe/switch.c b/arch/arm64/kvm/hyp/vhe/switch.c
index 395274532c20..f315058a50ca 100644
--- a/arch/arm64/kvm/hyp/vhe/switch.c
+++ b/arch/arm64/kvm/hyp/vhe/switch.c
@@ -111,6 +111,7 @@ static int __kvm_vcpu_run_vhe(struct kvm_vcpu *vcpu)
 {
        struct vcpu_hyp_state *vcpu_hyps = &hyp_state(vcpu);
        struct kvm_cpu_context *vcpu_ctxt = &vcpu_ctxt(vcpu);
+       struct vgic_dist *vgic = &vcpu->kvm->arch.vgic;
        struct kvm_cpu_context *host_ctxt;
        struct kvm_cpu_context *guest_ctxt;
        u64 exit_code;
@@ -145,7 +146,7 @@ static int __kvm_vcpu_run_vhe(struct kvm_vcpu *vcpu)
                exit_code = __guest_enter(vcpu);
 
                /* And we're baaack! */
-       } while (fixup_guest_exit(vcpu, &exit_code));
+       } while (fixup_guest_exit(vcpu, vgic, &exit_code));
 
        sysreg_save_guest_state_vhe(guest_ctxt);
 
-- 
2.33.0.685.g46640cef36-goog

_______________________________________________
kvmarm mailing list
[email protected]
https://lists.cs.columbia.edu/mailman/listinfo/kvmarm

Reply via email to