This should be no functional difference but makes the caller easier
to read.

Signed-off-by: Nicholas Piggin <npig...@gmail.com>
---
 arch/powerpc/kvm/book3s_hv.c | 65 +++++++++++++++++++++++-------------
 1 file changed, 41 insertions(+), 24 deletions(-)

diff --git a/arch/powerpc/kvm/book3s_hv.c b/arch/powerpc/kvm/book3s_hv.c
index c2c72875fca9..45211458ac05 100644
--- a/arch/powerpc/kvm/book3s_hv.c
+++ b/arch/powerpc/kvm/book3s_hv.c
@@ -4062,6 +4062,44 @@ static void store_spr_state(struct kvm_vcpu *vcpu)
        vcpu->arch.ctrl = mfspr(SPRN_CTRLF);
 }
 
+/* Returns true if current MSR and/or guest MSR may have changed */
+static bool load_vcpu_state(struct kvm_vcpu *vcpu,
+                          struct p9_host_os_sprs *host_os_sprs)
+{
+       bool ret = false;
+
+       if (cpu_has_feature(CPU_FTR_TM) ||
+           cpu_has_feature(CPU_FTR_P9_TM_HV_ASSIST)) {
+               kvmppc_restore_tm_hv(vcpu, vcpu->arch.shregs.msr, true);
+               ret = true;
+       }
+
+       load_spr_state(vcpu, host_os_sprs);
+
+       load_fp_state(&vcpu->arch.fp);
+#ifdef CONFIG_ALTIVEC
+       load_vr_state(&vcpu->arch.vr);
+#endif
+       mtspr(SPRN_VRSAVE, vcpu->arch.vrsave);
+
+       return ret;
+}
+
+static void store_vcpu_state(struct kvm_vcpu *vcpu)
+{
+       store_spr_state(vcpu);
+
+       store_fp_state(&vcpu->arch.fp);
+#ifdef CONFIG_ALTIVEC
+       store_vr_state(&vcpu->arch.vr);
+#endif
+       vcpu->arch.vrsave = mfspr(SPRN_VRSAVE);
+
+       if (cpu_has_feature(CPU_FTR_TM) ||
+           cpu_has_feature(CPU_FTR_P9_TM_HV_ASSIST))
+               kvmppc_save_tm_hv(vcpu, vcpu->arch.shregs.msr, true);
+}
+
 static void save_p9_host_os_sprs(struct p9_host_os_sprs *host_os_sprs)
 {
        if (!cpu_has_feature(CPU_FTR_ARCH_31))
@@ -4169,19 +4207,8 @@ static int kvmhv_p9_guest_entry(struct kvm_vcpu *vcpu, 
u64 time_limit,
 
        vcpu_vpa_increment_dispatch(vcpu);
 
-       if (cpu_has_feature(CPU_FTR_TM) ||
-           cpu_has_feature(CPU_FTR_P9_TM_HV_ASSIST)) {
-               kvmppc_restore_tm_hv(vcpu, vcpu->arch.shregs.msr, true);
-               msr = mfmsr(); /* TM restore can update msr */
-       }
-
-       load_spr_state(vcpu, &host_os_sprs);
-
-       load_fp_state(&vcpu->arch.fp);
-#ifdef CONFIG_ALTIVEC
-       load_vr_state(&vcpu->arch.vr);
-#endif
-       mtspr(SPRN_VRSAVE, vcpu->arch.vrsave);
+       if (unlikely(load_vcpu_state(vcpu, &host_os_sprs)))
+               msr = mfmsr(); /* MSR may have been updated */
 
        switch_pmu_to_guest(vcpu, &host_os_sprs);
 
@@ -4285,17 +4312,7 @@ static int kvmhv_p9_guest_entry(struct kvm_vcpu *vcpu, 
u64 time_limit,
 
        switch_pmu_to_host(vcpu, &host_os_sprs);
 
-       store_spr_state(vcpu);
-
-       store_fp_state(&vcpu->arch.fp);
-#ifdef CONFIG_ALTIVEC
-       store_vr_state(&vcpu->arch.vr);
-#endif
-       vcpu->arch.vrsave = mfspr(SPRN_VRSAVE);
-
-       if (cpu_has_feature(CPU_FTR_TM) ||
-           cpu_has_feature(CPU_FTR_P9_TM_HV_ASSIST))
-               kvmppc_save_tm_hv(vcpu, vcpu->arch.shregs.msr, true);
+       store_vcpu_state(vcpu);
 
        vcpu_vpa_increment_dispatch(vcpu);
 
-- 
2.23.0

Reply via email to