use PPC_KVM_LD and PPC_KVM_STD to fix gcc warnings on redefinition as we consolidate all ppc instruction encoding in ppc-opcode.h
Signed-off-by: Balamuruhan S <bal...@linux.ibm.com> --- arch/powerpc/include/asm/kvm_asm.h | 8 ++++---- arch/powerpc/kvm/booke_interrupts.S | 8 ++++---- arch/powerpc/kvm/bookehv_interrupts.S | 28 +++++++++++++-------------- 3 files changed, 22 insertions(+), 22 deletions(-) diff --git a/arch/powerpc/include/asm/kvm_asm.h b/arch/powerpc/include/asm/kvm_asm.h index 635fb154b33f..d3f607c57856 100644 --- a/arch/powerpc/include/asm/kvm_asm.h +++ b/arch/powerpc/include/asm/kvm_asm.h @@ -11,11 +11,11 @@ #ifdef __ASSEMBLY__ #ifdef CONFIG_64BIT -#define PPC_STD(sreg, offset, areg) std sreg, (offset)(areg) -#define PPC_LD(treg, offset, areg) ld treg, (offset)(areg) +#define PPC_KVM_STD(sreg, offset, areg) (std sreg, (offset)(areg)) +#define PPC_KVM_LD(treg, offset, areg) (ld treg, (offset)(areg)) #else -#define PPC_STD(sreg, offset, areg) stw sreg, (offset+4)(areg) -#define PPC_LD(treg, offset, areg) lwz treg, (offset+4)(areg) +#define PPC_KVM_STD(sreg, offset, areg) (stw sreg, ((offset) + 4)(areg)) +#define PPC_KVM_LD(treg, offset, areg) (lwz treg, ((offset) + 4)(areg)) #endif #endif diff --git a/arch/powerpc/kvm/booke_interrupts.S b/arch/powerpc/kvm/booke_interrupts.S index 2e56ab5a5f55..3a343da95ea5 100644 --- a/arch/powerpc/kvm/booke_interrupts.S +++ b/arch/powerpc/kvm/booke_interrupts.S @@ -443,13 +443,13 @@ lightweight_exit: * written directly to the shared area, so we * need to reload them here with the guest's values. */ - PPC_LD(r3, VCPU_SHARED_SPRG4, r5) + PPC_KVM_LD(r3, VCPU_SHARED_SPRG4, r5) mtspr SPRN_SPRG4W, r3 - PPC_LD(r3, VCPU_SHARED_SPRG5, r5) + PPC_KVM_LD(r3, VCPU_SHARED_SPRG5, r5) mtspr SPRN_SPRG5W, r3 - PPC_LD(r3, VCPU_SHARED_SPRG6, r5) + PPC_KVM_LD(r3, VCPU_SHARED_SPRG6, r5) mtspr SPRN_SPRG6W, r3 - PPC_LD(r3, VCPU_SHARED_SPRG7, r5) + PPC_KVM_LD(r3, VCPU_SHARED_SPRG7, r5) mtspr SPRN_SPRG7W, r3 #ifdef CONFIG_KVM_EXIT_TIMING diff --git a/arch/powerpc/kvm/bookehv_interrupts.S b/arch/powerpc/kvm/bookehv_interrupts.S index c577ba4b3169..97e9b3289c7b 100644 --- a/arch/powerpc/kvm/bookehv_interrupts.S +++ b/arch/powerpc/kvm/bookehv_interrupts.S @@ -89,7 +89,7 @@ END_BTB_FLUSH_SECTION #endif oris r8, r6, MSR_CE@h - PPC_STD(r6, VCPU_SHARED_MSR, r11) + PPC_KVM_STD(r6, VCPU_SHARED_MSR, r11) ori r8, r8, MSR_ME | MSR_RI PPC_STL r5, VCPU_PC(r4) @@ -386,17 +386,17 @@ _GLOBAL(kvmppc_resume_host) PPC_LL r3, PACA_SPRG_VDSO(r13) #endif mfspr r5, SPRN_SPRG9 - PPC_STD(r6, VCPU_SHARED_SPRG4, r11) + PPC_KVM_STD(r6, VCPU_SHARED_SPRG4, r11) mfspr r8, SPRN_SPRG6 - PPC_STD(r7, VCPU_SHARED_SPRG5, r11) + PPC_KVM_STD(r7, VCPU_SHARED_SPRG5, r11) mfspr r9, SPRN_SPRG7 #ifdef CONFIG_64BIT mtspr SPRN_SPRG_VDSO_WRITE, r3 #endif - PPC_STD(r5, VCPU_SPRG9, r4) - PPC_STD(r8, VCPU_SHARED_SPRG6, r11) + PPC_KVM_STD(r5, VCPU_SPRG9, r4) + PPC_KVM_STD(r8, VCPU_SHARED_SPRG6, r11) mfxer r3 - PPC_STD(r9, VCPU_SHARED_SPRG7, r11) + PPC_KVM_STD(r9, VCPU_SHARED_SPRG7, r11) /* save guest MAS registers and restore host mas4 & mas6 */ mfspr r5, SPRN_MAS0 @@ -405,7 +405,7 @@ _GLOBAL(kvmppc_resume_host) stw r5, VCPU_SHARED_MAS0(r11) mfspr r7, SPRN_MAS2 stw r6, VCPU_SHARED_MAS1(r11) - PPC_STD(r7, VCPU_SHARED_MAS2, r11) + PPC_KVM_STD(r7, VCPU_SHARED_MAS2, r11) mfspr r5, SPRN_MAS3 mfspr r6, SPRN_MAS4 stw r5, VCPU_SHARED_MAS7_3+4(r11) @@ -602,7 +602,7 @@ lightweight_exit: stw r3, VCPU_HOST_MAS6(r4) lwz r3, VCPU_SHARED_MAS0(r11) lwz r5, VCPU_SHARED_MAS1(r11) - PPC_LD(r6, VCPU_SHARED_MAS2, r11) + PPC_KVM_LD(r6, VCPU_SHARED_MAS2, r11) lwz r7, VCPU_SHARED_MAS7_3+4(r11) lwz r8, VCPU_SHARED_MAS4(r11) mtspr SPRN_MAS0, r3 @@ -620,15 +620,15 @@ lightweight_exit: * SPRGs, so we need to reload them here with the guest's values. */ lwz r3, VCPU_VRSAVE(r4) - PPC_LD(r5, VCPU_SHARED_SPRG4, r11) + PPC_KVM_LD(r5, VCPU_SHARED_SPRG4, r11) mtspr SPRN_VRSAVE, r3 - PPC_LD(r6, VCPU_SHARED_SPRG5, r11) + PPC_KVM_LD(r6, VCPU_SHARED_SPRG5, r11) mtspr SPRN_SPRG4W, r5 - PPC_LD(r7, VCPU_SHARED_SPRG6, r11) + PPC_KVM_LD(r7, VCPU_SHARED_SPRG6, r11) mtspr SPRN_SPRG5W, r6 - PPC_LD(r8, VCPU_SHARED_SPRG7, r11) + PPC_KVM_LD(r8, VCPU_SHARED_SPRG7, r11) mtspr SPRN_SPRG6W, r7 - PPC_LD(r5, VCPU_SPRG9, r4) + PPC_KVM_LD(r5, VCPU_SPRG9, r4) mtspr SPRN_SPRG7W, r8 mtspr SPRN_SPRG9, r5 @@ -638,7 +638,7 @@ lightweight_exit: PPC_LL r6, VCPU_CTR(r4) PPC_LL r7, VCPU_CR(r4) PPC_LL r8, VCPU_PC(r4) - PPC_LD(r9, VCPU_SHARED_MSR, r11) + PPC_KVM_LD(r9, VCPU_SHARED_MSR, r11) PPC_LL r0, VCPU_GPR(R0)(r4) PPC_LL r1, VCPU_GPR(R1)(r4) PPC_LL r2, VCPU_GPR(R2)(r4) -- 2.24.1