[RFC PATCH 4/4] powerpc kvm_asm: rename PPC_LD and PPC_STD macros to avoid redefinition
Balamuruhan S
bala24 at linux.ibm.com
Fri Mar 20 19:18:37 AEDT 2020
use PPC_KVM_LD and PPC_KVM_STD to fix gcc warnings on redefinition as
we consolidate all ppc instruction encoding in ppc-opcode.h
Signed-off-by: Balamuruhan S <bala24 at linux.ibm.com>
---
arch/powerpc/include/asm/kvm_asm.h | 8 ++++----
arch/powerpc/kvm/booke_interrupts.S | 8 ++++----
arch/powerpc/kvm/bookehv_interrupts.S | 28 +++++++++++++--------------
3 files changed, 22 insertions(+), 22 deletions(-)
diff --git a/arch/powerpc/include/asm/kvm_asm.h b/arch/powerpc/include/asm/kvm_asm.h
index 635fb154b33f..d3f607c57856 100644
--- a/arch/powerpc/include/asm/kvm_asm.h
+++ b/arch/powerpc/include/asm/kvm_asm.h
@@ -11,11 +11,11 @@
#ifdef __ASSEMBLY__
#ifdef CONFIG_64BIT
-#define PPC_STD(sreg, offset, areg) std sreg, (offset)(areg)
-#define PPC_LD(treg, offset, areg) ld treg, (offset)(areg)
+#define PPC_KVM_STD(sreg, offset, areg) (std sreg, (offset)(areg))
+#define PPC_KVM_LD(treg, offset, areg) (ld treg, (offset)(areg))
#else
-#define PPC_STD(sreg, offset, areg) stw sreg, (offset+4)(areg)
-#define PPC_LD(treg, offset, areg) lwz treg, (offset+4)(areg)
+#define PPC_KVM_STD(sreg, offset, areg) (stw sreg, ((offset) + 4)(areg))
+#define PPC_KVM_LD(treg, offset, areg) (lwz treg, ((offset) + 4)(areg))
#endif
#endif
diff --git a/arch/powerpc/kvm/booke_interrupts.S b/arch/powerpc/kvm/booke_interrupts.S
index 2e56ab5a5f55..3a343da95ea5 100644
--- a/arch/powerpc/kvm/booke_interrupts.S
+++ b/arch/powerpc/kvm/booke_interrupts.S
@@ -443,13 +443,13 @@ lightweight_exit:
* written directly to the shared area, so we
* need to reload them here with the guest's values.
*/
- PPC_LD(r3, VCPU_SHARED_SPRG4, r5)
+ PPC_KVM_LD(r3, VCPU_SHARED_SPRG4, r5)
mtspr SPRN_SPRG4W, r3
- PPC_LD(r3, VCPU_SHARED_SPRG5, r5)
+ PPC_KVM_LD(r3, VCPU_SHARED_SPRG5, r5)
mtspr SPRN_SPRG5W, r3
- PPC_LD(r3, VCPU_SHARED_SPRG6, r5)
+ PPC_KVM_LD(r3, VCPU_SHARED_SPRG6, r5)
mtspr SPRN_SPRG6W, r3
- PPC_LD(r3, VCPU_SHARED_SPRG7, r5)
+ PPC_KVM_LD(r3, VCPU_SHARED_SPRG7, r5)
mtspr SPRN_SPRG7W, r3
#ifdef CONFIG_KVM_EXIT_TIMING
diff --git a/arch/powerpc/kvm/bookehv_interrupts.S b/arch/powerpc/kvm/bookehv_interrupts.S
index c577ba4b3169..97e9b3289c7b 100644
--- a/arch/powerpc/kvm/bookehv_interrupts.S
+++ b/arch/powerpc/kvm/bookehv_interrupts.S
@@ -89,7 +89,7 @@ END_BTB_FLUSH_SECTION
#endif
oris r8, r6, MSR_CE at h
- PPC_STD(r6, VCPU_SHARED_MSR, r11)
+ PPC_KVM_STD(r6, VCPU_SHARED_MSR, r11)
ori r8, r8, MSR_ME | MSR_RI
PPC_STL r5, VCPU_PC(r4)
@@ -386,17 +386,17 @@ _GLOBAL(kvmppc_resume_host)
PPC_LL r3, PACA_SPRG_VDSO(r13)
#endif
mfspr r5, SPRN_SPRG9
- PPC_STD(r6, VCPU_SHARED_SPRG4, r11)
+ PPC_KVM_STD(r6, VCPU_SHARED_SPRG4, r11)
mfspr r8, SPRN_SPRG6
- PPC_STD(r7, VCPU_SHARED_SPRG5, r11)
+ PPC_KVM_STD(r7, VCPU_SHARED_SPRG5, r11)
mfspr r9, SPRN_SPRG7
#ifdef CONFIG_64BIT
mtspr SPRN_SPRG_VDSO_WRITE, r3
#endif
- PPC_STD(r5, VCPU_SPRG9, r4)
- PPC_STD(r8, VCPU_SHARED_SPRG6, r11)
+ PPC_KVM_STD(r5, VCPU_SPRG9, r4)
+ PPC_KVM_STD(r8, VCPU_SHARED_SPRG6, r11)
mfxer r3
- PPC_STD(r9, VCPU_SHARED_SPRG7, r11)
+ PPC_KVM_STD(r9, VCPU_SHARED_SPRG7, r11)
/* save guest MAS registers and restore host mas4 & mas6 */
mfspr r5, SPRN_MAS0
@@ -405,7 +405,7 @@ _GLOBAL(kvmppc_resume_host)
stw r5, VCPU_SHARED_MAS0(r11)
mfspr r7, SPRN_MAS2
stw r6, VCPU_SHARED_MAS1(r11)
- PPC_STD(r7, VCPU_SHARED_MAS2, r11)
+ PPC_KVM_STD(r7, VCPU_SHARED_MAS2, r11)
mfspr r5, SPRN_MAS3
mfspr r6, SPRN_MAS4
stw r5, VCPU_SHARED_MAS7_3+4(r11)
@@ -602,7 +602,7 @@ lightweight_exit:
stw r3, VCPU_HOST_MAS6(r4)
lwz r3, VCPU_SHARED_MAS0(r11)
lwz r5, VCPU_SHARED_MAS1(r11)
- PPC_LD(r6, VCPU_SHARED_MAS2, r11)
+ PPC_KVM_LD(r6, VCPU_SHARED_MAS2, r11)
lwz r7, VCPU_SHARED_MAS7_3+4(r11)
lwz r8, VCPU_SHARED_MAS4(r11)
mtspr SPRN_MAS0, r3
@@ -620,15 +620,15 @@ lightweight_exit:
* SPRGs, so we need to reload them here with the guest's values.
*/
lwz r3, VCPU_VRSAVE(r4)
- PPC_LD(r5, VCPU_SHARED_SPRG4, r11)
+ PPC_KVM_LD(r5, VCPU_SHARED_SPRG4, r11)
mtspr SPRN_VRSAVE, r3
- PPC_LD(r6, VCPU_SHARED_SPRG5, r11)
+ PPC_KVM_LD(r6, VCPU_SHARED_SPRG5, r11)
mtspr SPRN_SPRG4W, r5
- PPC_LD(r7, VCPU_SHARED_SPRG6, r11)
+ PPC_KVM_LD(r7, VCPU_SHARED_SPRG6, r11)
mtspr SPRN_SPRG5W, r6
- PPC_LD(r8, VCPU_SHARED_SPRG7, r11)
+ PPC_KVM_LD(r8, VCPU_SHARED_SPRG7, r11)
mtspr SPRN_SPRG6W, r7
- PPC_LD(r5, VCPU_SPRG9, r4)
+ PPC_KVM_LD(r5, VCPU_SPRG9, r4)
mtspr SPRN_SPRG7W, r8
mtspr SPRN_SPRG9, r5
@@ -638,7 +638,7 @@ lightweight_exit:
PPC_LL r6, VCPU_CTR(r4)
PPC_LL r7, VCPU_CR(r4)
PPC_LL r8, VCPU_PC(r4)
- PPC_LD(r9, VCPU_SHARED_MSR, r11)
+ PPC_KVM_LD(r9, VCPU_SHARED_MSR, r11)
PPC_LL r0, VCPU_GPR(R0)(r4)
PPC_LL r1, VCPU_GPR(R1)(r4)
PPC_LL r2, VCPU_GPR(R2)(r4)
--
2.24.1
More information about the Linuxppc-dev
mailing list