[RFC PATCH 4/8] powerpc/64s: put io_sync bit into r14
Nicholas Piggin
npiggin at gmail.com
Thu Dec 21 01:52:02 AEDT 2017
This simplifies spin unlock code and mmio primitives. This
may not be the best use of an r14 bit, but it was a simple
first proof of concept after the per-cpu data_offset, and
so it can stay until we get low on bits.
---
arch/powerpc/include/asm/io.h | 11 ++++------
arch/powerpc/include/asm/paca.h | 44 ++++++++++++++++++++++++++++++++++++-
arch/powerpc/include/asm/spinlock.h | 21 ++++++++++--------
arch/powerpc/xmon/xmon.c | 1 -
4 files changed, 59 insertions(+), 18 deletions(-)
diff --git a/arch/powerpc/include/asm/io.h b/arch/powerpc/include/asm/io.h
index 422f99cf9924..c817f3a83fcc 100644
--- a/arch/powerpc/include/asm/io.h
+++ b/arch/powerpc/include/asm/io.h
@@ -104,8 +104,8 @@ extern bool isa_io_special;
*
*/
-#ifdef CONFIG_PPC64
-#define IO_SET_SYNC_FLAG() do { local_paca->io_sync = 1; } while(0)
+#if defined(CONFIG_PPC64) && defined(CONFIG_SMP)
+#define IO_SET_SYNC_FLAG() do { r14_set_bits(R14_BIT_IO_SYNC); } while(0)
#else
#define IO_SET_SYNC_FLAG()
#endif
@@ -673,11 +673,8 @@ static inline void name at \
*/
static inline void mmiowb(void)
{
- unsigned long tmp;
-
- __asm__ __volatile__("sync; li %0,0; stb %0,%1(13)"
- : "=&r" (tmp) : "i" (offsetof(struct paca_struct, io_sync))
- : "memory");
+ __asm__ __volatile__("sync" : : : "memory");
+ r14_clear_bits(R14_BIT_IO_SYNC);
}
#endif /* !CONFIG_PPC32 */
diff --git a/arch/powerpc/include/asm/paca.h b/arch/powerpc/include/asm/paca.h
index 4dd4ac69e84f..408fa079e00d 100644
--- a/arch/powerpc/include/asm/paca.h
+++ b/arch/powerpc/include/asm/paca.h
@@ -35,12 +35,55 @@
register struct paca_struct *local_paca asm("r13");
#ifdef CONFIG_PPC_BOOK3S
+
+#define R14_BIT_IO_SYNC 0x0001
+
/*
* The top 32-bits of r14 is used as the per-cpu offset, shifted by PAGE_SHIFT.
* The per-cpu could be moved completely to vmalloc space if we had large
* vmalloc page mapping? (no, must access it in real mode).
*/
register u64 local_r14 asm("r14");
+
+/*
+ * r14 should not be modified by C code, because we can not guarantee it
+ * will be done with non-atomic (vs interrupts) read-modify-write sequences.
+ * All updates must be of the form `op r14,r14,xxx` or similar (i.e., atomic
+ * updates).
+ *
+ * Make asm statements have r14 for input and output so that the compiler
+ * does not re-order it with respect to other r14 manipulations.
+ */
+static inline void r14_set_bits(unsigned long mask)
+{
+ if (__builtin_constant_p(mask))
+ asm volatile("ori %0,%0,%2\n"
+ : "=r" (local_r14)
+ : "0" (local_r14), "i" (mask));
+ else
+ asm volatile("or %0,%0,%2\n"
+ : "=r" (local_r14)
+ : "0" (local_r14), "r" (mask));
+}
+
+static inline void r14_flip_bits(unsigned long mask)
+{
+ if (__builtin_constant_p(mask))
+ asm volatile("xori %0,%0,%2\n"
+ : "=r" (local_r14)
+ : "0" (local_r14), "i" (mask));
+ else
+ asm volatile("xor %0,%0,%2\n"
+ : "=r" (local_r14)
+ : "0" (local_r14), "r" (mask));
+}
+
+static inline void r14_clear_bits(unsigned long mask)
+{
+ asm volatile("andc %0,%0,%2\n"
+ : "=r" (local_r14)
+ : "0" (local_r14), "r" (mask));
+}
#endif
#if defined(CONFIG_DEBUG_PREEMPT) && defined(CONFIG_SMP)
@@ -169,7 +212,6 @@ struct paca_struct {
u16 trap_save; /* Used when bad stack is encountered */
u8 irq_soft_mask; /* mask for irq soft masking */
u8 irq_happened; /* irq happened while soft-disabled */
- u8 io_sync; /* writel() needs spin_unlock sync */
u8 irq_work_pending; /* IRQ_WORK interrupt while soft-disable */
u8 nap_state_lost; /* NV GPR values lost in power7_idle */
u64 sprg_vdso; /* Saved user-visible sprg */
diff --git a/arch/powerpc/include/asm/spinlock.h b/arch/powerpc/include/asm/spinlock.h
index b9ebc3085fb7..182bb9304c79 100644
--- a/arch/powerpc/include/asm/spinlock.h
+++ b/arch/powerpc/include/asm/spinlock.h
@@ -40,16 +40,9 @@
#endif
#if defined(CONFIG_PPC64) && defined(CONFIG_SMP)
-#define CLEAR_IO_SYNC (get_paca()->io_sync = 0)
-#define SYNC_IO do { \
- if (unlikely(get_paca()->io_sync)) { \
- mb(); \
- get_paca()->io_sync = 0; \
- } \
- } while (0)
+#define CLEAR_IO_SYNC do { r14_clear_bits(R14_BIT_IO_SYNC); } while(0)
#else
#define CLEAR_IO_SYNC
-#define SYNC_IO
#endif
#ifdef CONFIG_PPC_PSERIES
@@ -165,9 +158,19 @@ void arch_spin_lock_flags(arch_spinlock_t *lock, unsigned long flags)
static inline void arch_spin_unlock(arch_spinlock_t *lock)
{
- SYNC_IO;
+#if defined(CONFIG_PPC64) && defined(CONFIG_SMP)
+ bool io_sync = local_r14 & R14_BIT_IO_SYNC;
+ if (unlikely(io_sync)) {
+ mb();
+ CLEAR_IO_SYNC;
+ } else {
+ __asm__ __volatile__("# arch_spin_unlock\n\t"
+ PPC_RELEASE_BARRIER: : :"memory");
+ }
+#else
__asm__ __volatile__("# arch_spin_unlock\n\t"
PPC_RELEASE_BARRIER: : :"memory");
+#endif
lock->slock = 0;
}
diff --git a/arch/powerpc/xmon/xmon.c b/arch/powerpc/xmon/xmon.c
index a53454f61d09..40f0d02ae92d 100644
--- a/arch/powerpc/xmon/xmon.c
+++ b/arch/powerpc/xmon/xmon.c
@@ -2393,7 +2393,6 @@ static void dump_one_paca(int cpu)
DUMP(p, trap_save, "x");
DUMP(p, irq_soft_mask, "x");
DUMP(p, irq_happened, "x");
- DUMP(p, io_sync, "x");
DUMP(p, irq_work_pending, "x");
DUMP(p, nap_state_lost, "x");
DUMP(p, sprg_vdso, "llx");
--
2.15.0
More information about the Linuxppc-dev
mailing list