powerpc: Implement {cmp}xchg for u8 and u16
authorPan Xinhui <xinhui.pan@linux.vnet.ibm.com>
Wed, 27 Apr 2016 09:16:45 +0000 (17:16 +0800)
committerMichael Ellerman <mpe@ellerman.id.au>
Wed, 23 Nov 2016 11:56:26 +0000 (22:56 +1100)
Implement xchg{u8,u16}{local,relaxed}, and
cmpxchg{u8,u16}{,local,acquire,relaxed}.

It works on all ppc.

remove volatile of first parameter in __cmpxchg_local and __cmpxchg

Suggested-by: Peter Zijlstra (Intel) <peterz@infradead.org>
Signed-off-by: Pan Xinhui <xinhui.pan@linux.vnet.ibm.com>
Acked-by: Boqun Feng <boqun.feng@gmail.com>
Acked-by: Peter Zijlstra (Intel) <peterz@infradead.org>
Signed-off-by: Michael Ellerman <mpe@ellerman.id.au>
arch/powerpc/include/asm/cmpxchg.h

index 44efe739b6b9b4979f95b3a26d8abf52dcda3c26..c12f110261b218df24bf91c389927c5fd0778adf 100644 (file)
@@ -7,6 +7,71 @@
 #include <asm/asm-compat.h>
 #include <linux/bug.h>
 
+#ifdef __BIG_ENDIAN
+#define BITOFF_CAL(size, off)  ((sizeof(u32) - size - off) * BITS_PER_BYTE)
+#else
+#define BITOFF_CAL(size, off)  (off * BITS_PER_BYTE)
+#endif
+
+#define XCHG_GEN(type, sfx, cl)                                \
+static inline u32 __xchg_##type##sfx(void *p, u32 val)         \
+{                                                              \
+       unsigned int prev, prev_mask, tmp, bitoff, off;         \
+                                                               \
+       off = (unsigned long)p % sizeof(u32);                   \
+       bitoff = BITOFF_CAL(sizeof(type), off);                 \
+       p -= off;                                               \
+       val <<= bitoff;                                         \
+       prev_mask = (u32)(type)-1 << bitoff;                    \
+                                                               \
+       __asm__ __volatile__(                                   \
+"1:    lwarx   %0,0,%3\n"                                      \
+"      andc    %1,%0,%5\n"                                     \
+"      or      %1,%1,%4\n"                                     \
+       PPC405_ERR77(0,%3)                                      \
+"      stwcx.  %1,0,%3\n"                                      \
+"      bne-    1b\n"                                           \
+       : "=&r" (prev), "=&r" (tmp), "+m" (*(u32*)p)            \
+       : "r" (p), "r" (val), "r" (prev_mask)                   \
+       : "cc", cl);                                            \
+                                                               \
+       return prev >> bitoff;                                  \
+}
+
+#define CMPXCHG_GEN(type, sfx, br, br2, cl)                    \
+static inline                                                  \
+u32 __cmpxchg_##type##sfx(void *p, u32 old, u32 new)           \
+{                                                              \
+       unsigned int prev, prev_mask, tmp, bitoff, off;         \
+                                                               \
+       off = (unsigned long)p % sizeof(u32);                   \
+       bitoff = BITOFF_CAL(sizeof(type), off);                 \
+       p -= off;                                               \
+       old <<= bitoff;                                         \
+       new <<= bitoff;                                         \
+       prev_mask = (u32)(type)-1 << bitoff;                    \
+                                                               \
+       __asm__ __volatile__(                                   \
+       br                                                      \
+"1:    lwarx   %0,0,%3\n"                                      \
+"      and     %1,%0,%6\n"                                     \
+"      cmpw    0,%1,%4\n"                                      \
+"      bne-    2f\n"                                           \
+"      andc    %1,%0,%6\n"                                     \
+"      or      %1,%1,%5\n"                                     \
+       PPC405_ERR77(0,%3)                                      \
+"      stwcx.  %1,0,%3\n"                                      \
+"      bne-    1b\n"                                           \
+       br2                                                     \
+       "\n"                                                    \
+"2:"                                                           \
+       : "=&r" (prev), "=&r" (tmp), "+m" (*(u32*)p)            \
+       : "r" (p), "r" (old), "r" (new), "r" (prev_mask)        \
+       : "cc", cl);                                            \
+                                                               \
+       return prev >> bitoff;                                  \
+}
+
 /*
  * Atomic exchange
  *
  * the previous value stored there.
  */
 
+XCHG_GEN(u8, _local, "memory");
+XCHG_GEN(u8, _relaxed, "cc");
+XCHG_GEN(u16, _local, "memory");
+XCHG_GEN(u16, _relaxed, "cc");
+
 static __always_inline unsigned long
 __xchg_u32_local(volatile void *p, unsigned long val)
 {
@@ -85,9 +155,13 @@ __xchg_u64_relaxed(u64 *p, unsigned long val)
 #endif
 
 static __always_inline unsigned long
-__xchg_local(volatile void *ptr, unsigned long x, unsigned int size)
+__xchg_local(void *ptr, unsigned long x, unsigned int size)
 {
        switch (size) {
+       case 1:
+               return __xchg_u8_local(ptr, x);
+       case 2:
+               return __xchg_u16_local(ptr, x);
        case 4:
                return __xchg_u32_local(ptr, x);
 #ifdef CONFIG_PPC64
@@ -103,6 +177,10 @@ static __always_inline unsigned long
 __xchg_relaxed(void *ptr, unsigned long x, unsigned int size)
 {
        switch (size) {
+       case 1:
+               return __xchg_u8_relaxed(ptr, x);
+       case 2:
+               return __xchg_u16_relaxed(ptr, x);
        case 4:
                return __xchg_u32_relaxed(ptr, x);
 #ifdef CONFIG_PPC64
@@ -131,6 +209,15 @@ __xchg_relaxed(void *ptr, unsigned long x, unsigned int size)
  * and return the old value of *p.
  */
 
+CMPXCHG_GEN(u8, , PPC_ATOMIC_ENTRY_BARRIER, PPC_ATOMIC_EXIT_BARRIER, "memory");
+CMPXCHG_GEN(u8, _local, , , "memory");
+CMPXCHG_GEN(u8, _acquire, , PPC_ACQUIRE_BARRIER, "memory");
+CMPXCHG_GEN(u8, _relaxed, , , "cc");
+CMPXCHG_GEN(u16, , PPC_ATOMIC_ENTRY_BARRIER, PPC_ATOMIC_EXIT_BARRIER, "memory");
+CMPXCHG_GEN(u16, _local, , , "memory");
+CMPXCHG_GEN(u16, _acquire, , PPC_ACQUIRE_BARRIER, "memory");
+CMPXCHG_GEN(u16, _relaxed, , , "cc");
+
 static __always_inline unsigned long
 __cmpxchg_u32(volatile unsigned int *p, unsigned long old, unsigned long new)
 {
@@ -312,10 +399,14 @@ __cmpxchg_u64_acquire(u64 *p, unsigned long old, unsigned long new)
 #endif
 
 static __always_inline unsigned long
-__cmpxchg(volatile void *ptr, unsigned long old, unsigned long new,
+__cmpxchg(void *ptr, unsigned long old, unsigned long new,
          unsigned int size)
 {
        switch (size) {
+       case 1:
+               return __cmpxchg_u8(ptr, old, new);
+       case 2:
+               return __cmpxchg_u16(ptr, old, new);
        case 4:
                return __cmpxchg_u32(ptr, old, new);
 #ifdef CONFIG_PPC64
@@ -328,10 +419,14 @@ __cmpxchg(volatile void *ptr, unsigned long old, unsigned long new,
 }
 
 static __always_inline unsigned long
-__cmpxchg_local(volatile void *ptr, unsigned long old, unsigned long new,
+__cmpxchg_local(void *ptr, unsigned long old, unsigned long new,
          unsigned int size)
 {
        switch (size) {
+       case 1:
+               return __cmpxchg_u8_local(ptr, old, new);
+       case 2:
+               return __cmpxchg_u16_local(ptr, old, new);
        case 4:
                return __cmpxchg_u32_local(ptr, old, new);
 #ifdef CONFIG_PPC64
@@ -348,6 +443,10 @@ __cmpxchg_relaxed(void *ptr, unsigned long old, unsigned long new,
                  unsigned int size)
 {
        switch (size) {
+       case 1:
+               return __cmpxchg_u8_relaxed(ptr, old, new);
+       case 2:
+               return __cmpxchg_u16_relaxed(ptr, old, new);
        case 4:
                return __cmpxchg_u32_relaxed(ptr, old, new);
 #ifdef CONFIG_PPC64
@@ -364,6 +463,10 @@ __cmpxchg_acquire(void *ptr, unsigned long old, unsigned long new,
                  unsigned int size)
 {
        switch (size) {
+       case 1:
+               return __cmpxchg_u8_acquire(ptr, old, new);
+       case 2:
+               return __cmpxchg_u16_acquire(ptr, old, new);
        case 4:
                return __cmpxchg_u32_acquire(ptr, old, new);
 #ifdef CONFIG_PPC64