diff options
author | Rich Felker <dalias@aerifal.cx> | 2016-01-23 14:03:40 -0500 |
---|---|---|
committer | Rich Felker <dalias@aerifal.cx> | 2016-01-23 14:03:40 -0500 |
commit | b17fbd3520640745051d2fd9f1fd334ceb8b8783 (patch) | |
tree | e4b6b233e866b6baef822554855f23c342e9559a /arch/aarch64 | |
parent | 0f5eb3de2955e2eaf5ee4b9637d9216f82a97fbd (diff) | |
download | musl-b17fbd3520640745051d2fd9f1fd334ceb8b8783.tar.gz musl-b17fbd3520640745051d2fd9f1fd334ceb8b8783.tar.bz2 musl-b17fbd3520640745051d2fd9f1fd334ceb8b8783.tar.xz musl-b17fbd3520640745051d2fd9f1fd334ceb8b8783.zip |
improve aarch64 atomics
aarch64 provides ll/sc variants with acquire/release memory order,
freeing us from the need to have full barriers both before and after
the ll/sc operation. previously they were not used because the a_cas
can fail without performing a_sc, in which case half of the barrier
would be omitted. instead, define a custom version of a_cas for
aarch64 which uses a_barrier explicitly when aborting the cas
operation. aside from cas, other operations built on top of ll/sc are
not affected since they never abort but rather loop until they
succeed.
a split ll/sc version of the pointer-sized a_cas_p is also introduced
using the same technique.
patch by Szabolcs Nagy.
Diffstat (limited to 'arch/aarch64')
-rw-r--r-- | arch/aarch64/atomic_arch.h | 52 |
1 files changed, 36 insertions, 16 deletions
diff --git a/arch/aarch64/atomic_arch.h b/arch/aarch64/atomic_arch.h index 84499e2c..14fea03b 100644 --- a/arch/aarch64/atomic_arch.h +++ b/arch/aarch64/atomic_arch.h @@ -2,7 +2,7 @@ static inline int a_ll(volatile int *p) { int v; - __asm__ __volatile__ ("ldxr %0, %1" : "=r"(v) : "Q"(*p)); + __asm__ __volatile__ ("ldaxr %0, %1" : "=r"(v) : "Q"(*p)); return v; } @@ -10,7 +10,7 @@ static inline int a_ll(volatile int *p) static inline int a_sc(volatile int *p, int v) { int r; - __asm__ __volatile__ ("stxr %w0,%1,%2" : "=&r"(r) : "r"(v), "Q"(*p) : "memory"); + __asm__ __volatile__ ("stlxr %w0,%1,%2" : "=&r"(r) : "r"(v), "Q"(*p) : "memory"); return !r; } @@ -20,25 +20,45 @@ static inline void a_barrier() __asm__ __volatile__ ("dmb ish" : : : "memory"); } -#define a_pre_llsc a_barrier -#define a_post_llsc a_barrier +#define a_cas a_cas +static inline int a_cas(volatile int *p, int t, int s) +{ + int old; + do { + old = a_ll(p); + if (old != t) { + a_barrier(); + break; + } + } while (!a_sc(p, s)); + return old; +} + +static inline void *a_ll_p(volatile void *p) +{ + void *v; + __asm__ __volatile__ ("ldaxr %0, %1" : "=r"(v) : "Q"(*(void *volatile *)p)); + return v; +} + +static inline int a_sc_p(volatile int *p, void *v) +{ + int r; + __asm__ __volatile__ ("stlxr %w0,%1,%2" : "=&r"(r) : "r"(v), "Q"(*(void *volatile *)p) : "memory"); + return !r; +} #define a_cas_p a_cas_p static inline void *a_cas_p(volatile void *p, void *t, void *s) { void *old; - __asm__ __volatile__( - " dmb ish\n" - "1: ldxr %0,%3\n" - " cmp %0,%1\n" - " b.ne 1f\n" - " stxr %w0,%2,%3\n" - " cbnz %w0,1b\n" - " mov %0,%1\n" - "1: dmb ish\n" - : "=&r"(old) - : "r"(t), "r"(s), "Q"(*(void *volatile *)p) - : "memory", "cc"); + do { + old = a_ll_p(p); + if (old != t) { + a_barrier(); + break; + } + } while (!a_sc_p(p, s)); return old; } |