summaryrefslogtreecommitdiffstats
path: root/cpukit
diff options
context:
space:
mode:
authorWeiY <wei.a.yang@gmail.com>2013-01-26 00:34:57 +0800
committerGedare Bloom <gedare@rtems.org>2013-02-07 17:02:12 -0500
commit4c96372591ebd02dd6c0f612d29b915ba99d3747 (patch)
tree564f9d8709b7d81e3bf42ee1b8e1e072a9bdf792 /cpukit
parenttests: atomic support for RTEMS. SMP tests for atomic operations. (diff)
downloadrtems-4c96372591ebd02dd6c0f612d29b915ba99d3747.tar.bz2
score: atomic support for RTEMS. Cleanup.
Diffstat (limited to 'cpukit')
-rw-r--r--cpukit/score/cpu/i386/rtems/score/cpuatomic.h212
-rw-r--r--cpukit/score/cpu/powerpc/rtems/score/cpuatomic.h76
-rw-r--r--cpukit/score/inline/rtems/score/atomic.inl156
3 files changed, 218 insertions, 226 deletions
diff --git a/cpukit/score/cpu/i386/rtems/score/cpuatomic.h b/cpukit/score/cpu/i386/rtems/score/cpuatomic.h
index ae57584de4..eff6a35534 100644
--- a/cpukit/score/cpu/i386/rtems/score/cpuatomic.h
+++ b/cpukit/score/cpu/i386/rtems/score/cpuatomic.h
@@ -67,91 +67,91 @@ extern "C" {
* is always consistent, so we only need to take care of compiler.
*/
#define ATOMIC_STORE_LOAD(NAME, TYPE, LOP, SOP) \
-static inline Atomic_##TYPE \
-_CPU_Atomic_Load_##NAME(volatile Atomic_##TYPE *p) \
-{ \
- Atomic_##TYPE tmp; \
- \
- tmp = *p; \
- __asm __volatile("" : : : "memory"); \
- return (tmp); \
-} \
- \
+static inline Atomic_##TYPE \
+_CPU_Atomic_Load_##NAME(volatile Atomic_##TYPE *p) \
+{ \
+ Atomic_##TYPE tmp; \
+ \
+ tmp = *p; \
+ __asm __volatile("" : : : "memory"); \
+ return (tmp); \
+} \
+ \
static inline _CPU_Atomic_Load_acq_##NAME(volatile Atomic_##TYPE *p) \
-{ \
- Atomic_##TYPE tmp; \
- \
- tmp = *p; \
- __asm __volatile("" : : : "memory"); \
- return (tmp); \
-} \
- \
-static inline void \
+{ \
+ Atomic_##TYPE tmp; \
+ \
+ tmp = *p; \
+ __asm __volatile("" : : : "memory"); \
+ return (tmp); \
+} \
+ \
+static inline void \
_CPU_Atomic_Store_##NAME(volatile Atomic_##TYPE *p, Atomic_##TYPE v) \
-{ \
- __asm __volatile("" : : : "memory"); \
- *p = v; \
-} \
- \
-static inline void \
+{ \
+ __asm __volatile("" : : : "memory"); \
+ *p = v; \
+} \
+ \
+static inline void \
_CPU_Atomic_Store_rel_##NAME(volatile Atomic_##TYPE *p, Atomic_##TYPE v) \
-{ \
- __asm __volatile("" : : : "memory"); \
- *p = v; \
-} \
+{ \
+ __asm __volatile("" : : : "memory"); \
+ *p = v; \
+} \
#else /* !(!SMP) */
#define ATOMIC_STORE_LOAD(NAME, TYPE, LOP, SOP) \
-static inline Atomic_##TYPE \
-_CPU_Atomic_Load_##NAME(volatile Atomic_##TYPE *p) \
-{ \
- Atomic_##TYPE res; \
- \
- __asm __volatile(MPLOCKED LOP \
- : "=a" (res), /* 0 */ \
- "=m" (*p) /* 1 */ \
- : "m" (*p) /* 2 */ \
- : "memory", "cc"); \
- \
- return (res); \
-} \
- \
-static inline Atomic_##TYPE \
-_CPU_Atomic_Load_acq_##NAME(volatile Atomic_##TYPE *p) \
-{ \
- Atomic_##TYPE res; \
- \
- __asm __volatile(MPLOCKED LOP \
- : "=a" (res), /* 0 */ \
- "=m" (*p) /* 1 */ \
- : "m" (*p) /* 2 */ \
- : "memory", "cc"); \
- \
- return (res); \
-} \
- \
-/* \
- * The XCHG instruction asserts LOCK automagically. \
- */ \
-static inline void \
+static inline Atomic_##TYPE \
+_CPU_Atomic_Load_##NAME(volatile Atomic_##TYPE *p) \
+{ \
+ Atomic_##TYPE res; \
+ \
+ __asm __volatile(MPLOCKED LOP \
+ : "=a" (res), /* 0 */ \
+ "=m" (*p) /* 1 */ \
+ : "m" (*p) /* 2 */ \
+ : "memory", "cc"); \
+ \
+ return (res); \
+} \
+ \
+static inline Atomic_##TYPE \
+_CPU_Atomic_Load_acq_##NAME(volatile Atomic_##TYPE *p) \
+{ \
+ Atomic_##TYPE res; \
+ \
+ __asm __volatile(MPLOCKED LOP \
+ : "=a" (res), /* 0 */ \
+ "=m" (*p) /* 1 */ \
+ : "m" (*p) /* 2 */ \
+ : "memory", "cc"); \
+ \
+ return (res); \
+} \
+ \
+/* \
+ * The XCHG instruction asserts LOCK automagically. \
+ */ \
+static inline void \
_CPU_Atomic_Store_##NAME(volatile Atomic_##TYPE *p, Atomic_##TYPE v) \
-{ \
- __asm __volatile(SOP \
- : "=m" (*p), /* 0 */ \
- "+r" (v) /* 1 */ \
- : "m" (*p) /* 2 */ \
- : "memory"); \
-} \
-static inline void \
+{ \
+ __asm __volatile(SOP \
+ : "=m" (*p), /* 0 */ \
+ "+r" (v) /* 1 */ \
+ : "m" (*p) /* 2 */ \
+ : "memory"); \
+} \
+static inline void \
_CPU_Atomic_Store_rel_##NAME(volatile Atomic_##TYPE *p, Atomic_##TYPE v) \
-{ \
- __asm __volatile(SOP \
- : "=m" (*p), /* 0 */ \
- "+r" (v) /* 1 */ \
- : "m" (*p) /* 2 */ \
- : "memory"); \
-} \
+{ \
+ __asm __volatile(SOP \
+ : "=m" (*p), /* 0 */ \
+ "+r" (v) /* 1 */ \
+ : "m" (*p) /* 2 */ \
+ : "memory"); \
+} \
#endif /* !SMP */
@@ -160,8 +160,8 @@ _CPU_Atomic_Store_rel_##NAME(volatile Atomic_##TYPE *p, Atomic_##TYPE v) \
* GCC aggressively reorders operations and memory clobbering is necessary
* in order to avoid that for memory barriers.
*/
-#define ATOMIC_FETCH_GENERIC(NAME, TYPENAME, TYPE, OP, CONS, V) \
-static inline void \
+#define ATOMIC_FETCH_GENERIC(NAME, TYPENAME, TYPE, OP, CONS, V) \
+static inline void \
_CPU_Atomic_Fetch_##NAME##_##TYPENAME(volatile Atomic_##TYPE *p, Atomic_##TYPE v) \
{ \
__asm __volatile(MPLOCKED OP \
@@ -251,37 +251,37 @@ ATOMIC_FETCH_GENERIC(and, long, Long, "andl %1,%0", "ir", v);
#define _CPU_Atomic_Compare_exchange_rel_long _CPU_Atomic_Compare_exchange_long
/* Operations on 32-bit double words. */
-#define _CPU_Atomic_Fetch_or_32(p, v) \
+#define _CPU_Atomic_Fetch_or_32(p, v) \
_CPU_Atomic_Fetch_or_int((volatile Atomic_Int *)(p), (Atomic_Int)(v))
#define _CPU_Atomic_Fetch_or_acq_32(p, v) \
_CPU_Atomic_Fetch_or_acq_int((volatile Atomic_Int *)(p), (Atomic_Int)(v))
#define _CPU_Atomic_Fetch_or_rel_32(p, v) \
_CPU_Atomic_Fetch_or_rel_int((volatile Atomic_Int *)(p), (Atomic_Int)(v))
-#define _CPU_Atomic_Fetch_and_32(p, v) \
+#define _CPU_Atomic_Fetch_and_32(p, v) \
_CPU_Atomic_Fetch_and_int((volatile Atomic_Int *)(p), (Atomic_Int)(v))
-#define _CPU_Atomic_Fetch_and_acq_32(p, v) \
+#define _CPU_Atomic_Fetch_and_acq_32(p, v) \
_CPU_Atomic_Fetch_and_acq_int((volatile Atomic_Int *)(p), (Atomic_Int)(v))
-#define _CPU_Atomic_Fetch_and_rel_32(p, v) \
+#define _CPU_Atomic_Fetch_and_rel_32(p, v) \
_CPU_Atomic_Fetch_and_rel_int((volatile Atomic_Int *)(p), (Atomic_Int)(v))
-#define _CPU_Atomic_Fetch_add_32(p, v) \
+#define _CPU_Atomic_Fetch_add_32(p, v) \
_CPU_Atomic_Fetch_add_int((volatile Atomic_Int *)(p), (Atomic_Int)(v))
-#define _CPU_Atomic_Fetch_add_acq_32(p, v) \
+#define _CPU_Atomic_Fetch_add_acq_32(p, v) \
_CPU_Atomic_Fetch_add_acq_int((volatile Atomic_Int *)(p), (Atomic_Int)(v))
-#define _CPU_Atomic_Fetch_add_rel_32(p, v) \
+#define _CPU_Atomic_Fetch_add_rel_32(p, v) \
_CPU_Atomic_Fetch_add_rel_int((volatile Atomic_Int *)(p), (Atomic_Int)(v))
-#define _CPU_Atomic_Fetch_sub_32(p, v) \
+#define _CPU_Atomic_Fetch_sub_32(p, v) \
_CPU_Atomic_Fetch_sub_int((volatile Atomic_Int *)(p), (Atomic_Int)(v))
-#define _CPU_Atomic_Fetch_sub_acq_32(p, v) \
+#define _CPU_Atomic_Fetch_sub_acq_32(p, v) \
_CPU_Atomic_Fetch_sub_acq_int((volatile Atomic_Int *)(p), (Atomic_Int)(v))
-#define _CPU_Atomic_Fetch_sub_rel_32(p, v) \
+#define _CPU_Atomic_Fetch_sub_rel_32(p, v) \
_CPU_Atomic_Fetch_sub_rel_int((volatile Atomic_Int *)(p), (Atomic_Int)(v))
-#define _CPU_Atomic_Load_32(p) \
+#define _CPU_Atomic_Load_32(p) \
_CPU_Atomic_Load_int((volatile Atomic_Int *)(p))
-#define _CPU_Atomic_Load_acq_32(p) \
+#define _CPU_Atomic_Load_acq_32(p) \
_CPU_Atomic_Load_acq_int((volatile Atomic_Int *)(p))
-#define _CPU_Atomic_Store_32(p, v) \
+#define _CPU_Atomic_Store_32(p, v) \
_CPU_Atomic_Store_int((volatile Atomic_Int *)(p), (Atomic_Int)(v))
-#define _CPU_Atomic_Store_rel_32(p, v) \
+#define _CPU_Atomic_Store_rel_32(p, v) \
_CPU_Atomic_Store_rel_int((volatile Atomic_Int *)(p), (Atomic_Int)(v))
#define _CPU_Atomic_Compare_exchange_32(dst, old, new) \
_CPU_Atomic_Compare_exchange_int((volatile Atomic_Int *)(dst), (Atomic_Int)(old), (Atomic_Int)(new))
@@ -291,37 +291,37 @@ ATOMIC_FETCH_GENERIC(and, long, Long, "andl %1,%0", "ir", v);
_CPU_Atomic_Compare_exchange_rel_int((volatile Atomic_Int *)(dst), (Atomic_Int)(old), (Atomic_Int)(new))
/* Operations on pointers. */
-#define _CPU_Atomic_Fetch_or_ptr(p, v) \
+#define _CPU_Atomic_Fetch_or_ptr(p, v) \
_CPU_Atomic_Fetch_or_int((volatile Atomic_Int *)(p), (Atomic_Int)(v))
#define _CPU_Atomic_Fetch_or_acq_ptr(p, v) \
_CPU_Atomic_Fetch_or_acq_int((volatile Atomic_Int *)(p), (Atomic_Int)(v))
#define _CPU_Atomic_Fetch_or_rel_ptr(p, v) \
_CPU_Atomic_Fetch_or_rel_int((volatile Atomic_Int *)(p), (Atomic_Int)(v))
-#define _CPU_Atomic_Fetch_and_ptr(p, v) \
+#define _CPU_Atomic_Fetch_and_ptr(p, v) \
_CPU_Atomic_Fetch_and_int((volatile Atomic_Int *)(p), (Atomic_Int)(v))
-#define _CPU_Atomic_Fetch_and_acq_ptr(p, v) \
+#define _CPU_Atomic_Fetch_and_acq_ptr(p, v)\
_CPU_Atomic_Fetch_and_acq_int((volatile Atomic_Int *)(p), (Atomic_Int)(v))
-#define _CPU_Atomic_Fetch_and_rel_ptr(p, v) \
+#define _CPU_Atomic_Fetch_and_rel_ptr(p, v)\
_CPU_Atomic_Fetch_and_rel_int((volatile Atomic_Int *)(p), (Atomic_Int)(v))
-#define _CPU_Atomic_Fetch_add_ptr(p, v) \
+#define _CPU_Atomic_Fetch_add_ptr(p, v) \
_CPU_Atomic_Fetch_add_int((volatile Atomic_Int *)(p), (Atomic_Int)(v))
-#define _CPU_Atomic_Fetch_add_acq_ptr(p, v) \
+#define _CPU_Atomic_Fetch_add_acq_ptr(p, v)\
_CPU_Atomic_Fetch_add_acq_int((volatile Atomic_Int *)(p), (Atomic_Int)(v))
-#define _CPU_Atomic_Fetch_add_rel_ptr(p, v) \
+#define _CPU_Atomic_Fetch_add_rel_ptr(p, v)\
_CPU_Atomic_Fetch_add_rel_int((volatile Atomic_Int *)(p), (Atomic_Int)(v))
-#define _CPU_Atomic_Fetch_sub_ptr(p, v) \
+#define _CPU_Atomic_Fetch_sub_ptr(p, v) \
_CPU_Atomic_Fetch_sub_int((volatile Atomic_Int *)(p), (Atomic_Int)(v))
-#define _CPU_Atomic_Fetch_sub_acq_ptr(p, v) \
+#define _CPU_Atomic_Fetch_sub_acq_ptr(p, v)\
_CPU_Atomic_Fetch_sub_acq_int((volatile Atomic_Int *)(p), (Atomic_Int)(v))
-#define _CPU_Atomic_Fetch_sub_rel_ptr(p, v) \
+#define _CPU_Atomic_Fetch_sub_rel_ptr(p, v)\
_CPU_Atomic_Fetch_sub_rel_int((volatile Atomic_Int *)(p), (Atomic_Int)(v))
-#define _CPU_Atomic_Load_ptr(p) \
+#define _CPU_Atomic_Load_ptr(p) \
_CPU_Atomic_Load_int((volatile Atomic_Int *)(p))
-#define _CPU_Atomic_Load_acq_ptr(p) \
+#define _CPU_Atomic_Load_acq_ptr(p) \
_CPU_Atomic_Load_acq_int((volatile Atomic_Int *)(p))
-#define _CPU_Atomic_Store_ptr(p, v) \
+#define _CPU_Atomic_Store_ptr(p, v) \
_CPU_Atomic_Store_int((volatile Atomic_Int *)(p), (v))
-#define _CPU_Atomic_Store_rel_ptr(p, v) \
+#define _CPU_Atomic_Store_rel_ptr(p, v) \
_CPU_Atomic_Store_rel_int((volatile Atomic_Int *)(p), (v))
#define _CPU_Atomic_Compare_exchange_ptr(dst, old, new) \
_CPU_Atomic_Compare_exchange_int((volatile Atomic_Int *)(dst), (Atomic_Int)(old), (Atomic_Int)(new))
diff --git a/cpukit/score/cpu/powerpc/rtems/score/cpuatomic.h b/cpukit/score/cpu/powerpc/rtems/score/cpuatomic.h
index eeefa360f6..0ffb447d41 100644
--- a/cpukit/score/cpu/powerpc/rtems/score/cpuatomic.h
+++ b/cpukit/score/cpu/powerpc/rtems/score/cpuatomic.h
@@ -88,39 +88,39 @@ extern "C" {
: "cc", "memory") \
/* __CPU_Atomic_Fetch_add_long */
-#define _ATOMIC_ADD(typename, type) \
+#define _ATOMIC_ADD(typename, type) \
static __inline void \
_CPU_Atomic_Fetch_add_##typename(volatile Atomic_##type *p, Atomic_##type v) { \
Atomic_##type t; \
- __CPU_Atomic_Fetch_add_##typename(p, v, t); \
+ __CPU_Atomic_Fetch_add_##typename(p, v, t); \
} \
\
static __inline void \
_CPU_Atomic_Fetch_add_acq_##typename(volatile Atomic_##type *p, Atomic_##type v) { \
- Atomic_##type t; \
- __CPU_Atomic_Fetch_add_##typename(p, v, t); \
- __ATOMIC_BARRIER; \
- } \
- \
- static __inline void \
+ Atomic_##type t; \
+ __CPU_Atomic_Fetch_add_##typename(p, v, t); \
+ __ATOMIC_BARRIER; \
+ } \
+ \
+ static __inline void \
_CPU_Atomic_Fetch_add_rel_##typename(volatile Atomic_##type *p, Atomic_##type v) { \
- Atomic_##type t; \
- __ATOMIC_BARRIER; \
- __CPU_Atomic_Fetch_add_##typename(p, v, t); \
- } \
+ Atomic_##type t; \
+ __ATOMIC_BARRIER; \
+ __CPU_Atomic_Fetch_add_##typename(p, v, t); \
+ } \
/* _ATOMIC_ADD */
_ATOMIC_ADD(int, Int)
_ATOMIC_ADD(long, Long)
-#define _CPU_Atomic_Fetch_add_32(p, v) \
+#define _CPU_Atomic_Fetch_add_32(p, v) \
_CPU_Atomic_Fetch_add_int((volatile Atomic_Int *)(p), (Atomic_Int)(v))
#define _CPU_Atomic_Fetch_add_acq_32(p, v) \
_CPU_Atomic_Fetch_add_acq_int((volatile Atomic_Int *)(p), (Atomic_Int)(v))
#define _CPU_Atomic_Fetch_add_rel_32(p, v) \
_CPU_Atomic_Fetch_add_rel_int((volatile Atomic_Int *)(p), (Atomic_Int)(v))
-#define _CPU_Atomic_Fetch_add_ptr(p, v) \
+#define _CPU_Atomic_Fetch_add_ptr(p, v) \
_CPU_Atomic_Fetch_add_int((volatile Atomic_Int *)(p), (Atomic_Int)(v))
#define _CPU_Atomic_Fetch_add_acq_ptr(p, v) \
_CPU_Atomic_Fetch_add_acq_int((volatile Atomic_Int *)(p), (Atomic_Int)(v))
@@ -158,17 +158,17 @@ _ATOMIC_ADD(long, Long)
: "cc", "memory") \
/* _CPU_Atomic_Fetch_and_long */
-#define _ATOMIC_AND(typename, type) \
+#define _ATOMIC_AND(typename, type) \
static __inline void \
_CPU_Atomic_Fetch_and_##typename(volatile Atomic_##type *p, Atomic_##type v) { \
Atomic_##type t; \
- __CPU_Atomic_Fetch_and_##typename(p, v, t); \
+ __CPU_Atomic_Fetch_and_##typename(p, v, t); \
} \
\
static __inline void \
_CPU_Atomic_Fetch_and_acq_##typename(volatile Atomic_##type *p, Atomic_##type v) { \
Atomic_##type t; \
- __CPU_Atomic_Fetch_and_##typename(p, v, t); \
+ __CPU_Atomic_Fetch_and_##typename(p, v, t); \
__ATOMIC_BARRIER; \
} \
\
@@ -176,7 +176,7 @@ _ATOMIC_ADD(long, Long)
_CPU_Atomic_Fetch_and_rel_##typename(volatile Atomic_##type *p, Atomic_##type v) { \
Atomic_##type t; \
__ATOMIC_BARRIER; \
- __CPU_Atomic_Fetch_and_##typename(p, v, t); \
+ __CPU_Atomic_Fetch_and_##typename(p, v, t); \
} \
/* _ATOMIC_AND */
@@ -184,14 +184,14 @@ _ATOMIC_ADD(long, Long)
_ATOMIC_AND(int, Int)
_ATOMIC_AND(long, Long)
-#define _CPU_Atomic_Fetch_and_32(p, v) \
+#define _CPU_Atomic_Fetch_and_32(p, v) \
_CPU_Atomic_Fetch_and_int((volatile Atomic_Int *)(p), (Atomic_Int)(v))
#define _CPU_Atomic_Fetch_and_acq_32(p, v) \
_CPU_Atomic_Fetch_and_acq_int((volatile Atomic_Int *)(p), (Atomic_Int)(v))
#define _CPU_Atomic_Fetch_and_rel_32(p, v) \
_CPU_Atomic_Fetch_and_rel_int((volatile Atomic_Int *)(p), (Atomic_Int)(v))
-#define _CPU_Atomic_Fetch_and_ptr(p, v) \
+#define _CPU_Atomic_Fetch_and_ptr(p, v) \
_CPU_Atomic_Fetch_and_int((volatile Atomic_Int *)(p), (Atomic_Int)(v))
#define _CPU_Atomic_Fetch_and_acq_ptr(p, v) \
_CPU_Atomic_Fetch_and_acq_int((volatile Atomic_Int *)(p), (Atomic_Int)(v))
@@ -230,17 +230,17 @@ _ATOMIC_AND(long, Long)
: "cc", "memory") \
/* __CPU_Atomic_Fetch_or_long */
-#define _ATOMIC_OR(typename, type) \
+#define _ATOMIC_OR(typename, type) \
static __inline void \
_CPU_Atomic_Fetch_or_##typename(volatile Atomic_##type *p, Atomic_##type v) { \
Atomic_##type t; \
- __CPU_Atomic_Fetch_or_##typename(p, v, t); \
+ __CPU_Atomic_Fetch_or_##typename(p, v, t); \
} \
\
static __inline void \
_CPU_Atomic_Fetch_or_acq_##typename(volatile Atomic_##type *p, Atomic_##type v) { \
Atomic_##type t; \
- __CPU_Atomic_Fetch_or_##typename(p, v, t); \
+ __CPU_Atomic_Fetch_or_##typename(p, v, t); \
__ATOMIC_BARRIER; \
} \
\
@@ -248,21 +248,21 @@ _ATOMIC_AND(long, Long)
_CPU_Atomic_Fetch_or_rel_##typename(volatile Atomic_##type *p, Atomic_##type v) { \
Atomic_##type t; \
__ATOMIC_BARRIER; \
- __CPU_Atomic_Fetch_or_##typename(p, v, t); \
+ __CPU_Atomic_Fetch_or_##typename(p, v, t); \
} \
/* _ATOMIC_OR */
_ATOMIC_OR(int, Int)
_ATOMIC_OR(long, Long)
-#define _CPU_Atomic_Fetch_or_32(p, v) \
+#define _CPU_Atomic_Fetch_or_32(p, v) \
_CPU_Atomic_Fetch_or_int((volatile Atomic_Int *)(p), (Atomic_Int)(v))
#define _CPU_Atomic_Fetch_or_acq_32(p, v) \
_CPU_Atomic_Fetch_or_acq_int((volatile Atomic_Int *)(p), (Atomic_Int)(v))
#define _CPU_Atomic_Fetch_or_rel_32(p, v) \
_CPU_Atomic_Fetch_or_rel_int((volatile Atomic_Int *)(p), (Atomic_Int)(v))
-#define _CPU_Atomic_Fetch_or_ptr(p, v) \
+#define _CPU_Atomic_Fetch_or_ptr(p, v) \
_CPU_Atomic_Fetch_or_int((volatile Atomic_Int *)(p), (Atomic_Int)(v))
#define _CPU_Atomic_Fetch_or_acq_ptr(p, v) \
_CPU_Atomic_Fetch_or_acq_int((volatile Atomic_Int *)(p), (Atomic_Int)(v))
@@ -300,17 +300,17 @@ _ATOMIC_OR(long, Long)
: "cc", "memory") \
/* __CPU_Atomic_Fetch_sub_long */
-#define _ATOMIC_SUB(typename, type) \
+#define _ATOMIC_SUB(typename, type) \
static __inline void \
- _CPU_Atomic_Fetch_sub_##typename(volatile Atomic_##type *p, Atomic_##type v) { \
+ _CPU_Atomic_Fetch_sub_##typename(volatile Atomic_##type *p, Atomic_##type v) { \
Atomic_##type t; \
- __CPU_Atomic_Fetch_sub_##typename(p, v, t); \
+ __CPU_Atomic_Fetch_sub_##typename(p, v, t); \
} \
\
static __inline void \
_CPU_Atomic_Fetch_sub_acq_##typename(volatile Atomic_##type *p, Atomic_##type v) { \
Atomic_##type t; \
- __CPU_Atomic_Fetch_sub_##typename(p, v, t); \
+ __CPU_Atomic_Fetch_sub_##typename(p, v, t); \
__ATOMIC_BARRIER; \
} \
\
@@ -318,7 +318,7 @@ _ATOMIC_OR(long, Long)
_CPU_Atomic_Fetch_sub_rel_##typename(volatile Atomic_##type *p, Atomic_##type v) { \
Atomic_##type t; \
__ATOMIC_BARRIER; \
- __CPU_Atomic_Fetch_sub_##typename(p, v, t); \
+ __CPU_Atomic_Fetch_sub_##typename(p, v, t); \
} \
/* _ATOMIC_SUB */
@@ -326,14 +326,14 @@ _ATOMIC_OR(long, Long)
_ATOMIC_SUB(int, Int)
_ATOMIC_SUB(long, Long)
-#define _CPU_Atomic_Fetch_sub_32(p, v) \
+#define _CPU_Atomic_Fetch_sub_32(p, v) \
_CPU_Atomic_Fetch_sub_int((volatile Atomic_Int *)(p), (Atomic_Int)(v))
#define _CPU_Atomic_Fetch_sub_acq_32(p, v) \
_CPU_Atomic_Fetch_sub_acq_int((volatile Atomic_Int *)(p), (Atomic_Int)(v))
#define _CPU_Atomic_Fetch_sub_rel_32(p, v) \
_CPU_Atomic_Fetch_sub_rel_int((volatile Atomic_Int *)(p), (Atomic_Int)(v))
-#define _CPU_Atomic_Fetch_sub_ptr(p, v) \
+#define _CPU_Atomic_Fetch_sub_ptr(p, v) \
_CPU_Atomic_Fetch_sub_int((volatile Atomic_Int *)(p), (Atomic_Int)(v))
#define _CPU_Atomic_Fetch_sub_acq_ptr(p, v) \
_CPU_Atomic_Fetch_sub_acq_int((volatile Atomic_Int *)(p), (Atomic_Int)(v))
@@ -367,7 +367,7 @@ _CPU_Atomic_Load_acq_##TYPENAME(volatile Atomic_##TYPE *p) \
} \
\
static __inline void \
-_CPU_Atomic_Store_##TYPENAME(volatile Atomic_##TYPE *p, Atomic_##TYPE v) \
+_CPU_Atomic_Store_##TYPENAME(volatile Atomic_##TYPE *p, Atomic_##TYPE v) \
{ \
*p = v; \
} \
@@ -380,7 +380,7 @@ _CPU_Atomic_Store_rel_##TYPENAME(volatile Atomic_##TYPE *p, Atomic_##TYPE v) \
ATOMIC_STORE_LOAD(int, Int)
-#define _CPU_Atomic_Load_32(p) \
+#define _CPU_Atomic_Load_32(p) \
_CPU_Atomic_Load_int((volatile Atomic_Int *)(p))
#define _CPU_Atomic_Load_acq_32(p) \
_CPU_Atomic_Load_acq_int((volatile Atomic_Int *)(p))
@@ -413,7 +413,7 @@ _CPU_Atomic_Store_rel_long(volatile Atomic_Long *addr, Atomic_Long val)
_CPU_Atomic_Store_rel_int((volatile Atomic_Int *)addr, (Atomic_Int)val);
}
-#define _CPU_Atomic_Load_ptr(p) \
+#define _CPU_Atomic_Load_ptr(p) \
_CPU_Atomic_Load_int((volatile Atomic_Int *)(p))
#define _CPU_Atomic_Load_acq_ptr(p) \
_CPU_Atomic_Load_acq_int((volatile Atomic_Int *)(p))
@@ -511,14 +511,14 @@ _CPU_Atomic_Compare_exchange_rel_long(volatile Atomic_Long *p, Atomic_Long cmpva
return (_CPU_Atomic_Compare_exchange_long(p, cmpval, newval));
}
-#define _CPU_Atomic_Compare_exchange_32(dst, old, new) \
+#define _CPU_Atomic_Compare_exchange_32(dst, old, new) \
_CPU_Atomic_Compare_exchange_int((volatile Atomic_Int *)(dst), (Atomic_Int)(old), (Atomic_Int)(new))
#define _CPU_Atomic_Compare_exchange_acq_32(dst, old, new) \
_CPU_Atomic_Compare_exchange_acq_int((volatile Atomic_Int *)(dst), (Atomic_Int)(old), (Atomic_Int)(new))
#define _CPU_Atomic_Compare_exchange_rel_32(dst, old, new) \
_CPU_Atomic_Compare_exchange_rel_int((volatile Atomic_Int *)(dst), (Atomic_Int)(old), (Atomic_Int)(new))
-#define _CPU_Atomic_Compare_exchange_ptr(dst, old, new) \
+#define _CPU_Atomic_Compare_exchange_ptr(dst, old, new) \
_CPU_Atomic_Compare_exchange_int((volatile Atomic_Int *)(dst), (Atomic_Int)(old), (Atomic_Int)(new))
#define _CPU_Atomic_Compare_exchange_acq_ptr(dst, old, new) \
_CPU_Atomic_Compare_exchange_acq_int((volatile Atomic_Int *)(dst), (Atomic_Int)(old), \
diff --git a/cpukit/score/inline/rtems/score/atomic.inl b/cpukit/score/inline/rtems/score/atomic.inl
index 7deec01729..313366c79f 100644
--- a/cpukit/score/inline/rtems/score/atomic.inl
+++ b/cpukit/score/inline/rtems/score/atomic.inl
@@ -6,7 +6,9 @@
* The license and distribution terms for this file may be
* found in the file LICENSE in this distribution or at
* http://www.rtems.com/license/LICENSE.
- *
+ */
+
+/*
*
* The functions in this file implement the API to the RTEMS Atomic Manager and
* The API is designed to be compatable with C1X atomic definition as far as
@@ -16,8 +18,8 @@
*
* rtems/cpukit/score/cpu/xxx/rtems/score/cpuatomic.h
*
- * In the event that a CPU does not support a specific atomic function it has, the
- * CPU dependent routine does nothing (but does exist).
+ * In the event that a CPU does not support a specific atomic function it has,
+ * the CPU dependent routine does nothing (but does exist).
*/
#ifndef _RTEMS_SCORE_ATOMIC_H
@@ -36,8 +38,7 @@ RTEMS_INLINE_ROUTINE Atomic_Int _Atomic_Load_int(
{
if(ATOMIC_ACQUIRE_BARRIER == memory_barrier)
return _CPU_Atomic_Load_acq_int(address);
- else
- return _CPU_Atomic_Load_int(address);
+ return _CPU_Atomic_Load_int(address);
}
RTEMS_INLINE_ROUTINE Atomic_Long _Atomic_Load_long(
@@ -45,10 +46,9 @@ RTEMS_INLINE_ROUTINE Atomic_Long _Atomic_Load_long(
Atomic_Memory_barrier memory_barrier
)
{
- if(ATOMIC_ACQUIRE_BARRIER == memory_barrier)
+ if (ATOMIC_ACQUIRE_BARRIER == memory_barrier)
return _CPU_Atomic_Load_acq_long(address);
- else
- return _CPU_Atomic_Load_long(address);
+ return _CPU_Atomic_Load_long(address);
}
RTEMS_INLINE_ROUTINE Atomic_Pointer _Atomic_Load_ptr(
@@ -56,10 +56,9 @@ RTEMS_INLINE_ROUTINE Atomic_Pointer _Atomic_Load_ptr(
Atomic_Memory_barrier memory_barrier
)
{
- if(ATOMIC_ACQUIRE_BARRIER == memory_barrier)
+ if (ATOMIC_ACQUIRE_BARRIER == memory_barrier)
return _CPU_Atomic_Load_acq_ptr(address);
- else
- return _CPU_Atomic_Load_ptr(address);
+ return _CPU_Atomic_Load_ptr(address);
}
RTEMS_INLINE_ROUTINE Atomic_Int32 _Atomic_Load_32(
@@ -67,10 +66,9 @@ RTEMS_INLINE_ROUTINE Atomic_Int32 _Atomic_Load_32(
Atomic_Memory_barrier memory_barrier
)
{
- if(ATOMIC_ACQUIRE_BARRIER == memory_barrier)
+ if (ATOMIC_ACQUIRE_BARRIER == memory_barrier)
return _CPU_Atomic_Load_acq_32(address);
- else
- return _CPU_Atomic_Load_32(address);
+ return _CPU_Atomic_Load_32(address);
}
RTEMS_INLINE_ROUTINE Atomic_Int64 _Atomic_Load_64(
@@ -78,10 +76,9 @@ RTEMS_INLINE_ROUTINE Atomic_Int64 _Atomic_Load_64(
Atomic_Memory_barrier memory_barrier
)
{
- if(ATOMIC_ACQUIRE_BARRIER == memory_barrier)
+ if (ATOMIC_ACQUIRE_BARRIER == memory_barrier)
return _CPU_Atomic_Load_acq_64(address);
- else
- return _CPU_Atomic_Load_64(address);
+ return _CPU_Atomic_Load_64(address);
}
@@ -91,10 +88,9 @@ RTEMS_INLINE_ROUTINE void _Atomic_Store_int(
Atomic_Memory_barrier memory_barrier
)
{
- if(ATOMIC_RELEASE_BARRIER == memory_barrier)
+ if (ATOMIC_RELEASE_BARRIER == memory_barrier)
return _CPU_Atomic_Store_rel_int(address, value);
- else
- return _CPU_Atomic_Store_int(address, value);
+ return _CPU_Atomic_Store_int(address, value);
}
RTEMS_INLINE_ROUTINE void _Atomic_Store_long(
@@ -103,10 +99,9 @@ RTEMS_INLINE_ROUTINE void _Atomic_Store_long(
Atomic_Memory_barrier memory_barrier
)
{
- if(ATOMIC_RELEASE_BARRIER == memory_barrier)
+ if (ATOMIC_RELEASE_BARRIER == memory_barrier)
return _CPU_Atomic_Store_rel_long(address, value);
- else
- return _CPU_Atomic_Store_long(address, value);
+ return _CPU_Atomic_Store_long(address, value);
}
RTEMS_INLINE_ROUTINE void _Atomic_Store_ptr(
@@ -115,10 +110,9 @@ RTEMS_INLINE_ROUTINE void _Atomic_Store_ptr(
Atomic_Memory_barrier memory_barrier
)
{
- if(ATOMIC_RELEASE_BARRIER == memory_barrier)
+ if (ATOMIC_RELEASE_BARRIER == memory_barrier)
return _CPU_Atomic_Store_rel_ptr(address, value);
- else
- return _CPU_Atomic_Store_ptr(address, value);
+ return _CPU_Atomic_Store_ptr(address, value);
}
RTEMS_INLINE_ROUTINE void _Atomic_Store_32(
@@ -127,10 +121,9 @@ RTEMS_INLINE_ROUTINE void _Atomic_Store_32(
Atomic_Memory_barrier memory_barrier
)
{
- if(ATOMIC_RELEASE_BARRIER == memory_barrier)
+ if (ATOMIC_RELEASE_BARRIER == memory_barrier)
return _CPU_Atomic_Store_rel_32(address, value);
- else
- return _CPU_Atomic_Store_32(address, value);
+ return _CPU_Atomic_Store_32(address, value);
}
RTEMS_INLINE_ROUTINE void _Atomic_Store_64(
@@ -139,10 +132,9 @@ RTEMS_INLINE_ROUTINE void _Atomic_Store_64(
Atomic_Memory_barrier memory_barrier
)
{
- if(ATOMIC_RELEASE_BARRIER == memory_barrier)
+ if (ATOMIC_RELEASE_BARRIER == memory_barrier)
return _CPU_Atomic_Store_rel_64(address, value);
- else
- return _CPU_Atomic_Store_64(address, value);
+ return _CPU_Atomic_Store_64(address, value);
}
RTEMS_INLINE_ROUTINE void _Atomic_Fetch_add_int(
@@ -151,9 +143,9 @@ RTEMS_INLINE_ROUTINE void _Atomic_Fetch_add_int(
Atomic_Memory_barrier memory_barrier
)
{
- if(ATOMIC_ACQUIRE_BARRIER == memory_barrier)
+ if (ATOMIC_ACQUIRE_BARRIER == memory_barrier)
return _CPU_Atomic_Fetch_add_acq_int(address, value);
- else if(ATOMIC_RELEASE_BARRIER == memory_barrier)
+ else if (ATOMIC_RELEASE_BARRIER == memory_barrier)
return _CPU_Atomic_Fetch_add_rel_int(address, value);
else
return _CPU_Atomic_Fetch_add_int(address, value);
@@ -165,9 +157,9 @@ RTEMS_INLINE_ROUTINE void _Atomic_Fetch_add_long(
Atomic_Memory_barrier memory_barrier
)
{
- if(ATOMIC_ACQUIRE_BARRIER == memory_barrier)
+ if (ATOMIC_ACQUIRE_BARRIER == memory_barrier)
return _CPU_Atomic_Fetch_add_acq_long(address, value);
- else if(ATOMIC_RELEASE_BARRIER == memory_barrier)
+ else if (ATOMIC_RELEASE_BARRIER == memory_barrier)
return _CPU_Atomic_Fetch_add_rel_long(address, value);
else
return _CPU_Atomic_Fetch_add_long(address, value);
@@ -179,9 +171,9 @@ RTEMS_INLINE_ROUTINE void _Atomic_Fetch_add_ptr(
Atomic_Memory_barrier memory_barrier
)
{
- if(ATOMIC_ACQUIRE_BARRIER == memory_barrier)
+ if (ATOMIC_ACQUIRE_BARRIER == memory_barrier)
return _CPU_Atomic_Fetch_add_acq_ptr(address, value);
- else if(ATOMIC_RELEASE_BARRIER == memory_barrier)
+ else if (ATOMIC_RELEASE_BARRIER == memory_barrier)
return _CPU_Atomic_Fetch_add_rel_ptr(address, value);
else
return _CPU_Atomic_Fetch_add_ptr(address, value);
@@ -193,9 +185,9 @@ RTEMS_INLINE_ROUTINE void _Atomic_Fetch_add_32(
Atomic_Memory_barrier memory_barrier
)
{
- if(ATOMIC_ACQUIRE_BARRIER == memory_barrier)
+ if (ATOMIC_ACQUIRE_BARRIER == memory_barrier)
return _CPU_Atomic_Fetch_add_acq_32(address, value);
- else if(ATOMIC_RELEASE_BARRIER == memory_barrier)
+ else if (ATOMIC_RELEASE_BARRIER == memory_barrier)
return _CPU_Atomic_Fetch_add_rel_32(address, value);
else
return _CPU_Atomic_Fetch_add_32(address, value);
@@ -207,9 +199,9 @@ RTEMS_INLINE_ROUTINE void _Atomic_Fetch_add_64(
Atomic_Memory_barrier memory_barrier
)
{
- if(ATOMIC_ACQUIRE_BARRIER == memory_barrier)
+ if (ATOMIC_ACQUIRE_BARRIER == memory_barrier)
return _CPU_Atomic_Fetch_add_acq_64(address, value);
- else if(ATOMIC_RELEASE_BARRIER == memory_barrier)
+ else if (ATOMIC_RELEASE_BARRIER == memory_barrier)
return _CPU_Atomic_Fetch_add_rel_64(address, value);
else
return _CPU_Atomic_Fetch_add_64(address, value);
@@ -221,9 +213,9 @@ RTEMS_INLINE_ROUTINE void _Atomic_Fetch_sub_int(
Atomic_Memory_barrier memory_barrier
)
{
- if(ATOMIC_ACQUIRE_BARRIER == memory_barrier)
+ if (ATOMIC_ACQUIRE_BARRIER == memory_barrier)
return _CPU_Atomic_Fetch_sub_acq_int(address, value);
- else if(ATOMIC_RELEASE_BARRIER == memory_barrier)
+ else if (ATOMIC_RELEASE_BARRIER == memory_barrier)
return _CPU_Atomic_Fetch_sub_rel_int(address, value);
else
return _CPU_Atomic_Fetch_sub_int(address, value);
@@ -235,9 +227,9 @@ RTEMS_INLINE_ROUTINE void _Atomic_Fetch_sub_long(
Atomic_Memory_barrier memory_barrier
)
{
- if(ATOMIC_ACQUIRE_BARRIER == memory_barrier)
+ if (ATOMIC_ACQUIRE_BARRIER == memory_barrier)
return _CPU_Atomic_Fetch_sub_acq_long(address, value);
- else if(ATOMIC_RELEASE_BARRIER == memory_barrier)
+ else if (ATOMIC_RELEASE_BARRIER == memory_barrier)
return _CPU_Atomic_Fetch_sub_rel_long(address, value);
else
return _CPU_Atomic_Fetch_sub_long(address, value);
@@ -249,9 +241,9 @@ RTEMS_INLINE_ROUTINE void _Atomic_Fetch_sub_ptr(
Atomic_Memory_barrier memory_barrier
)
{
- if(ATOMIC_ACQUIRE_BARRIER == memory_barrier)
+ if (ATOMIC_ACQUIRE_BARRIER == memory_barrier)
return _CPU_Atomic_Fetch_sub_acq_ptr(address, value);
- else if(ATOMIC_RELEASE_BARRIER == memory_barrier)
+ else if (ATOMIC_RELEASE_BARRIER == memory_barrier)
return _CPU_Atomic_Fetch_sub_rel_ptr(address, value);
else
return _CPU_Atomic_Fetch_sub_ptr(address, value);
@@ -263,9 +255,9 @@ RTEMS_INLINE_ROUTINE void _Atomic_Fetch_sub_32(
Atomic_Memory_barrier memory_barrier
)
{
- if(ATOMIC_ACQUIRE_BARRIER == memory_barrier)
+ if (ATOMIC_ACQUIRE_BARRIER == memory_barrier)
return _CPU_Atomic_Fetch_sub_acq_32(address, value);
- else if(ATOMIC_RELEASE_BARRIER == memory_barrier)
+ else if (ATOMIC_RELEASE_BARRIER == memory_barrier)
return _CPU_Atomic_Fetch_sub_rel_32(address, value);
else
return _CPU_Atomic_Fetch_sub_32(address, value);
@@ -277,9 +269,9 @@ RTEMS_INLINE_ROUTINE void _Atomic_Fetch_sub_64(
Atomic_Memory_barrier memory_barrier
)
{
- if(ATOMIC_ACQUIRE_BARRIER == memory_barrier)
+ if (ATOMIC_ACQUIRE_BARRIER == memory_barrier)
return _CPU_Atomic_Fetch_sub_acq_64(address, value);
- else if(ATOMIC_RELEASE_BARRIER == memory_barrier)
+ else if (ATOMIC_RELEASE_BARRIER == memory_barrier)
return _CPU_Atomic_Fetch_sub_rel_64(address, value);
else
return _CPU_Atomic_Fetch_sub_64(address, value);
@@ -291,9 +283,9 @@ RTEMS_INLINE_ROUTINE void _Atomic_Fetch_or_int(
Atomic_Memory_barrier memory_barrier
)
{
- if(ATOMIC_ACQUIRE_BARRIER == memory_barrier)
+ if (ATOMIC_ACQUIRE_BARRIER == memory_barrier)
return _CPU_Atomic_Fetch_or_acq_int(address, value);
- else if(ATOMIC_RELEASE_BARRIER == memory_barrier)
+ else if (ATOMIC_RELEASE_BARRIER == memory_barrier)
return _CPU_Atomic_Fetch_or_rel_int(address, value);
else
return _CPU_Atomic_Fetch_or_int(address, value);
@@ -305,9 +297,9 @@ RTEMS_INLINE_ROUTINE void _Atomic_Fetch_or_long(
Atomic_Memory_barrier memory_barrier
)
{
- if(ATOMIC_ACQUIRE_BARRIER == memory_barrier)
+ if (ATOMIC_ACQUIRE_BARRIER == memory_barrier)
return _CPU_Atomic_Fetch_or_acq_long(address, value);
- else if(ATOMIC_RELEASE_BARRIER == memory_barrier)
+ else if (ATOMIC_RELEASE_BARRIER == memory_barrier)
return _CPU_Atomic_Fetch_or_rel_long(address, value);
else
return _CPU_Atomic_Fetch_or_long(address, value);
@@ -319,9 +311,9 @@ RTEMS_INLINE_ROUTINE void _Atomic_Fetch_or_ptr(
Atomic_Memory_barrier memory_barrier
)
{
- if(ATOMIC_ACQUIRE_BARRIER == memory_barrier)
+ if (ATOMIC_ACQUIRE_BARRIER == memory_barrier)
return _CPU_Atomic_Fetch_or_acq_ptr(address, value);
- else if(ATOMIC_RELEASE_BARRIER == memory_barrier)
+ else if (ATOMIC_RELEASE_BARRIER == memory_barrier)
return _CPU_Atomic_Fetch_or_rel_ptr(address, value);
else
return _CPU_Atomic_Fetch_or_ptr(address, value);
@@ -333,9 +325,9 @@ RTEMS_INLINE_ROUTINE void _Atomic_Fetch_or_32(
Atomic_Memory_barrier memory_barrier
)
{
- if(ATOMIC_ACQUIRE_BARRIER == memory_barrier)
+ if (ATOMIC_ACQUIRE_BARRIER == memory_barrier)
return _CPU_Atomic_Fetch_or_acq_32(address, value);
- else if(ATOMIC_RELEASE_BARRIER == memory_barrier)
+ else if (ATOMIC_RELEASE_BARRIER == memory_barrier)
return _CPU_Atomic_Fetch_or_rel_32(address, value);
else
return _CPU_Atomic_Fetch_or_32(address, value);
@@ -347,9 +339,9 @@ RTEMS_INLINE_ROUTINE void _Atomic_Fetch_or_64(
Atomic_Memory_barrier memory_barrier
)
{
- if(ATOMIC_ACQUIRE_BARRIER == memory_barrier)
+ if (ATOMIC_ACQUIRE_BARRIER == memory_barrier)
return _CPU_Atomic_Fetch_or_acq_64(address, value);
- else if(ATOMIC_RELEASE_BARRIER == memory_barrier)
+ else if (ATOMIC_RELEASE_BARRIER == memory_barrier)
return _CPU_Atomic_Fetch_or_rel_64(address, value);
else
return _CPU_Atomic_Fetch_or_64(address, value);
@@ -361,9 +353,9 @@ RTEMS_INLINE_ROUTINE void _Atomic_Fetch_and_int(
Atomic_Memory_barrier memory_barrier
)
{
- if(ATOMIC_ACQUIRE_BARRIER == memory_barrier)
+ if (ATOMIC_ACQUIRE_BARRIER == memory_barrier)
return _CPU_Atomic_Fetch_and_acq_int(address, value);
- else if(ATOMIC_RELEASE_BARRIER == memory_barrier)
+ else if (ATOMIC_RELEASE_BARRIER == memory_barrier)
return _CPU_Atomic_Fetch_and_rel_int(address, value);
else
return _CPU_Atomic_Fetch_and_int(address, value);
@@ -375,9 +367,9 @@ RTEMS_INLINE_ROUTINE void _Atomic_Fetch_and_long(
Atomic_Memory_barrier memory_barrier
)
{
- if(ATOMIC_ACQUIRE_BARRIER == memory_barrier)
+ if (ATOMIC_ACQUIRE_BARRIER == memory_barrier)
return _CPU_Atomic_Fetch_and_acq_long(address, value);
- else if(ATOMIC_RELEASE_BARRIER == memory_barrier)
+ else if (ATOMIC_RELEASE_BARRIER == memory_barrier)
return _CPU_Atomic_Fetch_and_rel_long(address, value);
else
return _CPU_Atomic_Fetch_and_long(address, value);
@@ -389,9 +381,9 @@ RTEMS_INLINE_ROUTINE void _Atomic_Fetch_and_ptr(
Atomic_Memory_barrier memory_barrier
)
{
- if(ATOMIC_ACQUIRE_BARRIER == memory_barrier)
+ if (ATOMIC_ACQUIRE_BARRIER == memory_barrier)
return _CPU_Atomic_Fetch_and_acq_ptr(address, value);
- else if(ATOMIC_RELEASE_BARRIER == memory_barrier)
+ else if (ATOMIC_RELEASE_BARRIER == memory_barrier)
return _CPU_Atomic_Fetch_and_rel_ptr(address, value);
else
return _CPU_Atomic_Fetch_and_ptr(address, value);
@@ -403,9 +395,9 @@ RTEMS_INLINE_ROUTINE void _Atomic_Fetch_and_32(
Atomic_Memory_barrier memory_barrier
)
{
- if(ATOMIC_ACQUIRE_BARRIER == memory_barrier)
+ if (ATOMIC_ACQUIRE_BARRIER == memory_barrier)
return _CPU_Atomic_Fetch_and_acq_32(address, value);
- else if(ATOMIC_RELEASE_BARRIER == memory_barrier)
+ else if (ATOMIC_RELEASE_BARRIER == memory_barrier)
return _CPU_Atomic_Fetch_and_rel_32(address, value);
else
return _CPU_Atomic_Fetch_and_32(address, value);
@@ -417,9 +409,9 @@ RTEMS_INLINE_ROUTINE void _Atomic_Fetch_and_64(
Atomic_Memory_barrier memory_barrier
)
{
- if(ATOMIC_ACQUIRE_BARRIER == memory_barrier)
+ if (ATOMIC_ACQUIRE_BARRIER == memory_barrier)
return _CPU_Atomic_Fetch_and_acq_64(address, value);
- else if(ATOMIC_RELEASE_BARRIER == memory_barrier)
+ else if (ATOMIC_RELEASE_BARRIER == memory_barrier)
return _CPU_Atomic_Fetch_and_rel_64(address, value);
else
return _CPU_Atomic_Fetch_and_64(address, value);
@@ -432,9 +424,9 @@ RTEMS_INLINE_ROUTINE int _Atomic_Compare_exchange_int(
Atomic_Memory_barrier memory_barrier
)
{
- if(ATOMIC_ACQUIRE_BARRIER == memory_barrier)
+ if (ATOMIC_ACQUIRE_BARRIER == memory_barrier)
return _CPU_Atomic_Compare_exchange_acq_int(address, old_value, new_value);
- else if(ATOMIC_RELEASE_BARRIER == memory_barrier)
+ else if (ATOMIC_RELEASE_BARRIER == memory_barrier)
return _CPU_Atomic_Compare_exchange_rel_int(address, old_value, new_value);
else
return _CPU_Atomic_Compare_exchange_int(address, old_value, new_value);
@@ -447,9 +439,9 @@ RTEMS_INLINE_ROUTINE int _Atomic_Compare_exchange_long(
Atomic_Memory_barrier memory_barrier
)
{
- if(ATOMIC_ACQUIRE_BARRIER == memory_barrier)
+ if (ATOMIC_ACQUIRE_BARRIER == memory_barrier)
return _CPU_Atomic_Compare_exchange_acq_long(address, old_value, new_value);
- else if(ATOMIC_RELEASE_BARRIER == memory_barrier)
+ else if (ATOMIC_RELEASE_BARRIER == memory_barrier)
return _CPU_Atomic_Compare_exchange_rel_long(address, old_value, new_value);
else
return _CPU_Atomic_Compare_exchange_long(address, old_value, new_value);
@@ -462,9 +454,9 @@ RTEMS_INLINE_ROUTINE int _Atomic_Compare_exchange_ptr(
Atomic_Memory_barrier memory_barrier
)
{
- if(ATOMIC_ACQUIRE_BARRIER == memory_barrier)
+ if (ATOMIC_ACQUIRE_BARRIER == memory_barrier)
return _CPU_Atomic_Compare_exchange_acq_ptr(address, old_value, new_value);
- else if(ATOMIC_RELEASE_BARRIER == memory_barrier)
+ else if (ATOMIC_RELEASE_BARRIER == memory_barrier)
return _CPU_Atomic_Compare_exchange_rel_ptr(address, old_value, new_value);
else
return _CPU_Atomic_Compare_exchange_ptr(address, old_value, new_value);
@@ -477,9 +469,9 @@ RTEMS_INLINE_ROUTINE int _Atomic_Compare_exchange_32(
Atomic_Memory_barrier memory_barrier
)
{
- if(ATOMIC_ACQUIRE_BARRIER == memory_barrier)
+ if (ATOMIC_ACQUIRE_BARRIER == memory_barrier)
return _CPU_Atomic_Compare_exchange_acq_32(address, old_value, new_value);
- else if(ATOMIC_RELEASE_BARRIER == memory_barrier)
+ else if (ATOMIC_RELEASE_BARRIER == memory_barrier)
return _CPU_Atomic_Compare_exchange_rel_32(address, old_value, new_value);
else
return _CPU_Atomic_Compare_exchange_32(address, old_value, new_value);
@@ -492,9 +484,9 @@ RTEMS_INLINE_ROUTINE int _Atomic_Compare_exchange_64(
Atomic_Memory_barrier memory_barrier
)
{
- if(ATOMIC_ACQUIRE_BARRIER == memory_barrier)
+ if (ATOMIC_ACQUIRE_BARRIER == memory_barrier)
return _CPU_Atomic_Compare_exchange_acq_64(address, old_value, new_value);
- else if(ATOMIC_RELEASE_BARRIER == memory_barrier)
+ else if (ATOMIC_RELEASE_BARRIER == memory_barrier)
return _CPU_Atomic_Compare_exchange_rel_64(address, old_value, new_value);
else
return _CPU_Atomic_Compare_exchange_64(address, old_value, new_value);