diff options
author | WeiY <wei.a.yang@gmail.com> | 2013-07-15 23:31:10 +0800 |
---|---|---|
committer | Sebastian Huber <sebastian.huber@embedded-brains.de> | 2013-07-17 13:07:33 +0200 |
commit | d61e54c2e0a475d4544f16b947edce2e07edbe35 (patch) | |
tree | 824b3b711ee99fefd1918ebd0f5def5676518c31 /cpukit/score/cpu/i386/rtems/score/cpuatomic.h | |
parent | A generic atomic implementation for smp architectures (diff) | |
download | rtems-d61e54c2e0a475d4544f16b947edce2e07edbe35.tar.bz2 |
update-all-architectures-to-new-atomic-implementation
Diffstat (limited to 'cpukit/score/cpu/i386/rtems/score/cpuatomic.h')
-rw-r--r-- | cpukit/score/cpu/i386/rtems/score/cpuatomic.h | 319 |
1 files changed, 8 insertions, 311 deletions
diff --git a/cpukit/score/cpu/i386/rtems/score/cpuatomic.h b/cpukit/score/cpu/i386/rtems/score/cpuatomic.h index eff6a35534..1656a8f49b 100644 --- a/cpukit/score/cpu/i386/rtems/score/cpuatomic.h +++ b/cpukit/score/cpu/i386/rtems/score/cpuatomic.h @@ -1,46 +1,17 @@ /** * @file rtems/score/cpuatomic.h - * - * This include file implements the atomic operations for i386 and defines + * + * This include file implements the atomic operations for i386 and defines * atomic data types which are used by the atomic operations API file. This * file should use fixed name cpuatomic.h and should be included in atomic - * operations API file atomic.h. Most of the parts of implementations are - * imported from FreeBSD kernel. - */ - -/* - * Copyright (c) 1998 Doug Rabson - * All rights reserved. - * - * Redistribution and use in source and binary forms, with or without - * modification, are permitted provided that the following conditions - * are met: - * 1. Redistributions of source code must retain the above copyright - * notice, this list of conditions and the following disclaimer. - * 2. Redistributions in binary form must reproduce the above copyright - * notice, this list of conditions and the following disclaimer in the - * documentation and/or other materials provided with the distribution. - * - * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND - * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE - * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE - * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE - * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL - * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS - * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) - * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT - * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY - * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF - * SUCH DAMAGE. - * - * $FreeBSD$ + * operations API file atomic.h. If the architecture works at the UP mode it + * will not define atomic ops. If the architecture works at SMP mode, most of + * the parts of implementations are based on stdatomic.h. */ #ifndef _RTEMS_SCORE_ATOMIC_CPU_H #define _RTEMS_SCORE_ATOMIC_CPU_H -#include <rtems/score/genericcpuatomic.h> - #ifdef __cplusplus extern "C" { #endif @@ -52,286 +23,12 @@ extern "C" { /**@{*/ -#if defined(RTEMS_SMP) -#define MPLOCKED "lock ; " +#if !defined(RTEMS_SMP) +# error "Now atomic implementation only supports SMP mode." #else -#define MPLOCKED +#include <rtems/score/cpustdatomic.h> #endif -#if !defined(RTEMS_SMP) -/* - * We assume that a = b will do atomic loads and stores. However, on a - * PentiumPro or higher, reads may pass writes, so for that case we have - * to use a serializing instruction (i.e. with LOCK) to do the load in - * SMP kernels. For UP kernels, however, the cache of the single processor - * is always consistent, so we only need to take care of compiler. - */ -#define ATOMIC_STORE_LOAD(NAME, TYPE, LOP, SOP) \ -static inline Atomic_##TYPE \ -_CPU_Atomic_Load_##NAME(volatile Atomic_##TYPE *p) \ -{ \ - Atomic_##TYPE tmp; \ - \ - tmp = *p; \ - __asm __volatile("" : : : "memory"); \ - return (tmp); \ -} \ - \ -static inline _CPU_Atomic_Load_acq_##NAME(volatile Atomic_##TYPE *p) \ -{ \ - Atomic_##TYPE tmp; \ - \ - tmp = *p; \ - __asm __volatile("" : : : "memory"); \ - return (tmp); \ -} \ - \ -static inline void \ -_CPU_Atomic_Store_##NAME(volatile Atomic_##TYPE *p, Atomic_##TYPE v) \ -{ \ - __asm __volatile("" : : : "memory"); \ - *p = v; \ -} \ - \ -static inline void \ -_CPU_Atomic_Store_rel_##NAME(volatile Atomic_##TYPE *p, Atomic_##TYPE v) \ -{ \ - __asm __volatile("" : : : "memory"); \ - *p = v; \ -} \ - -#else /* !(!SMP) */ - -#define ATOMIC_STORE_LOAD(NAME, TYPE, LOP, SOP) \ -static inline Atomic_##TYPE \ -_CPU_Atomic_Load_##NAME(volatile Atomic_##TYPE *p) \ -{ \ - Atomic_##TYPE res; \ - \ - __asm __volatile(MPLOCKED LOP \ - : "=a" (res), /* 0 */ \ - "=m" (*p) /* 1 */ \ - : "m" (*p) /* 2 */ \ - : "memory", "cc"); \ - \ - return (res); \ -} \ - \ -static inline Atomic_##TYPE \ -_CPU_Atomic_Load_acq_##NAME(volatile Atomic_##TYPE *p) \ -{ \ - Atomic_##TYPE res; \ - \ - __asm __volatile(MPLOCKED LOP \ - : "=a" (res), /* 0 */ \ - "=m" (*p) /* 1 */ \ - : "m" (*p) /* 2 */ \ - : "memory", "cc"); \ - \ - return (res); \ -} \ - \ -/* \ - * The XCHG instruction asserts LOCK automagically. \ - */ \ -static inline void \ -_CPU_Atomic_Store_##NAME(volatile Atomic_##TYPE *p, Atomic_##TYPE v) \ -{ \ - __asm __volatile(SOP \ - : "=m" (*p), /* 0 */ \ - "+r" (v) /* 1 */ \ - : "m" (*p) /* 2 */ \ - : "memory"); \ -} \ -static inline void \ -_CPU_Atomic_Store_rel_##NAME(volatile Atomic_##TYPE *p, Atomic_##TYPE v) \ -{ \ - __asm __volatile(SOP \ - : "=m" (*p), /* 0 */ \ - "+r" (v) /* 1 */ \ - : "m" (*p) /* 2 */ \ - : "memory"); \ -} \ - -#endif /* !SMP */ - -/* - * The assembly is volatilized to avoid code chunk removal by the compiler. - * GCC aggressively reorders operations and memory clobbering is necessary - * in order to avoid that for memory barriers. - */ -#define ATOMIC_FETCH_GENERIC(NAME, TYPENAME, TYPE, OP, CONS, V) \ -static inline void \ -_CPU_Atomic_Fetch_##NAME##_##TYPENAME(volatile Atomic_##TYPE *p, Atomic_##TYPE v) \ -{ \ - __asm __volatile(MPLOCKED OP \ - : "=m" (*p) \ - : CONS (V), "m" (*p) \ - : "cc"); \ -} \ - \ -static inline void \ -_CPU_Atomic_Fetch_##NAME##_barr_##TYPENAME(volatile Atomic_##TYPE *p, Atomic_##TYPE v)\ -{ \ - __asm __volatile(MPLOCKED OP \ - : "=m" (*p) \ - : CONS (V), "m" (*p) \ - : "memory", "cc"); \ -} \ - -/* - * Atomic compare and set, used by the mutex functions - * - * if (*dst == expect) *dst = src (all 32 bit words) - * - * Returns 0 on failure, non-zero on success - */ -static inline int -_CPU_Atomic_Compare_exchange_int(volatile Atomic_Int *dst, Atomic_Int expect, Atomic_Int src) -{ - unsigned char res; - - __asm __volatile( - " " MPLOCKED " " - " cmpxchgl %2,%1 ; " - " sete %0 ; " - "1: " - "# atomic_cmpset_int" - : "=a" (res), /* 0 */ - "=m" (*dst) /* 1 */ - : "r" (src), /* 2 */ - "a" (expect), /* 3 */ - "m" (*dst) /* 4 */ - : "memory", "cc"); - - return (res); -} - -static inline int -_CPU_Atomic_Compare_exchange_long(volatile Atomic_Long *dst, Atomic_Long expect, Atomic_Long src) -{ - - return (_CPU_Atomic_Compare_exchange_int((volatile Atomic_Int *)dst, (Atomic_Int)expect, - (Atomic_Int)src)); -} - -ATOMIC_STORE_LOAD(int, Int, "cmpxchgl %0,%1", "xchgl %1,%0"); -ATOMIC_STORE_LOAD(long, Long, "cmpxchgl %0,%1", "xchgl %1,%0"); - -ATOMIC_FETCH_GENERIC(add, int, Int, "addl %1,%0", "ir", v); -ATOMIC_FETCH_GENERIC(sub, int, Int, "subl %1,%0", "ir", v); -ATOMIC_FETCH_GENERIC(or, int, Int, "orl %1,%0", "ir", v); -ATOMIC_FETCH_GENERIC(and, int, Int, "andl %1,%0", "ir", v); - -ATOMIC_FETCH_GENERIC(add, long, Long, "addl %1,%0", "ir", v); -ATOMIC_FETCH_GENERIC(sub, long, Long, "subl %1,%0", "ir", v); -ATOMIC_FETCH_GENERIC(or, long, Long, "orl %1,%0", "ir", v); -ATOMIC_FETCH_GENERIC(and, long, Long, "andl %1,%0", "ir", v); - -#define _CPU_Atomic_Fetch_or_acq_int _CPU_Atomic_Fetch_or_barr_int -#define _CPU_Atomic_Fetch_or_rel_int _CPU_Atomic_Fetch_or_barr_int -#define _CPU_Atomic_Fetch_and_acq_int _CPU_Atomic_Fetch_and_barr_int -#define _CPU_Atomic_Fetch_and_rel_int _CPU_Atomic_Fetch_and_barr_int -#define _CPU_Atomic_Fetch_add_acq_int _CPU_Atomic_Fetch_add_barr_int -#define _CPU_Atomic_Fetch_add_rel_int _CPU_Atomic_Fetch_add_barr_int -#define _CPU_Atomic_Fetch_sub_acq_int _CPU_Atomic_Fetch_sub_barr_int -#define _CPU_Atomic_Fetch_sub_rel_int _CPU_Atomic_Fetch_sub_barr_int -#define _CPU_Atomic_Compare_exchange_acq_int _CPU_Atomic_Compare_exchange_int -#define _CPU_Atomic_Compare_exchange_rel_int _CPU_Atomic_Compare_exchange_int - -#define _CPU_Atomic_Fetch_or_acq_long _CPU_Atomic_Fetch_or_barr_long -#define _CPU_Atomic_Fetch_or_rel_long _CPU_Atomic_Fetch_or_barr_long -#define _CPU_Atomic_Fetch_and_acq_long _CPU_Atomic_Fetch_and_barr_long -#define _CPU_Atomic_Fetch_and_rel_long _CPU_Atomic_Fetch_and_barr_long -#define _CPU_Atomic_Fetch_add_acq_long _CPU_Atomic_Fetch_add_barr_long -#define _CPU_Atomic_Fetch_add_rel_long _CPU_Atomic_Fetch_add_barr_long -#define _CPU_Atomic_Fetch_sub_acq_long _CPU_Atomic_Fetch_sub_barr_long -#define _CPU_Atomic_Fetch_sub_rel_long _CPU_Atomic_Fetch_sub_barr_long -#define _CPU_Atomic_Compare_exchange_acq_long _CPU_Atomic_Compare_exchange_long -#define _CPU_Atomic_Compare_exchange_rel_long _CPU_Atomic_Compare_exchange_long - -/* Operations on 32-bit double words. */ -#define _CPU_Atomic_Fetch_or_32(p, v) \ - _CPU_Atomic_Fetch_or_int((volatile Atomic_Int *)(p), (Atomic_Int)(v)) -#define _CPU_Atomic_Fetch_or_acq_32(p, v) \ - _CPU_Atomic_Fetch_or_acq_int((volatile Atomic_Int *)(p), (Atomic_Int)(v)) -#define _CPU_Atomic_Fetch_or_rel_32(p, v) \ - _CPU_Atomic_Fetch_or_rel_int((volatile Atomic_Int *)(p), (Atomic_Int)(v)) -#define _CPU_Atomic_Fetch_and_32(p, v) \ - _CPU_Atomic_Fetch_and_int((volatile Atomic_Int *)(p), (Atomic_Int)(v)) -#define _CPU_Atomic_Fetch_and_acq_32(p, v) \ - _CPU_Atomic_Fetch_and_acq_int((volatile Atomic_Int *)(p), (Atomic_Int)(v)) -#define _CPU_Atomic_Fetch_and_rel_32(p, v) \ - _CPU_Atomic_Fetch_and_rel_int((volatile Atomic_Int *)(p), (Atomic_Int)(v)) -#define _CPU_Atomic_Fetch_add_32(p, v) \ - _CPU_Atomic_Fetch_add_int((volatile Atomic_Int *)(p), (Atomic_Int)(v)) -#define _CPU_Atomic_Fetch_add_acq_32(p, v) \ - _CPU_Atomic_Fetch_add_acq_int((volatile Atomic_Int *)(p), (Atomic_Int)(v)) -#define _CPU_Atomic_Fetch_add_rel_32(p, v) \ - _CPU_Atomic_Fetch_add_rel_int((volatile Atomic_Int *)(p), (Atomic_Int)(v)) -#define _CPU_Atomic_Fetch_sub_32(p, v) \ - _CPU_Atomic_Fetch_sub_int((volatile Atomic_Int *)(p), (Atomic_Int)(v)) -#define _CPU_Atomic_Fetch_sub_acq_32(p, v) \ - _CPU_Atomic_Fetch_sub_acq_int((volatile Atomic_Int *)(p), (Atomic_Int)(v)) -#define _CPU_Atomic_Fetch_sub_rel_32(p, v) \ - _CPU_Atomic_Fetch_sub_rel_int((volatile Atomic_Int *)(p), (Atomic_Int)(v)) -#define _CPU_Atomic_Load_32(p) \ - _CPU_Atomic_Load_int((volatile Atomic_Int *)(p)) -#define _CPU_Atomic_Load_acq_32(p) \ - _CPU_Atomic_Load_acq_int((volatile Atomic_Int *)(p)) -#define _CPU_Atomic_Store_32(p, v) \ - _CPU_Atomic_Store_int((volatile Atomic_Int *)(p), (Atomic_Int)(v)) -#define _CPU_Atomic_Store_rel_32(p, v) \ - _CPU_Atomic_Store_rel_int((volatile Atomic_Int *)(p), (Atomic_Int)(v)) -#define _CPU_Atomic_Compare_exchange_32(dst, old, new) \ - _CPU_Atomic_Compare_exchange_int((volatile Atomic_Int *)(dst), (Atomic_Int)(old), (Atomic_Int)(new)) -#define _CPU_Atomic_Compare_exchange_acq_32(dst, old, new) \ - _CPU_Atomic_Compare_exchange_acq_int((volatile Atomic_Int *)(dst), (Atomic_Int)(old), (Atomic_Int)(new)) -#define _CPU_Atomic_Compare_exchange_rel_32(dst, old, new) \ - _CPU_Atomic_Compare_exchange_rel_int((volatile Atomic_Int *)(dst), (Atomic_Int)(old), (Atomic_Int)(new)) - -/* Operations on pointers. */ -#define _CPU_Atomic_Fetch_or_ptr(p, v) \ - _CPU_Atomic_Fetch_or_int((volatile Atomic_Int *)(p), (Atomic_Int)(v)) -#define _CPU_Atomic_Fetch_or_acq_ptr(p, v) \ - _CPU_Atomic_Fetch_or_acq_int((volatile Atomic_Int *)(p), (Atomic_Int)(v)) -#define _CPU_Atomic_Fetch_or_rel_ptr(p, v) \ - _CPU_Atomic_Fetch_or_rel_int((volatile Atomic_Int *)(p), (Atomic_Int)(v)) -#define _CPU_Atomic_Fetch_and_ptr(p, v) \ - _CPU_Atomic_Fetch_and_int((volatile Atomic_Int *)(p), (Atomic_Int)(v)) -#define _CPU_Atomic_Fetch_and_acq_ptr(p, v)\ - _CPU_Atomic_Fetch_and_acq_int((volatile Atomic_Int *)(p), (Atomic_Int)(v)) -#define _CPU_Atomic_Fetch_and_rel_ptr(p, v)\ - _CPU_Atomic_Fetch_and_rel_int((volatile Atomic_Int *)(p), (Atomic_Int)(v)) -#define _CPU_Atomic_Fetch_add_ptr(p, v) \ - _CPU_Atomic_Fetch_add_int((volatile Atomic_Int *)(p), (Atomic_Int)(v)) -#define _CPU_Atomic_Fetch_add_acq_ptr(p, v)\ - _CPU_Atomic_Fetch_add_acq_int((volatile Atomic_Int *)(p), (Atomic_Int)(v)) -#define _CPU_Atomic_Fetch_add_rel_ptr(p, v)\ - _CPU_Atomic_Fetch_add_rel_int((volatile Atomic_Int *)(p), (Atomic_Int)(v)) -#define _CPU_Atomic_Fetch_sub_ptr(p, v) \ - _CPU_Atomic_Fetch_sub_int((volatile Atomic_Int *)(p), (Atomic_Int)(v)) -#define _CPU_Atomic_Fetch_sub_acq_ptr(p, v)\ - _CPU_Atomic_Fetch_sub_acq_int((volatile Atomic_Int *)(p), (Atomic_Int)(v)) -#define _CPU_Atomic_Fetch_sub_rel_ptr(p, v)\ - _CPU_Atomic_Fetch_sub_rel_int((volatile Atomic_Int *)(p), (Atomic_Int)(v)) -#define _CPU_Atomic_Load_ptr(p) \ - _CPU_Atomic_Load_int((volatile Atomic_Int *)(p)) -#define _CPU_Atomic_Load_acq_ptr(p) \ - _CPU_Atomic_Load_acq_int((volatile Atomic_Int *)(p)) -#define _CPU_Atomic_Store_ptr(p, v) \ - _CPU_Atomic_Store_int((volatile Atomic_Int *)(p), (v)) -#define _CPU_Atomic_Store_rel_ptr(p, v) \ - _CPU_Atomic_Store_rel_int((volatile Atomic_Int *)(p), (v)) -#define _CPU_Atomic_Compare_exchange_ptr(dst, old, new) \ - _CPU_Atomic_Compare_exchange_int((volatile Atomic_Int *)(dst), (Atomic_Int)(old), (Atomic_Int)(new)) -#define _CPU_Atomic_Compare_exchange_acq_ptr(dst, old, new) \ - _CPU_Atomic_Compare_exchange_acq_int((volatile Atomic_Int *)(dst), (Atomic_Int)(old), \ - (Atomic_Int)(new)) -#define _CPU_Atomic_Compare_exchange_rel_ptr(dst, old, new) \ - _CPU_Atomic_Compare_exchange_rel_int((volatile Atomic_Int *)(dst), (Atomic_Int)(old), \ - (Atomic_Int)(new)) - #ifdef __cplusplus } #endif |