22
22
* Author: Rickard Green
25
#ifndef ETHR_GCC_ATOMIC_H__
26
#define ETHR_GCC_ATOMIC_H__
28
#if !defined(ETHR_HAVE_NATIVE_ATOMICS) && defined(ETHR_HAVE_GCC_ATOMIC_OPS)
29
#define ETHR_HAVE_NATIVE_ATOMICS 1
31
#define ETHR_IMMED_ATOMIC_SET_GET_SAFE__ 0
32
/* Enable immediate read/write on platforms where we know it is safe */
25
#undef ETHR_INCLUDE_ATOMIC_IMPL__
26
#if !defined(ETHR_GCC_ATOMIC32_H__) && defined(ETHR_ATOMIC_WANT_32BIT_IMPL__)
27
#define ETHR_GCC_ATOMIC32_H__
28
#if defined(ETHR_HAVE___SYNC_VAL_COMPARE_AND_SWAP32)
29
# define ETHR_INCLUDE_ATOMIC_IMPL__ 4
31
#undef ETHR_ATOMIC_WANT_32BIT_IMPL__
32
#elif !defined(ETHR_GCC_ATOMIC64_H__) && defined(ETHR_ATOMIC_WANT_64BIT_IMPL__)
33
#define ETHR_GCC_ATOMIC64_H__
34
#if defined(ETHR_HAVE___SYNC_VAL_COMPARE_AND_SWAP64)
35
# define ETHR_INCLUDE_ATOMIC_IMPL__ 8
37
#undef ETHR_ATOMIC_WANT_64BIT_IMPL__
40
#ifdef ETHR_INCLUDE_ATOMIC_IMPL__
42
#ifndef ETHR_GCC_ATOMIC_COMMON__
43
#define ETHR_GCC_ATOMIC_COMMON__
45
#define ETHR_READ_AND_SET_WITHOUT_SYNC_OP__ 0
33
46
#if defined(__i386__) || defined(__x86_64__) || defined(__sparc__) \
34
47
|| defined(__powerpc__) || defined(__ppc__) || defined(__mips__)
35
# undef ETHR_IMMED_ATOMIC_SET_GET_SAFE__
36
# define ETHR_IMMED_ATOMIC_SET_GET_SAFE__ 1
48
# undef ETHR_READ_AND_SET_WITHOUT_SYNC_OP__
49
# define ETHR_READ_AND_SET_WITHOUT_SYNC_OP__ 1
52
#endif /* ETHR_GCC_ATOMIC_COMMON__ */
54
#if ETHR_INCLUDE_ATOMIC_IMPL__ == 4
55
#define ETHR_HAVE_NATIVE_ATOMIC32 1
56
#define ETHR_NATIVE_ATOMIC32_IMPL "gcc"
57
#define ETHR_NATMC_FUNC__(X) ethr_native_atomic32_ ## X
58
#define ETHR_ATMC_T__ ethr_native_atomic32_t
59
#define ETHR_AINT_T__ ethr_sint32_t
60
#if defined(ETHR_HAVE___SYNC_ADD_AND_FETCH32)
61
# define ETHR_HAVE___SYNC_ADD_AND_FETCH
63
#if defined(ETHR_HAVE___SYNC_FETCH_AND_AND32)
64
# define ETHR_HAVE___SYNC_FETCH_AND_AND
66
#if defined(ETHR_HAVE___SYNC_FETCH_AND_OR32)
67
# define ETHR_HAVE___SYNC_FETCH_AND_OR
69
#elif ETHR_INCLUDE_ATOMIC_IMPL__ == 8
70
#define ETHR_HAVE_NATIVE_ATOMIC64 1
71
#define ETHR_NATIVE_ATOMIC64_IMPL "gcc"
72
#define ETHR_NATMC_FUNC__(X) ethr_native_atomic64_ ## X
73
#define ETHR_ATMC_T__ ethr_native_atomic64_t
74
#define ETHR_AINT_T__ ethr_sint64_t
75
#if defined(ETHR_HAVE___SYNC_ADD_AND_FETCH64)
76
# define ETHR_HAVE___SYNC_ADD_AND_FETCH
78
#if defined(ETHR_HAVE___SYNC_FETCH_AND_AND64)
79
# define ETHR_HAVE___SYNC_FETCH_AND_AND
81
#if defined(ETHR_HAVE___SYNC_FETCH_AND_OR64)
82
# define ETHR_HAVE___SYNC_FETCH_AND_OR
85
#error "Unsupported integer size"
40
volatile long counter;
41
} ethr_native_atomic_t;
45
* According to the documentation this is what we want:
46
* #define ETHR_MEMORY_BARRIER __sync_synchronize()
47
* However, __sync_synchronize() is known to erroneously be
48
* a noop on at least some platforms with some gcc versions.
49
* This has suposedly been fixed in some gcc version, but we
50
* don't know from which version. Therefore, we use the
51
* workaround implemented below on all gcc versions except
52
* for gcc 4.2 or above for MIPS, where it's been verified.
54
#if defined(__mips__) && (__GNUC__ > 4 || (__GNUC__ == 4 && __GNUC_MINOR__ >= 2))
55
#define ETHR_MEMORY_BARRIER __sync_synchronize()
57
#define ETHR_MEMORY_BARRIER \
59
volatile long x___ = 0; \
60
(void) __sync_val_compare_and_swap(&x___, (long) 0, (long) 1); \
63
#define ETHR_READ_DEPEND_MEMORY_BARRIER ETHR_MEMORY_BARRIER
65
#if defined(ETHR_TRY_INLINE_FUNCS) || defined(ETHR_AUX_IMPL__)
89
volatile ETHR_AINT_T__ counter;
93
#if defined(ETHR_TRY_INLINE_FUNCS) || defined(ETHR_ATOMIC_IMPL__)
95
#if ETHR_INCLUDE_ATOMIC_IMPL__ == 4
96
# define ETHR_HAVE_ETHR_NATIVE_ATOMIC32_ADDR 1
98
# define ETHR_HAVE_ETHR_NATIVE_ATOMIC64_ADDR 1
101
static ETHR_INLINE ETHR_AINT_T__ *
102
ETHR_NATMC_FUNC__(addr)(ETHR_ATMC_T__ *var)
104
return (ETHR_AINT_T__ *) &var->counter;
107
#if ETHR_READ_AND_SET_WITHOUT_SYNC_OP__
109
#if ETHR_INCLUDE_ATOMIC_IMPL__ == 4
110
# define ETHR_HAVE_ETHR_NATIVE_ATOMIC32_SET 1
112
# define ETHR_HAVE_ETHR_NATIVE_ATOMIC64_SET 1
67
115
static ETHR_INLINE void
68
ethr_native_atomic_set(ethr_native_atomic_t *var, long value)
116
ETHR_NATMC_FUNC__(set)(ETHR_ATMC_T__ *var, ETHR_AINT_T__ value)
70
#if ETHR_IMMED_ATOMIC_SET_GET_SAFE__
71
118
var->counter = value;
121
#endif /* ETHR_READ_AND_SET_WITHOUT_SYNC_OP__ */
123
#if ETHR_READ_AND_SET_WITHOUT_SYNC_OP__
125
#if ETHR_INCLUDE_ATOMIC_IMPL__ == 4
126
# define ETHR_HAVE_ETHR_NATIVE_ATOMIC32_READ 1
74
* Unfortunately no __sync_store() or similar exist in the gcc atomic
75
* op interface. We therefore have to simulate it this way...
80
act = __sync_val_compare_and_swap(&var->counter, exp, value);
128
# define ETHR_HAVE_ETHR_NATIVE_ATOMIC64_READ 1
85
#define ethr_native_atomic_init ethr_native_atomic_set
87
static ETHR_INLINE long
88
ethr_native_atomic_read(ethr_native_atomic_t *var)
131
static ETHR_INLINE ETHR_AINT_T__
132
ETHR_NATMC_FUNC__(read)(ETHR_ATMC_T__ *var)
90
#if ETHR_IMMED_ATOMIC_SET_GET_SAFE__
91
134
return var->counter;
137
#endif /* ETHR_READ_AND_SET_WITHOUT_SYNC_OP__ */
139
#if defined(ETHR_HAVE___SYNC_ADD_AND_FETCH)
141
#if ETHR_INCLUDE_ATOMIC_IMPL__ == 4
142
# define ETHR_HAVE_ETHR_NATIVE_ATOMIC32_ADD_RETURN_MB 1
94
* Unfortunately no __sync_fetch() or similar exist in the gcc atomic
95
* op interface. We therefore have to simulate it this way...
97
return __sync_add_and_fetch(&var->counter, (long) 0);
144
# define ETHR_HAVE_ETHR_NATIVE_ATOMIC64_ADD_RETURN_MB 1
101
static ETHR_INLINE void
102
ethr_native_atomic_add(ethr_native_atomic_t *var, long incr)
104
(void) __sync_add_and_fetch(&var->counter, incr);
107
static ETHR_INLINE long
108
ethr_native_atomic_add_return(ethr_native_atomic_t *var, long incr)
147
static ETHR_INLINE ETHR_AINT_T__
148
ETHR_NATMC_FUNC__(add_return_mb)(ETHR_ATMC_T__ *var, ETHR_AINT_T__ incr)
110
150
return __sync_add_and_fetch(&var->counter, incr);
113
static ETHR_INLINE void
114
ethr_native_atomic_inc(ethr_native_atomic_t *var)
116
(void) __sync_add_and_fetch(&var->counter, (long) 1);
119
static ETHR_INLINE void
120
ethr_native_atomic_dec(ethr_native_atomic_t *var)
122
(void) __sync_sub_and_fetch(&var->counter, (long) 1);
125
static ETHR_INLINE long
126
ethr_native_atomic_inc_return(ethr_native_atomic_t *var)
128
return __sync_add_and_fetch(&var->counter, (long) 1);
131
static ETHR_INLINE long
132
ethr_native_atomic_dec_return(ethr_native_atomic_t *var)
134
return __sync_sub_and_fetch(&var->counter, (long) 1);
137
static ETHR_INLINE long
138
ethr_native_atomic_and_retold(ethr_native_atomic_t *var, long mask)
155
#if defined(ETHR_HAVE___SYNC_FETCH_AND_AND)
157
#if ETHR_INCLUDE_ATOMIC_IMPL__ == 4
158
# define ETHR_HAVE_ETHR_NATIVE_ATOMIC32_AND_RETOLD_MB 1
160
# define ETHR_HAVE_ETHR_NATIVE_ATOMIC64_AND_RETOLD_MB 1
163
static ETHR_INLINE ETHR_AINT_T__
164
ETHR_NATMC_FUNC__(and_retold_mb)(ETHR_ATMC_T__ *var, ETHR_AINT_T__ mask)
140
166
return __sync_fetch_and_and(&var->counter, mask);
143
static ETHR_INLINE long
144
ethr_native_atomic_or_retold(ethr_native_atomic_t *var, long mask)
171
#if defined(ETHR_HAVE___SYNC_FETCH_AND_OR)
173
#if ETHR_INCLUDE_ATOMIC_IMPL__ == 4
174
# define ETHR_HAVE_ETHR_NATIVE_ATOMIC32_OR_RETOLD_MB 1
176
# define ETHR_HAVE_ETHR_NATIVE_ATOMIC64_OR_RETOLD_MB 1
179
static ETHR_INLINE ETHR_AINT_T__
180
ETHR_NATMC_FUNC__(or_retold_mb)(ETHR_ATMC_T__ *var, ETHR_AINT_T__ mask)
146
return (long) __sync_fetch_and_or(&var->counter, mask);
182
return (ETHR_AINT_T__) __sync_fetch_and_or(&var->counter, mask);
149
static ETHR_INLINE long
150
ethr_native_atomic_cmpxchg(ethr_native_atomic_t *var, long new, long old)
187
#if ETHR_INCLUDE_ATOMIC_IMPL__ == 4
188
# define ETHR_HAVE_ETHR_NATIVE_ATOMIC32_CMPXCHG_MB 1
190
# define ETHR_HAVE_ETHR_NATIVE_ATOMIC64_CMPXCHG_MB 1
193
static ETHR_INLINE ETHR_AINT_T__
194
ETHR_NATMC_FUNC__(cmpxchg_mb)(ETHR_ATMC_T__ *var,
152
198
return __sync_val_compare_and_swap(&var->counter, old, new);
155
static ETHR_INLINE long
156
ethr_native_atomic_xchg(ethr_native_atomic_t *var, long new)
161
act = __sync_val_compare_and_swap(&var->counter, exp, new);
162
} while (act != exp);
167
* Atomic ops with at least specified barriers.
170
static ETHR_INLINE long
171
ethr_native_atomic_read_acqb(ethr_native_atomic_t *var)
173
return __sync_add_and_fetch(&var->counter, (long) 0);
176
#define ethr_native_atomic_inc_return_acqb ethr_native_atomic_inc_return
177
#define ethr_native_atomic_set_relb ethr_native_atomic_xchg
178
#define ethr_native_atomic_dec_relb ethr_native_atomic_dec_return
179
#define ethr_native_atomic_dec_return_relb ethr_native_atomic_dec_return
181
#define ethr_native_atomic_cmpxchg_acqb ethr_native_atomic_cmpxchg
182
#define ethr_native_atomic_cmpxchg_relb ethr_native_atomic_cmpxchg
201
#endif /* ETHR_TRY_INLINE_FUNCS */
203
#undef ETHR_NATMC_FUNC__
206
#undef ETHR_AINT_SUFFIX__
207
#undef ETHR_HAVE___SYNC_ADD_AND_FETCH
208
#undef ETHR_HAVE___SYNC_FETCH_AND_AND
209
#undef ETHR_HAVE___SYNC_FETCH_AND_OR