atomic.h revision 1.11 1 1.8 riastrad /* $NetBSD: atomic.h,v 1.11 2018/08/27 13:40:41 riastradh Exp $ */
2 1.2 riastrad
3 1.2 riastrad /*-
4 1.2 riastrad * Copyright (c) 2013 The NetBSD Foundation, Inc.
5 1.2 riastrad * All rights reserved.
6 1.2 riastrad *
7 1.2 riastrad * This code is derived from software contributed to The NetBSD Foundation
8 1.2 riastrad * by Taylor R. Campbell.
9 1.2 riastrad *
10 1.2 riastrad * Redistribution and use in source and binary forms, with or without
11 1.2 riastrad * modification, are permitted provided that the following conditions
12 1.2 riastrad * are met:
13 1.2 riastrad * 1. Redistributions of source code must retain the above copyright
14 1.2 riastrad * notice, this list of conditions and the following disclaimer.
15 1.2 riastrad * 2. Redistributions in binary form must reproduce the above copyright
16 1.2 riastrad * notice, this list of conditions and the following disclaimer in the
17 1.2 riastrad * documentation and/or other materials provided with the distribution.
18 1.2 riastrad *
19 1.2 riastrad * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
20 1.2 riastrad * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
21 1.2 riastrad * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
22 1.2 riastrad * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
23 1.2 riastrad * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
24 1.2 riastrad * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
25 1.2 riastrad * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
26 1.2 riastrad * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
27 1.2 riastrad * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
28 1.2 riastrad * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
29 1.2 riastrad * POSSIBILITY OF SUCH DAMAGE.
30 1.2 riastrad */
31 1.2 riastrad
32 1.2 riastrad #ifndef _LINUX_ATOMIC_H_
33 1.2 riastrad #define _LINUX_ATOMIC_H_
34 1.2 riastrad
35 1.2 riastrad #include <sys/atomic.h>
36 1.2 riastrad
37 1.2 riastrad #include <machine/limits.h>
38 1.2 riastrad
39 1.2 riastrad struct atomic {
40 1.2 riastrad union {
41 1.3 riastrad volatile int au_int;
42 1.3 riastrad volatile unsigned int au_uint;
43 1.2 riastrad } a_u;
44 1.2 riastrad };
45 1.2 riastrad
46 1.2 riastrad #define ATOMIC_INIT(i) { .a_u = { .au_int = (i) } }
47 1.2 riastrad
48 1.2 riastrad typedef struct atomic atomic_t;
49 1.2 riastrad
50 1.2 riastrad static inline int
51 1.2 riastrad atomic_read(atomic_t *atomic)
52 1.2 riastrad {
53 1.3 riastrad return atomic->a_u.au_int;
54 1.2 riastrad }
55 1.2 riastrad
56 1.2 riastrad static inline void
57 1.2 riastrad atomic_set(atomic_t *atomic, int value)
58 1.2 riastrad {
59 1.2 riastrad atomic->a_u.au_int = value;
60 1.2 riastrad }
61 1.2 riastrad
62 1.2 riastrad static inline void
63 1.2 riastrad atomic_add(int addend, atomic_t *atomic)
64 1.2 riastrad {
65 1.2 riastrad atomic_add_int(&atomic->a_u.au_uint, addend);
66 1.2 riastrad }
67 1.2 riastrad
68 1.2 riastrad static inline void
69 1.2 riastrad atomic_sub(int subtrahend, atomic_t *atomic)
70 1.2 riastrad {
71 1.2 riastrad atomic_add_int(&atomic->a_u.au_uint, -subtrahend);
72 1.2 riastrad }
73 1.2 riastrad
74 1.2 riastrad static inline int
75 1.2 riastrad atomic_add_return(int addend, atomic_t *atomic)
76 1.2 riastrad {
77 1.2 riastrad return (int)atomic_add_int_nv(&atomic->a_u.au_uint, addend);
78 1.2 riastrad }
79 1.2 riastrad
80 1.2 riastrad static inline void
81 1.2 riastrad atomic_inc(atomic_t *atomic)
82 1.2 riastrad {
83 1.2 riastrad atomic_inc_uint(&atomic->a_u.au_uint);
84 1.2 riastrad }
85 1.2 riastrad
86 1.2 riastrad static inline void
87 1.2 riastrad atomic_dec(atomic_t *atomic)
88 1.2 riastrad {
89 1.2 riastrad atomic_dec_uint(&atomic->a_u.au_uint);
90 1.2 riastrad }
91 1.2 riastrad
92 1.2 riastrad static inline int
93 1.2 riastrad atomic_inc_return(atomic_t *atomic)
94 1.2 riastrad {
95 1.2 riastrad return (int)atomic_inc_uint_nv(&atomic->a_u.au_uint);
96 1.2 riastrad }
97 1.2 riastrad
98 1.2 riastrad static inline int
99 1.2 riastrad atomic_dec_return(atomic_t *atomic)
100 1.2 riastrad {
101 1.2 riastrad return (int)atomic_dec_uint_nv(&atomic->a_u.au_uint);
102 1.2 riastrad }
103 1.2 riastrad
104 1.2 riastrad static inline int
105 1.2 riastrad atomic_dec_and_test(atomic_t *atomic)
106 1.2 riastrad {
107 1.4 riastrad return (0 == (int)atomic_dec_uint_nv(&atomic->a_u.au_uint));
108 1.2 riastrad }
109 1.2 riastrad
110 1.2 riastrad static inline void
111 1.8 riastrad atomic_or(int value, atomic_t *atomic)
112 1.8 riastrad {
113 1.8 riastrad atomic_or_uint(&atomic->a_u.au_uint, value);
114 1.8 riastrad }
115 1.8 riastrad
116 1.8 riastrad static inline void
117 1.2 riastrad atomic_set_mask(unsigned long mask, atomic_t *atomic)
118 1.2 riastrad {
119 1.2 riastrad atomic_or_uint(&atomic->a_u.au_uint, mask);
120 1.2 riastrad }
121 1.2 riastrad
122 1.2 riastrad static inline void
123 1.2 riastrad atomic_clear_mask(unsigned long mask, atomic_t *atomic)
124 1.2 riastrad {
125 1.2 riastrad atomic_and_uint(&atomic->a_u.au_uint, ~mask);
126 1.2 riastrad }
127 1.2 riastrad
128 1.2 riastrad static inline int
129 1.2 riastrad atomic_add_unless(atomic_t *atomic, int addend, int zero)
130 1.2 riastrad {
131 1.2 riastrad int value;
132 1.2 riastrad
133 1.2 riastrad do {
134 1.2 riastrad value = atomic->a_u.au_int;
135 1.2 riastrad if (value == zero)
136 1.2 riastrad return 0;
137 1.2 riastrad } while (atomic_cas_uint(&atomic->a_u.au_uint, value, (value + addend))
138 1.2 riastrad != value);
139 1.2 riastrad
140 1.2 riastrad return 1;
141 1.2 riastrad }
142 1.2 riastrad
143 1.2 riastrad static inline int
144 1.2 riastrad atomic_inc_not_zero(atomic_t *atomic)
145 1.2 riastrad {
146 1.2 riastrad return atomic_add_unless(atomic, 1, 0);
147 1.2 riastrad }
148 1.2 riastrad
149 1.5 riastrad static inline int
150 1.5 riastrad atomic_xchg(atomic_t *atomic, int new)
151 1.5 riastrad {
152 1.5 riastrad return (int)atomic_swap_uint(&atomic->a_u.au_uint, (unsigned)new);
153 1.5 riastrad }
154 1.5 riastrad
155 1.5 riastrad static inline int
156 1.5 riastrad atomic_cmpxchg(atomic_t *atomic, int old, int new)
157 1.5 riastrad {
158 1.5 riastrad return (int)atomic_cas_uint(&atomic->a_u.au_uint, (unsigned)old,
159 1.5 riastrad (unsigned)new);
160 1.5 riastrad }
161 1.5 riastrad
162 1.6 riastrad struct atomic64 {
163 1.6 riastrad volatile uint64_t a_v;
164 1.6 riastrad };
165 1.6 riastrad
166 1.6 riastrad typedef struct atomic64 atomic64_t;
167 1.6 riastrad
168 1.6 riastrad static inline uint64_t
169 1.6 riastrad atomic64_read(const struct atomic64 *a)
170 1.6 riastrad {
171 1.6 riastrad return a->a_v;
172 1.6 riastrad }
173 1.6 riastrad
174 1.6 riastrad static inline void
175 1.6 riastrad atomic64_set(struct atomic64 *a, uint64_t v)
176 1.6 riastrad {
177 1.6 riastrad a->a_v = v;
178 1.6 riastrad }
179 1.6 riastrad
180 1.6 riastrad static inline void
181 1.6 riastrad atomic64_add(long long d, struct atomic64 *a)
182 1.6 riastrad {
183 1.6 riastrad atomic_add_64(&a->a_v, d);
184 1.6 riastrad }
185 1.6 riastrad
186 1.6 riastrad static inline void
187 1.6 riastrad atomic64_sub(long long d, struct atomic64 *a)
188 1.6 riastrad {
189 1.6 riastrad atomic_add_64(&a->a_v, -d);
190 1.6 riastrad }
191 1.6 riastrad
192 1.6 riastrad static inline uint64_t
193 1.6 riastrad atomic64_xchg(struct atomic64 *a, uint64_t v)
194 1.6 riastrad {
195 1.6 riastrad return atomic_swap_64(&a->a_v, v);
196 1.6 riastrad }
197 1.6 riastrad
198 1.9 riastrad static inline uint64_t
199 1.9 riastrad atomic64_cmpxchg(struct atomic64 *atomic, uint64_t old, uint64_t new)
200 1.9 riastrad {
201 1.9 riastrad return atomic_cas_64(&atomic->a_v, old, new);
202 1.9 riastrad }
203 1.9 riastrad
204 1.2 riastrad static inline void
205 1.2 riastrad set_bit(unsigned int bit, volatile unsigned long *ptr)
206 1.2 riastrad {
207 1.2 riastrad const unsigned int units = (sizeof(*ptr) * CHAR_BIT);
208 1.2 riastrad
209 1.2 riastrad atomic_or_ulong(&ptr[bit / units], (1UL << (bit % units)));
210 1.2 riastrad }
211 1.2 riastrad
212 1.2 riastrad static inline void
213 1.2 riastrad clear_bit(unsigned int bit, volatile unsigned long *ptr)
214 1.2 riastrad {
215 1.2 riastrad const unsigned int units = (sizeof(*ptr) * CHAR_BIT);
216 1.2 riastrad
217 1.2 riastrad atomic_and_ulong(&ptr[bit / units], ~(1UL << (bit % units)));
218 1.2 riastrad }
219 1.2 riastrad
220 1.2 riastrad static inline void
221 1.2 riastrad change_bit(unsigned int bit, volatile unsigned long *ptr)
222 1.2 riastrad {
223 1.2 riastrad const unsigned int units = (sizeof(*ptr) * CHAR_BIT);
224 1.2 riastrad volatile unsigned long *const p = &ptr[bit / units];
225 1.2 riastrad const unsigned long mask = (1UL << (bit % units));
226 1.2 riastrad unsigned long v;
227 1.2 riastrad
228 1.2 riastrad do v = *p; while (atomic_cas_ulong(p, v, (v ^ mask)) != v);
229 1.2 riastrad }
230 1.2 riastrad
231 1.11 riastrad static inline int
232 1.2 riastrad test_and_set_bit(unsigned int bit, volatile unsigned long *ptr)
233 1.2 riastrad {
234 1.2 riastrad const unsigned int units = (sizeof(*ptr) * CHAR_BIT);
235 1.2 riastrad volatile unsigned long *const p = &ptr[bit / units];
236 1.2 riastrad const unsigned long mask = (1UL << (bit % units));
237 1.2 riastrad unsigned long v;
238 1.2 riastrad
239 1.2 riastrad do v = *p; while (atomic_cas_ulong(p, v, (v | mask)) != v);
240 1.2 riastrad
241 1.7 riastrad return ((v & mask) != 0);
242 1.2 riastrad }
243 1.2 riastrad
244 1.11 riastrad static inline int
245 1.2 riastrad test_and_clear_bit(unsigned int bit, volatile unsigned long *ptr)
246 1.2 riastrad {
247 1.2 riastrad const unsigned int units = (sizeof(*ptr) * CHAR_BIT);
248 1.2 riastrad volatile unsigned long *const p = &ptr[bit / units];
249 1.2 riastrad const unsigned long mask = (1UL << (bit % units));
250 1.2 riastrad unsigned long v;
251 1.2 riastrad
252 1.2 riastrad do v = *p; while (atomic_cas_ulong(p, v, (v & ~mask)) != v);
253 1.2 riastrad
254 1.7 riastrad return ((v & mask) != 0);
255 1.2 riastrad }
256 1.2 riastrad
257 1.11 riastrad static inline int
258 1.2 riastrad test_and_change_bit(unsigned int bit, volatile unsigned long *ptr)
259 1.2 riastrad {
260 1.2 riastrad const unsigned int units = (sizeof(*ptr) * CHAR_BIT);
261 1.2 riastrad volatile unsigned long *const p = &ptr[bit / units];
262 1.2 riastrad const unsigned long mask = (1UL << (bit % units));
263 1.2 riastrad unsigned long v;
264 1.2 riastrad
265 1.2 riastrad do v = *p; while (atomic_cas_ulong(p, v, (v ^ mask)) != v);
266 1.2 riastrad
267 1.7 riastrad return ((v & mask) != 0);
268 1.2 riastrad }
269 1.2 riastrad
270 1.2 riastrad #if defined(MULTIPROCESSOR) && !defined(__HAVE_ATOMIC_AS_MEMBAR)
271 1.2 riastrad /*
272 1.2 riastrad * XXX These memory barriers are doubtless overkill, but I am having
273 1.2 riastrad * trouble understanding the intent and use of the Linux atomic membar
274 1.2 riastrad * API. I think that for reference counting purposes, the sequences
275 1.2 riastrad * should be insn/inc/enter and exit/dec/insn, but the use of the
276 1.2 riastrad * before/after memory barriers is not consistent throughout Linux.
277 1.2 riastrad */
278 1.10 riastrad # define smp_mb__before_atomic() membar_sync()
279 1.2 riastrad # define smp_mb__before_atomic_inc() membar_sync()
280 1.2 riastrad # define smp_mb__after_atomic_inc() membar_sync()
281 1.2 riastrad # define smp_mb__before_atomic_dec() membar_sync()
282 1.2 riastrad # define smp_mb__after_atomic_dec() membar_sync()
283 1.2 riastrad #else
284 1.10 riastrad # define smp_mb__before_atomic() __insn_barrier()
285 1.2 riastrad # define smp_mb__before_atomic_inc() __insn_barrier()
286 1.2 riastrad # define smp_mb__after_atomic_inc() __insn_barrier()
287 1.2 riastrad # define smp_mb__before_atomic_dec() __insn_barrier()
288 1.2 riastrad # define smp_mb__after_atomic_dec() __insn_barrier()
289 1.2 riastrad #endif
290 1.2 riastrad
291 1.2 riastrad #endif /* _LINUX_ATOMIC_H_ */
292