atomic.h revision 1.1.2.10 1 1.1.2.3 riastrad /* $NetBSD: atomic.h,v 1.1.2.10 2013/09/08 15:37:04 riastradh Exp $ */
2 1.1.2.1 riastrad
3 1.1.2.1 riastrad /*-
4 1.1.2.1 riastrad * Copyright (c) 2013 The NetBSD Foundation, Inc.
5 1.1.2.1 riastrad * All rights reserved.
6 1.1.2.1 riastrad *
7 1.1.2.1 riastrad * This code is derived from software contributed to The NetBSD Foundation
8 1.1.2.1 riastrad * by Taylor R. Campbell.
9 1.1.2.1 riastrad *
10 1.1.2.1 riastrad * Redistribution and use in source and binary forms, with or without
11 1.1.2.1 riastrad * modification, are permitted provided that the following conditions
12 1.1.2.1 riastrad * are met:
13 1.1.2.1 riastrad * 1. Redistributions of source code must retain the above copyright
14 1.1.2.1 riastrad * notice, this list of conditions and the following disclaimer.
15 1.1.2.1 riastrad * 2. Redistributions in binary form must reproduce the above copyright
16 1.1.2.1 riastrad * notice, this list of conditions and the following disclaimer in the
17 1.1.2.1 riastrad * documentation and/or other materials provided with the distribution.
18 1.1.2.1 riastrad *
19 1.1.2.1 riastrad * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
20 1.1.2.1 riastrad * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
21 1.1.2.1 riastrad * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
22 1.1.2.1 riastrad * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
23 1.1.2.1 riastrad * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
24 1.1.2.1 riastrad * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
25 1.1.2.1 riastrad * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
26 1.1.2.1 riastrad * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
27 1.1.2.1 riastrad * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
28 1.1.2.1 riastrad * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
29 1.1.2.1 riastrad * POSSIBILITY OF SUCH DAMAGE.
30 1.1.2.1 riastrad */
31 1.1.2.1 riastrad
32 1.1.2.1 riastrad #ifndef _LINUX_ATOMIC_H_
33 1.1.2.1 riastrad #define _LINUX_ATOMIC_H_
34 1.1.2.1 riastrad
35 1.1.2.3 riastrad #include <sys/atomic.h>
36 1.1.2.3 riastrad
37 1.1.2.10 riastrad #include <machine/limits.h>
38 1.1.2.10 riastrad
39 1.1.2.6 riastrad struct atomic {
40 1.1.2.4 riastrad union {
41 1.1.2.4 riastrad int au_int;
42 1.1.2.4 riastrad unsigned int au_uint;
43 1.1.2.4 riastrad } a_u;
44 1.1.2.6 riastrad };
45 1.1.2.6 riastrad
46 1.1.2.6 riastrad typedef struct atomic atomic_t;
47 1.1.2.2 riastrad
48 1.1.2.2 riastrad static inline int
49 1.1.2.2 riastrad atomic_read(atomic_t *atomic)
50 1.1.2.2 riastrad {
51 1.1.2.4 riastrad return *(volatile int *)&atomic->a_u.au_int;
52 1.1.2.2 riastrad }
53 1.1.2.2 riastrad
54 1.1.2.3 riastrad static inline void
55 1.1.2.3 riastrad atomic_set(atomic_t *atomic, int value)
56 1.1.2.3 riastrad {
57 1.1.2.4 riastrad atomic->a_u.au_int = value;
58 1.1.2.3 riastrad }
59 1.1.2.3 riastrad
60 1.1.2.3 riastrad static inline void
61 1.1.2.7 riastrad atomic_add(int addend, atomic_t *atomic)
62 1.1.2.7 riastrad {
63 1.1.2.7 riastrad atomic_add_int(&atomic->a_u.au_uint, addend);
64 1.1.2.7 riastrad }
65 1.1.2.7 riastrad
66 1.1.2.9 riastrad static inline void
67 1.1.2.9 riastrad atomic_sub(int subtrahend, atomic_t *atomic)
68 1.1.2.9 riastrad {
69 1.1.2.9 riastrad atomic_add_int(&atomic->a_u.au_uint, -subtrahend);
70 1.1.2.9 riastrad }
71 1.1.2.9 riastrad
72 1.1.2.7 riastrad static inline int
73 1.1.2.7 riastrad atomic_add_return(int addend, atomic_t *atomic)
74 1.1.2.7 riastrad {
75 1.1.2.7 riastrad return (int)atomic_add_int_nv(&atomic->a_u.au_uint, addend);
76 1.1.2.7 riastrad }
77 1.1.2.7 riastrad
78 1.1.2.7 riastrad static inline void
79 1.1.2.3 riastrad atomic_inc(atomic_t *atomic)
80 1.1.2.3 riastrad {
81 1.1.2.4 riastrad atomic_inc_uint(&atomic->a_u.au_uint);
82 1.1.2.4 riastrad }
83 1.1.2.4 riastrad
84 1.1.2.4 riastrad static inline void
85 1.1.2.4 riastrad atomic_dec(atomic_t *atomic)
86 1.1.2.4 riastrad {
87 1.1.2.4 riastrad atomic_dec_uint(&atomic->a_u.au_uint);
88 1.1.2.3 riastrad }
89 1.1.2.3 riastrad
90 1.1.2.3 riastrad static inline int
91 1.1.2.3 riastrad atomic_dec_and_test(atomic_t *atomic)
92 1.1.2.3 riastrad {
93 1.1.2.4 riastrad return (-1 == (int)atomic_dec_uint_nv(&atomic->a_u.au_uint));
94 1.1.2.3 riastrad }
95 1.1.2.3 riastrad
96 1.1.2.5 riastrad static inline void
97 1.1.2.9 riastrad atomic_set_mask(unsigned long mask, atomic_t *atomic)
98 1.1.2.9 riastrad {
99 1.1.2.9 riastrad atomic_or_uint(&atomic->a_u.au_uint, mask);
100 1.1.2.9 riastrad }
101 1.1.2.9 riastrad
102 1.1.2.9 riastrad static inline void
103 1.1.2.9 riastrad atomic_clear_mask(unsigned long mask, atomic_t *atomic)
104 1.1.2.9 riastrad {
105 1.1.2.9 riastrad atomic_and_uint(&atomic->a_u.au_uint, ~mask);
106 1.1.2.9 riastrad }
107 1.1.2.9 riastrad
108 1.1.2.9 riastrad static inline int
109 1.1.2.9 riastrad atomic_add_unless(atomic_t *atomic, int addend, int zero)
110 1.1.2.9 riastrad {
111 1.1.2.9 riastrad int value;
112 1.1.2.9 riastrad
113 1.1.2.9 riastrad do {
114 1.1.2.9 riastrad value = atomic->a_u.au_int;
115 1.1.2.9 riastrad if (value == zero)
116 1.1.2.9 riastrad return 0;
117 1.1.2.9 riastrad } while (atomic_cas_uint(&atomic->a_u.au_uint, value, (value + addend))
118 1.1.2.9 riastrad != value);
119 1.1.2.9 riastrad
120 1.1.2.9 riastrad return 1;
121 1.1.2.9 riastrad }
122 1.1.2.9 riastrad
123 1.1.2.9 riastrad static inline int
124 1.1.2.9 riastrad atomic_inc_not_zero(atomic_t *atomic)
125 1.1.2.9 riastrad {
126 1.1.2.9 riastrad return atomic_add_unless(atomic, 1, 0);
127 1.1.2.9 riastrad }
128 1.1.2.9 riastrad
129 1.1.2.9 riastrad static inline void
130 1.1.2.10 riastrad set_bit(unsigned int bit, volatile unsigned long *ptr)
131 1.1.2.5 riastrad {
132 1.1.2.10 riastrad const unsigned int units = (sizeof(*ptr) * CHAR_BIT);
133 1.1.2.10 riastrad
134 1.1.2.10 riastrad atomic_or_ulong(&ptr[bit / units], (1UL << (bit % units)));
135 1.1.2.5 riastrad }
136 1.1.2.5 riastrad
137 1.1.2.5 riastrad static inline void
138 1.1.2.10 riastrad clear_bit(unsigned int bit, volatile unsigned long *ptr)
139 1.1.2.5 riastrad {
140 1.1.2.10 riastrad const unsigned int units = (sizeof(*ptr) * CHAR_BIT);
141 1.1.2.10 riastrad
142 1.1.2.10 riastrad atomic_and_ulong(&ptr[bit / units], ~(1UL << (bit % units)));
143 1.1.2.5 riastrad }
144 1.1.2.5 riastrad
145 1.1.2.5 riastrad static inline void
146 1.1.2.10 riastrad change_bit(unsigned int bit, volatile unsigned long *ptr)
147 1.1.2.5 riastrad {
148 1.1.2.10 riastrad const unsigned int units = (sizeof(*ptr) * CHAR_BIT);
149 1.1.2.10 riastrad volatile unsigned long *const p = &ptr[bit / units];
150 1.1.2.10 riastrad const unsigned long mask = (1UL << (bit % units));
151 1.1.2.5 riastrad unsigned long v;
152 1.1.2.5 riastrad
153 1.1.2.10 riastrad do v = *p; while (atomic_cas_ulong(p, v, (v ^ mask)) != v);
154 1.1.2.5 riastrad }
155 1.1.2.5 riastrad
156 1.1.2.5 riastrad static inline unsigned long
157 1.1.2.10 riastrad test_and_set_bit(unsigned int bit, volatile unsigned long *ptr)
158 1.1.2.5 riastrad {
159 1.1.2.10 riastrad const unsigned int units = (sizeof(*ptr) * CHAR_BIT);
160 1.1.2.10 riastrad volatile unsigned long *const p = &ptr[bit / units];
161 1.1.2.10 riastrad const unsigned long mask = (1UL << (bit % units));
162 1.1.2.5 riastrad unsigned long v;
163 1.1.2.5 riastrad
164 1.1.2.10 riastrad do v = *p; while (atomic_cas_ulong(p, v, (v | mask)) != v);
165 1.1.2.5 riastrad
166 1.1.2.10 riastrad return (v & mask);
167 1.1.2.5 riastrad }
168 1.1.2.5 riastrad
169 1.1.2.5 riastrad static inline unsigned long
170 1.1.2.10 riastrad test_and_clear_bit(unsigned int bit, volatile unsigned long *ptr)
171 1.1.2.5 riastrad {
172 1.1.2.10 riastrad const unsigned int units = (sizeof(*ptr) * CHAR_BIT);
173 1.1.2.10 riastrad volatile unsigned long *const p = &ptr[bit / units];
174 1.1.2.10 riastrad const unsigned long mask = (1UL << (bit % units));
175 1.1.2.5 riastrad unsigned long v;
176 1.1.2.5 riastrad
177 1.1.2.10 riastrad do v = *p; while (atomic_cas_ulong(p, v, (v & ~mask)) != v);
178 1.1.2.5 riastrad
179 1.1.2.10 riastrad return (v & mask);
180 1.1.2.5 riastrad }
181 1.1.2.5 riastrad
182 1.1.2.5 riastrad static inline unsigned long
183 1.1.2.10 riastrad test_and_change_bit(unsigned int bit, volatile unsigned long *ptr)
184 1.1.2.5 riastrad {
185 1.1.2.10 riastrad const unsigned int units = (sizeof(*ptr) * CHAR_BIT);
186 1.1.2.10 riastrad volatile unsigned long *const p = &ptr[bit / units];
187 1.1.2.10 riastrad const unsigned long mask = (1UL << (bit % units));
188 1.1.2.5 riastrad unsigned long v;
189 1.1.2.5 riastrad
190 1.1.2.10 riastrad do v = *p; while (atomic_cas_ulong(p, v, (v ^ mask)) != v);
191 1.1.2.5 riastrad
192 1.1.2.10 riastrad return (v & mask);
193 1.1.2.5 riastrad }
194 1.1.2.5 riastrad
195 1.1.2.8 riastrad #if defined(MULTIPROCESSOR) && !defined(__HAVE_ATOMIC_AS_MEMBAR)
196 1.1.2.8 riastrad /*
197 1.1.2.8 riastrad * XXX These memory barriers are doubtless overkill, but I am having
198 1.1.2.8 riastrad * trouble understanding the intent and use of the Linux atomic membar
199 1.1.2.8 riastrad * API. I think that for reference counting purposes, the sequences
200 1.1.2.8 riastrad * should be insn/inc/enter and exit/dec/insn, but the use of the
201 1.1.2.8 riastrad * before/after memory barriers is not consistent throughout Linux.
202 1.1.2.8 riastrad */
203 1.1.2.8 riastrad # define smp_mb__before_atomic_inc() membar_sync()
204 1.1.2.8 riastrad # define smp_mb__after_atomic_inc() membar_sync()
205 1.1.2.8 riastrad # define smp_mb__before_atomic_dec() membar_sync()
206 1.1.2.8 riastrad # define smp_mb__after_atomic_dec() membar_sync()
207 1.1.2.8 riastrad #else
208 1.1.2.8 riastrad # define smp_mb__before_atomic_inc() __insn_barrier()
209 1.1.2.8 riastrad # define smp_mb__after_atomic_inc() __insn_barrier()
210 1.1.2.8 riastrad # define smp_mb__before_atomic_dec() __insn_barrier()
211 1.1.2.8 riastrad # define smp_mb__after_atomic_dec() __insn_barrier()
212 1.1.2.8 riastrad #endif
213 1.1.2.8 riastrad
214 1.1.2.1 riastrad #endif /* _LINUX_ATOMIC_H_ */
215