Home | History | Annotate | Line # | Download | only in linux
atomic.h revision 1.3.2.1
      1  1.3.2.1       tls /*	$NetBSD: atomic.h,v 1.3.2.1 2014/08/10 06:55:39 tls Exp $	*/
      2      1.2  riastrad 
      3      1.2  riastrad /*-
      4      1.2  riastrad  * Copyright (c) 2013 The NetBSD Foundation, Inc.
      5      1.2  riastrad  * All rights reserved.
      6      1.2  riastrad  *
      7      1.2  riastrad  * This code is derived from software contributed to The NetBSD Foundation
      8      1.2  riastrad  * by Taylor R. Campbell.
      9      1.2  riastrad  *
     10      1.2  riastrad  * Redistribution and use in source and binary forms, with or without
     11      1.2  riastrad  * modification, are permitted provided that the following conditions
     12      1.2  riastrad  * are met:
     13      1.2  riastrad  * 1. Redistributions of source code must retain the above copyright
     14      1.2  riastrad  *    notice, this list of conditions and the following disclaimer.
     15      1.2  riastrad  * 2. Redistributions in binary form must reproduce the above copyright
     16      1.2  riastrad  *    notice, this list of conditions and the following disclaimer in the
     17      1.2  riastrad  *    documentation and/or other materials provided with the distribution.
     18      1.2  riastrad  *
     19      1.2  riastrad  * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
     20      1.2  riastrad  * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
     21      1.2  riastrad  * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
     22      1.2  riastrad  * PURPOSE ARE DISCLAIMED.  IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
     23      1.2  riastrad  * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
     24      1.2  riastrad  * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
     25      1.2  riastrad  * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
     26      1.2  riastrad  * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
     27      1.2  riastrad  * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
     28      1.2  riastrad  * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
     29      1.2  riastrad  * POSSIBILITY OF SUCH DAMAGE.
     30      1.2  riastrad  */
     31      1.2  riastrad 
     32      1.2  riastrad #ifndef _LINUX_ATOMIC_H_
     33      1.2  riastrad #define _LINUX_ATOMIC_H_
     34      1.2  riastrad 
     35      1.2  riastrad #include <sys/atomic.h>
     36      1.2  riastrad 
     37      1.2  riastrad #include <machine/limits.h>
     38      1.2  riastrad 
     39      1.2  riastrad struct atomic {
     40      1.2  riastrad 	union {
     41      1.3  riastrad 		volatile int au_int;
     42      1.3  riastrad 		volatile unsigned int au_uint;
     43      1.2  riastrad 	} a_u;
     44      1.2  riastrad };
     45      1.2  riastrad 
     46      1.2  riastrad #define	ATOMIC_INIT(i)	{ .a_u = { .au_int = (i) } }
     47      1.2  riastrad 
     48      1.2  riastrad typedef struct atomic atomic_t;
     49      1.2  riastrad 
     50      1.2  riastrad static inline int
     51      1.2  riastrad atomic_read(atomic_t *atomic)
     52      1.2  riastrad {
     53      1.3  riastrad 	return atomic->a_u.au_int;
     54      1.2  riastrad }
     55      1.2  riastrad 
     56      1.2  riastrad static inline void
     57      1.2  riastrad atomic_set(atomic_t *atomic, int value)
     58      1.2  riastrad {
     59      1.2  riastrad 	atomic->a_u.au_int = value;
     60      1.2  riastrad }
     61      1.2  riastrad 
     62      1.2  riastrad static inline void
     63      1.2  riastrad atomic_add(int addend, atomic_t *atomic)
     64      1.2  riastrad {
     65      1.2  riastrad 	atomic_add_int(&atomic->a_u.au_uint, addend);
     66      1.2  riastrad }
     67      1.2  riastrad 
     68      1.2  riastrad static inline void
     69      1.2  riastrad atomic_sub(int subtrahend, atomic_t *atomic)
     70      1.2  riastrad {
     71      1.2  riastrad 	atomic_add_int(&atomic->a_u.au_uint, -subtrahend);
     72      1.2  riastrad }
     73      1.2  riastrad 
     74      1.2  riastrad static inline int
     75      1.2  riastrad atomic_add_return(int addend, atomic_t *atomic)
     76      1.2  riastrad {
     77      1.2  riastrad 	return (int)atomic_add_int_nv(&atomic->a_u.au_uint, addend);
     78      1.2  riastrad }
     79      1.2  riastrad 
     80      1.2  riastrad static inline void
     81      1.2  riastrad atomic_inc(atomic_t *atomic)
     82      1.2  riastrad {
     83      1.2  riastrad 	atomic_inc_uint(&atomic->a_u.au_uint);
     84      1.2  riastrad }
     85      1.2  riastrad 
     86      1.2  riastrad static inline void
     87      1.2  riastrad atomic_dec(atomic_t *atomic)
     88      1.2  riastrad {
     89      1.2  riastrad 	atomic_dec_uint(&atomic->a_u.au_uint);
     90      1.2  riastrad }
     91      1.2  riastrad 
     92      1.2  riastrad static inline int
     93      1.2  riastrad atomic_inc_return(atomic_t *atomic)
     94      1.2  riastrad {
     95      1.2  riastrad 	return (int)atomic_inc_uint_nv(&atomic->a_u.au_uint);
     96      1.2  riastrad }
     97      1.2  riastrad 
     98      1.2  riastrad static inline int
     99      1.2  riastrad atomic_dec_return(atomic_t *atomic)
    100      1.2  riastrad {
    101      1.2  riastrad 	return (int)atomic_dec_uint_nv(&atomic->a_u.au_uint);
    102      1.2  riastrad }
    103      1.2  riastrad 
    104      1.2  riastrad static inline int
    105      1.2  riastrad atomic_dec_and_test(atomic_t *atomic)
    106      1.2  riastrad {
    107  1.3.2.1       tls 	return (0 == (int)atomic_dec_uint_nv(&atomic->a_u.au_uint));
    108      1.2  riastrad }
    109      1.2  riastrad 
    110      1.2  riastrad static inline void
    111      1.2  riastrad atomic_set_mask(unsigned long mask, atomic_t *atomic)
    112      1.2  riastrad {
    113      1.2  riastrad 	atomic_or_uint(&atomic->a_u.au_uint, mask);
    114      1.2  riastrad }
    115      1.2  riastrad 
    116      1.2  riastrad static inline void
    117      1.2  riastrad atomic_clear_mask(unsigned long mask, atomic_t *atomic)
    118      1.2  riastrad {
    119      1.2  riastrad 	atomic_and_uint(&atomic->a_u.au_uint, ~mask);
    120      1.2  riastrad }
    121      1.2  riastrad 
    122      1.2  riastrad static inline int
    123      1.2  riastrad atomic_add_unless(atomic_t *atomic, int addend, int zero)
    124      1.2  riastrad {
    125      1.2  riastrad 	int value;
    126      1.2  riastrad 
    127      1.2  riastrad 	do {
    128      1.2  riastrad 		value = atomic->a_u.au_int;
    129      1.2  riastrad 		if (value == zero)
    130      1.2  riastrad 			return 0;
    131      1.2  riastrad 	} while (atomic_cas_uint(&atomic->a_u.au_uint, value, (value + addend))
    132      1.2  riastrad 	    != value);
    133      1.2  riastrad 
    134      1.2  riastrad 	return 1;
    135      1.2  riastrad }
    136      1.2  riastrad 
    137      1.2  riastrad static inline int
    138      1.2  riastrad atomic_inc_not_zero(atomic_t *atomic)
    139      1.2  riastrad {
    140      1.2  riastrad 	return atomic_add_unless(atomic, 1, 0);
    141      1.2  riastrad }
    142      1.2  riastrad 
    143  1.3.2.1       tls static inline int
    144  1.3.2.1       tls atomic_xchg(atomic_t *atomic, int new)
    145  1.3.2.1       tls {
    146  1.3.2.1       tls 	return (int)atomic_swap_uint(&atomic->a_u.au_uint, (unsigned)new);
    147  1.3.2.1       tls }
    148  1.3.2.1       tls 
    149  1.3.2.1       tls static inline int
    150  1.3.2.1       tls atomic_cmpxchg(atomic_t *atomic, int old, int new)
    151  1.3.2.1       tls {
    152  1.3.2.1       tls 	return (int)atomic_cas_uint(&atomic->a_u.au_uint, (unsigned)old,
    153  1.3.2.1       tls 	    (unsigned)new);
    154  1.3.2.1       tls }
    155  1.3.2.1       tls 
    156  1.3.2.1       tls struct atomic64 {
    157  1.3.2.1       tls 	volatile uint64_t	a_v;
    158  1.3.2.1       tls };
    159  1.3.2.1       tls 
    160  1.3.2.1       tls typedef struct atomic64 atomic64_t;
    161  1.3.2.1       tls 
    162  1.3.2.1       tls static inline uint64_t
    163  1.3.2.1       tls atomic64_read(const struct atomic64 *a)
    164  1.3.2.1       tls {
    165  1.3.2.1       tls 	return a->a_v;
    166  1.3.2.1       tls }
    167  1.3.2.1       tls 
    168  1.3.2.1       tls static inline void
    169  1.3.2.1       tls atomic64_set(struct atomic64 *a, uint64_t v)
    170  1.3.2.1       tls {
    171  1.3.2.1       tls 	a->a_v = v;
    172  1.3.2.1       tls }
    173  1.3.2.1       tls 
    174  1.3.2.1       tls static inline void
    175  1.3.2.1       tls atomic64_add(long long d, struct atomic64 *a)
    176  1.3.2.1       tls {
    177  1.3.2.1       tls 	atomic_add_64(&a->a_v, d);
    178  1.3.2.1       tls }
    179  1.3.2.1       tls 
    180  1.3.2.1       tls static inline void
    181  1.3.2.1       tls atomic64_sub(long long d, struct atomic64 *a)
    182  1.3.2.1       tls {
    183  1.3.2.1       tls 	atomic_add_64(&a->a_v, -d);
    184  1.3.2.1       tls }
    185  1.3.2.1       tls 
    186  1.3.2.1       tls static inline uint64_t
    187  1.3.2.1       tls atomic64_xchg(struct atomic64 *a, uint64_t v)
    188  1.3.2.1       tls {
    189  1.3.2.1       tls 	return atomic_swap_64(&a->a_v, v);
    190  1.3.2.1       tls }
    191  1.3.2.1       tls 
    192      1.2  riastrad static inline void
    193      1.2  riastrad set_bit(unsigned int bit, volatile unsigned long *ptr)
    194      1.2  riastrad {
    195      1.2  riastrad 	const unsigned int units = (sizeof(*ptr) * CHAR_BIT);
    196      1.2  riastrad 
    197      1.2  riastrad 	atomic_or_ulong(&ptr[bit / units], (1UL << (bit % units)));
    198      1.2  riastrad }
    199      1.2  riastrad 
    200      1.2  riastrad static inline void
    201      1.2  riastrad clear_bit(unsigned int bit, volatile unsigned long *ptr)
    202      1.2  riastrad {
    203      1.2  riastrad 	const unsigned int units = (sizeof(*ptr) * CHAR_BIT);
    204      1.2  riastrad 
    205      1.2  riastrad 	atomic_and_ulong(&ptr[bit / units], ~(1UL << (bit % units)));
    206      1.2  riastrad }
    207      1.2  riastrad 
    208      1.2  riastrad static inline void
    209      1.2  riastrad change_bit(unsigned int bit, volatile unsigned long *ptr)
    210      1.2  riastrad {
    211      1.2  riastrad 	const unsigned int units = (sizeof(*ptr) * CHAR_BIT);
    212      1.2  riastrad 	volatile unsigned long *const p = &ptr[bit / units];
    213      1.2  riastrad 	const unsigned long mask = (1UL << (bit % units));
    214      1.2  riastrad 	unsigned long v;
    215      1.2  riastrad 
    216      1.2  riastrad 	do v = *p; while (atomic_cas_ulong(p, v, (v ^ mask)) != v);
    217      1.2  riastrad }
    218      1.2  riastrad 
    219      1.2  riastrad static inline unsigned long
    220      1.2  riastrad test_and_set_bit(unsigned int bit, volatile unsigned long *ptr)
    221      1.2  riastrad {
    222      1.2  riastrad 	const unsigned int units = (sizeof(*ptr) * CHAR_BIT);
    223      1.2  riastrad 	volatile unsigned long *const p = &ptr[bit / units];
    224      1.2  riastrad 	const unsigned long mask = (1UL << (bit % units));
    225      1.2  riastrad 	unsigned long v;
    226      1.2  riastrad 
    227      1.2  riastrad 	do v = *p; while (atomic_cas_ulong(p, v, (v | mask)) != v);
    228      1.2  riastrad 
    229  1.3.2.1       tls 	return ((v & mask) != 0);
    230      1.2  riastrad }
    231      1.2  riastrad 
    232      1.2  riastrad static inline unsigned long
    233      1.2  riastrad test_and_clear_bit(unsigned int bit, volatile unsigned long *ptr)
    234      1.2  riastrad {
    235      1.2  riastrad 	const unsigned int units = (sizeof(*ptr) * CHAR_BIT);
    236      1.2  riastrad 	volatile unsigned long *const p = &ptr[bit / units];
    237      1.2  riastrad 	const unsigned long mask = (1UL << (bit % units));
    238      1.2  riastrad 	unsigned long v;
    239      1.2  riastrad 
    240      1.2  riastrad 	do v = *p; while (atomic_cas_ulong(p, v, (v & ~mask)) != v);
    241      1.2  riastrad 
    242  1.3.2.1       tls 	return ((v & mask) != 0);
    243      1.2  riastrad }
    244      1.2  riastrad 
    245      1.2  riastrad static inline unsigned long
    246      1.2  riastrad test_and_change_bit(unsigned int bit, volatile unsigned long *ptr)
    247      1.2  riastrad {
    248      1.2  riastrad 	const unsigned int units = (sizeof(*ptr) * CHAR_BIT);
    249      1.2  riastrad 	volatile unsigned long *const p = &ptr[bit / units];
    250      1.2  riastrad 	const unsigned long mask = (1UL << (bit % units));
    251      1.2  riastrad 	unsigned long v;
    252      1.2  riastrad 
    253      1.2  riastrad 	do v = *p; while (atomic_cas_ulong(p, v, (v ^ mask)) != v);
    254      1.2  riastrad 
    255  1.3.2.1       tls 	return ((v & mask) != 0);
    256      1.2  riastrad }
    257      1.2  riastrad 
    258      1.2  riastrad #if defined(MULTIPROCESSOR) && !defined(__HAVE_ATOMIC_AS_MEMBAR)
    259      1.2  riastrad /*
    260      1.2  riastrad  * XXX These memory barriers are doubtless overkill, but I am having
    261      1.2  riastrad  * trouble understanding the intent and use of the Linux atomic membar
    262      1.2  riastrad  * API.  I think that for reference counting purposes, the sequences
    263      1.2  riastrad  * should be insn/inc/enter and exit/dec/insn, but the use of the
    264      1.2  riastrad  * before/after memory barriers is not consistent throughout Linux.
    265      1.2  riastrad  */
    266      1.2  riastrad #  define	smp_mb__before_atomic_inc()	membar_sync()
    267      1.2  riastrad #  define	smp_mb__after_atomic_inc()	membar_sync()
    268      1.2  riastrad #  define	smp_mb__before_atomic_dec()	membar_sync()
    269      1.2  riastrad #  define	smp_mb__after_atomic_dec()	membar_sync()
    270      1.2  riastrad #else
    271      1.2  riastrad #  define	smp_mb__before_atomic_inc()	__insn_barrier()
    272      1.2  riastrad #  define	smp_mb__after_atomic_inc()	__insn_barrier()
    273      1.2  riastrad #  define	smp_mb__before_atomic_dec()	__insn_barrier()
    274      1.2  riastrad #  define	smp_mb__after_atomic_dec()	__insn_barrier()
    275      1.2  riastrad #endif
    276      1.2  riastrad 
    277      1.2  riastrad #endif  /* _LINUX_ATOMIC_H_ */
    278