| /src/sys/kern/ |
| kern_mutex_obj.c | 58 struct kmutexobj *mo; local 60 mo = kmem_intr_alloc(sizeof(*mo), KM_SLEEP); 61 KASSERT(ALIGNED_POINTER(mo, coherency_unit)); 62 _mutex_init(&mo->mo_lock, type, ipl, 64 mo->mo_magic = MUTEX_OBJ_MAGIC; 65 mo->mo_refcnt = 1; 67 return (kmutex_t *)mo; 78 struct kmutexobj *mo; local 80 mo = kmem_intr_alloc(sizeof(*mo), KM_NOSLEEP) 101 struct kmutexobj *mo = (struct kmutexobj *)lock; local 122 struct kmutexobj *mo = (struct kmutexobj *)lock; local 149 struct kmutexobj *mo = (struct kmutexobj *)lock; local [all...] |
| /src/external/lgpl2/userspace-rcu/dist/include/urcu/uatomic/ |
| api.h | 19 #define _uatomic_default_mo(dflt, mo...) \ 20 _uatomic_select_arg1(_, ##mo, dflt) 22 #define _uatomic_default_mo2(dflt, mo...) \ 23 _uatomic_select_arg2(_, ##mo, dflt, dflt) 25 #define uatomic_load(addr, mo...) \ 26 uatomic_load_mo(addr, _uatomic_default_mo(CMM_RELAXED, ##mo)) 28 #define uatomic_read(addr, mo...) \ 29 uatomic_load_mo(addr, _uatomic_default_mo(CMM_RELAXED, ##mo)) 31 #define uatomic_store(addr, value, mo...) \ 32 uatomic_store_mo(addr, value, _uatomic_default_mo(CMM_RELAXED, ##mo)) [all...] |
| builtins-generic.h | 15 #define uatomic_store_mo(addr, v, mo) \ 18 cmm_to_c11(mo)); \ 19 cmm_seq_cst_fence_after_atomic(mo); \ 22 #define uatomic_load_mo(addr, mo) \ 27 cmm_to_c11(mo)); \ 28 cmm_seq_cst_fence_after_atomic(mo); \ 49 #define uatomic_xchg_mo(addr, v, mo) \ 54 cmm_to_c11(mo)); \ 55 cmm_seq_cst_fence_after_atomic(mo); \ 59 #define uatomic_add_return_mo(addr, v, mo) \ [all...] |
| arm.h | 36 # define uatomic_xchg_mo(addr, v, mo) \ 41 cmm_to_c11(mo)); \ 42 cmm_seq_cst_fence_after_atomic(mo); \ 47 static inline void _cmm_compat_c11_smp_mb__before_xchg_mo(enum cmm_memorder mo) 49 switch (mo) { 73 # define uatomic_xchg_mo(addr, v, mo) \ 75 _cmm_compat_c11_smp_mb__before_xchg_mo(mo); \
|
| /src/sys/external/bsd/compiler_rt/dist/include/sanitizer/ |
| tsan_interface_atomic.h | 45 __tsan_memory_order mo); 47 __tsan_memory_order mo); 49 __tsan_memory_order mo); 51 __tsan_memory_order mo); 54 __tsan_memory_order mo); 58 __tsan_memory_order mo); 60 __tsan_memory_order mo); 62 __tsan_memory_order mo); 64 __tsan_memory_order mo); 67 __tsan_memory_order mo); [all...] |
| /src/external/gpl3/gcc/dist/libsanitizer/include/sanitizer/ |
| tsan_interface_atomic.h | 46 __tsan_atomic8_load(const volatile __tsan_atomic8 *a, __tsan_memory_order mo); 48 __tsan_atomic16_load(const volatile __tsan_atomic16 *a, __tsan_memory_order mo); 50 __tsan_atomic32_load(const volatile __tsan_atomic32 *a, __tsan_memory_order mo); 52 __tsan_atomic64_load(const volatile __tsan_atomic64 *a, __tsan_memory_order mo); 55 const volatile __tsan_atomic128 *a, __tsan_memory_order mo); 60 __tsan_memory_order mo); 63 __tsan_memory_order mo); 66 __tsan_memory_order mo); 69 __tsan_memory_order mo); 73 __tsan_memory_order mo); [all...] |
| /src/external/gpl3/gcc.old/dist/libsanitizer/include/sanitizer/ |
| tsan_interface_atomic.h | 44 __tsan_memory_order mo); 46 __tsan_memory_order mo); 48 __tsan_memory_order mo); 50 __tsan_memory_order mo); 53 __tsan_memory_order mo); 57 __tsan_memory_order mo); 59 __tsan_memory_order mo); 61 __tsan_memory_order mo); 63 __tsan_memory_order mo); 66 __tsan_memory_order mo); [all...] |
| /src/sys/external/bsd/compiler_rt/dist/lib/tsan/rtl/ |
| tsan_interface_atomic.cc | 36 static bool IsLoadOrder(morder mo) { 37 return mo == mo_relaxed || mo == mo_consume 38 || mo == mo_acquire || mo == mo_seq_cst; 41 static bool IsStoreOrder(morder mo) { 42 return mo == mo_relaxed || mo == mo_release || mo == mo_seq_cst; 45 static bool IsReleaseOrder(morder mo) { [all...] |
| /src/external/gpl3/gcc/dist/libsanitizer/tsan/ |
| tsan_interface_atomic.cpp | 36 static bool IsLoadOrder(morder mo) { 37 return mo == mo_relaxed || mo == mo_consume 38 || mo == mo_acquire || mo == mo_seq_cst; 41 static bool IsStoreOrder(morder mo) { 42 return mo == mo_relaxed || mo == mo_release || mo == mo_seq_cst; 46 static bool IsReleaseOrder(morder mo) { [all...] |
| /src/external/gpl3/gcc.old/dist/libsanitizer/tsan/ |
| tsan_interface_atomic.cpp | 36 static bool IsLoadOrder(morder mo) { 37 return mo == mo_relaxed || mo == mo_consume 38 || mo == mo_acquire || mo == mo_seq_cst; 41 static bool IsStoreOrder(morder mo) { 42 return mo == mo_relaxed || mo == mo_release || mo == mo_seq_cst; 46 static bool IsReleaseOrder(morder mo) { [all...] |
| /src/external/bsd/jemalloc/dist/include/jemalloc/internal/ |
| atomic_c11.h | 23 atomic_memory_order_t mo) { \ 31 return atomic_load_explicit(a_nonconst, mo); \ 36 type val, atomic_memory_order_t mo) { \ 37 atomic_store_explicit(a, val, mo); \ 42 atomic_memory_order_t mo) { \ 43 return atomic_exchange_explicit(a, val, mo); \ 72 type val, atomic_memory_order_t mo) { \ 73 return atomic_fetch_add_explicit(a, val, mo); \ 78 type val, atomic_memory_order_t mo) { \ 79 return atomic_fetch_sub_explicit(a, val, mo); \ [all...] |
| atomic_gcc_atomic.h | 17 atomic_enum_to_builtin(atomic_memory_order_t mo) { 18 switch (mo) { 35 atomic_fence(atomic_memory_order_t mo) { 36 __atomic_thread_fence(atomic_enum_to_builtin(mo)); 47 atomic_memory_order_t mo) { \ 49 __atomic_load(&a->repr, &result, atomic_enum_to_builtin(mo)); \ 55 atomic_memory_order_t mo) { \ 56 __atomic_store(&a->repr, &val, atomic_enum_to_builtin(mo)); \ 61 atomic_memory_order_t mo) { \ 64 atomic_enum_to_builtin(mo)); \ [all...] |
| /src/external/bsd/jemalloc/include/jemalloc/internal/ |
| atomic_c11.h | 23 atomic_memory_order_t mo) { \ 31 return atomic_load_explicit(a_nonconst, mo); \ 36 type val, atomic_memory_order_t mo) { \ 37 atomic_store_explicit(a, val, mo); \ 42 atomic_memory_order_t mo) { \ 43 return atomic_exchange_explicit(a, val, mo); \ 72 type val, atomic_memory_order_t mo) { \ 73 return atomic_fetch_add_explicit(a, val, mo); \ 78 type val, atomic_memory_order_t mo) { \ 79 return atomic_fetch_sub_explicit(a, val, mo); \ [all...] |
| atomic_gcc_atomic.h | 17 atomic_enum_to_builtin(atomic_memory_order_t mo) { 18 switch (mo) { 35 atomic_fence(atomic_memory_order_t mo) { 36 __atomic_thread_fence(atomic_enum_to_builtin(mo)); 47 atomic_memory_order_t mo) { \ 49 __atomic_load(&a->repr, &result, atomic_enum_to_builtin(mo)); \ 55 atomic_memory_order_t mo) { \ 56 __atomic_store(&a->repr, &val, atomic_enum_to_builtin(mo)); \ 61 atomic_memory_order_t mo) { \ 64 atomic_enum_to_builtin(mo)); \ [all...] |
| /src/external/bsd/jemalloc.old/dist/include/jemalloc/internal/ |
| atomic_c11.h | 23 atomic_memory_order_t mo) { \ 31 return atomic_load_explicit(a_nonconst, mo); \ 36 type val, atomic_memory_order_t mo) { \ 37 atomic_store_explicit(a, val, mo); \ 42 atomic_memory_order_t mo) { \ 43 return atomic_exchange_explicit(a, val, mo); \ 72 type val, atomic_memory_order_t mo) { \ 73 return atomic_fetch_add_explicit(a, val, mo); \ 78 type val, atomic_memory_order_t mo) { \ 79 return atomic_fetch_sub_explicit(a, val, mo); \ [all...] |
| atomic_gcc_atomic.h | 17 atomic_enum_to_builtin(atomic_memory_order_t mo) { 18 switch (mo) { 35 atomic_fence(atomic_memory_order_t mo) { 36 __atomic_thread_fence(atomic_enum_to_builtin(mo)); 46 atomic_memory_order_t mo) { \ 48 __atomic_load(&a->repr, &result, atomic_enum_to_builtin(mo)); \ 54 atomic_memory_order_t mo) { \ 55 __atomic_store(&a->repr, &val, atomic_enum_to_builtin(mo)); \ 60 atomic_memory_order_t mo) { \ 63 atomic_enum_to_builtin(mo)); \ [all...] |
| /src/external/bsd/jemalloc.old/include/jemalloc/internal/ |
| atomic_c11.h | 23 atomic_memory_order_t mo) { \ 31 return atomic_load_explicit(a_nonconst, mo); \ 36 type val, atomic_memory_order_t mo) { \ 37 atomic_store_explicit(a, val, mo); \ 42 atomic_memory_order_t mo) { \ 43 return atomic_exchange_explicit(a, val, mo); \ 72 type val, atomic_memory_order_t mo) { \ 73 return atomic_fetch_add_explicit(a, val, mo); \ 78 type val, atomic_memory_order_t mo) { \ 79 return atomic_fetch_sub_explicit(a, val, mo); \ [all...] |
| atomic_gcc_atomic.h | 17 atomic_enum_to_builtin(atomic_memory_order_t mo) { 18 switch (mo) { 35 atomic_fence(atomic_memory_order_t mo) { 36 __atomic_thread_fence(atomic_enum_to_builtin(mo)); 46 atomic_memory_order_t mo) { \ 48 __atomic_load(&a->repr, &result, atomic_enum_to_builtin(mo)); \ 54 atomic_memory_order_t mo) { \ 55 __atomic_store(&a->repr, &val, atomic_enum_to_builtin(mo)); \ 60 atomic_memory_order_t mo) { \ 63 atomic_enum_to_builtin(mo)); \ [all...] |
| /src/external/gpl3/gcc/dist/libsanitizer/sanitizer_common/ |
| sanitizer_atomic_clang_other.h | 26 const volatile T *a, memory_order mo) { 27 DCHECK(mo & (memory_order_relaxed | memory_order_consume 34 if (mo == memory_order_relaxed) { 36 } else if (mo == memory_order_consume) { 42 } else if (mo == memory_order_acquire) { 60 inline void atomic_store(volatile T *a, typename T::Type v, memory_order mo) { 61 DCHECK(mo & (memory_order_relaxed | memory_order_release 67 if (mo == memory_order_relaxed) { 69 } else if (mo == memory_order_release) {
|
| sanitizer_atomic_clang.h | 47 typename T::Type v, memory_order mo) { 48 (void)mo; 55 typename T::Type v, memory_order mo) { 56 (void)mo; 63 typename T::Type v, memory_order mo) { 65 if (mo & (memory_order_release | memory_order_acq_rel | memory_order_seq_cst)) 68 if (mo == memory_order_seq_cst) 76 memory_order mo) { 89 memory_order mo) { 90 return atomic_compare_exchange_strong(a, cmp, xchg, mo); [all...] |
| sanitizer_atomic_clang_x86.h | 28 const volatile T *a, memory_order mo) { 29 DCHECK(mo & (memory_order_relaxed | memory_order_consume 36 if (mo == memory_order_relaxed) { 38 } else if (mo == memory_order_consume) { 44 } else if (mo == memory_order_acquire) { 73 inline void atomic_store(volatile T *a, typename T::Type v, memory_order mo) { 74 DCHECK(mo & (memory_order_relaxed | memory_order_release 80 if (mo == memory_order_relaxed) { 82 } else if (mo == memory_order_release) { 106 if (mo == memory_order_seq_cst [all...] |
| /src/external/gpl3/gcc.old/dist/libsanitizer/sanitizer_common/ |
| sanitizer_atomic_clang_other.h | 26 const volatile T *a, memory_order mo) { 27 DCHECK(mo & (memory_order_relaxed | memory_order_consume 34 if (mo == memory_order_relaxed) { 36 } else if (mo == memory_order_consume) { 42 } else if (mo == memory_order_acquire) { 60 inline void atomic_store(volatile T *a, typename T::Type v, memory_order mo) { 61 DCHECK(mo & (memory_order_relaxed | memory_order_release 67 if (mo == memory_order_relaxed) { 69 } else if (mo == memory_order_release) {
|
| /src/lib/libintl/ |
| gettext.c | 206 * the code reads in *.mo files generated by GNU gettext. *.mo is a host- 338 int rv = snprintf(buf, len, "%s/%s/%s/%s.mo", dir, p, 457 mohandle->mo.mo_sysdep_segs[str->segs[i].ref].len; 471 memcpy(dst, mohandle->mo.mo_sysdep_segs[str->segs[i].ref].str, 472 mohandle->mo.mo_sysdep_segs[str->segs[i].ref].len); 473 dst += mohandle->mo.mo_sysdep_segs[str->segs[i].ref].len; 496 setup_sysdep_stuffs(struct mo *mo, struct mohandle *mohandle, char *base) 505 magic = mo->mo_magic 581 struct mo *mo; local [all...] |
| /src/sys/external/bsd/compiler_rt/dist/lib/sanitizer_common/ |
| sanitizer_atomic_clang.h | 48 typename T::Type v, memory_order mo) { 49 (void)mo; 56 typename T::Type v, memory_order mo) { 57 (void)mo; 64 typename T::Type v, memory_order mo) { 66 if (mo & (memory_order_release | memory_order_acq_rel | memory_order_seq_cst)) 69 if (mo == memory_order_seq_cst) 77 memory_order mo) { 91 memory_order mo) { 92 return atomic_compare_exchange_strong(a, cmp, xchg, mo); [all...] |
| sanitizer_atomic_clang_x86.h | 29 const volatile T *a, memory_order mo) { 30 DCHECK(mo & (memory_order_relaxed | memory_order_consume 37 if (mo == memory_order_relaxed) { 39 } else if (mo == memory_order_consume) { 45 } else if (mo == memory_order_acquire) { 74 INLINE void atomic_store(volatile T *a, typename T::Type v, memory_order mo) { 75 DCHECK(mo & (memory_order_relaxed | memory_order_release 81 if (mo == memory_order_relaxed) { 83 } else if (mo == memory_order_release) { 107 if (mo == memory_order_seq_cst [all...] |