| /src/external/bsd/jemalloc/dist/include/jemalloc/internal/ |
| arena_inlines_a.h | 11 atomic_fetch_add_zu(&arena->stats.internal, size, ATOMIC_RELAXED); 16 atomic_fetch_sub_zu(&arena->stats.internal, size, ATOMIC_RELAXED); 21 return atomic_load_zu(&arena->stats.internal, ATOMIC_RELAXED);
|
| lockedint.h | 61 return atomic_load_u64(&p->val, ATOMIC_RELAXED); 72 atomic_fetch_add_u64(&p->val, x, ATOMIC_RELAXED); 83 uint64_t r = atomic_fetch_sub_u64(&p->val, x, ATOMIC_RELAXED); 99 before = atomic_load_u64(&p->val, ATOMIC_RELAXED); 108 ATOMIC_RELAXED, ATOMIC_RELAXED)); 129 uint64_t cur_dst = atomic_load_u64(&dst->val, ATOMIC_RELAXED); 130 atomic_store_u64(&dst->val, src + cur_dst, ATOMIC_RELAXED); 139 return atomic_load_u64(&p->val, ATOMIC_RELAXED); 148 atomic_store_u64(&p->val, x, ATOMIC_RELAXED); [all...] |
| atomic.h | 33 * atomic_store_pi(&atomic_ptr_to_int, some_ptr, ATOMIC_RELAXED); 48 #define ATOMIC_RELAXED atomic_memory_order_relaxed 63 type oldval = atomic_load_##short_type(a, ATOMIC_RELAXED); \ 65 atomic_store_##short_type(a, newval, ATOMIC_RELAXED); \ 70 type oldval = atomic_load_##short_type(a, ATOMIC_RELAXED); \ 72 atomic_store_##short_type(a, newval, ATOMIC_RELAXED); \
|
| seq.h | 26 size_t old_seq = atomic_load_zu(&dst->seq, ATOMIC_RELAXED); \ 27 atomic_store_zu(&dst->seq, old_seq + 1, ATOMIC_RELAXED); \ 30 atomic_store_zu(&dst->data[i], buf[i], ATOMIC_RELAXED); \ 44 buf[i] = atomic_load_zu(&src->data[i], ATOMIC_RELAXED); \ 47 size_t seq2 = atomic_load_zu(&src->seq, ATOMIC_RELAXED); \
|
| mpsc_queue.h | 56 atomic_store_p(&queue->tail, NULL, ATOMIC_RELAXED); \ 66 void* cur_tail = atomic_load_p(&queue->tail, ATOMIC_RELAXED); \ 79 &cur_tail, last, ATOMIC_RELEASE, ATOMIC_RELAXED)); \ 92 a_type *tail = atomic_load_p(&queue->tail, ATOMIC_RELAXED); \
|
| /src/external/bsd/jemalloc/include/jemalloc/internal/ |
| arena_inlines_a.h | 11 atomic_fetch_add_zu(&arena->stats.internal, size, ATOMIC_RELAXED); 16 atomic_fetch_sub_zu(&arena->stats.internal, size, ATOMIC_RELAXED); 21 return atomic_load_zu(&arena->stats.internal, ATOMIC_RELAXED);
|
| prof_inlines_a.h | 20 a0 = atomic_load_u64(&prof_accum->accumbytes, ATOMIC_RELAXED); 29 a1, ATOMIC_RELAXED, ATOMIC_RELAXED)); 56 a0 = atomic_load_u64(&prof_accum->accumbytes, ATOMIC_RELAXED); 61 a1, ATOMIC_RELAXED, ATOMIC_RELAXED));
|
| lockedint.h | 61 return atomic_load_u64(&p->val, ATOMIC_RELAXED); 72 atomic_fetch_add_u64(&p->val, x, ATOMIC_RELAXED); 83 uint64_t r = atomic_fetch_sub_u64(&p->val, x, ATOMIC_RELAXED); 99 before = atomic_load_u64(&p->val, ATOMIC_RELAXED); 108 ATOMIC_RELAXED, ATOMIC_RELAXED)); 129 uint64_t cur_dst = atomic_load_u64(&dst->val, ATOMIC_RELAXED); 130 atomic_store_u64(&dst->val, src + cur_dst, ATOMIC_RELAXED); 139 return atomic_load_u64(&p->val, ATOMIC_RELAXED); 148 atomic_store_u64(&p->val, x, ATOMIC_RELAXED); [all...] |
| atomic.h | 33 * atomic_store_pi(&atomic_ptr_to_int, some_ptr, ATOMIC_RELAXED); 48 #define ATOMIC_RELAXED atomic_memory_order_relaxed 63 type oldval = atomic_load_##short_type(a, ATOMIC_RELAXED); \ 65 atomic_store_##short_type(a, newval, ATOMIC_RELAXED); \ 70 type oldval = atomic_load_##short_type(a, ATOMIC_RELAXED); \ 72 atomic_store_##short_type(a, newval, ATOMIC_RELAXED); \
|
| seq.h | 26 size_t old_seq = atomic_load_zu(&dst->seq, ATOMIC_RELAXED); \ 27 atomic_store_zu(&dst->seq, old_seq + 1, ATOMIC_RELAXED); \ 30 atomic_store_zu(&dst->data[i], buf[i], ATOMIC_RELAXED); \ 44 buf[i] = atomic_load_zu(&src->data[i], ATOMIC_RELAXED); \ 47 size_t seq2 = atomic_load_zu(&src->seq, ATOMIC_RELAXED); \
|
| /src/external/bsd/jemalloc/dist/test/unit/ |
| atomic.c | 26 val = atomic_load_##ta(&atom, ATOMIC_RELAXED); \ 30 atomic_store_##ta(&atom, val1, ATOMIC_RELAXED); \ 31 atomic_store_##ta(&atom, val2, ATOMIC_RELAXED); \ 32 val = atomic_load_##ta(&atom, ATOMIC_RELAXED); \ 36 atomic_store_##ta(&atom, val1, ATOMIC_RELAXED); \ 37 val = atomic_exchange_##ta(&atom, val2, ATOMIC_RELAXED); \ 39 val = atomic_load_##ta(&atom, ATOMIC_RELAXED); \ 46 atomic_store_##ta(&atom, val1, ATOMIC_RELAXED); \ 51 &expected, val3, ATOMIC_RELAXED, ATOMIC_RELAXED); \ [all...] |
| edata_cache.c | 28 assert_zu_eq(atomic_load_zu(&ec.count, ATOMIC_RELAXED), 0, ""); 35 assert_zu_eq(atomic_load_zu(&ec.count, ATOMIC_RELAXED), 0, ""); 39 assert_zu_eq(atomic_load_zu(&ec.count, ATOMIC_RELAXED), 1, ""); 44 assert_zu_eq(atomic_load_zu(&ec.count, ATOMIC_RELAXED), 0, ""); 70 expect_zu_eq(atomic_load_zu(&ec.count, ATOMIC_RELAXED), 0, ""); 75 expect_zu_eq(atomic_load_zu(&ec.count, ATOMIC_RELAXED), 0, ""); 79 expect_zu_eq(atomic_load_zu(&ec.count, ATOMIC_RELAXED), 0, ""); 83 expect_zu_eq(atomic_load_zu(&ec.count, ATOMIC_RELAXED), 0, ""); 88 expect_zu_eq(atomic_load_zu(&ec.count, ATOMIC_RELAXED), 0, ""); 92 expect_zu_eq(atomic_load_zu(&ec.count, ATOMIC_RELAXED), 0, "") [all...] |
| /src/external/bsd/jemalloc.old/dist/test/unit/ |
| atomic.c | 26 val = atomic_load_##ta(&atom, ATOMIC_RELAXED); \ 30 atomic_store_##ta(&atom, val1, ATOMIC_RELAXED); \ 31 atomic_store_##ta(&atom, val2, ATOMIC_RELAXED); \ 32 val = atomic_load_##ta(&atom, ATOMIC_RELAXED); \ 36 atomic_store_##ta(&atom, val1, ATOMIC_RELAXED); \ 37 val = atomic_exchange_##ta(&atom, val2, ATOMIC_RELAXED); \ 39 val = atomic_load_##ta(&atom, ATOMIC_RELAXED); \ 46 atomic_store_##ta(&atom, val1, ATOMIC_RELAXED); \ 51 &expected, val3, ATOMIC_RELAXED, ATOMIC_RELAXED); \ [all...] |
| prng.c | 9 atomic_store_u32(&sa, 42, ATOMIC_RELAXED); 11 atomic_store_u32(&sa, 42, ATOMIC_RELAXED); 16 atomic_store_u32(&sb, 42, ATOMIC_RELAXED); 21 atomic_store_u32(&sa, 42, ATOMIC_RELAXED); 27 atomic_store_u32(&sa, 42, ATOMIC_RELAXED); 30 atomic_store_u32(&sb, 42, ATOMIC_RELAXED); 82 atomic_store_zu(&sa, 42, ATOMIC_RELAXED); 84 atomic_store_zu(&sa, 42, ATOMIC_RELAXED); 89 atomic_store_zu(&sb, 42, ATOMIC_RELAXED); 94 atomic_store_zu(&sa, 42, ATOMIC_RELAXED); [all...] |
| /src/external/bsd/jemalloc.old/dist/include/jemalloc/internal/ |
| prof_inlines_a.h | 20 a0 = atomic_load_u64(&prof_accum->accumbytes, ATOMIC_RELAXED); 29 a1, ATOMIC_RELAXED, ATOMIC_RELAXED)); 56 a0 = atomic_load_u64(&prof_accum->accumbytes, ATOMIC_RELAXED); 61 a1, ATOMIC_RELAXED, ATOMIC_RELAXED));
|
| prng.h | 73 state0 = atomic_load_u32(state, ATOMIC_RELAXED); 79 state1, ATOMIC_RELAXED, ATOMIC_RELAXED)); 82 atomic_store_u32(state, state1, ATOMIC_RELAXED); 110 state0 = atomic_load_zu(state, ATOMIC_RELAXED); 116 state1, ATOMIC_RELAXED, ATOMIC_RELAXED)); 119 atomic_store_zu(state, state1, ATOMIC_RELAXED);
|
| arena_stats.h | 132 return atomic_load_u64(p, ATOMIC_RELAXED); 143 atomic_fetch_add_u64(p, x, ATOMIC_RELAXED); 154 UNUSED uint64_t r = atomic_fetch_sub_u64(p, x, ATOMIC_RELAXED); 171 uint64_t cur_dst = atomic_load_u64(dst, ATOMIC_RELAXED); 172 atomic_store_u64(dst, src + cur_dst, ATOMIC_RELAXED); 181 return atomic_load_zu(p, ATOMIC_RELAXED); 184 return atomic_load_zu(p, ATOMIC_RELAXED); 192 atomic_fetch_add_zu(p, x, ATOMIC_RELAXED); 195 size_t cur = atomic_load_zu(p, ATOMIC_RELAXED); 196 atomic_store_zu(p, cur + x, ATOMIC_RELAXED); [all...] |
| arena_inlines_a.h | 11 atomic_fetch_add_zu(&arena->stats.internal, size, ATOMIC_RELAXED); 16 atomic_fetch_sub_zu(&arena->stats.internal, size, ATOMIC_RELAXED); 21 return atomic_load_zu(&arena->stats.internal, ATOMIC_RELAXED);
|
| atomic.h | 26 * atomic_store_pi(&atomic_ptr_to_int, some_ptr, ATOMIC_RELAXED); 41 #define ATOMIC_RELAXED atomic_memory_order_relaxed
|
| /src/external/bsd/jemalloc.old/include/jemalloc/internal/ |
| prof_inlines_a.h | 20 a0 = atomic_load_u64(&prof_accum->accumbytes, ATOMIC_RELAXED); 29 a1, ATOMIC_RELAXED, ATOMIC_RELAXED)); 56 a0 = atomic_load_u64(&prof_accum->accumbytes, ATOMIC_RELAXED); 61 a1, ATOMIC_RELAXED, ATOMIC_RELAXED));
|
| prng.h | 73 state0 = atomic_load_u32(state, ATOMIC_RELAXED); 79 state1, ATOMIC_RELAXED, ATOMIC_RELAXED)); 82 atomic_store_u32(state, state1, ATOMIC_RELAXED); 110 state0 = atomic_load_zu(state, ATOMIC_RELAXED); 116 state1, ATOMIC_RELAXED, ATOMIC_RELAXED)); 119 atomic_store_zu(state, state1, ATOMIC_RELAXED);
|
| arena_stats.h | 132 return atomic_load_u64(p, ATOMIC_RELAXED); 143 atomic_fetch_add_u64(p, x, ATOMIC_RELAXED); 154 UNUSED uint64_t r = atomic_fetch_sub_u64(p, x, ATOMIC_RELAXED); 171 uint64_t cur_dst = atomic_load_u64(dst, ATOMIC_RELAXED); 172 atomic_store_u64(dst, src + cur_dst, ATOMIC_RELAXED); 181 return atomic_load_zu(p, ATOMIC_RELAXED); 184 return atomic_load_zu(p, ATOMIC_RELAXED); 192 atomic_fetch_add_zu(p, x, ATOMIC_RELAXED); 195 size_t cur = atomic_load_zu(p, ATOMIC_RELAXED); 196 atomic_store_zu(p, cur + x, ATOMIC_RELAXED); [all...] |
| arena_inlines_a.h | 11 atomic_fetch_add_zu(&arena->stats.internal, size, ATOMIC_RELAXED); 16 atomic_fetch_sub_zu(&arena->stats.internal, size, ATOMIC_RELAXED); 21 return atomic_load_zu(&arena->stats.internal, ATOMIC_RELAXED);
|
| atomic.h | 26 * atomic_store_pi(&atomic_ptr_to_int, some_ptr, ATOMIC_RELAXED); 41 #define ATOMIC_RELAXED atomic_memory_order_relaxed
|
| /src/external/bsd/jemalloc/dist/src/ |
| eset.c | 19 atomic_store_zu(&bin_stats->nextents, 0, ATOMIC_RELAXED); 20 atomic_store_zu(&bin_stats->nbytes, 0, ATOMIC_RELAXED); 36 return atomic_load_zu(&eset->npages, ATOMIC_RELAXED); 41 return atomic_load_zu(&eset->bin_stats[pind].nextents, ATOMIC_RELAXED); 46 return atomic_load_zu(&eset->bin_stats[pind].nbytes, ATOMIC_RELAXED); 52 ATOMIC_RELAXED); 54 ATOMIC_RELAXED); 55 cur = atomic_load_zu(&eset->bin_stats[pind].nbytes, ATOMIC_RELAXED); 57 ATOMIC_RELAXED); 63 ATOMIC_RELAXED); [all...] |