HomeSort by: relevance | last modified time | path
    Searched refs:ATOMIC_RELAXED (Results 1 - 25 of 82) sorted by relevancy

1 2 3 4

  /src/external/bsd/jemalloc/dist/include/jemalloc/internal/
arena_inlines_a.h 11 atomic_fetch_add_zu(&arena->stats.internal, size, ATOMIC_RELAXED);
16 atomic_fetch_sub_zu(&arena->stats.internal, size, ATOMIC_RELAXED);
21 return atomic_load_zu(&arena->stats.internal, ATOMIC_RELAXED);
lockedint.h 61 return atomic_load_u64(&p->val, ATOMIC_RELAXED);
72 atomic_fetch_add_u64(&p->val, x, ATOMIC_RELAXED);
83 uint64_t r = atomic_fetch_sub_u64(&p->val, x, ATOMIC_RELAXED);
99 before = atomic_load_u64(&p->val, ATOMIC_RELAXED);
108 ATOMIC_RELAXED, ATOMIC_RELAXED));
129 uint64_t cur_dst = atomic_load_u64(&dst->val, ATOMIC_RELAXED);
130 atomic_store_u64(&dst->val, src + cur_dst, ATOMIC_RELAXED);
139 return atomic_load_u64(&p->val, ATOMIC_RELAXED);
148 atomic_store_u64(&p->val, x, ATOMIC_RELAXED);
    [all...]
atomic.h 33 * atomic_store_pi(&atomic_ptr_to_int, some_ptr, ATOMIC_RELAXED);
48 #define ATOMIC_RELAXED atomic_memory_order_relaxed
63 type oldval = atomic_load_##short_type(a, ATOMIC_RELAXED); \
65 atomic_store_##short_type(a, newval, ATOMIC_RELAXED); \
70 type oldval = atomic_load_##short_type(a, ATOMIC_RELAXED); \
72 atomic_store_##short_type(a, newval, ATOMIC_RELAXED); \
seq.h 26 size_t old_seq = atomic_load_zu(&dst->seq, ATOMIC_RELAXED); \
27 atomic_store_zu(&dst->seq, old_seq + 1, ATOMIC_RELAXED); \
30 atomic_store_zu(&dst->data[i], buf[i], ATOMIC_RELAXED); \
44 buf[i] = atomic_load_zu(&src->data[i], ATOMIC_RELAXED); \
47 size_t seq2 = atomic_load_zu(&src->seq, ATOMIC_RELAXED); \
mpsc_queue.h 56 atomic_store_p(&queue->tail, NULL, ATOMIC_RELAXED); \
66 void* cur_tail = atomic_load_p(&queue->tail, ATOMIC_RELAXED); \
79 &cur_tail, last, ATOMIC_RELEASE, ATOMIC_RELAXED)); \
92 a_type *tail = atomic_load_p(&queue->tail, ATOMIC_RELAXED); \
  /src/external/bsd/jemalloc/include/jemalloc/internal/
arena_inlines_a.h 11 atomic_fetch_add_zu(&arena->stats.internal, size, ATOMIC_RELAXED);
16 atomic_fetch_sub_zu(&arena->stats.internal, size, ATOMIC_RELAXED);
21 return atomic_load_zu(&arena->stats.internal, ATOMIC_RELAXED);
prof_inlines_a.h 20 a0 = atomic_load_u64(&prof_accum->accumbytes, ATOMIC_RELAXED);
29 a1, ATOMIC_RELAXED, ATOMIC_RELAXED));
56 a0 = atomic_load_u64(&prof_accum->accumbytes, ATOMIC_RELAXED);
61 a1, ATOMIC_RELAXED, ATOMIC_RELAXED));
lockedint.h 61 return atomic_load_u64(&p->val, ATOMIC_RELAXED);
72 atomic_fetch_add_u64(&p->val, x, ATOMIC_RELAXED);
83 uint64_t r = atomic_fetch_sub_u64(&p->val, x, ATOMIC_RELAXED);
99 before = atomic_load_u64(&p->val, ATOMIC_RELAXED);
108 ATOMIC_RELAXED, ATOMIC_RELAXED));
129 uint64_t cur_dst = atomic_load_u64(&dst->val, ATOMIC_RELAXED);
130 atomic_store_u64(&dst->val, src + cur_dst, ATOMIC_RELAXED);
139 return atomic_load_u64(&p->val, ATOMIC_RELAXED);
148 atomic_store_u64(&p->val, x, ATOMIC_RELAXED);
    [all...]
atomic.h 33 * atomic_store_pi(&atomic_ptr_to_int, some_ptr, ATOMIC_RELAXED);
48 #define ATOMIC_RELAXED atomic_memory_order_relaxed
63 type oldval = atomic_load_##short_type(a, ATOMIC_RELAXED); \
65 atomic_store_##short_type(a, newval, ATOMIC_RELAXED); \
70 type oldval = atomic_load_##short_type(a, ATOMIC_RELAXED); \
72 atomic_store_##short_type(a, newval, ATOMIC_RELAXED); \
seq.h 26 size_t old_seq = atomic_load_zu(&dst->seq, ATOMIC_RELAXED); \
27 atomic_store_zu(&dst->seq, old_seq + 1, ATOMIC_RELAXED); \
30 atomic_store_zu(&dst->data[i], buf[i], ATOMIC_RELAXED); \
44 buf[i] = atomic_load_zu(&src->data[i], ATOMIC_RELAXED); \
47 size_t seq2 = atomic_load_zu(&src->seq, ATOMIC_RELAXED); \
  /src/external/bsd/jemalloc/dist/test/unit/
atomic.c 26 val = atomic_load_##ta(&atom, ATOMIC_RELAXED); \
30 atomic_store_##ta(&atom, val1, ATOMIC_RELAXED); \
31 atomic_store_##ta(&atom, val2, ATOMIC_RELAXED); \
32 val = atomic_load_##ta(&atom, ATOMIC_RELAXED); \
36 atomic_store_##ta(&atom, val1, ATOMIC_RELAXED); \
37 val = atomic_exchange_##ta(&atom, val2, ATOMIC_RELAXED); \
39 val = atomic_load_##ta(&atom, ATOMIC_RELAXED); \
46 atomic_store_##ta(&atom, val1, ATOMIC_RELAXED); \
51 &expected, val3, ATOMIC_RELAXED, ATOMIC_RELAXED); \
    [all...]
edata_cache.c 28 assert_zu_eq(atomic_load_zu(&ec.count, ATOMIC_RELAXED), 0, "");
35 assert_zu_eq(atomic_load_zu(&ec.count, ATOMIC_RELAXED), 0, "");
39 assert_zu_eq(atomic_load_zu(&ec.count, ATOMIC_RELAXED), 1, "");
44 assert_zu_eq(atomic_load_zu(&ec.count, ATOMIC_RELAXED), 0, "");
70 expect_zu_eq(atomic_load_zu(&ec.count, ATOMIC_RELAXED), 0, "");
75 expect_zu_eq(atomic_load_zu(&ec.count, ATOMIC_RELAXED), 0, "");
79 expect_zu_eq(atomic_load_zu(&ec.count, ATOMIC_RELAXED), 0, "");
83 expect_zu_eq(atomic_load_zu(&ec.count, ATOMIC_RELAXED), 0, "");
88 expect_zu_eq(atomic_load_zu(&ec.count, ATOMIC_RELAXED), 0, "");
92 expect_zu_eq(atomic_load_zu(&ec.count, ATOMIC_RELAXED), 0, "")
    [all...]
  /src/external/bsd/jemalloc.old/dist/test/unit/
atomic.c 26 val = atomic_load_##ta(&atom, ATOMIC_RELAXED); \
30 atomic_store_##ta(&atom, val1, ATOMIC_RELAXED); \
31 atomic_store_##ta(&atom, val2, ATOMIC_RELAXED); \
32 val = atomic_load_##ta(&atom, ATOMIC_RELAXED); \
36 atomic_store_##ta(&atom, val1, ATOMIC_RELAXED); \
37 val = atomic_exchange_##ta(&atom, val2, ATOMIC_RELAXED); \
39 val = atomic_load_##ta(&atom, ATOMIC_RELAXED); \
46 atomic_store_##ta(&atom, val1, ATOMIC_RELAXED); \
51 &expected, val3, ATOMIC_RELAXED, ATOMIC_RELAXED); \
    [all...]
prng.c 9 atomic_store_u32(&sa, 42, ATOMIC_RELAXED);
11 atomic_store_u32(&sa, 42, ATOMIC_RELAXED);
16 atomic_store_u32(&sb, 42, ATOMIC_RELAXED);
21 atomic_store_u32(&sa, 42, ATOMIC_RELAXED);
27 atomic_store_u32(&sa, 42, ATOMIC_RELAXED);
30 atomic_store_u32(&sb, 42, ATOMIC_RELAXED);
82 atomic_store_zu(&sa, 42, ATOMIC_RELAXED);
84 atomic_store_zu(&sa, 42, ATOMIC_RELAXED);
89 atomic_store_zu(&sb, 42, ATOMIC_RELAXED);
94 atomic_store_zu(&sa, 42, ATOMIC_RELAXED);
    [all...]
  /src/external/bsd/jemalloc.old/dist/include/jemalloc/internal/
prof_inlines_a.h 20 a0 = atomic_load_u64(&prof_accum->accumbytes, ATOMIC_RELAXED);
29 a1, ATOMIC_RELAXED, ATOMIC_RELAXED));
56 a0 = atomic_load_u64(&prof_accum->accumbytes, ATOMIC_RELAXED);
61 a1, ATOMIC_RELAXED, ATOMIC_RELAXED));
prng.h 73 state0 = atomic_load_u32(state, ATOMIC_RELAXED);
79 state1, ATOMIC_RELAXED, ATOMIC_RELAXED));
82 atomic_store_u32(state, state1, ATOMIC_RELAXED);
110 state0 = atomic_load_zu(state, ATOMIC_RELAXED);
116 state1, ATOMIC_RELAXED, ATOMIC_RELAXED));
119 atomic_store_zu(state, state1, ATOMIC_RELAXED);
arena_stats.h 132 return atomic_load_u64(p, ATOMIC_RELAXED);
143 atomic_fetch_add_u64(p, x, ATOMIC_RELAXED);
154 UNUSED uint64_t r = atomic_fetch_sub_u64(p, x, ATOMIC_RELAXED);
171 uint64_t cur_dst = atomic_load_u64(dst, ATOMIC_RELAXED);
172 atomic_store_u64(dst, src + cur_dst, ATOMIC_RELAXED);
181 return atomic_load_zu(p, ATOMIC_RELAXED);
184 return atomic_load_zu(p, ATOMIC_RELAXED);
192 atomic_fetch_add_zu(p, x, ATOMIC_RELAXED);
195 size_t cur = atomic_load_zu(p, ATOMIC_RELAXED);
196 atomic_store_zu(p, cur + x, ATOMIC_RELAXED);
    [all...]
arena_inlines_a.h 11 atomic_fetch_add_zu(&arena->stats.internal, size, ATOMIC_RELAXED);
16 atomic_fetch_sub_zu(&arena->stats.internal, size, ATOMIC_RELAXED);
21 return atomic_load_zu(&arena->stats.internal, ATOMIC_RELAXED);
atomic.h 26 * atomic_store_pi(&atomic_ptr_to_int, some_ptr, ATOMIC_RELAXED);
41 #define ATOMIC_RELAXED atomic_memory_order_relaxed
  /src/external/bsd/jemalloc.old/include/jemalloc/internal/
prof_inlines_a.h 20 a0 = atomic_load_u64(&prof_accum->accumbytes, ATOMIC_RELAXED);
29 a1, ATOMIC_RELAXED, ATOMIC_RELAXED));
56 a0 = atomic_load_u64(&prof_accum->accumbytes, ATOMIC_RELAXED);
61 a1, ATOMIC_RELAXED, ATOMIC_RELAXED));
prng.h 73 state0 = atomic_load_u32(state, ATOMIC_RELAXED);
79 state1, ATOMIC_RELAXED, ATOMIC_RELAXED));
82 atomic_store_u32(state, state1, ATOMIC_RELAXED);
110 state0 = atomic_load_zu(state, ATOMIC_RELAXED);
116 state1, ATOMIC_RELAXED, ATOMIC_RELAXED));
119 atomic_store_zu(state, state1, ATOMIC_RELAXED);
arena_stats.h 132 return atomic_load_u64(p, ATOMIC_RELAXED);
143 atomic_fetch_add_u64(p, x, ATOMIC_RELAXED);
154 UNUSED uint64_t r = atomic_fetch_sub_u64(p, x, ATOMIC_RELAXED);
171 uint64_t cur_dst = atomic_load_u64(dst, ATOMIC_RELAXED);
172 atomic_store_u64(dst, src + cur_dst, ATOMIC_RELAXED);
181 return atomic_load_zu(p, ATOMIC_RELAXED);
184 return atomic_load_zu(p, ATOMIC_RELAXED);
192 atomic_fetch_add_zu(p, x, ATOMIC_RELAXED);
195 size_t cur = atomic_load_zu(p, ATOMIC_RELAXED);
196 atomic_store_zu(p, cur + x, ATOMIC_RELAXED);
    [all...]
arena_inlines_a.h 11 atomic_fetch_add_zu(&arena->stats.internal, size, ATOMIC_RELAXED);
16 atomic_fetch_sub_zu(&arena->stats.internal, size, ATOMIC_RELAXED);
21 return atomic_load_zu(&arena->stats.internal, ATOMIC_RELAXED);
atomic.h 26 * atomic_store_pi(&atomic_ptr_to_int, some_ptr, ATOMIC_RELAXED);
41 #define ATOMIC_RELAXED atomic_memory_order_relaxed
  /src/external/bsd/jemalloc/dist/src/
eset.c 19 atomic_store_zu(&bin_stats->nextents, 0, ATOMIC_RELAXED);
20 atomic_store_zu(&bin_stats->nbytes, 0, ATOMIC_RELAXED);
36 return atomic_load_zu(&eset->npages, ATOMIC_RELAXED);
41 return atomic_load_zu(&eset->bin_stats[pind].nextents, ATOMIC_RELAXED);
46 return atomic_load_zu(&eset->bin_stats[pind].nbytes, ATOMIC_RELAXED);
52 ATOMIC_RELAXED);
54 ATOMIC_RELAXED);
55 cur = atomic_load_zu(&eset->bin_stats[pind].nbytes, ATOMIC_RELAXED);
57 ATOMIC_RELAXED);
63 ATOMIC_RELAXED);
    [all...]

Completed in 52 milliseconds

1 2 3 4