HomeSort by: relevance | last modified time | path
    Searched refs:tcache (Results 1 - 25 of 65) sorted by relevancy

1 2 3

  /src/external/bsd/jemalloc.old/dist/src/
tcache.c 16 static unsigned stack_nelms; /* Total stack elms per tcache. */
40 tcache_event_hard(tsd_t *tsd, tcache_t *tcache) {
41 szind_t binind = tcache->next_gc_bin;
45 tbin = tcache_small_bin_get(tcache, binind);
47 tbin = tcache_large_bin_get(tcache, binind);
54 tcache_bin_flush_small(tsd, tcache, tbin, binind,
63 (tcache->lg_fill_div[binind] + 1)) >= 1) {
64 tcache->lg_fill_div[binind]++;
68 - tbin->low_water + (tbin->low_water >> 2), tcache);
75 if (binind < NBINS && tcache->lg_fill_div[binind] > 1)
393 tcache_t *tcache = tsd_tcachep_get_unsafe(tsd); local
419 tcache_arena_associate(tsd_tsdn(tsd), tcache, arena); local
424 tcache_arena_associate(tsd_tsdn(tsd), tcache, arena); local
435 tcache_t *tcache; local
454 tcache_arena_associate(tsd_tsdn(tsd), tcache, arena_ichoose(tsd, NULL)); local
496 tcache_arena_dissociate(tsd_tsdn(tsd), tcache); local
506 idalloctm(tsd_tsdn(tsd), tcache, NULL, NULL, true, true); local
513 tcache_t *tcache = tsd_tcachep_get(tsd); local
591 tcache_t *tcache = tcache_create_explicit(tsd); local
625 tcache_t *tcache = elm->tcache; local
633 tcache_t *tcache = tcaches_elm_remove(tsd, &tcaches[ind]); local
644 tcache_t *tcache = tcaches_elm_remove(tsd, elm); local
    [all...]
  /src/external/bsd/jemalloc.old/dist/include/jemalloc/internal/
tcache_externs.h 12 * Number of tcache bins. There are NBINS small-object bins, plus 0 or more
21 * Explicit tcaches, managed via the tcache.{create,flush,destroy} mallctls and
26 * explicit tcache is created without a disproportionate impact on memory usage.
31 void tcache_event_hard(tsd_t *tsd, tcache_t *tcache);
32 void *tcache_alloc_small_hard(tsdn_t *tsdn, arena_t *arena, tcache_t *tcache,
34 void tcache_bin_flush_small(tsd_t *tsd, tcache_t *tcache, cache_bin_t *tbin,
37 unsigned rem, tcache_t *tcache);
38 void tcache_arena_reassociate(tsdn_t *tsdn, tcache_t *tcache,
42 void tcache_stats_merge(tsdn_t *tsdn, tcache_t *tcache, arena_t *arena);
47 void tcache_arena_associate(tsdn_t *tsdn, tcache_t *tcache, arena_t *arena)
    [all...]
tcache_inlines.h 31 tcache_event(tsd_t *tsd, tcache_t *tcache) {
36 if (unlikely(ticker_tick(&tcache->gc_ticker))) {
37 tcache_event_hard(tsd, tcache);
42 tcache_alloc_small(tsd_t *tsd, arena_t *arena, tcache_t *tcache,
50 bin = tcache_small_bin_get(tcache, binind);
60 ret = tcache_alloc_small_hard(tsd_tsdn(tsd), arena, tcache,
97 tcache->prof_accumbytes += usize;
99 tcache_event(tsd, tcache);
104 tcache_alloc_large(tsd_t *tsd, arena_t *arena, tcache_t *tcache, size_t size,
111 bin = tcache_large_bin_get(tcache, binind)
    [all...]
jemalloc_internal_inlines_c.h 41 iallocztm(tsdn_t *tsdn, size_t size, szind_t ind, bool zero, tcache_t *tcache,
46 assert(!is_internal || tcache == NULL);
53 ret = arena_malloc(tsdn, arena, size, ind, zero, tcache, slow_path);
68 tcache_t *tcache, bool is_internal, arena_t *arena) {
73 assert(!is_internal || tcache == NULL);
78 ret = arena_palloc(tsdn, arena, usize, alignment, zero, tcache);
88 tcache_t *tcache, arena_t *arena) {
89 return ipallocztm(tsdn, usize, alignment, zero, tcache, false, arena);
104 idalloctm(tsdn_t *tsdn, void *ptr, tcache_t *tcache, alloc_ctx_t *alloc_ctx,
107 assert(!is_internal || tcache == NULL)
    [all...]
jemalloc_internal_inlines_b.h 25 tcache_t *tcache = tcache_get(tsd); local
26 if (tcache->arena != NULL) {
28 assert(tcache->arena ==
30 if (tcache->arena != ret) {
32 tcache, ret); local
35 tcache_arena_associate(tsd_tsdn(tsd), tcache, local
tcache_structs.h 21 * During tcache initialization, the avail pointer in each element of
34 * tcache definition. This enables arenas to aggregate stats across
35 * tcaches without having a tcache dependency.
39 /* The arena this tcache is associated with. */
53 /* Linkage for list of available (previously used) explicit tcache IDs. */
56 tcache_t *tcache; member in union:tcaches_s::__anon6872
arena_inlines_a.h 48 /* Set new arena/tcache associations. */
50 tcache_t *tcache = tcache_get(tsd); local
51 if (tcache != NULL) {
52 tcache_arena_reassociate(tsd_tsdn(tsd), tcache, local
arena_inlines_b.h 88 tcache_t *tcache, bool slow_path) {
89 assert(!tsdn_null(tsdn) || tcache == NULL);
92 if (likely(tcache != NULL)) {
95 tcache, size, ind, zero, slow_path);
99 tcache, size, ind, zero, slow_path);
190 arena_dalloc(tsdn_t *tsdn, void *ptr, tcache_t *tcache,
192 assert(!tsdn_null(tsdn) || tcache == NULL);
195 if (unlikely(tcache == NULL)) {
224 tcache_dalloc_small(tsdn_tsd(tsdn), tcache, ptr, szind,
229 arena_dalloc_promoted(tsdn, ptr, tcache,
    [all...]
  /src/external/bsd/jemalloc.old/include/jemalloc/internal/
tcache_externs.h 12 * Number of tcache bins. There are NBINS small-object bins, plus 0 or more
21 * Explicit tcaches, managed via the tcache.{create,flush,destroy} mallctls and
26 * explicit tcache is created without a disproportionate impact on memory usage.
31 void tcache_event_hard(tsd_t *tsd, tcache_t *tcache);
32 void *tcache_alloc_small_hard(tsdn_t *tsdn, arena_t *arena, tcache_t *tcache,
34 void tcache_bin_flush_small(tsd_t *tsd, tcache_t *tcache, cache_bin_t *tbin,
37 unsigned rem, tcache_t *tcache);
38 void tcache_arena_reassociate(tsdn_t *tsdn, tcache_t *tcache,
42 void tcache_stats_merge(tsdn_t *tsdn, tcache_t *tcache, arena_t *arena);
47 void tcache_arena_associate(tsdn_t *tsdn, tcache_t *tcache, arena_t *arena)
    [all...]
tcache_inlines.h 31 tcache_event(tsd_t *tsd, tcache_t *tcache) {
36 if (unlikely(ticker_tick(&tcache->gc_ticker))) {
37 tcache_event_hard(tsd, tcache);
42 tcache_alloc_small(tsd_t *tsd, arena_t *arena, tcache_t *tcache,
50 bin = tcache_small_bin_get(tcache, binind);
60 ret = tcache_alloc_small_hard(tsd_tsdn(tsd), arena, tcache,
97 tcache->prof_accumbytes += usize;
99 tcache_event(tsd, tcache);
104 tcache_alloc_large(tsd_t *tsd, arena_t *arena, tcache_t *tcache, size_t size,
111 bin = tcache_large_bin_get(tcache, binind)
    [all...]
jemalloc_internal_inlines_c.h 41 iallocztm(tsdn_t *tsdn, size_t size, szind_t ind, bool zero, tcache_t *tcache,
46 assert(!is_internal || tcache == NULL);
53 ret = arena_malloc(tsdn, arena, size, ind, zero, tcache, slow_path);
68 tcache_t *tcache, bool is_internal, arena_t *arena) {
73 assert(!is_internal || tcache == NULL);
78 ret = arena_palloc(tsdn, arena, usize, alignment, zero, tcache);
88 tcache_t *tcache, arena_t *arena) {
89 return ipallocztm(tsdn, usize, alignment, zero, tcache, false, arena);
104 idalloctm(tsdn_t *tsdn, void *ptr, tcache_t *tcache, alloc_ctx_t *alloc_ctx,
107 assert(!is_internal || tcache == NULL)
    [all...]
jemalloc_internal_inlines_b.h 25 tcache_t *tcache = tcache_get(tsd); local
26 if (tcache->arena != NULL) {
28 assert(tcache->arena ==
30 if (tcache->arena != ret) {
32 tcache, ret); local
35 tcache_arena_associate(tsd_tsdn(tsd), tcache, local
tcache_structs.h 21 * During tcache initialization, the avail pointer in each element of
34 * tcache definition. This enables arenas to aggregate stats across
35 * tcaches without having a tcache dependency.
39 /* The arena this tcache is associated with. */
53 /* Linkage for list of available (previously used) explicit tcache IDs. */
56 tcache_t *tcache; member in union:tcaches_s::__anon6925
arena_inlines_a.h 48 /* Set new arena/tcache associations. */
50 tcache_t *tcache = tcache_get(tsd); local
51 if (tcache != NULL) {
52 tcache_arena_reassociate(tsd_tsdn(tsd), tcache, local
  /src/external/bsd/jemalloc/dist/include/jemalloc/internal/
tcache_structs.h 11 * The tcache state is split into the slow and hot path data. Each has a
14 * tcache, and as part of a dynamic allocation for manual allocations. Keeping
16 * splitting up the tcache [de]allocation code into those paths called with the
17 * TSD tcache and those called with a manual tcache.
26 * tcache definition. This enables arenas to aggregate stats across
27 * tcaches without having a tcache dependency.
31 /* The arena this tcache is associated with. */
52 tcache_t *tcache; member in struct:tcache_slow_s
60 /* Linkage for list of available (previously used) explicit tcache IDs. *
63 tcache_t *tcache; member in union:tcaches_s::__anon6764
    [all...]
tcache_externs.h 17 * Number of tcache bins. There are SC_NBINS small-object bins, plus 0 or more
28 * Explicit tcaches, managed via the tcache.{create,flush,destroy} mallctls and
33 * explicit tcache is created without a disproportionate impact on memory usage.
38 void *tcache_alloc_small_hard(tsdn_t *tsdn, arena_t *arena, tcache_t *tcache,
41 void tcache_bin_flush_small(tsd_t *tsd, tcache_t *tcache, cache_bin_t *tbin,
43 void tcache_bin_flush_large(tsd_t *tsd, tcache_t *tcache, cache_bin_t *tbin,
45 void tcache_bin_flush_stashed(tsd_t *tsd, tcache_t *tcache, cache_bin_t *bin,
48 tcache_t *tcache, arena_t *arena);
51 void tcache_stats_merge(tsdn_t *tsdn, tcache_t *tcache, arena_t *arena);
57 tcache_t *tcache, arena_t *arena)
    [all...]
jemalloc_internal_inlines_c.h 44 iallocztm(tsdn_t *tsdn, size_t size, szind_t ind, bool zero, tcache_t *tcache,
48 assert(!is_internal || tcache == NULL);
55 ret = arena_malloc(tsdn, arena, size, ind, zero, tcache, slow_path);
70 tcache_t *tcache, bool is_internal, arena_t *arena) {
75 assert(!is_internal || tcache == NULL);
80 ret = arena_palloc(tsdn, arena, usize, alignment, zero, tcache);
90 tcache_t *tcache, arena_t *arena) {
91 return ipallocztm(tsdn, usize, alignment, zero, tcache, false, arena);
106 idalloctm(tsdn_t *tsdn, void *ptr, tcache_t *tcache,
109 assert(!is_internal || tcache == NULL)
    [all...]
tcache_inlines.h 45 tcache_alloc_small(tsd_t *tsd, arena_t *arena, tcache_t *tcache,
51 cache_bin_t *bin = &tcache->bins[binind];
65 tcache_bin_flush_stashed(tsd, tcache, bin, binind,
68 ret = tcache_alloc_small_hard(tsd_tsdn(tsd), arena, tcache,
88 tcache_alloc_large(tsd_t *tsd, arena_t *arena, tcache_t *tcache, size_t size,
94 cache_bin_t *bin = &tcache->bins[binind];
106 tcache_bin_flush_stashed(tsd, tcache, bin, binind,
129 tcache_dalloc_small(tsd_t *tsd, tcache_t *tcache, void *ptr, szind_t binind,
133 cache_bin_t *bin = &tcache->bins[binind];
157 tcache_bin_flush_small(tsd, tcache, bin, binind, remain)
    [all...]
jemalloc_internal_inlines_b.h 18 /* Set new arena/tcache associations. */
20 tcache_t *tcache = tcache_get(tsd); local
21 if (tcache != NULL) {
24 tcache, newarena); local
50 tcache_t *tcache = tsd_tcachep_get(tsd); local
57 tcache_slow, tcache, ret); local
61 tcache_slow, tcache, ret); local
  /src/external/bsd/jemalloc/include/jemalloc/internal/
tcache_structs.h 11 * The tcache state is split into the slow and hot path data. Each has a
14 * tcache, and as part of a dynamic allocation for manual allocations. Keeping
16 * splitting up the tcache [de]allocation code into those paths called with the
17 * TSD tcache and those called with a manual tcache.
26 * tcache definition. This enables arenas to aggregate stats across
27 * tcaches without having a tcache dependency.
31 /* The arena this tcache is associated with. */
52 tcache_t *tcache; member in struct:tcache_slow_s
60 /* Linkage for list of available (previously used) explicit tcache IDs. *
63 tcache_t *tcache; member in union:tcaches_s::__anon6843
    [all...]
tcache_externs.h 17 * Number of tcache bins. There are SC_NBINS small-object bins, plus 0 or more
28 * Explicit tcaches, managed via the tcache.{create,flush,destroy} mallctls and
33 * explicit tcache is created without a disproportionate impact on memory usage.
38 void *tcache_alloc_small_hard(tsdn_t *tsdn, arena_t *arena, tcache_t *tcache,
41 void tcache_bin_flush_small(tsd_t *tsd, tcache_t *tcache, cache_bin_t *tbin,
43 void tcache_bin_flush_large(tsd_t *tsd, tcache_t *tcache, cache_bin_t *tbin,
45 void tcache_bin_flush_stashed(tsd_t *tsd, tcache_t *tcache, cache_bin_t *bin,
48 tcache_t *tcache, arena_t *arena);
51 void tcache_stats_merge(tsdn_t *tsdn, tcache_t *tcache, arena_t *arena);
57 tcache_t *tcache, arena_t *arena)
    [all...]
jemalloc_internal_inlines_c.h 44 iallocztm(tsdn_t *tsdn, size_t size, szind_t ind, bool zero, tcache_t *tcache,
48 assert(!is_internal || tcache == NULL);
55 ret = arena_malloc(tsdn, arena, size, ind, zero, tcache, slow_path);
70 tcache_t *tcache, bool is_internal, arena_t *arena) {
75 assert(!is_internal || tcache == NULL);
80 ret = arena_palloc(tsdn, arena, usize, alignment, zero, tcache);
90 tcache_t *tcache, arena_t *arena) {
91 return ipallocztm(tsdn, usize, alignment, zero, tcache, false, arena);
106 idalloctm(tsdn_t *tsdn, void *ptr, tcache_t *tcache,
109 assert(!is_internal || tcache == NULL)
    [all...]
tcache_inlines.h 45 tcache_alloc_small(tsd_t *tsd, arena_t *arena, tcache_t *tcache,
51 cache_bin_t *bin = &tcache->bins[binind];
65 tcache_bin_flush_stashed(tsd, tcache, bin, binind,
68 ret = tcache_alloc_small_hard(tsd_tsdn(tsd), arena, tcache,
88 tcache_alloc_large(tsd_t *tsd, arena_t *arena, tcache_t *tcache, size_t size,
94 cache_bin_t *bin = &tcache->bins[binind];
106 tcache_bin_flush_stashed(tsd, tcache, bin, binind,
129 tcache_dalloc_small(tsd_t *tsd, tcache_t *tcache, void *ptr, szind_t binind,
133 cache_bin_t *bin = &tcache->bins[binind];
157 tcache_bin_flush_small(tsd, tcache, bin, binind, remain)
    [all...]
jemalloc_internal_inlines_b.h 18 /* Set new arena/tcache associations. */
20 tcache_t *tcache = tcache_get(tsd); local
21 if (tcache != NULL) {
24 tcache, newarena); local
50 tcache_t *tcache = tsd_tcachep_get(tsd); local
57 tcache_slow, tcache, ret); local
61 tcache_slow, tcache, ret); local
  /src/external/bsd/jemalloc/dist/src/
tcache.c 24 * We attempt to make the number of slots in a tcache bin for a given size class
26 * the multiplier is 2 (i.e. we set the maximum number of objects in the tcache
34 * Number of allocation bytes between tcache incremental GCs. Again, this
62 /* Total stack size required (per tcache). Include the padding above. */
123 tcache_gc_small(tsd_t *tsd, tcache_slow_t *tcache_slow, tcache_t *tcache,
128 cache_bin_t *cache_bin = &tcache->bins[szind];
148 tcache_bin_flush_small(tsd, tcache, cache_bin, szind,
162 tcache_gc_large(tsd_t *tsd, tcache_slow_t *tcache_slow, tcache_t *tcache,
166 cache_bin_t *cache_bin = &tcache->bins[szind];
171 tcache_bin_flush_large(tsd, tcache, cache_bin, szind
177 tcache_t *tcache = tcache_get(tsd); local
711 tcache_t *tcache = tsd_tcachep_get_unsafe(tsd); local
737 tcache_arena_associate(tsd_tsdn(tsd), tcache_slow, tcache, local
744 tcache, arena); local
771 tcache_t *tcache = (void *)((uintptr_t)mem + tcache_bin_alloc_size); local
776 tcache_arena_associate(tsd_tsdn(tsd), tcache_slow, tcache, local
811 tcache_arena_dissociate(tsd_tsdn(tsd), tcache_slow, tcache); local
843 tcache_t *tcache = tsd_tcachep_get(tsd); local
920 tcache_t *tcache = tcache_create_explicit(tsd); local
953 tcache_t *tcache = elm->tcache; local
969 tcache_t *tcache = tcaches_elm_remove(tsd, &tcaches[ind], true); local
981 tcache_t *tcache = tcaches_elm_remove(tsd, elm, false); local
    [all...]

Completed in 23 milliseconds

1 2 3