| /src/external/bsd/jemalloc.old/dist/src/ |
| tcache.c | 16 static unsigned stack_nelms; /* Total stack elms per tcache. */ 40 tcache_event_hard(tsd_t *tsd, tcache_t *tcache) { 41 szind_t binind = tcache->next_gc_bin; 45 tbin = tcache_small_bin_get(tcache, binind); 47 tbin = tcache_large_bin_get(tcache, binind); 54 tcache_bin_flush_small(tsd, tcache, tbin, binind, 63 (tcache->lg_fill_div[binind] + 1)) >= 1) { 64 tcache->lg_fill_div[binind]++; 68 - tbin->low_water + (tbin->low_water >> 2), tcache); 75 if (binind < NBINS && tcache->lg_fill_div[binind] > 1) 393 tcache_t *tcache = tsd_tcachep_get_unsafe(tsd); local 419 tcache_arena_associate(tsd_tsdn(tsd), tcache, arena); local 424 tcache_arena_associate(tsd_tsdn(tsd), tcache, arena); local 435 tcache_t *tcache; local 454 tcache_arena_associate(tsd_tsdn(tsd), tcache, arena_ichoose(tsd, NULL)); local 496 tcache_arena_dissociate(tsd_tsdn(tsd), tcache); local 506 idalloctm(tsd_tsdn(tsd), tcache, NULL, NULL, true, true); local 513 tcache_t *tcache = tsd_tcachep_get(tsd); local 591 tcache_t *tcache = tcache_create_explicit(tsd); local 625 tcache_t *tcache = elm->tcache; local 633 tcache_t *tcache = tcaches_elm_remove(tsd, &tcaches[ind]); local 644 tcache_t *tcache = tcaches_elm_remove(tsd, elm); local [all...] |
| /src/external/bsd/jemalloc.old/dist/include/jemalloc/internal/ |
| tcache_externs.h | 12 * Number of tcache bins. There are NBINS small-object bins, plus 0 or more 21 * Explicit tcaches, managed via the tcache.{create,flush,destroy} mallctls and 26 * explicit tcache is created without a disproportionate impact on memory usage. 31 void tcache_event_hard(tsd_t *tsd, tcache_t *tcache); 32 void *tcache_alloc_small_hard(tsdn_t *tsdn, arena_t *arena, tcache_t *tcache, 34 void tcache_bin_flush_small(tsd_t *tsd, tcache_t *tcache, cache_bin_t *tbin, 37 unsigned rem, tcache_t *tcache); 38 void tcache_arena_reassociate(tsdn_t *tsdn, tcache_t *tcache, 42 void tcache_stats_merge(tsdn_t *tsdn, tcache_t *tcache, arena_t *arena); 47 void tcache_arena_associate(tsdn_t *tsdn, tcache_t *tcache, arena_t *arena) [all...] |
| tcache_inlines.h | 31 tcache_event(tsd_t *tsd, tcache_t *tcache) { 36 if (unlikely(ticker_tick(&tcache->gc_ticker))) { 37 tcache_event_hard(tsd, tcache); 42 tcache_alloc_small(tsd_t *tsd, arena_t *arena, tcache_t *tcache, 50 bin = tcache_small_bin_get(tcache, binind); 60 ret = tcache_alloc_small_hard(tsd_tsdn(tsd), arena, tcache, 97 tcache->prof_accumbytes += usize; 99 tcache_event(tsd, tcache); 104 tcache_alloc_large(tsd_t *tsd, arena_t *arena, tcache_t *tcache, size_t size, 111 bin = tcache_large_bin_get(tcache, binind) [all...] |
| jemalloc_internal_inlines_c.h | 41 iallocztm(tsdn_t *tsdn, size_t size, szind_t ind, bool zero, tcache_t *tcache, 46 assert(!is_internal || tcache == NULL); 53 ret = arena_malloc(tsdn, arena, size, ind, zero, tcache, slow_path); 68 tcache_t *tcache, bool is_internal, arena_t *arena) { 73 assert(!is_internal || tcache == NULL); 78 ret = arena_palloc(tsdn, arena, usize, alignment, zero, tcache); 88 tcache_t *tcache, arena_t *arena) { 89 return ipallocztm(tsdn, usize, alignment, zero, tcache, false, arena); 104 idalloctm(tsdn_t *tsdn, void *ptr, tcache_t *tcache, alloc_ctx_t *alloc_ctx, 107 assert(!is_internal || tcache == NULL) [all...] |
| jemalloc_internal_inlines_b.h | 25 tcache_t *tcache = tcache_get(tsd); local 26 if (tcache->arena != NULL) { 28 assert(tcache->arena == 30 if (tcache->arena != ret) { 32 tcache, ret); local 35 tcache_arena_associate(tsd_tsdn(tsd), tcache, local
|
| tcache_structs.h | 21 * During tcache initialization, the avail pointer in each element of 34 * tcache definition. This enables arenas to aggregate stats across 35 * tcaches without having a tcache dependency. 39 /* The arena this tcache is associated with. */ 53 /* Linkage for list of available (previously used) explicit tcache IDs. */ 56 tcache_t *tcache; member in union:tcaches_s::__anon6872
|
| arena_inlines_a.h | 48 /* Set new arena/tcache associations. */ 50 tcache_t *tcache = tcache_get(tsd); local 51 if (tcache != NULL) { 52 tcache_arena_reassociate(tsd_tsdn(tsd), tcache, local
|
| arena_inlines_b.h | 88 tcache_t *tcache, bool slow_path) { 89 assert(!tsdn_null(tsdn) || tcache == NULL); 92 if (likely(tcache != NULL)) { 95 tcache, size, ind, zero, slow_path); 99 tcache, size, ind, zero, slow_path); 190 arena_dalloc(tsdn_t *tsdn, void *ptr, tcache_t *tcache, 192 assert(!tsdn_null(tsdn) || tcache == NULL); 195 if (unlikely(tcache == NULL)) { 224 tcache_dalloc_small(tsdn_tsd(tsdn), tcache, ptr, szind, 229 arena_dalloc_promoted(tsdn, ptr, tcache, [all...] |
| /src/external/bsd/jemalloc.old/include/jemalloc/internal/ |
| tcache_externs.h | 12 * Number of tcache bins. There are NBINS small-object bins, plus 0 or more 21 * Explicit tcaches, managed via the tcache.{create,flush,destroy} mallctls and 26 * explicit tcache is created without a disproportionate impact on memory usage. 31 void tcache_event_hard(tsd_t *tsd, tcache_t *tcache); 32 void *tcache_alloc_small_hard(tsdn_t *tsdn, arena_t *arena, tcache_t *tcache, 34 void tcache_bin_flush_small(tsd_t *tsd, tcache_t *tcache, cache_bin_t *tbin, 37 unsigned rem, tcache_t *tcache); 38 void tcache_arena_reassociate(tsdn_t *tsdn, tcache_t *tcache, 42 void tcache_stats_merge(tsdn_t *tsdn, tcache_t *tcache, arena_t *arena); 47 void tcache_arena_associate(tsdn_t *tsdn, tcache_t *tcache, arena_t *arena) [all...] |
| tcache_inlines.h | 31 tcache_event(tsd_t *tsd, tcache_t *tcache) { 36 if (unlikely(ticker_tick(&tcache->gc_ticker))) { 37 tcache_event_hard(tsd, tcache); 42 tcache_alloc_small(tsd_t *tsd, arena_t *arena, tcache_t *tcache, 50 bin = tcache_small_bin_get(tcache, binind); 60 ret = tcache_alloc_small_hard(tsd_tsdn(tsd), arena, tcache, 97 tcache->prof_accumbytes += usize; 99 tcache_event(tsd, tcache); 104 tcache_alloc_large(tsd_t *tsd, arena_t *arena, tcache_t *tcache, size_t size, 111 bin = tcache_large_bin_get(tcache, binind) [all...] |
| jemalloc_internal_inlines_c.h | 41 iallocztm(tsdn_t *tsdn, size_t size, szind_t ind, bool zero, tcache_t *tcache, 46 assert(!is_internal || tcache == NULL); 53 ret = arena_malloc(tsdn, arena, size, ind, zero, tcache, slow_path); 68 tcache_t *tcache, bool is_internal, arena_t *arena) { 73 assert(!is_internal || tcache == NULL); 78 ret = arena_palloc(tsdn, arena, usize, alignment, zero, tcache); 88 tcache_t *tcache, arena_t *arena) { 89 return ipallocztm(tsdn, usize, alignment, zero, tcache, false, arena); 104 idalloctm(tsdn_t *tsdn, void *ptr, tcache_t *tcache, alloc_ctx_t *alloc_ctx, 107 assert(!is_internal || tcache == NULL) [all...] |
| jemalloc_internal_inlines_b.h | 25 tcache_t *tcache = tcache_get(tsd); local 26 if (tcache->arena != NULL) { 28 assert(tcache->arena == 30 if (tcache->arena != ret) { 32 tcache, ret); local 35 tcache_arena_associate(tsd_tsdn(tsd), tcache, local
|
| tcache_structs.h | 21 * During tcache initialization, the avail pointer in each element of 34 * tcache definition. This enables arenas to aggregate stats across 35 * tcaches without having a tcache dependency. 39 /* The arena this tcache is associated with. */ 53 /* Linkage for list of available (previously used) explicit tcache IDs. */ 56 tcache_t *tcache; member in union:tcaches_s::__anon6925
|
| arena_inlines_a.h | 48 /* Set new arena/tcache associations. */ 50 tcache_t *tcache = tcache_get(tsd); local 51 if (tcache != NULL) { 52 tcache_arena_reassociate(tsd_tsdn(tsd), tcache, local
|
| /src/external/bsd/jemalloc/dist/include/jemalloc/internal/ |
| tcache_structs.h | 11 * The tcache state is split into the slow and hot path data. Each has a 14 * tcache, and as part of a dynamic allocation for manual allocations. Keeping 16 * splitting up the tcache [de]allocation code into those paths called with the 17 * TSD tcache and those called with a manual tcache. 26 * tcache definition. This enables arenas to aggregate stats across 27 * tcaches without having a tcache dependency. 31 /* The arena this tcache is associated with. */ 52 tcache_t *tcache; member in struct:tcache_slow_s 60 /* Linkage for list of available (previously used) explicit tcache IDs. * 63 tcache_t *tcache; member in union:tcaches_s::__anon6764 [all...] |
| tcache_externs.h | 17 * Number of tcache bins. There are SC_NBINS small-object bins, plus 0 or more 28 * Explicit tcaches, managed via the tcache.{create,flush,destroy} mallctls and 33 * explicit tcache is created without a disproportionate impact on memory usage. 38 void *tcache_alloc_small_hard(tsdn_t *tsdn, arena_t *arena, tcache_t *tcache, 41 void tcache_bin_flush_small(tsd_t *tsd, tcache_t *tcache, cache_bin_t *tbin, 43 void tcache_bin_flush_large(tsd_t *tsd, tcache_t *tcache, cache_bin_t *tbin, 45 void tcache_bin_flush_stashed(tsd_t *tsd, tcache_t *tcache, cache_bin_t *bin, 48 tcache_t *tcache, arena_t *arena); 51 void tcache_stats_merge(tsdn_t *tsdn, tcache_t *tcache, arena_t *arena); 57 tcache_t *tcache, arena_t *arena) [all...] |
| jemalloc_internal_inlines_c.h | 44 iallocztm(tsdn_t *tsdn, size_t size, szind_t ind, bool zero, tcache_t *tcache, 48 assert(!is_internal || tcache == NULL); 55 ret = arena_malloc(tsdn, arena, size, ind, zero, tcache, slow_path); 70 tcache_t *tcache, bool is_internal, arena_t *arena) { 75 assert(!is_internal || tcache == NULL); 80 ret = arena_palloc(tsdn, arena, usize, alignment, zero, tcache); 90 tcache_t *tcache, arena_t *arena) { 91 return ipallocztm(tsdn, usize, alignment, zero, tcache, false, arena); 106 idalloctm(tsdn_t *tsdn, void *ptr, tcache_t *tcache, 109 assert(!is_internal || tcache == NULL) [all...] |
| tcache_inlines.h | 45 tcache_alloc_small(tsd_t *tsd, arena_t *arena, tcache_t *tcache, 51 cache_bin_t *bin = &tcache->bins[binind]; 65 tcache_bin_flush_stashed(tsd, tcache, bin, binind, 68 ret = tcache_alloc_small_hard(tsd_tsdn(tsd), arena, tcache, 88 tcache_alloc_large(tsd_t *tsd, arena_t *arena, tcache_t *tcache, size_t size, 94 cache_bin_t *bin = &tcache->bins[binind]; 106 tcache_bin_flush_stashed(tsd, tcache, bin, binind, 129 tcache_dalloc_small(tsd_t *tsd, tcache_t *tcache, void *ptr, szind_t binind, 133 cache_bin_t *bin = &tcache->bins[binind]; 157 tcache_bin_flush_small(tsd, tcache, bin, binind, remain) [all...] |
| jemalloc_internal_inlines_b.h | 18 /* Set new arena/tcache associations. */ 20 tcache_t *tcache = tcache_get(tsd); local 21 if (tcache != NULL) { 24 tcache, newarena); local 50 tcache_t *tcache = tsd_tcachep_get(tsd); local 57 tcache_slow, tcache, ret); local 61 tcache_slow, tcache, ret); local
|
| /src/external/bsd/jemalloc/include/jemalloc/internal/ |
| tcache_structs.h | 11 * The tcache state is split into the slow and hot path data. Each has a 14 * tcache, and as part of a dynamic allocation for manual allocations. Keeping 16 * splitting up the tcache [de]allocation code into those paths called with the 17 * TSD tcache and those called with a manual tcache. 26 * tcache definition. This enables arenas to aggregate stats across 27 * tcaches without having a tcache dependency. 31 /* The arena this tcache is associated with. */ 52 tcache_t *tcache; member in struct:tcache_slow_s 60 /* Linkage for list of available (previously used) explicit tcache IDs. * 63 tcache_t *tcache; member in union:tcaches_s::__anon6843 [all...] |
| tcache_externs.h | 17 * Number of tcache bins. There are SC_NBINS small-object bins, plus 0 or more 28 * Explicit tcaches, managed via the tcache.{create,flush,destroy} mallctls and 33 * explicit tcache is created without a disproportionate impact on memory usage. 38 void *tcache_alloc_small_hard(tsdn_t *tsdn, arena_t *arena, tcache_t *tcache, 41 void tcache_bin_flush_small(tsd_t *tsd, tcache_t *tcache, cache_bin_t *tbin, 43 void tcache_bin_flush_large(tsd_t *tsd, tcache_t *tcache, cache_bin_t *tbin, 45 void tcache_bin_flush_stashed(tsd_t *tsd, tcache_t *tcache, cache_bin_t *bin, 48 tcache_t *tcache, arena_t *arena); 51 void tcache_stats_merge(tsdn_t *tsdn, tcache_t *tcache, arena_t *arena); 57 tcache_t *tcache, arena_t *arena) [all...] |
| jemalloc_internal_inlines_c.h | 44 iallocztm(tsdn_t *tsdn, size_t size, szind_t ind, bool zero, tcache_t *tcache, 48 assert(!is_internal || tcache == NULL); 55 ret = arena_malloc(tsdn, arena, size, ind, zero, tcache, slow_path); 70 tcache_t *tcache, bool is_internal, arena_t *arena) { 75 assert(!is_internal || tcache == NULL); 80 ret = arena_palloc(tsdn, arena, usize, alignment, zero, tcache); 90 tcache_t *tcache, arena_t *arena) { 91 return ipallocztm(tsdn, usize, alignment, zero, tcache, false, arena); 106 idalloctm(tsdn_t *tsdn, void *ptr, tcache_t *tcache, 109 assert(!is_internal || tcache == NULL) [all...] |
| tcache_inlines.h | 45 tcache_alloc_small(tsd_t *tsd, arena_t *arena, tcache_t *tcache, 51 cache_bin_t *bin = &tcache->bins[binind]; 65 tcache_bin_flush_stashed(tsd, tcache, bin, binind, 68 ret = tcache_alloc_small_hard(tsd_tsdn(tsd), arena, tcache, 88 tcache_alloc_large(tsd_t *tsd, arena_t *arena, tcache_t *tcache, size_t size, 94 cache_bin_t *bin = &tcache->bins[binind]; 106 tcache_bin_flush_stashed(tsd, tcache, bin, binind, 129 tcache_dalloc_small(tsd_t *tsd, tcache_t *tcache, void *ptr, szind_t binind, 133 cache_bin_t *bin = &tcache->bins[binind]; 157 tcache_bin_flush_small(tsd, tcache, bin, binind, remain) [all...] |
| jemalloc_internal_inlines_b.h | 18 /* Set new arena/tcache associations. */ 20 tcache_t *tcache = tcache_get(tsd); local 21 if (tcache != NULL) { 24 tcache, newarena); local 50 tcache_t *tcache = tsd_tcachep_get(tsd); local 57 tcache_slow, tcache, ret); local 61 tcache_slow, tcache, ret); local
|
| /src/external/bsd/jemalloc/dist/src/ |
| tcache.c | 24 * We attempt to make the number of slots in a tcache bin for a given size class 26 * the multiplier is 2 (i.e. we set the maximum number of objects in the tcache 34 * Number of allocation bytes between tcache incremental GCs. Again, this 62 /* Total stack size required (per tcache). Include the padding above. */ 123 tcache_gc_small(tsd_t *tsd, tcache_slow_t *tcache_slow, tcache_t *tcache, 128 cache_bin_t *cache_bin = &tcache->bins[szind]; 148 tcache_bin_flush_small(tsd, tcache, cache_bin, szind, 162 tcache_gc_large(tsd_t *tsd, tcache_slow_t *tcache_slow, tcache_t *tcache, 166 cache_bin_t *cache_bin = &tcache->bins[szind]; 171 tcache_bin_flush_large(tsd, tcache, cache_bin, szind 177 tcache_t *tcache = tcache_get(tsd); local 711 tcache_t *tcache = tsd_tcachep_get_unsafe(tsd); local 737 tcache_arena_associate(tsd_tsdn(tsd), tcache_slow, tcache, local 744 tcache, arena); local 771 tcache_t *tcache = (void *)((uintptr_t)mem + tcache_bin_alloc_size); local 776 tcache_arena_associate(tsd_tsdn(tsd), tcache_slow, tcache, local 811 tcache_arena_dissociate(tsd_tsdn(tsd), tcache_slow, tcache); local 843 tcache_t *tcache = tsd_tcachep_get(tsd); local 920 tcache_t *tcache = tcache_create_explicit(tsd); local 953 tcache_t *tcache = elm->tcache; local 969 tcache_t *tcache = tcaches_elm_remove(tsd, &tcaches[ind], true); local 981 tcache_t *tcache = tcaches_elm_remove(tsd, elm, false); local [all...] |