HomeSort by: relevance | last modified time | path
    Searched defs:tcache (Results 1 - 25 of 29) sorted by relevancy

1 2

  /src/external/bsd/jemalloc.old/dist/include/jemalloc/internal/
jemalloc_internal_inlines_b.h 18 /* Set new arena/tcache associations. */
20 tcache_t *tcache = tcache_get(tsd); local
21 if (tcache != NULL) {
24 tcache, newarena); local
50 tcache_t *tcache = tsd_tcachep_get(tsd); local
57 tcache_slow, tcache, ret); local
61 tcache_slow, tcache, ret); local
tcache_structs.h 11 * The tcache state is split into the slow and hot path data. Each has a
14 * tcache, and as part of a dynamic allocation for manual allocations. Keeping
16 * splitting up the tcache [de]allocation code into those paths called with the
17 * TSD tcache and those called with a manual tcache.
26 * tcache definition. This enables arenas to aggregate stats across
27 * tcaches without having a tcache dependency.
31 /* The arena this tcache is associated with. */
52 tcache_t *tcache; member in struct:tcache_slow_s
60 /* Linkage for list of available (previously used) explicit tcache IDs. *
63 tcache_t *tcache; member in union:tcaches_s::__anon23
    [all...]
arena_inlines_b.h 146 tcache_t *tcache, bool slow_path) {
147 assert(!tsdn_null(tsdn) || tcache == NULL);
149 if (likely(tcache != NULL)) {
152 tcache, size, ind, zero, slow_path);
156 tcache, size, ind, zero, slow_path);
281 arena_dalloc_large(tsdn_t *tsdn, void *ptr, tcache_t *tcache, szind_t szind,
285 arena_dalloc_promoted(tsdn, ptr, tcache, slow_path);
287 tcache_dalloc_large(tsdn_tsd(tsdn), tcache, ptr, szind, local
302 arena_dalloc(tsdn_t *tsdn, void *ptr, tcache_t *tcache,
304 assert(!tsdn_null(tsdn) || tcache == NULL)
331 tcache_dalloc_small(tsdn_tsd(tsdn), tcache, ptr, local
418 tcache_dalloc_small(tsdn_tsd(tsdn), tcache, ptr, local
    [all...]
  /src/external/bsd/jemalloc.old/include/jemalloc/internal/
jemalloc_internal_inlines_b.h 18 /* Set new arena/tcache associations. */
20 tcache_t *tcache = tcache_get(tsd); local
21 if (tcache != NULL) {
24 tcache, newarena); local
50 tcache_t *tcache = tsd_tcachep_get(tsd); local
57 tcache_slow, tcache, ret); local
61 tcache_slow, tcache, ret); local
tcache_structs.h 11 * The tcache state is split into the slow and hot path data. Each has a
14 * tcache, and as part of a dynamic allocation for manual allocations. Keeping
16 * splitting up the tcache [de]allocation code into those paths called with the
17 * TSD tcache and those called with a manual tcache.
26 * tcache definition. This enables arenas to aggregate stats across
27 * tcaches without having a tcache dependency.
31 /* The arena this tcache is associated with. */
52 tcache_t *tcache; member in struct:tcache_slow_s
60 /* Linkage for list of available (previously used) explicit tcache IDs. *
63 tcache_t *tcache; member in union:tcaches_s::__anon91
    [all...]
arena_inlines_b.h 146 tcache_t *tcache, bool slow_path) {
147 assert(!tsdn_null(tsdn) || tcache == NULL);
149 if (likely(tcache != NULL)) {
152 tcache, size, ind, zero, slow_path);
156 tcache, size, ind, zero, slow_path);
281 arena_dalloc_large(tsdn_t *tsdn, void *ptr, tcache_t *tcache, szind_t szind,
285 arena_dalloc_promoted(tsdn, ptr, tcache, slow_path);
287 tcache_dalloc_large(tsdn_tsd(tsdn), tcache, ptr, szind, local
302 arena_dalloc(tsdn_t *tsdn, void *ptr, tcache_t *tcache,
304 assert(!tsdn_null(tsdn) || tcache == NULL)
331 tcache_dalloc_small(tsdn_tsd(tsdn), tcache, ptr, local
418 tcache_dalloc_small(tsdn_tsd(tsdn), tcache, ptr, local
    [all...]
  /src/external/bsd/jemalloc/dist/include/jemalloc/internal/
jemalloc_internal_inlines_b.h 21 /* Set new arena/tcache associations. */
23 tcache_t *tcache = tcache_get(tsd); local
24 if (tcache != NULL) {
28 tsd_tsdn(tsd), tcache_slow, tcache, newarena); local
53 tcache_t *tcache = tsd_tcachep_get(tsd); local
60 tcache_slow, tcache, ret); local
64 tsd_tsdn(tsd), tcache_slow, tcache, ret); local
tcache_structs.h 12 * The tcache state is split into the slow and hot path data. Each has a
15 * tcache, and as part of a dynamic allocation for manual allocations. Keeping
17 * splitting up the tcache [de]allocation code into those paths called with the
18 * TSD tcache and those called with a manual tcache.
27 * tcache definition. This enables arenas to aggregate stats across
28 * tcaches without having a tcache dependency.
32 /* The arena this tcache is associated with. */
34 /* The number of bins activated in the tcache. */
59 tcache_t *tcache; member in struct:tcache_slow_s
70 tcache_t *tcache; member in union:tcaches_s::__anon39
    [all...]
arena_inlines_b.h 198 bool slab, tcache_t *tcache, bool slow_path) {
199 assert(!tsdn_null(tsdn) || tcache == NULL);
201 if (likely(tcache != NULL)) {
204 return tcache_alloc_small(tsdn_tsd(tsdn), arena, tcache,
206 } else if (likely(ind < tcache_nbins_get(tcache->tcache_slow)
207 && !tcache_bin_disabled(ind, &tcache->bins[ind],
208 tcache->tcache_slow))) {
209 return tcache_alloc_large(tsdn_tsd(tsdn), arena, tcache,
315 arena_dalloc_large(tsdn_t *tsdn, void *ptr, tcache_t *tcache, szind_t szind,
317 assert(!tsdn_null(tsdn) && tcache != NULL)
326 tsdn_tsd(tsdn), tcache, ptr, szind, slow_path); local
    [all...]
  /src/external/bsd/jemalloc/include/jemalloc/internal/
jemalloc_internal_inlines_b.h 21 /* Set new arena/tcache associations. */
23 tcache_t *tcache = tcache_get(tsd); local
24 if (tcache != NULL) {
28 tsd_tsdn(tsd), tcache_slow, tcache, newarena); local
53 tcache_t *tcache = tsd_tcachep_get(tsd); local
60 tcache_slow, tcache, ret); local
64 tsd_tsdn(tsd), tcache_slow, tcache, ret); local
tcache_structs.h 12 * The tcache state is split into the slow and hot path data. Each has a
15 * tcache, and as part of a dynamic allocation for manual allocations. Keeping
17 * splitting up the tcache [de]allocation code into those paths called with the
18 * TSD tcache and those called with a manual tcache.
27 * tcache definition. This enables arenas to aggregate stats across
28 * tcaches without having a tcache dependency.
32 /* The arena this tcache is associated with. */
34 /* The number of bins activated in the tcache. */
59 tcache_t *tcache; member in struct:tcache_slow_s
70 tcache_t *tcache; member in union:tcaches_s::__anon141
    [all...]
arena_inlines_b.h 198 bool slab, tcache_t *tcache, bool slow_path) {
199 assert(!tsdn_null(tsdn) || tcache == NULL);
201 if (likely(tcache != NULL)) {
204 return tcache_alloc_small(tsdn_tsd(tsdn), arena, tcache,
206 } else if (likely(ind < tcache_nbins_get(tcache->tcache_slow)
207 && !tcache_bin_disabled(ind, &tcache->bins[ind],
208 tcache->tcache_slow))) {
209 return tcache_alloc_large(tsdn_tsd(tsdn), arena, tcache,
315 arena_dalloc_large(tsdn_t *tsdn, void *ptr, tcache_t *tcache, szind_t szind,
317 assert(!tsdn_null(tsdn) && tcache != NULL)
326 tsdn_tsd(tsdn), tcache, ptr, szind, slow_path); local
    [all...]
  /src/external/bsd/jemalloc/dist/test/unit/
ncached_max.c 15 const char *get_name = "thread.tcache.ncached_max.read_sizeclass";
46 tcache_t *tcache = tsd_tcachep_get(tsd); local
47 assert(tcache != NULL);
48 tcache_slow_t *tcache_slow = tcache->tcache_slow;
86 const char *set_name = "thread.tcache.ncached_max.write";
90 /* Test the ncached_max set with tcache on. */
116 * Close the tcache and set ncached_max of some bins. It will be
117 * set properly but thread.tcache.ncached_max.read still returns 0
118 * since the bin is not available yet. After enabling the tcache,
124 expect_d_eq(mallctl("thread.tcache.enabled", (void *)&e1, &bool_sz
    [all...]
tcache_max.c 74 tcache_t *tcache = tcache_get(tsd); local
75 for (szind_t i = 0; i < tcache_nbins_get(tcache->tcache_slow); i++) {
76 cache_bin_t *cache_bin = &tcache->bins[i];
77 if (tcache_bin_disabled(i, cache_bin, tcache->tcache_slow)) {
89 expect_zu_eq(tcache_bytes, *prev + diff, "tcache bytes not expected");
96 expect_d_eq(mallctl("thread.tcache.flush", NULL, NULL, NULL, 0), 0,
97 "Unexpected tcache flush failure");
149 assert_d_eq(mallctl("thread.tcache.max", (void *)&tcache_max,
220 expect_d_eq(mallctl("thread.tcache.max", NULL, NULL, (void *)&large,
223 expect_d_eq(mallctl("thread.tcache.flush", NULL, NULL, NULL, 0), 0
228 tcache_t *tcache = tcache_get(tsd); local
287 tcache_t *tcache = tsd_tcachep_get(tsd); local
    [all...]
  /src/external/bsd/jemalloc.old/dist/src/
tcache.c 24 * We attempt to make the number of slots in a tcache bin for a given size class
26 * the multiplier is 2 (i.e. we set the maximum number of objects in the tcache
34 * Number of allocation bytes between tcache incremental GCs. Again, this
62 /* Total stack size required (per tcache). Include the padding above. */
123 tcache_gc_small(tsd_t *tsd, tcache_slow_t *tcache_slow, tcache_t *tcache,
128 cache_bin_t *cache_bin = &tcache->bins[szind];
148 tcache_bin_flush_small(tsd, tcache, cache_bin, szind,
162 tcache_gc_large(tsd_t *tsd, tcache_slow_t *tcache_slow, tcache_t *tcache,
166 cache_bin_t *cache_bin = &tcache->bins[szind];
171 tcache_bin_flush_large(tsd, tcache, cache_bin, szind
177 tcache_t *tcache = tcache_get(tsd); local
711 tcache_t *tcache = tsd_tcachep_get_unsafe(tsd); local
737 tcache_arena_associate(tsd_tsdn(tsd), tcache_slow, tcache, local
744 tcache, arena); local
771 tcache_t *tcache = (void *)((uintptr_t)mem + tcache_bin_alloc_size); local
776 tcache_arena_associate(tsd_tsdn(tsd), tcache_slow, tcache, local
811 tcache_arena_dissociate(tsd_tsdn(tsd), tcache_slow, tcache); local
843 tcache_t *tcache = tsd_tcachep_get(tsd); local
920 tcache_t *tcache = tcache_create_explicit(tsd); local
953 tcache_t *tcache = elm->tcache; local
969 tcache_t *tcache = tcaches_elm_remove(tsd, &tcaches[ind], true); local
981 tcache_t *tcache = tcaches_elm_remove(tsd, elm, false); local
    [all...]
arena.c 151 /* Currently cached bytes and sanitizer-stashed bytes in tcache. */
1210 bool zero, tcache_t *tcache) {
1225 zero, tcache, true);
1270 arena_dalloc_promoted(tsdn_t *tsdn, void *ptr, tcache_t *tcache,
1286 if (bumped_usize <= tcache_maxclass && tcache != NULL) {
1287 tcache_dalloc_large(tsdn_tsd(tsdn), tcache, ptr, local
1444 size_t alignment, bool zero, tcache_t *tcache) {
1447 zero, tcache, true);
1453 return ipalloct(tsdn, usize, alignment, zero, tcache, arena);
1458 size_t size, size_t alignment, bool zero, tcache_t *tcache,
1870 tcache_t *tcache = tcache_slow->tcache; local
    [all...]
jemalloc.c 672 * individual threads do not lock when recording tcache stats
686 tcache_slow->tcache, arena);
1400 CONF_HANDLE_BOOL(opt_tcache, "tcache")
2359 tcache_t *tcache; local
2362 /* Getting tcache ptr unconditionally. */
2363 tcache = tsd_tcachep_get(tsd);
2364 assert(tcache == tcache_get(tsd));
2367 tcache = tcache_get(tsd);
2369 tcache = NULL;
2373 * Should not specify tcache on deallocation path when bein
2411 tcache_t *tcache = tcache_get_from_ind(tsd, dopts->tcache_ind, local
2914 idalloctm(tsd_tsdn(tsd), ptr, tcache, &alloc_ctx, false, local
2920 idalloctm(tsd_tsdn(tsd), ptr, tcache, &alloc_ctx, false, local
3009 isdalloct(tsd_tsdn(tsd), ptr, usize, tcache, &alloc_ctx, local
3015 isdalloct(tsd_tsdn(tsd), ptr, usize, tcache, &alloc_ctx, local
3038 tcache_t *tcache = tcache_get_from_ind(tsd, local
3043 tcache_t *tcache = tcache_get_from_ind(tsd, local
3164 tcache_t *tcache = tcache_get_from_ind(tsd, TCACHE_IND_AUTOMATIC, local
3542 tcache_t *tcache = tcache_get_from_ind(tsd, tcache_ind, local
3626 tcache_t *tcache = tcache_get_from_ind(tsd, local
3907 tcache_t *tcache = tcache_get_from_ind(tsd, tcache_ind, !fast, local
3945 tcache_t *tcache = tcache_get_from_ind(tsd, tcache_ind, !fast, local
4236 tcache_t *tcache = tcache_get_from_ind(tsd, local
    [all...]
  /src/external/bsd/jemalloc/dist/src/
tcache.c 25 * We attempt to make the number of slots in a tcache bin for a given size class
27 * the multiplier is 2 (i.e. we set the maximum number of objects in the tcache
35 * Number of allocation bytes between tcache incremental GCs. Again, this
63 * is only used to initialize tcache_nbins in the per-thread tcache.
69 * to initialize tcache_max in the per-thread tcache. Directly modifying it
364 tsd_t *tsd, tcache_slow_t *tcache_slow, tcache_t *tcache, szind_t szind) {
371 cache_bin_t *cache_bin = &tcache->bins[szind];
372 assert(!tcache_bin_disabled(szind, cache_bin, tcache->tcache_slow));
396 * When the new tcache gc is not enabled, keep the flush delay logic,
470 tsd, tcache, cache_bin, szind, (unsigned)(ncached - nflush))
514 tcache_t *tcache = tcache_get(tsd); local
732 tcache_t *tcache = tsd_tcachep_get(tsd); local
961 tcache_t *tcache = tsd_tcachep_get_unsafe(tsd); local
988 tsd_tsdn(tsd), tcache_slow, tcache, arena); local
996 tsd_tsdn(tsd), tcache_slow, tcache, arena); local
1048 tcache_t *tcache = (void *)((byte_t *)mem + tcache_size); local
1056 tsd_tsdn(tsd), tcache_slow, tcache, arena_ichoose(tsd, NULL)); local
1102 tcache_t *tcache = tsd_tcachep_get(tsd); local
1181 tcache_t *tcache = tsd_tcachep_get(tsd); local
1228 tcache_arena_dissociate(tsd_tsdn(tsd), tcache_slow, tcache); local
1264 tcache_t *tcache = tsd_tcachep_get(tsd); local
1339 tcache_t *tcache = tcache_create_explicit(tsd); local
1372 tcache_t *tcache = elm->tcache; local
1388 tcache_t *tcache = tcaches_elm_remove(tsd, &tcaches[ind], true); local
1400 tcache_t *tcache = tcaches_elm_remove(tsd, elm, false); local
    [all...]
arena.c 164 /* Currently cached bytes and sanitizer-stashed bytes in tcache. */
631 tsdn_t *tsdn, void *ptr, tcache_t *tcache, bool slow_path, edata_t *edata) {
645 if (bumped_usize >= SC_LARGE_MINCLASS && tcache != NULL
648 bumped_ind, &tcache->bins[bumped_ind], tcache->tcache_slow)) {
650 tsdn_tsd(tsdn), tcache, ptr, bumped_ind, slow_path); local
658 tsdn_t *tsdn, void *ptr, tcache_t *tcache, bool slow_path) {
660 arena_dalloc_promoted_impl(tsdn, ptr, tcache, slow_path, edata);
703 /* tcache */ NULL, /* slow_path */ true, edata);
1142 bool zero, bool slab, tcache_t *tcache) {
2081 tcache_t *tcache = tcache_slow->tcache; local
    [all...]
ctl.c 442 {NAME("tcache"), CHILD(named, thread_tcache)},
510 {NAME("tcache"), CTL(opt_tcache)},
923 {NAME("tcache"), CHILD(named, tcache)},
2307 /* Set new arena/tcache associations. */
2423 /* pointer to tcache_t always exists even with tcache disabled. */
2424 tcache_t *tcache = tsd_tcachep_get(tsd); local
2425 assert(tcache != NULL);
2426 oldval = tcache_max_get(tcache->tcache_slow);
3914 /* tcache bin mutex *
    [all...]
jemalloc.c 705 * individual threads do not lock when recording tcache stats
719 tsdn, tcache_slow->tcache, arena);
1459 * allocation out of the tcache.
1475 /* Fill in the tcache. */
1476 tcache_t *tcache = tcache_get_from_ind( local
1487 dopts->alignment, dopts->zero, slab, tcache, arena);
1491 slab, tcache, false, arena, sopts->slow);
1957 ifree(tsd_t *tsd, void *ptr, tcache_t *tcache, bool slow_path) {
1980 idalloctm(tsd_tsdn(tsd), ptr, tcache, &alloc_ctx, false, false); local
1985 idalloctm(tsd_tsdn(tsd), ptr, tcache, &alloc_ctx, false, true) local
2056 isdalloct(tsd_tsdn(tsd), ptr, usize, tcache, &alloc_ctx, false); local
2061 isdalloct(tsd_tsdn(tsd), ptr, usize, tcache, &alloc_ctx, true); local
2083 tcache_t *tcache = tcache_get_from_ind(tsd, local
2088 tcache_t *tcache = tcache_get_from_ind(tsd, local
2527 tcache_t *tcache = tcache_get_from_ind(tsd, tcache_ind, local
2613 tcache_t *tcache = tcache_get_from_ind(tsd, local
2903 tcache_t *tcache = tcache_get_from_ind(tsd, tcache_ind, !fast, local
2941 tcache_t *tcache = tcache_get_from_ind(tsd, tcache_ind, !fast, local
3215 tcache_t *tcache = tcache_get_from_ind(tsd, tcache_ind, local
    [all...]
  /src/sys/compat/linux/arch/aarch64/
linux_syscallargs.h 775 syscallarg(struct linux_getcpu_cache *) tcache; member in struct:linux_sys_getcpu_args
  /src/sys/compat/linux/arch/alpha/
linux_syscallargs.h 1246 syscallarg(struct linux_getcpu_cache *) tcache; member in struct:linux_sys_getcpu_args
  /src/sys/compat/linux/arch/amd64/
linux_syscallargs.h 1271 syscallarg(struct linux_getcpu_cache *) tcache; member in struct:linux_sys_getcpu_args
  /src/sys/compat/linux/arch/arm/
linux_syscallargs.h 1210 syscallarg(struct linux_getcpu_cache *) tcache; member in struct:linux_sys_getcpu_args

Completed in 50 milliseconds

1 2