| /src/external/bsd/jemalloc/dist/include/jemalloc/internal/ |
| psset.h | 39 size_t nactive; member in struct:psset_bin_stats_s 123 return psset->merged_stats.nactive;
|
| pa.h | 77 atomic_zu_t nactive; member in struct:pa_shard_s 226 void pa_shard_basic_stats_merge(pa_shard_t *shard, size_t *nactive,
|
| arena_externs.h | 39 ssize_t *muzzy_decay_ms, size_t *nactive, size_t *ndirty, size_t *nmuzzy); 42 size_t *nactive, size_t *ndirty, size_t *nmuzzy, arena_stats_t *astats,
|
| /src/external/bsd/jemalloc/include/jemalloc/internal/ |
| psset.h | 39 size_t nactive; member in struct:psset_bin_stats_s 123 return psset->merged_stats.nactive;
|
| arena_structs_b.h | 145 atomic_zu_t nactive; member in struct:arena_s
|
| pa.h | 77 atomic_zu_t nactive; member in struct:pa_shard_s 226 void pa_shard_basic_stats_merge(pa_shard_t *shard, size_t *nactive,
|
| arena_externs.h | 39 ssize_t *muzzy_decay_ms, size_t *nactive, size_t *ndirty, size_t *nmuzzy); 42 size_t *nactive, size_t *ndirty, size_t *nmuzzy, arena_stats_t *astats,
|
| /src/external/bsd/atf/dist/tools/ |
| io.cpp | 323 size_t nactive = m_nfds; 324 while (nactive > 0 && !terminate) { 337 assert(nactive >= 1); 338 nactive--;
|
| /src/external/bsd/jemalloc.old/dist/include/jemalloc/internal/ |
| arena_structs_b.h | 145 atomic_zu_t nactive; member in struct:arena_s
|
| arena_externs.h | 21 ssize_t *muzzy_decay_ms, size_t *nactive, size_t *ndirty, size_t *nmuzzy); 24 size_t *nactive, size_t *ndirty, size_t *nmuzzy, arena_stats_t *astats,
|
| /src/external/bsd/jemalloc.old/include/jemalloc/internal/ |
| arena_structs_b.h | 145 atomic_zu_t nactive; member in struct:arena_s
|
| arena_externs.h | 21 ssize_t *muzzy_decay_ms, size_t *nactive, size_t *ndirty, size_t *nmuzzy); 24 size_t *nactive, size_t *ndirty, size_t *nmuzzy, arena_stats_t *astats,
|
| /src/external/bsd/jemalloc/dist/src/ |
| pa.c | 9 atomic_fetch_add_zu(&shard->nactive, add_pages, ATOMIC_RELAXED); 14 assert(atomic_load_zu(&shard->nactive, ATOMIC_RELAXED) >= sub_pages); 15 atomic_fetch_sub_zu(&shard->nactive, sub_pages, ATOMIC_RELAXED); 54 atomic_store_zu(&shard->nactive, 0, ATOMIC_RELAXED); 95 atomic_store_zu(&shard->nactive, 0, ATOMIC_RELAXED);
|
| psset.c | 27 dst->nactive += src->nactive; 57 binstats[huge_idx].nactive += mul * hpdata_nactive_get(ps); 61 psset->merged_stats.nactive += mul * hpdata_nactive_get(ps); 78 assert(psset->merged_stats.nactive == check_stats.nactive);
|
| pa_extra.c | 78 pa_shard_basic_stats_merge(pa_shard_t *shard, size_t *nactive, size_t *ndirty, 80 *nactive += atomic_load_zu(&shard->nactive, ATOMIC_RELAXED); 98 resident_pgs += atomic_load_zu(&shard->nactive, ATOMIC_RELAXED);
|
| arena.c | 76 size_t *nactive, size_t *ndirty, size_t *nmuzzy) { 81 pa_shard_basic_stats_merge(&arena->pa_shard, nactive, ndirty, nmuzzy); 87 size_t *nactive, size_t *ndirty, size_t *nmuzzy, arena_stats_t *astats, 93 muzzy_decay_ms, nactive, ndirty, nmuzzy);
|
| ctl.c | 3761 arenas_i(mib[2])->astats->hpastats.psset_stats.full_slabs[0].nactive, size_t) 3770 arenas_i(mib[2])->astats->hpastats.psset_stats.full_slabs[1].nactive, size_t) 3779 arenas_i(mib[2])->astats->hpastats.psset_stats.empty_slabs[0].nactive, size_t) 3788 arenas_i(mib[2])->astats->hpastats.psset_stats.empty_slabs[1].nactive, size_t) 3797 arenas_i(mib[2])->astats->hpastats.psset_stats.nonfull_slabs[mib[5]][0].nactive, 3808 arenas_i(mib[2])->astats->hpastats.psset_stats.nonfull_slabs[mib[5]][1].nactive, 4182 pactivep = (size_t *)&(arena->pa_shard.nactive.repr);
|
| hpa.c | 954 assert(bin_stats->nactive == 0);
|
| /src/sys/dev/pci/cxgb/ |
| cxgb_lro.c | 96 int active = l->nactive; 202 l->nactive--; 221 if (l->nactive > MAX_LRO_SES) 224 if (l->nactive == MAX_LRO_SES) { 238 l->nactive++; 379 while (state->nactive) {
|
| cxgb_adapter.h | 181 unsigned int nactive; member in struct:lro_state
|
| /src/external/bsd/jemalloc/dist/test/unit/ |
| psset.c | 346 expect_zu_eq(0, stats->nactive, "Unexpected nonempty bin" 347 "Supposedly empty bin had positive nactive"); 351 stats_expect(psset_t *psset, size_t nactive) { 352 if (nactive == HUGEPAGE_PAGES) { 356 psset->stats.full_slabs[0].nactive, 361 size_t ninactive = HUGEPAGE_PAGES - nactive; 372 expect_zu_eq(nactive, 373 psset->stats.nonfull_slabs[i][0].nactive, 379 expect_zu_eq(nactive, psset_nactive(psset), "");
|
| /src/external/bsd/tmux/dist/ |
| image-sixel.c | 521 u_int y, u_int *active, u_int *nactive) 551 active[(*nactive)++] = c; 577 u_int *active, nactive; local 621 nactive = 0; 622 sixel_print_compress_colors(si, chunks, y, active, &nactive); 624 for (i = 0; i < nactive; i++) {
|
| /src/external/bsd/jemalloc.old/dist/src/ |
| arena.c | 66 size_t *nactive, size_t *ndirty, size_t *nmuzzy) { 71 *nactive += atomic_load_zu(&arena->nactive, ATOMIC_RELAXED); 79 size_t *nactive, size_t *ndirty, size_t *nmuzzy, arena_stats_t *astats, 84 muzzy_decay_ms, nactive, ndirty, nmuzzy); 121 (((atomic_load_zu(&arena->nactive, ATOMIC_RELAXED) + 279 atomic_fetch_add_zu(&arena->nactive, add_pages, ATOMIC_RELAXED); 284 assert(atomic_load_zu(&arena->nactive, ATOMIC_RELAXED) >= sub_pages); 285 atomic_fetch_sub_zu(&arena->nactive, sub_pages, ATOMIC_RELAXED); 1036 atomic_store_zu(&arena->nactive, 0, ATOMIC_RELAXED) [all...] |
| /src/sys/altq/ |
| altq_hfsc.h | 103 int nactive; /* number of active children */ member in struct:hfsc_classstats
|
| /src/usr.sbin/altq/altqstat/ |
| qdisc_hfsc.c | 139 printf(" vtperiod:%u parentperiod:%u nactive:%d\n", 140 sp->vtperiod, sp->parentperiod, sp->nactive);
|