HomeSort by: relevance | last modified time | path
    Searched defs:bins (Results 1 - 22 of 22) sorted by relevancy

  /src/tests/net/bpf/
t_div-by-zero.c 29 struct bpf_insn bins[] = { local
37 bp.bf_len = __arraycount(bins);
38 bp.bf_insns = bins;
  /src/sys/arch/hppa/hppa/
in_cksum.c 71 #define ADDBYTE {ROL; sum += *w++; bins++; mlen--;}
81 u_int bins = 0; local
121 if (bins & 1)
  /src/external/bsd/jemalloc/dist/include/jemalloc/internal/
tcache_structs.h 25 * The descriptor lets the arena find our cache bins without seeing the
35 /* For small bins, fill (ncached_max >> lg_fill_div). */
37 /* For small bins, whether has been refilled since last GC. */
40 * For small bins, the number of items we can pretend to flush before
46 * either the cache bins alone, or the cache bin memory as well as this
51 /* The associated bins. */
57 cache_bin_t bins[TCACHE_NBINS_MAX]; member in struct:tcache_s
arena_structs.h 95 * The arena is allocated alongside its bins; really this is a
98 bin_t bins[0]; member in struct:arena_s
eset.h 43 eset_bin_t bins[SC_NPSIZES + 1]; member in struct:eset_s
sec.h 37 * lock while doing so, to allow operations on other bins to succeed.
46 * bins (e.g. a mutex per bin, which would also be more scalable
80 sec_bin_t *bins; member in struct:sec_shard_s
81 /* Number of bytes in all bins in the shard. */
cache_bin.h 36 * individual bin. The cache bins track their bounds looking just at the low
130 * associated bins, we keep (with the tcache) a linked list of
131 * cache_bin_array_descriptor_ts that tell the arena how to find the bins.
136 * The arena keeps a list of the cache bins associated with it, for
140 /* Pointers to the tcache bins. */
141 cache_bin_t *bins; member in struct:cache_bin_array_descriptor_s
146 cache_bin_t *bins) {
148 descriptor->bins = bins;
650 * Actually initialize some cache bins. Callers should allocate the backin
    [all...]
  /src/external/bsd/jemalloc/include/jemalloc/internal/
tcache_structs.h 25 * The descriptor lets the arena find our cache bins without seeing the
35 /* For small bins, fill (ncached_max >> lg_fill_div). */
37 /* For small bins, whether has been refilled since last GC. */
40 * For small bins, the number of items we can pretend to flush before
46 * either the cache bins alone, or the cache bin memory as well as this
51 /* The associated bins. */
57 cache_bin_t bins[TCACHE_NBINS_MAX]; member in struct:tcache_s
arena_structs.h 95 * The arena is allocated alongside its bins; really this is a
98 bin_t bins[0]; member in struct:arena_s
eset.h 43 eset_bin_t bins[SC_NPSIZES + 1]; member in struct:eset_s
sec.h 37 * lock while doing so, to allow operations on other bins to succeed.
46 * bins (e.g. a mutex per bin, which would also be more scalable
80 sec_bin_t *bins; member in struct:sec_shard_s
81 /* Number of bytes in all bins in the shard. */
arena_structs_b.h 202 * bins is used to store heaps of free regions.
206 bin_t bins[NBINS]; member in struct:arena_s
cache_bin.h 36 * individual bin. The cache bins track their bounds looking just at the low
130 * associated bins, we keep (with the tcache) a linked list of
131 * cache_bin_array_descriptor_ts that tell the arena how to find the bins.
136 * The arena keeps a list of the cache bins associated with it, for
140 /* Pointers to the tcache bins. */
141 cache_bin_t *bins; member in struct:cache_bin_array_descriptor_s
146 cache_bin_t *bins) {
148 descriptor->bins = bins;
650 * Actually initialize some cache bins. Callers should allocate the backin
    [all...]
  /src/external/bsd/jemalloc.old/dist/include/jemalloc/internal/
arena_structs_b.h 202 * bins is used to store heaps of free regions.
206 bin_t bins[NBINS]; member in struct:arena_s
  /src/external/bsd/jemalloc.old/include/jemalloc/internal/
arena_structs_b.h 202 * bins is used to store heaps of free regions.
206 bin_t bins[NBINS]; member in struct:arena_s
  /src/external/gpl3/gdb/dist/sim/ft32/
interp.c 295 /* Implement the BINS machine instruction.
297 static uint32_t bins (uint32_t d, uint32_t f, uint32_t len, uint32_t pos) function
444 case 0xb: result = bins (r_1v, rimmv >> 10, bit_len, bit_pos); break;
  /src/external/gpl3/gdb.old/dist/sim/ft32/
interp.c 295 /* Implement the BINS machine instruction.
297 static uint32_t bins (uint32_t d, uint32_t f, uint32_t len, uint32_t pos) function
444 case 0xb: result = bins (r_1v, rimmv >> 10, bit_len, bit_pos); break;
  /src/sys/external/bsd/drm2/dist/drm/nouveau/nvkm/subdev/fb/
nouveau_nvkm_subdev_fb_ramgt215.c 108 u8 median[8], bins[4] = {0, 0, 0, 0}, bin = 0, qty = 0; local
131 bins[(median[i] & 0xf0) >> 4]++;
137 if (bins[i] > qty) {
139 qty = bins[i];
  /src/external/mpl/bind/dist/lib/dns/include/dns/
rrl.h 166 dns_rrl_bin_t bins[1]; member in struct:dns_rrl_hash
  /src/external/mpl/dhcp/bind/dist/lib/dns/include/dns/
rrl.h 167 dns_rrl_bin_t bins[1]; member in struct:dns_rrl_hash
  /src/lib/libc/stdlib/
jemalloc.c 631 * bins is used to store rings of free regions of the following sizes,
634 * bins[i] | size |
653 arena_bin_t bins[1]; /* Dynamically sized. */ member in struct:arena_s
670 static size_t bin_maxclass; /* Max size class for bins. */
671 static unsigned ntbins; /* Number of (2^n)-spaced tiny bins. */
672 static unsigned nqbins; /* Number of quantum-spaced bins. */
673 static unsigned nsbins; /* Number of (2^n)-spaced sub-page bins. */
1144 malloc_printf("bins: bin size regs pgs requests newruns"
1147 if (arena->bins[i].stats.nrequests == 0) {
1167 arena->bins[i].reg_size
    [all...]
  /src/external/bsd/jemalloc/dist/src/
jemalloc.c 489 tsd_binshards_t *bins = tsd_binshardsp_get(tsd); local
493 bins->binshard[i] = shard % bin_infos[i].n_shards;
3166 cache_bin_t *bin = &tcache->bins[alloc_ctx.szind];
4240 bin = &tcache->bins[ind];
4263 * the arena bins here, the overhead of our

Completed in 55 milliseconds