| /src/tests/net/bpf/ |
| t_div-by-zero.c | 29 struct bpf_insn bins[] = { local 37 bp.bf_len = __arraycount(bins); 38 bp.bf_insns = bins;
|
| /src/sys/arch/hppa/hppa/ |
| in_cksum.c | 71 #define ADDBYTE {ROL; sum += *w++; bins++; mlen--;} 81 u_int bins = 0; local 121 if (bins & 1)
|
| /src/external/bsd/jemalloc/dist/include/jemalloc/internal/ |
| tcache_structs.h | 25 * The descriptor lets the arena find our cache bins without seeing the 35 /* For small bins, fill (ncached_max >> lg_fill_div). */ 37 /* For small bins, whether has been refilled since last GC. */ 40 * For small bins, the number of items we can pretend to flush before 46 * either the cache bins alone, or the cache bin memory as well as this 51 /* The associated bins. */ 57 cache_bin_t bins[TCACHE_NBINS_MAX]; member in struct:tcache_s
|
| arena_structs.h | 95 * The arena is allocated alongside its bins; really this is a 98 bin_t bins[0]; member in struct:arena_s
|
| eset.h | 43 eset_bin_t bins[SC_NPSIZES + 1]; member in struct:eset_s
|
| sec.h | 37 * lock while doing so, to allow operations on other bins to succeed. 46 * bins (e.g. a mutex per bin, which would also be more scalable 80 sec_bin_t *bins; member in struct:sec_shard_s 81 /* Number of bytes in all bins in the shard. */
|
| cache_bin.h | 36 * individual bin. The cache bins track their bounds looking just at the low 130 * associated bins, we keep (with the tcache) a linked list of 131 * cache_bin_array_descriptor_ts that tell the arena how to find the bins. 136 * The arena keeps a list of the cache bins associated with it, for 140 /* Pointers to the tcache bins. */ 141 cache_bin_t *bins; member in struct:cache_bin_array_descriptor_s 146 cache_bin_t *bins) { 148 descriptor->bins = bins; 650 * Actually initialize some cache bins. Callers should allocate the backin [all...] |
| /src/external/bsd/jemalloc/include/jemalloc/internal/ |
| tcache_structs.h | 25 * The descriptor lets the arena find our cache bins without seeing the 35 /* For small bins, fill (ncached_max >> lg_fill_div). */ 37 /* For small bins, whether has been refilled since last GC. */ 40 * For small bins, the number of items we can pretend to flush before 46 * either the cache bins alone, or the cache bin memory as well as this 51 /* The associated bins. */ 57 cache_bin_t bins[TCACHE_NBINS_MAX]; member in struct:tcache_s
|
| arena_structs.h | 95 * The arena is allocated alongside its bins; really this is a 98 bin_t bins[0]; member in struct:arena_s
|
| eset.h | 43 eset_bin_t bins[SC_NPSIZES + 1]; member in struct:eset_s
|
| sec.h | 37 * lock while doing so, to allow operations on other bins to succeed. 46 * bins (e.g. a mutex per bin, which would also be more scalable 80 sec_bin_t *bins; member in struct:sec_shard_s 81 /* Number of bytes in all bins in the shard. */
|
| arena_structs_b.h | 202 * bins is used to store heaps of free regions. 206 bin_t bins[NBINS]; member in struct:arena_s
|
| cache_bin.h | 36 * individual bin. The cache bins track their bounds looking just at the low 130 * associated bins, we keep (with the tcache) a linked list of 131 * cache_bin_array_descriptor_ts that tell the arena how to find the bins. 136 * The arena keeps a list of the cache bins associated with it, for 140 /* Pointers to the tcache bins. */ 141 cache_bin_t *bins; member in struct:cache_bin_array_descriptor_s 146 cache_bin_t *bins) { 148 descriptor->bins = bins; 650 * Actually initialize some cache bins. Callers should allocate the backin [all...] |
| /src/external/bsd/jemalloc.old/dist/include/jemalloc/internal/ |
| arena_structs_b.h | 202 * bins is used to store heaps of free regions. 206 bin_t bins[NBINS]; member in struct:arena_s
|
| /src/external/bsd/jemalloc.old/include/jemalloc/internal/ |
| arena_structs_b.h | 202 * bins is used to store heaps of free regions. 206 bin_t bins[NBINS]; member in struct:arena_s
|
| /src/external/gpl3/gdb/dist/sim/ft32/ |
| interp.c | 295 /* Implement the BINS machine instruction. 297 static uint32_t bins (uint32_t d, uint32_t f, uint32_t len, uint32_t pos) function 444 case 0xb: result = bins (r_1v, rimmv >> 10, bit_len, bit_pos); break;
|
| /src/external/gpl3/gdb.old/dist/sim/ft32/ |
| interp.c | 295 /* Implement the BINS machine instruction. 297 static uint32_t bins (uint32_t d, uint32_t f, uint32_t len, uint32_t pos) function 444 case 0xb: result = bins (r_1v, rimmv >> 10, bit_len, bit_pos); break;
|
| /src/sys/external/bsd/drm2/dist/drm/nouveau/nvkm/subdev/fb/ |
| nouveau_nvkm_subdev_fb_ramgt215.c | 108 u8 median[8], bins[4] = {0, 0, 0, 0}, bin = 0, qty = 0; local 131 bins[(median[i] & 0xf0) >> 4]++; 137 if (bins[i] > qty) { 139 qty = bins[i];
|
| /src/external/mpl/bind/dist/lib/dns/include/dns/ |
| rrl.h | 166 dns_rrl_bin_t bins[1]; member in struct:dns_rrl_hash
|
| /src/external/mpl/dhcp/bind/dist/lib/dns/include/dns/ |
| rrl.h | 167 dns_rrl_bin_t bins[1]; member in struct:dns_rrl_hash
|
| /src/lib/libc/stdlib/ |
| jemalloc.c | 631 * bins is used to store rings of free regions of the following sizes, 634 * bins[i] | size | 653 arena_bin_t bins[1]; /* Dynamically sized. */ member in struct:arena_s 670 static size_t bin_maxclass; /* Max size class for bins. */ 671 static unsigned ntbins; /* Number of (2^n)-spaced tiny bins. */ 672 static unsigned nqbins; /* Number of quantum-spaced bins. */ 673 static unsigned nsbins; /* Number of (2^n)-spaced sub-page bins. */ 1144 malloc_printf("bins: bin size regs pgs requests newruns" 1147 if (arena->bins[i].stats.nrequests == 0) { 1167 arena->bins[i].reg_size [all...] |
| /src/external/bsd/jemalloc/dist/src/ |
| jemalloc.c | 489 tsd_binshards_t *bins = tsd_binshardsp_get(tsd); local 493 bins->binshard[i] = shard % bin_infos[i].n_shards; 3166 cache_bin_t *bin = &tcache->bins[alloc_ctx.szind]; 4240 bin = &tcache->bins[ind]; 4263 * the arena bins here, the overhead of our
|