| /src/tests/net/bpf/ |
| t_div-by-zero.c | 29 struct bpf_insn bins[] = { local 37 bp.bf_len = __arraycount(bins); 38 bp.bf_insns = bins;
|
| /src/external/bsd/jemalloc/dist/src/ |
| eset.c | 26 eset_bin_init(&eset->bins[i]); 80 if (edata_heap_empty(&eset->bins[pind].heap)) { 83 eset->bins[pind].heap_min = edata_cmp_summary; 90 eset->bins[pind].heap_min) < 0) { 91 eset->bins[pind].heap_min = edata_cmp_summary; 94 edata_heap_insert(&eset->bins[pind].heap, edata); 126 edata_heap_remove(&eset->bins[pind].heap, edata); 127 if (edata_heap_empty(&eset->bins[pind].heap)) { 140 eset->bins[pind].heap_min) == 0) { 141 eset->bins[pind].heap_min = edata_cmp_summary_get [all...] |
| /src/external/bsd/jemalloc.old/dist/src/ |
| eset.c | 26 eset_bin_init(&eset->bins[i]); 80 if (edata_heap_empty(&eset->bins[pind].heap)) { 83 eset->bins[pind].heap_min = edata_cmp_summary; 90 eset->bins[pind].heap_min) < 0) { 91 eset->bins[pind].heap_min = edata_cmp_summary; 94 edata_heap_insert(&eset->bins[pind].heap, edata); 126 edata_heap_remove(&eset->bins[pind].heap, edata); 127 if (edata_heap_empty(&eset->bins[pind].heap)) { 140 eset->bins[pind].heap_min) == 0) { 141 eset->bins[pind].heap_min = edata_cmp_summary_get [all...] |
| /src/external/bsd/jemalloc/dist/include/jemalloc/internal/ |
| arena_structs.h | 95 * The arena is allocated alongside its bins; really this is a 98 bin_t bins[0]; member in struct:arena_s
|
| tcache_structs.h | 25 * The descriptor lets the arena find our cache bins without seeing the 35 /* For small bins, fill (ncached_max >> lg_fill_div). */ 37 /* For small bins, whether has been refilled since last GC. */ 40 * For small bins, the number of items we can pretend to flush before 46 * either the cache bins alone, or the cache bin memory as well as this 51 /* The associated bins. */ 57 cache_bin_t bins[TCACHE_NBINS_MAX]; member in struct:tcache_s
|
| eset.h | 43 eset_bin_t bins[SC_NPSIZES + 1]; member in struct:eset_s
|
| stats.h | 11 OPTION('b', bins, true, false) \
|
| sec.h | 37 * lock while doing so, to allow operations on other bins to succeed. 46 * bins (e.g. a mutex per bin, which would also be more scalable 80 sec_bin_t *bins; member in struct:sec_shard_s 81 /* Number of bytes in all bins in the shard. */
|
| tcache_inlines.h | 51 cache_bin_t *bin = &tcache->bins[binind]; 94 cache_bin_t *bin = &tcache->bins[binind]; 133 cache_bin_t *bin = &tcache->bins[binind]; 171 cache_bin_t *bin = &tcache->bins[binind];
|
| /src/external/bsd/jemalloc/include/jemalloc/internal/ |
| arena_structs.h | 95 * The arena is allocated alongside its bins; really this is a 98 bin_t bins[0]; member in struct:arena_s
|
| tcache_structs.h | 25 * The descriptor lets the arena find our cache bins without seeing the 35 /* For small bins, fill (ncached_max >> lg_fill_div). */ 37 /* For small bins, whether has been refilled since last GC. */ 40 * For small bins, the number of items we can pretend to flush before 46 * either the cache bins alone, or the cache bin memory as well as this 51 /* The associated bins. */ 57 cache_bin_t bins[TCACHE_NBINS_MAX]; member in struct:tcache_s
|
| eset.h | 43 eset_bin_t bins[SC_NPSIZES + 1]; member in struct:eset_s
|
| stats.h | 11 OPTION('b', bins, true, false) \
|
| arena_structs_b.h | 202 * bins is used to store heaps of free regions. 206 bin_t bins[NBINS]; member in struct:arena_s
|
| sec.h | 37 * lock while doing so, to allow operations on other bins to succeed. 46 * bins (e.g. a mutex per bin, which would also be more scalable 80 sec_bin_t *bins; member in struct:sec_shard_s 81 /* Number of bytes in all bins in the shard. */
|
| tcache_inlines.h | 51 cache_bin_t *bin = &tcache->bins[binind]; 94 cache_bin_t *bin = &tcache->bins[binind]; 133 cache_bin_t *bin = &tcache->bins[binind]; 171 cache_bin_t *bin = &tcache->bins[binind];
|
| /src/external/bsd/jemalloc.old/dist/include/jemalloc/internal/ |
| arena_structs.h | 95 * The arena is allocated alongside its bins; really this is a 98 bin_t bins[0]; member in struct:arena_s
|
| tcache_structs.h | 25 * The descriptor lets the arena find our cache bins without seeing the 35 /* For small bins, fill (ncached_max >> lg_fill_div). */ 37 /* For small bins, whether has been refilled since last GC. */ 40 * For small bins, the number of items we can pretend to flush before 46 * either the cache bins alone, or the cache bin memory as well as this 51 /* The associated bins. */ 57 cache_bin_t bins[TCACHE_NBINS_MAX]; member in struct:tcache_s
|
| eset.h | 43 eset_bin_t bins[SC_NPSIZES + 1]; member in struct:eset_s
|
| stats.h | 11 OPTION('b', bins, true, false) \
|
| /src/external/bsd/jemalloc.old/include/jemalloc/internal/ |
| arena_structs.h | 95 * The arena is allocated alongside its bins; really this is a 98 bin_t bins[0]; member in struct:arena_s
|
| tcache_structs.h | 25 * The descriptor lets the arena find our cache bins without seeing the 35 /* For small bins, fill (ncached_max >> lg_fill_div). */ 37 /* For small bins, whether has been refilled since last GC. */ 40 * For small bins, the number of items we can pretend to flush before 46 * either the cache bins alone, or the cache bin memory as well as this 51 /* The associated bins. */ 57 cache_bin_t bins[TCACHE_NBINS_MAX]; member in struct:tcache_s
|
| eset.h | 43 eset_bin_t bins[SC_NPSIZES + 1]; member in struct:eset_s
|
| stats.h | 11 OPTION('b', bins, true, false) \
|
| /src/sys/arch/hppa/hppa/ |
| in_cksum.c | 71 #define ADDBYTE {ROL; sum += *w++; bins++; mlen--;} 81 u_int bins = 0; local 121 if (bins & 1)
|