| /src/usr.sbin/timed/timed/ |
| correct.c | 122 long ndelta; local 143 ndelta = (unsigned long)delta >> smoother; 148 ndelta |= mask; 153 delta, ndelta); 154 adj.tv_usec = ndelta;
|
| /src/external/bsd/jemalloc/dist/src/ |
| sc.c | 17 reg_size_compute(int lg_base, int lg_delta, int ndelta) { 18 return (ZU(1) << lg_base) + (ZU(ndelta) << lg_delta); 23 slab_size(int lg_page, int lg_base, int lg_delta, int ndelta) { 25 size_t reg_size = reg_size_compute(lg_base, lg_delta, ndelta); 34 * base + ndelta * delta == (ndelta + base/ndelta) * delta, which is 35 * (ndelta + ngroup) * delta. The way we choose slabbing strategies 36 * means that delta is at most the page size and ndelta < ngroup. So 60 int index, int lg_base, int lg_delta, int ndelta) { 102 int ndelta = 0; local [all...] |
| /src/external/bsd/jemalloc.old/dist/src/ |
| sc.c | 17 reg_size_compute(int lg_base, int lg_delta, int ndelta) { 18 return (ZU(1) << lg_base) + (ZU(ndelta) << lg_delta); 23 slab_size(int lg_page, int lg_base, int lg_delta, int ndelta) { 25 size_t reg_size = reg_size_compute(lg_base, lg_delta, ndelta); 34 * base + ndelta * delta == (ndelta + base/ndelta) * delta, which is 35 * (ndelta + ngroup) * delta. The way we choose slabbing strategies 36 * means that delta is at most the page size and ndelta < ngroup. So 60 int index, int lg_base, int lg_delta, int ndelta) { 102 int ndelta = 0; local [all...] |
| /src/external/bsd/jemalloc/dist/include/jemalloc/internal/ |
| sc.h | 139 * lg_delta, and ndelta (i.e. number of deltas above the base) on a 141 * classes, size == (1 << lg_base) + ndelta * (1 << lg_delta). 144 * lg_delta is lg_base - SC_LG_NGROUP, and ndelta goes from 1 to SC_NGROUP. 148 * subsequent ones. ndelta is always 0. 151 * lg_base == LG_QUANTUM, lg_delta == LG_QUANTUM, and have ndelta range from 0 157 * lg_delta == LG_QUANTUM - 1, and ndelta == 1, keeping the rest of the 298 /* Delta multiplier. size == 1<<lg_base + ndelta<<lg_delta */ 299 int ndelta; member in struct:sc_s 347 size_t reg_size_compute(int lg_base, int lg_delta, int ndelta);
|
| /src/external/bsd/jemalloc/include/jemalloc/internal/ |
| sc.h | 139 * lg_delta, and ndelta (i.e. number of deltas above the base) on a 141 * classes, size == (1 << lg_base) + ndelta * (1 << lg_delta). 144 * lg_delta is lg_base - SC_LG_NGROUP, and ndelta goes from 1 to SC_NGROUP. 148 * subsequent ones. ndelta is always 0. 151 * lg_base == LG_QUANTUM, lg_delta == LG_QUANTUM, and have ndelta range from 0 157 * lg_delta == LG_QUANTUM - 1, and ndelta == 1, keeping the rest of the 298 /* Delta multiplier. size == 1<<lg_base + ndelta<<lg_delta */ 299 int ndelta; member in struct:sc_s 347 size_t reg_size_compute(int lg_base, int lg_delta, int ndelta);
|
| /src/external/bsd/jemalloc.old/dist/include/jemalloc/internal/ |
| sc.h | 139 * lg_delta, and ndelta (i.e. number of deltas above the base) on a 141 * classes, size == (1 << lg_base) + ndelta * (1 << lg_delta). 144 * lg_delta is lg_base - SC_LG_NGROUP, and ndelta goes from 1 to SC_NGROUP. 148 * subsequent ones. ndelta is always 0. 151 * lg_base == LG_QUANTUM, lg_delta == LG_QUANTUM, and have ndelta range from 0 157 * lg_delta == LG_QUANTUM - 1, and ndelta == 1, keeping the rest of the 298 /* Delta multiplier. size == 1<<lg_base + ndelta<<lg_delta */ 299 int ndelta; member in struct:sc_s 347 size_t reg_size_compute(int lg_base, int lg_delta, int ndelta);
|
| /src/external/bsd/jemalloc.old/include/jemalloc/internal/ |
| sc.h | 139 * lg_delta, and ndelta (i.e. number of deltas above the base) on a 141 * classes, size == (1 << lg_base) + ndelta * (1 << lg_delta). 144 * lg_delta is lg_base - SC_LG_NGROUP, and ndelta goes from 1 to SC_NGROUP. 148 * subsequent ones. ndelta is always 0. 151 * lg_base == LG_QUANTUM, lg_delta == LG_QUANTUM, and have ndelta range from 0 157 * lg_delta == LG_QUANTUM - 1, and ndelta == 1, keeping the rest of the 298 /* Delta multiplier. size == 1<<lg_base + ndelta<<lg_delta */ 299 int ndelta; member in struct:sc_s 347 size_t reg_size_compute(int lg_base, int lg_delta, int ndelta);
|