HomeSort by: relevance | last modified time | path
    Searched defs:bucket (Results 1 - 13 of 13) sorted by relevancy

  /src/sys/dev/raidframe/
rf_debugMem.c 157 size_t bucket = (size_t)HASHADDR(addr); local
163 for (p = mh_table[bucket]; p && (p->address != addr); p = p->next)
168 p->next = mh_table[bucket];
169 mh_table[bucket] = p;
189 size_t bucket = HASHADDR(addr); local
193 for (p = mh_table[bucket]; p && (p->address != addr); p = p->next)
  /src/usr.sbin/bootp/common/
hash.c 101 * Frees an entire linked list of bucket members (used in the open
122 * memory and resets all bucket pointers to NULL.
205 * to determine the bucket number, and "compare" and "key" to determine
209 * already exists in the given bucket of the hash table, or some other error
331 unsigned bucket;
344 * until we find a new chain (non-empty bucket) or run out of buckets.
346 bucket = hashtable->bucketnum + 1;
347 while ((bucket < hashtable->size) &&
348 !(memberptr = (hashtable->table)[bucket])) {
349 bucket++
323 unsigned bucket; local
    [all...]
  /src/libexec/ld.elf_so/
symbol.c 333 Elf32_Word bucket; local
349 bucket = obj->buckets_gnu[fast_remainder32(hash, obj->nbuckets_gnu,
351 if (bucket == 0)
354 hashval = &obj->chains_gnu[bucket];
xmalloc.c 111 uint16_t ovu_index; /* bucket # */
171 size_t bucket; local
200 bucket = 0;
203 bucket++;
204 if (amt == 0 || bucket >= NBUCKETS)
208 * If nothing in hash bucket right now,
211 if ((op = nextf[bucket]) == NULL) {
212 morecore(bucket);
213 if ((op = nextf[bucket]) == NULL)
217 nextf[bucket] = op->ov_next
    [all...]
  /src/sys/external/bsd/drm2/dist/drm/amd/amdgpu/
amdgpu_bo_list.c 204 /* This is based on the bucket sort with O(n) time complexity.
205 * An item with priority "i" is added to bucket[i]. The lists are then
208 struct list_head bucket[AMDGPU_BO_LIST_NUM_BUCKETS]; local
213 INIT_LIST_HEAD(&bucket[i]);
225 list_add_tail(&e->tv.head, &bucket[priority]);
232 list_splice(&bucket[i], validated);
  /src/lib/libbsdmalloc/
malloc.c 85 u_char ovu_index; /* bucket # */
117 static int pagebucket; /* page size bucket */
181 int bucket; local
203 bucket = 0;
207 bucket++;
209 pagebucket = bucket;
218 amt = 8; /* size of first bucket */
219 bucket = 0;
221 amt = 16; /* size of first bucket */
222 bucket = 1
    [all...]
  /src/sys/external/bsd/drm2/dist/drm/radeon/
radeon_cs.c 48 /* This is based on the bucket sort with O(n) time complexity.
49 * An item with priority "i" is added to bucket[i]. The lists are then
53 struct list_head bucket[RADEON_CS_NUM_BUCKETS]; member in struct:radeon_cs_buckets
61 INIT_LIST_HEAD(&b->bucket[i]);
72 list_add_tail(item, &b->bucket[min(priority, RADEON_CS_MAX_PRIORITY)]);
82 list_splice(&b->bucket[i], out_list);
  /src/sys/net/npf/
lpm.c 88 lpm_ent_t ** bucket; member in struct:__anon7045
115 KASSERT(!hmap->bucket);
119 lpm_ent_t *entry = hmap->bucket[i];
133 kmem_free(hmap->bucket, hmap->hashsize * sizeof(lpm_ent_t *));
134 hmap->bucket = NULL;
172 lpm_ent_t **bucket; local
178 bucket = kmem_zalloc(hashsize * sizeof(lpm_ent_t *), flags);
179 if (bucket == NULL)
182 lpm_ent_t *list = hmap->bucket[n];
190 entry->next = bucket[i]
    [all...]
  /src/lib/libc/db/hash/
hash.c 166 * maximum bucket number, so the number of buckets is
217 "BUCKET SIZE ", hashp->BSIZE,
218 "BUCKET SHIFT ", hashp->BSHIFT,
223 "MAX BUCKET ", hashp->MAX_BUCKET,
297 /* Fix bucket size to be optimal for file system */
597 /* Pin the bucket chain */
726 uint32_t bucket; local
748 for (bucket = hashp->cbucket;
749 bucket <= (uint32_t)hashp->MAX_BUCKET;
750 bucket++)
900 int n, bucket; local
    [all...]
  /src/sys/kern/
uipc_sem.c 333 u_long bucket = KSEM_PSHARED_HASH(id); local
338 LIST_FOREACH(ksem, &ksem_pshared_hashtab[bucket], ks_entry) {
387 u_long bucket = KSEM_PSHARED_HASH(ksem->ks_pshared_id); local
388 LIST_INSERT_HEAD(&ksem_pshared_hashtab[bucket], ksem, ks_entry);
  /src/usr.sbin/lockstat/
main.c 73 SLIST_HEAD(bucket, lockstruct);
74 typedef struct bucket bucket_t;
78 SLIST_ENTRY(lockstruct) bucket;
134 static bucket_t bucket[256]; variable
136 #define HASH(a) (&bucket[((a) >> 6) & (__arraycount(bucket) - 1)])
621 for (i = 0; i < __arraycount(bucket); i++) {
622 SLIST_INIT(&bucket[i]);
638 SLIST_FOREACH(l, bp, bucket) {
655 SLIST_INSERT_HEAD(bp, l, bucket);
    [all...]
  /src/sys/uvm/
uvm_page.c 160 * freelist locks - one per bucket.
392 /* We always start with only 1 bucket. */
798 * bucket for the page. For NUMA
864 * uvm_page_rebucket: Determine a bucket structure and redim the free
917 * Now tell each CPU which bucket to use. In the outer loop, scroll
926 * and assign the same bucket ID.
938 "%d package(s) per bucket\n", 1 << shift);
998 * specific freelist and specific bucket only.
1013 * Skip the bucket if empty, no lock needed. There could be many
1021 /* Skip bucket if low on memory. *
1417 int bucket, s; local
    [all...]
uvm_pdpolicy_clockpro.c 351 struct bucket { struct
359 static struct bucket static_bucket;
360 static struct bucket *buckets = &static_bucket;
406 struct bucket *newbuckets;
407 struct bucket *oldbuckets;
419 struct bucket *b = &newbuckets[i];
439 static struct bucket *
449 nonresident_rotate(struct bucket *b)
480 struct bucket *b = nonresident_getbucket(obj, idx);
546 struct bucket *b = nonresident_getbucket(obj, idx)
    [all...]

Completed in 29 milliseconds