| /src/external/bsd/jemalloc/dist/test/unit/ |
| slab.c | 10 edata_t slab; local 12 edata_init(&slab, INVALID_ARENA_IND, 17 expect_ptr_not_null(edata_addr_get(&slab), 22 void *reg = (void *)((uintptr_t)edata_addr_get(&slab) + 25 &slab, reg), 30 free(edata_addr_get(&slab));
|
| /src/external/bsd/jemalloc.old/dist/test/unit/ |
| slab.c | 10 edata_t slab; local 12 edata_init(&slab, INVALID_ARENA_IND, 17 expect_ptr_not_null(edata_addr_get(&slab), 22 void *reg = (void *)((uintptr_t)edata_addr_get(&slab) + 25 &slab, reg), 30 free(edata_addr_get(&slab));
|
| /src/external/bsd/unbound/dist/testcode/ |
| unitinfra.c | 66 static void test_keep_probing(struct infra_cache* slab, 79 slab->infra_keep_probing = keep_probing; 83 unit_assert( infra_host(slab, &one, onelen, zone, zonelen, 88 unit_assert( infra_rtt_update(slab, &one, onelen, zone, zonelen, 90 unit_assert( infra_host(slab, &one, onelen, zone, zonelen, 97 unit_assert( (d=infra_lookup_host(slab, &one, onelen, zone, zonelen, 0, *now, &k)) ); 108 unit_assert( infra_get_lame_rtt(slab, &one, onelen, zone, zonelen, 121 struct infra_cache* slab; local 135 slab = infra_create(cfg); 137 unit_assert( infra_host(slab, &one, onelen, zone, zonelen, now [all...] |
| /src/external/mpl/dhcp/bind/dist/lib/dns/include/dns/ |
| rdataslab.h | 74 * Slabify a rdataset. The slab area will be allocated and returned 92 dns_rdataslab_size(unsigned char *slab, unsigned int reservelen); 97 *\li 'slab' points to a slab. 100 *\li The number of bytes in the slab, including the reservelen. 104 dns_rdataslab_rdatasize(unsigned char *slab, unsigned int reservelen); 109 *\li 'slab' points to a slab. 113 dns_rdataslab_count(unsigned char *slab, unsigned int reservelen); 118 *\li 'slab' points to a slab [all...] |
| /src/sys/external/bsd/drm2/include/drm/ |
| drm_mem_util.h | 35 #include <linux/slab.h>
|
| /src/external/bsd/unbound/dist/util/storage/ |
| slabhash.c | 243 size_t slab, cnt = 0; local 245 for(slab=0; slab<sh->size; slab++) { 246 lock_quick_lock(&sh->array[slab]->lock); 247 cnt += sh->array[slab]->num; 248 lock_quick_unlock(&sh->array[slab]->lock); 255 size_t slab, cnt = 0, max_collisions = 0; local 257 for(slab=0; slab<sh->size; slab++) [all...] |
| /src/external/bsd/unbound/dist/validator/ |
| val_kcache.c | 62 kcache->slab = slabhash_create(numtables, start_size, maxmem, 65 if(!kcache->slab) { 78 slabhash_delete(kcache->slab); 90 slabhash_insert(kcache->slab, k->entry.hash, &k->entry, 116 e = slabhash_lookup(kcache->slab, lookfor.entry.hash, &lookfor, wr); 154 return sizeof(*kcache) + slabhash_get_mem(kcache->slab); 166 slabhash_remove(kcache->slab, lookfor.entry.hash, &lookfor);
|
| /src/sys/external/bsd/drm2/dist/drm/amd/acp/ |
| amdgpu_acp_hw.c | 30 #include <linux/slab.h>
|
| /src/external/mit/libuv/dist/test/ |
| test-close-fd.c | 31 static char slab[1]; local 32 buf->base = slab; 33 buf->len = sizeof(slab);
|
| test-not-readable-nor-writable-on-read-error.c | 44 static char slab[64]; local 45 buf->base = slab; 46 buf->len = sizeof(slab);
|
| test-tcp-close-after-read-timeout.c | 61 static char slab[8]; local 62 buf->base = slab; 63 buf->len = sizeof(slab); 86 static char slab[8]; local 87 buf->base = slab; 88 buf->len = sizeof(slab);
|
| test-iouring-pollhup.c | 45 static char slab[32]; local 46 *buf = uv_buf_init(slab, sizeof(slab));
|
| /src/external/bsd/jemalloc/dist/src/ |
| inspect.c | 70 edata_t *slab; local 72 slab = bin->slabcur; 74 slab = edata_heap_first(&bin->slabs_nonfull); 76 *slabcur_addr = slab != NULL ? edata_addr_get(slab) : NULL;
|
| arena.c | 65 static void arena_bin_lower_slab(tsdn_t *tsdn, arena_t *arena, edata_t *slab, 227 arena_slab_reg_alloc(edata_t *slab, const bin_info_t *bin_info) { 229 slab_data_t *slab_data = edata_slab_data_get(slab); 232 assert(edata_nfree_get(slab) > 0); 236 ret = (void *)((uintptr_t)edata_addr_get(slab) + 238 edata_nfree_dec(slab); 243 arena_slab_reg_alloc_batch(edata_t *slab, const bin_info_t *bin_info, 245 slab_data_t *slab_data = edata_slab_data_get(slab); 247 assert(edata_nfree_get(slab) >= cnt); 254 *(ptrs + i) = (void *)((uintptr_t)edata_addr_get(slab) 595 edata_t *slab = edata_heap_remove_first(&bin->slabs_nonfull); local 630 edata_t *slab; local 637 arena_slab_dalloc(tsd_tsdn(tsd), arena, slab); local 642 arena_slab_dalloc(tsd_tsdn(tsd), arena, slab); local 649 arena_slab_dalloc(tsd_tsdn(tsd), arena, slab); local 839 edata_t *slab = pa_alloc(tsdn, &arena->pa_shard, bin_info->slab_size, local 1070 edata_t *slab = NULL; local [all...] |
| /src/external/bsd/jemalloc.old/dist/src/ |
| inspect.c | 70 edata_t *slab; local 72 slab = bin->slabcur; 74 slab = edata_heap_first(&bin->slabs_nonfull); 76 *slabcur_addr = slab != NULL ? edata_addr_get(slab) : NULL;
|
| arena.c | 65 static void arena_bin_lower_slab(tsdn_t *tsdn, arena_t *arena, edata_t *slab, 227 arena_slab_reg_alloc(edata_t *slab, const bin_info_t *bin_info) { 229 slab_data_t *slab_data = edata_slab_data_get(slab); 232 assert(edata_nfree_get(slab) > 0); 236 ret = (void *)((uintptr_t)edata_addr_get(slab) + 238 edata_nfree_dec(slab); 243 arena_slab_reg_alloc_batch(edata_t *slab, const bin_info_t *bin_info, 245 slab_data_t *slab_data = edata_slab_data_get(slab); 247 assert(edata_nfree_get(slab) >= cnt); 254 *(ptrs + i) = (void *)((uintptr_t)edata_addr_get(slab) 595 edata_t *slab = edata_heap_remove_first(&bin->slabs_nonfull); local 630 edata_t *slab; local 637 arena_slab_dalloc(tsd_tsdn(tsd), arena, slab); local 642 arena_slab_dalloc(tsd_tsdn(tsd), arena, slab); local 649 arena_slab_dalloc(tsd_tsdn(tsd), arena, slab); local 839 edata_t *slab = pa_alloc(tsdn, &arena->pa_shard, bin_info->slab_size, local 1070 edata_t *slab = NULL; local [all...] |
| /src/external/bsd/jemalloc/dist/include/jemalloc/internal/ |
| emap.h | 25 bool slab; member in struct:emap_alloc_ctx_t 31 bool slab; member in struct:emap_full_alloc_ctx_s 38 bool slab); 68 * szind and slab info appropriately. 72 szind_t szind, bool slab); 75 * Does the same thing, but with the interior of the range, for slab 79 * does both depending on the value of 'slab'. The answer is twofold: 85 * used, and passing a 'slab' parameter everywhere makes that more 110 * addresses to edatas. If the result is going to be used as a slab, you 113 * Remap simply changes the szind and slab status of an extent's boundar [all...] |
| arena_inlines_b.h | 56 } else if (unlikely(!(is_slab = alloc_ctx->slab))) { 88 if (unlikely(!alloc_ctx->slab)) { 204 /* Only slab members should be looked up via interior pointers. */ 269 assert(alloc_ctx.slab == edata_slab_get(edata)); 272 if (likely(alloc_ctx.slab)) { 326 assert(alloc_ctx.slab == edata_slab_get(edata)); 329 if (likely(alloc_ctx.slab)) { 348 * object, so base szind and slab on the given size. 351 alloc_ctx.slab = (alloc_ctx.szind < SC_NBINS); 360 || alloc_ctx.slab == (alloc_ctx.szind < SC_NBINS)) [all...] |
| /src/external/bsd/jemalloc/include/jemalloc/internal/ |
| emap.h | 25 bool slab; member in struct:emap_alloc_ctx_t 31 bool slab; member in struct:emap_full_alloc_ctx_s 38 bool slab); 68 * szind and slab info appropriately. 72 szind_t szind, bool slab); 75 * Does the same thing, but with the interior of the range, for slab 79 * does both depending on the value of 'slab'. The answer is twofold: 85 * used, and passing a 'slab' parameter everywhere makes that more 110 * addresses to edatas. If the result is going to be used as a slab, you 113 * Remap simply changes the szind and slab status of an extent's boundar [all...] |
| arena_inlines_b.h | 56 } else if (unlikely(!(is_slab = alloc_ctx->slab))) { 88 if (unlikely(!alloc_ctx->slab)) { 204 /* Only slab members should be looked up via interior pointers. */ 269 assert(alloc_ctx.slab == edata_slab_get(edata)); 272 if (likely(alloc_ctx.slab)) { 326 assert(alloc_ctx.slab == edata_slab_get(edata)); 329 if (likely(alloc_ctx.slab)) { 348 * object, so base szind and slab on the given size. 351 alloc_ctx.slab = (alloc_ctx.szind < SC_NBINS); 360 || alloc_ctx.slab == (alloc_ctx.szind < SC_NBINS)) [all...] |
| /src/external/bsd/jemalloc.old/dist/include/jemalloc/internal/ |
| emap.h | 25 bool slab; member in struct:emap_alloc_ctx_t 31 bool slab; member in struct:emap_full_alloc_ctx_s 38 bool slab); 68 * szind and slab info appropriately. 72 szind_t szind, bool slab); 75 * Does the same thing, but with the interior of the range, for slab 79 * does both depending on the value of 'slab'. The answer is twofold: 85 * used, and passing a 'slab' parameter everywhere makes that more 110 * addresses to edatas. If the result is going to be used as a slab, you 113 * Remap simply changes the szind and slab status of an extent's boundar [all...] |
| arena_inlines_b.h | 56 } else if (unlikely(!(is_slab = alloc_ctx->slab))) { 88 if (unlikely(!alloc_ctx->slab)) { 204 /* Only slab members should be looked up via interior pointers. */ 269 assert(alloc_ctx.slab == edata_slab_get(edata)); 272 if (likely(alloc_ctx.slab)) { 326 assert(alloc_ctx.slab == edata_slab_get(edata)); 329 if (likely(alloc_ctx.slab)) { 348 * object, so base szind and slab on the given size. 351 alloc_ctx.slab = (alloc_ctx.szind < SC_NBINS); 360 || alloc_ctx.slab == (alloc_ctx.szind < SC_NBINS)) [all...] |
| /src/external/bsd/jemalloc.old/include/jemalloc/internal/ |
| emap.h | 25 bool slab; member in struct:emap_alloc_ctx_t 31 bool slab; member in struct:emap_full_alloc_ctx_s 38 bool slab); 68 * szind and slab info appropriately. 72 szind_t szind, bool slab); 75 * Does the same thing, but with the interior of the range, for slab 79 * does both depending on the value of 'slab'. The answer is twofold: 85 * used, and passing a 'slab' parameter everywhere makes that more 110 * addresses to edatas. If the result is going to be used as a slab, you 113 * Remap simply changes the szind and slab status of an extent's boundar [all...] |
| arena_inlines_b.h | 56 } else if (unlikely(!(is_slab = alloc_ctx->slab))) { 88 if (unlikely(!alloc_ctx->slab)) { 204 /* Only slab members should be looked up via interior pointers. */ 269 assert(alloc_ctx.slab == edata_slab_get(edata)); 272 if (likely(alloc_ctx.slab)) { 326 assert(alloc_ctx.slab == edata_slab_get(edata)); 329 if (likely(alloc_ctx.slab)) { 348 * object, so base szind and slab on the given size. 351 alloc_ctx.slab = (alloc_ctx.szind < SC_NBINS); 360 || alloc_ctx.slab == (alloc_ctx.szind < SC_NBINS)) [all...] |
| /src/external/mpl/bind/dist/lib/dns/ |
| rdataslab.c | 82 * rbtdb.c directly interacts with the slab's raw structures. If the 398 dns_rdataslab_size(unsigned char *slab, unsigned int reservelen) { 399 REQUIRE(slab != NULL); 401 unsigned char *current = slab + reservelen; 416 return (unsigned int)(current - slab); 420 dns_rdataslab_rdatasize(unsigned char *slab, unsigned int reservelen) { 421 REQUIRE(slab != NULL); 424 unsigned char *current = slab + reservelen; 444 dns_rdataslab_count(unsigned char *slab, unsigned int reservelen) { 445 REQUIRE(slab != NULL) [all...] |