HomeSort by: relevance | last modified time | path
    Searched refs:slab (Results 1 - 25 of 341) sorted by relevancy

1 2 3 4 5 6 7 8 91011>>

  /src/external/bsd/jemalloc/dist/test/unit/
slab.c 10 edata_t slab; local
12 edata_init(&slab, INVALID_ARENA_IND,
17 expect_ptr_not_null(edata_addr_get(&slab),
22 void *reg = (void *)((uintptr_t)edata_addr_get(&slab) +
25 &slab, reg),
30 free(edata_addr_get(&slab));
  /src/external/bsd/jemalloc.old/dist/test/unit/
slab.c 10 edata_t slab; local
12 edata_init(&slab, INVALID_ARENA_IND,
17 expect_ptr_not_null(edata_addr_get(&slab),
22 void *reg = (void *)((uintptr_t)edata_addr_get(&slab) +
25 &slab, reg),
30 free(edata_addr_get(&slab));
  /src/external/bsd/unbound/dist/testcode/
unitinfra.c 66 static void test_keep_probing(struct infra_cache* slab,
79 slab->infra_keep_probing = keep_probing;
83 unit_assert( infra_host(slab, &one, onelen, zone, zonelen,
88 unit_assert( infra_rtt_update(slab, &one, onelen, zone, zonelen,
90 unit_assert( infra_host(slab, &one, onelen, zone, zonelen,
97 unit_assert( (d=infra_lookup_host(slab, &one, onelen, zone, zonelen, 0, *now, &k)) );
108 unit_assert( infra_get_lame_rtt(slab, &one, onelen, zone, zonelen,
121 struct infra_cache* slab; local
135 slab = infra_create(cfg);
137 unit_assert( infra_host(slab, &one, onelen, zone, zonelen, now
    [all...]
  /src/external/mpl/dhcp/bind/dist/lib/dns/include/dns/
rdataslab.h 74 * Slabify a rdataset. The slab area will be allocated and returned
92 dns_rdataslab_size(unsigned char *slab, unsigned int reservelen);
97 *\li 'slab' points to a slab.
100 *\li The number of bytes in the slab, including the reservelen.
104 dns_rdataslab_rdatasize(unsigned char *slab, unsigned int reservelen);
109 *\li 'slab' points to a slab.
113 dns_rdataslab_count(unsigned char *slab, unsigned int reservelen);
118 *\li 'slab' points to a slab
    [all...]
  /src/sys/external/bsd/drm2/include/drm/
drm_mem_util.h 35 #include <linux/slab.h>
  /src/external/bsd/unbound/dist/util/storage/
slabhash.c 243 size_t slab, cnt = 0; local
245 for(slab=0; slab<sh->size; slab++) {
246 lock_quick_lock(&sh->array[slab]->lock);
247 cnt += sh->array[slab]->num;
248 lock_quick_unlock(&sh->array[slab]->lock);
255 size_t slab, cnt = 0, max_collisions = 0; local
257 for(slab=0; slab<sh->size; slab++)
    [all...]
  /src/external/bsd/unbound/dist/validator/
val_kcache.c 62 kcache->slab = slabhash_create(numtables, start_size, maxmem,
65 if(!kcache->slab) {
78 slabhash_delete(kcache->slab);
90 slabhash_insert(kcache->slab, k->entry.hash, &k->entry,
116 e = slabhash_lookup(kcache->slab, lookfor.entry.hash, &lookfor, wr);
154 return sizeof(*kcache) + slabhash_get_mem(kcache->slab);
166 slabhash_remove(kcache->slab, lookfor.entry.hash, &lookfor);
  /src/sys/external/bsd/drm2/dist/drm/amd/acp/
amdgpu_acp_hw.c 30 #include <linux/slab.h>
  /src/external/mit/libuv/dist/test/
test-close-fd.c 31 static char slab[1]; local
32 buf->base = slab;
33 buf->len = sizeof(slab);
test-not-readable-nor-writable-on-read-error.c 44 static char slab[64]; local
45 buf->base = slab;
46 buf->len = sizeof(slab);
test-tcp-close-after-read-timeout.c 61 static char slab[8]; local
62 buf->base = slab;
63 buf->len = sizeof(slab);
86 static char slab[8]; local
87 buf->base = slab;
88 buf->len = sizeof(slab);
test-iouring-pollhup.c 45 static char slab[32]; local
46 *buf = uv_buf_init(slab, sizeof(slab));
  /src/external/bsd/jemalloc/dist/src/
inspect.c 70 edata_t *slab; local
72 slab = bin->slabcur;
74 slab = edata_heap_first(&bin->slabs_nonfull);
76 *slabcur_addr = slab != NULL ? edata_addr_get(slab) : NULL;
arena.c 65 static void arena_bin_lower_slab(tsdn_t *tsdn, arena_t *arena, edata_t *slab,
227 arena_slab_reg_alloc(edata_t *slab, const bin_info_t *bin_info) {
229 slab_data_t *slab_data = edata_slab_data_get(slab);
232 assert(edata_nfree_get(slab) > 0);
236 ret = (void *)((uintptr_t)edata_addr_get(slab) +
238 edata_nfree_dec(slab);
243 arena_slab_reg_alloc_batch(edata_t *slab, const bin_info_t *bin_info,
245 slab_data_t *slab_data = edata_slab_data_get(slab);
247 assert(edata_nfree_get(slab) >= cnt);
254 *(ptrs + i) = (void *)((uintptr_t)edata_addr_get(slab)
595 edata_t *slab = edata_heap_remove_first(&bin->slabs_nonfull); local
630 edata_t *slab; local
637 arena_slab_dalloc(tsd_tsdn(tsd), arena, slab); local
642 arena_slab_dalloc(tsd_tsdn(tsd), arena, slab); local
649 arena_slab_dalloc(tsd_tsdn(tsd), arena, slab); local
839 edata_t *slab = pa_alloc(tsdn, &arena->pa_shard, bin_info->slab_size, local
1070 edata_t *slab = NULL; local
    [all...]
  /src/external/bsd/jemalloc.old/dist/src/
inspect.c 70 edata_t *slab; local
72 slab = bin->slabcur;
74 slab = edata_heap_first(&bin->slabs_nonfull);
76 *slabcur_addr = slab != NULL ? edata_addr_get(slab) : NULL;
arena.c 65 static void arena_bin_lower_slab(tsdn_t *tsdn, arena_t *arena, edata_t *slab,
227 arena_slab_reg_alloc(edata_t *slab, const bin_info_t *bin_info) {
229 slab_data_t *slab_data = edata_slab_data_get(slab);
232 assert(edata_nfree_get(slab) > 0);
236 ret = (void *)((uintptr_t)edata_addr_get(slab) +
238 edata_nfree_dec(slab);
243 arena_slab_reg_alloc_batch(edata_t *slab, const bin_info_t *bin_info,
245 slab_data_t *slab_data = edata_slab_data_get(slab);
247 assert(edata_nfree_get(slab) >= cnt);
254 *(ptrs + i) = (void *)((uintptr_t)edata_addr_get(slab)
595 edata_t *slab = edata_heap_remove_first(&bin->slabs_nonfull); local
630 edata_t *slab; local
637 arena_slab_dalloc(tsd_tsdn(tsd), arena, slab); local
642 arena_slab_dalloc(tsd_tsdn(tsd), arena, slab); local
649 arena_slab_dalloc(tsd_tsdn(tsd), arena, slab); local
839 edata_t *slab = pa_alloc(tsdn, &arena->pa_shard, bin_info->slab_size, local
1070 edata_t *slab = NULL; local
    [all...]
  /src/external/bsd/jemalloc/dist/include/jemalloc/internal/
emap.h 25 bool slab; member in struct:emap_alloc_ctx_t
31 bool slab; member in struct:emap_full_alloc_ctx_s
38 bool slab);
68 * szind and slab info appropriately.
72 szind_t szind, bool slab);
75 * Does the same thing, but with the interior of the range, for slab
79 * does both depending on the value of 'slab'. The answer is twofold:
85 * used, and passing a 'slab' parameter everywhere makes that more
110 * addresses to edatas. If the result is going to be used as a slab, you
113 * Remap simply changes the szind and slab status of an extent's boundar
    [all...]
arena_inlines_b.h 56 } else if (unlikely(!(is_slab = alloc_ctx->slab))) {
88 if (unlikely(!alloc_ctx->slab)) {
204 /* Only slab members should be looked up via interior pointers. */
269 assert(alloc_ctx.slab == edata_slab_get(edata));
272 if (likely(alloc_ctx.slab)) {
326 assert(alloc_ctx.slab == edata_slab_get(edata));
329 if (likely(alloc_ctx.slab)) {
348 * object, so base szind and slab on the given size.
351 alloc_ctx.slab = (alloc_ctx.szind < SC_NBINS);
360 || alloc_ctx.slab == (alloc_ctx.szind < SC_NBINS))
    [all...]
  /src/external/bsd/jemalloc/include/jemalloc/internal/
emap.h 25 bool slab; member in struct:emap_alloc_ctx_t
31 bool slab; member in struct:emap_full_alloc_ctx_s
38 bool slab);
68 * szind and slab info appropriately.
72 szind_t szind, bool slab);
75 * Does the same thing, but with the interior of the range, for slab
79 * does both depending on the value of 'slab'. The answer is twofold:
85 * used, and passing a 'slab' parameter everywhere makes that more
110 * addresses to edatas. If the result is going to be used as a slab, you
113 * Remap simply changes the szind and slab status of an extent's boundar
    [all...]
arena_inlines_b.h 56 } else if (unlikely(!(is_slab = alloc_ctx->slab))) {
88 if (unlikely(!alloc_ctx->slab)) {
204 /* Only slab members should be looked up via interior pointers. */
269 assert(alloc_ctx.slab == edata_slab_get(edata));
272 if (likely(alloc_ctx.slab)) {
326 assert(alloc_ctx.slab == edata_slab_get(edata));
329 if (likely(alloc_ctx.slab)) {
348 * object, so base szind and slab on the given size.
351 alloc_ctx.slab = (alloc_ctx.szind < SC_NBINS);
360 || alloc_ctx.slab == (alloc_ctx.szind < SC_NBINS))
    [all...]
  /src/external/bsd/jemalloc.old/dist/include/jemalloc/internal/
emap.h 25 bool slab; member in struct:emap_alloc_ctx_t
31 bool slab; member in struct:emap_full_alloc_ctx_s
38 bool slab);
68 * szind and slab info appropriately.
72 szind_t szind, bool slab);
75 * Does the same thing, but with the interior of the range, for slab
79 * does both depending on the value of 'slab'. The answer is twofold:
85 * used, and passing a 'slab' parameter everywhere makes that more
110 * addresses to edatas. If the result is going to be used as a slab, you
113 * Remap simply changes the szind and slab status of an extent's boundar
    [all...]
arena_inlines_b.h 56 } else if (unlikely(!(is_slab = alloc_ctx->slab))) {
88 if (unlikely(!alloc_ctx->slab)) {
204 /* Only slab members should be looked up via interior pointers. */
269 assert(alloc_ctx.slab == edata_slab_get(edata));
272 if (likely(alloc_ctx.slab)) {
326 assert(alloc_ctx.slab == edata_slab_get(edata));
329 if (likely(alloc_ctx.slab)) {
348 * object, so base szind and slab on the given size.
351 alloc_ctx.slab = (alloc_ctx.szind < SC_NBINS);
360 || alloc_ctx.slab == (alloc_ctx.szind < SC_NBINS))
    [all...]
  /src/external/bsd/jemalloc.old/include/jemalloc/internal/
emap.h 25 bool slab; member in struct:emap_alloc_ctx_t
31 bool slab; member in struct:emap_full_alloc_ctx_s
38 bool slab);
68 * szind and slab info appropriately.
72 szind_t szind, bool slab);
75 * Does the same thing, but with the interior of the range, for slab
79 * does both depending on the value of 'slab'. The answer is twofold:
85 * used, and passing a 'slab' parameter everywhere makes that more
110 * addresses to edatas. If the result is going to be used as a slab, you
113 * Remap simply changes the szind and slab status of an extent's boundar
    [all...]
arena_inlines_b.h 56 } else if (unlikely(!(is_slab = alloc_ctx->slab))) {
88 if (unlikely(!alloc_ctx->slab)) {
204 /* Only slab members should be looked up via interior pointers. */
269 assert(alloc_ctx.slab == edata_slab_get(edata));
272 if (likely(alloc_ctx.slab)) {
326 assert(alloc_ctx.slab == edata_slab_get(edata));
329 if (likely(alloc_ctx.slab)) {
348 * object, so base szind and slab on the given size.
351 alloc_ctx.slab = (alloc_ctx.szind < SC_NBINS);
360 || alloc_ctx.slab == (alloc_ctx.szind < SC_NBINS))
    [all...]
  /src/external/mpl/bind/dist/lib/dns/
rdataslab.c 82 * rbtdb.c directly interacts with the slab's raw structures. If the
398 dns_rdataslab_size(unsigned char *slab, unsigned int reservelen) {
399 REQUIRE(slab != NULL);
401 unsigned char *current = slab + reservelen;
416 return (unsigned int)(current - slab);
420 dns_rdataslab_rdatasize(unsigned char *slab, unsigned int reservelen) {
421 REQUIRE(slab != NULL);
424 unsigned char *current = slab + reservelen;
444 dns_rdataslab_count(unsigned char *slab, unsigned int reservelen) {
445 REQUIRE(slab != NULL)
    [all...]

Completed in 39 milliseconds

1 2 3 4 5 6 7 8 91011>>