| /src/external/bsd/jemalloc.old/dist/include/jemalloc/internal/ |
| sec_opts.h | 49 (32 * 1024) < PAGE ? PAGE : (32 * 1024), \
|
| /src/external/bsd/jemalloc.old/include/jemalloc/internal/ |
| sec_opts.h | 49 (32 * 1024) < PAGE ? PAGE : (32 * 1024), \
|
| /src/external/bsd/jemalloc.old/dist/test/unit/ |
| hpdata.c | 10 /* Allocating a page at a time, we should do first fit. */ 15 void *alloc = hpdata_reserve_alloc(&hpdata, PAGE); 16 expect_ptr_eq((char *)HPDATA_ADDR + i * PAGE, alloc, ""); 27 hpdata_unreserve(&hpdata, (char *)HPDATA_ADDR + 10 * PAGE, 2 * PAGE); 31 hpdata_unreserve(&hpdata, (char *)HPDATA_ADDR + 12 * PAGE, 2 * PAGE); 35 hpdata_unreserve(&hpdata, (char *)HPDATA_ADDR + 8 * PAGE, 2 * PAGE); 40 * Leave page 14 reserved, but free page 15 (this test the case wher [all...] |
| sc.c | 10 size_t pgs_too_big = (tiny_size * BITMAP_MAXBITS + PAGE - 1) / PAGE + 1; 14 sc_data_update_slab_size(&data, 1, 10 * PAGE, 1); 19 if (reg_size <= PAGE) { 20 expect_d_eq(sc->pgs, 1, "Ignored valid page size hint"); 23 "Allowed invalid page size hint");
|
| sec.c | 143 ta->next_ptr = 10 * PAGE; 168 * 11 allocs apiece of 1-PAGE and 2-PAGE objects means that we should be 178 test_sec_init(&sec, &ta.pai, /* nshards */ 1, /* max_alloc */ 2 * PAGE, 179 /* max_bytes */ 2 * (NALLOCS * PAGE + NALLOCS * 2 * PAGE)); 181 one_page[i] = pai_alloc(tsdn, &sec.pai, PAGE, PAGE, 185 two_page[i] = pai_alloc(tsdn, &sec.pai, 2 * PAGE, PAGE, [all...] |
| pages.c | 8 alloc_size = HUGEPAGE * 2 - PAGE; 10 pages = pages_map(NULL, alloc_size, PAGE, &commit);
|
| san_bump.c | 21 size_t alloc_size = PAGE * 16; 42 size_t alloc_size2 = PAGE * 28; 63 expect_zu_ge(between_allocs, PAGE, 64 "Guard page between allocs is missing"); 100 for (unsigned i = 0; i < alloc_size / PAGE; ++i) { 101 *((char *)ptr + PAGE * i) = 1;
|
| san.c | 17 * This test allocates page sized slabs and checks that every two slabs have 18 * at least one page in between them. That page is supposed to be the guard 19 * page. 29 size_t slab_sz = PAGE; 37 if ((uintptr_t)ptr % PAGE == 0) { 39 "Unexpectedly large number of page aligned allocs"); 42 if (((uintptr_t)ptr + (uintptr_t)sz) % PAGE == 0) { 59 expect_zu_ge((size_t)ptr_diff, slab_sz + PAGE, 90 expect_zu_ge((size_t)ptr_diff, large_sz + 2 * PAGE, [all...] |
| /src/external/bsd/jemalloc/dist/include/jemalloc/internal/ |
| sec_opts.h | 40 #define SEC_OPTS_MAX_ALLOC_DEFAULT ((32 * 1024) < PAGE ? PAGE : (32 * 1024))
|
| hpa_opts.h | 27 * purge_threshold=HUGEPAGE bytes (2097152 for 2Mb page), although other 34 * that memory at page-fault time. This is usually more efficient than doing 38 * for page not to be purged quickly, etc. This is a good choice if you can 117 * Minimum number of inactive bytes needed for a non-empty page to be 121 * >= purge_threshold, the page is purgable. Empty pages are always 128 * bytes, and not just up to the threshold. If this is PAGE bytes, then 131 * dirty_mult, IOW, if no page meets purge_threshold, we will not purge 137 * Minimum number of ms that needs to elapse between HP page becoming 141 * that memory. Setting it to 0 means that page is eligible for purging 143 * purgability of the page changes (page goes from being non-purgable t [all...] |
| pages.h | 7 /* Actual operating system page size, detected during bootstrap, <= PAGE. */ 10 /* Page size. LG_PAGE is determined by the configure script. */ 14 #define PAGE ((size_t)(1U << LG_PAGE)) 15 #define PAGE_MASK ((size_t)(PAGE - 1)) 16 /* Return the page base address for the page containing address a. */ 17 #define PAGE_ADDR2BASE(a) ALIGNMENT_ADDR2BASE(a, PAGE) 23 /* Huge page size. LG_HUGEPAGE is determined by the configure script. */ 29 * page features (HPA, metadata_thp) are primarily designed with a 2M THP siz [all...] |
| /src/external/bsd/jemalloc/dist/test/unit/ |
| hpdata.c | 10 /* Allocating a page at a time, we should do first fit. */ 15 void *alloc = hpdata_reserve_alloc(&hpdata, PAGE); 16 expect_ptr_eq((char *)HPDATA_ADDR + i * PAGE, alloc, ""); 27 hpdata_unreserve(&hpdata, (char *)HPDATA_ADDR + 10 * PAGE, 2 * PAGE); 31 hpdata_unreserve(&hpdata, (char *)HPDATA_ADDR + 12 * PAGE, 2 * PAGE); 35 hpdata_unreserve(&hpdata, (char *)HPDATA_ADDR + 8 * PAGE, 2 * PAGE); 40 * Leave page 14 reserved, but free page 15 (this test the case wher [all...] |
| sc.c | 10 size_t pgs_too_big = (tiny_size * BITMAP_MAXBITS + PAGE - 1) / PAGE + 1; 14 sc_data_update_slab_size(&data, 1, 10 * PAGE, 1); 19 if (reg_size <= PAGE) { 20 expect_d_eq(sc->pgs, 1, "Ignored valid page size hint"); 23 sc->pgs, 1, "Allowed invalid page size hint");
|
| sec.c | 38 opts.max_alloc = PAGE; 39 opts.max_bytes = 512 * PAGE; 44 edata_t *edata = sec_alloc(tsdn, &tdata.sec, PAGE); 54 opts.max_alloc = 2 * PAGE; 55 opts.max_bytes = 512 * PAGE; 60 edata_t *edata = sec_alloc(tsdn, &tdata.sec, 3 * PAGE); 61 expect_ptr_null(edata, "max_alloc is 2*PAGE, should not alloc 3*PAGE"); 70 opts.max_alloc = 2 * PAGE; 71 opts.max_bytes = 4 * PAGE; [all...] |
| pages.c | 8 alloc_size = HUGEPAGE * 2 - PAGE; 10 pages = pages_map(NULL, alloc_size, PAGE, &commit);
|
| hpa_sec_integration.c | 42 PAGE, 162 sec_opts.max_alloc = 2 * PAGE; 163 sec_opts.max_bytes = NALLOCS * PAGE; 170 /* alloc 1 PAGE, confirm sec has fill_extra bytes. */ 171 edata_t *edata1 = pai_alloc(tsdn, &shard->pai, PAGE, PAGE, false, false, 179 expect_zu_eq(hpa_stats.secstats.bytes, PAGE * sec_opts.batch_fill_extra, 185 edatas[i] = pai_alloc(tsdn, &shard->pai, PAGE, PAGE, false, 192 expect_zu_eq(hpa_stats.secstats.bytes, PAGE, "2 refills (at 0 and 4)") [all...] |
| /src/external/bsd/jemalloc/include/jemalloc/internal/ |
| sec_opts.h | 40 #define SEC_OPTS_MAX_ALLOC_DEFAULT ((32 * 1024) < PAGE ? PAGE : (32 * 1024))
|
| hpa_opts.h | 27 * purge_threshold=HUGEPAGE bytes (2097152 for 2Mb page), although other 34 * that memory at page-fault time. This is usually more efficient than doing 38 * for page not to be purged quickly, etc. This is a good choice if you can 117 * Minimum number of inactive bytes needed for a non-empty page to be 121 * >= purge_threshold, the page is purgable. Empty pages are always 128 * bytes, and not just up to the threshold. If this is PAGE bytes, then 131 * dirty_mult, IOW, if no page meets purge_threshold, we will not purge 137 * Minimum number of ms that needs to elapse between HP page becoming 141 * that memory. Setting it to 0 means that page is eligible for purging 143 * purgability of the page changes (page goes from being non-purgable t [all...] |
| pages.h | 7 /* Actual operating system page size, detected during bootstrap, <= PAGE. */ 10 /* Page size. LG_PAGE is determined by the configure script. */ 14 #define PAGE ((size_t)(1U << LG_PAGE)) 15 #define PAGE_MASK ((size_t)(PAGE - 1)) 16 /* Return the page base address for the page containing address a. */ 17 #define PAGE_ADDR2BASE(a) ALIGNMENT_ADDR2BASE(a, PAGE) 23 /* Huge page size. LG_HUGEPAGE is determined by the configure script. */ 29 * page features (HPA, metadata_thp) are primarily designed with a 2M THP siz [all...] |
| /src/lib/libc/db/btree/ |
| extern.h | 42 int __bt_dleaf(BTREE *, const DBT *, PAGE *, unsigned int); 44 int __bt_free(BTREE *, PAGE *); 46 PAGE *__bt_new(BTREE *, pgno_t *); 51 int __bt_relink(BTREE *, PAGE *); 56 int __bt_split(BTREE *, PAGE *, 65 void __bt_dmpage(PAGE *); 67 void __bt_dpage(PAGE *);
|
| bt_page.c | 50 * Put a page on the freelist. 54 * h: page to free 60 * mpool_put's the page. 63 __bt_free(BTREE *t, PAGE *h) 65 /* Insert the page at the head of the free list. */ 71 /* Make sure the page gets written back. */ 77 * Get a new page, preferably from the freelist. 81 * npg: storage for page number. 84 * Pointer to a page, NULL on error. 86 PAGE * [all...] |
| bt_split.c | 54 static int bt_broot(BTREE *, PAGE *, PAGE *, PAGE *); 55 static PAGE *bt_page(BTREE *, PAGE *, PAGE **, PAGE **, indx_t *, size_t); 57 static PAGE *bt_psplit(BTREE *, PAGE *, PAGE *, PAGE *, indx_t *, size_t) [all...] |
| /src/external/gpl2/groff/font/devlbp/ |
| Makefile | 18 if test "${PAGE}" = A4; then \
|
| /src/external/bsd/jemalloc.old/dist/src/ |
| extent_mmap.c | 23 assert(alignment == ALIGNMENT_CEILING(alignment, PAGE));
|
| /src/external/bsd/jemalloc/dist/src/ |
| extent_mmap.c | 23 assert(alignment == ALIGNMENT_CEILING(alignment, PAGE));
|