HomeSort by: relevance | last modified time | path
    Searched refs:PAGE (Results 1 - 25 of 199) sorted by relevancy

1 2 3 4 5 6 7 8

  /src/external/bsd/jemalloc.old/dist/include/jemalloc/internal/
sec_opts.h 49 (32 * 1024) < PAGE ? PAGE : (32 * 1024), \
  /src/external/bsd/jemalloc.old/include/jemalloc/internal/
sec_opts.h 49 (32 * 1024) < PAGE ? PAGE : (32 * 1024), \
  /src/external/bsd/jemalloc.old/dist/test/unit/
hpdata.c 10 /* Allocating a page at a time, we should do first fit. */
15 void *alloc = hpdata_reserve_alloc(&hpdata, PAGE);
16 expect_ptr_eq((char *)HPDATA_ADDR + i * PAGE, alloc, "");
27 hpdata_unreserve(&hpdata, (char *)HPDATA_ADDR + 10 * PAGE, 2 * PAGE);
31 hpdata_unreserve(&hpdata, (char *)HPDATA_ADDR + 12 * PAGE, 2 * PAGE);
35 hpdata_unreserve(&hpdata, (char *)HPDATA_ADDR + 8 * PAGE, 2 * PAGE);
40 * Leave page 14 reserved, but free page 15 (this test the case wher
    [all...]
sc.c 10 size_t pgs_too_big = (tiny_size * BITMAP_MAXBITS + PAGE - 1) / PAGE + 1;
14 sc_data_update_slab_size(&data, 1, 10 * PAGE, 1);
19 if (reg_size <= PAGE) {
20 expect_d_eq(sc->pgs, 1, "Ignored valid page size hint");
23 "Allowed invalid page size hint");
sec.c 143 ta->next_ptr = 10 * PAGE;
168 * 11 allocs apiece of 1-PAGE and 2-PAGE objects means that we should be
178 test_sec_init(&sec, &ta.pai, /* nshards */ 1, /* max_alloc */ 2 * PAGE,
179 /* max_bytes */ 2 * (NALLOCS * PAGE + NALLOCS * 2 * PAGE));
181 one_page[i] = pai_alloc(tsdn, &sec.pai, PAGE, PAGE,
185 two_page[i] = pai_alloc(tsdn, &sec.pai, 2 * PAGE, PAGE,
    [all...]
pages.c 8 alloc_size = HUGEPAGE * 2 - PAGE;
10 pages = pages_map(NULL, alloc_size, PAGE, &commit);
san_bump.c 21 size_t alloc_size = PAGE * 16;
42 size_t alloc_size2 = PAGE * 28;
63 expect_zu_ge(between_allocs, PAGE,
64 "Guard page between allocs is missing");
100 for (unsigned i = 0; i < alloc_size / PAGE; ++i) {
101 *((char *)ptr + PAGE * i) = 1;
san.c 17 * This test allocates page sized slabs and checks that every two slabs have
18 * at least one page in between them. That page is supposed to be the guard
19 * page.
29 size_t slab_sz = PAGE;
37 if ((uintptr_t)ptr % PAGE == 0) {
39 "Unexpectedly large number of page aligned allocs");
42 if (((uintptr_t)ptr + (uintptr_t)sz) % PAGE == 0) {
59 expect_zu_ge((size_t)ptr_diff, slab_sz + PAGE,
90 expect_zu_ge((size_t)ptr_diff, large_sz + 2 * PAGE,
    [all...]
  /src/external/bsd/jemalloc/dist/include/jemalloc/internal/
sec_opts.h 40 #define SEC_OPTS_MAX_ALLOC_DEFAULT ((32 * 1024) < PAGE ? PAGE : (32 * 1024))
hpa_opts.h 27 * purge_threshold=HUGEPAGE bytes (2097152 for 2Mb page), although other
34 * that memory at page-fault time. This is usually more efficient than doing
38 * for page not to be purged quickly, etc. This is a good choice if you can
117 * Minimum number of inactive bytes needed for a non-empty page to be
121 * >= purge_threshold, the page is purgable. Empty pages are always
128 * bytes, and not just up to the threshold. If this is PAGE bytes, then
131 * dirty_mult, IOW, if no page meets purge_threshold, we will not purge
137 * Minimum number of ms that needs to elapse between HP page becoming
141 * that memory. Setting it to 0 means that page is eligible for purging
143 * purgability of the page changes (page goes from being non-purgable t
    [all...]
pages.h 7 /* Actual operating system page size, detected during bootstrap, <= PAGE. */
10 /* Page size. LG_PAGE is determined by the configure script. */
14 #define PAGE ((size_t)(1U << LG_PAGE))
15 #define PAGE_MASK ((size_t)(PAGE - 1))
16 /* Return the page base address for the page containing address a. */
17 #define PAGE_ADDR2BASE(a) ALIGNMENT_ADDR2BASE(a, PAGE)
23 /* Huge page size. LG_HUGEPAGE is determined by the configure script. */
29 * page features (HPA, metadata_thp) are primarily designed with a 2M THP siz
    [all...]
  /src/external/bsd/jemalloc/dist/test/unit/
hpdata.c 10 /* Allocating a page at a time, we should do first fit. */
15 void *alloc = hpdata_reserve_alloc(&hpdata, PAGE);
16 expect_ptr_eq((char *)HPDATA_ADDR + i * PAGE, alloc, "");
27 hpdata_unreserve(&hpdata, (char *)HPDATA_ADDR + 10 * PAGE, 2 * PAGE);
31 hpdata_unreserve(&hpdata, (char *)HPDATA_ADDR + 12 * PAGE, 2 * PAGE);
35 hpdata_unreserve(&hpdata, (char *)HPDATA_ADDR + 8 * PAGE, 2 * PAGE);
40 * Leave page 14 reserved, but free page 15 (this test the case wher
    [all...]
sc.c 10 size_t pgs_too_big = (tiny_size * BITMAP_MAXBITS + PAGE - 1) / PAGE + 1;
14 sc_data_update_slab_size(&data, 1, 10 * PAGE, 1);
19 if (reg_size <= PAGE) {
20 expect_d_eq(sc->pgs, 1, "Ignored valid page size hint");
23 sc->pgs, 1, "Allowed invalid page size hint");
sec.c 38 opts.max_alloc = PAGE;
39 opts.max_bytes = 512 * PAGE;
44 edata_t *edata = sec_alloc(tsdn, &tdata.sec, PAGE);
54 opts.max_alloc = 2 * PAGE;
55 opts.max_bytes = 512 * PAGE;
60 edata_t *edata = sec_alloc(tsdn, &tdata.sec, 3 * PAGE);
61 expect_ptr_null(edata, "max_alloc is 2*PAGE, should not alloc 3*PAGE");
70 opts.max_alloc = 2 * PAGE;
71 opts.max_bytes = 4 * PAGE;
    [all...]
pages.c 8 alloc_size = HUGEPAGE * 2 - PAGE;
10 pages = pages_map(NULL, alloc_size, PAGE, &commit);
hpa_sec_integration.c 42 PAGE,
162 sec_opts.max_alloc = 2 * PAGE;
163 sec_opts.max_bytes = NALLOCS * PAGE;
170 /* alloc 1 PAGE, confirm sec has fill_extra bytes. */
171 edata_t *edata1 = pai_alloc(tsdn, &shard->pai, PAGE, PAGE, false, false,
179 expect_zu_eq(hpa_stats.secstats.bytes, PAGE * sec_opts.batch_fill_extra,
185 edatas[i] = pai_alloc(tsdn, &shard->pai, PAGE, PAGE, false,
192 expect_zu_eq(hpa_stats.secstats.bytes, PAGE, "2 refills (at 0 and 4)")
    [all...]
  /src/external/bsd/jemalloc/include/jemalloc/internal/
sec_opts.h 40 #define SEC_OPTS_MAX_ALLOC_DEFAULT ((32 * 1024) < PAGE ? PAGE : (32 * 1024))
hpa_opts.h 27 * purge_threshold=HUGEPAGE bytes (2097152 for 2Mb page), although other
34 * that memory at page-fault time. This is usually more efficient than doing
38 * for page not to be purged quickly, etc. This is a good choice if you can
117 * Minimum number of inactive bytes needed for a non-empty page to be
121 * >= purge_threshold, the page is purgable. Empty pages are always
128 * bytes, and not just up to the threshold. If this is PAGE bytes, then
131 * dirty_mult, IOW, if no page meets purge_threshold, we will not purge
137 * Minimum number of ms that needs to elapse between HP page becoming
141 * that memory. Setting it to 0 means that page is eligible for purging
143 * purgability of the page changes (page goes from being non-purgable t
    [all...]
pages.h 7 /* Actual operating system page size, detected during bootstrap, <= PAGE. */
10 /* Page size. LG_PAGE is determined by the configure script. */
14 #define PAGE ((size_t)(1U << LG_PAGE))
15 #define PAGE_MASK ((size_t)(PAGE - 1))
16 /* Return the page base address for the page containing address a. */
17 #define PAGE_ADDR2BASE(a) ALIGNMENT_ADDR2BASE(a, PAGE)
23 /* Huge page size. LG_HUGEPAGE is determined by the configure script. */
29 * page features (HPA, metadata_thp) are primarily designed with a 2M THP siz
    [all...]
  /src/lib/libc/db/btree/
extern.h 42 int __bt_dleaf(BTREE *, const DBT *, PAGE *, unsigned int);
44 int __bt_free(BTREE *, PAGE *);
46 PAGE *__bt_new(BTREE *, pgno_t *);
51 int __bt_relink(BTREE *, PAGE *);
56 int __bt_split(BTREE *, PAGE *,
65 void __bt_dmpage(PAGE *);
67 void __bt_dpage(PAGE *);
bt_page.c 50 * Put a page on the freelist.
54 * h: page to free
60 * mpool_put's the page.
63 __bt_free(BTREE *t, PAGE *h)
65 /* Insert the page at the head of the free list. */
71 /* Make sure the page gets written back. */
77 * Get a new page, preferably from the freelist.
81 * npg: storage for page number.
84 * Pointer to a page, NULL on error.
86 PAGE *
    [all...]
bt_split.c 54 static int bt_broot(BTREE *, PAGE *, PAGE *, PAGE *);
55 static PAGE *bt_page(BTREE *, PAGE *, PAGE **, PAGE **, indx_t *, size_t);
57 static PAGE *bt_psplit(BTREE *, PAGE *, PAGE *, PAGE *, indx_t *, size_t)
    [all...]
  /src/external/gpl2/groff/font/devlbp/
Makefile 18 if test "${PAGE}" = A4; then \
  /src/external/bsd/jemalloc.old/dist/src/
extent_mmap.c 23 assert(alignment == ALIGNMENT_CEILING(alignment, PAGE));
  /src/external/bsd/jemalloc/dist/src/
extent_mmap.c 23 assert(alignment == ALIGNMENT_CEILING(alignment, PAGE));

Completed in 80 milliseconds

1 2 3 4 5 6 7 8