Lines Matching refs:npages
274 * This macro was added to convert uvmexp.npages from int to psize_t
286 uvmexp.npages = 0;
492 ATF_REQUIRE_EQ(0, uvmexp.npages);
497 ATF_REQUIRE_EQ(0, uvmexp.npages);
506 , INT_TO_PSIZE_T(uvmexp.npages));
515 + npages3, INT_TO_PSIZE_T(uvmexp.npages));
559 ATF_REQUIRE_EQ(0, uvmexp.npages); /* Boot time sanity */
584 ATF_CHECK_EQ(atop(FIVEONETWO_KILO), uvmexp.npages);
653 ATF_REQUIRE_EQ(0, uvmexp.npages);
666 ATF_REQUIRE_EQ(0, uvmexp.npages);
697 ATF_REQUIRE_EQ(0, uvmexp.npages);
714 ATF_REQUIRE_EQ(npages1 + npages2, INT_TO_PSIZE_T(uvmexp.npages));
737 ATF_REQUIRE_EQ(0, uvmexp.npages);
748 ATF_REQUIRE_EQ(0, uvmexp.npages);
781 const size_t npages = UVM_PHYSSEG_BOOT_UNPLUG_MAX; /* Number of pages */
786 slab = malloc(sizeof(struct vm_page) * npages * 2);
790 uvm_physseg_seg_chomp_slab(PHYSSEG_NODE_TO_HANDLE(seg), slab, npages * 2);
805 for (i = 0; i < npages; i++) {
815 uvm_physseg_seg_chomp_slab(PHYSSEG_NODE_TO_HANDLE(seg), slab, npages * 2);
818 for (i = 0; i < npages; i++) {
836 const size_t npages = UVM_PHYSSEG_BOOT_UNPLUG_MAX; /* Number of pages */
841 slab = malloc(sizeof(struct vm_page) * npages * 2);
845 uvm_physseg_seg_chomp_slab(PHYSSEG_NODE_TO_HANDLE(seg), slab, npages * 2);
847 pgs = uvm_physseg_seg_alloc_from_slab(PHYSSEG_NODE_TO_HANDLE(seg), npages);
855 pgs = uvm_physseg_seg_alloc_from_slab(PHYSSEG_NODE_TO_HANDLE(seg), npages);
860 pgs = uvm_physseg_seg_alloc_from_slab(PHYSSEG_NODE_TO_HANDLE(seg), npages);
868 "uvmexp.npages");
874 const size_t npages = UVM_PHYSSEG_BOOT_UNPLUG_MAX; /* Number of pages */
879 slab = malloc(sizeof(struct vm_page) * npages * 2);
883 uvm_physseg_seg_chomp_slab(PHYSSEG_NODE_TO_HANDLE(seg), slab, npages * 2);
885 pgs = uvm_physseg_seg_alloc_from_slab(PHYSSEG_NODE_TO_HANDLE(seg), npages);
887 ATF_REQUIRE_EQ(0, uvmexp.npages);
890 seg->end = npages;
893 seg->avail_end = npages;
897 ATF_REQUIRE_EQ(npages, INT_TO_PSIZE_T(uvmexp.npages));
910 psize_t npages = (VALID_END_PFN_1 - VALID_START_PFN_1);
911 struct vm_page *pgs = malloc(sizeof(struct vm_page) * npages);
919 ATF_CHECK_EQ(0, uvmexp.npages);
928 uvm_physseg_seg_chomp_slab(upm, pgs, npages);
933 uvm_physseg_alloc_from_slab(upm, npages);
943 ATF_CHECK_EQ(PAGE_COUNT_1M, uvmexp.npages);
950 ATF_CHECK_EQ(PAGE_COUNT_1M + PAGE_COUNT_1M, uvmexp.npages);
971 ATF_REQUIRE_EQ(0, uvmexp.npages);
984 ATF_REQUIRE_EQ(0, uvmexp.npages);
1007 ATF_REQUIRE_EQ(0, uvmexp.npages);
1032 ATF_REQUIRE_EQ(0, uvmexp.npages);
1045 ATF_REQUIRE_EQ(0, uvmexp.npages);
1068 ATF_REQUIRE_EQ(0, uvmexp.npages);
1093 ATF_REQUIRE_EQ(0, uvmexp.npages);
1104 ATF_REQUIRE_EQ(0, uvmexp.npages);
1129 ATF_REQUIRE_EQ(0, uvmexp.npages);
1154 ATF_REQUIRE_EQ(0, uvmexp.npages);
1167 ATF_REQUIRE_EQ(0, uvmexp.npages);
1190 ATF_REQUIRE_EQ(0, uvmexp.npages);
1220 ATF_REQUIRE_EQ(0, uvmexp.npages);
1232 ATF_REQUIRE_EQ(0, uvmexp.npages);
1248 ATF_REQUIRE_EQ(0, uvmexp.npages);
1292 ATF_REQUIRE_EQ(0, uvmexp.npages);
1304 ATF_REQUIRE_EQ(0, uvmexp.npages);
1321 ATF_REQUIRE_EQ(0, uvmexp.npages);
1369 ATF_REQUIRE_EQ(0, uvmexp.npages);
1387 ATF_CHECK_EQ(0, uvmexp.npages);
1406 ATF_CHECK_EQ(0, uvmexp.npages);
1439 ATF_REQUIRE_EQ(0, uvmexp.npages);
1457 ATF_REQUIRE_EQ(0, uvmexp.npages);
1476 ATF_REQUIRE_EQ(0, uvmexp.npages);
1498 psize_t npages = (VALID_END_PFN_1 - VALID_START_PFN_1);
1500 struct vm_page *pgs = malloc(sizeof(struct vm_page) * npages);
1508 ATF_REQUIRE_EQ(0, uvmexp.npages);
1514 ATF_REQUIRE_EQ(PAGE_COUNT_1M, uvmexp.npages);
1647 psize_t npages = (VALID_END_PFN_1 - VALID_START_PFN_1);
1649 struct vm_page *pgs = malloc(sizeof(struct vm_page) * npages);
1704 psize_t npages = (VALID_END_PFN_1 - VALID_START_PFN_1);
1706 struct vm_page *pgs = malloc(sizeof(struct vm_page) * npages);
1718 ATF_REQUIRE_EQ(0, uvmexp.npages);
1723 ATF_REQUIRE_EQ(PAGE_COUNT_1M, uvmexp.npages);
1748 psize_t npages = (VALID_END_PFN_1 - VALID_START_PFN_1);
1750 struct vm_page *pgs = malloc(sizeof(struct vm_page) * npages);
1764 ATF_REQUIRE_EQ(0, uvmexp.npages);
1769 ATF_REQUIRE_EQ(PAGE_COUNT_1M, uvmexp.npages);
1823 ATF_REQUIRE_EQ(0, uvmexp.npages);
1832 ATF_REQUIRE_EQ(0, uvmexp.npages);
1902 ATF_REQUIRE_EQ(0, uvmexp.npages);
1926 * Would uvmexp.npages reduce everytime an uvm_page_physunload is called?
1928 psize_t npages = (VALID_END_PFN_2 - VALID_START_PFN_2);
1930 struct vm_page *pgs = malloc(sizeof(struct vm_page) * npages);
1942 ATF_REQUIRE_EQ(0, uvmexp.npages);
1974 * Would uvmexp.npages reduce everytime an uvm_page_physunload is called?
1990 ATF_REQUIRE_EQ(0, uvmexp.npages);
2033 psize_t npages = (VALID_END_PFN_2 - VALID_START_PFN_2);
2035 struct vm_page *pgs = malloc(sizeof(struct vm_page) * npages);
2055 ATF_REQUIRE_EQ(0, uvmexp.npages);
2086 * Would uvmexp.npages reduce everytime an uvm_page_physunload is called?
2104 ATF_REQUIRE_EQ(0, uvmexp.npages);
2144 * Would uvmexp.npages reduce everytime an uvm_page_physunload is called?
2163 ATF_REQUIRE_EQ(0, uvmexp.npages);
2207 psize_t npages = (VALID_END_PFN_2 - VALID_START_PFN_2);
2209 struct vm_page *pgs = malloc(sizeof(struct vm_page) * npages);
2221 ATF_REQUIRE_EQ(0, uvmexp.npages);
2237 * Would uvmexp.npages reduce everytime an uvm_page_physunload is called?
2249 ATF_REQUIRE_EQ(0, uvmexp.npages);
2316 ATF_REQUIRE_EQ(0, uvmexp.npages);