/src/sys/external/bsd/drm2/dist/drm/vmwgfx/ |
vmwgfx_reg.h | 46 u32 num_pages; member in struct:svga_guest_mem_descriptor
|
vmwgfx_mob.c | 59 * @num_pages Number of pages that make up the page table. 65 unsigned long num_pages; member in struct:vmw_mob 421 mob->num_pages = vmw_mob_calculate_pt_pages(data_pages); 448 ret = ttm_bo_create(&dev_priv->bdev, mob->num_pages * PAGE_SIZE,
|
vmwgfx_page_dirty.c | 240 pgoff_t num_pages = vbo->base.num_pages; local in function:vmw_bo_dirty_add 253 size = sizeof(*dirty) + BITS_TO_LONGS(num_pages) * sizeof(long); 268 dirty->bitmap_size = num_pages; 272 if (num_pages < PAGE_SIZE / sizeof(pte_t)) { 281 wp_shared_mapping_range(mapping, offset, num_pages); 282 clean_record_shared_mapping_range(mapping, offset, num_pages, 421 if (unlikely(page_offset >= bo->num_pages)) { 464 if (page_offset >= bo->num_pages ||
|
vmwgfx_bo.c | 231 place.lpfn = bo->num_pages; 252 bo->mem.start < bo->num_pages && 403 ret = ttm_bo_kmap(bo, 0, bo->num_pages, &vbo->map); 439 size_t num_pages = PAGE_ALIGN(size) >> PAGE_SHIFT; local in function:vmw_bo_acc_size 440 size_t page_array_size = ttm_round_pot(num_pages * sizeof(void *)); 454 ttm_round_pot(num_pages * sizeof(dma_addr_t));
|
vmwgfx_drv.h | 319 unsigned long num_pages; member in struct:vmw_sg_table 331 * @num_pages: Number of pages total. 343 unsigned long num_pages; member in struct:vmw_piter 714 unsigned long num_pages,
|
/src/sys/external/bsd/drm2/dist/drm/ |
drm_memory.c | 68 unsigned long i, num_pages = local in function:agp_remap 94 /* note: use vmalloc() because num_pages could be large... */ 95 page_map = vmalloc(array_size(num_pages, sizeof(struct page *))); 100 for (i = 0; i < num_pages; ++i) 102 addr = vmap(page_map, num_pages, VM_IOREMAP, PAGE_AGP);
|
/src/sys/external/bsd/drm2/dist/drm/xen/ |
xen_drm_front_gem.c | 33 size_t num_pages; member in struct:xen_gem_object 52 xen_obj->num_pages = DIV_ROUND_UP(buf_size, PAGE_SIZE); 53 xen_obj->pages = kvmalloc_array(xen_obj->num_pages, 107 ret = alloc_xenballooned_pages(xen_obj->num_pages, 111 xen_obj->num_pages, ret); 123 xen_obj->num_pages = DIV_ROUND_UP(size, PAGE_SIZE); 160 free_xenballooned_pages(xen_obj->num_pages, 187 return drm_prime_pages_to_sg(xen_obj->pages, xen_obj->num_pages); 212 NULL, xen_obj->num_pages); 258 ret = vm_map_pages(vma, xen_obj->pages, xen_obj->num_pages); [all...] |
/src/sys/external/bsd/drm2/dist/include/drm/ttm/ |
ttm_tt.h | 96 * @num_pages: Number of pages in the page array. 112 unsigned long num_pages; member in struct:ttm_tt
|
ttm_bo_api.h | 101 * @num_pages: Actual size of memory region in pages. 114 unsigned long num_pages; member in struct:ttm_mem_reg 150 * @num_pages: Actual number of pages. 195 unsigned long num_pages; member in struct:ttm_buffer_object 663 * @num_pages: Number of pages to map. 675 unsigned long num_pages, struct ttm_bo_kmap_obj *map);
|
/src/sys/external/bsd/drm2/dist/drm/i915/gem/ |
i915_gem_userptr.c | 442 struct page **pvec, unsigned long num_pages) 454 ret = __sg_alloc_table_from_pages(st, pvec, num_pages, 455 0, num_pages << PAGE_SHIFT, 603 const unsigned long num_pages = obj->base.size >> PAGE_SHIFT; local in function:i915_gem_userptr_get_pages 648 pvec = kvmalloc_array(num_pages, sizeof(struct page *), 654 num_pages, 663 } else if (pinned < num_pages) { 667 pages = __i915_gem_userptr_alloc_pages(obj, pvec, num_pages);
|
/src/sys/external/bsd/drm2/dist/drm/vgem/ |
vgem_drv.c | 83 loff_t num_pages; local in function:vgem_gem_fault 87 num_pages = DIV_ROUND_UP(obj->base.size, PAGE_SIZE); 89 if (page_offset >= num_pages)
|
/src/sys/external/bsd/drm2/dist/drm/via/ |
via_dmablit.h | 51 unsigned long num_pages; member in struct:_drm_via_sg_info
|
/src/sys/arch/arm/acpi/ |
acpi_machdep.c | 107 const uint64_t num_pages = be64dec(&map[3]); local in function:acpi_md_pmapflags 110 if (pa >= phys_start && pa < phys_start + (num_pages * EFI_PAGE_SIZE)) {
|
/src/sys/external/bsd/drm2/dist/drm/radeon/ |
radeon_ttm.c | 226 unsigned num_pages; local in function:radeon_move_blit 263 num_pages = new_mem->num_pages * (PAGE_SIZE / RADEON_GPU_PAGE_SIZE); 264 fence = radeon_copy(rdev, old_start, new_start, num_pages, bo->base.resv); 415 atomic64_add((u64)bo->num_pages << PAGE_SHIFT, &rdev->num_bytes_moved); 426 mem->bus.size = mem->num_pages << PAGE_SHIFT; 542 unsigned long end = gtt->userptr + ttm->num_pages * PAGE_SIZE; 567 .iov_len = ttm->num_pages << PAGE_SHIFT, 573 .uio_resid = ttm->num_pages << PAGE_SHIFT, 583 ttm->num_pages << PAGE_SHIFT 620 unsigned num_pages = ttm->num_pages - pinned; local in function:radeon_ttm_tt_pin_userptr [all...] |
/src/sys/external/bsd/drm2/dist/drm/ttm/ |
ttm_page_alloc.c | 1050 ttm_put_pages(ttm->pages, ttm->num_pages, ttm->page_flags, 1064 if (ttm_check_under_lowerlimit(mem_glob, ttm->num_pages, ctx)) 1067 ret = ttm_get_pages(ttm->pages, ttm->num_pages, ttm->page_flags, 1074 for (i = 0; i < ttm->num_pages; ++i) { 1098 ttm_pool_unpopulate_helper(ttm, ttm->num_pages); 1112 for (i = 0; i < tt->ttm.num_pages; ++i) { 1114 size_t num_pages = 1; local in function:ttm_populate_and_map_pages 1116 for (j = i + 1; j < tt->ttm.num_pages; ++j) { 1120 ++num_pages; 1124 0, num_pages * PAGE_SIZE 1151 size_t num_pages = 1; local in function:ttm_unmap_and_unpopulate_pages [all...] |
ttm_bo.c | 127 bo, bo->mem.num_pages, bo->mem.size >> 10, 416 ctx->bytes_moved += bo->num_pages << PAGE_SHIFT; 1194 mem.num_pages = bo->num_pages; 1195 mem.size = mem.num_pages << PAGE_SHIFT; 1224 (heap->lpfn != 0 && (mem->start + mem->num_pages) > heap->lpfn))) 1305 unsigned long num_pages; local in function:ttm_bo_init_reserved 1323 num_pages = (size + PAGE_SIZE - 1) >> PAGE_SHIFT; 1324 if (num_pages == 0) { 1343 bo->num_pages = num_pages [all...] |
ttm_page_alloc_dma.c | 360 unsigned num_pages; local in function:ttm_dma_page_put 364 num_pages = pool->size / PAGE_SIZE; 365 if (ttm_set_pages_wb(page, num_pages)) 367 pool->dev_name, num_pages); 895 unsigned long num_pages = ttm->num_pages; local in function:ttm_dma_populate 905 if (ttm_check_under_lowerlimit(mem_glob, num_pages, ctx)) 926 while (num_pages >= HPAGE_PMD_NR) { 948 num_pages -= HPAGE_PMD_NR; 963 while (num_pages) { [all...] |
/src/sys/external/bsd/drm2/dist/drm/i915/ |
i915_gpu_error.c | 286 if (dst->page_count >= dst->num_pages) 984 unsigned long num_pages; local in function:i915_vma_coredump_create 993 num_pages = min_t(u64, vma->size, vma->obj->base.size) >> PAGE_SHIFT; 994 num_pages = DIV_ROUND_UP(10 * num_pages, 8); /* worstcase zlib growth */ 995 dst = kmalloc(sizeof(*dst) + num_pages * sizeof(u32 *), ALLOW_FAIL); 1010 dst->num_pages = num_pages;
|
i915_gpu_error.h | 44 int num_pages; member in struct:i915_vma_coredump
|
/src/sys/dev/pci/ |
pvscsi.h | 124 uint32_t num_pages; member in struct:pvscsi_cmd_desc_setup_msg_ring
|
/src/sys/external/bsd/drm2/dist/drm/amd/amdgpu/ |
amdgpu_ttm.c | 71 struct ttm_mem_reg *mem, unsigned num_pages, 458 new_mem->num_pages << PAGE_SHIFT, 614 if (nodes->size != mem->num_pages) 705 atomic64_add((u64)bo->num_pages << PAGE_SHIFT, &adev->num_bytes_moved); 722 mem->bus.size = mem->num_pages << PAGE_SHIFT; 743 (mm_node->size == mem->num_pages)) 867 range->pfns = kvmalloc_array(ttm->num_pages, sizeof(*range->pfns), 904 for (i = 0; i < ttm->num_pages; i++) { 947 gtt->userptr, ttm->num_pages); 979 for (i = 0; i < ttm->num_pages; ++i 1614 unsigned long num_pages = bo->mem.num_pages; local in function:amdgpu_ttm_bo_eviction_valuable 2322 unsigned long num_pages; local in function:amdgpu_fill_buffer [all...] |
/src/sys/external/bsd/vchiq/dist/interface/vchiq_arm/ |
vchiq_arm.c | 1477 int num_pages; local in function:dump_phys_mem 1495 num_pages = (offset + num_bytes + PAGE_SIZE - 1) / PAGE_SIZE; 1497 pages = kmalloc(sizeof(struct page *) * num_pages, GFP_KERNEL); 1501 num_pages); 1509 num_pages, /* len */ 1545 for (page_idx = 0; page_idx < num_pages; page_idx++)
|