HomeSort by: relevance | last modified time | path
    Searched defs:num_pages (Results 1 - 22 of 22) sorted by relevancy

  /src/sys/external/bsd/drm2/dist/drm/vmwgfx/
vmwgfx_reg.h 46 u32 num_pages; member in struct:svga_guest_mem_descriptor
vmwgfx_mob.c 59 * @num_pages Number of pages that make up the page table.
65 unsigned long num_pages; member in struct:vmw_mob
421 mob->num_pages = vmw_mob_calculate_pt_pages(data_pages);
448 ret = ttm_bo_create(&dev_priv->bdev, mob->num_pages * PAGE_SIZE,
vmwgfx_page_dirty.c 240 pgoff_t num_pages = vbo->base.num_pages; local in function:vmw_bo_dirty_add
253 size = sizeof(*dirty) + BITS_TO_LONGS(num_pages) * sizeof(long);
268 dirty->bitmap_size = num_pages;
272 if (num_pages < PAGE_SIZE / sizeof(pte_t)) {
281 wp_shared_mapping_range(mapping, offset, num_pages);
282 clean_record_shared_mapping_range(mapping, offset, num_pages,
421 if (unlikely(page_offset >= bo->num_pages)) {
464 if (page_offset >= bo->num_pages ||
vmwgfx_bo.c 231 place.lpfn = bo->num_pages;
252 bo->mem.start < bo->num_pages &&
403 ret = ttm_bo_kmap(bo, 0, bo->num_pages, &vbo->map);
439 size_t num_pages = PAGE_ALIGN(size) >> PAGE_SHIFT; local in function:vmw_bo_acc_size
440 size_t page_array_size = ttm_round_pot(num_pages * sizeof(void *));
454 ttm_round_pot(num_pages * sizeof(dma_addr_t));
vmwgfx_drv.h 319 unsigned long num_pages; member in struct:vmw_sg_table
331 * @num_pages: Number of pages total.
343 unsigned long num_pages; member in struct:vmw_piter
714 unsigned long num_pages,
  /src/sys/external/bsd/drm2/dist/drm/
drm_memory.c 68 unsigned long i, num_pages = local in function:agp_remap
94 /* note: use vmalloc() because num_pages could be large... */
95 page_map = vmalloc(array_size(num_pages, sizeof(struct page *)));
100 for (i = 0; i < num_pages; ++i)
102 addr = vmap(page_map, num_pages, VM_IOREMAP, PAGE_AGP);
  /src/sys/external/bsd/drm2/dist/drm/xen/
xen_drm_front_gem.c 33 size_t num_pages; member in struct:xen_gem_object
52 xen_obj->num_pages = DIV_ROUND_UP(buf_size, PAGE_SIZE);
53 xen_obj->pages = kvmalloc_array(xen_obj->num_pages,
107 ret = alloc_xenballooned_pages(xen_obj->num_pages,
111 xen_obj->num_pages, ret);
123 xen_obj->num_pages = DIV_ROUND_UP(size, PAGE_SIZE);
160 free_xenballooned_pages(xen_obj->num_pages,
187 return drm_prime_pages_to_sg(xen_obj->pages, xen_obj->num_pages);
212 NULL, xen_obj->num_pages);
258 ret = vm_map_pages(vma, xen_obj->pages, xen_obj->num_pages);
    [all...]
  /src/sys/external/bsd/drm2/dist/include/drm/ttm/
ttm_tt.h 96 * @num_pages: Number of pages in the page array.
112 unsigned long num_pages; member in struct:ttm_tt
ttm_bo_api.h 101 * @num_pages: Actual size of memory region in pages.
114 unsigned long num_pages; member in struct:ttm_mem_reg
150 * @num_pages: Actual number of pages.
195 unsigned long num_pages; member in struct:ttm_buffer_object
663 * @num_pages: Number of pages to map.
675 unsigned long num_pages, struct ttm_bo_kmap_obj *map);
  /src/sys/external/bsd/drm2/dist/drm/i915/gem/
i915_gem_userptr.c 442 struct page **pvec, unsigned long num_pages)
454 ret = __sg_alloc_table_from_pages(st, pvec, num_pages,
455 0, num_pages << PAGE_SHIFT,
603 const unsigned long num_pages = obj->base.size >> PAGE_SHIFT; local in function:i915_gem_userptr_get_pages
648 pvec = kvmalloc_array(num_pages, sizeof(struct page *),
654 num_pages,
663 } else if (pinned < num_pages) {
667 pages = __i915_gem_userptr_alloc_pages(obj, pvec, num_pages);
  /src/sys/external/bsd/drm2/dist/drm/vgem/
vgem_drv.c 83 loff_t num_pages; local in function:vgem_gem_fault
87 num_pages = DIV_ROUND_UP(obj->base.size, PAGE_SIZE);
89 if (page_offset >= num_pages)
  /src/sys/external/bsd/drm2/dist/drm/via/
via_dmablit.h 51 unsigned long num_pages; member in struct:_drm_via_sg_info
  /src/sys/arch/arm/acpi/
acpi_machdep.c 107 const uint64_t num_pages = be64dec(&map[3]); local in function:acpi_md_pmapflags
110 if (pa >= phys_start && pa < phys_start + (num_pages * EFI_PAGE_SIZE)) {
  /src/sys/external/bsd/drm2/dist/drm/radeon/
radeon_ttm.c 226 unsigned num_pages; local in function:radeon_move_blit
263 num_pages = new_mem->num_pages * (PAGE_SIZE / RADEON_GPU_PAGE_SIZE);
264 fence = radeon_copy(rdev, old_start, new_start, num_pages, bo->base.resv);
415 atomic64_add((u64)bo->num_pages << PAGE_SHIFT, &rdev->num_bytes_moved);
426 mem->bus.size = mem->num_pages << PAGE_SHIFT;
542 unsigned long end = gtt->userptr + ttm->num_pages * PAGE_SIZE;
567 .iov_len = ttm->num_pages << PAGE_SHIFT,
573 .uio_resid = ttm->num_pages << PAGE_SHIFT,
583 ttm->num_pages << PAGE_SHIFT
620 unsigned num_pages = ttm->num_pages - pinned; local in function:radeon_ttm_tt_pin_userptr
    [all...]
  /src/sys/external/bsd/drm2/dist/drm/ttm/
ttm_page_alloc.c 1050 ttm_put_pages(ttm->pages, ttm->num_pages, ttm->page_flags,
1064 if (ttm_check_under_lowerlimit(mem_glob, ttm->num_pages, ctx))
1067 ret = ttm_get_pages(ttm->pages, ttm->num_pages, ttm->page_flags,
1074 for (i = 0; i < ttm->num_pages; ++i) {
1098 ttm_pool_unpopulate_helper(ttm, ttm->num_pages);
1112 for (i = 0; i < tt->ttm.num_pages; ++i) {
1114 size_t num_pages = 1; local in function:ttm_populate_and_map_pages
1116 for (j = i + 1; j < tt->ttm.num_pages; ++j) {
1120 ++num_pages;
1124 0, num_pages * PAGE_SIZE
1151 size_t num_pages = 1; local in function:ttm_unmap_and_unpopulate_pages
    [all...]
ttm_bo.c 127 bo, bo->mem.num_pages, bo->mem.size >> 10,
416 ctx->bytes_moved += bo->num_pages << PAGE_SHIFT;
1194 mem.num_pages = bo->num_pages;
1195 mem.size = mem.num_pages << PAGE_SHIFT;
1224 (heap->lpfn != 0 && (mem->start + mem->num_pages) > heap->lpfn)))
1305 unsigned long num_pages; local in function:ttm_bo_init_reserved
1323 num_pages = (size + PAGE_SIZE - 1) >> PAGE_SHIFT;
1324 if (num_pages == 0) {
1343 bo->num_pages = num_pages
    [all...]
ttm_page_alloc_dma.c 360 unsigned num_pages; local in function:ttm_dma_page_put
364 num_pages = pool->size / PAGE_SIZE;
365 if (ttm_set_pages_wb(page, num_pages))
367 pool->dev_name, num_pages);
895 unsigned long num_pages = ttm->num_pages; local in function:ttm_dma_populate
905 if (ttm_check_under_lowerlimit(mem_glob, num_pages, ctx))
926 while (num_pages >= HPAGE_PMD_NR) {
948 num_pages -= HPAGE_PMD_NR;
963 while (num_pages) {
    [all...]
  /src/sys/external/bsd/drm2/dist/drm/i915/
i915_gpu_error.c 286 if (dst->page_count >= dst->num_pages)
984 unsigned long num_pages; local in function:i915_vma_coredump_create
993 num_pages = min_t(u64, vma->size, vma->obj->base.size) >> PAGE_SHIFT;
994 num_pages = DIV_ROUND_UP(10 * num_pages, 8); /* worstcase zlib growth */
995 dst = kmalloc(sizeof(*dst) + num_pages * sizeof(u32 *), ALLOW_FAIL);
1010 dst->num_pages = num_pages;
i915_gpu_error.h 44 int num_pages; member in struct:i915_vma_coredump
  /src/sys/dev/pci/
pvscsi.h 124 uint32_t num_pages; member in struct:pvscsi_cmd_desc_setup_msg_ring
  /src/sys/external/bsd/drm2/dist/drm/amd/amdgpu/
amdgpu_ttm.c 71 struct ttm_mem_reg *mem, unsigned num_pages,
458 new_mem->num_pages << PAGE_SHIFT,
614 if (nodes->size != mem->num_pages)
705 atomic64_add((u64)bo->num_pages << PAGE_SHIFT, &adev->num_bytes_moved);
722 mem->bus.size = mem->num_pages << PAGE_SHIFT;
743 (mm_node->size == mem->num_pages))
867 range->pfns = kvmalloc_array(ttm->num_pages, sizeof(*range->pfns),
904 for (i = 0; i < ttm->num_pages; i++) {
947 gtt->userptr, ttm->num_pages);
979 for (i = 0; i < ttm->num_pages; ++i
1614 unsigned long num_pages = bo->mem.num_pages; local in function:amdgpu_ttm_bo_eviction_valuable
2322 unsigned long num_pages; local in function:amdgpu_fill_buffer
    [all...]
  /src/sys/external/bsd/vchiq/dist/interface/vchiq_arm/
vchiq_arm.c 1477 int num_pages; local in function:dump_phys_mem
1495 num_pages = (offset + num_bytes + PAGE_SIZE - 1) / PAGE_SIZE;
1497 pages = kmalloc(sizeof(struct page *) * num_pages, GFP_KERNEL);
1501 num_pages);
1509 num_pages, /* len */
1545 for (page_idx = 0; page_idx < num_pages; page_idx++)

Completed in 26 milliseconds