/src/sys/external/bsd/drm2/drm/ |
drm_vm.c | 98 const off_t page_offset = (byte_offset >> PAGE_SHIFT); local in function:drm_legacy_mmap_paddr_locked 106 (page_offset <= dev->dma->page_count)) 109 if (drm_ht_find_item(&dev->map_hash, page_offset, &hash)) 138 const off_t page_offset = (byte_offset >> PAGE_SHIFT); local in function:drm_legacy_mmap_dma_paddr 142 KASSERT(page_offset <= dev->dma->page_count); 147 return dev->dma->pagelist[page_offset];
|
/src/sys/external/bsd/drm2/dist/drm/vmwgfx/ |
vmwgfx_page_dirty.c | 404 unsigned long page_offset; local in function:vmw_bo_vm_mkwrite 420 page_offset = vmf->pgoff - drm_vma_node_start(&bo->base.vma_node); 421 if (unlikely(page_offset >= bo->num_pages)) { 427 !test_bit(page_offset, &vbo->dirty->bitmap[0])) { 430 __set_bit(page_offset, &dirty->bitmap[0]); 431 dirty->start = min(dirty->start, page_offset); 432 dirty->end = max(dirty->end, page_offset + 1); 460 unsigned long page_offset; local in function:vmw_bo_vm_fault 462 page_offset = vmf->pgoff - 464 if (page_offset >= bo->num_pages | [all...] |
/src/sys/external/bsd/drm2/dist/drm/ttm/ |
ttm_bo_vm.c | 101 unsigned long page_offset) 106 return bdev->driver->io_mem_pfn(bo, page_offset); 109 + page_offset; 189 unsigned long page_offset; local in function:ttm_bo_vm_fault_reserved 247 page_offset = ((address - vma->vm_start) >> PAGE_SHIFT) + 252 if (unlikely(page_offset >= bo->num_pages)) { 282 pfn = ttm_bo_io_mem_pfn(bo, page_offset); 284 page = ttm->pages[page_offset]; 292 page_offset; 320 if (unlikely(++page_offset >= page_last) [all...] |
/src/sys/external/bsd/drm2/dist/drm/qxl/ |
qxl_image.c | 171 unsigned int page_base, page_offset, out_offset; local in function:qxl_image_init_helper 180 page_offset = offset_in_page(out_offset); 181 size = min((int)(PAGE_SIZE - page_offset), remain); 184 k_data = ptr + page_offset;
|
qxl_object.c | 173 struct qxl_bo *bo, int page_offset) 188 return io_mapping_map_atomic_wc(map, bo->tbo.mem.bus.offset + page_offset); 191 rptr = bo->kptr + (page_offset * PAGE_SIZE); 199 rptr += page_offset * PAGE_SIZE;
|
qxl_object.h | 98 void *qxl_bo_kmap_atomic_page(struct qxl_device *qdev, struct qxl_bo *bo, int page_offset);
|
/src/sys/external/bsd/drm2/dist/drm/vgem/ |
vgem_drv.c | 84 pgoff_t page_offset; local in function:vgem_gem_fault 85 page_offset = (vaddr - vma->vm_start) >> PAGE_SHIFT; 89 if (page_offset >= num_pages) 94 get_page(obj->pages[page_offset]); 95 vmf->page = obj->pages[page_offset]; 104 page_offset);
|
/src/sys/external/bsd/drm2/dist/include/drm/ |
drm_vma_manager.h | 67 unsigned long page_offset, unsigned long size);
|
/src/sys/external/bsd/drm2/dist/drm/ |
drm_vma_manager.c | 78 * @page_offset: Offset of available memory area (page-based) 82 * manager are given as @page_offset and @size. Both are interpreted as 91 unsigned long page_offset, unsigned long size) 94 drm_mm_init(&mgr->vm_addr_space_mm, page_offset, size);
|
drm_vm.c | 356 unsigned long page_offset; local in function:drm_vm_sg_fault 366 page_offset = (offset >> PAGE_SHIFT) + (map_offset >> PAGE_SHIFT); 367 page = entry->pagelist[page_offset];
|
/src/sys/external/bsd/drm2/dist/drm/i915/gem/ |
i915_gem_mman.c | 224 pgoff_t page_offset, 233 view.partial.offset = rounddown(page_offset, chunk); 427 pgoff_t page_offset; local in function:vm_fault_gtt 440 page_offset = (ufi->entry->offset + (vaddr - ufi->entry->start)) 444 page_offset = (vmf->address - area->vm_start) >> PAGE_SHIFT; 447 trace_i915_gem_object_fault(obj, page_offset, true, write); 467 compute_partial_view(obj, page_offset, MIN_CHUNK_PAGES); 506 unsigned startpage = page_offset;
|
/src/sys/external/bsd/drm2/dist/drm/i915/ |
i915_gem.c | 475 * page_offset = offset within page 479 unsigned page_offset = offset_in_page(offset); local in function:i915_gem_gtt_pread 480 unsigned page_length = PAGE_SIZE - page_offset; 490 if (gtt_user_read(&ggtt->iomap, page_base, page_offset, 679 * page_offset = offset within page 683 unsigned int page_offset = offset_in_page(offset); local in function:i915_gem_gtt_pwrite_fast 684 unsigned int page_length = PAGE_SIZE - page_offset; 702 if (ggtt_write(&ggtt->iomap, page_base, page_offset,
|
/src/sys/dev/pci/cxgb/ |
cxgb_lro.c | 314 skb_shinfo(s->skb)->frags[nr].page_offset = 315 frag->page_offset + IPH_OFFSET + tcpiphlen;
|
/src/sys/external/bsd/drm2/include/ |
i915_trace.h | 141 "pgoff_t"/*page_offset*/, 146 pgoff_t page_offset, bool gtt, bool write) 148 TRACE4(i915,, gem__object__fault, obj, page_offset, gtt, write);
|
/src/sys/external/bsd/drm2/dist/include/drm/ttm/ |
ttm_bo_driver.h | 375 * Return the pfn for a given page_offset inside the BO. 378 * @page_offset: the offset to look up 381 unsigned long page_offset);
|
/src/sys/arch/sparc/sparc/ |
locore.s | 1247 #define PTE_OF_ADDR4_4C(addr, pte, bad, page_offset) \ 1250 be,a 1f; andn addr, page_offset, pte; \ 1253 andn addr, page_offset, pte; \ 1275 #define PTE_OF_ADDR4M(addr, pte, bad, page_offset) \ 1276 andn addr, page_offset, pte 1316 #define PTE_OF_ADDR(addr, pte, bad, page_offset, label) \ 1317 PTE_OF_ADDR4M(addr, pte, bad, page_offset) 1323 #define PTE_OF_ADDR(addr, pte, bad, page_offset,label) \ 1324 PTE_OF_ADDR4_4C(addr, pte, bad, page_offset) 1330 #define PTE_OF_ADDR(addr, pte, bad, page_offset, label) [all...] |
/src/sys/external/bsd/drm2/dist/drm/amd/amdgpu/ |
amdgpu_ttm.c | 761 unsigned long page_offset) 764 unsigned long offset = (page_offset << PAGE_SHIFT); 779 (mm->start + page_offset) << PAGE_SHIFT, /*vm_prot*/0, /*flags*/0);
|
/src/sys/external/bsd/vchiq/dist/interface/vchiq_arm/ |
vchiq_arm.c | 1521 int page_offset = offset % PAGE_SIZE; local in function:dump_phys_mem 1537 page_offset], 1538 &kmapped_virt_ptr[page_offset], 16);
|