/src/sys/external/bsd/drm2/dist/drm/nouveau/nvkm/subdev/bios/ |
nouveau_nvkm_subdev_bios_vmap.c | 31 #include <subdev/bios/vmap.h> 37 u32 vmap = 0; local in function:nvbios_vmap_table 41 vmap = nvbios_rd32(bios, bit_P.offset + 0x20); 42 if (vmap) { 43 *ver = nvbios_rd08(bios, vmap + 0); 47 *hdr = nvbios_rd08(bios, vmap + 1); 48 *cnt = nvbios_rd08(bios, vmap + 3); 49 *len = nvbios_rd08(bios, vmap + 2); 50 return vmap; 65 u32 vmap = nvbios_vmap_table(bios, ver, hdr, cnt, len) local in function:nvbios_vmap_parse 89 u32 vmap = nvbios_vmap_table(bios, ver, &hdr, &cnt, len); local in function:nvbios_vmap_entry 101 u32 vmap = nvbios_vmap_entry(bios, idx, ver, len); local in function:nvbios_vmap_entry_parse [all...] |
/src/sys/external/bsd/drm2/dist/drm/nouveau/nvkm/subdev/volt/ |
nouveau_nvkm_subdev_volt_base.c | 32 #include <subdev/bios/vmap.h> 92 u32 vmap; local in function:nvkm_volt_map_min 94 vmap = nvbios_vmap_entry_parse(bios, id, &ver, &len, &info); 95 if (vmap) { 114 u32 vmap; local in function:nvkm_volt_map 116 vmap = nvbios_vmap_entry_parse(bios, id, &ver, &len, &info); 117 if (vmap) { 300 struct nvbios_vmap vmap; local in function:nvkm_volt_ctor 306 if (nvbios_vmap_parse(bios, &ver, &hdr, &cnt, &len, &vmap)) { 307 volt->max0_id = vmap.max0 [all...] |
/src/sys/dev/mvme/ |
mvmebus.c | 473 struct mvmebus_dmamap *vmap; local in function:mvmebus_dmamap_create 498 if ((vmap = malloc(sizeof(*vmap), M_DMAMAP, 506 free(vmap, M_DMAMAP); 510 vmap->vm_am = am; 511 vmap->vm_datasize = datasize; 512 vmap->vm_swap = swap; 513 vmap->vm_slave = vr; 515 (*mapp)->_dm_cookie = vmap; 532 struct mvmebus_dmamap *vmap = map->_dm_cookie local in function:mvmebus_dmamap_load_common [all...] |
/src/sys/external/bsd/drm2/include/linux/ |
vmalloc.h | 85 * vmap(pages, npages, flags, prot) 95 vmap(struct page **pages, unsigned npages, unsigned long flags, function in typeref:typename:void * 124 * to vmap with the same npages parameter.
|
dma-buf.h | 70 void * (*vmap)(struct dma_buf *); member in struct:dma_buf_ops
|
/src/sys/external/bsd/drm2/dist/drm/i915/gem/selftests/ |
mock_dmabuf.c | 94 .vmap = mock_dmabuf_vmap,
|
/src/sys/external/bsd/drm2/dist/drm/virtio/ |
virtgpu_object.c | 92 .vmap = drm_gem_shmem_vmap,
|
/src/sys/external/bsd/drm2/dist/drm/ |
drm_memory.c | 102 addr = vmap(page_map, num_pages, VM_IOREMAP, PAGE_AGP);
|
drm_gem.c | 1378 if (obj->funcs && obj->funcs->vmap) 1379 vaddr = obj->funcs->vmap(obj);
|
drm_prime.c | 800 * drm_gem_dmabuf_vmap - dma_buf vmap implementation for GEM 803 * Sets up a kernel virtual mapping. This can be used as the &dma_buf_ops.vmap 804 * callback. Calls into &drm_gem_object_funcs.vmap for device specific handling. 967 .vmap = drm_gem_dmabuf_vmap,
|
/src/sys/external/bsd/drm2/dist/include/drm/ |
drm_gem.h | 145 * @vmap: 152 void *(*vmap)(struct drm_gem_object *obj); member in struct:drm_gem_object_funcs 157 * Releases the the address previously returned by @vmap. Used by the
|
/src/sys/external/bsd/drm2/dist/drm/xen/ |
xen_drm_front_gem.c | 288 return vmap(xen_obj->pages, xen_obj->num_pages,
|
/src/sys/external/bsd/drm2/dist/drm/amd/amdgpu/ |
amdgpu_dma_buf.c | 48 * amdgpu_gem_prime_vmap - &dma_buf_ops.vmap implementation 368 .vmap = drm_gem_dmabuf_vmap,
|
/src/sys/external/bsd/drm2/dist/drm/qxl/ |
qxl_object.c | 97 .vmap = qxl_gem_prime_vmap,
|
/src/sys/external/bsd/drm2/dist/drm/i915/gem/ |
i915_gem_dmabuf.c | 199 .vmap = i915_gem_dmabuf_vmap,
|
/src/sys/arch/i386/i386/ |
machdep.c | 1783 struct vm_page *vmap; local in function:cpu_alloc_l3_page 1791 vmap = TAILQ_FIRST(&pg); 1793 if (ret != 0 || vmap == NULL) 1797 ci->ci_pae_l3_pdirpa = VM_PAGE_TO_PHYS(vmap);
|
/src/sys/external/bsd/drm2/dist/drm/nouveau/nvkm/subdev/mmu/ |
nouveau_nvkm_subdev_mmu_mem.c | 187 *pmap = vmap(mem->mem, mem->pages, VM_MAP, PAGE_KERNEL);
|
/src/sys/external/bsd/drm2/dist/drm/vgem/ |
vgem_drv.c | 400 return vmap(pages, n_pages, 0, pgprot_writecombine(PAGE_KERNEL));
|
/src/sys/external/bsd/drm2/dist/drm/ttm/ |
ttm_bo_util.c | 325 #define __ttm_kmap_atomic_prot(__page, __prot) vmap(&__page, 1, 0, __prot) 338 * otherwise falls back to vmap. The user must make sure that the 716 * We need to use vmap to get the desired page protection 721 map->virtual = vmap(ttm->pages + start_page, num_pages,
|
/src/sys/external/bsd/drm2/dist/drm/nouveau/nvkm/subdev/instmem/ |
nouveau_nvkm_subdev_instmem_gk20a.c | 249 node->base.vaddr = vmap(node->pages, size >> PAGE_SHIFT, VM_MAP, 663 nvkm_warn(&base->subdev, "instobj vmap area not empty! "
|