HomeSort by: relevance | last modified time | path
    Searched defs:kva (Results 1 - 25 of 61) sorted by relevancy

1 2 3

  /src/sys/uvm/
uvm_io.c 58 vaddr_t baseva, endva, pageoffset, kva; local in function:uvm_io
104 error = uvm_map_extract(map, baseva, chunksz, kernel_map, &kva,
126 error = uiomove((void *) (kva + pageoffset), sz, uio);
135 uvm_unmap_remove(kernel_map, kva, kva + chunksz, &dead_entries,
uvm_pager.c 77 * the pager map: provides KVA for I/O
161 * this is an abuse of pmap_direct_process(), since the kva is being grabbed
166 uvm_pagermapdirect(void *kva, size_t sz, void *cookie)
170 *(vaddr_t *)cookie = (vaddr_t)kva;
176 * uvm_pagermapin: map pages into KVA (pager_map) for I/O that needs mappings
186 vaddr_t kva; local in function:uvm_pagermapin
205 PAGE_SIZE, uvm_pagermapdirect, &kva);
207 UVMHIST_LOG(maphist, "<- done, direct (KVA=%#jx)", kva,0,0,0);
208 return kva;
    [all...]
uvm_km.c 128 * Main arena controlling the kernel KVA used by other arenas.
368 kmem_va_arena = vmem_init(&kmem_va_arena_store, "kva",
448 * uvm_km_pgremove: remove pages from a kernel uvm_object and KVA.
606 * => we return KVA of memory allocated
612 vaddr_t kva, loopva; local in function:uvm_km_alloc
632 kva = vm_map_min(map); /* hint */
643 if (__predict_false(uvm_map(map, &kva, size, obj, UVM_UNKNOWN_OFFSET,
657 UVMHIST_LOG(maphist,"<- done valloc (kva=%#jx)", kva,0,0,0);
658 return(kva);
    [all...]
  /src/sys/external/bsd/drm2/dist/drm/amd/amdgpu/
amdgpu_ih.c 72 void *kva; local in function:amdgpu_ih_ring_init
87 &kva, BUS_DMA_WAITOK|BUS_DMA_COHERENT);
93 r = -bus_dmamap_load(adev->ddev->dmat, ih->ring_map, kva, size,
96 fail3: __unused bus_dmamem_unmap(adev->ddev->dmat, kva, size);
99 ih->ring = kva;
164 void *kva = __UNVOLATILE(ih->ring); local in function:amdgpu_ih_ring_fini
166 bus_dmamem_unmap(adev->ddev->dmat, kva, size);
  /src/sys/arch/hp300/hp300/
bus_space.c 54 vaddr_t kva; local in function:bus_space_map
78 &kva);
85 physaccess((void *)kva, (void *)bpa, size, PG_RW|PG_CI);
90 *bshp = (bus_space_handle_t)(kva + offset);
119 vaddr_t kva; local in function:bus_space_unmap
134 kva = m68k_trunc_page(bsh);
147 physunaccess((void *)kva, size);
152 if (extent_free(extio_ex, kva, size,
154 printf("%s: kva 0x%lx size 0x%lx: "
  /src/sys/arch/aarch64/aarch64/
efi_machdep.c 76 /* even if TBI is disabled, AARCH64_ADDRTOP_TAG means KVA */
77 bool kva = (va & AARCH64_ADDRTOP_TAG) != 0; local in function:cpu_efirt_map_range
78 if (kva) {
93 if (kva) {
102 if (kva)
  /src/sys/arch/m68k/m68k/
vm_machdep.c 153 vaddr_t kva; /* Kernel VA (new to) */ local in function:vmapbuf
163 kva = uvm_km_alloc(phys_map, len, 0, UVM_KMF_VAONLY | UVM_KMF_WAITVA);
164 bp->b_data = (void *)(kva + off);
172 pmap_enter(kpmap, kva, pa, VM_PROT_READ | VM_PROT_WRITE,
175 pmap_kenter_pa(kva, pa, VM_PROT_READ | VM_PROT_WRITE, 0);
178 kva += PAGE_SIZE;
192 vaddr_t kva; local in function:vunmapbuf
198 kva = m68k_trunc_page(bp->b_data);
199 off = (vaddr_t)bp->b_data - kva;
203 pmap_remove(vm_map_pmap(phys_map), kva, kva + len)
    [all...]
  /src/sys/arch/hppa/hppa/
vm_machdep.c 208 vaddr_t uva, kva; local in function:vmapbuf
224 kva = uvm_km_alloc(phys_map, len, 0, UVM_KMF_VAONLY | UVM_KMF_WAITVA);
225 bp->b_data = (void *)(kva + off);
230 pmap_enter(kpmap, kva, pa,
233 kva += PAGE_SIZE;
247 vaddr_t kva; local in function:vunmapbuf
254 kva = trunc_page((vaddr_t)bp->b_data);
255 off = (vaddr_t)bp->b_data - kva;
258 pmap_remove(pmap, kva, kva + len)
    [all...]
  /src/sys/arch/sparc/sparc/
vm_machdep.c 85 vaddr_t kva; /* Kernel VA (new to) */ local in function:vmapbuf
100 kva = uvm_km_alloc(kernel_map, len, 0, UVM_KMF_VAONLY | UVM_KMF_WAITVA);
101 bp->b_data = (void *)(kva + off);
117 pmap_enter(kpmap, kva, pa,
120 kva += PAGE_SIZE;
134 vaddr_t kva; local in function:vunmapbuf
140 kva = trunc_page((vaddr_t)bp->b_data);
141 off = (vaddr_t)bp->b_data - kva;
143 pmap_remove(vm_map_pmap(kernel_map), kva, kva + len)
    [all...]
  /src/sys/arch/riscv/riscv/
vm_machdep.c 215 vaddr_t kva; /* Kernel VA (new to) */ local in function:vmapbuf
224 kva = uvm_km_alloc(phys_map, len, atop(uva) & uvmexp.colormask,
226 KASSERT((atop(kva ^ uva) & uvmexp.colormask) == 0);
228 bp->b_data = (void *)(kva + off);
234 pmap_kenter_pa(kva, pa, VM_PROT_READ | VM_PROT_WRITE,
237 kva += PAGE_SIZE;
251 vaddr_t kva; local in function:vunmapbuf
255 kva = trunc_page((vaddr_t)bp->b_data);
256 len = round_page((vaddr_t)bp->b_data - kva + len);
257 pmap_kremove(kva, len)
    [all...]
  /src/sys/arch/sparc64/sparc64/
vm_machdep.c 84 vaddr_t kva; /* Kernel VA (new to) */ local in function:vmapbuf
95 kva = uvm_km_alloc(kernel_map, len, 0, UVM_KMF_VAONLY | UVM_KMF_WAITVA);
96 bp->b_data = (void *)(kva + off);
103 pmap_kenter_pa(kva, pa, VM_PROT_READ | VM_PROT_WRITE, 0);
106 kva += PAGE_SIZE;
120 vaddr_t kva; local in function:vunmapbuf
126 kva = trunc_page((vaddr_t)bp->b_data);
127 off = (vaddr_t)bp->b_data - kva;
129 pmap_kremove(kva, len);
130 uvm_km_free(kernel_map, kva, len, UVM_KMF_VAONLY)
    [all...]
  /src/sys/dev/isa/
isadmavar.h 89 void *kva; member in struct:isa_mem
isadma.c 683 _isa_dmamem_unmap(struct isa_dma_state *ids, int chan, void *kva, size_t size)
691 bus_dmamem_unmap(ids->ids_dmat, kva, size);
729 void *kva; local in function:_isa_malloc
737 if (_isa_dmamem_map(ids, chan, addr, size, &kva, bflags)) {
743 _isa_dmamem_unmap(ids, chan, kva, size);
751 m->kva = kva;
754 return (void *)kva;
761 void *kva = (void *)addr; local in function:_isa_free
763 for(mp = &isa_mem_head; *mp && (*mp)->kva != kva
    [all...]
  /src/sys/kern/
subr_physmap.c 268 * If there is currently a non-direct mapped KVA region allocated,
318 * If we can't direct map it, we have to allocate some KVA
339 * Return the starting KVA (including offset into the page) and
352 * If there is currently a non-direct mapped KVA region allocated,
370 * sparse files). But since we are using physmaps, there is no kva to use with
371 * memset so we need a helper to obtain a kva and memset the desired memory.
380 vaddr_t kva; local in function:physmap_zero
381 size_t seglen = physmap_map(cookie, &kva);
385 memset((void *)kva, 0, seglen);
  /src/sys/arch/mips/mips/
vm_machdep.c 294 vaddr_t kva; /* Kernel VA (new to) */ local in function:vmapbuf
303 kva = uvm_km_alloc(phys_map, len, atop(uva) & uvmexp.colormask,
305 KASSERT((atop(kva ^ uva) & uvmexp.colormask) == 0);
307 bp->b_data = (void *)(kva + off);
313 pmap_kenter_pa(kva, pa, VM_PROT_READ | VM_PROT_WRITE,
316 kva += PAGE_SIZE;
330 vaddr_t kva; local in function:vunmapbuf
335 kva = mips_trunc_page(bp->b_data);
336 len = mips_round_page((vaddr_t)bp->b_data - kva + len);
337 pmap_kremove(kva, len)
    [all...]
  /src/sys/arch/amiga/dev/
zbus.c 71 vaddr_t ZBUSADDR; /* kva of Zorro bus I/O pages */
410 * Zorro devices) to have enough kva-space available, so there is no extra
418 vaddr_t kva; local in function:zbusmap
427 kva = nextkva;
431 physaccess((void *)kva, (void *)pa, size, PG_RW|PG_CI);
432 return((void *)kva);
  /src/sys/arch/sun3/sun3/
dvma.c 149 dvma_kvtopa(void *kva, int bustype)
153 addr = (u_long)kva;
177 dvma_mapin(void *kva, int len, int canwait /* ignored */)
186 seg_kva = (vaddr_t)kva;
289 vaddr_t kva; local in function:_bus_dmamap_load
305 kva = (vaddr_t)buf;
306 off = kva & PGOFSET;
341 rv = pmap_extract(pmap, kva, &pa);
348 kva += PAGE_SIZE;
  /src/sys/dev/marvell/
mvxpbm.c 172 char *kva, *ptr, *ptr_next, *ptr_data; local in function:mvxpbm_alloc_buffer
251 (void **)&kva, BUS_DMA_NOWAIT)) {
264 kva, sc->sc_buf_size, NULL, BUS_DMA_NOWAIT)) {
269 sc->sc_buf = (void *)kva;
271 bm_buf_end = (void *)(kva + sc->sc_buf_size);
305 bus_dmamem_unmap(sc->sc_dmat, kva, sc->sc_buf_size);
  /src/sys/arch/sun3/sun3x/
dvma.c 137 dvma_kvtopa(void *kva, int bustype)
141 addr = (u_long)kva;
167 vaddr_t kva; local in function:dvma_mapin
174 kva = (vaddr_t)kmem_va;
175 KASSERT(kva >= VM_MIN_KERNEL_ADDRESS);
180 off = kva & PGOFSET;
181 kva -= off; /* Truncate starting address to nearest page. */
210 for (; npf--; kva += PAGE_SIZE, tva += PAGE_SIZE) {
217 rv = pmap_extract(pmap_kernel(), kva, &pa);
233 * Remove double map of `va' in DVMA space at `kva'
242 u_long kva; local in function:dvma_mapout
302 vaddr_t kva; local in function:_bus_dmamap_load
    [all...]
  /src/sys/arch/mips/cavium/
octeon_cpunode.c 266 const vaddr_t kva = PMAP_MAP_POOLPAGE(VM_PAGE_TO_PHYS(pg)); local in function:cpu_cpunode_attach_common
267 KASSERT(kva != 0);
268 ci->ci_nmi_stack = (void *)(kva + PAGE_SIZE - sizeof(struct kernframe));
  /src/sys/dev/tc/
bba.c 75 void *kva; member in struct:bba_mem
272 void *kva; local in function:bba_allocm
289 &kva, BUS_DMA_WAITOK | BUS_DMA_COHERENT)) {
298 m->kva = kva;
302 return (void *)kva;
306 bus_dmamem_unmap(sc->sc_dmat, kva, size);
319 void *kva; local in function:bba_freem
322 kva = (void *)addr;
323 for (mp = &sc->sc_mem_head; *mp && (*mp)->kva != kva
    [all...]
px.c 184 char *kva, *bva; local in function:px_init
187 kva = (void *)si->si_slotbase;
212 si->si_vdac = (uint32_t *)(kva + PX_VDAC_OFFSET);
213 si->si_vdac_reset = (uint32_t *)(kva + PX_VDAC_RESET_OFFSET);
214 si->si_stic = (volatile struct stic_regs *)(kva + PX_STIC_OFFSET);
215 si->si_stamp = (uint32_t *)(kva + PX_STAMP_OFFSET);
pxg.c 187 char *kva; local in function:pxg_init
189 kva = (void *)si->si_slotbase;
191 si->si_vdac = (uint32_t *)(kva + PXG_VDAC_OFFSET);
192 si->si_vdac_reset = (uint32_t *)(kva + PXG_VDAC_RESET_OFFSET);
193 si->si_stic = (volatile struct stic_regs *)(kva + PXG_STIC_OFFSET);
194 si->si_stamp = (uint32_t *)(kva + PXG_STAMP_OFFSET);
195 si->si_buf = (uint32_t *)(kva + PXG_SRAM_OFFSET);
206 slot = (volatile uint32_t *)kva;
  /src/sys/external/bsd/drm2/dist/drm/i915/gt/
intel_gtt.c 391 void *kva; local in function:setup_scratch_page
420 size, &kva, BUS_DMA_NOWAIT|BUS_DMA_NOCACHE);
423 memset(kva, 0, size);
426 bus_dmamem_unmap(vm->dmat, kva, size);
  /src/sys/external/bsd/drm2/dist/drm/nouveau/nvkm/subdev/instmem/
nouveau_nvkm_subdev_instmem_gk20a.c 242 void *kva; local in function:gk20a_instobj_acquire_iommu
244 &kva, BUS_DMA_WAITOK|BUS_DMA_PREFETCHABLE))
247 node->base.vaddr = kva;

Completed in 22 milliseconds

1 2 3