/src/lib/libkvm/ |
kvm_i386pae.c | 72 pd_entry_t pde; local in function:_kvm_kvatop_i386pae 80 * Find and read the PDE. Ignore the L3, as it is only a per-CPU 83 * to increment pdppaddr to compute the address of the PDE. 86 pde_pa = (cpu_kh->pdppaddr & PTE_FRAME) + (pl2_pi(va) * sizeof(pde)); 87 if (_kvm_pread(kd, kd->pmfd, (void *)&pde, sizeof(pde), 88 _kvm_pa2off(kd, pde_pa)) != sizeof(pde)) { 89 _kvm_syserr(kd, 0, "could not read PDE"); 96 if ((pde & PTE_P) == 0) { 97 _kvm_err(kd, 0, "invalid translation (invalid PDE)"); [all...] |
kvm_arm.c | 79 pd_entry_t pde; local in function:_kvm_kvatop 110 if (_kvm_pread(kd, kd->pmfd, (void *) &pde, sizeof(pd_entry_t), 118 switch (pde & L1_TYPE_MASK) { 120 *pa = (pde & L1_S_FRAME) | (va & L1_S_OFFSET); 123 pte_pa = (pde & L1_C_ADDR_MASK) 127 pte_pa = (pde & L1_S_ADDR_MASK)
|
kvm_i386.c | 140 pd_entry_t pde; local in function:_kvm_kvatop_i386 151 pde_pa = (cpu_kh->pdppaddr & PTE_FRAME) + (pl2_pi(va) * sizeof(pde)); 152 if (_kvm_pread(kd, kd->pmfd, (void *)&pde, sizeof(pde), 153 _kvm_pa2off(kd, pde_pa)) != sizeof(pde)) { 154 _kvm_syserr(kd, 0, "could not read PDE"); 161 if ((pde & PTE_P) == 0) { 162 _kvm_err(kd, 0, "invalid translation (invalid PDE)"); 165 if ((pde & PTE_PS) != 0) { 170 *pa = (pde & PTE_LGFRAME) + page_off [all...] |
kvm_x86_64.c | 97 pd_entry_t pde; local in function:_kvm_kvatop 116 if (_kvm_pread(kd, kd->pmfd, (void *)&pde, sizeof(pde), 117 _kvm_pa2off(kd, pde_pa)) != sizeof(pde)) { 121 if ((pde & PTE_P) == 0) { 122 _kvm_err(kd, 0, "invalid translation (invalid level 4 PDE)"); 129 pde_pa = (pde & PTE_FRAME) + (pl3_pi(va) * sizeof(pd_entry_t)); 130 if (_kvm_pread(kd, kd->pmfd, (void *)&pde, sizeof(pde), 131 _kvm_pa2off(kd, pde_pa)) != sizeof(pde)) { [all...] |
/src/sys/arch/xen/x86/ |
xen_pmap.c | 212 pd_entry_t pde; local in function:pmap_extract_ma 221 if (!pmap_pdes_valid(va, pdes, &pde, &lvl)) {
|
/src/sys/arch/aarch64/include/ |
asan.h | 121 pd_entry_t pde; local in function:kasan_md_shadow_map_page 132 pde = l0[idx]; 133 if (!l0pde_valid(pde)) { 137 pa = l0pde_pa(pde); 146 pde = l1[idx]; 147 if (!l1pde_valid(pde)) { 151 pa = l1pde_pa(pde); 160 pde = l2[idx]; 161 if (!l2pde_valid(pde)) { 173 } else if (l2pde_is_block(pde)) { [all...] |
/src/sys/arch/arm/arm32/ |
vm_machdep.c | 280 pd_entry_t *pde, oldpde, tmppde; local in function:ktext_write 297 /* Get the PDE of the current VA. */ 298 if (pmap_get_pde_pte(pmap, (vaddr_t) dst, &pde, &pte) == false) 300 switch ((oldpde = *pde) & L1_TYPE_MASK) { 306 *pde = tmppde; 307 PTE_SYNC(pde); 347 *pde = oldpde; 348 PTE_SYNC(pde);
|
pmap.c | 1612 pd_entry_t pde __diagused = *pdep; 1618 KASSERT((pde & L1_TYPE_MASK) == L1_TYPE_C); 1623 if ((pde & (L1_C_DOM_MASK|L1_TYPE_MASK)) 3428 pd_entry_t pde = L1_C_PROTO | l2b->l2b_pa local in function:pmap_enter 3430 if (*pdep != pde) { 3431 l1pte_setone(pdep, pde); 3998 pd_entry_t *pdep, pde; local in function:pmap_extract_coherency 4009 pde = *pdep; 4011 if (l1pte_section_p(pde)) { 4020 if (l1pte_supersection_p(pde)) { 4703 pd_entry_t pde = L1_C_PROTO | l2b->l2b_pa | L1_C_DOM(pmap_domain(pm)); local in function:pmap_fault_fixup 6236 pd_entry_t *pdep, pde; local in function:pmap_get_pde_pte 6401 pd_entry_t pde = l1pt[l1slot]; local in function:pmap_bootstrap 7597 pd_entry_t *pde = (pd_entry_t *) l1pt; local in function:xscale_setup_minidata [all...] |
/src/sys/external/bsd/drm2/dist/drm/i915/gt/ |
gen6_ppgtt.c | 20 /* Write pde (index) from the page directory @pd to the page table @pt */ 22 const unsigned int pde, 28 bus_space_write_4(ppgtt->pd_bst, ppgtt->pd_bsh, pde*sizeof(gen6_pte_t), 32 ppgtt->pd_addr + pde); 95 unsigned int pde = first_entry / GEN6_PTES; local in function:gen6_ppgtt_clear_range 101 i915_pt_entry(ppgtt->base.pd, pde++); 114 * Note that the hw doesn't support removing PDE on the fly 194 unsigned int pde; local in function:gen6_flush_pd 201 gen6_for_each_pde(pt, pd, start, end, pde) 202 gen6_write_pde(ppgtt, pde, pt) 224 unsigned int pde; local in function:gen6_alloc_va_range 304 u32 pde; local in function:gen6_ppgtt_free_pd 388 unsigned int pde; local in function:pd_vma_unbind [all...] |
gen8_ppgtt.c | 23 u64 pde = addr | _PAGE_PRESENT | _PAGE_RW; local in function:gen8_pde_encode 26 pde |= PPAT_CACHED_PDE; 28 pde |= PPAT_UNCACHED; 30 return pde; 154 void **pde = pd->entry; local in function:__gen8_ppgtt_cleanup 157 if (!*pde) 160 __gen8_ppgtt_cleanup(vm, *pde, GEN8_PDES, lvl - 1); 161 } while (pde++, --count); 713 struct i915_page_directory *pde; local in function:gen8_preallocate_top_level_pdp 715 pde = alloc_pd(vm) [all...] |
/src/sys/arch/x86/x86/ |
svs.c | 360 pd_entry_t *srcpde, *dstpde, pde; local in function:svs_page_add 382 pde = (srcpde[idx] & ~(PTE_PS|PTE_2MFRAME)) | pa; 387 dstpde[pidx] = pde; 392 * Normal page, just copy the PDE.
|
pmap.c | 296 * a PTP's index is the PD index of the PDE that points to it 1847 pd_entry_t *pde; local in function:pmap_remap_largepages 1865 pde = &L2_BASE[pl2_i(kva)]; 1866 *pde = pa | pmap_pg_g | PTE_PS | PTE_P; 1884 pde = &L2_BASE[pl2_i(kva)]; 1885 *pde = pa | pmap_pg_g | PTE_PS | pmap_pg_nx | PTE_P; 1903 pde = &L2_BASE[pl2_i(kva)]; 1904 *pde = pa | pmap_pg_g | PTE_PS | pmap_pg_nx | PTE_W | PTE_P; 2746 /* Put in recursive PDE to map the PTEs */ 2754 /* Copy the kernel's top level PDE */ 3876 pd_entry_t pde; local in function:pmap_pdes_valid 3902 pd_entry_t pde; local in function:pmap_extract 4287 pd_entry_t pde; local in function:pmap_remove_locked 5534 pd_entry_t pde; local in function:pmap_remove_gnt 6530 pd_entry_t pde; local in function:pmap_ept_extract 6677 pd_entry_t pde; local in function:pmap_ept_remove 6822 pd_entry_t pde; local in function:pmap_ept_write_protect 6876 pd_entry_t pde; local in function:pmap_ept_unwire [all...] |
/src/sys/arch/aarch64/aarch64/ |
db_interface.c | 464 pd_entry_t pde; variable in typeref:typename:pd_entry_t 500 pde = l0[idx]; 502 pr("L0[%3d]=%016"PRIx64":", idx, pde); 503 db_pte_print(pde, 0, pr); 505 if (!l0pde_valid(pde)) 508 l1 = (pd_entry_t *)AARCH64_PA_TO_KVA(l0pde_pa(pde)); 510 pde = l1[idx]; 512 pr(" L1[%3d]=%016"PRIx64":", idx, pde); 513 db_pte_print(pde, 1, pr); 515 if (!l1pde_valid(pde) || l1pde_is_block(pde) 559 pd_entry_t pde; variable in typeref:typename:pd_entry_t [all...] |
pmap.c | 815 pd_entry_t pde; local in function:_pmap_pte_lookup_bs 828 pde = *ptep; 829 if (!l0pde_valid(pde)) 833 l1 = (pd_entry_t *)AARCH64_PA_TO_KVA(l0pde_pa(pde)); 836 pde = *ptep; 837 if (!l1pde_valid(pde) || l1pde_is_block(pde)) 841 l2 = (pd_entry_t *)AARCH64_PA_TO_KVA(l1pde_pa(pde)); 844 pde = *ptep; 845 if (!l2pde_valid(pde) || l2pde_is_block(pde) 1717 pd_entry_t pde; local in function:_pmap_get_pdp [all...] |
/src/sys/arch/riscv/riscv/ |
pmap_machdep.c | 238 pd_entry_t pde = pmap_kernel()->pm_pdetab->pde_pde[i]; local in function:pmap_md_pdetab_init 241 if (pde) { 242 pmap->pm_pdetab->pde_pde[i] = pde;
|
/src/sys/external/bsd/drm2/dist/drm/radeon/ |
radeon_vm.c | 677 uint64_t pde, pt; local in function:radeon_vm_update_page_directory 687 pde = pd_addr + pt_idx * 8; 688 if (((last_pde + 8 * count) != pde) || 698 last_pde = pde;
|
/src/sys/external/bsd/drm2/dist/drm/nouveau/nvkm/subdev/mmu/ |
vmm.h | 34 * The array is indexed by PDE, and will either point to the 35 * child page table, or indicate the PDE is marked as sparse. 37 #define NVKM_VMM_PDE_INVALID(pde) IS_ERR_OR_NULL(pde) 38 #define NVKM_VMM_PDE_SPARSED(pde) IS_ERR(pde) 40 struct nvkm_vmm_pt **pde; member in struct:nvkm_vmm_pt 67 nvkm_vmm_pde_func pde; member in struct:nvkm_vmm_desc_func
|
/src/sys/arch/x86/include/ |
pmap_private.h | 250 #define pmap_valid_entry(E) ((E) & PTE_P) /* is PDE or PTE valid? */ 329 pd_entry_t *pde; local in function:kvtopte 333 pde = L2_BASE + pl2_i(va); 334 if (*pde & PTE_PS) 335 return ((pt_entry_t *)pde);
|
/src/common/lib/libprop/ |
prop_dictionary.c | 923 struct _prop_dict_entry *pde; local in function:_prop_dict_lookup 934 pde = &pd->pd_array[idx]; 935 _PROP_ASSERT(pde->pde_key != NULL); 936 res = strcmp(key, pde->pde_key->pdk_key); 940 return (pde); 957 const struct _prop_dict_entry *pde; local in function:_prop_dictionary_get 966 pde = _prop_dict_lookup(pd, key, NULL); 967 if (pde != NULL) { 968 _PROP_ASSERT(pde->pde_objref != NULL); 969 po = pde->pde_objref 1025 struct _prop_dict_entry *pde; local in function:prop_dictionary_set 1165 struct _prop_dict_entry *pde; local in function:prop_dictionary_remove [all...] |
/src/sys/arch/hppa/hppa/ |
pmap.c | 366 UVMHIST_LOG(maphist, "pde %#jx", pa, 0, 0, 0); 379 pmap_pde_ptp(pmap_t pm, volatile pt_entry_t *pde) 381 paddr_t pa = (paddr_t)pde; 385 (uintptr_t)pde, 0, 0); 420 pmap_pte_get(volatile pt_entry_t *pde, vaddr_t va) 423 return (pde[(va >> 12) & 0x3ff]); 427 pmap_pte_set(volatile pt_entry_t *pde, vaddr_t va, pt_entry_t pte) 433 UVMHIST_CALLARGS(maphist, "pdep %#jx va %#jx pte %#jx", (uintptr_t)pde, 437 KASSERT(pde != NULL); 438 KASSERT(((paddr_t)pde & PGOFSET) == 0) 474 volatile pt_entry_t *pde; local in function:pmap_vp_find 487 volatile pt_entry_t *pde = NULL; local in function:pmap_dump_table 565 volatile pt_entry_t *pde; local in function:pmap_resolve_alias 1134 volatile pt_entry_t *pde; local in function:pmap_init 1265 pt_entry_t *pde, *epde; local in function:pmap_destroy 1369 volatile pt_entry_t *pde; local in function:pmap_enter 1491 volatile pt_entry_t *pde = NULL; local in function:pmap_remove 1558 volatile pt_entry_t *pde = NULL; local in function:pmap_write_protect 1626 volatile pt_entry_t *pde; local in function:pmap_page_remove 1718 volatile pt_entry_t *pde; local in function:pmap_unwire 1761 volatile pt_entry_t *pde; local in function:pmap_changebit 1987 volatile pt_entry_t *pde; local in function:pmap_kenter_pa 2047 volatile pt_entry_t *pde = NULL; local in function:pmap_kremove [all...] |
/src/sys/external/bsd/drm2/dist/drm/amd/amdgpu/ |
amdgpu_vm.c | 955 * @entry: PDE to free 1260 uint64_t pde, pt, flags; local in function:amdgpu_vm_update_pde 1268 pde = (entry - parent->entries) * 8; 1269 return vm->update_funcs->update(params, bo, pde, pt, 1, 0, flags);
|