HomeSort by: relevance | last modified time | path
    Searched defs:pde (Results 1 - 21 of 21) sorted by relevancy

  /src/lib/libkvm/
kvm_i386pae.c 72 pd_entry_t pde; local in function:_kvm_kvatop_i386pae
80 * Find and read the PDE. Ignore the L3, as it is only a per-CPU
83 * to increment pdppaddr to compute the address of the PDE.
86 pde_pa = (cpu_kh->pdppaddr & PTE_FRAME) + (pl2_pi(va) * sizeof(pde));
87 if (_kvm_pread(kd, kd->pmfd, (void *)&pde, sizeof(pde),
88 _kvm_pa2off(kd, pde_pa)) != sizeof(pde)) {
89 _kvm_syserr(kd, 0, "could not read PDE");
96 if ((pde & PTE_P) == 0) {
97 _kvm_err(kd, 0, "invalid translation (invalid PDE)");
    [all...]
kvm_arm.c 79 pd_entry_t pde; local in function:_kvm_kvatop
110 if (_kvm_pread(kd, kd->pmfd, (void *) &pde, sizeof(pd_entry_t),
118 switch (pde & L1_TYPE_MASK) {
120 *pa = (pde & L1_S_FRAME) | (va & L1_S_OFFSET);
123 pte_pa = (pde & L1_C_ADDR_MASK)
127 pte_pa = (pde & L1_S_ADDR_MASK)
kvm_i386.c 140 pd_entry_t pde; local in function:_kvm_kvatop_i386
151 pde_pa = (cpu_kh->pdppaddr & PTE_FRAME) + (pl2_pi(va) * sizeof(pde));
152 if (_kvm_pread(kd, kd->pmfd, (void *)&pde, sizeof(pde),
153 _kvm_pa2off(kd, pde_pa)) != sizeof(pde)) {
154 _kvm_syserr(kd, 0, "could not read PDE");
161 if ((pde & PTE_P) == 0) {
162 _kvm_err(kd, 0, "invalid translation (invalid PDE)");
165 if ((pde & PTE_PS) != 0) {
170 *pa = (pde & PTE_LGFRAME) + page_off
    [all...]
kvm_x86_64.c 97 pd_entry_t pde; local in function:_kvm_kvatop
116 if (_kvm_pread(kd, kd->pmfd, (void *)&pde, sizeof(pde),
117 _kvm_pa2off(kd, pde_pa)) != sizeof(pde)) {
121 if ((pde & PTE_P) == 0) {
122 _kvm_err(kd, 0, "invalid translation (invalid level 4 PDE)");
129 pde_pa = (pde & PTE_FRAME) + (pl3_pi(va) * sizeof(pd_entry_t));
130 if (_kvm_pread(kd, kd->pmfd, (void *)&pde, sizeof(pde),
131 _kvm_pa2off(kd, pde_pa)) != sizeof(pde)) {
    [all...]
  /src/sys/arch/xen/x86/
xen_pmap.c 212 pd_entry_t pde; local in function:pmap_extract_ma
221 if (!pmap_pdes_valid(va, pdes, &pde, &lvl)) {
  /src/sys/arch/aarch64/include/
asan.h 121 pd_entry_t pde; local in function:kasan_md_shadow_map_page
132 pde = l0[idx];
133 if (!l0pde_valid(pde)) {
137 pa = l0pde_pa(pde);
146 pde = l1[idx];
147 if (!l1pde_valid(pde)) {
151 pa = l1pde_pa(pde);
160 pde = l2[idx];
161 if (!l2pde_valid(pde)) {
173 } else if (l2pde_is_block(pde)) {
    [all...]
  /src/sys/arch/arm/arm32/
vm_machdep.c 280 pd_entry_t *pde, oldpde, tmppde; local in function:ktext_write
297 /* Get the PDE of the current VA. */
298 if (pmap_get_pde_pte(pmap, (vaddr_t) dst, &pde, &pte) == false)
300 switch ((oldpde = *pde) & L1_TYPE_MASK) {
306 *pde = tmppde;
307 PTE_SYNC(pde);
347 *pde = oldpde;
348 PTE_SYNC(pde);
pmap.c 1612 pd_entry_t pde __diagused = *pdep;
1618 KASSERT((pde & L1_TYPE_MASK) == L1_TYPE_C);
1623 if ((pde & (L1_C_DOM_MASK|L1_TYPE_MASK))
3428 pd_entry_t pde = L1_C_PROTO | l2b->l2b_pa local in function:pmap_enter
3430 if (*pdep != pde) {
3431 l1pte_setone(pdep, pde);
3998 pd_entry_t *pdep, pde; local in function:pmap_extract_coherency
4009 pde = *pdep;
4011 if (l1pte_section_p(pde)) {
4020 if (l1pte_supersection_p(pde)) {
4703 pd_entry_t pde = L1_C_PROTO | l2b->l2b_pa | L1_C_DOM(pmap_domain(pm)); local in function:pmap_fault_fixup
6236 pd_entry_t *pdep, pde; local in function:pmap_get_pde_pte
6401 pd_entry_t pde = l1pt[l1slot]; local in function:pmap_bootstrap
7597 pd_entry_t *pde = (pd_entry_t *) l1pt; local in function:xscale_setup_minidata
    [all...]
  /src/sys/external/bsd/drm2/dist/drm/i915/gt/
gen6_ppgtt.c 20 /* Write pde (index) from the page directory @pd to the page table @pt */
22 const unsigned int pde,
28 bus_space_write_4(ppgtt->pd_bst, ppgtt->pd_bsh, pde*sizeof(gen6_pte_t),
32 ppgtt->pd_addr + pde);
95 unsigned int pde = first_entry / GEN6_PTES; local in function:gen6_ppgtt_clear_range
101 i915_pt_entry(ppgtt->base.pd, pde++);
114 * Note that the hw doesn't support removing PDE on the fly
194 unsigned int pde; local in function:gen6_flush_pd
201 gen6_for_each_pde(pt, pd, start, end, pde)
202 gen6_write_pde(ppgtt, pde, pt)
224 unsigned int pde; local in function:gen6_alloc_va_range
304 u32 pde; local in function:gen6_ppgtt_free_pd
388 unsigned int pde; local in function:pd_vma_unbind
    [all...]
gen8_ppgtt.c 23 u64 pde = addr | _PAGE_PRESENT | _PAGE_RW; local in function:gen8_pde_encode
26 pde |= PPAT_CACHED_PDE;
28 pde |= PPAT_UNCACHED;
30 return pde;
154 void **pde = pd->entry; local in function:__gen8_ppgtt_cleanup
157 if (!*pde)
160 __gen8_ppgtt_cleanup(vm, *pde, GEN8_PDES, lvl - 1);
161 } while (pde++, --count);
713 struct i915_page_directory *pde; local in function:gen8_preallocate_top_level_pdp
715 pde = alloc_pd(vm)
    [all...]
  /src/sys/arch/x86/x86/
svs.c 360 pd_entry_t *srcpde, *dstpde, pde; local in function:svs_page_add
382 pde = (srcpde[idx] & ~(PTE_PS|PTE_2MFRAME)) | pa;
387 dstpde[pidx] = pde;
392 * Normal page, just copy the PDE.
pmap.c 296 * a PTP's index is the PD index of the PDE that points to it
1847 pd_entry_t *pde; local in function:pmap_remap_largepages
1865 pde = &L2_BASE[pl2_i(kva)];
1866 *pde = pa | pmap_pg_g | PTE_PS | PTE_P;
1884 pde = &L2_BASE[pl2_i(kva)];
1885 *pde = pa | pmap_pg_g | PTE_PS | pmap_pg_nx | PTE_P;
1903 pde = &L2_BASE[pl2_i(kva)];
1904 *pde = pa | pmap_pg_g | PTE_PS | pmap_pg_nx | PTE_W | PTE_P;
2746 /* Put in recursive PDE to map the PTEs */
2754 /* Copy the kernel's top level PDE */
3876 pd_entry_t pde; local in function:pmap_pdes_valid
3902 pd_entry_t pde; local in function:pmap_extract
4287 pd_entry_t pde; local in function:pmap_remove_locked
5534 pd_entry_t pde; local in function:pmap_remove_gnt
6530 pd_entry_t pde; local in function:pmap_ept_extract
6677 pd_entry_t pde; local in function:pmap_ept_remove
6822 pd_entry_t pde; local in function:pmap_ept_write_protect
6876 pd_entry_t pde; local in function:pmap_ept_unwire
    [all...]
  /src/sys/arch/aarch64/aarch64/
db_interface.c 464 pd_entry_t pde; variable in typeref:typename:pd_entry_t
500 pde = l0[idx];
502 pr("L0[%3d]=%016"PRIx64":", idx, pde);
503 db_pte_print(pde, 0, pr);
505 if (!l0pde_valid(pde))
508 l1 = (pd_entry_t *)AARCH64_PA_TO_KVA(l0pde_pa(pde));
510 pde = l1[idx];
512 pr(" L1[%3d]=%016"PRIx64":", idx, pde);
513 db_pte_print(pde, 1, pr);
515 if (!l1pde_valid(pde) || l1pde_is_block(pde)
559 pd_entry_t pde; variable in typeref:typename:pd_entry_t
    [all...]
pmap.c 815 pd_entry_t pde; local in function:_pmap_pte_lookup_bs
828 pde = *ptep;
829 if (!l0pde_valid(pde))
833 l1 = (pd_entry_t *)AARCH64_PA_TO_KVA(l0pde_pa(pde));
836 pde = *ptep;
837 if (!l1pde_valid(pde) || l1pde_is_block(pde))
841 l2 = (pd_entry_t *)AARCH64_PA_TO_KVA(l1pde_pa(pde));
844 pde = *ptep;
845 if (!l2pde_valid(pde) || l2pde_is_block(pde)
1717 pd_entry_t pde; local in function:_pmap_get_pdp
    [all...]
  /src/sys/arch/riscv/riscv/
pmap_machdep.c 238 pd_entry_t pde = pmap_kernel()->pm_pdetab->pde_pde[i]; local in function:pmap_md_pdetab_init
241 if (pde) {
242 pmap->pm_pdetab->pde_pde[i] = pde;
  /src/sys/external/bsd/drm2/dist/drm/radeon/
radeon_vm.c 677 uint64_t pde, pt; local in function:radeon_vm_update_page_directory
687 pde = pd_addr + pt_idx * 8;
688 if (((last_pde + 8 * count) != pde) ||
698 last_pde = pde;
  /src/sys/external/bsd/drm2/dist/drm/nouveau/nvkm/subdev/mmu/
vmm.h 34 * The array is indexed by PDE, and will either point to the
35 * child page table, or indicate the PDE is marked as sparse.
37 #define NVKM_VMM_PDE_INVALID(pde) IS_ERR_OR_NULL(pde)
38 #define NVKM_VMM_PDE_SPARSED(pde) IS_ERR(pde)
40 struct nvkm_vmm_pt **pde; member in struct:nvkm_vmm_pt
67 nvkm_vmm_pde_func pde; member in struct:nvkm_vmm_desc_func
  /src/sys/arch/x86/include/
pmap_private.h 250 #define pmap_valid_entry(E) ((E) & PTE_P) /* is PDE or PTE valid? */
329 pd_entry_t *pde; local in function:kvtopte
333 pde = L2_BASE + pl2_i(va);
334 if (*pde & PTE_PS)
335 return ((pt_entry_t *)pde);
  /src/common/lib/libprop/
prop_dictionary.c 923 struct _prop_dict_entry *pde; local in function:_prop_dict_lookup
934 pde = &pd->pd_array[idx];
935 _PROP_ASSERT(pde->pde_key != NULL);
936 res = strcmp(key, pde->pde_key->pdk_key);
940 return (pde);
957 const struct _prop_dict_entry *pde; local in function:_prop_dictionary_get
966 pde = _prop_dict_lookup(pd, key, NULL);
967 if (pde != NULL) {
968 _PROP_ASSERT(pde->pde_objref != NULL);
969 po = pde->pde_objref
1025 struct _prop_dict_entry *pde; local in function:prop_dictionary_set
1165 struct _prop_dict_entry *pde; local in function:prop_dictionary_remove
    [all...]
  /src/sys/arch/hppa/hppa/
pmap.c 366 UVMHIST_LOG(maphist, "pde %#jx", pa, 0, 0, 0);
379 pmap_pde_ptp(pmap_t pm, volatile pt_entry_t *pde)
381 paddr_t pa = (paddr_t)pde;
385 (uintptr_t)pde, 0, 0);
420 pmap_pte_get(volatile pt_entry_t *pde, vaddr_t va)
423 return (pde[(va >> 12) & 0x3ff]);
427 pmap_pte_set(volatile pt_entry_t *pde, vaddr_t va, pt_entry_t pte)
433 UVMHIST_CALLARGS(maphist, "pdep %#jx va %#jx pte %#jx", (uintptr_t)pde,
437 KASSERT(pde != NULL);
438 KASSERT(((paddr_t)pde & PGOFSET) == 0)
474 volatile pt_entry_t *pde; local in function:pmap_vp_find
487 volatile pt_entry_t *pde = NULL; local in function:pmap_dump_table
565 volatile pt_entry_t *pde; local in function:pmap_resolve_alias
1134 volatile pt_entry_t *pde; local in function:pmap_init
1265 pt_entry_t *pde, *epde; local in function:pmap_destroy
1369 volatile pt_entry_t *pde; local in function:pmap_enter
1491 volatile pt_entry_t *pde = NULL; local in function:pmap_remove
1558 volatile pt_entry_t *pde = NULL; local in function:pmap_write_protect
1626 volatile pt_entry_t *pde; local in function:pmap_page_remove
1718 volatile pt_entry_t *pde; local in function:pmap_unwire
1761 volatile pt_entry_t *pde; local in function:pmap_changebit
1987 volatile pt_entry_t *pde; local in function:pmap_kenter_pa
2047 volatile pt_entry_t *pde = NULL; local in function:pmap_kremove
    [all...]
  /src/sys/external/bsd/drm2/dist/drm/amd/amdgpu/
amdgpu_vm.c 955 * @entry: PDE to free
1260 uint64_t pde, pt, flags; local in function:amdgpu_vm_update_pde
1268 pde = (entry - parent->entries) * 8;
1269 return vm->update_funcs->update(params, bo, pde, pt, 1, 0, flags);

Completed in 78 milliseconds