| /src/lib/libkvm/ |
| kvm_i386pae.c | 57 #include <i386/pte.h> 73 pt_entry_t pte; local 110 if (_kvm_pread(kd, kd->pmfd, (void *) &pte, sizeof(pte), 111 _kvm_pa2off(kd, pte_pa)) != sizeof(pte)) { 112 _kvm_syserr(kd, 0, "could not read PTE"); 117 * Validate the PTE and return the physical address. 119 if ((pte & PTE_P) == 0) { 120 _kvm_err(kd, 0, "invalid translation (invalid PTE)"); 123 *pa = (pte & PTE_FRAME) + page_off [all...] |
| kvm_alpha.c | 30 #define __KVM_ALPHA_PRIVATE /* see <machine/pte.h> */ 73 alpha_pt_entry_t pte; local 97 /* Find and read the L1 PTE. */ 100 if (_kvm_pread(kd, kd->pmfd, &pte, sizeof(pte), 101 _kvm_pa2off(kd, pteoff)) != sizeof(pte)) { 102 _kvm_syserr(kd, 0, "could not read L1 PTE"); 106 /* Find and read the L2 PTE. */ 107 if ((pte & ALPHA_PTE_VALID) == 0) { 108 _kvm_err(kd, 0, "invalid translation (invalid L1 PTE)"); [all...] |
| kvm_arm.c | 51 #include <arm/arm32/pte.h> 80 pt_entry_t pte; local 136 * locate the pte and load it 138 if (_kvm_pread(kd, kd->pmfd, (void *) &pte, sizeof(pt_entry_t), 143 switch (pte & L2_TYPE_MASK) { 145 *pa = (pte & L2_L_FRAME) | (va & L2_L_OFFSET); 148 *pa = (pte & L2_S_FRAME) | (va & L2_S_OFFSET); 151 *pa = (pte & L2_T_FRAME) | (va & L2_T_OFFSET);
|
| kvm_mips.c | 91 u_int pte; local 177 * Step 2: Locate and read the PTE. 181 if (_kvm_pread(kd, kd->pmfd, &pte, sizeof(pte), 182 _kvm_pa2off(kd, pte_pa)) != sizeof(pte)) { 183 _kvm_syserr(kd, 0, "could not read PTE"); 188 * Step 3: Validate the PTE and return the physical address. 190 if ((pte & cpu_kh->pg_v) == 0) { 191 _kvm_err(kd, 0, "invalid translation (invalid PTE)"); 194 *pa = (((pte & cpu_kh->pg_frame) >> cpu_kh->pg_shift) << PGSHIFT) [all...] |
| kvm_i386.c | 68 #include <i386/pte.h> 141 pt_entry_t pte; local 174 if (_kvm_pread(kd, kd->pmfd, (void *) &pte, sizeof(pte), 175 _kvm_pa2off(kd, pte_pa)) != sizeof(pte)) { 176 _kvm_syserr(kd, 0, "could not read PTE"); 181 * Validate the PTE and return the physical address. 183 if ((pte & PTE_P) == 0) { 184 _kvm_err(kd, 0, "invalid translation (invalid PTE)"); 187 *pa = (pte & PTE_FRAME) + page_off [all...] |
| kvm_m68k_cmn.c | 194 uint32_t ste, pte; local 232 if (_kvm_pread(kd, kd->pmfd, &pte, sizeof(pte), 233 _kvm_cmn_pa2off(kd, addr)) != sizeof(pte)) 235 addr = pte & m->pg_frame; 236 if ((pte & m->pg_v) == 0) { 256 uint32_t ste, pte; local 312 if (_kvm_pread(kd, kd->pmfd, &pte, sizeof(pte), 313 _kvm_cmn_pa2off(kd, addr)) != sizeof(pte)) [all...] |
| kvm_sun3x.c | 77 #define _kvm_pg_pa(pte, h) \ 78 ((pte) & (h)->pg_frame) 107 int idx, len, offset, pte; local 134 * read the PTE for this KVA from the page table. 138 if (kvm_read(kd, pteva, &pte, 4) != 4) { 139 _kvm_err(kd, 0, "can not read PTE!"); 142 if ((pte & s->pg_valid) == 0) { 148 pa = _kvm_pg_pa(pte, s) + offset;
|
| /src/sys/arch/sh3/sh3/ |
| kgdb_machdep.c | 94 pt_entry_t *pte; local 111 pte = __pmap_kpte_lookup(kva); 112 if (!(*pte & PG_V))
|
| locore_c.c | 176 * Prepare kernel stack PTE table. sh4_switch_resume wires these PTEs. 183 pt_entry_t *pte; local 191 pte = __pmap_kpte_lookup(vpn); 192 KDASSERT(pte && *pte != 0); 201 md_upte->data = (*pte & PG_HW_BITS) |
|
| /src/sys/arch/sun2/sun2/ |
| control.c | 37 #include <machine/pte.h> 55 u_int pte; local 57 pte = get_control_word(CONTROL_ADDR_BUILD(PGMAP_BASE, va)); 58 if (pte & PG_VALID) { 67 * readable bit). See pte.h for more about this 70 pte &= ~(0x56000000); 75 pte ^= (PG_SYSTEM); 77 return (pte); 81 set_pte(vaddr_t va, u_int pte) 83 if (pte & PG_VALID) 106 set_control_word(CONTROL_ADDR_BUILD(PGMAP_BASE, va), pte); local [all...] |
| db_machdep.c | 46 #include <machine/pte.h> 66 "Prints the PTE and segmap values", "virtual-address", NULL) }, 77 * pgmap: Given addr, Print addr, segmap, pagemap, pte 109 int pte; local 113 if (sme == 0xFF) pte = 0; 114 else pte = get_pte(va); 115 db_printf("0x%08lx [%02x] 0x%08x", va, sme, pte); 117 pte_print(pte); 122 pte_print(int pte) 129 if (pte & PG_VALID) [all...] |
| obio.c | 43 #include <machine/pte.h> 174 u_int pte; local 178 pte = get_pte((vaddr_t) (h + o)); 179 if ((pte & PG_VALID) == 0) 181 pa = PG_PA(pte);
|
| /src/sys/arch/sun3/sun3/ |
| control.c | 37 #include <machine/pte.h> 59 set_pte(vaddr_t va, u_int pte) 61 set_control_word(CONTROL_ADDR_BUILD(PGMAP_BASE, va), pte); local
|
| bus_subr.c | 50 #include <machine/pte.h> 91 int off, pte; local 102 pte = PA_PGNUM(pa); 103 pte |= (bus_info[bustype].type << PG_MOD_SHIFT); 104 pte |= (PG_VALID | PG_WRITE | PG_SYSTEM | PG_NC); 111 set_pte(pgva, pte);
|
| db_machdep.c | 43 #include <machine/pte.h> 65 "Prints the PTE and segmap values", "virtual-address", NULL) }, 76 * pgmap: Given addr, Print addr, segmap, pagemap, pte 110 int pte; local 116 pte = 0; 118 pte = get_pte(va); 120 db_printf("0x%08lx [%02x] 0x%08x", va, sme, pte); 123 pte = get_pte(va); 124 db_printf("0x%08lx 0x%08x", va, pte); 127 pte_print(pte); [all...] |
| /src/sys/arch/x86/x86/ |
| kgdb_machdep.c | 74 #include <machine/pte.h> 84 pt_entry_t *pte; local 92 pte = vtopte(va); 94 pte = kvtopte(va); 95 if ((*pte & PTE_P) == 0) 97 if (*pte & PTE_PS)
|
| db_memrw.c | 38 * the page writable, do the write, then restore the PTE. 136 pt_entry_t *ppte, pte; local 148 * Get the PTE for the page. 151 pte = *ppte; 153 if ((pte & PTE_P) == 0) { 165 if (pte & PTE_PS) 180 * MULTIPROCESSOR: no shootdown required as the PTE continues to 199 * should be in CPUF_PAUSE state and will not cache the PTE
|
| /src/sys/arch/alpha/alpha/ |
| kgdb_machdep.c | 119 pt_entry_t *pte; local 127 pte = pmap_l3pte(pmap_kernel(), va, NULL); 128 if (pte == NULL || pmap_pte_v(pte) == 0)
|
| /src/sys/arch/cesfic/cesfic/ |
| autoconf.c | 50 #include <machine/pte.h> 127 pt_entry_t *pte = kvtopte(va); local 128 *pte |= PG_CI; 129 *pte &= ~PG_CCB;
|
| /src/sys/arch/ia64/stand/efi/libefi/ |
| exec.c | 79 #include <machine/pte.h> 136 pt_entry_t pte; local 190 pte = PTE_PRESENT | PTE_MA_WB | PTE_ACCESSED | PTE_DIRTY | 199 __asm __volatile("itr.d dtr[%0]=%1;;" :: "r"(0), "r"(pte)); 204 __asm __volatile("itr.i itr[%0]=%1;;" :: "r"(0), "r"(pte));
|
| /src/sys/arch/m68k/m68k/ |
| db_memrw.c | 63 * the page writable, do the write, then restore the PTE. 112 volatile pt_entry_t *pte; local 142 * prevent a cache hit on the old, read-only PTE. 149 * N.B. we use the 68851 PTE bit names here, but in 151 * a vis the 68040 PTE bits. 158 pte = kvtopte(pgva); 159 oldpte = *pte; 165 *pte = tmppte; 175 * Restore the old PTE. 177 *pte = oldpte [all...] |
| /src/sys/arch/vax/uba/ |
| uba_dma.c | 80 struct pte *pte; local 110 pte = sc->uv_uba->uba_map; 112 pte = (struct pte *)vax_map_physmem(sc->uv_addr, 113 vax_btoc(vax_btoc(sc->uv_size) * sizeof(struct pte))); 114 if (pte == 0) 120 vax_sgmap_init(t, &sc->uv_sgmap, "uba_sgmap", 0, sc->uv_size, pte, 0);
|
| /src/sys/arch/arm/s3c2xx0/ |
| s3c2xx0_space.c | 175 pt_entry_t *pte; local 197 pte = vtopte(va); 199 *pte &= ~L2_S_CACHE_MASK;
|
| /src/sys/arch/sparc/stand/common/ |
| mmu.c | 37 #include <machine/pte.h> 100 u_int pte; local 109 pte = PG_S | PG_V | PG_W | PG_NC | ((pa >> PGSHIFT) & PG_PFNUM); 110 setpte4(va, pte); 126 u_int pte; local 129 pte = getpte4(va); 130 if ((pte & PG_V) == 0) 133 *ppa = (pte & PG_PFNUM) << PGSHIFT; 159 u_int pte; local 162 snprintf(buf, sizeof(buf), "%lx pgmap@ %lx L!", va, (u_long)&pte); [all...] |
| /src/sys/arch/vax/vax/ |
| disksubr.c | 254 disk_reallymapin(struct buf *bp, struct pte *map, int reg, int flag) 258 pt_entry_t *pte; local 268 * Get a pointer to the pte pointing out the first virtual address. 272 pte = kvtopte(addr); 278 pte = &p->p_vmspace->vm_map.pmap->pm_p1br[xaddr & 281 pte = &p->p_vmspace->vm_map.pmap->pm_p0br[xaddr]; 287 pfnum = pte->pg_pfn; 290 pte++;
|