/src/sys/arch/hppa/hppa/ |
hpt.h | 117 struct pv_entry *hpt_next; /* Pointer to associated hash list */
|
pmap.c | 176 static inline struct pv_entry *pmap_pv_alloc(void); 177 static inline void pmap_pv_free(struct pv_entry *); 178 static inline void pmap_pv_enter(struct vm_page *, struct pv_entry *, pmap_t, 180 static inline struct pv_entry *pmap_pv_remove(struct vm_page *, pmap_t, 527 struct pv_entry *pve; 551 struct pv_entry *pve, *npve, **pvp; 689 * This allocates and returns a new struct pv_entry. 691 static inline struct pv_entry * 694 struct pv_entry *pv; 702 pmap_pv_free(struct pv_entry *pv [all...] |
/src/sys/arch/x86/include/ |
pmap_pv.h | 55 * pv_entry: plug pv_pte into lists. 32 bytes on i386, 64 on amd64. 58 struct pv_entry { struct 60 LIST_ENTRY(pv_entry) pve_list; /* on pmap_page::pp_pvlist */ 81 LIST_HEAD(, pv_entry) pvlist;
|
pmap_private.h | 182 struct pv_entry *pm_pve; /* spare pv_entry */
|
/src/sys/arch/vax/include/ |
pmap.h | 119 struct pv_entry { struct 120 struct pv_entry *pv_next; /* next pv_entry */ 126 extern struct pv_entry *pv_table; 192 void pmap_page_protect_long(struct pv_entry *, vm_prot_t); 195 static __inline struct pv_entry * 204 const struct pv_entry * const pv = pmap_pg_to_pv(pg); 212 struct pv_entry * const pv = pmap_pg_to_pv(pg);
|
/src/sys/arch/ia64/include/ |
pmap.h | 108 struct pv_entry; 138 typedef struct pv_entry { struct 140 TAILQ_ENTRY(pv_entry) pv_list; 179 TAILQ_HEAD(,pv_entry) pv_list;
|
/src/sys/arch/usermode/usermode/ |
pmap.c | 49 struct pv_entry { struct 50 struct pv_entry *pv_next; 65 #define PMAP_L2_NENTRY (PMAP_L2_SIZE / sizeof(struct pv_entry *)) 68 struct pv_entry *pm_l2[PMAP_L2_NENTRY]; 80 * pv_table is list of pv_entry structs completely spanning the total memory. 82 * with pv_entry records for each usage in all the pmaps. 88 static struct pv_entry **kernel_pm_entries; 89 static struct pv_entry *pv_table; /* physical pages info (direct mapped) */ 90 static struct pv_entry **tlb; /* current tlb mappings (direct mapped) */ 112 static void pmap_page_activate(struct pv_entry *pv) [all...] |
/src/sys/arch/sh3/include/ |
pmap.h | 93 struct pv_entry; 95 SLIST_HEAD(, pv_entry) pvh_head;
|
/src/sys/uvm/pmap/ |
vmpagemd.h | 59 typedef struct pv_entry { struct 60 struct pv_entry *pv_next; 93 struct pv_entry mdpg_first; /* pv_entry first */
|
/src/sys/arch/hppa/include/ |
pmap.h | 83 struct pv_entry { /* locked by its list's pvh_lock */ struct 84 struct pv_entry *pv_next; 214 struct pv_entry; 217 struct pv_entry *pvh_list; /* head of list */
|
/src/sys/arch/powerpc/include/ibm4xx/ |
vmparam.h | 98 struct pv_entry *pvent;
|
/src/sys/arch/alpha/include/ |
pmap.h | 139 LIST_HEAD(pmap_pvlist, pv_entry); 177 typedef struct pv_entry { struct 178 struct pv_entry *pv_next; /* next pv_entry on page list */ 179 LIST_ENTRY(pv_entry) pv_link; /* link on owning pmap's list */ 361 uintptr_t pvh_listx; /* pv_entry list + attrs */ 386 ((struct pv_entry *)((pg)->mdpage.pvh_listx & ~3UL))
|
/src/sys/arch/sparc64/include/ |
pmap.h | 256 typedef struct pv_entry { struct 257 struct pv_entry *pv_next; /* next pv_entry */ 261 /* PV flags encoded in the low bits of the VA of the first pv_entry */ 264 struct pv_entry mdpg_pvh;
|
/src/sys/arch/m68k/include/ |
pmap_motorola.h | 197 * mappings of that page. An entry is a pv_entry, the list is pv_table. 199 struct pv_entry { struct 200 struct pv_entry *pv_next; /* next pv_entry */
|
/src/sys/arch/aarch64/include/ |
pmap.h | 273 LIST_HEAD(, pv_entry) pm_pvlist; /* all pv of this process */ 297 struct pv_entry { struct 298 struct pv_entry *pv_next; 302 LIST_ENTRY(pv_entry) pv_proc; /* belonging to the process */ 307 struct pv_entry pp_pv;
|
/src/sys/arch/vax/vax/ |
pmap.c | 92 struct pv_entry *pv_table; /* array of entries, one per LOGICAL page */ 177 struct pv_entry *get_pventry(void); 178 void free_pventry(struct pv_entry *); 330 pv_table = (struct pv_entry *)(scratch + SCRATCHPAGES * VAX_NBPG); 333 sizeof(struct pv_entry) - KERNBASE; 517 struct pv_entry *pv, *pl, *pf; 971 vaddrtopte(const struct pv_entry *pv) 1059 struct pv_entry *pv, *tmp; 1378 struct pv_entry *pv; 1433 struct pv_entry *pv = pmap_pg_to_pv(pg) [all...] |
/src/sys/arch/powerpc/ibm4xx/ |
pmap.c | 163 struct pv_entry { struct 164 struct pv_entry *pv_next; /* Linked list of mappings */ 181 struct pv_entry *pv_table; 188 struct pv_entry *pa_to_pv(paddr_t); 202 struct pv_entry * 465 struct pv_entry *pv; 471 sz = (vsize_t)((sizeof(struct pv_entry) + 1) * npgs); 477 pv = pv_table = (struct pv_entry *)addr; 499 pool_init(&pv_pool, sizeof(struct pv_entry), 0, 0, 0, "pv_entry", [all...] |
/src/sys/arch/sh3/sh3/ |
pmap.c | 72 /* pv_entry ops. */ 73 struct pv_entry { struct 76 SLIST_ENTRY(pv_entry) pv_link; 214 pool_init(&__pmap_pv_pool, sizeof(struct pv_entry), 0, 0, 0, "pvpl", 472 struct pv_entry *pv; 552 struct pv_entry *pv; 570 SLIST_REMOVE(&pvh->pvh_head, pv, pv_entry, pv_link); 706 struct pv_entry *pv; 810 struct pv_entry *pv; 853 struct pv_entry *pv [all...] |
/src/sys/arch/aarch64/aarch64/ |
pmap.c | 113 PMAP_COUNTER(pv_enter, "pv_entry fill"); 114 PMAP_COUNTER(pv_remove_dyn, "pv_entry free and unlink dynamic"); 115 PMAP_COUNTER(pv_remove_emb, "pv_entry clear embedded"); 116 PMAP_COUNTER(pv_remove_nopv, "no pv_entry found when removing pv"); 154 PMAP_COUNTER(pv_entry_cannotalloc, "pv_entry allocation failure"); 196 struct pv_entry **); 472 memset(v, 0, sizeof(struct pv_entry)); 491 pool_cache_bootstrap(&_pmap_pv_pool, sizeof(struct pv_entry), 1050 static struct pv_entry * 1054 struct pv_entry *pv, *ppv [all...] |
/src/sys/arch/m68k/m68k/ |
pmap_motorola.c | 267 struct pv_entry pvh_first; /* first PV entry */ 314 struct pv_entry **); 580 * Initialize the pv_entry pools. 582 pool_init(&pmap_pv_pool, sizeof(struct pv_entry), 0, 0, 0, "pvpl", 971 struct pv_entry *pv; 1128 struct pv_entry *opv = NULL; 1229 struct pv_entry *pv, *npv; 1705 struct pv_entry *pv; 2080 struct pv_entry **opvp) 2084 struct pv_entry *pv, *npv, *opv = NULL [all...] |
/src/sys/arch/x86/x86/ |
pmap.c | 226 * - struct pv_entry: describes one <PMAP,VA> mapping of a PA 228 * physical memory. the pp_pvlist points to a list of pv_entry 467 LIST_HEAD(, pv_entry) pvp_pves; 473 #define PVE_PER_PVP ((PAGE_SIZE / sizeof(struct pv_entry)) - 1) 486 .rbto_node_offset = offsetof(struct pv_entry, pve_rb), 523 static struct pv_entry *pmap_alloc_pv(struct pmap *); 524 static void pmap_free_pv(struct pmap *, struct pv_entry *); 589 pve_to_pvpte(struct pv_entry *pve) 598 static inline struct pv_entry * 601 struct pv_entry *pve = (void *)pvpte [all...] |
/src/sys/arch/arm/arm32/ |
pmap.c | 665 * Real definition of pv_entry. 667 struct pv_entry { struct 668 SLIST_ENTRY(pv_entry) pv_link; /* next pv_entry */ 702 static void pmap_enter_pv(struct vm_page_md *, paddr_t, struct pv_entry *, 704 static struct pv_entry *pmap_find_pv(struct vm_page_md *, pmap_t, vaddr_t); 705 static struct pv_entry *pmap_remove_pv(struct vm_page_md *, paddr_t, pmap_t, vaddr_t); 963 * main pv_entry manipulation functions: 976 * => we will gain the lock on the vm_page and allocate the new pv_entry 981 pmap_enter_pv(struct vm_page_md *md, paddr_t pa, struct pv_entry *pv, pmap_t pm [all...] |
/src/sys/arch/powerpc/powerpc/ |
db_interface.c | 551 struct pv_entry { struct in function:db_ppc4xx_pv 552 struct pv_entry *pv_next; /* Linked list of mappings */ 556 struct pv_entry *pa_to_pv(paddr_t); 557 struct pv_entry *pv;
|
/src/sys/arch/riscv/riscv/ |
pmap_machdep.c | 457 pool_init(&pmap_pv_pool, sizeof(struct pv_entry), 0, 0, 0, "pvpl",
|
/src/sys/arch/sun2/sun2/ |
pmap.c | 263 struct pv_entry { struct 264 struct pv_entry *pv_next; 268 typedef struct pv_entry *pv_entry_t; 271 static struct pv_entry **pv_head_tbl; 274 static struct pv_entry *pv_free_list; 423 static inline struct pv_entry ** 1051 * One pv_entry for each page frame. 2078 /* Was not main memory, so no pv_entry for it. */ 2086 * If not changing the type or pfnum then re-use pv_entry. 2090 do_pv = false; /* re-use pv_entry */ [all...] |