pte.h revision 1.11 1 1.11 skrll /* $NetBSD: pte.h,v 1.11 2020/08/22 15:34:51 skrll Exp $ */
2 1.1 matt /*-
3 1.2 matt * Copyright (c) 2010, 2011 The NetBSD Foundation, Inc.
4 1.1 matt * All rights reserved.
5 1.1 matt *
6 1.1 matt * This code is derived from software contributed to The NetBSD Foundation
7 1.2 matt * by Raytheon BBN Technologies Corp and Defense Advanced Research Projects
8 1.2 matt * Agency and which was developed by Matt Thomas of 3am Software Foundry.
9 1.2 matt *
10 1.2 matt * This material is based upon work supported by the Defense Advanced Research
11 1.2 matt * Projects Agency and Space and Naval Warfare Systems Center, Pacific, under
12 1.2 matt * Contract No. N66001-09-C-2073.
13 1.2 matt * Approved for Public Release, Distribution Unlimited
14 1.1 matt *
15 1.1 matt * Redistribution and use in source and binary forms, with or without
16 1.1 matt * modification, are permitted provided that the following conditions
17 1.1 matt * are met:
18 1.1 matt * 1. Redistributions of source code must retain the above copyright
19 1.1 matt * notice, this list of conditions and the following disclaimer.
20 1.1 matt * 2. Redistributions in binary form must reproduce the above copyright
21 1.1 matt * notice, this list of conditions and the following disclaimer in the
22 1.1 matt * documentation and/or other materials provided with the distribution.
23 1.1 matt *
24 1.1 matt * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
25 1.1 matt * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
26 1.1 matt * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
27 1.1 matt * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
28 1.1 matt * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
29 1.1 matt * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
30 1.1 matt * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
31 1.1 matt * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
32 1.1 matt * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
33 1.1 matt * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
34 1.1 matt * POSSIBILITY OF SUCH DAMAGE.
35 1.1 matt */
36 1.1 matt
37 1.1 matt #ifndef _POWERPC_BOOKE_PTE_H_
38 1.1 matt #define _POWERPC_BOOKE_PTE_H_
39 1.1 matt
40 1.1 matt #ifndef _LOCORE
41 1.4 matt #ifndef __BSD_PT_ENTRY_T
42 1.4 matt #define __BSD_PT_ENTRY_T __uint32_t
43 1.4 matt typedef __BSD_PT_ENTRY_T pt_entry_t;
44 1.7 matt #define PRIxPTE PRIx32
45 1.4 matt #endif
46 1.1 matt #endif
47 1.1 matt
48 1.2 matt #include <powerpc/booke/spr.h>
49 1.2 matt
50 1.1 matt /*
51 1.1 matt * The PTE format is software and must be translated into the various portions
52 1.1 matt * X W R are separted by single bits so that they can map to the MAS2 bits
53 1.1 matt * UX/UW/UR or SX/SW/SR by a mask and a shift.
54 1.1 matt */
55 1.2 matt #define PTE_IO (PTE_I|PTE_G|PTE_xW|PTE_xR)
56 1.2 matt #define PTE_DEFAULT (PTE_M|PTE_xX|PTE_xW|PTE_xR)
57 1.2 matt #define PTE_MAS3_MASK (MAS3_RPN|MAS3_U2|MAS3_U0)
58 1.1 matt #define PTE_MAS2_MASK (MAS2_WIMGE)
59 1.1 matt #define PTE_RPN_MASK MAS3_RPN /* MAS3[RPN] */
60 1.1 matt #define PTE_RWX_MASK (PTE_xX|PTE_xW|PTE_xR)
61 1.1 matt #define PTE_WIRED (MAS3_U0 << 2) /* page is wired (PTE only) */
62 1.1 matt #define PTE_xX (MAS3_U0 << 1) /* MAS2[UX] | MAS2[SX] */
63 1.1 matt #define PTE_UNSYNCED MAS3_U0 /* page needs isync */
64 1.1 matt #define PTE_xW MAS3_U1 /* MAS2[UW] | MAS2[SW] */
65 1.1 matt #define PTE_UNMODIFIED MAS3_U2 /* page is unmodified */
66 1.1 matt #define PTE_xR MAS3_U3 /* MAS2[UR] | MAS2[SR] */
67 1.2 matt #define PTE_RWX_SHIFT 6
68 1.2 matt #define PTE_UNUSED 0x00000020
69 1.1 matt #define PTE_WIMGE_MASK MAS2_WIMGE
70 1.2 matt #define PTE_WIG (PTE_W|PTE_I|PTE_G)
71 1.1 matt #define PTE_W MAS2_W /* Write-through */
72 1.1 matt #define PTE_I MAS2_I /* cache-Inhibited */
73 1.1 matt #define PTE_M MAS2_M /* Memory coherence */
74 1.1 matt #define PTE_G MAS2_G /* Guarded */
75 1.1 matt #define PTE_E MAS2_E /* [Little] Endian */
76 1.1 matt
77 1.2 matt #ifndef _LOCORE
78 1.2 matt #ifdef _KERNEL
79 1.2 matt
80 1.10 christos static __inline uint32_t
81 1.7 matt pte_value(pt_entry_t pt_entry)
82 1.7 matt {
83 1.7 matt return pt_entry;
84 1.7 matt }
85 1.7 matt
86 1.10 christos static __inline bool
87 1.2 matt pte_cached_p(pt_entry_t pt_entry)
88 1.2 matt {
89 1.2 matt return (pt_entry & PTE_I) == 0;
90 1.2 matt }
91 1.2 matt
92 1.10 christos static __inline bool
93 1.2 matt pte_modified_p(pt_entry_t pt_entry)
94 1.2 matt {
95 1.2 matt return (pt_entry & (PTE_UNMODIFIED|PTE_xW)) == PTE_xW;
96 1.2 matt }
97 1.2 matt
98 1.10 christos static __inline bool
99 1.2 matt pte_valid_p(pt_entry_t pt_entry)
100 1.2 matt {
101 1.2 matt return pt_entry != 0;
102 1.2 matt }
103 1.2 matt
104 1.10 christos static __inline bool
105 1.2 matt pte_exec_p(pt_entry_t pt_entry)
106 1.2 matt {
107 1.2 matt return (pt_entry & PTE_xX) != 0;
108 1.2 matt }
109 1.2 matt
110 1.10 christos static __inline bool
111 1.8 matt pte_readonly_p(pt_entry_t pt_entry)
112 1.8 matt {
113 1.8 matt return (pt_entry & PTE_xW) == 0;
114 1.8 matt }
115 1.8 matt
116 1.10 christos static __inline bool
117 1.2 matt pte_deferred_exec_p(pt_entry_t pt_entry)
118 1.2 matt {
119 1.2 matt //return (pt_entry & (PTE_xX|PTE_UNSYNCED)) == (PTE_xX|PTE_UNSYNCED);
120 1.2 matt return (pt_entry & PTE_UNSYNCED) == PTE_UNSYNCED;
121 1.2 matt }
122 1.2 matt
123 1.10 christos static __inline bool
124 1.2 matt pte_wired_p(pt_entry_t pt_entry)
125 1.2 matt {
126 1.2 matt return (pt_entry & PTE_WIRED) != 0;
127 1.2 matt }
128 1.2 matt
129 1.10 christos static __inline pt_entry_t
130 1.2 matt pte_nv_entry(bool kernel)
131 1.2 matt {
132 1.2 matt return 0;
133 1.2 matt }
134 1.2 matt
135 1.10 christos static __inline paddr_t
136 1.2 matt pte_to_paddr(pt_entry_t pt_entry)
137 1.2 matt {
138 1.2 matt return (paddr_t)(pt_entry & PTE_RPN_MASK);
139 1.2 matt }
140 1.2 matt
141 1.10 christos static __inline pt_entry_t
142 1.2 matt pte_ionocached_bits(void)
143 1.2 matt {
144 1.6 matt return PTE_I|PTE_G;
145 1.2 matt }
146 1.2 matt
147 1.10 christos static __inline pt_entry_t
148 1.2 matt pte_iocached_bits(void)
149 1.2 matt {
150 1.2 matt return PTE_G;
151 1.2 matt }
152 1.2 matt
153 1.10 christos static __inline pt_entry_t
154 1.2 matt pte_nocached_bits(void)
155 1.2 matt {
156 1.2 matt return PTE_M|PTE_I;
157 1.2 matt }
158 1.2 matt
159 1.10 christos static __inline pt_entry_t
160 1.2 matt pte_cached_bits(void)
161 1.2 matt {
162 1.2 matt return PTE_M;
163 1.2 matt }
164 1.2 matt
165 1.10 christos static __inline pt_entry_t
166 1.2 matt pte_cached_change(pt_entry_t pt_entry, bool cached)
167 1.2 matt {
168 1.2 matt return (pt_entry & ~PTE_I) | (cached ? 0 : PTE_I);
169 1.2 matt }
170 1.2 matt
171 1.10 christos static __inline pt_entry_t
172 1.5 matt pte_wire_entry(pt_entry_t pt_entry)
173 1.2 matt {
174 1.5 matt return pt_entry | PTE_WIRED;
175 1.5 matt }
176 1.5 matt
177 1.10 christos static __inline pt_entry_t
178 1.5 matt pte_unwire_entry(pt_entry_t pt_entry)
179 1.5 matt {
180 1.5 matt return pt_entry & ~PTE_WIRED;
181 1.2 matt }
182 1.2 matt
183 1.10 christos static __inline pt_entry_t
184 1.2 matt pte_prot_nowrite(pt_entry_t pt_entry)
185 1.2 matt {
186 1.2 matt return pt_entry & ~(PTE_xW|PTE_UNMODIFIED);
187 1.2 matt }
188 1.2 matt
189 1.10 christos static __inline pt_entry_t
190 1.2 matt pte_prot_downgrade(pt_entry_t pt_entry, vm_prot_t newprot)
191 1.2 matt {
192 1.2 matt pt_entry &= ~(PTE_xW|PTE_UNMODIFIED);
193 1.2 matt if ((newprot & VM_PROT_EXECUTE) == 0)
194 1.2 matt pt_entry &= ~(PTE_xX|PTE_UNSYNCED);
195 1.2 matt return pt_entry;
196 1.2 matt }
197 1.2 matt
198 1.10 christos static __inline pt_entry_t
199 1.3 matt pte_prot_bits(struct vm_page_md *mdpg, vm_prot_t prot)
200 1.2 matt {
201 1.2 matt KASSERT(prot & VM_PROT_READ);
202 1.2 matt pt_entry_t pt_entry = PTE_xR;
203 1.2 matt if (prot & VM_PROT_EXECUTE) {
204 1.2 matt #if 0
205 1.2 matt pt_entry |= PTE_xX;
206 1.3 matt if (mdpg != NULL && !VM_PAGEMD_EXECPAGE_P(mdpg))
207 1.2 matt pt_entry |= PTE_UNSYNCED;
208 1.2 matt #elif 1
209 1.3 matt if (mdpg != NULL && !VM_PAGEMD_EXECPAGE_P(mdpg))
210 1.2 matt pt_entry |= PTE_UNSYNCED;
211 1.2 matt else
212 1.2 matt pt_entry |= PTE_xX;
213 1.2 matt #else
214 1.2 matt pt_entry |= PTE_UNSYNCED;
215 1.2 matt #endif
216 1.2 matt }
217 1.2 matt if (prot & VM_PROT_WRITE) {
218 1.2 matt pt_entry |= PTE_xW;
219 1.3 matt if (mdpg != NULL && !VM_PAGEMD_MODIFIED_P(mdpg))
220 1.2 matt pt_entry |= PTE_UNMODIFIED;
221 1.2 matt }
222 1.2 matt return pt_entry;
223 1.2 matt }
224 1.2 matt
225 1.10 christos static __inline pt_entry_t
226 1.3 matt pte_flag_bits(struct vm_page_md *mdpg, int flags)
227 1.2 matt {
228 1.6 matt if (__predict_false(flags & PMAP_NOCACHE)) {
229 1.3 matt if (__predict_true(mdpg != NULL)) {
230 1.2 matt return pte_nocached_bits();
231 1.2 matt } else {
232 1.2 matt return pte_ionocached_bits();
233 1.2 matt }
234 1.2 matt } else {
235 1.3 matt if (__predict_false(mdpg != NULL)) {
236 1.2 matt return pte_cached_bits();
237 1.2 matt } else {
238 1.2 matt return pte_iocached_bits();
239 1.2 matt }
240 1.2 matt }
241 1.2 matt }
242 1.2 matt
243 1.10 christos static __inline pt_entry_t
244 1.3 matt pte_make_enter(paddr_t pa, struct vm_page_md *mdpg, vm_prot_t prot,
245 1.2 matt int flags, bool kernel)
246 1.2 matt {
247 1.2 matt pt_entry_t pt_entry = (pt_entry_t) pa & PTE_RPN_MASK;
248 1.2 matt
249 1.3 matt pt_entry |= pte_flag_bits(mdpg, flags);
250 1.3 matt pt_entry |= pte_prot_bits(mdpg, prot);
251 1.2 matt
252 1.2 matt return pt_entry;
253 1.2 matt }
254 1.2 matt
255 1.10 christos static __inline pt_entry_t
256 1.3 matt pte_make_kenter_pa(paddr_t pa, struct vm_page_md *mdpg, vm_prot_t prot,
257 1.2 matt int flags)
258 1.2 matt {
259 1.2 matt pt_entry_t pt_entry = (pt_entry_t) pa & PTE_RPN_MASK;
260 1.2 matt
261 1.5 matt pt_entry |= PTE_WIRED;
262 1.3 matt pt_entry |= pte_flag_bits(mdpg, flags);
263 1.2 matt pt_entry |= pte_prot_bits(NULL, prot); /* pretend unmanaged */
264 1.2 matt
265 1.2 matt return pt_entry;
266 1.2 matt }
267 1.7 matt
268 1.10 christos static __inline void
269 1.9 skrll pte_set(pt_entry_t *ptep, pt_entry_t pte)
270 1.9 skrll {
271 1.9 skrll *ptep = pte;
272 1.9 skrll }
273 1.9 skrll
274 1.2 matt #endif /* _KERNEL */
275 1.2 matt #endif /* !_LOCORE */
276 1.2 matt
277 1.1 matt #endif /* !_POWERPC_BOOKE_PTE_H_ */
278