pmap_segtab.c revision 1.20 1 1.20 mrg /* $NetBSD: pmap_segtab.c,v 1.20 2020/08/20 23:36:45 mrg Exp $ */
2 1.1 christos
3 1.1 christos /*-
4 1.1 christos * Copyright (c) 1998, 2001 The NetBSD Foundation, Inc.
5 1.1 christos * All rights reserved.
6 1.1 christos *
7 1.1 christos * This code is derived from software contributed to The NetBSD Foundation
8 1.1 christos * by Jason R. Thorpe of the Numerical Aerospace Simulation Facility,
9 1.1 christos * NASA Ames Research Center and by Chris G. Demetriou.
10 1.1 christos *
11 1.1 christos * Redistribution and use in source and binary forms, with or without
12 1.1 christos * modification, are permitted provided that the following conditions
13 1.1 christos * are met:
14 1.1 christos * 1. Redistributions of source code must retain the above copyright
15 1.1 christos * notice, this list of conditions and the following disclaimer.
16 1.1 christos * 2. Redistributions in binary form must reproduce the above copyright
17 1.1 christos * notice, this list of conditions and the following disclaimer in the
18 1.1 christos * documentation and/or other materials provided with the distribution.
19 1.1 christos *
20 1.1 christos * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
21 1.1 christos * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
22 1.1 christos * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
23 1.1 christos * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
24 1.1 christos * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
25 1.1 christos * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
26 1.1 christos * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
27 1.1 christos * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
28 1.1 christos * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
29 1.1 christos * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
30 1.1 christos * POSSIBILITY OF SUCH DAMAGE.
31 1.1 christos */
32 1.1 christos
33 1.1 christos /*
34 1.1 christos * Copyright (c) 1992, 1993
35 1.1 christos * The Regents of the University of California. All rights reserved.
36 1.1 christos *
37 1.1 christos * This code is derived from software contributed to Berkeley by
38 1.1 christos * the Systems Programming Group of the University of Utah Computer
39 1.1 christos * Science Department and Ralph Campbell.
40 1.1 christos *
41 1.1 christos * Redistribution and use in source and binary forms, with or without
42 1.1 christos * modification, are permitted provided that the following conditions
43 1.1 christos * are met:
44 1.1 christos * 1. Redistributions of source code must retain the above copyright
45 1.1 christos * notice, this list of conditions and the following disclaimer.
46 1.1 christos * 2. Redistributions in binary form must reproduce the above copyright
47 1.1 christos * notice, this list of conditions and the following disclaimer in the
48 1.1 christos * documentation and/or other materials provided with the distribution.
49 1.1 christos * 3. Neither the name of the University nor the names of its contributors
50 1.1 christos * may be used to endorse or promote products derived from this software
51 1.1 christos * without specific prior written permission.
52 1.1 christos *
53 1.1 christos * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
54 1.1 christos * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
55 1.1 christos * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
56 1.1 christos * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
57 1.1 christos * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
58 1.1 christos * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
59 1.1 christos * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
60 1.1 christos * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
61 1.1 christos * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
62 1.1 christos * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
63 1.1 christos * SUCH DAMAGE.
64 1.1 christos *
65 1.1 christos * @(#)pmap.c 8.4 (Berkeley) 1/26/94
66 1.1 christos */
67 1.1 christos
68 1.1 christos #include <sys/cdefs.h>
69 1.1 christos
70 1.20 mrg __KERNEL_RCSID(0, "$NetBSD: pmap_segtab.c,v 1.20 2020/08/20 23:36:45 mrg Exp $");
71 1.1 christos
72 1.1 christos /*
73 1.1 christos * Manages physical address maps.
74 1.1 christos *
75 1.1 christos * In addition to hardware address maps, this
76 1.1 christos * module is called upon to provide software-use-only
77 1.1 christos * maps which may or may not be stored in the same
78 1.1 christos * form as hardware maps. These pseudo-maps are
79 1.1 christos * used to store intermediate results from copy
80 1.1 christos * operations to and from address spaces.
81 1.1 christos *
82 1.1 christos * Since the information managed by this module is
83 1.1 christos * also stored by the logical address mapping module,
84 1.1 christos * this module may throw away valid virtual-to-physical
85 1.1 christos * mappings at almost any time. However, invalidations
86 1.1 christos * of virtual-to-physical mappings must be done as
87 1.1 christos * requested.
88 1.1 christos *
89 1.1 christos * In order to cope with hardware architectures which
90 1.1 christos * make virtual-to-physical map invalidates expensive,
91 1.1 christos * this module may delay invalidate or reduced protection
92 1.1 christos * operations until such time as they are actually
93 1.1 christos * necessary. This module is given full information as
94 1.1 christos * to which processors are currently using which maps,
95 1.1 christos * and to when physical maps must be made correct.
96 1.1 christos */
97 1.1 christos
98 1.1 christos #define __PMAP_PRIVATE
99 1.1 christos
100 1.1 christos #include "opt_multiprocessor.h"
101 1.1 christos
102 1.1 christos #include <sys/param.h>
103 1.13 skrll
104 1.13 skrll #include <sys/atomic.h>
105 1.13 skrll #include <sys/mutex.h>
106 1.13 skrll #include <sys/proc.h>
107 1.1 christos #include <sys/systm.h>
108 1.1 christos
109 1.1 christos #include <uvm/uvm.h>
110 1.1 christos
111 1.1 christos CTASSERT(NBPG >= sizeof(pmap_segtab_t));
112 1.1 christos
113 1.1 christos struct pmap_segtab_info {
114 1.1 christos pmap_segtab_t *free_segtab; /* free list kept locally */
115 1.1 christos #ifdef DEBUG
116 1.1 christos uint32_t nget_segtab;
117 1.1 christos uint32_t nput_segtab;
118 1.1 christos uint32_t npage_segtab;
119 1.1 christos #define SEGTAB_ADD(n, v) (pmap_segtab_info.n ## _segtab += (v))
120 1.1 christos #else
121 1.1 christos #define SEGTAB_ADD(n, v) ((void) 0)
122 1.1 christos #endif
123 1.1 christos #ifdef PMAP_PTP_CACHE
124 1.1 christos struct pgflist ptp_pgflist; /* Keep a list of idle page tables. */
125 1.1 christos #endif
126 1.1 christos } pmap_segtab_info = {
127 1.1 christos #ifdef PMAP_PTP_CACHE
128 1.1 christos .ptp_pgflist = LIST_HEAD_INITIALIZER(pmap_segtab_info.ptp_pgflist),
129 1.1 christos #endif
130 1.1 christos };
131 1.1 christos
132 1.1 christos kmutex_t pmap_segtab_lock __cacheline_aligned;
133 1.1 christos
134 1.19 mrg /*
135 1.19 mrg * Check that a seg_tab[] array is empty.
136 1.19 mrg *
137 1.19 mrg * This is used when allocating or freeing a pmap_segtab_t. The stp
138 1.19 mrg * should be unused -- meaning, none of the seg_tab[] pointers are
139 1.19 mrg * not NULL, as it transitions from either freshly allocated segtab from
140 1.19 mrg * pmap pool, an unused allocated page segtab alloc from the SMP case,
141 1.19 mrg * where two CPUs attempt to allocate the same underlying segtab, the
142 1.19 mrg * release of a segtab entry to the freelist, or for SMP, where reserve
143 1.19 mrg * also frees a freshly allocated but unused entry.
144 1.19 mrg */
145 1.4 mrg static void
146 1.4 mrg pmap_check_stp(pmap_segtab_t *stp, const char *caller, const char *why)
147 1.4 mrg {
148 1.4 mrg #ifdef DEBUG
149 1.4 mrg for (size_t i = 0; i < PMAP_SEGTABSIZE; i++) {
150 1.19 mrg if (stp->seg_tab[i] != NULL) {
151 1.19 mrg #define DEBUG_NOISY
152 1.4 mrg #ifdef DEBUG_NOISY
153 1.19 mrg UVMHIST_FUNC(__func__);
154 1.19 mrg UVMHIST_CALLARGS(pmapsegtabhist, "stp=%#jx",
155 1.19 mrg (uintptr_t)stp, 0, 0, 0);
156 1.4 mrg for (size_t j = i; j < PMAP_SEGTABSIZE; j++)
157 1.19 mrg if (stp->seg_tab[j] != NULL)
158 1.19 mrg printf("%s: stp->seg_tab[%zu] = %p\n",
159 1.19 mrg caller, j, stp->seg_tab[j]);
160 1.4 mrg #endif
161 1.14 rin panic("%s: pm_segtab.seg_tab[%zu] != 0 (%p): %s",
162 1.11 skrll caller, i, stp->seg_tab[i], why);
163 1.4 mrg }
164 1.4 mrg }
165 1.4 mrg #endif
166 1.4 mrg }
167 1.4 mrg
168 1.19 mrg /*
169 1.19 mrg * Check that an array of ptes is actually zero.
170 1.19 mrg */
171 1.19 mrg static void
172 1.19 mrg pmap_check_ptes(pt_entry_t *pte, const char *caller)
173 1.19 mrg {
174 1.19 mrg #ifdef DEBUG
175 1.19 mrg for (size_t i = 0; i < NPTEPG; i++)
176 1.19 mrg if (!pte_zero_p(pte[i])) {
177 1.19 mrg #ifdef DEBUG_NOISY
178 1.19 mrg UVMHIST_FUNC(__func__);
179 1.19 mrg UVMHIST_CALLARGS(pmapsegtabhist, "pte=%#jx",
180 1.19 mrg (uintptr_t)pte, 0, 0, 0);
181 1.19 mrg for (size_t j = i + 1; j < NPTEPG; j++)
182 1.19 mrg if (!pte_zero_p(pte[j]))
183 1.19 mrg UVMHIST_LOG(pmapsegtabhist,
184 1.19 mrg "pte[%zu] = %#"PRIxPTE,
185 1.19 mrg j, pte_value(pte[j]), 0, 0);
186 1.19 mrg #endif
187 1.20 mrg panic("%s: pte[%zu] entry at %p not 0 (%#"PRIxPTE")",
188 1.19 mrg caller, i, &pte[i], pte_value(pte[i]));
189 1.19 mrg }
190 1.19 mrg #endif
191 1.19 mrg }
192 1.19 mrg
193 1.1 christos static inline struct vm_page *
194 1.1 christos pmap_pte_pagealloc(void)
195 1.1 christos {
196 1.1 christos struct vm_page *pg;
197 1.1 christos
198 1.2 matt pg = PMAP_ALLOC_POOLPAGE(UVM_PGA_ZERO|UVM_PGA_USERESERVE);
199 1.1 christos if (pg) {
200 1.1 christos #ifdef UVM_PAGE_TRKOWN
201 1.1 christos pg->owner_tag = NULL;
202 1.1 christos #endif
203 1.1 christos UVM_PAGE_OWN(pg, "pmap-ptp");
204 1.1 christos }
205 1.1 christos
206 1.1 christos return pg;
207 1.1 christos }
208 1.1 christos
209 1.1 christos static inline pt_entry_t *
210 1.1 christos pmap_segmap(struct pmap *pmap, vaddr_t va)
211 1.1 christos {
212 1.1 christos pmap_segtab_t *stp = pmap->pm_segtab;
213 1.6 skrll KASSERTMSG(pmap != pmap_kernel() || !pmap_md_direct_mapped_vaddr_p(va),
214 1.6 skrll "pmap %p va %#" PRIxVADDR, pmap, va);
215 1.1 christos #ifdef _LP64
216 1.1 christos stp = stp->seg_seg[(va >> XSEGSHIFT) & (NSEGPG - 1)];
217 1.1 christos if (stp == NULL)
218 1.1 christos return NULL;
219 1.1 christos #endif
220 1.1 christos
221 1.1 christos return stp->seg_tab[(va >> SEGSHIFT) & (PMAP_SEGTABSIZE - 1)];
222 1.1 christos }
223 1.1 christos
224 1.1 christos pt_entry_t *
225 1.1 christos pmap_pte_lookup(pmap_t pmap, vaddr_t va)
226 1.1 christos {
227 1.1 christos pt_entry_t *pte = pmap_segmap(pmap, va);
228 1.1 christos if (pte == NULL)
229 1.1 christos return NULL;
230 1.1 christos
231 1.1 christos return pte + ((va >> PGSHIFT) & (NPTEPG - 1));
232 1.1 christos }
233 1.1 christos
234 1.19 mrg /*
235 1.19 mrg * Insert the segtab into the segtab freelist.
236 1.19 mrg */
237 1.1 christos static void
238 1.1 christos pmap_segtab_free(pmap_segtab_t *stp)
239 1.1 christos {
240 1.19 mrg UVMHIST_FUNC(__func__);
241 1.19 mrg
242 1.19 mrg UVMHIST_CALLARGS(pmapsegtabhist, "stp=%#jx", stp, 0, 0, 0);
243 1.19 mrg
244 1.1 christos mutex_spin_enter(&pmap_segtab_lock);
245 1.1 christos stp->seg_seg[0] = pmap_segtab_info.free_segtab;
246 1.1 christos pmap_segtab_info.free_segtab = stp;
247 1.1 christos SEGTAB_ADD(nput, 1);
248 1.1 christos mutex_spin_exit(&pmap_segtab_lock);
249 1.1 christos }
250 1.1 christos
251 1.1 christos static void
252 1.1 christos pmap_segtab_release(pmap_t pmap, pmap_segtab_t **stp_p, bool free_stp,
253 1.1 christos pte_callback_t callback, uintptr_t flags,
254 1.1 christos vaddr_t va, vsize_t vinc)
255 1.1 christos {
256 1.1 christos pmap_segtab_t *stp = *stp_p;
257 1.1 christos
258 1.16 mrg UVMHIST_FUNC(__func__);
259 1.19 mrg UVMHIST_CALLARGS(pmapsegtabhist, "pm=%#jx stpp=%#jx free=%jd",
260 1.16 mrg (uintptr_t)pmap, (uintptr_t)stp_p, free_stp, 0);
261 1.19 mrg UVMHIST_LOG(pmapsegtabhist, " callback=%jx flags=%jx va=%jx vinc=%jx",
262 1.16 mrg (uintptr_t)callback, flags, (uintptr_t)va, (uintptr_t)vinc);
263 1.4 mrg for (size_t i = (va / vinc) & (PMAP_SEGTABSIZE - 1);
264 1.4 mrg i < PMAP_SEGTABSIZE;
265 1.4 mrg i++, va += vinc) {
266 1.1 christos #ifdef _LP64
267 1.1 christos if (vinc > NBSEG) {
268 1.1 christos if (stp->seg_seg[i] != NULL) {
269 1.19 mrg UVMHIST_LOG(pmapsegtabhist,
270 1.19 mrg " recursing %jd", i, 0, 0, 0);
271 1.1 christos pmap_segtab_release(pmap, &stp->seg_seg[i],
272 1.1 christos true, callback, flags, va, vinc / NSEGPG);
273 1.1 christos KASSERT(stp->seg_seg[i] == NULL);
274 1.1 christos }
275 1.1 christos continue;
276 1.1 christos }
277 1.1 christos #endif
278 1.1 christos KASSERT(vinc == NBSEG);
279 1.1 christos
280 1.1 christos /* get pointer to segment map */
281 1.1 christos pt_entry_t *pte = stp->seg_tab[i];
282 1.1 christos if (pte == NULL)
283 1.1 christos continue;
284 1.19 mrg pmap_check_ptes(pte, __func__);
285 1.19 mrg
286 1.19 mrg #if defined(__mips_n64) && PAGE_SIZE == 8192
287 1.19 mrg /*
288 1.19 mrg * XXX This is evil. If vinc is 1000000 we are in
289 1.19 mrg * the last level, and this pte should be page aligned.
290 1.19 mrg */
291 1.19 mrg if (vinc == 0x1000000 && ((uintptr_t)pte & PAGE_MASK) != 0) {
292 1.19 mrg panic("%s: pte entry at %p not page aligned",
293 1.19 mrg __func__, pte);
294 1.19 mrg }
295 1.19 mrg #endif
296 1.1 christos
297 1.1 christos /*
298 1.18 simonb * If our caller wants a callback, do so.
299 1.1 christos */
300 1.1 christos if (callback != NULL) {
301 1.1 christos (*callback)(pmap, va, va + vinc, pte, flags);
302 1.1 christos }
303 1.19 mrg
304 1.2 matt // PMAP_UNMAP_POOLPAGE should handle any VCA issues itself
305 1.1 christos paddr_t pa = PMAP_UNMAP_POOLPAGE((vaddr_t)pte);
306 1.1 christos struct vm_page *pg = PHYS_TO_VM_PAGE(pa);
307 1.1 christos #ifdef PMAP_PTP_CACHE
308 1.1 christos mutex_spin_enter(&pmap_segtab_lock);
309 1.12 ad LIST_INSERT_HEAD(&pmap_segtab_info.ptp_pgflist, pg, pageq.list);
310 1.1 christos mutex_spin_exit(&pmap_segtab_lock);
311 1.1 christos #else
312 1.1 christos uvm_pagefree(pg);
313 1.1 christos #endif
314 1.1 christos
315 1.1 christos stp->seg_tab[i] = NULL;
316 1.19 mrg UVMHIST_LOG(pmapsegtabhist, " zeroing tab[%jd]", i, 0, 0, 0);
317 1.1 christos }
318 1.1 christos
319 1.1 christos if (free_stp) {
320 1.5 skrll pmap_check_stp(stp, __func__,
321 1.4 mrg vinc == NBSEG ? "release seg" : "release xseg");
322 1.1 christos pmap_segtab_free(stp);
323 1.1 christos *stp_p = NULL;
324 1.1 christos }
325 1.1 christos }
326 1.1 christos
327 1.1 christos /*
328 1.1 christos * Create and return a physical map.
329 1.1 christos *
330 1.1 christos * If the size specified for the map
331 1.1 christos * is zero, the map is an actual physical
332 1.1 christos * map, and may be referenced by the
333 1.1 christos * hardware.
334 1.1 christos *
335 1.1 christos * If the size specified is non-zero,
336 1.1 christos * the map will be used in software only, and
337 1.1 christos * is bounded by that size.
338 1.1 christos */
339 1.1 christos static pmap_segtab_t *
340 1.1 christos pmap_segtab_alloc(void)
341 1.1 christos {
342 1.1 christos pmap_segtab_t *stp;
343 1.4 mrg bool found_on_freelist = false;
344 1.1 christos
345 1.19 mrg UVMHIST_FUNC(__func__);
346 1.1 christos again:
347 1.1 christos mutex_spin_enter(&pmap_segtab_lock);
348 1.1 christos if (__predict_true((stp = pmap_segtab_info.free_segtab) != NULL)) {
349 1.1 christos pmap_segtab_info.free_segtab = stp->seg_seg[0];
350 1.1 christos stp->seg_seg[0] = NULL;
351 1.1 christos SEGTAB_ADD(nget, 1);
352 1.4 mrg found_on_freelist = true;
353 1.19 mrg UVMHIST_CALLARGS(pmapsegtabhist, "freelist stp=%#jx", stp, 0, 0, 0);
354 1.1 christos }
355 1.1 christos mutex_spin_exit(&pmap_segtab_lock);
356 1.1 christos
357 1.1 christos if (__predict_false(stp == NULL)) {
358 1.1 christos struct vm_page * const stp_pg = pmap_pte_pagealloc();
359 1.1 christos
360 1.1 christos if (__predict_false(stp_pg == NULL)) {
361 1.1 christos /*
362 1.1 christos * XXX What else can we do? Could we deadlock here?
363 1.1 christos */
364 1.10 skrll uvm_wait("segtab");
365 1.1 christos goto again;
366 1.1 christos }
367 1.1 christos SEGTAB_ADD(npage, 1);
368 1.1 christos const paddr_t stp_pa = VM_PAGE_TO_PHYS(stp_pg);
369 1.1 christos
370 1.1 christos stp = (pmap_segtab_t *)PMAP_MAP_POOLPAGE(stp_pa);
371 1.19 mrg UVMHIST_CALLARGS(pmapsegtabhist, "new stp=%#jx", stp, 0, 0, 0);
372 1.1 christos const size_t n = NBPG / sizeof(*stp);
373 1.1 christos if (n > 1) {
374 1.1 christos /*
375 1.1 christos * link all the segtabs in this page together
376 1.1 christos */
377 1.1 christos for (size_t i = 1; i < n - 1; i++) {
378 1.1 christos stp[i].seg_seg[0] = &stp[i+1];
379 1.1 christos }
380 1.1 christos /*
381 1.1 christos * Now link the new segtabs into the free segtab list.
382 1.1 christos */
383 1.1 christos mutex_spin_enter(&pmap_segtab_lock);
384 1.1 christos stp[n-1].seg_seg[0] = pmap_segtab_info.free_segtab;
385 1.1 christos pmap_segtab_info.free_segtab = stp + 1;
386 1.1 christos SEGTAB_ADD(nput, n - 1);
387 1.1 christos mutex_spin_exit(&pmap_segtab_lock);
388 1.1 christos }
389 1.1 christos }
390 1.1 christos
391 1.4 mrg pmap_check_stp(stp, __func__,
392 1.4 mrg found_on_freelist ? "from free list" : "allocated");
393 1.4 mrg
394 1.1 christos return stp;
395 1.1 christos }
396 1.1 christos
397 1.1 christos /*
398 1.1 christos * Allocate the top segment table for the pmap.
399 1.1 christos */
400 1.1 christos void
401 1.1 christos pmap_segtab_init(pmap_t pmap)
402 1.1 christos {
403 1.1 christos
404 1.1 christos pmap->pm_segtab = pmap_segtab_alloc();
405 1.1 christos }
406 1.1 christos
407 1.1 christos /*
408 1.1 christos * Retire the given physical map from service.
409 1.1 christos * Should only be called if the map contains
410 1.1 christos * no valid mappings.
411 1.1 christos */
412 1.1 christos void
413 1.1 christos pmap_segtab_destroy(pmap_t pmap, pte_callback_t func, uintptr_t flags)
414 1.1 christos {
415 1.1 christos if (pmap->pm_segtab == NULL)
416 1.1 christos return;
417 1.1 christos
418 1.1 christos #ifdef _LP64
419 1.1 christos const vsize_t vinc = NBXSEG;
420 1.1 christos #else
421 1.1 christos const vsize_t vinc = NBSEG;
422 1.1 christos #endif
423 1.1 christos pmap_segtab_release(pmap, &pmap->pm_segtab,
424 1.1 christos func == NULL, func, flags, pmap->pm_minaddr, vinc);
425 1.1 christos }
426 1.1 christos
427 1.1 christos /*
428 1.1 christos * Make a new pmap (vmspace) active for the given process.
429 1.1 christos */
430 1.1 christos void
431 1.1 christos pmap_segtab_activate(struct pmap *pm, struct lwp *l)
432 1.1 christos {
433 1.1 christos if (l == curlwp) {
434 1.3 matt struct cpu_info * const ci = l->l_cpu;
435 1.15 skrll pmap_md_xtab_activate(pm, l);
436 1.1 christos KASSERT(pm == l->l_proc->p_vmspace->vm_map.pmap);
437 1.1 christos if (pm == pmap_kernel()) {
438 1.3 matt ci->ci_pmap_user_segtab = PMAP_INVALID_SEGTAB_ADDRESS;
439 1.1 christos #ifdef _LP64
440 1.3 matt ci->ci_pmap_user_seg0tab = PMAP_INVALID_SEGTAB_ADDRESS;
441 1.1 christos #endif
442 1.1 christos } else {
443 1.3 matt ci->ci_pmap_user_segtab = pm->pm_segtab;
444 1.1 christos #ifdef _LP64
445 1.3 matt ci->ci_pmap_user_seg0tab = pm->pm_segtab->seg_seg[0];
446 1.1 christos #endif
447 1.1 christos }
448 1.1 christos }
449 1.1 christos }
450 1.1 christos
451 1.15 skrll
452 1.15 skrll void
453 1.15 skrll pmap_segtab_deactivate(pmap_t pm)
454 1.15 skrll {
455 1.15 skrll
456 1.15 skrll pmap_md_xtab_deactivate(pm);
457 1.15 skrll
458 1.15 skrll curcpu()->ci_pmap_user_segtab = PMAP_INVALID_SEGTAB_ADDRESS;
459 1.15 skrll #ifdef _LP64
460 1.15 skrll curcpu()->ci_pmap_user_seg0tab = NULL;
461 1.15 skrll #endif
462 1.15 skrll
463 1.15 skrll }
464 1.15 skrll
465 1.1 christos /*
466 1.1 christos * Act on the given range of addresses from the specified map.
467 1.1 christos *
468 1.1 christos * It is assumed that the start and end are properly rounded to
469 1.1 christos * the page size.
470 1.1 christos */
471 1.1 christos void
472 1.1 christos pmap_pte_process(pmap_t pmap, vaddr_t sva, vaddr_t eva,
473 1.8 skrll pte_callback_t callback, uintptr_t flags)
474 1.1 christos {
475 1.1 christos #if 0
476 1.1 christos printf("%s: %p, %"PRIxVADDR", %"PRIxVADDR", %p, %"PRIxPTR"\n",
477 1.1 christos __func__, pmap, sva, eva, callback, flags);
478 1.1 christos #endif
479 1.1 christos while (sva < eva) {
480 1.1 christos vaddr_t lastseg_va = pmap_trunc_seg(sva) + NBSEG;
481 1.1 christos if (lastseg_va == 0 || lastseg_va > eva)
482 1.1 christos lastseg_va = eva;
483 1.1 christos
484 1.1 christos /*
485 1.1 christos * If VA belongs to an unallocated segment,
486 1.1 christos * skip to the next segment boundary.
487 1.1 christos */
488 1.9 skrll pt_entry_t * const ptep = pmap_pte_lookup(pmap, sva);
489 1.9 skrll if (ptep != NULL) {
490 1.1 christos /*
491 1.1 christos * Callback to deal with the ptes for this segment.
492 1.1 christos */
493 1.9 skrll (*callback)(pmap, sva, lastseg_va, ptep, flags);
494 1.1 christos }
495 1.1 christos /*
496 1.1 christos * In theory we could release pages with no entries,
497 1.1 christos * but that takes more effort than we want here.
498 1.1 christos */
499 1.1 christos sva = lastseg_va;
500 1.1 christos }
501 1.1 christos }
502 1.1 christos
503 1.1 christos /*
504 1.1 christos * Return a pointer for the pte that corresponds to the specified virtual
505 1.1 christos * address (va) in the target physical map, allocating if needed.
506 1.1 christos */
507 1.1 christos pt_entry_t *
508 1.1 christos pmap_pte_reserve(pmap_t pmap, vaddr_t va, int flags)
509 1.1 christos {
510 1.1 christos pmap_segtab_t *stp = pmap->pm_segtab;
511 1.1 christos pt_entry_t *pte;
512 1.16 mrg UVMHIST_FUNC(__func__);
513 1.16 mrg
514 1.1 christos pte = pmap_pte_lookup(pmap, va);
515 1.1 christos if (__predict_false(pte == NULL)) {
516 1.1 christos #ifdef _LP64
517 1.1 christos pmap_segtab_t ** const stp_p =
518 1.1 christos &stp->seg_seg[(va >> XSEGSHIFT) & (NSEGPG - 1)];
519 1.1 christos if (__predict_false((stp = *stp_p) == NULL)) {
520 1.1 christos pmap_segtab_t *nstp = pmap_segtab_alloc();
521 1.1 christos #ifdef MULTIPROCESSOR
522 1.1 christos pmap_segtab_t *ostp = atomic_cas_ptr(stp_p, NULL, nstp);
523 1.1 christos if (__predict_false(ostp != NULL)) {
524 1.4 mrg pmap_check_stp(nstp, __func__, "reserve");
525 1.1 christos pmap_segtab_free(nstp);
526 1.1 christos nstp = ostp;
527 1.1 christos }
528 1.1 christos #else
529 1.1 christos *stp_p = nstp;
530 1.1 christos #endif /* MULTIPROCESSOR */
531 1.1 christos stp = nstp;
532 1.1 christos }
533 1.1 christos KASSERT(stp == pmap->pm_segtab->seg_seg[(va >> XSEGSHIFT) & (NSEGPG - 1)]);
534 1.1 christos #endif /* _LP64 */
535 1.1 christos struct vm_page *pg = NULL;
536 1.1 christos #ifdef PMAP_PTP_CACHE
537 1.1 christos mutex_spin_enter(&pmap_segtab_lock);
538 1.1 christos if ((pg = LIST_FIRST(&pmap_segtab_info.ptp_pgflist)) != NULL) {
539 1.12 ad LIST_REMOVE(pg, pageq.list);
540 1.1 christos KASSERT(LIST_FIRST(&pmap_segtab_info.ptp_pgflist) != pg);
541 1.1 christos }
542 1.1 christos mutex_spin_exit(&pmap_segtab_lock);
543 1.1 christos #endif
544 1.1 christos if (pg == NULL)
545 1.1 christos pg = pmap_pte_pagealloc();
546 1.1 christos if (pg == NULL) {
547 1.1 christos if (flags & PMAP_CANFAIL)
548 1.1 christos return NULL;
549 1.1 christos panic("%s: cannot allocate page table page "
550 1.1 christos "for va %" PRIxVADDR, __func__, va);
551 1.1 christos }
552 1.1 christos
553 1.1 christos const paddr_t pa = VM_PAGE_TO_PHYS(pg);
554 1.2 matt pte = (pt_entry_t *)PMAP_MAP_POOLPAGE(pa);
555 1.1 christos pt_entry_t ** const pte_p =
556 1.1 christos &stp->seg_tab[(va >> SEGSHIFT) & (PMAP_SEGTABSIZE - 1)];
557 1.1 christos #ifdef MULTIPROCESSOR
558 1.1 christos pt_entry_t *opte = atomic_cas_ptr(pte_p, NULL, pte);
559 1.1 christos /*
560 1.1 christos * If another thread allocated the segtab needed for this va
561 1.1 christos * free the page we just allocated.
562 1.1 christos */
563 1.1 christos if (__predict_false(opte != NULL)) {
564 1.1 christos #ifdef PMAP_PTP_CACHE
565 1.1 christos mutex_spin_enter(&pmap_segtab_lock);
566 1.1 christos LIST_INSERT_HEAD(&pmap_segtab_info.ptp_pgflist,
567 1.12 ad pg, pageq.list);
568 1.1 christos mutex_spin_exit(&pmap_segtab_lock);
569 1.1 christos #else
570 1.2 matt PMAP_UNMAP_POOLPAGE((vaddr_t)pte);
571 1.1 christos uvm_pagefree(pg);
572 1.1 christos #endif
573 1.1 christos pte = opte;
574 1.1 christos }
575 1.1 christos #else
576 1.1 christos *pte_p = pte;
577 1.1 christos #endif
578 1.1 christos KASSERT(pte == stp->seg_tab[(va >> SEGSHIFT) & (PMAP_SEGTABSIZE - 1)]);
579 1.19 mrg UVMHIST_CALLARGS(pmapsegtabhist, "pm=%#jx va=%#jx -> tab[%jd]=%jx",
580 1.19 mrg (uintptr_t)pmap, (uintptr_t)va,
581 1.19 mrg (va >> SEGSHIFT) & (PMAP_SEGTABSIZE - 1), pte);
582 1.1 christos
583 1.19 mrg pmap_check_ptes(pte, __func__);
584 1.2 matt pte += (va >> PGSHIFT) & (NPTEPG - 1);
585 1.1 christos }
586 1.1 christos
587 1.1 christos return pte;
588 1.1 christos }
589