Home | History | Annotate | Line # | Download | only in pmap
pmap_segtab.c revision 1.29
      1  1.29     skrll /*	$NetBSD: pmap_segtab.c,v 1.29 2022/10/26 07:35:20 skrll Exp $	*/
      2   1.1  christos 
      3   1.1  christos /*-
      4   1.1  christos  * Copyright (c) 1998, 2001 The NetBSD Foundation, Inc.
      5   1.1  christos  * All rights reserved.
      6   1.1  christos  *
      7   1.1  christos  * This code is derived from software contributed to The NetBSD Foundation
      8   1.1  christos  * by Jason R. Thorpe of the Numerical Aerospace Simulation Facility,
      9   1.1  christos  * NASA Ames Research Center and by Chris G. Demetriou.
     10   1.1  christos  *
     11   1.1  christos  * Redistribution and use in source and binary forms, with or without
     12   1.1  christos  * modification, are permitted provided that the following conditions
     13   1.1  christos  * are met:
     14   1.1  christos  * 1. Redistributions of source code must retain the above copyright
     15   1.1  christos  *    notice, this list of conditions and the following disclaimer.
     16   1.1  christos  * 2. Redistributions in binary form must reproduce the above copyright
     17   1.1  christos  *    notice, this list of conditions and the following disclaimer in the
     18   1.1  christos  *    documentation and/or other materials provided with the distribution.
     19   1.1  christos  *
     20   1.1  christos  * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
     21   1.1  christos  * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
     22   1.1  christos  * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
     23   1.1  christos  * PURPOSE ARE DISCLAIMED.  IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
     24   1.1  christos  * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
     25   1.1  christos  * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
     26   1.1  christos  * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
     27   1.1  christos  * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
     28   1.1  christos  * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
     29   1.1  christos  * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
     30   1.1  christos  * POSSIBILITY OF SUCH DAMAGE.
     31   1.1  christos  */
     32   1.1  christos 
     33   1.1  christos /*
     34   1.1  christos  * Copyright (c) 1992, 1993
     35   1.1  christos  *	The Regents of the University of California.  All rights reserved.
     36   1.1  christos  *
     37   1.1  christos  * This code is derived from software contributed to Berkeley by
     38   1.1  christos  * the Systems Programming Group of the University of Utah Computer
     39   1.1  christos  * Science Department and Ralph Campbell.
     40   1.1  christos  *
     41   1.1  christos  * Redistribution and use in source and binary forms, with or without
     42   1.1  christos  * modification, are permitted provided that the following conditions
     43   1.1  christos  * are met:
     44   1.1  christos  * 1. Redistributions of source code must retain the above copyright
     45   1.1  christos  *    notice, this list of conditions and the following disclaimer.
     46   1.1  christos  * 2. Redistributions in binary form must reproduce the above copyright
     47   1.1  christos  *    notice, this list of conditions and the following disclaimer in the
     48   1.1  christos  *    documentation and/or other materials provided with the distribution.
     49   1.1  christos  * 3. Neither the name of the University nor the names of its contributors
     50   1.1  christos  *    may be used to endorse or promote products derived from this software
     51   1.1  christos  *    without specific prior written permission.
     52   1.1  christos  *
     53   1.1  christos  * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
     54   1.1  christos  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
     55   1.1  christos  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
     56   1.1  christos  * ARE DISCLAIMED.  IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
     57   1.1  christos  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
     58   1.1  christos  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
     59   1.1  christos  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
     60   1.1  christos  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
     61   1.1  christos  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
     62   1.1  christos  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
     63   1.1  christos  * SUCH DAMAGE.
     64   1.1  christos  *
     65   1.1  christos  *	@(#)pmap.c	8.4 (Berkeley) 1/26/94
     66   1.1  christos  */
     67   1.1  christos 
     68   1.1  christos #include <sys/cdefs.h>
     69   1.1  christos 
     70  1.29     skrll __KERNEL_RCSID(0, "$NetBSD: pmap_segtab.c,v 1.29 2022/10/26 07:35:20 skrll Exp $");
     71   1.1  christos 
     72   1.1  christos /*
     73   1.1  christos  *	Manages physical address maps.
     74   1.1  christos  *
     75   1.1  christos  *	In addition to hardware address maps, this
     76   1.1  christos  *	module is called upon to provide software-use-only
     77   1.1  christos  *	maps which may or may not be stored in the same
     78   1.1  christos  *	form as hardware maps.  These pseudo-maps are
     79   1.1  christos  *	used to store intermediate results from copy
     80   1.1  christos  *	operations to and from address spaces.
     81   1.1  christos  *
     82   1.1  christos  *	Since the information managed by this module is
     83   1.1  christos  *	also stored by the logical address mapping module,
     84   1.1  christos  *	this module may throw away valid virtual-to-physical
     85   1.1  christos  *	mappings at almost any time.  However, invalidations
     86   1.1  christos  *	of virtual-to-physical mappings must be done as
     87   1.1  christos  *	requested.
     88   1.1  christos  *
     89   1.1  christos  *	In order to cope with hardware architectures which
     90   1.1  christos  *	make virtual-to-physical map invalidates expensive,
     91   1.1  christos  *	this module may delay invalidate or reduced protection
     92   1.1  christos  *	operations until such time as they are actually
     93   1.1  christos  *	necessary.  This module is given full information as
     94   1.1  christos  *	to which processors are currently using which maps,
     95   1.1  christos  *	and to when physical maps must be made correct.
     96   1.1  christos  */
     97   1.1  christos 
     98   1.1  christos #define __PMAP_PRIVATE
     99   1.1  christos 
    100   1.1  christos #include "opt_multiprocessor.h"
    101   1.1  christos 
    102   1.1  christos #include <sys/param.h>
    103  1.13     skrll 
    104  1.13     skrll #include <sys/atomic.h>
    105  1.13     skrll #include <sys/mutex.h>
    106  1.13     skrll #include <sys/proc.h>
    107   1.1  christos #include <sys/systm.h>
    108   1.1  christos 
    109   1.1  christos #include <uvm/uvm.h>
    110  1.29     skrll #include <uvm/pmap/pmap.h>
    111   1.1  christos 
    112  1.29     skrll #if defined(XSEGSHIFT) && XSEGSHIFT == SEGSHIFT
    113  1.29     skrll #undef XSEGSHIFT
    114  1.29     skrll #undef XSEGLENGTH
    115  1.29     skrll #undef NBXSEG
    116  1.29     skrll #undef NXSEGPG
    117  1.29     skrll #endif
    118  1.29     skrll 
    119  1.29     skrll #define MULT_CTASSERT(a,b)	__CTASSERT((a) < (b) || ((a) % (b) == 0))
    120  1.29     skrll 
    121  1.29     skrll __CTASSERT(sizeof(pmap_ptpage_t) == NBPG);
    122  1.29     skrll 
    123  1.29     skrll #if defined(PMAP_HWPAGEWALKER)
    124  1.29     skrll #ifdef _LP64
    125  1.29     skrll MULT_CTASSERT(PMAP_PDETABSIZE, NPDEPG);
    126  1.29     skrll MULT_CTASSERT(NPDEPG, PMAP_PDETABSIZE);
    127  1.29     skrll #endif /* _LP64 */
    128  1.29     skrll MULT_CTASSERT(sizeof(pmap_pdetab_t *), sizeof(pd_entry_t));
    129  1.29     skrll MULT_CTASSERT(sizeof(pd_entry_t), sizeof(pmap_pdetab_t));
    130  1.29     skrll 
    131  1.29     skrll #if 0
    132  1.29     skrll #ifdef _LP64
    133  1.29     skrll static const bool separate_pdetab_root_p = NPDEPG != PMAP_PDETABSIZE;
    134  1.29     skrll #else
    135  1.29     skrll static const bool separate_pdetab_root_p = true;
    136  1.29     skrll #endif /* _LP64 */
    137  1.29     skrll #endif
    138   1.1  christos 
    139  1.29     skrll typedef struct {
    140  1.29     skrll 	pmap_pdetab_t *free_pdetab0;	/* free list kept locally */
    141  1.29     skrll 	pmap_pdetab_t *free_pdetab;	/* free list kept locally */
    142  1.29     skrll #ifdef DEBUG
    143  1.29     skrll 	uint32_t nget;
    144  1.29     skrll 	uint32_t nput;
    145  1.29     skrll 	uint32_t npage;
    146  1.29     skrll #define	PDETAB_ADD(n, v)	(pmap_segtab_info.pdealloc.n += (v))
    147  1.29     skrll #else
    148  1.29     skrll #define	PDETAB_ADD(n, v)	((void) 0)
    149  1.29     skrll #endif /* DEBUG */
    150  1.29     skrll } pmap_pdetab_alloc_t;
    151  1.29     skrll #endif /* PMAP_HWPAGEWALKER */
    152  1.29     skrll 
    153  1.29     skrll #if !defined(PMAP_HWPAGEWALKER) || !defined(PMAP_MAP_PDETABPAGE)
    154  1.29     skrll #ifdef _LP64
    155  1.29     skrll __CTASSERT(NSEGPG >= PMAP_SEGTABSIZE);
    156  1.29     skrll __CTASSERT(NSEGPG % PMAP_SEGTABSIZE == 0);
    157  1.29     skrll #endif
    158  1.29     skrll __CTASSERT(NBPG >= sizeof(pmap_segtab_t));
    159  1.29     skrll 
    160  1.29     skrll typedef struct {
    161  1.29     skrll 	pmap_segtab_t *free_segtab0;	/* free list kept locally */
    162   1.1  christos 	pmap_segtab_t *free_segtab;	/* free list kept locally */
    163   1.1  christos #ifdef DEBUG
    164  1.29     skrll 	uint32_t nget;
    165  1.29     skrll 	uint32_t nput;
    166  1.29     skrll 	uint32_t npage;
    167  1.29     skrll #define	SEGTAB_ADD(n, v)	(pmap_segtab_info.segalloc.n += (v))
    168   1.1  christos #else
    169   1.1  christos #define	SEGTAB_ADD(n, v)	((void) 0)
    170   1.1  christos #endif
    171  1.29     skrll } pmap_segtab_alloc_t;
    172  1.29     skrll #endif /* !PMAP_HWPAGEWALKER || !PMAP_MAP_PDETABPAGE */
    173  1.29     skrll 
    174  1.29     skrll struct pmap_segtab_info {
    175  1.29     skrll #if defined(PMAP_HWPAGEWALKER)
    176  1.29     skrll 	pmap_pdetab_alloc_t pdealloc;
    177  1.29     skrll #endif
    178  1.29     skrll #if !defined(PMAP_HWPAGEWALKER) || !defined(PMAP_MAP_PDETABPAGE)
    179  1.29     skrll 	pmap_segtab_alloc_t segalloc;
    180  1.29     skrll #endif
    181  1.29     skrll #ifdef PMAP_PPG_CACHE
    182   1.1  christos 	struct pgflist ptp_pgflist;	/* Keep a list of idle page tables. */
    183   1.1  christos #endif
    184   1.1  christos } pmap_segtab_info = {
    185  1.29     skrll #ifdef PMAP_PPG_CACHE
    186   1.1  christos 	.ptp_pgflist = LIST_HEAD_INITIALIZER(pmap_segtab_info.ptp_pgflist),
    187   1.1  christos #endif
    188   1.1  christos };
    189   1.1  christos 
    190   1.1  christos kmutex_t pmap_segtab_lock __cacheline_aligned;
    191   1.1  christos 
    192  1.29     skrll #ifndef PMAP_HWPAGEWALKER
    193  1.19       mrg /*
    194  1.29     skrll  * Check that a seg_ppg[] array is empty.
    195  1.19       mrg  *
    196  1.28     skrll  * This is used when allocating or freeing a pmap_segtab_t.  The stb
    197  1.29     skrll  * should be unused -- meaning, none of the seg_ppg[] pointers are
    198  1.19       mrg  * not NULL, as it transitions from either freshly allocated segtab from
    199  1.19       mrg  * pmap pool, an unused allocated page segtab alloc from the SMP case,
    200  1.19       mrg  * where two CPUs attempt to allocate the same underlying segtab, the
    201  1.19       mrg  * release of a segtab entry to the freelist, or for SMP, where reserve
    202  1.19       mrg  * also frees a freshly allocated but unused entry.
    203  1.19       mrg  */
    204   1.4       mrg static void
    205  1.28     skrll pmap_check_stb(pmap_segtab_t *stb, const char *caller, const char *why)
    206   1.4       mrg {
    207   1.4       mrg #ifdef DEBUG
    208   1.4       mrg 	for (size_t i = 0; i < PMAP_SEGTABSIZE; i++) {
    209  1.29     skrll 		if (stb->seg_ppg[i] != NULL) {
    210  1.19       mrg #define DEBUG_NOISY
    211   1.4       mrg #ifdef DEBUG_NOISY
    212  1.19       mrg 			UVMHIST_FUNC(__func__);
    213  1.29     skrll 			UVMHIST_CALLARGS(pmapxtabhist, "stb=%#jx",
    214  1.28     skrll 			    (uintptr_t)stb, 0, 0, 0);
    215   1.4       mrg 			for (size_t j = i; j < PMAP_SEGTABSIZE; j++)
    216  1.29     skrll 				if (stb->seg_ppg[j] != NULL)
    217  1.29     skrll 					printf("%s: stb->seg_ppg[%zu] = %p\n",
    218  1.29     skrll 					    caller, j, stb->seg_ppg[j]);
    219   1.4       mrg #endif
    220  1.29     skrll 			panic("%s: pm_segtab.seg_ppg[%zu] != 0 (%p): %s",
    221  1.29     skrll 			    caller, i, stb->seg_ppg[i], why);
    222   1.4       mrg 		}
    223   1.4       mrg 	}
    224   1.4       mrg #endif
    225   1.4       mrg }
    226  1.29     skrll #endif /* PMAP_HWPAGEWALKER */
    227  1.19       mrg 
    228   1.1  christos static inline struct vm_page *
    229   1.1  christos pmap_pte_pagealloc(void)
    230   1.1  christos {
    231   1.1  christos 	struct vm_page *pg;
    232   1.1  christos 
    233  1.29     skrll 	pg = pmap_md_alloc_poolpage(UVM_PGA_ZERO | UVM_PGA_USERESERVE);
    234   1.1  christos 	if (pg) {
    235   1.1  christos #ifdef UVM_PAGE_TRKOWN
    236   1.1  christos 		pg->owner_tag = NULL;
    237   1.1  christos #endif
    238   1.1  christos 		UVM_PAGE_OWN(pg, "pmap-ptp");
    239   1.1  christos 	}
    240   1.1  christos 
    241   1.1  christos 	return pg;
    242   1.1  christos }
    243   1.1  christos 
    244  1.29     skrll #if defined(PMAP_HWPAGEWALKER) && defined(PMAP_MAP_PDETABPAGE)
    245  1.29     skrll static vaddr_t
    246  1.29     skrll pmap_pde_to_va(pd_entry_t pde)
    247  1.29     skrll {
    248  1.29     skrll 	if (!pte_pde_valid_p(pde))
    249  1.29     skrll 		return 0;
    250  1.29     skrll 
    251  1.29     skrll 	paddr_t pa = pte_pde_to_paddr(pde);
    252  1.29     skrll 	return pmap_md_direct_map_paddr(pa);
    253  1.29     skrll }
    254  1.29     skrll 
    255  1.29     skrll #ifdef _LP64
    256  1.29     skrll static pmap_pdetab_t *
    257  1.29     skrll pmap_pde_to_pdetab(pd_entry_t pde)
    258   1.1  christos {
    259  1.29     skrll 
    260  1.29     skrll 	return (pmap_pdetab_t *)pmap_pde_to_va(pde);
    261  1.29     skrll }
    262  1.29     skrll #endif
    263  1.29     skrll 
    264  1.29     skrll static pmap_ptpage_t *
    265  1.29     skrll pmap_pde_to_ptpage(pd_entry_t pde)
    266  1.29     skrll {
    267  1.29     skrll 
    268  1.29     skrll 	return (pmap_ptpage_t *)pmap_pde_to_va(pde);
    269  1.29     skrll }
    270  1.29     skrll #endif
    271  1.29     skrll 
    272  1.29     skrll #ifdef _LP64
    273  1.29     skrll __CTASSERT((XSEGSHIFT - SEGSHIFT) % (PGSHIFT-3) == 0);
    274  1.29     skrll #endif
    275  1.29     skrll 
    276  1.29     skrll static inline pmap_ptpage_t *
    277  1.29     skrll pmap_ptpage(struct pmap *pmap, vaddr_t va)
    278  1.29     skrll {
    279  1.29     skrll #if defined(PMAP_HWPAGEWALKER) && defined(PMAP_MAP_PDETABPAGE)
    280  1.29     skrll 	vaddr_t pdetab_mask = PMAP_PDETABSIZE - 1;
    281  1.29     skrll 	pmap_pdetab_t *ptb = pmap->pm_pdetab;
    282  1.29     skrll 
    283  1.29     skrll //	UVMHIST_LOG(pmaphist, "pm_pdetab %#jx", ptb, 0, 0, 0);
    284  1.29     skrll 
    285  1.29     skrll 	KASSERT(pmap != pmap_kernel() || !pmap_md_direct_mapped_vaddr_p(va));
    286  1.29     skrll 
    287  1.29     skrll #ifdef _LP64
    288  1.29     skrll 	for (size_t segshift = XSEGSHIFT;
    289  1.29     skrll 	    segshift > SEGSHIFT;
    290  1.29     skrll 	    segshift -= PGSHIFT - 3, pdetab_mask = NSEGPG - 1) {
    291  1.29     skrll 		ptb = pmap_pde_to_pdetab(ptb->pde_pde[(va >> segshift) & pdetab_mask]);
    292  1.29     skrll 		if (ptb == NULL)
    293  1.29     skrll 			return NULL;
    294  1.29     skrll 	}
    295  1.29     skrll #endif
    296  1.29     skrll 	return pmap_pde_to_ptpage(ptb->pde_pde[(va >> SEGSHIFT) & pdetab_mask]);
    297  1.29     skrll #else
    298  1.29     skrll 	vaddr_t segtab_mask = PMAP_SEGTABSIZE - 1;
    299  1.28     skrll 	pmap_segtab_t *stb = pmap->pm_segtab;
    300  1.29     skrll 
    301   1.6     skrll 	KASSERTMSG(pmap != pmap_kernel() || !pmap_md_direct_mapped_vaddr_p(va),
    302   1.6     skrll 	    "pmap %p va %#" PRIxVADDR, pmap, va);
    303   1.1  christos #ifdef _LP64
    304  1.29     skrll 	for (size_t segshift = XSEGSHIFT;
    305  1.29     skrll 	    segshift > SEGSHIFT;
    306  1.29     skrll 	    segshift -= PGSHIFT - 3, segtab_mask = NSEGPG - 1) {
    307  1.29     skrll 		stb = stb->seg_seg[(va >> segshift) & segtab_mask];
    308  1.29     skrll 		if (stb == NULL)
    309  1.29     skrll 			return NULL;
    310  1.29     skrll 	}
    311  1.29     skrll #endif
    312  1.29     skrll 	return stb->seg_ppg[(va >> SEGSHIFT) & segtab_mask];
    313  1.29     skrll #endif
    314  1.29     skrll }
    315  1.29     skrll 
    316  1.29     skrll #if defined(PMAP_HWPAGEWALKER)
    317  1.29     skrll bool
    318  1.29     skrll pmap_pdetab_fixup(struct pmap *pmap, vaddr_t va)
    319  1.29     skrll {
    320  1.29     skrll 	struct pmap * const kpm = pmap_kernel();
    321  1.29     skrll 	pmap_pdetab_t * const kptb = kpm->pm_pdetab;
    322  1.29     skrll 	pmap_pdetab_t * const uptb = pmap->pm_pdetab;
    323  1.29     skrll 	size_t idx = PMAP_PDETABSIZE - 1;
    324  1.29     skrll #if !defined(PMAP_MAP_PDETABPAGE)
    325  1.29     skrll 	__CTASSERT(PMAP_PDETABSIZE == PMAP_SEGTABSIZE);
    326  1.29     skrll 	pmap_segtab_t * const kstb = &pmap_kern_segtab;
    327  1.29     skrll 	pmap_segtab_t * const ustb = pmap->pm_segtab;
    328  1.29     skrll #endif
    329  1.29     skrll 
    330  1.29     skrll 	// Regardless of how many levels deep this page table is, we only
    331  1.29     skrll 	// need to verify the first level PDEs match up.
    332  1.29     skrll #ifdef XSEGSHIFT
    333  1.29     skrll 	idx &= va >> XSEGSHIFT;
    334  1.29     skrll #else
    335  1.29     skrll 	idx &= va >> SEGSHIFT;
    336  1.29     skrll #endif
    337  1.29     skrll 	if (uptb->pde_pde[idx] != kptb->pde_pde[idx]) {
    338  1.29     skrll 		pte_pde_set(&uptb->pde_pde[idx], kptb->pde_pde[idx]);
    339  1.29     skrll #if !defined(PMAP_MAP_PDETABPAGE)
    340  1.29     skrll 		ustb->seg_seg[idx] = kstb->seg_seg[idx]; // copy KVA of PTP
    341  1.29     skrll #endif
    342  1.29     skrll 		return true;
    343  1.29     skrll 	}
    344  1.29     skrll 	return false;
    345  1.29     skrll }
    346  1.29     skrll #endif /* PMAP_HWPAGEWALKER */
    347  1.29     skrll 
    348  1.29     skrll 
    349  1.29     skrll static void
    350  1.29     skrll pmap_page_attach(pmap_t pmap, vaddr_t kva, struct vm_page *pg,
    351  1.29     skrll     struct pglist *pglist, voff_t off)
    352  1.29     skrll {
    353  1.29     skrll 	UVMHIST_FUNC(__func__);
    354  1.29     skrll 	UVMHIST_CALLARGS(pmapxtabhist, "pm %#jx kva %#jx pg %#jx list %#jx",
    355  1.29     skrll 	    (uintptr_t)pmap, (uintptr_t)kva, (uintptr_t)pg, (uintptr_t)pglist);
    356  1.29     skrll 
    357  1.29     skrll 	struct uvm_object * const uobj = &pmap->pm_uobject;
    358  1.29     skrll 	if (pg == NULL) {
    359  1.29     skrll 		paddr_t pa;
    360  1.29     skrll 
    361  1.29     skrll 		bool ok __diagused = pmap_extract(pmap_kernel(), kva, &pa);
    362  1.29     skrll 		KASSERT(ok);
    363  1.29     skrll 
    364  1.29     skrll 		pg = PHYS_TO_VM_PAGE(pa);
    365  1.29     skrll 		KASSERT(pg != NULL);
    366  1.29     skrll 	}
    367  1.29     skrll 
    368  1.29     skrll 	UVMHIST_LOG(pmapxtabhist, "kva %#jx uobj %#jx pg %#jx list %#jx",
    369  1.29     skrll 	    (uintptr_t)kva, (uintptr_t)uobj, (uintptr_t)pg, (uintptr_t)pglist);
    370  1.29     skrll 
    371  1.29     skrll 	pmap_lock(pmap);
    372  1.29     skrll 	TAILQ_INSERT_TAIL(pglist, pg, pageq.queue);
    373  1.29     skrll 	uobj->uo_npages++;
    374  1.29     skrll 	pmap_unlock(pmap);
    375  1.29     skrll 
    376  1.29     skrll 	/*
    377  1.29     skrll 	 * Now set each vm_page that maps this page to point to the
    378  1.29     skrll 	 * pmap and set the offset to what we want.
    379  1.29     skrll 	 */
    380  1.29     skrll 	KASSERTMSG(pg->uobject == NULL, "pg %p pg->uobject %p", pg, pg->uobject);
    381  1.29     skrll 	pg->uobject = uobj;
    382  1.29     skrll 	pg->offset = off;
    383  1.29     skrll }
    384  1.29     skrll 
    385  1.29     skrll static struct vm_page *
    386  1.29     skrll pmap_page_detach(pmap_t pmap, struct pglist *list, vaddr_t va)
    387  1.29     skrll {
    388  1.29     skrll 	UVMHIST_FUNC(__func__);
    389  1.29     skrll 	UVMHIST_CALLARGS(pmapxtabhist, "pm %#jx kva %#jx list %#jx",
    390  1.29     skrll 	    (uintptr_t)pmap, (uintptr_t)va, (uintptr_t)list, 0);
    391  1.29     skrll 
    392  1.29     skrll 	paddr_t pa;
    393  1.29     skrll 	bool ok __diagused = pmap_extract(pmap_kernel(), va, &pa);
    394  1.29     skrll 	KASSERT(ok);
    395  1.29     skrll 
    396  1.29     skrll 	struct vm_page * const pg = PHYS_TO_VM_PAGE(pa);
    397  1.29     skrll 	struct uvm_object * const uobj = &pmap->pm_uobject;
    398  1.29     skrll 
    399  1.29     skrll 	UVMHIST_LOG(pmapxtabhist, "kva %#jx uobj %#jx pg %#jx list %#jx",
    400  1.29     skrll 	    (uintptr_t)va, (uintptr_t)uobj, (uintptr_t)pg, (uintptr_t)list);
    401  1.29     skrll 
    402  1.29     skrll 	KASSERTMSG(pg->uobject == uobj, "pg->uobject %p vs uobj %p",
    403  1.29     skrll 	    pg->uobject, uobj);
    404  1.29     skrll 
    405  1.29     skrll 	pmap_lock(pmap);
    406  1.29     skrll 	TAILQ_REMOVE(list, pg, pageq.queue);
    407  1.29     skrll 	uobj->uo_npages--;
    408  1.29     skrll 	pmap_unlock(pmap);
    409  1.29     skrll 
    410  1.29     skrll 	pg->uobject = NULL;
    411  1.29     skrll 	pg->offset = 0;
    412  1.29     skrll 
    413  1.29     skrll 	return pg;
    414  1.29     skrll }
    415  1.29     skrll 
    416  1.29     skrll #ifndef PMAP_PPG_CACHE
    417  1.29     skrll static void
    418  1.29     skrll pmap_segtab_pagefree(pmap_t pmap, struct pglist *list, vaddr_t kva, size_t size)
    419  1.29     skrll {
    420  1.29     skrll #ifdef PMAP_MAP_PTEPAGE
    421  1.29     skrll 	UVMHIST_FUNC(__func__);
    422  1.29     skrll 	UVMHIST_CALLARGS(pmapxtabhist, "pm %#jx list %#jx kva %#jx size %#jx",
    423  1.29     skrll 	    (uintptr_t)pmap, (uintptr_t)list, kva, size);
    424  1.29     skrll 	KASSERT(size == PAGE_SIZE);
    425  1.29     skrll 	if (size == PAGE_SIZE) {
    426  1.29     skrll 		UVMHIST_LOG(pmapxtabhist, "about to detach (kva %#jx)",
    427  1.29     skrll 		    kva, 0, 0, 0);
    428  1.29     skrll 		uvm_pagefree(pmap_page_detach(pmap, list, kva));
    429  1.29     skrll 		return;
    430  1.29     skrll 	}
    431   1.1  christos #endif
    432  1.29     skrll 	for (size_t i = 0; i < size; i += PAGE_SIZE) {
    433  1.29     skrll 		(void)pmap_page_detach(pmap, list, kva + i);
    434  1.29     skrll 	}
    435   1.1  christos 
    436  1.29     skrll 	uvm_km_free(kernel_map, kva, size, UVM_KMF_WIRED);
    437   1.1  christos }
    438  1.29     skrll #endif
    439   1.1  christos 
    440   1.1  christos pt_entry_t *
    441   1.1  christos pmap_pte_lookup(pmap_t pmap, vaddr_t va)
    442   1.1  christos {
    443  1.29     skrll 	pmap_ptpage_t * const ppg = pmap_ptpage(pmap, va);
    444  1.29     skrll 	if (ppg == NULL)
    445   1.1  christos 		return NULL;
    446   1.1  christos 
    447  1.29     skrll 	const size_t pte_idx = (va >> PGSHIFT) & (NPTEPG - 1);
    448  1.29     skrll 
    449  1.29     skrll 	return ppg->ppg_ptes + pte_idx;
    450   1.1  christos }
    451   1.1  christos 
    452  1.29     skrll 
    453  1.29     skrll static pmap_ptpage_t *
    454  1.29     skrll pmap_ptpage_alloc(pmap_t pmap, int flags, paddr_t *pa_p)
    455   1.1  christos {
    456  1.19       mrg 	UVMHIST_FUNC(__func__);
    457  1.29     skrll 	UVMHIST_CALLARGS(pmapxtabhist, "pm %#jx flags %#jx pa_p %#jx", (uintptr_t)pmap,
    458  1.29     skrll 	    (uintptr_t)flags, (uintptr_t)pa_p, 0);
    459  1.29     skrll 
    460  1.29     skrll 	pmap_ptpage_t *ppg = NULL;
    461  1.29     skrll 
    462  1.29     skrll #ifdef PMAP_MAP_PTEPAGE
    463  1.29     skrll 	struct vm_page *pg = NULL;
    464  1.29     skrll 	paddr_t pa;
    465  1.29     skrll #ifdef PMAP_PPG_CACHE
    466  1.29     skrll 	ppg = pmap_pgcache_alloc(&pmap_segtab_info.ppg_flist);
    467  1.29     skrll #endif
    468  1.29     skrll 	if (ppg == NULL) {
    469  1.29     skrll 		pg = pmap_pte_pagealloc();
    470  1.29     skrll 		if (pg == NULL) {
    471  1.29     skrll 			if (flags & PMAP_CANFAIL)
    472  1.29     skrll 				return NULL;
    473  1.29     skrll 			panic("%s: cannot allocate page table page ",
    474  1.29     skrll 			    __func__);
    475  1.29     skrll 		}
    476  1.29     skrll 		pa = VM_PAGE_TO_PHYS(pg);
    477  1.29     skrll 		ppg = (pmap_ptpage_t *)PMAP_MAP_PTEPAGE(pa);
    478  1.29     skrll 	} else {
    479  1.29     skrll 		bool ok __diagused = pmap_extract(pmap_kernel(), (vaddr_t)ppg, &pa);
    480  1.29     skrll 		KASSERT(ok);
    481  1.29     skrll 	}
    482  1.19       mrg 
    483  1.29     skrll 	UVMHIST_LOG(pmapxtabhist, "about to attach",  0, 0, 0, 0);
    484  1.29     skrll 	pmap_page_attach(pmap, (vaddr_t)ppg, pg, &pmap->pm_ppg_list, 0);
    485  1.19       mrg 
    486  1.29     skrll 	*pa_p = pa;
    487  1.29     skrll #else
    488  1.29     skrll 	vaddr_t kva = uvm_km_alloc(kernel_map, PAGE_SIZE, PAGE_SIZE,
    489  1.29     skrll 	    UVM_KMF_WIRED | UVM_KMF_WAITVA
    490  1.29     skrll 	    | (flags & PMAP_CANFAIL ? UVM_KMF_CANFAIL : 0));
    491  1.29     skrll 	if (kva == 0) {
    492  1.29     skrll 		if (flags & PMAP_CANFAIL)
    493  1.29     skrll 			return NULL;
    494  1.29     skrll 		panic("%s: cannot allocate page table page", __func__);
    495  1.29     skrll 	}
    496  1.29     skrll 	UVMHIST_LOG(pmapxtabhist, "about to attach",  0, 0, 0, 0);
    497  1.29     skrll 	pmap_page_attach(pmap, kva, NULL, &pmap->pm_ppg_list, 0);
    498  1.29     skrll 	ppg = (pmap_ptpage_t *)kva;
    499  1.29     skrll #endif
    500  1.29     skrll 
    501  1.29     skrll 	UVMHIST_LOG(pmapxtabhist, "... ppg %#jx", (uintptr_t)ppg, 0, 0, 0);
    502  1.29     skrll 
    503  1.29     skrll 	return ppg;
    504   1.1  christos }
    505   1.1  christos 
    506   1.1  christos static void
    507  1.29     skrll pmap_ptpage_free(pmap_t pmap, pmap_ptpage_t *ppg, const char *caller)
    508   1.1  christos {
    509  1.29     skrll 	UVMHIST_FUNC(__func__);
    510  1.29     skrll 	UVMHIST_CALLARGS(pmapxtabhist, "pm %#jx va %#jx", (uintptr_t)pmap,
    511  1.29     skrll 	    (uintptr_t)ppg, 0, 0);
    512  1.29     skrll 
    513  1.29     skrll 	const vaddr_t kva = (vaddr_t)ppg;
    514  1.29     skrll 	/*
    515  1.29     skrll 	 * All pte arrays should be page aligned.
    516  1.29     skrll 	 */
    517  1.29     skrll 	if ((kva & PAGE_MASK) != 0) {
    518  1.29     skrll 		panic("%s: pte entry at %p not page aligned", caller, ppg);
    519  1.29     skrll 	}
    520  1.29     skrll 
    521  1.29     skrll #ifdef DEBUG
    522  1.29     skrll 	for (size_t j = 0; j < NPTEPG; j++) {
    523  1.29     skrll 		if (ppg->ppg_ptes[j] != 0) {
    524  1.29     skrll 			UVMHIST_LOG(pmapxtabhist,
    525  1.29     skrll 			    "pte entry %#jx not 0 (%#jx)",
    526  1.29     skrll 			    (uintptr_t)&ppg->ppg_ptes[j],
    527  1.29     skrll 			    (uintptr_t)ppg->ppg_ptes[j], 0, 0);
    528  1.29     skrll 			for (size_t i = j + 1; i < NPTEPG; i++)
    529  1.29     skrll 				if (ppg->ppg_ptes[i] != 0)
    530  1.29     skrll 					UVMHIST_LOG(pmapxtabhist,
    531  1.29     skrll 					    "pte[%zu] = %#"PRIxPTE,
    532  1.29     skrll 					    i, ppg->ppg_ptes[i], 0, 0);
    533   1.1  christos 
    534  1.29     skrll 			panic("%s: pte entry at %p not 0 (%#" PRIxPTE ")",
    535  1.29     skrll 			    __func__, &ppg->ppg_ptes[j],
    536  1.29     skrll 			    ppg->ppg_ptes[j]);
    537   1.1  christos 		}
    538  1.29     skrll 	}
    539   1.1  christos #endif
    540  1.29     skrll 	//pmap_md_vca_clean(pg, (vaddr_t)ppg, NBPG);
    541  1.29     skrll #ifdef PMAP_PPG_CACHE
    542  1.29     skrll 	UVMHIST_LOG(pmapxtabhist, "about to detach",  0, 0, 0, 0);
    543  1.29     skrll 	pmap_page_detach(pmap, &pmap->pm_ppg_list, kva);
    544  1.29     skrll 	pmap_segtab_pagecache(&pmap_segtab_info.ppg_flist, ppg);
    545  1.29     skrll #else
    546  1.29     skrll 	pmap_segtab_pagefree(pmap, &pmap->pm_ppg_list, kva, PAGE_SIZE);
    547  1.29     skrll #endif /* PMAP_PPG_CACHE */
    548  1.29     skrll }
    549  1.29     skrll 
    550   1.1  christos 
    551  1.29     skrll #if defined(PMAP_HWPAGEWALKER) && defined(PMAP_MAP_PDETABPAGE)
    552  1.19       mrg 
    553  1.29     skrll static pmap_pdetab_t *
    554  1.29     skrll pmap_pdetab_alloc(struct pmap *pmap)
    555  1.29     skrll {
    556  1.29     skrll 	UVMHIST_FUNC(__func__);
    557  1.29     skrll 	UVMHIST_CALLARGS(pmapxtabhist, "pm %#jx", (uintptr_t)pmap, 0, 0, 0);
    558  1.19       mrg 
    559  1.29     skrll 	pmap_pdetab_t *ptb;
    560  1.29     skrll #ifdef KERNHIST
    561  1.29     skrll 	bool found_on_freelist = false;
    562   1.1  christos #endif
    563   1.1  christos 
    564  1.29     skrll  again:
    565  1.29     skrll 	mutex_spin_enter(&pmap_segtab_lock);
    566  1.29     skrll 	UVMHIST_LOG(pmapxtabhist, "free_pdetab %#jx",
    567  1.29     skrll 	    (uintptr_t)pmap_segtab_info.pdealloc.free_pdetab, 0, 0, 0);
    568  1.29     skrll 	if (__predict_true((ptb = pmap_segtab_info.pdealloc.free_pdetab) != NULL)) {
    569  1.29     skrll 		pmap_segtab_info.pdealloc.free_pdetab = ptb->pde_next;
    570  1.29     skrll 
    571  1.29     skrll 		UVMHIST_LOG(pmapxtabhist, "freelist ptb=%#jx",
    572  1.29     skrll 		    (uintptr_t)ptb, 0, 0, 0);
    573  1.29     skrll 
    574  1.29     skrll 		PDETAB_ADD(nget, 1);
    575  1.29     skrll 		ptb->pde_next = NULL;
    576  1.29     skrll #ifdef KERNHIST
    577  1.29     skrll 		found_on_freelist = true;
    578  1.29     skrll #endif
    579   1.1  christos 	}
    580  1.29     skrll 	mutex_spin_exit(&pmap_segtab_lock);
    581   1.1  christos 
    582  1.29     skrll 	struct vm_page *ptb_pg = NULL;
    583  1.29     skrll 	if (__predict_false(ptb == NULL)) {
    584  1.29     skrll 		ptb_pg = pmap_pte_pagealloc();
    585  1.29     skrll 
    586  1.29     skrll 		UVMHIST_LOG(pmapxtabhist, "ptb_pg=%#jx",
    587  1.29     skrll 		    (uintptr_t)ptb_pg, 0, 0, 0);
    588  1.29     skrll 		if (__predict_false(ptb_pg == NULL)) {
    589  1.29     skrll 			/*
    590  1.29     skrll 			 * XXX What else can we do?  Could we deadlock here?
    591  1.29     skrll 			 */
    592  1.29     skrll 			uvm_wait("pdetab");
    593  1.29     skrll 			goto again;
    594  1.29     skrll 		}
    595  1.29     skrll 
    596  1.29     skrll 		UVMHIST_LOG(pmapxtabhist, "ptb_pg=%#jx 2",
    597  1.29     skrll 		    (uintptr_t)ptb_pg, 0, 0, 0);
    598  1.29     skrll 		PDETAB_ADD(npage, 1);
    599  1.29     skrll 		const paddr_t ptb_pa = VM_PAGE_TO_PHYS(ptb_pg);
    600  1.29     skrll 		UVMHIST_LOG(pmapxtabhist, "ptb_pa=%#jx",  (uintptr_t)ptb_pa, 0, 0, 0);
    601  1.29     skrll 		ptb = (pmap_pdetab_t *)PMAP_MAP_PDETABPAGE(ptb_pa);
    602  1.29     skrll 		UVMHIST_LOG(pmapxtabhist, "new ptb=%#jx", (uintptr_t)ptb, 0,
    603  1.29     skrll 		    0, 0);
    604  1.29     skrll 
    605  1.29     skrll 		if (pte_invalid_pde() != 0) {
    606  1.29     skrll 			for (size_t i = 0; i < NPDEPG; i++) {
    607  1.29     skrll 				ptb->pde_pde[i] = pte_invalid_pde();
    608  1.29     skrll 			}
    609  1.29     skrll 		}
    610   1.1  christos 	}
    611  1.29     skrll 
    612  1.29     skrll 	UVMHIST_LOG(pmapxtabhist, "about to attach",  0, 0, 0, 0);
    613  1.29     skrll 	pmap_page_attach(pmap, (vaddr_t)ptb, ptb_pg, &pmap->pm_pdetab_list, 0);
    614  1.29     skrll 
    615  1.29     skrll 	UVMHIST_LOG(pmapxtabhist, "... ptb %#jx found on freelist %d",
    616  1.29     skrll 	    (uintptr_t)ptb, found_on_freelist, 0, 0);
    617  1.29     skrll 
    618  1.29     skrll 	return ptb;
    619   1.1  christos }
    620   1.1  christos 
    621  1.29     skrll 
    622  1.29     skrll #else
    623   1.1  christos /*
    624   1.1  christos  *	Create and return a physical map.
    625   1.1  christos  *
    626   1.1  christos  *	If the size specified for the map
    627   1.1  christos  *	is zero, the map is an actual physical
    628   1.1  christos  *	map, and may be referenced by the
    629   1.1  christos  *	hardware.
    630   1.1  christos  *
    631   1.1  christos  *	If the size specified is non-zero,
    632   1.1  christos  *	the map will be used in software only, and
    633   1.1  christos  *	is bounded by that size.
    634   1.1  christos  */
    635   1.1  christos static pmap_segtab_t *
    636  1.29     skrll pmap_segtab_alloc(struct pmap *pmap)
    637   1.1  christos {
    638  1.29     skrll 	UVMHIST_FUNC(__func__);
    639  1.29     skrll 	UVMHIST_CALLARGS(pmapxtabhist, "pm %#jx", (uintptr_t)pmap, 0, 0, 0);
    640  1.29     skrll 
    641  1.28     skrll 	pmap_segtab_t *stb;
    642   1.4       mrg 	bool found_on_freelist = false;
    643   1.1  christos 
    644   1.1  christos  again:
    645   1.1  christos 	mutex_spin_enter(&pmap_segtab_lock);
    646  1.29     skrll 	if (__predict_true((stb = pmap_segtab_info.segalloc.free_segtab) != NULL)) {
    647  1.29     skrll 		pmap_segtab_info.segalloc.free_segtab = stb->seg_next;
    648   1.1  christos 		SEGTAB_ADD(nget, 1);
    649  1.29     skrll 		stb->seg_next = NULL;
    650   1.4       mrg 		found_on_freelist = true;
    651  1.29     skrll 		UVMHIST_LOG(pmapxtabhist, "freelist stb=%#jx",
    652  1.28     skrll 		    (uintptr_t)stb, 0, 0, 0);
    653   1.1  christos 	}
    654   1.1  christos 	mutex_spin_exit(&pmap_segtab_lock);
    655   1.1  christos 
    656  1.29     skrll 	struct vm_page *stb_pg = NULL;
    657  1.28     skrll 	if (__predict_false(stb == NULL)) {
    658  1.29     skrll 		stb_pg = pmap_pte_pagealloc();
    659   1.1  christos 
    660  1.28     skrll 		if (__predict_false(stb_pg == NULL)) {
    661   1.1  christos 			/*
    662   1.1  christos 			 * XXX What else can we do?  Could we deadlock here?
    663   1.1  christos 			 */
    664  1.10     skrll 			uvm_wait("segtab");
    665   1.1  christos 			goto again;
    666   1.1  christos 		}
    667   1.1  christos 		SEGTAB_ADD(npage, 1);
    668  1.28     skrll 		const paddr_t stb_pa = VM_PAGE_TO_PHYS(stb_pg);
    669   1.1  christos 
    670  1.29     skrll 		stb = (pmap_segtab_t *)PMAP_MAP_SEGTABPAGE(stb_pa);
    671  1.29     skrll 		UVMHIST_LOG(pmapxtabhist, "new stb=%#jx", (uintptr_t)stb, 0,
    672  1.29     skrll 		    0, 0);
    673  1.29     skrll #if 0
    674  1.29     skrll CTASSERT(NBPG / sizeof(*stb) == 1);
    675  1.28     skrll 		const size_t n = NBPG / sizeof(*stb);
    676   1.1  christos 		if (n > 1) {
    677   1.1  christos 			/*
    678   1.1  christos 			 * link all the segtabs in this page together
    679   1.1  christos 			 */
    680   1.1  christos 			for (size_t i = 1; i < n - 1; i++) {
    681  1.29     skrll 				stb[i].seg_next = &stb[i + 1];
    682   1.1  christos 			}
    683   1.1  christos 			/*
    684   1.1  christos 			 * Now link the new segtabs into the free segtab list.
    685   1.1  christos 			 */
    686   1.1  christos 			mutex_spin_enter(&pmap_segtab_lock);
    687  1.29     skrll 			stb[n - 1].seg_next = pmap_segtab_info.segalloc.free_segtab;
    688  1.29     skrll 			pmap_segtab_info.segalloc.free_segtab = stb + 1;
    689   1.1  christos 			SEGTAB_ADD(nput, n - 1);
    690   1.1  christos 			mutex_spin_exit(&pmap_segtab_lock);
    691   1.1  christos 		}
    692  1.29     skrll #endif
    693   1.1  christos 	}
    694   1.1  christos 
    695  1.29     skrll 	UVMHIST_LOG(pmapxtabhist, "about to attach",  0, 0, 0, 0);
    696  1.29     skrll 	pmap_page_attach(pmap, (vaddr_t)stb, stb_pg, &pmap->pm_segtab_list, 0);
    697  1.29     skrll 
    698  1.28     skrll 	pmap_check_stb(stb, __func__,
    699  1.29     skrll 	    found_on_freelist ? "from free list" : "allocated");
    700  1.29     skrll 
    701  1.29     skrll 	UVMHIST_LOG(pmapxtabhist, "... stb %#jx found on freelist %zu",
    702  1.29     skrll 	    (uintptr_t)stb, found_on_freelist, 0, 0);
    703   1.4       mrg 
    704  1.28     skrll 	return stb;
    705   1.1  christos }
    706  1.29     skrll #endif
    707  1.29     skrll 
    708  1.29     skrll #if defined(PMAP_HWPAGEWALKER)
    709  1.29     skrll static void
    710  1.29     skrll pmap_pdetab_free(pmap_pdetab_t *ptb)
    711  1.29     skrll {
    712  1.29     skrll 	UVMHIST_FUNC(__func__);
    713  1.29     skrll 	UVMHIST_CALLARGS(pmaphist, "ptb %#jx", (uintptr_t)ptb, 0, 0, 0);
    714  1.29     skrll 	/*
    715  1.29     skrll 	 * Insert the pdetab into the pdetab freelist.
    716  1.29     skrll 	 */
    717  1.29     skrll 	mutex_spin_enter(&pmap_segtab_lock);
    718  1.29     skrll 	ptb->pde_next = pmap_segtab_info.pdealloc.free_pdetab;
    719  1.29     skrll 	pmap_segtab_info.pdealloc.free_pdetab = ptb;
    720  1.29     skrll 	PDETAB_ADD(nput, 1);
    721  1.29     skrll 	mutex_spin_exit(&pmap_segtab_lock);
    722  1.29     skrll 
    723  1.29     skrll }
    724  1.29     skrll #endif
    725  1.29     skrll 
    726  1.29     skrll 
    727  1.29     skrll #if !defined(PMAP_HWPAGEWALKER) || !defined(PMAP_MAP_PDETABPAGE)
    728  1.29     skrll /*
    729  1.29     skrll  * Insert the segtab into the segtab freelist.
    730  1.29     skrll  */
    731  1.29     skrll static void
    732  1.29     skrll pmap_segtab_free(pmap_segtab_t *stb)
    733  1.29     skrll {
    734  1.29     skrll 	UVMHIST_FUNC(__func__);
    735  1.29     skrll 	UVMHIST_CALLARGS(pmaphist, "stb %#jx", (uintptr_t)stb, 0, 0, 0);
    736  1.29     skrll 
    737  1.29     skrll 	/*
    738  1.29     skrll 	 * Insert the segtab into the segtab freelist.
    739  1.29     skrll 	 */
    740  1.29     skrll 	mutex_spin_enter(&pmap_segtab_lock);
    741  1.29     skrll 	stb->seg_next = pmap_segtab_info.segalloc.free_segtab;
    742  1.29     skrll 	pmap_segtab_info.segalloc.free_segtab = stb;
    743  1.29     skrll 	SEGTAB_ADD(nput, 1);
    744  1.29     skrll 	mutex_spin_exit(&pmap_segtab_lock);
    745  1.29     skrll }
    746  1.29     skrll #endif
    747  1.29     skrll 
    748  1.29     skrll #if defined(PMAP_HWPAGEWALKER)
    749  1.29     skrll static void
    750  1.29     skrll pmap_pdetab_release(pmap_t pmap, pmap_pdetab_t **ptb_p, bool free_ptb,
    751  1.29     skrll     vaddr_t va, vsize_t vinc)
    752  1.29     skrll {
    753  1.29     skrll 	const vaddr_t pdetab_mask = PMAP_PDETABSIZE - 1;
    754  1.29     skrll 	pmap_pdetab_t *ptb = *ptb_p;
    755  1.29     skrll 
    756  1.29     skrll 	UVMHIST_FUNC(__func__);
    757  1.29     skrll 	UVMHIST_CALLARGS(pmapxtabhist, "pm %#jx ptb_p %#jx ptb %#jx free %jd",
    758  1.29     skrll 	    (uintptr_t)pmap, (uintptr_t)ptb_p, (uintptr_t)ptb, free_ptb);
    759  1.29     skrll 	UVMHIST_LOG(pmapxtabhist, " va=%#jx vinc=%#jx",
    760  1.29     skrll 	    (uintptr_t)va, (uintptr_t)vinc, 0, 0);
    761  1.29     skrll 
    762  1.29     skrll 	for (size_t i = (va / vinc) & pdetab_mask;
    763  1.29     skrll 	    i < PMAP_PDETABSIZE;
    764  1.29     skrll 	    i++, va += vinc) {
    765  1.29     skrll #ifdef _LP64
    766  1.29     skrll 		if (vinc > NBSEG) {
    767  1.29     skrll 			if (pte_pde_valid_p(ptb->pde_pde[i])) {
    768  1.29     skrll 				pmap_pdetab_t *nptb =
    769  1.29     skrll 				    pmap_pde_to_pdetab(ptb->pde_pde[i]);
    770  1.29     skrll 				UVMHIST_LOG(pmapxtabhist,
    771  1.29     skrll 				    " va %#jx ptp->pde_pde[%jd] (*%#jx) = %#jx "
    772  1.29     skrll 				    "recursing", va, i, &ptb->pde_pde[i],
    773  1.29     skrll 				    ptb->pde_pde[i]);
    774  1.29     skrll 				pmap_pdetab_release(pmap, &nptb, true,
    775  1.29     skrll 				    va, vinc / NPDEPG);
    776  1.29     skrll 				ptb->pde_pde[i] = pte_invalid_pde();
    777  1.29     skrll 				KASSERT(nptb == NULL);
    778  1.29     skrll 			}
    779  1.29     skrll 			continue;
    780  1.29     skrll 		}
    781  1.29     skrll #endif
    782  1.29     skrll 		KASSERT(vinc == NBSEG);
    783  1.29     skrll 
    784  1.29     skrll 		/* get pointer to PT page */
    785  1.29     skrll 		pmap_ptpage_t *ppg = pmap_pde_to_ptpage(ptb->pde_pde[i]);
    786  1.29     skrll 		UVMHIST_LOG(pmapxtabhist,
    787  1.29     skrll 		    "   va %#jx ptb->pde_pde[%jd] (*%#jx) = %#jx", va, i,
    788  1.29     skrll 		    (uintptr_t)&ptb->pde_pde[i], ptb->pde_pde[i]);
    789  1.29     skrll 		if (ppg == NULL)
    790  1.29     skrll 			continue;
    791  1.29     skrll 
    792  1.29     skrll 		UVMHIST_LOG(pmapxtabhist, " zeroing tab (%#jx)[%jd] (%#jx)",
    793  1.29     skrll 		    (uintptr_t)ptb->pde_pde, i, (uintptr_t)&ptb->pde_pde[i], 0);
    794  1.29     skrll 
    795  1.29     skrll 		ptb->pde_pde[i] = pte_invalid_pde();
    796  1.29     skrll 
    797  1.29     skrll 		pmap_ptpage_free(pmap, ppg, __func__);
    798  1.29     skrll 	}
    799  1.29     skrll 
    800  1.29     skrll 	if (free_ptb) {
    801  1.29     skrll 		UVMHIST_LOG(pmapxtabhist, " ptbp %#jx ptb %#jx",
    802  1.29     skrll 		    (uintptr_t)ptb_p, (uintptr_t)ptb, 0, 0);
    803  1.29     skrll 		const vaddr_t kva = (vaddr_t)ptb;
    804  1.29     skrll 		UVMHIST_LOG(pmapxtabhist, "about to detach",  0, 0, 0, 0);
    805  1.29     skrll 		pmap_page_detach(pmap, &pmap->pm_pdetab_list, kva);
    806  1.29     skrll 		pmap_pdetab_free(ptb);
    807  1.29     skrll 		*ptb_p = NULL;
    808  1.29     skrll 	}
    809  1.29     skrll }
    810  1.29     skrll #endif
    811  1.29     skrll 
    812  1.29     skrll #if !defined(PMAP_HWPAGEWALKER) || !defined(PMAP_MAP_PDETABPAGE)
    813  1.29     skrll static void
    814  1.29     skrll pmap_segtab_release(pmap_t pmap, pmap_segtab_t **stb_p, bool free_stb,
    815  1.29     skrll     pte_callback_t callback, uintptr_t flags, vaddr_t va, vsize_t vinc)
    816  1.29     skrll {
    817  1.29     skrll 	pmap_segtab_t *stb = *stb_p;
    818  1.29     skrll 
    819  1.29     skrll 	UVMHIST_FUNC(__func__);
    820  1.29     skrll 	UVMHIST_CALLARGS(pmapxtabhist, "pm=%#jx stb_p=%#jx free=%jd",
    821  1.29     skrll 	    (uintptr_t)pmap, (uintptr_t)stb, free_stb, 0);
    822  1.29     skrll 	UVMHIST_LOG(pmapxtabhist, " callback=%#jx flags=%#jx va=%#jx vinc=%#jx",
    823  1.29     skrll 	    (uintptr_t)callback, flags, (uintptr_t)va, (uintptr_t)vinc);
    824  1.29     skrll 
    825  1.29     skrll 	for (size_t i = (va / vinc) & (PMAP_SEGTABSIZE - 1);
    826  1.29     skrll 	    i < PMAP_SEGTABSIZE;
    827  1.29     skrll 	    i++, va += vinc) {
    828  1.29     skrll #ifdef _LP64
    829  1.29     skrll 		if (vinc > NBSEG) {
    830  1.29     skrll 			if (stb->seg_seg[i] != NULL) {
    831  1.29     skrll 				UVMHIST_LOG(pmapxtabhist,
    832  1.29     skrll 				    " recursing %jd", i, 0, 0, 0);
    833  1.29     skrll 				pmap_segtab_release(pmap, &stb->seg_seg[i],
    834  1.29     skrll 				    true, callback, flags, va, vinc / NSEGPG);
    835  1.29     skrll 				KASSERT(stb->seg_seg[i] == NULL);
    836  1.29     skrll 			}
    837  1.29     skrll 			continue;
    838  1.29     skrll 		}
    839  1.29     skrll #endif
    840  1.29     skrll 		KASSERT(vinc == NBSEG);
    841  1.29     skrll 
    842  1.29     skrll 		/* get pointer to segment map */
    843  1.29     skrll 		pmap_ptpage_t *ppg = stb->seg_ppg[i];
    844  1.29     skrll 		if (ppg == NULL)
    845  1.29     skrll 			continue;
    846  1.29     skrll 
    847  1.29     skrll 		/*
    848  1.29     skrll 		 * If our caller wants a callback, do so.
    849  1.29     skrll 		 */
    850  1.29     skrll 		if (callback != NULL) {
    851  1.29     skrll 			(*callback)(pmap, va, va + vinc, ppg->ppg_ptes, flags);
    852  1.29     skrll 		}
    853  1.29     skrll 		pmap_ptpage_free(pmap, ppg, __func__);
    854  1.29     skrll 		stb->seg_ppg[i] = NULL;
    855  1.29     skrll 		UVMHIST_LOG(pmapxtabhist, " zeroing tab[%jd]", i, 0, 0, 0);
    856  1.29     skrll 	}
    857  1.29     skrll 
    858  1.29     skrll 	if (free_stb) {
    859  1.29     skrll 		pmap_check_stb(stb, __func__,
    860  1.29     skrll 		    vinc == NBSEG ? "release seg" : "release xseg");
    861  1.29     skrll 
    862  1.29     skrll 		const vaddr_t kva = (vaddr_t)stb;
    863  1.29     skrll 		UVMHIST_LOG(pmapxtabhist, "about to detach",  0, 0, 0, 0);
    864  1.29     skrll 		pmap_page_detach(pmap, &pmap->pm_segtab_list, kva);
    865  1.29     skrll 		pmap_segtab_free(stb);
    866  1.29     skrll 		*stb_p = NULL;
    867  1.29     skrll 	}
    868  1.29     skrll }
    869  1.29     skrll #endif
    870  1.29     skrll 
    871  1.29     skrll 
    872   1.1  christos 
    873   1.1  christos /*
    874   1.1  christos  * Allocate the top segment table for the pmap.
    875   1.1  christos  */
    876   1.1  christos void
    877   1.1  christos pmap_segtab_init(pmap_t pmap)
    878   1.1  christos {
    879  1.29     skrll 	UVMHIST_FUNC(__func__);
    880  1.29     skrll 	UVMHIST_CALLARGS(pmaphist, "pm %#jx", (uintptr_t)pmap, 0, 0, 0);
    881   1.1  christos 
    882  1.29     skrll #if !defined(PMAP_HWPAGEWALKER) || !defined(PMAP_MAP_PDETABPAGE)
    883  1.29     skrll 	/*
    884  1.29     skrll 	 * Constantly converting from extracted PA to VA is somewhat expensive
    885  1.29     skrll 	 * for systems with hardware page walkers and without an inexpensive
    886  1.29     skrll 	 * way to access arbitrary virtual addresses, so we allocate an extra
    887  1.29     skrll 	 * root segtab so that it can contain non-virtual addresses.
    888  1.29     skrll 	 */
    889  1.29     skrll 	pmap->pm_segtab = pmap_segtab_alloc(pmap);
    890  1.29     skrll #endif
    891  1.29     skrll #if defined(PMAP_HWPAGEWALKER)
    892  1.29     skrll 	pmap->pm_pdetab = pmap_pdetab_alloc(pmap);
    893  1.29     skrll 	pmap_md_pdetab_init(pmap);
    894  1.29     skrll #endif
    895   1.1  christos }
    896   1.1  christos 
    897   1.1  christos /*
    898   1.1  christos  *	Retire the given physical map from service.
    899   1.1  christos  *	Should only be called if the map contains
    900   1.1  christos  *	no valid mappings.
    901   1.1  christos  */
    902   1.1  christos void
    903   1.1  christos pmap_segtab_destroy(pmap_t pmap, pte_callback_t func, uintptr_t flags)
    904   1.1  christos {
    905  1.29     skrll 	KASSERT(pmap != pmap_kernel());
    906   1.1  christos #ifdef _LP64
    907   1.1  christos 	const vsize_t vinc = NBXSEG;
    908   1.1  christos #else
    909   1.1  christos 	const vsize_t vinc = NBSEG;
    910   1.1  christos #endif
    911  1.29     skrll 
    912  1.29     skrll #if defined(PMAP_HWPAGEWALKER)
    913  1.29     skrll 	if (pmap->pm_pdetab != NULL) {
    914  1.29     skrll 		pmap_pdetab_release(pmap, &pmap->pm_pdetab,
    915  1.29     skrll 		    true, pmap->pm_minaddr, vinc);
    916  1.29     skrll 	}
    917  1.29     skrll #endif
    918  1.29     skrll #if !defined(PMAP_HWPAGEWALKER) || !defined(PMAP_MAP_PDETABPAGE)
    919  1.29     skrll 	if (pmap->pm_segtab != NULL) {
    920  1.29     skrll 		pmap_segtab_release(pmap, &pmap->pm_segtab,
    921  1.29     skrll 		    func == NULL, func, flags, pmap->pm_minaddr, vinc);
    922  1.29     skrll 	}
    923  1.29     skrll #endif
    924  1.29     skrll 
    925  1.29     skrll #if defined(PMAP_HWPAGEWALKER)
    926  1.29     skrll #if !defined(PMAP_MAP_PDETABPAGE)
    927  1.29     skrll 	KASSERT((pmap->pm_segtab == NULL) == (pmap->pm_pdetab == NULL));
    928  1.29     skrll #endif
    929  1.29     skrll 	KASSERT(pmap->pm_pdetab == NULL);
    930  1.29     skrll #endif
    931  1.29     skrll #if !defined(PMAP_HWPAGEWALKER) || !defined(PMAP_MAP_PDETABPAGE)
    932  1.29     skrll 	KASSERT(pmap->pm_segtab == NULL);
    933  1.29     skrll #endif
    934  1.29     skrll 
    935   1.1  christos }
    936   1.1  christos 
    937   1.1  christos /*
    938   1.1  christos  *	Make a new pmap (vmspace) active for the given process.
    939   1.1  christos  */
    940   1.1  christos void
    941   1.1  christos pmap_segtab_activate(struct pmap *pm, struct lwp *l)
    942   1.1  christos {
    943   1.1  christos 	if (l == curlwp) {
    944  1.29     skrll 		KASSERT(pm == l->l_proc->p_vmspace->vm_map.pmap);
    945  1.29     skrll 		pmap_md_xtab_activate(pm, l);
    946  1.29     skrll #if !defined(PMAP_HWPAGEWALKER) || !defined(PMAP_MAP_PDETABPAGE)
    947   1.3      matt 		struct cpu_info * const ci = l->l_cpu;
    948   1.1  christos 		if (pm == pmap_kernel()) {
    949   1.3      matt 			ci->ci_pmap_user_segtab = PMAP_INVALID_SEGTAB_ADDRESS;
    950   1.1  christos #ifdef _LP64
    951   1.3      matt 			ci->ci_pmap_user_seg0tab = PMAP_INVALID_SEGTAB_ADDRESS;
    952   1.1  christos #endif
    953   1.1  christos 		} else {
    954   1.3      matt 			ci->ci_pmap_user_segtab = pm->pm_segtab;
    955   1.1  christos #ifdef _LP64
    956   1.3      matt 			ci->ci_pmap_user_seg0tab = pm->pm_segtab->seg_seg[0];
    957   1.1  christos #endif
    958   1.1  christos 		}
    959  1.29     skrll #endif
    960   1.1  christos 	}
    961   1.1  christos }
    962   1.1  christos 
    963  1.15     skrll void
    964  1.15     skrll pmap_segtab_deactivate(pmap_t pm)
    965  1.15     skrll {
    966  1.15     skrll 	pmap_md_xtab_deactivate(pm);
    967  1.15     skrll 
    968  1.29     skrll #if !defined(PMAP_HWPAGEWALKER) || !defined(PMAP_MAP_PDETABPAGE)
    969  1.15     skrll 	curcpu()->ci_pmap_user_segtab = PMAP_INVALID_SEGTAB_ADDRESS;
    970  1.15     skrll #ifdef _LP64
    971  1.15     skrll 	curcpu()->ci_pmap_user_seg0tab = NULL;
    972  1.15     skrll #endif
    973  1.29     skrll #endif
    974  1.15     skrll }
    975  1.15     skrll 
    976   1.1  christos /*
    977   1.1  christos  *	Act on the given range of addresses from the specified map.
    978   1.1  christos  *
    979   1.1  christos  *	It is assumed that the start and end are properly rounded to
    980   1.1  christos  *	the page size.
    981   1.1  christos  */
    982   1.1  christos void
    983   1.1  christos pmap_pte_process(pmap_t pmap, vaddr_t sva, vaddr_t eva,
    984   1.8     skrll     pte_callback_t callback, uintptr_t flags)
    985   1.1  christos {
    986   1.1  christos #if 0
    987   1.1  christos 	printf("%s: %p, %"PRIxVADDR", %"PRIxVADDR", %p, %"PRIxPTR"\n",
    988   1.1  christos 	    __func__, pmap, sva, eva, callback, flags);
    989   1.1  christos #endif
    990   1.1  christos 	while (sva < eva) {
    991   1.1  christos 		vaddr_t lastseg_va = pmap_trunc_seg(sva) + NBSEG;
    992   1.1  christos 		if (lastseg_va == 0 || lastseg_va > eva)
    993   1.1  christos 			lastseg_va = eva;
    994   1.1  christos 
    995   1.1  christos 		/*
    996   1.1  christos 		 * If VA belongs to an unallocated segment,
    997   1.1  christos 		 * skip to the next segment boundary.
    998   1.1  christos 		 */
    999   1.9     skrll 		pt_entry_t * const ptep = pmap_pte_lookup(pmap, sva);
   1000   1.9     skrll 		if (ptep != NULL) {
   1001   1.1  christos 			/*
   1002   1.1  christos 			 * Callback to deal with the ptes for this segment.
   1003   1.1  christos 			 */
   1004   1.9     skrll 			(*callback)(pmap, sva, lastseg_va, ptep, flags);
   1005   1.1  christos 		}
   1006   1.1  christos 		/*
   1007   1.1  christos 		 * In theory we could release pages with no entries,
   1008   1.1  christos 		 * but that takes more effort than we want here.
   1009   1.1  christos 		 */
   1010   1.1  christos 		sva = lastseg_va;
   1011   1.1  christos 	}
   1012   1.1  christos }
   1013   1.1  christos 
   1014  1.29     skrll #if defined(PMAP_HWPAGEWALKER) && defined(PMAP_MAP_PDETABPAGE)
   1015  1.29     skrll static pd_entry_t *
   1016  1.29     skrll pmap_pdetab_reserve(struct pmap *pmap, vaddr_t va)
   1017  1.29     skrll #elif defined(PMAP_HWPAGEWALKER)
   1018  1.29     skrll static pmap_ptpage_t **
   1019  1.29     skrll pmap_segtab_reserve(struct pmap *pmap, vaddr_t va, pd_entry_t **pde_p)
   1020  1.29     skrll #else
   1021  1.29     skrll static pmap_ptpage_t **
   1022  1.29     skrll pmap_segtab_reserve(struct pmap *pmap, vaddr_t va)
   1023  1.29     skrll #endif
   1024   1.1  christos {
   1025  1.16       mrg 	UVMHIST_FUNC(__func__);
   1026  1.29     skrll 	UVMHIST_CALLARGS(pmaphist, "pm %#jx va %#jx", (uintptr_t)pmap,
   1027  1.29     skrll 	    (uintptr_t)va, 0, 0);
   1028  1.29     skrll 
   1029  1.29     skrll #if defined(PMAP_HWPAGEWALKER)
   1030  1.29     skrll 	pmap_pdetab_t *ptb = pmap->pm_pdetab;
   1031  1.29     skrll 	UVMHIST_LOG(pmaphist, "pm_pdetab %#jx", (uintptr_t)ptb, 0, 0, 0);
   1032  1.29     skrll #endif
   1033  1.29     skrll #if defined(PMAP_HWPAGEWALKER) && defined(PMAP_MAP_PDETABPAGE)
   1034  1.29     skrll 	vaddr_t segtab_mask = PMAP_PDETABSIZE - 1;
   1035  1.29     skrll #ifdef _LP64
   1036  1.29     skrll 	for (size_t segshift = XSEGSHIFT;
   1037  1.29     skrll 	    segshift > SEGSHIFT;
   1038  1.29     skrll 	    segshift -= PGSHIFT - 3, segtab_mask = NSEGPG - 1) {
   1039  1.29     skrll 		pd_entry_t * const pde_p =
   1040  1.29     skrll 		    &ptb->pde_pde[(va >> segshift) & segtab_mask];
   1041  1.29     skrll 		pd_entry_t opde = *pde_p;
   1042  1.29     skrll 
   1043  1.29     skrll 		UVMHIST_LOG(pmaphist,
   1044  1.29     skrll 		    "ptb %#jx segshift %jd pde_p %#jx opde %#jx",
   1045  1.29     skrll 		    ptb, segshift, pde_p, opde);
   1046  1.29     skrll 
   1047  1.29     skrll 		if (__predict_false(!pte_pde_valid_p(opde))) {
   1048  1.29     skrll 			ptb = pmap_pdetab_alloc(pmap);
   1049  1.29     skrll 			pd_entry_t npde = pte_pde_pdetab(
   1050  1.29     skrll 			    pmap_md_direct_mapped_vaddr_to_paddr((vaddr_t)ptb),
   1051  1.29     skrll 			    pmap == pmap_kernel());
   1052  1.29     skrll 			opde = pte_pde_cas(pde_p, opde, npde);
   1053  1.29     skrll 			if (__predict_false(pte_pde_valid_p(opde))) {
   1054  1.29     skrll 				const vaddr_t kva = (vaddr_t)ptb;
   1055  1.29     skrll 				UVMHIST_LOG(pmapxtabhist, "about to detach",
   1056  1.29     skrll 				    0, 0, 0, 0);
   1057  1.29     skrll 				pmap_page_detach(pmap, &pmap->pm_pdetab_list,
   1058  1.29     skrll 				    kva);
   1059  1.29     skrll 				pmap_pdetab_free(ptb);
   1060  1.29     skrll 			} else {
   1061  1.29     skrll 				opde = npde;
   1062  1.29     skrll 			}
   1063  1.29     skrll 		}
   1064  1.29     skrll 		ptb = pmap_pde_to_pdetab(opde);
   1065  1.29     skrll 		UVMHIST_LOG(pmaphist, "opde %#jx ptb %#jx", opde, ptb, 0, 0);
   1066  1.29     skrll 	}
   1067  1.29     skrll #elif defined(XSEGSHIFT)
   1068  1.29     skrll 	size_t segshift = XSEGSHIFT;
   1069  1.29     skrll 
   1070  1.29     skrll 	pd_entry_t opde = ptb->pde_pde[(va >> segshift) & segtab_mask];
   1071  1.29     skrll 	KASSERT(pte_pde_valid_p(opde));
   1072  1.29     skrll 	ptb = pmap_pde_to_pdetab(opde);
   1073  1.29     skrll 	segtab_mask = NSEGPG - 1;
   1074  1.29     skrll #endif /* _LP64 */
   1075  1.29     skrll 	const size_t idx = (va >> SEGSHIFT) & segtab_mask;
   1076  1.16       mrg 
   1077  1.29     skrll 	UVMHIST_LOG(pmaphist, "... returning %#jx (idx %jd)", (uintptr_t)&ptb->pde_pde[idx], idx, 0, 0);
   1078  1.29     skrll 
   1079  1.29     skrll 	return &ptb->pde_pde[idx];
   1080  1.29     skrll #else /* PMAP_HWPAGEWALKER && PMAP_MAP_PDETABPAGE */
   1081  1.29     skrll 	pmap_segtab_t *stb = pmap->pm_segtab;
   1082  1.29     skrll 	vaddr_t segtab_mask = PMAP_SEGTABSIZE - 1;
   1083   1.1  christos #ifdef _LP64
   1084  1.29     skrll 	for (size_t segshift = XSEGSHIFT;
   1085  1.29     skrll 	    segshift > SEGSHIFT;
   1086  1.29     skrll 	    segshift -= PGSHIFT - 3, segtab_mask = NSEGPG - 1) {
   1087  1.29     skrll 		size_t idx = (va >> segshift) & segtab_mask;
   1088  1.29     skrll 		pmap_segtab_t ** const stb_p = &stb->seg_seg[idx];
   1089  1.29     skrll #if defined(PMAP_HWPAGEWALKER)
   1090  1.29     skrll 		pmap_pdetab_t ** const ptb_p = &ptb->pde_pde[idx];
   1091  1.29     skrll #endif	/* PMAP_HWPAGEWALKER */
   1092  1.28     skrll 		if (__predict_false((stb = *stb_p) == NULL)) {
   1093  1.29     skrll 			stb = pmap_segtab_alloc(pmap);
   1094   1.1  christos #ifdef MULTIPROCESSOR
   1095  1.29     skrll 			pmap_segtab_t *ostb = atomic_cas_ptr(stb_p, NULL, stb);
   1096  1.28     skrll 			if (__predict_false(ostb != NULL)) {
   1097  1.29     skrll 				const vaddr_t kva = (vaddr_t)stb;
   1098  1.29     skrll 				UVMHIST_LOG(pmapxtabhist, "about to detach",
   1099  1.29     skrll 				    0, 0, 0, 0);
   1100  1.29     skrll 				pmap_page_detach(pmap, &pmap->pm_segtab_list,
   1101  1.29     skrll 				    kva);
   1102  1.29     skrll 				pmap_segtab_free(stb);
   1103  1.29     skrll 				stb = ostb;
   1104   1.1  christos 			}
   1105   1.1  christos #else
   1106  1.29     skrll 			*stb_p = stb;
   1107   1.1  christos #endif /* MULTIPROCESSOR */
   1108   1.1  christos 		}
   1109  1.29     skrll 	}
   1110  1.29     skrll #elif defined(PMAP_HWPAGEWALKER)
   1111  1.29     skrll 	pmap_segtab_t opde = ptb->pde_pde[(va >> segshift) & segtab_mask];
   1112  1.29     skrll 	KASSERT(pte_pde_valid_p(opde));
   1113  1.29     skrll 	ptb = pmap_pde_to_pdetab(opde);
   1114  1.29     skrll 	segtab_mask = NSEGPG - 1;
   1115  1.29     skrll 
   1116   1.1  christos #endif /* _LP64 */
   1117  1.29     skrll 	size_t idx = (va >> SEGSHIFT) & segtab_mask;
   1118  1.29     skrll #if defined(PMAP_HWPAGEWALKER)
   1119  1.29     skrll #if defined(XSEGSHIFT) && (XSEGSHIFT != SEGSHIFT)
   1120  1.29     skrll 	*pte_p = &pmap->pm_segtab
   1121  1.29     skrll #else /* XSEGSHIFT */
   1122  1.29     skrll 	*pde_p = &ptb->pde_pde[idx];
   1123  1.29     skrll #endif /* XSEGSHIFT */
   1124  1.29     skrll #endif /* PMAP_HWPAGEWALKER */
   1125  1.29     skrll 	return &stb->seg_ppg[idx];
   1126  1.29     skrll #endif
   1127  1.29     skrll }
   1128  1.29     skrll 
   1129  1.29     skrll 
   1130  1.29     skrll /*
   1131  1.29     skrll  *	Return a pointer for the pte that corresponds to the specified virtual
   1132  1.29     skrll  *	address (va) in the target physical map, allocating if needed.
   1133  1.29     skrll  */
   1134  1.29     skrll pt_entry_t *
   1135  1.29     skrll pmap_pte_reserve(pmap_t pmap, vaddr_t va, int flags)
   1136  1.29     skrll {
   1137  1.29     skrll 	UVMHIST_FUNC(__func__);
   1138  1.29     skrll 	UVMHIST_CALLARGS(pmaphist, "pm=%#jx va=%#jx flags=%#jx",
   1139  1.29     skrll 	    (uintptr_t)pmap, (uintptr_t)va, flags, 0);
   1140  1.29     skrll 	pmap_ptpage_t *ppg;
   1141  1.29     skrll 	paddr_t pa = 0;
   1142  1.29     skrll 
   1143  1.29     skrll #if defined(PMAP_HWPAGEWALKER) && defined(PMAP_MAP_PDETABPAGE)
   1144  1.29     skrll 	pd_entry_t * const pde_p = pmap_pdetab_reserve(pmap, va);
   1145  1.29     skrll 	ppg = pmap_pde_to_ptpage(*pde_p);
   1146  1.29     skrll #elif defined(PMAP_HWPAGEWALKER)
   1147  1.29     skrll 	pd_entry_t *pde_p;
   1148  1.29     skrll 	pmap_ptpage_t ** const ppg_p = pmap_segtab_reserve(pmap, va, &pde_p);
   1149  1.29     skrll 	ppg = *ppg_p;
   1150  1.29     skrll #else
   1151  1.29     skrll 	pmap_ptpage_t ** const ppg_p = pmap_segtab_reserve(pmap, va);
   1152  1.29     skrll 	ppg = *ppg_p;
   1153  1.29     skrll #endif
   1154  1.29     skrll 
   1155  1.29     skrll 	if (__predict_false(ppg == NULL)) {
   1156  1.29     skrll 		ppg = pmap_ptpage_alloc(pmap, flags, &pa);
   1157  1.29     skrll 		if (__predict_false(ppg == NULL))
   1158  1.29     skrll 			return NULL;
   1159  1.29     skrll 
   1160  1.29     skrll #if defined(PMAP_HWPAGEWALKER)
   1161  1.29     skrll 		pd_entry_t npde = pte_pde_ptpage(pa, pmap == pmap_kernel());
   1162   1.1  christos #endif
   1163  1.29     skrll #if defined(PMAP_HWPAGEWALKER) && defined(PMAP_MAP_PDETABPAGE)
   1164  1.29     skrll 		pd_entry_t opde = *pde_p;
   1165  1.29     skrll 		opde = pte_pde_cas(pde_p, opde, npde);
   1166  1.29     skrll 		if (__predict_false(pte_pde_valid_p(opde))) {
   1167  1.29     skrll 			pmap_ptpage_free(pmap, ppg, __func__);
   1168  1.29     skrll 			ppg = pmap_pde_to_ptpage(opde);
   1169   1.1  christos 		}
   1170  1.29     skrll #else
   1171   1.1  christos #ifdef MULTIPROCESSOR
   1172  1.29     skrll 		pmap_ptpage_t *oppg = atomic_cas_ptr(ppg_p, NULL, ppg);
   1173   1.1  christos 		/*
   1174   1.1  christos 		 * If another thread allocated the segtab needed for this va
   1175   1.1  christos 		 * free the page we just allocated.
   1176   1.1  christos 		 */
   1177  1.29     skrll 		if (__predict_false(oppg != NULL)) {
   1178  1.29     skrll 			pmap_ptpage_free(pmap, ppg, __func__);
   1179  1.29     skrll 			ppg = oppg;
   1180  1.29     skrll #if defined(PMAP_HWPAGEWALKER)
   1181  1.29     skrll 		} else {
   1182  1.29     skrll 			pte_pde_set(pde_p, npde);
   1183   1.1  christos #endif
   1184   1.1  christos 		}
   1185  1.29     skrll #else /* !MULTIPROCESSOR */
   1186  1.29     skrll 		*ppg_p = ppg;
   1187  1.29     skrll #endif /* MULTIPROCESSOR */
   1188  1.29     skrll #endif /* PMAP_HWPAGEWALKER && PMAP_MAP_PDETABPAGE */
   1189  1.29     skrll 	}
   1190   1.1  christos 
   1191  1.29     skrll 	const size_t pte_idx = (va >> PGSHIFT) & (NPTEPG - 1);
   1192   1.1  christos 
   1193  1.29     skrll 	return ppg->ppg_ptes + pte_idx;
   1194   1.1  christos }
   1195