Home | History | Annotate | Line # | Download | only in include
vmparam.h revision 1.7
      1  1.7       ryo /* $NetBSD: vmparam.h,v 1.7 2018/10/12 01:28:58 ryo Exp $ */
      2  1.1      matt 
      3  1.1      matt /*-
      4  1.1      matt  * Copyright (c) 2014 The NetBSD Foundation, Inc.
      5  1.1      matt  * All rights reserved.
      6  1.1      matt  *
      7  1.1      matt  * This code is derived from software contributed to The NetBSD Foundation
      8  1.1      matt  * by Matt Thomas of 3am Software Foundry.
      9  1.1      matt  *
     10  1.1      matt  * Redistribution and use in source and binary forms, with or without
     11  1.1      matt  * modification, are permitted provided that the following conditions
     12  1.1      matt  * are met:
     13  1.1      matt  * 1. Redistributions of source code must retain the above copyright
     14  1.1      matt  *    notice, this list of conditions and the following disclaimer.
     15  1.1      matt  * 2. Redistributions in binary form must reproduce the above copyright
     16  1.1      matt  *    notice, this list of conditions and the following disclaimer in the
     17  1.1      matt  *    documentation and/or other materials provided with the distribution.
     18  1.1      matt  *
     19  1.1      matt  * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
     20  1.1      matt  * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
     21  1.1      matt  * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
     22  1.1      matt  * PURPOSE ARE DISCLAIMED.  IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
     23  1.1      matt  * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
     24  1.1      matt  * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
     25  1.1      matt  * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
     26  1.1      matt  * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
     27  1.1      matt  * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
     28  1.1      matt  * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
     29  1.1      matt  * POSSIBILITY OF SUCH DAMAGE.
     30  1.1      matt  */
     31  1.1      matt 
     32  1.1      matt #ifndef _AARCH64_VMPARAM_H_
     33  1.1      matt #define _AARCH64_VMPARAM_H_
     34  1.1      matt 
     35  1.1      matt #ifdef __aarch64__
     36  1.1      matt 
     37  1.3       ryo #define	__USE_TOPDOWN_VM
     38  1.3       ryo 
     39  1.1      matt /*
     40  1.4  jmcneill  * Default pager_map of 16MB is small and we have plenty of VA to burn.
     41  1.4  jmcneill  */
     42  1.4  jmcneill #define	PAGER_MAP_DEFAULT_SIZE	(512 * 1024 * 1024)
     43  1.4  jmcneill 
     44  1.4  jmcneill /*
     45  1.1      matt  * AARCH64 supports 3 page sizes: 4KB, 16KB, 64KB.  Each page table can
     46  1.1      matt  * even have its own page size.
     47  1.1      matt  */
     48  1.1      matt 
     49  1.1      matt #ifdef AARCH64_PAGE_SHIFT
     50  1.1      matt #if (1 << AARCH64_PAGE_SHIFT) & ~0x141000
     51  1.1      matt #error AARCH64_PAGE_SHIFT contains an unsupported value.
     52  1.1      matt #endif
     53  1.1      matt #define PAGE_SHIFT	AARCH64_PAGE_SHIFT
     54  1.1      matt #else
     55  1.1      matt #define PAGE_SHIFT	12
     56  1.1      matt #endif
     57  1.1      matt #define PAGE_SIZE	(1 << PAGE_SHIFT)
     58  1.1      matt #define PAGE_MASK	(PAGE_SIZE - 1)
     59  1.1      matt 
     60  1.1      matt #if PAGE_SHIFT <= 14
     61  1.3       ryo #define USPACE		32768
     62  1.1      matt #else
     63  1.1      matt #define USPACE		65536
     64  1.1      matt #endif
     65  1.1      matt #define	UPAGES		(USPACE >> PAGE_SHIFT)
     66  1.1      matt 
     67  1.1      matt /*
     68  1.1      matt  * USRSTACK is the top (end) of the user stack.  The user VA space is a
     69  1.1      matt  * 48-bit address space starting at 0.  Place the stack at its top end.
     70  1.1      matt  */
     71  1.3       ryo #define USRSTACK	VM_MAXUSER_ADDRESS
     72  1.1      matt 
     73  1.1      matt #ifndef MAXTSIZ
     74  1.1      matt #define	MAXTSIZ		(1L << 30)	/* max text size (1GB) */
     75  1.1      matt #endif
     76  1.1      matt 
     77  1.1      matt #ifndef MAXDSIZ
     78  1.1      matt #define	MAXDSIZ		(1L << 36)	/* max data size (64GB) */
     79  1.1      matt #endif
     80  1.1      matt 
     81  1.1      matt #ifndef MAXSSIZ
     82  1.1      matt #define	MAXSSIZ		(1L << 26)	/* max stack size (64MB) */
     83  1.1      matt #endif
     84  1.1      matt 
     85  1.1      matt #ifndef DFLDSIZ
     86  1.1      matt #define	DFLDSIZ		(1L << 32)	/* default data size (4GB) */
     87  1.1      matt #endif
     88  1.1      matt 
     89  1.3       ryo #ifndef DFLSSIZ
     90  1.3       ryo #define	DFLSSIZ		(1L << 23)	/* default stack size (8MB) */
     91  1.3       ryo #endif
     92  1.3       ryo 
     93  1.3       ryo #define USRSTACK32	VM_MAXUSER_ADDRESS32
     94  1.3       ryo 
     95  1.3       ryo #ifndef MAXTSIZ32
     96  1.3       ryo #define	MAXTSIZ32	(1L << 26)	/* 32bit max text size (64MB) */
     97  1.3       ryo #endif
     98  1.3       ryo 
     99  1.3       ryo #ifndef	MAXDSIZ32
    100  1.7       ryo #define	MAXDSIZ32	(3U*1024*1024*1024)	/* max data size */
    101  1.3       ryo #endif
    102  1.3       ryo 
    103  1.3       ryo #ifndef	MAXSSIZ32
    104  1.3       ryo #define	MAXSSIZ32	(64*1024*1024)		/* max stack size */
    105  1.3       ryo #endif
    106  1.3       ryo 
    107  1.1      matt #ifndef DFLDSIZ32
    108  1.1      matt #define	DFLDSIZ32	(1L << 27)	/* 32bit default data size (128MB) */
    109  1.1      matt #endif
    110  1.1      matt 
    111  1.1      matt #ifndef DFLSSIZ32
    112  1.1      matt #define	DFLSSIZ32	(1L << 21)	/* 32bit default stack size (2MB) */
    113  1.1      matt #endif
    114  1.1      matt 
    115  1.1      matt #define	VM_MIN_ADDRESS		((vaddr_t) 0x0)
    116  1.1      matt #define	VM_MAXUSER_ADDRESS	((vaddr_t) (1L << 48) - PAGE_SIZE)
    117  1.1      matt #define	VM_MAX_ADDRESS		VM_MAXUSER_ADDRESS
    118  1.1      matt 
    119  1.7       ryo #define VM_MAXUSER_ADDRESS32	((vaddr_t) 0xfffff000)
    120  1.1      matt 
    121  1.1      matt /*
    122  1.1      matt  * Give ourselves 64GB of mappable kernel space.  That leaves the rest
    123  1.3       ryo  * to be user for directly mapped (block addressable) addresses.
    124  1.1      matt  */
    125  1.1      matt #define VM_MIN_KERNEL_ADDRESS	((vaddr_t) 0xffffffc000000000L)
    126  1.3       ryo #define VM_MAX_KERNEL_ADDRESS	((vaddr_t) 0xffffffffffe00000L)
    127  1.3       ryo 
    128  1.3       ryo /*
    129  1.3       ryo  * last 254MB of kernel vm area (0xfffffffff0000000-0xffffffffffe00000)
    130  1.3       ryo  * may be used for devmap. address must be aligned 2MB (L2_SIZE)
    131  1.3       ryo  * see also aarch64/pmap.c:pmap_devmap_*
    132  1.3       ryo  */
    133  1.3       ryo #define VM_KERNEL_IO_ADDRESS	0xfffffffff0000000L
    134  1.6       ryo #define VM_KERNEL_IO_SIZE	(VM_MAX_KERNEL_ADDRESS - VM_KERNEL_IO_ADDRESS)
    135  1.1      matt 
    136  1.1      matt /* virtual sizes (bytes) for various kernel submaps */
    137  1.1      matt #define USRIOSIZE		(PAGE_SIZE / 8)
    138  1.1      matt #define VM_PHYS_SIZE		(USRIOSIZE * PAGE_SIZE)
    139  1.1      matt 
    140  1.7       ryo #define VM_DEFAULT_ADDRESS32_TOPDOWN(da, sz) \
    141  1.7       ryo 	trunc_page(USRSTACK32 - MAXSSIZ32 - (sz) - user_stack_guard_size)
    142  1.7       ryo #define VM_DEFAULT_ADDRESS32_BOTTOMUP(da, sz) \
    143  1.7       ryo 	round_page((vaddr_t)(da) + (vsize_t)MAXDSIZ32)
    144  1.7       ryo 
    145  1.1      matt /*
    146  1.1      matt  * Since we have the address space, we map all of physical memory (RAM)
    147  1.1      matt  * using block page table entries.
    148  1.1      matt  */
    149  1.1      matt #define AARCH64_KSEG_MASK	((vaddr_t) 0xffff000000000000L)
    150  1.3       ryo #define AARCH64_KSEG_SIZE	(1UL << 39)	/* 512GB */
    151  1.1      matt #define AARCH64_KSEG_START	AARCH64_KSEG_MASK
    152  1.3       ryo #define AARCH64_KSEG_END	(AARCH64_KSEG_START + AARCH64_KSEG_SIZE)
    153  1.3       ryo #define AARCH64_KMEMORY_BASE	AARCH64_KSEG_MASK
    154  1.1      matt #define AARCH64_KVA_P(va)	(((vaddr_t) (va) & AARCH64_KSEG_MASK) != 0)
    155  1.1      matt #define AARCH64_PA_TO_KVA(pa)	((vaddr_t) ((pa) | AARCH64_KSEG_START))
    156  1.3       ryo #define AARCH64_KVA_TO_PA(va)	((paddr_t) ((va) & ~AARCH64_KSEG_MASK))
    157  1.1      matt 
    158  1.1      matt /* */
    159  1.5  jmcneill #define VM_PHYSSEG_MAX		64              /* XXX */
    160  1.1      matt #define VM_PHYSSEG_STRAT	VM_PSTRAT_BSEARCH
    161  1.1      matt 
    162  1.3       ryo #define VM_NFREELIST		3
    163  1.3       ryo #define VM_FREELIST_DEFAULT	0
    164  1.1      matt #define VM_FREELIST_FIRST4GB	1
    165  1.3       ryo #define VM_FREELIST_HIGHMEM	2
    166  1.1      matt 
    167  1.1      matt #elif defined(__arm__)
    168  1.1      matt 
    169  1.1      matt #include <arm/vmparam.h>
    170  1.1      matt 
    171  1.1      matt #endif /* __aarch64__/__arm__ */
    172  1.1      matt 
    173  1.1      matt #endif /* _AARCH64_VMPARAM_H_ */
    174