uvm_amap.h revision 1.13 1 /* $NetBSD: uvm_amap.h,v 1.13 2000/11/25 06:27:59 chs Exp $ */
2
3 /*
4 *
5 * Copyright (c) 1997 Charles D. Cranor and Washington University.
6 * All rights reserved.
7 *
8 * Redistribution and use in source and binary forms, with or without
9 * modification, are permitted provided that the following conditions
10 * are met:
11 * 1. Redistributions of source code must retain the above copyright
12 * notice, this list of conditions and the following disclaimer.
13 * 2. Redistributions in binary form must reproduce the above copyright
14 * notice, this list of conditions and the following disclaimer in the
15 * documentation and/or other materials provided with the distribution.
16 * 3. All advertising materials mentioning features or use of this software
17 * must display the following acknowledgement:
18 * This product includes software developed by Charles D. Cranor and
19 * Washington University.
20 * 4. The name of the author may not be used to endorse or promote products
21 * derived from this software without specific prior written permission.
22 *
23 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
24 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
25 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
26 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
27 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
28 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
29 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
30 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
31 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
32 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
33 */
34
35 #ifndef _UVM_UVM_AMAP_H_
36 #define _UVM_UVM_AMAP_H_
37
38 /*
39 * uvm_amap.h: general amap interface and amap implementation-specific info
40 */
41
42 /*
43 * an amap structure contains pointers to a set of anons that are
44 * mapped together in virtual memory (an anon is a single page of
45 * anonymous virtual memory -- see uvm_anon.h). in uvm we hide the
46 * details of the implementation of amaps behind a general amap
47 * interface. this allows us to change the amap implementation
48 * without having to touch the rest of the code. this file is divided
49 * into two parts: the definition of the uvm amap interface and the
50 * amap implementation-specific definitions.
51 */
52
53 #ifdef _KERNEL
54
55 /*
56 * part 1: amap interface
57 */
58
59 /*
60 * forward definition of vm_amap structure. only amap
61 * implementation-specific code should directly access the fields of
62 * this structure.
63 */
64
65 struct vm_amap;
66
67 /*
68 * handle inline options... we allow amap ops to be inline, but we also
69 * provide a hook to turn this off. macros can also be used.
70 */
71
72 #ifdef UVM_AMAP_INLINE /* defined/undef'd in uvm_amap.c */
73 #define AMAP_INLINE static __inline /* inline enabled */
74 #else
75 #define AMAP_INLINE /* inline disabled */
76 #endif /* UVM_AMAP_INLINE */
77
78
79 /*
80 * prototypes for the amap interface
81 */
82
83 AMAP_INLINE
84 void amap_add /* add an anon to an amap */
85 __P((struct vm_aref *, vaddr_t,
86 struct vm_anon *, boolean_t));
87 struct vm_amap *amap_alloc /* allocate a new amap */
88 __P((vaddr_t, vaddr_t, int));
89 void amap_copy /* clear amap needs-copy flag */
90 __P((vm_map_t, vm_map_entry_t, int,
91 boolean_t, vaddr_t, vaddr_t));
92 void amap_cow_now /* resolve all COW faults now */
93 __P((vm_map_t, vm_map_entry_t));
94 void amap_extend /* make amap larger */
95 __P((vm_map_entry_t, vsize_t));
96 int amap_flags /* get amap's flags */
97 __P((struct vm_amap *));
98 void amap_free /* free amap */
99 __P((struct vm_amap *));
100 void amap_init /* init amap module (at boot time) */
101 __P((void));
102 void amap_lock /* lock amap */
103 __P((struct vm_amap *));
104 AMAP_INLINE
105 struct vm_anon *amap_lookup /* lookup an anon @ offset in amap */
106 __P((struct vm_aref *, vaddr_t));
107 AMAP_INLINE
108 void amap_lookups /* lookup multiple anons */
109 __P((struct vm_aref *, vaddr_t,
110 struct vm_anon **, int));
111 AMAP_INLINE
112 void amap_ref /* add a reference to an amap */
113 __P((struct vm_amap *, vaddr_t, vsize_t, int));
114 int amap_refs /* get number of references of amap */
115 __P((struct vm_amap *));
116 void amap_share_protect /* protect pages in a shared amap */
117 __P((vm_map_entry_t, vm_prot_t));
118 void amap_splitref /* split reference to amap into two */
119 __P((struct vm_aref *, struct vm_aref *,
120 vaddr_t));
121 AMAP_INLINE
122 void amap_unadd /* remove an anon from an amap */
123 __P((struct vm_aref *, vaddr_t));
124 void amap_unlock /* unlock amap */
125 __P((struct vm_amap *));
126 AMAP_INLINE
127 void amap_unref /* drop reference to an amap */
128 __P((struct vm_amap *, vaddr_t, vsize_t, int));
129 void amap_wipeout /* remove all anons from amap */
130 __P((struct vm_amap *));
131
132 /*
133 * amap flag values
134 */
135
136 #define AMAP_SHARED 0x1 /* amap is shared */
137 #define AMAP_REFALL 0x2 /* amap_ref: reference entire amap */
138
139 #endif /* _KERNEL */
140
141 /**********************************************************************/
142
143 /*
144 * part 2: amap implementation-specific info
145 */
146
147 /*
148 * we currently provide an array-based amap implementation. in this
149 * implementation we provide the option of tracking split references
150 * so that we don't lose track of references during partial unmaps
151 * ... this is enabled with the "UVM_AMAP_PPREF" define.
152 */
153
154 #define UVM_AMAP_PPREF /* track partial references */
155
156 /*
157 * here is the definition of the vm_amap structure for this implementation.
158 */
159
160 struct vm_amap {
161 simple_lock_data_t am_l; /* simple lock [locks all vm_amap fields] */
162 int am_ref; /* reference count */
163 int am_flags; /* flags */
164 int am_maxslot; /* max # of slots allocated */
165 int am_nslot; /* # of slots currently in map ( <= maxslot) */
166 int am_nused; /* # of slots currently in use */
167 int *am_slots; /* contig array of active slots */
168 int *am_bckptr; /* back pointer array to am_slots */
169 struct vm_anon **am_anon; /* array of anonymous pages */
170 #ifdef UVM_AMAP_PPREF
171 int *am_ppref; /* per page reference count (if !NULL) */
172 #endif
173 };
174
175 /*
176 * note that am_slots, am_bckptr, and am_anon are arrays. this allows
177 * fast lookup of pages based on their virual address at the expense of
178 * some extra memory. in the future we should be smarter about memory
179 * usage and fall back to a non-array based implementation on systems
180 * that are short of memory (XXXCDC).
181 *
182 * the entries in the array are called slots... for example an amap that
183 * covers four pages of virtual memory is said to have four slots. here
184 * is an example of the array usage for a four slot amap. note that only
185 * slots one and three have anons assigned to them. "D/C" means that we
186 * "don't care" about the value.
187 *
188 * 0 1 2 3
189 * am_anon: NULL, anon0, NULL, anon1 (actual pointers to anons)
190 * am_bckptr: D/C, 1, D/C, 0 (points to am_slots entry)
191 *
192 * am_slots: 3, 1, D/C, D/C (says slots 3 and 1 are in use)
193 *
194 * note that am_bckptr is D/C if the slot in am_anon is set to NULL.
195 * to find the entry in am_slots for an anon, look at am_bckptr[slot],
196 * thus the entry for slot 3 in am_slots[] is at am_slots[am_bckptr[3]].
197 * in general, if am_anon[X] is non-NULL, then the following must be
198 * true: am_slots[am_bckptr[X]] == X
199 *
200 * note that am_slots is always contig-packed.
201 */
202
203 /*
204 * defines for handling of large sparce amaps:
205 *
206 * one of the problems of array-based amaps is that if you allocate a
207 * large sparcely-used area of virtual memory you end up allocating
208 * large arrays that, for the most part, don't get used. this is a
209 * problem for BSD in that the kernel likes to make these types of
210 * allocations to "reserve" memory for possible future use.
211 *
212 * for example, the kernel allocates (reserves) a large chunk of user
213 * VM for possible stack growth. most of the time only a page or two
214 * of this VM is actually used. since the stack is anonymous memory
215 * it makes sense for it to live in an amap, but if we allocated an
216 * amap for the entire stack range we could end up wasting a large
217 * amount of malloc'd KVM.
218 *
219 * for example, on the i386 at boot time we allocate two amaps for the stack
220 * of /sbin/init:
221 * 1. a 7680 slot amap at protection 0 (reserve space for stack)
222 * 2. a 512 slot amap at protection 7 (top of stack)
223 *
224 * most of the array allocated for the amaps for this is never used.
225 * the amap interface provides a way for us to avoid this problem by
226 * allowing amap_copy() to break larger amaps up into smaller sized
227 * chunks (controlled by the "canchunk" option). we use this feature
228 * to reduce our memory usage with the BSD stack management. if we
229 * are asked to create an amap with more than UVM_AMAP_LARGE slots in it,
230 * we attempt to break it up into a UVM_AMAP_CHUNK sized amap if the
231 * "canchunk" flag is set.
232 *
233 * so, in the i386 example, the 7680 slot area is never referenced so
234 * nothing gets allocated (amap_copy is never called because the protection
235 * is zero). the 512 slot area for the top of the stack is referenced.
236 * the chunking code breaks it up into 16 slot chunks (hopefully a single
237 * 16 slot chunk is enough to handle the whole stack).
238 */
239
240 #define UVM_AMAP_LARGE 256 /* # of slots in "large" amap */
241 #define UVM_AMAP_CHUNK 16 /* # of slots to chunk large amaps in */
242
243 #ifdef _KERNEL
244
245 /*
246 * macros
247 */
248
249 /* AMAP_B2SLOT: convert byte offset to slot */
250 #ifdef DIAGNOSTIC
251 #define AMAP_B2SLOT(S,B) { \
252 if ((B) & (PAGE_SIZE - 1)) \
253 panic("AMAP_B2SLOT: invalid byte count"); \
254 (S) = (B) >> PAGE_SHIFT; \
255 }
256 #else
257 #define AMAP_B2SLOT(S,B) (S) = (B) >> PAGE_SHIFT
258 #endif
259
260 /*
261 * lock/unlock/refs/flags macros
262 */
263
264 #define amap_flags(AMAP) ((AMAP)->am_flags)
265 #define amap_lock(AMAP) simple_lock(&(AMAP)->am_l)
266 #define amap_refs(AMAP) ((AMAP)->am_ref)
267 #define amap_unlock(AMAP) simple_unlock(&(AMAP)->am_l)
268
269 /*
270 * if we enable PPREF, then we have a couple of extra functions that
271 * we need to prototype here...
272 */
273
274 #ifdef UVM_AMAP_PPREF
275
276 #define PPREF_NONE ((int *) -1) /* not using ppref */
277
278 void amap_pp_adjref /* adjust references */
279 __P((struct vm_amap *, int, vsize_t, int));
280 void amap_pp_establish /* establish ppref */
281 __P((struct vm_amap *));
282 void amap_wiperange /* wipe part of an amap */
283 __P((struct vm_amap *, int, int));
284 #endif /* UVM_AMAP_PPREF */
285
286 #endif /* _KERNEL */
287
288 #endif /* _UVM_UVM_AMAP_H_ */
289