uvm_fault_i.h revision 1.30 1 1.30 christos /* $NetBSD: uvm_fault_i.h,v 1.30 2018/05/07 21:00:14 christos Exp $ */
2 1.1 mrg
3 1.1 mrg /*
4 1.1 mrg * Copyright (c) 1997 Charles D. Cranor and Washington University.
5 1.1 mrg * All rights reserved.
6 1.1 mrg *
7 1.1 mrg * Redistribution and use in source and binary forms, with or without
8 1.1 mrg * modification, are permitted provided that the following conditions
9 1.1 mrg * are met:
10 1.1 mrg * 1. Redistributions of source code must retain the above copyright
11 1.1 mrg * notice, this list of conditions and the following disclaimer.
12 1.1 mrg * 2. Redistributions in binary form must reproduce the above copyright
13 1.1 mrg * notice, this list of conditions and the following disclaimer in the
14 1.1 mrg * documentation and/or other materials provided with the distribution.
15 1.1 mrg *
16 1.1 mrg * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
17 1.1 mrg * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
18 1.1 mrg * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
19 1.1 mrg * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
20 1.1 mrg * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
21 1.1 mrg * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
22 1.1 mrg * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
23 1.1 mrg * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
24 1.1 mrg * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
25 1.1 mrg * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
26 1.3 mrg *
27 1.3 mrg * from: Id: uvm_fault_i.h,v 1.1.6.1 1997/12/08 16:07:12 chuck Exp
28 1.1 mrg */
29 1.1 mrg
30 1.4 perry #ifndef _UVM_UVM_FAULT_I_H_
31 1.4 perry #define _UVM_UVM_FAULT_I_H_
32 1.4 perry
33 1.1 mrg /*
34 1.1 mrg * uvm_fault_i.h: fault inline functions
35 1.1 mrg */
36 1.30 christos void uvmfault_update_stats(struct uvm_faultinfo *);
37 1.30 christos
38 1.1 mrg
39 1.1 mrg /*
40 1.1 mrg * uvmfault_unlockmaps: unlock the maps
41 1.1 mrg */
42 1.1 mrg
43 1.29 christos static __inline void
44 1.22 thorpej uvmfault_unlockmaps(struct uvm_faultinfo *ufi, bool write_locked)
45 1.5 mrg {
46 1.10 chs /*
47 1.10 chs * ufi can be NULL when this isn't really a fault,
48 1.10 chs * but merely paging in anon data.
49 1.10 chs */
50 1.10 chs
51 1.10 chs if (ufi == NULL) {
52 1.10 chs return;
53 1.10 chs }
54 1.1 mrg
55 1.30 christos uvmfault_update_stats(ufi);
56 1.5 mrg if (write_locked) {
57 1.5 mrg vm_map_unlock(ufi->map);
58 1.5 mrg } else {
59 1.5 mrg vm_map_unlock_read(ufi->map);
60 1.5 mrg }
61 1.1 mrg }
62 1.1 mrg
63 1.1 mrg /*
64 1.1 mrg * uvmfault_unlockall: unlock everything passed in.
65 1.1 mrg *
66 1.1 mrg * => maps must be read-locked (not write-locked).
67 1.1 mrg */
68 1.1 mrg
69 1.29 christos static __inline void
70 1.18 thorpej uvmfault_unlockall(struct uvm_faultinfo *ufi, struct vm_amap *amap,
71 1.27 rmind struct uvm_object *uobj)
72 1.5 mrg {
73 1.1 mrg
74 1.5 mrg if (uobj)
75 1.27 rmind mutex_exit(uobj->vmobjlock);
76 1.5 mrg if (amap)
77 1.7 chuck amap_unlock(amap);
78 1.23 thorpej uvmfault_unlockmaps(ufi, false);
79 1.9 thorpej }
80 1.9 thorpej
81 1.9 thorpej /*
82 1.1 mrg * uvmfault_lookup: lookup a virtual address in a map
83 1.1 mrg *
84 1.6 chuck * => caller must provide a uvm_faultinfo structure with the IN
85 1.1 mrg * params properly filled in
86 1.6 chuck * => we will lookup the map entry (handling submaps) as we go
87 1.1 mrg * => if the lookup is a success we will return with the maps locked
88 1.23 thorpej * => if "write_lock" is true, we write_lock the map, otherwise we only
89 1.1 mrg * get a read lock.
90 1.12 chs * => note that submaps can only appear in the kernel and they are
91 1.6 chuck * required to use the same virtual addresses as the map they
92 1.6 chuck * are referenced by (thus address translation between the main
93 1.6 chuck * map and the submap is unnecessary).
94 1.1 mrg */
95 1.1 mrg
96 1.29 christos static __inline bool
97 1.22 thorpej uvmfault_lookup(struct uvm_faultinfo *ufi, bool write_lock)
98 1.1 mrg {
99 1.13 chs struct vm_map *tmpmap;
100 1.1 mrg
101 1.5 mrg /*
102 1.5 mrg * init ufi values for lookup.
103 1.5 mrg */
104 1.5 mrg
105 1.5 mrg ufi->map = ufi->orig_map;
106 1.5 mrg ufi->size = ufi->orig_size;
107 1.5 mrg
108 1.5 mrg /*
109 1.5 mrg * keep going down levels until we are done. note that there can
110 1.5 mrg * only be two levels so we won't loop very long.
111 1.5 mrg */
112 1.5 mrg
113 1.28 rmind for (;;) {
114 1.5 mrg /*
115 1.5 mrg * lock map
116 1.5 mrg */
117 1.5 mrg if (write_lock) {
118 1.5 mrg vm_map_lock(ufi->map);
119 1.5 mrg } else {
120 1.5 mrg vm_map_lock_read(ufi->map);
121 1.5 mrg }
122 1.5 mrg
123 1.5 mrg /*
124 1.5 mrg * lookup
125 1.5 mrg */
126 1.12 chs if (!uvm_map_lookup_entry(ufi->map, ufi->orig_rvaddr,
127 1.27 rmind &ufi->entry)) {
128 1.5 mrg uvmfault_unlockmaps(ufi, write_lock);
129 1.23 thorpej return(false);
130 1.5 mrg }
131 1.5 mrg
132 1.5 mrg /*
133 1.5 mrg * reduce size if necessary
134 1.5 mrg */
135 1.6 chuck if (ufi->entry->end - ufi->orig_rvaddr < ufi->size)
136 1.6 chuck ufi->size = ufi->entry->end - ufi->orig_rvaddr;
137 1.5 mrg
138 1.5 mrg /*
139 1.5 mrg * submap? replace map with the submap and lookup again.
140 1.5 mrg * note: VAs in submaps must match VAs in main map.
141 1.5 mrg */
142 1.5 mrg if (UVM_ET_ISSUBMAP(ufi->entry)) {
143 1.5 mrg tmpmap = ufi->entry->object.sub_map;
144 1.5 mrg if (write_lock) {
145 1.5 mrg vm_map_unlock(ufi->map);
146 1.5 mrg } else {
147 1.5 mrg vm_map_unlock_read(ufi->map);
148 1.5 mrg }
149 1.5 mrg ufi->map = tmpmap;
150 1.5 mrg continue;
151 1.5 mrg }
152 1.5 mrg
153 1.5 mrg /*
154 1.5 mrg * got it!
155 1.5 mrg */
156 1.1 mrg
157 1.5 mrg ufi->mapv = ufi->map->timestamp;
158 1.23 thorpej return(true);
159 1.1 mrg
160 1.5 mrg } /* while loop */
161 1.1 mrg
162 1.5 mrg /*NOTREACHED*/
163 1.1 mrg }
164 1.1 mrg
165 1.1 mrg /*
166 1.1 mrg * uvmfault_relock: attempt to relock the same version of the map
167 1.1 mrg *
168 1.1 mrg * => fault data structures should be unlocked before calling.
169 1.23 thorpej * => if a success (true) maps will be locked after call.
170 1.1 mrg */
171 1.1 mrg
172 1.29 christos static __inline bool
173 1.18 thorpej uvmfault_relock(struct uvm_faultinfo *ufi)
174 1.5 mrg {
175 1.10 chs /*
176 1.10 chs * ufi can be NULL when this isn't really a fault,
177 1.10 chs * but merely paging in anon data.
178 1.10 chs */
179 1.10 chs
180 1.10 chs if (ufi == NULL) {
181 1.23 thorpej return true;
182 1.10 chs }
183 1.1 mrg
184 1.5 mrg uvmexp.fltrelck++;
185 1.10 chs
186 1.5 mrg /*
187 1.12 chs * relock map. fail if version mismatch (in which case nothing
188 1.6 chuck * gets locked).
189 1.5 mrg */
190 1.5 mrg
191 1.5 mrg vm_map_lock_read(ufi->map);
192 1.5 mrg if (ufi->mapv != ufi->map->timestamp) {
193 1.5 mrg vm_map_unlock_read(ufi->map);
194 1.23 thorpej return(false);
195 1.5 mrg }
196 1.1 mrg
197 1.5 mrg uvmexp.fltrelckok++;
198 1.23 thorpej return(true);
199 1.1 mrg }
200 1.4 perry
201 1.4 perry #endif /* _UVM_UVM_FAULT_I_H_ */
202