subr_devsw.c revision 1.43 1 1.43 riastrad /* $NetBSD: subr_devsw.c,v 1.43 2022/03/28 12:38:33 riastradh Exp $ */
2 1.11 ad
3 1.2 gehenna /*-
4 1.20 ad * Copyright (c) 2001, 2002, 2007, 2008 The NetBSD Foundation, Inc.
5 1.2 gehenna * All rights reserved.
6 1.2 gehenna *
7 1.2 gehenna * This code is derived from software contributed to The NetBSD Foundation
8 1.11 ad * by MAEKAWA Masahide <gehenna (at) NetBSD.org>, and by Andrew Doran.
9 1.2 gehenna *
10 1.2 gehenna * Redistribution and use in source and binary forms, with or without
11 1.2 gehenna * modification, are permitted provided that the following conditions
12 1.2 gehenna * are met:
13 1.2 gehenna * 1. Redistributions of source code must retain the above copyright
14 1.2 gehenna * notice, this list of conditions and the following disclaimer.
15 1.2 gehenna * 2. Redistributions in binary form must reproduce the above copyright
16 1.2 gehenna * notice, this list of conditions and the following disclaimer in the
17 1.2 gehenna * documentation and/or other materials provided with the distribution.
18 1.2 gehenna *
19 1.2 gehenna * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
20 1.2 gehenna * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
21 1.2 gehenna * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
22 1.2 gehenna * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
23 1.2 gehenna * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
24 1.2 gehenna * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
25 1.2 gehenna * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
26 1.2 gehenna * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
27 1.2 gehenna * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
28 1.2 gehenna * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
29 1.2 gehenna * POSSIBILITY OF SUCH DAMAGE.
30 1.2 gehenna */
31 1.11 ad
32 1.11 ad /*
33 1.11 ad * Overview
34 1.11 ad *
35 1.11 ad * subr_devsw.c: registers device drivers by name and by major
36 1.11 ad * number, and provides wrapper methods for performing I/O and
37 1.11 ad * other tasks on device drivers, keying on the device number
38 1.11 ad * (dev_t).
39 1.11 ad *
40 1.11 ad * When the system is built, the config(8) command generates
41 1.11 ad * static tables of device drivers built into the kernel image
42 1.11 ad * along with their associated methods. These are recorded in
43 1.11 ad * the cdevsw0 and bdevsw0 tables. Drivers can also be added to
44 1.11 ad * and removed from the system dynamically.
45 1.11 ad *
46 1.11 ad * Allocation
47 1.11 ad *
48 1.11 ad * When the system initially boots only the statically allocated
49 1.11 ad * indexes (bdevsw0, cdevsw0) are used. If these overflow due to
50 1.11 ad * allocation, we allocate a fixed block of memory to hold the new,
51 1.11 ad * expanded index. This "fork" of the table is only ever performed
52 1.11 ad * once in order to guarantee that other threads may safely access
53 1.11 ad * the device tables:
54 1.11 ad *
55 1.11 ad * o Once a thread has a "reference" to the table via an earlier
56 1.11 ad * open() call, we know that the entry in the table must exist
57 1.11 ad * and so it is safe to access it.
58 1.11 ad *
59 1.11 ad * o Regardless of whether other threads see the old or new
60 1.11 ad * pointers, they will point to a correct device switch
61 1.11 ad * structure for the operation being performed.
62 1.11 ad *
63 1.11 ad * XXX Currently, the wrapper methods such as cdev_read() verify
64 1.11 ad * that a device driver does in fact exist before calling the
65 1.11 ad * associated driver method. This should be changed so that
66 1.11 ad * once the device is has been referenced by a vnode (opened),
67 1.11 ad * calling the other methods should be valid until that reference
68 1.11 ad * is dropped.
69 1.11 ad */
70 1.7 lukem
71 1.7 lukem #include <sys/cdefs.h>
72 1.43 riastrad __KERNEL_RCSID(0, "$NetBSD: subr_devsw.c,v 1.43 2022/03/28 12:38:33 riastradh Exp $");
73 1.34 riz
74 1.34 riz #ifdef _KERNEL_OPT
75 1.34 riz #include "opt_dtrace.h"
76 1.34 riz #endif
77 1.2 gehenna
78 1.2 gehenna #include <sys/param.h>
79 1.2 gehenna #include <sys/conf.h>
80 1.11 ad #include <sys/kmem.h>
81 1.2 gehenna #include <sys/systm.h>
82 1.11 ad #include <sys/poll.h>
83 1.11 ad #include <sys/tty.h>
84 1.15 matt #include <sys/cpu.h>
85 1.11 ad #include <sys/buf.h>
86 1.29 mrg #include <sys/reboot.h>
87 1.34 riz #include <sys/sdt.h>
88 1.40 riastrad #include <sys/atomic.h>
89 1.40 riastrad #include <sys/localcount.h>
90 1.40 riastrad #include <sys/pserialize.h>
91 1.40 riastrad #include <sys/xcall.h>
92 1.41 riastrad #include <sys/device.h>
93 1.2 gehenna
94 1.2 gehenna #ifdef DEVSW_DEBUG
95 1.2 gehenna #define DPRINTF(x) printf x
96 1.2 gehenna #else /* DEVSW_DEBUG */
97 1.2 gehenna #define DPRINTF(x)
98 1.2 gehenna #endif /* DEVSW_DEBUG */
99 1.2 gehenna
100 1.11 ad #define MAXDEVSW 512 /* the maximum of major device number */
101 1.2 gehenna #define BDEVSW_SIZE (sizeof(struct bdevsw *))
102 1.2 gehenna #define CDEVSW_SIZE (sizeof(struct cdevsw *))
103 1.2 gehenna #define DEVSWCONV_SIZE (sizeof(struct devsw_conv))
104 1.2 gehenna
105 1.40 riastrad struct devswref {
106 1.40 riastrad struct localcount *dr_lc;
107 1.40 riastrad };
108 1.40 riastrad
109 1.40 riastrad /* XXX bdevsw, cdevsw, max_bdevsws, and max_cdevsws should be volatile */
110 1.2 gehenna extern const struct bdevsw **bdevsw, *bdevsw0[];
111 1.2 gehenna extern const struct cdevsw **cdevsw, *cdevsw0[];
112 1.2 gehenna extern struct devsw_conv *devsw_conv, devsw_conv0[];
113 1.2 gehenna extern const int sys_bdevsws, sys_cdevsws;
114 1.2 gehenna extern int max_bdevsws, max_cdevsws, max_devsw_convs;
115 1.2 gehenna
116 1.40 riastrad static struct devswref *cdevswref;
117 1.40 riastrad static struct devswref *bdevswref;
118 1.40 riastrad static kcondvar_t devsw_cv;
119 1.40 riastrad
120 1.24 drochner static int bdevsw_attach(const struct bdevsw *, devmajor_t *);
121 1.24 drochner static int cdevsw_attach(const struct cdevsw *, devmajor_t *);
122 1.11 ad static void devsw_detach_locked(const struct bdevsw *, const struct cdevsw *);
123 1.11 ad
124 1.23 pooka kmutex_t device_lock;
125 1.23 pooka
126 1.31 pooka void (*biodone_vfs)(buf_t *) = (void *)nullop;
127 1.31 pooka
128 1.11 ad void
129 1.11 ad devsw_init(void)
130 1.11 ad {
131 1.11 ad
132 1.11 ad KASSERT(sys_bdevsws < MAXDEVSW - 1);
133 1.11 ad KASSERT(sys_cdevsws < MAXDEVSW - 1);
134 1.23 pooka mutex_init(&device_lock, MUTEX_DEFAULT, IPL_NONE);
135 1.40 riastrad
136 1.40 riastrad cv_init(&devsw_cv, "devsw");
137 1.11 ad }
138 1.2 gehenna
139 1.2 gehenna int
140 1.24 drochner devsw_attach(const char *devname,
141 1.24 drochner const struct bdevsw *bdev, devmajor_t *bmajor,
142 1.24 drochner const struct cdevsw *cdev, devmajor_t *cmajor)
143 1.2 gehenna {
144 1.2 gehenna struct devsw_conv *conv;
145 1.2 gehenna char *name;
146 1.2 gehenna int error, i;
147 1.2 gehenna
148 1.2 gehenna if (devname == NULL || cdev == NULL)
149 1.2 gehenna return (EINVAL);
150 1.2 gehenna
151 1.23 pooka mutex_enter(&device_lock);
152 1.11 ad
153 1.2 gehenna for (i = 0 ; i < max_devsw_convs ; i++) {
154 1.2 gehenna conv = &devsw_conv[i];
155 1.2 gehenna if (conv->d_name == NULL || strcmp(devname, conv->d_name) != 0)
156 1.2 gehenna continue;
157 1.2 gehenna
158 1.2 gehenna if (*bmajor < 0)
159 1.2 gehenna *bmajor = conv->d_bmajor;
160 1.2 gehenna if (*cmajor < 0)
161 1.2 gehenna *cmajor = conv->d_cmajor;
162 1.2 gehenna
163 1.11 ad if (*bmajor != conv->d_bmajor || *cmajor != conv->d_cmajor) {
164 1.11 ad error = EINVAL;
165 1.11 ad goto fail;
166 1.11 ad }
167 1.11 ad if ((*bmajor >= 0 && bdev == NULL) || *cmajor < 0) {
168 1.11 ad error = EINVAL;
169 1.11 ad goto fail;
170 1.11 ad }
171 1.2 gehenna
172 1.2 gehenna if ((*bmajor >= 0 && bdevsw[*bmajor] != NULL) ||
173 1.11 ad cdevsw[*cmajor] != NULL) {
174 1.11 ad error = EEXIST;
175 1.11 ad goto fail;
176 1.11 ad }
177 1.40 riastrad break;
178 1.2 gehenna }
179 1.2 gehenna
180 1.40 riastrad /*
181 1.40 riastrad * XXX This should allocate what it needs up front so we never
182 1.40 riastrad * need to flail around trying to unwind.
183 1.40 riastrad */
184 1.14 pooka error = bdevsw_attach(bdev, bmajor);
185 1.11 ad if (error != 0)
186 1.11 ad goto fail;
187 1.14 pooka error = cdevsw_attach(cdev, cmajor);
188 1.2 gehenna if (error != 0) {
189 1.11 ad devsw_detach_locked(bdev, NULL);
190 1.11 ad goto fail;
191 1.2 gehenna }
192 1.2 gehenna
193 1.40 riastrad /*
194 1.40 riastrad * If we already found a conv, we're done. Otherwise, find an
195 1.40 riastrad * empty slot or extend the table.
196 1.40 riastrad */
197 1.40 riastrad if (i == max_devsw_convs)
198 1.40 riastrad goto fail;
199 1.40 riastrad
200 1.2 gehenna for (i = 0 ; i < max_devsw_convs ; i++) {
201 1.2 gehenna if (devsw_conv[i].d_name == NULL)
202 1.2 gehenna break;
203 1.2 gehenna }
204 1.2 gehenna if (i == max_devsw_convs) {
205 1.2 gehenna struct devsw_conv *newptr;
206 1.33 matt int old_convs, new_convs;
207 1.2 gehenna
208 1.33 matt old_convs = max_devsw_convs;
209 1.33 matt new_convs = old_convs + 1;
210 1.2 gehenna
211 1.33 matt newptr = kmem_zalloc(new_convs * DEVSWCONV_SIZE, KM_NOSLEEP);
212 1.2 gehenna if (newptr == NULL) {
213 1.11 ad devsw_detach_locked(bdev, cdev);
214 1.11 ad error = ENOMEM;
215 1.11 ad goto fail;
216 1.2 gehenna }
217 1.33 matt newptr[old_convs].d_name = NULL;
218 1.33 matt newptr[old_convs].d_bmajor = -1;
219 1.33 matt newptr[old_convs].d_cmajor = -1;
220 1.33 matt memcpy(newptr, devsw_conv, old_convs * DEVSWCONV_SIZE);
221 1.2 gehenna if (devsw_conv != devsw_conv0)
222 1.33 matt kmem_free(devsw_conv, old_convs * DEVSWCONV_SIZE);
223 1.2 gehenna devsw_conv = newptr;
224 1.33 matt max_devsw_convs = new_convs;
225 1.2 gehenna }
226 1.2 gehenna
227 1.38 christos name = kmem_strdupsize(devname, NULL, KM_NOSLEEP);
228 1.2 gehenna if (name == NULL) {
229 1.11 ad devsw_detach_locked(bdev, cdev);
230 1.25 enami error = ENOMEM;
231 1.11 ad goto fail;
232 1.2 gehenna }
233 1.2 gehenna
234 1.2 gehenna devsw_conv[i].d_name = name;
235 1.2 gehenna devsw_conv[i].d_bmajor = *bmajor;
236 1.2 gehenna devsw_conv[i].d_cmajor = *cmajor;
237 1.2 gehenna
238 1.23 pooka mutex_exit(&device_lock);
239 1.2 gehenna return (0);
240 1.11 ad fail:
241 1.23 pooka mutex_exit(&device_lock);
242 1.11 ad return (error);
243 1.2 gehenna }
244 1.2 gehenna
245 1.2 gehenna static int
246 1.24 drochner bdevsw_attach(const struct bdevsw *devsw, devmajor_t *devmajor)
247 1.2 gehenna {
248 1.40 riastrad const struct bdevsw **newbdevsw = NULL;
249 1.40 riastrad struct devswref *newbdevswref = NULL;
250 1.40 riastrad struct localcount *lc;
251 1.24 drochner devmajor_t bmajor;
252 1.24 drochner int i;
253 1.2 gehenna
254 1.23 pooka KASSERT(mutex_owned(&device_lock));
255 1.11 ad
256 1.2 gehenna if (devsw == NULL)
257 1.2 gehenna return (0);
258 1.2 gehenna
259 1.2 gehenna if (*devmajor < 0) {
260 1.2 gehenna for (bmajor = sys_bdevsws ; bmajor < max_bdevsws ; bmajor++) {
261 1.2 gehenna if (bdevsw[bmajor] != NULL)
262 1.2 gehenna continue;
263 1.2 gehenna for (i = 0 ; i < max_devsw_convs ; i++) {
264 1.2 gehenna if (devsw_conv[i].d_bmajor == bmajor)
265 1.2 gehenna break;
266 1.2 gehenna }
267 1.2 gehenna if (i != max_devsw_convs)
268 1.2 gehenna continue;
269 1.2 gehenna break;
270 1.2 gehenna }
271 1.3 gehenna *devmajor = bmajor;
272 1.2 gehenna }
273 1.11 ad
274 1.2 gehenna if (*devmajor >= MAXDEVSW) {
275 1.37 pgoyette printf("%s: block majors exhausted", __func__);
276 1.2 gehenna return (ENOMEM);
277 1.2 gehenna }
278 1.2 gehenna
279 1.40 riastrad if (bdevswref == NULL) {
280 1.40 riastrad newbdevswref = kmem_zalloc(MAXDEVSW * sizeof(newbdevswref[0]),
281 1.40 riastrad KM_NOSLEEP);
282 1.40 riastrad if (newbdevswref == NULL)
283 1.40 riastrad return ENOMEM;
284 1.40 riastrad atomic_store_release(&bdevswref, newbdevswref);
285 1.40 riastrad }
286 1.40 riastrad
287 1.2 gehenna if (*devmajor >= max_bdevsws) {
288 1.11 ad KASSERT(bdevsw == bdevsw0);
289 1.40 riastrad newbdevsw = kmem_zalloc(MAXDEVSW * sizeof(newbdevsw[0]),
290 1.40 riastrad KM_NOSLEEP);
291 1.40 riastrad if (newbdevsw == NULL)
292 1.40 riastrad return ENOMEM;
293 1.40 riastrad memcpy(newbdevsw, bdevsw, max_bdevsws * sizeof(bdevsw[0]));
294 1.40 riastrad atomic_store_release(&bdevsw, newbdevsw);
295 1.40 riastrad atomic_store_release(&max_bdevsws, MAXDEVSW);
296 1.2 gehenna }
297 1.2 gehenna
298 1.2 gehenna if (bdevsw[*devmajor] != NULL)
299 1.2 gehenna return (EEXIST);
300 1.2 gehenna
301 1.40 riastrad KASSERT(bdevswref[*devmajor].dr_lc == NULL);
302 1.40 riastrad lc = kmem_zalloc(sizeof(*lc), KM_SLEEP);
303 1.40 riastrad localcount_init(lc);
304 1.40 riastrad bdevswref[*devmajor].dr_lc = lc;
305 1.40 riastrad
306 1.40 riastrad atomic_store_release(&bdevsw[*devmajor], devsw);
307 1.2 gehenna
308 1.2 gehenna return (0);
309 1.2 gehenna }
310 1.2 gehenna
311 1.2 gehenna static int
312 1.24 drochner cdevsw_attach(const struct cdevsw *devsw, devmajor_t *devmajor)
313 1.2 gehenna {
314 1.40 riastrad const struct cdevsw **newcdevsw = NULL;
315 1.40 riastrad struct devswref *newcdevswref = NULL;
316 1.40 riastrad struct localcount *lc;
317 1.24 drochner devmajor_t cmajor;
318 1.24 drochner int i;
319 1.2 gehenna
320 1.23 pooka KASSERT(mutex_owned(&device_lock));
321 1.11 ad
322 1.2 gehenna if (*devmajor < 0) {
323 1.2 gehenna for (cmajor = sys_cdevsws ; cmajor < max_cdevsws ; cmajor++) {
324 1.2 gehenna if (cdevsw[cmajor] != NULL)
325 1.2 gehenna continue;
326 1.2 gehenna for (i = 0 ; i < max_devsw_convs ; i++) {
327 1.2 gehenna if (devsw_conv[i].d_cmajor == cmajor)
328 1.2 gehenna break;
329 1.2 gehenna }
330 1.2 gehenna if (i != max_devsw_convs)
331 1.2 gehenna continue;
332 1.2 gehenna break;
333 1.2 gehenna }
334 1.3 gehenna *devmajor = cmajor;
335 1.2 gehenna }
336 1.11 ad
337 1.2 gehenna if (*devmajor >= MAXDEVSW) {
338 1.37 pgoyette printf("%s: character majors exhausted", __func__);
339 1.2 gehenna return (ENOMEM);
340 1.2 gehenna }
341 1.2 gehenna
342 1.40 riastrad if (cdevswref == NULL) {
343 1.40 riastrad newcdevswref = kmem_zalloc(MAXDEVSW * sizeof(newcdevswref[0]),
344 1.40 riastrad KM_NOSLEEP);
345 1.40 riastrad if (newcdevswref == NULL)
346 1.40 riastrad return ENOMEM;
347 1.40 riastrad atomic_store_release(&cdevswref, newcdevswref);
348 1.40 riastrad }
349 1.40 riastrad
350 1.2 gehenna if (*devmajor >= max_cdevsws) {
351 1.11 ad KASSERT(cdevsw == cdevsw0);
352 1.40 riastrad newcdevsw = kmem_zalloc(MAXDEVSW * sizeof(newcdevsw[0]),
353 1.40 riastrad KM_NOSLEEP);
354 1.40 riastrad if (newcdevsw == NULL)
355 1.40 riastrad return ENOMEM;
356 1.40 riastrad memcpy(newcdevsw, cdevsw, max_cdevsws * sizeof(cdevsw[0]));
357 1.40 riastrad atomic_store_release(&cdevsw, newcdevsw);
358 1.40 riastrad atomic_store_release(&max_cdevsws, MAXDEVSW);
359 1.2 gehenna }
360 1.2 gehenna
361 1.2 gehenna if (cdevsw[*devmajor] != NULL)
362 1.2 gehenna return (EEXIST);
363 1.2 gehenna
364 1.40 riastrad KASSERT(cdevswref[*devmajor].dr_lc == NULL);
365 1.40 riastrad lc = kmem_zalloc(sizeof(*lc), KM_SLEEP);
366 1.40 riastrad localcount_init(lc);
367 1.40 riastrad cdevswref[*devmajor].dr_lc = lc;
368 1.40 riastrad
369 1.40 riastrad atomic_store_release(&cdevsw[*devmajor], devsw);
370 1.2 gehenna
371 1.2 gehenna return (0);
372 1.2 gehenna }
373 1.2 gehenna
374 1.11 ad static void
375 1.11 ad devsw_detach_locked(const struct bdevsw *bdev, const struct cdevsw *cdev)
376 1.2 gehenna {
377 1.40 riastrad int bi, ci = -1/*XXXGCC*/;
378 1.2 gehenna
379 1.23 pooka KASSERT(mutex_owned(&device_lock));
380 1.11 ad
381 1.40 riastrad /* Prevent new references. */
382 1.2 gehenna if (bdev != NULL) {
383 1.40 riastrad for (bi = 0; bi < max_bdevsws; bi++) {
384 1.40 riastrad if (bdevsw[bi] != bdev)
385 1.2 gehenna continue;
386 1.40 riastrad atomic_store_relaxed(&bdevsw[bi], NULL);
387 1.2 gehenna break;
388 1.2 gehenna }
389 1.40 riastrad KASSERT(bi < max_bdevsws);
390 1.2 gehenna }
391 1.2 gehenna if (cdev != NULL) {
392 1.40 riastrad for (ci = 0; ci < max_cdevsws; ci++) {
393 1.40 riastrad if (cdevsw[ci] != cdev)
394 1.2 gehenna continue;
395 1.40 riastrad atomic_store_relaxed(&cdevsw[ci], NULL);
396 1.2 gehenna break;
397 1.2 gehenna }
398 1.40 riastrad KASSERT(ci < max_cdevsws);
399 1.40 riastrad }
400 1.40 riastrad
401 1.40 riastrad if (bdev == NULL && cdev == NULL) /* XXX possible? */
402 1.40 riastrad return;
403 1.40 riastrad
404 1.40 riastrad /*
405 1.40 riastrad * Wait for all bdevsw_lookup_acquire, cdevsw_lookup_acquire
406 1.40 riastrad * calls to notice that the devsw is gone.
407 1.40 riastrad *
408 1.40 riastrad * XXX Despite the use of the pserialize_read_enter/exit API
409 1.40 riastrad * elsewhere in this file, we use xc_barrier here instead of
410 1.40 riastrad * pserialize_perform -- because devsw_init is too early for
411 1.40 riastrad * pserialize_create. Either pserialize_create should be made
412 1.40 riastrad * to work earlier, or it should be nixed altogether. Until
413 1.40 riastrad * that is fixed, xc_barrier will serve the same purpose.
414 1.40 riastrad */
415 1.40 riastrad xc_barrier(0);
416 1.40 riastrad
417 1.40 riastrad /*
418 1.40 riastrad * Wait for all references to drain. It is the caller's
419 1.40 riastrad * responsibility to ensure that at this point, there are no
420 1.40 riastrad * extant open instances and all new d_open calls will fail.
421 1.40 riastrad *
422 1.40 riastrad * Note that localcount_drain may release and reacquire
423 1.40 riastrad * device_lock.
424 1.40 riastrad */
425 1.40 riastrad if (bdev != NULL) {
426 1.40 riastrad localcount_drain(bdevswref[bi].dr_lc,
427 1.40 riastrad &devsw_cv, &device_lock);
428 1.40 riastrad localcount_fini(bdevswref[bi].dr_lc);
429 1.40 riastrad kmem_free(bdevswref[bi].dr_lc, sizeof(*bdevswref[bi].dr_lc));
430 1.40 riastrad bdevswref[bi].dr_lc = NULL;
431 1.40 riastrad }
432 1.40 riastrad if (cdev != NULL) {
433 1.40 riastrad localcount_drain(cdevswref[ci].dr_lc,
434 1.40 riastrad &devsw_cv, &device_lock);
435 1.40 riastrad localcount_fini(cdevswref[ci].dr_lc);
436 1.40 riastrad kmem_free(cdevswref[ci].dr_lc, sizeof(*cdevswref[ci].dr_lc));
437 1.40 riastrad cdevswref[ci].dr_lc = NULL;
438 1.2 gehenna }
439 1.2 gehenna }
440 1.2 gehenna
441 1.39 riastrad void
442 1.11 ad devsw_detach(const struct bdevsw *bdev, const struct cdevsw *cdev)
443 1.11 ad {
444 1.11 ad
445 1.23 pooka mutex_enter(&device_lock);
446 1.11 ad devsw_detach_locked(bdev, cdev);
447 1.23 pooka mutex_exit(&device_lock);
448 1.11 ad }
449 1.11 ad
450 1.11 ad /*
451 1.11 ad * Look up a block device by number.
452 1.11 ad *
453 1.11 ad * => Caller must ensure that the device is attached.
454 1.11 ad */
455 1.2 gehenna const struct bdevsw *
456 1.2 gehenna bdevsw_lookup(dev_t dev)
457 1.2 gehenna {
458 1.24 drochner devmajor_t bmajor;
459 1.2 gehenna
460 1.2 gehenna if (dev == NODEV)
461 1.2 gehenna return (NULL);
462 1.2 gehenna bmajor = major(dev);
463 1.40 riastrad if (bmajor < 0 || bmajor >= atomic_load_relaxed(&max_bdevsws))
464 1.2 gehenna return (NULL);
465 1.2 gehenna
466 1.40 riastrad return atomic_load_consume(&bdevsw)[bmajor];
467 1.40 riastrad }
468 1.40 riastrad
469 1.40 riastrad static const struct bdevsw *
470 1.40 riastrad bdevsw_lookup_acquire(dev_t dev, struct localcount **lcp)
471 1.40 riastrad {
472 1.40 riastrad devmajor_t bmajor;
473 1.40 riastrad const struct bdevsw *bdev = NULL, *const *curbdevsw;
474 1.40 riastrad struct devswref *curbdevswref;
475 1.40 riastrad int s;
476 1.40 riastrad
477 1.40 riastrad if (dev == NODEV)
478 1.40 riastrad return NULL;
479 1.40 riastrad bmajor = major(dev);
480 1.40 riastrad if (bmajor < 0)
481 1.40 riastrad return NULL;
482 1.40 riastrad
483 1.40 riastrad s = pserialize_read_enter();
484 1.40 riastrad
485 1.40 riastrad /*
486 1.40 riastrad * max_bdevsws never goes down, so it is safe to rely on this
487 1.40 riastrad * condition without any locking for the array access below.
488 1.40 riastrad * Test sys_bdevsws first so we can avoid the memory barrier in
489 1.40 riastrad * that case.
490 1.40 riastrad */
491 1.40 riastrad if (bmajor >= sys_bdevsws &&
492 1.40 riastrad bmajor >= atomic_load_acquire(&max_bdevsws))
493 1.40 riastrad goto out;
494 1.40 riastrad curbdevsw = atomic_load_consume(&bdevsw);
495 1.40 riastrad if ((bdev = atomic_load_consume(&curbdevsw[bmajor])) == NULL)
496 1.40 riastrad goto out;
497 1.40 riastrad
498 1.40 riastrad curbdevswref = atomic_load_consume(&bdevswref);
499 1.40 riastrad if (curbdevswref == NULL) {
500 1.40 riastrad *lcp = NULL;
501 1.40 riastrad } else if ((*lcp = curbdevswref[bmajor].dr_lc) != NULL) {
502 1.40 riastrad localcount_acquire(*lcp);
503 1.40 riastrad }
504 1.40 riastrad out:
505 1.40 riastrad pserialize_read_exit(s);
506 1.40 riastrad return bdev;
507 1.40 riastrad }
508 1.40 riastrad
509 1.40 riastrad static void
510 1.40 riastrad bdevsw_release(const struct bdevsw *bdev, struct localcount *lc)
511 1.40 riastrad {
512 1.40 riastrad
513 1.40 riastrad if (lc == NULL)
514 1.40 riastrad return;
515 1.40 riastrad localcount_release(lc, &devsw_cv, &device_lock);
516 1.2 gehenna }
517 1.2 gehenna
518 1.11 ad /*
519 1.11 ad * Look up a character device by number.
520 1.11 ad *
521 1.11 ad * => Caller must ensure that the device is attached.
522 1.11 ad */
523 1.2 gehenna const struct cdevsw *
524 1.2 gehenna cdevsw_lookup(dev_t dev)
525 1.2 gehenna {
526 1.24 drochner devmajor_t cmajor;
527 1.2 gehenna
528 1.2 gehenna if (dev == NODEV)
529 1.2 gehenna return (NULL);
530 1.2 gehenna cmajor = major(dev);
531 1.40 riastrad if (cmajor < 0 || cmajor >= atomic_load_relaxed(&max_cdevsws))
532 1.2 gehenna return (NULL);
533 1.2 gehenna
534 1.40 riastrad return atomic_load_consume(&cdevsw)[cmajor];
535 1.40 riastrad }
536 1.40 riastrad
537 1.40 riastrad static const struct cdevsw *
538 1.40 riastrad cdevsw_lookup_acquire(dev_t dev, struct localcount **lcp)
539 1.40 riastrad {
540 1.40 riastrad devmajor_t cmajor;
541 1.40 riastrad const struct cdevsw *cdev = NULL, *const *curcdevsw;
542 1.40 riastrad struct devswref *curcdevswref;
543 1.40 riastrad int s;
544 1.40 riastrad
545 1.40 riastrad if (dev == NODEV)
546 1.40 riastrad return NULL;
547 1.40 riastrad cmajor = major(dev);
548 1.40 riastrad if (cmajor < 0)
549 1.40 riastrad return NULL;
550 1.40 riastrad
551 1.40 riastrad s = pserialize_read_enter();
552 1.40 riastrad
553 1.40 riastrad /*
554 1.40 riastrad * max_cdevsws never goes down, so it is safe to rely on this
555 1.40 riastrad * condition without any locking for the array access below.
556 1.40 riastrad * Test sys_cdevsws first so we can avoid the memory barrier in
557 1.40 riastrad * that case.
558 1.40 riastrad */
559 1.40 riastrad if (cmajor >= sys_cdevsws &&
560 1.40 riastrad cmajor >= atomic_load_acquire(&max_cdevsws))
561 1.40 riastrad goto out;
562 1.40 riastrad curcdevsw = atomic_load_consume(&cdevsw);
563 1.40 riastrad if ((cdev = atomic_load_consume(&curcdevsw[cmajor])) == NULL)
564 1.40 riastrad goto out;
565 1.40 riastrad
566 1.40 riastrad curcdevswref = atomic_load_consume(&cdevswref);
567 1.40 riastrad if (curcdevswref == NULL) {
568 1.40 riastrad *lcp = NULL;
569 1.40 riastrad } else if ((*lcp = curcdevswref[cmajor].dr_lc) != NULL) {
570 1.40 riastrad localcount_acquire(*lcp);
571 1.40 riastrad }
572 1.40 riastrad out:
573 1.40 riastrad pserialize_read_exit(s);
574 1.40 riastrad return cdev;
575 1.40 riastrad }
576 1.40 riastrad
577 1.40 riastrad static void
578 1.40 riastrad cdevsw_release(const struct cdevsw *cdev, struct localcount *lc)
579 1.40 riastrad {
580 1.40 riastrad
581 1.40 riastrad if (lc == NULL)
582 1.40 riastrad return;
583 1.40 riastrad localcount_release(lc, &devsw_cv, &device_lock);
584 1.2 gehenna }
585 1.2 gehenna
586 1.11 ad /*
587 1.11 ad * Look up a block device by reference to its operations set.
588 1.11 ad *
589 1.11 ad * => Caller must ensure that the device is not detached, and therefore
590 1.11 ad * that the returned major is still valid when dereferenced.
591 1.11 ad */
592 1.24 drochner devmajor_t
593 1.2 gehenna bdevsw_lookup_major(const struct bdevsw *bdev)
594 1.2 gehenna {
595 1.40 riastrad const struct bdevsw *const *curbdevsw;
596 1.40 riastrad devmajor_t bmajor, bmax;
597 1.2 gehenna
598 1.40 riastrad bmax = atomic_load_acquire(&max_bdevsws);
599 1.40 riastrad curbdevsw = atomic_load_consume(&bdevsw);
600 1.40 riastrad for (bmajor = 0; bmajor < bmax; bmajor++) {
601 1.40 riastrad if (atomic_load_relaxed(&curbdevsw[bmajor]) == bdev)
602 1.2 gehenna return (bmajor);
603 1.2 gehenna }
604 1.2 gehenna
605 1.24 drochner return (NODEVMAJOR);
606 1.2 gehenna }
607 1.2 gehenna
608 1.11 ad /*
609 1.11 ad * Look up a character device by reference to its operations set.
610 1.11 ad *
611 1.11 ad * => Caller must ensure that the device is not detached, and therefore
612 1.11 ad * that the returned major is still valid when dereferenced.
613 1.11 ad */
614 1.24 drochner devmajor_t
615 1.2 gehenna cdevsw_lookup_major(const struct cdevsw *cdev)
616 1.2 gehenna {
617 1.40 riastrad const struct cdevsw *const *curcdevsw;
618 1.40 riastrad devmajor_t cmajor, cmax;
619 1.2 gehenna
620 1.40 riastrad cmax = atomic_load_acquire(&max_cdevsws);
621 1.40 riastrad curcdevsw = atomic_load_consume(&cdevsw);
622 1.40 riastrad for (cmajor = 0; cmajor < cmax; cmajor++) {
623 1.40 riastrad if (atomic_load_relaxed(&curcdevsw[cmajor]) == cdev)
624 1.2 gehenna return (cmajor);
625 1.2 gehenna }
626 1.2 gehenna
627 1.24 drochner return (NODEVMAJOR);
628 1.2 gehenna }
629 1.2 gehenna
630 1.2 gehenna /*
631 1.2 gehenna * Convert from block major number to name.
632 1.11 ad *
633 1.11 ad * => Caller must ensure that the device is not detached, and therefore
634 1.11 ad * that the name pointer is still valid when dereferenced.
635 1.2 gehenna */
636 1.2 gehenna const char *
637 1.24 drochner devsw_blk2name(devmajor_t bmajor)
638 1.2 gehenna {
639 1.11 ad const char *name;
640 1.24 drochner devmajor_t cmajor;
641 1.24 drochner int i;
642 1.2 gehenna
643 1.11 ad name = NULL;
644 1.11 ad cmajor = -1;
645 1.11 ad
646 1.23 pooka mutex_enter(&device_lock);
647 1.11 ad if (bmajor < 0 || bmajor >= max_bdevsws || bdevsw[bmajor] == NULL) {
648 1.23 pooka mutex_exit(&device_lock);
649 1.2 gehenna return (NULL);
650 1.2 gehenna }
651 1.11 ad for (i = 0 ; i < max_devsw_convs; i++) {
652 1.11 ad if (devsw_conv[i].d_bmajor == bmajor) {
653 1.11 ad cmajor = devsw_conv[i].d_cmajor;
654 1.11 ad break;
655 1.11 ad }
656 1.11 ad }
657 1.11 ad if (cmajor >= 0 && cmajor < max_cdevsws && cdevsw[cmajor] != NULL)
658 1.11 ad name = devsw_conv[i].d_name;
659 1.23 pooka mutex_exit(&device_lock);
660 1.2 gehenna
661 1.11 ad return (name);
662 1.2 gehenna }
663 1.2 gehenna
664 1.2 gehenna /*
665 1.26 haad * Convert char major number to device driver name.
666 1.26 haad */
667 1.27 yamt const char *
668 1.26 haad cdevsw_getname(devmajor_t major)
669 1.26 haad {
670 1.26 haad const char *name;
671 1.26 haad int i;
672 1.26 haad
673 1.26 haad name = NULL;
674 1.26 haad
675 1.26 haad if (major < 0)
676 1.26 haad return (NULL);
677 1.26 haad
678 1.26 haad mutex_enter(&device_lock);
679 1.26 haad for (i = 0 ; i < max_devsw_convs; i++) {
680 1.26 haad if (devsw_conv[i].d_cmajor == major) {
681 1.26 haad name = devsw_conv[i].d_name;
682 1.26 haad break;
683 1.26 haad }
684 1.26 haad }
685 1.26 haad mutex_exit(&device_lock);
686 1.26 haad return (name);
687 1.26 haad }
688 1.26 haad
689 1.26 haad /*
690 1.26 haad * Convert block major number to device driver name.
691 1.26 haad */
692 1.27 yamt const char *
693 1.26 haad bdevsw_getname(devmajor_t major)
694 1.26 haad {
695 1.26 haad const char *name;
696 1.26 haad int i;
697 1.26 haad
698 1.26 haad name = NULL;
699 1.26 haad
700 1.26 haad if (major < 0)
701 1.26 haad return (NULL);
702 1.26 haad
703 1.26 haad mutex_enter(&device_lock);
704 1.26 haad for (i = 0 ; i < max_devsw_convs; i++) {
705 1.26 haad if (devsw_conv[i].d_bmajor == major) {
706 1.26 haad name = devsw_conv[i].d_name;
707 1.26 haad break;
708 1.26 haad }
709 1.26 haad }
710 1.26 haad mutex_exit(&device_lock);
711 1.26 haad return (name);
712 1.26 haad }
713 1.26 haad
714 1.26 haad /*
715 1.2 gehenna * Convert from device name to block major number.
716 1.11 ad *
717 1.11 ad * => Caller must ensure that the device is not detached, and therefore
718 1.11 ad * that the major number is still valid when dereferenced.
719 1.2 gehenna */
720 1.24 drochner devmajor_t
721 1.2 gehenna devsw_name2blk(const char *name, char *devname, size_t devnamelen)
722 1.2 gehenna {
723 1.2 gehenna struct devsw_conv *conv;
724 1.24 drochner devmajor_t bmajor;
725 1.24 drochner int i;
726 1.2 gehenna
727 1.2 gehenna if (name == NULL)
728 1.24 drochner return (NODEVMAJOR);
729 1.2 gehenna
730 1.23 pooka mutex_enter(&device_lock);
731 1.2 gehenna for (i = 0 ; i < max_devsw_convs ; i++) {
732 1.5 mrg size_t len;
733 1.5 mrg
734 1.2 gehenna conv = &devsw_conv[i];
735 1.2 gehenna if (conv->d_name == NULL)
736 1.2 gehenna continue;
737 1.5 mrg len = strlen(conv->d_name);
738 1.5 mrg if (strncmp(conv->d_name, name, len) != 0)
739 1.5 mrg continue;
740 1.5 mrg if (*(name +len) && !isdigit(*(name + len)))
741 1.2 gehenna continue;
742 1.2 gehenna bmajor = conv->d_bmajor;
743 1.2 gehenna if (bmajor < 0 || bmajor >= max_bdevsws ||
744 1.2 gehenna bdevsw[bmajor] == NULL)
745 1.5 mrg break;
746 1.2 gehenna if (devname != NULL) {
747 1.2 gehenna #ifdef DEVSW_DEBUG
748 1.2 gehenna if (strlen(conv->d_name) >= devnamelen)
749 1.37 pgoyette printf("%s: too short buffer", __func__);
750 1.2 gehenna #endif /* DEVSW_DEBUG */
751 1.4 tsutsui strncpy(devname, conv->d_name, devnamelen);
752 1.2 gehenna devname[devnamelen - 1] = '\0';
753 1.2 gehenna }
754 1.23 pooka mutex_exit(&device_lock);
755 1.2 gehenna return (bmajor);
756 1.2 gehenna }
757 1.2 gehenna
758 1.23 pooka mutex_exit(&device_lock);
759 1.24 drochner return (NODEVMAJOR);
760 1.2 gehenna }
761 1.2 gehenna
762 1.2 gehenna /*
763 1.16 plunky * Convert from device name to char major number.
764 1.16 plunky *
765 1.16 plunky * => Caller must ensure that the device is not detached, and therefore
766 1.16 plunky * that the major number is still valid when dereferenced.
767 1.16 plunky */
768 1.24 drochner devmajor_t
769 1.16 plunky devsw_name2chr(const char *name, char *devname, size_t devnamelen)
770 1.16 plunky {
771 1.16 plunky struct devsw_conv *conv;
772 1.24 drochner devmajor_t cmajor;
773 1.24 drochner int i;
774 1.16 plunky
775 1.16 plunky if (name == NULL)
776 1.24 drochner return (NODEVMAJOR);
777 1.16 plunky
778 1.23 pooka mutex_enter(&device_lock);
779 1.16 plunky for (i = 0 ; i < max_devsw_convs ; i++) {
780 1.16 plunky size_t len;
781 1.16 plunky
782 1.16 plunky conv = &devsw_conv[i];
783 1.16 plunky if (conv->d_name == NULL)
784 1.16 plunky continue;
785 1.16 plunky len = strlen(conv->d_name);
786 1.16 plunky if (strncmp(conv->d_name, name, len) != 0)
787 1.16 plunky continue;
788 1.16 plunky if (*(name +len) && !isdigit(*(name + len)))
789 1.16 plunky continue;
790 1.16 plunky cmajor = conv->d_cmajor;
791 1.16 plunky if (cmajor < 0 || cmajor >= max_cdevsws ||
792 1.16 plunky cdevsw[cmajor] == NULL)
793 1.16 plunky break;
794 1.16 plunky if (devname != NULL) {
795 1.16 plunky #ifdef DEVSW_DEBUG
796 1.16 plunky if (strlen(conv->d_name) >= devnamelen)
797 1.37 pgoyette printf("%s: too short buffer", __func__);
798 1.16 plunky #endif /* DEVSW_DEBUG */
799 1.16 plunky strncpy(devname, conv->d_name, devnamelen);
800 1.16 plunky devname[devnamelen - 1] = '\0';
801 1.16 plunky }
802 1.23 pooka mutex_exit(&device_lock);
803 1.16 plunky return (cmajor);
804 1.16 plunky }
805 1.16 plunky
806 1.23 pooka mutex_exit(&device_lock);
807 1.24 drochner return (NODEVMAJOR);
808 1.16 plunky }
809 1.16 plunky
810 1.16 plunky /*
811 1.2 gehenna * Convert from character dev_t to block dev_t.
812 1.11 ad *
813 1.11 ad * => Caller must ensure that the device is not detached, and therefore
814 1.11 ad * that the major number is still valid when dereferenced.
815 1.2 gehenna */
816 1.2 gehenna dev_t
817 1.2 gehenna devsw_chr2blk(dev_t cdev)
818 1.2 gehenna {
819 1.24 drochner devmajor_t bmajor, cmajor;
820 1.24 drochner int i;
821 1.11 ad dev_t rv;
822 1.2 gehenna
823 1.2 gehenna cmajor = major(cdev);
824 1.24 drochner bmajor = NODEVMAJOR;
825 1.11 ad rv = NODEV;
826 1.2 gehenna
827 1.23 pooka mutex_enter(&device_lock);
828 1.11 ad if (cmajor < 0 || cmajor >= max_cdevsws || cdevsw[cmajor] == NULL) {
829 1.23 pooka mutex_exit(&device_lock);
830 1.11 ad return (NODEV);
831 1.11 ad }
832 1.2 gehenna for (i = 0 ; i < max_devsw_convs ; i++) {
833 1.11 ad if (devsw_conv[i].d_cmajor == cmajor) {
834 1.11 ad bmajor = devsw_conv[i].d_bmajor;
835 1.11 ad break;
836 1.11 ad }
837 1.2 gehenna }
838 1.11 ad if (bmajor >= 0 && bmajor < max_bdevsws && bdevsw[bmajor] != NULL)
839 1.11 ad rv = makedev(bmajor, minor(cdev));
840 1.23 pooka mutex_exit(&device_lock);
841 1.2 gehenna
842 1.11 ad return (rv);
843 1.2 gehenna }
844 1.2 gehenna
845 1.2 gehenna /*
846 1.2 gehenna * Convert from block dev_t to character dev_t.
847 1.11 ad *
848 1.11 ad * => Caller must ensure that the device is not detached, and therefore
849 1.11 ad * that the major number is still valid when dereferenced.
850 1.2 gehenna */
851 1.2 gehenna dev_t
852 1.2 gehenna devsw_blk2chr(dev_t bdev)
853 1.2 gehenna {
854 1.24 drochner devmajor_t bmajor, cmajor;
855 1.24 drochner int i;
856 1.11 ad dev_t rv;
857 1.2 gehenna
858 1.11 ad bmajor = major(bdev);
859 1.24 drochner cmajor = NODEVMAJOR;
860 1.11 ad rv = NODEV;
861 1.11 ad
862 1.23 pooka mutex_enter(&device_lock);
863 1.11 ad if (bmajor < 0 || bmajor >= max_bdevsws || bdevsw[bmajor] == NULL) {
864 1.23 pooka mutex_exit(&device_lock);
865 1.2 gehenna return (NODEV);
866 1.11 ad }
867 1.11 ad for (i = 0 ; i < max_devsw_convs ; i++) {
868 1.11 ad if (devsw_conv[i].d_bmajor == bmajor) {
869 1.11 ad cmajor = devsw_conv[i].d_cmajor;
870 1.11 ad break;
871 1.11 ad }
872 1.11 ad }
873 1.11 ad if (cmajor >= 0 && cmajor < max_cdevsws && cdevsw[cmajor] != NULL)
874 1.11 ad rv = makedev(cmajor, minor(bdev));
875 1.23 pooka mutex_exit(&device_lock);
876 1.2 gehenna
877 1.11 ad return (rv);
878 1.11 ad }
879 1.11 ad
880 1.11 ad /*
881 1.11 ad * Device access methods.
882 1.11 ad */
883 1.11 ad
884 1.11 ad #define DEV_LOCK(d) \
885 1.17 ad if ((mpflag = (d->d_flag & D_MPSAFE)) == 0) { \
886 1.17 ad KERNEL_LOCK(1, NULL); \
887 1.11 ad }
888 1.2 gehenna
889 1.11 ad #define DEV_UNLOCK(d) \
890 1.17 ad if (mpflag == 0) { \
891 1.17 ad KERNEL_UNLOCK_ONE(NULL); \
892 1.2 gehenna }
893 1.2 gehenna
894 1.11 ad int
895 1.11 ad bdev_open(dev_t dev, int flag, int devtype, lwp_t *l)
896 1.11 ad {
897 1.11 ad const struct bdevsw *d;
898 1.40 riastrad struct localcount *lc;
899 1.41 riastrad device_t dv = NULL/*XXXGCC*/;
900 1.41 riastrad int unit, rv, mpflag;
901 1.11 ad
902 1.40 riastrad d = bdevsw_lookup_acquire(dev, &lc);
903 1.11 ad if (d == NULL)
904 1.11 ad return ENXIO;
905 1.11 ad
906 1.41 riastrad if (d->d_devtounit) {
907 1.41 riastrad /*
908 1.41 riastrad * If the device node corresponds to an autoconf device
909 1.41 riastrad * instance, acquire a reference to it so that during
910 1.41 riastrad * d_open, device_lookup is stable.
911 1.41 riastrad *
912 1.41 riastrad * XXX This should also arrange to instantiate cloning
913 1.41 riastrad * pseudo-devices if appropriate, but that requires
914 1.41 riastrad * reviewing them all to find and verify a common
915 1.41 riastrad * pattern.
916 1.41 riastrad */
917 1.41 riastrad if ((unit = (*d->d_devtounit)(dev)) == -1)
918 1.41 riastrad return ENXIO;
919 1.41 riastrad if ((dv = device_lookup_acquire(d->d_cfdriver, unit)) == NULL)
920 1.41 riastrad return ENXIO;
921 1.41 riastrad }
922 1.41 riastrad
923 1.11 ad DEV_LOCK(d);
924 1.11 ad rv = (*d->d_open)(dev, flag, devtype, l);
925 1.11 ad DEV_UNLOCK(d);
926 1.11 ad
927 1.41 riastrad if (d->d_devtounit) {
928 1.41 riastrad device_release(dv);
929 1.41 riastrad }
930 1.41 riastrad
931 1.40 riastrad bdevsw_release(d, lc);
932 1.40 riastrad
933 1.11 ad return rv;
934 1.11 ad }
935 1.11 ad
936 1.11 ad int
937 1.11 ad bdev_close(dev_t dev, int flag, int devtype, lwp_t *l)
938 1.11 ad {
939 1.11 ad const struct bdevsw *d;
940 1.17 ad int rv, mpflag;
941 1.11 ad
942 1.11 ad if ((d = bdevsw_lookup(dev)) == NULL)
943 1.11 ad return ENXIO;
944 1.11 ad
945 1.11 ad DEV_LOCK(d);
946 1.11 ad rv = (*d->d_close)(dev, flag, devtype, l);
947 1.11 ad DEV_UNLOCK(d);
948 1.11 ad
949 1.11 ad return rv;
950 1.11 ad }
951 1.11 ad
952 1.34 riz SDT_PROVIDER_DECLARE(io);
953 1.34 riz SDT_PROBE_DEFINE1(io, kernel, , start, "struct buf *"/*bp*/);
954 1.34 riz
955 1.11 ad void
956 1.11 ad bdev_strategy(struct buf *bp)
957 1.11 ad {
958 1.11 ad const struct bdevsw *d;
959 1.17 ad int mpflag;
960 1.11 ad
961 1.34 riz SDT_PROBE1(io, kernel, , start, bp);
962 1.34 riz
963 1.28 jmcneill if ((d = bdevsw_lookup(bp->b_dev)) == NULL) {
964 1.28 jmcneill bp->b_error = ENXIO;
965 1.28 jmcneill bp->b_resid = bp->b_bcount;
966 1.31 pooka biodone_vfs(bp); /* biodone() iff vfs present */
967 1.28 jmcneill return;
968 1.28 jmcneill }
969 1.11 ad
970 1.11 ad DEV_LOCK(d);
971 1.11 ad (*d->d_strategy)(bp);
972 1.11 ad DEV_UNLOCK(d);
973 1.11 ad }
974 1.11 ad
975 1.11 ad int
976 1.11 ad bdev_ioctl(dev_t dev, u_long cmd, void *data, int flag, lwp_t *l)
977 1.11 ad {
978 1.11 ad const struct bdevsw *d;
979 1.17 ad int rv, mpflag;
980 1.11 ad
981 1.11 ad if ((d = bdevsw_lookup(dev)) == NULL)
982 1.11 ad return ENXIO;
983 1.11 ad
984 1.11 ad DEV_LOCK(d);
985 1.11 ad rv = (*d->d_ioctl)(dev, cmd, data, flag, l);
986 1.11 ad DEV_UNLOCK(d);
987 1.11 ad
988 1.11 ad return rv;
989 1.11 ad }
990 1.11 ad
991 1.11 ad int
992 1.11 ad bdev_dump(dev_t dev, daddr_t addr, void *data, size_t sz)
993 1.11 ad {
994 1.11 ad const struct bdevsw *d;
995 1.11 ad int rv;
996 1.11 ad
997 1.11 ad /*
998 1.11 ad * Dump can be called without the device open. Since it can
999 1.11 ad * currently only be called with the system paused (and in a
1000 1.11 ad * potentially unstable state), we don't perform any locking.
1001 1.11 ad */
1002 1.11 ad if ((d = bdevsw_lookup(dev)) == NULL)
1003 1.11 ad return ENXIO;
1004 1.11 ad
1005 1.11 ad /* DEV_LOCK(d); */
1006 1.11 ad rv = (*d->d_dump)(dev, addr, data, sz);
1007 1.11 ad /* DEV_UNLOCK(d); */
1008 1.11 ad
1009 1.11 ad return rv;
1010 1.11 ad }
1011 1.11 ad
1012 1.11 ad int
1013 1.35 nat bdev_flags(dev_t dev)
1014 1.35 nat {
1015 1.35 nat const struct bdevsw *d;
1016 1.35 nat
1017 1.35 nat if ((d = bdevsw_lookup(dev)) == NULL)
1018 1.35 nat return 0;
1019 1.35 nat return d->d_flag & ~D_TYPEMASK;
1020 1.35 nat }
1021 1.35 nat
1022 1.35 nat int
1023 1.11 ad bdev_type(dev_t dev)
1024 1.11 ad {
1025 1.11 ad const struct bdevsw *d;
1026 1.11 ad
1027 1.11 ad if ((d = bdevsw_lookup(dev)) == NULL)
1028 1.11 ad return D_OTHER;
1029 1.11 ad return d->d_flag & D_TYPEMASK;
1030 1.11 ad }
1031 1.11 ad
1032 1.11 ad int
1033 1.29 mrg bdev_size(dev_t dev)
1034 1.29 mrg {
1035 1.29 mrg const struct bdevsw *d;
1036 1.29 mrg int rv, mpflag = 0;
1037 1.29 mrg
1038 1.29 mrg if ((d = bdevsw_lookup(dev)) == NULL ||
1039 1.29 mrg d->d_psize == NULL)
1040 1.29 mrg return -1;
1041 1.29 mrg
1042 1.29 mrg /*
1043 1.29 mrg * Don't to try lock the device if we're dumping.
1044 1.30 mrg * XXX: is there a better way to test this?
1045 1.29 mrg */
1046 1.29 mrg if ((boothowto & RB_DUMP) == 0)
1047 1.29 mrg DEV_LOCK(d);
1048 1.29 mrg rv = (*d->d_psize)(dev);
1049 1.29 mrg if ((boothowto & RB_DUMP) == 0)
1050 1.29 mrg DEV_UNLOCK(d);
1051 1.29 mrg
1052 1.29 mrg return rv;
1053 1.29 mrg }
1054 1.29 mrg
1055 1.29 mrg int
1056 1.32 dholland bdev_discard(dev_t dev, off_t pos, off_t len)
1057 1.32 dholland {
1058 1.32 dholland const struct bdevsw *d;
1059 1.32 dholland int rv, mpflag;
1060 1.32 dholland
1061 1.32 dholland if ((d = bdevsw_lookup(dev)) == NULL)
1062 1.32 dholland return ENXIO;
1063 1.32 dholland
1064 1.32 dholland DEV_LOCK(d);
1065 1.32 dholland rv = (*d->d_discard)(dev, pos, len);
1066 1.32 dholland DEV_UNLOCK(d);
1067 1.32 dholland
1068 1.32 dholland return rv;
1069 1.32 dholland }
1070 1.32 dholland
1071 1.43 riastrad void
1072 1.43 riastrad bdev_detached(dev_t dev)
1073 1.43 riastrad {
1074 1.43 riastrad const struct bdevsw *d;
1075 1.43 riastrad device_t dv;
1076 1.43 riastrad int unit;
1077 1.43 riastrad
1078 1.43 riastrad if ((d = bdevsw_lookup(dev)) == NULL)
1079 1.43 riastrad return;
1080 1.43 riastrad if (d->d_devtounit == NULL)
1081 1.43 riastrad return;
1082 1.43 riastrad if ((unit = (*d->d_devtounit)(dev)) == -1)
1083 1.43 riastrad return;
1084 1.43 riastrad if ((dv = device_lookup(d->d_cfdriver, unit)) == NULL)
1085 1.43 riastrad return;
1086 1.43 riastrad config_detach_commit(dv);
1087 1.43 riastrad }
1088 1.43 riastrad
1089 1.32 dholland int
1090 1.11 ad cdev_open(dev_t dev, int flag, int devtype, lwp_t *l)
1091 1.11 ad {
1092 1.11 ad const struct cdevsw *d;
1093 1.40 riastrad struct localcount *lc;
1094 1.41 riastrad device_t dv = NULL/*XXXGCC*/;
1095 1.41 riastrad int unit, rv, mpflag;
1096 1.11 ad
1097 1.40 riastrad d = cdevsw_lookup_acquire(dev, &lc);
1098 1.11 ad if (d == NULL)
1099 1.11 ad return ENXIO;
1100 1.11 ad
1101 1.41 riastrad if (d->d_devtounit) {
1102 1.41 riastrad /*
1103 1.41 riastrad * If the device node corresponds to an autoconf device
1104 1.41 riastrad * instance, acquire a reference to it so that during
1105 1.41 riastrad * d_open, device_lookup is stable.
1106 1.41 riastrad *
1107 1.41 riastrad * XXX This should also arrange to instantiate cloning
1108 1.41 riastrad * pseudo-devices if appropriate, but that requires
1109 1.41 riastrad * reviewing them all to find and verify a common
1110 1.41 riastrad * pattern.
1111 1.41 riastrad */
1112 1.41 riastrad if ((unit = (*d->d_devtounit)(dev)) == -1)
1113 1.41 riastrad return ENXIO;
1114 1.41 riastrad if ((dv = device_lookup_acquire(d->d_cfdriver, unit)) == NULL)
1115 1.41 riastrad return ENXIO;
1116 1.41 riastrad }
1117 1.41 riastrad
1118 1.11 ad DEV_LOCK(d);
1119 1.11 ad rv = (*d->d_open)(dev, flag, devtype, l);
1120 1.11 ad DEV_UNLOCK(d);
1121 1.11 ad
1122 1.41 riastrad if (d->d_devtounit) {
1123 1.41 riastrad device_release(dv);
1124 1.41 riastrad }
1125 1.41 riastrad
1126 1.40 riastrad cdevsw_release(d, lc);
1127 1.40 riastrad
1128 1.11 ad return rv;
1129 1.11 ad }
1130 1.11 ad
1131 1.11 ad int
1132 1.11 ad cdev_close(dev_t dev, int flag, int devtype, lwp_t *l)
1133 1.11 ad {
1134 1.11 ad const struct cdevsw *d;
1135 1.17 ad int rv, mpflag;
1136 1.11 ad
1137 1.11 ad if ((d = cdevsw_lookup(dev)) == NULL)
1138 1.11 ad return ENXIO;
1139 1.11 ad
1140 1.11 ad DEV_LOCK(d);
1141 1.11 ad rv = (*d->d_close)(dev, flag, devtype, l);
1142 1.11 ad DEV_UNLOCK(d);
1143 1.11 ad
1144 1.11 ad return rv;
1145 1.11 ad }
1146 1.11 ad
1147 1.11 ad int
1148 1.11 ad cdev_read(dev_t dev, struct uio *uio, int flag)
1149 1.11 ad {
1150 1.11 ad const struct cdevsw *d;
1151 1.17 ad int rv, mpflag;
1152 1.11 ad
1153 1.11 ad if ((d = cdevsw_lookup(dev)) == NULL)
1154 1.11 ad return ENXIO;
1155 1.11 ad
1156 1.11 ad DEV_LOCK(d);
1157 1.11 ad rv = (*d->d_read)(dev, uio, flag);
1158 1.11 ad DEV_UNLOCK(d);
1159 1.11 ad
1160 1.11 ad return rv;
1161 1.11 ad }
1162 1.11 ad
1163 1.11 ad int
1164 1.11 ad cdev_write(dev_t dev, struct uio *uio, int flag)
1165 1.11 ad {
1166 1.11 ad const struct cdevsw *d;
1167 1.17 ad int rv, mpflag;
1168 1.11 ad
1169 1.11 ad if ((d = cdevsw_lookup(dev)) == NULL)
1170 1.11 ad return ENXIO;
1171 1.11 ad
1172 1.11 ad DEV_LOCK(d);
1173 1.11 ad rv = (*d->d_write)(dev, uio, flag);
1174 1.11 ad DEV_UNLOCK(d);
1175 1.11 ad
1176 1.11 ad return rv;
1177 1.11 ad }
1178 1.11 ad
1179 1.11 ad int
1180 1.11 ad cdev_ioctl(dev_t dev, u_long cmd, void *data, int flag, lwp_t *l)
1181 1.11 ad {
1182 1.11 ad const struct cdevsw *d;
1183 1.17 ad int rv, mpflag;
1184 1.11 ad
1185 1.11 ad if ((d = cdevsw_lookup(dev)) == NULL)
1186 1.11 ad return ENXIO;
1187 1.11 ad
1188 1.11 ad DEV_LOCK(d);
1189 1.11 ad rv = (*d->d_ioctl)(dev, cmd, data, flag, l);
1190 1.11 ad DEV_UNLOCK(d);
1191 1.11 ad
1192 1.11 ad return rv;
1193 1.11 ad }
1194 1.11 ad
1195 1.11 ad void
1196 1.11 ad cdev_stop(struct tty *tp, int flag)
1197 1.11 ad {
1198 1.11 ad const struct cdevsw *d;
1199 1.17 ad int mpflag;
1200 1.11 ad
1201 1.11 ad if ((d = cdevsw_lookup(tp->t_dev)) == NULL)
1202 1.11 ad return;
1203 1.11 ad
1204 1.11 ad DEV_LOCK(d);
1205 1.11 ad (*d->d_stop)(tp, flag);
1206 1.11 ad DEV_UNLOCK(d);
1207 1.11 ad }
1208 1.11 ad
1209 1.11 ad struct tty *
1210 1.11 ad cdev_tty(dev_t dev)
1211 1.11 ad {
1212 1.11 ad const struct cdevsw *d;
1213 1.11 ad
1214 1.11 ad if ((d = cdevsw_lookup(dev)) == NULL)
1215 1.11 ad return NULL;
1216 1.11 ad
1217 1.12 ad /* XXX Check if necessary. */
1218 1.12 ad if (d->d_tty == NULL)
1219 1.12 ad return NULL;
1220 1.12 ad
1221 1.21 ad return (*d->d_tty)(dev);
1222 1.11 ad }
1223 1.11 ad
1224 1.11 ad int
1225 1.11 ad cdev_poll(dev_t dev, int flag, lwp_t *l)
1226 1.11 ad {
1227 1.11 ad const struct cdevsw *d;
1228 1.17 ad int rv, mpflag;
1229 1.11 ad
1230 1.11 ad if ((d = cdevsw_lookup(dev)) == NULL)
1231 1.11 ad return POLLERR;
1232 1.11 ad
1233 1.11 ad DEV_LOCK(d);
1234 1.11 ad rv = (*d->d_poll)(dev, flag, l);
1235 1.11 ad DEV_UNLOCK(d);
1236 1.11 ad
1237 1.11 ad return rv;
1238 1.11 ad }
1239 1.11 ad
1240 1.11 ad paddr_t
1241 1.11 ad cdev_mmap(dev_t dev, off_t off, int flag)
1242 1.11 ad {
1243 1.11 ad const struct cdevsw *d;
1244 1.11 ad paddr_t rv;
1245 1.17 ad int mpflag;
1246 1.11 ad
1247 1.11 ad if ((d = cdevsw_lookup(dev)) == NULL)
1248 1.11 ad return (paddr_t)-1LL;
1249 1.11 ad
1250 1.11 ad DEV_LOCK(d);
1251 1.11 ad rv = (*d->d_mmap)(dev, off, flag);
1252 1.11 ad DEV_UNLOCK(d);
1253 1.11 ad
1254 1.11 ad return rv;
1255 1.11 ad }
1256 1.11 ad
1257 1.11 ad int
1258 1.11 ad cdev_kqfilter(dev_t dev, struct knote *kn)
1259 1.11 ad {
1260 1.11 ad const struct cdevsw *d;
1261 1.17 ad int rv, mpflag;
1262 1.11 ad
1263 1.11 ad if ((d = cdevsw_lookup(dev)) == NULL)
1264 1.11 ad return ENXIO;
1265 1.11 ad
1266 1.11 ad DEV_LOCK(d);
1267 1.11 ad rv = (*d->d_kqfilter)(dev, kn);
1268 1.11 ad DEV_UNLOCK(d);
1269 1.11 ad
1270 1.11 ad return rv;
1271 1.11 ad }
1272 1.11 ad
1273 1.11 ad int
1274 1.32 dholland cdev_discard(dev_t dev, off_t pos, off_t len)
1275 1.32 dholland {
1276 1.32 dholland const struct cdevsw *d;
1277 1.32 dholland int rv, mpflag;
1278 1.32 dholland
1279 1.32 dholland if ((d = cdevsw_lookup(dev)) == NULL)
1280 1.32 dholland return ENXIO;
1281 1.32 dholland
1282 1.32 dholland DEV_LOCK(d);
1283 1.32 dholland rv = (*d->d_discard)(dev, pos, len);
1284 1.32 dholland DEV_UNLOCK(d);
1285 1.32 dholland
1286 1.32 dholland return rv;
1287 1.32 dholland }
1288 1.32 dholland
1289 1.32 dholland int
1290 1.35 nat cdev_flags(dev_t dev)
1291 1.35 nat {
1292 1.35 nat const struct cdevsw *d;
1293 1.35 nat
1294 1.35 nat if ((d = cdevsw_lookup(dev)) == NULL)
1295 1.35 nat return 0;
1296 1.35 nat return d->d_flag & ~D_TYPEMASK;
1297 1.35 nat }
1298 1.35 nat
1299 1.35 nat int
1300 1.11 ad cdev_type(dev_t dev)
1301 1.11 ad {
1302 1.11 ad const struct cdevsw *d;
1303 1.11 ad
1304 1.11 ad if ((d = cdevsw_lookup(dev)) == NULL)
1305 1.11 ad return D_OTHER;
1306 1.11 ad return d->d_flag & D_TYPEMASK;
1307 1.2 gehenna }
1308 1.36 riastrad
1309 1.43 riastrad void
1310 1.43 riastrad cdev_detached(dev_t dev)
1311 1.43 riastrad {
1312 1.43 riastrad const struct cdevsw *d;
1313 1.43 riastrad device_t dv;
1314 1.43 riastrad int unit;
1315 1.43 riastrad
1316 1.43 riastrad if ((d = cdevsw_lookup(dev)) == NULL)
1317 1.43 riastrad return;
1318 1.43 riastrad if (d->d_devtounit == NULL)
1319 1.43 riastrad return;
1320 1.43 riastrad if ((unit = (*d->d_devtounit)(dev)) == -1)
1321 1.43 riastrad return;
1322 1.43 riastrad if ((dv = device_lookup(d->d_cfdriver, unit)) == NULL)
1323 1.43 riastrad return;
1324 1.43 riastrad config_detach_commit(dv);
1325 1.43 riastrad }
1326 1.43 riastrad
1327 1.36 riastrad /*
1328 1.36 riastrad * nommap(dev, off, prot)
1329 1.36 riastrad *
1330 1.36 riastrad * mmap routine that always fails, for non-mmappable devices.
1331 1.36 riastrad */
1332 1.36 riastrad paddr_t
1333 1.36 riastrad nommap(dev_t dev, off_t off, int prot)
1334 1.36 riastrad {
1335 1.36 riastrad
1336 1.36 riastrad return (paddr_t)-1;
1337 1.36 riastrad }
1338 1.42 riastrad
1339 1.42 riastrad /*
1340 1.42 riastrad * dev_minor_unit(dev)
1341 1.42 riastrad *
1342 1.42 riastrad * Returns minor(dev) as an int. Intended for use with struct
1343 1.42 riastrad * bdevsw, cdevsw::d_devtounit for drivers whose /dev nodes are
1344 1.42 riastrad * implemented by reference to an autoconf instance with the minor
1345 1.42 riastrad * number.
1346 1.42 riastrad */
1347 1.42 riastrad int
1348 1.42 riastrad dev_minor_unit(dev_t dev)
1349 1.42 riastrad {
1350 1.42 riastrad
1351 1.42 riastrad return minor(dev);
1352 1.42 riastrad }
1353