spec_vnops.c revision 1.2 1 1.1 cgd /*
2 1.1 cgd * Copyright (c) 1989 The Regents of the University of California.
3 1.1 cgd * All rights reserved.
4 1.1 cgd *
5 1.1 cgd * Redistribution and use in source and binary forms, with or without
6 1.1 cgd * modification, are permitted provided that the following conditions
7 1.1 cgd * are met:
8 1.1 cgd * 1. Redistributions of source code must retain the above copyright
9 1.1 cgd * notice, this list of conditions and the following disclaimer.
10 1.1 cgd * 2. Redistributions in binary form must reproduce the above copyright
11 1.1 cgd * notice, this list of conditions and the following disclaimer in the
12 1.1 cgd * documentation and/or other materials provided with the distribution.
13 1.1 cgd * 3. All advertising materials mentioning features or use of this software
14 1.1 cgd * must display the following acknowledgement:
15 1.1 cgd * This product includes software developed by the University of
16 1.1 cgd * California, Berkeley and its contributors.
17 1.1 cgd * 4. Neither the name of the University nor the names of its contributors
18 1.1 cgd * may be used to endorse or promote products derived from this software
19 1.1 cgd * without specific prior written permission.
20 1.1 cgd *
21 1.1 cgd * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
22 1.1 cgd * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
23 1.1 cgd * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
24 1.1 cgd * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
25 1.1 cgd * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
26 1.1 cgd * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
27 1.1 cgd * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
28 1.1 cgd * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
29 1.1 cgd * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
30 1.1 cgd * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
31 1.1 cgd * SUCH DAMAGE.
32 1.1 cgd *
33 1.1 cgd * @(#)spec_vnops.c 7.37 (Berkeley) 5/30/91
34 1.2 cgd *
35 1.2 cgd * PATCHES MAGIC LEVEL PATCH THAT GOT US HERE
36 1.2 cgd * -------------------- ----- ----------------------
37 1.2 cgd * CURRENT PATCH LEVEL: 1 00007
38 1.2 cgd * -------------------- ----- ----------------------
39 1.2 cgd *
40 1.2 cgd * 20 Aug 92 David Greenman Fixed incorrect setting of B_AGE on
41 1.1 cgd */
42 1.1 cgd
43 1.1 cgd #include "param.h"
44 1.1 cgd #include "proc.h"
45 1.1 cgd #include "systm.h"
46 1.1 cgd #include "kernel.h"
47 1.1 cgd #include "conf.h"
48 1.1 cgd #include "buf.h"
49 1.1 cgd #include "mount.h"
50 1.1 cgd #include "namei.h"
51 1.1 cgd #include "vnode.h"
52 1.1 cgd #include "specdev.h"
53 1.1 cgd #include "stat.h"
54 1.1 cgd #include "errno.h"
55 1.1 cgd #include "ioctl.h"
56 1.1 cgd #include "file.h"
57 1.1 cgd #include "dkbad.h" /* XXX */
58 1.1 cgd #include "disklabel.h"
59 1.1 cgd
60 1.1 cgd /* symbolic sleep message strings for devices */
61 1.1 cgd char devopn[] = "devopn";
62 1.1 cgd char devio[] = "devio";
63 1.1 cgd char devwait[] = "devwait";
64 1.1 cgd char devin[] = "devin";
65 1.1 cgd char devout[] = "devout";
66 1.1 cgd char devioc[] = "devioc";
67 1.1 cgd char devcls[] = "devcls";
68 1.1 cgd
69 1.1 cgd struct vnodeops spec_vnodeops = {
70 1.1 cgd spec_lookup, /* lookup */
71 1.1 cgd spec_create, /* create */
72 1.1 cgd spec_mknod, /* mknod */
73 1.1 cgd spec_open, /* open */
74 1.1 cgd spec_close, /* close */
75 1.1 cgd spec_access, /* access */
76 1.1 cgd spec_getattr, /* getattr */
77 1.1 cgd spec_setattr, /* setattr */
78 1.1 cgd spec_read, /* read */
79 1.1 cgd spec_write, /* write */
80 1.1 cgd spec_ioctl, /* ioctl */
81 1.1 cgd spec_select, /* select */
82 1.1 cgd spec_mmap, /* mmap */
83 1.1 cgd spec_fsync, /* fsync */
84 1.1 cgd spec_seek, /* seek */
85 1.1 cgd spec_remove, /* remove */
86 1.1 cgd spec_link, /* link */
87 1.1 cgd spec_rename, /* rename */
88 1.1 cgd spec_mkdir, /* mkdir */
89 1.1 cgd spec_rmdir, /* rmdir */
90 1.1 cgd spec_symlink, /* symlink */
91 1.1 cgd spec_readdir, /* readdir */
92 1.1 cgd spec_readlink, /* readlink */
93 1.1 cgd spec_abortop, /* abortop */
94 1.1 cgd spec_inactive, /* inactive */
95 1.1 cgd spec_reclaim, /* reclaim */
96 1.1 cgd spec_lock, /* lock */
97 1.1 cgd spec_unlock, /* unlock */
98 1.1 cgd spec_bmap, /* bmap */
99 1.1 cgd spec_strategy, /* strategy */
100 1.1 cgd spec_print, /* print */
101 1.1 cgd spec_islocked, /* islocked */
102 1.1 cgd spec_advlock, /* advlock */
103 1.1 cgd };
104 1.1 cgd
105 1.1 cgd /*
106 1.1 cgd * Trivial lookup routine that always fails.
107 1.1 cgd */
108 1.1 cgd spec_lookup(vp, ndp, p)
109 1.1 cgd struct vnode *vp;
110 1.1 cgd struct nameidata *ndp;
111 1.1 cgd struct proc *p;
112 1.1 cgd {
113 1.1 cgd
114 1.1 cgd ndp->ni_dvp = vp;
115 1.1 cgd ndp->ni_vp = NULL;
116 1.1 cgd return (ENOTDIR);
117 1.1 cgd }
118 1.1 cgd
119 1.1 cgd /*
120 1.1 cgd * Open a special file: Don't allow open if fs is mounted -nodev,
121 1.1 cgd * and don't allow opens of block devices that are currently mounted.
122 1.1 cgd * Otherwise, call device driver open function.
123 1.1 cgd */
124 1.1 cgd /* ARGSUSED */
125 1.1 cgd spec_open(vp, mode, cred, p)
126 1.1 cgd register struct vnode *vp;
127 1.1 cgd int mode;
128 1.1 cgd struct ucred *cred;
129 1.1 cgd struct proc *p;
130 1.1 cgd {
131 1.1 cgd dev_t dev = (dev_t)vp->v_rdev;
132 1.1 cgd register int maj = major(dev);
133 1.1 cgd int error;
134 1.1 cgd
135 1.1 cgd if (vp->v_mount && (vp->v_mount->mnt_flag & MNT_NODEV))
136 1.1 cgd return (ENXIO);
137 1.1 cgd
138 1.1 cgd switch (vp->v_type) {
139 1.1 cgd
140 1.1 cgd case VCHR:
141 1.1 cgd if ((u_int)maj >= nchrdev)
142 1.1 cgd return (ENXIO);
143 1.1 cgd VOP_UNLOCK(vp);
144 1.1 cgd error = (*cdevsw[maj].d_open)(dev, mode, S_IFCHR, p);
145 1.1 cgd VOP_LOCK(vp);
146 1.1 cgd return (error);
147 1.1 cgd
148 1.1 cgd case VBLK:
149 1.1 cgd if ((u_int)maj >= nblkdev)
150 1.1 cgd return (ENXIO);
151 1.1 cgd if (error = mountedon(vp))
152 1.1 cgd return (error);
153 1.1 cgd return ((*bdevsw[maj].d_open)(dev, mode, S_IFBLK, p));
154 1.1 cgd }
155 1.1 cgd return (0);
156 1.1 cgd }
157 1.1 cgd
158 1.1 cgd /*
159 1.1 cgd * Vnode op for read
160 1.1 cgd */
161 1.1 cgd /* ARGSUSED */
162 1.1 cgd spec_read(vp, uio, ioflag, cred)
163 1.1 cgd register struct vnode *vp;
164 1.1 cgd register struct uio *uio;
165 1.1 cgd int ioflag;
166 1.1 cgd struct ucred *cred;
167 1.1 cgd {
168 1.1 cgd struct proc *p = uio->uio_procp;
169 1.1 cgd struct buf *bp;
170 1.1 cgd daddr_t bn;
171 1.1 cgd long bsize, bscale;
172 1.1 cgd struct partinfo dpart;
173 1.1 cgd register int n, on;
174 1.1 cgd int error = 0;
175 1.1 cgd extern int mem_no;
176 1.1 cgd
177 1.1 cgd #ifdef DIAGNOSTIC
178 1.1 cgd if (uio->uio_rw != UIO_READ)
179 1.1 cgd panic("spec_read mode");
180 1.1 cgd if (uio->uio_segflg == UIO_USERSPACE && uio->uio_procp != curproc)
181 1.1 cgd panic("spec_read proc");
182 1.1 cgd #endif
183 1.1 cgd if (uio->uio_resid == 0)
184 1.1 cgd return (0);
185 1.1 cgd
186 1.1 cgd switch (vp->v_type) {
187 1.1 cgd
188 1.1 cgd case VCHR:
189 1.1 cgd /*
190 1.1 cgd * Negative offsets allowed only for /dev/kmem
191 1.1 cgd */
192 1.1 cgd if (uio->uio_offset < 0 && major(vp->v_rdev) != mem_no)
193 1.1 cgd return (EINVAL);
194 1.1 cgd VOP_UNLOCK(vp);
195 1.1 cgd error = (*cdevsw[major(vp->v_rdev)].d_read)
196 1.1 cgd (vp->v_rdev, uio, ioflag);
197 1.1 cgd VOP_LOCK(vp);
198 1.1 cgd return (error);
199 1.1 cgd
200 1.1 cgd case VBLK:
201 1.1 cgd if (uio->uio_offset < 0)
202 1.1 cgd return (EINVAL);
203 1.1 cgd bsize = BLKDEV_IOSIZE;
204 1.1 cgd if ((*bdevsw[major(vp->v_rdev)].d_ioctl)(vp->v_rdev, DIOCGPART,
205 1.1 cgd (caddr_t)&dpart, FREAD, p) == 0) {
206 1.1 cgd if (dpart.part->p_fstype == FS_BSDFFS &&
207 1.1 cgd dpart.part->p_frag != 0 && dpart.part->p_fsize != 0)
208 1.1 cgd bsize = dpart.part->p_frag *
209 1.1 cgd dpart.part->p_fsize;
210 1.1 cgd }
211 1.1 cgd bscale = bsize / DEV_BSIZE;
212 1.1 cgd do {
213 1.1 cgd bn = (uio->uio_offset / DEV_BSIZE) &~ (bscale - 1);
214 1.1 cgd on = uio->uio_offset % bsize;
215 1.1 cgd n = MIN((unsigned)(bsize - on), uio->uio_resid);
216 1.1 cgd if (vp->v_lastr + bscale == bn)
217 1.1 cgd error = breada(vp, bn, (int)bsize, bn + bscale,
218 1.1 cgd (int)bsize, NOCRED, &bp);
219 1.1 cgd else
220 1.1 cgd error = bread(vp, bn, (int)bsize, NOCRED, &bp);
221 1.1 cgd vp->v_lastr = bn;
222 1.1 cgd n = MIN(n, bsize - bp->b_resid);
223 1.1 cgd if (error) {
224 1.1 cgd brelse(bp);
225 1.1 cgd return (error);
226 1.1 cgd }
227 1.1 cgd error = uiomove(bp->b_un.b_addr + on, n, uio);
228 1.2 cgd #ifdef OMIT /* 20 Aug 92*/
229 1.1 cgd if (n + on == bsize)
230 1.1 cgd bp->b_flags |= B_AGE;
231 1.2 cgd #endif /* OMIT*/
232 1.1 cgd brelse(bp);
233 1.1 cgd } while (error == 0 && uio->uio_resid > 0 && n != 0);
234 1.1 cgd return (error);
235 1.1 cgd
236 1.1 cgd default:
237 1.1 cgd panic("spec_read type");
238 1.1 cgd }
239 1.1 cgd /* NOTREACHED */
240 1.1 cgd }
241 1.1 cgd
242 1.1 cgd /*
243 1.1 cgd * Vnode op for write
244 1.1 cgd */
245 1.1 cgd /* ARGSUSED */
246 1.1 cgd spec_write(vp, uio, ioflag, cred)
247 1.1 cgd register struct vnode *vp;
248 1.1 cgd register struct uio *uio;
249 1.1 cgd int ioflag;
250 1.1 cgd struct ucred *cred;
251 1.1 cgd {
252 1.1 cgd struct proc *p = uio->uio_procp;
253 1.1 cgd struct buf *bp;
254 1.1 cgd daddr_t bn;
255 1.1 cgd int bsize, blkmask;
256 1.1 cgd struct partinfo dpart;
257 1.1 cgd register int n, on;
258 1.1 cgd int error = 0;
259 1.1 cgd extern int mem_no;
260 1.1 cgd
261 1.1 cgd #ifdef DIAGNOSTIC
262 1.1 cgd if (uio->uio_rw != UIO_WRITE)
263 1.1 cgd panic("spec_write mode");
264 1.1 cgd if (uio->uio_segflg == UIO_USERSPACE && uio->uio_procp != curproc)
265 1.1 cgd panic("spec_write proc");
266 1.1 cgd #endif
267 1.1 cgd
268 1.1 cgd switch (vp->v_type) {
269 1.1 cgd
270 1.1 cgd case VCHR:
271 1.1 cgd /*
272 1.1 cgd * Negative offsets allowed only for /dev/kmem
273 1.1 cgd */
274 1.1 cgd if (uio->uio_offset < 0 && major(vp->v_rdev) != mem_no)
275 1.1 cgd return (EINVAL);
276 1.1 cgd VOP_UNLOCK(vp);
277 1.1 cgd error = (*cdevsw[major(vp->v_rdev)].d_write)
278 1.1 cgd (vp->v_rdev, uio, ioflag);
279 1.1 cgd VOP_LOCK(vp);
280 1.1 cgd return (error);
281 1.1 cgd
282 1.1 cgd case VBLK:
283 1.1 cgd if (uio->uio_resid == 0)
284 1.1 cgd return (0);
285 1.1 cgd if (uio->uio_offset < 0)
286 1.1 cgd return (EINVAL);
287 1.1 cgd bsize = BLKDEV_IOSIZE;
288 1.1 cgd if ((*bdevsw[major(vp->v_rdev)].d_ioctl)(vp->v_rdev, DIOCGPART,
289 1.1 cgd (caddr_t)&dpart, FREAD, p) == 0) {
290 1.1 cgd if (dpart.part->p_fstype == FS_BSDFFS &&
291 1.1 cgd dpart.part->p_frag != 0 && dpart.part->p_fsize != 0)
292 1.1 cgd bsize = dpart.part->p_frag *
293 1.1 cgd dpart.part->p_fsize;
294 1.1 cgd }
295 1.1 cgd blkmask = (bsize / DEV_BSIZE) - 1;
296 1.1 cgd do {
297 1.1 cgd bn = (uio->uio_offset / DEV_BSIZE) &~ blkmask;
298 1.1 cgd on = uio->uio_offset % bsize;
299 1.1 cgd n = MIN((unsigned)(bsize - on), uio->uio_resid);
300 1.1 cgd if (n == bsize)
301 1.1 cgd bp = getblk(vp, bn, bsize);
302 1.1 cgd else
303 1.1 cgd error = bread(vp, bn, bsize, NOCRED, &bp);
304 1.1 cgd n = MIN(n, bsize - bp->b_resid);
305 1.1 cgd if (error) {
306 1.1 cgd brelse(bp);
307 1.1 cgd return (error);
308 1.1 cgd }
309 1.1 cgd error = uiomove(bp->b_un.b_addr + on, n, uio);
310 1.1 cgd if (n + on == bsize) {
311 1.1 cgd bp->b_flags |= B_AGE;
312 1.1 cgd bawrite(bp);
313 1.1 cgd } else
314 1.1 cgd bdwrite(bp);
315 1.1 cgd } while (error == 0 && uio->uio_resid > 0 && n != 0);
316 1.1 cgd return (error);
317 1.1 cgd
318 1.1 cgd default:
319 1.1 cgd panic("spec_write type");
320 1.1 cgd }
321 1.1 cgd /* NOTREACHED */
322 1.1 cgd }
323 1.1 cgd
324 1.1 cgd /*
325 1.1 cgd * Device ioctl operation.
326 1.1 cgd */
327 1.1 cgd /* ARGSUSED */
328 1.1 cgd spec_ioctl(vp, com, data, fflag, cred, p)
329 1.1 cgd struct vnode *vp;
330 1.1 cgd int com;
331 1.1 cgd caddr_t data;
332 1.1 cgd int fflag;
333 1.1 cgd struct ucred *cred;
334 1.1 cgd struct proc *p;
335 1.1 cgd {
336 1.1 cgd dev_t dev = vp->v_rdev;
337 1.1 cgd
338 1.1 cgd switch (vp->v_type) {
339 1.1 cgd
340 1.1 cgd case VCHR:
341 1.1 cgd return ((*cdevsw[major(dev)].d_ioctl)(dev, com, data,
342 1.1 cgd fflag, p));
343 1.1 cgd
344 1.1 cgd case VBLK:
345 1.1 cgd if (com == 0 && (int)data == B_TAPE)
346 1.1 cgd if (bdevsw[major(dev)].d_flags & B_TAPE)
347 1.1 cgd return (0);
348 1.1 cgd else
349 1.1 cgd return (1);
350 1.1 cgd return ((*bdevsw[major(dev)].d_ioctl)(dev, com, data,
351 1.1 cgd fflag, p));
352 1.1 cgd
353 1.1 cgd default:
354 1.1 cgd panic("spec_ioctl");
355 1.1 cgd /* NOTREACHED */
356 1.1 cgd }
357 1.1 cgd }
358 1.1 cgd
359 1.1 cgd /* ARGSUSED */
360 1.1 cgd spec_select(vp, which, fflags, cred, p)
361 1.1 cgd struct vnode *vp;
362 1.1 cgd int which, fflags;
363 1.1 cgd struct ucred *cred;
364 1.1 cgd struct proc *p;
365 1.1 cgd {
366 1.1 cgd register dev_t dev;
367 1.1 cgd
368 1.1 cgd switch (vp->v_type) {
369 1.1 cgd
370 1.1 cgd default:
371 1.1 cgd return (1); /* XXX */
372 1.1 cgd
373 1.1 cgd case VCHR:
374 1.1 cgd dev = vp->v_rdev;
375 1.1 cgd return (*cdevsw[major(dev)].d_select)(dev, which, p);
376 1.1 cgd }
377 1.1 cgd }
378 1.1 cgd
379 1.1 cgd /*
380 1.1 cgd * Just call the device strategy routine
381 1.1 cgd */
382 1.1 cgd spec_strategy(bp)
383 1.1 cgd register struct buf *bp;
384 1.1 cgd {
385 1.1 cgd
386 1.1 cgd (*bdevsw[major(bp->b_dev)].d_strategy)(bp);
387 1.1 cgd return (0);
388 1.1 cgd }
389 1.1 cgd
390 1.1 cgd /*
391 1.1 cgd * This is a noop, simply returning what one has been given.
392 1.1 cgd */
393 1.1 cgd spec_bmap(vp, bn, vpp, bnp)
394 1.1 cgd struct vnode *vp;
395 1.1 cgd daddr_t bn;
396 1.1 cgd struct vnode **vpp;
397 1.1 cgd daddr_t *bnp;
398 1.1 cgd {
399 1.1 cgd
400 1.1 cgd if (vpp != NULL)
401 1.1 cgd *vpp = vp;
402 1.1 cgd if (bnp != NULL)
403 1.1 cgd *bnp = bn;
404 1.1 cgd return (0);
405 1.1 cgd }
406 1.1 cgd
407 1.1 cgd /*
408 1.1 cgd * At the moment we do not do any locking.
409 1.1 cgd */
410 1.1 cgd /* ARGSUSED */
411 1.1 cgd spec_lock(vp)
412 1.1 cgd struct vnode *vp;
413 1.1 cgd {
414 1.1 cgd
415 1.1 cgd return (0);
416 1.1 cgd }
417 1.1 cgd
418 1.1 cgd /* ARGSUSED */
419 1.1 cgd spec_unlock(vp)
420 1.1 cgd struct vnode *vp;
421 1.1 cgd {
422 1.1 cgd
423 1.1 cgd return (0);
424 1.1 cgd }
425 1.1 cgd
426 1.1 cgd /*
427 1.1 cgd * Device close routine
428 1.1 cgd */
429 1.1 cgd /* ARGSUSED */
430 1.1 cgd spec_close(vp, flag, cred, p)
431 1.1 cgd register struct vnode *vp;
432 1.1 cgd int flag;
433 1.1 cgd struct ucred *cred;
434 1.1 cgd struct proc *p;
435 1.1 cgd {
436 1.1 cgd dev_t dev = vp->v_rdev;
437 1.1 cgd int (*devclose) __P((dev_t, int, int, struct proc *));
438 1.1 cgd int mode;
439 1.1 cgd
440 1.1 cgd switch (vp->v_type) {
441 1.1 cgd
442 1.1 cgd case VCHR:
443 1.1 cgd /*
444 1.1 cgd * If the vnode is locked, then we are in the midst
445 1.1 cgd * of forcably closing the device, otherwise we only
446 1.1 cgd * close on last reference.
447 1.1 cgd */
448 1.1 cgd if (vcount(vp) > 1 && (vp->v_flag & VXLOCK) == 0)
449 1.1 cgd return (0);
450 1.1 cgd devclose = cdevsw[major(dev)].d_close;
451 1.1 cgd mode = S_IFCHR;
452 1.1 cgd break;
453 1.1 cgd
454 1.1 cgd case VBLK:
455 1.1 cgd /*
456 1.1 cgd * On last close of a block device (that isn't mounted)
457 1.1 cgd * we must invalidate any in core blocks, so that
458 1.1 cgd * we can, for instance, change floppy disks.
459 1.1 cgd */
460 1.1 cgd vflushbuf(vp, 0);
461 1.1 cgd if (vinvalbuf(vp, 1))
462 1.1 cgd return (0);
463 1.1 cgd /*
464 1.1 cgd * We do not want to really close the device if it
465 1.1 cgd * is still in use unless we are trying to close it
466 1.1 cgd * forcibly. Since every use (buffer, vnode, swap, cmap)
467 1.1 cgd * holds a reference to the vnode, and because we mark
468 1.1 cgd * any other vnodes that alias this device, when the
469 1.1 cgd * sum of the reference counts on all the aliased
470 1.1 cgd * vnodes descends to one, we are on last close.
471 1.1 cgd */
472 1.1 cgd if (vcount(vp) > 1 && (vp->v_flag & VXLOCK) == 0)
473 1.1 cgd return (0);
474 1.1 cgd devclose = bdevsw[major(dev)].d_close;
475 1.1 cgd mode = S_IFBLK;
476 1.1 cgd break;
477 1.1 cgd
478 1.1 cgd default:
479 1.1 cgd panic("spec_close: not special");
480 1.1 cgd }
481 1.1 cgd
482 1.1 cgd return ((*devclose)(dev, flag, mode, p));
483 1.1 cgd }
484 1.1 cgd
485 1.1 cgd /*
486 1.1 cgd * Print out the contents of a special device vnode.
487 1.1 cgd */
488 1.1 cgd spec_print(vp)
489 1.1 cgd struct vnode *vp;
490 1.1 cgd {
491 1.1 cgd
492 1.1 cgd printf("tag VT_NON, dev %d, %d\n", major(vp->v_rdev),
493 1.1 cgd minor(vp->v_rdev));
494 1.1 cgd }
495 1.1 cgd
496 1.1 cgd /*
497 1.1 cgd * Special device advisory byte-level locks.
498 1.1 cgd */
499 1.1 cgd /* ARGSUSED */
500 1.1 cgd spec_advlock(vp, id, op, fl, flags)
501 1.1 cgd struct vnode *vp;
502 1.1 cgd caddr_t id;
503 1.1 cgd int op;
504 1.1 cgd struct flock *fl;
505 1.1 cgd int flags;
506 1.1 cgd {
507 1.1 cgd
508 1.1 cgd return (EOPNOTSUPP);
509 1.1 cgd }
510 1.1 cgd
511 1.1 cgd /*
512 1.1 cgd * Special device failed operation
513 1.1 cgd */
514 1.1 cgd spec_ebadf()
515 1.1 cgd {
516 1.1 cgd
517 1.1 cgd return (EBADF);
518 1.1 cgd }
519 1.1 cgd
520 1.1 cgd /*
521 1.1 cgd * Special device bad operation
522 1.1 cgd */
523 1.1 cgd spec_badop()
524 1.1 cgd {
525 1.1 cgd
526 1.1 cgd panic("spec_badop called");
527 1.1 cgd /* NOTREACHED */
528 1.1 cgd }
529