spec_vnops.c revision 1.44 1 /* $NetBSD: spec_vnops.c,v 1.44 1999/10/16 23:53:27 wrstuden Exp $ */
2
3 /*
4 * Copyright (c) 1989, 1993
5 * The Regents of the University of California. All rights reserved.
6 *
7 * Redistribution and use in source and binary forms, with or without
8 * modification, are permitted provided that the following conditions
9 * are met:
10 * 1. Redistributions of source code must retain the above copyright
11 * notice, this list of conditions and the following disclaimer.
12 * 2. Redistributions in binary form must reproduce the above copyright
13 * notice, this list of conditions and the following disclaimer in the
14 * documentation and/or other materials provided with the distribution.
15 * 3. All advertising materials mentioning features or use of this software
16 * must display the following acknowledgement:
17 * This product includes software developed by the University of
18 * California, Berkeley and its contributors.
19 * 4. Neither the name of the University nor the names of its contributors
20 * may be used to endorse or promote products derived from this software
21 * without specific prior written permission.
22 *
23 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
24 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
25 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
26 * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
27 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
28 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
29 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
30 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
31 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
32 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
33 * SUCH DAMAGE.
34 *
35 * @(#)spec_vnops.c 8.15 (Berkeley) 7/14/95
36 */
37
38 #include <sys/param.h>
39 #include <sys/proc.h>
40 #include <sys/systm.h>
41 #include <sys/kernel.h>
42 #include <sys/conf.h>
43 #include <sys/buf.h>
44 #include <sys/mount.h>
45 #include <sys/namei.h>
46 #include <sys/vnode.h>
47 #include <sys/stat.h>
48 #include <sys/errno.h>
49 #include <sys/ioctl.h>
50 #include <sys/file.h>
51 #include <sys/disklabel.h>
52 #include <sys/lockf.h>
53
54 #include <miscfs/genfs/genfs.h>
55 #include <miscfs/specfs/specdev.h>
56
57 /* symbolic sleep message strings for devices */
58 const char devopn[] = "devopn";
59 const char devio[] = "devio";
60 const char devwait[] = "devwait";
61 const char devin[] = "devin";
62 const char devout[] = "devout";
63 const char devioc[] = "devioc";
64 const char devcls[] = "devcls";
65
66 int (**spec_vnodeop_p) __P((void *));
67 struct vnodeopv_entry_desc spec_vnodeop_entries[] = {
68 { &vop_default_desc, vn_default_error },
69 { &vop_lookup_desc, spec_lookup }, /* lookup */
70 { &vop_create_desc, spec_create }, /* create */
71 { &vop_mknod_desc, spec_mknod }, /* mknod */
72 { &vop_open_desc, spec_open }, /* open */
73 { &vop_close_desc, spec_close }, /* close */
74 { &vop_access_desc, spec_access }, /* access */
75 { &vop_getattr_desc, spec_getattr }, /* getattr */
76 { &vop_setattr_desc, spec_setattr }, /* setattr */
77 { &vop_read_desc, spec_read }, /* read */
78 { &vop_write_desc, spec_write }, /* write */
79 { &vop_lease_desc, spec_lease_check }, /* lease */
80 { &vop_ioctl_desc, spec_ioctl }, /* ioctl */
81 { &vop_poll_desc, spec_poll }, /* poll */
82 { &vop_revoke_desc, spec_revoke }, /* revoke */
83 { &vop_mmap_desc, spec_mmap }, /* mmap */
84 { &vop_fsync_desc, spec_fsync }, /* fsync */
85 { &vop_seek_desc, spec_seek }, /* seek */
86 { &vop_remove_desc, spec_remove }, /* remove */
87 { &vop_link_desc, spec_link }, /* link */
88 { &vop_rename_desc, spec_rename }, /* rename */
89 { &vop_mkdir_desc, spec_mkdir }, /* mkdir */
90 { &vop_rmdir_desc, spec_rmdir }, /* rmdir */
91 { &vop_symlink_desc, spec_symlink }, /* symlink */
92 { &vop_readdir_desc, spec_readdir }, /* readdir */
93 { &vop_readlink_desc, spec_readlink }, /* readlink */
94 { &vop_abortop_desc, spec_abortop }, /* abortop */
95 { &vop_inactive_desc, spec_inactive }, /* inactive */
96 { &vop_reclaim_desc, spec_reclaim }, /* reclaim */
97 { &vop_lock_desc, spec_lock }, /* lock */
98 { &vop_unlock_desc, spec_unlock }, /* unlock */
99 { &vop_bmap_desc, spec_bmap }, /* bmap */
100 { &vop_strategy_desc, spec_strategy }, /* strategy */
101 { &vop_print_desc, spec_print }, /* print */
102 { &vop_islocked_desc, spec_islocked }, /* islocked */
103 { &vop_pathconf_desc, spec_pathconf }, /* pathconf */
104 { &vop_advlock_desc, spec_advlock }, /* advlock */
105 { &vop_blkatoff_desc, spec_blkatoff }, /* blkatoff */
106 { &vop_valloc_desc, spec_valloc }, /* valloc */
107 { &vop_vfree_desc, spec_vfree }, /* vfree */
108 { &vop_truncate_desc, spec_truncate }, /* truncate */
109 { &vop_update_desc, spec_update }, /* update */
110 { &vop_bwrite_desc, spec_bwrite }, /* bwrite */
111 { (struct vnodeop_desc*)NULL, (int(*) __P((void *)))NULL }
112 };
113 struct vnodeopv_desc spec_vnodeop_opv_desc =
114 { &spec_vnodeop_p, spec_vnodeop_entries };
115
116 /*
117 * Trivial lookup routine that always fails.
118 */
119 int
120 spec_lookup(v)
121 void *v;
122 {
123 struct vop_lookup_args /* {
124 struct vnode *a_dvp;
125 struct vnode **a_vpp;
126 struct componentname *a_cnp;
127 } */ *ap = v;
128
129 *ap->a_vpp = NULL;
130 return (ENOTDIR);
131 }
132
133 /*
134 * Open a special file.
135 */
136 /* ARGSUSED */
137 int
138 spec_open(v)
139 void *v;
140 {
141 struct vop_open_args /* {
142 struct vnode *a_vp;
143 int a_mode;
144 struct ucred *a_cred;
145 struct proc *a_p;
146 } */ *ap = v;
147 struct proc *p = ap->a_p;
148 struct vnode *bvp, *vp = ap->a_vp;
149 dev_t bdev, dev = (dev_t)vp->v_rdev;
150 register int maj = major(dev);
151 int error;
152
153 /*
154 * Don't allow open if fs is mounted -nodev.
155 */
156 if (vp->v_mount && (vp->v_mount->mnt_flag & MNT_NODEV))
157 return (ENXIO);
158
159 switch (vp->v_type) {
160
161 case VCHR:
162 if ((u_int)maj >= nchrdev)
163 return (ENXIO);
164 if (ap->a_cred != FSCRED && (ap->a_mode & FWRITE)) {
165 /*
166 * When running in very secure mode, do not allow
167 * opens for writing of any disk character devices.
168 */
169 if (securelevel >= 2 && cdevsw[maj].d_type == D_DISK)
170 return (EPERM);
171 /*
172 * When running in secure mode, do not allow opens
173 * for writing of /dev/mem, /dev/kmem, or character
174 * devices whose corresponding block devices are
175 * currently mounted.
176 */
177 if (securelevel >= 1) {
178 if ((bdev = chrtoblk(dev)) != (dev_t)NODEV &&
179 vfinddev(bdev, VBLK, &bvp) &&
180 bvp->v_usecount > 0 &&
181 (error = vfs_mountedon(bvp)))
182 return (error);
183 if (iskmemdev(dev))
184 return (EPERM);
185 }
186 }
187 if (cdevsw[maj].d_type == D_TTY)
188 vp->v_flag |= VISTTY;
189 VOP_UNLOCK(vp, 0);
190 error = (*cdevsw[maj].d_open)(dev, ap->a_mode, S_IFCHR, p);
191 vn_lock(vp, LK_EXCLUSIVE | LK_RETRY);
192 return (error);
193
194 case VBLK:
195 if ((u_int)maj >= nblkdev)
196 return (ENXIO);
197 /*
198 * When running in very secure mode, do not allow
199 * opens for writing of any disk block devices.
200 */
201 if (securelevel >= 2 && ap->a_cred != FSCRED &&
202 (ap->a_mode & FWRITE) && bdevsw[maj].d_type == D_DISK)
203 return (EPERM);
204 /*
205 * Do not allow opens of block devices that are
206 * currently mounted.
207 */
208 if ((error = vfs_mountedon(vp)) != 0)
209 return (error);
210 return ((*bdevsw[maj].d_open)(dev, ap->a_mode, S_IFBLK, p));
211 case VNON:
212 case VLNK:
213 case VDIR:
214 case VREG:
215 case VBAD:
216 case VFIFO:
217 case VSOCK:
218 break;
219 }
220 return (0);
221 }
222
223 /*
224 * Vnode op for read
225 */
226 /* ARGSUSED */
227 int
228 spec_read(v)
229 void *v;
230 {
231 struct vop_read_args /* {
232 struct vnode *a_vp;
233 struct uio *a_uio;
234 int a_ioflag;
235 struct ucred *a_cred;
236 } */ *ap = v;
237 register struct vnode *vp = ap->a_vp;
238 register struct uio *uio = ap->a_uio;
239 struct proc *p = uio->uio_procp;
240 struct buf *bp;
241 daddr_t bn, nextbn;
242 long bsize, bscale, ssize;
243 struct partinfo dpart;
244 int n, on, majordev;
245 int (*ioctl) __P((dev_t, u_long, caddr_t, int, struct proc *));
246 int error = 0;
247
248 #ifdef DIAGNOSTIC
249 if (uio->uio_rw != UIO_READ)
250 panic("spec_read mode");
251 if (uio->uio_segflg == UIO_USERSPACE && uio->uio_procp != curproc)
252 panic("spec_read proc");
253 #endif
254 if (uio->uio_resid == 0)
255 return (0);
256
257 switch (vp->v_type) {
258
259 case VCHR:
260 VOP_UNLOCK(vp, 0);
261 error = (*cdevsw[major(vp->v_rdev)].d_read)
262 (vp->v_rdev, uio, ap->a_ioflag);
263 vn_lock(vp, LK_EXCLUSIVE | LK_RETRY);
264 return (error);
265
266 case VBLK:
267 if (uio->uio_offset < 0)
268 return (EINVAL);
269 bsize = BLKDEV_IOSIZE;
270 ssize = DEV_BSIZE;
271 if ((majordev = major(vp->v_rdev)) < nblkdev &&
272 (ioctl = bdevsw[majordev].d_ioctl) != NULL &&
273 (*ioctl)(vp->v_rdev, DIOCGPART, (caddr_t)&dpart, FREAD, p) == 0) {
274 if (dpart.part->p_fstype == FS_BSDFFS &&
275 dpart.part->p_frag != 0 && dpart.part->p_fsize != 0)
276 bsize = dpart.part->p_frag *
277 dpart.part->p_fsize;
278 if (dpart.disklab->d_secsize != 0)
279 ssize = dpart.disklab->d_secsize;
280 }
281 bscale = bsize / ssize;
282 do {
283 bn = (uio->uio_offset / ssize) &~ (bscale - 1);
284 on = uio->uio_offset % bsize;
285 n = min((unsigned)(bsize - on), uio->uio_resid);
286 if (vp->v_lastr + bscale == bn) {
287 nextbn = bn + bscale;
288 error = breadn(vp, bn, (int)bsize, &nextbn,
289 (int *)&bsize, 1, NOCRED, &bp);
290 } else
291 error = bread(vp, bn, (int)bsize, NOCRED, &bp);
292 vp->v_lastr = bn;
293 n = min(n, bsize - bp->b_resid);
294 if (error) {
295 brelse(bp);
296 return (error);
297 }
298 error = uiomove((char *)bp->b_data + on, n, uio);
299 brelse(bp);
300 } while (error == 0 && uio->uio_resid > 0 && n != 0);
301 return (error);
302
303 default:
304 panic("spec_read type");
305 }
306 /* NOTREACHED */
307 }
308
309 /*
310 * Vnode op for write
311 */
312 /* ARGSUSED */
313 int
314 spec_write(v)
315 void *v;
316 {
317 struct vop_write_args /* {
318 struct vnode *a_vp;
319 struct uio *a_uio;
320 int a_ioflag;
321 struct ucred *a_cred;
322 } */ *ap = v;
323 register struct vnode *vp = ap->a_vp;
324 register struct uio *uio = ap->a_uio;
325 struct proc *p = uio->uio_procp;
326 struct buf *bp;
327 daddr_t bn;
328 long bsize, bscale, ssize;
329 struct partinfo dpart;
330 int n, on, majordev;
331 int (*ioctl) __P((dev_t, u_long, caddr_t, int, struct proc *));
332 int error = 0;
333
334 #ifdef DIAGNOSTIC
335 if (uio->uio_rw != UIO_WRITE)
336 panic("spec_write mode");
337 if (uio->uio_segflg == UIO_USERSPACE && uio->uio_procp != curproc)
338 panic("spec_write proc");
339 #endif
340
341 switch (vp->v_type) {
342
343 case VCHR:
344 VOP_UNLOCK(vp, 0);
345 error = (*cdevsw[major(vp->v_rdev)].d_write)
346 (vp->v_rdev, uio, ap->a_ioflag);
347 vn_lock(vp, LK_EXCLUSIVE | LK_RETRY);
348 return (error);
349
350 case VBLK:
351 if (uio->uio_resid == 0)
352 return (0);
353 if (uio->uio_offset < 0)
354 return (EINVAL);
355 bsize = BLKDEV_IOSIZE;
356 ssize = DEV_BSIZE;
357 if ((majordev = major(vp->v_rdev)) < nblkdev &&
358 (ioctl = bdevsw[majordev].d_ioctl) != NULL &&
359 (*ioctl)(vp->v_rdev, DIOCGPART, (caddr_t)&dpart, FREAD, p) == 0) {
360 if (dpart.part->p_fstype == FS_BSDFFS &&
361 dpart.part->p_frag != 0 && dpart.part->p_fsize != 0)
362 bsize = dpart.part->p_frag *
363 dpart.part->p_fsize;
364 if (dpart.disklab->d_secsize != 0)
365 ssize = dpart.disklab->d_secsize;
366 }
367 bscale = bsize / ssize;
368 do {
369 bn = (uio->uio_offset / ssize) &~ (bscale - 1);
370 on = uio->uio_offset % bsize;
371 n = min((unsigned)(bsize - on), uio->uio_resid);
372 if (n == bsize)
373 bp = getblk(vp, bn, bsize, 0, 0);
374 else
375 error = bread(vp, bn, bsize, NOCRED, &bp);
376 if (error) {
377 brelse(bp);
378 return (error);
379 }
380 n = min(n, bsize - bp->b_resid);
381 error = uiomove((char *)bp->b_data + on, n, uio);
382 if (error)
383 brelse(bp);
384 else {
385 if (n + on == bsize)
386 bawrite(bp);
387 else
388 bdwrite(bp);
389 if (bp->b_flags & B_ERROR)
390 error = bp->b_error;
391 }
392 } while (error == 0 && uio->uio_resid > 0 && n != 0);
393 return (error);
394
395 default:
396 panic("spec_write type");
397 }
398 /* NOTREACHED */
399 }
400
401 /*
402 * Device ioctl operation.
403 */
404 /* ARGSUSED */
405 int
406 spec_ioctl(v)
407 void *v;
408 {
409 struct vop_ioctl_args /* {
410 struct vnode *a_vp;
411 u_long a_command;
412 caddr_t a_data;
413 int a_fflag;
414 struct ucred *a_cred;
415 struct proc *a_p;
416 } */ *ap = v;
417 dev_t dev = ap->a_vp->v_rdev;
418 int maj = major(dev);
419
420 switch (ap->a_vp->v_type) {
421
422 case VCHR:
423 return ((*cdevsw[maj].d_ioctl)(dev, ap->a_command, ap->a_data,
424 ap->a_fflag, ap->a_p));
425
426 case VBLK:
427 if (ap->a_command == 0 && (long)ap->a_data == B_TAPE) {
428 if (bdevsw[maj].d_type == D_TAPE)
429 return (0);
430 else
431 return (1);
432 }
433 return ((*bdevsw[maj].d_ioctl)(dev, ap->a_command, ap->a_data,
434 ap->a_fflag, ap->a_p));
435
436 default:
437 panic("spec_ioctl");
438 /* NOTREACHED */
439 }
440 }
441
442 /* ARGSUSED */
443 int
444 spec_poll(v)
445 void *v;
446 {
447 struct vop_poll_args /* {
448 struct vnode *a_vp;
449 int a_events;
450 struct proc *a_p;
451 } */ *ap = v;
452 register dev_t dev;
453
454 switch (ap->a_vp->v_type) {
455
456 case VCHR:
457 dev = ap->a_vp->v_rdev;
458 return (*cdevsw[major(dev)].d_poll)(dev, ap->a_events, ap->a_p);
459
460 default:
461 return (genfs_poll(v));
462 }
463 }
464 /*
465 * Synch buffers associated with a block device
466 */
467 /* ARGSUSED */
468 int
469 spec_fsync(v)
470 void *v;
471 {
472 struct vop_fsync_args /* {
473 struct vnode *a_vp;
474 struct ucred *a_cred;
475 int a_flags;
476 struct proc *a_p;
477 } */ *ap = v;
478 register struct vnode *vp = ap->a_vp;
479
480 if (vp->v_type == VBLK)
481 vflushbuf(vp, (ap->a_flags & FSYNC_WAIT) != 0);
482 return (0);
483 }
484
485 /*
486 * Just call the device strategy routine
487 */
488 int
489 spec_strategy(v)
490 void *v;
491 {
492 struct vop_strategy_args /* {
493 struct buf *a_bp;
494 } */ *ap = v;
495
496 (*bdevsw[major(ap->a_bp->b_dev)].d_strategy)(ap->a_bp);
497 return (0);
498 }
499
500 int
501 spec_inactive(v)
502 void *v;
503 {
504 struct vop_inactive_args /* {
505 struct vnode *a_vp;
506 struct proc *a_p;
507 } */ *ap = v;
508
509 VOP_UNLOCK(ap->a_vp, 0);
510 return (0);
511 }
512
513 /*
514 * This is a noop, simply returning what one has been given.
515 */
516 int
517 spec_bmap(v)
518 void *v;
519 {
520 struct vop_bmap_args /* {
521 struct vnode *a_vp;
522 daddr_t a_bn;
523 struct vnode **a_vpp;
524 daddr_t *a_bnp;
525 int *a_runp;
526 } */ *ap = v;
527
528 if (ap->a_vpp != NULL)
529 *ap->a_vpp = ap->a_vp;
530 if (ap->a_bnp != NULL)
531 *ap->a_bnp = ap->a_bn;
532 if (ap->a_runp != NULL)
533 *ap->a_runp = 0;
534 return (0);
535 }
536
537 /*
538 * Device close routine
539 */
540 /* ARGSUSED */
541 int
542 spec_close(v)
543 void *v;
544 {
545 struct vop_close_args /* {
546 struct vnode *a_vp;
547 int a_fflag;
548 struct ucred *a_cred;
549 struct proc *a_p;
550 } */ *ap = v;
551 register struct vnode *vp = ap->a_vp;
552 dev_t dev = vp->v_rdev;
553 int (*devclose) __P((dev_t, int, int, struct proc *));
554 int mode, error, count, flags, flags1;
555
556 simple_lock(&vp->v_interlock);
557 count = vcount(vp);
558 flags = vp->v_flag;
559 simple_unlock(&vp->v_interlock);
560
561 switch (vp->v_type) {
562
563 case VCHR:
564 /*
565 * Hack: a tty device that is a controlling terminal
566 * has a reference from the session structure.
567 * We cannot easily tell that a character device is
568 * a controlling terminal, unless it is the closing
569 * process' controlling terminal. In that case,
570 * if the reference count is 2 (this last descriptor
571 * plus the session), release the reference from the session.
572 */
573 if (count == 2 && ap->a_p &&
574 vp == ap->a_p->p_session->s_ttyvp) {
575 vrele(vp);
576 count--;
577 ap->a_p->p_session->s_ttyvp = NULL;
578 }
579 /*
580 * If the vnode is locked, then we are in the midst
581 * of forcably closing the device, otherwise we only
582 * close on last reference.
583 */
584 if (count > 1 && (flags & VXLOCK) == 0)
585 return (0);
586 devclose = cdevsw[major(dev)].d_close;
587 mode = S_IFCHR;
588 break;
589
590 case VBLK:
591 /*
592 * On last close of a block device (that isn't mounted)
593 * we must invalidate any in core blocks, so that
594 * we can, for instance, change floppy disks.
595 */
596 error = vinvalbuf(vp, V_SAVE, ap->a_cred, ap->a_p, 0, 0);
597 if (error)
598 return (error);
599 /*
600 * We do not want to really close the device if it
601 * is still in use unless we are trying to close it
602 * forcibly. Since every use (buffer, vnode, swap, cmap)
603 * holds a reference to the vnode, and because we mark
604 * any other vnodes that alias this device, when the
605 * sum of the reference counts on all the aliased
606 * vnodes descends to one, we are on last close.
607 */
608 if (count > 1 && (flags & VXLOCK) == 0)
609 return (0);
610 devclose = bdevsw[major(dev)].d_close;
611 mode = S_IFBLK;
612 break;
613
614 default:
615 panic("spec_close: not special");
616 }
617
618 flags1 = ap->a_fflag;
619
620 /*
621 * if VXLOCK is set, then we're going away soon, so make this
622 * non-blocking. Also ensures that we won't wedge in vn_lock below.
623 */
624 if (flags & VXLOCK)
625 flags1 |= FNONBLOCK;
626
627 /*
628 * If we're able to block, release the vnode lock & reaquire. We
629 * might end up sleaping for someone else who wants our queues. They
630 * won't get them if we hold the vnode locked. Also, if VXLOCK is set,
631 * don't release the lock as we won't be able to regain it.
632 */
633 if (!(flags1 & FNONBLOCK))
634 VOP_UNLOCK(vp, 0);
635
636 error = (*devclose)(dev, flags1, mode, ap->a_p);
637
638 if (!(flags1 & FNONBLOCK))
639 vn_lock(vp, LK_EXCLUSIVE | LK_RETRY);
640
641 return (error);
642 }
643
644 /*
645 * Print out the contents of a special device vnode.
646 */
647 int
648 spec_print(v)
649 void *v;
650 {
651 struct vop_print_args /* {
652 struct vnode *a_vp;
653 } */ *ap = v;
654
655 printf("tag VT_NON, dev %d, %d\n", major(ap->a_vp->v_rdev),
656 minor(ap->a_vp->v_rdev));
657 return 0;
658 }
659
660 /*
661 * Return POSIX pathconf information applicable to special devices.
662 */
663 int
664 spec_pathconf(v)
665 void *v;
666 {
667 struct vop_pathconf_args /* {
668 struct vnode *a_vp;
669 int a_name;
670 register_t *a_retval;
671 } */ *ap = v;
672
673 switch (ap->a_name) {
674 case _PC_LINK_MAX:
675 *ap->a_retval = LINK_MAX;
676 return (0);
677 case _PC_MAX_CANON:
678 *ap->a_retval = MAX_CANON;
679 return (0);
680 case _PC_MAX_INPUT:
681 *ap->a_retval = MAX_INPUT;
682 return (0);
683 case _PC_PIPE_BUF:
684 *ap->a_retval = PIPE_BUF;
685 return (0);
686 case _PC_CHOWN_RESTRICTED:
687 *ap->a_retval = 1;
688 return (0);
689 case _PC_VDISABLE:
690 *ap->a_retval = _POSIX_VDISABLE;
691 return (0);
692 case _PC_SYNC_IO:
693 *ap->a_retval = 1;
694 return (0);
695 default:
696 return (EINVAL);
697 }
698 /* NOTREACHED */
699 }
700
701 /*
702 * Advisory record locking support.
703 */
704 int
705 spec_advlock(v)
706 void *v;
707 {
708 struct vop_advlock_args /* {
709 struct vnode *a_vp;
710 caddr_t a_id;
711 int a_op;
712 struct flock *a_fl;
713 int a_flags;
714 } */ *ap = v;
715 register struct vnode *vp = ap->a_vp;
716
717 return (lf_advlock(&vp->v_speclockf, (off_t)0, ap->a_id, ap->a_op,
718 ap->a_fl, ap->a_flags));
719 }
720