Home | History | Annotate | Line # | Download | only in lfs
lfs_vnops.c revision 1.25.2.1.2.5
      1 /*	$NetBSD: lfs_vnops.c,v 1.25.2.1.2.5 1999/08/31 21:03:46 perseant Exp $	*/
      2 
      3 /*-
      4  * Copyright (c) 1999 The NetBSD Foundation, Inc.
      5  * All rights reserved.
      6  *
      7  * This code is derived from software contributed to The NetBSD Foundation
      8  * by Konrad E. Schroder <perseant (at) hhhh.org>.
      9  *
     10  * Redistribution and use in source and binary forms, with or without
     11  * modification, are permitted provided that the following conditions
     12  * are met:
     13  * 1. Redistributions of source code must retain the above copyright
     14  *    notice, this list of conditions and the following disclaimer.
     15  * 2. Redistributions in binary form must reproduce the above copyright
     16  *    notice, this list of conditions and the following disclaimer in the
     17  *    documentation and/or other materials provided with the distribution.
     18  * 3. All advertising materials mentioning features or use of this software
     19  *    must display the following acknowledgement:
     20  *      This product includes software developed by the NetBSD
     21  *      Foundation, Inc. and its contributors.
     22  * 4. Neither the name of The NetBSD Foundation nor the names of its
     23  *    contributors may be used to endorse or promote products derived
     24  *    from this software without specific prior written permission.
     25  *
     26  * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
     27  * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
     28  * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
     29  * PURPOSE ARE DISCLAIMED.  IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
     30  * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
     31  * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
     32  * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
     33  * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
     34  * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
     35  * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
     36  * POSSIBILITY OF SUCH DAMAGE.
     37  */
     38 /*
     39  * Copyright (c) 1986, 1989, 1991, 1993, 1995
     40  *	The Regents of the University of California.  All rights reserved.
     41  *
     42  * Redistribution and use in source and binary forms, with or without
     43  * modification, are permitted provided that the following conditions
     44  * are met:
     45  * 1. Redistributions of source code must retain the above copyright
     46  *    notice, this list of conditions and the following disclaimer.
     47  * 2. Redistributions in binary form must reproduce the above copyright
     48  *    notice, this list of conditions and the following disclaimer in the
     49  *    documentation and/or other materials provided with the distribution.
     50  * 3. All advertising materials mentioning features or use of this software
     51  *    must display the following acknowledgement:
     52  *	This product includes software developed by the University of
     53  *	California, Berkeley and its contributors.
     54  * 4. Neither the name of the University nor the names of its contributors
     55  *    may be used to endorse or promote products derived from this software
     56  *    without specific prior written permission.
     57  *
     58  * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
     59  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
     60  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
     61  * ARE DISCLAIMED.  IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
     62  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
     63  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
     64  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
     65  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
     66  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
     67  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
     68  * SUCH DAMAGE.
     69  *
     70  *	@(#)lfs_vnops.c	8.13 (Berkeley) 6/10/95
     71  */
     72 
     73 #include <sys/param.h>
     74 #include <sys/systm.h>
     75 #include <sys/namei.h>
     76 #include <sys/resourcevar.h>
     77 #include <sys/kernel.h>
     78 #include <sys/file.h>
     79 #include <sys/stat.h>
     80 #include <sys/buf.h>
     81 #include <sys/proc.h>
     82 #include <sys/conf.h>
     83 #include <sys/mount.h>
     84 #include <sys/vnode.h>
     85 #include <sys/malloc.h>
     86 #include <sys/pool.h>
     87 #include <sys/signalvar.h>
     88 
     89 #include <vm/vm.h>
     90 #include <uvm/uvm.h>
     91 #include <uvm/uvm_pager.h> /* For PGO_* flags */
     92 
     93 #include <miscfs/fifofs/fifo.h>
     94 #include <miscfs/genfs/genfs.h>
     95 #include <miscfs/specfs/specdev.h>
     96 
     97 #include <ufs/ufs/quota.h>
     98 #include <ufs/ufs/inode.h>
     99 #include <ufs/ufs/dir.h>
    100 #include <ufs/ufs/ufsmount.h>
    101 #include <ufs/ufs/ufs_extern.h>
    102 
    103 #include <ufs/lfs/lfs.h>
    104 #include <ufs/lfs/lfs_extern.h>
    105 
    106 /* Global vfs data structures for lfs. */
    107 int (**lfs_vnodeop_p) __P((void *));
    108 struct vnodeopv_entry_desc lfs_vnodeop_entries[] = {
    109 	{ &vop_default_desc, vn_default_error },
    110 	{ &vop_lookup_desc, ufs_lookup },		/* lookup */
    111 	{ &vop_create_desc, lfs_create },		/* create */
    112 	{ &vop_whiteout_desc, lfs_whiteout },		/* whiteout */
    113 	{ &vop_mknod_desc, lfs_mknod },			/* mknod */
    114 	{ &vop_open_desc, ufs_open },			/* open */
    115 	{ &vop_close_desc, lfs_close },			/* close */
    116 	{ &vop_access_desc, ufs_access },		/* access */
    117 	{ &vop_getattr_desc, lfs_getattr },		/* getattr */
    118 	{ &vop_setattr_desc, ufs_setattr },		/* setattr */
    119 	{ &vop_read_desc, lfs_read },			/* read */
    120 	{ &vop_write_desc, lfs_write },			/* write */
    121 	{ &vop_lease_desc, ufs_lease_check },		/* lease */
    122 	{ &vop_ioctl_desc, ufs_ioctl },			/* ioctl */
    123 	{ &vop_poll_desc, ufs_poll },			/* poll */
    124 	{ &vop_revoke_desc, ufs_revoke },		/* revoke */
    125 	{ &vop_mmap_desc, ufs_mmap },			/* mmap */
    126 	{ &vop_fsync_desc, lfs_fsync },			/* fsync */
    127 	{ &vop_seek_desc, ufs_seek },			/* seek */
    128 	{ &vop_remove_desc, lfs_remove },		/* remove */
    129 	{ &vop_link_desc, lfs_link },			/* link */
    130 	{ &vop_rename_desc, lfs_rename },		/* rename */
    131 	{ &vop_mkdir_desc, lfs_mkdir },			/* mkdir */
    132 	{ &vop_rmdir_desc, lfs_rmdir },			/* rmdir */
    133 	{ &vop_symlink_desc, lfs_symlink },		/* symlink */
    134 	{ &vop_readdir_desc, ufs_readdir },		/* readdir */
    135 	{ &vop_readlink_desc, ufs_readlink },		/* readlink */
    136 	{ &vop_abortop_desc, ufs_abortop },		/* abortop */
    137 	{ &vop_inactive_desc, ufs_inactive },		/* inactive */
    138 	{ &vop_reclaim_desc, lfs_reclaim },		/* reclaim */
    139 	{ &vop_lock_desc, ufs_lock },			/* lock */
    140 	{ &vop_unlock_desc, ufs_unlock },		/* unlock */
    141 	{ &vop_bmap_desc, ufs_bmap },			/* bmap */
    142 	{ &vop_strategy_desc, ufs_strategy },		/* strategy */
    143 	{ &vop_print_desc, ufs_print },			/* print */
    144 	{ &vop_islocked_desc, ufs_islocked },		/* islocked */
    145 	{ &vop_pathconf_desc, ufs_pathconf },		/* pathconf */
    146 	{ &vop_advlock_desc, ufs_advlock },		/* advlock */
    147 	{ &vop_blkatoff_desc, lfs_blkatoff },		/* blkatoff */
    148 	{ &vop_valloc_desc, lfs_valloc },		/* valloc */
    149 	{ &vop_balloc_desc, lfs_balloc },		/* balloc */
    150 	{ &vop_vfree_desc, lfs_vfree },			/* vfree */
    151 	{ &vop_truncate_desc, lfs_truncate },		/* truncate */
    152 	{ &vop_update_desc, lfs_update },		/* update */
    153 	{ &vop_bwrite_desc, lfs_bwrite },		/* bwrite */
    154 	{ &vop_getpages_desc, lfs_getpages },		/* getpages */
    155 	{ &vop_putpages_desc, lfs_putpages },		/* putpages */
    156 	{ (struct vnodeop_desc*)NULL, (int(*) __P((void *)))NULL }
    157 };
    158 struct vnodeopv_desc lfs_vnodeop_opv_desc =
    159 	{ &lfs_vnodeop_p, lfs_vnodeop_entries };
    160 
    161 int (**lfs_specop_p) __P((void *));
    162 struct vnodeopv_entry_desc lfs_specop_entries[] = {
    163 	{ &vop_default_desc, vn_default_error },
    164 	{ &vop_lookup_desc, spec_lookup },		/* lookup */
    165 	{ &vop_create_desc, spec_create },		/* create */
    166 	{ &vop_mknod_desc, spec_mknod },		/* mknod */
    167 	{ &vop_open_desc, spec_open },			/* open */
    168 	{ &vop_close_desc, ufsspec_close },		/* close */
    169 	{ &vop_access_desc, ufs_access },		/* access */
    170 	{ &vop_getattr_desc, lfs_getattr },		/* getattr */
    171 	{ &vop_setattr_desc, ufs_setattr },		/* setattr */
    172 	{ &vop_read_desc, ufsspec_read },		/* read */
    173 	{ &vop_write_desc, ufsspec_write },		/* write */
    174 	{ &vop_lease_desc, spec_lease_check },		/* lease */
    175 	{ &vop_ioctl_desc, spec_ioctl },		/* ioctl */
    176 	{ &vop_poll_desc, spec_poll },			/* poll */
    177 	{ &vop_revoke_desc, spec_revoke },		/* revoke */
    178 	{ &vop_mmap_desc, spec_mmap },			/* mmap */
    179 	{ &vop_fsync_desc, spec_fsync },		/* fsync */
    180 	{ &vop_seek_desc, spec_seek },			/* seek */
    181 	{ &vop_remove_desc, spec_remove },		/* remove */
    182 	{ &vop_link_desc, spec_link },			/* link */
    183 	{ &vop_rename_desc, spec_rename },		/* rename */
    184 	{ &vop_mkdir_desc, spec_mkdir },		/* mkdir */
    185 	{ &vop_rmdir_desc, spec_rmdir },		/* rmdir */
    186 	{ &vop_symlink_desc, spec_symlink },		/* symlink */
    187 	{ &vop_readdir_desc, spec_readdir },		/* readdir */
    188 	{ &vop_readlink_desc, spec_readlink },		/* readlink */
    189 	{ &vop_abortop_desc, spec_abortop },		/* abortop */
    190 	{ &vop_inactive_desc, ufs_inactive },		/* inactive */
    191 	{ &vop_reclaim_desc, lfs_reclaim },		/* reclaim */
    192 	{ &vop_lock_desc, ufs_lock },			/* lock */
    193 	{ &vop_unlock_desc, ufs_unlock },		/* unlock */
    194 	{ &vop_bmap_desc, spec_bmap },			/* bmap */
    195 	{ &vop_strategy_desc, spec_strategy },		/* strategy */
    196 	{ &vop_print_desc, ufs_print },			/* print */
    197 	{ &vop_islocked_desc, ufs_islocked },		/* islocked */
    198 	{ &vop_pathconf_desc, spec_pathconf },		/* pathconf */
    199 	{ &vop_advlock_desc, spec_advlock },		/* advlock */
    200 	{ &vop_blkatoff_desc, spec_blkatoff },		/* blkatoff */
    201 	{ &vop_valloc_desc, spec_valloc },		/* valloc */
    202 	{ &vop_vfree_desc, lfs_vfree },			/* vfree */
    203 	{ &vop_truncate_desc, spec_truncate },		/* truncate */
    204 	{ &vop_update_desc, lfs_update },		/* update */
    205 	{ &vop_bwrite_desc, lfs_bwrite },		/* bwrite */
    206 	{ (struct vnodeop_desc*)NULL, (int(*) __P((void *)))NULL }
    207 };
    208 struct vnodeopv_desc lfs_specop_opv_desc =
    209 	{ &lfs_specop_p, lfs_specop_entries };
    210 
    211 int (**lfs_fifoop_p) __P((void *));
    212 struct vnodeopv_entry_desc lfs_fifoop_entries[] = {
    213 	{ &vop_default_desc, vn_default_error },
    214 	{ &vop_lookup_desc, fifo_lookup },		/* lookup */
    215 	{ &vop_create_desc, fifo_create },		/* create */
    216 	{ &vop_mknod_desc, fifo_mknod },		/* mknod */
    217 	{ &vop_open_desc, fifo_open },			/* open */
    218 	{ &vop_close_desc, ufsfifo_close },		/* close */
    219 	{ &vop_access_desc, ufs_access },		/* access */
    220 	{ &vop_getattr_desc, lfs_getattr },		/* getattr */
    221 	{ &vop_setattr_desc, ufs_setattr },		/* setattr */
    222 	{ &vop_read_desc, ufsfifo_read },		/* read */
    223 	{ &vop_write_desc, ufsfifo_write },		/* write */
    224 	{ &vop_lease_desc, fifo_lease_check },		/* lease */
    225 	{ &vop_ioctl_desc, fifo_ioctl },		/* ioctl */
    226 	{ &vop_poll_desc, fifo_poll },			/* poll */
    227 	{ &vop_revoke_desc, fifo_revoke },		/* revoke */
    228 	{ &vop_mmap_desc, fifo_mmap },			/* mmap */
    229 	{ &vop_fsync_desc, fifo_fsync },		/* fsync */
    230 	{ &vop_seek_desc, fifo_seek },			/* seek */
    231 	{ &vop_remove_desc, fifo_remove },		/* remove */
    232 	{ &vop_link_desc, fifo_link },			/* link */
    233 	{ &vop_rename_desc, fifo_rename },		/* rename */
    234 	{ &vop_mkdir_desc, fifo_mkdir },		/* mkdir */
    235 	{ &vop_rmdir_desc, fifo_rmdir },		/* rmdir */
    236 	{ &vop_symlink_desc, fifo_symlink },		/* symlink */
    237 	{ &vop_readdir_desc, fifo_readdir },		/* readdir */
    238 	{ &vop_readlink_desc, fifo_readlink },		/* readlink */
    239 	{ &vop_abortop_desc, fifo_abortop },		/* abortop */
    240 	{ &vop_inactive_desc, ufs_inactive },		/* inactive */
    241 	{ &vop_reclaim_desc, lfs_reclaim },		/* reclaim */
    242 	{ &vop_lock_desc, ufs_lock },			/* lock */
    243 	{ &vop_unlock_desc, ufs_unlock },		/* unlock */
    244 	{ &vop_bmap_desc, fifo_bmap },			/* bmap */
    245 	{ &vop_strategy_desc, fifo_strategy },		/* strategy */
    246 	{ &vop_print_desc, ufs_print },			/* print */
    247 	{ &vop_islocked_desc, ufs_islocked },		/* islocked */
    248 	{ &vop_pathconf_desc, fifo_pathconf },		/* pathconf */
    249 	{ &vop_advlock_desc, fifo_advlock },		/* advlock */
    250 	{ &vop_blkatoff_desc, fifo_blkatoff },		/* blkatoff */
    251 	{ &vop_valloc_desc, fifo_valloc },		/* valloc */
    252 	{ &vop_vfree_desc, lfs_vfree },			/* vfree */
    253 	{ &vop_truncate_desc, fifo_truncate },		/* truncate */
    254 	{ &vop_update_desc, lfs_update },		/* update */
    255 	{ &vop_bwrite_desc, lfs_bwrite },		/* bwrite */
    256 	{ (struct vnodeop_desc*)NULL, (int(*) __P((void *)))NULL }
    257 };
    258 struct vnodeopv_desc lfs_fifoop_opv_desc =
    259 	{ &lfs_fifoop_p, lfs_fifoop_entries };
    260 
    261 #define	LFS_READWRITE
    262 #include <ufs/ufs/ufs_readwrite.c>
    263 #undef	LFS_READWRITE
    264 
    265 /*
    266  * Synch an open file.
    267  */
    268 /* ARGSUSED */
    269 int
    270 lfs_fsync(v)
    271 	void *v;
    272 {
    273 	struct vop_fsync_args /* {
    274 		struct vnode *a_vp;
    275 		struct ucred *a_cred;
    276 		int a_flags;
    277 		struct proc *a_p;
    278 	} */ *ap = v;
    279 
    280 	return (VOP_UPDATE(ap->a_vp, NULL, NULL,
    281 			   (ap->a_flags & FSYNC_WAIT) != 0 ? LFS_SYNC : 0)); /* XXX */
    282 }
    283 
    284 /*
    285  * These macros are used to bracket UFS directory ops, so that we can
    286  * identify all the pages touched during directory ops which need to
    287  * be ordered and flushed atomically, so that they may be recovered.
    288  */
    289 /*
    290  * XXX KS - Because we have to mark nodes VDIROP in order to prevent
    291  * the cache from reclaiming them while a dirop is in progress, we must
    292  * also manage the number of nodes so marked (otherwise we can run out).
    293  * We do this by setting lfs_dirvcount to the number of marked vnodes; it
    294  * is decremented during segment write, when VDIROP is taken off.
    295  */
    296 #define	SET_DIROP(fs) lfs_set_dirop(fs)
    297 static int lfs_set_dirop __P((struct lfs *));
    298 
    299 static int lfs_set_dirop(fs)
    300 	struct lfs *fs;
    301 {
    302 	int error;
    303 
    304 	while (fs->lfs_writer || fs->lfs_dirvcount>LFS_MAXDIROP) {
    305 		if(fs->lfs_writer)
    306 			tsleep(&fs->lfs_dirops, PRIBIO + 1, "lfs_dirop", 0);
    307 		if(fs->lfs_dirvcount > LFS_MAXDIROP) {
    308 #ifdef DEBUG_LFS
    309 			printf("(dirvcount=%d)\n",fs->lfs_dirvcount);
    310 #endif
    311 			if((error=tsleep(&fs->lfs_dirvcount, PCATCH|PUSER, "lfs_maxdirop", 0))!=0)
    312 				return error;
    313 		}
    314 	}
    315 	++fs->lfs_dirops;
    316 	fs->lfs_doifile = 1;
    317 
    318 	return 0;
    319 }
    320 
    321 #define	SET_ENDOP(fs,vp,str) {						\
    322 	--(fs)->lfs_dirops;						\
    323 	if (!(fs)->lfs_dirops) {					\
    324 		wakeup(&(fs)->lfs_writer);				\
    325 		lfs_check((vp),LFS_UNUSED_LBN,0);			\
    326 	}								\
    327 }
    328 
    329 #define	MARK_VNODE(dvp)  do {                                           \
    330         if(!((dvp)->v_flag & VDIROP)) {					\
    331                 lfs_vref(dvp);						\
    332 		++VTOI((dvp))->i_lfs->lfs_dirvcount;			\
    333 	}								\
    334         (dvp)->v_flag |= VDIROP;					\
    335 } while(0)
    336 
    337 #define MAYBE_INACTIVE(fs,vp) do {                                      \
    338         if((vp) && ((vp)->v_flag & VDIROP) && (vp)->v_usecount == 1     \
    339            && VTOI(vp) && VTOI(vp)->i_ffs_nlink == 0)                   \
    340         {                                                               \
    341 		if (VOP_LOCK((vp), LK_EXCLUSIVE) == 0) { 		\
    342                         VOP_INACTIVE((vp),curproc);                     \
    343 		}                                                       \
    344         }                                                               \
    345 } while(0)
    346 
    347 int
    348 lfs_symlink(v)
    349 	void *v;
    350 {
    351 	struct vop_symlink_args /* {
    352 		struct vnode *a_dvp;
    353 		struct vnode **a_vpp;
    354 		struct componentname *a_cnp;
    355 		struct vattr *a_vap;
    356 		char *a_target;
    357 	} */ *ap = v;
    358 	int ret;
    359 
    360 	if((ret=SET_DIROP(VTOI(ap->a_dvp)->i_lfs))!=0)
    361 		return ret;
    362 	MARK_VNODE(ap->a_dvp);
    363 	ret = ufs_symlink(ap);
    364 	MAYBE_INACTIVE(VTOI(ap->a_dvp)->i_lfs,*(ap->a_vpp)); /* XXX KS */
    365 	SET_ENDOP(VTOI(ap->a_dvp)->i_lfs,ap->a_dvp,"symilnk");
    366 	return (ret);
    367 }
    368 
    369 int
    370 lfs_mknod(v)
    371 	void *v;
    372 {
    373 	struct vop_mknod_args	/* {
    374 		struct vnode *a_dvp;
    375 		struct vnode **a_vpp;
    376 		struct componentname *a_cnp;
    377 		struct vattr *a_vap;
    378 		} */ *ap = v;
    379 	int ret;
    380 
    381 	if((ret=SET_DIROP(VTOI(ap->a_dvp)->i_lfs))!=0)
    382 		return ret;
    383 	MARK_VNODE(ap->a_dvp);
    384 	ret = ufs_mknod(ap);
    385 	MAYBE_INACTIVE(VTOI(ap->a_dvp)->i_lfs,*(ap->a_vpp)); /* XXX KS */
    386 	SET_ENDOP(VTOI(ap->a_dvp)->i_lfs,ap->a_dvp,"mknod");
    387 	return (ret);
    388 }
    389 
    390 int
    391 lfs_create(v)
    392 	void *v;
    393 {
    394 	struct vop_create_args	/* {
    395 		struct vnode *a_dvp;
    396 		struct vnode **a_vpp;
    397 		struct componentname *a_cnp;
    398 		struct vattr *a_vap;
    399 	} */ *ap = v;
    400 	int ret;
    401 
    402 	if((ret=SET_DIROP(VTOI(ap->a_dvp)->i_lfs))!=0)
    403 		return ret;
    404 	MARK_VNODE(ap->a_dvp);
    405 	ret = ufs_create(ap);
    406 	MAYBE_INACTIVE(VTOI(ap->a_dvp)->i_lfs,*(ap->a_vpp)); /* XXX KS */
    407 	SET_ENDOP(VTOI(ap->a_dvp)->i_lfs,ap->a_dvp,"create");
    408 	return (ret);
    409 }
    410 
    411 int
    412 lfs_whiteout(v)
    413 	void *v;
    414 {
    415 	struct vop_whiteout_args /* {
    416 		struct vnode *a_dvp;
    417 		struct componentname *a_cnp;
    418 		int a_flags;
    419 	} */ *ap = v;
    420 	int ret;
    421 
    422 	if((ret=SET_DIROP(VTOI(ap->a_dvp)->i_lfs))!=0)
    423 		return ret;
    424 	MARK_VNODE(ap->a_dvp);
    425 	ret = ufs_whiteout(ap);
    426 	SET_ENDOP(VTOI(ap->a_dvp)->i_lfs,ap->a_dvp,"whiteout");
    427 	return (ret);
    428 }
    429 
    430 int
    431 lfs_mkdir(v)
    432 	void *v;
    433 {
    434 	struct vop_mkdir_args	/* {
    435 		struct vnode *a_dvp;
    436 		struct vnode **a_vpp;
    437 		struct componentname *a_cnp;
    438 		struct vattr *a_vap;
    439 	} */ *ap = v;
    440 	int ret;
    441 
    442 	if((ret=SET_DIROP(VTOI(ap->a_dvp)->i_lfs))!=0)
    443 		return ret;
    444 	MARK_VNODE(ap->a_dvp);
    445 	ret = ufs_mkdir(ap);
    446 	MAYBE_INACTIVE(VTOI(ap->a_dvp)->i_lfs,*(ap->a_vpp)); /* XXX KS */
    447 	SET_ENDOP(VTOI(ap->a_dvp)->i_lfs,ap->a_dvp,"mkdir");
    448 	return (ret);
    449 }
    450 
    451 int
    452 lfs_remove(v)
    453 	void *v;
    454 {
    455 	struct vop_remove_args	/* {
    456 		struct vnode *a_dvp;
    457 		struct vnode *a_vp;
    458 		struct componentname *a_cnp;
    459 	} */ *ap = v;
    460 	int ret;
    461 	if((ret=SET_DIROP(VTOI(ap->a_dvp)->i_lfs))!=0)
    462 		return ret;
    463 	MARK_VNODE(ap->a_dvp);
    464 	MARK_VNODE(ap->a_vp);
    465 	ret = ufs_remove(ap);
    466 	MAYBE_INACTIVE(VTOI(ap->a_dvp)->i_lfs,ap->a_vp);
    467 	SET_ENDOP(VTOI(ap->a_dvp)->i_lfs,ap->a_dvp,"remove");
    468 	return (ret);
    469 }
    470 
    471 int
    472 lfs_rmdir(v)
    473 	void *v;
    474 {
    475 	struct vop_rmdir_args	/* {
    476 		struct vnodeop_desc *a_desc;
    477 		struct vnode *a_dvp;
    478 		struct vnode *a_vp;
    479 		struct componentname *a_cnp;
    480 	} */ *ap = v;
    481 	int ret;
    482 
    483 	if((ret=SET_DIROP(VTOI(ap->a_dvp)->i_lfs))!=0)
    484 		return ret;
    485 	MARK_VNODE(ap->a_dvp);
    486 	MARK_VNODE(ap->a_vp);
    487 	ret = ufs_rmdir(ap);
    488 	MAYBE_INACTIVE(VTOI(ap->a_dvp)->i_lfs,ap->a_vp);
    489 	SET_ENDOP(VTOI(ap->a_dvp)->i_lfs,ap->a_dvp,"rmdir");
    490 	return (ret);
    491 }
    492 
    493 int
    494 lfs_link(v)
    495 	void *v;
    496 {
    497 	struct vop_link_args	/* {
    498 		struct vnode *a_dvp;
    499 		struct vnode *a_vp;
    500 		struct componentname *a_cnp;
    501 	} */ *ap = v;
    502 	int ret;
    503 
    504 	if((ret=SET_DIROP(VTOI(ap->a_dvp)->i_lfs))!=0)
    505 		return ret;
    506 	MARK_VNODE(ap->a_dvp);
    507 	ret = ufs_link(ap);
    508 	SET_ENDOP(VTOI(ap->a_dvp)->i_lfs,ap->a_dvp,"link");
    509 	return (ret);
    510 }
    511 
    512 int
    513 lfs_rename(v)
    514 	void *v;
    515 {
    516 	struct vop_rename_args	/* {
    517 		struct vnode *a_fdvp;
    518 		struct vnode *a_fvp;
    519 		struct componentname *a_fcnp;
    520 		struct vnode *a_tdvp;
    521 		struct vnode *a_tvp;
    522 		struct componentname *a_tcnp;
    523 	} */ *ap = v;
    524 	int ret;
    525 
    526 	if((ret=SET_DIROP(VTOI(ap->a_fdvp)->i_lfs))!=0)
    527 		return ret;
    528 	MARK_VNODE(ap->a_fdvp);
    529 	MARK_VNODE(ap->a_tdvp);
    530 	ret = ufs_rename(ap);
    531 	MAYBE_INACTIVE(VTOI(ap->a_dvp)->i_lfs,ap->a_fvp);
    532 	MAYBE_INACTIVE(VTOI(ap->a_dvp)->i_lfs,ap->a_tvp);
    533 	SET_ENDOP(VTOI(ap->a_fdvp)->i_lfs,ap->a_fdvp,"rename");
    534 	return (ret);
    535 }
    536 
    537 /* XXX hack to avoid calling ITIMES in getattr */
    538 int
    539 lfs_getattr(v)
    540 	void *v;
    541 {
    542 	struct vop_getattr_args /* {
    543 		struct vnode *a_vp;
    544 		struct vattr *a_vap;
    545 		struct ucred *a_cred;
    546 		struct proc *a_p;
    547 	} */ *ap = v;
    548 	register struct vnode *vp = ap->a_vp;
    549 	register struct inode *ip = VTOI(vp);
    550 	register struct vattr *vap = ap->a_vap;
    551 	/*
    552 	 * Copy from inode table
    553 	 */
    554 	vap->va_fsid = ip->i_dev;
    555 	vap->va_fileid = ip->i_number;
    556 	vap->va_mode = ip->i_ffs_mode & ~IFMT;
    557 	vap->va_nlink = ip->i_ffs_nlink;
    558 	vap->va_uid = ip->i_ffs_uid;
    559 	vap->va_gid = ip->i_ffs_gid;
    560 	vap->va_rdev = (dev_t)ip->i_ffs_rdev;
    561 	vap->va_size = ip->i_ffs_size;
    562 	vap->va_atime.tv_sec = ip->i_ffs_atime;
    563 	vap->va_atime.tv_nsec = ip->i_ffs_atimensec;
    564 	vap->va_mtime.tv_sec = ip->i_ffs_mtime;
    565 	vap->va_mtime.tv_nsec = ip->i_ffs_mtimensec;
    566 	vap->va_ctime.tv_sec = ip->i_ffs_ctime;
    567 	vap->va_ctime.tv_nsec = ip->i_ffs_ctimensec;
    568 	vap->va_flags = ip->i_ffs_flags;
    569 	vap->va_gen = ip->i_ffs_gen;
    570 	/* this doesn't belong here */
    571 	if (vp->v_type == VBLK)
    572 		vap->va_blocksize = BLKDEV_IOSIZE;
    573 	else if (vp->v_type == VCHR)
    574 		vap->va_blocksize = MAXBSIZE;
    575 	else
    576 		vap->va_blocksize = vp->v_mount->mnt_stat.f_iosize;
    577 	vap->va_bytes = dbtob((u_quad_t)ip->i_ffs_blocks);
    578 	vap->va_type = vp->v_type;
    579 	vap->va_filerev = ip->i_modrev;
    580 	return (0);
    581 }
    582 
    583 /*
    584  * Close called
    585  *
    586  * XXX -- we were using ufs_close, but since it updates the
    587  * times on the inode, we might need to bump the uinodes
    588  * count.
    589  */
    590 /* ARGSUSED */
    591 int
    592 lfs_close(v)
    593 	void *v;
    594 {
    595 	struct vop_close_args /* {
    596 		struct vnode *a_vp;
    597 		int  a_fflag;
    598 		struct ucred *a_cred;
    599 		struct proc *a_p;
    600 	} */ *ap = v;
    601 	register struct vnode *vp = ap->a_vp;
    602 	register struct inode *ip = VTOI(vp);
    603 	int mod;
    604 	struct timespec ts;
    605 
    606 	simple_lock(&vp->v_interlock);
    607 	if (vp->v_usecount > 1) {
    608 		mod = ip->i_flag & IN_MODIFIED;
    609 		TIMEVAL_TO_TIMESPEC(&time, &ts);
    610 		LFS_ITIMES(ip, &ts, &ts, &ts);
    611 		if (!mod && ip->i_flag & IN_MODIFIED)
    612 			ip->i_lfs->lfs_uinodes++;
    613 	}
    614 	simple_unlock(&vp->v_interlock);
    615 	return (0);
    616 }
    617 
    618 /*
    619  * Reclaim an inode so that it can be used for other purposes.
    620  */
    621 int lfs_no_inactive = 0;
    622 
    623 int
    624 lfs_reclaim(v)
    625 	void *v;
    626 {
    627 	struct vop_reclaim_args /* {
    628 		struct vnode *a_vp;
    629 		struct proc *a_p;
    630 	} */ *ap = v;
    631 	struct vnode *vp = ap->a_vp;
    632 	int error;
    633 
    634 	if ((error = ufs_reclaim(vp, ap->a_p)))
    635 		return (error);
    636 	pool_put(&lfs_inode_pool, vp->v_data);
    637 	vp->v_data = NULL;
    638 	return (0);
    639 }
    640 
    641 #define BUFHEAD_THROTTLE 128
    642 static unsigned long bufhead_count, bufhead_rem;
    643 
    644 static void
    645 lfs_bufcallback(struct buf *bp)
    646 {
    647 	int s;
    648 
    649 	s = splbio();
    650 	if(bp->b_vp)
    651 		brelvp(bp);
    652 	--bufhead_count;
    653 	splx(s);
    654 
    655 	uvm_aio_biodone1(bp); /* maybe set mbp->b_iodone = aio_aiodone */
    656 	return;
    657 }
    658 
    659 /*
    660  * LFS-specific VM putpages routine.
    661  * Encapsulate the pages in buffer headers, and call lfs_segwrite to write
    662  * them out.
    663  *
    664  * Mostly taken from genfs_putpages.
    665  *
    666  * XXX Assumes that fs block size is <= page size, in order to avoid
    667  * XXX having to read partial blocks in from disk.
    668  *
    669  * Ideally we would use pagemove() and not have this requirement
    670  * (or just call lfs_segwrite with some notation of which pages
    671  * were busy).
    672  */
    673 
    674 #define SET_FLUSHING(fs,vp) (fs)->lfs_flushvp = (vp)
    675 #define IS_FLUSHING(fs,vp)  ((fs)->lfs_flushvp == (vp))
    676 #define CLR_FLUSHING(fs,vp) (fs)->lfs_flushvp = NULL
    677 
    678 /* Statistics Counters */
    679 extern int lfs_dostats;
    680 extern struct lfs_stats lfs_stats;
    681 
    682 /*
    683  * Utility function for flushing a single vnode, in a case where
    684  * lfs_segwrite might deadlock.  Assumes that we hold the seglock.
    685  * Excerpted from lfs_vflush. (XXX should be combined)
    686  */
    687 static inline void
    688 lfs_writevnode(struct lfs *fs, struct vnode *vp)
    689 {
    690 	struct segment *sp;
    691 	struct inode *ip;
    692 
    693 	ip = VTOI(vp);
    694 	sp = fs->lfs_sp;
    695 
    696 	do {
    697 		do {
    698 			if (vp->v_dirtyblkhd.lh_first != NULL)
    699 				lfs_writefile(fs, sp, vp);
    700 		} while (lfs_writeinode(fs, sp, ip));
    701 	} while (lfs_writeseg(fs, sp) && ip->i_number == LFS_IFILE_INUM);
    702 
    703 	if(lfs_dostats) {
    704 		++lfs_stats.nwrites;
    705 		if (sp->seg_flags & SEGM_SYNC)
    706 			++lfs_stats.nsync_writes;
    707 		if (sp->seg_flags & SEGM_CKP)
    708 			++lfs_stats.ncheckpoints;
    709 	}
    710 }
    711 
    712 int
    713 lfs_putpages(v)
    714 	void *v;
    715 {
    716 	struct vop_putpages_args /* {
    717 		struct vnode *a_vp;
    718 		struct vm_page **a_m;
    719 		int a_count;
    720 		int a_flags;
    721 		int *a_rtvals;
    722 	} */ *ap = v;
    723 
    724 	int s, error, npages, bshift, dev_bshift, dev_bsize;
    725 	char * kva;
    726 	off_t offset, startoffset;
    727 	size_t bytes, iobytes, skipbytes;
    728 	daddr_t lbn;
    729 	struct vm_page *pg;
    730 #if 0
    731 	struct vm_page *vmpp, *vmpp2;
    732 #endif
    733 	struct buf *mbp, *bp;
    734 	struct vnode *vp = ap->a_vp;
    735 	struct lfs *fs;
    736 	struct inode *ip;
    737 
    738 	UVMHIST_FUNC("lfs_putpages"); UVMHIST_CALLED(ubchist);
    739 
    740 	ip = VTOI(vp);
    741 	fs = ip->i_lfs;
    742 
    743 	lfs_seglock(fs,((ap->a_flags & PGO_SYNCIO) ? SEGM_SYNC : 0));
    744 
    745 	/* If we're flushing, segwrite needs to know not to wait */
    746 	if(ap->a_flags & PGO_ALLPAGES) {
    747 		SET_FLUSHING(fs,vp);
    748 	}
    749 
    750 	error = 0;
    751 	npages = ap->a_count;
    752 	bshift = vp->v_mount->mnt_fs_bshift;
    753 	dev_bshift = vp->v_mount->mnt_dev_bshift;
    754 	dev_bsize = 1 << dev_bshift;
    755 
    756 	pg = ap->a_m[0];
    757 	startoffset = pg->offset;
    758 	bytes = npages << PAGE_SHIFT;
    759 	skipbytes = 0;
    760 
    761 	if (bytes == 0) {
    762 		panic("lfs_putpages: bytes == 0??? vp %p", vp);
    763 	}
    764 
    765 #if 0
    766 	/*
    767 	 * Fix start/end addresses to write entire blocks/fragments
    768 	 * XXX This can deadlock us; can the pager do this klustering
    769 	 * for us?
    770 	 */
    771 	lbn = startoffset >> bshift;
    772 	if(startoffset != (lbn << bshift)) {
    773 		printf("Reading %d bytes at beginning of putpages\n",
    774 			startoffset - (lbn<<bshift));
    775 		/* Make sure the request is valid - XXX deadlock */
    776 		VOP_GETPAGES(vp, (lbn<<bshift), &vmpp, (startoffset - (lbn<<bshift)) >> pgshift, &newpages,
    777 			0, VM_PROT_DEFAULT, 0, 0);
    778 		bytes += (startoffset - (lbn << bshift));
    779 	        startoffset = lbn << bshift;
    780 		pg = vmpp[0];
    781 		printf("Done reading beginning.\n");
    782 	}
    783 	if(bytes & fs->lfs_bmask) {
    784 		new_lastsize = blksize(fs, ip, startoffset + bytes);
    785 		diff = new_lastsize - (bytes & fs->lfs_bmask);
    786 		if(diff) {
    787 			printf("Reading %d bytes at end of putpages\n", diff);
    788 			/* Make sure request is valid - XXX deadlock */
    789 			VOP_GETPAGES(vp, startoffset+bytes, &vmpp2, diff, &newpages, 0,
    790 				VM_PROT_DEFAULT, 0, 0);
    791 			bytes += diff;
    792 			printf("Done reading end.\n");
    793 		}
    794 	}
    795 #endif
    796 
    797 	kva = (void *)uvm_pagermapin(ap->a_m, npages, M_WAITOK);
    798 
    799 	s = splbio();
    800 	vp->v_numoutput++;
    801 	mbp = pool_get(&bufpool, PR_WAITOK);
    802 	UVMHIST_LOG(ubchist, "master vp %p bp %p num now %d",
    803 		    vp, mbp, vp->v_numoutput, 0);
    804 	splx(s);
    805 	mbp->b_bufsize = npages << PAGE_SHIFT;
    806 	mbp->b_data = kva;
    807 	mbp->b_resid = mbp->b_bcount = bytes;
    808 	mbp->b_flags = B_BUSY|B_WRITE|(ap->a_flags & PGO_SYNCIO ? 0 : B_CALL)|
    809 		(curproc == uvm.pagedaemon_proc ? B_PDAEMON : 0);
    810 	mbp->b_iodone = uvm_aio_biodone;
    811 	mbp->b_vp = vp;
    812 
    813 	bp = NULL;
    814 
    815 	offset = startoffset;
    816 
    817 	for (; bytes > 0; offset += iobytes, bytes -= iobytes) {
    818 		lbn = offset >> bshift;
    819 		iobytes = min((1<<bshift), bytes);
    820 
    821 		if(bufhead_rem > BUFHEAD_THROTTLE) {
    822 			if(ap->a_flags & PGO_ALLPAGES)
    823 				lfs_writevnode(fs, vp);
    824 			else
    825 				lfs_segwrite(vp->v_mount, 0);
    826 			bufhead_rem = 0;
    827 			/* Don't sleep here, there's no point XXX ? */
    828 		}
    829 
    830 		++bufhead_rem;
    831 		s = splbio();
    832 		++bufhead_count;
    833 		/* vp->v_numoutput++; */
    834 		bp = pool_get(&bufpool, PR_WAITOK);
    835 		UVMHIST_LOG(ubchist, "vp %p bp %p num now %d",
    836 			    vp, bp, vp->v_numoutput, 0);
    837 		memset(bp,0,sizeof(*bp));
    838 		bgetvp(vp, bp);
    839 		splx(s);
    840 
    841 		bp->b_data = kva + offset - pg->offset;
    842 		bp->b_resid = bp->b_bcount = bp->b_bufsize = iobytes;
    843 		bp->b_flags = B_BUSY|B_WRITE|B_CALL;
    844 		bp->b_iodone = lfs_bufcallback /* uvm_aio_biodone1 */ ;
    845 		bp->b_lblkno = bp->b_blkno = lbn;
    846 		bp->b_private = mbp;
    847 
    848 #ifdef DIAGNOSTIC
    849 		if(offset != (lbn<<bshift)) {
    850 			printf("off=0x%lx, lbn<<bshift=0x%lx\n",
    851 				(unsigned long)offset,
    852 				(unsigned long)(lbn<<bshift));
    853 			panic("Oops, fs bsize is %u but page size is apparently %u\n", (unsigned)1<<bshift, (unsigned)bp->b_bcount);
    854 		}
    855 		if(iobytes != (1<<bshift)) {
    856 			printf("iobytes=0x%lx, 1<<bshift=0x%lx\n",
    857 				(unsigned long)iobytes,
    858 				(unsigned long)(1<<bshift));
    859 			panic("Oops, fs bsize is %u but page size is apparently %u\n", (unsigned)1<<bshift, (unsigned)bp->b_bcount);
    860 		}
    861 #endif
    862 		VOP_BWRITE(bp);
    863 	}
    864 
    865 	/*
    866 	 * The pager wants these buffers to be written right now (async may
    867 	 * be okay, but definitely start right away).
    868 	 *
    869 	 * Call segwrite to start the write.
    870 	 */
    871 	if (skipbytes) {
    872 		s = splbio();
    873 		mbp->b_resid -= skipbytes;
    874 		if (mbp->b_resid == 0)
    875 			biodone(mbp);
    876 		splx(s);
    877 	}
    878 	if(bufhead_rem) {
    879 		if(ap->a_flags & PGO_ALLPAGES)
    880 			lfs_writevnode(fs, vp);
    881 		else
    882 			lfs_segwrite(vp->v_mount,
    883 				     ((ap->a_flags & PGO_SYNCIO) ?
    884 				      SEGM_SYNC : 0));
    885 		bufhead_rem = 0;
    886 		if (!(ap->a_flags & PGO_SYNCIO)) {
    887 			CLR_FLUSHING(fs,vp);
    888 			lfs_segunlock(fs);
    889 			return EINPROGRESS;
    890 		}
    891 	}
    892 
    893 	/*
    894 	 * By the time we reach this point, all buffers should be recovered,
    895 	 * and the mbp should be waiting for aiodoned to reclaim it.  We're writing
    896 	 * synchronously, so we have to wait for this to happen.
    897 	 */
    898  /* errout: */
    899 	CLR_FLUSHING(fs,vp);
    900 	lfs_segunlock(fs);
    901 
    902 	if (bufhead_count) {
    903 		error = biowait(mbp);
    904 	}
    905 	s = splbio();
    906 	pool_put(&bufpool, mbp);
    907 	splx(s);
    908 
    909 	uvm_pagermapout((vaddr_t)kva, npages);
    910 	UVMHIST_LOG(ubchist, "returning, error %d", error,0,0,0);
    911 	return error;
    912 }
    913