Home | History | Annotate | Line # | Download | only in lfs
lfs_balloc.c revision 1.85
      1 /*	$NetBSD: lfs_balloc.c,v 1.85 2015/08/02 18:08:13 dholland Exp $	*/
      2 
      3 /*-
      4  * Copyright (c) 1999, 2000, 2001, 2002, 2003 The NetBSD Foundation, Inc.
      5  * All rights reserved.
      6  *
      7  * This code is derived from software contributed to The NetBSD Foundation
      8  * by Konrad E. Schroder <perseant (at) hhhh.org>.
      9  *
     10  * Redistribution and use in source and binary forms, with or without
     11  * modification, are permitted provided that the following conditions
     12  * are met:
     13  * 1. Redistributions of source code must retain the above copyright
     14  *    notice, this list of conditions and the following disclaimer.
     15  * 2. Redistributions in binary form must reproduce the above copyright
     16  *    notice, this list of conditions and the following disclaimer in the
     17  *    documentation and/or other materials provided with the distribution.
     18  *
     19  * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
     20  * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
     21  * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
     22  * PURPOSE ARE DISCLAIMED.  IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
     23  * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
     24  * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
     25  * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
     26  * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
     27  * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
     28  * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
     29  * POSSIBILITY OF SUCH DAMAGE.
     30  */
     31 /*
     32  * Copyright (c) 1989, 1991, 1993
     33  *	The Regents of the University of California.  All rights reserved.
     34  *
     35  * Redistribution and use in source and binary forms, with or without
     36  * modification, are permitted provided that the following conditions
     37  * are met:
     38  * 1. Redistributions of source code must retain the above copyright
     39  *    notice, this list of conditions and the following disclaimer.
     40  * 2. Redistributions in binary form must reproduce the above copyright
     41  *    notice, this list of conditions and the following disclaimer in the
     42  *    documentation and/or other materials provided with the distribution.
     43  * 3. Neither the name of the University nor the names of its contributors
     44  *    may be used to endorse or promote products derived from this software
     45  *    without specific prior written permission.
     46  *
     47  * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
     48  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
     49  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
     50  * ARE DISCLAIMED.  IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
     51  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
     52  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
     53  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
     54  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
     55  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
     56  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
     57  * SUCH DAMAGE.
     58  *
     59  *	@(#)lfs_balloc.c	8.4 (Berkeley) 5/8/95
     60  */
     61 
     62 #include <sys/cdefs.h>
     63 __KERNEL_RCSID(0, "$NetBSD: lfs_balloc.c,v 1.85 2015/08/02 18:08:13 dholland Exp $");
     64 
     65 #if defined(_KERNEL_OPT)
     66 #include "opt_quota.h"
     67 #endif
     68 
     69 #include <sys/param.h>
     70 #include <sys/systm.h>
     71 #include <sys/buf.h>
     72 #include <sys/proc.h>
     73 #include <sys/vnode.h>
     74 #include <sys/mount.h>
     75 #include <sys/resourcevar.h>
     76 #include <sys/tree.h>
     77 #include <sys/trace.h>
     78 #include <sys/kauth.h>
     79 
     80 #include <miscfs/specfs/specdev.h>
     81 
     82 #include <ufs/lfs/ulfs_quotacommon.h>
     83 #include <ufs/lfs/ulfs_inode.h>
     84 #include <ufs/lfs/ulfsmount.h>
     85 #include <ufs/lfs/ulfs_extern.h>
     86 
     87 #include <ufs/lfs/lfs.h>
     88 #include <ufs/lfs/lfs_accessors.h>
     89 #include <ufs/lfs/lfs_extern.h>
     90 #include <ufs/lfs/lfs_kernel.h>
     91 
     92 #include <uvm/uvm.h>
     93 
     94 int lfs_fragextend(struct vnode *, int, int, daddr_t, struct buf **, kauth_cred_t);
     95 
     96 u_int64_t locked_fakequeue_count;
     97 
     98 /*
     99  * Allocate a block, and to inode and filesystem block accounting for it
    100  * and for any indirect blocks the may need to be created in order for
    101  * this block to be created.
    102  *
    103  * Blocks which have never been accounted for (i.e., which "do not exist")
    104  * have disk address 0, which is translated by ulfs_bmap to the special value
    105  * UNASSIGNED == -1, as in the historical ULFS.
    106  *
    107  * Blocks which have been accounted for but which have not yet been written
    108  * to disk are given the new special disk address UNWRITTEN == -2, so that
    109  * they can be differentiated from completely new blocks.
    110  */
    111 /* VOP_BWRITE ULFS_NIADDR+2 times */
    112 int
    113 lfs_balloc(struct vnode *vp, off_t startoffset, int iosize, kauth_cred_t cred,
    114     int flags, struct buf **bpp)
    115 {
    116 	int offset;
    117 	daddr_t daddr, idaddr;
    118 	struct buf *ibp, *bp;
    119 	struct inode *ip;
    120 	struct lfs *fs;
    121 	struct indir indirs[ULFS_NIADDR+2], *idp;
    122 	daddr_t	lbn, lastblock;
    123 	int bcount;
    124 	int error, frags, i, nsize, osize, num;
    125 
    126 	ip = VTOI(vp);
    127 	fs = ip->i_lfs;
    128 	offset = lfs_blkoff(fs, startoffset);
    129 	KASSERT(iosize <= lfs_sb_getbsize(fs));
    130 	lbn = lfs_lblkno(fs, startoffset);
    131 	/* (void)lfs_check(vp, lbn, 0); */
    132 
    133 	ASSERT_MAYBE_SEGLOCK(fs);
    134 
    135 	/*
    136 	 * Three cases: it's a block beyond the end of file, it's a block in
    137 	 * the file that may or may not have been assigned a disk address or
    138 	 * we're writing an entire block.
    139 	 *
    140 	 * Note, if the daddr is UNWRITTEN, the block already exists in
    141 	 * the cache (it was read or written earlier).	If so, make sure
    142 	 * we don't count it as a new block or zero out its contents. If
    143 	 * it did not, make sure we allocate any necessary indirect
    144 	 * blocks.
    145 	 *
    146 	 * If we are writing a block beyond the end of the file, we need to
    147 	 * check if the old last block was a fragment.	If it was, we need
    148 	 * to rewrite it.
    149 	 */
    150 
    151 	if (bpp)
    152 		*bpp = NULL;
    153 
    154 	/* Check for block beyond end of file and fragment extension needed. */
    155 	lastblock = lfs_lblkno(fs, ip->i_size);
    156 	if (lastblock < ULFS_NDADDR && lastblock < lbn) {
    157 		osize = lfs_blksize(fs, ip, lastblock);
    158 		if (osize < lfs_sb_getbsize(fs) && osize > 0) {
    159 			if ((error = lfs_fragextend(vp, osize, lfs_sb_getbsize(fs),
    160 						    lastblock,
    161 						    (bpp ? &bp : NULL), cred)))
    162 				return (error);
    163 			ip->i_ffs1_size = ip->i_size =
    164 			    (lastblock + 1) * lfs_sb_getbsize(fs);
    165 			uvm_vnp_setsize(vp, ip->i_size);
    166 			ip->i_flag |= IN_CHANGE | IN_UPDATE;
    167 			if (bpp)
    168 				(void) VOP_BWRITE(bp->b_vp, bp);
    169 		}
    170 	}
    171 
    172 	/*
    173 	 * If the block we are writing is a direct block, it's the last
    174 	 * block in the file, and offset + iosize is less than a full
    175 	 * block, we can write one or more fragments.  There are two cases:
    176 	 * the block is brand new and we should allocate it the correct
    177 	 * size or it already exists and contains some fragments and
    178 	 * may need to extend it.
    179 	 */
    180 	if (lbn < ULFS_NDADDR && lfs_lblkno(fs, ip->i_size) <= lbn) {
    181 		osize = lfs_blksize(fs, ip, lbn);
    182 		nsize = lfs_fragroundup(fs, offset + iosize);
    183 		if (lfs_lblktosize(fs, lbn) >= ip->i_size) {
    184 			/* Brand new block or fragment */
    185 			frags = lfs_numfrags(fs, nsize);
    186 			if (!ISSPACE(fs, frags, cred))
    187 				return ENOSPC;
    188 			if (bpp) {
    189 				*bpp = bp = getblk(vp, lbn, nsize, 0, 0);
    190 				bp->b_blkno = UNWRITTEN;
    191 				if (flags & B_CLRBUF)
    192 					clrbuf(bp);
    193 			}
    194 			ip->i_lfs_effnblks += frags;
    195 			mutex_enter(&lfs_lock);
    196 			lfs_sb_subbfree(fs, frags);
    197 			mutex_exit(&lfs_lock);
    198 			ip->i_ffs1_db[lbn] = UNWRITTEN;
    199 		} else {
    200 			if (nsize <= osize) {
    201 				/* No need to extend */
    202 				if (bpp && (error = bread(vp, lbn, osize,
    203 				    0, &bp)))
    204 					return error;
    205 			} else {
    206 				/* Extend existing block */
    207 				if ((error =
    208 				     lfs_fragextend(vp, osize, nsize, lbn,
    209 						    (bpp ? &bp : NULL), cred)))
    210 					return error;
    211 			}
    212 			if (bpp)
    213 				*bpp = bp;
    214 		}
    215 		return 0;
    216 	}
    217 
    218 	error = ulfs_bmaparray(vp, lbn, &daddr, &indirs[0], &num, NULL, NULL);
    219 	if (error)
    220 		return (error);
    221 
    222 	daddr = (daddr_t)((int32_t)daddr); /* XXX ondisk32 */
    223 	KASSERT(daddr <= LFS_MAX_DADDR);
    224 
    225 	/*
    226 	 * Do byte accounting all at once, so we can gracefully fail *before*
    227 	 * we start assigning blocks.
    228 	 */
    229 	frags = fs->um_seqinc;
    230 	bcount = 0;
    231 	if (daddr == UNASSIGNED) {
    232 		bcount = frags;
    233 	}
    234 	for (i = 1; i < num; ++i) {
    235 		if (!indirs[i].in_exists) {
    236 			bcount += frags;
    237 		}
    238 	}
    239 	if (ISSPACE(fs, bcount, cred)) {
    240 		mutex_enter(&lfs_lock);
    241 		lfs_sb_subbfree(fs, bcount);
    242 		mutex_exit(&lfs_lock);
    243 		ip->i_lfs_effnblks += bcount;
    244 	} else {
    245 		return ENOSPC;
    246 	}
    247 
    248 	if (daddr == UNASSIGNED) {
    249 		if (num > 0 && ip->i_ffs1_ib[indirs[0].in_off] == 0) {
    250 			ip->i_ffs1_ib[indirs[0].in_off] = UNWRITTEN;
    251 		}
    252 
    253 		/*
    254 		 * Create new indirect blocks if necessary
    255 		 */
    256 		if (num > 1) {
    257 			idaddr = ip->i_ffs1_ib[indirs[0].in_off];
    258 			for (i = 1; i < num; ++i) {
    259 				ibp = getblk(vp, indirs[i].in_lbn,
    260 				    lfs_sb_getbsize(fs), 0,0);
    261 				if (!indirs[i].in_exists) {
    262 					clrbuf(ibp);
    263 					ibp->b_blkno = UNWRITTEN;
    264 				} else if (!(ibp->b_oflags & (BO_DELWRI | BO_DONE))) {
    265 					ibp->b_blkno = LFS_FSBTODB(fs, idaddr);
    266 					ibp->b_flags |= B_READ;
    267 					VOP_STRATEGY(vp, ibp);
    268 					biowait(ibp);
    269 				}
    270 				/*
    271 				 * This block exists, but the next one may not.
    272 				 * If that is the case mark it UNWRITTEN to keep
    273 				 * the accounting straight.
    274 				 */
    275 				/* XXX ondisk32 */
    276 				if (((int32_t *)ibp->b_data)[indirs[i].in_off] == 0)
    277 					((int32_t *)ibp->b_data)[indirs[i].in_off] =
    278 						UNWRITTEN;
    279 				/* XXX ondisk32 */
    280 				idaddr = ((int32_t *)ibp->b_data)[indirs[i].in_off];
    281 #ifdef DEBUG
    282 				if (vp == fs->lfs_ivnode) {
    283 					LFS_ENTER_LOG("balloc", __FILE__,
    284 						__LINE__, indirs[i].in_lbn,
    285 						ibp->b_flags, curproc->p_pid);
    286 				}
    287 #endif
    288 				if ((error = VOP_BWRITE(ibp->b_vp, ibp)))
    289 					return error;
    290 			}
    291 		}
    292 	}
    293 
    294 
    295 	/*
    296 	 * Get the existing block from the cache, if requested.
    297 	 */
    298 	if (bpp)
    299 		*bpp = bp = getblk(vp, lbn, lfs_blksize(fs, ip, lbn), 0, 0);
    300 
    301 	/*
    302 	 * Do accounting on blocks that represent pages.
    303 	 */
    304 	if (!bpp)
    305 		lfs_register_block(vp, lbn);
    306 
    307 	/*
    308 	 * The block we are writing may be a brand new block
    309 	 * in which case we need to do accounting.
    310 	 *
    311 	 * We can tell a truly new block because ulfs_bmaparray will say
    312 	 * it is UNASSIGNED.  Once we allocate it we will assign it the
    313 	 * disk address UNWRITTEN.
    314 	 */
    315 	if (daddr == UNASSIGNED) {
    316 		if (bpp) {
    317 			if (flags & B_CLRBUF)
    318 				clrbuf(bp);
    319 
    320 			/* Note the new address */
    321 			bp->b_blkno = UNWRITTEN;
    322 		}
    323 
    324 		switch (num) {
    325 		    case 0:
    326 			ip->i_ffs1_db[lbn] = UNWRITTEN;
    327 			break;
    328 		    case 1:
    329 			ip->i_ffs1_ib[indirs[0].in_off] = UNWRITTEN;
    330 			break;
    331 		    default:
    332 			idp = &indirs[num - 1];
    333 			if (bread(vp, idp->in_lbn, lfs_sb_getbsize(fs),
    334 				  B_MODIFY, &ibp))
    335 				panic("lfs_balloc: bread bno %lld",
    336 				    (long long)idp->in_lbn);
    337 			/* XXX ondisk32 */
    338 			((int32_t *)ibp->b_data)[idp->in_off] = UNWRITTEN;
    339 #ifdef DEBUG
    340 			if (vp == fs->lfs_ivnode) {
    341 				LFS_ENTER_LOG("balloc", __FILE__,
    342 					__LINE__, idp->in_lbn,
    343 					ibp->b_flags, curproc->p_pid);
    344 			}
    345 #endif
    346 			VOP_BWRITE(ibp->b_vp, ibp);
    347 		}
    348 	} else if (bpp && !(bp->b_oflags & (BO_DONE|BO_DELWRI))) {
    349 		/*
    350 		 * Not a brand new block, also not in the cache;
    351 		 * read it in from disk.
    352 		 */
    353 		if (iosize == lfs_sb_getbsize(fs))
    354 			/* Optimization: I/O is unnecessary. */
    355 			bp->b_blkno = daddr;
    356 		else {
    357 			/*
    358 			 * We need to read the block to preserve the
    359 			 * existing bytes.
    360 			 */
    361 			bp->b_blkno = daddr;
    362 			bp->b_flags |= B_READ;
    363 			VOP_STRATEGY(vp, bp);
    364 			return (biowait(bp));
    365 		}
    366 	}
    367 
    368 	return (0);
    369 }
    370 
    371 /* VOP_BWRITE 1 time */
    372 int
    373 lfs_fragextend(struct vnode *vp, int osize, int nsize, daddr_t lbn, struct buf **bpp,
    374     kauth_cred_t cred)
    375 {
    376 	struct inode *ip;
    377 	struct lfs *fs;
    378 	long frags;
    379 	int error;
    380 	extern long locked_queue_bytes;
    381 	size_t obufsize;
    382 
    383 	ip = VTOI(vp);
    384 	fs = ip->i_lfs;
    385 	frags = (long)lfs_numfrags(fs, nsize - osize);
    386 	error = 0;
    387 
    388 	ASSERT_NO_SEGLOCK(fs);
    389 
    390 	/*
    391 	 * Get the seglock so we don't enlarge blocks while a segment
    392 	 * is being written.  If we're called with bpp==NULL, though,
    393 	 * we are only pretending to change a buffer, so we don't have to
    394 	 * lock.
    395 	 */
    396     top:
    397 	if (bpp) {
    398 		rw_enter(&fs->lfs_fraglock, RW_READER);
    399 		LFS_DEBUG_COUNTLOCKED("frag");
    400 	}
    401 
    402 	if (!ISSPACE(fs, frags, cred)) {
    403 		error = ENOSPC;
    404 		goto out;
    405 	}
    406 
    407 	/*
    408 	 * If we are not asked to actually return the block, all we need
    409 	 * to do is allocate space for it.  UBC will handle dirtying the
    410 	 * appropriate things and making sure it all goes to disk.
    411 	 * Don't bother to read in that case.
    412 	 */
    413 	if (bpp && (error = bread(vp, lbn, osize, 0, bpp))) {
    414 		goto out;
    415 	}
    416 #if defined(LFS_QUOTA) || defined(LFS_QUOTA2)
    417 	if ((error = lfs_chkdq(ip, frags, cred, 0))) {
    418 		if (bpp)
    419 			brelse(*bpp, 0);
    420 		goto out;
    421 	}
    422 #endif
    423 	/*
    424 	 * Adjust accounting for lfs_avail.  If there's not enough room,
    425 	 * we will have to wait for the cleaner, which we can't do while
    426 	 * holding a block busy or while holding the seglock.  In that case,
    427 	 * release both and start over after waiting.
    428 	 */
    429 
    430 	if (bpp && ((*bpp)->b_oflags & BO_DELWRI)) {
    431 		if (!lfs_fits(fs, frags)) {
    432 			if (bpp)
    433 				brelse(*bpp, 0);
    434 #if defined(LFS_QUOTA) || defined(LFS_QUOTA2)
    435 			lfs_chkdq(ip, -frags, cred, 0);
    436 #endif
    437 			rw_exit(&fs->lfs_fraglock);
    438 			lfs_availwait(fs, frags);
    439 			goto top;
    440 		}
    441 		lfs_sb_subavail(fs, frags);
    442 	}
    443 
    444 	mutex_enter(&lfs_lock);
    445 	lfs_sb_subbfree(fs, frags);
    446 	mutex_exit(&lfs_lock);
    447 	ip->i_lfs_effnblks += frags;
    448 	ip->i_flag |= IN_CHANGE | IN_UPDATE;
    449 
    450 	if (bpp) {
    451 		obufsize = (*bpp)->b_bufsize;
    452 		allocbuf(*bpp, nsize, 1);
    453 
    454 		/* Adjust locked-list accounting */
    455 		if (((*bpp)->b_flags & B_LOCKED) != 0 &&
    456 		    (*bpp)->b_iodone == NULL) {
    457 			mutex_enter(&lfs_lock);
    458 			locked_queue_bytes += (*bpp)->b_bufsize - obufsize;
    459 			mutex_exit(&lfs_lock);
    460 		}
    461 
    462 		memset((char *)((*bpp)->b_data) + osize, 0, (u_int)(nsize - osize));
    463 	}
    464 
    465     out:
    466 	if (bpp) {
    467 		rw_exit(&fs->lfs_fraglock);
    468 	}
    469 	return (error);
    470 }
    471 
    472 static inline int
    473 lge(struct lbnentry *a, struct lbnentry *b)
    474 {
    475 	return a->lbn - b->lbn;
    476 }
    477 
    478 SPLAY_PROTOTYPE(lfs_splay, lbnentry, entry, lge);
    479 
    480 SPLAY_GENERATE(lfs_splay, lbnentry, entry, lge);
    481 
    482 /*
    483  * Record this lbn as being "write pending".  We used to have this information
    484  * on the buffer headers, but since pages don't have buffer headers we
    485  * record it here instead.
    486  */
    487 void
    488 lfs_register_block(struct vnode *vp, daddr_t lbn)
    489 {
    490 	struct lfs *fs;
    491 	struct inode *ip;
    492 	struct lbnentry *lbp;
    493 
    494 	ip = VTOI(vp);
    495 
    496 	/* Don't count metadata */
    497 	if (lbn < 0 || vp->v_type != VREG || ip->i_number == LFS_IFILE_INUM)
    498 		return;
    499 
    500 	fs = ip->i_lfs;
    501 
    502 	ASSERT_NO_SEGLOCK(fs);
    503 
    504 	/* If no space, wait for the cleaner */
    505 	lfs_availwait(fs, lfs_btofsb(fs, 1 << lfs_sb_getbshift(fs)));
    506 
    507 	lbp = (struct lbnentry *)pool_get(&lfs_lbnentry_pool, PR_WAITOK);
    508 	lbp->lbn = lbn;
    509 	mutex_enter(&lfs_lock);
    510 	if (SPLAY_INSERT(lfs_splay, &ip->i_lfs_lbtree, lbp) != NULL) {
    511 		mutex_exit(&lfs_lock);
    512 		/* Already there */
    513 		pool_put(&lfs_lbnentry_pool, lbp);
    514 		return;
    515 	}
    516 
    517 	++ip->i_lfs_nbtree;
    518 	fs->lfs_favail += lfs_btofsb(fs, (1 << lfs_sb_getbshift(fs)));
    519 	fs->lfs_pages += lfs_sb_getbsize(fs) >> PAGE_SHIFT;
    520 	++locked_fakequeue_count;
    521 	lfs_subsys_pages += lfs_sb_getbsize(fs) >> PAGE_SHIFT;
    522 	mutex_exit(&lfs_lock);
    523 }
    524 
    525 static void
    526 lfs_do_deregister(struct lfs *fs, struct inode *ip, struct lbnentry *lbp)
    527 {
    528 	ASSERT_MAYBE_SEGLOCK(fs);
    529 
    530 	mutex_enter(&lfs_lock);
    531 	--ip->i_lfs_nbtree;
    532 	SPLAY_REMOVE(lfs_splay, &ip->i_lfs_lbtree, lbp);
    533 	if (fs->lfs_favail > lfs_btofsb(fs, (1 << lfs_sb_getbshift(fs))))
    534 		fs->lfs_favail -= lfs_btofsb(fs, (1 << lfs_sb_getbshift(fs)));
    535 	fs->lfs_pages -= lfs_sb_getbsize(fs) >> PAGE_SHIFT;
    536 	if (locked_fakequeue_count > 0)
    537 		--locked_fakequeue_count;
    538 	lfs_subsys_pages -= lfs_sb_getbsize(fs) >> PAGE_SHIFT;
    539 	mutex_exit(&lfs_lock);
    540 
    541 	pool_put(&lfs_lbnentry_pool, lbp);
    542 }
    543 
    544 void
    545 lfs_deregister_block(struct vnode *vp, daddr_t lbn)
    546 {
    547 	struct lfs *fs;
    548 	struct inode *ip;
    549 	struct lbnentry *lbp;
    550 	struct lbnentry tmp;
    551 
    552 	ip = VTOI(vp);
    553 
    554 	/* Don't count metadata */
    555 	if (lbn < 0 || vp->v_type != VREG || ip->i_number == LFS_IFILE_INUM)
    556 		return;
    557 
    558 	fs = ip->i_lfs;
    559 	tmp.lbn = lbn;
    560 	lbp = SPLAY_FIND(lfs_splay, &ip->i_lfs_lbtree, &tmp);
    561 	if (lbp == NULL)
    562 		return;
    563 
    564 	lfs_do_deregister(fs, ip, lbp);
    565 }
    566 
    567 void
    568 lfs_deregister_all(struct vnode *vp)
    569 {
    570 	struct lbnentry *lbp, *nlbp;
    571 	struct lfs_splay *hd;
    572 	struct lfs *fs;
    573 	struct inode *ip;
    574 
    575 	ip = VTOI(vp);
    576 	fs = ip->i_lfs;
    577 	hd = &ip->i_lfs_lbtree;
    578 
    579 	for (lbp = SPLAY_MIN(lfs_splay, hd); lbp != NULL; lbp = nlbp) {
    580 		nlbp = SPLAY_NEXT(lfs_splay, hd, lbp);
    581 		lfs_do_deregister(fs, ip, lbp);
    582 	}
    583 }
    584