Home | History | Annotate | Line # | Download | only in kern
subr_blist.c revision 1.2
      1  1.2  yamt /*	$NetBSD: subr_blist.c,v 1.2 2005/04/06 11:33:54 yamt Exp $	*/
      2  1.2  yamt 
      3  1.1  yamt /*-
      4  1.1  yamt  * Copyright (c) 1998 Matthew Dillon.  All Rights Reserved.
      5  1.1  yamt  * Redistribution and use in source and binary forms, with or without
      6  1.1  yamt  * modification, are permitted provided that the following conditions
      7  1.1  yamt  * are met:
      8  1.1  yamt  * 1. Redistributions of source code must retain the above copyright
      9  1.1  yamt  *    notice, this list of conditions and the following disclaimer.
     10  1.1  yamt  * 2. Redistributions in binary form must reproduce the above copyright
     11  1.1  yamt  *    notice, this list of conditions and the following disclaimer in the
     12  1.1  yamt  *    documentation and/or other materials provided with the distribution.
     13  1.1  yamt  * 4. Neither the name of the University nor the names of its contributors
     14  1.1  yamt  *    may be used to endorse or promote products derived from this software
     15  1.1  yamt  *    without specific prior written permission.
     16  1.1  yamt  *
     17  1.1  yamt  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS
     18  1.1  yamt  * OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
     19  1.1  yamt  * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
     20  1.1  yamt  * ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY
     21  1.1  yamt  * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
     22  1.1  yamt  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE
     23  1.1  yamt  * GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
     24  1.1  yamt  * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY,
     25  1.1  yamt  * WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING
     26  1.1  yamt  * NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
     27  1.1  yamt  * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
     28  1.1  yamt  */
     29  1.1  yamt /*
     30  1.1  yamt  * BLIST.C -	Bitmap allocator/deallocator, using a radix tree with hinting
     31  1.1  yamt  *
     32  1.1  yamt  *	This module implements a general bitmap allocator/deallocator.  The
     33  1.1  yamt  *	allocator eats around 2 bits per 'block'.  The module does not
     34  1.1  yamt  *	try to interpret the meaning of a 'block' other then to return
     35  1.1  yamt  *	SWAPBLK_NONE on an allocation failure.
     36  1.1  yamt  *
     37  1.1  yamt  *	A radix tree is used to maintain the bitmap.  Two radix constants are
     38  1.1  yamt  *	involved:  One for the bitmaps contained in the leaf nodes (typically
     39  1.1  yamt  *	32), and one for the meta nodes (typically 16).  Both meta and leaf
     40  1.1  yamt  *	nodes have a hint field.  This field gives us a hint as to the largest
     41  1.1  yamt  *	free contiguous range of blocks under the node.  It may contain a
     42  1.1  yamt  *	value that is too high, but will never contain a value that is too
     43  1.1  yamt  *	low.  When the radix tree is searched, allocation failures in subtrees
     44  1.1  yamt  *	update the hint.
     45  1.1  yamt  *
     46  1.1  yamt  *	The radix tree also implements two collapsed states for meta nodes:
     47  1.1  yamt  *	the ALL-ALLOCATED state and the ALL-FREE state.  If a meta node is
     48  1.1  yamt  *	in either of these two states, all information contained underneath
     49  1.1  yamt  *	the node is considered stale.  These states are used to optimize
     50  1.1  yamt  *	allocation and freeing operations.
     51  1.1  yamt  *
     52  1.1  yamt  * 	The hinting greatly increases code efficiency for allocations while
     53  1.1  yamt  *	the general radix structure optimizes both allocations and frees.  The
     54  1.1  yamt  *	radix tree should be able to operate well no matter how much
     55  1.1  yamt  *	fragmentation there is and no matter how large a bitmap is used.
     56  1.1  yamt  *
     57  1.1  yamt  *	Unlike the rlist code, the blist code wires all necessary memory at
     58  1.1  yamt  *	creation time.  Neither allocations nor frees require interaction with
     59  1.1  yamt  *	the memory subsystem.  In contrast, the rlist code may allocate memory
     60  1.1  yamt  *	on an rlist_free() call.  The non-blocking features of the blist code
     61  1.1  yamt  *	are used to great advantage in the swap code (vm/nswap_pager.c).  The
     62  1.1  yamt  *	rlist code uses a little less overall memory then the blist code (but
     63  1.1  yamt  *	due to swap interleaving not all that much less), but the blist code
     64  1.1  yamt  *	scales much, much better.
     65  1.1  yamt  *
     66  1.1  yamt  *	LAYOUT: The radix tree is layed out recursively using a
     67  1.1  yamt  *	linear array.  Each meta node is immediately followed (layed out
     68  1.1  yamt  *	sequentially in memory) by BLIST_META_RADIX lower level nodes.  This
     69  1.1  yamt  *	is a recursive structure but one that can be easily scanned through
     70  1.1  yamt  *	a very simple 'skip' calculation.  In order to support large radixes,
     71  1.1  yamt  *	portions of the tree may reside outside our memory allocation.  We
     72  1.1  yamt  *	handle this with an early-termination optimization (when bighint is
     73  1.1  yamt  *	set to -1) on the scan.  The memory allocation is only large enough
     74  1.1  yamt  *	to cover the number of blocks requested at creation time even if it
     75  1.1  yamt  *	must be encompassed in larger root-node radix.
     76  1.1  yamt  *
     77  1.1  yamt  *	NOTE: the allocator cannot currently allocate more then
     78  1.1  yamt  *	BLIST_BMAP_RADIX blocks per call.  It will panic with 'allocation too
     79  1.1  yamt  *	large' if you try.  This is an area that could use improvement.  The
     80  1.1  yamt  *	radix is large enough that this restriction does not effect the swap
     81  1.1  yamt  *	system, though.  Currently only the allocation code is effected by
     82  1.1  yamt  *	this algorithmic unfeature.  The freeing code can handle arbitrary
     83  1.1  yamt  *	ranges.
     84  1.1  yamt  *
     85  1.1  yamt  *	This code can be compiled stand-alone for debugging.
     86  1.1  yamt  */
     87  1.1  yamt 
     88  1.1  yamt #include <sys/cdefs.h>
     89  1.2  yamt __KERNEL_RCSID(0, "$NetBSD: subr_blist.c,v 1.2 2005/04/06 11:33:54 yamt Exp $");
     90  1.2  yamt #if 0
     91  1.1  yamt __FBSDID("$FreeBSD: src/sys/kern/subr_blist.c,v 1.17 2004/06/04 04:03:25 alc Exp $");
     92  1.2  yamt #endif
     93  1.1  yamt 
     94  1.1  yamt #ifdef _KERNEL
     95  1.1  yamt 
     96  1.1  yamt #include <sys/param.h>
     97  1.1  yamt #include <sys/systm.h>
     98  1.1  yamt #include <sys/lock.h>
     99  1.1  yamt #include <sys/kernel.h>
    100  1.1  yamt #include <sys/blist.h>
    101  1.1  yamt #include <sys/malloc.h>
    102  1.1  yamt #include <sys/proc.h>
    103  1.1  yamt 
    104  1.1  yamt #else
    105  1.1  yamt 
    106  1.1  yamt #ifndef BLIST_NO_DEBUG
    107  1.1  yamt #define BLIST_DEBUG
    108  1.1  yamt #endif
    109  1.1  yamt 
    110  1.1  yamt #define SWAPBLK_NONE ((daddr_t)-1)
    111  1.1  yamt 
    112  1.1  yamt #include <sys/types.h>
    113  1.1  yamt #include <stdio.h>
    114  1.1  yamt #include <string.h>
    115  1.1  yamt #include <stdlib.h>
    116  1.1  yamt #include <stdarg.h>
    117  1.1  yamt 
    118  1.1  yamt #define malloc(a,b,c)	calloc(a, 1)
    119  1.1  yamt #define free(a,b)	free(a)
    120  1.1  yamt 
    121  1.1  yamt typedef unsigned int u_daddr_t;
    122  1.1  yamt 
    123  1.1  yamt #include <sys/blist.h>
    124  1.1  yamt 
    125  1.1  yamt void panic(const char *ctl, ...);
    126  1.1  yamt 
    127  1.1  yamt #endif
    128  1.1  yamt 
    129  1.1  yamt /*
    130  1.1  yamt  * static support functions
    131  1.1  yamt  */
    132  1.1  yamt 
    133  1.1  yamt static daddr_t blst_leaf_alloc(blmeta_t *scan, daddr_t blk, int count);
    134  1.1  yamt static daddr_t blst_meta_alloc(blmeta_t *scan, daddr_t blk,
    135  1.1  yamt 				daddr_t count, daddr_t radix, int skip);
    136  1.1  yamt static void blst_leaf_free(blmeta_t *scan, daddr_t relblk, int count);
    137  1.1  yamt static void blst_meta_free(blmeta_t *scan, daddr_t freeBlk, daddr_t count,
    138  1.1  yamt 					daddr_t radix, int skip, daddr_t blk);
    139  1.1  yamt static void blst_copy(blmeta_t *scan, daddr_t blk, daddr_t radix,
    140  1.1  yamt 				daddr_t skip, blist_t dest, daddr_t count);
    141  1.1  yamt static int blst_leaf_fill(blmeta_t *scan, daddr_t blk, int count);
    142  1.1  yamt static int blst_meta_fill(blmeta_t *scan, daddr_t allocBlk, daddr_t count,
    143  1.1  yamt 				daddr_t radix, int skip, daddr_t blk);
    144  1.1  yamt static daddr_t	blst_radix_init(blmeta_t *scan, daddr_t radix,
    145  1.1  yamt 						int skip, daddr_t count);
    146  1.1  yamt #ifndef _KERNEL
    147  1.1  yamt static void	blst_radix_print(blmeta_t *scan, daddr_t blk,
    148  1.1  yamt 					daddr_t radix, int skip, int tab);
    149  1.1  yamt #endif
    150  1.1  yamt 
    151  1.1  yamt #ifdef _KERNEL
    152  1.1  yamt static MALLOC_DEFINE(M_SWAP, "SWAP", "Swap space");
    153  1.1  yamt #endif
    154  1.1  yamt 
    155  1.1  yamt /*
    156  1.1  yamt  * blist_create() - create a blist capable of handling up to the specified
    157  1.1  yamt  *		    number of blocks
    158  1.1  yamt  *
    159  1.1  yamt  *	blocks must be greater then 0
    160  1.1  yamt  *
    161  1.1  yamt  *	The smallest blist consists of a single leaf node capable of
    162  1.1  yamt  *	managing BLIST_BMAP_RADIX blocks.
    163  1.1  yamt  */
    164  1.1  yamt 
    165  1.1  yamt blist_t
    166  1.1  yamt blist_create(daddr_t blocks)
    167  1.1  yamt {
    168  1.1  yamt 	blist_t bl;
    169  1.1  yamt 	int radix;
    170  1.1  yamt 	int skip = 0;
    171  1.1  yamt 
    172  1.1  yamt 	/*
    173  1.1  yamt 	 * Calculate radix and skip field used for scanning.
    174  1.1  yamt 	 */
    175  1.1  yamt 	radix = BLIST_BMAP_RADIX;
    176  1.1  yamt 
    177  1.1  yamt 	while (radix < blocks) {
    178  1.1  yamt 		radix *= BLIST_META_RADIX;
    179  1.1  yamt 		skip = (skip + 1) * BLIST_META_RADIX;
    180  1.1  yamt 	}
    181  1.1  yamt 
    182  1.1  yamt 	bl = malloc(sizeof(struct blist), M_SWAP, M_WAITOK | M_ZERO);
    183  1.1  yamt 
    184  1.1  yamt 	bl->bl_blocks = blocks;
    185  1.1  yamt 	bl->bl_radix = radix;
    186  1.1  yamt 	bl->bl_skip = skip;
    187  1.1  yamt 	bl->bl_rootblks = 1 +
    188  1.1  yamt 	    blst_radix_init(NULL, bl->bl_radix, bl->bl_skip, blocks);
    189  1.1  yamt 	bl->bl_root = malloc(sizeof(blmeta_t) * bl->bl_rootblks, M_SWAP, M_WAITOK);
    190  1.1  yamt 
    191  1.1  yamt #if defined(BLIST_DEBUG)
    192  1.1  yamt 	printf(
    193  1.1  yamt 		"BLIST representing %lld blocks (%lld MB of swap)"
    194  1.1  yamt 		", requiring %lldK of ram\n",
    195  1.1  yamt 		(long long)bl->bl_blocks,
    196  1.1  yamt 		(long long)bl->bl_blocks * 4 / 1024,
    197  1.1  yamt 		(long long)(bl->bl_rootblks * sizeof(blmeta_t) + 1023) / 1024
    198  1.1  yamt 	);
    199  1.1  yamt 	printf("BLIST raw radix tree contains %lld records\n",
    200  1.1  yamt 	    (long long)bl->bl_rootblks);
    201  1.1  yamt #endif
    202  1.1  yamt 	blst_radix_init(bl->bl_root, bl->bl_radix, bl->bl_skip, blocks);
    203  1.1  yamt 
    204  1.1  yamt 	return(bl);
    205  1.1  yamt }
    206  1.1  yamt 
    207  1.1  yamt void
    208  1.1  yamt blist_destroy(blist_t bl)
    209  1.1  yamt {
    210  1.1  yamt 	free(bl->bl_root, M_SWAP);
    211  1.1  yamt 	free(bl, M_SWAP);
    212  1.1  yamt }
    213  1.1  yamt 
    214  1.1  yamt /*
    215  1.1  yamt  * blist_alloc() - reserve space in the block bitmap.  Return the base
    216  1.1  yamt  *		     of a contiguous region or SWAPBLK_NONE if space could
    217  1.1  yamt  *		     not be allocated.
    218  1.1  yamt  */
    219  1.1  yamt 
    220  1.1  yamt daddr_t
    221  1.1  yamt blist_alloc(blist_t bl, daddr_t count)
    222  1.1  yamt {
    223  1.1  yamt 	daddr_t blk = SWAPBLK_NONE;
    224  1.1  yamt 
    225  1.1  yamt 	if (bl) {
    226  1.1  yamt 		if (bl->bl_radix == BLIST_BMAP_RADIX)
    227  1.1  yamt 			blk = blst_leaf_alloc(bl->bl_root, 0, count);
    228  1.1  yamt 		else
    229  1.1  yamt 			blk = blst_meta_alloc(bl->bl_root, 0, count, bl->bl_radix, bl->bl_skip);
    230  1.1  yamt 		if (blk != SWAPBLK_NONE)
    231  1.1  yamt 			bl->bl_free -= count;
    232  1.1  yamt 	}
    233  1.1  yamt 	return(blk);
    234  1.1  yamt }
    235  1.1  yamt 
    236  1.1  yamt /*
    237  1.1  yamt  * blist_free() -	free up space in the block bitmap.  Return the base
    238  1.1  yamt  *		     	of a contiguous region.  Panic if an inconsistancy is
    239  1.1  yamt  *			found.
    240  1.1  yamt  */
    241  1.1  yamt 
    242  1.1  yamt void
    243  1.1  yamt blist_free(blist_t bl, daddr_t blkno, daddr_t count)
    244  1.1  yamt {
    245  1.1  yamt 	if (bl) {
    246  1.1  yamt 		if (bl->bl_radix == BLIST_BMAP_RADIX)
    247  1.1  yamt 			blst_leaf_free(bl->bl_root, blkno, count);
    248  1.1  yamt 		else
    249  1.1  yamt 			blst_meta_free(bl->bl_root, blkno, count, bl->bl_radix, bl->bl_skip, 0);
    250  1.1  yamt 		bl->bl_free += count;
    251  1.1  yamt 	}
    252  1.1  yamt }
    253  1.1  yamt 
    254  1.1  yamt /*
    255  1.1  yamt  * blist_fill() -	mark a region in the block bitmap as off-limits
    256  1.1  yamt  *			to the allocator (i.e. allocate it), ignoring any
    257  1.1  yamt  *			existing allocations.  Return the number of blocks
    258  1.1  yamt  *			actually filled that were free before the call.
    259  1.1  yamt  */
    260  1.1  yamt 
    261  1.1  yamt int
    262  1.1  yamt blist_fill(blist_t bl, daddr_t blkno, daddr_t count)
    263  1.1  yamt {
    264  1.1  yamt 	int filled;
    265  1.1  yamt 
    266  1.1  yamt 	if (bl) {
    267  1.1  yamt 		if (bl->bl_radix == BLIST_BMAP_RADIX)
    268  1.1  yamt 			filled = blst_leaf_fill(bl->bl_root, blkno, count);
    269  1.1  yamt 		else
    270  1.1  yamt 			filled = blst_meta_fill(bl->bl_root, blkno, count,
    271  1.1  yamt 			    bl->bl_radix, bl->bl_skip, 0);
    272  1.1  yamt 		bl->bl_free -= filled;
    273  1.1  yamt 		return filled;
    274  1.1  yamt 	} else
    275  1.1  yamt 		return 0;
    276  1.1  yamt }
    277  1.1  yamt 
    278  1.1  yamt /*
    279  1.1  yamt  * blist_resize() -	resize an existing radix tree to handle the
    280  1.1  yamt  *			specified number of blocks.  This will reallocate
    281  1.1  yamt  *			the tree and transfer the previous bitmap to the new
    282  1.1  yamt  *			one.  When extending the tree you can specify whether
    283  1.1  yamt  *			the new blocks are to left allocated or freed.
    284  1.1  yamt  */
    285  1.1  yamt 
    286  1.1  yamt void
    287  1.1  yamt blist_resize(blist_t *pbl, daddr_t count, int freenew)
    288  1.1  yamt {
    289  1.1  yamt     blist_t newbl = blist_create(count);
    290  1.1  yamt     blist_t save = *pbl;
    291  1.1  yamt 
    292  1.1  yamt     *pbl = newbl;
    293  1.1  yamt     if (count > save->bl_blocks)
    294  1.1  yamt 	    count = save->bl_blocks;
    295  1.1  yamt     blst_copy(save->bl_root, 0, save->bl_radix, save->bl_skip, newbl, count);
    296  1.1  yamt 
    297  1.1  yamt     /*
    298  1.1  yamt      * If resizing upwards, should we free the new space or not?
    299  1.1  yamt      */
    300  1.1  yamt     if (freenew && count < newbl->bl_blocks) {
    301  1.1  yamt 	    blist_free(newbl, count, newbl->bl_blocks - count);
    302  1.1  yamt     }
    303  1.1  yamt     blist_destroy(save);
    304  1.1  yamt }
    305  1.1  yamt 
    306  1.1  yamt #ifdef BLIST_DEBUG
    307  1.1  yamt 
    308  1.1  yamt /*
    309  1.1  yamt  * blist_print()    - dump radix tree
    310  1.1  yamt  */
    311  1.1  yamt 
    312  1.1  yamt void
    313  1.1  yamt blist_print(blist_t bl)
    314  1.1  yamt {
    315  1.1  yamt 	printf("BLIST {\n");
    316  1.1  yamt 	blst_radix_print(bl->bl_root, 0, bl->bl_radix, bl->bl_skip, 4);
    317  1.1  yamt 	printf("}\n");
    318  1.1  yamt }
    319  1.1  yamt 
    320  1.1  yamt #endif
    321  1.1  yamt 
    322  1.1  yamt /************************************************************************
    323  1.1  yamt  *			  ALLOCATION SUPPORT FUNCTIONS			*
    324  1.1  yamt  ************************************************************************
    325  1.1  yamt  *
    326  1.1  yamt  *	These support functions do all the actual work.  They may seem
    327  1.1  yamt  *	rather longish, but that's because I've commented them up.  The
    328  1.1  yamt  *	actual code is straight forward.
    329  1.1  yamt  *
    330  1.1  yamt  */
    331  1.1  yamt 
    332  1.1  yamt /*
    333  1.1  yamt  * blist_leaf_alloc() -	allocate at a leaf in the radix tree (a bitmap).
    334  1.1  yamt  *
    335  1.1  yamt  *	This is the core of the allocator and is optimized for the 1 block
    336  1.1  yamt  *	and the BLIST_BMAP_RADIX block allocation cases.  Other cases are
    337  1.1  yamt  *	somewhat slower.  The 1 block allocation case is log2 and extremely
    338  1.1  yamt  *	quick.
    339  1.1  yamt  */
    340  1.1  yamt 
    341  1.1  yamt static daddr_t
    342  1.1  yamt blst_leaf_alloc(
    343  1.1  yamt 	blmeta_t *scan,
    344  1.1  yamt 	daddr_t blk,
    345  1.1  yamt 	int count
    346  1.1  yamt ) {
    347  1.1  yamt 	u_daddr_t orig = scan->u.bmu_bitmap;
    348  1.1  yamt 
    349  1.1  yamt 	if (orig == 0) {
    350  1.1  yamt 		/*
    351  1.1  yamt 		 * Optimize bitmap all-allocated case.  Also, count = 1
    352  1.1  yamt 		 * case assumes at least 1 bit is free in the bitmap, so
    353  1.1  yamt 		 * we have to take care of this case here.
    354  1.1  yamt 		 */
    355  1.1  yamt 		scan->bm_bighint = 0;
    356  1.1  yamt 		return(SWAPBLK_NONE);
    357  1.1  yamt 	}
    358  1.1  yamt 	if (count == 1) {
    359  1.1  yamt 		/*
    360  1.1  yamt 		 * Optimized code to allocate one bit out of the bitmap
    361  1.1  yamt 		 */
    362  1.1  yamt 		u_daddr_t mask;
    363  1.1  yamt 		int j = BLIST_BMAP_RADIX/2;
    364  1.1  yamt 		int r = 0;
    365  1.1  yamt 
    366  1.1  yamt 		mask = (u_daddr_t)-1 >> (BLIST_BMAP_RADIX/2);
    367  1.1  yamt 
    368  1.1  yamt 		while (j) {
    369  1.1  yamt 			if ((orig & mask) == 0) {
    370  1.1  yamt 			    r += j;
    371  1.1  yamt 			    orig >>= j;
    372  1.1  yamt 			}
    373  1.1  yamt 			j >>= 1;
    374  1.1  yamt 			mask >>= j;
    375  1.1  yamt 		}
    376  1.1  yamt 		scan->u.bmu_bitmap &= ~(1 << r);
    377  1.1  yamt 		return(blk + r);
    378  1.1  yamt 	}
    379  1.1  yamt 	if (count <= BLIST_BMAP_RADIX) {
    380  1.1  yamt 		/*
    381  1.1  yamt 		 * non-optimized code to allocate N bits out of the bitmap.
    382  1.1  yamt 		 * The more bits, the faster the code runs.  It will run
    383  1.1  yamt 		 * the slowest allocating 2 bits, but since there aren't any
    384  1.1  yamt 		 * memory ops in the core loop (or shouldn't be, anyway),
    385  1.1  yamt 		 * you probably won't notice the difference.
    386  1.1  yamt 		 */
    387  1.1  yamt 		int j;
    388  1.1  yamt 		int n = BLIST_BMAP_RADIX - count;
    389  1.1  yamt 		u_daddr_t mask;
    390  1.1  yamt 
    391  1.1  yamt 		mask = (u_daddr_t)-1 >> n;
    392  1.1  yamt 
    393  1.1  yamt 		for (j = 0; j <= n; ++j) {
    394  1.1  yamt 			if ((orig & mask) == mask) {
    395  1.1  yamt 				scan->u.bmu_bitmap &= ~mask;
    396  1.1  yamt 				return(blk + j);
    397  1.1  yamt 			}
    398  1.1  yamt 			mask = (mask << 1);
    399  1.1  yamt 		}
    400  1.1  yamt 	}
    401  1.1  yamt 	/*
    402  1.1  yamt 	 * We couldn't allocate count in this subtree, update bighint.
    403  1.1  yamt 	 */
    404  1.1  yamt 	scan->bm_bighint = count - 1;
    405  1.1  yamt 	return(SWAPBLK_NONE);
    406  1.1  yamt }
    407  1.1  yamt 
    408  1.1  yamt /*
    409  1.1  yamt  * blist_meta_alloc() -	allocate at a meta in the radix tree.
    410  1.1  yamt  *
    411  1.1  yamt  *	Attempt to allocate at a meta node.  If we can't, we update
    412  1.1  yamt  *	bighint and return a failure.  Updating bighint optimize future
    413  1.1  yamt  *	calls that hit this node.  We have to check for our collapse cases
    414  1.1  yamt  *	and we have a few optimizations strewn in as well.
    415  1.1  yamt  */
    416  1.1  yamt 
    417  1.1  yamt static daddr_t
    418  1.1  yamt blst_meta_alloc(
    419  1.1  yamt 	blmeta_t *scan,
    420  1.1  yamt 	daddr_t blk,
    421  1.1  yamt 	daddr_t count,
    422  1.1  yamt 	daddr_t radix,
    423  1.1  yamt 	int skip
    424  1.1  yamt ) {
    425  1.1  yamt 	int i;
    426  1.1  yamt 	int next_skip = ((u_int)skip / BLIST_META_RADIX);
    427  1.1  yamt 
    428  1.1  yamt 	if (scan->u.bmu_avail == 0)  {
    429  1.1  yamt 		/*
    430  1.1  yamt 		 * ALL-ALLOCATED special case
    431  1.1  yamt 		 */
    432  1.1  yamt 		scan->bm_bighint = count;
    433  1.1  yamt 		return(SWAPBLK_NONE);
    434  1.1  yamt 	}
    435  1.1  yamt 
    436  1.1  yamt 	if (scan->u.bmu_avail == radix) {
    437  1.1  yamt 		radix /= BLIST_META_RADIX;
    438  1.1  yamt 
    439  1.1  yamt 		/*
    440  1.1  yamt 		 * ALL-FREE special case, initialize uninitialize
    441  1.1  yamt 		 * sublevel.
    442  1.1  yamt 		 */
    443  1.1  yamt 		for (i = 1; i <= skip; i += next_skip) {
    444  1.1  yamt 			if (scan[i].bm_bighint == (daddr_t)-1)
    445  1.1  yamt 				break;
    446  1.1  yamt 			if (next_skip == 1) {
    447  1.1  yamt 				scan[i].u.bmu_bitmap = (u_daddr_t)-1;
    448  1.1  yamt 				scan[i].bm_bighint = BLIST_BMAP_RADIX;
    449  1.1  yamt 			} else {
    450  1.1  yamt 				scan[i].bm_bighint = radix;
    451  1.1  yamt 				scan[i].u.bmu_avail = radix;
    452  1.1  yamt 			}
    453  1.1  yamt 		}
    454  1.1  yamt 	} else {
    455  1.1  yamt 		radix /= BLIST_META_RADIX;
    456  1.1  yamt 	}
    457  1.1  yamt 
    458  1.1  yamt 	for (i = 1; i <= skip; i += next_skip) {
    459  1.1  yamt 		if (count <= scan[i].bm_bighint) {
    460  1.1  yamt 			/*
    461  1.1  yamt 			 * count fits in object
    462  1.1  yamt 			 */
    463  1.1  yamt 			daddr_t r;
    464  1.1  yamt 			if (next_skip == 1) {
    465  1.1  yamt 				r = blst_leaf_alloc(&scan[i], blk, count);
    466  1.1  yamt 			} else {
    467  1.1  yamt 				r = blst_meta_alloc(&scan[i], blk, count, radix, next_skip - 1);
    468  1.1  yamt 			}
    469  1.1  yamt 			if (r != SWAPBLK_NONE) {
    470  1.1  yamt 				scan->u.bmu_avail -= count;
    471  1.1  yamt 				if (scan->bm_bighint > scan->u.bmu_avail)
    472  1.1  yamt 					scan->bm_bighint = scan->u.bmu_avail;
    473  1.1  yamt 				return(r);
    474  1.1  yamt 			}
    475  1.1  yamt 		} else if (scan[i].bm_bighint == (daddr_t)-1) {
    476  1.1  yamt 			/*
    477  1.1  yamt 			 * Terminator
    478  1.1  yamt 			 */
    479  1.1  yamt 			break;
    480  1.1  yamt 		} else if (count > radix) {
    481  1.1  yamt 			/*
    482  1.1  yamt 			 * count does not fit in object even if it were
    483  1.1  yamt 			 * complete free.
    484  1.1  yamt 			 */
    485  1.1  yamt 			panic("blist_meta_alloc: allocation too large");
    486  1.1  yamt 		}
    487  1.1  yamt 		blk += radix;
    488  1.1  yamt 	}
    489  1.1  yamt 
    490  1.1  yamt 	/*
    491  1.1  yamt 	 * We couldn't allocate count in this subtree, update bighint.
    492  1.1  yamt 	 */
    493  1.1  yamt 	if (scan->bm_bighint >= count)
    494  1.1  yamt 		scan->bm_bighint = count - 1;
    495  1.1  yamt 	return(SWAPBLK_NONE);
    496  1.1  yamt }
    497  1.1  yamt 
    498  1.1  yamt /*
    499  1.1  yamt  * BLST_LEAF_FREE() -	free allocated block from leaf bitmap
    500  1.1  yamt  *
    501  1.1  yamt  */
    502  1.1  yamt 
    503  1.1  yamt static void
    504  1.1  yamt blst_leaf_free(
    505  1.1  yamt 	blmeta_t *scan,
    506  1.1  yamt 	daddr_t blk,
    507  1.1  yamt 	int count
    508  1.1  yamt ) {
    509  1.1  yamt 	/*
    510  1.1  yamt 	 * free some data in this bitmap
    511  1.1  yamt 	 *
    512  1.1  yamt 	 * e.g.
    513  1.1  yamt 	 *	0000111111111110000
    514  1.1  yamt 	 *          \_________/\__/
    515  1.1  yamt 	 *		v        n
    516  1.1  yamt 	 */
    517  1.1  yamt 	int n = blk & (BLIST_BMAP_RADIX - 1);
    518  1.1  yamt 	u_daddr_t mask;
    519  1.1  yamt 
    520  1.1  yamt 	mask = ((u_daddr_t)-1 << n) &
    521  1.1  yamt 	    ((u_daddr_t)-1 >> (BLIST_BMAP_RADIX - count - n));
    522  1.1  yamt 
    523  1.1  yamt 	if (scan->u.bmu_bitmap & mask)
    524  1.1  yamt 		panic("blst_radix_free: freeing free block");
    525  1.1  yamt 	scan->u.bmu_bitmap |= mask;
    526  1.1  yamt 
    527  1.1  yamt 	/*
    528  1.1  yamt 	 * We could probably do a better job here.  We are required to make
    529  1.1  yamt 	 * bighint at least as large as the biggest contiguous block of
    530  1.1  yamt 	 * data.  If we just shoehorn it, a little extra overhead will
    531  1.1  yamt 	 * be incured on the next allocation (but only that one typically).
    532  1.1  yamt 	 */
    533  1.1  yamt 	scan->bm_bighint = BLIST_BMAP_RADIX;
    534  1.1  yamt }
    535  1.1  yamt 
    536  1.1  yamt /*
    537  1.1  yamt  * BLST_META_FREE() - free allocated blocks from radix tree meta info
    538  1.1  yamt  *
    539  1.1  yamt  *	This support routine frees a range of blocks from the bitmap.
    540  1.1  yamt  *	The range must be entirely enclosed by this radix node.  If a
    541  1.1  yamt  *	meta node, we break the range down recursively to free blocks
    542  1.1  yamt  *	in subnodes (which means that this code can free an arbitrary
    543  1.1  yamt  *	range whereas the allocation code cannot allocate an arbitrary
    544  1.1  yamt  *	range).
    545  1.1  yamt  */
    546  1.1  yamt 
    547  1.1  yamt static void
    548  1.1  yamt blst_meta_free(
    549  1.1  yamt 	blmeta_t *scan,
    550  1.1  yamt 	daddr_t freeBlk,
    551  1.1  yamt 	daddr_t count,
    552  1.1  yamt 	daddr_t radix,
    553  1.1  yamt 	int skip,
    554  1.1  yamt 	daddr_t blk
    555  1.1  yamt ) {
    556  1.1  yamt 	int i;
    557  1.1  yamt 	int next_skip = ((u_int)skip / BLIST_META_RADIX);
    558  1.1  yamt 
    559  1.1  yamt #if 0
    560  1.1  yamt 	printf("FREE (%llx,%lld) FROM (%llx,%lld)\n",
    561  1.1  yamt 	    (long long)freeBlk, (long long)count,
    562  1.1  yamt 	    (long long)blk, (long long)radix
    563  1.1  yamt 	);
    564  1.1  yamt #endif
    565  1.1  yamt 
    566  1.1  yamt 	if (scan->u.bmu_avail == 0) {
    567  1.1  yamt 		/*
    568  1.1  yamt 		 * ALL-ALLOCATED special case, with possible
    569  1.1  yamt 		 * shortcut to ALL-FREE special case.
    570  1.1  yamt 		 */
    571  1.1  yamt 		scan->u.bmu_avail = count;
    572  1.1  yamt 		scan->bm_bighint = count;
    573  1.1  yamt 
    574  1.1  yamt 		if (count != radix)  {
    575  1.1  yamt 			for (i = 1; i <= skip; i += next_skip) {
    576  1.1  yamt 				if (scan[i].bm_bighint == (daddr_t)-1)
    577  1.1  yamt 					break;
    578  1.1  yamt 				scan[i].bm_bighint = 0;
    579  1.1  yamt 				if (next_skip == 1) {
    580  1.1  yamt 					scan[i].u.bmu_bitmap = 0;
    581  1.1  yamt 				} else {
    582  1.1  yamt 					scan[i].u.bmu_avail = 0;
    583  1.1  yamt 				}
    584  1.1  yamt 			}
    585  1.1  yamt 			/* fall through */
    586  1.1  yamt 		}
    587  1.1  yamt 	} else {
    588  1.1  yamt 		scan->u.bmu_avail += count;
    589  1.1  yamt 		/* scan->bm_bighint = radix; */
    590  1.1  yamt 	}
    591  1.1  yamt 
    592  1.1  yamt 	/*
    593  1.1  yamt 	 * ALL-FREE special case.
    594  1.1  yamt 	 */
    595  1.1  yamt 
    596  1.1  yamt 	if (scan->u.bmu_avail == radix)
    597  1.1  yamt 		return;
    598  1.1  yamt 	if (scan->u.bmu_avail > radix)
    599  1.1  yamt 		panic("blst_meta_free: freeing already free blocks (%lld) %lld/%lld",
    600  1.1  yamt 		    (long long)count, (long long)scan->u.bmu_avail,
    601  1.1  yamt 		    (long long)radix);
    602  1.1  yamt 
    603  1.1  yamt 	/*
    604  1.1  yamt 	 * Break the free down into its components
    605  1.1  yamt 	 */
    606  1.1  yamt 
    607  1.1  yamt 	radix /= BLIST_META_RADIX;
    608  1.1  yamt 
    609  1.1  yamt 	i = (freeBlk - blk) / radix;
    610  1.1  yamt 	blk += i * radix;
    611  1.1  yamt 	i = i * next_skip + 1;
    612  1.1  yamt 
    613  1.1  yamt 	while (i <= skip && blk < freeBlk + count) {
    614  1.1  yamt 		daddr_t v;
    615  1.1  yamt 
    616  1.1  yamt 		v = blk + radix - freeBlk;
    617  1.1  yamt 		if (v > count)
    618  1.1  yamt 			v = count;
    619  1.1  yamt 
    620  1.1  yamt 		if (scan->bm_bighint == (daddr_t)-1)
    621  1.1  yamt 			panic("blst_meta_free: freeing unexpected range");
    622  1.1  yamt 
    623  1.1  yamt 		if (next_skip == 1) {
    624  1.1  yamt 			blst_leaf_free(&scan[i], freeBlk, v);
    625  1.1  yamt 		} else {
    626  1.1  yamt 			blst_meta_free(&scan[i], freeBlk, v, radix, next_skip - 1, blk);
    627  1.1  yamt 		}
    628  1.1  yamt 		if (scan->bm_bighint < scan[i].bm_bighint)
    629  1.1  yamt 		    scan->bm_bighint = scan[i].bm_bighint;
    630  1.1  yamt 		count -= v;
    631  1.1  yamt 		freeBlk += v;
    632  1.1  yamt 		blk += radix;
    633  1.1  yamt 		i += next_skip;
    634  1.1  yamt 	}
    635  1.1  yamt }
    636  1.1  yamt 
    637  1.1  yamt /*
    638  1.1  yamt  * BLIST_RADIX_COPY() - copy one radix tree to another
    639  1.1  yamt  *
    640  1.1  yamt  *	Locates free space in the source tree and frees it in the destination
    641  1.1  yamt  *	tree.  The space may not already be free in the destination.
    642  1.1  yamt  */
    643  1.1  yamt 
    644  1.1  yamt static void blst_copy(
    645  1.1  yamt 	blmeta_t *scan,
    646  1.1  yamt 	daddr_t blk,
    647  1.1  yamt 	daddr_t radix,
    648  1.1  yamt 	daddr_t skip,
    649  1.1  yamt 	blist_t dest,
    650  1.1  yamt 	daddr_t count
    651  1.1  yamt ) {
    652  1.1  yamt 	int next_skip;
    653  1.1  yamt 	int i;
    654  1.1  yamt 
    655  1.1  yamt 	/*
    656  1.1  yamt 	 * Leaf node
    657  1.1  yamt 	 */
    658  1.1  yamt 
    659  1.1  yamt 	if (radix == BLIST_BMAP_RADIX) {
    660  1.1  yamt 		u_daddr_t v = scan->u.bmu_bitmap;
    661  1.1  yamt 
    662  1.1  yamt 		if (v == (u_daddr_t)-1) {
    663  1.1  yamt 			blist_free(dest, blk, count);
    664  1.1  yamt 		} else if (v != 0) {
    665  1.1  yamt 			int i;
    666  1.1  yamt 
    667  1.1  yamt 			for (i = 0; i < BLIST_BMAP_RADIX && i < count; ++i) {
    668  1.1  yamt 				if (v & (1 << i))
    669  1.1  yamt 					blist_free(dest, blk + i, 1);
    670  1.1  yamt 			}
    671  1.1  yamt 		}
    672  1.1  yamt 		return;
    673  1.1  yamt 	}
    674  1.1  yamt 
    675  1.1  yamt 	/*
    676  1.1  yamt 	 * Meta node
    677  1.1  yamt 	 */
    678  1.1  yamt 
    679  1.1  yamt 	if (scan->u.bmu_avail == 0) {
    680  1.1  yamt 		/*
    681  1.1  yamt 		 * Source all allocated, leave dest allocated
    682  1.1  yamt 		 */
    683  1.1  yamt 		return;
    684  1.1  yamt 	}
    685  1.1  yamt 	if (scan->u.bmu_avail == radix) {
    686  1.1  yamt 		/*
    687  1.1  yamt 		 * Source all free, free entire dest
    688  1.1  yamt 		 */
    689  1.1  yamt 		if (count < radix)
    690  1.1  yamt 			blist_free(dest, blk, count);
    691  1.1  yamt 		else
    692  1.1  yamt 			blist_free(dest, blk, radix);
    693  1.1  yamt 		return;
    694  1.1  yamt 	}
    695  1.1  yamt 
    696  1.1  yamt 
    697  1.1  yamt 	radix /= BLIST_META_RADIX;
    698  1.1  yamt 	next_skip = ((u_int)skip / BLIST_META_RADIX);
    699  1.1  yamt 
    700  1.1  yamt 	for (i = 1; count && i <= skip; i += next_skip) {
    701  1.1  yamt 		if (scan[i].bm_bighint == (daddr_t)-1)
    702  1.1  yamt 			break;
    703  1.1  yamt 
    704  1.1  yamt 		if (count >= radix) {
    705  1.1  yamt 			blst_copy(
    706  1.1  yamt 			    &scan[i],
    707  1.1  yamt 			    blk,
    708  1.1  yamt 			    radix,
    709  1.1  yamt 			    next_skip - 1,
    710  1.1  yamt 			    dest,
    711  1.1  yamt 			    radix
    712  1.1  yamt 			);
    713  1.1  yamt 			count -= radix;
    714  1.1  yamt 		} else {
    715  1.1  yamt 			if (count) {
    716  1.1  yamt 				blst_copy(
    717  1.1  yamt 				    &scan[i],
    718  1.1  yamt 				    blk,
    719  1.1  yamt 				    radix,
    720  1.1  yamt 				    next_skip - 1,
    721  1.1  yamt 				    dest,
    722  1.1  yamt 				    count
    723  1.1  yamt 				);
    724  1.1  yamt 			}
    725  1.1  yamt 			count = 0;
    726  1.1  yamt 		}
    727  1.1  yamt 		blk += radix;
    728  1.1  yamt 	}
    729  1.1  yamt }
    730  1.1  yamt 
    731  1.1  yamt /*
    732  1.1  yamt  * BLST_LEAF_FILL() -	allocate specific blocks in leaf bitmap
    733  1.1  yamt  *
    734  1.1  yamt  *	This routine allocates all blocks in the specified range
    735  1.1  yamt  *	regardless of any existing allocations in that range.  Returns
    736  1.1  yamt  *	the number of blocks allocated by the call.
    737  1.1  yamt  */
    738  1.1  yamt 
    739  1.1  yamt static int
    740  1.1  yamt blst_leaf_fill(blmeta_t *scan, daddr_t blk, int count)
    741  1.1  yamt {
    742  1.1  yamt 	int n = blk & (BLIST_BMAP_RADIX - 1);
    743  1.1  yamt 	int nblks;
    744  1.1  yamt 	u_daddr_t mask, bitmap;
    745  1.1  yamt 
    746  1.1  yamt 	mask = ((u_daddr_t)-1 << n) &
    747  1.1  yamt 	    ((u_daddr_t)-1 >> (BLIST_BMAP_RADIX - count - n));
    748  1.1  yamt 
    749  1.1  yamt 	/* Count the number of blocks we're about to allocate */
    750  1.1  yamt 	bitmap = scan->u.bmu_bitmap & mask;
    751  1.1  yamt 	for (nblks = 0; bitmap != 0; nblks++)
    752  1.1  yamt 		bitmap &= bitmap - 1;
    753  1.1  yamt 
    754  1.1  yamt 	scan->u.bmu_bitmap &= ~mask;
    755  1.1  yamt 	return nblks;
    756  1.1  yamt }
    757  1.1  yamt 
    758  1.1  yamt /*
    759  1.1  yamt  * BLIST_META_FILL() -	allocate specific blocks at a meta node
    760  1.1  yamt  *
    761  1.1  yamt  *	This routine allocates the specified range of blocks,
    762  1.1  yamt  *	regardless of any existing allocations in the range.  The
    763  1.1  yamt  *	range must be within the extent of this node.  Returns the
    764  1.1  yamt  *	number of blocks allocated by the call.
    765  1.1  yamt  */
    766  1.1  yamt static int
    767  1.1  yamt blst_meta_fill(
    768  1.1  yamt 	blmeta_t *scan,
    769  1.1  yamt 	daddr_t allocBlk,
    770  1.1  yamt 	daddr_t count,
    771  1.1  yamt 	daddr_t radix,
    772  1.1  yamt 	int skip,
    773  1.1  yamt 	daddr_t blk
    774  1.1  yamt ) {
    775  1.1  yamt 	int i;
    776  1.1  yamt 	int next_skip = ((u_int)skip / BLIST_META_RADIX);
    777  1.1  yamt 	int nblks = 0;
    778  1.1  yamt 
    779  1.1  yamt 	if (count == radix || scan->u.bmu_avail == 0)  {
    780  1.1  yamt 		/*
    781  1.1  yamt 		 * ALL-ALLOCATED special case
    782  1.1  yamt 		 */
    783  1.1  yamt 		nblks = scan->u.bmu_avail;
    784  1.1  yamt 		scan->u.bmu_avail = 0;
    785  1.1  yamt 		scan->bm_bighint = count;
    786  1.1  yamt 		return nblks;
    787  1.1  yamt 	}
    788  1.1  yamt 
    789  1.1  yamt 	if (scan->u.bmu_avail == radix) {
    790  1.1  yamt 		radix /= BLIST_META_RADIX;
    791  1.1  yamt 
    792  1.1  yamt 		/*
    793  1.1  yamt 		 * ALL-FREE special case, initialize sublevel
    794  1.1  yamt 		 */
    795  1.1  yamt 		for (i = 1; i <= skip; i += next_skip) {
    796  1.1  yamt 			if (scan[i].bm_bighint == (daddr_t)-1)
    797  1.1  yamt 				break;
    798  1.1  yamt 			if (next_skip == 1) {
    799  1.1  yamt 				scan[i].u.bmu_bitmap = (u_daddr_t)-1;
    800  1.1  yamt 				scan[i].bm_bighint = BLIST_BMAP_RADIX;
    801  1.1  yamt 			} else {
    802  1.1  yamt 				scan[i].bm_bighint = radix;
    803  1.1  yamt 				scan[i].u.bmu_avail = radix;
    804  1.1  yamt 			}
    805  1.1  yamt 		}
    806  1.1  yamt 	} else {
    807  1.1  yamt 		radix /= BLIST_META_RADIX;
    808  1.1  yamt 	}
    809  1.1  yamt 
    810  1.1  yamt 	if (count > radix)
    811  1.1  yamt 		panic("blist_meta_fill: allocation too large");
    812  1.1  yamt 
    813  1.1  yamt 	i = (allocBlk - blk) / radix;
    814  1.1  yamt 	blk += i * radix;
    815  1.1  yamt 	i = i * next_skip + 1;
    816  1.1  yamt 
    817  1.1  yamt 	while (i <= skip && blk < allocBlk + count) {
    818  1.1  yamt 		daddr_t v;
    819  1.1  yamt 
    820  1.1  yamt 		v = blk + radix - allocBlk;
    821  1.1  yamt 		if (v > count)
    822  1.1  yamt 			v = count;
    823  1.1  yamt 
    824  1.1  yamt 		if (scan->bm_bighint == (daddr_t)-1)
    825  1.1  yamt 			panic("blst_meta_fill: filling unexpected range");
    826  1.1  yamt 
    827  1.1  yamt 		if (next_skip == 1) {
    828  1.1  yamt 			nblks += blst_leaf_fill(&scan[i], allocBlk, v);
    829  1.1  yamt 		} else {
    830  1.1  yamt 			nblks += blst_meta_fill(&scan[i], allocBlk, v,
    831  1.1  yamt 			    radix, next_skip - 1, blk);
    832  1.1  yamt 		}
    833  1.1  yamt 		count -= v;
    834  1.1  yamt 		allocBlk += v;
    835  1.1  yamt 		blk += radix;
    836  1.1  yamt 		i += next_skip;
    837  1.1  yamt 	}
    838  1.1  yamt 	scan->u.bmu_avail -= nblks;
    839  1.1  yamt 	return nblks;
    840  1.1  yamt }
    841  1.1  yamt 
    842  1.1  yamt /*
    843  1.1  yamt  * BLST_RADIX_INIT() - initialize radix tree
    844  1.1  yamt  *
    845  1.1  yamt  *	Initialize our meta structures and bitmaps and calculate the exact
    846  1.1  yamt  *	amount of space required to manage 'count' blocks - this space may
    847  1.1  yamt  *	be considerably less then the calculated radix due to the large
    848  1.1  yamt  *	RADIX values we use.
    849  1.1  yamt  */
    850  1.1  yamt 
    851  1.1  yamt static daddr_t
    852  1.1  yamt blst_radix_init(blmeta_t *scan, daddr_t radix, int skip, daddr_t count)
    853  1.1  yamt {
    854  1.1  yamt 	int i;
    855  1.1  yamt 	int next_skip;
    856  1.1  yamt 	daddr_t memindex = 0;
    857  1.1  yamt 
    858  1.1  yamt 	/*
    859  1.1  yamt 	 * Leaf node
    860  1.1  yamt 	 */
    861  1.1  yamt 
    862  1.1  yamt 	if (radix == BLIST_BMAP_RADIX) {
    863  1.1  yamt 		if (scan) {
    864  1.1  yamt 			scan->bm_bighint = 0;
    865  1.1  yamt 			scan->u.bmu_bitmap = 0;
    866  1.1  yamt 		}
    867  1.1  yamt 		return(memindex);
    868  1.1  yamt 	}
    869  1.1  yamt 
    870  1.1  yamt 	/*
    871  1.1  yamt 	 * Meta node.  If allocating the entire object we can special
    872  1.1  yamt 	 * case it.  However, we need to figure out how much memory
    873  1.1  yamt 	 * is required to manage 'count' blocks, so we continue on anyway.
    874  1.1  yamt 	 */
    875  1.1  yamt 
    876  1.1  yamt 	if (scan) {
    877  1.1  yamt 		scan->bm_bighint = 0;
    878  1.1  yamt 		scan->u.bmu_avail = 0;
    879  1.1  yamt 	}
    880  1.1  yamt 
    881  1.1  yamt 	radix /= BLIST_META_RADIX;
    882  1.1  yamt 	next_skip = ((u_int)skip / BLIST_META_RADIX);
    883  1.1  yamt 
    884  1.1  yamt 	for (i = 1; i <= skip; i += next_skip) {
    885  1.1  yamt 		if (count >= radix) {
    886  1.1  yamt 			/*
    887  1.1  yamt 			 * Allocate the entire object
    888  1.1  yamt 			 */
    889  1.1  yamt 			memindex = i + blst_radix_init(
    890  1.1  yamt 			    ((scan) ? &scan[i] : NULL),
    891  1.1  yamt 			    radix,
    892  1.1  yamt 			    next_skip - 1,
    893  1.1  yamt 			    radix
    894  1.1  yamt 			);
    895  1.1  yamt 			count -= radix;
    896  1.1  yamt 		} else if (count > 0) {
    897  1.1  yamt 			/*
    898  1.1  yamt 			 * Allocate a partial object
    899  1.1  yamt 			 */
    900  1.1  yamt 			memindex = i + blst_radix_init(
    901  1.1  yamt 			    ((scan) ? &scan[i] : NULL),
    902  1.1  yamt 			    radix,
    903  1.1  yamt 			    next_skip - 1,
    904  1.1  yamt 			    count
    905  1.1  yamt 			);
    906  1.1  yamt 			count = 0;
    907  1.1  yamt 		} else {
    908  1.1  yamt 			/*
    909  1.1  yamt 			 * Add terminator and break out
    910  1.1  yamt 			 */
    911  1.1  yamt 			if (scan)
    912  1.1  yamt 				scan[i].bm_bighint = (daddr_t)-1;
    913  1.1  yamt 			break;
    914  1.1  yamt 		}
    915  1.1  yamt 	}
    916  1.1  yamt 	if (memindex < i)
    917  1.1  yamt 		memindex = i;
    918  1.1  yamt 	return(memindex);
    919  1.1  yamt }
    920  1.1  yamt 
    921  1.1  yamt #ifdef BLIST_DEBUG
    922  1.1  yamt 
    923  1.1  yamt static void
    924  1.1  yamt blst_radix_print(blmeta_t *scan, daddr_t blk, daddr_t radix, int skip, int tab)
    925  1.1  yamt {
    926  1.1  yamt 	int i;
    927  1.1  yamt 	int next_skip;
    928  1.1  yamt 	int lastState = 0;
    929  1.1  yamt 
    930  1.1  yamt 	if (radix == BLIST_BMAP_RADIX) {
    931  1.1  yamt 		printf(
    932  1.1  yamt 		    "%*.*s(%08llx,%lld): bitmap %08llx big=%lld\n",
    933  1.1  yamt 		    tab, tab, "",
    934  1.1  yamt 		    (long long)blk, (long long)radix,
    935  1.1  yamt 		    (long long)scan->u.bmu_bitmap,
    936  1.1  yamt 		    (long long)scan->bm_bighint
    937  1.1  yamt 		);
    938  1.1  yamt 		return;
    939  1.1  yamt 	}
    940  1.1  yamt 
    941  1.1  yamt 	if (scan->u.bmu_avail == 0) {
    942  1.1  yamt 		printf(
    943  1.1  yamt 		    "%*.*s(%08llx,%lld) ALL ALLOCATED\n",
    944  1.1  yamt 		    tab, tab, "",
    945  1.1  yamt 		    (long long)blk,
    946  1.1  yamt 		    (long long)radix
    947  1.1  yamt 		);
    948  1.1  yamt 		return;
    949  1.1  yamt 	}
    950  1.1  yamt 	if (scan->u.bmu_avail == radix) {
    951  1.1  yamt 		printf(
    952  1.1  yamt 		    "%*.*s(%08llx,%lld) ALL FREE\n",
    953  1.1  yamt 		    tab, tab, "",
    954  1.1  yamt 		    (long long)blk,
    955  1.1  yamt 		    (long long)radix
    956  1.1  yamt 		);
    957  1.1  yamt 		return;
    958  1.1  yamt 	}
    959  1.1  yamt 
    960  1.1  yamt 	printf(
    961  1.1  yamt 	    "%*.*s(%08llx,%lld): subtree (%lld/%lld) big=%lld {\n",
    962  1.1  yamt 	    tab, tab, "",
    963  1.1  yamt 	    (long long)blk, (long long)radix,
    964  1.1  yamt 	    (long long)scan->u.bmu_avail,
    965  1.1  yamt 	    (long long)radix,
    966  1.1  yamt 	    (long long)scan->bm_bighint
    967  1.1  yamt 	);
    968  1.1  yamt 
    969  1.1  yamt 	radix /= BLIST_META_RADIX;
    970  1.1  yamt 	next_skip = ((u_int)skip / BLIST_META_RADIX);
    971  1.1  yamt 	tab += 4;
    972  1.1  yamt 
    973  1.1  yamt 	for (i = 1; i <= skip; i += next_skip) {
    974  1.1  yamt 		if (scan[i].bm_bighint == (daddr_t)-1) {
    975  1.1  yamt 			printf(
    976  1.1  yamt 			    "%*.*s(%08llx,%lld): Terminator\n",
    977  1.1  yamt 			    tab, tab, "",
    978  1.1  yamt 			    (long long)blk, (long long)radix
    979  1.1  yamt 			);
    980  1.1  yamt 			lastState = 0;
    981  1.1  yamt 			break;
    982  1.1  yamt 		}
    983  1.1  yamt 		blst_radix_print(
    984  1.1  yamt 		    &scan[i],
    985  1.1  yamt 		    blk,
    986  1.1  yamt 		    radix,
    987  1.1  yamt 		    next_skip - 1,
    988  1.1  yamt 		    tab
    989  1.1  yamt 		);
    990  1.1  yamt 		blk += radix;
    991  1.1  yamt 	}
    992  1.1  yamt 	tab -= 4;
    993  1.1  yamt 
    994  1.1  yamt 	printf(
    995  1.1  yamt 	    "%*.*s}\n",
    996  1.1  yamt 	    tab, tab, ""
    997  1.1  yamt 	);
    998  1.1  yamt }
    999  1.1  yamt 
   1000  1.1  yamt #endif
   1001  1.1  yamt 
   1002  1.1  yamt #ifdef BLIST_DEBUG
   1003  1.1  yamt 
   1004  1.1  yamt int
   1005  1.1  yamt main(int ac, char **av)
   1006  1.1  yamt {
   1007  1.1  yamt 	int size = 1024;
   1008  1.1  yamt 	int i;
   1009  1.1  yamt 	blist_t bl;
   1010  1.1  yamt 
   1011  1.1  yamt 	for (i = 1; i < ac; ++i) {
   1012  1.1  yamt 		const char *ptr = av[i];
   1013  1.1  yamt 		if (*ptr != '-') {
   1014  1.1  yamt 			size = strtol(ptr, NULL, 0);
   1015  1.1  yamt 			continue;
   1016  1.1  yamt 		}
   1017  1.1  yamt 		ptr += 2;
   1018  1.1  yamt 		fprintf(stderr, "Bad option: %s\n", ptr - 2);
   1019  1.1  yamt 		exit(1);
   1020  1.1  yamt 	}
   1021  1.1  yamt 	bl = blist_create(size);
   1022  1.1  yamt 	blist_free(bl, 0, size);
   1023  1.1  yamt 
   1024  1.1  yamt 	for (;;) {
   1025  1.1  yamt 		char buf[1024];
   1026  1.1  yamt 		daddr_t da = 0;
   1027  1.1  yamt 		daddr_t count = 0;
   1028  1.1  yamt 
   1029  1.1  yamt 
   1030  1.1  yamt 		printf("%lld/%lld/%lld> ", (long long)bl->bl_free,
   1031  1.1  yamt 		    (long long)size, (long long)bl->bl_radix);
   1032  1.1  yamt 		fflush(stdout);
   1033  1.1  yamt 		if (fgets(buf, sizeof(buf), stdin) == NULL)
   1034  1.1  yamt 			break;
   1035  1.1  yamt 		switch(buf[0]) {
   1036  1.1  yamt 		case 'r':
   1037  1.1  yamt 			if (sscanf(buf + 1, "%lld", &count) == 1) {
   1038  1.1  yamt 				blist_resize(&bl, count, 1);
   1039  1.1  yamt 			} else {
   1040  1.1  yamt 				printf("?\n");
   1041  1.1  yamt 			}
   1042  1.1  yamt 		case 'p':
   1043  1.1  yamt 			blist_print(bl);
   1044  1.1  yamt 			break;
   1045  1.1  yamt 		case 'a':
   1046  1.1  yamt 			if (sscanf(buf + 1, "%lld", &count) == 1) {
   1047  1.1  yamt 				daddr_t blk = blist_alloc(bl, count);
   1048  1.1  yamt 				printf("    R=%08llx\n", (long long)blk);
   1049  1.1  yamt 			} else {
   1050  1.1  yamt 				printf("?\n");
   1051  1.1  yamt 			}
   1052  1.1  yamt 			break;
   1053  1.1  yamt 		case 'f':
   1054  1.1  yamt 			if (sscanf(buf + 1, "%llx %lld",
   1055  1.1  yamt 			    (long long *)&da, (long long *)&count) == 2) {
   1056  1.1  yamt 				blist_free(bl, da, count);
   1057  1.1  yamt 			} else {
   1058  1.1  yamt 				printf("?\n");
   1059  1.1  yamt 			}
   1060  1.1  yamt 			break;
   1061  1.1  yamt 		case 'l':
   1062  1.1  yamt 			if (sscanf(buf + 1, "%llx %lld",
   1063  1.1  yamt 			    (long long *)&da, (long long *)&count) == 2) {
   1064  1.1  yamt 				printf("    n=%d\n",
   1065  1.1  yamt 				    blist_fill(bl, da, count));
   1066  1.1  yamt 			} else {
   1067  1.1  yamt 				printf("?\n");
   1068  1.1  yamt 			}
   1069  1.1  yamt 			break;
   1070  1.1  yamt 		case '?':
   1071  1.1  yamt 		case 'h':
   1072  1.1  yamt 			puts(
   1073  1.1  yamt 			    "p          -print\n"
   1074  1.1  yamt 			    "a %d       -allocate\n"
   1075  1.1  yamt 			    "f %x %d    -free\n"
   1076  1.1  yamt 			    "l %x %d    -fill\n"
   1077  1.1  yamt 			    "r %d       -resize\n"
   1078  1.1  yamt 			    "h/?        -help"
   1079  1.1  yamt 			);
   1080  1.1  yamt 			break;
   1081  1.1  yamt 		default:
   1082  1.1  yamt 			printf("?\n");
   1083  1.1  yamt 			break;
   1084  1.1  yamt 		}
   1085  1.1  yamt 	}
   1086  1.1  yamt 	return(0);
   1087  1.1  yamt }
   1088  1.1  yamt 
   1089  1.1  yamt void
   1090  1.1  yamt panic(const char *ctl, ...)
   1091  1.1  yamt {
   1092  1.1  yamt 	va_list va;
   1093  1.1  yamt 
   1094  1.1  yamt 	va_start(va, ctl);
   1095  1.1  yamt 	vfprintf(stderr, ctl, va);
   1096  1.1  yamt 	fprintf(stderr, "\n");
   1097  1.1  yamt 	va_end(va);
   1098  1.1  yamt 	exit(1);
   1099  1.1  yamt }
   1100  1.1  yamt 
   1101  1.1  yamt #endif
   1102  1.1  yamt 
   1103