Home | History | Annotate | Line # | Download | only in kern
subr_blist.c revision 1.15.10.1
      1  1.15.10.1  perseant /*	$NetBSD: subr_blist.c,v 1.15.10.1 2025/08/02 05:57:42 perseant Exp $	*/
      2        1.2      yamt 
      3        1.1      yamt /*-
      4        1.1      yamt  * Copyright (c) 1998 Matthew Dillon.  All Rights Reserved.
      5        1.1      yamt  * Redistribution and use in source and binary forms, with or without
      6        1.1      yamt  * modification, are permitted provided that the following conditions
      7        1.1      yamt  * are met:
      8        1.1      yamt  * 1. Redistributions of source code must retain the above copyright
      9        1.1      yamt  *    notice, this list of conditions and the following disclaimer.
     10        1.1      yamt  * 2. Redistributions in binary form must reproduce the above copyright
     11        1.1      yamt  *    notice, this list of conditions and the following disclaimer in the
     12        1.1      yamt  *    documentation and/or other materials provided with the distribution.
     13        1.1      yamt  * 4. Neither the name of the University nor the names of its contributors
     14        1.1      yamt  *    may be used to endorse or promote products derived from this software
     15        1.1      yamt  *    without specific prior written permission.
     16        1.1      yamt  *
     17        1.1      yamt  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS
     18        1.1      yamt  * OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
     19        1.1      yamt  * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
     20        1.1      yamt  * ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY
     21        1.1      yamt  * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
     22        1.1      yamt  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE
     23        1.1      yamt  * GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
     24        1.1      yamt  * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY,
     25        1.1      yamt  * WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING
     26        1.1      yamt  * NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
     27        1.1      yamt  * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
     28        1.1      yamt  */
     29        1.1      yamt /*
     30        1.1      yamt  * BLIST.C -	Bitmap allocator/deallocator, using a radix tree with hinting
     31        1.1      yamt  *
     32        1.1      yamt  *	This module implements a general bitmap allocator/deallocator.  The
     33        1.1      yamt  *	allocator eats around 2 bits per 'block'.  The module does not
     34       1.12       wiz  *	try to interpret the meaning of a 'block' other than to return
     35        1.3      yamt  *	BLIST_NONE on an allocation failure.
     36        1.1      yamt  *
     37        1.1      yamt  *	A radix tree is used to maintain the bitmap.  Two radix constants are
     38        1.1      yamt  *	involved:  One for the bitmaps contained in the leaf nodes (typically
     39        1.1      yamt  *	32), and one for the meta nodes (typically 16).  Both meta and leaf
     40        1.1      yamt  *	nodes have a hint field.  This field gives us a hint as to the largest
     41        1.1      yamt  *	free contiguous range of blocks under the node.  It may contain a
     42        1.1      yamt  *	value that is too high, but will never contain a value that is too
     43        1.1      yamt  *	low.  When the radix tree is searched, allocation failures in subtrees
     44        1.1      yamt  *	update the hint.
     45        1.1      yamt  *
     46        1.1      yamt  *	The radix tree also implements two collapsed states for meta nodes:
     47        1.1      yamt  *	the ALL-ALLOCATED state and the ALL-FREE state.  If a meta node is
     48        1.1      yamt  *	in either of these two states, all information contained underneath
     49        1.1      yamt  *	the node is considered stale.  These states are used to optimize
     50        1.1      yamt  *	allocation and freeing operations.
     51        1.1      yamt  *
     52        1.1      yamt  * 	The hinting greatly increases code efficiency for allocations while
     53        1.1      yamt  *	the general radix structure optimizes both allocations and frees.  The
     54        1.1      yamt  *	radix tree should be able to operate well no matter how much
     55        1.1      yamt  *	fragmentation there is and no matter how large a bitmap is used.
     56        1.1      yamt  *
     57        1.1      yamt  *	Unlike the rlist code, the blist code wires all necessary memory at
     58        1.1      yamt  *	creation time.  Neither allocations nor frees require interaction with
     59        1.1      yamt  *	the memory subsystem.  In contrast, the rlist code may allocate memory
     60        1.1      yamt  *	on an rlist_free() call.  The non-blocking features of the blist code
     61        1.1      yamt  *	are used to great advantage in the swap code (vm/nswap_pager.c).  The
     62       1.12       wiz  *	rlist code uses a little less overall memory than the blist code (but
     63        1.1      yamt  *	due to swap interleaving not all that much less), but the blist code
     64        1.1      yamt  *	scales much, much better.
     65        1.1      yamt  *
     66       1.14    andvar  *	LAYOUT: The radix tree is laid out recursively using a
     67       1.14    andvar  *	linear array.  Each meta node is immediately followed (laid out
     68        1.1      yamt  *	sequentially in memory) by BLIST_META_RADIX lower level nodes.  This
     69        1.1      yamt  *	is a recursive structure but one that can be easily scanned through
     70        1.1      yamt  *	a very simple 'skip' calculation.  In order to support large radixes,
     71        1.1      yamt  *	portions of the tree may reside outside our memory allocation.  We
     72        1.1      yamt  *	handle this with an early-termination optimization (when bighint is
     73        1.1      yamt  *	set to -1) on the scan.  The memory allocation is only large enough
     74        1.1      yamt  *	to cover the number of blocks requested at creation time even if it
     75        1.1      yamt  *	must be encompassed in larger root-node radix.
     76        1.1      yamt  *
     77       1.12       wiz  *	NOTE: the allocator cannot currently allocate more than
     78        1.1      yamt  *	BLIST_BMAP_RADIX blocks per call.  It will panic with 'allocation too
     79        1.1      yamt  *	large' if you try.  This is an area that could use improvement.  The
     80  1.15.10.1  perseant  *	radix is large enough that this restriction does not affect the swap
     81        1.1      yamt  *	system, though.  Currently only the allocation code is effected by
     82        1.1      yamt  *	this algorithmic unfeature.  The freeing code can handle arbitrary
     83        1.1      yamt  *	ranges.
     84        1.1      yamt  *
     85        1.1      yamt  *	This code can be compiled stand-alone for debugging.
     86        1.1      yamt  */
     87        1.1      yamt 
     88        1.1      yamt #include <sys/cdefs.h>
     89  1.15.10.1  perseant __KERNEL_RCSID(0, "$NetBSD: subr_blist.c,v 1.15.10.1 2025/08/02 05:57:42 perseant Exp $");
     90        1.2      yamt #if 0
     91        1.1      yamt __FBSDID("$FreeBSD: src/sys/kern/subr_blist.c,v 1.17 2004/06/04 04:03:25 alc Exp $");
     92        1.2      yamt #endif
     93        1.1      yamt 
     94        1.1      yamt #ifdef _KERNEL
     95        1.1      yamt 
     96        1.1      yamt #include <sys/param.h>
     97        1.1      yamt #include <sys/systm.h>
     98        1.1      yamt #include <sys/blist.h>
     99       1.10     rmind #include <sys/kmem.h>
    100        1.1      yamt 
    101        1.1      yamt #else
    102        1.1      yamt 
    103        1.1      yamt #ifndef BLIST_NO_DEBUG
    104        1.1      yamt #define BLIST_DEBUG
    105        1.1      yamt #endif
    106        1.1      yamt 
    107        1.1      yamt #include <sys/types.h>
    108        1.1      yamt #include <stdio.h>
    109        1.1      yamt #include <string.h>
    110        1.1      yamt #include <stdlib.h>
    111        1.1      yamt #include <stdarg.h>
    112        1.3      yamt #include <inttypes.h>
    113        1.1      yamt 
    114       1.10     rmind #define	KM_SLEEP 1
    115       1.13     zafer #define	kmem_zalloc(a,b) calloc(1, (a))
    116       1.13     zafer #define	kmem_alloc(a,b) malloc(a)
    117       1.10     rmind #define	kmem_free(a,b) free(a)
    118        1.1      yamt 
    119        1.3      yamt #include "../sys/blist.h"
    120        1.1      yamt 
    121       1.11  christos void panic(const char *ctl, ...) __printflike(1, 2);
    122        1.1      yamt 
    123        1.1      yamt #endif
    124        1.1      yamt 
    125        1.1      yamt /*
    126        1.4      yamt  * blmeta and bl_bitmap_t MUST be a power of 2 in size.
    127        1.4      yamt  */
    128        1.4      yamt 
    129        1.4      yamt typedef struct blmeta {
    130        1.4      yamt 	union {
    131        1.5      yamt 		blist_blkno_t	bmu_avail; /* space available under us	*/
    132        1.5      yamt 		blist_bitmap_t	bmu_bitmap; /* bitmap if we are a leaf	*/
    133        1.4      yamt 	} u;
    134        1.5      yamt 	blist_blkno_t	bm_bighint;	/* biggest contiguous block hint*/
    135        1.4      yamt } blmeta_t;
    136        1.4      yamt 
    137        1.4      yamt struct blist {
    138        1.5      yamt 	blist_blkno_t		bl_blocks;	/* area of coverage		*/
    139        1.5      yamt 	blist_blkno_t		bl_radix;	/* coverage radix		*/
    140        1.5      yamt 	blist_blkno_t		bl_skip;	/* starting skip		*/
    141        1.5      yamt 	blist_blkno_t		bl_free;	/* number of free blocks	*/
    142        1.4      yamt 	blmeta_t	*bl_root;	/* root of radix tree		*/
    143        1.5      yamt 	blist_blkno_t		bl_rootblks;	/* blks allocated for tree */
    144        1.4      yamt };
    145        1.4      yamt 
    146        1.4      yamt #define BLIST_META_RADIX	16
    147        1.4      yamt 
    148        1.4      yamt /*
    149        1.1      yamt  * static support functions
    150        1.1      yamt  */
    151        1.1      yamt 
    152        1.5      yamt static blist_blkno_t blst_leaf_alloc(blmeta_t *scan, blist_blkno_t blk,
    153        1.5      yamt     int count);
    154        1.5      yamt static blist_blkno_t blst_meta_alloc(blmeta_t *scan, blist_blkno_t blk,
    155        1.5      yamt     blist_blkno_t count, blist_blkno_t radix, blist_blkno_t skip);
    156        1.5      yamt static void blst_leaf_free(blmeta_t *scan, blist_blkno_t relblk, int count);
    157        1.5      yamt static void blst_meta_free(blmeta_t *scan, blist_blkno_t freeBlk,
    158        1.5      yamt     blist_blkno_t count, blist_blkno_t radix, blist_blkno_t skip,
    159        1.5      yamt     blist_blkno_t blk);
    160        1.5      yamt static void blst_copy(blmeta_t *scan, blist_blkno_t blk, blist_blkno_t radix,
    161        1.5      yamt     blist_blkno_t skip, blist_t dest, blist_blkno_t count);
    162        1.5      yamt static int blst_leaf_fill(blmeta_t *scan, blist_blkno_t blk, int count);
    163        1.5      yamt static blist_blkno_t blst_meta_fill(blmeta_t *scan, blist_blkno_t allocBlk,
    164        1.5      yamt     blist_blkno_t count, blist_blkno_t radix, blist_blkno_t skip,
    165        1.5      yamt     blist_blkno_t blk);
    166        1.5      yamt static blist_blkno_t blst_radix_init(blmeta_t *scan, blist_blkno_t radix,
    167        1.5      yamt     blist_blkno_t skip, blist_blkno_t count);
    168        1.1      yamt #ifndef _KERNEL
    169        1.5      yamt static void blst_radix_print(blmeta_t *scan, blist_blkno_t blk,
    170        1.5      yamt     blist_blkno_t radix, blist_blkno_t skip, int tab);
    171        1.1      yamt #endif
    172        1.1      yamt 
    173        1.1      yamt /*
    174        1.1      yamt  * blist_create() - create a blist capable of handling up to the specified
    175        1.1      yamt  *		    number of blocks
    176        1.1      yamt  *
    177       1.12       wiz  *	blocks must be greater than 0
    178        1.1      yamt  *
    179        1.1      yamt  *	The smallest blist consists of a single leaf node capable of
    180        1.1      yamt  *	managing BLIST_BMAP_RADIX blocks.
    181        1.1      yamt  */
    182        1.1      yamt 
    183        1.1      yamt blist_t
    184        1.5      yamt blist_create(blist_blkno_t blocks)
    185        1.1      yamt {
    186        1.1      yamt 	blist_t bl;
    187        1.5      yamt 	blist_blkno_t radix;
    188        1.5      yamt 	blist_blkno_t skip = 0;
    189        1.1      yamt 
    190        1.1      yamt 	/*
    191        1.1      yamt 	 * Calculate radix and skip field used for scanning.
    192        1.5      yamt 	 *
    193        1.5      yamt 	 * XXX check overflow
    194        1.1      yamt 	 */
    195        1.1      yamt 	radix = BLIST_BMAP_RADIX;
    196        1.1      yamt 
    197        1.1      yamt 	while (radix < blocks) {
    198        1.1      yamt 		radix *= BLIST_META_RADIX;
    199        1.1      yamt 		skip = (skip + 1) * BLIST_META_RADIX;
    200        1.1      yamt 	}
    201        1.1      yamt 
    202       1.10     rmind 	bl = kmem_zalloc(sizeof(struct blist), KM_SLEEP);
    203        1.1      yamt 
    204        1.1      yamt 	bl->bl_blocks = blocks;
    205        1.1      yamt 	bl->bl_radix = radix;
    206        1.1      yamt 	bl->bl_skip = skip;
    207        1.1      yamt 	bl->bl_rootblks = 1 +
    208        1.1      yamt 	    blst_radix_init(NULL, bl->bl_radix, bl->bl_skip, blocks);
    209       1.10     rmind 	bl->bl_root = kmem_alloc(sizeof(blmeta_t) * bl->bl_rootblks, KM_SLEEP);
    210        1.1      yamt 
    211        1.1      yamt #if defined(BLIST_DEBUG)
    212        1.1      yamt 	printf(
    213        1.3      yamt 		"BLIST representing %" PRIu64 " blocks (%" PRIu64 " MB of swap)"
    214        1.3      yamt 		", requiring %" PRIu64 "K of ram\n",
    215        1.5      yamt 		(uint64_t)bl->bl_blocks,
    216        1.5      yamt 		(uint64_t)bl->bl_blocks * 4 / 1024,
    217        1.5      yamt 		((uint64_t)bl->bl_rootblks * sizeof(blmeta_t) + 1023) / 1024
    218        1.1      yamt 	);
    219        1.3      yamt 	printf("BLIST raw radix tree contains %" PRIu64 " records\n",
    220        1.5      yamt 	    (uint64_t)bl->bl_rootblks);
    221        1.1      yamt #endif
    222        1.1      yamt 	blst_radix_init(bl->bl_root, bl->bl_radix, bl->bl_skip, blocks);
    223        1.1      yamt 
    224        1.1      yamt 	return(bl);
    225        1.1      yamt }
    226        1.1      yamt 
    227        1.1      yamt void
    228        1.1      yamt blist_destroy(blist_t bl)
    229        1.1      yamt {
    230       1.10     rmind 
    231       1.10     rmind 	kmem_free(bl->bl_root, sizeof(blmeta_t) * bl->bl_rootblks);
    232       1.10     rmind 	kmem_free(bl, sizeof(struct blist));
    233        1.1      yamt }
    234        1.1      yamt 
    235        1.1      yamt /*
    236        1.1      yamt  * blist_alloc() - reserve space in the block bitmap.  Return the base
    237        1.3      yamt  *		     of a contiguous region or BLIST_NONE if space could
    238        1.1      yamt  *		     not be allocated.
    239        1.1      yamt  */
    240        1.1      yamt 
    241        1.5      yamt blist_blkno_t
    242        1.5      yamt blist_alloc(blist_t bl, blist_blkno_t count)
    243        1.1      yamt {
    244        1.5      yamt 	blist_blkno_t blk = BLIST_NONE;
    245        1.1      yamt 
    246        1.1      yamt 	if (bl) {
    247        1.1      yamt 		if (bl->bl_radix == BLIST_BMAP_RADIX)
    248        1.1      yamt 			blk = blst_leaf_alloc(bl->bl_root, 0, count);
    249        1.1      yamt 		else
    250        1.1      yamt 			blk = blst_meta_alloc(bl->bl_root, 0, count, bl->bl_radix, bl->bl_skip);
    251        1.3      yamt 		if (blk != BLIST_NONE)
    252        1.1      yamt 			bl->bl_free -= count;
    253        1.1      yamt 	}
    254        1.1      yamt 	return(blk);
    255        1.1      yamt }
    256        1.1      yamt 
    257        1.1      yamt /*
    258        1.1      yamt  * blist_free() -	free up space in the block bitmap.  Return the base
    259       1.15    andvar  *		     	of a contiguous region.  Panic if an inconsistency is
    260        1.1      yamt  *			found.
    261        1.1      yamt  */
    262        1.1      yamt 
    263        1.1      yamt void
    264        1.5      yamt blist_free(blist_t bl, blist_blkno_t blkno, blist_blkno_t count)
    265        1.1      yamt {
    266        1.1      yamt 	if (bl) {
    267        1.1      yamt 		if (bl->bl_radix == BLIST_BMAP_RADIX)
    268        1.1      yamt 			blst_leaf_free(bl->bl_root, blkno, count);
    269        1.1      yamt 		else
    270        1.1      yamt 			blst_meta_free(bl->bl_root, blkno, count, bl->bl_radix, bl->bl_skip, 0);
    271        1.1      yamt 		bl->bl_free += count;
    272        1.1      yamt 	}
    273        1.1      yamt }
    274        1.1      yamt 
    275        1.1      yamt /*
    276        1.1      yamt  * blist_fill() -	mark a region in the block bitmap as off-limits
    277        1.1      yamt  *			to the allocator (i.e. allocate it), ignoring any
    278        1.1      yamt  *			existing allocations.  Return the number of blocks
    279        1.1      yamt  *			actually filled that were free before the call.
    280        1.1      yamt  */
    281        1.1      yamt 
    282        1.5      yamt blist_blkno_t
    283        1.5      yamt blist_fill(blist_t bl, blist_blkno_t blkno, blist_blkno_t count)
    284        1.1      yamt {
    285        1.5      yamt 	blist_blkno_t filled;
    286        1.1      yamt 
    287        1.1      yamt 	if (bl) {
    288        1.1      yamt 		if (bl->bl_radix == BLIST_BMAP_RADIX)
    289        1.1      yamt 			filled = blst_leaf_fill(bl->bl_root, blkno, count);
    290        1.1      yamt 		else
    291        1.1      yamt 			filled = blst_meta_fill(bl->bl_root, blkno, count,
    292        1.1      yamt 			    bl->bl_radix, bl->bl_skip, 0);
    293        1.1      yamt 		bl->bl_free -= filled;
    294        1.1      yamt 		return filled;
    295        1.1      yamt 	} else
    296        1.1      yamt 		return 0;
    297        1.1      yamt }
    298        1.1      yamt 
    299        1.1      yamt /*
    300        1.1      yamt  * blist_resize() -	resize an existing radix tree to handle the
    301        1.1      yamt  *			specified number of blocks.  This will reallocate
    302        1.1      yamt  *			the tree and transfer the previous bitmap to the new
    303        1.1      yamt  *			one.  When extending the tree you can specify whether
    304        1.1      yamt  *			the new blocks are to left allocated or freed.
    305        1.1      yamt  */
    306        1.1      yamt 
    307        1.1      yamt void
    308        1.5      yamt blist_resize(blist_t *pbl, blist_blkno_t count, int freenew)
    309        1.1      yamt {
    310        1.1      yamt     blist_t newbl = blist_create(count);
    311        1.1      yamt     blist_t save = *pbl;
    312        1.1      yamt 
    313        1.1      yamt     *pbl = newbl;
    314        1.1      yamt     if (count > save->bl_blocks)
    315        1.1      yamt 	    count = save->bl_blocks;
    316        1.1      yamt     blst_copy(save->bl_root, 0, save->bl_radix, save->bl_skip, newbl, count);
    317        1.1      yamt 
    318        1.1      yamt     /*
    319        1.1      yamt      * If resizing upwards, should we free the new space or not?
    320        1.1      yamt      */
    321        1.1      yamt     if (freenew && count < newbl->bl_blocks) {
    322        1.1      yamt 	    blist_free(newbl, count, newbl->bl_blocks - count);
    323        1.1      yamt     }
    324        1.1      yamt     blist_destroy(save);
    325        1.1      yamt }
    326        1.1      yamt 
    327        1.1      yamt #ifdef BLIST_DEBUG
    328        1.1      yamt 
    329        1.1      yamt /*
    330        1.1      yamt  * blist_print()    - dump radix tree
    331        1.1      yamt  */
    332        1.1      yamt 
    333        1.1      yamt void
    334        1.1      yamt blist_print(blist_t bl)
    335        1.1      yamt {
    336        1.1      yamt 	printf("BLIST {\n");
    337        1.1      yamt 	blst_radix_print(bl->bl_root, 0, bl->bl_radix, bl->bl_skip, 4);
    338        1.1      yamt 	printf("}\n");
    339        1.1      yamt }
    340        1.1      yamt 
    341        1.1      yamt #endif
    342        1.1      yamt 
    343        1.1      yamt /************************************************************************
    344        1.1      yamt  *			  ALLOCATION SUPPORT FUNCTIONS			*
    345        1.1      yamt  ************************************************************************
    346        1.1      yamt  *
    347        1.1      yamt  *	These support functions do all the actual work.  They may seem
    348        1.1      yamt  *	rather longish, but that's because I've commented them up.  The
    349        1.1      yamt  *	actual code is straight forward.
    350        1.1      yamt  *
    351        1.1      yamt  */
    352        1.1      yamt 
    353        1.1      yamt /*
    354        1.1      yamt  * blist_leaf_alloc() -	allocate at a leaf in the radix tree (a bitmap).
    355        1.1      yamt  *
    356        1.1      yamt  *	This is the core of the allocator and is optimized for the 1 block
    357        1.1      yamt  *	and the BLIST_BMAP_RADIX block allocation cases.  Other cases are
    358        1.1      yamt  *	somewhat slower.  The 1 block allocation case is log2 and extremely
    359        1.1      yamt  *	quick.
    360        1.1      yamt  */
    361        1.1      yamt 
    362        1.5      yamt static blist_blkno_t
    363        1.1      yamt blst_leaf_alloc(
    364        1.1      yamt 	blmeta_t *scan,
    365        1.5      yamt 	blist_blkno_t blk,
    366        1.1      yamt 	int count
    367        1.1      yamt ) {
    368        1.5      yamt 	blist_bitmap_t orig = scan->u.bmu_bitmap;
    369        1.1      yamt 
    370        1.1      yamt 	if (orig == 0) {
    371        1.1      yamt 		/*
    372        1.1      yamt 		 * Optimize bitmap all-allocated case.  Also, count = 1
    373        1.1      yamt 		 * case assumes at least 1 bit is free in the bitmap, so
    374        1.1      yamt 		 * we have to take care of this case here.
    375        1.1      yamt 		 */
    376        1.1      yamt 		scan->bm_bighint = 0;
    377        1.3      yamt 		return(BLIST_NONE);
    378        1.1      yamt 	}
    379        1.1      yamt 	if (count == 1) {
    380        1.1      yamt 		/*
    381        1.1      yamt 		 * Optimized code to allocate one bit out of the bitmap
    382        1.1      yamt 		 */
    383        1.5      yamt 		blist_bitmap_t mask;
    384        1.1      yamt 		int j = BLIST_BMAP_RADIX/2;
    385        1.1      yamt 		int r = 0;
    386        1.1      yamt 
    387        1.5      yamt 		mask = (blist_bitmap_t)-1 >> (BLIST_BMAP_RADIX/2);
    388        1.1      yamt 
    389        1.1      yamt 		while (j) {
    390        1.1      yamt 			if ((orig & mask) == 0) {
    391        1.1      yamt 			    r += j;
    392        1.1      yamt 			    orig >>= j;
    393        1.1      yamt 			}
    394        1.1      yamt 			j >>= 1;
    395        1.1      yamt 			mask >>= j;
    396        1.1      yamt 		}
    397        1.5      yamt 		scan->u.bmu_bitmap &= ~((blist_bitmap_t)1 << r);
    398        1.1      yamt 		return(blk + r);
    399        1.1      yamt 	}
    400        1.1      yamt 	if (count <= BLIST_BMAP_RADIX) {
    401        1.1      yamt 		/*
    402        1.1      yamt 		 * non-optimized code to allocate N bits out of the bitmap.
    403        1.1      yamt 		 * The more bits, the faster the code runs.  It will run
    404        1.1      yamt 		 * the slowest allocating 2 bits, but since there aren't any
    405        1.1      yamt 		 * memory ops in the core loop (or shouldn't be, anyway),
    406        1.1      yamt 		 * you probably won't notice the difference.
    407        1.1      yamt 		 */
    408        1.1      yamt 		int j;
    409        1.1      yamt 		int n = BLIST_BMAP_RADIX - count;
    410        1.5      yamt 		blist_bitmap_t mask;
    411        1.1      yamt 
    412        1.5      yamt 		mask = (blist_bitmap_t)-1 >> n;
    413        1.1      yamt 
    414        1.1      yamt 		for (j = 0; j <= n; ++j) {
    415        1.1      yamt 			if ((orig & mask) == mask) {
    416        1.1      yamt 				scan->u.bmu_bitmap &= ~mask;
    417        1.1      yamt 				return(blk + j);
    418        1.1      yamt 			}
    419        1.1      yamt 			mask = (mask << 1);
    420        1.1      yamt 		}
    421        1.1      yamt 	}
    422        1.1      yamt 	/*
    423        1.1      yamt 	 * We couldn't allocate count in this subtree, update bighint.
    424        1.1      yamt 	 */
    425        1.1      yamt 	scan->bm_bighint = count - 1;
    426        1.3      yamt 	return(BLIST_NONE);
    427        1.1      yamt }
    428        1.1      yamt 
    429        1.1      yamt /*
    430        1.1      yamt  * blist_meta_alloc() -	allocate at a meta in the radix tree.
    431        1.1      yamt  *
    432        1.1      yamt  *	Attempt to allocate at a meta node.  If we can't, we update
    433        1.1      yamt  *	bighint and return a failure.  Updating bighint optimize future
    434        1.1      yamt  *	calls that hit this node.  We have to check for our collapse cases
    435        1.1      yamt  *	and we have a few optimizations strewn in as well.
    436        1.1      yamt  */
    437        1.1      yamt 
    438        1.5      yamt static blist_blkno_t
    439        1.1      yamt blst_meta_alloc(
    440        1.1      yamt 	blmeta_t *scan,
    441        1.5      yamt 	blist_blkno_t blk,
    442        1.5      yamt 	blist_blkno_t count,
    443        1.5      yamt 	blist_blkno_t radix,
    444        1.5      yamt 	blist_blkno_t skip
    445        1.1      yamt ) {
    446        1.5      yamt 	blist_blkno_t i;
    447        1.5      yamt 	blist_blkno_t next_skip = (skip / BLIST_META_RADIX);
    448        1.1      yamt 
    449        1.1      yamt 	if (scan->u.bmu_avail == 0)  {
    450        1.1      yamt 		/*
    451        1.1      yamt 		 * ALL-ALLOCATED special case
    452        1.1      yamt 		 */
    453        1.1      yamt 		scan->bm_bighint = count;
    454        1.3      yamt 		return(BLIST_NONE);
    455        1.1      yamt 	}
    456        1.1      yamt 
    457        1.1      yamt 	if (scan->u.bmu_avail == radix) {
    458        1.1      yamt 		radix /= BLIST_META_RADIX;
    459        1.1      yamt 
    460        1.1      yamt 		/*
    461        1.1      yamt 		 * ALL-FREE special case, initialize uninitialize
    462        1.1      yamt 		 * sublevel.
    463        1.1      yamt 		 */
    464        1.1      yamt 		for (i = 1; i <= skip; i += next_skip) {
    465        1.5      yamt 			if (scan[i].bm_bighint == (blist_blkno_t)-1)
    466        1.1      yamt 				break;
    467        1.1      yamt 			if (next_skip == 1) {
    468        1.5      yamt 				scan[i].u.bmu_bitmap = (blist_bitmap_t)-1;
    469        1.1      yamt 				scan[i].bm_bighint = BLIST_BMAP_RADIX;
    470        1.1      yamt 			} else {
    471        1.1      yamt 				scan[i].bm_bighint = radix;
    472        1.1      yamt 				scan[i].u.bmu_avail = radix;
    473        1.1      yamt 			}
    474        1.1      yamt 		}
    475        1.1      yamt 	} else {
    476        1.1      yamt 		radix /= BLIST_META_RADIX;
    477        1.1      yamt 	}
    478        1.1      yamt 
    479        1.1      yamt 	for (i = 1; i <= skip; i += next_skip) {
    480        1.5      yamt 		if (scan[i].bm_bighint == (blist_blkno_t)-1) {
    481        1.3      yamt 			/*
    482        1.3      yamt 			 * Terminator
    483        1.3      yamt 			 */
    484        1.3      yamt 			break;
    485        1.3      yamt 		} else if (count <= scan[i].bm_bighint) {
    486        1.1      yamt 			/*
    487        1.1      yamt 			 * count fits in object
    488        1.1      yamt 			 */
    489        1.5      yamt 			blist_blkno_t r;
    490        1.1      yamt 			if (next_skip == 1) {
    491        1.1      yamt 				r = blst_leaf_alloc(&scan[i], blk, count);
    492        1.1      yamt 			} else {
    493        1.1      yamt 				r = blst_meta_alloc(&scan[i], blk, count, radix, next_skip - 1);
    494        1.1      yamt 			}
    495        1.3      yamt 			if (r != BLIST_NONE) {
    496        1.1      yamt 				scan->u.bmu_avail -= count;
    497        1.1      yamt 				if (scan->bm_bighint > scan->u.bmu_avail)
    498        1.1      yamt 					scan->bm_bighint = scan->u.bmu_avail;
    499        1.1      yamt 				return(r);
    500        1.1      yamt 			}
    501        1.1      yamt 		} else if (count > radix) {
    502        1.1      yamt 			/*
    503        1.1      yamt 			 * count does not fit in object even if it were
    504        1.1      yamt 			 * complete free.
    505        1.1      yamt 			 */
    506        1.1      yamt 			panic("blist_meta_alloc: allocation too large");
    507        1.1      yamt 		}
    508        1.1      yamt 		blk += radix;
    509        1.1      yamt 	}
    510        1.1      yamt 
    511        1.1      yamt 	/*
    512        1.1      yamt 	 * We couldn't allocate count in this subtree, update bighint.
    513        1.1      yamt 	 */
    514        1.1      yamt 	if (scan->bm_bighint >= count)
    515        1.1      yamt 		scan->bm_bighint = count - 1;
    516        1.3      yamt 	return(BLIST_NONE);
    517        1.1      yamt }
    518        1.1      yamt 
    519        1.1      yamt /*
    520        1.1      yamt  * BLST_LEAF_FREE() -	free allocated block from leaf bitmap
    521        1.1      yamt  *
    522        1.1      yamt  */
    523        1.1      yamt 
    524        1.1      yamt static void
    525        1.1      yamt blst_leaf_free(
    526        1.1      yamt 	blmeta_t *scan,
    527        1.5      yamt 	blist_blkno_t blk,
    528        1.1      yamt 	int count
    529        1.1      yamt ) {
    530        1.1      yamt 	/*
    531        1.1      yamt 	 * free some data in this bitmap
    532        1.1      yamt 	 *
    533        1.1      yamt 	 * e.g.
    534        1.1      yamt 	 *	0000111111111110000
    535        1.1      yamt 	 *          \_________/\__/
    536        1.1      yamt 	 *		v        n
    537        1.1      yamt 	 */
    538        1.1      yamt 	int n = blk & (BLIST_BMAP_RADIX - 1);
    539        1.5      yamt 	blist_bitmap_t mask;
    540        1.1      yamt 
    541        1.5      yamt 	mask = ((blist_bitmap_t)-1 << n) &
    542        1.5      yamt 	    ((blist_bitmap_t)-1 >> (BLIST_BMAP_RADIX - count - n));
    543        1.1      yamt 
    544        1.1      yamt 	if (scan->u.bmu_bitmap & mask)
    545        1.1      yamt 		panic("blst_radix_free: freeing free block");
    546        1.1      yamt 	scan->u.bmu_bitmap |= mask;
    547        1.1      yamt 
    548        1.1      yamt 	/*
    549        1.1      yamt 	 * We could probably do a better job here.  We are required to make
    550        1.1      yamt 	 * bighint at least as large as the biggest contiguous block of
    551        1.1      yamt 	 * data.  If we just shoehorn it, a little extra overhead will
    552        1.1      yamt 	 * be incured on the next allocation (but only that one typically).
    553        1.1      yamt 	 */
    554        1.1      yamt 	scan->bm_bighint = BLIST_BMAP_RADIX;
    555        1.1      yamt }
    556        1.1      yamt 
    557        1.1      yamt /*
    558        1.1      yamt  * BLST_META_FREE() - free allocated blocks from radix tree meta info
    559        1.1      yamt  *
    560        1.1      yamt  *	This support routine frees a range of blocks from the bitmap.
    561        1.1      yamt  *	The range must be entirely enclosed by this radix node.  If a
    562        1.1      yamt  *	meta node, we break the range down recursively to free blocks
    563        1.1      yamt  *	in subnodes (which means that this code can free an arbitrary
    564        1.1      yamt  *	range whereas the allocation code cannot allocate an arbitrary
    565        1.1      yamt  *	range).
    566        1.1      yamt  */
    567        1.1      yamt 
    568        1.1      yamt static void
    569        1.1      yamt blst_meta_free(
    570        1.1      yamt 	blmeta_t *scan,
    571        1.5      yamt 	blist_blkno_t freeBlk,
    572        1.5      yamt 	blist_blkno_t count,
    573        1.5      yamt 	blist_blkno_t radix,
    574        1.5      yamt 	blist_blkno_t skip,
    575        1.5      yamt 	blist_blkno_t blk
    576        1.1      yamt ) {
    577        1.5      yamt 	blist_blkno_t i;
    578        1.5      yamt 	blist_blkno_t next_skip = (skip / BLIST_META_RADIX);
    579        1.1      yamt 
    580        1.1      yamt #if 0
    581        1.3      yamt 	printf("FREE (%" PRIx64 ",%" PRIu64
    582        1.3      yamt 	    ") FROM (%" PRIx64 ",%" PRIu64 ")\n",
    583        1.5      yamt 	    (uint64_t)freeBlk, (uint64_t)count,
    584        1.5      yamt 	    (uint64_t)blk, (uint64_t)radix
    585        1.1      yamt 	);
    586        1.1      yamt #endif
    587        1.1      yamt 
    588        1.1      yamt 	if (scan->u.bmu_avail == 0) {
    589        1.1      yamt 		/*
    590        1.1      yamt 		 * ALL-ALLOCATED special case, with possible
    591        1.1      yamt 		 * shortcut to ALL-FREE special case.
    592        1.1      yamt 		 */
    593        1.1      yamt 		scan->u.bmu_avail = count;
    594        1.1      yamt 		scan->bm_bighint = count;
    595        1.1      yamt 
    596        1.1      yamt 		if (count != radix)  {
    597        1.1      yamt 			for (i = 1; i <= skip; i += next_skip) {
    598        1.5      yamt 				if (scan[i].bm_bighint == (blist_blkno_t)-1)
    599        1.1      yamt 					break;
    600        1.1      yamt 				scan[i].bm_bighint = 0;
    601        1.1      yamt 				if (next_skip == 1) {
    602        1.1      yamt 					scan[i].u.bmu_bitmap = 0;
    603        1.1      yamt 				} else {
    604        1.1      yamt 					scan[i].u.bmu_avail = 0;
    605        1.1      yamt 				}
    606        1.1      yamt 			}
    607        1.1      yamt 			/* fall through */
    608        1.1      yamt 		}
    609        1.1      yamt 	} else {
    610        1.1      yamt 		scan->u.bmu_avail += count;
    611        1.1      yamt 		/* scan->bm_bighint = radix; */
    612        1.1      yamt 	}
    613        1.1      yamt 
    614        1.1      yamt 	/*
    615        1.1      yamt 	 * ALL-FREE special case.
    616        1.1      yamt 	 */
    617        1.1      yamt 
    618        1.1      yamt 	if (scan->u.bmu_avail == radix)
    619        1.1      yamt 		return;
    620        1.1      yamt 	if (scan->u.bmu_avail > radix)
    621        1.3      yamt 		panic("blst_meta_free: freeing already free blocks (%"
    622        1.3      yamt 		    PRIu64 ") %" PRIu64 "/%" PRIu64,
    623        1.5      yamt 		    (uint64_t)count,
    624        1.5      yamt 		    (uint64_t)scan->u.bmu_avail,
    625        1.5      yamt 		    (uint64_t)radix);
    626        1.1      yamt 
    627        1.1      yamt 	/*
    628        1.1      yamt 	 * Break the free down into its components
    629        1.1      yamt 	 */
    630        1.1      yamt 
    631        1.1      yamt 	radix /= BLIST_META_RADIX;
    632        1.1      yamt 
    633        1.1      yamt 	i = (freeBlk - blk) / radix;
    634        1.1      yamt 	blk += i * radix;
    635        1.1      yamt 	i = i * next_skip + 1;
    636        1.1      yamt 
    637        1.1      yamt 	while (i <= skip && blk < freeBlk + count) {
    638        1.5      yamt 		blist_blkno_t v;
    639        1.1      yamt 
    640        1.1      yamt 		v = blk + radix - freeBlk;
    641        1.1      yamt 		if (v > count)
    642        1.1      yamt 			v = count;
    643        1.1      yamt 
    644        1.5      yamt 		if (scan->bm_bighint == (blist_blkno_t)-1)
    645        1.1      yamt 			panic("blst_meta_free: freeing unexpected range");
    646        1.1      yamt 
    647        1.1      yamt 		if (next_skip == 1) {
    648        1.1      yamt 			blst_leaf_free(&scan[i], freeBlk, v);
    649        1.1      yamt 		} else {
    650        1.1      yamt 			blst_meta_free(&scan[i], freeBlk, v, radix, next_skip - 1, blk);
    651        1.1      yamt 		}
    652        1.1      yamt 		if (scan->bm_bighint < scan[i].bm_bighint)
    653        1.1      yamt 		    scan->bm_bighint = scan[i].bm_bighint;
    654        1.1      yamt 		count -= v;
    655        1.1      yamt 		freeBlk += v;
    656        1.1      yamt 		blk += radix;
    657        1.1      yamt 		i += next_skip;
    658        1.1      yamt 	}
    659        1.1      yamt }
    660        1.1      yamt 
    661        1.1      yamt /*
    662        1.1      yamt  * BLIST_RADIX_COPY() - copy one radix tree to another
    663        1.1      yamt  *
    664        1.1      yamt  *	Locates free space in the source tree and frees it in the destination
    665        1.1      yamt  *	tree.  The space may not already be free in the destination.
    666        1.1      yamt  */
    667        1.1      yamt 
    668        1.1      yamt static void blst_copy(
    669        1.1      yamt 	blmeta_t *scan,
    670        1.5      yamt 	blist_blkno_t blk,
    671        1.5      yamt 	blist_blkno_t radix,
    672        1.5      yamt 	blist_blkno_t skip,
    673        1.1      yamt 	blist_t dest,
    674        1.5      yamt 	blist_blkno_t count
    675        1.1      yamt ) {
    676        1.5      yamt 	blist_blkno_t next_skip;
    677        1.5      yamt 	blist_blkno_t i;
    678        1.1      yamt 
    679        1.1      yamt 	/*
    680        1.1      yamt 	 * Leaf node
    681        1.1      yamt 	 */
    682        1.1      yamt 
    683        1.1      yamt 	if (radix == BLIST_BMAP_RADIX) {
    684        1.5      yamt 		blist_bitmap_t v = scan->u.bmu_bitmap;
    685        1.1      yamt 
    686        1.5      yamt 		if (v == (blist_bitmap_t)-1) {
    687        1.1      yamt 			blist_free(dest, blk, count);
    688        1.1      yamt 		} else if (v != 0) {
    689        1.7  christos 			int j;
    690        1.1      yamt 
    691        1.7  christos 			for (j = 0; j < BLIST_BMAP_RADIX && j < count; ++j) {
    692        1.7  christos 				if (v & (1 << j))
    693        1.7  christos 					blist_free(dest, blk + j, 1);
    694        1.1      yamt 			}
    695        1.1      yamt 		}
    696        1.1      yamt 		return;
    697        1.1      yamt 	}
    698        1.1      yamt 
    699        1.1      yamt 	/*
    700        1.1      yamt 	 * Meta node
    701        1.1      yamt 	 */
    702        1.1      yamt 
    703        1.1      yamt 	if (scan->u.bmu_avail == 0) {
    704        1.1      yamt 		/*
    705        1.1      yamt 		 * Source all allocated, leave dest allocated
    706        1.1      yamt 		 */
    707        1.1      yamt 		return;
    708        1.1      yamt 	}
    709        1.1      yamt 	if (scan->u.bmu_avail == radix) {
    710        1.1      yamt 		/*
    711        1.1      yamt 		 * Source all free, free entire dest
    712        1.1      yamt 		 */
    713        1.1      yamt 		if (count < radix)
    714        1.1      yamt 			blist_free(dest, blk, count);
    715        1.1      yamt 		else
    716        1.1      yamt 			blist_free(dest, blk, radix);
    717        1.1      yamt 		return;
    718        1.1      yamt 	}
    719        1.1      yamt 
    720        1.1      yamt 
    721        1.1      yamt 	radix /= BLIST_META_RADIX;
    722        1.5      yamt 	next_skip = (skip / BLIST_META_RADIX);
    723        1.1      yamt 
    724        1.1      yamt 	for (i = 1; count && i <= skip; i += next_skip) {
    725        1.5      yamt 		if (scan[i].bm_bighint == (blist_blkno_t)-1)
    726        1.1      yamt 			break;
    727        1.1      yamt 
    728        1.1      yamt 		if (count >= radix) {
    729        1.1      yamt 			blst_copy(
    730        1.1      yamt 			    &scan[i],
    731        1.1      yamt 			    blk,
    732        1.1      yamt 			    radix,
    733        1.1      yamt 			    next_skip - 1,
    734        1.1      yamt 			    dest,
    735        1.1      yamt 			    radix
    736        1.1      yamt 			);
    737        1.1      yamt 			count -= radix;
    738        1.1      yamt 		} else {
    739        1.1      yamt 			if (count) {
    740        1.1      yamt 				blst_copy(
    741        1.1      yamt 				    &scan[i],
    742        1.1      yamt 				    blk,
    743        1.1      yamt 				    radix,
    744        1.1      yamt 				    next_skip - 1,
    745        1.1      yamt 				    dest,
    746        1.1      yamt 				    count
    747        1.1      yamt 				);
    748        1.1      yamt 			}
    749        1.1      yamt 			count = 0;
    750        1.1      yamt 		}
    751        1.1      yamt 		blk += radix;
    752        1.1      yamt 	}
    753        1.1      yamt }
    754        1.1      yamt 
    755        1.1      yamt /*
    756        1.1      yamt  * BLST_LEAF_FILL() -	allocate specific blocks in leaf bitmap
    757        1.1      yamt  *
    758        1.1      yamt  *	This routine allocates all blocks in the specified range
    759        1.1      yamt  *	regardless of any existing allocations in that range.  Returns
    760        1.1      yamt  *	the number of blocks allocated by the call.
    761        1.1      yamt  */
    762        1.1      yamt 
    763        1.1      yamt static int
    764        1.5      yamt blst_leaf_fill(blmeta_t *scan, blist_blkno_t blk, int count)
    765        1.1      yamt {
    766        1.1      yamt 	int n = blk & (BLIST_BMAP_RADIX - 1);
    767        1.1      yamt 	int nblks;
    768        1.5      yamt 	blist_bitmap_t mask, bitmap;
    769        1.1      yamt 
    770        1.5      yamt 	mask = ((blist_bitmap_t)-1 << n) &
    771        1.5      yamt 	    ((blist_bitmap_t)-1 >> (BLIST_BMAP_RADIX - count - n));
    772        1.1      yamt 
    773        1.1      yamt 	/* Count the number of blocks we're about to allocate */
    774        1.1      yamt 	bitmap = scan->u.bmu_bitmap & mask;
    775        1.1      yamt 	for (nblks = 0; bitmap != 0; nblks++)
    776        1.1      yamt 		bitmap &= bitmap - 1;
    777        1.1      yamt 
    778        1.1      yamt 	scan->u.bmu_bitmap &= ~mask;
    779        1.1      yamt 	return nblks;
    780        1.1      yamt }
    781        1.1      yamt 
    782        1.1      yamt /*
    783        1.1      yamt  * BLIST_META_FILL() -	allocate specific blocks at a meta node
    784        1.1      yamt  *
    785        1.1      yamt  *	This routine allocates the specified range of blocks,
    786        1.1      yamt  *	regardless of any existing allocations in the range.  The
    787        1.1      yamt  *	range must be within the extent of this node.  Returns the
    788        1.1      yamt  *	number of blocks allocated by the call.
    789        1.1      yamt  */
    790        1.5      yamt static blist_blkno_t
    791        1.1      yamt blst_meta_fill(
    792        1.1      yamt 	blmeta_t *scan,
    793        1.5      yamt 	blist_blkno_t allocBlk,
    794        1.5      yamt 	blist_blkno_t count,
    795        1.5      yamt 	blist_blkno_t radix,
    796        1.5      yamt 	blist_blkno_t skip,
    797        1.5      yamt 	blist_blkno_t blk
    798        1.1      yamt ) {
    799        1.5      yamt 	blist_blkno_t i;
    800        1.5      yamt 	blist_blkno_t next_skip = (skip / BLIST_META_RADIX);
    801        1.5      yamt 	blist_blkno_t nblks = 0;
    802        1.1      yamt 
    803        1.1      yamt 	if (count == radix || scan->u.bmu_avail == 0)  {
    804        1.1      yamt 		/*
    805        1.1      yamt 		 * ALL-ALLOCATED special case
    806        1.1      yamt 		 */
    807        1.1      yamt 		nblks = scan->u.bmu_avail;
    808        1.1      yamt 		scan->u.bmu_avail = 0;
    809        1.1      yamt 		scan->bm_bighint = count;
    810        1.1      yamt 		return nblks;
    811        1.1      yamt 	}
    812        1.1      yamt 
    813        1.9      yamt 	if (count > radix)
    814        1.9      yamt 		panic("blist_meta_fill: allocation too large");
    815        1.9      yamt 
    816        1.1      yamt 	if (scan->u.bmu_avail == radix) {
    817        1.1      yamt 		radix /= BLIST_META_RADIX;
    818        1.1      yamt 
    819        1.1      yamt 		/*
    820        1.1      yamt 		 * ALL-FREE special case, initialize sublevel
    821        1.1      yamt 		 */
    822        1.1      yamt 		for (i = 1; i <= skip; i += next_skip) {
    823        1.5      yamt 			if (scan[i].bm_bighint == (blist_blkno_t)-1)
    824        1.1      yamt 				break;
    825        1.1      yamt 			if (next_skip == 1) {
    826        1.5      yamt 				scan[i].u.bmu_bitmap = (blist_bitmap_t)-1;
    827        1.1      yamt 				scan[i].bm_bighint = BLIST_BMAP_RADIX;
    828        1.1      yamt 			} else {
    829        1.1      yamt 				scan[i].bm_bighint = radix;
    830        1.1      yamt 				scan[i].u.bmu_avail = radix;
    831        1.1      yamt 			}
    832        1.1      yamt 		}
    833        1.1      yamt 	} else {
    834        1.1      yamt 		radix /= BLIST_META_RADIX;
    835        1.1      yamt 	}
    836        1.1      yamt 
    837        1.1      yamt 	i = (allocBlk - blk) / radix;
    838        1.1      yamt 	blk += i * radix;
    839        1.1      yamt 	i = i * next_skip + 1;
    840        1.1      yamt 
    841        1.1      yamt 	while (i <= skip && blk < allocBlk + count) {
    842        1.5      yamt 		blist_blkno_t v;
    843        1.1      yamt 
    844        1.1      yamt 		v = blk + radix - allocBlk;
    845        1.1      yamt 		if (v > count)
    846        1.1      yamt 			v = count;
    847        1.1      yamt 
    848        1.5      yamt 		if (scan->bm_bighint == (blist_blkno_t)-1)
    849        1.1      yamt 			panic("blst_meta_fill: filling unexpected range");
    850        1.1      yamt 
    851        1.1      yamt 		if (next_skip == 1) {
    852        1.1      yamt 			nblks += blst_leaf_fill(&scan[i], allocBlk, v);
    853        1.1      yamt 		} else {
    854        1.1      yamt 			nblks += blst_meta_fill(&scan[i], allocBlk, v,
    855        1.1      yamt 			    radix, next_skip - 1, blk);
    856        1.1      yamt 		}
    857        1.1      yamt 		count -= v;
    858        1.1      yamt 		allocBlk += v;
    859        1.1      yamt 		blk += radix;
    860        1.1      yamt 		i += next_skip;
    861        1.1      yamt 	}
    862        1.1      yamt 	scan->u.bmu_avail -= nblks;
    863        1.1      yamt 	return nblks;
    864        1.1      yamt }
    865        1.1      yamt 
    866        1.1      yamt /*
    867        1.1      yamt  * BLST_RADIX_INIT() - initialize radix tree
    868        1.1      yamt  *
    869        1.1      yamt  *	Initialize our meta structures and bitmaps and calculate the exact
    870        1.1      yamt  *	amount of space required to manage 'count' blocks - this space may
    871       1.12       wiz  *	be considerably less than the calculated radix due to the large
    872        1.1      yamt  *	RADIX values we use.
    873        1.1      yamt  */
    874        1.1      yamt 
    875        1.5      yamt static blist_blkno_t
    876        1.5      yamt blst_radix_init(blmeta_t *scan, blist_blkno_t radix, blist_blkno_t skip,
    877        1.5      yamt     blist_blkno_t count)
    878        1.1      yamt {
    879        1.5      yamt 	blist_blkno_t i;
    880        1.5      yamt 	blist_blkno_t next_skip;
    881        1.5      yamt 	blist_blkno_t memindex = 0;
    882        1.1      yamt 
    883        1.1      yamt 	/*
    884        1.1      yamt 	 * Leaf node
    885        1.1      yamt 	 */
    886        1.1      yamt 
    887        1.1      yamt 	if (radix == BLIST_BMAP_RADIX) {
    888        1.1      yamt 		if (scan) {
    889        1.1      yamt 			scan->bm_bighint = 0;
    890        1.1      yamt 			scan->u.bmu_bitmap = 0;
    891        1.1      yamt 		}
    892        1.1      yamt 		return(memindex);
    893        1.1      yamt 	}
    894        1.1      yamt 
    895        1.1      yamt 	/*
    896        1.1      yamt 	 * Meta node.  If allocating the entire object we can special
    897        1.1      yamt 	 * case it.  However, we need to figure out how much memory
    898        1.1      yamt 	 * is required to manage 'count' blocks, so we continue on anyway.
    899        1.1      yamt 	 */
    900        1.1      yamt 
    901        1.1      yamt 	if (scan) {
    902        1.1      yamt 		scan->bm_bighint = 0;
    903        1.1      yamt 		scan->u.bmu_avail = 0;
    904        1.1      yamt 	}
    905        1.1      yamt 
    906        1.1      yamt 	radix /= BLIST_META_RADIX;
    907        1.5      yamt 	next_skip = (skip / BLIST_META_RADIX);
    908        1.1      yamt 
    909        1.1      yamt 	for (i = 1; i <= skip; i += next_skip) {
    910        1.1      yamt 		if (count >= radix) {
    911        1.1      yamt 			/*
    912        1.1      yamt 			 * Allocate the entire object
    913        1.1      yamt 			 */
    914        1.1      yamt 			memindex = i + blst_radix_init(
    915        1.1      yamt 			    ((scan) ? &scan[i] : NULL),
    916        1.1      yamt 			    radix,
    917        1.1      yamt 			    next_skip - 1,
    918        1.1      yamt 			    radix
    919        1.1      yamt 			);
    920        1.1      yamt 			count -= radix;
    921        1.1      yamt 		} else if (count > 0) {
    922        1.1      yamt 			/*
    923        1.1      yamt 			 * Allocate a partial object
    924        1.1      yamt 			 */
    925        1.1      yamt 			memindex = i + blst_radix_init(
    926        1.1      yamt 			    ((scan) ? &scan[i] : NULL),
    927        1.1      yamt 			    radix,
    928        1.1      yamt 			    next_skip - 1,
    929        1.1      yamt 			    count
    930        1.1      yamt 			);
    931        1.1      yamt 			count = 0;
    932        1.1      yamt 		} else {
    933        1.1      yamt 			/*
    934        1.1      yamt 			 * Add terminator and break out
    935        1.1      yamt 			 */
    936        1.1      yamt 			if (scan)
    937        1.5      yamt 				scan[i].bm_bighint = (blist_blkno_t)-1;
    938        1.1      yamt 			break;
    939        1.1      yamt 		}
    940        1.1      yamt 	}
    941        1.1      yamt 	if (memindex < i)
    942        1.1      yamt 		memindex = i;
    943        1.1      yamt 	return(memindex);
    944        1.1      yamt }
    945        1.1      yamt 
    946        1.1      yamt #ifdef BLIST_DEBUG
    947        1.1      yamt 
    948        1.1      yamt static void
    949        1.5      yamt blst_radix_print(blmeta_t *scan, blist_blkno_t blk, blist_blkno_t radix,
    950        1.5      yamt     blist_blkno_t skip, int tab)
    951        1.1      yamt {
    952        1.5      yamt 	blist_blkno_t i;
    953        1.5      yamt 	blist_blkno_t next_skip;
    954        1.1      yamt 	int lastState = 0;
    955        1.1      yamt 
    956        1.1      yamt 	if (radix == BLIST_BMAP_RADIX) {
    957        1.1      yamt 		printf(
    958        1.5      yamt 		    "%*.*s(%0*" PRIx64 ",%" PRIu64
    959        1.5      yamt 		    "): bitmap %0*" PRIx64 " big=%" PRIu64 "\n",
    960        1.1      yamt 		    tab, tab, "",
    961        1.5      yamt 		    sizeof(blk) * 2,
    962        1.5      yamt 		    (uint64_t)blk,
    963        1.5      yamt 		    (uint64_t)radix,
    964        1.5      yamt 		    sizeof(scan->u.bmu_bitmap) * 2,
    965        1.5      yamt 		    (uint64_t)scan->u.bmu_bitmap,
    966        1.5      yamt 		    (uint64_t)scan->bm_bighint
    967        1.1      yamt 		);
    968        1.1      yamt 		return;
    969        1.1      yamt 	}
    970        1.1      yamt 
    971        1.1      yamt 	if (scan->u.bmu_avail == 0) {
    972        1.1      yamt 		printf(
    973        1.5      yamt 		    "%*.*s(%0*" PRIx64 ",%" PRIu64") ALL ALLOCATED\n",
    974        1.5      yamt 		    tab, tab, "",
    975        1.5      yamt 		    sizeof(blk) * 2,
    976        1.5      yamt 		    (uint64_t)blk,
    977        1.5      yamt 		    (uint64_t)radix
    978        1.1      yamt 		);
    979        1.1      yamt 		return;
    980        1.1      yamt 	}
    981        1.1      yamt 	if (scan->u.bmu_avail == radix) {
    982        1.1      yamt 		printf(
    983        1.5      yamt 		    "%*.*s(%0*" PRIx64 ",%" PRIu64 ") ALL FREE\n",
    984        1.5      yamt 		    tab, tab, "",
    985        1.5      yamt 		    sizeof(blk) * 2,
    986        1.5      yamt 		    (uint64_t)blk,
    987        1.5      yamt 		    (uint64_t)radix
    988        1.1      yamt 		);
    989        1.1      yamt 		return;
    990        1.1      yamt 	}
    991        1.1      yamt 
    992        1.1      yamt 	printf(
    993        1.5      yamt 	    "%*.*s(%0*" PRIx64 ",%" PRIu64 "): subtree (%" PRIu64 "/%"
    994        1.3      yamt 	    PRIu64 ") big=%" PRIu64 " {\n",
    995        1.1      yamt 	    tab, tab, "",
    996        1.5      yamt 	    sizeof(blk) * 2,
    997        1.5      yamt 	    (uint64_t)blk,
    998        1.5      yamt 	    (uint64_t)radix,
    999        1.5      yamt 	    (uint64_t)scan->u.bmu_avail,
   1000        1.5      yamt 	    (uint64_t)radix,
   1001        1.5      yamt 	    (uint64_t)scan->bm_bighint
   1002        1.1      yamt 	);
   1003        1.1      yamt 
   1004        1.1      yamt 	radix /= BLIST_META_RADIX;
   1005        1.5      yamt 	next_skip = (skip / BLIST_META_RADIX);
   1006        1.1      yamt 	tab += 4;
   1007        1.1      yamt 
   1008        1.1      yamt 	for (i = 1; i <= skip; i += next_skip) {
   1009        1.5      yamt 		if (scan[i].bm_bighint == (blist_blkno_t)-1) {
   1010        1.1      yamt 			printf(
   1011        1.5      yamt 			    "%*.*s(%0*" PRIx64 ",%" PRIu64 "): Terminator\n",
   1012        1.1      yamt 			    tab, tab, "",
   1013        1.5      yamt 			    sizeof(blk) * 2,
   1014        1.5      yamt 			    (uint64_t)blk,
   1015        1.5      yamt 			    (uint64_t)radix
   1016        1.1      yamt 			);
   1017        1.1      yamt 			lastState = 0;
   1018        1.1      yamt 			break;
   1019        1.1      yamt 		}
   1020        1.1      yamt 		blst_radix_print(
   1021        1.1      yamt 		    &scan[i],
   1022        1.1      yamt 		    blk,
   1023        1.1      yamt 		    radix,
   1024        1.1      yamt 		    next_skip - 1,
   1025        1.1      yamt 		    tab
   1026        1.1      yamt 		);
   1027        1.1      yamt 		blk += radix;
   1028        1.1      yamt 	}
   1029        1.1      yamt 	tab -= 4;
   1030        1.1      yamt 
   1031        1.1      yamt 	printf(
   1032        1.1      yamt 	    "%*.*s}\n",
   1033        1.1      yamt 	    tab, tab, ""
   1034        1.1      yamt 	);
   1035        1.1      yamt }
   1036        1.1      yamt 
   1037        1.1      yamt #endif
   1038        1.1      yamt 
   1039        1.1      yamt #ifdef BLIST_DEBUG
   1040        1.1      yamt 
   1041        1.1      yamt int
   1042        1.1      yamt main(int ac, char **av)
   1043        1.1      yamt {
   1044        1.5      yamt 	blist_blkno_t size = 1024;
   1045        1.1      yamt 	int i;
   1046        1.1      yamt 	blist_t bl;
   1047        1.1      yamt 
   1048        1.1      yamt 	for (i = 1; i < ac; ++i) {
   1049        1.1      yamt 		const char *ptr = av[i];
   1050        1.1      yamt 		if (*ptr != '-') {
   1051        1.1      yamt 			size = strtol(ptr, NULL, 0);
   1052        1.1      yamt 			continue;
   1053        1.1      yamt 		}
   1054        1.1      yamt 		ptr += 2;
   1055        1.1      yamt 		fprintf(stderr, "Bad option: %s\n", ptr - 2);
   1056        1.1      yamt 		exit(1);
   1057        1.1      yamt 	}
   1058        1.1      yamt 	bl = blist_create(size);
   1059        1.1      yamt 	blist_free(bl, 0, size);
   1060        1.1      yamt 
   1061        1.1      yamt 	for (;;) {
   1062        1.1      yamt 		char buf[1024];
   1063        1.3      yamt 		uint64_t da = 0;
   1064        1.3      yamt 		uint64_t count = 0;
   1065        1.1      yamt 
   1066        1.3      yamt 		printf("%" PRIu64 "/%" PRIu64 "/%" PRIu64 "> ",
   1067        1.5      yamt 		    (uint64_t)bl->bl_free,
   1068        1.5      yamt 		    (uint64_t)size,
   1069        1.5      yamt 		    (uint64_t)bl->bl_radix);
   1070        1.1      yamt 		fflush(stdout);
   1071        1.1      yamt 		if (fgets(buf, sizeof(buf), stdin) == NULL)
   1072        1.1      yamt 			break;
   1073        1.1      yamt 		switch(buf[0]) {
   1074        1.1      yamt 		case 'r':
   1075        1.3      yamt 			if (sscanf(buf + 1, "%" SCNu64, &count) == 1) {
   1076        1.1      yamt 				blist_resize(&bl, count, 1);
   1077        1.1      yamt 			} else {
   1078        1.1      yamt 				printf("?\n");
   1079        1.1      yamt 			}
   1080        1.1      yamt 		case 'p':
   1081        1.1      yamt 			blist_print(bl);
   1082        1.1      yamt 			break;
   1083        1.1      yamt 		case 'a':
   1084        1.3      yamt 			if (sscanf(buf + 1, "%" SCNu64, &count) == 1) {
   1085        1.5      yamt 				blist_blkno_t blk = blist_alloc(bl, count);
   1086        1.5      yamt 				printf("    R=%0*" PRIx64 "\n",
   1087        1.5      yamt 				    sizeof(blk) * 2,
   1088        1.5      yamt 				    (uint64_t)blk);
   1089        1.1      yamt 			} else {
   1090        1.1      yamt 				printf("?\n");
   1091        1.1      yamt 			}
   1092        1.1      yamt 			break;
   1093        1.1      yamt 		case 'f':
   1094        1.3      yamt 			if (sscanf(buf + 1, "%" SCNx64 " %" SCNu64,
   1095        1.3      yamt 			    &da, &count) == 2) {
   1096        1.1      yamt 				blist_free(bl, da, count);
   1097        1.1      yamt 			} else {
   1098        1.1      yamt 				printf("?\n");
   1099        1.1      yamt 			}
   1100        1.1      yamt 			break;
   1101        1.1      yamt 		case 'l':
   1102        1.3      yamt 			if (sscanf(buf + 1, "%" SCNx64 " %" SCNu64,
   1103        1.3      yamt 			    &da, &count) == 2) {
   1104        1.5      yamt 				printf("    n=%" PRIu64 "\n",
   1105        1.5      yamt 				    (uint64_t)blist_fill(bl, da, count));
   1106        1.1      yamt 			} else {
   1107        1.1      yamt 				printf("?\n");
   1108        1.1      yamt 			}
   1109        1.1      yamt 			break;
   1110        1.1      yamt 		case '?':
   1111        1.1      yamt 		case 'h':
   1112        1.1      yamt 			puts(
   1113        1.1      yamt 			    "p          -print\n"
   1114        1.1      yamt 			    "a %d       -allocate\n"
   1115        1.1      yamt 			    "f %x %d    -free\n"
   1116        1.1      yamt 			    "l %x %d    -fill\n"
   1117        1.1      yamt 			    "r %d       -resize\n"
   1118        1.1      yamt 			    "h/?        -help"
   1119        1.1      yamt 			);
   1120        1.1      yamt 			break;
   1121        1.1      yamt 		default:
   1122        1.1      yamt 			printf("?\n");
   1123        1.1      yamt 			break;
   1124        1.1      yamt 		}
   1125        1.1      yamt 	}
   1126        1.1      yamt 	return(0);
   1127        1.1      yamt }
   1128        1.1      yamt 
   1129        1.1      yamt void
   1130        1.1      yamt panic(const char *ctl, ...)
   1131        1.1      yamt {
   1132        1.1      yamt 	va_list va;
   1133        1.1      yamt 
   1134        1.1      yamt 	va_start(va, ctl);
   1135        1.1      yamt 	vfprintf(stderr, ctl, va);
   1136        1.1      yamt 	fprintf(stderr, "\n");
   1137        1.1      yamt 	va_end(va);
   1138        1.1      yamt 	exit(1);
   1139        1.1      yamt }
   1140        1.1      yamt 
   1141        1.1      yamt #endif
   1142        1.1      yamt 
   1143