Home | History | Annotate | Line # | Download | only in dev
ld.c revision 1.116
      1 /*	$NetBSD: ld.c,v 1.116 2025/04/13 02:34:02 rin Exp $	*/
      2 
      3 /*-
      4  * Copyright (c) 1998, 2000 The NetBSD Foundation, Inc.
      5  * All rights reserved.
      6  *
      7  * This code is derived from software contributed to The NetBSD Foundation
      8  * by Andrew Doran and Charles M. Hannum.
      9  *
     10  * Redistribution and use in source and binary forms, with or without
     11  * modification, are permitted provided that the following conditions
     12  * are met:
     13  * 1. Redistributions of source code must retain the above copyright
     14  *    notice, this list of conditions and the following disclaimer.
     15  * 2. Redistributions in binary form must reproduce the above copyright
     16  *    notice, this list of conditions and the following disclaimer in the
     17  *    documentation and/or other materials provided with the distribution.
     18  *
     19  * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
     20  * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
     21  * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
     22  * PURPOSE ARE DISCLAIMED.  IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
     23  * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
     24  * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
     25  * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
     26  * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
     27  * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
     28  * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
     29  * POSSIBILITY OF SUCH DAMAGE.
     30  */
     31 
     32 /*
     33  * Disk driver for use by RAID controllers.
     34  */
     35 
     36 #include <sys/cdefs.h>
     37 __KERNEL_RCSID(0, "$NetBSD: ld.c,v 1.116 2025/04/13 02:34:02 rin Exp $");
     38 
     39 #include <sys/param.h>
     40 #include <sys/systm.h>
     41 #include <sys/kernel.h>
     42 #include <sys/device.h>
     43 #include <sys/queue.h>
     44 #include <sys/proc.h>
     45 #include <sys/buf.h>
     46 #include <sys/bufq.h>
     47 #include <sys/endian.h>
     48 #include <sys/disklabel.h>
     49 #include <sys/disk.h>
     50 #include <sys/dkio.h>
     51 #include <sys/stat.h>
     52 #include <sys/conf.h>
     53 #include <sys/fcntl.h>
     54 #include <sys/vnode.h>
     55 #include <sys/syslog.h>
     56 #include <sys/mutex.h>
     57 #include <sys/module.h>
     58 #include <sys/reboot.h>
     59 
     60 #include <dev/ldvar.h>
     61 
     62 #include "ioconf.h"
     63 
     64 static void	ldminphys(struct buf *bp);
     65 static bool	ld_suspend(device_t, const pmf_qual_t *);
     66 static bool	ld_resume(device_t, const pmf_qual_t *);
     67 static bool	ld_shutdown(device_t, int);
     68 static int	ld_diskstart(device_t, struct buf *bp);
     69 static void	ld_iosize(device_t, int *);
     70 static int	ld_dumpblocks(device_t, void *, daddr_t, int);
     71 static void	ld_fake_geometry(struct ld_softc *);
     72 static void	ld_set_geometry(struct ld_softc *);
     73 static void	ld_config_interrupts (device_t);
     74 static int	ld_lastclose(device_t);
     75 static int	ld_discard(device_t, off_t, off_t);
     76 static int	ld_flush(device_t, bool);
     77 
     78 static dev_type_open(ldopen);
     79 static dev_type_close(ldclose);
     80 static dev_type_read(ldread);
     81 static dev_type_write(ldwrite);
     82 static dev_type_ioctl(ldioctl);
     83 static dev_type_strategy(ldstrategy);
     84 static dev_type_dump(lddump);
     85 static dev_type_size(ldsize);
     86 static dev_type_discard(lddiscard);
     87 
     88 const struct bdevsw ld_bdevsw = {
     89 	.d_open = ldopen,
     90 	.d_close = ldclose,
     91 	.d_strategy = ldstrategy,
     92 	.d_ioctl = ldioctl,
     93 	.d_dump = lddump,
     94 	.d_psize = ldsize,
     95 	.d_discard = lddiscard,
     96 	.d_flag = D_DISK | D_MPSAFE
     97 };
     98 
     99 const struct cdevsw ld_cdevsw = {
    100 	.d_open = ldopen,
    101 	.d_close = ldclose,
    102 	.d_read = ldread,
    103 	.d_write = ldwrite,
    104 	.d_ioctl = ldioctl,
    105 	.d_stop = nostop,
    106 	.d_tty = notty,
    107 	.d_poll = nopoll,
    108 	.d_mmap = nommap,
    109 	.d_kqfilter = nokqfilter,
    110 	.d_discard = lddiscard,
    111 	.d_flag = D_DISK | D_MPSAFE
    112 };
    113 
    114 static const struct	dkdriver lddkdriver = {
    115 	.d_open = ldopen,
    116 	.d_close = ldclose,
    117 	.d_strategy = ldstrategy,
    118 	.d_iosize = ld_iosize,
    119 	.d_minphys  = ldminphys,
    120 	.d_diskstart = ld_diskstart,
    121 	.d_dumpblocks = ld_dumpblocks,
    122 	.d_lastclose = ld_lastclose,
    123 	.d_discard = ld_discard
    124 };
    125 
    126 void
    127 ldattach(struct ld_softc *sc, const char *default_strategy)
    128 {
    129 	device_t self = sc->sc_dv;
    130 	struct dk_softc *dksc = &sc->sc_dksc;
    131 
    132 	mutex_init(&sc->sc_mutex, MUTEX_DEFAULT, IPL_VM);
    133 	cv_init(&sc->sc_drain, "lddrain");
    134 
    135 	if ((sc->sc_flags & LDF_ENABLED) == 0) {
    136 		return;
    137 	}
    138 
    139 	/* don't attach a disk that we cannot handle */
    140 	if (sc->sc_secsize < DEV_BSIZE) {
    141 		sc->sc_flags &= ~LDF_ENABLED;
    142 		return;
    143 	}
    144 
    145 	/* Initialise dk and disk structure. */
    146 	dk_init(dksc, self, DKTYPE_LD);
    147 	disk_init(&dksc->sc_dkdev, dksc->sc_xname, &lddkdriver);
    148 
    149 	if (sc->sc_maxxfer > MAXPHYS)
    150 		sc->sc_maxxfer = MAXPHYS;
    151 
    152 	/* Build synthetic geometry if necessary. */
    153 	if (sc->sc_nheads == 0 || sc->sc_nsectors == 0 ||
    154 	    sc->sc_ncylinders == 0)
    155 	    ld_fake_geometry(sc);
    156 
    157 	sc->sc_disksize512 = sc->sc_secperunit * sc->sc_secsize / DEV_BSIZE;
    158 
    159 	if (sc->sc_flags & LDF_NO_RND)
    160 		dksc->sc_flags |= DKF_NO_RND;
    161 
    162 	/* Attach dk and disk subsystems */
    163 	dk_attach(dksc);
    164 	disk_attach(&dksc->sc_dkdev);
    165 	ld_set_geometry(sc);
    166 
    167 	bufq_alloc(&dksc->sc_bufq, default_strategy, BUFQ_SORT_RAWBLOCK);
    168 
    169 	/* Register with PMF */
    170 	if (!pmf_device_register1(dksc->sc_dev, ld_suspend, ld_resume,
    171 		ld_shutdown))
    172 		aprint_error_dev(dksc->sc_dev,
    173 		    "couldn't establish power handler\n");
    174 
    175 	/* Discover wedges on this disk. */
    176 	config_interrupts(sc->sc_dv, ld_config_interrupts);
    177 }
    178 
    179 int
    180 ldadjqparam(struct ld_softc *sc, int xmax)
    181 {
    182 
    183 	mutex_enter(&sc->sc_mutex);
    184 	sc->sc_maxqueuecnt = xmax;
    185 	mutex_exit(&sc->sc_mutex);
    186 
    187 	return (0);
    188 }
    189 
    190 int
    191 ldbegindetach(struct ld_softc *sc, int flags)
    192 {
    193 	struct dk_softc *dksc = &sc->sc_dksc;
    194 	int error;
    195 
    196 	/* If we never attached properly, no problem with detaching.  */
    197 	if ((sc->sc_flags & LDF_ENABLED) == 0)
    198 		return 0;
    199 
    200 	/*
    201 	 * If the disk is still open, back out before we commit to
    202 	 * detaching.
    203 	 */
    204 	error = disk_begindetach(&dksc->sc_dkdev, ld_lastclose, dksc->sc_dev,
    205 	    flags);
    206 	if (error)
    207 		return error;
    208 
    209 	/* We are now committed to detaching.  Prevent new xfers.  */
    210 	ldadjqparam(sc, 0);
    211 
    212 	return 0;
    213 }
    214 
    215 void
    216 ldenddetach(struct ld_softc *sc)
    217 {
    218 	struct dk_softc *dksc = &sc->sc_dksc;
    219 	int bmaj, cmaj, i, mn;
    220 
    221 	if ((sc->sc_flags & LDF_ENABLED) == 0)
    222 		return;
    223 
    224 	/* Wait for commands queued with the hardware to complete. */
    225 	mutex_enter(&sc->sc_mutex);
    226 	while (sc->sc_queuecnt > 0) {
    227 		if (cv_timedwait(&sc->sc_drain, &sc->sc_mutex, 30 * hz)) {
    228 			/*
    229 			 * XXX This seems like a recipe for crashing on
    230 			 * use after free...
    231 			 */
    232 			printf("%s: not drained\n", dksc->sc_xname);
    233 			break;
    234 		}
    235 	}
    236 	mutex_exit(&sc->sc_mutex);
    237 
    238 	/* Kill off any queued buffers. */
    239 	dk_drain(dksc);
    240 	bufq_free(dksc->sc_bufq);
    241 
    242 	/* Locate the major numbers. */
    243 	bmaj = bdevsw_lookup_major(&ld_bdevsw);
    244 	cmaj = cdevsw_lookup_major(&ld_cdevsw);
    245 
    246 	/* Nuke the vnodes for any open instances. */
    247 	for (i = 0; i < MAXPARTITIONS; i++) {
    248 		mn = DISKMINOR(device_unit(dksc->sc_dev), i);
    249 		vdevgone(bmaj, mn, mn, VBLK);
    250 		vdevgone(cmaj, mn, mn, VCHR);
    251 	}
    252 
    253 	/* Delete all of our wedges. */
    254 	dkwedge_delall(&dksc->sc_dkdev);
    255 
    256 	/* Detach from the disk list. */
    257 	disk_detach(&dksc->sc_dkdev);
    258 	disk_destroy(&dksc->sc_dkdev);
    259 
    260 	dk_detach(dksc);
    261 
    262 	/* Deregister with PMF */
    263 	pmf_device_deregister(dksc->sc_dev);
    264 
    265 	/*
    266 	 * XXX We can't really flush the cache here, because the
    267 	 * XXX device may already be non-existent from the controller's
    268 	 * XXX perspective.
    269 	 */
    270 #if 0
    271 	ld_flush(dksc->sc_dev, false);
    272 #endif
    273 	cv_destroy(&sc->sc_drain);
    274 	mutex_destroy(&sc->sc_mutex);
    275 }
    276 
    277 /* ARGSUSED */
    278 static bool
    279 ld_suspend(device_t dev, const pmf_qual_t *qual)
    280 {
    281 	struct ld_softc *sc = device_private(dev);
    282 	int queuecnt;
    283 	bool ok = false;
    284 
    285 	/* Block new requests and wait for outstanding requests to drain.  */
    286 	mutex_enter(&sc->sc_mutex);
    287 	KASSERT((sc->sc_flags & LDF_SUSPEND) == 0);
    288 	sc->sc_flags |= LDF_SUSPEND;
    289 	while ((queuecnt = sc->sc_queuecnt) > 0) {
    290 		if (cv_timedwait(&sc->sc_drain, &sc->sc_mutex, 30 * hz))
    291 			break;
    292 	}
    293 	mutex_exit(&sc->sc_mutex);
    294 
    295 	/* Block suspend if we couldn't drain everything in 30sec.  */
    296 	if (queuecnt > 0) {
    297 		device_printf(dev, "timeout draining buffers\n");
    298 		goto out;
    299 	}
    300 
    301 	/* Flush cache before we lose power.  If we can't, block suspend.  */
    302 	if (ld_flush(dev, /*poll*/false) != 0) {
    303 		device_printf(dev, "failed to flush cache\n");
    304 		goto out;
    305 	}
    306 
    307 	/* Success!  */
    308 	ok = true;
    309 
    310 out:	if (!ok)
    311 		(void)ld_resume(dev, qual);
    312 	return ok;
    313 }
    314 
    315 static bool
    316 ld_resume(device_t dev, const pmf_qual_t *qual)
    317 {
    318 	struct ld_softc *sc = device_private(dev);
    319 
    320 	/* Allow new requests to come in.  */
    321 	mutex_enter(&sc->sc_mutex);
    322 	KASSERT(sc->sc_flags & LDF_SUSPEND);
    323 	sc->sc_flags &= ~LDF_SUSPEND;
    324 	mutex_exit(&sc->sc_mutex);
    325 
    326 	/* Restart any pending queued requests.  */
    327 	dk_start(&sc->sc_dksc, NULL);
    328 
    329 	return true;
    330 }
    331 
    332 /* ARGSUSED */
    333 static bool
    334 ld_shutdown(device_t dev, int flags)
    335 {
    336 	if ((flags & RB_NOSYNC) == 0 && ld_flush(dev, true) != 0)
    337 		return false;
    338 
    339 	return true;
    340 }
    341 
    342 /* ARGSUSED */
    343 static int
    344 ldopen(dev_t dev, int flags, int fmt, struct lwp *l)
    345 {
    346 	struct ld_softc *sc;
    347 	struct dk_softc *dksc;
    348 	int unit;
    349 
    350 	unit = DISKUNIT(dev);
    351 	if ((sc = device_lookup_private(&ld_cd, unit)) == NULL)
    352 		return (ENXIO);
    353 
    354 	if ((sc->sc_flags & LDF_ENABLED) == 0)
    355 		return (ENODEV);
    356 
    357 	dksc = &sc->sc_dksc;
    358 
    359 	return dk_open(dksc, dev, flags, fmt, l);
    360 }
    361 
    362 static int
    363 ld_lastclose(device_t self)
    364 {
    365 	ld_flush(self, false);
    366 
    367 	return 0;
    368 }
    369 
    370 /* ARGSUSED */
    371 static int
    372 ldclose(dev_t dev, int flags, int fmt, struct lwp *l)
    373 {
    374 	struct ld_softc *sc;
    375 	struct dk_softc *dksc;
    376 	int unit;
    377 
    378 	unit = DISKUNIT(dev);
    379 	sc = device_lookup_private(&ld_cd, unit);
    380 	dksc = &sc->sc_dksc;
    381 
    382 	return dk_close(dksc, dev, flags, fmt, l);
    383 }
    384 
    385 /* ARGSUSED */
    386 static int
    387 ldread(dev_t dev, struct uio *uio, int ioflag)
    388 {
    389 
    390 	return (physio(ldstrategy, NULL, dev, B_READ, ldminphys, uio));
    391 }
    392 
    393 /* ARGSUSED */
    394 static int
    395 ldwrite(dev_t dev, struct uio *uio, int ioflag)
    396 {
    397 
    398 	return (physio(ldstrategy, NULL, dev, B_WRITE, ldminphys, uio));
    399 }
    400 
    401 /* ARGSUSED */
    402 static int
    403 ldioctl(dev_t dev, u_long cmd, void *addr, int32_t flag, struct lwp *l)
    404 {
    405 	struct ld_softc *sc;
    406 	struct dk_softc *dksc;
    407 	int unit, part, error;
    408 
    409 	unit = DISKUNIT(dev);
    410 	part = DISKPART(dev);
    411 	sc = device_lookup_private(&ld_cd, unit);
    412 	dksc = &sc->sc_dksc;
    413 
    414 	error = 0;
    415 
    416 	/*
    417 	 * Some common checks so that individual attachments wouldn't need
    418 	 * to duplicate them.
    419 	 */
    420 	switch (cmd) {
    421 	case DIOCCACHESYNC:
    422 		/*
    423 		 * XXX Do we really need to care about having a writable
    424 		 * file descriptor here?
    425 		 */
    426 		if ((flag & FWRITE) == 0)
    427 			error = EBADF;
    428 		else
    429 			error = 0;
    430 		break;
    431 	case DIOCGSECTORALIGN: {
    432 		struct disk_sectoralign *dsa = addr;
    433 		dsa->dsa_alignment = sc->sc_physsecsize / sc->sc_secsize;
    434 		dsa->dsa_alignment = MAX(1, dsa->dsa_alignment);
    435 		dsa->dsa_firstaligned = sc->sc_alignedsec;
    436 		if (part != RAW_PART) {
    437 			struct disklabel *lp = dksc->sc_dkdev.dk_label;
    438 			daddr_t offset = lp->d_partitions[part].p_offset;
    439 			uint32_t r = offset % dsa->dsa_alignment;
    440 
    441 			if (r < dsa->dsa_firstaligned)
    442 				dsa->dsa_firstaligned = dsa->dsa_firstaligned
    443 				    - r;
    444 			else
    445 				dsa->dsa_firstaligned = (dsa->dsa_firstaligned
    446 				    + dsa->dsa_alignment) - r;
    447 		}
    448 		dsa->dsa_firstaligned %= dsa->dsa_alignment;
    449 		return 0;
    450 	}
    451 	}
    452 
    453 	if (error != 0)
    454 		return (error);
    455 
    456 	if (sc->sc_ioctl) {
    457 		if ((sc->sc_flags & LDF_MPSAFE) == 0)
    458 			KERNEL_LOCK(1, curlwp);
    459 		error = (*sc->sc_ioctl)(sc, cmd, addr, flag, 0);
    460 		if ((sc->sc_flags & LDF_MPSAFE) == 0)
    461 			KERNEL_UNLOCK_ONE(curlwp);
    462 		if (error != EPASSTHROUGH)
    463 			return (error);
    464 	}
    465 
    466 	/* something not handled by the attachment */
    467 	return dk_ioctl(dksc, dev, cmd, addr, flag, l);
    468 }
    469 
    470 /*
    471  * Flush the device's cache.
    472  */
    473 static int
    474 ld_flush(device_t self, bool poll)
    475 {
    476 	int error = 0;
    477 	struct ld_softc *sc = device_private(self);
    478 
    479 	if (sc->sc_ioctl) {
    480 		if ((sc->sc_flags & LDF_MPSAFE) == 0)
    481 			KERNEL_LOCK(1, curlwp);
    482 		error = (*sc->sc_ioctl)(sc, DIOCCACHESYNC, NULL, 0, poll);
    483 		if ((sc->sc_flags & LDF_MPSAFE) == 0)
    484 			KERNEL_UNLOCK_ONE(curlwp);
    485 		if (error != 0)
    486 			device_printf(self, "unable to flush cache\n");
    487 	}
    488 
    489 	return error;
    490 }
    491 
    492 static void
    493 ldstrategy(struct buf *bp)
    494 {
    495 	struct ld_softc *sc;
    496 	struct dk_softc *dksc;
    497 	int unit;
    498 
    499 	unit = DISKUNIT(bp->b_dev);
    500 	sc = device_lookup_private(&ld_cd, unit);
    501 	dksc = &sc->sc_dksc;
    502 
    503 	dk_strategy(dksc, bp);
    504 }
    505 
    506 static int
    507 ld_diskstart(device_t dev, struct buf *bp)
    508 {
    509 	struct ld_softc *sc = device_private(dev);
    510 	int error;
    511 
    512 	if (sc->sc_queuecnt >= sc->sc_maxqueuecnt ||
    513 	    sc->sc_flags & LDF_SUSPEND) {
    514 		if (sc->sc_flags & LDF_SUSPEND)
    515 			aprint_debug_dev(dev, "i/o blocked while suspended\n");
    516 		return EAGAIN;
    517 	}
    518 
    519 	if ((sc->sc_flags & LDF_MPSAFE) == 0)
    520 		KERNEL_LOCK(1, curlwp);
    521 
    522 	mutex_enter(&sc->sc_mutex);
    523 
    524 	if (sc->sc_queuecnt >= sc->sc_maxqueuecnt ||
    525 	    sc->sc_flags & LDF_SUSPEND) {
    526 		if (sc->sc_flags & LDF_SUSPEND)
    527 			aprint_debug_dev(dev, "i/o blocked while suspended\n");
    528 		error = EAGAIN;
    529 	} else {
    530 		error = (*sc->sc_start)(sc, bp);
    531 		if (error == 0)
    532 			sc->sc_queuecnt++;
    533 	}
    534 
    535 	mutex_exit(&sc->sc_mutex);
    536 
    537 	if ((sc->sc_flags & LDF_MPSAFE) == 0)
    538 		KERNEL_UNLOCK_ONE(curlwp);
    539 
    540 	return error;
    541 }
    542 
    543 void
    544 lddone(struct ld_softc *sc, struct buf *bp)
    545 {
    546 	struct dk_softc *dksc = &sc->sc_dksc;
    547 
    548 	dk_done(dksc, bp);
    549 
    550 	mutex_enter(&sc->sc_mutex);
    551 	if (--sc->sc_queuecnt <= sc->sc_maxqueuecnt) {
    552 		cv_broadcast(&sc->sc_drain);
    553 		mutex_exit(&sc->sc_mutex);
    554 		dk_start(dksc, NULL);
    555 	} else
    556 		mutex_exit(&sc->sc_mutex);
    557 }
    558 
    559 static int
    560 ldsize(dev_t dev)
    561 {
    562 	struct ld_softc *sc;
    563 	struct dk_softc *dksc;
    564 	int unit;
    565 
    566 	unit = DISKUNIT(dev);
    567 	if ((sc = device_lookup_private(&ld_cd, unit)) == NULL)
    568 		return (-1);
    569 	dksc = &sc->sc_dksc;
    570 
    571 	if ((sc->sc_flags & LDF_ENABLED) == 0)
    572 		return (-1);
    573 
    574 	return dk_size(dksc, dev);
    575 }
    576 
    577 /*
    578  * Take a dump.
    579  */
    580 static int
    581 lddump(dev_t dev, daddr_t blkno, void *va, size_t size)
    582 {
    583 	struct ld_softc *sc;
    584 	struct dk_softc *dksc;
    585 	int unit;
    586 
    587 	unit = DISKUNIT(dev);
    588 	if ((sc = device_lookup_private(&ld_cd, unit)) == NULL)
    589 		return (ENXIO);
    590 	dksc = &sc->sc_dksc;
    591 
    592 	if ((sc->sc_flags & LDF_ENABLED) == 0)
    593 		return (ENODEV);
    594 
    595 	return dk_dump(dksc, dev, blkno, va, size, 0);
    596 }
    597 
    598 static int
    599 ld_dumpblocks(device_t dev, void *va, daddr_t blkno, int nblk)
    600 {
    601 	struct ld_softc *sc = device_private(dev);
    602 
    603 	if (sc->sc_dump == NULL)
    604 		return (ENODEV);
    605 
    606 	/*
    607 	 * Minimum consistency check; sc_dump() should check
    608 	 * device-dependent constraints if necessary.
    609 	 */
    610 	if (blkno < 0)
    611 		return (EIO);
    612 
    613 	return (*sc->sc_dump)(sc, va, blkno, nblk);
    614 }
    615 
    616 /*
    617  * Adjust the size of a transfer.
    618  */
    619 static void
    620 ldminphys(struct buf *bp)
    621 {
    622 	int unit;
    623 	struct ld_softc *sc;
    624 
    625 	unit = DISKUNIT(bp->b_dev);
    626 	sc = device_lookup_private(&ld_cd, unit);
    627 
    628 	ld_iosize(sc->sc_dv, &bp->b_bcount);
    629 	minphys(bp);
    630 }
    631 
    632 static void
    633 ld_iosize(device_t d, int *countp)
    634 {
    635 	struct ld_softc *sc = device_private(d);
    636 
    637 	if (*countp > sc->sc_maxxfer)
    638 		*countp = sc->sc_maxxfer;
    639 }
    640 
    641 static void
    642 ld_fake_geometry(struct ld_softc *sc)
    643 {
    644 	uint64_t ncyl;
    645 
    646 	if (sc->sc_secperunit <= 528 * 2048)		/* 528MB */
    647 		sc->sc_nheads = 16;
    648 	else if (sc->sc_secperunit <= 1024 * 2048)	/* 1GB */
    649 		sc->sc_nheads = 32;
    650 	else if (sc->sc_secperunit <= 21504 * 2048)	/* 21GB */
    651 		sc->sc_nheads = 64;
    652 	else if (sc->sc_secperunit <= 43008 * 2048)	/* 42GB */
    653 		sc->sc_nheads = 128;
    654 	else
    655 		sc->sc_nheads = 255;
    656 
    657 	sc->sc_nsectors = 63;
    658 	sc->sc_ncylinders = INT_MAX;
    659 	ncyl = sc->sc_secperunit /
    660 	    (sc->sc_nheads * sc->sc_nsectors);
    661 	if (ncyl < INT_MAX)
    662 		sc->sc_ncylinders = (int)ncyl;
    663 }
    664 
    665 static void
    666 ld_set_geometry(struct ld_softc *sc)
    667 {
    668 	struct dk_softc *dksc = &sc->sc_dksc;
    669 	struct disk_geom *dg = &dksc->sc_dkdev.dk_geom;
    670 	char tbuf[9];
    671 
    672 	format_bytes(tbuf, sizeof(tbuf), sc->sc_secperunit *
    673 	    sc->sc_secsize);
    674 	aprint_normal_dev(dksc->sc_dev, "%s, %d cyl, %d head, %d sec, "
    675 	    "%d bytes/sect x %"PRIu64" sectors",
    676 	    tbuf, sc->sc_ncylinders, sc->sc_nheads,
    677 	    sc->sc_nsectors, sc->sc_secsize, sc->sc_secperunit);
    678 	if (sc->sc_physsecsize != sc->sc_secsize) {
    679 		aprint_normal(" (%d bytes/physsect", sc->sc_physsecsize);
    680 		if (sc->sc_alignedsec != 0)
    681 			aprint_normal("; first aligned sector %u",
    682 			    sc->sc_alignedsec);
    683 		aprint_normal(")");
    684 	}
    685 	aprint_normal("\n");
    686 
    687 	memset(dg, 0, sizeof(*dg));
    688 	dg->dg_secperunit = sc->sc_secperunit;
    689 	dg->dg_secsize = sc->sc_secsize;
    690 	dg->dg_nsectors = sc->sc_nsectors;
    691 	dg->dg_ntracks = sc->sc_nheads;
    692 	dg->dg_ncylinders = sc->sc_ncylinders;
    693 
    694 	disk_set_info(dksc->sc_dev, &dksc->sc_dkdev, sc->sc_typename);
    695 }
    696 
    697 static void
    698 ld_config_interrupts(device_t d)
    699 {
    700 	struct ld_softc *sc = device_private(d);
    701 	struct dk_softc *dksc = &sc->sc_dksc;
    702 
    703 	dkwedge_discover(&dksc->sc_dkdev);
    704 }
    705 
    706 static int
    707 ld_discard(device_t dev, off_t pos, off_t len)
    708 {
    709 	struct ld_softc *sc = device_private(dev);
    710 	struct buf dbuf, *bp = &dbuf;
    711 	int error = 0;
    712 
    713 	KASSERT(len <= INT_MAX);
    714 
    715 	if (sc->sc_discard == NULL)
    716 		return (ENODEV);
    717 
    718 	if ((sc->sc_flags & LDF_MPSAFE) == 0)
    719 		KERNEL_LOCK(1, curlwp);
    720 
    721 	buf_init(bp);
    722 	bp->b_vp = NULL;
    723 	bp->b_data = NULL;
    724 	bp->b_bufsize = 0;
    725 	bp->b_rawblkno = pos / sc->sc_secsize;
    726 	bp->b_bcount = len;
    727 	bp->b_flags = B_WRITE;
    728 	bp->b_cflags = BC_BUSY;
    729 
    730 	error = (*sc->sc_discard)(sc, bp);
    731 	if (error == 0)
    732 		error = biowait(bp);
    733 
    734 	buf_destroy(bp);
    735 
    736 	if ((sc->sc_flags & LDF_MPSAFE) == 0)
    737 		KERNEL_UNLOCK_ONE(curlwp);
    738 
    739 	return error;
    740 }
    741 
    742 void
    743 lddiscardend(struct ld_softc *sc, struct buf *bp)
    744 {
    745 
    746 	if (bp->b_error)
    747 		bp->b_resid = bp->b_bcount;
    748 	biodone(bp);
    749 }
    750 
    751 static int
    752 lddiscard(dev_t dev, off_t pos, off_t len)
    753 {
    754 	struct ld_softc *sc;
    755 	struct dk_softc *dksc;
    756 	int unit;
    757 
    758 	unit = DISKUNIT(dev);
    759 	sc = device_lookup_private(&ld_cd, unit);
    760 	dksc = &sc->sc_dksc;
    761 
    762 	return dk_discard(dksc, dev, pos, len);
    763 }
    764 
    765 MODULE(MODULE_CLASS_DRIVER, ld, "dk_subr");
    766 
    767 #ifdef _MODULE
    768 CFDRIVER_DECL(ld, DV_DISK, NULL);
    769 #endif
    770 
    771 static int
    772 ld_modcmd(modcmd_t cmd, void *opaque)
    773 {
    774 #ifdef _MODULE
    775 	devmajor_t bmajor, cmajor;
    776 #endif
    777 	int error = 0;
    778 
    779 #ifdef _MODULE
    780 	switch (cmd) {
    781 	case MODULE_CMD_INIT:
    782 		bmajor = cmajor = -1;
    783 		error = devsw_attach(ld_cd.cd_name, &ld_bdevsw, &bmajor,
    784 		    &ld_cdevsw, &cmajor);
    785 		if (error)
    786 			break;
    787 		error = config_cfdriver_attach(&ld_cd);
    788 		break;
    789 	case MODULE_CMD_FINI:
    790 		error = config_cfdriver_detach(&ld_cd);
    791 		if (error)
    792 			break;
    793 		devsw_detach(&ld_bdevsw, &ld_cdevsw);
    794 		break;
    795 	default:
    796 		error = ENOTTY;
    797 		break;
    798 	}
    799 #endif
    800 
    801 	return error;
    802 }
    803