Home | History | Annotate | Line # | Download | only in ic
icp.c revision 1.27
      1 /*	$NetBSD: icp.c,v 1.27 2008/04/08 12:07:26 cegger Exp $	*/
      2 
      3 /*-
      4  * Copyright (c) 2002, 2003 The NetBSD Foundation, Inc.
      5  * All rights reserved.
      6  *
      7  * This code is derived from software contributed to The NetBSD Foundation
      8  * by Andrew Doran, and by Jason R. Thorpe of Wasabi Systems, Inc.
      9  *
     10  * Redistribution and use in source and binary forms, with or without
     11  * modification, are permitted provided that the following conditions
     12  * are met:
     13  * 1. Redistributions of source code must retain the above copyright
     14  *    notice, this list of conditions and the following disclaimer.
     15  * 2. Redistributions in binary form must reproduce the above copyright
     16  *    notice, this list of conditions and the following disclaimer in the
     17  *    documentation and/or other materials provided with the distribution.
     18  * 3. All advertising materials mentioning features or use of this software
     19  *    must display the following acknowledgement:
     20  *        This product includes software developed by the NetBSD
     21  *        Foundation, Inc. and its contributors.
     22  * 4. Neither the name of The NetBSD Foundation nor the names of its
     23  *    contributors may be used to endorse or promote products derived
     24  *    from this software without specific prior written permission.
     25  *
     26  * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
     27  * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
     28  * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
     29  * PURPOSE ARE DISCLAIMED.  IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
     30  * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
     31  * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
     32  * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
     33  * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
     34  * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
     35  * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
     36  * POSSIBILITY OF SUCH DAMAGE.
     37  */
     38 
     39 /*
     40  * Copyright (c) 1999, 2000 Niklas Hallqvist.  All rights reserved.
     41  *
     42  * Redistribution and use in source and binary forms, with or without
     43  * modification, are permitted provided that the following conditions
     44  * are met:
     45  * 1. Redistributions of source code must retain the above copyright
     46  *    notice, this list of conditions and the following disclaimer.
     47  * 2. Redistributions in binary form must reproduce the above copyright
     48  *    notice, this list of conditions and the following disclaimer in the
     49  *    documentation and/or other materials provided with the distribution.
     50  * 3. All advertising materials mentioning features or use of this software
     51  *    must display the following acknowledgement:
     52  *	This product includes software developed by Niklas Hallqvist.
     53  * 4. The name of the author may not be used to endorse or promote products
     54  *    derived from this software without specific prior written permission.
     55  *
     56  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
     57  * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
     58  * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
     59  * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
     60  * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
     61   * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
     62  * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
     63  * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
     64  * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
     65  * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
     66  *
     67  * from OpenBSD: gdt_common.c,v 1.12 2001/07/04 06:43:18 niklas Exp
     68  */
     69 
     70 /*
     71  * This driver would not have written if it was not for the hardware donations
     72  * from both ICP-Vortex and ko.neT.  I want to thank them for their support.
     73  *
     74  * Re-worked for NetBSD by Andrew Doran.  Test hardware kindly supplied by
     75  * Intel.
     76  *
     77  * Support for the ICP-Vortex management tools added by
     78  * Jason R. Thorpe of Wasabi Systems, Inc., based on code
     79  * provided by Achim Leubner <achim.leubner (at) intel.com>.
     80  *
     81  * Additional support for dynamic rescan of cacheservice drives by
     82  * Jason R. Thorpe of Wasabi Systems, Inc.
     83  */
     84 
     85 #include <sys/cdefs.h>
     86 __KERNEL_RCSID(0, "$NetBSD: icp.c,v 1.27 2008/04/08 12:07:26 cegger Exp $");
     87 
     88 #include <sys/param.h>
     89 #include <sys/systm.h>
     90 #include <sys/kernel.h>
     91 #include <sys/device.h>
     92 #include <sys/queue.h>
     93 #include <sys/proc.h>
     94 #include <sys/buf.h>
     95 #include <sys/endian.h>
     96 #include <sys/malloc.h>
     97 #include <sys/disk.h>
     98 
     99 #include <uvm/uvm_extern.h>
    100 
    101 #include <sys/bswap.h>
    102 #include <sys/bus.h>
    103 
    104 #include <dev/pci/pcireg.h>
    105 #include <dev/pci/pcivar.h>
    106 #include <dev/pci/pcidevs.h>
    107 
    108 #include <dev/ic/icpreg.h>
    109 #include <dev/ic/icpvar.h>
    110 
    111 #include <dev/scsipi/scsipi_all.h>
    112 #include <dev/scsipi/scsiconf.h>
    113 
    114 #include "locators.h"
    115 
    116 int	icp_async_event(struct icp_softc *, int);
    117 void	icp_ccb_submit(struct icp_softc *icp, struct icp_ccb *ic);
    118 void	icp_chain(struct icp_softc *);
    119 int	icp_print(void *, const char *);
    120 void	icp_watchdog(void *);
    121 void	icp_ucmd_intr(struct icp_ccb *);
    122 void	icp_recompute_openings(struct icp_softc *);
    123 
    124 int	icp_count;	/* total # of controllers, for ioctl interface */
    125 
    126 /*
    127  * Statistics for the ioctl interface to query.
    128  *
    129  * XXX Global.  They should probably be made per-controller
    130  * XXX at some point.
    131  */
    132 gdt_statist_t icp_stats;
    133 
    134 int
    135 icp_init(struct icp_softc *icp, const char *intrstr)
    136 {
    137 	struct icp_attach_args icpa;
    138 	struct icp_binfo binfo;
    139 	struct icp_ccb *ic;
    140 	u_int16_t cdev_cnt;
    141 	int i, j, state, feat, nsegs, rv;
    142 	int locs[ICPCF_NLOCS];
    143 
    144 	state = 0;
    145 
    146 	if (intrstr != NULL)
    147 		aprint_normal_dev(&icp->icp_dv, "interrupting at %s\n",
    148 		    intrstr);
    149 
    150 	SIMPLEQ_INIT(&icp->icp_ccb_queue);
    151 	SIMPLEQ_INIT(&icp->icp_ccb_freelist);
    152 	SIMPLEQ_INIT(&icp->icp_ucmd_queue);
    153 	callout_init(&icp->icp_wdog_callout, 0);
    154 
    155 	/*
    156 	 * Allocate a scratch area.
    157 	 */
    158 	if (bus_dmamap_create(icp->icp_dmat, ICP_SCRATCH_SIZE, 1,
    159 	    ICP_SCRATCH_SIZE, 0, BUS_DMA_NOWAIT | BUS_DMA_ALLOCNOW,
    160 	    &icp->icp_scr_dmamap) != 0) {
    161 		aprint_error_dev(&icp->icp_dv, "cannot create scratch dmamap\n");
    162 		return (1);
    163 	}
    164 	state++;
    165 
    166 	if (bus_dmamem_alloc(icp->icp_dmat, ICP_SCRATCH_SIZE, PAGE_SIZE, 0,
    167 	    icp->icp_scr_seg, 1, &nsegs, BUS_DMA_NOWAIT) != 0) {
    168 		aprint_error_dev(&icp->icp_dv, "cannot alloc scratch dmamem\n");
    169 		goto bail_out;
    170 	}
    171 	state++;
    172 
    173 	if (bus_dmamem_map(icp->icp_dmat, icp->icp_scr_seg, nsegs,
    174 	    ICP_SCRATCH_SIZE, &icp->icp_scr, 0)) {
    175 		aprint_error_dev(&icp->icp_dv, "cannot map scratch dmamem\n");
    176 		goto bail_out;
    177 	}
    178 	state++;
    179 
    180 	if (bus_dmamap_load(icp->icp_dmat, icp->icp_scr_dmamap, icp->icp_scr,
    181 	    ICP_SCRATCH_SIZE, NULL, BUS_DMA_NOWAIT)) {
    182 		aprint_error_dev(&icp->icp_dv, "cannot load scratch dmamap\n");
    183 		goto bail_out;
    184 	}
    185 	state++;
    186 
    187 	/*
    188 	 * Allocate and initialize the command control blocks.
    189 	 */
    190 	ic = malloc(sizeof(*ic) * ICP_NCCBS, M_DEVBUF, M_NOWAIT | M_ZERO);
    191 	if ((icp->icp_ccbs = ic) == NULL) {
    192 		aprint_error_dev(&icp->icp_dv, "malloc() failed\n");
    193 		goto bail_out;
    194 	}
    195 	state++;
    196 
    197 	for (i = 0; i < ICP_NCCBS; i++, ic++) {
    198 		/*
    199 		 * The first two command indexes have special meanings, so
    200 		 * we can't use them.
    201 		 */
    202 		ic->ic_ident = i + 2;
    203 		rv = bus_dmamap_create(icp->icp_dmat, ICP_MAX_XFER,
    204 		    ICP_MAXSG, ICP_MAX_XFER, 0,
    205 		    BUS_DMA_NOWAIT | BUS_DMA_ALLOCNOW,
    206 		    &ic->ic_xfer_map);
    207 		if (rv != 0)
    208 			break;
    209 		icp->icp_nccbs++;
    210 		icp_ccb_free(icp, ic);
    211 	}
    212 #ifdef DIAGNOSTIC
    213 	if (icp->icp_nccbs != ICP_NCCBS)
    214 		aprint_error_dev(&icp->icp_dv, "%d/%d CCBs usable\n",
    215 		    icp->icp_nccbs, ICP_NCCBS);
    216 #endif
    217 
    218 	/*
    219 	 * Initalize the controller.
    220 	 */
    221 	if (!icp_cmd(icp, ICP_SCREENSERVICE, ICP_INIT, 0, 0, 0)) {
    222 		aprint_error_dev(&icp->icp_dv, "screen service init error %d\n",
    223 		    icp->icp_status);
    224 		goto bail_out;
    225 	}
    226 
    227 	if (!icp_cmd(icp, ICP_CACHESERVICE, ICP_INIT, ICP_LINUX_OS, 0, 0)) {
    228 		aprint_error_dev(&icp->icp_dv, "cache service init error %d\n",
    229 		    icp->icp_status);
    230 		goto bail_out;
    231 	}
    232 
    233 	icp_cmd(icp, ICP_CACHESERVICE, ICP_UNFREEZE_IO, 0, 0, 0);
    234 
    235 	if (!icp_cmd(icp, ICP_CACHESERVICE, ICP_MOUNT, 0xffff, 1, 0)) {
    236 		aprint_error_dev(&icp->icp_dv, "cache service mount error %d\n",
    237 		    icp->icp_status);
    238 		goto bail_out;
    239 	}
    240 
    241 	if (!icp_cmd(icp, ICP_CACHESERVICE, ICP_INIT, ICP_LINUX_OS, 0, 0)) {
    242 		aprint_error_dev(&icp->icp_dv, "cache service post-mount init error %d\n",
    243 		    icp->icp_status);
    244 		goto bail_out;
    245 	}
    246 	cdev_cnt = (u_int16_t)icp->icp_info;
    247 	icp->icp_fw_vers = icp->icp_service;
    248 
    249 	if (!icp_cmd(icp, ICP_SCSIRAWSERVICE, ICP_INIT, 0, 0, 0)) {
    250 		aprint_error_dev(&icp->icp_dv, "raw service init error %d\n",
    251 		    icp->icp_status);
    252 		goto bail_out;
    253 	}
    254 
    255 	/*
    256 	 * Set/get raw service features (scatter/gather).
    257 	 */
    258 	feat = 0;
    259 	if (icp_cmd(icp, ICP_SCSIRAWSERVICE, ICP_SET_FEAT, ICP_SCATTER_GATHER,
    260 	    0, 0))
    261 		if (icp_cmd(icp, ICP_SCSIRAWSERVICE, ICP_GET_FEAT, 0, 0, 0))
    262 			feat = icp->icp_info;
    263 
    264 	if ((feat & ICP_SCATTER_GATHER) == 0) {
    265 #ifdef DIAGNOSTIC
    266 		aprint_normal_dev(&icp->icp_dv,
    267 		    "scatter/gather not supported (raw service)\n");
    268 #endif
    269 	} else
    270 		icp->icp_features |= ICP_FEAT_RAWSERVICE;
    271 
    272 	/*
    273 	 * Set/get cache service features (scatter/gather).
    274 	 */
    275 	feat = 0;
    276 	if (icp_cmd(icp, ICP_CACHESERVICE, ICP_SET_FEAT, 0,
    277 	    ICP_SCATTER_GATHER, 0))
    278 		if (icp_cmd(icp, ICP_CACHESERVICE, ICP_GET_FEAT, 0, 0, 0))
    279 			feat = icp->icp_info;
    280 
    281 	if ((feat & ICP_SCATTER_GATHER) == 0) {
    282 #ifdef DIAGNOSTIC
    283 		aprint_normal_dev(&icp->icp_dv,
    284 		    "scatter/gather not supported (cache service)\n");
    285 #endif
    286 	} else
    287 		icp->icp_features |= ICP_FEAT_CACHESERVICE;
    288 
    289 	/*
    290 	 * Pull some information from the board and dump.
    291 	 */
    292 	if (!icp_cmd(icp, ICP_CACHESERVICE, ICP_IOCTL, ICP_BOARD_INFO,
    293 	    ICP_INVALID_CHANNEL, sizeof(struct icp_binfo))) {
    294 		aprint_error_dev(&icp->icp_dv, "unable to retrive board info\n");
    295 		goto bail_out;
    296 	}
    297 	memcpy(&binfo, icp->icp_scr, sizeof(binfo));
    298 
    299 	aprint_normal_dev(&icp->icp_dv,
    300 	    "model <%s>, firmware <%s>, %d channel(s), %dMB memory\n",
    301 	    binfo.bi_type_string, binfo.bi_raid_string,
    302 	    binfo.bi_chan_count, le32toh(binfo.bi_memsize) >> 20);
    303 
    304 	/*
    305 	 * Determine the number of devices, and number of openings per
    306 	 * device.
    307 	 */
    308 	if (icp->icp_features & ICP_FEAT_CACHESERVICE) {
    309 		for (j = 0; j < cdev_cnt && j < ICP_MAX_HDRIVES; j++) {
    310 			if (!icp_cmd(icp, ICP_CACHESERVICE, ICP_INFO, j, 0,
    311 			    0))
    312 				continue;
    313 
    314 			icp->icp_cdr[j].cd_size = icp->icp_info;
    315 			if (icp->icp_cdr[j].cd_size != 0)
    316 				icp->icp_ndevs++;
    317 
    318 			if (icp_cmd(icp, ICP_CACHESERVICE, ICP_DEVTYPE, j, 0,
    319 			    0))
    320 				icp->icp_cdr[j].cd_type = icp->icp_info;
    321 		}
    322 	}
    323 
    324 	if (icp->icp_features & ICP_FEAT_RAWSERVICE) {
    325 		icp->icp_nchan = binfo.bi_chan_count;
    326 		icp->icp_ndevs += icp->icp_nchan;
    327 	}
    328 
    329 	icp_recompute_openings(icp);
    330 
    331 	/*
    332 	 * Attach SCSI channels.
    333 	 */
    334 	if (icp->icp_features & ICP_FEAT_RAWSERVICE) {
    335 		struct icp_ioc_version *iv;
    336 		struct icp_rawioc *ri;
    337 		struct icp_getch *gc;
    338 
    339 		iv = (struct icp_ioc_version *)icp->icp_scr;
    340 		iv->iv_version = htole32(ICP_IOC_NEWEST);
    341 		iv->iv_listents = ICP_MAXBUS;
    342 		iv->iv_firstchan = 0;
    343 		iv->iv_lastchan = ICP_MAXBUS - 1;
    344 		iv->iv_listoffset = htole32(sizeof(*iv));
    345 
    346 		if (icp_cmd(icp, ICP_CACHESERVICE, ICP_IOCTL,
    347 		    ICP_IOCHAN_RAW_DESC, ICP_INVALID_CHANNEL,
    348 		    sizeof(*iv) + ICP_MAXBUS * sizeof(*ri))) {
    349 			ri = (struct icp_rawioc *)(iv + 1);
    350 			for (j = 0; j < binfo.bi_chan_count; j++, ri++)
    351 				icp->icp_bus_id[j] = ri->ri_procid;
    352 		} else {
    353 			/*
    354 			 * Fall back to the old method.
    355 			 */
    356 			gc = (struct icp_getch *)icp->icp_scr;
    357 
    358 			for (j = 0; j < binfo.bi_chan_count; j++) {
    359 				if (!icp_cmd(icp, ICP_CACHESERVICE, ICP_IOCTL,
    360 				    ICP_SCSI_CHAN_CNT | ICP_L_CTRL_PATTERN,
    361 				    ICP_IO_CHANNEL | ICP_INVALID_CHANNEL,
    362 				    sizeof(*gc))) {
    363 				    	aprint_error_dev(&icp->icp_dv,
    364 					    "unable to get chan info");
    365 					goto bail_out;
    366 				}
    367 				icp->icp_bus_id[j] = gc->gc_scsiid;
    368 			}
    369 		}
    370 
    371 		for (j = 0; j < binfo.bi_chan_count; j++) {
    372 			if (icp->icp_bus_id[j] > ICP_MAXID_FC)
    373 				icp->icp_bus_id[j] = ICP_MAXID_FC;
    374 
    375 			icpa.icpa_unit = j + ICPA_UNIT_SCSI;
    376 
    377 			locs[ICPCF_UNIT] = j + ICPA_UNIT_SCSI;
    378 
    379 			icp->icp_children[icpa.icpa_unit] =
    380 				config_found_sm_loc(&icp->icp_dv, "icp", locs,
    381 					&icpa, icp_print, config_stdsubmatch);
    382 		}
    383 	}
    384 
    385 	/*
    386 	 * Attach cache devices.
    387 	 */
    388 	if (icp->icp_features & ICP_FEAT_CACHESERVICE) {
    389 		for (j = 0; j < cdev_cnt && j < ICP_MAX_HDRIVES; j++) {
    390 			if (icp->icp_cdr[j].cd_size == 0)
    391 				continue;
    392 
    393 			icpa.icpa_unit = j;
    394 
    395 			locs[ICPCF_UNIT] = j;
    396 
    397 			icp->icp_children[icpa.icpa_unit] =
    398 			    config_found_sm_loc(&icp->icp_dv, "icp", locs,
    399 				&icpa, icp_print, config_stdsubmatch);
    400 		}
    401 	}
    402 
    403 	/*
    404 	 * Start the watchdog.
    405 	 */
    406 	icp_watchdog(icp);
    407 
    408 	/*
    409 	 * Count the controller, and we're done!
    410 	 */
    411 	if (icp_count++ == 0)
    412 		mutex_init(&icp_ioctl_mutex, MUTEX_DEFAULT, IPL_NONE);
    413 
    414 	return (0);
    415 
    416  bail_out:
    417 	if (state > 4)
    418 		for (j = 0; j < i; j++)
    419 			bus_dmamap_destroy(icp->icp_dmat,
    420 			    icp->icp_ccbs[j].ic_xfer_map);
    421  	if (state > 3)
    422 		free(icp->icp_ccbs, M_DEVBUF);
    423 	if (state > 2)
    424 		bus_dmamap_unload(icp->icp_dmat, icp->icp_scr_dmamap);
    425 	if (state > 1)
    426 		bus_dmamem_unmap(icp->icp_dmat, icp->icp_scr,
    427 		    ICP_SCRATCH_SIZE);
    428 	if (state > 0)
    429 		bus_dmamem_free(icp->icp_dmat, icp->icp_scr_seg, nsegs);
    430 	bus_dmamap_destroy(icp->icp_dmat, icp->icp_scr_dmamap);
    431 
    432 	return (1);
    433 }
    434 
    435 void
    436 icp_register_servicecb(struct icp_softc *icp, int unit,
    437     const struct icp_servicecb *cb)
    438 {
    439 
    440 	icp->icp_servicecb[unit] = cb;
    441 }
    442 
    443 void
    444 icp_rescan(struct icp_softc *icp, int unit)
    445 {
    446 	struct icp_attach_args icpa;
    447 	u_int newsize, newtype;
    448 	int locs[ICPCF_NLOCS];
    449 
    450 	/*
    451 	 * NOTE: It is very important that the queue be frozen and not
    452 	 * commands running when this is called.  The ioctl mutex must
    453 	 * also be held.
    454 	 */
    455 
    456 	KASSERT(icp->icp_qfreeze != 0);
    457 	KASSERT(icp->icp_running == 0);
    458 	KASSERT(unit < ICP_MAX_HDRIVES);
    459 
    460 	if (!icp_cmd(icp, ICP_CACHESERVICE, ICP_INFO, unit, 0, 0)) {
    461 #ifdef ICP_DEBUG
    462 		printf("%s: rescan: unit %d ICP_INFO failed -> 0x%04x\n",
    463 		    device_xname(&icp->icp_dv), unit, icp->icp_status);
    464 #endif
    465 		goto gone;
    466 	}
    467 	if ((newsize = icp->icp_info) == 0) {
    468 #ifdef ICP_DEBUG
    469 		printf("%s: rescan: unit %d has zero size\n",
    470 		    device-xname(&icp->icp_dv), unit);
    471 #endif
    472  gone:
    473 		/*
    474 		 * Host drive is no longer present; detach if a child
    475 		 * is currently there.
    476 		 */
    477 		if (icp->icp_cdr[unit].cd_size != 0)
    478 			icp->icp_ndevs--;
    479 		icp->icp_cdr[unit].cd_size = 0;
    480 		if (icp->icp_children[unit] != NULL) {
    481 			(void) config_detach(icp->icp_children[unit],
    482 			    DETACH_FORCE);
    483 			icp->icp_children[unit] = NULL;
    484 		}
    485 		return;
    486 	}
    487 
    488 	if (icp_cmd(icp, ICP_CACHESERVICE, ICP_DEVTYPE, unit, 0, 0))
    489 		newtype = icp->icp_info;
    490 	else {
    491 #ifdef ICP_DEBUG
    492 		printf("%s: rescan: unit %d ICP_DEVTYPE failed\n",
    493 		    device_xname(&icp->icp_dv), unit);
    494 #endif
    495 		newtype = 0;	/* XXX? */
    496 	}
    497 
    498 #ifdef ICP_DEBUG
    499 	printf("%s: rescan: unit %d old %u/%u, new %u/%u\n",
    500 	    device_xname(&icp->icp_dv), unit, icp->icp_cdr[unit].cd_size,
    501 	    icp->icp_cdr[unit].cd_type, newsize, newtype);
    502 #endif
    503 
    504 	/*
    505 	 * If the type or size changed, detach any old child (if it exists)
    506 	 * and attach a new one.
    507 	 */
    508 	if (icp->icp_children[unit] == NULL ||
    509 	    newsize != icp->icp_cdr[unit].cd_size ||
    510 	    newtype != icp->icp_cdr[unit].cd_type) {
    511 		if (icp->icp_cdr[unit].cd_size == 0)
    512 			icp->icp_ndevs++;
    513 		icp->icp_cdr[unit].cd_size = newsize;
    514 		icp->icp_cdr[unit].cd_type = newtype;
    515 		if (icp->icp_children[unit] != NULL)
    516 			(void) config_detach(icp->icp_children[unit],
    517 			    DETACH_FORCE);
    518 
    519 		icpa.icpa_unit = unit;
    520 
    521 		locs[ICPCF_UNIT] = unit;
    522 
    523 		icp->icp_children[unit] = config_found_sm_loc(&icp->icp_dv,
    524 			"icp", locs, &icpa, icp_print, config_stdsubmatch);
    525 	}
    526 
    527 	icp_recompute_openings(icp);
    528 }
    529 
    530 void
    531 icp_rescan_all(struct icp_softc *icp)
    532 {
    533 	int unit;
    534 	u_int16_t cdev_cnt;
    535 
    536 	/*
    537 	 * This is the old method of rescanning the host drives.  We
    538 	 * start by reinitializing the cache service.
    539 	 */
    540 	if (!icp_cmd(icp, ICP_CACHESERVICE, ICP_INIT, ICP_LINUX_OS, 0, 0)) {
    541 		printf("%s: unable to re-initialize cache service for rescan\n",
    542 		    device_xname(&icp->icp_dv));
    543 		return;
    544 	}
    545 	cdev_cnt = (u_int16_t) icp->icp_info;
    546 
    547 	/* For each host drive, do the new-style rescan. */
    548 	for (unit = 0; unit < cdev_cnt && unit < ICP_MAX_HDRIVES; unit++)
    549 		icp_rescan(icp, unit);
    550 
    551 	/* Now detach anything in the slots after cdev_cnt. */
    552 	for (; unit < ICP_MAX_HDRIVES; unit++) {
    553 		if (icp->icp_cdr[unit].cd_size != 0) {
    554 #ifdef ICP_DEBUG
    555 			printf("%s: rescan all: unit %d < new cdev_cnt (%d)\n",
    556 			    device_xname(&icp->icp_dv), unit, cdev_cnt);
    557 #endif
    558 			icp->icp_ndevs--;
    559 			icp->icp_cdr[unit].cd_size = 0;
    560 			if (icp->icp_children[unit] != NULL) {
    561 				(void) config_detach(icp->icp_children[unit],
    562 				    DETACH_FORCE);
    563 				icp->icp_children[unit] = NULL;
    564 			}
    565 		}
    566 	}
    567 
    568 	icp_recompute_openings(icp);
    569 }
    570 
    571 void
    572 icp_recompute_openings(struct icp_softc *icp)
    573 {
    574 	int unit, openings;
    575 
    576 	if (icp->icp_ndevs != 0)
    577 		openings =
    578 		    (icp->icp_nccbs - ICP_NCCB_RESERVE) / icp->icp_ndevs;
    579 	else
    580 		openings = 0;
    581 	if (openings == icp->icp_openings)
    582 		return;
    583 	icp->icp_openings = openings;
    584 
    585 #ifdef ICP_DEBUG
    586 	printf("%s: %d device%s, %d openings per device\n",
    587 	    device_xname(&icp->icp_dv), icp->icp_ndevs,
    588 	    icp->icp_ndevs == 1 ? "" : "s", icp->icp_openings);
    589 #endif
    590 
    591 	for (unit = 0; unit < ICP_MAX_HDRIVES + ICP_MAXBUS; unit++) {
    592 		if (icp->icp_children[unit] != NULL)
    593 			(*icp->icp_servicecb[unit]->iscb_openings)(
    594 			    icp->icp_children[unit], icp->icp_openings);
    595 	}
    596 }
    597 
    598 void
    599 icp_watchdog(void *cookie)
    600 {
    601 	struct icp_softc *icp;
    602 	int s;
    603 
    604 	icp = cookie;
    605 
    606 	s = splbio();
    607 	icp_intr(icp);
    608 	if (ICP_HAS_WORK(icp))
    609 		icp_ccb_enqueue(icp, NULL);
    610 	splx(s);
    611 
    612 	callout_reset(&icp->icp_wdog_callout, hz * ICP_WATCHDOG_FREQ,
    613 	    icp_watchdog, icp);
    614 }
    615 
    616 int
    617 icp_print(void *aux, const char *pnp)
    618 {
    619 	struct icp_attach_args *icpa;
    620 	const char *str;
    621 
    622 	icpa = (struct icp_attach_args *)aux;
    623 
    624 	if (pnp != NULL) {
    625 		if (icpa->icpa_unit < ICPA_UNIT_SCSI)
    626 			str = "block device";
    627 		else
    628 			str = "SCSI channel";
    629 		aprint_normal("%s at %s", str, pnp);
    630 	}
    631 	aprint_normal(" unit %d", icpa->icpa_unit);
    632 
    633 	return (UNCONF);
    634 }
    635 
    636 int
    637 icp_async_event(struct icp_softc *icp, int service)
    638 {
    639 
    640 	if (service == ICP_SCREENSERVICE) {
    641 		if (icp->icp_status == ICP_S_MSG_REQUEST) {
    642 			/* XXX */
    643 		}
    644 	} else {
    645 		if ((icp->icp_fw_vers & 0xff) >= 0x1a) {
    646 			icp->icp_evt.size = 0;
    647 			icp->icp_evt.eu.async.ionode =
    648 			    device_unit(&icp->icp_dv);
    649 			icp->icp_evt.eu.async.status = icp->icp_status;
    650 			/*
    651 			 * Severity and event string are filled in by the
    652 			 * hardware interface interrupt handler.
    653 			 */
    654 			printf("%s: %s\n", device_xname(&icp->icp_dv),
    655 			    icp->icp_evt.event_string);
    656 		} else {
    657 			icp->icp_evt.size = sizeof(icp->icp_evt.eu.async);
    658 			icp->icp_evt.eu.async.ionode =
    659 			    device_unit(&icp->icp_dv);
    660 			icp->icp_evt.eu.async.service = service;
    661 			icp->icp_evt.eu.async.status = icp->icp_status;
    662 			icp->icp_evt.eu.async.info = icp->icp_info;
    663 			/* XXXJRT FIX THIS */
    664 			*(u_int32_t *) icp->icp_evt.eu.async.scsi_coord =
    665 			    icp->icp_info2;
    666 		}
    667 		icp_store_event(icp, GDT_ES_ASYNC, service, &icp->icp_evt);
    668 	}
    669 
    670 	return (0);
    671 }
    672 
    673 int
    674 icp_intr(void *cookie)
    675 {
    676 	struct icp_softc *icp;
    677 	struct icp_intr_ctx ctx;
    678 	struct icp_ccb *ic;
    679 
    680 	icp = cookie;
    681 
    682 	ctx.istatus = (*icp->icp_get_status)(icp);
    683 	if (!ctx.istatus) {
    684 		icp->icp_status = ICP_S_NO_STATUS;
    685 		return (0);
    686 	}
    687 
    688 	(*icp->icp_intr)(icp, &ctx);
    689 
    690 	icp->icp_status = ctx.cmd_status;
    691 	icp->icp_service = ctx.service;
    692 	icp->icp_info = ctx.info;
    693 	icp->icp_info2 = ctx.info2;
    694 
    695 	switch (ctx.istatus) {
    696 	case ICP_ASYNCINDEX:
    697 		icp_async_event(icp, ctx.service);
    698 		return (1);
    699 
    700 	case ICP_SPEZINDEX:
    701 		aprint_error_dev(&icp->icp_dv, "uninitialized or unknown service (%d/%d)\n",
    702 		    ctx.info, ctx.info2);
    703 		icp->icp_evt.size = sizeof(icp->icp_evt.eu.driver);
    704 		icp->icp_evt.eu.driver.ionode = device_unit(&icp->icp_dv);
    705 		icp_store_event(icp, GDT_ES_DRIVER, 4, &icp->icp_evt);
    706 		return (1);
    707 	}
    708 
    709 	if ((ctx.istatus - 2) > icp->icp_nccbs)
    710 		panic("icp_intr: bad command index returned");
    711 
    712 	ic = &icp->icp_ccbs[ctx.istatus - 2];
    713 	ic->ic_status = icp->icp_status;
    714 
    715 	if ((ic->ic_flags & IC_ALLOCED) == 0) {
    716 		/* XXX ICP's "iir" driver just sends an event here. */
    717 		panic("icp_intr: inactive CCB identified");
    718 	}
    719 
    720 	/*
    721 	 * Try to protect ourselves from the running command count already
    722 	 * being 0 (e.g. if a polled command times out).
    723 	 */
    724 	KDASSERT(icp->icp_running != 0);
    725 	if (--icp->icp_running == 0 &&
    726 	    (icp->icp_flags & ICP_F_WAIT_FREEZE) != 0) {
    727 		icp->icp_flags &= ~ICP_F_WAIT_FREEZE;
    728 		wakeup(&icp->icp_qfreeze);
    729 	}
    730 
    731 	switch (icp->icp_status) {
    732 	case ICP_S_BSY:
    733 #ifdef ICP_DEBUG
    734 		printf("%s: ICP_S_BSY received\n", device_xname(&icp->icp_dv));
    735 #endif
    736 		if (__predict_false((ic->ic_flags & IC_UCMD) != 0))
    737 			SIMPLEQ_INSERT_HEAD(&icp->icp_ucmd_queue, ic, ic_chain);
    738 		else
    739 			SIMPLEQ_INSERT_HEAD(&icp->icp_ccb_queue, ic, ic_chain);
    740 		break;
    741 
    742 	default:
    743 		ic->ic_flags |= IC_COMPLETE;
    744 
    745 		if ((ic->ic_flags & IC_WAITING) != 0)
    746 			wakeup(ic);
    747 		else if (ic->ic_intr != NULL)
    748 			(*ic->ic_intr)(ic);
    749 
    750 		if (ICP_HAS_WORK(icp))
    751 			icp_ccb_enqueue(icp, NULL);
    752 
    753 		break;
    754 	}
    755 
    756 	return (1);
    757 }
    758 
    759 struct icp_ucmd_ctx {
    760 	gdt_ucmd_t *iu_ucmd;
    761 	u_int32_t iu_cnt;
    762 };
    763 
    764 void
    765 icp_ucmd_intr(struct icp_ccb *ic)
    766 {
    767 	struct icp_softc *icp = (void *) ic->ic_dv;
    768 	struct icp_ucmd_ctx *iu = ic->ic_context;
    769 	gdt_ucmd_t *ucmd = iu->iu_ucmd;
    770 
    771 	ucmd->status = icp->icp_status;
    772 	ucmd->info = icp->icp_info;
    773 
    774 	if (iu->iu_cnt != 0) {
    775 		bus_dmamap_sync(icp->icp_dmat,
    776 		    icp->icp_scr_dmamap,
    777 		    ICP_SCRATCH_UCMD, iu->iu_cnt,
    778 		    BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE);
    779 		memcpy(ucmd->data,
    780 		    (char *)icp->icp_scr + ICP_SCRATCH_UCMD, iu->iu_cnt);
    781 	}
    782 
    783 	icp->icp_ucmd_ccb = NULL;
    784 
    785 	ic->ic_flags |= IC_COMPLETE;
    786 	wakeup(ic);
    787 }
    788 
    789 /*
    790  * NOTE: We assume that it is safe to sleep here!
    791  */
    792 int
    793 icp_cmd(struct icp_softc *icp, u_int8_t service, u_int16_t opcode,
    794 	u_int32_t arg1, u_int32_t arg2, u_int32_t arg3)
    795 {
    796 	struct icp_ioctlcmd *icmd;
    797 	struct icp_cachecmd *cc;
    798 	struct icp_rawcmd *rc;
    799 	int retries, rv;
    800 	struct icp_ccb *ic;
    801 
    802 	retries = ICP_RETRIES;
    803 
    804 	do {
    805 		ic = icp_ccb_alloc_wait(icp);
    806 		memset(&ic->ic_cmd, 0, sizeof(ic->ic_cmd));
    807 		ic->ic_cmd.cmd_opcode = htole16(opcode);
    808 
    809 		switch (service) {
    810 		case ICP_CACHESERVICE:
    811 			if (opcode == ICP_IOCTL) {
    812 				icmd = &ic->ic_cmd.cmd_packet.ic;
    813 				icmd->ic_subfunc = htole16(arg1);
    814 				icmd->ic_channel = htole32(arg2);
    815 				icmd->ic_bufsize = htole32(arg3);
    816 				icmd->ic_addr =
    817 				    htole32(icp->icp_scr_seg[0].ds_addr);
    818 
    819 				bus_dmamap_sync(icp->icp_dmat,
    820 				    icp->icp_scr_dmamap, 0, arg3,
    821 				    BUS_DMASYNC_PREWRITE |
    822 				    BUS_DMASYNC_PREREAD);
    823 			} else {
    824 				cc = &ic->ic_cmd.cmd_packet.cc;
    825 				cc->cc_deviceno = htole16(arg1);
    826 				cc->cc_blockno = htole32(arg2);
    827 			}
    828 			break;
    829 
    830 		case ICP_SCSIRAWSERVICE:
    831 			rc = &ic->ic_cmd.cmd_packet.rc;
    832 			rc->rc_direction = htole32(arg1);
    833 			rc->rc_bus = arg2;
    834 			rc->rc_target = arg3;
    835 			rc->rc_lun = arg3 >> 8;
    836 			break;
    837 		}
    838 
    839 		ic->ic_service = service;
    840 		ic->ic_cmdlen = sizeof(ic->ic_cmd);
    841 		rv = icp_ccb_poll(icp, ic, 10000);
    842 
    843 		switch (service) {
    844 		case ICP_CACHESERVICE:
    845 			if (opcode == ICP_IOCTL) {
    846 				bus_dmamap_sync(icp->icp_dmat,
    847 				    icp->icp_scr_dmamap, 0, arg3,
    848 				    BUS_DMASYNC_POSTWRITE |
    849 				    BUS_DMASYNC_POSTREAD);
    850 			}
    851 			break;
    852 		}
    853 
    854 		icp_ccb_free(icp, ic);
    855 	} while (rv != 0 && --retries > 0);
    856 
    857 	return (icp->icp_status == ICP_S_OK);
    858 }
    859 
    860 int
    861 icp_ucmd(struct icp_softc *icp, gdt_ucmd_t *ucmd)
    862 {
    863 	struct icp_ccb *ic;
    864 	struct icp_ucmd_ctx iu;
    865 	u_int32_t cnt;
    866 	int error;
    867 
    868 	if (ucmd->service == ICP_CACHESERVICE) {
    869 		if (ucmd->command.cmd_opcode == ICP_IOCTL) {
    870 			cnt = ucmd->command.cmd_packet.ic.ic_bufsize;
    871 			if (cnt > GDT_SCRATCH_SZ) {
    872 				aprint_error_dev(&icp->icp_dv, "scratch buffer too small (%d/%d)\n",
    873 				    GDT_SCRATCH_SZ, cnt);
    874 				return (EINVAL);
    875 			}
    876 		} else {
    877 			cnt = ucmd->command.cmd_packet.cc.cc_blockcnt *
    878 			    ICP_SECTOR_SIZE;
    879 			if (cnt > GDT_SCRATCH_SZ) {
    880 				aprint_error_dev(&icp->icp_dv, "scratch buffer too small (%d/%d)\n",
    881 				    GDT_SCRATCH_SZ, cnt);
    882 				return (EINVAL);
    883 			}
    884 		}
    885 	} else {
    886 		cnt = ucmd->command.cmd_packet.rc.rc_sdlen +
    887 		    ucmd->command.cmd_packet.rc.rc_sense_len;
    888 		if (cnt > GDT_SCRATCH_SZ) {
    889 			aprint_error_dev(&icp->icp_dv, "scratch buffer too small (%d/%d)\n",
    890 			    GDT_SCRATCH_SZ, cnt);
    891 			return (EINVAL);
    892 		}
    893 	}
    894 
    895 	iu.iu_ucmd = ucmd;
    896 	iu.iu_cnt = cnt;
    897 
    898 	ic = icp_ccb_alloc_wait(icp);
    899 	memset(&ic->ic_cmd, 0, sizeof(ic->ic_cmd));
    900 	ic->ic_cmd.cmd_opcode = htole16(ucmd->command.cmd_opcode);
    901 
    902 	if (ucmd->service == ICP_CACHESERVICE) {
    903 		if (ucmd->command.cmd_opcode == ICP_IOCTL) {
    904 			struct icp_ioctlcmd *icmd, *uicmd;
    905 
    906 			icmd = &ic->ic_cmd.cmd_packet.ic;
    907 			uicmd = &ucmd->command.cmd_packet.ic;
    908 
    909 			icmd->ic_subfunc = htole16(uicmd->ic_subfunc);
    910 			icmd->ic_channel = htole32(uicmd->ic_channel);
    911 			icmd->ic_bufsize = htole32(uicmd->ic_bufsize);
    912 			icmd->ic_addr =
    913 			    htole32(icp->icp_scr_seg[0].ds_addr +
    914 				    ICP_SCRATCH_UCMD);
    915 		} else {
    916 			struct icp_cachecmd *cc, *ucc;
    917 
    918 			cc = &ic->ic_cmd.cmd_packet.cc;
    919 			ucc = &ucmd->command.cmd_packet.cc;
    920 
    921 			cc->cc_deviceno = htole16(ucc->cc_deviceno);
    922 			cc->cc_blockno = htole32(ucc->cc_blockno);
    923 			cc->cc_blockcnt = htole32(ucc->cc_blockcnt);
    924 			cc->cc_addr = htole32(0xffffffffU);
    925 			cc->cc_nsgent = htole32(1);
    926 			cc->cc_sg[0].sg_addr =
    927 			    htole32(icp->icp_scr_seg[0].ds_addr +
    928 				    ICP_SCRATCH_UCMD);
    929 			cc->cc_sg[0].sg_len = htole32(cnt);
    930 		}
    931 	} else {
    932 		struct icp_rawcmd *rc, *urc;
    933 
    934 		rc = &ic->ic_cmd.cmd_packet.rc;
    935 		urc = &ucmd->command.cmd_packet.rc;
    936 
    937 		rc->rc_direction = htole32(urc->rc_direction);
    938 		rc->rc_sdata = htole32(0xffffffffU);
    939 		rc->rc_sdlen = htole32(urc->rc_sdlen);
    940 		rc->rc_clen = htole32(urc->rc_clen);
    941 		memcpy(rc->rc_cdb, urc->rc_cdb, sizeof(rc->rc_cdb));
    942 		rc->rc_target = urc->rc_target;
    943 		rc->rc_lun = urc->rc_lun;
    944 		rc->rc_bus = urc->rc_bus;
    945 		rc->rc_sense_len = htole32(urc->rc_sense_len);
    946 		rc->rc_sense_addr =
    947 		    htole32(icp->icp_scr_seg[0].ds_addr +
    948 			    ICP_SCRATCH_UCMD + urc->rc_sdlen);
    949 		rc->rc_nsgent = htole32(1);
    950 		rc->rc_sg[0].sg_addr =
    951 		    htole32(icp->icp_scr_seg[0].ds_addr + ICP_SCRATCH_UCMD);
    952 		rc->rc_sg[0].sg_len = htole32(cnt - urc->rc_sense_len);
    953 	}
    954 
    955 	ic->ic_service = ucmd->service;
    956 	ic->ic_cmdlen = sizeof(ic->ic_cmd);
    957 	ic->ic_context = &iu;
    958 
    959 	/*
    960 	 * XXX What units are ucmd->timeout in?  Until we know, we
    961 	 * XXX just pull a number out of thin air.
    962 	 */
    963 	if (__predict_false((error = icp_ccb_wait_user(icp, ic, 30000)) != 0))
    964 		aprint_error_dev(&icp->icp_dv, "error %d waiting for ucmd to complete\n",
    965 		    error);
    966 
    967 	/* icp_ucmd_intr() has updated ucmd. */
    968 	icp_ccb_free(icp, ic);
    969 
    970 	return (error);
    971 }
    972 
    973 struct icp_ccb *
    974 icp_ccb_alloc(struct icp_softc *icp)
    975 {
    976 	struct icp_ccb *ic;
    977 	int s;
    978 
    979 	s = splbio();
    980 	if (__predict_false((ic =
    981 			     SIMPLEQ_FIRST(&icp->icp_ccb_freelist)) == NULL)) {
    982 		splx(s);
    983 		return (NULL);
    984 	}
    985 	SIMPLEQ_REMOVE_HEAD(&icp->icp_ccb_freelist, ic_chain);
    986 	splx(s);
    987 
    988 	ic->ic_flags = IC_ALLOCED;
    989 	return (ic);
    990 }
    991 
    992 struct icp_ccb *
    993 icp_ccb_alloc_wait(struct icp_softc *icp)
    994 {
    995 	struct icp_ccb *ic;
    996 	int s;
    997 
    998 	s = splbio();
    999 	while ((ic = SIMPLEQ_FIRST(&icp->icp_ccb_freelist)) == NULL) {
   1000 		icp->icp_flags |= ICP_F_WAIT_CCB;
   1001 		(void) tsleep(&icp->icp_ccb_freelist, PRIBIO, "icpccb", 0);
   1002 	}
   1003 	SIMPLEQ_REMOVE_HEAD(&icp->icp_ccb_freelist, ic_chain);
   1004 	splx(s);
   1005 
   1006 	ic->ic_flags = IC_ALLOCED;
   1007 	return (ic);
   1008 }
   1009 
   1010 void
   1011 icp_ccb_free(struct icp_softc *icp, struct icp_ccb *ic)
   1012 {
   1013 	int s;
   1014 
   1015 	s = splbio();
   1016 	ic->ic_flags = 0;
   1017 	ic->ic_intr = NULL;
   1018 	SIMPLEQ_INSERT_HEAD(&icp->icp_ccb_freelist, ic, ic_chain);
   1019 	if (__predict_false((icp->icp_flags & ICP_F_WAIT_CCB) != 0)) {
   1020 		icp->icp_flags &= ~ICP_F_WAIT_CCB;
   1021 		wakeup(&icp->icp_ccb_freelist);
   1022 	}
   1023 	splx(s);
   1024 }
   1025 
   1026 void
   1027 icp_ccb_enqueue(struct icp_softc *icp, struct icp_ccb *ic)
   1028 {
   1029 	int s;
   1030 
   1031 	s = splbio();
   1032 
   1033 	if (ic != NULL) {
   1034 		if (__predict_false((ic->ic_flags & IC_UCMD) != 0))
   1035 			SIMPLEQ_INSERT_TAIL(&icp->icp_ucmd_queue, ic, ic_chain);
   1036 		else
   1037 			SIMPLEQ_INSERT_TAIL(&icp->icp_ccb_queue, ic, ic_chain);
   1038 	}
   1039 
   1040 	for (; icp->icp_qfreeze == 0;) {
   1041 		if (__predict_false((ic =
   1042 			    SIMPLEQ_FIRST(&icp->icp_ucmd_queue)) != NULL)) {
   1043 			struct icp_ucmd_ctx *iu = ic->ic_context;
   1044 			gdt_ucmd_t *ucmd = iu->iu_ucmd;
   1045 
   1046 			/*
   1047 			 * All user-generated commands share the same
   1048 			 * scratch space, so if one is already running,
   1049 			 * we have to stall the command queue.
   1050 			 */
   1051 			if (icp->icp_ucmd_ccb != NULL)
   1052 				break;
   1053 			if ((*icp->icp_test_busy)(icp))
   1054 				break;
   1055 			icp->icp_ucmd_ccb = ic;
   1056 
   1057 			if (iu->iu_cnt != 0) {
   1058 				memcpy((char *)icp->icp_scr + ICP_SCRATCH_UCMD,
   1059 				    ucmd->data, iu->iu_cnt);
   1060 				bus_dmamap_sync(icp->icp_dmat,
   1061 				    icp->icp_scr_dmamap,
   1062 				    ICP_SCRATCH_UCMD, iu->iu_cnt,
   1063 				    BUS_DMASYNC_PREREAD |
   1064 				    BUS_DMASYNC_PREWRITE);
   1065 			}
   1066 		} else if (__predict_true((ic =
   1067 				SIMPLEQ_FIRST(&icp->icp_ccb_queue)) != NULL)) {
   1068 			if ((*icp->icp_test_busy)(icp))
   1069 				break;
   1070 		} else {
   1071 			/* no command found */
   1072 			break;
   1073 		}
   1074 		icp_ccb_submit(icp, ic);
   1075 		if (__predict_false((ic->ic_flags & IC_UCMD) != 0))
   1076 			SIMPLEQ_REMOVE_HEAD(&icp->icp_ucmd_queue, ic_chain);
   1077 		else
   1078 			SIMPLEQ_REMOVE_HEAD(&icp->icp_ccb_queue, ic_chain);
   1079 	}
   1080 
   1081 	splx(s);
   1082 }
   1083 
   1084 int
   1085 icp_ccb_map(struct icp_softc *icp, struct icp_ccb *ic, void *data, int size,
   1086 	    int dir)
   1087 {
   1088 	struct icp_sg *sg;
   1089 	int nsegs, i, rv;
   1090 	bus_dmamap_t xfer;
   1091 
   1092 	xfer = ic->ic_xfer_map;
   1093 
   1094 	rv = bus_dmamap_load(icp->icp_dmat, xfer, data, size, NULL,
   1095 	    BUS_DMA_NOWAIT | BUS_DMA_STREAMING |
   1096 	    ((dir & IC_XFER_IN) ? BUS_DMA_READ : BUS_DMA_WRITE));
   1097 	if (rv != 0)
   1098 		return (rv);
   1099 
   1100 	nsegs = xfer->dm_nsegs;
   1101 	ic->ic_xfer_size = size;
   1102 	ic->ic_nsgent = nsegs;
   1103 	ic->ic_flags |= dir;
   1104 	sg = ic->ic_sg;
   1105 
   1106 	if (sg != NULL) {
   1107 		for (i = 0; i < nsegs; i++, sg++) {
   1108 			sg->sg_addr = htole32(xfer->dm_segs[i].ds_addr);
   1109 			sg->sg_len = htole32(xfer->dm_segs[i].ds_len);
   1110 		}
   1111 	} else if (nsegs > 1)
   1112 		panic("icp_ccb_map: no SG list specified, but nsegs > 1");
   1113 
   1114 	if ((dir & IC_XFER_OUT) != 0)
   1115 		i = BUS_DMASYNC_PREWRITE;
   1116 	else /* if ((dir & IC_XFER_IN) != 0) */
   1117 		i = BUS_DMASYNC_PREREAD;
   1118 
   1119 	bus_dmamap_sync(icp->icp_dmat, xfer, 0, ic->ic_xfer_size, i);
   1120 	return (0);
   1121 }
   1122 
   1123 void
   1124 icp_ccb_unmap(struct icp_softc *icp, struct icp_ccb *ic)
   1125 {
   1126 	int i;
   1127 
   1128 	if ((ic->ic_flags & IC_XFER_OUT) != 0)
   1129 		i = BUS_DMASYNC_POSTWRITE;
   1130 	else /* if ((ic->ic_flags & IC_XFER_IN) != 0) */
   1131 		i = BUS_DMASYNC_POSTREAD;
   1132 
   1133 	bus_dmamap_sync(icp->icp_dmat, ic->ic_xfer_map, 0, ic->ic_xfer_size, i);
   1134 	bus_dmamap_unload(icp->icp_dmat, ic->ic_xfer_map);
   1135 }
   1136 
   1137 int
   1138 icp_ccb_poll(struct icp_softc *icp, struct icp_ccb *ic, int timo)
   1139 {
   1140 	int s, rv;
   1141 
   1142 	s = splbio();
   1143 
   1144 	for (timo = ICP_BUSY_WAIT_MS * 100; timo != 0; timo--) {
   1145 		if (!(*icp->icp_test_busy)(icp))
   1146 			break;
   1147 		DELAY(10);
   1148 	}
   1149 	if (timo == 0) {
   1150 		printf("%s: submit: busy\n", device_xname(&icp->icp_dv));
   1151 		return (EAGAIN);
   1152 	}
   1153 
   1154 	icp_ccb_submit(icp, ic);
   1155 
   1156 	if (cold) {
   1157 		for (timo *= 10; timo != 0; timo--) {
   1158 			DELAY(100);
   1159 			icp_intr(icp);
   1160 			if ((ic->ic_flags & IC_COMPLETE) != 0)
   1161 				break;
   1162 		}
   1163 	} else {
   1164 		ic->ic_flags |= IC_WAITING;
   1165 		while ((ic->ic_flags & IC_COMPLETE) == 0) {
   1166 			if ((rv = tsleep(ic, PRIBIO, "icpwccb",
   1167 					 mstohz(timo))) != 0) {
   1168 				timo = 0;
   1169 				break;
   1170 			}
   1171 		}
   1172 	}
   1173 
   1174 	if (timo != 0) {
   1175 		if (ic->ic_status != ICP_S_OK) {
   1176 #ifdef ICP_DEBUG
   1177 			printf("%s: request failed; status=0x%04x\n",
   1178 			    device_xname(&icp->icp_dv), ic->ic_status);
   1179 #endif
   1180 			rv = EIO;
   1181 		} else
   1182 			rv = 0;
   1183 	} else {
   1184 		aprint_error_dev(&icp->icp_dv, "command timed out\n");
   1185 		rv = EIO;
   1186 	}
   1187 
   1188 	while ((*icp->icp_test_busy)(icp) != 0)
   1189 		DELAY(10);
   1190 
   1191 	splx(s);
   1192 
   1193 	return (rv);
   1194 }
   1195 
   1196 int
   1197 icp_ccb_wait(struct icp_softc *icp, struct icp_ccb *ic, int timo)
   1198 {
   1199 	int s, rv;
   1200 
   1201 	ic->ic_flags |= IC_WAITING;
   1202 
   1203 	s = splbio();
   1204 	icp_ccb_enqueue(icp, ic);
   1205 	while ((ic->ic_flags & IC_COMPLETE) == 0) {
   1206 		if ((rv = tsleep(ic, PRIBIO, "icpwccb", mstohz(timo))) != 0) {
   1207 			splx(s);
   1208 			return (rv);
   1209 		}
   1210 	}
   1211 	splx(s);
   1212 
   1213 	if (ic->ic_status != ICP_S_OK) {
   1214 		aprint_error_dev(&icp->icp_dv, "command failed; status=%x\n",
   1215 		    ic->ic_status);
   1216 		return (EIO);
   1217 	}
   1218 
   1219 	return (0);
   1220 }
   1221 
   1222 int
   1223 icp_ccb_wait_user(struct icp_softc *icp, struct icp_ccb *ic, int timo)
   1224 {
   1225 	int s, rv;
   1226 
   1227 	ic->ic_dv = &icp->icp_dv;
   1228 	ic->ic_intr = icp_ucmd_intr;
   1229 	ic->ic_flags |= IC_UCMD;
   1230 
   1231 	s = splbio();
   1232 	icp_ccb_enqueue(icp, ic);
   1233 	while ((ic->ic_flags & IC_COMPLETE) == 0) {
   1234 		if ((rv = tsleep(ic, PRIBIO, "icpwuccb", mstohz(timo))) != 0) {
   1235 			splx(s);
   1236 			return (rv);
   1237 		}
   1238 	}
   1239 	splx(s);
   1240 
   1241 	return (0);
   1242 }
   1243 
   1244 void
   1245 icp_ccb_submit(struct icp_softc *icp, struct icp_ccb *ic)
   1246 {
   1247 
   1248 	ic->ic_cmdlen = (ic->ic_cmdlen + 3) & ~3;
   1249 
   1250 	(*icp->icp_set_sema0)(icp);
   1251 	DELAY(10);
   1252 
   1253 	ic->ic_cmd.cmd_boardnode = htole32(ICP_LOCALBOARD);
   1254 	ic->ic_cmd.cmd_cmdindex = htole32(ic->ic_ident);
   1255 
   1256 	icp->icp_running++;
   1257 
   1258 	(*icp->icp_copy_cmd)(icp, ic);
   1259 	(*icp->icp_release_event)(icp, ic);
   1260 }
   1261 
   1262 int
   1263 icp_freeze(struct icp_softc *icp)
   1264 {
   1265 	int s, error = 0;
   1266 
   1267 	s = splbio();
   1268 	if (icp->icp_qfreeze++ == 0) {
   1269 		while (icp->icp_running != 0) {
   1270 			icp->icp_flags |= ICP_F_WAIT_FREEZE;
   1271 			error = tsleep(&icp->icp_qfreeze, PRIBIO|PCATCH,
   1272 			    "icpqfrz", 0);
   1273 			if (error != 0 && --icp->icp_qfreeze == 0 &&
   1274 			    ICP_HAS_WORK(icp)) {
   1275 				icp_ccb_enqueue(icp, NULL);
   1276 				break;
   1277 			}
   1278 		}
   1279 	}
   1280 	splx(s);
   1281 
   1282 	return (error);
   1283 }
   1284 
   1285 void
   1286 icp_unfreeze(struct icp_softc *icp)
   1287 {
   1288 	int s;
   1289 
   1290 	s = splbio();
   1291 	KDASSERT(icp->icp_qfreeze != 0);
   1292 	if (--icp->icp_qfreeze == 0 && ICP_HAS_WORK(icp))
   1293 		icp_ccb_enqueue(icp, NULL);
   1294 	splx(s);
   1295 }
   1296 
   1297 /* XXX Global - should be per-controller? XXX */
   1298 static gdt_evt_str icp_event_buffer[ICP_MAX_EVENTS];
   1299 static int icp_event_oldidx;
   1300 static int icp_event_lastidx;
   1301 
   1302 gdt_evt_str *
   1303 icp_store_event(struct icp_softc *icp, u_int16_t source, u_int16_t idx,
   1304     gdt_evt_data *evt)
   1305 {
   1306 	gdt_evt_str *e;
   1307 
   1308 	/* no source == no event */
   1309 	if (source == 0)
   1310 		return (NULL);
   1311 
   1312 	e = &icp_event_buffer[icp_event_lastidx];
   1313 	if (e->event_source == source && e->event_idx == idx &&
   1314 	    ((evt->size != 0 && e->event_data.size != 0 &&
   1315 	      memcmp(&e->event_data.eu, &evt->eu, evt->size) == 0) ||
   1316 	     (evt->size == 0 && e->event_data.size == 0 &&
   1317 	      strcmp((char *) e->event_data.event_string,
   1318 	      	     (char *) evt->event_string) == 0))) {
   1319 		e->last_stamp = time_second;
   1320 		e->same_count++;
   1321 	} else {
   1322 		if (icp_event_buffer[icp_event_lastidx].event_source != 0) {
   1323 			icp_event_lastidx++;
   1324 			if (icp_event_lastidx == ICP_MAX_EVENTS)
   1325 				icp_event_lastidx = 0;
   1326 			if (icp_event_lastidx == icp_event_oldidx) {
   1327 				icp_event_oldidx++;
   1328 				if (icp_event_oldidx == ICP_MAX_EVENTS)
   1329 					icp_event_oldidx = 0;
   1330 			}
   1331 		}
   1332 		e = &icp_event_buffer[icp_event_lastidx];
   1333 		e->event_source = source;
   1334 		e->event_idx = idx;
   1335 		e->first_stamp = e->last_stamp = time_second;
   1336 		e->same_count = 1;
   1337 		e->event_data = *evt;
   1338 		e->application = 0;
   1339 	}
   1340 	return (e);
   1341 }
   1342 
   1343 int
   1344 icp_read_event(struct icp_softc *icp, int handle, gdt_evt_str *estr)
   1345 {
   1346 	gdt_evt_str *e;
   1347 	int eindex, s;
   1348 
   1349 	s = splbio();
   1350 
   1351 	if (handle == -1)
   1352 		eindex = icp_event_oldidx;
   1353 	else
   1354 		eindex = handle;
   1355 
   1356 	estr->event_source = 0;
   1357 
   1358 	if (eindex < 0 || eindex >= ICP_MAX_EVENTS) {
   1359 		splx(s);
   1360 		return (eindex);
   1361 	}
   1362 
   1363 	e = &icp_event_buffer[eindex];
   1364 	if (e->event_source != 0) {
   1365 		if (eindex != icp_event_lastidx) {
   1366 			eindex++;
   1367 			if (eindex == ICP_MAX_EVENTS)
   1368 				eindex = 0;
   1369 		} else
   1370 			eindex = -1;
   1371 		memcpy(estr, e, sizeof(gdt_evt_str));
   1372 	}
   1373 
   1374 	splx(s);
   1375 
   1376 	return (eindex);
   1377 }
   1378 
   1379 void
   1380 icp_readapp_event(struct icp_softc *icp, u_int8_t application,
   1381     gdt_evt_str *estr)
   1382 {
   1383 	gdt_evt_str *e;
   1384 	int found = 0, eindex, s;
   1385 
   1386 	s = splbio();
   1387 
   1388 	eindex = icp_event_oldidx;
   1389 	for (;;) {
   1390 		e = &icp_event_buffer[eindex];
   1391 		if (e->event_source == 0)
   1392 			break;
   1393 		if ((e->application & application) == 0) {
   1394 			e->application |= application;
   1395 			found = 1;
   1396 			break;
   1397 		}
   1398 		if (eindex == icp_event_lastidx)
   1399 			break;
   1400 		eindex++;
   1401 		if (eindex == ICP_MAX_EVENTS)
   1402 			eindex = 0;
   1403 	}
   1404 	if (found)
   1405 		memcpy(estr, e, sizeof(gdt_evt_str));
   1406 	else
   1407 		estr->event_source = 0;
   1408 
   1409 	splx(s);
   1410 }
   1411 
   1412 void
   1413 icp_clear_events(struct icp_softc *icp)
   1414 {
   1415 	int s;
   1416 
   1417 	s = splbio();
   1418 	icp_event_oldidx = icp_event_lastidx = 0;
   1419 	memset(icp_event_buffer, 0, sizeof(icp_event_buffer));
   1420 	splx(s);
   1421 }
   1422