Home | History | Annotate | Line # | Download | only in ic
mpt_netbsd.c revision 1.16.6.1
      1 /*	$NetBSD: mpt_netbsd.c,v 1.16.6.1 2012/02/18 07:34:22 mrg Exp $	*/
      2 
      3 /*
      4  * Copyright (c) 2003 Wasabi Systems, Inc.
      5  * All rights reserved.
      6  *
      7  * Written by Jason R. Thorpe for Wasabi Systems, Inc.
      8  *
      9  * Redistribution and use in source and binary forms, with or without
     10  * modification, are permitted provided that the following conditions
     11  * are met:
     12  * 1. Redistributions of source code must retain the above copyright
     13  *    notice, this list of conditions and the following disclaimer.
     14  * 2. Redistributions in binary form must reproduce the above copyright
     15  *    notice, this list of conditions and the following disclaimer in the
     16  *    documentation and/or other materials provided with the distribution.
     17  * 3. All advertising materials mentioning features or use of this software
     18  *    must display the following acknowledgement:
     19  *	This product includes software developed for the NetBSD Project by
     20  *	Wasabi Systems, Inc.
     21  * 4. The name of Wasabi Systems, Inc. may not be used to endorse
     22  *    or promote products derived from this software without specific prior
     23  *    written permission.
     24  *
     25  * THIS SOFTWARE IS PROVIDED BY WASABI SYSTEMS, INC. ``AS IS'' AND
     26  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
     27  * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
     28  * PURPOSE ARE DISCLAIMED.  IN NO EVENT SHALL WASABI SYSTEMS, INC
     29  * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
     30  * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
     31  * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
     32  * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
     33  * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
     34  * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
     35  * POSSIBILITY OF SUCH DAMAGE.
     36  */
     37 
     38 /*
     39  * Copyright (c) 2000, 2001 by Greg Ansley
     40  * Partially derived from Matt Jacob's ISP driver.
     41  *
     42  * Redistribution and use in source and binary forms, with or without
     43  * modification, are permitted provided that the following conditions
     44  * are met:
     45  * 1. Redistributions of source code must retain the above copyright
     46  *    notice immediately at the beginning of the file, without modification,
     47  *    this list of conditions, and the following disclaimer.
     48  * 2. The name of the author may not be used to endorse or promote products
     49  *    derived from this software without specific prior written permission.
     50  *
     51  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
     52  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
     53  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
     54  * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE FOR
     55  * ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
     56  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
     57  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
     58  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
     59  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
     60  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
     61  * SUCH DAMAGE.
     62  */
     63 /*
     64  * Additional Copyright (c) 2002 by Matthew Jacob under same license.
     65  */
     66 
     67 /*
     68  * mpt_netbsd.c:
     69  *
     70  * NetBSD-specific routines for LSI Fusion adapters.  Includes some
     71  * bus_dma glue, and SCSIPI glue.
     72  *
     73  * Adapted from the FreeBSD "mpt" driver by Jason R. Thorpe for
     74  * Wasabi Systems, Inc.
     75  *
     76  * Additional contributions by Garrett D'Amore on behalf of TELES AG.
     77  */
     78 
     79 #include <sys/cdefs.h>
     80 __KERNEL_RCSID(0, "$NetBSD: mpt_netbsd.c,v 1.16.6.1 2012/02/18 07:34:22 mrg Exp $");
     81 
     82 #include <dev/ic/mpt.h>			/* pulls in all headers */
     83 
     84 static int	mpt_poll(mpt_softc_t *, struct scsipi_xfer *, int);
     85 static void	mpt_timeout(void *);
     86 static void	mpt_done(mpt_softc_t *, uint32_t);
     87 static void	mpt_run_xfer(mpt_softc_t *, struct scsipi_xfer *);
     88 static void	mpt_set_xfer_mode(mpt_softc_t *, struct scsipi_xfer_mode *);
     89 static void	mpt_get_xfer_mode(mpt_softc_t *, struct scsipi_periph *);
     90 static void	mpt_ctlop(mpt_softc_t *, void *vmsg, uint32_t);
     91 static void	mpt_event_notify_reply(mpt_softc_t *, MSG_EVENT_NOTIFY_REPLY *);
     92 
     93 static void	mpt_scsipi_request(struct scsipi_channel *,
     94 		    scsipi_adapter_req_t, void *);
     95 static void	mpt_minphys(struct buf *);
     96 
     97 void
     98 mpt_scsipi_attach(mpt_softc_t *mpt)
     99 {
    100 	struct scsipi_adapter *adapt = &mpt->sc_adapter;
    101 	struct scsipi_channel *chan = &mpt->sc_channel;
    102 	int maxq;
    103 
    104 	mpt->bus = 0;		/* XXX ?? */
    105 
    106 	maxq = (mpt->mpt_global_credits < MPT_MAX_REQUESTS(mpt)) ?
    107 	    mpt->mpt_global_credits : MPT_MAX_REQUESTS(mpt);
    108 
    109 	/* Fill in the scsipi_adapter. */
    110 	memset(adapt, 0, sizeof(*adapt));
    111 	adapt->adapt_dev = &mpt->sc_dev;
    112 	adapt->adapt_nchannels = 1;
    113 	adapt->adapt_openings = maxq - 2;	/* Reserve 2 for driver use*/
    114 	adapt->adapt_max_periph = maxq - 2;
    115 	adapt->adapt_request = mpt_scsipi_request;
    116 	adapt->adapt_minphys = mpt_minphys;
    117 
    118 	/* Fill in the scsipi_channel. */
    119 	memset(chan, 0, sizeof(*chan));
    120 	chan->chan_adapter = adapt;
    121 	chan->chan_bustype = &scsi_bustype;
    122 	chan->chan_channel = 0;
    123 	chan->chan_flags = 0;
    124 	chan->chan_nluns = 8;
    125 	chan->chan_ntargets = mpt->mpt_max_devices;
    126 	chan->chan_id = mpt->mpt_ini_id;
    127 
    128 	(void) config_found(&mpt->sc_dev, &mpt->sc_channel, scsiprint);
    129 }
    130 
    131 int
    132 mpt_dma_mem_alloc(mpt_softc_t *mpt)
    133 {
    134 	bus_dma_segment_t reply_seg, request_seg;
    135 	int reply_rseg, request_rseg;
    136 	bus_addr_t pptr, end;
    137 	char *vptr;
    138 	size_t len;
    139 	int error, i;
    140 
    141 	/* Check if we have already allocated the reply memory. */
    142 	if (mpt->reply != NULL)
    143 		return (0);
    144 
    145 	/*
    146 	 * Allocate the request pool.  This isn't really DMA'd memory,
    147 	 * but it's a convenient place to do it.
    148 	 */
    149 	len = sizeof(request_t) * MPT_MAX_REQUESTS(mpt);
    150 	mpt->request_pool = malloc(len, M_DEVBUF, M_WAITOK | M_ZERO);
    151 	if (mpt->request_pool == NULL) {
    152 		aprint_error_dev(&mpt->sc_dev, "unable to allocate request pool\n");
    153 		return (ENOMEM);
    154 	}
    155 
    156 	/*
    157 	 * Allocate DMA resources for reply buffers.
    158 	 */
    159 	error = bus_dmamem_alloc(mpt->sc_dmat, PAGE_SIZE, PAGE_SIZE, 0,
    160 	    &reply_seg, 1, &reply_rseg, 0);
    161 	if (error) {
    162 		aprint_error_dev(&mpt->sc_dev, "unable to allocate reply area, error = %d\n",
    163 		    error);
    164 		goto fail_0;
    165 	}
    166 
    167 	error = bus_dmamem_map(mpt->sc_dmat, &reply_seg, reply_rseg, PAGE_SIZE,
    168 	    (void **) &mpt->reply, BUS_DMA_COHERENT/*XXX*/);
    169 	if (error) {
    170 		aprint_error_dev(&mpt->sc_dev, "unable to map reply area, error = %d\n",
    171 		    error);
    172 		goto fail_1;
    173 	}
    174 
    175 	error = bus_dmamap_create(mpt->sc_dmat, PAGE_SIZE, 1, PAGE_SIZE,
    176 	    0, 0, &mpt->reply_dmap);
    177 	if (error) {
    178 		aprint_error_dev(&mpt->sc_dev, "unable to create reply DMA map, error = %d\n",
    179 		    error);
    180 		goto fail_2;
    181 	}
    182 
    183 	error = bus_dmamap_load(mpt->sc_dmat, mpt->reply_dmap, mpt->reply,
    184 	    PAGE_SIZE, NULL, 0);
    185 	if (error) {
    186 		aprint_error_dev(&mpt->sc_dev, "unable to load reply DMA map, error = %d\n",
    187 		    error);
    188 		goto fail_3;
    189 	}
    190 	mpt->reply_phys = mpt->reply_dmap->dm_segs[0].ds_addr;
    191 
    192 	/*
    193 	 * Allocate DMA resources for request buffers.
    194 	 */
    195 	error = bus_dmamem_alloc(mpt->sc_dmat, MPT_REQ_MEM_SIZE(mpt),
    196 	    PAGE_SIZE, 0, &request_seg, 1, &request_rseg, 0);
    197 	if (error) {
    198 		aprint_error_dev(&mpt->sc_dev, "unable to allocate request area, "
    199 		    "error = %d\n", error);
    200 		goto fail_4;
    201 	}
    202 
    203 	error = bus_dmamem_map(mpt->sc_dmat, &request_seg, request_rseg,
    204 	    MPT_REQ_MEM_SIZE(mpt), (void **) &mpt->request, 0);
    205 	if (error) {
    206 		aprint_error_dev(&mpt->sc_dev, "unable to map request area, error = %d\n",
    207 		    error);
    208 		goto fail_5;
    209 	}
    210 
    211 	error = bus_dmamap_create(mpt->sc_dmat, MPT_REQ_MEM_SIZE(mpt), 1,
    212 	    MPT_REQ_MEM_SIZE(mpt), 0, 0, &mpt->request_dmap);
    213 	if (error) {
    214 		aprint_error_dev(&mpt->sc_dev, "unable to create request DMA map, "
    215 		    "error = %d\n", error);
    216 		goto fail_6;
    217 	}
    218 
    219 	error = bus_dmamap_load(mpt->sc_dmat, mpt->request_dmap, mpt->request,
    220 	    MPT_REQ_MEM_SIZE(mpt), NULL, 0);
    221 	if (error) {
    222 		aprint_error_dev(&mpt->sc_dev, "unable to load request DMA map, error = %d\n",
    223 		    error);
    224 		goto fail_7;
    225 	}
    226 	mpt->request_phys = mpt->request_dmap->dm_segs[0].ds_addr;
    227 
    228 	pptr = mpt->request_phys;
    229 	vptr = (void *) mpt->request;
    230 	end = pptr + MPT_REQ_MEM_SIZE(mpt);
    231 
    232 	for (i = 0; pptr < end; i++) {
    233 		request_t *req = &mpt->request_pool[i];
    234 		req->index = i;
    235 
    236 		/* Store location of Request Data */
    237 		req->req_pbuf = pptr;
    238 		req->req_vbuf = vptr;
    239 
    240 		pptr += MPT_REQUEST_AREA;
    241 		vptr += MPT_REQUEST_AREA;
    242 
    243 		req->sense_pbuf = (pptr - MPT_SENSE_SIZE);
    244 		req->sense_vbuf = (vptr - MPT_SENSE_SIZE);
    245 
    246 		error = bus_dmamap_create(mpt->sc_dmat, MAXPHYS,
    247 		    MPT_SGL_MAX, MAXPHYS, 0, 0, &req->dmap);
    248 		if (error) {
    249 			aprint_error_dev(&mpt->sc_dev, "unable to create req %d DMA map, "
    250 			    "error = %d\n", i, error);
    251 			goto fail_8;
    252 		}
    253 	}
    254 
    255 	return (0);
    256 
    257  fail_8:
    258 	for (--i; i >= 0; i--) {
    259 		request_t *req = &mpt->request_pool[i];
    260 		if (req->dmap != NULL)
    261 			bus_dmamap_destroy(mpt->sc_dmat, req->dmap);
    262 	}
    263 	bus_dmamap_unload(mpt->sc_dmat, mpt->request_dmap);
    264  fail_7:
    265 	bus_dmamap_destroy(mpt->sc_dmat, mpt->request_dmap);
    266  fail_6:
    267 	bus_dmamem_unmap(mpt->sc_dmat, (void *)mpt->request, PAGE_SIZE);
    268  fail_5:
    269 	bus_dmamem_free(mpt->sc_dmat, &request_seg, request_rseg);
    270  fail_4:
    271 	bus_dmamap_unload(mpt->sc_dmat, mpt->reply_dmap);
    272  fail_3:
    273 	bus_dmamap_destroy(mpt->sc_dmat, mpt->reply_dmap);
    274  fail_2:
    275 	bus_dmamem_unmap(mpt->sc_dmat, (void *)mpt->reply, PAGE_SIZE);
    276  fail_1:
    277 	bus_dmamem_free(mpt->sc_dmat, &reply_seg, reply_rseg);
    278  fail_0:
    279 	free(mpt->request_pool, M_DEVBUF);
    280 
    281 	mpt->reply = NULL;
    282 	mpt->request = NULL;
    283 	mpt->request_pool = NULL;
    284 
    285 	return (error);
    286 }
    287 
    288 int
    289 mpt_intr(void *arg)
    290 {
    291 	mpt_softc_t *mpt = arg;
    292 	int nrepl = 0;
    293 	uint32_t reply;
    294 
    295 	if ((mpt_read(mpt, MPT_OFFSET_INTR_STATUS) & MPT_INTR_REPLY_READY) == 0)
    296 		return (0);
    297 
    298 	reply = mpt_pop_reply_queue(mpt);
    299 	while (reply != MPT_REPLY_EMPTY) {
    300 		nrepl++;
    301 		if (mpt->verbose > 1) {
    302 			if ((reply & MPT_CONTEXT_REPLY) != 0) {
    303 				/* Address reply; IOC has something to say */
    304 				mpt_print_reply(MPT_REPLY_PTOV(mpt, reply));
    305 			} else {
    306 				/* Context reply; all went well */
    307 				mpt_prt(mpt, "context %u reply OK", reply);
    308 			}
    309 		}
    310 		mpt_done(mpt, reply);
    311 		reply = mpt_pop_reply_queue(mpt);
    312 	}
    313 	return (nrepl != 0);
    314 }
    315 
    316 void
    317 mpt_prt(mpt_softc_t *mpt, const char *fmt, ...)
    318 {
    319 	va_list ap;
    320 
    321 	printf("%s: ", device_xname(&mpt->sc_dev));
    322 	va_start(ap, fmt);
    323 	vprintf(fmt, ap);
    324 	va_end(ap);
    325 	printf("\n");
    326 }
    327 
    328 static int
    329 mpt_poll(mpt_softc_t *mpt, struct scsipi_xfer *xs, int count)
    330 {
    331 
    332 	/* Timeouts are in msec, so we loop in 1000usec cycles */
    333 	while (count) {
    334 		mpt_intr(mpt);
    335 		if (xs->xs_status & XS_STS_DONE)
    336 			return (0);
    337 		delay(1000);		/* only happens in boot, so ok */
    338 		count--;
    339 	}
    340 	return (1);
    341 }
    342 
    343 static void
    344 mpt_timeout(void *arg)
    345 {
    346 	request_t *req = arg;
    347 	struct scsipi_xfer *xs = req->xfer;
    348 	struct scsipi_periph *periph = xs->xs_periph;
    349 	mpt_softc_t *mpt =
    350 	    (void *) periph->periph_channel->chan_adapter->adapt_dev;
    351 	uint32_t oseq;
    352 	int s;
    353 
    354 	scsipi_printaddr(periph);
    355 	printf("command timeout\n");
    356 
    357 	s = splbio();
    358 
    359 	oseq = req->sequence;
    360 	mpt->timeouts++;
    361 	if (mpt_intr(mpt)) {
    362 		if (req->sequence != oseq) {
    363 			mpt_prt(mpt, "recovered from command timeout");
    364 			splx(s);
    365 			return;
    366 		}
    367 	}
    368 	mpt_prt(mpt,
    369 	    "timeout on request index = 0x%x, seq = 0x%08x",
    370 	    req->index, req->sequence);
    371 	mpt_check_doorbell(mpt);
    372 	mpt_prt(mpt, "Status 0x%08x, Mask 0x%08x, Doorbell 0x%08x",
    373 	    mpt_read(mpt, MPT_OFFSET_INTR_STATUS),
    374 	    mpt_read(mpt, MPT_OFFSET_INTR_MASK),
    375 	    mpt_read(mpt, MPT_OFFSET_DOORBELL));
    376 	mpt_prt(mpt, "request state: %s", mpt_req_state(req->debug));
    377 	if (mpt->verbose > 1)
    378 		mpt_print_scsi_io_request((MSG_SCSI_IO_REQUEST *)req->req_vbuf);
    379 
    380 	/* XXX WHAT IF THE IOC IS STILL USING IT?? */
    381 	req->xfer = NULL;
    382 	mpt_free_request(mpt, req);
    383 
    384 	xs->error = XS_TIMEOUT;
    385 	scsipi_done(xs);
    386 
    387 	splx(s);
    388 }
    389 
    390 static void
    391 mpt_done(mpt_softc_t *mpt, uint32_t reply)
    392 {
    393 	struct scsipi_xfer *xs = NULL;
    394 	struct scsipi_periph *periph;
    395 	int index;
    396 	request_t *req;
    397 	MSG_REQUEST_HEADER *mpt_req;
    398 	MSG_SCSI_IO_REPLY *mpt_reply;
    399 
    400 	if (__predict_true((reply & MPT_CONTEXT_REPLY) == 0)) {
    401 		/* context reply (ok) */
    402 		mpt_reply = NULL;
    403 		index = reply & MPT_CONTEXT_MASK;
    404 	} else {
    405 		/* address reply (error) */
    406 
    407 		/* XXX BUS_DMASYNC_POSTREAD XXX */
    408 		mpt_reply = MPT_REPLY_PTOV(mpt, reply);
    409 		if (mpt->verbose > 1) {
    410 			uint32_t *pReply = (uint32_t *) mpt_reply;
    411 
    412 			mpt_prt(mpt, "Address Reply (index %u):",
    413 			    le32toh(mpt_reply->MsgContext) & 0xffff);
    414 			mpt_prt(mpt, "%08x %08x %08x %08x",
    415 			    pReply[0], pReply[1], pReply[2], pReply[3]);
    416 			mpt_prt(mpt, "%08x %08x %08x %08x",
    417 			    pReply[4], pReply[5], pReply[6], pReply[7]);
    418 			mpt_prt(mpt, "%08x %08x %08x %08x",
    419 			    pReply[8], pReply[9], pReply[10], pReply[11]);
    420 		}
    421 		index = le32toh(mpt_reply->MsgContext);
    422 	}
    423 
    424 	/*
    425 	 * Address reply with MessageContext high bit set.
    426 	 * This is most likely a notify message, so we try
    427 	 * to process it, then free it.
    428 	 */
    429 	if (__predict_false((index & 0x80000000) != 0)) {
    430 		if (mpt_reply != NULL)
    431 			mpt_ctlop(mpt, mpt_reply, reply);
    432 		else
    433 			mpt_prt(mpt, "mpt_done: index 0x%x, NULL reply", index);
    434 		return;
    435 	}
    436 
    437 	/* Did we end up with a valid index into the table? */
    438 	if (__predict_false(index < 0 || index >= MPT_MAX_REQUESTS(mpt))) {
    439 		mpt_prt(mpt, "mpt_done: invalid index (0x%x) in reply", index);
    440 		return;
    441 	}
    442 
    443 	req = &mpt->request_pool[index];
    444 
    445 	/* Make sure memory hasn't been trashed. */
    446 	if (__predict_false(req->index != index)) {
    447 		mpt_prt(mpt, "mpt_done: corrupted request_t (0x%x)", index);
    448 		return;
    449 	}
    450 
    451 	MPT_SYNC_REQ(mpt, req, BUS_DMASYNC_POSTREAD|BUS_DMASYNC_POSTWRITE);
    452 	mpt_req = req->req_vbuf;
    453 
    454 	/* Short cut for task management replies; nothing more for us to do. */
    455 	if (__predict_false(mpt_req->Function == MPI_FUNCTION_SCSI_TASK_MGMT)) {
    456 		if (mpt->verbose > 1)
    457 			mpt_prt(mpt, "mpt_done: TASK MGMT");
    458 		goto done;
    459 	}
    460 
    461 	if (__predict_false(mpt_req->Function == MPI_FUNCTION_PORT_ENABLE))
    462 		goto done;
    463 
    464 	/*
    465 	 * At this point, it had better be a SCSI I/O command, but don't
    466 	 * crash if it isn't.
    467 	 */
    468 	if (__predict_false(mpt_req->Function !=
    469 			    MPI_FUNCTION_SCSI_IO_REQUEST)) {
    470 		if (mpt->verbose > 1)
    471 			mpt_prt(mpt, "mpt_done: unknown Function 0x%x (0x%x)",
    472 			    mpt_req->Function, index);
    473 		goto done;
    474 	}
    475 
    476 	/* Recover scsipi_xfer from the request structure. */
    477 	xs = req->xfer;
    478 
    479 	/* Can't have a SCSI command without a scsipi_xfer. */
    480 	if (__predict_false(xs == NULL)) {
    481 		mpt_prt(mpt,
    482 		    "mpt_done: no scsipi_xfer, index = 0x%x, seq = 0x%08x",
    483 		    req->index, req->sequence);
    484 		mpt_prt(mpt, "request state: %s", mpt_req_state(req->debug));
    485 		mpt_prt(mpt, "mpt_request:");
    486 		mpt_print_scsi_io_request((MSG_SCSI_IO_REQUEST *)req->req_vbuf);
    487 
    488 		if (mpt_reply != NULL) {
    489 			mpt_prt(mpt, "mpt_reply:");
    490 			mpt_print_reply(mpt_reply);
    491 		} else {
    492 			mpt_prt(mpt, "context reply: 0x%08x", reply);
    493 		}
    494 		goto done;
    495 	}
    496 
    497 	callout_stop(&xs->xs_callout);
    498 
    499 	periph = xs->xs_periph;
    500 
    501 	/*
    502 	 * If we were a data transfer, unload the map that described
    503 	 * the data buffer.
    504 	 */
    505 	if (__predict_true(xs->datalen != 0)) {
    506 		bus_dmamap_sync(mpt->sc_dmat, req->dmap, 0,
    507 		    req->dmap->dm_mapsize,
    508 		    (xs->xs_control & XS_CTL_DATA_IN) ? BUS_DMASYNC_POSTREAD
    509 						      : BUS_DMASYNC_POSTWRITE);
    510 		bus_dmamap_unload(mpt->sc_dmat, req->dmap);
    511 	}
    512 
    513 	if (__predict_true(mpt_reply == NULL)) {
    514 		/*
    515 		 * Context reply; report that the command was
    516 		 * successful!
    517 		 *
    518 		 * Also report the xfer mode, if necessary.
    519 		 */
    520 		if (__predict_false(mpt->mpt_report_xfer_mode != 0)) {
    521 			if ((mpt->mpt_report_xfer_mode &
    522 			     (1 << periph->periph_target)) != 0)
    523 				mpt_get_xfer_mode(mpt, periph);
    524 		}
    525 		xs->error = XS_NOERROR;
    526 		xs->status = SCSI_OK;
    527 		xs->resid = 0;
    528 		mpt_free_request(mpt, req);
    529 		scsipi_done(xs);
    530 		return;
    531 	}
    532 
    533 	xs->status = mpt_reply->SCSIStatus;
    534 	switch (le16toh(mpt_reply->IOCStatus)) {
    535 	case MPI_IOCSTATUS_SCSI_DATA_OVERRUN:
    536 		xs->error = XS_DRIVER_STUFFUP;
    537 		break;
    538 
    539 	case MPI_IOCSTATUS_SCSI_DATA_UNDERRUN:
    540 		/*
    541 		 * Yikes!  Tagged queue full comes through this path!
    542 		 *
    543 		 * So we'll change it to a status error and anything
    544 		 * that returns status should probably be a status
    545 		 * error as well.
    546 		 */
    547 		xs->resid = xs->datalen - le32toh(mpt_reply->TransferCount);
    548 		if (mpt_reply->SCSIState &
    549 		    MPI_SCSI_STATE_NO_SCSI_STATUS) {
    550 			xs->error = XS_DRIVER_STUFFUP;
    551 			break;
    552 		}
    553 		/* FALLTHROUGH */
    554 	case MPI_IOCSTATUS_SUCCESS:
    555 	case MPI_IOCSTATUS_SCSI_RECOVERED_ERROR:
    556 		switch (xs->status) {
    557 		case SCSI_OK:
    558 			/* Report the xfer mode, if necessary. */
    559 			if ((mpt->mpt_report_xfer_mode &
    560 			     (1 << periph->periph_target)) != 0)
    561 				mpt_get_xfer_mode(mpt, periph);
    562 			xs->resid = 0;
    563 			break;
    564 
    565 		case SCSI_CHECK:
    566 			xs->error = XS_SENSE;
    567 			break;
    568 
    569 		case SCSI_BUSY:
    570 		case SCSI_QUEUE_FULL:
    571 			xs->error = XS_BUSY;
    572 			break;
    573 
    574 		default:
    575 			scsipi_printaddr(periph);
    576 			printf("invalid status code %d\n", xs->status);
    577 			xs->error = XS_DRIVER_STUFFUP;
    578 			break;
    579 		}
    580 		break;
    581 
    582 	case MPI_IOCSTATUS_BUSY:
    583 	case MPI_IOCSTATUS_INSUFFICIENT_RESOURCES:
    584 		xs->error = XS_RESOURCE_SHORTAGE;
    585 		break;
    586 
    587 	case MPI_IOCSTATUS_SCSI_INVALID_BUS:
    588 	case MPI_IOCSTATUS_SCSI_INVALID_TARGETID:
    589 	case MPI_IOCSTATUS_SCSI_DEVICE_NOT_THERE:
    590 		xs->error = XS_SELTIMEOUT;
    591 		break;
    592 
    593 	case MPI_IOCSTATUS_SCSI_RESIDUAL_MISMATCH:
    594 		xs->error = XS_DRIVER_STUFFUP;
    595 		break;
    596 
    597 	case MPI_IOCSTATUS_SCSI_TASK_TERMINATED:
    598 		/* XXX What should we do here? */
    599 		break;
    600 
    601 	case MPI_IOCSTATUS_SCSI_TASK_MGMT_FAILED:
    602 		/* XXX */
    603 		xs->error = XS_DRIVER_STUFFUP;
    604 		break;
    605 
    606 	case MPI_IOCSTATUS_SCSI_IOC_TERMINATED:
    607 		/* XXX */
    608 		xs->error = XS_DRIVER_STUFFUP;
    609 		break;
    610 
    611 	case MPI_IOCSTATUS_SCSI_EXT_TERMINATED:
    612 		/* XXX This is a bus-reset */
    613 		xs->error = XS_DRIVER_STUFFUP;
    614 		break;
    615 
    616 	default:
    617 		/* XXX unrecognized HBA error */
    618 		xs->error = XS_DRIVER_STUFFUP;
    619 		break;
    620 	}
    621 
    622 	if (mpt_reply->SCSIState & MPI_SCSI_STATE_AUTOSENSE_VALID) {
    623 		memcpy(&xs->sense.scsi_sense, req->sense_vbuf,
    624 		    sizeof(xs->sense.scsi_sense));
    625 	} else if (mpt_reply->SCSIState & MPI_SCSI_STATE_AUTOSENSE_FAILED) {
    626 		/*
    627 		 * This will cause the scsipi layer to issue
    628 		 * a REQUEST SENSE.
    629 		 */
    630 		if (xs->status == SCSI_CHECK)
    631 			xs->error = XS_BUSY;
    632 	}
    633 
    634  done:
    635 	/* If IOC done with this requeset, free it up. */
    636 	if (mpt_reply == NULL || (mpt_reply->MsgFlags & 0x80) == 0)
    637 		mpt_free_request(mpt, req);
    638 
    639 	/* If address reply, give the buffer back to the IOC. */
    640 	if (mpt_reply != NULL)
    641 		mpt_free_reply(mpt, (reply << 1));
    642 
    643 	if (xs != NULL)
    644 		scsipi_done(xs);
    645 }
    646 
    647 static void
    648 mpt_run_xfer(mpt_softc_t *mpt, struct scsipi_xfer *xs)
    649 {
    650 	struct scsipi_periph *periph = xs->xs_periph;
    651 	request_t *req;
    652 	MSG_SCSI_IO_REQUEST *mpt_req;
    653 	int error, s;
    654 
    655 	s = splbio();
    656 	req = mpt_get_request(mpt);
    657 	if (__predict_false(req == NULL)) {
    658 		/* This should happen very infrequently. */
    659 		xs->error = XS_RESOURCE_SHORTAGE;
    660 		scsipi_done(xs);
    661 		splx(s);
    662 		return;
    663 	}
    664 	splx(s);
    665 
    666 	/* Link the req and the scsipi_xfer. */
    667 	req->xfer = xs;
    668 
    669 	/* Now we build the command for the IOC */
    670 	mpt_req = req->req_vbuf;
    671 	memset(mpt_req, 0, sizeof(*mpt_req));
    672 
    673 	mpt_req->Function = MPI_FUNCTION_SCSI_IO_REQUEST;
    674 	mpt_req->Bus = mpt->bus;
    675 
    676 	mpt_req->SenseBufferLength =
    677 	    (sizeof(xs->sense.scsi_sense) < MPT_SENSE_SIZE) ?
    678 	    sizeof(xs->sense.scsi_sense) : MPT_SENSE_SIZE;
    679 
    680 	/*
    681 	 * We use the message context to find the request structure when
    682 	 * we get the command completion interrupt from the IOC.
    683 	 */
    684 	mpt_req->MsgContext = htole32(req->index);
    685 
    686 	/* Which physical device to do the I/O on. */
    687 	mpt_req->TargetID = periph->periph_target;
    688 	mpt_req->LUN[1] = periph->periph_lun;
    689 
    690 	/* Set the direction of the transfer. */
    691 	if (xs->xs_control & XS_CTL_DATA_IN)
    692 		mpt_req->Control = MPI_SCSIIO_CONTROL_READ;
    693 	else if (xs->xs_control & XS_CTL_DATA_OUT)
    694 		mpt_req->Control = MPI_SCSIIO_CONTROL_WRITE;
    695 	else
    696 		mpt_req->Control = MPI_SCSIIO_CONTROL_NODATATRANSFER;
    697 
    698 	/* Set the queue behavior. */
    699 	if (__predict_true((!mpt->is_scsi) ||
    700 			   (mpt->mpt_tag_enable &
    701 			    (1 << periph->periph_target)))) {
    702 		switch (XS_CTL_TAGTYPE(xs)) {
    703 		case XS_CTL_HEAD_TAG:
    704 			mpt_req->Control |= MPI_SCSIIO_CONTROL_HEADOFQ;
    705 			break;
    706 
    707 #if 0	/* XXX */
    708 		case XS_CTL_ACA_TAG:
    709 			mpt_req->Control |= MPI_SCSIIO_CONTROL_ACAQ;
    710 			break;
    711 #endif
    712 
    713 		case XS_CTL_ORDERED_TAG:
    714 			mpt_req->Control |= MPI_SCSIIO_CONTROL_ORDEREDQ;
    715 			break;
    716 
    717 		case XS_CTL_SIMPLE_TAG:
    718 			mpt_req->Control |= MPI_SCSIIO_CONTROL_SIMPLEQ;
    719 			break;
    720 
    721 		default:
    722 			if (mpt->is_scsi)
    723 				mpt_req->Control |= MPI_SCSIIO_CONTROL_UNTAGGED;
    724 			else
    725 				mpt_req->Control |= MPI_SCSIIO_CONTROL_SIMPLEQ;
    726 			break;
    727 		}
    728 	} else
    729 		mpt_req->Control |= MPI_SCSIIO_CONTROL_UNTAGGED;
    730 
    731 	if (__predict_false(mpt->is_scsi &&
    732 			    (mpt->mpt_disc_enable &
    733 			     (1 << periph->periph_target)) == 0))
    734 		mpt_req->Control |= MPI_SCSIIO_CONTROL_NO_DISCONNECT;
    735 
    736 	mpt_req->Control = htole32(mpt_req->Control);
    737 
    738 	/* Copy the SCSI command block into place. */
    739 	memcpy(mpt_req->CDB, xs->cmd, xs->cmdlen);
    740 
    741 	mpt_req->CDBLength = xs->cmdlen;
    742 	mpt_req->DataLength = htole32(xs->datalen);
    743 	mpt_req->SenseBufferLowAddr = htole32(req->sense_pbuf);
    744 
    745 	/*
    746 	 * Map the DMA transfer.
    747 	 */
    748 	if (xs->datalen) {
    749 		SGE_SIMPLE32 *se;
    750 
    751 		error = bus_dmamap_load(mpt->sc_dmat, req->dmap, xs->data,
    752 		    xs->datalen, NULL,
    753 		    ((xs->xs_control & XS_CTL_NOSLEEP) ? BUS_DMA_NOWAIT
    754 						       : BUS_DMA_WAITOK) |
    755 		    BUS_DMA_STREAMING |
    756 		    ((xs->xs_control & XS_CTL_DATA_IN) ? BUS_DMA_READ
    757 						       : BUS_DMA_WRITE));
    758 		switch (error) {
    759 		case 0:
    760 			break;
    761 
    762 		case ENOMEM:
    763 		case EAGAIN:
    764 			xs->error = XS_RESOURCE_SHORTAGE;
    765 			goto out_bad;
    766 
    767 		default:
    768 			xs->error = XS_DRIVER_STUFFUP;
    769 			mpt_prt(mpt, "error %d loading DMA map", error);
    770  out_bad:
    771 			s = splbio();
    772 			mpt_free_request(mpt, req);
    773 			scsipi_done(xs);
    774 			splx(s);
    775 			return;
    776 		}
    777 
    778 		if (req->dmap->dm_nsegs > MPT_NSGL_FIRST(mpt)) {
    779 			int seg, i, nleft = req->dmap->dm_nsegs;
    780 			uint32_t flags;
    781 			SGE_CHAIN32 *ce;
    782 
    783 			seg = 0;
    784 			flags = MPI_SGE_FLAGS_SIMPLE_ELEMENT;
    785 			if (xs->xs_control & XS_CTL_DATA_OUT)
    786 				flags |= MPI_SGE_FLAGS_HOST_TO_IOC;
    787 
    788 			se = (SGE_SIMPLE32 *) &mpt_req->SGL;
    789 			for (i = 0; i < MPT_NSGL_FIRST(mpt) - 1;
    790 			     i++, se++, seg++) {
    791 				uint32_t tf;
    792 
    793 				memset(se, 0, sizeof(*se));
    794 				se->Address =
    795 				    htole32(req->dmap->dm_segs[seg].ds_addr);
    796 				MPI_pSGE_SET_LENGTH(se,
    797 				    req->dmap->dm_segs[seg].ds_len);
    798 				tf = flags;
    799 				if (i == MPT_NSGL_FIRST(mpt) - 2)
    800 					tf |= MPI_SGE_FLAGS_LAST_ELEMENT;
    801 				MPI_pSGE_SET_FLAGS(se, tf);
    802 				se->FlagsLength = htole32(se->FlagsLength);
    803 				nleft--;
    804 			}
    805 
    806 			/*
    807 			 * Tell the IOC where to find the first chain element.
    808 			 */
    809 			mpt_req->ChainOffset =
    810 			    ((char *)se - (char *)mpt_req) >> 2;
    811 
    812 			/*
    813 			 * Until we're finished with all segments...
    814 			 */
    815 			while (nleft) {
    816 				int ntodo;
    817 
    818 				/*
    819 				 * Construct the chain element that points to
    820 				 * the next segment.
    821 				 */
    822 				ce = (SGE_CHAIN32 *) se++;
    823 				if (nleft > MPT_NSGL(mpt)) {
    824 					ntodo = MPT_NSGL(mpt) - 1;
    825 					ce->NextChainOffset = (MPT_RQSL(mpt) -
    826 					    sizeof(SGE_SIMPLE32)) >> 2;
    827 					ce->Length = htole16(MPT_NSGL(mpt)
    828 						* sizeof(SGE_SIMPLE32));
    829 				} else {
    830 					ntodo = nleft;
    831 					ce->NextChainOffset = 0;
    832 					ce->Length = htole16(ntodo
    833 						* sizeof(SGE_SIMPLE32));
    834 				}
    835 				ce->Address = htole32(req->req_pbuf +
    836 				    ((char *)se - (char *)mpt_req));
    837 				ce->Flags = MPI_SGE_FLAGS_CHAIN_ELEMENT;
    838 				for (i = 0; i < ntodo; i++, se++, seg++) {
    839 					uint32_t tf;
    840 
    841 					memset(se, 0, sizeof(*se));
    842 					se->Address = htole32(
    843 					    req->dmap->dm_segs[seg].ds_addr);
    844 					MPI_pSGE_SET_LENGTH(se,
    845 					    req->dmap->dm_segs[seg].ds_len);
    846 					tf = flags;
    847 					if (i == ntodo - 1) {
    848 						tf |=
    849 						    MPI_SGE_FLAGS_LAST_ELEMENT;
    850 						if (ce->NextChainOffset == 0) {
    851 							tf |=
    852 						    MPI_SGE_FLAGS_END_OF_LIST |
    853 						    MPI_SGE_FLAGS_END_OF_BUFFER;
    854 						}
    855 					}
    856 					MPI_pSGE_SET_FLAGS(se, tf);
    857 					se->FlagsLength =
    858 					    htole32(se->FlagsLength);
    859 					nleft--;
    860 				}
    861 			}
    862 			bus_dmamap_sync(mpt->sc_dmat, req->dmap, 0,
    863 			    req->dmap->dm_mapsize,
    864 			    (xs->xs_control & XS_CTL_DATA_IN) ?
    865 			    				BUS_DMASYNC_PREREAD
    866 						      : BUS_DMASYNC_PREWRITE);
    867 		} else {
    868 			int i;
    869 			uint32_t flags;
    870 
    871 			flags = MPI_SGE_FLAGS_SIMPLE_ELEMENT;
    872 			if (xs->xs_control & XS_CTL_DATA_OUT)
    873 				flags |= MPI_SGE_FLAGS_HOST_TO_IOC;
    874 
    875 			/* Copy the segments into our SG list. */
    876 			se = (SGE_SIMPLE32 *) &mpt_req->SGL;
    877 			for (i = 0; i < req->dmap->dm_nsegs;
    878 			     i++, se++) {
    879 				uint32_t tf;
    880 
    881 				memset(se, 0, sizeof(*se));
    882 				se->Address =
    883 				    htole32(req->dmap->dm_segs[i].ds_addr);
    884 				MPI_pSGE_SET_LENGTH(se,
    885 				    req->dmap->dm_segs[i].ds_len);
    886 				tf = flags;
    887 				if (i == req->dmap->dm_nsegs - 1) {
    888 					tf |=
    889 					    MPI_SGE_FLAGS_LAST_ELEMENT |
    890 					    MPI_SGE_FLAGS_END_OF_BUFFER |
    891 					    MPI_SGE_FLAGS_END_OF_LIST;
    892 				}
    893 				MPI_pSGE_SET_FLAGS(se, tf);
    894 				se->FlagsLength = htole32(se->FlagsLength);
    895 			}
    896 			bus_dmamap_sync(mpt->sc_dmat, req->dmap, 0,
    897 			    req->dmap->dm_mapsize,
    898 			    (xs->xs_control & XS_CTL_DATA_IN) ?
    899 			    				BUS_DMASYNC_PREREAD
    900 						      : BUS_DMASYNC_PREWRITE);
    901 		}
    902 	} else {
    903 		/*
    904 		 * No data to transfer; just make a single simple SGL
    905 		 * with zero length.
    906 		 */
    907 		SGE_SIMPLE32 *se = (SGE_SIMPLE32 *) &mpt_req->SGL;
    908 		memset(se, 0, sizeof(*se));
    909 		MPI_pSGE_SET_FLAGS(se,
    910 		    (MPI_SGE_FLAGS_LAST_ELEMENT | MPI_SGE_FLAGS_END_OF_BUFFER |
    911 		     MPI_SGE_FLAGS_SIMPLE_ELEMENT | MPI_SGE_FLAGS_END_OF_LIST));
    912 		se->FlagsLength = htole32(se->FlagsLength);
    913 	}
    914 
    915 	if (mpt->verbose > 1)
    916 		mpt_print_scsi_io_request(mpt_req);
    917 
    918 	s = splbio();
    919 	if (__predict_true((xs->xs_control & XS_CTL_POLL) == 0))
    920 		callout_reset(&xs->xs_callout,
    921 		    mstohz(xs->timeout), mpt_timeout, req);
    922 	mpt_send_cmd(mpt, req);
    923 	splx(s);
    924 
    925 	if (__predict_true((xs->xs_control & XS_CTL_POLL) == 0))
    926 		return;
    927 
    928 	/*
    929 	 * If we can't use interrupts, poll on completion.
    930 	 */
    931 	if (mpt_poll(mpt, xs, xs->timeout))
    932 		mpt_timeout(req);
    933 }
    934 
    935 static void
    936 mpt_set_xfer_mode(mpt_softc_t *mpt, struct scsipi_xfer_mode *xm)
    937 {
    938 	fCONFIG_PAGE_SCSI_DEVICE_1 tmp;
    939 
    940 	/*
    941 	 * Always allow disconnect; we don't have a way to disable
    942 	 * it right now, in any case.
    943 	 */
    944 	mpt->mpt_disc_enable |= (1 << xm->xm_target);
    945 
    946 	if (xm->xm_mode & PERIPH_CAP_TQING)
    947 		mpt->mpt_tag_enable |= (1 << xm->xm_target);
    948 	else
    949 		mpt->mpt_tag_enable &= ~(1 << xm->xm_target);
    950 
    951 	if (mpt->is_scsi) {
    952 		/*
    953 		 * SCSI transport settings only make any sense for
    954 		 * SCSI
    955 		 */
    956 
    957 		tmp = mpt->mpt_dev_page1[xm->xm_target];
    958 
    959 		/*
    960 		 * Set the wide/narrow parameter for the target.
    961 		 */
    962 		if (xm->xm_mode & PERIPH_CAP_WIDE16)
    963 			tmp.RequestedParameters |= MPI_SCSIDEVPAGE1_RP_WIDE;
    964 		else
    965 			tmp.RequestedParameters &= ~MPI_SCSIDEVPAGE1_RP_WIDE;
    966 
    967 		/*
    968 		 * Set the synchronous parameters for the target.
    969 		 *
    970 		 * XXX If we request sync transfers, we just go ahead and
    971 		 * XXX request the maximum available.  We need finer control
    972 		 * XXX in order to implement Domain Validation.
    973 		 */
    974 		tmp.RequestedParameters &= ~(MPI_SCSIDEVPAGE1_RP_MIN_SYNC_PERIOD_MASK |
    975 		    MPI_SCSIDEVPAGE1_RP_MAX_SYNC_OFFSET_MASK |
    976 		    MPI_SCSIDEVPAGE1_RP_DT | MPI_SCSIDEVPAGE1_RP_QAS |
    977 		    MPI_SCSIDEVPAGE1_RP_IU);
    978 		if (xm->xm_mode & PERIPH_CAP_SYNC) {
    979 			int factor, offset, np;
    980 
    981 			factor = (mpt->mpt_port_page0.Capabilities >> 8) & 0xff;
    982 			offset = (mpt->mpt_port_page0.Capabilities >> 16) & 0xff;
    983 			np = 0;
    984 			if (factor < 0x9) {
    985 				/* Ultra320 */
    986 				np |= MPI_SCSIDEVPAGE1_RP_QAS | MPI_SCSIDEVPAGE1_RP_IU;
    987 			}
    988 			if (factor < 0xa) {
    989 				/* at least Ultra160 */
    990 				np |= MPI_SCSIDEVPAGE1_RP_DT;
    991 			}
    992 			np |= (factor << 8) | (offset << 16);
    993 			tmp.RequestedParameters |= np;
    994 		}
    995 
    996 		host2mpt_config_page_scsi_device_1(&tmp);
    997 		if (mpt_write_cfg_page(mpt, xm->xm_target, &tmp.Header)) {
    998 			mpt_prt(mpt, "unable to write Device Page 1");
    999 			return;
   1000 		}
   1001 
   1002 		if (mpt_read_cfg_page(mpt, xm->xm_target, &tmp.Header)) {
   1003 			mpt_prt(mpt, "unable to read back Device Page 1");
   1004 			return;
   1005 		}
   1006 
   1007 		mpt2host_config_page_scsi_device_1(&tmp);
   1008 		mpt->mpt_dev_page1[xm->xm_target] = tmp;
   1009 		if (mpt->verbose > 1) {
   1010 			mpt_prt(mpt,
   1011 			    "SPI Target %d Page 1: RequestedParameters %x Config %x",
   1012 			    xm->xm_target,
   1013 			    mpt->mpt_dev_page1[xm->xm_target].RequestedParameters,
   1014 			    mpt->mpt_dev_page1[xm->xm_target].Configuration);
   1015 		}
   1016 	}
   1017 
   1018 	/*
   1019 	 * Make a note that we should perform an async callback at the
   1020 	 * end of the next successful command completion to report the
   1021 	 * negotiated transfer mode.
   1022 	 */
   1023 	mpt->mpt_report_xfer_mode |= (1 << xm->xm_target);
   1024 }
   1025 
   1026 static void
   1027 mpt_get_xfer_mode(mpt_softc_t *mpt, struct scsipi_periph *periph)
   1028 {
   1029 	fCONFIG_PAGE_SCSI_DEVICE_0 tmp;
   1030 	struct scsipi_xfer_mode xm;
   1031 	int period, offset;
   1032 
   1033 	tmp = mpt->mpt_dev_page0[periph->periph_target];
   1034 	host2mpt_config_page_scsi_device_0(&tmp);
   1035 	if (mpt_read_cfg_page(mpt, periph->periph_target, &tmp.Header)) {
   1036 		mpt_prt(mpt, "unable to read Device Page 0");
   1037 		return;
   1038 	}
   1039 	mpt2host_config_page_scsi_device_0(&tmp);
   1040 
   1041 	if (mpt->verbose > 1) {
   1042 		mpt_prt(mpt,
   1043 		    "SPI Tgt %d Page 0: NParms %x Information %x",
   1044 		    periph->periph_target,
   1045 		    tmp.NegotiatedParameters, tmp.Information);
   1046 	}
   1047 
   1048 	xm.xm_target = periph->periph_target;
   1049 	xm.xm_mode = 0;
   1050 
   1051 	if (tmp.NegotiatedParameters & MPI_SCSIDEVPAGE0_NP_WIDE)
   1052 		xm.xm_mode |= PERIPH_CAP_WIDE16;
   1053 
   1054 	period = (tmp.NegotiatedParameters >> 8) & 0xff;
   1055 	offset = (tmp.NegotiatedParameters >> 16) & 0xff;
   1056 	if (offset) {
   1057 		xm.xm_period = period;
   1058 		xm.xm_offset = offset;
   1059 		xm.xm_mode |= PERIPH_CAP_SYNC;
   1060 	}
   1061 
   1062 	/*
   1063 	 * Tagged queueing is all controlled by us; there is no
   1064 	 * other setting to query.
   1065 	 */
   1066 	if (mpt->mpt_tag_enable & (1 << periph->periph_target))
   1067 		xm.xm_mode |= PERIPH_CAP_TQING;
   1068 
   1069 	/*
   1070 	 * We're going to deliver the async event, so clear the marker.
   1071 	 */
   1072 	mpt->mpt_report_xfer_mode &= ~(1 << periph->periph_target);
   1073 
   1074 	scsipi_async_event(&mpt->sc_channel, ASYNC_EVENT_XFER_MODE, &xm);
   1075 }
   1076 
   1077 static void
   1078 mpt_ctlop(mpt_softc_t *mpt, void *vmsg, uint32_t reply)
   1079 {
   1080 	MSG_DEFAULT_REPLY *dmsg = vmsg;
   1081 
   1082 	switch (dmsg->Function) {
   1083 	case MPI_FUNCTION_EVENT_NOTIFICATION:
   1084 		mpt_event_notify_reply(mpt, vmsg);
   1085 		mpt_free_reply(mpt, (reply << 1));
   1086 		break;
   1087 
   1088 	case MPI_FUNCTION_EVENT_ACK:
   1089 		mpt_free_reply(mpt, (reply << 1));
   1090 		break;
   1091 
   1092 	case MPI_FUNCTION_PORT_ENABLE:
   1093 	    {
   1094 		MSG_PORT_ENABLE_REPLY *msg = vmsg;
   1095 		int index = le32toh(msg->MsgContext) & ~0x80000000;
   1096 		if (mpt->verbose > 1)
   1097 			mpt_prt(mpt, "enable port reply index %d", index);
   1098 		if (index >= 0 && index < MPT_MAX_REQUESTS(mpt)) {
   1099 			request_t *req = &mpt->request_pool[index];
   1100 			req->debug = REQ_DONE;
   1101 		}
   1102 		mpt_free_reply(mpt, (reply << 1));
   1103 		break;
   1104 	    }
   1105 
   1106 	case MPI_FUNCTION_CONFIG:
   1107 	    {
   1108 		MSG_CONFIG_REPLY *msg = vmsg;
   1109 		int index = le32toh(msg->MsgContext) & ~0x80000000;
   1110 		if (index >= 0 && index < MPT_MAX_REQUESTS(mpt)) {
   1111 			request_t *req = &mpt->request_pool[index];
   1112 			req->debug = REQ_DONE;
   1113 			req->sequence = reply;
   1114 		} else
   1115 			mpt_free_reply(mpt, (reply << 1));
   1116 		break;
   1117 	    }
   1118 
   1119 	default:
   1120 		mpt_prt(mpt, "unknown ctlop: 0x%x", dmsg->Function);
   1121 	}
   1122 }
   1123 
   1124 static void
   1125 mpt_event_notify_reply(mpt_softc_t *mpt, MSG_EVENT_NOTIFY_REPLY *msg)
   1126 {
   1127 
   1128 	switch (le32toh(msg->Event)) {
   1129 	case MPI_EVENT_LOG_DATA:
   1130 	    {
   1131 		int i;
   1132 
   1133 		/* Some error occurrerd that the Fusion wants logged. */
   1134 		mpt_prt(mpt, "EvtLogData: IOCLogInfo: 0x%08x", msg->IOCLogInfo);
   1135 		mpt_prt(mpt, "EvtLogData: Event Data:");
   1136 		for (i = 0; i < msg->EventDataLength; i++) {
   1137 			if ((i % 4) == 0)
   1138 				printf("%s:\t", device_xname(&mpt->sc_dev));
   1139 			printf("0x%08x%c", msg->Data[i],
   1140 			    ((i % 4) == 3) ? '\n' : ' ');
   1141 		}
   1142 		if ((i % 4) != 0)
   1143 			printf("\n");
   1144 		break;
   1145 	    }
   1146 
   1147 	case MPI_EVENT_UNIT_ATTENTION:
   1148 		mpt_prt(mpt, "Unit Attn: Bus 0x%02x Target 0x%02x",
   1149 		    (msg->Data[0] >> 8) & 0xff, msg->Data[0] & 0xff);
   1150 		break;
   1151 
   1152 	case MPI_EVENT_IOC_BUS_RESET:
   1153 		/* We generated a bus reset. */
   1154 		mpt_prt(mpt, "IOC Bus Reset Port %d",
   1155 		    (msg->Data[0] >> 8) & 0xff);
   1156 		break;
   1157 
   1158 	case MPI_EVENT_EXT_BUS_RESET:
   1159 		/* Someone else generated a bus reset. */
   1160 		mpt_prt(mpt, "External Bus Reset");
   1161 		/*
   1162 		 * These replies don't return EventData like the MPI
   1163 		 * spec says they do.
   1164 		 */
   1165 		/* XXX Send an async event? */
   1166 		break;
   1167 
   1168 	case MPI_EVENT_RESCAN:
   1169 		/*
   1170 		 * In general, thise means a device has been added
   1171 		 * to the loop.
   1172 		 */
   1173 		mpt_prt(mpt, "Rescan Port %d", (msg->Data[0] >> 8) & 0xff);
   1174 		/* XXX Send an async event? */
   1175 		break;
   1176 
   1177 	case MPI_EVENT_LINK_STATUS_CHANGE:
   1178 		mpt_prt(mpt, "Port %d: Link state %s",
   1179 		    (msg->Data[1] >> 8) & 0xff,
   1180 		    (msg->Data[0] & 0xff) == 0 ? "Failed" : "Active");
   1181 		break;
   1182 
   1183 	case MPI_EVENT_LOOP_STATE_CHANGE:
   1184 		switch ((msg->Data[0] >> 16) & 0xff) {
   1185 		case 0x01:
   1186 			mpt_prt(mpt,
   1187 			    "Port %d: FC Link Event: LIP(%02x,%02x) "
   1188 			    "(Loop Initialization)",
   1189 			    (msg->Data[1] >> 8) & 0xff,
   1190 			    (msg->Data[0] >> 8) & 0xff,
   1191 			    (msg->Data[0]     ) & 0xff);
   1192 			switch ((msg->Data[0] >> 8) & 0xff) {
   1193 			case 0xf7:
   1194 				if ((msg->Data[0] & 0xff) == 0xf7)
   1195 					mpt_prt(mpt, "\tDevice needs AL_PA");
   1196 				else
   1197 					mpt_prt(mpt, "\tDevice %02x doesn't "
   1198 					    "like FC performance",
   1199 					    msg->Data[0] & 0xff);
   1200 				break;
   1201 
   1202 			case 0xf8:
   1203 				if ((msg->Data[0] & 0xff) == 0xf7)
   1204 					mpt_prt(mpt, "\tDevice detected loop "
   1205 					    "failure before acquiring AL_PA");
   1206 				else
   1207 					mpt_prt(mpt, "\tDevice %02x detected "
   1208 					    "loop failure",
   1209 					    msg->Data[0] & 0xff);
   1210 				break;
   1211 
   1212 			default:
   1213 				mpt_prt(mpt, "\tDevice %02x requests that "
   1214 				    "device %02x reset itself",
   1215 				    msg->Data[0] & 0xff,
   1216 				    (msg->Data[0] >> 8) & 0xff);
   1217 				break;
   1218 			}
   1219 			break;
   1220 
   1221 		case 0x02:
   1222 			mpt_prt(mpt, "Port %d: FC Link Event: LPE(%02x,%02x) "
   1223 			    "(Loop Port Enable)",
   1224 			    (msg->Data[1] >> 8) & 0xff,
   1225 			    (msg->Data[0] >> 8) & 0xff,
   1226 			    (msg->Data[0]     ) & 0xff);
   1227 			break;
   1228 
   1229 		case 0x03:
   1230 			mpt_prt(mpt, "Port %d: FC Link Event: LPB(%02x,%02x) "
   1231 			    "(Loop Port Bypass)",
   1232 			    (msg->Data[1] >> 8) & 0xff,
   1233 			    (msg->Data[0] >> 8) & 0xff,
   1234 			    (msg->Data[0]     ) & 0xff);
   1235 			break;
   1236 
   1237 		default:
   1238 			mpt_prt(mpt, "Port %d: FC Link Event: "
   1239 			    "Unknown event (%02x %02x %02x)",
   1240 			    (msg->Data[1] >>  8) & 0xff,
   1241 			    (msg->Data[0] >> 16) & 0xff,
   1242 			    (msg->Data[0] >>  8) & 0xff,
   1243 			    (msg->Data[0]      ) & 0xff);
   1244 			break;
   1245 		}
   1246 		break;
   1247 
   1248 	case MPI_EVENT_LOGOUT:
   1249 		mpt_prt(mpt, "Port %d: FC Logout: N_PortID: %02x",
   1250 		    (msg->Data[1] >> 8) & 0xff, msg->Data[0]);
   1251 		break;
   1252 
   1253 	case MPI_EVENT_EVENT_CHANGE:
   1254 		/*
   1255 		 * This is just an acknowledgement of our
   1256 		 * mpt_send_event_request().
   1257 		 */
   1258 		break;
   1259 
   1260 	case MPI_EVENT_SAS_PHY_LINK_STATUS:
   1261 		switch ((msg->Data[0] >> 12) & 0x0f) {
   1262 		case 0x00:
   1263 			mpt_prt(mpt, "Phy %d: Link Status Unknown",
   1264 			    msg->Data[0] & 0xff);
   1265 			break;
   1266 		case 0x01:
   1267 			mpt_prt(mpt, "Phy %d: Link Disabled",
   1268 			    msg->Data[0] & 0xff);
   1269 			break;
   1270 		case 0x02:
   1271 			mpt_prt(mpt, "Phy %d: Failed Speed Negotiation",
   1272 			    msg->Data[0] & 0xff);
   1273 			break;
   1274 		case 0x03:
   1275 			mpt_prt(mpt, "Phy %d: SATA OOB Complete",
   1276 			    msg->Data[0] & 0xff);
   1277 			break;
   1278 		case 0x08:
   1279 			mpt_prt(mpt, "Phy %d: Link Rate 1.5 Gbps",
   1280 			    msg->Data[0] & 0xff);
   1281 			break;
   1282 		case 0x09:
   1283 			mpt_prt(mpt, "Phy %d: Link Rate 3.0 Gbps",
   1284 			    msg->Data[0] & 0xff);
   1285 			break;
   1286 		default:
   1287 			mpt_prt(mpt, "Phy %d: SAS Phy Link Status Event: "
   1288 			    "Unknown event (%0x)",
   1289 			    msg->Data[0] & 0xff, (msg->Data[0] >> 8) & 0xff);
   1290 		}
   1291 		break;
   1292 
   1293 	case MPI_EVENT_SAS_DEVICE_STATUS_CHANGE:
   1294 	case MPI_EVENT_SAS_DISCOVERY:
   1295 		/* ignore these events for now */
   1296 		break;
   1297 
   1298 	case MPI_EVENT_QUEUE_FULL:
   1299 		/* This can get a little chatty */
   1300 		if (mpt->verbose > 0)
   1301 			mpt_prt(mpt, "Queue Full Event");
   1302 		break;
   1303 
   1304 	default:
   1305 		mpt_prt(mpt, "Unknown async event: 0x%x", msg->Event);
   1306 		break;
   1307 	}
   1308 
   1309 	if (msg->AckRequired) {
   1310 		MSG_EVENT_ACK *ackp;
   1311 		request_t *req;
   1312 
   1313 		if ((req = mpt_get_request(mpt)) == NULL) {
   1314 			/* XXX XXX XXX XXXJRT */
   1315 			panic("mpt_event_notify_reply: unable to allocate "
   1316 			    "request structure");
   1317 		}
   1318 
   1319 		ackp = (MSG_EVENT_ACK *) req->req_vbuf;
   1320 		memset(ackp, 0, sizeof(*ackp));
   1321 		ackp->Function = MPI_FUNCTION_EVENT_ACK;
   1322 		ackp->Event = msg->Event;
   1323 		ackp->EventContext = msg->EventContext;
   1324 		ackp->MsgContext = htole32(req->index | 0x80000000);
   1325 		mpt_check_doorbell(mpt);
   1326 		mpt_send_cmd(mpt, req);
   1327 	}
   1328 }
   1329 
   1330 /* XXXJRT mpt_bus_reset() */
   1331 
   1332 /*****************************************************************************
   1333  * SCSI interface routines
   1334  *****************************************************************************/
   1335 
   1336 static void
   1337 mpt_scsipi_request(struct scsipi_channel *chan, scsipi_adapter_req_t req,
   1338     void *arg)
   1339 {
   1340 	struct scsipi_adapter *adapt = chan->chan_adapter;
   1341 	mpt_softc_t *mpt = (void *) adapt->adapt_dev;
   1342 
   1343 	switch (req) {
   1344 	case ADAPTER_REQ_RUN_XFER:
   1345 		mpt_run_xfer(mpt, (struct scsipi_xfer *) arg);
   1346 		return;
   1347 
   1348 	case ADAPTER_REQ_GROW_RESOURCES:
   1349 		/* Not supported. */
   1350 		return;
   1351 
   1352 	case ADAPTER_REQ_SET_XFER_MODE:
   1353 		mpt_set_xfer_mode(mpt, (struct scsipi_xfer_mode *) arg);
   1354 		return;
   1355 	}
   1356 }
   1357 
   1358 static void
   1359 mpt_minphys(struct buf *bp)
   1360 {
   1361 
   1362 /*
   1363  * Subtract one from the SGL limit, since we need an extra one to handle
   1364  * an non-page-aligned transfer.
   1365  */
   1366 #define	MPT_MAX_XFER	((MPT_SGL_MAX - 1) * PAGE_SIZE)
   1367 
   1368 	if (bp->b_bcount > MPT_MAX_XFER)
   1369 		bp->b_bcount = MPT_MAX_XFER;
   1370 	minphys(bp);
   1371 }
   1372