icp.c revision 1.33 1 /* $NetBSD: icp.c,v 1.33 2019/11/10 21:16:35 chs Exp $ */
2
3 /*-
4 * Copyright (c) 2002, 2003 The NetBSD Foundation, Inc.
5 * All rights reserved.
6 *
7 * This code is derived from software contributed to The NetBSD Foundation
8 * by Andrew Doran, and by Jason R. Thorpe of Wasabi Systems, Inc.
9 *
10 * Redistribution and use in source and binary forms, with or without
11 * modification, are permitted provided that the following conditions
12 * are met:
13 * 1. Redistributions of source code must retain the above copyright
14 * notice, this list of conditions and the following disclaimer.
15 * 2. Redistributions in binary form must reproduce the above copyright
16 * notice, this list of conditions and the following disclaimer in the
17 * documentation and/or other materials provided with the distribution.
18 *
19 * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
20 * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
21 * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
22 * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
23 * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
24 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
25 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
26 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
27 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
28 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
29 * POSSIBILITY OF SUCH DAMAGE.
30 */
31
32 /*
33 * Copyright (c) 1999, 2000 Niklas Hallqvist. All rights reserved.
34 *
35 * Redistribution and use in source and binary forms, with or without
36 * modification, are permitted provided that the following conditions
37 * are met:
38 * 1. Redistributions of source code must retain the above copyright
39 * notice, this list of conditions and the following disclaimer.
40 * 2. Redistributions in binary form must reproduce the above copyright
41 * notice, this list of conditions and the following disclaimer in the
42 * documentation and/or other materials provided with the distribution.
43 * 3. All advertising materials mentioning features or use of this software
44 * must display the following acknowledgement:
45 * This product includes software developed by Niklas Hallqvist.
46 * 4. The name of the author may not be used to endorse or promote products
47 * derived from this software without specific prior written permission.
48 *
49 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
50 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
51 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
52 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
53 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
54 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
55 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
56 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
57 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
58 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
59 *
60 * from OpenBSD: gdt_common.c,v 1.12 2001/07/04 06:43:18 niklas Exp
61 */
62
63 /*
64 * This driver would not have written if it was not for the hardware donations
65 * from both ICP-Vortex and ko.neT. I want to thank them for their support.
66 *
67 * Re-worked for NetBSD by Andrew Doran. Test hardware kindly supplied by
68 * Intel.
69 *
70 * Support for the ICP-Vortex management tools added by
71 * Jason R. Thorpe of Wasabi Systems, Inc., based on code
72 * provided by Achim Leubner <achim.leubner (at) intel.com>.
73 *
74 * Additional support for dynamic rescan of cacheservice drives by
75 * Jason R. Thorpe of Wasabi Systems, Inc.
76 */
77
78 #include <sys/cdefs.h>
79 __KERNEL_RCSID(0, "$NetBSD: icp.c,v 1.33 2019/11/10 21:16:35 chs Exp $");
80
81 #include <sys/param.h>
82 #include <sys/systm.h>
83 #include <sys/kernel.h>
84 #include <sys/device.h>
85 #include <sys/queue.h>
86 #include <sys/proc.h>
87 #include <sys/buf.h>
88 #include <sys/endian.h>
89 #include <sys/malloc.h>
90 #include <sys/disk.h>
91
92 #include <sys/bswap.h>
93 #include <sys/bus.h>
94
95 #include <dev/pci/pcireg.h>
96 #include <dev/pci/pcivar.h>
97 #include <dev/pci/pcidevs.h>
98
99 #include <dev/ic/icpreg.h>
100 #include <dev/ic/icpvar.h>
101
102 #include <dev/scsipi/scsipi_all.h>
103 #include <dev/scsipi/scsiconf.h>
104
105 #include "locators.h"
106
107 int icp_async_event(struct icp_softc *, int);
108 void icp_ccb_submit(struct icp_softc *icp, struct icp_ccb *ic);
109 void icp_chain(struct icp_softc *);
110 int icp_print(void *, const char *);
111 void icp_watchdog(void *);
112 void icp_ucmd_intr(struct icp_ccb *);
113 void icp_recompute_openings(struct icp_softc *);
114
115 int icp_count; /* total # of controllers, for ioctl interface */
116
117 /*
118 * Statistics for the ioctl interface to query.
119 *
120 * XXX Global. They should probably be made per-controller
121 * XXX at some point.
122 */
123 gdt_statist_t icp_stats;
124
125 int
126 icp_init(struct icp_softc *icp, const char *intrstr)
127 {
128 struct icp_attach_args icpa;
129 struct icp_binfo binfo;
130 struct icp_ccb *ic;
131 u_int16_t cdev_cnt;
132 int i, j, state, feat, nsegs, rv;
133 int locs[ICPCF_NLOCS];
134
135 state = 0;
136
137 if (intrstr != NULL)
138 aprint_normal_dev(icp->icp_dv, "interrupting at %s\n",
139 intrstr);
140
141 SIMPLEQ_INIT(&icp->icp_ccb_queue);
142 SIMPLEQ_INIT(&icp->icp_ccb_freelist);
143 SIMPLEQ_INIT(&icp->icp_ucmd_queue);
144 callout_init(&icp->icp_wdog_callout, 0);
145
146 /*
147 * Allocate a scratch area.
148 */
149 if (bus_dmamap_create(icp->icp_dmat, ICP_SCRATCH_SIZE, 1,
150 ICP_SCRATCH_SIZE, 0, BUS_DMA_NOWAIT | BUS_DMA_ALLOCNOW,
151 &icp->icp_scr_dmamap) != 0) {
152 aprint_error_dev(icp->icp_dv, "cannot create scratch dmamap\n");
153 return (1);
154 }
155 state++;
156
157 if (bus_dmamem_alloc(icp->icp_dmat, ICP_SCRATCH_SIZE, PAGE_SIZE, 0,
158 icp->icp_scr_seg, 1, &nsegs, BUS_DMA_NOWAIT) != 0) {
159 aprint_error_dev(icp->icp_dv, "cannot alloc scratch dmamem\n");
160 goto bail_out;
161 }
162 state++;
163
164 if (bus_dmamem_map(icp->icp_dmat, icp->icp_scr_seg, nsegs,
165 ICP_SCRATCH_SIZE, &icp->icp_scr, 0)) {
166 aprint_error_dev(icp->icp_dv, "cannot map scratch dmamem\n");
167 goto bail_out;
168 }
169 state++;
170
171 if (bus_dmamap_load(icp->icp_dmat, icp->icp_scr_dmamap, icp->icp_scr,
172 ICP_SCRATCH_SIZE, NULL, BUS_DMA_NOWAIT)) {
173 aprint_error_dev(icp->icp_dv, "cannot load scratch dmamap\n");
174 goto bail_out;
175 }
176 state++;
177
178 /*
179 * Allocate and initialize the command control blocks.
180 */
181 ic = malloc(sizeof(*ic) * ICP_NCCBS, M_DEVBUF, M_WAITOK | M_ZERO);
182 state++;
183
184 for (i = 0; i < ICP_NCCBS; i++, ic++) {
185 /*
186 * The first two command indexes have special meanings, so
187 * we can't use them.
188 */
189 ic->ic_ident = i + 2;
190 rv = bus_dmamap_create(icp->icp_dmat, ICP_MAX_XFER,
191 ICP_MAXSG, ICP_MAX_XFER, 0,
192 BUS_DMA_NOWAIT | BUS_DMA_ALLOCNOW,
193 &ic->ic_xfer_map);
194 if (rv != 0)
195 break;
196 icp->icp_nccbs++;
197 icp_ccb_free(icp, ic);
198 }
199 #ifdef DIAGNOSTIC
200 if (icp->icp_nccbs != ICP_NCCBS)
201 aprint_error_dev(icp->icp_dv, "%d/%d CCBs usable\n",
202 icp->icp_nccbs, ICP_NCCBS);
203 #endif
204
205 /*
206 * Initialize the controller.
207 */
208 if (!icp_cmd(icp, ICP_SCREENSERVICE, ICP_INIT, 0, 0, 0)) {
209 aprint_error_dev(icp->icp_dv, "screen service init error %d\n",
210 icp->icp_status);
211 goto bail_out;
212 }
213
214 if (!icp_cmd(icp, ICP_CACHESERVICE, ICP_INIT, ICP_LINUX_OS, 0, 0)) {
215 aprint_error_dev(icp->icp_dv, "cache service init error %d\n",
216 icp->icp_status);
217 goto bail_out;
218 }
219
220 icp_cmd(icp, ICP_CACHESERVICE, ICP_UNFREEZE_IO, 0, 0, 0);
221
222 if (!icp_cmd(icp, ICP_CACHESERVICE, ICP_MOUNT, 0xffff, 1, 0)) {
223 aprint_error_dev(icp->icp_dv, "cache service mount error %d\n",
224 icp->icp_status);
225 goto bail_out;
226 }
227
228 if (!icp_cmd(icp, ICP_CACHESERVICE, ICP_INIT, ICP_LINUX_OS, 0, 0)) {
229 aprint_error_dev(icp->icp_dv, "cache service post-mount init error %d\n",
230 icp->icp_status);
231 goto bail_out;
232 }
233 cdev_cnt = (u_int16_t)icp->icp_info;
234 icp->icp_fw_vers = icp->icp_service;
235
236 if (!icp_cmd(icp, ICP_SCSIRAWSERVICE, ICP_INIT, 0, 0, 0)) {
237 aprint_error_dev(icp->icp_dv, "raw service init error %d\n",
238 icp->icp_status);
239 goto bail_out;
240 }
241
242 /*
243 * Set/get raw service features (scatter/gather).
244 */
245 feat = 0;
246 if (icp_cmd(icp, ICP_SCSIRAWSERVICE, ICP_SET_FEAT, ICP_SCATTER_GATHER,
247 0, 0))
248 if (icp_cmd(icp, ICP_SCSIRAWSERVICE, ICP_GET_FEAT, 0, 0, 0))
249 feat = icp->icp_info;
250
251 if ((feat & ICP_SCATTER_GATHER) == 0) {
252 #ifdef DIAGNOSTIC
253 aprint_normal_dev(icp->icp_dv,
254 "scatter/gather not supported (raw service)\n");
255 #endif
256 } else
257 icp->icp_features |= ICP_FEAT_RAWSERVICE;
258
259 /*
260 * Set/get cache service features (scatter/gather).
261 */
262 feat = 0;
263 if (icp_cmd(icp, ICP_CACHESERVICE, ICP_SET_FEAT, 0,
264 ICP_SCATTER_GATHER, 0))
265 if (icp_cmd(icp, ICP_CACHESERVICE, ICP_GET_FEAT, 0, 0, 0))
266 feat = icp->icp_info;
267
268 if ((feat & ICP_SCATTER_GATHER) == 0) {
269 #ifdef DIAGNOSTIC
270 aprint_normal_dev(icp->icp_dv,
271 "scatter/gather not supported (cache service)\n");
272 #endif
273 } else
274 icp->icp_features |= ICP_FEAT_CACHESERVICE;
275
276 /*
277 * Pull some information from the board and dump.
278 */
279 if (!icp_cmd(icp, ICP_CACHESERVICE, ICP_IOCTL, ICP_BOARD_INFO,
280 ICP_INVALID_CHANNEL, sizeof(struct icp_binfo))) {
281 aprint_error_dev(icp->icp_dv, "unable to retrive board info\n");
282 goto bail_out;
283 }
284 memcpy(&binfo, icp->icp_scr, sizeof(binfo));
285
286 aprint_normal_dev(icp->icp_dv,
287 "model <%s>, firmware <%s>, %d channel(s), %dMB memory\n",
288 binfo.bi_type_string, binfo.bi_raid_string,
289 binfo.bi_chan_count, le32toh(binfo.bi_memsize) >> 20);
290
291 /*
292 * Determine the number of devices, and number of openings per
293 * device.
294 */
295 if (icp->icp_features & ICP_FEAT_CACHESERVICE) {
296 for (j = 0; j < cdev_cnt && j < ICP_MAX_HDRIVES; j++) {
297 if (!icp_cmd(icp, ICP_CACHESERVICE, ICP_INFO, j, 0,
298 0))
299 continue;
300
301 icp->icp_cdr[j].cd_size = icp->icp_info;
302 if (icp->icp_cdr[j].cd_size != 0)
303 icp->icp_ndevs++;
304
305 if (icp_cmd(icp, ICP_CACHESERVICE, ICP_DEVTYPE, j, 0,
306 0))
307 icp->icp_cdr[j].cd_type = icp->icp_info;
308 }
309 }
310
311 if (icp->icp_features & ICP_FEAT_RAWSERVICE) {
312 icp->icp_nchan = binfo.bi_chan_count;
313 icp->icp_ndevs += icp->icp_nchan;
314 }
315
316 icp_recompute_openings(icp);
317
318 /*
319 * Attach SCSI channels.
320 */
321 if (icp->icp_features & ICP_FEAT_RAWSERVICE) {
322 struct icp_ioc_version *iv;
323 struct icp_rawioc *ri;
324 struct icp_getch *gc;
325
326 iv = (struct icp_ioc_version *)icp->icp_scr;
327 iv->iv_version = htole32(ICP_IOC_NEWEST);
328 iv->iv_listents = ICP_MAXBUS;
329 iv->iv_firstchan = 0;
330 iv->iv_lastchan = ICP_MAXBUS - 1;
331 iv->iv_listoffset = htole32(sizeof(*iv));
332
333 if (icp_cmd(icp, ICP_CACHESERVICE, ICP_IOCTL,
334 ICP_IOCHAN_RAW_DESC, ICP_INVALID_CHANNEL,
335 sizeof(*iv) + ICP_MAXBUS * sizeof(*ri))) {
336 ri = (struct icp_rawioc *)(iv + 1);
337 for (j = 0; j < binfo.bi_chan_count; j++, ri++)
338 icp->icp_bus_id[j] = ri->ri_procid;
339 } else {
340 /*
341 * Fall back to the old method.
342 */
343 gc = (struct icp_getch *)icp->icp_scr;
344
345 for (j = 0; j < binfo.bi_chan_count; j++) {
346 if (!icp_cmd(icp, ICP_CACHESERVICE, ICP_IOCTL,
347 ICP_SCSI_CHAN_CNT | ICP_L_CTRL_PATTERN,
348 ICP_IO_CHANNEL | ICP_INVALID_CHANNEL,
349 sizeof(*gc))) {
350 aprint_error_dev(icp->icp_dv,
351 "unable to get chan info");
352 goto bail_out;
353 }
354 icp->icp_bus_id[j] = gc->gc_scsiid;
355 }
356 }
357
358 for (j = 0; j < binfo.bi_chan_count; j++) {
359 if (icp->icp_bus_id[j] > ICP_MAXID_FC)
360 icp->icp_bus_id[j] = ICP_MAXID_FC;
361
362 icpa.icpa_unit = j + ICPA_UNIT_SCSI;
363
364 locs[ICPCF_UNIT] = j + ICPA_UNIT_SCSI;
365
366 icp->icp_children[icpa.icpa_unit] =
367 config_found_sm_loc(icp->icp_dv, "icp", locs,
368 &icpa, icp_print, config_stdsubmatch);
369 }
370 }
371
372 /*
373 * Attach cache devices.
374 */
375 if (icp->icp_features & ICP_FEAT_CACHESERVICE) {
376 for (j = 0; j < cdev_cnt && j < ICP_MAX_HDRIVES; j++) {
377 if (icp->icp_cdr[j].cd_size == 0)
378 continue;
379
380 icpa.icpa_unit = j;
381
382 locs[ICPCF_UNIT] = j;
383
384 icp->icp_children[icpa.icpa_unit] =
385 config_found_sm_loc(icp->icp_dv, "icp", locs,
386 &icpa, icp_print, config_stdsubmatch);
387 }
388 }
389
390 /*
391 * Start the watchdog.
392 */
393 icp_watchdog(icp);
394
395 /*
396 * Count the controller, and we're done!
397 */
398 if (icp_count++ == 0)
399 mutex_init(&icp_ioctl_mutex, MUTEX_DEFAULT, IPL_NONE);
400
401 return (0);
402
403 bail_out:
404 if (state > 4)
405 for (j = 0; j < i; j++)
406 bus_dmamap_destroy(icp->icp_dmat,
407 icp->icp_ccbs[j].ic_xfer_map);
408 if (state > 3)
409 free(icp->icp_ccbs, M_DEVBUF);
410 if (state > 2)
411 bus_dmamap_unload(icp->icp_dmat, icp->icp_scr_dmamap);
412 if (state > 1)
413 bus_dmamem_unmap(icp->icp_dmat, icp->icp_scr,
414 ICP_SCRATCH_SIZE);
415 if (state > 0)
416 bus_dmamem_free(icp->icp_dmat, icp->icp_scr_seg, nsegs);
417 bus_dmamap_destroy(icp->icp_dmat, icp->icp_scr_dmamap);
418
419 return (1);
420 }
421
422 void
423 icp_register_servicecb(struct icp_softc *icp, int unit,
424 const struct icp_servicecb *cb)
425 {
426
427 icp->icp_servicecb[unit] = cb;
428 }
429
430 void
431 icp_rescan(struct icp_softc *icp, int unit)
432 {
433 struct icp_attach_args icpa;
434 u_int newsize, newtype;
435 int locs[ICPCF_NLOCS];
436
437 /*
438 * NOTE: It is very important that the queue be frozen and not
439 * commands running when this is called. The ioctl mutex must
440 * also be held.
441 */
442
443 KASSERT(icp->icp_qfreeze != 0);
444 KASSERT(icp->icp_running == 0);
445 KASSERT(unit < ICP_MAX_HDRIVES);
446
447 if (!icp_cmd(icp, ICP_CACHESERVICE, ICP_INFO, unit, 0, 0)) {
448 #ifdef ICP_DEBUG
449 printf("%s: rescan: unit %d ICP_INFO failed -> 0x%04x\n",
450 device_xname(icp->icp_dv), unit, icp->icp_status);
451 #endif
452 goto gone;
453 }
454 if ((newsize = icp->icp_info) == 0) {
455 #ifdef ICP_DEBUG
456 printf("%s: rescan: unit %d has zero size\n",
457 device_xname(icp->icp_dv), unit);
458 #endif
459 gone:
460 /*
461 * Host drive is no longer present; detach if a child
462 * is currently there.
463 */
464 if (icp->icp_cdr[unit].cd_size != 0)
465 icp->icp_ndevs--;
466 icp->icp_cdr[unit].cd_size = 0;
467 if (icp->icp_children[unit] != NULL) {
468 (void) config_detach(icp->icp_children[unit],
469 DETACH_FORCE);
470 icp->icp_children[unit] = NULL;
471 }
472 return;
473 }
474
475 if (icp_cmd(icp, ICP_CACHESERVICE, ICP_DEVTYPE, unit, 0, 0))
476 newtype = icp->icp_info;
477 else {
478 #ifdef ICP_DEBUG
479 printf("%s: rescan: unit %d ICP_DEVTYPE failed\n",
480 device_xname(icp->icp_dv), unit);
481 #endif
482 newtype = 0; /* XXX? */
483 }
484
485 #ifdef ICP_DEBUG
486 printf("%s: rescan: unit %d old %u/%u, new %u/%u\n",
487 device_xname(icp->icp_dv), unit, icp->icp_cdr[unit].cd_size,
488 icp->icp_cdr[unit].cd_type, newsize, newtype);
489 #endif
490
491 /*
492 * If the type or size changed, detach any old child (if it exists)
493 * and attach a new one.
494 */
495 if (icp->icp_children[unit] == NULL ||
496 newsize != icp->icp_cdr[unit].cd_size ||
497 newtype != icp->icp_cdr[unit].cd_type) {
498 if (icp->icp_cdr[unit].cd_size == 0)
499 icp->icp_ndevs++;
500 icp->icp_cdr[unit].cd_size = newsize;
501 icp->icp_cdr[unit].cd_type = newtype;
502 if (icp->icp_children[unit] != NULL)
503 (void) config_detach(icp->icp_children[unit],
504 DETACH_FORCE);
505
506 icpa.icpa_unit = unit;
507
508 locs[ICPCF_UNIT] = unit;
509
510 icp->icp_children[unit] = config_found_sm_loc(icp->icp_dv,
511 "icp", locs, &icpa, icp_print, config_stdsubmatch);
512 }
513
514 icp_recompute_openings(icp);
515 }
516
517 void
518 icp_rescan_all(struct icp_softc *icp)
519 {
520 int unit;
521 u_int16_t cdev_cnt;
522
523 /*
524 * This is the old method of rescanning the host drives. We
525 * start by reinitializing the cache service.
526 */
527 if (!icp_cmd(icp, ICP_CACHESERVICE, ICP_INIT, ICP_LINUX_OS, 0, 0)) {
528 printf("%s: unable to re-initialize cache service for rescan\n",
529 device_xname(icp->icp_dv));
530 return;
531 }
532 cdev_cnt = (u_int16_t) icp->icp_info;
533
534 /* For each host drive, do the new-style rescan. */
535 for (unit = 0; unit < cdev_cnt && unit < ICP_MAX_HDRIVES; unit++)
536 icp_rescan(icp, unit);
537
538 /* Now detach anything in the slots after cdev_cnt. */
539 for (; unit < ICP_MAX_HDRIVES; unit++) {
540 if (icp->icp_cdr[unit].cd_size != 0) {
541 #ifdef ICP_DEBUG
542 printf("%s: rescan all: unit %d < new cdev_cnt (%d)\n",
543 device_xname(icp->icp_dv), unit, cdev_cnt);
544 #endif
545 icp->icp_ndevs--;
546 icp->icp_cdr[unit].cd_size = 0;
547 if (icp->icp_children[unit] != NULL) {
548 (void) config_detach(icp->icp_children[unit],
549 DETACH_FORCE);
550 icp->icp_children[unit] = NULL;
551 }
552 }
553 }
554
555 icp_recompute_openings(icp);
556 }
557
558 void
559 icp_recompute_openings(struct icp_softc *icp)
560 {
561 int unit, openings;
562
563 if (icp->icp_ndevs != 0)
564 openings =
565 (icp->icp_nccbs - ICP_NCCB_RESERVE) / icp->icp_ndevs;
566 else
567 openings = 0;
568 if (openings == icp->icp_openings)
569 return;
570 icp->icp_openings = openings;
571
572 #ifdef ICP_DEBUG
573 printf("%s: %d device%s, %d openings per device\n",
574 device_xname(icp->icp_dv), icp->icp_ndevs,
575 icp->icp_ndevs == 1 ? "" : "s", icp->icp_openings);
576 #endif
577
578 for (unit = 0; unit < ICP_MAX_HDRIVES + ICP_MAXBUS; unit++) {
579 if (icp->icp_children[unit] != NULL)
580 (*icp->icp_servicecb[unit]->iscb_openings)(
581 icp->icp_children[unit], icp->icp_openings);
582 }
583 }
584
585 void
586 icp_watchdog(void *cookie)
587 {
588 struct icp_softc *icp;
589 int s;
590
591 icp = cookie;
592
593 s = splbio();
594 icp_intr(icp);
595 if (ICP_HAS_WORK(icp))
596 icp_ccb_enqueue(icp, NULL);
597 splx(s);
598
599 callout_reset(&icp->icp_wdog_callout, hz * ICP_WATCHDOG_FREQ,
600 icp_watchdog, icp);
601 }
602
603 int
604 icp_print(void *aux, const char *pnp)
605 {
606 struct icp_attach_args *icpa;
607 const char *str;
608
609 icpa = (struct icp_attach_args *)aux;
610
611 if (pnp != NULL) {
612 if (icpa->icpa_unit < ICPA_UNIT_SCSI)
613 str = "block device";
614 else
615 str = "SCSI channel";
616 aprint_normal("%s at %s", str, pnp);
617 }
618 aprint_normal(" unit %d", icpa->icpa_unit);
619
620 return (UNCONF);
621 }
622
623 int
624 icp_async_event(struct icp_softc *icp, int service)
625 {
626
627 if (service == ICP_SCREENSERVICE) {
628 if (icp->icp_status == ICP_S_MSG_REQUEST) {
629 /* XXX */
630 }
631 } else {
632 if ((icp->icp_fw_vers & 0xff) >= 0x1a) {
633 icp->icp_evt.size = 0;
634 icp->icp_evt.eu.async.ionode =
635 device_unit(icp->icp_dv);
636 icp->icp_evt.eu.async.status = icp->icp_status;
637 /*
638 * Severity and event string are filled in by the
639 * hardware interface interrupt handler.
640 */
641 printf("%s: %s\n", device_xname(icp->icp_dv),
642 icp->icp_evt.event_string);
643 } else {
644 icp->icp_evt.size = sizeof(icp->icp_evt.eu.async);
645 icp->icp_evt.eu.async.ionode =
646 device_unit(icp->icp_dv);
647 icp->icp_evt.eu.async.service = service;
648 icp->icp_evt.eu.async.status = icp->icp_status;
649 icp->icp_evt.eu.async.info = icp->icp_info;
650 /* XXXJRT FIX THIS */
651 *(u_int32_t *) icp->icp_evt.eu.async.scsi_coord =
652 icp->icp_info2;
653 }
654 icp_store_event(icp, GDT_ES_ASYNC, service, &icp->icp_evt);
655 }
656
657 return (0);
658 }
659
660 int
661 icp_intr(void *cookie)
662 {
663 struct icp_softc *icp;
664 struct icp_intr_ctx ctx;
665 struct icp_ccb *ic;
666
667 icp = cookie;
668
669 ctx.istatus = (*icp->icp_get_status)(icp);
670 if (!ctx.istatus) {
671 icp->icp_status = ICP_S_NO_STATUS;
672 return (0);
673 }
674
675 (*icp->icp_intr)(icp, &ctx);
676
677 icp->icp_status = ctx.cmd_status;
678 icp->icp_service = ctx.service;
679 icp->icp_info = ctx.info;
680 icp->icp_info2 = ctx.info2;
681
682 switch (ctx.istatus) {
683 case ICP_ASYNCINDEX:
684 icp_async_event(icp, ctx.service);
685 return (1);
686
687 case ICP_SPEZINDEX:
688 aprint_error_dev(icp->icp_dv, "uninitialized or unknown service (%d/%d)\n",
689 ctx.info, ctx.info2);
690 icp->icp_evt.size = sizeof(icp->icp_evt.eu.driver);
691 icp->icp_evt.eu.driver.ionode = device_unit(icp->icp_dv);
692 icp_store_event(icp, GDT_ES_DRIVER, 4, &icp->icp_evt);
693 return (1);
694 }
695
696 if ((ctx.istatus - 2) > icp->icp_nccbs)
697 panic("icp_intr: bad command index returned");
698
699 ic = &icp->icp_ccbs[ctx.istatus - 2];
700 ic->ic_status = icp->icp_status;
701
702 if ((ic->ic_flags & IC_ALLOCED) == 0) {
703 /* XXX ICP's "iir" driver just sends an event here. */
704 panic("icp_intr: inactive CCB identified");
705 }
706
707 /*
708 * Try to protect ourselves from the running command count already
709 * being 0 (e.g. if a polled command times out).
710 */
711 KDASSERT(icp->icp_running != 0);
712 if (--icp->icp_running == 0 &&
713 (icp->icp_flags & ICP_F_WAIT_FREEZE) != 0) {
714 icp->icp_flags &= ~ICP_F_WAIT_FREEZE;
715 wakeup(&icp->icp_qfreeze);
716 }
717
718 switch (icp->icp_status) {
719 case ICP_S_BSY:
720 #ifdef ICP_DEBUG
721 printf("%s: ICP_S_BSY received\n", device_xname(icp->icp_dv));
722 #endif
723 if (__predict_false((ic->ic_flags & IC_UCMD) != 0))
724 SIMPLEQ_INSERT_HEAD(&icp->icp_ucmd_queue, ic, ic_chain);
725 else
726 SIMPLEQ_INSERT_HEAD(&icp->icp_ccb_queue, ic, ic_chain);
727 break;
728
729 default:
730 ic->ic_flags |= IC_COMPLETE;
731
732 if ((ic->ic_flags & IC_WAITING) != 0)
733 wakeup(ic);
734 else if (ic->ic_intr != NULL)
735 (*ic->ic_intr)(ic);
736
737 if (ICP_HAS_WORK(icp))
738 icp_ccb_enqueue(icp, NULL);
739
740 break;
741 }
742
743 return (1);
744 }
745
746 struct icp_ucmd_ctx {
747 gdt_ucmd_t *iu_ucmd;
748 u_int32_t iu_cnt;
749 };
750
751 void
752 icp_ucmd_intr(struct icp_ccb *ic)
753 {
754 struct icp_softc *icp = device_private(ic->ic_dv);
755 struct icp_ucmd_ctx *iu = ic->ic_context;
756 gdt_ucmd_t *ucmd = iu->iu_ucmd;
757
758 ucmd->status = icp->icp_status;
759 ucmd->info = icp->icp_info;
760
761 if (iu->iu_cnt != 0) {
762 bus_dmamap_sync(icp->icp_dmat,
763 icp->icp_scr_dmamap,
764 ICP_SCRATCH_UCMD, iu->iu_cnt,
765 BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE);
766 memcpy(ucmd->data,
767 (char *)icp->icp_scr + ICP_SCRATCH_UCMD, iu->iu_cnt);
768 }
769
770 icp->icp_ucmd_ccb = NULL;
771
772 ic->ic_flags |= IC_COMPLETE;
773 wakeup(ic);
774 }
775
776 /*
777 * NOTE: We assume that it is safe to sleep here!
778 */
779 int
780 icp_cmd(struct icp_softc *icp, u_int8_t service, u_int16_t opcode,
781 u_int32_t arg1, u_int32_t arg2, u_int32_t arg3)
782 {
783 struct icp_ioctlcmd *icmd;
784 struct icp_cachecmd *cc;
785 struct icp_rawcmd *rc;
786 int retries, rv;
787 struct icp_ccb *ic;
788
789 retries = ICP_RETRIES;
790
791 do {
792 ic = icp_ccb_alloc_wait(icp);
793 memset(&ic->ic_cmd, 0, sizeof(ic->ic_cmd));
794 ic->ic_cmd.cmd_opcode = htole16(opcode);
795
796 switch (service) {
797 case ICP_CACHESERVICE:
798 if (opcode == ICP_IOCTL) {
799 icmd = &ic->ic_cmd.cmd_packet.ic;
800 icmd->ic_subfunc = htole16(arg1);
801 icmd->ic_channel = htole32(arg2);
802 icmd->ic_bufsize = htole32(arg3);
803 icmd->ic_addr =
804 htole32(icp->icp_scr_seg[0].ds_addr);
805
806 bus_dmamap_sync(icp->icp_dmat,
807 icp->icp_scr_dmamap, 0, arg3,
808 BUS_DMASYNC_PREWRITE |
809 BUS_DMASYNC_PREREAD);
810 } else {
811 cc = &ic->ic_cmd.cmd_packet.cc;
812 cc->cc_deviceno = htole16(arg1);
813 cc->cc_blockno = htole32(arg2);
814 }
815 break;
816
817 case ICP_SCSIRAWSERVICE:
818 rc = &ic->ic_cmd.cmd_packet.rc;
819 rc->rc_direction = htole32(arg1);
820 rc->rc_bus = arg2;
821 rc->rc_target = arg3;
822 rc->rc_lun = arg3 >> 8;
823 break;
824 }
825
826 ic->ic_service = service;
827 ic->ic_cmdlen = sizeof(ic->ic_cmd);
828 rv = icp_ccb_poll(icp, ic, 10000);
829
830 switch (service) {
831 case ICP_CACHESERVICE:
832 if (opcode == ICP_IOCTL) {
833 bus_dmamap_sync(icp->icp_dmat,
834 icp->icp_scr_dmamap, 0, arg3,
835 BUS_DMASYNC_POSTWRITE |
836 BUS_DMASYNC_POSTREAD);
837 }
838 break;
839 }
840
841 icp_ccb_free(icp, ic);
842 } while (rv != 0 && --retries > 0);
843
844 return (icp->icp_status == ICP_S_OK);
845 }
846
847 int
848 icp_ucmd(struct icp_softc *icp, gdt_ucmd_t *ucmd)
849 {
850 struct icp_ccb *ic;
851 struct icp_ucmd_ctx iu;
852 u_int32_t cnt;
853 int error;
854
855 if (ucmd->service == ICP_CACHESERVICE) {
856 if (ucmd->command.cmd_opcode == ICP_IOCTL) {
857 cnt = ucmd->command.cmd_packet.ic.ic_bufsize;
858 if (cnt > GDT_SCRATCH_SZ) {
859 aprint_error_dev(icp->icp_dv, "scratch buffer too small (%d/%d)\n",
860 GDT_SCRATCH_SZ, cnt);
861 return (EINVAL);
862 }
863 } else {
864 cnt = ucmd->command.cmd_packet.cc.cc_blockcnt *
865 ICP_SECTOR_SIZE;
866 if (cnt > GDT_SCRATCH_SZ) {
867 aprint_error_dev(icp->icp_dv, "scratch buffer too small (%d/%d)\n",
868 GDT_SCRATCH_SZ, cnt);
869 return (EINVAL);
870 }
871 }
872 } else {
873 cnt = ucmd->command.cmd_packet.rc.rc_sdlen +
874 ucmd->command.cmd_packet.rc.rc_sense_len;
875 if (cnt > GDT_SCRATCH_SZ) {
876 aprint_error_dev(icp->icp_dv, "scratch buffer too small (%d/%d)\n",
877 GDT_SCRATCH_SZ, cnt);
878 return (EINVAL);
879 }
880 }
881
882 iu.iu_ucmd = ucmd;
883 iu.iu_cnt = cnt;
884
885 ic = icp_ccb_alloc_wait(icp);
886 memset(&ic->ic_cmd, 0, sizeof(ic->ic_cmd));
887 ic->ic_cmd.cmd_opcode = htole16(ucmd->command.cmd_opcode);
888
889 if (ucmd->service == ICP_CACHESERVICE) {
890 if (ucmd->command.cmd_opcode == ICP_IOCTL) {
891 struct icp_ioctlcmd *icmd, *uicmd;
892
893 icmd = &ic->ic_cmd.cmd_packet.ic;
894 uicmd = &ucmd->command.cmd_packet.ic;
895
896 icmd->ic_subfunc = htole16(uicmd->ic_subfunc);
897 icmd->ic_channel = htole32(uicmd->ic_channel);
898 icmd->ic_bufsize = htole32(uicmd->ic_bufsize);
899 icmd->ic_addr =
900 htole32(icp->icp_scr_seg[0].ds_addr +
901 ICP_SCRATCH_UCMD);
902 } else {
903 struct icp_cachecmd *cc, *ucc;
904
905 cc = &ic->ic_cmd.cmd_packet.cc;
906 ucc = &ucmd->command.cmd_packet.cc;
907
908 cc->cc_deviceno = htole16(ucc->cc_deviceno);
909 cc->cc_blockno = htole32(ucc->cc_blockno);
910 cc->cc_blockcnt = htole32(ucc->cc_blockcnt);
911 cc->cc_addr = htole32(0xffffffffU);
912 cc->cc_nsgent = htole32(1);
913 cc->cc_sg[0].sg_addr =
914 htole32(icp->icp_scr_seg[0].ds_addr +
915 ICP_SCRATCH_UCMD);
916 cc->cc_sg[0].sg_len = htole32(cnt);
917 }
918 } else {
919 struct icp_rawcmd *rc, *urc;
920
921 rc = &ic->ic_cmd.cmd_packet.rc;
922 urc = &ucmd->command.cmd_packet.rc;
923
924 rc->rc_direction = htole32(urc->rc_direction);
925 rc->rc_sdata = htole32(0xffffffffU);
926 rc->rc_sdlen = htole32(urc->rc_sdlen);
927 rc->rc_clen = htole32(urc->rc_clen);
928 memcpy(rc->rc_cdb, urc->rc_cdb, sizeof(rc->rc_cdb));
929 rc->rc_target = urc->rc_target;
930 rc->rc_lun = urc->rc_lun;
931 rc->rc_bus = urc->rc_bus;
932 rc->rc_sense_len = htole32(urc->rc_sense_len);
933 rc->rc_sense_addr =
934 htole32(icp->icp_scr_seg[0].ds_addr +
935 ICP_SCRATCH_UCMD + urc->rc_sdlen);
936 rc->rc_nsgent = htole32(1);
937 rc->rc_sg[0].sg_addr =
938 htole32(icp->icp_scr_seg[0].ds_addr + ICP_SCRATCH_UCMD);
939 rc->rc_sg[0].sg_len = htole32(cnt - urc->rc_sense_len);
940 }
941
942 ic->ic_service = ucmd->service;
943 ic->ic_cmdlen = sizeof(ic->ic_cmd);
944 ic->ic_context = &iu;
945
946 /*
947 * XXX What units are ucmd->timeout in? Until we know, we
948 * XXX just pull a number out of thin air.
949 */
950 if (__predict_false((error = icp_ccb_wait_user(icp, ic, 30000)) != 0))
951 aprint_error_dev(icp->icp_dv, "error %d waiting for ucmd to complete\n",
952 error);
953
954 /* icp_ucmd_intr() has updated ucmd. */
955 icp_ccb_free(icp, ic);
956
957 return (error);
958 }
959
960 struct icp_ccb *
961 icp_ccb_alloc(struct icp_softc *icp)
962 {
963 struct icp_ccb *ic;
964 int s;
965
966 s = splbio();
967 if (__predict_false((ic =
968 SIMPLEQ_FIRST(&icp->icp_ccb_freelist)) == NULL)) {
969 splx(s);
970 return (NULL);
971 }
972 SIMPLEQ_REMOVE_HEAD(&icp->icp_ccb_freelist, ic_chain);
973 splx(s);
974
975 ic->ic_flags = IC_ALLOCED;
976 return (ic);
977 }
978
979 struct icp_ccb *
980 icp_ccb_alloc_wait(struct icp_softc *icp)
981 {
982 struct icp_ccb *ic;
983 int s;
984
985 s = splbio();
986 while ((ic = SIMPLEQ_FIRST(&icp->icp_ccb_freelist)) == NULL) {
987 icp->icp_flags |= ICP_F_WAIT_CCB;
988 (void) tsleep(&icp->icp_ccb_freelist, PRIBIO, "icpccb", 0);
989 }
990 SIMPLEQ_REMOVE_HEAD(&icp->icp_ccb_freelist, ic_chain);
991 splx(s);
992
993 ic->ic_flags = IC_ALLOCED;
994 return (ic);
995 }
996
997 void
998 icp_ccb_free(struct icp_softc *icp, struct icp_ccb *ic)
999 {
1000 int s;
1001
1002 s = splbio();
1003 ic->ic_flags = 0;
1004 ic->ic_intr = NULL;
1005 SIMPLEQ_INSERT_HEAD(&icp->icp_ccb_freelist, ic, ic_chain);
1006 if (__predict_false((icp->icp_flags & ICP_F_WAIT_CCB) != 0)) {
1007 icp->icp_flags &= ~ICP_F_WAIT_CCB;
1008 wakeup(&icp->icp_ccb_freelist);
1009 }
1010 splx(s);
1011 }
1012
1013 void
1014 icp_ccb_enqueue(struct icp_softc *icp, struct icp_ccb *ic)
1015 {
1016 int s;
1017
1018 s = splbio();
1019
1020 if (ic != NULL) {
1021 if (__predict_false((ic->ic_flags & IC_UCMD) != 0))
1022 SIMPLEQ_INSERT_TAIL(&icp->icp_ucmd_queue, ic, ic_chain);
1023 else
1024 SIMPLEQ_INSERT_TAIL(&icp->icp_ccb_queue, ic, ic_chain);
1025 }
1026
1027 for (; icp->icp_qfreeze == 0;) {
1028 if (__predict_false((ic =
1029 SIMPLEQ_FIRST(&icp->icp_ucmd_queue)) != NULL)) {
1030 struct icp_ucmd_ctx *iu = ic->ic_context;
1031 gdt_ucmd_t *ucmd = iu->iu_ucmd;
1032
1033 /*
1034 * All user-generated commands share the same
1035 * scratch space, so if one is already running,
1036 * we have to stall the command queue.
1037 */
1038 if (icp->icp_ucmd_ccb != NULL)
1039 break;
1040 if ((*icp->icp_test_busy)(icp))
1041 break;
1042 icp->icp_ucmd_ccb = ic;
1043
1044 if (iu->iu_cnt != 0) {
1045 memcpy((char *)icp->icp_scr + ICP_SCRATCH_UCMD,
1046 ucmd->data, iu->iu_cnt);
1047 bus_dmamap_sync(icp->icp_dmat,
1048 icp->icp_scr_dmamap,
1049 ICP_SCRATCH_UCMD, iu->iu_cnt,
1050 BUS_DMASYNC_PREREAD |
1051 BUS_DMASYNC_PREWRITE);
1052 }
1053 } else if (__predict_true((ic =
1054 SIMPLEQ_FIRST(&icp->icp_ccb_queue)) != NULL)) {
1055 if ((*icp->icp_test_busy)(icp))
1056 break;
1057 } else {
1058 /* no command found */
1059 break;
1060 }
1061 icp_ccb_submit(icp, ic);
1062 if (__predict_false((ic->ic_flags & IC_UCMD) != 0))
1063 SIMPLEQ_REMOVE_HEAD(&icp->icp_ucmd_queue, ic_chain);
1064 else
1065 SIMPLEQ_REMOVE_HEAD(&icp->icp_ccb_queue, ic_chain);
1066 }
1067
1068 splx(s);
1069 }
1070
1071 int
1072 icp_ccb_map(struct icp_softc *icp, struct icp_ccb *ic, void *data, int size,
1073 int dir)
1074 {
1075 struct icp_sg *sg;
1076 int nsegs, i, rv;
1077 bus_dmamap_t xfer;
1078
1079 xfer = ic->ic_xfer_map;
1080
1081 rv = bus_dmamap_load(icp->icp_dmat, xfer, data, size, NULL,
1082 BUS_DMA_NOWAIT | BUS_DMA_STREAMING |
1083 ((dir & IC_XFER_IN) ? BUS_DMA_READ : BUS_DMA_WRITE));
1084 if (rv != 0)
1085 return (rv);
1086
1087 nsegs = xfer->dm_nsegs;
1088 ic->ic_xfer_size = size;
1089 ic->ic_nsgent = nsegs;
1090 ic->ic_flags |= dir;
1091 sg = ic->ic_sg;
1092
1093 if (sg != NULL) {
1094 for (i = 0; i < nsegs; i++, sg++) {
1095 sg->sg_addr = htole32(xfer->dm_segs[i].ds_addr);
1096 sg->sg_len = htole32(xfer->dm_segs[i].ds_len);
1097 }
1098 } else if (nsegs > 1)
1099 panic("icp_ccb_map: no SG list specified, but nsegs > 1");
1100
1101 if ((dir & IC_XFER_OUT) != 0)
1102 i = BUS_DMASYNC_PREWRITE;
1103 else /* if ((dir & IC_XFER_IN) != 0) */
1104 i = BUS_DMASYNC_PREREAD;
1105
1106 bus_dmamap_sync(icp->icp_dmat, xfer, 0, ic->ic_xfer_size, i);
1107 return (0);
1108 }
1109
1110 void
1111 icp_ccb_unmap(struct icp_softc *icp, struct icp_ccb *ic)
1112 {
1113 int i;
1114
1115 if ((ic->ic_flags & IC_XFER_OUT) != 0)
1116 i = BUS_DMASYNC_POSTWRITE;
1117 else /* if ((ic->ic_flags & IC_XFER_IN) != 0) */
1118 i = BUS_DMASYNC_POSTREAD;
1119
1120 bus_dmamap_sync(icp->icp_dmat, ic->ic_xfer_map, 0, ic->ic_xfer_size, i);
1121 bus_dmamap_unload(icp->icp_dmat, ic->ic_xfer_map);
1122 }
1123
1124 int
1125 icp_ccb_poll(struct icp_softc *icp, struct icp_ccb *ic, int timo)
1126 {
1127 int s, rv;
1128
1129 s = splbio();
1130
1131 for (timo = ICP_BUSY_WAIT_MS * 100; timo != 0; timo--) {
1132 if (!(*icp->icp_test_busy)(icp))
1133 break;
1134 DELAY(10);
1135 }
1136 if (timo == 0) {
1137 printf("%s: submit: busy\n", device_xname(icp->icp_dv));
1138 return (EAGAIN);
1139 }
1140
1141 icp_ccb_submit(icp, ic);
1142
1143 if (cold) {
1144 for (timo *= 10; timo != 0; timo--) {
1145 DELAY(100);
1146 icp_intr(icp);
1147 if ((ic->ic_flags & IC_COMPLETE) != 0)
1148 break;
1149 }
1150 } else {
1151 ic->ic_flags |= IC_WAITING;
1152 while ((ic->ic_flags & IC_COMPLETE) == 0) {
1153 if ((rv = tsleep(ic, PRIBIO, "icpwccb",
1154 mstohz(timo))) != 0) {
1155 timo = 0;
1156 break;
1157 }
1158 }
1159 }
1160
1161 if (timo != 0) {
1162 if (ic->ic_status != ICP_S_OK) {
1163 #ifdef ICP_DEBUG
1164 printf("%s: request failed; status=0x%04x\n",
1165 device_xname(icp->icp_dv), ic->ic_status);
1166 #endif
1167 rv = EIO;
1168 } else
1169 rv = 0;
1170 } else {
1171 aprint_error_dev(icp->icp_dv, "command timed out\n");
1172 rv = EIO;
1173 }
1174
1175 while ((*icp->icp_test_busy)(icp) != 0)
1176 DELAY(10);
1177
1178 splx(s);
1179
1180 return (rv);
1181 }
1182
1183 int
1184 icp_ccb_wait(struct icp_softc *icp, struct icp_ccb *ic, int timo)
1185 {
1186 int s, rv;
1187
1188 ic->ic_flags |= IC_WAITING;
1189
1190 s = splbio();
1191 icp_ccb_enqueue(icp, ic);
1192 while ((ic->ic_flags & IC_COMPLETE) == 0) {
1193 if ((rv = tsleep(ic, PRIBIO, "icpwccb", mstohz(timo))) != 0) {
1194 splx(s);
1195 return (rv);
1196 }
1197 }
1198 splx(s);
1199
1200 if (ic->ic_status != ICP_S_OK) {
1201 aprint_error_dev(icp->icp_dv, "command failed; status=%x\n",
1202 ic->ic_status);
1203 return (EIO);
1204 }
1205
1206 return (0);
1207 }
1208
1209 int
1210 icp_ccb_wait_user(struct icp_softc *icp, struct icp_ccb *ic, int timo)
1211 {
1212 int s, rv;
1213
1214 ic->ic_dv = icp->icp_dv;
1215 ic->ic_intr = icp_ucmd_intr;
1216 ic->ic_flags |= IC_UCMD;
1217
1218 s = splbio();
1219 icp_ccb_enqueue(icp, ic);
1220 while ((ic->ic_flags & IC_COMPLETE) == 0) {
1221 if ((rv = tsleep(ic, PRIBIO, "icpwuccb", mstohz(timo))) != 0) {
1222 splx(s);
1223 return (rv);
1224 }
1225 }
1226 splx(s);
1227
1228 return (0);
1229 }
1230
1231 void
1232 icp_ccb_submit(struct icp_softc *icp, struct icp_ccb *ic)
1233 {
1234
1235 ic->ic_cmdlen = (ic->ic_cmdlen + 3) & ~3;
1236
1237 (*icp->icp_set_sema0)(icp);
1238 DELAY(10);
1239
1240 ic->ic_cmd.cmd_boardnode = htole32(ICP_LOCALBOARD);
1241 ic->ic_cmd.cmd_cmdindex = htole32(ic->ic_ident);
1242
1243 icp->icp_running++;
1244
1245 (*icp->icp_copy_cmd)(icp, ic);
1246 (*icp->icp_release_event)(icp, ic);
1247 }
1248
1249 int
1250 icp_freeze(struct icp_softc *icp)
1251 {
1252 int s, error = 0;
1253
1254 s = splbio();
1255 if (icp->icp_qfreeze++ == 0) {
1256 while (icp->icp_running != 0) {
1257 icp->icp_flags |= ICP_F_WAIT_FREEZE;
1258 error = tsleep(&icp->icp_qfreeze, PRIBIO|PCATCH,
1259 "icpqfrz", 0);
1260 if (error != 0 && --icp->icp_qfreeze == 0 &&
1261 ICP_HAS_WORK(icp)) {
1262 icp_ccb_enqueue(icp, NULL);
1263 break;
1264 }
1265 }
1266 }
1267 splx(s);
1268
1269 return (error);
1270 }
1271
1272 void
1273 icp_unfreeze(struct icp_softc *icp)
1274 {
1275 int s;
1276
1277 s = splbio();
1278 KDASSERT(icp->icp_qfreeze != 0);
1279 if (--icp->icp_qfreeze == 0 && ICP_HAS_WORK(icp))
1280 icp_ccb_enqueue(icp, NULL);
1281 splx(s);
1282 }
1283
1284 /* XXX Global - should be per-controller? XXX */
1285 static gdt_evt_str icp_event_buffer[ICP_MAX_EVENTS];
1286 static int icp_event_oldidx;
1287 static int icp_event_lastidx;
1288
1289 gdt_evt_str *
1290 icp_store_event(struct icp_softc *icp, u_int16_t source, u_int16_t idx,
1291 gdt_evt_data *evt)
1292 {
1293 gdt_evt_str *e;
1294
1295 /* no source == no event */
1296 if (source == 0)
1297 return (NULL);
1298
1299 e = &icp_event_buffer[icp_event_lastidx];
1300 if (e->event_source == source && e->event_idx == idx &&
1301 ((evt->size != 0 && e->event_data.size != 0 &&
1302 memcmp(&e->event_data.eu, &evt->eu, evt->size) == 0) ||
1303 (evt->size == 0 && e->event_data.size == 0 &&
1304 strcmp((char *) e->event_data.event_string,
1305 (char *) evt->event_string) == 0))) {
1306 e->last_stamp = time_second;
1307 e->same_count++;
1308 } else {
1309 if (icp_event_buffer[icp_event_lastidx].event_source != 0) {
1310 icp_event_lastidx++;
1311 if (icp_event_lastidx == ICP_MAX_EVENTS)
1312 icp_event_lastidx = 0;
1313 if (icp_event_lastidx == icp_event_oldidx) {
1314 icp_event_oldidx++;
1315 if (icp_event_oldidx == ICP_MAX_EVENTS)
1316 icp_event_oldidx = 0;
1317 }
1318 }
1319 e = &icp_event_buffer[icp_event_lastidx];
1320 e->event_source = source;
1321 e->event_idx = idx;
1322 e->first_stamp = e->last_stamp = time_second;
1323 e->same_count = 1;
1324 e->event_data = *evt;
1325 e->application = 0;
1326 }
1327 return (e);
1328 }
1329
1330 int
1331 icp_read_event(struct icp_softc *icp, int handle, gdt_evt_str *estr)
1332 {
1333 gdt_evt_str *e;
1334 int eindex, s;
1335
1336 s = splbio();
1337
1338 if (handle == -1)
1339 eindex = icp_event_oldidx;
1340 else
1341 eindex = handle;
1342
1343 estr->event_source = 0;
1344
1345 if (eindex < 0 || eindex >= ICP_MAX_EVENTS) {
1346 splx(s);
1347 return (eindex);
1348 }
1349
1350 e = &icp_event_buffer[eindex];
1351 if (e->event_source != 0) {
1352 if (eindex != icp_event_lastidx) {
1353 eindex++;
1354 if (eindex == ICP_MAX_EVENTS)
1355 eindex = 0;
1356 } else
1357 eindex = -1;
1358 memcpy(estr, e, sizeof(gdt_evt_str));
1359 }
1360
1361 splx(s);
1362
1363 return (eindex);
1364 }
1365
1366 void
1367 icp_readapp_event(struct icp_softc *icp, u_int8_t application,
1368 gdt_evt_str *estr)
1369 {
1370 gdt_evt_str *e;
1371 int found = 0, eindex, s;
1372
1373 s = splbio();
1374
1375 eindex = icp_event_oldidx;
1376 for (;;) {
1377 e = &icp_event_buffer[eindex];
1378 if (e->event_source == 0)
1379 break;
1380 if ((e->application & application) == 0) {
1381 e->application |= application;
1382 found = 1;
1383 break;
1384 }
1385 if (eindex == icp_event_lastidx)
1386 break;
1387 eindex++;
1388 if (eindex == ICP_MAX_EVENTS)
1389 eindex = 0;
1390 }
1391 if (found)
1392 memcpy(estr, e, sizeof(gdt_evt_str));
1393 else
1394 estr->event_source = 0;
1395
1396 splx(s);
1397 }
1398
1399 void
1400 icp_clear_events(struct icp_softc *icp)
1401 {
1402 int s;
1403
1404 s = splbio();
1405 icp_event_oldidx = icp_event_lastidx = 0;
1406 memset(icp_event_buffer, 0, sizeof(icp_event_buffer));
1407 splx(s);
1408 }
1409