ata.c revision 1.146 1 /* $NetBSD: ata.c,v 1.146 2018/11/12 18:51:01 jdolecek Exp $ */
2
3 /*
4 * Copyright (c) 1998, 2001 Manuel Bouyer. All rights reserved.
5 *
6 * Redistribution and use in source and binary forms, with or without
7 * modification, are permitted provided that the following conditions
8 * are met:
9 * 1. Redistributions of source code must retain the above copyright
10 * notice, this list of conditions and the following disclaimer.
11 * 2. Redistributions in binary form must reproduce the above copyright
12 * notice, this list of conditions and the following disclaimer in the
13 * documentation and/or other materials provided with the distribution.
14 *
15 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
16 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
17 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
18 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
19 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
20 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
21 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
22 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
23 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
24 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
25 */
26
27 #include <sys/cdefs.h>
28 __KERNEL_RCSID(0, "$NetBSD: ata.c,v 1.146 2018/11/12 18:51:01 jdolecek Exp $");
29
30 #include "opt_ata.h"
31
32 #include <sys/param.h>
33 #include <sys/systm.h>
34 #include <sys/kernel.h>
35 #include <sys/device.h>
36 #include <sys/conf.h>
37 #include <sys/fcntl.h>
38 #include <sys/proc.h>
39 #include <sys/kthread.h>
40 #include <sys/errno.h>
41 #include <sys/ataio.h>
42 #include <sys/kmem.h>
43 #include <sys/intr.h>
44 #include <sys/bus.h>
45 #include <sys/once.h>
46 #include <sys/bitops.h>
47
48 #define ATABUS_PRIVATE
49
50 #include <dev/ata/ataconf.h>
51 #include <dev/ata/atareg.h>
52 #include <dev/ata/atavar.h>
53 #include <dev/ic/wdcvar.h> /* for PIOBM */
54
55 #include "ioconf.h"
56 #include "locators.h"
57
58 #include "atapibus.h"
59 #include "ataraid.h"
60 #include "sata_pmp.h"
61
62 #if NATARAID > 0
63 #include <dev/ata/ata_raidvar.h>
64 #endif
65 #if NSATA_PMP > 0
66 #include <dev/ata/satapmpvar.h>
67 #endif
68 #include <dev/ata/satapmpreg.h>
69
70 #define DEBUG_FUNCS 0x08
71 #define DEBUG_PROBE 0x10
72 #define DEBUG_DETACH 0x20
73 #define DEBUG_XFERS 0x40
74 #ifdef ATADEBUG
75 #ifndef ATADEBUG_MASK
76 #define ATADEBUG_MASK 0
77 #endif
78 int atadebug_mask = ATADEBUG_MASK;
79 #define ATADEBUG_PRINT(args, level) \
80 if (atadebug_mask & (level)) \
81 printf args
82 #else
83 #define ATADEBUG_PRINT(args, level)
84 #endif
85
86 static ONCE_DECL(ata_init_ctrl);
87 static struct pool ata_xfer_pool;
88
89 /*
90 * A queue of atabus instances, used to ensure the same bus probe order
91 * for a given hardware configuration at each boot. Kthread probing
92 * devices on a atabus. Only one probing at once.
93 */
94 static TAILQ_HEAD(, atabus_initq) atabus_initq_head;
95 static kmutex_t atabus_qlock;
96 static kcondvar_t atabus_qcv;
97 static lwp_t * atabus_cfg_lwp;
98
99 /*****************************************************************************
100 * ATA bus layer.
101 *
102 * ATA controllers attach an atabus instance, which handles probing the bus
103 * for drives, etc.
104 *****************************************************************************/
105
106 dev_type_open(atabusopen);
107 dev_type_close(atabusclose);
108 dev_type_ioctl(atabusioctl);
109
110 const struct cdevsw atabus_cdevsw = {
111 .d_open = atabusopen,
112 .d_close = atabusclose,
113 .d_read = noread,
114 .d_write = nowrite,
115 .d_ioctl = atabusioctl,
116 .d_stop = nostop,
117 .d_tty = notty,
118 .d_poll = nopoll,
119 .d_mmap = nommap,
120 .d_kqfilter = nokqfilter,
121 .d_discard = nodiscard,
122 .d_flag = D_OTHER
123 };
124
125 static void atabus_childdetached(device_t, device_t);
126 static int atabus_rescan(device_t, const char *, const int *);
127 static bool atabus_resume(device_t, const pmf_qual_t *);
128 static bool atabus_suspend(device_t, const pmf_qual_t *);
129 static void atabusconfig_thread(void *);
130
131 static void ata_channel_idle(struct ata_channel *);
132 static void ata_activate_xfer_locked(struct ata_channel *, struct ata_xfer *);
133 static void ata_channel_freeze_locked(struct ata_channel *);
134 static void ata_thread_wake_locked(struct ata_channel *);
135
136 /*
137 * atabus_init:
138 *
139 * Initialize ATA subsystem structures.
140 */
141 static int
142 atabus_init(void)
143 {
144
145 pool_init(&ata_xfer_pool, sizeof(struct ata_xfer), 0, 0, 0,
146 "ataspl", NULL, IPL_BIO);
147 TAILQ_INIT(&atabus_initq_head);
148 mutex_init(&atabus_qlock, MUTEX_DEFAULT, IPL_NONE);
149 cv_init(&atabus_qcv, "atainitq");
150 return 0;
151 }
152
153 /*
154 * atabusprint:
155 *
156 * Autoconfiguration print routine used by ATA controllers when
157 * attaching an atabus instance.
158 */
159 int
160 atabusprint(void *aux, const char *pnp)
161 {
162 struct ata_channel *chan = aux;
163
164 if (pnp)
165 aprint_normal("atabus at %s", pnp);
166 aprint_normal(" channel %d", chan->ch_channel);
167
168 return (UNCONF);
169 }
170
171 /*
172 * ataprint:
173 *
174 * Autoconfiguration print routine.
175 */
176 int
177 ataprint(void *aux, const char *pnp)
178 {
179 struct ata_device *adev = aux;
180
181 if (pnp)
182 aprint_normal("wd at %s", pnp);
183 aprint_normal(" drive %d", adev->adev_drv_data->drive);
184
185 return (UNCONF);
186 }
187
188 /*
189 * ata_channel_attach:
190 *
191 * Common parts of attaching an atabus to an ATA controller channel.
192 */
193 void
194 ata_channel_attach(struct ata_channel *chp)
195 {
196 if (chp->ch_flags & ATACH_DISABLED)
197 return;
198
199 ata_channel_init(chp);
200
201 KASSERT(chp->ch_queue != NULL);
202
203 chp->atabus = config_found_ia(chp->ch_atac->atac_dev, "ata", chp,
204 atabusprint);
205 }
206
207 /*
208 * ata_channel_detach:
209 *
210 * Common parts of detaching an atabus to an ATA controller channel.
211 */
212 void
213 ata_channel_detach(struct ata_channel *chp)
214 {
215 if (chp->ch_flags & ATACH_DISABLED)
216 return;
217
218 ata_channel_destroy(chp);
219
220 chp->ch_flags |= ATACH_DETACHED;
221 }
222
223 static void
224 atabusconfig(struct atabus_softc *atabus_sc)
225 {
226 struct ata_channel *chp = atabus_sc->sc_chan;
227 struct atac_softc *atac = chp->ch_atac;
228 struct atabus_initq *atabus_initq = NULL;
229 int i, error;
230
231 /* we are in the atabus's thread context */
232 ata_channel_lock(chp);
233 chp->ch_flags |= ATACH_TH_RUN;
234 ata_channel_unlock(chp);
235
236 /*
237 * Probe for the drives attached to controller, unless a PMP
238 * is already known
239 */
240 /* XXX for SATA devices we will power up all drives at once */
241 if (chp->ch_satapmp_nports == 0)
242 (*atac->atac_probe)(chp);
243
244 if (chp->ch_ndrives >= 2) {
245 ATADEBUG_PRINT(("atabusattach: ch_drive_type 0x%x 0x%x\n",
246 chp->ch_drive[0].drive_type, chp->ch_drive[1].drive_type),
247 DEBUG_PROBE);
248 }
249
250 /* next operations will occurs in a separate thread */
251 ata_channel_lock(chp);
252 chp->ch_flags &= ~ATACH_TH_RUN;
253 ata_channel_unlock(chp);
254
255 /* Make sure the devices probe in atabus order to avoid jitter. */
256 mutex_enter(&atabus_qlock);
257 for (;;) {
258 atabus_initq = TAILQ_FIRST(&atabus_initq_head);
259 if (atabus_initq->atabus_sc == atabus_sc)
260 break;
261 cv_wait(&atabus_qcv, &atabus_qlock);
262 }
263 mutex_exit(&atabus_qlock);
264
265 ata_channel_lock(chp);
266
267 /* If no drives, abort here */
268 if (chp->ch_drive == NULL)
269 goto out;
270 KASSERT(chp->ch_ndrives == 0 || chp->ch_drive != NULL);
271 for (i = 0; i < chp->ch_ndrives; i++)
272 if (chp->ch_drive[i].drive_type != ATA_DRIVET_NONE)
273 break;
274 if (i == chp->ch_ndrives)
275 goto out;
276
277 /* Shortcut in case we've been shutdown */
278 if (chp->ch_flags & ATACH_SHUTDOWN)
279 goto out;
280
281 ata_channel_unlock(chp);
282
283 if ((error = kthread_create(PRI_NONE, 0, NULL, atabusconfig_thread,
284 atabus_sc, &atabus_cfg_lwp,
285 "%scnf", device_xname(atac->atac_dev))) != 0)
286 aprint_error_dev(atac->atac_dev,
287 "unable to create config thread: error %d\n", error);
288 return;
289
290 out:
291 ata_channel_unlock(chp);
292
293 mutex_enter(&atabus_qlock);
294 TAILQ_REMOVE(&atabus_initq_head, atabus_initq, atabus_initq);
295 cv_broadcast(&atabus_qcv);
296 mutex_exit(&atabus_qlock);
297
298 kmem_free(atabus_initq, sizeof(*atabus_initq));
299
300 ata_delref(chp);
301
302 config_pending_decr(atac->atac_dev);
303 }
304
305 /*
306 * atabus_configthread: finish attach of atabus's childrens, in a separate
307 * kernel thread.
308 */
309 static void
310 atabusconfig_thread(void *arg)
311 {
312 struct atabus_softc *atabus_sc = arg;
313 struct ata_channel *chp = atabus_sc->sc_chan;
314 struct atac_softc *atac = chp->ch_atac;
315 struct atabus_initq *atabus_initq = NULL;
316 int i, s;
317
318 /* XXX seems wrong */
319 mutex_enter(&atabus_qlock);
320 atabus_initq = TAILQ_FIRST(&atabus_initq_head);
321 KASSERT(atabus_initq->atabus_sc == atabus_sc);
322 mutex_exit(&atabus_qlock);
323
324 /*
325 * First look for a port multiplier
326 */
327 if (chp->ch_ndrives == PMP_MAX_DRIVES &&
328 chp->ch_drive[PMP_PORT_CTL].drive_type == ATA_DRIVET_PM) {
329 #if NSATA_PMP > 0
330 satapmp_attach(chp);
331 #else
332 aprint_error_dev(atabus_sc->sc_dev,
333 "SATA port multiplier not supported\n");
334 /* no problems going on, all drives are ATA_DRIVET_NONE */
335 #endif
336 }
337
338 /*
339 * Attach an ATAPI bus, if needed.
340 */
341 KASSERT(chp->ch_ndrives == 0 || chp->ch_drive != NULL);
342 for (i = 0; i < chp->ch_ndrives && chp->atapibus == NULL; i++) {
343 if (chp->ch_drive[i].drive_type == ATA_DRIVET_ATAPI) {
344 #if NATAPIBUS > 0
345 (*atac->atac_atapibus_attach)(atabus_sc);
346 #else
347 /*
348 * Fake the autoconfig "not configured" message
349 */
350 aprint_normal("atapibus at %s not configured\n",
351 device_xname(atac->atac_dev));
352 chp->atapibus = NULL;
353 s = splbio();
354 for (i = 0; i < chp->ch_ndrives; i++) {
355 if (chp->ch_drive[i].drive_type == ATA_DRIVET_ATAPI)
356 chp->ch_drive[i].drive_type = ATA_DRIVET_NONE;
357 }
358 splx(s);
359 #endif
360 break;
361 }
362 }
363
364 for (i = 0; i < chp->ch_ndrives; i++) {
365 struct ata_device adev;
366 if (chp->ch_drive[i].drive_type != ATA_DRIVET_ATA &&
367 chp->ch_drive[i].drive_type != ATA_DRIVET_OLD) {
368 continue;
369 }
370 if (chp->ch_drive[i].drv_softc != NULL)
371 continue;
372 memset(&adev, 0, sizeof(struct ata_device));
373 adev.adev_bustype = atac->atac_bustype_ata;
374 adev.adev_channel = chp->ch_channel;
375 adev.adev_drv_data = &chp->ch_drive[i];
376 chp->ch_drive[i].drv_softc = config_found_ia(atabus_sc->sc_dev,
377 "ata_hl", &adev, ataprint);
378 if (chp->ch_drive[i].drv_softc != NULL) {
379 ata_probe_caps(&chp->ch_drive[i]);
380 } else {
381 s = splbio();
382 chp->ch_drive[i].drive_type = ATA_DRIVET_NONE;
383 splx(s);
384 }
385 }
386
387 /* now that we know the drives, the controller can set its modes */
388 if (atac->atac_set_modes) {
389 (*atac->atac_set_modes)(chp);
390 ata_print_modes(chp);
391 }
392 #if NATARAID > 0
393 if (atac->atac_cap & ATAC_CAP_RAID) {
394 for (i = 0; i < chp->ch_ndrives; i++) {
395 if (chp->ch_drive[i].drive_type == ATA_DRIVET_ATA) {
396 ata_raid_check_component(
397 chp->ch_drive[i].drv_softc);
398 }
399 }
400 }
401 #endif /* NATARAID > 0 */
402
403 /*
404 * reset drive_flags for unattached devices, reset state for attached
405 * ones
406 */
407 s = splbio();
408 for (i = 0; i < chp->ch_ndrives; i++) {
409 if (chp->ch_drive[i].drive_type == ATA_DRIVET_PM)
410 continue;
411 if (chp->ch_drive[i].drv_softc == NULL) {
412 chp->ch_drive[i].drive_flags = 0;
413 chp->ch_drive[i].drive_type = ATA_DRIVET_NONE;
414 } else
415 chp->ch_drive[i].state = 0;
416 }
417 splx(s);
418
419 mutex_enter(&atabus_qlock);
420 TAILQ_REMOVE(&atabus_initq_head, atabus_initq, atabus_initq);
421 cv_broadcast(&atabus_qcv);
422 mutex_exit(&atabus_qlock);
423
424 kmem_free(atabus_initq, sizeof(*atabus_initq));
425
426 ata_delref(chp);
427
428 config_pending_decr(atac->atac_dev);
429 kthread_exit(0);
430 }
431
432 /*
433 * atabus_thread:
434 *
435 * Worker thread for the ATA bus.
436 */
437 static void
438 atabus_thread(void *arg)
439 {
440 struct atabus_softc *sc = arg;
441 struct ata_channel *chp = sc->sc_chan;
442 struct ata_queue *chq = chp->ch_queue;
443 struct ata_xfer *xfer;
444 int i, rv;
445
446 ata_channel_lock(chp);
447 chp->ch_flags |= ATACH_TH_RUN;
448
449 /*
450 * Probe the drives. Reset type to indicate to controllers
451 * that can re-probe that all drives must be probed..
452 *
453 * Note: ch_ndrives may be changed during the probe.
454 */
455 KASSERT(chp->ch_ndrives == 0 || chp->ch_drive != NULL);
456 for (i = 0; i < chp->ch_ndrives; i++) {
457 chp->ch_drive[i].drive_flags = 0;
458 chp->ch_drive[i].drive_type = ATA_DRIVET_NONE;
459 }
460 ata_channel_unlock(chp);
461
462 atabusconfig(sc);
463
464 ata_channel_lock(chp);
465 for (;;) {
466 if ((chp->ch_flags & (ATACH_TH_RESET | ATACH_TH_DRIVE_RESET
467 | ATACH_TH_RECOVERY | ATACH_SHUTDOWN)) == 0 &&
468 (chq->queue_active == 0 || chq->queue_freeze == 0)) {
469 chp->ch_flags &= ~ATACH_TH_RUN;
470 cv_wait(&chp->ch_thr_idle, &chp->ch_lock);
471 chp->ch_flags |= ATACH_TH_RUN;
472 }
473 if (chp->ch_flags & ATACH_SHUTDOWN) {
474 break;
475 }
476 if (chp->ch_flags & ATACH_TH_RESCAN) {
477 chp->ch_flags &= ~ATACH_TH_RESCAN;
478 ata_channel_unlock(chp);
479 atabusconfig(sc);
480 ata_channel_lock(chp);
481 }
482 if (chp->ch_flags & ATACH_TH_RESET) {
483 /* this will unfreeze the channel */
484 ata_thread_run(chp, AT_WAIT,
485 ATACH_TH_RESET, ATACH_NODRIVE);
486 } else if (chp->ch_flags & ATACH_TH_DRIVE_RESET) {
487 /* this will unfreeze the channel */
488 for (i = 0; i < chp->ch_ndrives; i++) {
489 struct ata_drive_datas *drvp;
490
491 drvp = &chp->ch_drive[i];
492
493 if (drvp->drive_flags & ATA_DRIVE_TH_RESET) {
494 ata_thread_run(chp,
495 AT_WAIT, ATACH_TH_DRIVE_RESET, i);
496 }
497 }
498 chp->ch_flags &= ~ATACH_TH_DRIVE_RESET;
499 } else if (chp->ch_flags & ATACH_TH_RECOVERY) {
500 /*
501 * This will unfreeze the channel; drops locks during
502 * run, so must wrap in splbio()/splx() to avoid
503 * spurious interrupts. XXX MPSAFE
504 */
505 int s = splbio();
506 ata_thread_run(chp, AT_WAIT, ATACH_TH_RECOVERY,
507 chp->recovery_tfd);
508 splx(s);
509 } else if (chq->queue_active > 0 && chq->queue_freeze == 1) {
510 /*
511 * Caller has bumped queue_freeze, decrease it. This
512 * flow shalt never be executed for NCQ commands.
513 */
514 KASSERT((chp->ch_flags & ATACH_NCQ) == 0);
515 KASSERT(chq->queue_active == 1);
516
517 ata_channel_thaw_locked(chp);
518 xfer = ata_queue_get_active_xfer_locked(chp);
519
520 KASSERT(xfer != NULL);
521 KASSERT((xfer->c_flags & C_POLL) == 0);
522
523 switch ((rv = ata_xfer_start(xfer))) {
524 case ATASTART_STARTED:
525 case ATASTART_POLL:
526 case ATASTART_ABORT:
527 break;
528 case ATASTART_TH:
529 default:
530 panic("%s: ata_xfer_start() unexpected rv %d",
531 __func__, rv);
532 /* NOTREACHED */
533 }
534 } else if (chq->queue_freeze > 1)
535 panic("%s: queue_freeze", __func__);
536
537 /* Try to run down the queue once channel is unfrozen */
538 if (chq->queue_freeze == 0) {
539 ata_channel_unlock(chp);
540 atastart(chp);
541 ata_channel_lock(chp);
542 }
543 }
544 chp->ch_thread = NULL;
545 cv_signal(&chp->ch_thr_idle);
546 ata_channel_unlock(chp);
547 kthread_exit(0);
548 }
549
550 static void
551 ata_thread_wake_locked(struct ata_channel *chp)
552 {
553 KASSERT(mutex_owned(&chp->ch_lock));
554 ata_channel_freeze_locked(chp);
555 cv_signal(&chp->ch_thr_idle);
556 }
557
558 /*
559 * atabus_match:
560 *
561 * Autoconfiguration match routine.
562 */
563 static int
564 atabus_match(device_t parent, cfdata_t cf, void *aux)
565 {
566 struct ata_channel *chp = aux;
567
568 if (chp == NULL)
569 return (0);
570
571 if (cf->cf_loc[ATACF_CHANNEL] != chp->ch_channel &&
572 cf->cf_loc[ATACF_CHANNEL] != ATACF_CHANNEL_DEFAULT)
573 return (0);
574
575 return (1);
576 }
577
578 /*
579 * atabus_attach:
580 *
581 * Autoconfiguration attach routine.
582 */
583 static void
584 atabus_attach(device_t parent, device_t self, void *aux)
585 {
586 struct atabus_softc *sc = device_private(self);
587 struct ata_channel *chp = aux;
588 struct atabus_initq *initq;
589 int error;
590
591 sc->sc_chan = chp;
592
593 aprint_normal("\n");
594 aprint_naive("\n");
595
596 sc->sc_dev = self;
597
598 if (ata_addref(chp))
599 return;
600
601 RUN_ONCE(&ata_init_ctrl, atabus_init);
602
603 initq = kmem_zalloc(sizeof(*initq), KM_SLEEP);
604 initq->atabus_sc = sc;
605 mutex_enter(&atabus_qlock);
606 TAILQ_INSERT_TAIL(&atabus_initq_head, initq, atabus_initq);
607 mutex_exit(&atabus_qlock);
608 config_pending_incr(sc->sc_dev);
609
610 /* XXX MPSAFE - no KTHREAD_MPSAFE, so protected by KERNEL_LOCK() */
611 if ((error = kthread_create(PRI_NONE, 0, NULL, atabus_thread, sc,
612 &chp->ch_thread, "%s", device_xname(self))) != 0)
613 aprint_error_dev(self,
614 "unable to create kernel thread: error %d\n", error);
615
616 if (!pmf_device_register(self, atabus_suspend, atabus_resume))
617 aprint_error_dev(self, "couldn't establish power handler\n");
618 }
619
620 /*
621 * atabus_detach:
622 *
623 * Autoconfiguration detach routine.
624 */
625 static int
626 atabus_detach(device_t self, int flags)
627 {
628 struct atabus_softc *sc = device_private(self);
629 struct ata_channel *chp = sc->sc_chan;
630 device_t dev = NULL;
631 int i, error = 0;
632
633 /* Shutdown the channel. */
634 ata_channel_lock(chp);
635 chp->ch_flags |= ATACH_SHUTDOWN;
636 while (chp->ch_thread != NULL) {
637 cv_signal(&chp->ch_thr_idle);
638 cv_wait(&chp->ch_thr_idle, &chp->ch_lock);
639 }
640 ata_channel_unlock(chp);
641
642 /*
643 * Detach atapibus and its children.
644 */
645 if ((dev = chp->atapibus) != NULL) {
646 ATADEBUG_PRINT(("atabus_detach: %s: detaching %s\n",
647 device_xname(self), device_xname(dev)), DEBUG_DETACH);
648
649 error = config_detach(dev, flags);
650 if (error)
651 goto out;
652 KASSERT(chp->atapibus == NULL);
653 }
654
655 KASSERT(chp->ch_ndrives == 0 || chp->ch_drive != NULL);
656
657 /*
658 * Detach our other children.
659 */
660 for (i = 0; i < chp->ch_ndrives; i++) {
661 if (chp->ch_drive[i].drive_type == ATA_DRIVET_ATAPI)
662 continue;
663 if (chp->ch_drive[i].drive_type == ATA_DRIVET_PM)
664 chp->ch_drive[i].drive_type = ATA_DRIVET_NONE;
665 if ((dev = chp->ch_drive[i].drv_softc) != NULL) {
666 ATADEBUG_PRINT(("%s.%d: %s: detaching %s\n", __func__,
667 __LINE__, device_xname(self), device_xname(dev)),
668 DEBUG_DETACH);
669 error = config_detach(dev, flags);
670 if (error)
671 goto out;
672 KASSERT(chp->ch_drive[i].drv_softc == NULL);
673 KASSERT(chp->ch_drive[i].drive_type == 0);
674 }
675 }
676 atabus_free_drives(chp);
677
678 out:
679 #ifdef ATADEBUG
680 if (dev != NULL && error != 0)
681 ATADEBUG_PRINT(("%s: %s: error %d detaching %s\n", __func__,
682 device_xname(self), error, device_xname(dev)),
683 DEBUG_DETACH);
684 #endif /* ATADEBUG */
685
686 return (error);
687 }
688
689 void
690 atabus_childdetached(device_t self, device_t child)
691 {
692 bool found = false;
693 struct atabus_softc *sc = device_private(self);
694 struct ata_channel *chp = sc->sc_chan;
695 int i;
696
697 KASSERT(chp->ch_ndrives == 0 || chp->ch_drive != NULL);
698 /*
699 * atapibus detached.
700 */
701 if (child == chp->atapibus) {
702 chp->atapibus = NULL;
703 found = true;
704 for (i = 0; i < chp->ch_ndrives; i++) {
705 if (chp->ch_drive[i].drive_type != ATA_DRIVET_ATAPI)
706 continue;
707 KASSERT(chp->ch_drive[i].drv_softc != NULL);
708 chp->ch_drive[i].drv_softc = NULL;
709 chp->ch_drive[i].drive_flags = 0;
710 chp->ch_drive[i].drive_type = ATA_DRIVET_NONE;
711 }
712 }
713
714 /*
715 * Detach our other children.
716 */
717 for (i = 0; i < chp->ch_ndrives; i++) {
718 if (chp->ch_drive[i].drive_type == ATA_DRIVET_ATAPI)
719 continue;
720 if (child == chp->ch_drive[i].drv_softc) {
721 chp->ch_drive[i].drv_softc = NULL;
722 chp->ch_drive[i].drive_flags = 0;
723 if (chp->ch_drive[i].drive_type == ATA_DRIVET_PM)
724 chp->ch_satapmp_nports = 0;
725 chp->ch_drive[i].drive_type = ATA_DRIVET_NONE;
726 found = true;
727 }
728 }
729
730 if (!found)
731 panic("%s: unknown child %p", device_xname(self),
732 (const void *)child);
733 }
734
735 CFATTACH_DECL3_NEW(atabus, sizeof(struct atabus_softc),
736 atabus_match, atabus_attach, atabus_detach, NULL, atabus_rescan,
737 atabus_childdetached, DVF_DETACH_SHUTDOWN);
738
739 /*****************************************************************************
740 * Common ATA bus operations.
741 *****************************************************************************/
742
743 /* allocate/free the channel's ch_drive[] array */
744 int
745 atabus_alloc_drives(struct ata_channel *chp, int ndrives)
746 {
747 int i;
748 if (chp->ch_ndrives != ndrives)
749 atabus_free_drives(chp);
750 if (chp->ch_drive == NULL) {
751 chp->ch_drive = kmem_zalloc(
752 sizeof(struct ata_drive_datas) * ndrives, KM_NOSLEEP);
753 }
754 if (chp->ch_drive == NULL) {
755 aprint_error_dev(chp->ch_atac->atac_dev,
756 "can't alloc drive array\n");
757 chp->ch_ndrives = 0;
758 return ENOMEM;
759 };
760 for (i = 0; i < ndrives; i++) {
761 chp->ch_drive[i].chnl_softc = chp;
762 chp->ch_drive[i].drive = i;
763 }
764 chp->ch_ndrives = ndrives;
765 return 0;
766 }
767
768 void
769 atabus_free_drives(struct ata_channel *chp)
770 {
771 #ifdef DIAGNOSTIC
772 int i;
773 int dopanic = 0;
774 KASSERT(chp->ch_ndrives == 0 || chp->ch_drive != NULL);
775 for (i = 0; i < chp->ch_ndrives; i++) {
776 if (chp->ch_drive[i].drive_type != ATA_DRIVET_NONE) {
777 printf("%s: ch_drive[%d] type %d != ATA_DRIVET_NONE\n",
778 device_xname(chp->atabus), i,
779 chp->ch_drive[i].drive_type);
780 dopanic = 1;
781 }
782 if (chp->ch_drive[i].drv_softc != NULL) {
783 printf("%s: ch_drive[%d] attached to %s\n",
784 device_xname(chp->atabus), i,
785 device_xname(chp->ch_drive[i].drv_softc));
786 dopanic = 1;
787 }
788 }
789 if (dopanic)
790 panic("atabus_free_drives");
791 #endif
792
793 if (chp->ch_drive == NULL)
794 return;
795 kmem_free(chp->ch_drive,
796 sizeof(struct ata_drive_datas) * chp->ch_ndrives);
797 chp->ch_ndrives = 0;
798 chp->ch_drive = NULL;
799 }
800
801 /* Get the disk's parameters */
802 int
803 ata_get_params(struct ata_drive_datas *drvp, uint8_t flags,
804 struct ataparams *prms)
805 {
806 struct ata_xfer *xfer;
807 struct ata_channel *chp = drvp->chnl_softc;
808 struct atac_softc *atac = chp->ch_atac;
809 char *tb;
810 int i, rv;
811 uint16_t *p;
812
813 ATADEBUG_PRINT(("%s\n", __func__), DEBUG_FUNCS);
814
815 xfer = ata_get_xfer(chp, false);
816 if (xfer == NULL) {
817 ATADEBUG_PRINT(("%s: no xfer\n", __func__),
818 DEBUG_FUNCS|DEBUG_PROBE);
819 return CMD_AGAIN;
820 }
821
822 tb = kmem_zalloc(ATA_BSIZE, KM_SLEEP);
823 memset(prms, 0, sizeof(struct ataparams));
824
825 if (drvp->drive_type == ATA_DRIVET_ATA) {
826 xfer->c_ata_c.r_command = WDCC_IDENTIFY;
827 xfer->c_ata_c.r_st_bmask = WDCS_DRDY;
828 xfer->c_ata_c.r_st_pmask = WDCS_DRQ;
829 xfer->c_ata_c.timeout = 3000; /* 3s */
830 } else if (drvp->drive_type == ATA_DRIVET_ATAPI) {
831 xfer->c_ata_c.r_command = ATAPI_IDENTIFY_DEVICE;
832 xfer->c_ata_c.r_st_bmask = 0;
833 xfer->c_ata_c.r_st_pmask = WDCS_DRQ;
834 xfer->c_ata_c.timeout = 10000; /* 10s */
835 } else {
836 ATADEBUG_PRINT(("ata_get_parms: no disks\n"),
837 DEBUG_FUNCS|DEBUG_PROBE);
838 rv = CMD_ERR;
839 goto out;
840 }
841 xfer->c_ata_c.flags = AT_READ | flags;
842 xfer->c_ata_c.data = tb;
843 xfer->c_ata_c.bcount = ATA_BSIZE;
844 if ((*atac->atac_bustype_ata->ata_exec_command)(drvp,
845 xfer) != ATACMD_COMPLETE) {
846 ATADEBUG_PRINT(("ata_get_parms: wdc_exec_command failed\n"),
847 DEBUG_FUNCS|DEBUG_PROBE);
848 rv = CMD_AGAIN;
849 goto out;
850 }
851 if (xfer->c_ata_c.flags & (AT_ERROR | AT_TIMEOU | AT_DF)) {
852 ATADEBUG_PRINT(("ata_get_parms: ata_c.flags=0x%x\n",
853 xfer->c_ata_c.flags), DEBUG_FUNCS|DEBUG_PROBE);
854 rv = CMD_ERR;
855 goto out;
856 }
857 /* if we didn't read any data something is wrong */
858 if ((xfer->c_ata_c.flags & AT_XFDONE) == 0) {
859 rv = CMD_ERR;
860 goto out;
861 }
862
863 /* Read in parameter block. */
864 memcpy(prms, tb, sizeof(struct ataparams));
865
866 /*
867 * Shuffle string byte order.
868 * ATAPI NEC, Mitsumi and Pioneer drives and
869 * old ATA TDK CompactFlash cards
870 * have different byte order.
871 */
872 #if BYTE_ORDER == BIG_ENDIAN
873 # define M(n) prms->atap_model[(n) ^ 1]
874 #else
875 # define M(n) prms->atap_model[n]
876 #endif
877 if (
878 #if BYTE_ORDER == BIG_ENDIAN
879 !
880 #endif
881 ((drvp->drive_type == ATA_DRIVET_ATAPI) ?
882 ((M(0) == 'N' && M(1) == 'E') ||
883 (M(0) == 'F' && M(1) == 'X') ||
884 (M(0) == 'P' && M(1) == 'i')) :
885 ((M(0) == 'T' && M(1) == 'D' && M(2) == 'K')))) {
886 rv = CMD_OK;
887 goto out;
888 }
889 #undef M
890 for (i = 0; i < sizeof(prms->atap_model); i += 2) {
891 p = (uint16_t *)(prms->atap_model + i);
892 *p = bswap16(*p);
893 }
894 for (i = 0; i < sizeof(prms->atap_serial); i += 2) {
895 p = (uint16_t *)(prms->atap_serial + i);
896 *p = bswap16(*p);
897 }
898 for (i = 0; i < sizeof(prms->atap_revision); i += 2) {
899 p = (uint16_t *)(prms->atap_revision + i);
900 *p = bswap16(*p);
901 }
902
903 rv = CMD_OK;
904 out:
905 kmem_free(tb, ATA_BSIZE);
906 ata_free_xfer(chp, xfer);
907 return rv;
908 }
909
910 int
911 ata_set_mode(struct ata_drive_datas *drvp, uint8_t mode, uint8_t flags)
912 {
913 struct ata_xfer *xfer;
914 int rv;
915 struct ata_channel *chp = drvp->chnl_softc;
916 struct atac_softc *atac = chp->ch_atac;
917
918 ATADEBUG_PRINT(("ata_set_mode=0x%x\n", mode), DEBUG_FUNCS);
919
920 xfer = ata_get_xfer(chp, false);
921 if (xfer == NULL) {
922 ATADEBUG_PRINT(("%s: no xfer\n", __func__),
923 DEBUG_FUNCS|DEBUG_PROBE);
924 return CMD_AGAIN;
925 }
926
927 xfer->c_ata_c.r_command = SET_FEATURES;
928 xfer->c_ata_c.r_st_bmask = 0;
929 xfer->c_ata_c.r_st_pmask = 0;
930 xfer->c_ata_c.r_features = WDSF_SET_MODE;
931 xfer->c_ata_c.r_count = mode;
932 xfer->c_ata_c.flags = flags;
933 xfer->c_ata_c.timeout = 1000; /* 1s */
934 if ((*atac->atac_bustype_ata->ata_exec_command)(drvp,
935 xfer) != ATACMD_COMPLETE) {
936 rv = CMD_AGAIN;
937 goto out;
938 }
939 if (xfer->c_ata_c.flags & (AT_ERROR | AT_TIMEOU | AT_DF)) {
940 rv = CMD_ERR;
941 goto out;
942 }
943
944 rv = CMD_OK;
945
946 out:
947 ata_free_xfer(chp, xfer);
948 return rv;
949 }
950
951 #if NATA_DMA
952 void
953 ata_dmaerr(struct ata_drive_datas *drvp, int flags)
954 {
955 ata_channel_lock_owned(drvp->chnl_softc);
956
957 /*
958 * Downgrade decision: if we get NERRS_MAX in NXFER.
959 * We start with n_dmaerrs set to NERRS_MAX-1 so that the
960 * first error within the first NXFER ops will immediatly trigger
961 * a downgrade.
962 * If we got an error and n_xfers is bigger than NXFER reset counters.
963 */
964 drvp->n_dmaerrs++;
965 if (drvp->n_dmaerrs >= NERRS_MAX && drvp->n_xfers <= NXFER) {
966 ata_downgrade_mode(drvp, flags);
967 drvp->n_dmaerrs = NERRS_MAX-1;
968 drvp->n_xfers = 0;
969 return;
970 }
971 if (drvp->n_xfers > NXFER) {
972 drvp->n_dmaerrs = 1; /* just got an error */
973 drvp->n_xfers = 1; /* restart counting from this error */
974 }
975 }
976 #endif /* NATA_DMA */
977
978 /*
979 * freeze the queue and wait for the controller to be idle. Caller has to
980 * unfreeze/restart the queue
981 */
982 static void
983 ata_channel_idle(struct ata_channel *chp)
984 {
985 ata_channel_lock(chp);
986 ata_channel_freeze_locked(chp);
987 while (chp->ch_queue->queue_active > 0) {
988 chp->ch_queue->queue_flags |= QF_IDLE_WAIT;
989 cv_timedwait(&chp->ch_queue->queue_idle, &chp->ch_lock, 1);
990 }
991 ata_channel_unlock(chp);
992 }
993
994 /*
995 * Add a command to the queue and start controller.
996 *
997 * MUST BE CALLED AT splbio()!
998 */
999 void
1000 ata_exec_xfer(struct ata_channel *chp, struct ata_xfer *xfer)
1001 {
1002
1003 ATADEBUG_PRINT(("ata_exec_xfer %p channel %d drive %d\n", xfer,
1004 chp->ch_channel, xfer->c_drive), DEBUG_XFERS);
1005
1006 /* complete xfer setup */
1007 xfer->c_chp = chp;
1008
1009 ata_channel_lock(chp);
1010
1011 /*
1012 * Standard commands are added to the end of command list, but
1013 * recovery commands must be run immediatelly.
1014 */
1015 if ((xfer->c_flags & C_SKIP_QUEUE) == 0)
1016 SIMPLEQ_INSERT_TAIL(&chp->ch_queue->queue_xfer, xfer,
1017 c_xferchain);
1018 else
1019 SIMPLEQ_INSERT_HEAD(&chp->ch_queue->queue_xfer, xfer,
1020 c_xferchain);
1021
1022 /*
1023 * if polling and can sleep, wait for the xfer to be at head of queue
1024 */
1025 if ((xfer->c_flags & (C_POLL | C_WAIT)) == (C_POLL | C_WAIT)) {
1026 while (chp->ch_queue->queue_active > 0 ||
1027 SIMPLEQ_FIRST(&chp->ch_queue->queue_xfer) != xfer) {
1028 xfer->c_flags |= C_WAITACT;
1029 cv_wait(&chp->ch_queue->c_active, &chp->ch_lock);
1030 xfer->c_flags &= ~C_WAITACT;
1031 }
1032
1033 /*
1034 * Free xfer now if it there was attempt to free it
1035 * while we were waiting.
1036 */
1037 if ((xfer->c_flags & (C_FREE|C_WAITTIMO)) == C_FREE) {
1038 ata_channel_unlock(chp);
1039
1040 ata_free_xfer(chp, xfer);
1041 return;
1042 }
1043 }
1044
1045 ata_channel_unlock(chp);
1046
1047 ATADEBUG_PRINT(("atastart from ata_exec_xfer, flags 0x%x\n",
1048 chp->ch_flags), DEBUG_XFERS);
1049 atastart(chp);
1050 }
1051
1052 /*
1053 * Start I/O on a controller, for the given channel.
1054 * The first xfer may be not for our channel if the channel queues
1055 * are shared.
1056 *
1057 * MUST BE CALLED AT splbio()!
1058 *
1059 * XXX FIS-based switching with PMP
1060 * Currently atastart() never schedules concurrent NCQ transfers to more than
1061 * one drive, even when channel has several SATA drives attached via PMP.
1062 * To support concurrent transfers to different drives with PMP, it would be
1063 * necessary to implement FIS-based switching support in controller driver,
1064 * and then adjust error handling and recovery to stop assuming at most
1065 * one active drive.
1066 */
1067 void
1068 atastart(struct ata_channel *chp)
1069 {
1070 struct atac_softc *atac = chp->ch_atac;
1071 struct ata_queue *chq = chp->ch_queue;
1072 struct ata_xfer *xfer, *axfer;
1073 bool skipq;
1074
1075 #ifdef ATA_DEBUG
1076 int spl1, spl2;
1077
1078 spl1 = splbio();
1079 spl2 = splbio();
1080 if (spl2 != spl1) {
1081 printf("atastart: not at splbio()\n");
1082 panic("atastart");
1083 }
1084 splx(spl2);
1085 splx(spl1);
1086 #endif /* ATA_DEBUG */
1087
1088 ata_channel_lock(chp);
1089
1090 again:
1091 /* is there a xfer ? */
1092 if ((xfer = SIMPLEQ_FIRST(&chp->ch_queue->queue_xfer)) == NULL) {
1093 ATADEBUG_PRINT(("%s(chp=%p): channel %d queue_xfer is empty\n",
1094 __func__, chp, chp->ch_channel), DEBUG_XFERS);
1095 goto out;
1096 }
1097
1098 /*
1099 * if someone is waiting for the command to be active, wake it up
1100 * and let it process the command
1101 */
1102 if (__predict_false(xfer->c_flags & C_WAITACT)) {
1103 ATADEBUG_PRINT(("atastart: xfer %p channel %d drive %d "
1104 "wait active\n", xfer, chp->ch_channel, xfer->c_drive),
1105 DEBUG_XFERS);
1106 cv_broadcast(&chp->ch_queue->c_active);
1107 goto out;
1108 }
1109
1110 skipq = ISSET(xfer->c_flags, C_SKIP_QUEUE);
1111
1112 /* is the queue frozen? */
1113 if (__predict_false(!skipq && chq->queue_freeze > 0)) {
1114 if (chq->queue_flags & QF_IDLE_WAIT) {
1115 chq->queue_flags &= ~QF_IDLE_WAIT;
1116 cv_signal(&chp->ch_queue->queue_idle);
1117 }
1118 ATADEBUG_PRINT(("%s(chp=%p): channel %d drive %d "
1119 "queue frozen: %d\n",
1120 __func__, chp, chp->ch_channel, xfer->c_drive,
1121 chq->queue_freeze),
1122 DEBUG_XFERS);
1123 goto out;
1124 }
1125
1126 /* all xfers on same queue must belong to the same channel */
1127 KASSERT(xfer->c_chp == chp);
1128
1129 /*
1130 * Can only take the command if there are no current active
1131 * commands, or if the command is NCQ and the active commands are also
1132 * NCQ. If PM is in use and HBA driver doesn't support/use FIS-based
1133 * switching, can only send commands to single drive.
1134 * Need only check first xfer.
1135 * XXX FIS-based switching - revisit
1136 */
1137 if (!skipq && (axfer = TAILQ_FIRST(&chp->ch_queue->active_xfers))) {
1138 if (!ISSET(xfer->c_flags, C_NCQ) ||
1139 !ISSET(axfer->c_flags, C_NCQ) ||
1140 xfer->c_drive != axfer->c_drive)
1141 goto out;
1142 }
1143
1144 struct ata_drive_datas * const drvp = &chp->ch_drive[xfer->c_drive];
1145
1146 /*
1147 * Are we on limit of active xfers ? If the queue has more
1148 * than 1 openings, we keep one slot reserved for recovery or dump.
1149 */
1150 KASSERT(chq->queue_active <= chq->queue_openings);
1151 const uint8_t chq_openings = (!skipq && chq->queue_openings > 1)
1152 ? (chq->queue_openings - 1) : chq->queue_openings;
1153 const uint8_t drv_openings = ISSET(xfer->c_flags, C_NCQ)
1154 ? drvp->drv_openings : ATA_MAX_OPENINGS;
1155 if (chq->queue_active >= MIN(chq_openings, drv_openings)) {
1156 if (skipq) {
1157 panic("%s: channel %d busy, xfer not possible",
1158 __func__, chp->ch_channel);
1159 }
1160
1161 ATADEBUG_PRINT(("%s(chp=%p): channel %d completely busy\n",
1162 __func__, chp, chp->ch_channel), DEBUG_XFERS);
1163 goto out;
1164 }
1165
1166 /* Slot allocation can fail if drv_openings < ch_openings */
1167 if (!ata_queue_alloc_slot(chp, &xfer->c_slot, drv_openings))
1168 goto out;
1169
1170 if (__predict_false(atac->atac_claim_hw)) {
1171 if (!atac->atac_claim_hw(chp, 0)) {
1172 ata_queue_free_slot(chp, xfer->c_slot);
1173 goto out;
1174 }
1175 }
1176
1177 /* Now committed to start the xfer */
1178
1179 ATADEBUG_PRINT(("%s(chp=%p): xfer %p channel %d drive %d\n",
1180 __func__, chp, xfer, chp->ch_channel, xfer->c_drive), DEBUG_XFERS);
1181 if (drvp->drive_flags & ATA_DRIVE_RESET) {
1182 drvp->drive_flags &= ~ATA_DRIVE_RESET;
1183 drvp->state = 0;
1184 }
1185
1186 if (ISSET(xfer->c_flags, C_NCQ))
1187 SET(chp->ch_flags, ATACH_NCQ);
1188 else
1189 CLR(chp->ch_flags, ATACH_NCQ);
1190
1191 SIMPLEQ_REMOVE_HEAD(&chq->queue_xfer, c_xferchain);
1192
1193 ata_activate_xfer_locked(chp, xfer);
1194
1195 if (atac->atac_cap & ATAC_CAP_NOIRQ)
1196 KASSERT(xfer->c_flags & C_POLL);
1197
1198 switch (ata_xfer_start(xfer)) {
1199 case ATASTART_TH:
1200 case ATASTART_ABORT:
1201 /* don't start any further commands in this case */
1202 goto out;
1203 default:
1204 /* nothing to do */
1205 break;
1206 }
1207
1208 /* Queue more commands if possible, but not during recovery or dump */
1209 if (!skipq && chq->queue_active < chq->queue_openings)
1210 goto again;
1211
1212 out:
1213 ata_channel_unlock(chp);
1214 }
1215
1216 int
1217 ata_xfer_start(struct ata_xfer *xfer)
1218 {
1219 struct ata_channel *chp = xfer->c_chp;
1220 int rv;
1221
1222 KASSERT(mutex_owned(&chp->ch_lock));
1223
1224 rv = xfer->ops->c_start(chp, xfer);
1225 switch (rv) {
1226 case ATASTART_STARTED:
1227 /* nothing to do */
1228 break;
1229 case ATASTART_TH:
1230 /* postpone xfer to thread */
1231 ata_thread_wake_locked(chp);
1232 break;
1233 case ATASTART_POLL:
1234 /* can happen even in thread context for some ATAPI devices */
1235 ata_channel_unlock(chp);
1236 KASSERT(xfer->ops != NULL && xfer->ops->c_poll != NULL);
1237 xfer->ops->c_poll(chp, xfer);
1238 ata_channel_lock(chp);
1239 break;
1240 case ATASTART_ABORT:
1241 ata_channel_unlock(chp);
1242 KASSERT(xfer->ops != NULL && xfer->ops->c_abort != NULL);
1243 xfer->ops->c_abort(chp, xfer);
1244 ata_channel_lock(chp);
1245 break;
1246 }
1247
1248 return rv;
1249 }
1250
1251 static void
1252 ata_activate_xfer_locked(struct ata_channel *chp, struct ata_xfer *xfer)
1253 {
1254 struct ata_queue * const chq = chp->ch_queue;
1255
1256 KASSERT(mutex_owned(&chp->ch_lock));
1257 KASSERT((chq->active_xfers_used & __BIT(xfer->c_slot)) == 0);
1258
1259 if ((xfer->c_flags & C_SKIP_QUEUE) == 0)
1260 TAILQ_INSERT_TAIL(&chq->active_xfers, xfer, c_activechain);
1261 else {
1262 /*
1263 * Must go to head, so that ata_queue_get_active_xfer()
1264 * returns the recovery command, and not some other
1265 * random active transfer.
1266 */
1267 TAILQ_INSERT_HEAD(&chq->active_xfers, xfer, c_activechain);
1268 }
1269 chq->active_xfers_used |= __BIT(xfer->c_slot);
1270 chq->queue_active++;
1271 }
1272
1273 /*
1274 * Does it's own locking, does not require splbio().
1275 * flags - whether to block waiting for free xfer
1276 */
1277 struct ata_xfer *
1278 ata_get_xfer(struct ata_channel *chp, bool waitok)
1279 {
1280 struct ata_xfer *xfer;
1281
1282 xfer = pool_get(&ata_xfer_pool, waitok ? PR_WAITOK : PR_NOWAIT);
1283 KASSERT(!waitok || xfer != NULL);
1284
1285 if (xfer != NULL) {
1286 /* zero everything */
1287 memset(xfer, 0, sizeof(*xfer));
1288 }
1289
1290 return xfer;
1291 }
1292
1293 /*
1294 * ata_deactivate_xfer() must be always called prior to ata_free_xfer()
1295 */
1296 void
1297 ata_free_xfer(struct ata_channel *chp, struct ata_xfer *xfer)
1298 {
1299 struct ata_queue *chq = chp->ch_queue;
1300
1301 ata_channel_lock(chp);
1302
1303 if (__predict_false(xfer->c_flags & (C_WAITACT|C_WAITTIMO))) {
1304 /* Someone is waiting for this xfer, so we can't free now */
1305 xfer->c_flags |= C_FREE;
1306 cv_broadcast(&chq->c_active);
1307 ata_channel_unlock(chp);
1308 return;
1309 }
1310
1311 /* XXX move PIOBM and free_gw to deactivate? */
1312 #if NATA_PIOBM /* XXX wdc dependent code */
1313 if (__predict_false(xfer->c_flags & C_PIOBM)) {
1314 struct wdc_softc *wdc = CHAN_TO_WDC(chp);
1315
1316 /* finish the busmastering PIO */
1317 (*wdc->piobm_done)(wdc->dma_arg,
1318 chp->ch_channel, xfer->c_drive);
1319 chp->ch_flags &= ~(ATACH_DMA_WAIT | ATACH_PIOBM_WAIT | ATACH_IRQ_WAIT);
1320 }
1321 #endif
1322
1323 if (__predict_false(chp->ch_atac->atac_free_hw))
1324 chp->ch_atac->atac_free_hw(chp);
1325
1326 ata_channel_unlock(chp);
1327
1328 if (__predict_true(!ISSET(xfer->c_flags, C_PRIVATE_ALLOC)))
1329 pool_put(&ata_xfer_pool, xfer);
1330 }
1331
1332 void
1333 ata_deactivate_xfer(struct ata_channel *chp, struct ata_xfer *xfer)
1334 {
1335 struct ata_queue * const chq = chp->ch_queue;
1336
1337 ata_channel_lock(chp);
1338
1339 KASSERT(chq->queue_active > 0);
1340 KASSERT((chq->active_xfers_used & __BIT(xfer->c_slot)) != 0);
1341
1342 /* Stop only when this is last active xfer */
1343 if (chq->queue_active == 1)
1344 callout_stop(&chp->c_timo_callout);
1345
1346 if (callout_invoking(&chp->c_timo_callout))
1347 xfer->c_flags |= C_WAITTIMO;
1348
1349 TAILQ_REMOVE(&chq->active_xfers, xfer, c_activechain);
1350 chq->active_xfers_used &= ~__BIT(xfer->c_slot);
1351 chq->queue_active--;
1352
1353 ata_queue_free_slot(chp, xfer->c_slot);
1354
1355 if (xfer->c_flags & C_WAIT)
1356 cv_broadcast(&chq->c_cmd_finish);
1357
1358 ata_channel_unlock(chp);
1359 }
1360
1361 /*
1362 * Called in c_intr hook. Must be called before before any deactivations
1363 * are done - if there is drain pending, it calls c_kill_xfer hook which
1364 * deactivates the xfer.
1365 * Calls c_kill_xfer with channel lock free.
1366 * Returns true if caller should just exit without further processing.
1367 * Caller must not further access any part of xfer or any related controller
1368 * structures in that case, it should just return.
1369 */
1370 bool
1371 ata_waitdrain_xfer_check(struct ata_channel *chp, struct ata_xfer *xfer)
1372 {
1373 int drive = xfer->c_drive;
1374 bool draining = false;
1375
1376 ata_channel_lock(chp);
1377
1378 if (chp->ch_drive[drive].drive_flags & ATA_DRIVE_WAITDRAIN) {
1379 ata_channel_unlock(chp);
1380
1381 xfer->ops->c_kill_xfer(chp, xfer, KILL_GONE);
1382
1383 ata_channel_lock(chp);
1384 chp->ch_drive[drive].drive_flags &= ~ATA_DRIVE_WAITDRAIN;
1385 cv_signal(&chp->ch_queue->queue_drain);
1386 draining = true;
1387 }
1388
1389 ata_channel_unlock(chp);
1390
1391 return draining;
1392 }
1393
1394 /*
1395 * Check for race of normal transfer handling vs. timeout.
1396 */
1397 bool
1398 ata_timo_xfer_check(struct ata_xfer *xfer)
1399 {
1400 struct ata_channel *chp = xfer->c_chp;
1401 struct ata_drive_datas *drvp = &chp->ch_drive[xfer->c_drive];
1402
1403 ata_channel_lock(chp);
1404
1405 if (xfer->c_flags & C_WAITTIMO) {
1406 xfer->c_flags &= ~C_WAITTIMO;
1407
1408 /* Handle race vs. ata_free_xfer() */
1409 if (xfer->c_flags & C_FREE) {
1410 xfer->c_flags &= ~C_FREE;
1411 ata_channel_unlock(chp);
1412
1413 device_printf(drvp->drv_softc,
1414 "xfer %"PRIxPTR" freed while invoking timeout\n",
1415 (intptr_t)xfer & PAGE_MASK);
1416
1417 ata_free_xfer(chp, xfer);
1418 return true;
1419 }
1420
1421 /* Race vs. callout_stop() in ata_deactivate_xfer() */
1422 ata_channel_unlock(chp);
1423
1424 device_printf(drvp->drv_softc,
1425 "xfer %"PRIxPTR" deactivated while invoking timeout\n",
1426 (intptr_t)xfer & PAGE_MASK);
1427 return true;
1428 }
1429
1430 ata_channel_unlock(chp);
1431
1432 /* No race, proceed with timeout handling */
1433 return false;
1434 }
1435
1436 /*
1437 * Kill off all active xfers for a ata_channel.
1438 *
1439 * Must be called with channel lock held.
1440 */
1441 void
1442 ata_kill_active(struct ata_channel *chp, int reason, int flags)
1443 {
1444 struct ata_queue * const chq = chp->ch_queue;
1445 struct ata_xfer *xfer, *xfernext;
1446
1447 KASSERT(mutex_owned(&chp->ch_lock));
1448
1449 TAILQ_FOREACH_SAFE(xfer, &chq->active_xfers, c_activechain, xfernext) {
1450 ata_channel_unlock(chp);
1451 xfer->ops->c_kill_xfer(xfer->c_chp, xfer, reason);
1452 ata_channel_lock(chp);
1453 }
1454 }
1455
1456 /*
1457 * Kill off all pending xfers for a drive.
1458 */
1459 void
1460 ata_kill_pending(struct ata_drive_datas *drvp)
1461 {
1462 struct ata_channel * const chp = drvp->chnl_softc;
1463 struct ata_queue * const chq = chp->ch_queue;
1464 struct ata_xfer *xfer;
1465
1466 ata_channel_lock(chp);
1467
1468 /* Kill all pending transfers */
1469 while ((xfer = SIMPLEQ_FIRST(&chq->queue_xfer))) {
1470 KASSERT(xfer->c_chp == chp);
1471
1472 if (xfer->c_drive != drvp->drive)
1473 continue;
1474
1475 SIMPLEQ_REMOVE_HEAD(&chp->ch_queue->queue_xfer, c_xferchain);
1476
1477 /*
1478 * Keep the lock, so that we get deadlock (and 'locking against
1479 * myself' with LOCKDEBUG), instead of silent
1480 * data corruption, if the hook tries to call back into
1481 * middle layer for inactive xfer.
1482 */
1483 xfer->ops->c_kill_xfer(chp, xfer, KILL_GONE_INACTIVE);
1484 }
1485
1486 /* Wait until all active transfers on the drive finish */
1487 while (chq->queue_active > 0) {
1488 bool drv_active = false;
1489
1490 TAILQ_FOREACH(xfer, &chq->active_xfers, c_activechain) {
1491 KASSERT(xfer->c_chp == chp);
1492
1493 if (xfer->c_drive == drvp->drive) {
1494 drv_active = true;
1495 break;
1496 }
1497 }
1498
1499 if (!drv_active) {
1500 /* all finished */
1501 break;
1502 }
1503
1504 drvp->drive_flags |= ATA_DRIVE_WAITDRAIN;
1505 cv_wait(&chq->queue_drain, &chp->ch_lock);
1506 }
1507
1508 ata_channel_unlock(chp);
1509 }
1510
1511 static void
1512 ata_channel_freeze_locked(struct ata_channel *chp)
1513 {
1514 chp->ch_queue->queue_freeze++;
1515
1516 ATADEBUG_PRINT(("%s(chp=%p) -> %d\n", __func__, chp,
1517 chp->ch_queue->queue_freeze), DEBUG_FUNCS | DEBUG_XFERS);
1518 }
1519
1520 void
1521 ata_channel_freeze(struct ata_channel *chp)
1522 {
1523 ata_channel_lock(chp);
1524 ata_channel_freeze_locked(chp);
1525 ata_channel_unlock(chp);
1526 }
1527
1528 void
1529 ata_channel_thaw_locked(struct ata_channel *chp)
1530 {
1531 KASSERT(mutex_owned(&chp->ch_lock));
1532 KASSERT(chp->ch_queue->queue_freeze > 0);
1533
1534 chp->ch_queue->queue_freeze--;
1535
1536 ATADEBUG_PRINT(("%s(chp=%p) -> %d\n", __func__, chp,
1537 chp->ch_queue->queue_freeze), DEBUG_FUNCS | DEBUG_XFERS);
1538 }
1539
1540 /*
1541 * ata_thread_run:
1542 *
1543 * Reset and ATA channel. Channel lock must be held. arg is type-specific.
1544 */
1545 void
1546 ata_thread_run(struct ata_channel *chp, int flags, int type, int arg)
1547 {
1548 struct atac_softc *atac = chp->ch_atac;
1549 bool threset = false;
1550 struct ata_drive_datas *drvp;
1551
1552 ata_channel_lock_owned(chp);
1553
1554 /*
1555 * If we can poll or wait it's OK, otherwise wake up the
1556 * kernel thread to do it for us.
1557 */
1558 ATADEBUG_PRINT(("%s flags 0x%x ch_flags 0x%x\n",
1559 __func__, flags, chp->ch_flags), DEBUG_FUNCS | DEBUG_XFERS);
1560 if ((flags & (AT_POLL | AT_WAIT)) == 0) {
1561 switch (type) {
1562 case ATACH_TH_RESET:
1563 if (chp->ch_flags & ATACH_TH_RESET) {
1564 /* No need to schedule another reset */
1565 return;
1566 }
1567 break;
1568 case ATACH_TH_DRIVE_RESET:
1569 {
1570 int drive = arg;
1571
1572 KASSERT(drive <= chp->ch_ndrives);
1573 drvp = &chp->ch_drive[drive];
1574
1575 if (drvp->drive_flags & ATA_DRIVE_TH_RESET) {
1576 /* No need to schedule another reset */
1577 return;
1578 }
1579 drvp->drive_flags |= ATA_DRIVE_TH_RESET;
1580 break;
1581 }
1582 case ATACH_TH_RECOVERY:
1583 {
1584 uint32_t tfd = (uint32_t)arg;
1585
1586 KASSERT((chp->ch_flags & ATACH_RECOVERING) == 0);
1587 chp->recovery_tfd = tfd;
1588 break;
1589 }
1590 default:
1591 panic("%s: unknown type: %x", __func__, type);
1592 /* NOTREACHED */
1593 }
1594
1595 /*
1596 * Block execution of other commands while reset is scheduled
1597 * to a thread.
1598 */
1599 ata_channel_freeze_locked(chp);
1600 chp->ch_flags |= type;
1601
1602 cv_signal(&chp->ch_thr_idle);
1603 return;
1604 }
1605
1606 /* Block execution of other commands during reset */
1607 ata_channel_freeze_locked(chp);
1608
1609 /*
1610 * If reset has been scheduled to a thread, then clear
1611 * the flag now so that the thread won't try to execute it if
1612 * we happen to sleep, and thaw one more time after the reset.
1613 */
1614 if (chp->ch_flags & type) {
1615 chp->ch_flags &= ~type;
1616 threset = true;
1617 }
1618
1619 switch (type) {
1620 case ATACH_TH_RESET:
1621 (*atac->atac_bustype_ata->ata_reset_channel)(chp, flags);
1622
1623 KASSERT(chp->ch_ndrives == 0 || chp->ch_drive != NULL);
1624 for (int drive = 0; drive < chp->ch_ndrives; drive++)
1625 chp->ch_drive[drive].state = 0;
1626 break;
1627
1628 case ATACH_TH_DRIVE_RESET:
1629 {
1630 int drive = arg;
1631
1632 KASSERT(drive <= chp->ch_ndrives);
1633 drvp = &chp->ch_drive[drive];
1634 (*atac->atac_bustype_ata->ata_reset_drive)(drvp, flags, NULL);
1635 drvp->state = 0;
1636 break;
1637 }
1638
1639 case ATACH_TH_RECOVERY:
1640 {
1641 uint32_t tfd = (uint32_t)arg;
1642
1643 KASSERT((chp->ch_flags & ATACH_RECOVERING) == 0);
1644 KASSERT(atac->atac_bustype_ata->ata_recovery != NULL);
1645
1646 SET(chp->ch_flags, ATACH_RECOVERING);
1647 (*atac->atac_bustype_ata->ata_recovery)(chp, flags, tfd);
1648 CLR(chp->ch_flags, ATACH_RECOVERING);
1649 break;
1650 }
1651
1652 default:
1653 panic("%s: unknown type: %x", __func__, type);
1654 /* NOTREACHED */
1655 }
1656
1657 /*
1658 * Thaw one extra time to clear the freeze done when the reset has
1659 * been scheduled to the thread.
1660 */
1661 if (threset)
1662 ata_channel_thaw_locked(chp);
1663
1664 /* Allow commands to run again */
1665 ata_channel_thaw_locked(chp);
1666
1667 /* Signal the thread in case there is an xfer to run */
1668 cv_signal(&chp->ch_thr_idle);
1669 }
1670
1671 int
1672 ata_addref(struct ata_channel *chp)
1673 {
1674 struct atac_softc *atac = chp->ch_atac;
1675 struct scsipi_adapter *adapt = &atac->atac_atapi_adapter._generic;
1676 int s, error = 0;
1677
1678 s = splbio();
1679 if (adapt->adapt_refcnt++ == 0 &&
1680 adapt->adapt_enable != NULL) {
1681 error = (*adapt->adapt_enable)(atac->atac_dev, 1);
1682 if (error)
1683 adapt->adapt_refcnt--;
1684 }
1685 splx(s);
1686 return (error);
1687 }
1688
1689 void
1690 ata_delref(struct ata_channel *chp)
1691 {
1692 struct atac_softc *atac = chp->ch_atac;
1693 struct scsipi_adapter *adapt = &atac->atac_atapi_adapter._generic;
1694 int s;
1695
1696 s = splbio();
1697 if (adapt->adapt_refcnt-- == 1 &&
1698 adapt->adapt_enable != NULL)
1699 (void) (*adapt->adapt_enable)(atac->atac_dev, 0);
1700 splx(s);
1701 }
1702
1703 void
1704 ata_print_modes(struct ata_channel *chp)
1705 {
1706 struct atac_softc *atac = chp->ch_atac;
1707 int drive;
1708 struct ata_drive_datas *drvp;
1709
1710 KASSERT(chp->ch_ndrives == 0 || chp->ch_drive != NULL);
1711 for (drive = 0; drive < chp->ch_ndrives; drive++) {
1712 drvp = &chp->ch_drive[drive];
1713 if (drvp->drive_type == ATA_DRIVET_NONE ||
1714 drvp->drv_softc == NULL)
1715 continue;
1716 aprint_verbose("%s(%s:%d:%d): using PIO mode %d",
1717 device_xname(drvp->drv_softc),
1718 device_xname(atac->atac_dev),
1719 chp->ch_channel, drvp->drive, drvp->PIO_mode);
1720 #if NATA_DMA
1721 if (drvp->drive_flags & ATA_DRIVE_DMA)
1722 aprint_verbose(", DMA mode %d", drvp->DMA_mode);
1723 #if NATA_UDMA
1724 if (drvp->drive_flags & ATA_DRIVE_UDMA) {
1725 aprint_verbose(", Ultra-DMA mode %d", drvp->UDMA_mode);
1726 if (drvp->UDMA_mode == 2)
1727 aprint_verbose(" (Ultra/33)");
1728 else if (drvp->UDMA_mode == 4)
1729 aprint_verbose(" (Ultra/66)");
1730 else if (drvp->UDMA_mode == 5)
1731 aprint_verbose(" (Ultra/100)");
1732 else if (drvp->UDMA_mode == 6)
1733 aprint_verbose(" (Ultra/133)");
1734 }
1735 #endif /* NATA_UDMA */
1736 #endif /* NATA_DMA */
1737 #if NATA_DMA || NATA_PIOBM
1738 if (0
1739 #if NATA_DMA
1740 || (drvp->drive_flags & (ATA_DRIVE_DMA | ATA_DRIVE_UDMA))
1741 #endif
1742 #if NATA_PIOBM
1743 /* PIOBM capable controllers use DMA for PIO commands */
1744 || (atac->atac_cap & ATAC_CAP_PIOBM)
1745 #endif
1746 )
1747 aprint_verbose(" (using DMA)");
1748
1749 if (drvp->drive_flags & ATA_DRIVE_NCQ) {
1750 aprint_verbose(", NCQ (%d tags)%s",
1751 ATA_REAL_OPENINGS(chp->ch_queue->queue_openings),
1752 (drvp->drive_flags & ATA_DRIVE_NCQ_PRIO)
1753 ? " w/PRIO" : "");
1754 } else if (drvp->drive_flags & ATA_DRIVE_WFUA)
1755 aprint_verbose(", WRITE DMA FUA EXT");
1756
1757 #endif /* NATA_DMA || NATA_PIOBM */
1758 aprint_verbose("\n");
1759 }
1760 }
1761
1762 #if NATA_DMA
1763 /*
1764 * downgrade the transfer mode of a drive after an error. return 1 if
1765 * downgrade was possible, 0 otherwise.
1766 *
1767 * MUST BE CALLED AT splbio()!
1768 */
1769 int
1770 ata_downgrade_mode(struct ata_drive_datas *drvp, int flags)
1771 {
1772 struct ata_channel *chp = drvp->chnl_softc;
1773 struct atac_softc *atac = chp->ch_atac;
1774 device_t drv_dev = drvp->drv_softc;
1775 int cf_flags = device_cfdata(drv_dev)->cf_flags;
1776
1777 ata_channel_lock_owned(drvp->chnl_softc);
1778
1779 /* if drive or controller don't know its mode, we can't do much */
1780 if ((drvp->drive_flags & ATA_DRIVE_MODE) == 0 ||
1781 (atac->atac_set_modes == NULL))
1782 return 0;
1783 /* current drive mode was set by a config flag, let it this way */
1784 if ((cf_flags & ATA_CONFIG_PIO_SET) ||
1785 (cf_flags & ATA_CONFIG_DMA_SET) ||
1786 (cf_flags & ATA_CONFIG_UDMA_SET))
1787 return 0;
1788
1789 #if NATA_UDMA
1790 /*
1791 * If we were using Ultra-DMA mode, downgrade to the next lower mode.
1792 */
1793 if ((drvp->drive_flags & ATA_DRIVE_UDMA) && drvp->UDMA_mode >= 2) {
1794 drvp->UDMA_mode--;
1795 aprint_error_dev(drv_dev,
1796 "transfer error, downgrading to Ultra-DMA mode %d\n",
1797 drvp->UDMA_mode);
1798 }
1799 #endif
1800
1801 /*
1802 * If we were using ultra-DMA, don't downgrade to multiword DMA.
1803 */
1804 else if (drvp->drive_flags & (ATA_DRIVE_DMA | ATA_DRIVE_UDMA)) {
1805 drvp->drive_flags &= ~(ATA_DRIVE_DMA | ATA_DRIVE_UDMA);
1806 drvp->PIO_mode = drvp->PIO_cap;
1807 aprint_error_dev(drv_dev,
1808 "transfer error, downgrading to PIO mode %d\n",
1809 drvp->PIO_mode);
1810 } else /* already using PIO, can't downgrade */
1811 return 0;
1812
1813 (*atac->atac_set_modes)(chp);
1814 ata_print_modes(chp);
1815 /* reset the channel, which will schedule all drives for setup */
1816 ata_thread_run(chp, flags, ATACH_TH_RESET, ATACH_NODRIVE);
1817 return 1;
1818 }
1819 #endif /* NATA_DMA */
1820
1821 /*
1822 * Probe drive's capabilities, for use by the controller later
1823 * Assumes drvp points to an existing drive.
1824 */
1825 void
1826 ata_probe_caps(struct ata_drive_datas *drvp)
1827 {
1828 struct ataparams params, params2;
1829 struct ata_channel *chp = drvp->chnl_softc;
1830 struct atac_softc *atac = chp->ch_atac;
1831 device_t drv_dev = drvp->drv_softc;
1832 int i, printed = 0;
1833 const char *sep = "";
1834 int cf_flags;
1835
1836 if (ata_get_params(drvp, AT_WAIT, ¶ms) != CMD_OK) {
1837 /* IDENTIFY failed. Can't tell more about the device */
1838 return;
1839 }
1840 if ((atac->atac_cap & (ATAC_CAP_DATA16 | ATAC_CAP_DATA32)) ==
1841 (ATAC_CAP_DATA16 | ATAC_CAP_DATA32)) {
1842 /*
1843 * Controller claims 16 and 32 bit transfers.
1844 * Re-do an IDENTIFY with 32-bit transfers,
1845 * and compare results.
1846 */
1847 ata_channel_lock(chp);
1848 drvp->drive_flags |= ATA_DRIVE_CAP32;
1849 ata_channel_unlock(chp);
1850 ata_get_params(drvp, AT_WAIT, ¶ms2);
1851 if (memcmp(¶ms, ¶ms2, sizeof(struct ataparams)) != 0) {
1852 /* Not good. fall back to 16bits */
1853 ata_channel_lock(chp);
1854 drvp->drive_flags &= ~ATA_DRIVE_CAP32;
1855 ata_channel_unlock(chp);
1856 } else {
1857 aprint_verbose_dev(drv_dev, "32-bit data port\n");
1858 }
1859 }
1860 #if 0 /* Some ultra-DMA drives claims to only support ATA-3. sigh */
1861 if (params.atap_ata_major > 0x01 &&
1862 params.atap_ata_major != 0xffff) {
1863 for (i = 14; i > 0; i--) {
1864 if (params.atap_ata_major & (1 << i)) {
1865 aprint_verbose_dev(drv_dev,
1866 "ATA version %d\n", i);
1867 drvp->ata_vers = i;
1868 break;
1869 }
1870 }
1871 }
1872 #endif
1873
1874 /* An ATAPI device is at last PIO mode 3 */
1875 if (drvp->drive_type == ATA_DRIVET_ATAPI)
1876 drvp->PIO_mode = 3;
1877
1878 /*
1879 * It's not in the specs, but it seems that some drive
1880 * returns 0xffff in atap_extensions when this field is invalid
1881 */
1882 if (params.atap_extensions != 0xffff &&
1883 (params.atap_extensions & WDC_EXT_MODES)) {
1884 /*
1885 * XXX some drives report something wrong here (they claim to
1886 * support PIO mode 8 !). As mode is coded on 3 bits in
1887 * SET FEATURE, limit it to 7 (so limit i to 4).
1888 * If higher mode than 7 is found, abort.
1889 */
1890 for (i = 7; i >= 0; i--) {
1891 if ((params.atap_piomode_supp & (1 << i)) == 0)
1892 continue;
1893 if (i > 4)
1894 return;
1895 /*
1896 * See if mode is accepted.
1897 * If the controller can't set its PIO mode,
1898 * assume the defaults are good, so don't try
1899 * to set it
1900 */
1901 if (atac->atac_set_modes)
1902 /*
1903 * It's OK to pool here, it's fast enough
1904 * to not bother waiting for interrupt
1905 */
1906 if (ata_set_mode(drvp, 0x08 | (i + 3),
1907 AT_WAIT) != CMD_OK)
1908 continue;
1909 if (!printed) {
1910 aprint_verbose_dev(drv_dev,
1911 "drive supports PIO mode %d", i + 3);
1912 sep = ",";
1913 printed = 1;
1914 }
1915 /*
1916 * If controller's driver can't set its PIO mode,
1917 * get the highter one for the drive.
1918 */
1919 if (atac->atac_set_modes == NULL ||
1920 atac->atac_pio_cap >= i + 3) {
1921 drvp->PIO_mode = i + 3;
1922 drvp->PIO_cap = i + 3;
1923 break;
1924 }
1925 }
1926 if (!printed) {
1927 /*
1928 * We didn't find a valid PIO mode.
1929 * Assume the values returned for DMA are buggy too
1930 */
1931 return;
1932 }
1933 ata_channel_lock(chp);
1934 drvp->drive_flags |= ATA_DRIVE_MODE;
1935 ata_channel_unlock(chp);
1936 printed = 0;
1937 for (i = 7; i >= 0; i--) {
1938 if ((params.atap_dmamode_supp & (1 << i)) == 0)
1939 continue;
1940 #if NATA_DMA
1941 if ((atac->atac_cap & ATAC_CAP_DMA) &&
1942 atac->atac_set_modes != NULL)
1943 if (ata_set_mode(drvp, 0x20 | i, AT_WAIT)
1944 != CMD_OK)
1945 continue;
1946 #endif
1947 if (!printed) {
1948 aprint_verbose("%s DMA mode %d", sep, i);
1949 sep = ",";
1950 printed = 1;
1951 }
1952 #if NATA_DMA
1953 if (atac->atac_cap & ATAC_CAP_DMA) {
1954 if (atac->atac_set_modes != NULL &&
1955 atac->atac_dma_cap < i)
1956 continue;
1957 drvp->DMA_mode = i;
1958 drvp->DMA_cap = i;
1959 ata_channel_lock(chp);
1960 drvp->drive_flags |= ATA_DRIVE_DMA;
1961 ata_channel_unlock(chp);
1962 }
1963 #endif
1964 break;
1965 }
1966 if (params.atap_extensions & WDC_EXT_UDMA_MODES) {
1967 printed = 0;
1968 for (i = 7; i >= 0; i--) {
1969 if ((params.atap_udmamode_supp & (1 << i))
1970 == 0)
1971 continue;
1972 #if NATA_UDMA
1973 if (atac->atac_set_modes != NULL &&
1974 (atac->atac_cap & ATAC_CAP_UDMA))
1975 if (ata_set_mode(drvp, 0x40 | i,
1976 AT_WAIT) != CMD_OK)
1977 continue;
1978 #endif
1979 if (!printed) {
1980 aprint_verbose("%s Ultra-DMA mode %d",
1981 sep, i);
1982 if (i == 2)
1983 aprint_verbose(" (Ultra/33)");
1984 else if (i == 4)
1985 aprint_verbose(" (Ultra/66)");
1986 else if (i == 5)
1987 aprint_verbose(" (Ultra/100)");
1988 else if (i == 6)
1989 aprint_verbose(" (Ultra/133)");
1990 sep = ",";
1991 printed = 1;
1992 }
1993 #if NATA_UDMA
1994 if (atac->atac_cap & ATAC_CAP_UDMA) {
1995 if (atac->atac_set_modes != NULL &&
1996 atac->atac_udma_cap < i)
1997 continue;
1998 drvp->UDMA_mode = i;
1999 drvp->UDMA_cap = i;
2000 ata_channel_lock(chp);
2001 drvp->drive_flags |= ATA_DRIVE_UDMA;
2002 ata_channel_unlock(chp);
2003 }
2004 #endif
2005 break;
2006 }
2007 }
2008 }
2009
2010 ata_channel_lock(chp);
2011 drvp->drive_flags &= ~ATA_DRIVE_NOSTREAM;
2012 if (drvp->drive_type == ATA_DRIVET_ATAPI) {
2013 if (atac->atac_cap & ATAC_CAP_ATAPI_NOSTREAM)
2014 drvp->drive_flags |= ATA_DRIVE_NOSTREAM;
2015 } else {
2016 if (atac->atac_cap & ATAC_CAP_ATA_NOSTREAM)
2017 drvp->drive_flags |= ATA_DRIVE_NOSTREAM;
2018 }
2019 ata_channel_unlock(chp);
2020
2021 /* Try to guess ATA version here, if it didn't get reported */
2022 if (drvp->ata_vers == 0) {
2023 #if NATA_UDMA
2024 if (drvp->drive_flags & ATA_DRIVE_UDMA)
2025 drvp->ata_vers = 4; /* should be at last ATA-4 */
2026 else
2027 #endif
2028 if (drvp->PIO_cap > 2)
2029 drvp->ata_vers = 2; /* should be at last ATA-2 */
2030 }
2031 cf_flags = device_cfdata(drv_dev)->cf_flags;
2032 if (cf_flags & ATA_CONFIG_PIO_SET) {
2033 ata_channel_lock(chp);
2034 drvp->PIO_mode =
2035 (cf_flags & ATA_CONFIG_PIO_MODES) >> ATA_CONFIG_PIO_OFF;
2036 drvp->drive_flags |= ATA_DRIVE_MODE;
2037 ata_channel_unlock(chp);
2038 }
2039 #if NATA_DMA
2040 if ((atac->atac_cap & ATAC_CAP_DMA) == 0) {
2041 /* don't care about DMA modes */
2042 return;
2043 }
2044 if (cf_flags & ATA_CONFIG_DMA_SET) {
2045 ata_channel_lock(chp);
2046 if ((cf_flags & ATA_CONFIG_DMA_MODES) ==
2047 ATA_CONFIG_DMA_DISABLE) {
2048 drvp->drive_flags &= ~ATA_DRIVE_DMA;
2049 } else {
2050 drvp->DMA_mode = (cf_flags & ATA_CONFIG_DMA_MODES) >>
2051 ATA_CONFIG_DMA_OFF;
2052 drvp->drive_flags |= ATA_DRIVE_DMA | ATA_DRIVE_MODE;
2053 }
2054 ata_channel_unlock(chp);
2055 }
2056
2057 /*
2058 * Probe WRITE DMA FUA EXT. Support is mandatory for devices
2059 * supporting LBA48, but nevertheless confirm with the feature flag.
2060 */
2061 if (drvp->drive_flags & ATA_DRIVE_DMA) {
2062 if ((params.atap_cmd2_en & ATA_CMD2_LBA48) != 0
2063 && (params.atap_cmd_def & ATA_CMDE_WFE)) {
2064 drvp->drive_flags |= ATA_DRIVE_WFUA;
2065 aprint_verbose("%s WRITE DMA FUA", sep);
2066 sep = ",";
2067 }
2068 }
2069
2070 /* Probe NCQ support - READ/WRITE FPDMA QUEUED command support */
2071 ata_channel_lock(chp);
2072 drvp->drv_openings = 1;
2073 if (params.atap_sata_caps & SATA_NATIVE_CMDQ) {
2074 if (atac->atac_cap & ATAC_CAP_NCQ)
2075 drvp->drive_flags |= ATA_DRIVE_NCQ;
2076 drvp->drv_openings =
2077 (params.atap_queuedepth & WDC_QUEUE_DEPTH_MASK) + 1;
2078 aprint_verbose("%s NCQ (%d tags)", sep, drvp->drv_openings);
2079 sep = ",";
2080
2081 if (params.atap_sata_caps & SATA_NCQ_PRIO) {
2082 drvp->drive_flags |= ATA_DRIVE_NCQ_PRIO;
2083 aprint_verbose(" w/PRIO");
2084 }
2085 }
2086 ata_channel_unlock(chp);
2087
2088 if (printed)
2089 aprint_verbose("\n");
2090
2091 #if NATA_UDMA
2092 if ((atac->atac_cap & ATAC_CAP_UDMA) == 0) {
2093 /* don't care about UDMA modes */
2094 return;
2095 }
2096 if (cf_flags & ATA_CONFIG_UDMA_SET) {
2097 ata_channel_lock(chp);
2098 if ((cf_flags & ATA_CONFIG_UDMA_MODES) ==
2099 ATA_CONFIG_UDMA_DISABLE) {
2100 drvp->drive_flags &= ~ATA_DRIVE_UDMA;
2101 } else {
2102 drvp->UDMA_mode = (cf_flags & ATA_CONFIG_UDMA_MODES) >>
2103 ATA_CONFIG_UDMA_OFF;
2104 drvp->drive_flags |= ATA_DRIVE_UDMA | ATA_DRIVE_MODE;
2105 }
2106 ata_channel_unlock(chp);
2107 }
2108 #endif /* NATA_UDMA */
2109 #endif /* NATA_DMA */
2110 }
2111
2112 /* management of the /dev/atabus* devices */
2113 int
2114 atabusopen(dev_t dev, int flag, int fmt, struct lwp *l)
2115 {
2116 struct atabus_softc *sc;
2117 int error;
2118
2119 sc = device_lookup_private(&atabus_cd, minor(dev));
2120 if (sc == NULL)
2121 return (ENXIO);
2122
2123 if (sc->sc_flags & ATABUSCF_OPEN)
2124 return (EBUSY);
2125
2126 if ((error = ata_addref(sc->sc_chan)) != 0)
2127 return (error);
2128
2129 sc->sc_flags |= ATABUSCF_OPEN;
2130
2131 return (0);
2132 }
2133
2134
2135 int
2136 atabusclose(dev_t dev, int flag, int fmt, struct lwp *l)
2137 {
2138 struct atabus_softc *sc =
2139 device_lookup_private(&atabus_cd, minor(dev));
2140
2141 ata_delref(sc->sc_chan);
2142
2143 sc->sc_flags &= ~ATABUSCF_OPEN;
2144
2145 return (0);
2146 }
2147
2148 int
2149 atabusioctl(dev_t dev, u_long cmd, void *addr, int flag, struct lwp *l)
2150 {
2151 struct atabus_softc *sc =
2152 device_lookup_private(&atabus_cd, minor(dev));
2153 struct ata_channel *chp = sc->sc_chan;
2154 int min_drive, max_drive, drive;
2155 int error;
2156
2157 /*
2158 * Enforce write permission for ioctls that change the
2159 * state of the bus. Host adapter specific ioctls must
2160 * be checked by the adapter driver.
2161 */
2162 switch (cmd) {
2163 case ATABUSIOSCAN:
2164 case ATABUSIODETACH:
2165 case ATABUSIORESET:
2166 if ((flag & FWRITE) == 0)
2167 return (EBADF);
2168 }
2169
2170 switch (cmd) {
2171 case ATABUSIORESET:
2172 ata_channel_lock(chp);
2173 ata_thread_run(sc->sc_chan, AT_WAIT | AT_POLL,
2174 ATACH_TH_RESET, ATACH_NODRIVE);
2175 ata_channel_unlock(chp);
2176 return 0;
2177 case ATABUSIOSCAN:
2178 {
2179 #if 0
2180 struct atabusioscan_args *a=
2181 (struct atabusioscan_args *)addr;
2182 #endif
2183 if ((chp->ch_drive[0].drive_type == ATA_DRIVET_OLD) ||
2184 (chp->ch_drive[1].drive_type == ATA_DRIVET_OLD))
2185 return (EOPNOTSUPP);
2186 return (EOPNOTSUPP);
2187 }
2188 case ATABUSIODETACH:
2189 {
2190 struct atabusiodetach_args *a=
2191 (struct atabusiodetach_args *)addr;
2192 if ((chp->ch_drive[0].drive_type == ATA_DRIVET_OLD) ||
2193 (chp->ch_drive[1].drive_type == ATA_DRIVET_OLD))
2194 return (EOPNOTSUPP);
2195 switch (a->at_dev) {
2196 case -1:
2197 min_drive = 0;
2198 max_drive = 1;
2199 break;
2200 case 0:
2201 case 1:
2202 min_drive = max_drive = a->at_dev;
2203 break;
2204 default:
2205 return (EINVAL);
2206 }
2207 for (drive = min_drive; drive <= max_drive; drive++) {
2208 if (chp->ch_drive[drive].drv_softc != NULL) {
2209 error = config_detach(
2210 chp->ch_drive[drive].drv_softc, 0);
2211 if (error)
2212 return (error);
2213 KASSERT(chp->ch_drive[drive].drv_softc == NULL);
2214 }
2215 }
2216 return 0;
2217 }
2218 default:
2219 return ENOTTY;
2220 }
2221 }
2222
2223 static bool
2224 atabus_suspend(device_t dv, const pmf_qual_t *qual)
2225 {
2226 struct atabus_softc *sc = device_private(dv);
2227 struct ata_channel *chp = sc->sc_chan;
2228
2229 ata_channel_idle(chp);
2230
2231 return true;
2232 }
2233
2234 static bool
2235 atabus_resume(device_t dv, const pmf_qual_t *qual)
2236 {
2237 struct atabus_softc *sc = device_private(dv);
2238 struct ata_channel *chp = sc->sc_chan;
2239
2240 /*
2241 * XXX joerg: with wdc, the first channel unfreezes the controler.
2242 * Move this the reset and queue idling into wdc.
2243 */
2244 ata_channel_lock(chp);
2245 if (chp->ch_queue->queue_freeze == 0) {
2246 ata_channel_unlock(chp);
2247 goto out;
2248 }
2249
2250 /* unfreeze the queue and reset drives */
2251 ata_channel_thaw_locked(chp);
2252
2253 /* reset channel only if there are drives attached */
2254 if (chp->ch_ndrives > 0)
2255 ata_thread_run(chp, AT_WAIT, ATACH_TH_RESET, ATACH_NODRIVE);
2256
2257 ata_channel_unlock(chp);
2258
2259 out:
2260 return true;
2261 }
2262
2263 static int
2264 atabus_rescan(device_t self, const char *ifattr, const int *locators)
2265 {
2266 struct atabus_softc *sc = device_private(self);
2267 struct ata_channel *chp = sc->sc_chan;
2268 struct atabus_initq *initq;
2269 int i;
2270
2271 /*
2272 * we can rescan a port multiplier atabus, even if some devices are
2273 * still attached
2274 */
2275 if (chp->ch_satapmp_nports == 0) {
2276 if (chp->atapibus != NULL) {
2277 return EBUSY;
2278 }
2279
2280 KASSERT(chp->ch_ndrives == 0 || chp->ch_drive != NULL);
2281 for (i = 0; i < chp->ch_ndrives; i++) {
2282 if (chp->ch_drive[i].drv_softc != NULL) {
2283 return EBUSY;
2284 }
2285 }
2286 }
2287
2288 initq = kmem_zalloc(sizeof(*initq), KM_SLEEP);
2289 initq->atabus_sc = sc;
2290 mutex_enter(&atabus_qlock);
2291 TAILQ_INSERT_TAIL(&atabus_initq_head, initq, atabus_initq);
2292 mutex_exit(&atabus_qlock);
2293 config_pending_incr(sc->sc_dev);
2294
2295 ata_channel_lock(chp);
2296 chp->ch_flags |= ATACH_TH_RESCAN;
2297 cv_signal(&chp->ch_thr_idle);
2298 ata_channel_unlock(chp);
2299
2300 return 0;
2301 }
2302
2303 void
2304 ata_delay(struct ata_channel *chp, int ms, const char *msg, int flags)
2305 {
2306 KASSERT(mutex_owned(&chp->ch_lock));
2307
2308 if ((flags & (AT_WAIT | AT_POLL)) == AT_POLL) {
2309 /*
2310 * can't use kpause(), we may be in interrupt context
2311 * or taking a crash dump
2312 */
2313 delay(ms * 1000);
2314 } else {
2315 int pause = mstohz(ms);
2316
2317 kpause(msg, false, pause > 0 ? pause : 1, &chp->ch_lock);
2318 }
2319 }
2320
2321 void
2322 atacmd_toncq(struct ata_xfer *xfer, uint8_t *cmd, uint16_t *count,
2323 uint16_t *features, uint8_t *device)
2324 {
2325 if ((xfer->c_flags & C_NCQ) == 0) {
2326 /* FUA handling for non-NCQ drives */
2327 if (xfer->c_bio.flags & ATA_FUA
2328 && *cmd == WDCC_WRITEDMA_EXT)
2329 *cmd = WDCC_WRITEDMA_FUA_EXT;
2330
2331 return;
2332 }
2333
2334 *cmd = (xfer->c_bio.flags & ATA_READ) ?
2335 WDCC_READ_FPDMA_QUEUED : WDCC_WRITE_FPDMA_QUEUED;
2336
2337 /* for FPDMA the block count is in features */
2338 *features = *count;
2339
2340 /* NCQ tag */
2341 *count = (xfer->c_slot << 3);
2342
2343 if (xfer->c_bio.flags & ATA_PRIO_HIGH)
2344 *count |= WDSC_PRIO_HIGH;
2345
2346 /* other device flags */
2347 if (xfer->c_bio.flags & ATA_FUA)
2348 *device |= WDSD_FUA;
2349 }
2350
2351 void
2352 ata_wait_cmd(struct ata_channel *chp, struct ata_xfer *xfer)
2353 {
2354 struct ata_queue *chq = chp->ch_queue;
2355 struct ata_command *ata_c = &xfer->c_ata_c;
2356
2357 ata_channel_lock(chp);
2358
2359 while ((ata_c->flags & AT_DONE) == 0)
2360 cv_wait(&chq->c_cmd_finish, &chp->ch_lock);
2361
2362 ata_channel_unlock(chp);
2363
2364 KASSERT((ata_c->flags & AT_DONE) != 0);
2365 }
2366