wd.c revision 1.468 1 /* $NetBSD: wd.c,v 1.468 2023/01/24 08:34:18 mlelstv Exp $ */
2
3 /*
4 * Copyright (c) 1998, 2001 Manuel Bouyer. All rights reserved.
5 *
6 * Redistribution and use in source and binary forms, with or without
7 * modification, are permitted provided that the following conditions
8 * are met:
9 * 1. Redistributions of source code must retain the above copyright
10 * notice, this list of conditions and the following disclaimer.
11 * 2. Redistributions in binary form must reproduce the above copyright
12 * notice, this list of conditions and the following disclaimer in the
13 * documentation and/or other materials provided with the distribution.
14 *
15 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
16 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
17 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
18 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
19 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
20 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
21 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
22 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
23 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
24 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
25 */
26
27 /*-
28 * Copyright (c) 1998, 2003, 2004 The NetBSD Foundation, Inc.
29 * All rights reserved.
30 *
31 * This code is derived from software contributed to The NetBSD Foundation
32 * by Charles M. Hannum and by Onno van der Linden.
33 *
34 * Redistribution and use in source and binary forms, with or without
35 * modification, are permitted provided that the following conditions
36 * are met:
37 * 1. Redistributions of source code must retain the above copyright
38 * notice, this list of conditions and the following disclaimer.
39 * 2. Redistributions in binary form must reproduce the above copyright
40 * notice, this list of conditions and the following disclaimer in the
41 * documentation and/or other materials provided with the distribution.
42 *
43 * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
44 * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
45 * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
46 * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
47 * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
48 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
49 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
50 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
51 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
52 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
53 * POSSIBILITY OF SUCH DAMAGE.
54 */
55
56 #include <sys/cdefs.h>
57 __KERNEL_RCSID(0, "$NetBSD: wd.c,v 1.468 2023/01/24 08:34:18 mlelstv Exp $");
58
59 #include "opt_ata.h"
60 #include "opt_wd.h"
61
62 #include <sys/param.h>
63 #include <sys/systm.h>
64 #include <sys/kernel.h>
65 #include <sys/conf.h>
66 #include <sys/file.h>
67 #include <sys/stat.h>
68 #include <sys/ioctl.h>
69 #include <sys/buf.h>
70 #include <sys/bufq.h>
71 #include <sys/uio.h>
72 #include <sys/device.h>
73 #include <sys/disklabel.h>
74 #include <sys/disk.h>
75 #include <sys/syslog.h>
76 #include <sys/proc.h>
77 #include <sys/reboot.h>
78 #include <sys/vnode.h>
79 #include <sys/rndsource.h>
80
81 #include <sys/intr.h>
82 #include <sys/bus.h>
83
84 #include <dev/ata/atareg.h>
85 #include <dev/ata/atavar.h>
86 #include <dev/ata/wdvar.h>
87 #include <dev/ic/wdcreg.h>
88 #include <sys/ataio.h>
89 #include "locators.h"
90
91 #include <prop/proplib.h>
92
93 #define WDIORETRIES_SINGLE 4 /* number of retries for single-sector */
94 #define WDIORETRIES 5 /* number of retries before giving up */
95 #define RECOVERYTIME hz/2 /* time to wait before retrying a cmd */
96
97 #define WDUNIT(dev) DISKUNIT(dev)
98 #define WDPART(dev) DISKPART(dev)
99 #define WDMINOR(unit, part) DISKMINOR(unit, part)
100 #define MAKEWDDEV(maj, unit, part) MAKEDISKDEV(maj, unit, part)
101
102 #define WDLABELDEV(dev) (MAKEWDDEV(major(dev), WDUNIT(dev), RAW_PART))
103
104 #define DEBUG_FUNCS 0x08
105 #define DEBUG_PROBE 0x10
106 #define DEBUG_DETACH 0x20
107 #define DEBUG_XFERS 0x40
108 #ifdef ATADEBUG
109 #ifndef ATADEBUG_WD_MASK
110 #define ATADEBUG_WD_MASK 0x0
111 #endif
112 int wdcdebug_wd_mask = ATADEBUG_WD_MASK;
113 #define ATADEBUG_PRINT(args, level) \
114 if (wdcdebug_wd_mask & (level)) \
115 printf args
116 #else
117 #define ATADEBUG_PRINT(args, level)
118 #endif
119
120 static int wdprobe(device_t, cfdata_t, void *);
121 static void wdattach(device_t, device_t, void *);
122 static int wddetach(device_t, int);
123 static void wdperror(const struct wd_softc *, struct ata_xfer *);
124
125 static void wdminphys(struct buf *);
126
127 static int wd_firstopen(device_t, dev_t, int, int);
128 static int wd_lastclose(device_t);
129 static bool wd_suspend(device_t, const pmf_qual_t *);
130 static int wd_standby(struct wd_softc *, int);
131
132 CFATTACH_DECL3_NEW(wd, sizeof(struct wd_softc),
133 wdprobe, wdattach, wddetach, NULL, NULL, NULL, DVF_DETACH_SHUTDOWN);
134
135 extern struct cfdriver wd_cd;
136
137 static dev_type_open(wdopen);
138 static dev_type_close(wdclose);
139 static dev_type_read(wdread);
140 static dev_type_write(wdwrite);
141 static dev_type_ioctl(wdioctl);
142 static dev_type_strategy(wdstrategy);
143 static dev_type_dump(wddump);
144 static dev_type_size(wdsize);
145 static dev_type_discard(wddiscard);
146
147 const struct bdevsw wd_bdevsw = {
148 .d_open = wdopen,
149 .d_close = wdclose,
150 .d_strategy = wdstrategy,
151 .d_ioctl = wdioctl,
152 .d_dump = wddump,
153 .d_psize = wdsize,
154 .d_discard = wddiscard,
155 .d_cfdriver = &wd_cd,
156 .d_devtounit = disklabel_dev_unit,
157 .d_flag = D_DISK
158 };
159
160 const struct cdevsw wd_cdevsw = {
161 .d_open = wdopen,
162 .d_close = wdclose,
163 .d_read = wdread,
164 .d_write = wdwrite,
165 .d_ioctl = wdioctl,
166 .d_stop = nostop,
167 .d_tty = notty,
168 .d_poll = nopoll,
169 .d_mmap = nommap,
170 .d_kqfilter = nokqfilter,
171 .d_discard = wddiscard,
172 .d_cfdriver = &wd_cd,
173 .d_devtounit = disklabel_dev_unit,
174 .d_flag = D_DISK
175 };
176
177 /* #define WD_DUMP_NOT_TRUSTED if you just want to watch */
178 static int wddoingadump = 0;
179 static int wddumprecalibrated = 0;
180
181 /*
182 * Glue necessary to hook WDCIOCCOMMAND into physio
183 */
184
185 struct wd_ioctl {
186 LIST_ENTRY(wd_ioctl) wi_list;
187 struct buf wi_bp;
188 struct uio wi_uio;
189 struct iovec wi_iov;
190 atareq_t wi_atareq;
191 struct wd_softc *wi_softc;
192 };
193
194 static struct wd_ioctl *wi_find(struct buf *);
195 static void wi_free(struct wd_ioctl *);
196 static struct wd_ioctl *wi_get(struct wd_softc *);
197 static void wdioctlstrategy(struct buf *);
198
199 static void wdrestart(void *);
200 static void wdstart1(struct wd_softc *, struct buf *, struct ata_xfer *);
201 static int wd_diskstart(device_t, struct buf *);
202 static int wd_dumpblocks(device_t, void *, daddr_t, int);
203 static void wd_iosize(device_t, int *);
204 static int wd_discard(device_t, off_t, off_t);
205 static void wdbioretry(void *);
206 static void wdbiorequeue(void *);
207 static void wddone(device_t, struct ata_xfer *);
208 static int wd_get_params(struct wd_softc *, struct ataparams *);
209 static void wd_set_geometry(struct wd_softc *);
210 static int wd_flushcache(struct wd_softc *, int);
211 static int wd_trim(struct wd_softc *, daddr_t, long);
212 static bool wd_shutdown(device_t, int);
213
214 static int wd_getcache(struct wd_softc *, int *);
215 static int wd_setcache(struct wd_softc *, int);
216
217 static void wd_sysctl_attach(struct wd_softc *);
218 static void wd_sysctl_detach(struct wd_softc *);
219
220 static const struct dkdriver wddkdriver = {
221 .d_open = wdopen,
222 .d_close = wdclose,
223 .d_strategy = wdstrategy,
224 .d_minphys = wdminphys,
225 .d_diskstart = wd_diskstart,
226 .d_dumpblocks = wd_dumpblocks,
227 .d_iosize = wd_iosize,
228 .d_firstopen = wd_firstopen,
229 .d_lastclose = wd_lastclose,
230 .d_discard = wd_discard
231 };
232
233 #ifdef HAS_BAD144_HANDLING
234 static void bad144intern(struct wd_softc *);
235 #endif
236
237 #define WD_QUIRK_SPLIT_MOD15_WRITE 0x0001 /* must split certain writes */
238
239 #define WD_QUIRK_FMT "\20\1SPLIT_MOD15_WRITE"
240
241 /*
242 * Quirk table for IDE drives. Put more-specific matches first, since
243 * a simple globing routine is used for matching.
244 */
245 static const struct wd_quirk {
246 const char *wdq_match; /* inquiry pattern to match */
247 int wdq_quirks; /* drive quirks */
248 } wd_quirk_table[] = {
249 /*
250 * Some Seagate S-ATA drives have a PHY which can get confused
251 * with the way data is packetized by some S-ATA controllers.
252 *
253 * The work-around is to split in two any write transfer whose
254 * sector count % 15 == 1 (assuming 512 byte sectors).
255 *
256 * XXX This is an incomplete list. There are at least a couple
257 * XXX more model numbers. If you have trouble with such transfers
258 * XXX (8K is the most common) on Seagate S-ATA drives, please
259 * XXX notify thorpej (at) NetBSD.org.
260 *
261 * The ST360015AS has not yet been confirmed to have this
262 * issue, however, it is the only other drive in the
263 * Seagate Barracuda Serial ATA V family.
264 *
265 */
266 { "ST3120023AS", WD_QUIRK_SPLIT_MOD15_WRITE },
267 { "ST380023AS", WD_QUIRK_SPLIT_MOD15_WRITE },
268 { "ST360015AS", WD_QUIRK_SPLIT_MOD15_WRITE },
269 { NULL,
270 0 }
271 };
272
273 static const struct wd_quirk *
274 wd_lookup_quirks(const char *name)
275 {
276 const struct wd_quirk *wdq;
277 const char *estr;
278
279 for (wdq = wd_quirk_table; wdq->wdq_match != NULL; wdq++) {
280 /*
281 * We only want exact matches (which include matches
282 * against globbing characters).
283 */
284 if (pmatch(name, wdq->wdq_match, &estr) == 2)
285 return (wdq);
286 }
287 return (NULL);
288 }
289
290 static int
291 wdprobe(device_t parent, cfdata_t match, void *aux)
292 {
293 struct ata_device *adev = aux;
294
295 if (adev == NULL)
296 return 0;
297 if (adev->adev_bustype->bustype_type != SCSIPI_BUSTYPE_ATA)
298 return 0;
299
300 if (match->cf_loc[ATA_HLCF_DRIVE] != ATA_HLCF_DRIVE_DEFAULT &&
301 match->cf_loc[ATA_HLCF_DRIVE] != adev->adev_drv_data->drive)
302 return 0;
303 return 1;
304 }
305
306 static void
307 wdattach(device_t parent, device_t self, void *aux)
308 {
309 struct wd_softc *wd = device_private(self);
310 struct dk_softc *dksc = &wd->sc_dksc;
311 struct ata_device *adev= aux;
312 int i, blank;
313 char tbuf[41],pbuf[9], c, *p, *q;
314 const struct wd_quirk *wdq;
315 int dtype = DKTYPE_UNKNOWN;
316
317 dksc->sc_dev = self;
318
319 ATADEBUG_PRINT(("wdattach\n"), DEBUG_FUNCS | DEBUG_PROBE);
320 mutex_init(&wd->sc_lock, MUTEX_DEFAULT, IPL_BIO);
321 #ifdef WD_SOFTBADSECT
322 SLIST_INIT(&wd->sc_bslist);
323 cv_init(&wd->sc_bslist_cv, "wdbadsect");
324 #endif
325 wd->atabus = adev->adev_bustype;
326 wd->inflight = 0;
327 wd->drvp = adev->adev_drv_data;
328
329 wd->drvp->drv_openings = 1;
330 wd->drvp->drv_done = wddone;
331 wd->drvp->drv_softc = dksc->sc_dev; /* done in atabusconfig_thread()
332 but too late */
333
334 SLIST_INIT(&wd->sc_retry_list);
335 SLIST_INIT(&wd->sc_requeue_list);
336 callout_init(&wd->sc_retry_callout, 0); /* XXX MPSAFE */
337 callout_init(&wd->sc_requeue_callout, 0); /* XXX MPSAFE */
338 callout_init(&wd->sc_restart_diskqueue, 0); /* XXX MPSAFE */
339
340 aprint_naive("\n");
341 aprint_normal("\n");
342
343 /* read our drive info */
344 if (wd_get_params(wd, &wd->sc_params) != 0) {
345 aprint_error_dev(self, "IDENTIFY failed\n");
346 goto out;
347 }
348
349 for (blank = 0, p = wd->sc_params.atap_model, q = tbuf, i = 0;
350 i < sizeof(wd->sc_params.atap_model); i++) {
351 c = *p++;
352 if (c == '\0')
353 break;
354 if (c != ' ') {
355 if (blank) {
356 *q++ = ' ';
357 blank = 0;
358 }
359 *q++ = c;
360 } else
361 blank = 1;
362 }
363 *q++ = '\0';
364
365 wd->sc_typename = kmem_asprintf("%s", tbuf);
366 aprint_normal_dev(self, "<%s>\n", wd->sc_typename);
367
368 wdq = wd_lookup_quirks(tbuf);
369 if (wdq != NULL)
370 wd->sc_quirks = wdq->wdq_quirks;
371
372 if (wd->sc_quirks != 0) {
373 char sbuf[sizeof(WD_QUIRK_FMT) + 64];
374 snprintb(sbuf, sizeof(sbuf), WD_QUIRK_FMT, wd->sc_quirks);
375 aprint_normal_dev(self, "quirks %s\n", sbuf);
376
377 if (wd->sc_quirks & WD_QUIRK_SPLIT_MOD15_WRITE) {
378 aprint_error_dev(self, "drive corrupts write transfers with certain controllers, consider replacing\n");
379 }
380 }
381
382 if ((wd->sc_params.atap_multi & 0xff) > 1) {
383 wd->drvp->multi = wd->sc_params.atap_multi & 0xff;
384 } else {
385 wd->drvp->multi = 1;
386 }
387
388 aprint_verbose_dev(self, "drive supports %d-sector PIO transfers,",
389 wd->drvp->multi);
390
391 /* 48-bit LBA addressing */
392 if ((wd->sc_params.atap_cmd2_en & ATA_CMD2_LBA48) != 0)
393 wd->sc_flags |= WDF_LBA48;
394
395 /* Prior to ATA-4, LBA was optional. */
396 if ((wd->sc_params.atap_capabilities1 & WDC_CAP_LBA) != 0)
397 wd->sc_flags |= WDF_LBA;
398 #if 0
399 /* ATA-4 requires LBA. */
400 if (wd->sc_params.atap_ataversion != 0xffff &&
401 wd->sc_params.atap_ataversion >= WDC_VER_ATA4)
402 wd->sc_flags |= WDF_LBA;
403 #endif
404
405 if ((wd->sc_flags & WDF_LBA48) != 0) {
406 aprint_verbose(" LBA48 addressing\n");
407 wd->sc_capacity =
408 ((uint64_t) wd->sc_params.atap_max_lba[3] << 48) |
409 ((uint64_t) wd->sc_params.atap_max_lba[2] << 32) |
410 ((uint64_t) wd->sc_params.atap_max_lba[1] << 16) |
411 ((uint64_t) wd->sc_params.atap_max_lba[0] << 0);
412 wd->sc_capacity28 =
413 (wd->sc_params.atap_capacity[1] << 16) |
414 wd->sc_params.atap_capacity[0];
415 /*
416 * Force LBA48 addressing for invalid numbers.
417 */
418 if (wd->sc_capacity28 > 0xfffffff)
419 wd->sc_capacity28 = 0xfffffff;
420 } else if ((wd->sc_flags & WDF_LBA) != 0) {
421 aprint_verbose(" LBA addressing\n");
422 wd->sc_capacity28 =
423 (wd->sc_params.atap_capacity[1] << 16) |
424 wd->sc_params.atap_capacity[0];
425 /*
426 * Limit capacity to LBA28 numbers to avoid overflow.
427 */
428 if (wd->sc_capacity28 > 0xfffffff)
429 wd->sc_capacity28 = 0xfffffff;
430 wd->sc_capacity = wd->sc_capacity28;
431 } else {
432 aprint_verbose(" chs addressing\n");
433 wd->sc_capacity =
434 wd->sc_params.atap_cylinders *
435 wd->sc_params.atap_heads *
436 wd->sc_params.atap_sectors;
437 /*
438 * LBA28 size is ignored for CHS addressing. Use a reasonable
439 * value for debugging. The CHS values may be artifical and
440 * are mostly ignored.
441 */
442 if (wd->sc_capacity < 0xfffffff)
443 wd->sc_capacity28 = wd->sc_capacity;
444 else
445 wd->sc_capacity28 = 0xfffffff;
446 }
447 if ((wd->sc_params.atap_secsz & ATA_SECSZ_VALID_MASK) == ATA_SECSZ_VALID
448 && ((wd->sc_params.atap_secsz & ATA_SECSZ_LLS) != 0)) {
449 wd->sc_blksize = 2ULL *
450 ((uint32_t)((wd->sc_params.atap_lls_secsz[1] << 16) |
451 wd->sc_params.atap_lls_secsz[0]));
452 } else {
453 wd->sc_blksize = 512;
454 }
455 wd->sc_sectoralign.dsa_firstaligned = 0;
456 wd->sc_sectoralign.dsa_alignment = 1;
457 if ((wd->sc_params.atap_secsz & ATA_SECSZ_VALID_MASK) == ATA_SECSZ_VALID
458 && ((wd->sc_params.atap_secsz & ATA_SECSZ_LPS) != 0)) {
459 wd->sc_sectoralign.dsa_alignment = 1 <<
460 (wd->sc_params.atap_secsz & ATA_SECSZ_LPS_SZMSK);
461 if ((wd->sc_params.atap_logical_align & ATA_LA_VALID_MASK) ==
462 ATA_LA_VALID) {
463 wd->sc_sectoralign.dsa_firstaligned =
464 (wd->sc_sectoralign.dsa_alignment -
465 (wd->sc_params.atap_logical_align &
466 ATA_LA_MASK));
467 }
468 }
469 wd->sc_capacity512 = (wd->sc_capacity * wd->sc_blksize) / DEV_BSIZE;
470 format_bytes(pbuf, sizeof(pbuf), wd->sc_capacity * wd->sc_blksize);
471 aprint_normal_dev(self, "%s, %d cyl, %d head, %d sec, "
472 "%d bytes/sect x %llu sectors",
473 pbuf,
474 (wd->sc_flags & WDF_LBA) ? (int)(wd->sc_capacity /
475 (wd->sc_params.atap_heads * wd->sc_params.atap_sectors)) :
476 wd->sc_params.atap_cylinders,
477 wd->sc_params.atap_heads, wd->sc_params.atap_sectors,
478 wd->sc_blksize, (unsigned long long)wd->sc_capacity);
479 if (wd->sc_sectoralign.dsa_alignment != 1) {
480 aprint_normal(" (%d bytes/physsect",
481 wd->sc_sectoralign.dsa_alignment * wd->sc_blksize);
482 if (wd->sc_sectoralign.dsa_firstaligned != 0) {
483 aprint_normal("; first aligned sector: %jd",
484 (intmax_t)wd->sc_sectoralign.dsa_firstaligned);
485 }
486 aprint_normal(")");
487 }
488 aprint_normal("\n");
489
490 ATADEBUG_PRINT(("%s: atap_dmatiming_mimi=%d, atap_dmatiming_recom=%d\n",
491 device_xname(self), wd->sc_params.atap_dmatiming_mimi,
492 wd->sc_params.atap_dmatiming_recom), DEBUG_PROBE);
493
494 if (wd->sc_blksize <= 0 || !powerof2(wd->sc_blksize) ||
495 wd->sc_blksize < DEV_BSIZE || wd->sc_blksize > MAXPHYS) {
496 aprint_normal_dev(self, "WARNING: block size %u "
497 "might not actually work\n", wd->sc_blksize);
498 }
499
500 if (strcmp(wd->sc_params.atap_model, "ST506") == 0)
501 dtype = DKTYPE_ST506;
502 else
503 dtype = DKTYPE_ESDI;
504
505 out:
506 /*
507 * Initialize and attach the disk structure.
508 */
509 dk_init(dksc, self, dtype);
510 disk_init(&dksc->sc_dkdev, dksc->sc_xname, &wddkdriver);
511
512 /* Attach dk and disk subsystems */
513 dk_attach(dksc);
514 disk_attach(&dksc->sc_dkdev);
515 wd_set_geometry(wd);
516
517 bufq_alloc(&dksc->sc_bufq, BUFQ_DISK_DEFAULT_STRAT, BUFQ_SORT_RAWBLOCK);
518
519 /* reference to label structure, used by ata code */
520 wd->drvp->lp = dksc->sc_dkdev.dk_label;
521
522 /* Discover wedges on this disk. */
523 dkwedge_discover(&dksc->sc_dkdev);
524
525 if (!pmf_device_register1(self, wd_suspend, NULL, wd_shutdown))
526 aprint_error_dev(self, "couldn't establish power handler\n");
527
528 wd_sysctl_attach(wd);
529 }
530
531 static bool
532 wd_suspend(device_t dv, const pmf_qual_t *qual)
533 {
534 struct wd_softc *sc = device_private(dv);
535
536 /* the adapter needs to be enabled */
537 if (sc->atabus->ata_addref(sc->drvp))
538 return true; /* no need to complain */
539
540 wd_flushcache(sc, AT_WAIT);
541 wd_standby(sc, AT_WAIT);
542
543 sc->atabus->ata_delref(sc->drvp);
544 return true;
545 }
546
547 static int
548 wddetach(device_t self, int flags)
549 {
550 struct wd_softc *wd = device_private(self);
551 struct dk_softc *dksc = &wd->sc_dksc;
552 int bmaj, cmaj, i, mn, rc;
553
554 if ((rc = disk_begindetach(&dksc->sc_dkdev, wd_lastclose, self, flags)) != 0)
555 return rc;
556
557 /* locate the major number */
558 bmaj = bdevsw_lookup_major(&wd_bdevsw);
559 cmaj = cdevsw_lookup_major(&wd_cdevsw);
560
561 /* Nuke the vnodes for any open instances. */
562 for (i = 0; i < MAXPARTITIONS; i++) {
563 mn = WDMINOR(device_unit(self), i);
564 vdevgone(bmaj, mn, mn, VBLK);
565 vdevgone(cmaj, mn, mn, VCHR);
566 }
567
568 dk_drain(dksc);
569
570 /* Kill off any pending commands. */
571 mutex_enter(&wd->sc_lock);
572 wd->atabus->ata_killpending(wd->drvp);
573
574 callout_halt(&wd->sc_retry_callout, &wd->sc_lock);
575 callout_destroy(&wd->sc_retry_callout);
576 callout_halt(&wd->sc_requeue_callout, &wd->sc_lock);
577 callout_destroy(&wd->sc_requeue_callout);
578 callout_halt(&wd->sc_restart_diskqueue, &wd->sc_lock);
579 callout_destroy(&wd->sc_restart_diskqueue);
580
581 mutex_exit(&wd->sc_lock);
582
583 bufq_free(dksc->sc_bufq);
584
585 /* Delete all of our wedges. */
586 dkwedge_delall(&dksc->sc_dkdev);
587
588 if (flags & DETACH_POWEROFF)
589 wd_standby(wd, AT_POLL);
590
591 /* Detach from the disk list. */
592 disk_detach(&dksc->sc_dkdev);
593 disk_destroy(&dksc->sc_dkdev);
594
595 dk_detach(dksc);
596
597 #ifdef WD_SOFTBADSECT
598 /* Clean out the bad sector list */
599 while (!SLIST_EMPTY(&wd->sc_bslist)) {
600 struct disk_badsectors *dbs = SLIST_FIRST(&wd->sc_bslist);
601 SLIST_REMOVE_HEAD(&wd->sc_bslist, dbs_next);
602 kmem_free(dbs, sizeof(*dbs));
603 }
604 wd->sc_bscount = 0;
605 #endif
606 if (wd->sc_typename != NULL) {
607 kmem_free(wd->sc_typename, strlen(wd->sc_typename) + 1);
608 wd->sc_typename = NULL;
609 }
610
611 pmf_device_deregister(self);
612
613 wd_sysctl_detach(wd);
614
615 #ifdef WD_SOFTBADSECT
616 KASSERT(SLIST_EMPTY(&wd->sc_bslist));
617 cv_destroy(&wd->sc_bslist_cv);
618 #endif
619
620 mutex_destroy(&wd->sc_lock);
621
622 wd->drvp->drive_type = ATA_DRIVET_NONE; /* no drive any more here */
623 wd->drvp->drive_flags = 0;
624
625 return (0);
626 }
627
628 /*
629 * Read/write routine for a buffer. Validates the arguments and schedules the
630 * transfer. Does not wait for the transfer to complete.
631 */
632 static void
633 wdstrategy(struct buf *bp)
634 {
635 struct wd_softc *wd =
636 device_lookup_private(&wd_cd, WDUNIT(bp->b_dev));
637 struct dk_softc *dksc = &wd->sc_dksc;
638
639 ATADEBUG_PRINT(("wdstrategy (%s)\n", dksc->sc_xname),
640 DEBUG_XFERS);
641
642 /* If device invalidated (e.g. media change, door open,
643 * device detachment), then error.
644 */
645 if ((wd->sc_flags & WDF_LOADED) == 0 ||
646 !device_is_enabled(dksc->sc_dev))
647 goto err;
648
649 #ifdef WD_SOFTBADSECT
650 /*
651 * If the transfer about to be attempted contains only a block that
652 * is known to be bad then return an error for the transfer without
653 * even attempting to start a transfer up under the premis that we
654 * will just end up doing more retries for a transfer that will end
655 * up failing again.
656 */
657 if (__predict_false(!SLIST_EMPTY(&wd->sc_bslist))) {
658 struct disklabel *lp = dksc->sc_dkdev.dk_label;
659 struct disk_badsectors *dbs;
660 daddr_t blkno, maxblk;
661
662 /* convert the block number to absolute */
663 if (lp->d_secsize >= DEV_BSIZE)
664 blkno = bp->b_blkno / (lp->d_secsize / DEV_BSIZE);
665 else
666 blkno = bp->b_blkno * (DEV_BSIZE / lp->d_secsize);
667 if (WDPART(bp->b_dev) != RAW_PART)
668 blkno += lp->d_partitions[WDPART(bp->b_dev)].p_offset;
669 maxblk = blkno + (bp->b_bcount / wd->sc_blksize) - 1;
670
671 mutex_enter(&wd->sc_lock);
672 SLIST_FOREACH(dbs, &wd->sc_bslist, dbs_next)
673 if ((dbs->dbs_min <= bp->b_rawblkno &&
674 bp->b_rawblkno <= dbs->dbs_max) ||
675 (dbs->dbs_min <= maxblk && maxblk <= dbs->dbs_max)){
676 mutex_exit(&wd->sc_lock);
677 goto err;
678 }
679 mutex_exit(&wd->sc_lock);
680 }
681 #endif
682
683 dk_strategy(dksc, bp);
684 return;
685
686 err:
687 bp->b_error = EIO;
688 bp->b_resid = bp->b_bcount;
689 biodone(bp);
690 }
691
692 static void
693 wdstart1(struct wd_softc *wd, struct buf *bp, struct ata_xfer *xfer)
694 {
695 struct dk_softc *dksc = &wd->sc_dksc;
696 const uint32_t secsize = dksc->sc_dkdev.dk_geom.dg_secsize;
697
698 KASSERT(bp == xfer->c_bio.bp || xfer->c_bio.bp == NULL);
699 KASSERT((xfer->c_flags & (C_WAITACT|C_FREE)) == 0);
700 KASSERT(mutex_owned(&wd->sc_lock));
701
702 /* Reset state, so that retries don't use stale info */
703 if (__predict_false(xfer->c_retries > 0)) {
704 xfer->c_flags = 0;
705 memset(&xfer->c_bio, 0, sizeof(xfer->c_bio));
706 }
707
708 xfer->c_bio.blkno = bp->b_rawblkno;
709 xfer->c_bio.bcount = bp->b_bcount;
710 xfer->c_bio.databuf = bp->b_data;
711 xfer->c_bio.blkdone = 0;
712 xfer->c_bio.bp = bp;
713
714 /* Adjust blkno and bcount if xfer has been already partially done */
715 if (__predict_false(xfer->c_skip > 0)) {
716 KASSERT(xfer->c_skip < xfer->c_bio.bcount);
717 KASSERT((xfer->c_skip % secsize) == 0);
718 xfer->c_bio.bcount -= xfer->c_skip;
719 xfer->c_bio.blkno += xfer->c_skip / secsize;
720 }
721
722 #ifdef WD_CHAOS_MONKEY
723 /*
724 * Override blkno to be over device capacity to trigger error,
725 * but only if it's read, to avoid trashing disk contents should
726 * the command be clipped, or otherwise misinterpreted, by the
727 * driver or controller.
728 */
729 if (BUF_ISREAD(bp) && xfer->c_retries == 0 && wd->drv_chaos_freq > 0 &&
730 (++wd->drv_chaos_cnt % wd->drv_chaos_freq) == 0) {
731 device_printf(dksc->sc_dev, "%s: chaos xfer %"PRIxPTR"\n",
732 __func__, (intptr_t)xfer & PAGE_MASK);
733 xfer->c_bio.blkno = 7777777 + wd->sc_capacity;
734 xfer->c_flags |= C_CHAOS;
735 }
736 #endif
737
738 /*
739 * If we're retrying, retry in single-sector mode. This will give us
740 * the sector number of the problem, and will eventually allow the
741 * transfer to succeed. If FUA is requested, we can't actually
742 * do this, as ATA_SINGLE is usually executed as PIO transfer by drivers
743 * which support it, and that isn't compatible with NCQ/FUA.
744 */
745 if (xfer->c_retries >= WDIORETRIES_SINGLE &&
746 (bp->b_flags & B_MEDIA_FUA) == 0)
747 xfer->c_bio.flags = ATA_SINGLE;
748 else
749 xfer->c_bio.flags = 0;
750
751 /*
752 * request LBA48 transfers when supported by the controller
753 * and needed by transfer offset or size.
754 */
755 if (wd->sc_flags & WDF_LBA48 &&
756 (((xfer->c_bio.blkno + xfer->c_bio.bcount / secsize) >
757 wd->sc_capacity28) ||
758 ((xfer->c_bio.bcount / secsize) > 128)))
759 xfer->c_bio.flags |= ATA_LBA48;
760
761 /*
762 * If NCQ was negotiated, always use it for the first several attempts.
763 * Since device cancels all outstanding requests on error, downgrade
764 * to non-NCQ on retry, so that the retried transfer would not cause
765 * cascade failure for the other transfers if it fails again.
766 * If FUA was requested, we can't downgrade, as that would violate
767 * the semantics - FUA would not be honored. In that case, continue
768 * retrying with NCQ.
769 */
770 if (WD_USE_NCQ(wd) && (xfer->c_retries < WDIORETRIES_SINGLE ||
771 (bp->b_flags & B_MEDIA_FUA) != 0)) {
772 xfer->c_bio.flags |= ATA_LBA48;
773 xfer->c_flags |= C_NCQ;
774
775 if (WD_USE_NCQ_PRIO(wd) &&
776 BIO_GETPRIO(bp) == BPRIO_TIMECRITICAL)
777 xfer->c_bio.flags |= ATA_PRIO_HIGH;
778 }
779
780 if (wd->sc_flags & WDF_LBA)
781 xfer->c_bio.flags |= ATA_LBA;
782 if (bp->b_flags & B_READ) {
783 xfer->c_bio.flags |= ATA_READ;
784 } else {
785 /* it's a write */
786 wd->sc_flags |= WDF_DIRTY;
787 }
788 if (bp->b_flags & B_MEDIA_FUA) {
789 /* If not using NCQ, the command WRITE DMA FUA EXT is LBA48 */
790 KASSERT((wd->sc_flags & WDF_LBA48) != 0);
791 if ((xfer->c_flags & C_NCQ) == 0)
792 xfer->c_bio.flags |= ATA_LBA48;
793
794 xfer->c_bio.flags |= ATA_FUA;
795 }
796
797 if (xfer->c_retries == 0)
798 wd->inflight++;
799 mutex_exit(&wd->sc_lock);
800
801 /* Queue the xfer */
802 wd->atabus->ata_bio(wd->drvp, xfer);
803
804 mutex_enter(&wd->sc_lock);
805 }
806
807 static int
808 wd_diskstart(device_t dev, struct buf *bp)
809 {
810 struct wd_softc *wd = device_private(dev);
811 #ifdef ATADEBUG
812 struct dk_softc *dksc = &wd->sc_dksc;
813 #endif
814 struct ata_xfer *xfer;
815 struct ata_channel *chp;
816 unsigned openings;
817 int ticks;
818
819 mutex_enter(&wd->sc_lock);
820
821 chp = wd->drvp->chnl_softc;
822
823 ata_channel_lock(chp);
824 openings = ata_queue_openings(chp);
825 ata_channel_unlock(chp);
826
827 openings = uimin(openings, wd->drvp->drv_openings);
828
829 if (wd->inflight >= openings) {
830 /*
831 * pretend we run out of memory when the queue is full,
832 * so that the operation is retried after a minimal
833 * delay.
834 */
835 xfer = NULL;
836 ticks = 1;
837 } else {
838 /*
839 * If there is no available memory, retry later. This
840 * happens very rarely and only under memory pressure,
841 * so wait relatively long before retry.
842 */
843 xfer = ata_get_xfer(chp, false);
844 ticks = hz/2;
845 }
846
847 if (xfer == NULL) {
848 ATADEBUG_PRINT(("wd_diskstart %s no xfer\n",
849 dksc->sc_xname), DEBUG_XFERS);
850
851 /*
852 * The disk queue is pushed automatically when an I/O
853 * operation finishes or another one is queued. We
854 * need this extra timeout because an ATA channel
855 * might be shared by more than one disk queue and
856 * all queues need to be restarted when another slot
857 * becomes available.
858 */
859 if (!callout_pending(&wd->sc_restart_diskqueue)) {
860 callout_reset(&wd->sc_restart_diskqueue, ticks,
861 wdrestart, dev);
862 }
863
864 mutex_exit(&wd->sc_lock);
865 return EAGAIN;
866 }
867
868 wdstart1(wd, bp, xfer);
869
870 mutex_exit(&wd->sc_lock);
871
872 return 0;
873 }
874
875 /*
876 * Queue a drive for I/O.
877 */
878 static void
879 wdrestart(void *x)
880 {
881 device_t self = x;
882 struct wd_softc *wd = device_private(self);
883 struct dk_softc *dksc = &wd->sc_dksc;
884
885 ATADEBUG_PRINT(("wdstart %s\n", dksc->sc_xname),
886 DEBUG_XFERS);
887
888 if (!device_is_active(dksc->sc_dev))
889 return;
890
891 dk_start(dksc, NULL);
892 }
893
894 static void
895 wddone(device_t self, struct ata_xfer *xfer)
896 {
897 struct wd_softc *wd = device_private(self);
898 struct dk_softc *dksc = &wd->sc_dksc;
899 const char *errmsg;
900 int do_perror = 0;
901 struct buf *bp;
902
903 ATADEBUG_PRINT(("wddone %s\n", dksc->sc_xname),
904 DEBUG_XFERS);
905
906 if (__predict_false(wddoingadump)) {
907 /* just drop it to the floor */
908 ata_free_xfer(wd->drvp->chnl_softc, xfer);
909 return;
910 }
911
912 bp = xfer->c_bio.bp;
913 KASSERT(bp != NULL);
914
915 bp->b_resid = xfer->c_bio.bcount;
916 switch (xfer->c_bio.error) {
917 case ERR_DMA:
918 errmsg = "DMA error";
919 goto retry;
920 case ERR_DF:
921 errmsg = "device fault";
922 goto retry;
923 case TIMEOUT:
924 errmsg = "device timeout";
925 goto retry;
926 case REQUEUE:
927 errmsg = "requeue";
928 goto retry2;
929 case ERR_RESET:
930 errmsg = "channel reset";
931 goto retry2;
932 case ERROR:
933 /* Don't care about media change bits */
934 if (xfer->c_bio.r_error != 0 &&
935 (xfer->c_bio.r_error & ~(WDCE_MC | WDCE_MCR)) == 0)
936 goto noerror;
937 errmsg = "error";
938 do_perror = 1;
939 retry: /* Just reset and retry. Can we do more ? */
940 if ((xfer->c_flags & C_RECOVERED) == 0) {
941 int wflags = (xfer->c_flags & C_POLL) ? AT_POLL : 0;
942 ata_channel_lock(wd->drvp->chnl_softc);
943 ata_thread_run(wd->drvp->chnl_softc, wflags,
944 ATACH_TH_DRIVE_RESET, wd->drvp->drive);
945 ata_channel_unlock(wd->drvp->chnl_softc);
946 }
947 retry2:
948 mutex_enter(&wd->sc_lock);
949
950 diskerr(bp, "wd", errmsg, LOG_PRINTF,
951 xfer->c_bio.blkdone, dksc->sc_dkdev.dk_label);
952 if (xfer->c_retries < WDIORETRIES)
953 printf(", xfer %"PRIxPTR", retry %d",
954 (intptr_t)xfer & PAGE_MASK,
955 xfer->c_retries);
956 printf("\n");
957 if (do_perror)
958 wdperror(wd, xfer);
959
960 if (xfer->c_retries < WDIORETRIES) {
961 xfer->c_retries++;
962
963 /* Rerun ASAP if just requeued */
964 if (xfer->c_bio.error == REQUEUE) {
965 SLIST_INSERT_HEAD(&wd->sc_requeue_list, xfer,
966 c_retrychain);
967 callout_reset(&wd->sc_requeue_callout,
968 1, wdbiorequeue, wd);
969 } else {
970 SLIST_INSERT_HEAD(&wd->sc_retry_list, xfer,
971 c_retrychain);
972 callout_reset(&wd->sc_retry_callout,
973 RECOVERYTIME, wdbioretry, wd);
974 }
975
976 mutex_exit(&wd->sc_lock);
977 return;
978 }
979
980 mutex_exit(&wd->sc_lock);
981
982 #ifdef WD_SOFTBADSECT
983 /*
984 * Not all errors indicate a failed block but those that do,
985 * put the block on the bad-block list for the device. Only
986 * do this for reads because the drive should do it for writes,
987 * itself, according to Manuel.
988 */
989 if ((bp->b_flags & B_READ) &&
990 ((wd->drvp->ata_vers >= 4 && xfer->c_bio.r_error & 64) ||
991 (wd->drvp->ata_vers < 4 && xfer->c_bio.r_error & 192))) {
992 struct disk_badsectors *dbs;
993
994 dbs = kmem_zalloc(sizeof *dbs, KM_NOSLEEP);
995 if (dbs == NULL) {
996 aprint_error_dev(dksc->sc_dev,
997 "failed to add bad block to list\n");
998 goto out;
999 }
1000
1001 dbs->dbs_min = bp->b_rawblkno;
1002 dbs->dbs_max = dbs->dbs_min +
1003 (bp->b_bcount /wd->sc_blksize) - 1;
1004 microtime(&dbs->dbs_failedat);
1005
1006 mutex_enter(&wd->sc_lock);
1007 SLIST_INSERT_HEAD(&wd->sc_bslist, dbs, dbs_next);
1008 wd->sc_bscount++;
1009 mutex_exit(&wd->sc_lock);
1010 }
1011 out:
1012 #endif
1013 bp->b_error = EIO;
1014 break;
1015 case NOERROR:
1016 #ifdef WD_CHAOS_MONKEY
1017 /*
1018 * For example Parallels AHCI emulation doesn't actually
1019 * return error for the invalid I/O, so just re-run
1020 * the request and do not panic.
1021 */
1022 if (__predict_false(xfer->c_flags & C_CHAOS)) {
1023 xfer->c_bio.error = REQUEUE;
1024 errmsg = "chaos noerror";
1025 goto retry2;
1026 }
1027 #endif
1028
1029 noerror: if ((xfer->c_bio.flags & ATA_CORR) || xfer->c_retries > 0)
1030 device_printf(dksc->sc_dev,
1031 "soft error (corrected) xfer %"PRIxPTR"\n",
1032 (intptr_t)xfer & PAGE_MASK);
1033 break;
1034 case ERR_NODEV:
1035 bp->b_error = EIO;
1036 break;
1037 }
1038 if (__predict_false(bp->b_error != 0) && bp->b_resid == 0) {
1039 /*
1040 * the disk or controller sometimes report a complete
1041 * xfer, when there has been an error. This is wrong,
1042 * assume nothing got transferred in this case
1043 */
1044 bp->b_resid = bp->b_bcount;
1045 }
1046
1047 ata_free_xfer(wd->drvp->chnl_softc, xfer);
1048
1049 mutex_enter(&wd->sc_lock);
1050 wd->inflight--;
1051 mutex_exit(&wd->sc_lock);
1052 dk_done(dksc, bp);
1053 dk_start(dksc, NULL);
1054 }
1055
1056 static void
1057 wdbioretry(void *v)
1058 {
1059 struct wd_softc *wd = v;
1060 struct ata_xfer *xfer;
1061
1062 ATADEBUG_PRINT(("%s %s\n", __func__, wd->sc_dksc.sc_xname),
1063 DEBUG_XFERS);
1064
1065 mutex_enter(&wd->sc_lock);
1066 while ((xfer = SLIST_FIRST(&wd->sc_retry_list))) {
1067 SLIST_REMOVE_HEAD(&wd->sc_retry_list, c_retrychain);
1068 wdstart1(wd, xfer->c_bio.bp, xfer);
1069 }
1070 mutex_exit(&wd->sc_lock);
1071 }
1072
1073 static void
1074 wdbiorequeue(void *v)
1075 {
1076 struct wd_softc *wd = v;
1077 struct ata_xfer *xfer;
1078
1079 ATADEBUG_PRINT(("%s %s\n", __func__, wd->sc_dksc.sc_xname),
1080 DEBUG_XFERS);
1081
1082 mutex_enter(&wd->sc_lock);
1083 while ((xfer = SLIST_FIRST(&wd->sc_requeue_list))) {
1084 SLIST_REMOVE_HEAD(&wd->sc_requeue_list, c_retrychain);
1085 wdstart1(wd, xfer->c_bio.bp, xfer);
1086 }
1087 mutex_exit(&wd->sc_lock);
1088 }
1089
1090 static void
1091 wdminphys(struct buf *bp)
1092 {
1093 const struct wd_softc * const wd =
1094 device_lookup_private(&wd_cd, WDUNIT(bp->b_dev));
1095 int maxsectors;
1096
1097 /*
1098 * The limit is actually 65536 for LBA48 and 256 for non-LBA48,
1099 * but that requires to set the count for the ATA command
1100 * to 0, which is somewhat error prone, so better stay safe.
1101 */
1102 if (wd->sc_flags & WDF_LBA48)
1103 maxsectors = 65535;
1104 else
1105 maxsectors = 128;
1106
1107 if (bp->b_bcount > (wd->sc_blksize * maxsectors))
1108 bp->b_bcount = (wd->sc_blksize * maxsectors);
1109
1110 minphys(bp);
1111 }
1112
1113 static void
1114 wd_iosize(device_t dev, int *count)
1115 {
1116 struct buf B;
1117 int bmaj;
1118
1119 bmaj = bdevsw_lookup_major(&wd_bdevsw);
1120 B.b_dev = MAKEWDDEV(bmaj,device_unit(dev),RAW_PART);
1121 B.b_bcount = *count;
1122
1123 wdminphys(&B);
1124
1125 *count = B.b_bcount;
1126 }
1127
1128 static int
1129 wdread(dev_t dev, struct uio *uio, int flags)
1130 {
1131
1132 ATADEBUG_PRINT(("wdread\n"), DEBUG_XFERS);
1133 return (physio(wdstrategy, NULL, dev, B_READ, wdminphys, uio));
1134 }
1135
1136 static int
1137 wdwrite(dev_t dev, struct uio *uio, int flags)
1138 {
1139
1140 ATADEBUG_PRINT(("wdwrite\n"), DEBUG_XFERS);
1141 return (physio(wdstrategy, NULL, dev, B_WRITE, wdminphys, uio));
1142 }
1143
1144 static int
1145 wdopen(dev_t dev, int flag, int fmt, struct lwp *l)
1146 {
1147 struct wd_softc *wd;
1148 struct dk_softc *dksc;
1149 int unit, part, error;
1150
1151 ATADEBUG_PRINT(("wdopen\n"), DEBUG_FUNCS);
1152 unit = WDUNIT(dev);
1153 wd = device_lookup_private(&wd_cd, unit);
1154 if (wd == NULL)
1155 return (ENXIO);
1156 dksc = &wd->sc_dksc;
1157
1158 if (! device_is_active(dksc->sc_dev))
1159 return (ENODEV);
1160
1161 part = WDPART(dev);
1162
1163 if (wd->sc_capacity == 0)
1164 return (ENODEV);
1165
1166 /*
1167 * If any partition is open, but the disk has been invalidated,
1168 * disallow further opens.
1169 */
1170 if ((wd->sc_flags & (WDF_OPEN | WDF_LOADED)) == WDF_OPEN) {
1171 if (part != RAW_PART || fmt != S_IFCHR)
1172 return EIO;
1173 }
1174
1175 error = dk_open(dksc, dev, flag, fmt, l);
1176
1177 return error;
1178 }
1179
1180 /*
1181 * Serialized by caller
1182 */
1183 static int
1184 wd_firstopen(device_t self, dev_t dev, int flag, int fmt)
1185 {
1186 struct wd_softc *wd = device_private(self);
1187 struct dk_softc *dksc = &wd->sc_dksc;
1188 int error;
1189
1190 error = wd->atabus->ata_addref(wd->drvp);
1191 if (error)
1192 return error;
1193
1194 if ((wd->sc_flags & WDF_LOADED) == 0) {
1195 int param_error;
1196
1197 /* Load the physical device parameters. */
1198 param_error = wd_get_params(wd, &wd->sc_params);
1199 if (param_error != 0) {
1200 aprint_error_dev(dksc->sc_dev, "IDENTIFY failed\n");
1201 error = EIO;
1202 goto bad;
1203 }
1204 wd_set_geometry(wd);
1205 wd->sc_flags |= WDF_LOADED;
1206 }
1207
1208 wd->sc_flags |= WDF_OPEN;
1209 return 0;
1210
1211 bad:
1212 wd->atabus->ata_delref(wd->drvp);
1213 return error;
1214 }
1215
1216 /*
1217 * Caller must hold wd->sc_dk.dk_openlock.
1218 */
1219 static int
1220 wd_lastclose(device_t self)
1221 {
1222 struct wd_softc *wd = device_private(self);
1223
1224 KASSERTMSG(bufq_peek(wd->sc_dksc.sc_bufq) == NULL, "bufq not empty");
1225
1226 if (wd->sc_flags & WDF_DIRTY)
1227 wd_flushcache(wd, AT_WAIT);
1228
1229 wd->atabus->ata_delref(wd->drvp);
1230 wd->sc_flags &= ~WDF_OPEN;
1231
1232 return 0;
1233 }
1234
1235 static int
1236 wdclose(dev_t dev, int flag, int fmt, struct lwp *l)
1237 {
1238 struct wd_softc *wd;
1239 struct dk_softc *dksc;
1240 int unit;
1241
1242 unit = WDUNIT(dev);
1243 wd = device_lookup_private(&wd_cd, unit);
1244 dksc = &wd->sc_dksc;
1245
1246 return dk_close(dksc, dev, flag, fmt, l);
1247 }
1248
1249 void
1250 wdperror(const struct wd_softc *wd, struct ata_xfer *xfer)
1251 {
1252 static const char *const errstr0_3[] = {"address mark not found",
1253 "track 0 not found", "aborted command", "media change requested",
1254 "id not found", "media changed", "uncorrectable data error",
1255 "bad block detected"};
1256 static const char *const errstr4_5[] = {
1257 "obsolete (address mark not found)",
1258 "no media/write protected", "aborted command",
1259 "media change requested", "id not found", "media changed",
1260 "uncorrectable data error", "interface CRC error"};
1261 const char *const *errstr;
1262 int i;
1263 const char *sep = "";
1264
1265 const struct dk_softc *dksc = &wd->sc_dksc;
1266 const char *devname = dksc->sc_xname;
1267 struct ata_drive_datas *drvp = wd->drvp;
1268 int errno = xfer->c_bio.r_error;
1269
1270 if (drvp->ata_vers >= 4)
1271 errstr = errstr4_5;
1272 else
1273 errstr = errstr0_3;
1274
1275 printf("%s: (", devname);
1276
1277 if (errno == 0)
1278 printf("error not notified");
1279
1280 for (i = 0; i < 8; i++) {
1281 if (errno & (1 << i)) {
1282 printf("%s%s", sep, errstr[i]);
1283 sep = ", ";
1284 }
1285 }
1286 printf(")\n");
1287 }
1288
1289 int
1290 wdioctl(dev_t dev, u_long cmd, void *addr, int flag, struct lwp *l)
1291 {
1292 struct wd_softc *wd =
1293 device_lookup_private(&wd_cd, WDUNIT(dev));
1294 struct dk_softc *dksc = &wd->sc_dksc;
1295
1296 ATADEBUG_PRINT(("wdioctl\n"), DEBUG_FUNCS);
1297
1298 if ((wd->sc_flags & WDF_LOADED) == 0)
1299 return EIO;
1300
1301 switch (cmd) {
1302 #ifdef HAS_BAD144_HANDLING
1303 case DIOCSBAD:
1304 if ((flag & FWRITE) == 0)
1305 return EBADF;
1306 dksc->sc_dkdev.dk_cpulabel->bad = *(struct dkbad *)addr;
1307 dksc->sc_dkdev.dk_label->d_flags |= D_BADSECT;
1308 bad144intern(wd);
1309 return 0;
1310 #endif
1311 #ifdef WD_SOFTBADSECT
1312 case DIOCBSLIST: {
1313 uint32_t count, missing, skip;
1314 struct disk_badsecinfo dbsi;
1315 struct disk_badsectors *dbs, dbsbuf;
1316 size_t available;
1317 uint8_t *laddr;
1318 int error;
1319
1320 dbsi = *(struct disk_badsecinfo *)addr;
1321 missing = wd->sc_bscount;
1322 count = 0;
1323 available = dbsi.dbsi_bufsize;
1324 skip = dbsi.dbsi_skip;
1325 laddr = (uint8_t *)dbsi.dbsi_buffer;
1326
1327 /*
1328 * We start this loop with the expectation that all of the
1329 * entries will be missed and decrement this counter each
1330 * time we either skip over one (already copied out) or
1331 * we actually copy it back to user space. The structs
1332 * holding the bad sector information are copied directly
1333 * back to user space whilst the summary is returned via
1334 * the struct passed in via the ioctl.
1335 */
1336 error = 0;
1337 mutex_enter(&wd->sc_lock);
1338 wd->sc_bslist_inuse++;
1339 SLIST_FOREACH(dbs, &wd->sc_bslist, dbs_next) {
1340 if (skip > 0) {
1341 missing--;
1342 skip--;
1343 continue;
1344 }
1345 if (available < sizeof(*dbs))
1346 break;
1347 available -= sizeof(*dbs);
1348 memset(&dbsbuf, 0, sizeof(dbsbuf));
1349 dbsbuf.dbs_min = dbs->dbs_min;
1350 dbsbuf.dbs_max = dbs->dbs_max;
1351 dbsbuf.dbs_failedat = dbs->dbs_failedat;
1352 mutex_exit(&wd->sc_lock);
1353 error = copyout(&dbsbuf, laddr, sizeof(dbsbuf));
1354 mutex_enter(&wd->sc_lock);
1355 if (error)
1356 break;
1357 laddr += sizeof(*dbs);
1358 missing--;
1359 count++;
1360 }
1361 if (--wd->sc_bslist_inuse == 0)
1362 cv_broadcast(&wd->sc_bslist_cv);
1363 mutex_exit(&wd->sc_lock);
1364 dbsi.dbsi_left = missing;
1365 dbsi.dbsi_copied = count;
1366 *(struct disk_badsecinfo *)addr = dbsi;
1367
1368 /*
1369 * If we copied anything out, ignore error and return
1370 * success -- can't back it out.
1371 */
1372 return count ? 0 : error;
1373 }
1374
1375 case DIOCBSFLUSH: {
1376 int error;
1377
1378 /* Clean out the bad sector list */
1379 mutex_enter(&wd->sc_lock);
1380 while (wd->sc_bslist_inuse) {
1381 error = cv_wait_sig(&wd->sc_bslist_cv, &wd->sc_lock);
1382 if (error) {
1383 mutex_exit(&wd->sc_lock);
1384 return error;
1385 }
1386 }
1387 while (!SLIST_EMPTY(&wd->sc_bslist)) {
1388 struct disk_badsectors *dbs =
1389 SLIST_FIRST(&wd->sc_bslist);
1390 SLIST_REMOVE_HEAD(&wd->sc_bslist, dbs_next);
1391 mutex_exit(&wd->sc_lock);
1392 kmem_free(dbs, sizeof(*dbs));
1393 mutex_enter(&wd->sc_lock);
1394 }
1395 mutex_exit(&wd->sc_lock);
1396 wd->sc_bscount = 0;
1397 return 0;
1398 }
1399 #endif
1400
1401 #ifdef notyet
1402 case DIOCWFORMAT:
1403 if ((flag & FWRITE) == 0)
1404 return EBADF;
1405 {
1406 register struct format_op *fop;
1407 struct iovec aiov;
1408 struct uio auio;
1409 int error1;
1410
1411 fop = (struct format_op *)addr;
1412 aiov.iov_base = fop->df_buf;
1413 aiov.iov_len = fop->df_count;
1414 auio.uio_iov = &aiov;
1415 auio.uio_iovcnt = 1;
1416 auio.uio_resid = fop->df_count;
1417 auio.uio_offset =
1418 fop->df_startblk * wd->sc_dk.dk_label->d_secsize;
1419 auio.uio_vmspace = l->l_proc->p_vmspace;
1420 error1 = physio(wdformat, NULL, dev, B_WRITE, wdminphys,
1421 &auio);
1422 fop->df_count -= auio.uio_resid;
1423 fop->df_reg[0] = wdc->sc_status;
1424 fop->df_reg[1] = wdc->sc_error;
1425 return error1;
1426 }
1427 #endif
1428 case DIOCGCACHE:
1429 return wd_getcache(wd, (int *)addr);
1430
1431 case DIOCSCACHE:
1432 return wd_setcache(wd, *(int *)addr);
1433
1434 case DIOCCACHESYNC:
1435 return wd_flushcache(wd, AT_WAIT);
1436
1437 case ATAIOCCOMMAND:
1438 /*
1439 * Make sure this command is (relatively) safe first
1440 */
1441 if ((((atareq_t *) addr)->flags & ATACMD_READ) == 0 &&
1442 (flag & FWRITE) == 0)
1443 return (EBADF);
1444 {
1445 struct wd_ioctl *wi;
1446 atareq_t *atareq = (atareq_t *) addr;
1447 int error1;
1448
1449 wi = wi_get(wd);
1450 wi->wi_atareq = *atareq;
1451
1452 if (atareq->datalen && atareq->flags &
1453 (ATACMD_READ | ATACMD_WRITE)) {
1454 void *tbuf;
1455 if (atareq->datalen < DEV_BSIZE
1456 && atareq->command == WDCC_IDENTIFY) {
1457 tbuf = kmem_zalloc(DEV_BSIZE, KM_SLEEP);
1458 wi->wi_iov.iov_base = tbuf;
1459 wi->wi_iov.iov_len = DEV_BSIZE;
1460 UIO_SETUP_SYSSPACE(&wi->wi_uio);
1461 } else {
1462 tbuf = NULL;
1463 wi->wi_iov.iov_base = atareq->databuf;
1464 wi->wi_iov.iov_len = atareq->datalen;
1465 wi->wi_uio.uio_vmspace = l->l_proc->p_vmspace;
1466 }
1467 wi->wi_uio.uio_iov = &wi->wi_iov;
1468 wi->wi_uio.uio_iovcnt = 1;
1469 wi->wi_uio.uio_resid = atareq->datalen;
1470 wi->wi_uio.uio_offset = 0;
1471 wi->wi_uio.uio_rw =
1472 (atareq->flags & ATACMD_READ) ? B_READ : B_WRITE;
1473 error1 = physio(wdioctlstrategy, &wi->wi_bp, dev,
1474 (atareq->flags & ATACMD_READ) ? B_READ : B_WRITE,
1475 wdminphys, &wi->wi_uio);
1476 if (tbuf != NULL && error1 == 0) {
1477 error1 = copyout(tbuf, atareq->databuf,
1478 atareq->datalen);
1479 kmem_free(tbuf, DEV_BSIZE);
1480 }
1481 } else {
1482 /* No need to call physio if we don't have any
1483 user data */
1484 wi->wi_bp.b_flags = 0;
1485 wi->wi_bp.b_data = 0;
1486 wi->wi_bp.b_bcount = 0;
1487 wi->wi_bp.b_dev = dev;
1488 wi->wi_bp.b_proc = l->l_proc;
1489 wdioctlstrategy(&wi->wi_bp);
1490 error1 = wi->wi_bp.b_error;
1491 }
1492 *atareq = wi->wi_atareq;
1493 wi_free(wi);
1494 return(error1);
1495 }
1496
1497 case DIOCGSECTORALIGN: {
1498 struct disk_sectoralign *dsa = addr;
1499 int part = WDPART(dev);
1500
1501 *dsa = wd->sc_sectoralign;
1502 if (part != RAW_PART) {
1503 struct disklabel *lp = dksc->sc_dkdev.dk_label;
1504 daddr_t offset = lp->d_partitions[part].p_offset;
1505 uint32_t r = offset % dsa->dsa_alignment;
1506
1507 if (r < dsa->dsa_firstaligned)
1508 dsa->dsa_firstaligned = dsa->dsa_firstaligned
1509 - r;
1510 else
1511 dsa->dsa_firstaligned = (dsa->dsa_firstaligned
1512 + dsa->dsa_alignment) - r;
1513 }
1514
1515 return 0;
1516 }
1517
1518 default:
1519 return dk_ioctl(dksc, dev, cmd, addr, flag, l);
1520 }
1521
1522 #ifdef DIAGNOSTIC
1523 panic("wdioctl: impossible");
1524 #endif
1525 }
1526
1527 static int
1528 wd_discard(device_t dev, off_t pos, off_t len)
1529 {
1530 struct wd_softc *wd = device_private(dev);
1531 daddr_t bno;
1532 long size, done;
1533 long maxatonce, amount;
1534 int result;
1535
1536 if (!(wd->sc_params.atap_ata_major & WDC_VER_ATA7)
1537 || !(wd->sc_params.support_dsm & ATA_SUPPORT_DSM_TRIM)) {
1538 /* not supported; ignore request */
1539 ATADEBUG_PRINT(("wddiscard (unsupported)\n"), DEBUG_FUNCS);
1540 return 0;
1541 }
1542 maxatonce = 0xffff; /*wd->sc_params.max_dsm_blocks*/
1543
1544 ATADEBUG_PRINT(("wddiscard\n"), DEBUG_FUNCS);
1545
1546 if ((wd->sc_flags & WDF_LOADED) == 0)
1547 return EIO;
1548
1549 /* round the start up and the end down */
1550 bno = (pos + wd->sc_blksize - 1) / wd->sc_blksize;
1551 size = ((pos + len) / wd->sc_blksize) - bno;
1552
1553 done = 0;
1554 while (done < size) {
1555 amount = size - done;
1556 if (amount > maxatonce) {
1557 amount = maxatonce;
1558 }
1559 result = wd_trim(wd, bno + done, amount);
1560 if (result) {
1561 return result;
1562 }
1563 done += amount;
1564 }
1565 return 0;
1566 }
1567
1568 static int
1569 wddiscard(dev_t dev, off_t pos, off_t len)
1570 {
1571 struct wd_softc *wd;
1572 struct dk_softc *dksc;
1573 int unit;
1574
1575 unit = WDUNIT(dev);
1576 wd = device_lookup_private(&wd_cd, unit);
1577 dksc = &wd->sc_dksc;
1578
1579 return dk_discard(dksc, dev, pos, len);
1580 }
1581
1582 #ifdef B_FORMAT
1583 int
1584 wdformat(struct buf *bp)
1585 {
1586
1587 bp->b_flags |= B_FORMAT;
1588 return wdstrategy(bp);
1589 }
1590 #endif
1591
1592 int
1593 wdsize(dev_t dev)
1594 {
1595 struct wd_softc *wd;
1596 struct dk_softc *dksc;
1597 int unit;
1598
1599 ATADEBUG_PRINT(("wdsize\n"), DEBUG_FUNCS);
1600
1601 unit = WDUNIT(dev);
1602 wd = device_lookup_private(&wd_cd, unit);
1603 if (wd == NULL)
1604 return (-1);
1605 dksc = &wd->sc_dksc;
1606
1607 if (!device_is_active(dksc->sc_dev))
1608 return (-1);
1609
1610 return dk_size(dksc, dev);
1611 }
1612
1613 /*
1614 * Dump core after a system crash.
1615 */
1616 static int
1617 wddump(dev_t dev, daddr_t blkno, void *va, size_t size)
1618 {
1619 struct wd_softc *wd;
1620 struct dk_softc *dksc;
1621 int unit;
1622
1623 /* Check if recursive dump; if so, punt. */
1624 if (wddoingadump)
1625 return EFAULT;
1626 wddoingadump = 1;
1627
1628 unit = WDUNIT(dev);
1629 wd = device_lookup_private(&wd_cd, unit);
1630 if (wd == NULL)
1631 return (ENXIO);
1632 dksc = &wd->sc_dksc;
1633
1634 return dk_dump(dksc, dev, blkno, va, size, 0);
1635 }
1636
1637 static int
1638 wd_dumpblocks(device_t dev, void *va, daddr_t blkno, int nblk)
1639 {
1640 struct wd_softc *wd = device_private(dev);
1641 struct dk_softc *dksc = &wd->sc_dksc;
1642 struct disk_geom *dg = &dksc->sc_dkdev.dk_geom;
1643 struct ata_xfer *xfer = &wd->dump_xfer;
1644 int err;
1645
1646 /* Recalibrate, if first dump transfer. */
1647 if (wddumprecalibrated == 0) {
1648 wddumprecalibrated = 1;
1649 ata_channel_lock(wd->drvp->chnl_softc);
1650 /* This will directly execute the reset due to AT_POLL */
1651 ata_thread_run(wd->drvp->chnl_softc, AT_POLL,
1652 ATACH_TH_DRIVE_RESET, wd->drvp->drive);
1653
1654 wd->drvp->state = RESET;
1655 ata_channel_unlock(wd->drvp->chnl_softc);
1656 }
1657
1658 memset(xfer, 0, sizeof(*xfer));
1659 xfer->c_flags |= C_PRIVATE_ALLOC | C_SKIP_QUEUE;
1660
1661 xfer->c_bio.blkno = blkno;
1662 xfer->c_bio.flags = ATA_POLL;
1663 if (wd->sc_flags & WDF_LBA48 &&
1664 (xfer->c_bio.blkno + nblk) > wd->sc_capacity28)
1665 xfer->c_bio.flags |= ATA_LBA48;
1666 if (wd->sc_flags & WDF_LBA)
1667 xfer->c_bio.flags |= ATA_LBA;
1668 xfer->c_bio.bcount = nblk * dg->dg_secsize;
1669 xfer->c_bio.databuf = va;
1670 #ifndef WD_DUMP_NOT_TRUSTED
1671 /* This will poll until the bio is complete */
1672 wd->atabus->ata_bio(wd->drvp, xfer);
1673
1674 switch(err = xfer->c_bio.error) {
1675 case TIMEOUT:
1676 printf("wddump: device timed out");
1677 err = EIO;
1678 break;
1679 case ERR_DF:
1680 printf("wddump: drive fault");
1681 err = EIO;
1682 break;
1683 case ERR_DMA:
1684 printf("wddump: DMA error");
1685 err = EIO;
1686 break;
1687 case ERROR:
1688 printf("wddump: ");
1689 wdperror(wd, xfer);
1690 err = EIO;
1691 break;
1692 case NOERROR:
1693 err = 0;
1694 break;
1695 default:
1696 panic("wddump: unknown error type %x", err);
1697 }
1698
1699 if (err != 0) {
1700 printf("\n");
1701 return err;
1702 }
1703 #else /* WD_DUMP_NOT_TRUSTED */
1704 /* Let's just talk about this first... */
1705 printf("wd%d: dump addr 0x%x, cylin %d, head %d, sector %d\n",
1706 unit, va, cylin, head, sector);
1707 delay(500 * 1000); /* half a second */
1708 #endif
1709
1710 wddoingadump = 0;
1711 return 0;
1712 }
1713
1714 #ifdef HAS_BAD144_HANDLING
1715 /*
1716 * Internalize the bad sector table.
1717 */
1718 void
1719 bad144intern(struct wd_softc *wd)
1720 {
1721 struct dk_softc *dksc = &wd->sc_dksc;
1722 struct dkbad *bt = &dksc->sc_dkdev.dk_cpulabel->bad;
1723 struct disklabel *lp = dksc->sc_dkdev.dk_label;
1724 int i = 0;
1725
1726 ATADEBUG_PRINT(("bad144intern\n"), DEBUG_XFERS);
1727
1728 for (; i < NBT_BAD; i++) {
1729 if (bt->bt_bad[i].bt_cyl == 0xffff)
1730 break;
1731 wd->drvp->badsect[i] =
1732 bt->bt_bad[i].bt_cyl * lp->d_secpercyl +
1733 (bt->bt_bad[i].bt_trksec >> 8) * lp->d_nsectors +
1734 (bt->bt_bad[i].bt_trksec & 0xff);
1735 }
1736 for (; i < NBT_BAD+1; i++)
1737 wd->drvp->badsect[i] = -1;
1738 }
1739 #endif
1740
1741 static void
1742 wd_set_geometry(struct wd_softc *wd)
1743 {
1744 struct dk_softc *dksc = &wd->sc_dksc;
1745 struct disk_geom *dg = &dksc->sc_dkdev.dk_geom;
1746
1747 memset(dg, 0, sizeof(*dg));
1748
1749 dg->dg_secperunit = wd->sc_capacity;
1750 dg->dg_secsize = wd->sc_blksize;
1751 dg->dg_nsectors = wd->sc_params.atap_sectors;
1752 dg->dg_ntracks = wd->sc_params.atap_heads;
1753 if ((wd->sc_flags & WDF_LBA) == 0)
1754 dg->dg_ncylinders = wd->sc_params.atap_cylinders;
1755
1756 disk_set_info(dksc->sc_dev, &dksc->sc_dkdev, wd->sc_typename);
1757 }
1758
1759 int
1760 wd_get_params(struct wd_softc *wd, struct ataparams *params)
1761 {
1762 int retry = 0;
1763 struct ata_channel *chp = wd->drvp->chnl_softc;
1764 const int flags = AT_WAIT;
1765
1766 again:
1767 switch (wd->atabus->ata_get_params(wd->drvp, flags, params)) {
1768 case CMD_AGAIN:
1769 return 1;
1770 case CMD_ERR:
1771 if (retry == 0) {
1772 retry++;
1773 ata_channel_lock(chp);
1774 (*wd->atabus->ata_reset_drive)(wd->drvp, flags, NULL);
1775 ata_channel_unlock(chp);
1776 goto again;
1777 }
1778
1779 if (wd->drvp->drive_type != ATA_DRIVET_OLD)
1780 return 1;
1781 /*
1782 * We `know' there's a drive here; just assume it's old.
1783 * This geometry is only used to read the MBR and print a
1784 * (false) attach message.
1785 */
1786 strncpy(params->atap_model, "ST506",
1787 sizeof params->atap_model);
1788 params->atap_config = ATA_CFG_FIXED;
1789 params->atap_cylinders = 1024;
1790 params->atap_heads = 8;
1791 params->atap_sectors = 17;
1792 params->atap_multi = 1;
1793 params->atap_capabilities1 = params->atap_capabilities2 = 0;
1794 wd->drvp->ata_vers = -1; /* Mark it as pre-ATA */
1795 /* FALLTHROUGH */
1796 case CMD_OK:
1797 return 0;
1798 default:
1799 panic("wd_get_params: bad return code from ata_get_params");
1800 /* NOTREACHED */
1801 }
1802 }
1803
1804 int
1805 wd_getcache(struct wd_softc *wd, int *bitsp)
1806 {
1807 struct ataparams params;
1808
1809 if (wd_get_params(wd, ¶ms) != 0)
1810 return EIO;
1811 if (params.atap_cmd_set1 == 0x0000 ||
1812 params.atap_cmd_set1 == 0xffff ||
1813 (params.atap_cmd_set1 & WDC_CMD1_CACHE) == 0) {
1814 *bitsp = 0;
1815 return 0;
1816 }
1817 *bitsp = DKCACHE_WCHANGE | DKCACHE_READ;
1818 if (params.atap_cmd1_en & WDC_CMD1_CACHE)
1819 *bitsp |= DKCACHE_WRITE;
1820
1821 if (WD_USE_NCQ(wd) || (wd->drvp->drive_flags & ATA_DRIVE_WFUA))
1822 *bitsp |= DKCACHE_FUA;
1823
1824 return 0;
1825 }
1826
1827
1828 static int
1829 wd_check_error(const struct dk_softc *dksc, const struct ata_xfer *xfer,
1830 const char *func)
1831 {
1832 static const char at_errbits[] = "\20\10ERROR\11TIMEOU\12DF";
1833
1834 int flags = xfer->c_ata_c.flags;
1835
1836 if ((flags & AT_ERROR) != 0 && xfer->c_ata_c.r_error == WDCE_ABRT) {
1837 /* command not supported */
1838 aprint_debug_dev(dksc->sc_dev, "%s: not supported\n", func);
1839 return ENODEV;
1840 }
1841 if (flags & (AT_ERROR | AT_TIMEOU | AT_DF)) {
1842 char sbuf[sizeof(at_errbits) + 64];
1843 snprintb(sbuf, sizeof(sbuf), at_errbits, flags);
1844 aprint_error_dev(dksc->sc_dev, "%s: status=%s\n", func, sbuf);
1845 return EIO;
1846 }
1847 return 0;
1848 }
1849
1850 int
1851 wd_setcache(struct wd_softc *wd, int bits)
1852 {
1853 struct dk_softc *dksc = &wd->sc_dksc;
1854 struct ataparams params;
1855 struct ata_xfer *xfer;
1856 int error;
1857
1858 if (wd_get_params(wd, ¶ms) != 0)
1859 return EIO;
1860
1861 if (params.atap_cmd_set1 == 0x0000 ||
1862 params.atap_cmd_set1 == 0xffff ||
1863 (params.atap_cmd_set1 & WDC_CMD1_CACHE) == 0)
1864 return EOPNOTSUPP;
1865
1866 if ((bits & DKCACHE_READ) == 0 ||
1867 (bits & DKCACHE_SAVE) != 0)
1868 return EOPNOTSUPP;
1869
1870 xfer = ata_get_xfer(wd->drvp->chnl_softc, true);
1871
1872 xfer->c_ata_c.r_command = SET_FEATURES;
1873 xfer->c_ata_c.r_st_bmask = 0;
1874 xfer->c_ata_c.r_st_pmask = 0;
1875 xfer->c_ata_c.timeout = 30000; /* 30s timeout */
1876 xfer->c_ata_c.flags = AT_WAIT;
1877 if (bits & DKCACHE_WRITE)
1878 xfer->c_ata_c.r_features = WDSF_WRITE_CACHE_EN;
1879 else
1880 xfer->c_ata_c.r_features = WDSF_WRITE_CACHE_DS;
1881
1882 wd->atabus->ata_exec_command(wd->drvp, xfer);
1883 ata_wait_cmd(wd->drvp->chnl_softc, xfer);
1884
1885 error = wd_check_error(dksc, xfer, __func__);
1886 ata_free_xfer(wd->drvp->chnl_softc, xfer);
1887 return error;
1888 }
1889
1890 static int
1891 wd_standby(struct wd_softc *wd, int flags)
1892 {
1893 struct dk_softc *dksc = &wd->sc_dksc;
1894 struct ata_xfer *xfer;
1895 int error;
1896
1897 aprint_debug_dev(dksc->sc_dev, "standby immediate\n");
1898 xfer = ata_get_xfer(wd->drvp->chnl_softc, true);
1899
1900 xfer->c_ata_c.r_command = WDCC_STANDBY_IMMED;
1901 xfer->c_ata_c.r_st_bmask = WDCS_DRDY;
1902 xfer->c_ata_c.r_st_pmask = WDCS_DRDY;
1903 xfer->c_ata_c.flags = flags;
1904 xfer->c_ata_c.timeout = 30000; /* 30s timeout */
1905
1906 wd->atabus->ata_exec_command(wd->drvp, xfer);
1907 ata_wait_cmd(wd->drvp->chnl_softc, xfer);
1908
1909 error = wd_check_error(dksc, xfer, __func__);
1910 ata_free_xfer(wd->drvp->chnl_softc, xfer);
1911 return error;
1912 }
1913
1914 int
1915 wd_flushcache(struct wd_softc *wd, int flags)
1916 {
1917 struct dk_softc *dksc = &wd->sc_dksc;
1918 struct ata_xfer *xfer;
1919 int error;
1920
1921 /*
1922 * WDCC_FLUSHCACHE is here since ATA-4, but some drives report
1923 * only ATA-2 and still support it.
1924 */
1925 if (wd->drvp->ata_vers < 4 &&
1926 ((wd->sc_params.atap_cmd_set2 & WDC_CMD2_FC) == 0 ||
1927 wd->sc_params.atap_cmd_set2 == 0xffff))
1928 return ENODEV;
1929
1930 xfer = ata_get_xfer(wd->drvp->chnl_softc, true);
1931
1932 if ((wd->sc_params.atap_cmd2_en & ATA_CMD2_LBA48) != 0 &&
1933 (wd->sc_params.atap_cmd2_en & ATA_CMD2_FCE) != 0) {
1934 xfer->c_ata_c.r_command = WDCC_FLUSHCACHE_EXT;
1935 flags |= AT_LBA48;
1936 } else
1937 xfer->c_ata_c.r_command = WDCC_FLUSHCACHE;
1938 xfer->c_ata_c.r_st_bmask = WDCS_DRDY;
1939 xfer->c_ata_c.r_st_pmask = WDCS_DRDY;
1940 xfer->c_ata_c.flags = flags | AT_READREG;
1941 xfer->c_ata_c.timeout = 300000; /* 5m timeout */
1942
1943 wd->atabus->ata_exec_command(wd->drvp, xfer);
1944 ata_wait_cmd(wd->drvp->chnl_softc, xfer);
1945
1946 error = wd_check_error(dksc, xfer, __func__);
1947 wd->sc_flags &= ~WDF_DIRTY;
1948 ata_free_xfer(wd->drvp->chnl_softc, xfer);
1949 return error;
1950 }
1951
1952 /*
1953 * Execute TRIM command, assumes sleep context.
1954 */
1955 static int
1956 wd_trim(struct wd_softc *wd, daddr_t bno, long size)
1957 {
1958 struct dk_softc *dksc = &wd->sc_dksc;
1959 struct ata_xfer *xfer;
1960 int error;
1961 unsigned char *req;
1962
1963 xfer = ata_get_xfer(wd->drvp->chnl_softc, true);
1964
1965 req = kmem_zalloc(512, KM_SLEEP);
1966 req[0] = bno & 0xff;
1967 req[1] = (bno >> 8) & 0xff;
1968 req[2] = (bno >> 16) & 0xff;
1969 req[3] = (bno >> 24) & 0xff;
1970 req[4] = (bno >> 32) & 0xff;
1971 req[5] = (bno >> 40) & 0xff;
1972 req[6] = size & 0xff;
1973 req[7] = (size >> 8) & 0xff;
1974
1975 /*
1976 * XXX We could possibly use NCQ TRIM, which supports executing
1977 * this command concurrently. It would need some investigation, some
1978 * early or not so early disk firmware caused data loss with NCQ TRIM.
1979 * atastart() et.al would need to be adjusted to allow and support
1980 * running several non-I/O ATA commands in parallel.
1981 */
1982
1983 xfer->c_ata_c.r_command = ATA_DATA_SET_MANAGEMENT;
1984 xfer->c_ata_c.r_count = 1;
1985 xfer->c_ata_c.r_features = ATA_SUPPORT_DSM_TRIM;
1986 xfer->c_ata_c.r_st_bmask = WDCS_DRDY;
1987 xfer->c_ata_c.r_st_pmask = WDCS_DRDY;
1988 xfer->c_ata_c.timeout = 30000; /* 30s timeout */
1989 xfer->c_ata_c.data = req;
1990 xfer->c_ata_c.bcount = 512;
1991 xfer->c_ata_c.flags |= AT_WRITE | AT_WAIT;
1992
1993 wd->atabus->ata_exec_command(wd->drvp, xfer);
1994 ata_wait_cmd(wd->drvp->chnl_softc, xfer);
1995
1996 kmem_free(req, 512);
1997 error = wd_check_error(dksc, xfer, __func__);
1998 ata_free_xfer(wd->drvp->chnl_softc, xfer);
1999 return error;
2000 }
2001
2002 bool
2003 wd_shutdown(device_t dev, int how)
2004 {
2005 struct wd_softc *wd = device_private(dev);
2006
2007 /* the adapter needs to be enabled */
2008 if (wd->atabus->ata_addref(wd->drvp))
2009 return true; /* no need to complain */
2010
2011 wd_flushcache(wd, AT_POLL);
2012 if ((how & RB_POWERDOWN) == RB_POWERDOWN)
2013 wd_standby(wd, AT_POLL);
2014 return true;
2015 }
2016
2017 /*
2018 * Allocate space for a ioctl queue structure. Mostly taken from
2019 * scsipi_ioctl.c
2020 */
2021 struct wd_ioctl *
2022 wi_get(struct wd_softc *wd)
2023 {
2024 struct wd_ioctl *wi;
2025
2026 wi = kmem_zalloc(sizeof(struct wd_ioctl), KM_SLEEP);
2027 wi->wi_softc = wd;
2028 buf_init(&wi->wi_bp);
2029
2030 return (wi);
2031 }
2032
2033 /*
2034 * Free an ioctl structure and remove it from our list
2035 */
2036
2037 void
2038 wi_free(struct wd_ioctl *wi)
2039 {
2040 buf_destroy(&wi->wi_bp);
2041 kmem_free(wi, sizeof(*wi));
2042 }
2043
2044 /*
2045 * Find a wd_ioctl structure based on the struct buf.
2046 */
2047
2048 struct wd_ioctl *
2049 wi_find(struct buf *bp)
2050 {
2051 return container_of(bp, struct wd_ioctl, wi_bp);
2052 }
2053
2054 static uint
2055 wi_sector_size(const struct wd_ioctl * const wi)
2056 {
2057 switch (wi->wi_atareq.command) {
2058 case WDCC_READ:
2059 case WDCC_WRITE:
2060 case WDCC_READMULTI:
2061 case WDCC_WRITEMULTI:
2062 case WDCC_READDMA:
2063 case WDCC_WRITEDMA:
2064 case WDCC_READ_EXT:
2065 case WDCC_WRITE_EXT:
2066 case WDCC_READMULTI_EXT:
2067 case WDCC_WRITEMULTI_EXT:
2068 case WDCC_READDMA_EXT:
2069 case WDCC_WRITEDMA_EXT:
2070 case WDCC_READ_FPDMA_QUEUED:
2071 case WDCC_WRITE_FPDMA_QUEUED:
2072 return wi->wi_softc->sc_blksize;
2073 default:
2074 return 512;
2075 }
2076 }
2077
2078 /*
2079 * Ioctl pseudo strategy routine
2080 *
2081 * This is mostly stolen from scsipi_ioctl.c:scsistrategy(). What
2082 * happens here is:
2083 *
2084 * - wdioctl() queues a wd_ioctl structure.
2085 *
2086 * - wdioctl() calls physio/wdioctlstrategy based on whether or not
2087 * user space I/O is required. If physio() is called, physio() eventually
2088 * calls wdioctlstrategy().
2089 *
2090 * - In either case, wdioctlstrategy() calls wd->atabus->ata_exec_command()
2091 * to perform the actual command
2092 *
2093 * The reason for the use of the pseudo strategy routine is because
2094 * when doing I/O to/from user space, physio _really_ wants to be in
2095 * the loop. We could put the entire buffer into the ioctl request
2096 * structure, but that won't scale if we want to do things like download
2097 * microcode.
2098 */
2099
2100 void
2101 wdioctlstrategy(struct buf *bp)
2102 {
2103 struct wd_ioctl *wi;
2104 struct ata_xfer *xfer;
2105 int error = 0;
2106
2107 wi = wi_find(bp);
2108 if (wi == NULL) {
2109 printf("wdioctlstrategy: "
2110 "No matching ioctl request found in queue\n");
2111 error = EINVAL;
2112 goto out2;
2113 }
2114
2115 xfer = ata_get_xfer(wi->wi_softc->drvp->chnl_softc, true);
2116
2117 /*
2118 * Abort if physio broke up the transfer
2119 */
2120
2121 if (bp->b_bcount != wi->wi_atareq.datalen) {
2122 printf("physio split wd ioctl request... cannot proceed\n");
2123 error = EIO;
2124 goto out;
2125 }
2126
2127 /*
2128 * Abort if we didn't get a buffer size that was a multiple of
2129 * our sector size (or overflows CHS/LBA28 sector count)
2130 */
2131
2132 if ((bp->b_bcount % wi_sector_size(wi)) != 0 ||
2133 (bp->b_bcount / wi_sector_size(wi)) >=
2134 (1 << NBBY)) {
2135 error = EINVAL;
2136 goto out;
2137 }
2138
2139 /*
2140 * Make sure a timeout was supplied in the ioctl request
2141 */
2142
2143 if (wi->wi_atareq.timeout == 0) {
2144 error = EINVAL;
2145 goto out;
2146 }
2147
2148 if (wi->wi_atareq.flags & ATACMD_READ)
2149 xfer->c_ata_c.flags |= AT_READ;
2150 else if (wi->wi_atareq.flags & ATACMD_WRITE)
2151 xfer->c_ata_c.flags |= AT_WRITE;
2152
2153 if (wi->wi_atareq.flags & ATACMD_READREG)
2154 xfer->c_ata_c.flags |= AT_READREG;
2155
2156 if ((wi->wi_atareq.flags & ATACMD_LBA) != 0)
2157 xfer->c_ata_c.flags |= AT_LBA;
2158
2159 xfer->c_ata_c.flags |= AT_WAIT;
2160
2161 xfer->c_ata_c.timeout = wi->wi_atareq.timeout;
2162 xfer->c_ata_c.r_command = wi->wi_atareq.command;
2163 xfer->c_ata_c.r_lba = ((wi->wi_atareq.head & 0x0f) << 24) |
2164 (wi->wi_atareq.cylinder << 8) |
2165 wi->wi_atareq.sec_num;
2166 xfer->c_ata_c.r_count = wi->wi_atareq.sec_count;
2167 xfer->c_ata_c.r_features = wi->wi_atareq.features;
2168 xfer->c_ata_c.r_st_bmask = WDCS_DRDY;
2169 xfer->c_ata_c.r_st_pmask = WDCS_DRDY;
2170 xfer->c_ata_c.data = wi->wi_bp.b_data;
2171 xfer->c_ata_c.bcount = wi->wi_bp.b_bcount;
2172
2173 wi->wi_softc->atabus->ata_exec_command(wi->wi_softc->drvp, xfer);
2174 ata_wait_cmd(wi->wi_softc->drvp->chnl_softc, xfer);
2175
2176 if (xfer->c_ata_c.flags & (AT_ERROR | AT_TIMEOU | AT_DF)) {
2177 if (xfer->c_ata_c.flags & AT_ERROR) {
2178 wi->wi_atareq.retsts = ATACMD_ERROR;
2179 wi->wi_atareq.error = xfer->c_ata_c.r_error;
2180 } else if (xfer->c_ata_c.flags & AT_DF)
2181 wi->wi_atareq.retsts = ATACMD_DF;
2182 else
2183 wi->wi_atareq.retsts = ATACMD_TIMEOUT;
2184 } else {
2185 wi->wi_atareq.retsts = ATACMD_OK;
2186 if (wi->wi_atareq.flags & ATACMD_READREG) {
2187 wi->wi_atareq.command = xfer->c_ata_c.r_status;
2188 wi->wi_atareq.features = xfer->c_ata_c.r_error;
2189 wi->wi_atareq.sec_count = xfer->c_ata_c.r_count;
2190 wi->wi_atareq.sec_num = xfer->c_ata_c.r_lba & 0xff;
2191 wi->wi_atareq.head = (xfer->c_ata_c.r_device & 0xf0) |
2192 ((xfer->c_ata_c.r_lba >> 24) & 0x0f);
2193 wi->wi_atareq.cylinder =
2194 (xfer->c_ata_c.r_lba >> 8) & 0xffff;
2195 wi->wi_atareq.error = xfer->c_ata_c.r_error;
2196 }
2197 }
2198
2199 out:
2200 ata_free_xfer(wi->wi_softc->drvp->chnl_softc, xfer);
2201 out2:
2202 bp->b_error = error;
2203 if (error)
2204 bp->b_resid = bp->b_bcount;
2205 biodone(bp);
2206 }
2207
2208 static void
2209 wd_sysctl_attach(struct wd_softc *wd)
2210 {
2211 struct dk_softc *dksc = &wd->sc_dksc;
2212 const struct sysctlnode *node;
2213 int error;
2214
2215 /* sysctl set-up */
2216 if (sysctl_createv(&wd->nodelog, 0, NULL, &node,
2217 0, CTLTYPE_NODE, dksc->sc_xname,
2218 SYSCTL_DESCR("wd driver settings"),
2219 NULL, 0, NULL, 0,
2220 CTL_HW, CTL_CREATE, CTL_EOL) != 0) {
2221 aprint_error_dev(dksc->sc_dev,
2222 "could not create %s.%s sysctl node\n",
2223 "hw", dksc->sc_xname);
2224 return;
2225 }
2226
2227 wd->drv_ncq = true;
2228 if ((error = sysctl_createv(&wd->nodelog, 0, NULL, NULL,
2229 CTLFLAG_READWRITE, CTLTYPE_BOOL, "use_ncq",
2230 SYSCTL_DESCR("use NCQ if supported"),
2231 NULL, 0, &wd->drv_ncq, 0,
2232 CTL_HW, node->sysctl_num, CTL_CREATE, CTL_EOL))
2233 != 0) {
2234 aprint_error_dev(dksc->sc_dev,
2235 "could not create %s.%s.use_ncq sysctl - error %d\n",
2236 "hw", dksc->sc_xname, error);
2237 return;
2238 }
2239
2240 wd->drv_ncq_prio = false;
2241 if ((error = sysctl_createv(&wd->nodelog, 0, NULL, NULL,
2242 CTLFLAG_READWRITE, CTLTYPE_BOOL, "use_ncq_prio",
2243 SYSCTL_DESCR("use NCQ PRIORITY if supported"),
2244 NULL, 0, &wd->drv_ncq_prio, 0,
2245 CTL_HW, node->sysctl_num, CTL_CREATE, CTL_EOL))
2246 != 0) {
2247 aprint_error_dev(dksc->sc_dev,
2248 "could not create %s.%s.use_ncq_prio sysctl - error %d\n",
2249 "hw", dksc->sc_xname, error);
2250 return;
2251 }
2252
2253 #ifdef WD_CHAOS_MONKEY
2254 wd->drv_chaos_freq = 0;
2255 if ((error = sysctl_createv(&wd->nodelog, 0, NULL, NULL,
2256 CTLFLAG_READWRITE, CTLTYPE_INT, "chaos_freq",
2257 SYSCTL_DESCR("simulated bio read error rate"),
2258 NULL, 0, &wd->drv_chaos_freq, 0,
2259 CTL_HW, node->sysctl_num, CTL_CREATE, CTL_EOL))
2260 != 0) {
2261 aprint_error_dev(dksc->sc_dev,
2262 "could not create %s.%s.chaos_freq sysctl - error %d\n",
2263 "hw", dksc->sc_xname, error);
2264 return;
2265 }
2266
2267 wd->drv_chaos_cnt = 0;
2268 if ((error = sysctl_createv(&wd->nodelog, 0, NULL, NULL,
2269 CTLFLAG_READONLY, CTLTYPE_INT, "chaos_cnt",
2270 SYSCTL_DESCR("number of processed bio reads"),
2271 NULL, 0, &wd->drv_chaos_cnt, 0,
2272 CTL_HW, node->sysctl_num, CTL_CREATE, CTL_EOL))
2273 != 0) {
2274 aprint_error_dev(dksc->sc_dev,
2275 "could not create %s.%s.chaos_cnt sysctl - error %d\n",
2276 "hw", dksc->sc_xname, error);
2277 return;
2278 }
2279 #endif
2280
2281 }
2282
2283 static void
2284 wd_sysctl_detach(struct wd_softc *wd)
2285 {
2286 sysctl_teardown(&wd->nodelog);
2287 }
2288
2289 #ifdef ATADEBUG
2290 int wddebug(void);
2291
2292 int
2293 wddebug(void)
2294 {
2295 struct wd_softc *wd;
2296 struct dk_softc *dksc;
2297 int unit;
2298
2299 for (unit = 0; unit <= 3; unit++) {
2300 wd = device_lookup_private(&wd_cd, unit);
2301 if (wd == NULL)
2302 continue;
2303 dksc = &wd->sc_dksc;
2304 printf("%s fl %x bufq %p:\n",
2305 dksc->sc_xname, wd->sc_flags, bufq_peek(dksc->sc_bufq));
2306
2307 atachannel_debug(wd->drvp->chnl_softc);
2308 }
2309 return 0;
2310 }
2311 #endif /* ATADEBUG */
2312