subr_disk.c revision 1.59 1 /* $NetBSD: subr_disk.c,v 1.59 2004/02/28 06:28:48 yamt Exp $ */
2
3 /*-
4 * Copyright (c) 1996, 1997, 1999, 2000 The NetBSD Foundation, Inc.
5 * All rights reserved.
6 *
7 * This code is derived from software contributed to The NetBSD Foundation
8 * by Jason R. Thorpe of the Numerical Aerospace Simulation Facility,
9 * NASA Ames Research Center.
10 *
11 * Redistribution and use in source and binary forms, with or without
12 * modification, are permitted provided that the following conditions
13 * are met:
14 * 1. Redistributions of source code must retain the above copyright
15 * notice, this list of conditions and the following disclaimer.
16 * 2. Redistributions in binary form must reproduce the above copyright
17 * notice, this list of conditions and the following disclaimer in the
18 * documentation and/or other materials provided with the distribution.
19 * 3. All advertising materials mentioning features or use of this software
20 * must display the following acknowledgement:
21 * This product includes software developed by the NetBSD
22 * Foundation, Inc. and its contributors.
23 * 4. Neither the name of The NetBSD Foundation nor the names of its
24 * contributors may be used to endorse or promote products derived
25 * from this software without specific prior written permission.
26 *
27 * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
28 * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
29 * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
30 * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
31 * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
32 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
33 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
34 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
35 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
36 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
37 * POSSIBILITY OF SUCH DAMAGE.
38 */
39
40 /*
41 * Copyright (c) 1982, 1986, 1988, 1993
42 * The Regents of the University of California. All rights reserved.
43 * (c) UNIX System Laboratories, Inc.
44 * All or some portions of this file are derived from material licensed
45 * to the University of California by American Telephone and Telegraph
46 * Co. or Unix System Laboratories, Inc. and are reproduced herein with
47 * the permission of UNIX System Laboratories, Inc.
48 *
49 * Redistribution and use in source and binary forms, with or without
50 * modification, are permitted provided that the following conditions
51 * are met:
52 * 1. Redistributions of source code must retain the above copyright
53 * notice, this list of conditions and the following disclaimer.
54 * 2. Redistributions in binary form must reproduce the above copyright
55 * notice, this list of conditions and the following disclaimer in the
56 * documentation and/or other materials provided with the distribution.
57 * 3. Neither the name of the University nor the names of its contributors
58 * may be used to endorse or promote products derived from this software
59 * without specific prior written permission.
60 *
61 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
62 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
63 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
64 * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
65 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
66 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
67 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
68 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
69 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
70 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
71 * SUCH DAMAGE.
72 *
73 * @(#)ufs_disksubr.c 8.5 (Berkeley) 1/21/94
74 */
75
76 #include <sys/cdefs.h>
77 __KERNEL_RCSID(0, "$NetBSD: subr_disk.c,v 1.59 2004/02/28 06:28:48 yamt Exp $");
78
79 #include "opt_compat_netbsd.h"
80 #include "opt_bufq.h"
81
82 #include <sys/param.h>
83 #include <sys/kernel.h>
84 #include <sys/malloc.h>
85 #include <sys/buf.h>
86 #include <sys/syslog.h>
87 #include <sys/disklabel.h>
88 #include <sys/disk.h>
89 #include <sys/sysctl.h>
90 #include <lib/libkern/libkern.h>
91
92 /*
93 * A global list of all disks attached to the system. May grow or
94 * shrink over time.
95 */
96 struct disklist_head disklist; /* TAILQ_HEAD */
97 int disk_count; /* number of drives in global disklist */
98 struct simplelock disklist_slock = SIMPLELOCK_INITIALIZER;
99
100 #ifdef NEW_BUFQ_STRATEGY
101 int bufq_disk_default_strat = BUFQ_READ_PRIO;
102 #else /* NEW_BUFQ_STRATEGY */
103 int bufq_disk_default_strat = BUFQ_DISKSORT;
104 #endif /* NEW_BUFQ_STRATEGY */
105
106 /*
107 * Compute checksum for disk label.
108 */
109 u_int
110 dkcksum(struct disklabel *lp)
111 {
112 u_short *start, *end;
113 u_short sum = 0;
114
115 start = (u_short *)lp;
116 end = (u_short *)&lp->d_partitions[lp->d_npartitions];
117 while (start < end)
118 sum ^= *start++;
119 return (sum);
120 }
121
122 /*
123 * Disk error is the preface to plaintive error messages
124 * about failing disk transfers. It prints messages of the form
125
126 hp0g: hard error reading fsbn 12345 of 12344-12347 (hp0 bn %d cn %d tn %d sn %d)
127
128 * if the offset of the error in the transfer and a disk label
129 * are both available. blkdone should be -1 if the position of the error
130 * is unknown; the disklabel pointer may be null from drivers that have not
131 * been converted to use them. The message is printed with printf
132 * if pri is LOG_PRINTF, otherwise it uses log at the specified priority.
133 * The message should be completed (with at least a newline) with printf
134 * or addlog, respectively. There is no trailing space.
135 */
136 #ifndef PRIdaddr
137 #define PRIdaddr PRId64
138 #endif
139 void
140 diskerr(const struct buf *bp, const char *dname, const char *what, int pri,
141 int blkdone, const struct disklabel *lp)
142 {
143 int unit = DISKUNIT(bp->b_dev), part = DISKPART(bp->b_dev);
144 void (*pr)(const char *, ...);
145 char partname = 'a' + part;
146 daddr_t sn;
147
148 if (/*CONSTCOND*/0)
149 /* Compiler will error this is the format is wrong... */
150 printf("%" PRIdaddr, bp->b_blkno);
151
152 if (pri != LOG_PRINTF) {
153 static const char fmt[] = "";
154 log(pri, fmt);
155 pr = addlog;
156 } else
157 pr = printf;
158 (*pr)("%s%d%c: %s %sing fsbn ", dname, unit, partname, what,
159 bp->b_flags & B_READ ? "read" : "writ");
160 sn = bp->b_blkno;
161 if (bp->b_bcount <= DEV_BSIZE)
162 (*pr)("%" PRIdaddr, sn);
163 else {
164 if (blkdone >= 0) {
165 sn += blkdone;
166 (*pr)("%" PRIdaddr " of ", sn);
167 }
168 (*pr)("%" PRIdaddr "-%" PRIdaddr "", bp->b_blkno,
169 bp->b_blkno + (bp->b_bcount - 1) / DEV_BSIZE);
170 }
171 if (lp && (blkdone >= 0 || bp->b_bcount <= lp->d_secsize)) {
172 sn += lp->d_partitions[part].p_offset;
173 (*pr)(" (%s%d bn %" PRIdaddr "; cn %" PRIdaddr "",
174 dname, unit, sn, sn / lp->d_secpercyl);
175 sn %= lp->d_secpercyl;
176 (*pr)(" tn %" PRIdaddr " sn %" PRIdaddr ")",
177 sn / lp->d_nsectors, sn % lp->d_nsectors);
178 }
179 }
180
181 /*
182 * Initialize the disklist. Called by main() before autoconfiguration.
183 */
184 void
185 disk_init(void)
186 {
187
188 TAILQ_INIT(&disklist);
189 disk_count = 0;
190 }
191
192 /*
193 * Searches the disklist for the disk corresponding to the
194 * name provided.
195 */
196 struct disk *
197 disk_find(char *name)
198 {
199 struct disk *diskp;
200
201 if ((name == NULL) || (disk_count <= 0))
202 return (NULL);
203
204 simple_lock(&disklist_slock);
205 for (diskp = TAILQ_FIRST(&disklist); diskp != NULL;
206 diskp = TAILQ_NEXT(diskp, dk_link))
207 if (strcmp(diskp->dk_name, name) == 0) {
208 simple_unlock(&disklist_slock);
209 return (diskp);
210 }
211 simple_unlock(&disklist_slock);
212
213 return (NULL);
214 }
215
216 /*
217 * Attach a disk.
218 */
219 void
220 disk_attach(struct disk *diskp)
221 {
222 int s;
223
224 /*
225 * Allocate and initialize the disklabel structures. Note that
226 * it's not safe to sleep here, since we're probably going to be
227 * called during autoconfiguration.
228 */
229 diskp->dk_label = malloc(sizeof(struct disklabel), M_DEVBUF, M_NOWAIT);
230 diskp->dk_cpulabel = malloc(sizeof(struct cpu_disklabel), M_DEVBUF,
231 M_NOWAIT);
232 if ((diskp->dk_label == NULL) || (diskp->dk_cpulabel == NULL))
233 panic("disk_attach: can't allocate storage for disklabel");
234
235 memset(diskp->dk_label, 0, sizeof(struct disklabel));
236 memset(diskp->dk_cpulabel, 0, sizeof(struct cpu_disklabel));
237
238 /*
239 * Set the attached timestamp.
240 */
241 s = splclock();
242 diskp->dk_attachtime = mono_time;
243 splx(s);
244
245 /*
246 * Link into the disklist.
247 */
248 simple_lock(&disklist_slock);
249 TAILQ_INSERT_TAIL(&disklist, diskp, dk_link);
250 simple_unlock(&disklist_slock);
251 ++disk_count;
252 }
253
254 /*
255 * Detach a disk.
256 */
257 void
258 disk_detach(struct disk *diskp)
259 {
260
261 /*
262 * Remove from the disklist.
263 */
264 if (--disk_count < 0)
265 panic("disk_detach: disk_count < 0");
266 simple_lock(&disklist_slock);
267 TAILQ_REMOVE(&disklist, diskp, dk_link);
268 simple_unlock(&disklist_slock);
269
270 /*
271 * Free the space used by the disklabel structures.
272 */
273 free(diskp->dk_label, M_DEVBUF);
274 free(diskp->dk_cpulabel, M_DEVBUF);
275 }
276
277 /*
278 * Increment a disk's busy counter. If the counter is going from
279 * 0 to 1, set the timestamp.
280 */
281 void
282 disk_busy(struct disk *diskp)
283 {
284 int s;
285
286 /*
287 * XXX We'd like to use something as accurate as microtime(),
288 * but that doesn't depend on the system TOD clock.
289 */
290 if (diskp->dk_busy++ == 0) {
291 s = splclock();
292 diskp->dk_timestamp = mono_time;
293 splx(s);
294 }
295 }
296
297 /*
298 * Decrement a disk's busy counter, increment the byte count, total busy
299 * time, and reset the timestamp.
300 */
301 void
302 disk_unbusy(struct disk *diskp, long bcount, int read)
303 {
304 int s;
305 struct timeval dv_time, diff_time;
306
307 if (diskp->dk_busy-- == 0) {
308 printf("%s: dk_busy < 0\n", diskp->dk_name);
309 panic("disk_unbusy");
310 }
311
312 s = splclock();
313 dv_time = mono_time;
314 splx(s);
315
316 timersub(&dv_time, &diskp->dk_timestamp, &diff_time);
317 timeradd(&diskp->dk_time, &diff_time, &diskp->dk_time);
318
319 diskp->dk_timestamp = dv_time;
320 if (bcount > 0) {
321 if (read) {
322 diskp->dk_rbytes += bcount;
323 diskp->dk_rxfer++;
324 } else {
325 diskp->dk_wbytes += bcount;
326 diskp->dk_wxfer++;
327 }
328 }
329 }
330
331 /*
332 * Reset the metrics counters on the given disk. Note that we cannot
333 * reset the busy counter, as it may case a panic in disk_unbusy().
334 * We also must avoid playing with the timestamp information, as it
335 * may skew any pending transfer results.
336 */
337 void
338 disk_resetstat(struct disk *diskp)
339 {
340 int s = splbio(), t;
341
342 diskp->dk_rxfer = 0;
343 diskp->dk_rbytes = 0;
344 diskp->dk_wxfer = 0;
345 diskp->dk_wbytes = 0;
346
347 t = splclock();
348 diskp->dk_attachtime = mono_time;
349 splx(t);
350
351 timerclear(&diskp->dk_time);
352
353 splx(s);
354 }
355
356 int
357 sysctl_hw_disknames(SYSCTLFN_ARGS)
358 {
359 char buf[DK_DISKNAMELEN + 1];
360 char *where = oldp;
361 struct disk *diskp;
362 size_t needed, left, slen;
363 int error, first;
364
365 if (newp != NULL)
366 return (EPERM);
367 if (namelen != 0)
368 return (EINVAL);
369
370 first = 1;
371 error = 0;
372 needed = 0;
373 left = *oldlenp;
374
375 simple_lock(&disklist_slock);
376 for (diskp = TAILQ_FIRST(&disklist); diskp != NULL;
377 diskp = TAILQ_NEXT(diskp, dk_link)) {
378 if (where == NULL)
379 needed += strlen(diskp->dk_name) + 1;
380 else {
381 memset(buf, 0, sizeof(buf));
382 if (first) {
383 strncpy(buf, diskp->dk_name, sizeof(buf));
384 first = 0;
385 } else {
386 buf[0] = ' ';
387 strncpy(buf + 1, diskp->dk_name,
388 sizeof(buf) - 1);
389 }
390 buf[DK_DISKNAMELEN] = '\0';
391 slen = strlen(buf);
392 if (left < slen + 1)
393 break;
394 /* +1 to copy out the trailing NUL byte */
395 error = copyout(buf, where, slen + 1);
396 if (error)
397 break;
398 where += slen;
399 needed += slen;
400 left -= slen;
401 }
402 }
403 simple_unlock(&disklist_slock);
404 *oldlenp = needed;
405 return (error);
406 }
407
408 int
409 sysctl_hw_diskstats(SYSCTLFN_ARGS)
410 {
411 struct disk_sysctl sdisk;
412 struct disk *diskp;
413 char *where = oldp;
414 size_t tocopy, left;
415 int error;
416
417 if (newp != NULL)
418 return (EPERM);
419
420 /*
421 * The original hw.diskstats call was broken and did not require
422 * the userland to pass in it's size of struct disk_sysctl. This
423 * was fixed after NetBSD 1.6 was released, and any applications
424 * that do not pass in the size are given an error only, unless
425 * we care about 1.6 compatibility.
426 */
427 if (namelen == 0)
428 #ifdef COMPAT_16
429 tocopy = offsetof(struct disk_sysctl, dk_rxfer);
430 #else
431 return (EINVAL);
432 #endif
433 else
434 tocopy = name[0];
435
436 if (where == NULL) {
437 *oldlenp = disk_count * tocopy;
438 return (0);
439 }
440
441 error = 0;
442 left = *oldlenp;
443 memset(&sdisk, 0, sizeof(sdisk));
444 *oldlenp = 0;
445
446 simple_lock(&disklist_slock);
447 TAILQ_FOREACH(diskp, &disklist, dk_link) {
448 if (left < tocopy)
449 break;
450 strncpy(sdisk.dk_name, diskp->dk_name, sizeof(sdisk.dk_name));
451 sdisk.dk_xfer = diskp->dk_rxfer + diskp->dk_wxfer;
452 sdisk.dk_rxfer = diskp->dk_rxfer;
453 sdisk.dk_wxfer = diskp->dk_wxfer;
454 sdisk.dk_seek = diskp->dk_seek;
455 sdisk.dk_bytes = diskp->dk_rbytes + diskp->dk_wbytes;
456 sdisk.dk_rbytes = diskp->dk_rbytes;
457 sdisk.dk_wbytes = diskp->dk_wbytes;
458 sdisk.dk_attachtime_sec = diskp->dk_attachtime.tv_sec;
459 sdisk.dk_attachtime_usec = diskp->dk_attachtime.tv_usec;
460 sdisk.dk_timestamp_sec = diskp->dk_timestamp.tv_sec;
461 sdisk.dk_timestamp_usec = diskp->dk_timestamp.tv_usec;
462 sdisk.dk_time_sec = diskp->dk_time.tv_sec;
463 sdisk.dk_time_usec = diskp->dk_time.tv_usec;
464 sdisk.dk_busy = diskp->dk_busy;
465
466 error = copyout(&sdisk, where, min(tocopy, sizeof(sdisk)));
467 if (error)
468 break;
469 where += tocopy;
470 *oldlenp += tocopy;
471 left -= tocopy;
472 }
473 simple_unlock(&disklist_slock);
474 return (error);
475 }
476
477 struct bufq_fcfs {
478 TAILQ_HEAD(, buf) bq_head; /* actual list of buffers */
479 };
480
481 struct bufq_disksort {
482 TAILQ_HEAD(, buf) bq_head; /* actual list of buffers */
483 };
484
485 #define PRIO_READ_BURST 48
486 #define PRIO_WRITE_REQ 16
487
488 struct bufq_prio {
489 TAILQ_HEAD(, buf) bq_read, bq_write; /* actual list of buffers */
490 struct buf *bq_write_next; /* next request in bq_write */
491 struct buf *bq_next; /* current request */
492 int bq_read_burst; /* # of consecutive reads */
493 };
494
495
496 /*
497 * Check if two buf's are in ascending order.
498 */
499 static __inline int
500 buf_inorder(struct buf *bp, struct buf *bq, int sortby)
501 {
502
503 if (bp == NULL || bq == NULL)
504 return (bq == NULL);
505
506 if (sortby == BUFQ_SORT_CYLINDER) {
507 if (bp->b_cylinder != bq->b_cylinder)
508 return bp->b_cylinder < bq->b_cylinder;
509 else
510 return bp->b_rawblkno < bq->b_rawblkno;
511 } else
512 return bp->b_rawblkno < bq->b_rawblkno;
513 }
514
515
516 /*
517 * First-come first-served sort for disks.
518 *
519 * Requests are appended to the queue without any reordering.
520 */
521 static void
522 bufq_fcfs_put(struct bufq_state *bufq, struct buf *bp)
523 {
524 struct bufq_fcfs *fcfs = bufq->bq_private;
525
526 TAILQ_INSERT_TAIL(&fcfs->bq_head, bp, b_actq);
527 }
528
529 static struct buf *
530 bufq_fcfs_get(struct bufq_state *bufq, int remove)
531 {
532 struct bufq_fcfs *fcfs = bufq->bq_private;
533 struct buf *bp;
534
535 bp = TAILQ_FIRST(&fcfs->bq_head);
536
537 if (bp != NULL && remove)
538 TAILQ_REMOVE(&fcfs->bq_head, bp, b_actq);
539
540 return (bp);
541 }
542
543
544 /*
545 * Seek sort for disks.
546 *
547 * There are actually two queues, sorted in ascendening order. The first
548 * queue holds those requests which are positioned after the current block;
549 * the second holds requests which came in after their position was passed.
550 * Thus we implement a one-way scan, retracting after reaching the end of
551 * the drive to the first request on the second queue, at which time it
552 * becomes the first queue.
553 *
554 * A one-way scan is natural because of the way UNIX read-ahead blocks are
555 * allocated.
556 */
557 static void
558 bufq_disksort_put(struct bufq_state *bufq, struct buf *bp)
559 {
560 struct bufq_disksort *disksort = bufq->bq_private;
561 struct buf *bq, *nbq;
562 int sortby;
563
564 sortby = bufq->bq_flags & BUFQ_SORT_MASK;
565
566 bq = TAILQ_FIRST(&disksort->bq_head);
567
568 /*
569 * If the queue is empty it's easy; we just go on the end.
570 */
571 if (bq == NULL) {
572 TAILQ_INSERT_TAIL(&disksort->bq_head, bp, b_actq);
573 return;
574 }
575
576 /*
577 * If we lie before the currently active request, then we
578 * must locate the second request list and add ourselves to it.
579 */
580 if (buf_inorder(bp, bq, sortby)) {
581 while ((nbq = TAILQ_NEXT(bq, b_actq)) != NULL) {
582 /*
583 * Check for an ``inversion'' in the normally ascending
584 * block numbers, indicating the start of the second
585 * request list.
586 */
587 if (buf_inorder(nbq, bq, sortby)) {
588 /*
589 * Search the second request list for the first
590 * request at a larger block number. We go
591 * after that; if there is no such request, we
592 * go at the end.
593 */
594 do {
595 if (buf_inorder(bp, nbq, sortby))
596 goto insert;
597 bq = nbq;
598 } while ((nbq =
599 TAILQ_NEXT(bq, b_actq)) != NULL);
600 goto insert; /* after last */
601 }
602 bq = nbq;
603 }
604 /*
605 * No inversions... we will go after the last, and
606 * be the first request in the second request list.
607 */
608 goto insert;
609 }
610 /*
611 * Request is at/after the current request...
612 * sort in the first request list.
613 */
614 while ((nbq = TAILQ_NEXT(bq, b_actq)) != NULL) {
615 /*
616 * We want to go after the current request if there is an
617 * inversion after it (i.e. it is the end of the first
618 * request list), or if the next request is a larger cylinder
619 * than our request.
620 */
621 if (buf_inorder(nbq, bq, sortby) ||
622 buf_inorder(bp, nbq, sortby))
623 goto insert;
624 bq = nbq;
625 }
626 /*
627 * Neither a second list nor a larger request... we go at the end of
628 * the first list, which is the same as the end of the whole schebang.
629 */
630 insert: TAILQ_INSERT_AFTER(&disksort->bq_head, bq, bp, b_actq);
631 }
632
633 static struct buf *
634 bufq_disksort_get(struct bufq_state *bufq, int remove)
635 {
636 struct bufq_disksort *disksort = bufq->bq_private;
637 struct buf *bp;
638
639 bp = TAILQ_FIRST(&disksort->bq_head);
640
641 if (bp != NULL && remove)
642 TAILQ_REMOVE(&disksort->bq_head, bp, b_actq);
643
644 return (bp);
645 }
646
647
648 /*
649 * Seek sort for disks.
650 *
651 * There are two queues. The first queue holds read requests; the second
652 * holds write requests. The read queue is first-come first-served; the
653 * write queue is sorted in ascendening block order.
654 * The read queue is processed first. After PRIO_READ_BURST consecutive
655 * read requests with non-empty write queue PRIO_WRITE_REQ requests from
656 * the write queue will be processed.
657 */
658 static void
659 bufq_prio_put(struct bufq_state *bufq, struct buf *bp)
660 {
661 struct bufq_prio *prio = bufq->bq_private;
662 struct buf *bq;
663 int sortby;
664
665 sortby = bufq->bq_flags & BUFQ_SORT_MASK;
666
667 /*
668 * If it's a read request append it to the list.
669 */
670 if ((bp->b_flags & B_READ) == B_READ) {
671 TAILQ_INSERT_TAIL(&prio->bq_read, bp, b_actq);
672 return;
673 }
674
675 bq = TAILQ_FIRST(&prio->bq_write);
676
677 /*
678 * If the write list is empty, simply append it to the list.
679 */
680 if (bq == NULL) {
681 TAILQ_INSERT_TAIL(&prio->bq_write, bp, b_actq);
682 prio->bq_write_next = bp;
683 return;
684 }
685
686 /*
687 * If we lie after the next request, insert after this request.
688 */
689 if (buf_inorder(prio->bq_write_next, bp, sortby))
690 bq = prio->bq_write_next;
691
692 /*
693 * Search for the first request at a larger block number.
694 * We go before this request if it exists.
695 */
696 while (bq != NULL && buf_inorder(bq, bp, sortby))
697 bq = TAILQ_NEXT(bq, b_actq);
698
699 if (bq != NULL)
700 TAILQ_INSERT_BEFORE(bq, bp, b_actq);
701 else
702 TAILQ_INSERT_TAIL(&prio->bq_write, bp, b_actq);
703 }
704
705 static struct buf *
706 bufq_prio_get(struct bufq_state *bufq, int remove)
707 {
708 struct bufq_prio *prio = bufq->bq_private;
709 struct buf *bp;
710
711 /*
712 * If no current request, get next from the lists.
713 */
714 if (prio->bq_next == NULL) {
715 /*
716 * If at least one list is empty, select the other.
717 */
718 if (TAILQ_FIRST(&prio->bq_read) == NULL) {
719 prio->bq_next = prio->bq_write_next;
720 prio->bq_read_burst = 0;
721 } else if (prio->bq_write_next == NULL) {
722 prio->bq_next = TAILQ_FIRST(&prio->bq_read);
723 prio->bq_read_burst = 0;
724 } else {
725 /*
726 * Both list have requests. Select the read list up
727 * to PRIO_READ_BURST times, then select the write
728 * list PRIO_WRITE_REQ times.
729 */
730 if (prio->bq_read_burst++ < PRIO_READ_BURST)
731 prio->bq_next = TAILQ_FIRST(&prio->bq_read);
732 else if (prio->bq_read_burst <
733 PRIO_READ_BURST + PRIO_WRITE_REQ)
734 prio->bq_next = prio->bq_write_next;
735 else {
736 prio->bq_next = TAILQ_FIRST(&prio->bq_read);
737 prio->bq_read_burst = 0;
738 }
739 }
740 }
741
742 bp = prio->bq_next;
743
744 if (bp != NULL && remove) {
745 if ((bp->b_flags & B_READ) == B_READ)
746 TAILQ_REMOVE(&prio->bq_read, bp, b_actq);
747 else {
748 /*
749 * Advance the write pointer before removing
750 * bp since it is actually prio->bq_write_next.
751 */
752 prio->bq_write_next =
753 TAILQ_NEXT(prio->bq_write_next, b_actq);
754 TAILQ_REMOVE(&prio->bq_write, bp, b_actq);
755 if (prio->bq_write_next == NULL)
756 prio->bq_write_next =
757 TAILQ_FIRST(&prio->bq_write);
758 }
759
760 prio->bq_next = NULL;
761 }
762
763 return (bp);
764 }
765
766
767 /*
768 * Cyclical scan (CSCAN)
769 */
770 TAILQ_HEAD(bqhead, buf);
771 struct cscan_queue {
772 struct bqhead cq_head[2]; /* actual lists of buffers */
773 int cq_idx; /* current list index */
774 int cq_lastcylinder; /* b_cylinder of the last request */
775 daddr_t cq_lastrawblkno; /* b_rawblkno of the last request */
776 };
777
778 static int __inline cscan_empty(const struct cscan_queue *);
779 static void cscan_put(struct cscan_queue *, struct buf *, int);
780 static struct buf *cscan_get(struct cscan_queue *, int);
781 static void cscan_init(struct cscan_queue *);
782
783 static __inline int
784 cscan_empty(const struct cscan_queue *q)
785 {
786
787 return TAILQ_EMPTY(&q->cq_head[0]) && TAILQ_EMPTY(&q->cq_head[1]);
788 }
789
790 static void
791 cscan_put(struct cscan_queue *q, struct buf *bp, int sortby)
792 {
793 struct buf tmp;
794 struct buf *it;
795 struct bqhead *bqh;
796 int idx;
797
798 tmp.b_cylinder = q->cq_lastcylinder;
799 tmp.b_rawblkno = q->cq_lastrawblkno;
800
801 if (buf_inorder(bp, &tmp, sortby))
802 idx = 1 - q->cq_idx;
803 else
804 idx = q->cq_idx;
805
806 bqh = &q->cq_head[idx];
807
808 TAILQ_FOREACH(it, bqh, b_actq)
809 if (buf_inorder(bp, it, sortby))
810 break;
811
812 if (it != NULL)
813 TAILQ_INSERT_BEFORE(it, bp, b_actq);
814 else
815 TAILQ_INSERT_TAIL(bqh, bp, b_actq);
816 }
817
818 static struct buf *
819 cscan_get(struct cscan_queue *q, int remove)
820 {
821 int idx = q->cq_idx;
822 struct bqhead *bqh;
823 struct buf *bp;
824
825 bqh = &q->cq_head[idx];
826 bp = TAILQ_FIRST(bqh);
827
828 if (bp == NULL) {
829 /* switch queue */
830 idx = 1 - idx;
831 bqh = &q->cq_head[idx];
832 bp = TAILQ_FIRST(bqh);
833 }
834
835 KDASSERT((bp != NULL && !cscan_empty(q)) ||
836 (bp == NULL && cscan_empty(q)));
837
838 if (bp != NULL && remove) {
839 q->cq_idx = idx;
840 TAILQ_REMOVE(bqh, bp, b_actq);
841
842 q->cq_lastcylinder = bp->b_cylinder;
843 q->cq_lastrawblkno =
844 bp->b_rawblkno + (bp->b_bcount >> DEV_BSHIFT);
845 }
846
847 return (bp);
848 }
849
850 static void
851 cscan_init(struct cscan_queue *q)
852 {
853
854 TAILQ_INIT(&q->cq_head[0]);
855 TAILQ_INIT(&q->cq_head[1]);
856 }
857
858
859 /*
860 * Per-prioritiy CSCAN.
861 *
862 * XXX probably we should have a way to raise
863 * priority of the on-queue requests.
864 */
865 #define PRIOCSCAN_NQUEUE 3
866
867 struct priocscan_queue {
868 struct cscan_queue q_queue;
869 int q_burst;
870 };
871
872 struct bufq_priocscan {
873 struct priocscan_queue bq_queue[PRIOCSCAN_NQUEUE];
874
875 #if 0
876 /*
877 * XXX using "global" head position can reduce positioning time
878 * when switching between queues.
879 * although it might affect against fairness.
880 */
881 daddr_t bq_lastrawblkno;
882 int bq_lastcylinder;
883 #endif
884 };
885
886 /*
887 * how many requests to serve when having pending requests on other queues.
888 *
889 * XXX tune
890 */
891 const int priocscan_burst[] = {
892 64, 16, 4
893 };
894
895 static void bufq_priocscan_put(struct bufq_state *, struct buf *);
896 static struct buf *bufq_priocscan_get(struct bufq_state *, int);
897 static void bufq_priocscan_init(struct bufq_state *);
898 static __inline struct cscan_queue *bufq_priocscan_selectqueue(
899 struct bufq_priocscan *, const struct buf *);
900
901 static __inline struct cscan_queue *
902 bufq_priocscan_selectqueue(struct bufq_priocscan *q, const struct buf *bp)
903 {
904 static const int priocscan_priomap[] = {
905 [BPRIO_TIMENONCRITICAL] = 2,
906 [BPRIO_TIMELIMITED] = 1,
907 [BPRIO_TIMECRITICAL] = 0
908 };
909
910 return &q->bq_queue[priocscan_priomap[BIO_GETPRIO(bp)]].q_queue;
911 }
912
913 static void
914 bufq_priocscan_put(struct bufq_state *bufq, struct buf *bp)
915 {
916 struct bufq_priocscan *q = bufq->bq_private;
917 struct cscan_queue *cq;
918 const int sortby = bufq->bq_flags & BUFQ_SORT_MASK;
919
920 cq = bufq_priocscan_selectqueue(q, bp);
921 cscan_put(cq, bp, sortby);
922 }
923
924 static struct buf *
925 bufq_priocscan_get(struct bufq_state *bufq, int remove)
926 {
927 struct bufq_priocscan *q = bufq->bq_private;
928 struct priocscan_queue *pq, *npq;
929 struct priocscan_queue *first; /* first non-empty queue */
930 const struct priocscan_queue *epq;
931 const struct cscan_queue *cq;
932 struct buf *bp;
933 boolean_t single; /* true if there's only one non-empty queue */
934
935 pq = &q->bq_queue[0];
936 epq = pq + PRIOCSCAN_NQUEUE;
937 for (; pq < epq; pq++) {
938 cq = &pq->q_queue;
939 if (!cscan_empty(cq))
940 break;
941 }
942 if (pq == epq) {
943 /* there's no requests */
944 return NULL;
945 }
946
947 first = pq;
948 single = TRUE;
949 for (npq = first + 1; npq < epq; npq++) {
950 cq = &npq->q_queue;
951 if (!cscan_empty(cq)) {
952 single = FALSE;
953 if (pq->q_burst > 0)
954 break;
955 pq = npq;
956 }
957 }
958 if (single) {
959 /*
960 * there's only a non-empty queue. just serve it.
961 */
962 pq = first;
963 } else if (pq->q_burst > 0) {
964 /*
965 * XXX account only by number of requests. is it good enough?
966 */
967 pq->q_burst--;
968 } else {
969 /*
970 * no queue was selected due to burst counts
971 */
972 int i;
973 #ifdef DEBUG
974 for (i = 0; i < PRIOCSCAN_NQUEUE; i++) {
975 pq = &q->bq_queue[i];
976 cq = &pq->q_queue;
977 if (!cscan_empty(cq) && pq->q_burst)
978 panic("%s: inconsist", __func__);
979 }
980 #endif /* DEBUG */
981
982 /*
983 * reset burst counts
984 */
985 for (i = 0; i < PRIOCSCAN_NQUEUE; i++) {
986 pq = &q->bq_queue[i];
987 pq->q_burst = priocscan_burst[i];
988 }
989
990 /*
991 * serve first non-empty queue.
992 */
993 pq = first;
994 }
995
996 KDASSERT(!cscan_empty(&pq->q_queue));
997 bp = cscan_get(&pq->q_queue, remove);
998 KDASSERT(bp != NULL);
999 KDASSERT(&pq->q_queue == bufq_priocscan_selectqueue(q, bp));
1000
1001 return bp;
1002 }
1003
1004 static void
1005 bufq_priocscan_init(struct bufq_state *bufq)
1006 {
1007 struct bufq_priocscan *q;
1008 int i;
1009
1010 bufq->bq_get = bufq_priocscan_get;
1011 bufq->bq_put = bufq_priocscan_put;
1012 bufq->bq_private = malloc(sizeof(struct bufq_priocscan),
1013 M_DEVBUF, M_ZERO);
1014
1015 q = bufq->bq_private;
1016 for (i = 0; i < PRIOCSCAN_NQUEUE; i++) {
1017 struct cscan_queue *cq = &q->bq_queue[i].q_queue;
1018
1019 cscan_init(cq);
1020 }
1021 }
1022
1023
1024 /*
1025 * Create a device buffer queue.
1026 */
1027 void
1028 bufq_alloc(struct bufq_state *bufq, int flags)
1029 {
1030 struct bufq_fcfs *fcfs;
1031 struct bufq_disksort *disksort;
1032 struct bufq_prio *prio;
1033
1034 bufq->bq_flags = flags;
1035
1036 switch (flags & BUFQ_SORT_MASK) {
1037 case BUFQ_SORT_RAWBLOCK:
1038 case BUFQ_SORT_CYLINDER:
1039 break;
1040 case 0:
1041 if ((flags & BUFQ_METHOD_MASK) == BUFQ_FCFS)
1042 break;
1043 /* FALLTHROUGH */
1044 default:
1045 panic("bufq_alloc: sort out of range");
1046 }
1047
1048 switch (flags & BUFQ_METHOD_MASK) {
1049 case BUFQ_FCFS:
1050 bufq->bq_get = bufq_fcfs_get;
1051 bufq->bq_put = bufq_fcfs_put;
1052 MALLOC(bufq->bq_private, struct bufq_fcfs *,
1053 sizeof(struct bufq_fcfs), M_DEVBUF, M_ZERO);
1054 fcfs = (struct bufq_fcfs *)bufq->bq_private;
1055 TAILQ_INIT(&fcfs->bq_head);
1056 break;
1057 case BUFQ_DISKSORT:
1058 bufq->bq_get = bufq_disksort_get;
1059 bufq->bq_put = bufq_disksort_put;
1060 MALLOC(bufq->bq_private, struct bufq_disksort *,
1061 sizeof(struct bufq_disksort), M_DEVBUF, M_ZERO);
1062 disksort = (struct bufq_disksort *)bufq->bq_private;
1063 TAILQ_INIT(&disksort->bq_head);
1064 break;
1065 case BUFQ_READ_PRIO:
1066 bufq->bq_get = bufq_prio_get;
1067 bufq->bq_put = bufq_prio_put;
1068 MALLOC(bufq->bq_private, struct bufq_prio *,
1069 sizeof(struct bufq_prio), M_DEVBUF, M_ZERO);
1070 prio = (struct bufq_prio *)bufq->bq_private;
1071 TAILQ_INIT(&prio->bq_read);
1072 TAILQ_INIT(&prio->bq_write);
1073 break;
1074 case BUFQ_PRIOCSCAN:
1075 bufq_priocscan_init(bufq);
1076 break;
1077 default:
1078 panic("bufq_alloc: method out of range");
1079 }
1080 }
1081
1082 /*
1083 * Destroy a device buffer queue.
1084 */
1085 void
1086 bufq_free(struct bufq_state *bufq)
1087 {
1088
1089 KASSERT(bufq->bq_private != NULL);
1090 KASSERT(BUFQ_PEEK(bufq) == NULL);
1091
1092 FREE(bufq->bq_private, M_DEVBUF);
1093 bufq->bq_get = NULL;
1094 bufq->bq_put = NULL;
1095 }
1096
1097 /*
1098 * Bounds checking against the media size, used for the raw partition.
1099 * The sector size passed in should currently always be DEV_BSIZE,
1100 * and the media size the size of the device in DEV_BSIZE sectors.
1101 */
1102 int
1103 bounds_check_with_mediasize(struct buf *bp, int secsize, u_int64_t mediasize)
1104 {
1105 int sz;
1106
1107 sz = howmany(bp->b_bcount, secsize);
1108
1109 if (bp->b_blkno + sz > mediasize) {
1110 sz = mediasize - bp->b_blkno;
1111 if (sz == 0) {
1112 /* If exactly at end of disk, return EOF. */
1113 bp->b_resid = bp->b_bcount;
1114 goto done;
1115 }
1116 if (sz < 0) {
1117 /* If past end of disk, return EINVAL. */
1118 bp->b_error = EINVAL;
1119 goto bad;
1120 }
1121 /* Otherwise, truncate request. */
1122 bp->b_bcount = sz << DEV_BSHIFT;
1123 }
1124
1125 return 1;
1126
1127 bad:
1128 bp->b_flags |= B_ERROR;
1129 done:
1130 return 0;
1131 }
1132