virtio_pci.c revision 1.54 1 /* $NetBSD: virtio_pci.c,v 1.54 2024/06/25 14:55:23 riastradh Exp $ */
2
3 /*
4 * Copyright (c) 2020 The NetBSD Foundation, Inc.
5 * Copyright (c) 2012 Stefan Fritsch.
6 * Copyright (c) 2010 Minoura Makoto.
7 * All rights reserved.
8 *
9 * Redistribution and use in source and binary forms, with or without
10 * modification, are permitted provided that the following conditions
11 * are met:
12 * 1. Redistributions of source code must retain the above copyright
13 * notice, this list of conditions and the following disclaimer.
14 * 2. Redistributions in binary form must reproduce the above copyright
15 * notice, this list of conditions and the following disclaimer in the
16 * documentation and/or other materials provided with the distribution.
17 *
18 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
19 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
20 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
21 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
22 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
23 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
24 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
25 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
26 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
27 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
28 */
29
30 #include <sys/cdefs.h>
31 __KERNEL_RCSID(0, "$NetBSD: virtio_pci.c,v 1.54 2024/06/25 14:55:23 riastradh Exp $");
32
33 #include <sys/param.h>
34 #include <sys/types.h>
35
36 #include <sys/device.h>
37 #include <sys/endian.h>
38 #include <sys/interrupt.h>
39 #include <sys/kmem.h>
40 #include <sys/module.h>
41 #include <sys/syslog.h>
42 #include <sys/systm.h>
43
44 #include <dev/pci/pcidevs.h>
45 #include <dev/pci/pcireg.h>
46 #include <dev/pci/pcivar.h>
47
48 #include <dev/pci/virtioreg.h> /* XXX: move to non-pci */
49 #include <dev/pci/virtio_pcireg.h>
50
51 #define VIRTIO_PRIVATE
52 #include <dev/pci/virtiovar.h> /* XXX: move to non-pci */
53
54 #if defined(__alpha__) || defined(__sparc64__)
55 /*
56 * XXX VIRTIO_F_ACCESS_PLATFORM is required for standard PCI DMA
57 * XXX to work on these platforms, at least by Qemu.
58 * XXX
59 * XXX Generalize this later.
60 */
61 #define __NEED_VIRTIO_F_ACCESS_PLATFORM
62 #endif /* __alpha__ || __sparc64__ */
63
64 #define VIRTIO_PCI_LOG(_sc, _use_log, _fmt, _args...) \
65 do { \
66 if ((_use_log)) { \
67 log(LOG_DEBUG, "%s: " _fmt, \
68 device_xname((_sc)->sc_dev), \
69 ##_args); \
70 } else { \
71 aprint_error_dev((_sc)->sc_dev, \
72 _fmt, ##_args); \
73 } \
74 } while(0)
75
76 static int virtio_pci_match(device_t, cfdata_t, void *);
77 static void virtio_pci_attach(device_t, device_t, void *);
78 static int virtio_pci_rescan(device_t, const char *, const int *);
79 static int virtio_pci_detach(device_t, int);
80
81 #define NMAPREG ((PCI_MAPREG_END - PCI_MAPREG_START) / \
82 sizeof(pcireg_t))
83 struct virtio_pci_softc {
84 struct virtio_softc sc_sc;
85 bool sc_intr_pervq;
86
87 /* IO space */
88 bus_space_tag_t sc_iot;
89 bus_space_handle_t sc_ioh;
90 bus_size_t sc_iosize;
91
92 /* BARs */
93 bus_space_tag_t sc_bars_iot[NMAPREG];
94 bus_space_handle_t sc_bars_ioh[NMAPREG];
95 bus_size_t sc_bars_iosize[NMAPREG];
96
97 /* notify space */
98 bus_space_tag_t sc_notify_iot;
99 bus_space_handle_t sc_notify_ioh;
100 bus_size_t sc_notify_iosize;
101 uint32_t sc_notify_off_multiplier;
102
103 /* isr space */
104 bus_space_tag_t sc_isr_iot;
105 bus_space_handle_t sc_isr_ioh;
106 bus_size_t sc_isr_iosize;
107
108 /* generic */
109 struct pci_attach_args sc_pa;
110 pci_intr_handle_t *sc_ihp;
111 void **sc_ihs;
112 int sc_ihs_num;
113 int sc_devcfg_offset; /* for 0.9 */
114 };
115
116 static int virtio_pci_attach_09(device_t, void *);
117 static void virtio_pci_kick_09(struct virtio_softc *, uint16_t);
118 static uint16_t virtio_pci_read_queue_size_09(struct virtio_softc *, uint16_t);
119 static void virtio_pci_setup_queue_09(struct virtio_softc *, uint16_t,
120 uint64_t);
121 static void virtio_pci_set_status_09(struct virtio_softc *, int);
122 static void virtio_pci_negotiate_features_09(struct virtio_softc *,
123 uint64_t);
124
125 static int virtio_pci_attach_10(device_t, void *);
126 static void virtio_pci_kick_10(struct virtio_softc *, uint16_t);
127 static uint16_t virtio_pci_read_queue_size_10(struct virtio_softc *, uint16_t);
128 static void virtio_pci_setup_queue_10(struct virtio_softc *, uint16_t,
129 uint64_t);
130 static void virtio_pci_set_status_10(struct virtio_softc *, int);
131 static void virtio_pci_negotiate_features_10(struct virtio_softc *,
132 uint64_t);
133 static int virtio_pci_find_cap(struct virtio_pci_softc *, int, void *,
134 int);
135
136 static int virtio_pci_alloc_interrupts(struct virtio_softc *);
137 static void virtio_pci_free_interrupts(struct virtio_softc *);
138 static int virtio_pci_adjust_config_region(struct virtio_pci_softc *);
139 static int virtio_pci_intr(void *);
140 static int virtio_pci_msix_queue_intr(void *);
141 static int virtio_pci_msix_config_intr(void *);
142 static int virtio_pci_setup_interrupts_09(struct virtio_softc *, int);
143 static int virtio_pci_setup_interrupts_10(struct virtio_softc *, int);
144 static int virtio_pci_establish_msix_interrupts(struct virtio_softc *,
145 const struct pci_attach_args *);
146 static int virtio_pci_establish_intx_interrupt(struct virtio_softc *,
147 const struct pci_attach_args *);
148 static bool virtio_pci_msix_enabled(struct virtio_pci_softc *);
149
150 #define VIRTIO_MSIX_CONFIG_VECTOR_INDEX 0
151 #define VIRTIO_MSIX_QUEUE_VECTOR_INDEX 1
152
153 /*
154 * For big-endian aarch64/armv7 on QEMU (and most real HW), only CPU cores
155 * are running in big-endian mode, with all peripheral being configured to
156 * little-endian mode. Their default bus_space(9) functions forcibly swap
157 * byte-order. This guarantees that PIO'ed data from pci(4), e.g., are
158 * correctly handled by bus_space(9), while DMA'ed ones should be swapped
159 * by hand, in violation of virtio(4) specifications.
160 */
161
162 #if (defined(__aarch64__) || defined(__arm__)) && BYTE_ORDER == BIG_ENDIAN
163 # define READ_ENDIAN_09 BIG_ENDIAN
164 # define READ_ENDIAN_10 BIG_ENDIAN
165 # define STRUCT_ENDIAN_09 BIG_ENDIAN
166 # define STRUCT_ENDIAN_10 LITTLE_ENDIAN
167 #elif BYTE_ORDER == BIG_ENDIAN
168 # define READ_ENDIAN_09 LITTLE_ENDIAN
169 # define READ_ENDIAN_10 BIG_ENDIAN
170 # define STRUCT_ENDIAN_09 BIG_ENDIAN
171 # define STRUCT_ENDIAN_10 LITTLE_ENDIAN
172 #else /* little endian */
173 # define READ_ENDIAN_09 LITTLE_ENDIAN
174 # define READ_ENDIAN_10 LITTLE_ENDIAN
175 # define STRUCT_ENDIAN_09 LITTLE_ENDIAN
176 # define STRUCT_ENDIAN_10 LITTLE_ENDIAN
177 #endif
178
179 CFATTACH_DECL3_NEW(virtio_pci, sizeof(struct virtio_pci_softc),
180 virtio_pci_match, virtio_pci_attach, virtio_pci_detach, NULL,
181 virtio_pci_rescan, NULL, 0);
182
183 static const struct virtio_ops virtio_pci_ops_09 = {
184 .kick = virtio_pci_kick_09,
185 .read_queue_size = virtio_pci_read_queue_size_09,
186 .setup_queue = virtio_pci_setup_queue_09,
187 .set_status = virtio_pci_set_status_09,
188 .neg_features = virtio_pci_negotiate_features_09,
189 .alloc_interrupts = virtio_pci_alloc_interrupts,
190 .free_interrupts = virtio_pci_free_interrupts,
191 .setup_interrupts = virtio_pci_setup_interrupts_09,
192 };
193
194 static const struct virtio_ops virtio_pci_ops_10 = {
195 .kick = virtio_pci_kick_10,
196 .read_queue_size = virtio_pci_read_queue_size_10,
197 .setup_queue = virtio_pci_setup_queue_10,
198 .set_status = virtio_pci_set_status_10,
199 .neg_features = virtio_pci_negotiate_features_10,
200 .alloc_interrupts = virtio_pci_alloc_interrupts,
201 .free_interrupts = virtio_pci_free_interrupts,
202 .setup_interrupts = virtio_pci_setup_interrupts_10,
203 };
204
205 static int
206 virtio_pci_match(device_t parent, cfdata_t match, void *aux)
207 {
208 const struct pci_attach_args * const pa = aux;
209
210 switch (PCI_VENDOR(pa->pa_id)) {
211 case PCI_VENDOR_QUMRANET:
212 /* Transitional devices MUST have a PCI Revision ID of 0. */
213 if (((PCI_PRODUCT_QUMRANET_VIRTIO_1000 <=
214 PCI_PRODUCT(pa->pa_id)) &&
215 (PCI_PRODUCT(pa->pa_id) <=
216 PCI_PRODUCT_QUMRANET_VIRTIO_103F)) &&
217 PCI_REVISION(pa->pa_class) == 0)
218 return 1;
219 /*
220 * Non-transitional devices SHOULD have a PCI Revision
221 * ID of 1 or higher. Drivers MUST match any PCI
222 * Revision ID value.
223 */
224 if (((PCI_PRODUCT_QUMRANET_VIRTIO_1040 <=
225 PCI_PRODUCT(pa->pa_id)) &&
226 (PCI_PRODUCT(pa->pa_id) <=
227 PCI_PRODUCT_QUMRANET_VIRTIO_107F)) &&
228 /* XXX: TODO */
229 PCI_REVISION(pa->pa_class) == 1)
230 return 1;
231 break;
232 }
233
234 return 0;
235 }
236
237 static void
238 virtio_pci_attach(device_t parent, device_t self, void *aux)
239 {
240 struct virtio_pci_softc * const psc = device_private(self);
241 struct virtio_softc * const sc = &psc->sc_sc;
242 const struct pci_attach_args * const pa = aux;
243 pci_chipset_tag_t pc = pa->pa_pc;
244 pcitag_t tag = pa->pa_tag;
245 int revision;
246 int ret;
247 pcireg_t id;
248 pcireg_t csr;
249
250 revision = PCI_REVISION(pa->pa_class);
251 switch (revision) {
252 case 0:
253 /* subsystem ID shows what I am */
254 id = PCI_SUBSYS_ID(pci_conf_read(pc, tag, PCI_SUBSYS_ID_REG));
255 break;
256 case 1:
257 /* pci product number shows what I am */
258 id = PCI_PRODUCT(pa->pa_id) - PCI_PRODUCT_QUMRANET_VIRTIO_1040;
259 break;
260 default:
261 aprint_normal(": unknown revision 0x%02x; giving up\n",
262 revision);
263 return;
264 }
265
266 aprint_normal("\n");
267 aprint_naive("\n");
268 virtio_print_device_type(self, id, revision);
269
270 csr = pci_conf_read(pc, tag, PCI_COMMAND_STATUS_REG);
271 csr |= PCI_COMMAND_MASTER_ENABLE | PCI_COMMAND_IO_ENABLE;
272 pci_conf_write(pc, tag, PCI_COMMAND_STATUS_REG, csr);
273
274 sc->sc_dev = self;
275 psc->sc_pa = *pa;
276 psc->sc_iot = pa->pa_iot;
277
278 sc->sc_dmat = pa->pa_dmat;
279 if (pci_dma64_available(pa))
280 sc->sc_dmat = pa->pa_dmat64;
281
282 /* attach is dependent on revision */
283 ret = 0;
284 if (revision == 1) {
285 /* try to attach 1.0 */
286 ret = virtio_pci_attach_10(self, aux);
287 }
288 if (ret == 0 && revision == 0) {
289 /*
290 * revision 0 means 0.9 only or both 0.9 and 1.0. The
291 * latter are so-called "Transitional Devices". For
292 * those devices, we want to use the 1.0 interface if
293 * possible.
294 *
295 * XXX Currently only on platforms that require 1.0
296 * XXX features, such as VIRTIO_F_ACCESS_PLATFORM.
297 */
298 #ifdef __NEED_VIRTIO_F_ACCESS_PLATFORM
299 /* First, try to attach 1.0 */
300 ret = virtio_pci_attach_10(self, aux);
301 if (ret != 0) {
302 aprint_error_dev(self,
303 "VirtIO 1.0 error = %d, falling back to 0.9\n",
304 ret);
305 /* Fall back to 0.9. */
306 ret = virtio_pci_attach_09(self, aux);
307 }
308 #else
309 ret = virtio_pci_attach_09(self, aux);
310 #endif /* __NEED_VIRTIO_F_ACCESS_PLATFORM */
311 }
312 if (ret) {
313 aprint_error_dev(self, "cannot attach (%d)\n", ret);
314 return;
315 }
316 KASSERT(sc->sc_ops);
317
318 /* preset config region */
319 psc->sc_devcfg_offset = VIRTIO_CONFIG_DEVICE_CONFIG_NOMSI;
320 if (virtio_pci_adjust_config_region(psc))
321 return;
322
323 /* generic */
324 virtio_device_reset(sc);
325 virtio_set_status(sc, VIRTIO_CONFIG_DEVICE_STATUS_ACK);
326 virtio_set_status(sc, VIRTIO_CONFIG_DEVICE_STATUS_DRIVER);
327
328 sc->sc_childdevid = id;
329 sc->sc_child = NULL;
330 virtio_pci_rescan(self, NULL, NULL);
331 return;
332 }
333
334 /* ARGSUSED */
335 static int
336 virtio_pci_rescan(device_t self, const char *ifattr, const int *locs)
337 {
338 struct virtio_pci_softc * const psc = device_private(self);
339 struct virtio_softc * const sc = &psc->sc_sc;
340 struct virtio_attach_args va;
341
342 if (sc->sc_child) /* Child already attached? */
343 return 0;
344
345 memset(&va, 0, sizeof(va));
346 va.sc_childdevid = sc->sc_childdevid;
347
348 config_found(self, &va, NULL, CFARGS_NONE);
349
350 if (virtio_attach_failed(sc))
351 return 0;
352
353 return 0;
354 }
355
356 static int
357 virtio_pci_detach(device_t self, int flags)
358 {
359 struct virtio_pci_softc * const psc = device_private(self);
360 struct virtio_softc * const sc = &psc->sc_sc;
361 unsigned i;
362 int r;
363
364 r = config_detach_children(self, flags);
365 if (r != 0)
366 return r;
367
368 /* Check that child never attached, or detached properly */
369 KASSERT(sc->sc_child == NULL);
370 KASSERT(sc->sc_vqs == NULL);
371 KASSERT(psc->sc_ihs_num == 0);
372
373 if (sc->sc_version_1) {
374 for (i = 0; i < __arraycount(psc->sc_bars_iot); i++) {
375 if (psc->sc_bars_iosize[i] == 0)
376 continue;
377 bus_space_unmap(psc->sc_bars_iot[i],
378 psc->sc_bars_ioh[i], psc->sc_bars_iosize[i]);
379 psc->sc_bars_iosize[i] = 0;
380 }
381 } else {
382 if (psc->sc_iosize) {
383 bus_space_unmap(psc->sc_iot, psc->sc_ioh,
384 psc->sc_iosize);
385 psc->sc_iosize = 0;
386 }
387 }
388
389 return 0;
390 }
391
392 static int
393 virtio_pci_attach_09(device_t self, void *aux)
394 {
395 struct virtio_pci_softc * const psc = device_private(self);
396 const struct pci_attach_args * const pa = aux;
397 struct virtio_softc * const sc = &psc->sc_sc;
398
399 /* complete IO region */
400 if (pci_mapreg_map(pa, PCI_MAPREG_START, PCI_MAPREG_TYPE_IO, 0,
401 &psc->sc_iot, &psc->sc_ioh, NULL, &psc->sc_iosize)) {
402 aprint_error_dev(self, "can't map i/o space\n");
403 return EIO;
404 }
405
406 /* queue space */
407 if (bus_space_subregion(psc->sc_iot, psc->sc_ioh,
408 VIRTIO_CONFIG_QUEUE_NOTIFY, 2, &psc->sc_notify_ioh)) {
409 aprint_error_dev(self, "can't map notify i/o space\n");
410 return EIO;
411 }
412 psc->sc_notify_iosize = 2;
413 psc->sc_notify_iot = psc->sc_iot;
414
415 /* ISR space */
416 if (bus_space_subregion(psc->sc_iot, psc->sc_ioh,
417 VIRTIO_CONFIG_ISR_STATUS, 1, &psc->sc_isr_ioh)) {
418 aprint_error_dev(self, "can't map isr i/o space\n");
419 return EIO;
420 }
421 psc->sc_isr_iosize = 1;
422 psc->sc_isr_iot = psc->sc_iot;
423
424 /* set our version 0.9 ops */
425 sc->sc_ops = &virtio_pci_ops_09;
426 sc->sc_bus_endian = READ_ENDIAN_09;
427 sc->sc_struct_endian = STRUCT_ENDIAN_09;
428 return 0;
429 }
430
431 static int
432 virtio_pci_attach_10(device_t self, void *aux)
433 {
434 struct virtio_pci_softc * const psc = device_private(self);
435 const struct pci_attach_args * const pa = aux;
436 struct virtio_softc * const sc = &psc->sc_sc;
437 const pci_chipset_tag_t pc = pa->pa_pc;
438 const pcitag_t tag = pa->pa_tag;
439
440 struct virtio_pci_cap common, isr, device;
441 struct virtio_pci_notify_cap notify;
442 int have_device_cfg = 0;
443 bus_size_t bars[NMAPREG] = { 0 };
444 int bars_idx[NMAPREG] = { 0 };
445 struct virtio_pci_cap * const caps[] =
446 { &common, &isr, &device, ¬ify.cap };
447 int i, j, ret = 0;
448
449 if (virtio_pci_find_cap(psc, VIRTIO_PCI_CAP_COMMON_CFG,
450 &common, sizeof(common)))
451 return ENODEV;
452 if (virtio_pci_find_cap(psc, VIRTIO_PCI_CAP_NOTIFY_CFG,
453 ¬ify, sizeof(notify)))
454 return ENODEV;
455 if (virtio_pci_find_cap(psc, VIRTIO_PCI_CAP_ISR_CFG,
456 &isr, sizeof(isr)))
457 return ENODEV;
458 if (virtio_pci_find_cap(psc, VIRTIO_PCI_CAP_DEVICE_CFG,
459 &device, sizeof(device)))
460 memset(&device, 0, sizeof(device));
461 else
462 have_device_cfg = 1;
463
464 /* Figure out which bars we need to map */
465 for (i = 0; i < __arraycount(caps); i++) {
466 int bar = caps[i]->bar;
467 bus_size_t len = caps[i]->offset + caps[i]->length;
468
469 if (caps[i]->length == 0)
470 continue;
471 if (bars[bar] < len)
472 bars[bar] = len;
473 }
474
475 for (i = j = 0; i < __arraycount(bars); i++) {
476 int reg;
477 pcireg_t type;
478
479 if (bars[i] == 0)
480 continue;
481 reg = PCI_BAR(i);
482 type = pci_mapreg_type(pc, tag, reg);
483 if (pci_mapreg_map(pa, reg, type, 0,
484 &psc->sc_bars_iot[j], &psc->sc_bars_ioh[j],
485 NULL, &psc->sc_bars_iosize[j])) {
486 aprint_error_dev(self, "can't map bar %u \n", i);
487 ret = EIO;
488 goto err;
489 }
490 aprint_debug_dev(self,
491 "bar[%d]: iot %p, size 0x%" PRIxBUSSIZE "\n",
492 j, psc->sc_bars_iot[j], psc->sc_bars_iosize[j]);
493 bars_idx[i] = j;
494 j++;
495 }
496
497 i = bars_idx[notify.cap.bar];
498 if (bus_space_subregion(psc->sc_bars_iot[i], psc->sc_bars_ioh[i],
499 notify.cap.offset, notify.cap.length, &psc->sc_notify_ioh)) {
500 aprint_error_dev(self, "can't map notify i/o space\n");
501 ret = EIO;
502 goto err;
503 }
504 psc->sc_notify_iosize = notify.cap.length;
505 psc->sc_notify_iot = psc->sc_bars_iot[i];
506 psc->sc_notify_off_multiplier = le32toh(notify.notify_off_multiplier);
507
508 if (have_device_cfg) {
509 i = bars_idx[device.bar];
510 if (bus_space_subregion(psc->sc_bars_iot[i],
511 psc->sc_bars_ioh[i], device.offset, device.length,
512 &sc->sc_devcfg_ioh)) {
513 aprint_error_dev(self, "can't map devcfg i/o space\n");
514 ret = EIO;
515 goto err;
516 }
517 aprint_debug_dev(self,
518 "device.offset = 0x%x, device.length = 0x%x\n",
519 device.offset, device.length);
520 sc->sc_devcfg_iosize = device.length;
521 sc->sc_devcfg_iot = psc->sc_bars_iot[i];
522 }
523
524 i = bars_idx[isr.bar];
525 if (bus_space_subregion(psc->sc_bars_iot[i], psc->sc_bars_ioh[i],
526 isr.offset, isr.length, &psc->sc_isr_ioh)) {
527 aprint_error_dev(self, "can't map isr i/o space\n");
528 ret = EIO;
529 goto err;
530 }
531 psc->sc_isr_iosize = isr.length;
532 psc->sc_isr_iot = psc->sc_bars_iot[i];
533
534 i = bars_idx[common.bar];
535 if (bus_space_subregion(psc->sc_bars_iot[i], psc->sc_bars_ioh[i],
536 common.offset, common.length, &psc->sc_ioh)) {
537 aprint_error_dev(self, "can't map common i/o space\n");
538 ret = EIO;
539 goto err;
540 }
541 psc->sc_iosize = common.length;
542 psc->sc_iot = psc->sc_bars_iot[i];
543
544 psc->sc_sc.sc_version_1 = 1;
545
546 /* set our version 1.0 ops */
547 sc->sc_ops = &virtio_pci_ops_10;
548 sc->sc_bus_endian = READ_ENDIAN_10;
549 sc->sc_struct_endian = STRUCT_ENDIAN_10;
550 return 0;
551
552 err:
553 /* undo our pci_mapreg_map()s */
554 for (i = 0; i < __arraycount(bars); i++) {
555 if (psc->sc_bars_iosize[i] == 0)
556 continue;
557 bus_space_unmap(psc->sc_bars_iot[i], psc->sc_bars_ioh[i],
558 psc->sc_bars_iosize[i]);
559 psc->sc_bars_iosize[i] = 0;
560 }
561 return ret;
562 }
563
564 /* v1.0 attach helper */
565 static int
566 virtio_pci_find_cap(struct virtio_pci_softc *psc, int cfg_type, void *buf,
567 int buflen)
568 {
569 device_t self = psc->sc_sc.sc_dev;
570 pci_chipset_tag_t pc = psc->sc_pa.pa_pc;
571 pcitag_t tag = psc->sc_pa.pa_tag;
572 unsigned int offset, i, len;
573 union {
574 pcireg_t reg[8];
575 struct virtio_pci_cap vcap;
576 } *v = buf;
577
578 if (buflen < sizeof(struct virtio_pci_cap))
579 return ERANGE;
580
581 if (!pci_get_capability(pc, tag, PCI_CAP_VENDSPEC, &offset,
582 &v->reg[0]))
583 return ENOENT;
584
585 do {
586 for (i = 0; i < 4; i++)
587 v->reg[i] =
588 le32toh(pci_conf_read(pc, tag, offset + i * 4));
589 if (v->vcap.cfg_type == cfg_type)
590 break;
591 offset = v->vcap.cap_next;
592 } while (offset != 0);
593
594 if (offset == 0)
595 return ENOENT;
596
597 if (v->vcap.cap_len > sizeof(struct virtio_pci_cap)) {
598 len = roundup(v->vcap.cap_len, sizeof(pcireg_t));
599 if (len > buflen) {
600 aprint_error_dev(self, "%s cap too large\n", __func__);
601 return ERANGE;
602 }
603 for (i = 4; i < len / sizeof(pcireg_t); i++)
604 v->reg[i] =
605 le32toh(pci_conf_read(pc, tag, offset + i * 4));
606 }
607
608 /* endian fixup */
609 v->vcap.offset = le32toh(v->vcap.offset);
610 v->vcap.length = le32toh(v->vcap.length);
611 return 0;
612 }
613
614 /* -------------------------------------
615 * Version 0.9 support
616 * -------------------------------------*/
617
618 static void
619 virtio_pci_kick_09(struct virtio_softc *sc, uint16_t idx)
620 {
621 struct virtio_pci_softc * const psc = container_of(sc,
622 struct virtio_pci_softc, sc_sc);
623
624 bus_space_write_2(psc->sc_notify_iot, psc->sc_notify_ioh, 0, idx);
625 }
626
627 /* only applicable for v 0.9 but also called for 1.0 */
628 static int
629 virtio_pci_adjust_config_region(struct virtio_pci_softc *psc)
630 {
631 struct virtio_softc * const sc = &psc->sc_sc;
632 device_t self = sc->sc_dev;
633
634 if (psc->sc_sc.sc_version_1)
635 return 0;
636
637 sc->sc_devcfg_iosize = psc->sc_iosize - psc->sc_devcfg_offset;
638 sc->sc_devcfg_iot = psc->sc_iot;
639 if (bus_space_subregion(psc->sc_iot, psc->sc_ioh,
640 psc->sc_devcfg_offset, sc->sc_devcfg_iosize,
641 &sc->sc_devcfg_ioh)) {
642 aprint_error_dev(self, "can't map config i/o space\n");
643 return EIO;
644 }
645
646 return 0;
647 }
648
649 static uint16_t
650 virtio_pci_read_queue_size_09(struct virtio_softc *sc, uint16_t idx)
651 {
652 struct virtio_pci_softc * const psc = container_of(sc,
653 struct virtio_pci_softc, sc_sc);
654
655 bus_space_write_2(psc->sc_iot, psc->sc_ioh,
656 VIRTIO_CONFIG_QUEUE_SELECT, idx);
657 return bus_space_read_2(psc->sc_iot, psc->sc_ioh,
658 VIRTIO_CONFIG_QUEUE_SIZE);
659 }
660
661 static void
662 virtio_pci_setup_queue_09(struct virtio_softc *sc, uint16_t idx, uint64_t addr)
663 {
664 struct virtio_pci_softc * const psc = container_of(sc,
665 struct virtio_pci_softc, sc_sc);
666
667 bus_space_write_2(psc->sc_iot, psc->sc_ioh,
668 VIRTIO_CONFIG_QUEUE_SELECT, idx);
669 bus_space_write_4(psc->sc_iot, psc->sc_ioh,
670 VIRTIO_CONFIG_QUEUE_ADDRESS, addr / VIRTIO_PAGE_SIZE);
671
672 if (psc->sc_ihs_num > 1) {
673 int vec = VIRTIO_MSIX_QUEUE_VECTOR_INDEX;
674 if (psc->sc_intr_pervq)
675 vec += idx;
676 bus_space_write_2(psc->sc_iot, psc->sc_ioh,
677 VIRTIO_CONFIG_MSI_QUEUE_VECTOR, vec);
678 }
679 }
680
681 static void
682 virtio_pci_set_status_09(struct virtio_softc *sc, int status)
683 {
684 struct virtio_pci_softc * const psc = container_of(sc,
685 struct virtio_pci_softc, sc_sc);
686 int old = 0;
687
688 if (status != 0) {
689 old = bus_space_read_1(psc->sc_iot, psc->sc_ioh,
690 VIRTIO_CONFIG_DEVICE_STATUS);
691 }
692 bus_space_write_1(psc->sc_iot, psc->sc_ioh,
693 VIRTIO_CONFIG_DEVICE_STATUS, status|old);
694 }
695
696 static void
697 virtio_pci_negotiate_features_09(struct virtio_softc *sc,
698 uint64_t guest_features)
699 {
700 struct virtio_pci_softc * const psc = container_of(sc,
701 struct virtio_pci_softc, sc_sc);
702 uint32_t r;
703
704 r = bus_space_read_4(psc->sc_iot, psc->sc_ioh,
705 VIRTIO_CONFIG_DEVICE_FEATURES);
706
707 r &= guest_features;
708
709 bus_space_write_4(psc->sc_iot, psc->sc_ioh,
710 VIRTIO_CONFIG_GUEST_FEATURES, r);
711
712 sc->sc_active_features = r;
713 }
714
715 /* -------------------------------------
716 * Version 1.0 support
717 * -------------------------------------*/
718
719 static void
720 virtio_pci_kick_10(struct virtio_softc *sc, uint16_t idx)
721 {
722 struct virtio_pci_softc * const psc = container_of(sc,
723 struct virtio_pci_softc, sc_sc);
724 unsigned offset = sc->sc_vqs[idx].vq_notify_off *
725 psc->sc_notify_off_multiplier;
726
727 bus_space_write_2(psc->sc_notify_iot, psc->sc_notify_ioh, offset, idx);
728 }
729
730 static uint16_t
731 virtio_pci_read_queue_size_10(struct virtio_softc *sc, uint16_t idx)
732 {
733 struct virtio_pci_softc * const psc = container_of(sc,
734 struct virtio_pci_softc, sc_sc);
735 bus_space_tag_t iot = psc->sc_iot;
736 bus_space_handle_t ioh = psc->sc_ioh;
737
738 bus_space_write_2(iot, ioh, VIRTIO_CONFIG1_QUEUE_SELECT, idx);
739 return bus_space_read_2(iot, ioh, VIRTIO_CONFIG1_QUEUE_SIZE);
740 }
741
742 /*
743 * By definition little endian only in v1.0. NB: "MAY" in the text
744 * below refers to "independently" (i.e. the order of accesses) not
745 * "32-bit" (which is restricted by the earlier "MUST").
746 *
747 * 4.1.3.1 Driver Requirements: PCI Device Layout
748 *
749 * For device configuration access, the driver MUST use ... 32-bit
750 * wide and aligned accesses for ... 64-bit wide fields. For 64-bit
751 * fields, the driver MAY access each of the high and low 32-bit parts
752 * of the field independently.
753 */
754 static __inline void
755 virtio_pci_bus_space_write_8(bus_space_tag_t iot, bus_space_handle_t ioh,
756 bus_size_t offset, uint64_t value)
757 {
758 #if _QUAD_HIGHWORD
759 bus_space_write_4(iot, ioh, offset, BUS_ADDR_LO32(value));
760 bus_space_write_4(iot, ioh, offset + 4, BUS_ADDR_HI32(value));
761 #else
762 bus_space_write_4(iot, ioh, offset, BUS_ADDR_HI32(value));
763 bus_space_write_4(iot, ioh, offset + 4, BUS_ADDR_LO32(value));
764 #endif
765 }
766
767 static void
768 virtio_pci_setup_queue_10(struct virtio_softc *sc, uint16_t idx, uint64_t addr)
769 {
770 struct virtio_pci_softc * const psc = container_of(sc,
771 struct virtio_pci_softc, sc_sc);
772 struct virtqueue *vq = &sc->sc_vqs[idx];
773 bus_space_tag_t iot = psc->sc_iot;
774 bus_space_handle_t ioh = psc->sc_ioh;
775 KASSERT(vq->vq_index == idx);
776
777 bus_space_write_2(iot, ioh, VIRTIO_CONFIG1_QUEUE_SELECT, vq->vq_index);
778 if (addr == 0) {
779 bus_space_write_2(iot, ioh, VIRTIO_CONFIG1_QUEUE_ENABLE, 0);
780 virtio_pci_bus_space_write_8(iot, ioh,
781 VIRTIO_CONFIG1_QUEUE_DESC, 0);
782 virtio_pci_bus_space_write_8(iot, ioh,
783 VIRTIO_CONFIG1_QUEUE_AVAIL, 0);
784 virtio_pci_bus_space_write_8(iot, ioh,
785 VIRTIO_CONFIG1_QUEUE_USED, 0);
786 } else {
787 virtio_pci_bus_space_write_8(iot, ioh,
788 VIRTIO_CONFIG1_QUEUE_DESC, addr);
789 virtio_pci_bus_space_write_8(iot, ioh,
790 VIRTIO_CONFIG1_QUEUE_AVAIL, addr + vq->vq_availoffset);
791 virtio_pci_bus_space_write_8(iot, ioh,
792 VIRTIO_CONFIG1_QUEUE_USED, addr + vq->vq_usedoffset);
793 bus_space_write_2(iot, ioh,
794 VIRTIO_CONFIG1_QUEUE_ENABLE, 1);
795 vq->vq_notify_off = bus_space_read_2(iot, ioh,
796 VIRTIO_CONFIG1_QUEUE_NOTIFY_OFF);
797 }
798
799 if (psc->sc_ihs_num > 1) {
800 int vec = VIRTIO_MSIX_QUEUE_VECTOR_INDEX;
801 if (psc->sc_intr_pervq)
802 vec += idx;
803 bus_space_write_2(iot, ioh,
804 VIRTIO_CONFIG1_QUEUE_MSIX_VECTOR, vec);
805 }
806 }
807
808 static void
809 virtio_pci_set_status_10(struct virtio_softc *sc, int status)
810 {
811 struct virtio_pci_softc * const psc = container_of(sc,
812 struct virtio_pci_softc, sc_sc);
813 bus_space_tag_t iot = psc->sc_iot;
814 bus_space_handle_t ioh = psc->sc_ioh;
815 int old = 0;
816
817 if (status)
818 old = bus_space_read_1(iot, ioh, VIRTIO_CONFIG1_DEVICE_STATUS);
819 bus_space_write_1(iot, ioh, VIRTIO_CONFIG1_DEVICE_STATUS,
820 status | old);
821 }
822
823 void
824 virtio_pci_negotiate_features_10(struct virtio_softc *sc,
825 uint64_t guest_features)
826 {
827 struct virtio_pci_softc * const psc = container_of(sc,
828 struct virtio_pci_softc, sc_sc);
829 device_t self = sc->sc_dev;
830 bus_space_tag_t iot = psc->sc_iot;
831 bus_space_handle_t ioh = psc->sc_ioh;
832 uint64_t host, negotiated, device_status;
833
834 guest_features |= VIRTIO_F_VERSION_1;
835 #ifdef __NEED_VIRTIO_F_ACCESS_PLATFORM
836 /* XXX This could use some work. */
837 guest_features |= VIRTIO_F_ACCESS_PLATFORM;
838 #endif /* __NEED_VIRTIO_F_ACCESS_PLATFORM */
839 /* notify on empty is 0.9 only */
840 guest_features &= ~VIRTIO_F_NOTIFY_ON_EMPTY;
841 sc->sc_active_features = 0;
842
843 bus_space_write_4(iot, ioh, VIRTIO_CONFIG1_DEVICE_FEATURE_SELECT, 0);
844 host = bus_space_read_4(iot, ioh, VIRTIO_CONFIG1_DEVICE_FEATURE);
845 bus_space_write_4(iot, ioh, VIRTIO_CONFIG1_DEVICE_FEATURE_SELECT, 1);
846 host |= (uint64_t)bus_space_read_4(iot, ioh,
847 VIRTIO_CONFIG1_DEVICE_FEATURE) << 32;
848
849 negotiated = host & guest_features;
850
851 bus_space_write_4(iot, ioh, VIRTIO_CONFIG1_DRIVER_FEATURE_SELECT, 0);
852 bus_space_write_4(iot, ioh, VIRTIO_CONFIG1_DRIVER_FEATURE,
853 negotiated & 0xffffffff);
854 bus_space_write_4(iot, ioh, VIRTIO_CONFIG1_DRIVER_FEATURE_SELECT, 1);
855 bus_space_write_4(iot, ioh, VIRTIO_CONFIG1_DRIVER_FEATURE,
856 negotiated >> 32);
857 virtio_pci_set_status_10(sc, VIRTIO_CONFIG_DEVICE_STATUS_FEATURES_OK);
858
859 device_status = bus_space_read_1(iot, ioh,
860 VIRTIO_CONFIG1_DEVICE_STATUS);
861 if ((device_status & VIRTIO_CONFIG_DEVICE_STATUS_FEATURES_OK) == 0) {
862 aprint_error_dev(self, "feature negotiation failed\n");
863 bus_space_write_1(iot, ioh, VIRTIO_CONFIG1_DEVICE_STATUS,
864 VIRTIO_CONFIG_DEVICE_STATUS_FAILED);
865 return;
866 }
867
868 if ((negotiated & VIRTIO_F_VERSION_1) == 0) {
869 aprint_error_dev(self, "host rejected version 1\n");
870 bus_space_write_1(iot, ioh, VIRTIO_CONFIG1_DEVICE_STATUS,
871 VIRTIO_CONFIG_DEVICE_STATUS_FAILED);
872 return;
873 }
874
875 sc->sc_active_features = negotiated;
876 return;
877 }
878
879 /* -------------------------------------
880 * Generic PCI interrupt code
881 * -------------------------------------*/
882
883 static int
884 virtio_pci_setup_interrupts_10(struct virtio_softc *sc, int reinit)
885 {
886 struct virtio_pci_softc * const psc = container_of(sc,
887 struct virtio_pci_softc, sc_sc);
888 bus_space_tag_t iot = psc->sc_iot;
889 bus_space_handle_t ioh = psc->sc_ioh;
890 int vector, ret, qid;
891
892 if (!virtio_pci_msix_enabled(psc))
893 return 0;
894
895 vector = VIRTIO_MSIX_CONFIG_VECTOR_INDEX;
896 bus_space_write_2(iot, ioh, VIRTIO_CONFIG1_CONFIG_MSIX_VECTOR, vector);
897 ret = bus_space_read_2(iot, ioh, VIRTIO_CONFIG1_CONFIG_MSIX_VECTOR);
898 if (ret != vector) {
899 VIRTIO_PCI_LOG(sc, reinit, "can't set config msix vector\n");
900 return -1;
901 }
902
903 for (qid = 0; qid < sc->sc_nvqs; qid++) {
904 vector = VIRTIO_MSIX_QUEUE_VECTOR_INDEX;
905
906 if (psc->sc_intr_pervq)
907 vector += qid;
908 bus_space_write_2(iot, ioh, VIRTIO_CONFIG1_QUEUE_SELECT, qid);
909 bus_space_write_2(iot, ioh, VIRTIO_CONFIG1_QUEUE_MSIX_VECTOR,
910 vector);
911 ret = bus_space_read_2(iot, ioh,
912 VIRTIO_CONFIG1_QUEUE_MSIX_VECTOR);
913 if (ret != vector) {
914 VIRTIO_PCI_LOG(sc, reinit, "can't set queue %d "
915 "msix vector\n", qid);
916 return -1;
917 }
918 }
919
920 return 0;
921 }
922
923 static int
924 virtio_pci_setup_interrupts_09(struct virtio_softc *sc, int reinit)
925 {
926 struct virtio_pci_softc * const psc = container_of(sc,
927 struct virtio_pci_softc, sc_sc);
928 int offset, vector, ret, qid;
929
930 if (!virtio_pci_msix_enabled(psc))
931 return 0;
932
933 offset = VIRTIO_CONFIG_MSI_CONFIG_VECTOR;
934 vector = VIRTIO_MSIX_CONFIG_VECTOR_INDEX;
935
936 bus_space_write_2(psc->sc_iot, psc->sc_ioh, offset, vector);
937 ret = bus_space_read_2(psc->sc_iot, psc->sc_ioh, offset);
938 if (ret != vector) {
939 aprint_debug_dev(sc->sc_dev, "%s: expected=%d, actual=%d\n",
940 __func__, vector, ret);
941 VIRTIO_PCI_LOG(sc, reinit,
942 "can't set config msix vector\n");
943 return -1;
944 }
945
946 for (qid = 0; qid < sc->sc_nvqs; qid++) {
947 offset = VIRTIO_CONFIG_QUEUE_SELECT;
948 bus_space_write_2(psc->sc_iot, psc->sc_ioh, offset, qid);
949
950 offset = VIRTIO_CONFIG_MSI_QUEUE_VECTOR;
951 vector = VIRTIO_MSIX_QUEUE_VECTOR_INDEX;
952
953 if (psc->sc_intr_pervq)
954 vector += qid;
955
956 bus_space_write_2(psc->sc_iot, psc->sc_ioh, offset, vector);
957 ret = bus_space_read_2(psc->sc_iot, psc->sc_ioh, offset);
958 if (ret != vector) {
959 aprint_debug_dev(sc->sc_dev, "%s[qid=%d]:"
960 " expected=%d, actual=%d\n",
961 __func__, qid, vector, ret);
962 VIRTIO_PCI_LOG(sc, reinit, "can't set queue %d "
963 "msix vector\n", qid);
964 return -1;
965 }
966 }
967
968 return 0;
969 }
970
971 static int
972 virtio_pci_establish_msix_interrupts(struct virtio_softc *sc,
973 const struct pci_attach_args *pa)
974 {
975 struct virtio_pci_softc * const psc = container_of(sc,
976 struct virtio_pci_softc, sc_sc);
977 device_t self = sc->sc_dev;
978 pci_chipset_tag_t pc = pa->pa_pc;
979 struct virtqueue *vq;
980 char intrbuf[PCI_INTRSTR_LEN];
981 char intr_xname[INTRDEVNAMEBUF];
982 char const *intrstr;
983 int idx, qid, n;
984
985 idx = VIRTIO_MSIX_CONFIG_VECTOR_INDEX;
986 if (sc->sc_flags & VIRTIO_F_INTR_MPSAFE)
987 pci_intr_setattr(pc, &psc->sc_ihp[idx], PCI_INTR_MPSAFE, true);
988
989 snprintf(intr_xname, sizeof(intr_xname), "%s config",
990 device_xname(sc->sc_dev));
991
992 psc->sc_ihs[idx] = pci_intr_establish_xname(pc, psc->sc_ihp[idx],
993 sc->sc_ipl, virtio_pci_msix_config_intr, sc, intr_xname);
994 if (psc->sc_ihs[idx] == NULL) {
995 aprint_error_dev(self,
996 "couldn't establish MSI-X for config\n");
997 goto error;
998 }
999
1000 idx = VIRTIO_MSIX_QUEUE_VECTOR_INDEX;
1001 if (psc->sc_intr_pervq) {
1002 for (qid = 0; qid < sc->sc_nvqs; qid++) {
1003 n = idx + qid;
1004 vq = &sc->sc_vqs[qid];
1005
1006 snprintf(intr_xname, sizeof(intr_xname), "%s vq#%d",
1007 device_xname(sc->sc_dev), qid);
1008
1009 if (sc->sc_flags & VIRTIO_F_INTR_MPSAFE) {
1010 pci_intr_setattr(pc, &psc->sc_ihp[n],
1011 PCI_INTR_MPSAFE, true);
1012 }
1013
1014 psc->sc_ihs[n] = pci_intr_establish_xname(pc,
1015 psc->sc_ihp[n], sc->sc_ipl,
1016 vq->vq_intrhand, vq->vq_intrhand_arg, intr_xname);
1017 if (psc->sc_ihs[n] == NULL) {
1018 aprint_error_dev(self,
1019 "couldn't establish MSI-X for a vq\n");
1020 goto error;
1021 }
1022 }
1023 } else {
1024 if (sc->sc_flags & VIRTIO_F_INTR_MPSAFE) {
1025 pci_intr_setattr(pc, &psc->sc_ihp[idx],
1026 PCI_INTR_MPSAFE, true);
1027 }
1028
1029 snprintf(intr_xname, sizeof(intr_xname), "%s queues",
1030 device_xname(sc->sc_dev));
1031 psc->sc_ihs[idx] = pci_intr_establish_xname(pc,
1032 psc->sc_ihp[idx], sc->sc_ipl,
1033 virtio_pci_msix_queue_intr, sc, intr_xname);
1034 if (psc->sc_ihs[idx] == NULL) {
1035 aprint_error_dev(self,
1036 "couldn't establish MSI-X for queues\n");
1037 goto error;
1038 }
1039 }
1040
1041 idx = VIRTIO_MSIX_CONFIG_VECTOR_INDEX;
1042 intrstr = pci_intr_string(pc, psc->sc_ihp[idx], intrbuf,
1043 sizeof(intrbuf));
1044 aprint_normal_dev(self, "config interrupting at %s\n", intrstr);
1045 idx = VIRTIO_MSIX_QUEUE_VECTOR_INDEX;
1046 if (psc->sc_intr_pervq) {
1047 kcpuset_t *affinity;
1048 int affinity_to, r;
1049
1050 kcpuset_create(&affinity, false);
1051
1052 for (qid = 0; qid < sc->sc_nvqs; qid++) {
1053 n = idx + qid;
1054 affinity_to = (qid / 2) % ncpu;
1055
1056 intrstr = pci_intr_string(pc, psc->sc_ihp[n],
1057 intrbuf, sizeof(intrbuf));
1058
1059 kcpuset_zero(affinity);
1060 kcpuset_set(affinity, affinity_to);
1061 r = interrupt_distribute(psc->sc_ihs[n], affinity,
1062 NULL);
1063 if (r == 0) {
1064 aprint_normal_dev(self,
1065 "for vq #%d interrupting at %s"
1066 " affinity to %u\n",
1067 qid, intrstr, affinity_to);
1068 } else {
1069 aprint_normal_dev(self,
1070 "for vq #%d interrupting at %s\n",
1071 qid, intrstr);
1072 }
1073 }
1074
1075 kcpuset_destroy(affinity);
1076 } else {
1077 intrstr = pci_intr_string(pc, psc->sc_ihp[idx], intrbuf,
1078 sizeof(intrbuf));
1079 aprint_normal_dev(self, "queues interrupting at %s\n",
1080 intrstr);
1081 }
1082
1083 return 0;
1084
1085 error:
1086 idx = VIRTIO_MSIX_CONFIG_VECTOR_INDEX;
1087 if (psc->sc_ihs[idx] != NULL)
1088 pci_intr_disestablish(psc->sc_pa.pa_pc, psc->sc_ihs[idx]);
1089 idx = VIRTIO_MSIX_QUEUE_VECTOR_INDEX;
1090 if (psc->sc_intr_pervq) {
1091 for (qid = 0; qid < sc->sc_nvqs; qid++) {
1092 n = idx + qid;
1093 if (psc->sc_ihs[n] == NULL)
1094 continue;
1095 pci_intr_disestablish(psc->sc_pa.pa_pc,
1096 psc->sc_ihs[n]);
1097 }
1098
1099 } else {
1100 if (psc->sc_ihs[idx] != NULL) {
1101 pci_intr_disestablish(psc->sc_pa.pa_pc,
1102 psc->sc_ihs[idx]);
1103 }
1104 }
1105
1106 return -1;
1107 }
1108
1109 static int
1110 virtio_pci_establish_intx_interrupt(struct virtio_softc *sc,
1111 const struct pci_attach_args *pa)
1112 {
1113 struct virtio_pci_softc * const psc = container_of(sc,
1114 struct virtio_pci_softc, sc_sc);
1115 device_t self = sc->sc_dev;
1116 pci_chipset_tag_t pc = pa->pa_pc;
1117 char intrbuf[PCI_INTRSTR_LEN];
1118 char const *intrstr;
1119
1120 if (sc->sc_flags & VIRTIO_F_INTR_MPSAFE)
1121 pci_intr_setattr(pc, &psc->sc_ihp[0], PCI_INTR_MPSAFE, true);
1122
1123 psc->sc_ihs[0] = pci_intr_establish_xname(pc, psc->sc_ihp[0],
1124 sc->sc_ipl, virtio_pci_intr, sc, device_xname(sc->sc_dev));
1125 if (psc->sc_ihs[0] == NULL) {
1126 aprint_error_dev(self, "couldn't establish INTx\n");
1127 return -1;
1128 }
1129
1130 intrstr = pci_intr_string(pc, psc->sc_ihp[0], intrbuf,
1131 sizeof(intrbuf));
1132 aprint_normal_dev(self, "interrupting at %s\n", intrstr);
1133
1134 return 0;
1135 }
1136
1137 static int
1138 virtio_pci_alloc_interrupts(struct virtio_softc *sc)
1139 {
1140 struct virtio_pci_softc * const psc = container_of(sc,
1141 struct virtio_pci_softc, sc_sc);
1142 device_t self = sc->sc_dev;
1143 pci_chipset_tag_t pc = psc->sc_pa.pa_pc;
1144 pcitag_t tag = psc->sc_pa.pa_tag;
1145 int error;
1146 int nmsix;
1147 int off;
1148 int counts[PCI_INTR_TYPE_SIZE];
1149 pci_intr_type_t max_type;
1150 pcireg_t ctl;
1151
1152 nmsix = pci_msix_count(psc->sc_pa.pa_pc, psc->sc_pa.pa_tag);
1153 aprint_debug_dev(self, "pci_msix_count=%d\n", nmsix);
1154
1155 /* We need at least two: one for config and the other for queues */
1156 if ((sc->sc_flags & VIRTIO_F_INTR_MSIX) == 0 || nmsix < 2) {
1157 /* Try INTx only */
1158 max_type = PCI_INTR_TYPE_INTX;
1159 counts[PCI_INTR_TYPE_INTX] = 1;
1160 } else {
1161 /* Try MSI-X first and INTx second */
1162 if (ISSET(sc->sc_flags, VIRTIO_F_INTR_PERVQ) &&
1163 sc->sc_nvqs + VIRTIO_MSIX_QUEUE_VECTOR_INDEX <= nmsix) {
1164 nmsix = sc->sc_nvqs + VIRTIO_MSIX_QUEUE_VECTOR_INDEX;
1165 } else {
1166 nmsix = 2;
1167 }
1168
1169 max_type = PCI_INTR_TYPE_MSIX;
1170 counts[PCI_INTR_TYPE_MSIX] = nmsix;
1171 counts[PCI_INTR_TYPE_MSI] = 0;
1172 counts[PCI_INTR_TYPE_INTX] = 1;
1173 }
1174
1175 retry:
1176 error = pci_intr_alloc(&psc->sc_pa, &psc->sc_ihp, counts, max_type);
1177 if (error != 0) {
1178 aprint_error_dev(self, "couldn't map interrupt\n");
1179 return -1;
1180 }
1181
1182 if (pci_intr_type(pc, psc->sc_ihp[0]) == PCI_INTR_TYPE_MSIX) {
1183 psc->sc_intr_pervq = nmsix > 2 ? true : false;
1184 psc->sc_ihs = kmem_zalloc(sizeof(*psc->sc_ihs) * nmsix,
1185 KM_SLEEP);
1186
1187 error = virtio_pci_establish_msix_interrupts(sc, &psc->sc_pa);
1188 if (error != 0) {
1189 kmem_free(psc->sc_ihs, sizeof(*psc->sc_ihs) * nmsix);
1190 pci_intr_release(pc, psc->sc_ihp, nmsix);
1191
1192 /* Retry INTx */
1193 max_type = PCI_INTR_TYPE_INTX;
1194 counts[PCI_INTR_TYPE_INTX] = 1;
1195 goto retry;
1196 }
1197
1198 psc->sc_ihs_num = nmsix;
1199 psc->sc_devcfg_offset = VIRTIO_CONFIG_DEVICE_CONFIG_MSI;
1200 virtio_pci_adjust_config_region(psc);
1201 } else if (pci_intr_type(pc, psc->sc_ihp[0]) == PCI_INTR_TYPE_INTX) {
1202 psc->sc_intr_pervq = false;
1203 psc->sc_ihs = kmem_zalloc(sizeof(*psc->sc_ihs) * 1,
1204 KM_SLEEP);
1205
1206 error = virtio_pci_establish_intx_interrupt(sc, &psc->sc_pa);
1207 if (error != 0) {
1208 kmem_free(psc->sc_ihs, sizeof(*psc->sc_ihs) * 1);
1209 pci_intr_release(pc, psc->sc_ihp, 1);
1210 return -1;
1211 }
1212
1213 psc->sc_ihs_num = 1;
1214 psc->sc_devcfg_offset = VIRTIO_CONFIG_DEVICE_CONFIG_NOMSI;
1215 virtio_pci_adjust_config_region(psc);
1216
1217 error = pci_get_capability(pc, tag, PCI_CAP_MSIX, &off, NULL);
1218 if (error != 0) {
1219 ctl = pci_conf_read(pc, tag, off + PCI_MSIX_CTL);
1220 ctl &= ~PCI_MSIX_CTL_ENABLE;
1221 pci_conf_write(pc, tag, off + PCI_MSIX_CTL, ctl);
1222 }
1223 }
1224
1225 if (!psc->sc_intr_pervq)
1226 CLR(sc->sc_flags, VIRTIO_F_INTR_PERVQ);
1227 return 0;
1228 }
1229
1230 static void
1231 virtio_pci_free_interrupts(struct virtio_softc *sc)
1232 {
1233 struct virtio_pci_softc * const psc = container_of(sc,
1234 struct virtio_pci_softc, sc_sc);
1235
1236 for (int i = 0; i < psc->sc_ihs_num; i++) {
1237 if (psc->sc_ihs[i] == NULL)
1238 continue;
1239 pci_intr_disestablish(psc->sc_pa.pa_pc, psc->sc_ihs[i]);
1240 psc->sc_ihs[i] = NULL;
1241 }
1242
1243 if (psc->sc_ihs_num > 0) {
1244 pci_intr_release(psc->sc_pa.pa_pc, psc->sc_ihp,
1245 psc->sc_ihs_num);
1246 }
1247
1248 if (psc->sc_ihs != NULL) {
1249 kmem_free(psc->sc_ihs, sizeof(*psc->sc_ihs) * psc->sc_ihs_num);
1250 psc->sc_ihs = NULL;
1251 }
1252 psc->sc_ihs_num = 0;
1253 }
1254
1255 static bool
1256 virtio_pci_msix_enabled(struct virtio_pci_softc *psc)
1257 {
1258 pci_chipset_tag_t pc = psc->sc_pa.pa_pc;
1259
1260 if (pci_intr_type(pc, psc->sc_ihp[0]) == PCI_INTR_TYPE_MSIX)
1261 return true;
1262
1263 return false;
1264 }
1265
1266 /*
1267 * Interrupt handler.
1268 */
1269 static int
1270 virtio_pci_intr(void *arg)
1271 {
1272 struct virtio_softc *sc = arg;
1273 struct virtio_pci_softc * const psc = container_of(sc,
1274 struct virtio_pci_softc, sc_sc);
1275 int isr, r = 0;
1276
1277 /* check and ack the interrupt */
1278 isr = bus_space_read_1(psc->sc_isr_iot, psc->sc_isr_ioh, 0);
1279 if (isr == 0)
1280 return 0;
1281 if ((isr & VIRTIO_CONFIG_ISR_CONFIG_CHANGE) &&
1282 (sc->sc_config_change != NULL))
1283 r = (sc->sc_config_change)(sc);
1284 if (sc->sc_intrhand != NULL) {
1285 if (sc->sc_soft_ih != NULL)
1286 softint_schedule(sc->sc_soft_ih);
1287 else
1288 r |= (sc->sc_intrhand)(sc);
1289 }
1290
1291 return r;
1292 }
1293
1294 static int
1295 virtio_pci_msix_queue_intr(void *arg)
1296 {
1297 struct virtio_softc *sc = arg;
1298 int r = 0;
1299
1300 if (sc->sc_intrhand != NULL) {
1301 if (sc->sc_soft_ih != NULL)
1302 softint_schedule(sc->sc_soft_ih);
1303 else
1304 r |= (sc->sc_intrhand)(sc);
1305 }
1306
1307 return r;
1308 }
1309
1310 static int
1311 virtio_pci_msix_config_intr(void *arg)
1312 {
1313 struct virtio_softc *sc = arg;
1314 int r = 0;
1315
1316 if (sc->sc_config_change != NULL)
1317 r = (sc->sc_config_change)(sc);
1318 return r;
1319 }
1320
1321 MODULE(MODULE_CLASS_DRIVER, virtio_pci, "pci,virtio");
1322
1323 #ifdef _MODULE
1324 #include "ioconf.c"
1325 #endif
1326
1327 static int
1328 virtio_pci_modcmd(modcmd_t cmd, void *opaque)
1329 {
1330 int error = 0;
1331
1332 #ifdef _MODULE
1333 switch (cmd) {
1334 case MODULE_CMD_INIT:
1335 error = config_init_component(cfdriver_ioconf_virtio_pci,
1336 cfattach_ioconf_virtio_pci, cfdata_ioconf_virtio_pci);
1337 break;
1338 case MODULE_CMD_FINI:
1339 error = config_fini_component(cfdriver_ioconf_virtio_pci,
1340 cfattach_ioconf_virtio_pci, cfdata_ioconf_virtio_pci);
1341 break;
1342 default:
1343 error = ENOTTY;
1344 break;
1345 }
1346 #endif
1347
1348 return error;
1349 }
1350