gicv3_its.c revision 1.38 1 /* $NetBSD: gicv3_its.c,v 1.38 2024/12/10 21:33:53 jmcneill Exp $ */
2
3 /*-
4 * Copyright (c) 2018 The NetBSD Foundation, Inc.
5 * All rights reserved.
6 *
7 * This code is derived from software contributed to The NetBSD Foundation
8 * by Jared McNeill <jmcneill (at) invisible.ca>.
9 *
10 * Redistribution and use in source and binary forms, with or without
11 * modification, are permitted provided that the following conditions
12 * are met:
13 * 1. Redistributions of source code must retain the above copyright
14 * notice, this list of conditions and the following disclaimer.
15 * 2. Redistributions in binary form must reproduce the above copyright
16 * notice, this list of conditions and the following disclaimer in the
17 * documentation and/or other materials provided with the distribution.
18 *
19 * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
20 * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
21 * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
22 * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
23 * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
24 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
25 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
26 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
27 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
28 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
29 * POSSIBILITY OF SUCH DAMAGE.
30 */
31
32 #define _INTR_PRIVATE
33
34 #include <sys/cdefs.h>
35 __KERNEL_RCSID(0, "$NetBSD: gicv3_its.c,v 1.38 2024/12/10 21:33:53 jmcneill Exp $");
36
37 #include <sys/param.h>
38 #include <sys/kmem.h>
39 #include <sys/bus.h>
40 #include <sys/cpu.h>
41 #include <sys/bitops.h>
42
43 #include <uvm/uvm.h>
44
45 #include <dev/pci/pcireg.h>
46 #include <dev/pci/pcivar.h>
47
48 #include <machine/cpufunc.h>
49
50 #include <arm/pic/picvar.h>
51 #include <arm/cortex/gicv3_its.h>
52
53 #ifdef ITS_DEBUG
54 #define DPRINTF(x) printf x
55 #else
56 #define DPRINTF(x) __nothing
57 #endif
58
59 /*
60 * ITS translation table sizes
61 */
62 #define GITS_COMMANDS_SIZE 0x1000
63 #define GITS_COMMANDS_ALIGN 0x10000
64
65 #define GITS_ITT_ALIGN 0x100
66
67 #define GITS_INDIRECT_ENTRY_SIZE 8
68
69 /*
70 * IIDR values used for errata
71 */
72 #define GITS_IIDR_PID_CAVIUM_THUNDERX 0xa1
73 #define GITS_IIDR_IMP_CAVIUM 0x34c
74 #define GITS_IIDR_CAVIUM_ERRATA_MASK (GITS_IIDR_Implementor|GITS_IIDR_ProductID|GITS_IIDR_Variant)
75 #define GITS_IIDR_CAVIUM_ERRATA_VALUE \
76 (__SHIFTIN(GITS_IIDR_IMP_CAVIUM, GITS_IIDR_Implementor) | \
77 __SHIFTIN(GITS_IIDR_PID_CAVIUM_THUNDERX, GITS_IIDR_ProductID) | \
78 __SHIFTIN(0, GITS_IIDR_Variant))
79
80 static const char * gits_cache_type[] = {
81 [GITS_Cache_DEVICE_nGnRnE] = "Device-nGnRnE",
82 [GITS_Cache_NORMAL_NC] = "Non-cacheable",
83 [GITS_Cache_NORMAL_RA_WT] = "Cacheable RA WT",
84 [GITS_Cache_NORMAL_RA_WB] = "Cacheable RA WB",
85 [GITS_Cache_NORMAL_WA_WT] = "Cacheable WA WT",
86 [GITS_Cache_NORMAL_WA_WB] = "Cacheable WA WB",
87 [GITS_Cache_NORMAL_RA_WA_WT] = "Cacheable RA WA WT",
88 [GITS_Cache_NORMAL_RA_WA_WB] = "Cacheable RA WA WB",
89 };
90
91 static const char * gits_share_type[] = {
92 [GITS_Shareability_NS] = "Non-shareable",
93 [GITS_Shareability_IS] = "Inner shareable",
94 [GITS_Shareability_OS] = "Outer shareable",
95 [3] = "(Reserved)",
96 };
97
98 static inline uint32_t
99 gits_read_4(struct gicv3_its *its, bus_size_t reg)
100 {
101 return bus_space_read_4(its->its_bst, its->its_bsh, reg);
102 }
103
104 static inline void
105 gits_write_4(struct gicv3_its *its, bus_size_t reg, uint32_t val)
106 {
107 bus_space_write_4(its->its_bst, its->its_bsh, reg, val);
108 }
109
110 static inline uint64_t
111 gits_read_8(struct gicv3_its *its, bus_size_t reg)
112 {
113 return bus_space_read_8(its->its_bst, its->its_bsh, reg);
114 }
115
116 static inline void
117 gits_write_8(struct gicv3_its *its, bus_size_t reg, uint64_t val)
118 {
119 bus_space_write_8(its->its_bst, its->its_bsh, reg, val);
120 }
121
122 static int
123 gits_command(struct gicv3_its *its, const struct gicv3_its_command *cmd)
124 {
125 uint64_t cwriter, creadr;
126 u_int woff;
127
128 creadr = gits_read_8(its, GITS_CREADR);
129 if (ISSET(creadr, GITS_CREADR_Stalled)) {
130 DPRINTF(("ITS: stalled! GITS_CREADR = 0x%lx\n", creadr));
131 return EIO;
132 }
133
134 cwriter = gits_read_8(its, GITS_CWRITER);
135 woff = cwriter & GITS_CWRITER_Offset;
136
137 uint64_t *dw = (uint64_t *)(its->its_cmd.base + woff);
138 for (int i = 0; i < __arraycount(cmd->dw); i++) {
139 dw[i] = htole64(cmd->dw[i]);
140 DPRINTF(("ITS: dw[%u] = 0x%016lx\n", i, cmd->dw[i]));
141 }
142
143 if (its->its_cmd_flush) {
144 cpu_dcache_wb_range((vaddr_t)dw, sizeof(cmd->dw));
145 }
146 dsb(sy);
147
148 woff += sizeof(cmd->dw);
149 if (woff == its->its_cmd.len)
150 woff = 0;
151
152 gits_write_8(its, GITS_CWRITER, woff);
153
154 return 0;
155 }
156
157 static int
158 gits_command_mapc(struct gicv3_its *its, uint16_t icid, uint64_t rdbase, bool v)
159 {
160 struct gicv3_its_command cmd;
161
162 KASSERT((rdbase & 0xffff) == 0);
163
164 /*
165 * Map a collection table entry (ICID) to the target redistributor (RDbase).
166 */
167 memset(&cmd, 0, sizeof(cmd));
168 cmd.dw[0] = GITS_CMD_MAPC;
169 cmd.dw[2] = icid;
170 if (v) {
171 cmd.dw[2] |= rdbase;
172 cmd.dw[2] |= __BIT(63);
173 }
174
175 DPRINTF(("ITS #%u: MAPC icid 0x%x rdbase 0x%lx valid %u\n",
176 its->its_id, icid, rdbase, v));
177
178 return gits_command(its, &cmd);
179 }
180
181 static int
182 gits_command_mapd(struct gicv3_its *its, uint32_t deviceid, uint64_t itt_addr, u_int size, bool v)
183 {
184 struct gicv3_its_command cmd;
185
186 KASSERT((itt_addr & 0xff) == 0);
187
188 /*
189 * Map a device table entry (DeviceID) to its associated ITT (ITT_addr).
190 */
191 memset(&cmd, 0, sizeof(cmd));
192 cmd.dw[0] = GITS_CMD_MAPD | ((uint64_t)deviceid << 32);
193 if (v) {
194 cmd.dw[1] = uimax(1, size) - 1;
195 cmd.dw[2] = itt_addr | __BIT(63);
196 }
197
198 DPRINTF(("ITS #%u: MAPD deviceid 0x%x itt_addr 0x%lx size %u valid %u\n",
199 its->its_id, deviceid, itt_addr, size, v));
200
201 return gits_command(its, &cmd);
202 }
203
204 static int
205 gits_command_mapti(struct gicv3_its *its, uint32_t deviceid, uint32_t eventid, uint32_t pintid, uint16_t icid)
206 {
207 struct gicv3_its_command cmd;
208
209 /*
210 * Map the event defined by EventID and DeviceID to its associated ITE, defined by ICID and pINTID
211 * in the ITT associated with DeviceID.
212 */
213 memset(&cmd, 0, sizeof(cmd));
214 cmd.dw[0] = GITS_CMD_MAPTI | ((uint64_t)deviceid << 32);
215 cmd.dw[1] = eventid | ((uint64_t)pintid << 32);
216 cmd.dw[2] = icid;
217
218 DPRINTF(("ITS #%u: MAPTI deviceid 0x%x eventid 0x%x pintid 0x%x icid 0x%x\n",
219 its->its_id, deviceid, eventid, pintid, icid));
220
221 return gits_command(its, &cmd);
222 }
223
224 static int
225 gits_command_movi(struct gicv3_its *its, uint32_t deviceid, uint32_t eventid, uint16_t icid)
226 {
227 struct gicv3_its_command cmd;
228
229 /*
230 * Update the ICID field in the ITT entry for the event defined by DeviceID and
231 * EventID.
232 */
233 memset(&cmd, 0, sizeof(cmd));
234 cmd.dw[0] = GITS_CMD_MOVI | ((uint64_t)deviceid << 32);
235 cmd.dw[1] = eventid;
236 cmd.dw[2] = icid;
237
238 DPRINTF(("ITS #%u: MOVI deviceid 0x%x eventid 0x%x icid 0x%x\n",
239 its->its_id, deviceid, eventid, icid));
240
241 return gits_command(its, &cmd);
242 }
243
244 static int
245 gits_command_inv(struct gicv3_its *its, uint32_t deviceid, uint32_t eventid)
246 {
247 struct gicv3_its_command cmd;
248
249 /*
250 * Ensure any caching in the redistributors associated with the specified
251 * EventID is consistent with the LPI configuration tables.
252 */
253 memset(&cmd, 0, sizeof(cmd));
254 cmd.dw[0] = GITS_CMD_INV | ((uint64_t)deviceid << 32);
255 cmd.dw[1] = eventid;
256
257 DPRINTF(("ITS #%u: INV deviceid 0x%x eventid 0x%x\n",
258 its->its_id, deviceid, eventid));
259
260 return gits_command(its, &cmd);
261 }
262
263 static int
264 gits_command_invall(struct gicv3_its *its, uint16_t icid)
265 {
266 struct gicv3_its_command cmd;
267
268 /*
269 * Ensure any caching associated with this ICID is consistent with LPI
270 * configuration tables for all redistributors.
271 */
272 memset(&cmd, 0, sizeof(cmd));
273 cmd.dw[0] = GITS_CMD_INVALL;
274 cmd.dw[2] = icid;
275
276 DPRINTF(("ITS #%u: INVALL icid 0x%x\n", its->its_id, icid));
277
278 return gits_command(its, &cmd);
279 }
280
281 static int
282 gits_command_sync(struct gicv3_its *its, uint64_t rdbase)
283 {
284 struct gicv3_its_command cmd;
285
286 KASSERT((rdbase & 0xffff) == 0);
287
288 /*
289 * Ensure all outstanding ITS operations associated with physical interrupts
290 * for the specified redistributor (RDbase) are globally observed before
291 * further ITS commands are executed.
292 */
293 memset(&cmd, 0, sizeof(cmd));
294 cmd.dw[0] = GITS_CMD_SYNC;
295 cmd.dw[2] = rdbase;
296
297 DPRINTF(("ITS #%u: SYNC rdbase 0x%lx\n", its->its_id, rdbase));
298
299 return gits_command(its, &cmd);
300 }
301
302 #if 0
303 static int
304 gits_command_int(struct gicv3_its *its, uint32_t deviceid, uint32_t eventid)
305 {
306 struct gicv3_its_command cmd;
307
308 /*
309 * Translate the deviceid and eventid into an icid and pintid through
310 * the device table and ITT. Mark the pintid as pending
311 * on the redistributor.
312 * If the interrupt is not configured the command queue stalls.
313 */
314 memset(&cmd, 0, sizeof(cmd));
315 cmd.dw[0] = GITS_CMD_INT | ((uint64_t)deviceid << 32);
316 cmd.dw[1] = eventid;
317
318 DPRINTF(("ITS #%u: INT deviceid 0x%x eventid 0x%x\n",
319 its->its_id, deviceid, eventid));
320
321 return gits_command(its, &cmd);
322 }
323 #endif
324
325 static int
326 gits_wait(struct gicv3_its *its)
327 {
328 u_int woff, roff;
329 int retry = 100000;
330
331 /*
332 * The ITS command queue is empty when CWRITER and CREADR specify the
333 * same base address offset value.
334 */
335 for (retry = 1000; retry > 0; retry--) {
336 woff = gits_read_8(its, GITS_CWRITER) & GITS_CWRITER_Offset;
337 roff = gits_read_8(its, GITS_CREADR) & GITS_CREADR_Offset;
338 if (woff == roff)
339 break;
340 delay(100);
341 }
342 if (retry == 0) {
343 device_printf(its->its_gic->sc_dev,
344 "ITS command queue timeout! CREADR=0x%lx CWRITER=0x%lx\n",
345 gits_read_8(its, GITS_CREADR), gits_read_8(its, GITS_CWRITER));
346 return ETIMEDOUT;
347 }
348
349 return 0;
350 }
351
352 static int
353 gicv3_its_msi_alloc_lpi(struct gicv3_its *its,
354 const struct pci_attach_args *pa)
355 {
356 struct pci_attach_args *new_pa;
357 vmem_addr_t n;
358
359 KASSERT(its->its_gic->sc_lpi_pool != NULL);
360
361 if (vmem_alloc(its->its_gic->sc_lpi_pool, 1, VM_INSTANTFIT|VM_SLEEP, &n) != 0)
362 return -1;
363
364 KASSERT(its->its_pa[n] == NULL);
365
366 new_pa = kmem_alloc(sizeof(*new_pa), KM_SLEEP);
367 memcpy(new_pa, pa, sizeof(*new_pa));
368 its->its_pa[n] = new_pa;
369 return n + its->its_pic->pic_irqbase;
370 }
371
372 static void
373 gicv3_its_msi_free_lpi(struct gicv3_its *its, int lpi)
374 {
375 struct pci_attach_args *pa;
376
377 KASSERT(its->its_gic->sc_lpi_pool != NULL);
378 KASSERT(lpi >= its->its_pic->pic_irqbase);
379
380 pa = its->its_pa[lpi - its->its_pic->pic_irqbase];
381 its->its_pa[lpi - its->its_pic->pic_irqbase] = NULL;
382 kmem_free(pa, sizeof(*pa));
383
384 vmem_free(its->its_gic->sc_lpi_pool, lpi - its->its_pic->pic_irqbase, 1);
385 }
386
387 static uint32_t
388 gicv3_its_devid(pci_chipset_tag_t pc, pcitag_t tag)
389 {
390 uint32_t devid;
391 int b, d, f;
392
393 pci_decompose_tag(pc, tag, &b, &d, &f);
394
395 devid = (b << 8) | (d << 3) | f;
396
397 return pci_get_devid(pc, devid);
398 }
399
400 static int
401 gicv3_its_device_map(struct gicv3_its *its, uint32_t devid, u_int count)
402 {
403 struct gicv3_its_device *dev;
404 struct gicv3_its_table *itstab = &its->its_tab_device;
405 u_int vectors;
406 int error;
407
408 vectors = MAX(2, count);
409 while (!powerof2(vectors))
410 vectors++;
411
412 const uint64_t typer = gits_read_8(its, GITS_TYPER);
413 const u_int itt_entry_size = __SHIFTOUT(typer, GITS_TYPER_ITT_entry_size) + 1;
414 const u_int itt_size = roundup(uimax(vectors, 2) * itt_entry_size, GITS_ITT_ALIGN);
415
416 LIST_FOREACH(dev, &its->its_devices, dev_list)
417 if (dev->dev_id == devid) {
418 return itt_size <= dev->dev_size ? 0 : EEXIST;
419 }
420
421 if (itstab->tab_indirect) {
422 /* Need to allocate the L2 table. */
423 uint64_t *l1_tab = itstab->tab_l1;
424 struct gicv3_its_page_table *pt;
425 const u_int index = devid / itstab->tab_l2_num_ids;
426
427 pt = kmem_alloc(sizeof(*pt), KM_SLEEP);
428 pt->pt_dev_id = devid;
429 gicv3_dma_alloc(its->its_gic, &pt->pt_dma, itstab->tab_l2_entry_size,
430 itstab->tab_page_size);
431 LIST_INSERT_HEAD(&itstab->tab_pt, pt, pt_list);
432
433 if (!itstab->tab_shareable) {
434 cpu_dcache_wb_range((vaddr_t)pt->pt_dma.base,
435 itstab->tab_l2_entry_size);
436 }
437 l1_tab[index] = pt->pt_dma.segs[0].ds_addr | GITS_BASER_Valid;
438 if (!itstab->tab_shareable) {
439 cpu_dcache_wb_range((vaddr_t)&l1_tab[index],
440 sizeof(l1_tab[index]));
441 }
442 dsb(sy);
443
444 DPRINTF(("ITS: Allocated L2 entry at index %u for devid 0x%x\n",
445 index, devid));
446 }
447
448 dev = kmem_alloc(sizeof(*dev), KM_SLEEP);
449 dev->dev_id = devid;
450 dev->dev_size = itt_size;
451 gicv3_dma_alloc(its->its_gic, &dev->dev_itt, itt_size, GITS_ITT_ALIGN);
452 LIST_INSERT_HEAD(&its->its_devices, dev, dev_list);
453
454 if (its->its_cmd_flush) {
455 cpu_dcache_wb_range((vaddr_t)dev->dev_itt.base, itt_size);
456 }
457 dsb(sy);
458
459 /*
460 * Map the device to the ITT
461 */
462 const u_int size = uimax(1, fls32(vectors));
463 mutex_enter(its->its_lock);
464 error = gits_command_mapd(its, devid, dev->dev_itt.segs[0].ds_addr, size, true);
465 if (error == 0) {
466 error = gits_wait(its);
467 }
468 mutex_exit(its->its_lock);
469
470 return error;
471 }
472
473 static void
474 gicv3_its_msi_enable(struct gicv3_its *its, int lpi, int count)
475 {
476 const struct pci_attach_args *pa = its->its_pa[lpi - its->its_pic->pic_irqbase];
477 pci_chipset_tag_t pc = pa->pa_pc;
478 pcitag_t tag = pa->pa_tag;
479 pcireg_t ctl;
480 int off;
481
482 if (!pci_get_capability(pc, tag, PCI_CAP_MSI, &off, NULL))
483 panic("gicv3_its_msi_enable: device is not MSI-capable");
484
485 ctl = pci_conf_read(pc, tag, off + PCI_MSI_CTL);
486 ctl &= ~PCI_MSI_CTL_MME_MASK;
487 ctl |= __SHIFTIN(ilog2(count), PCI_MSI_CTL_MME_MASK);
488 pci_conf_write(pc, tag, off + PCI_MSI_CTL, ctl);
489
490 const uint64_t addr = its->its_base + GITS_TRANSLATER;
491 ctl = pci_conf_read(pc, tag, off + PCI_MSI_CTL);
492 if (ctl & PCI_MSI_CTL_64BIT_ADDR) {
493 pci_conf_write(pc, tag, off + PCI_MSI_MADDR64_LO,
494 addr & 0xffffffff);
495 pci_conf_write(pc, tag, off + PCI_MSI_MADDR64_HI,
496 (addr >> 32) & 0xffffffff);
497 pci_conf_write(pc, tag, off + PCI_MSI_MDATA64,
498 lpi - its->its_pic->pic_irqbase);
499 } else {
500 KASSERT((addr >> 32) == 0);
501 pci_conf_write(pc, tag, off + PCI_MSI_MADDR,
502 addr & 0xffffffff);
503 pci_conf_write(pc, tag, off + PCI_MSI_MDATA,
504 lpi - its->its_pic->pic_irqbase);
505 }
506 ctl |= PCI_MSI_CTL_MSI_ENABLE;
507 pci_conf_write(pc, tag, off + PCI_MSI_CTL, ctl);
508 }
509
510 static void
511 gicv3_its_msi_disable(struct gicv3_its *its, int lpi)
512 {
513 const struct pci_attach_args *pa = its->its_pa[lpi - its->its_pic->pic_irqbase];
514 pci_chipset_tag_t pc = pa->pa_pc;
515 pcitag_t tag = pa->pa_tag;
516 pcireg_t ctl;
517 int off;
518
519 if (!pci_get_capability(pc, tag, PCI_CAP_MSI, &off, NULL))
520 panic("gicv3_its_msi_enable: device is not MSI-capable");
521
522 ctl = pci_conf_read(pc, tag, off + PCI_MSI_CTL);
523 ctl &= ~PCI_MSI_CTL_MSI_ENABLE;
524 pci_conf_write(pc, tag, off + PCI_MSI_CTL, ctl);
525 }
526
527 static void
528 gicv3_its_msix_enable(struct gicv3_its *its, int lpi, int msix_vec,
529 bus_space_tag_t bst, bus_space_handle_t bsh)
530 {
531 const struct pci_attach_args *pa = its->its_pa[lpi - its->its_pic->pic_irqbase];
532 pci_chipset_tag_t pc = pa->pa_pc;
533 pcitag_t tag = pa->pa_tag;
534 pcireg_t ctl;
535 uint32_t val;
536 int off;
537
538 if (!pci_get_capability(pc, tag, PCI_CAP_MSIX, &off, NULL))
539 panic("gicv3_its_msix_enable: device is not MSI-X-capable");
540
541 const uint64_t addr = its->its_base + GITS_TRANSLATER;
542 const uint64_t entry_base = PCI_MSIX_TABLE_ENTRY_SIZE * msix_vec;
543 bus_space_write_4(bst, bsh, entry_base + PCI_MSIX_TABLE_ENTRY_ADDR_LO, (uint32_t)addr);
544 bus_space_write_4(bst, bsh, entry_base + PCI_MSIX_TABLE_ENTRY_ADDR_HI, (uint32_t)(addr >> 32));
545 bus_space_write_4(bst, bsh, entry_base + PCI_MSIX_TABLE_ENTRY_DATA, lpi - its->its_pic->pic_irqbase);
546 val = bus_space_read_4(bst, bsh, entry_base + PCI_MSIX_TABLE_ENTRY_VECTCTL);
547 val &= ~PCI_MSIX_VECTCTL_MASK;
548 bus_space_write_4(bst, bsh, entry_base + PCI_MSIX_TABLE_ENTRY_VECTCTL, val);
549
550 ctl = pci_conf_read(pc, tag, off + PCI_MSIX_CTL);
551 ctl |= PCI_MSIX_CTL_ENABLE;
552 pci_conf_write(pc, tag, off + PCI_MSIX_CTL, ctl);
553 }
554
555 static void
556 gicv3_its_msix_disable(struct gicv3_its *its, int lpi)
557 {
558 const struct pci_attach_args *pa = its->its_pa[lpi - its->its_pic->pic_irqbase];
559 pci_chipset_tag_t pc = pa->pa_pc;
560 pcitag_t tag = pa->pa_tag;
561 pcireg_t ctl;
562 int off;
563
564 if (!pci_get_capability(pc, tag, PCI_CAP_MSIX, &off, NULL))
565 panic("gicv3_its_msix_disable: device is not MSI-X-capable");
566
567 ctl = pci_conf_read(pc, tag, off + PCI_MSIX_CTL);
568 ctl &= ~PCI_MSIX_CTL_ENABLE;
569 pci_conf_write(pc, tag, off + PCI_MSIX_CTL, ctl);
570 }
571
572 static pci_intr_handle_t *
573 gicv3_its_msi_alloc(struct arm_pci_msi *msi, int *count,
574 const struct pci_attach_args *pa, bool exact)
575 {
576 struct gicv3_its * const its = msi->msi_priv;
577 struct cpu_info * const ci = cpu_lookup(0);
578 pci_intr_handle_t *vectors;
579 int n, off, error;
580
581 if (!pci_get_capability(pa->pa_pc, pa->pa_tag, PCI_CAP_MSI, &off, NULL))
582 return NULL;
583
584 const uint64_t typer = gits_read_8(its, GITS_TYPER);
585 const u_int id_bits = __SHIFTOUT(typer, GITS_TYPER_ID_bits) + 1;
586 if (*count == 0 || *count > (1 << id_bits))
587 return NULL;
588
589 const uint32_t devid = gicv3_its_devid(pa->pa_pc, pa->pa_tag);
590
591 if (gicv3_its_device_map(its, devid, *count) != 0)
592 return NULL;
593
594 vectors = kmem_alloc(sizeof(*vectors) * *count, KM_SLEEP);
595 mutex_enter(its->its_lock);
596 for (n = 0; n < *count; n++) {
597 const int lpi = gicv3_its_msi_alloc_lpi(its, pa);
598 KASSERT(lpi >= 0);
599 vectors[n] = ARM_PCI_INTR_MSI |
600 __SHIFTIN(lpi, ARM_PCI_INTR_IRQ) |
601 __SHIFTIN(n, ARM_PCI_INTR_MSI_VEC) |
602 __SHIFTIN(msi->msi_id, ARM_PCI_INTR_FRAME);
603
604 if (n == 0)
605 gicv3_its_msi_enable(its, lpi, *count);
606
607 /*
608 * Record devid and target PE
609 */
610 its->its_devid[lpi - its->its_pic->pic_irqbase] = devid;
611 its->its_targets[lpi - its->its_pic->pic_irqbase] = ci;
612
613 /*
614 * Map event
615 */
616 gits_command_mapti(its, devid, lpi - its->its_pic->pic_irqbase, lpi, cpu_index(ci));
617 gits_command_sync(its, its->its_rdbase[cpu_index(ci)]);
618 }
619 error = gits_wait(its);
620 mutex_exit(its->its_lock);
621
622 if (error != 0) {
623 kmem_free(vectors, sizeof(*vectors) * *count);
624 vectors = NULL;
625 }
626
627 return vectors;
628 }
629
630 static pci_intr_handle_t *
631 gicv3_its_msix_alloc(struct arm_pci_msi *msi, u_int *table_indexes, int *count,
632 const struct pci_attach_args *pa, bool exact)
633 {
634 struct gicv3_its * const its = msi->msi_priv;
635 struct cpu_info *ci = cpu_lookup(0);
636 pci_intr_handle_t *vectors;
637 bus_space_tag_t bst;
638 bus_space_handle_t bsh;
639 bus_size_t bsz;
640 uint32_t table_offset, table_size;
641 int n, off, bar, error;
642 pcireg_t tbl;
643
644 if (!pci_get_capability(pa->pa_pc, pa->pa_tag, PCI_CAP_MSIX, &off, NULL))
645 return NULL;
646
647 const uint64_t typer = gits_read_8(its, GITS_TYPER);
648 const u_int id_bits = __SHIFTOUT(typer, GITS_TYPER_ID_bits) + 1;
649 if (*count == 0 || *count > (1 << id_bits))
650 return NULL;
651
652 tbl = pci_conf_read(pa->pa_pc, pa->pa_tag, off + PCI_MSIX_TBLOFFSET);
653 bar = PCI_BAR0 + (4 * (tbl & PCI_MSIX_TBLBIR_MASK));
654 table_offset = tbl & PCI_MSIX_TBLOFFSET_MASK;
655 table_size = pci_msix_count(pa->pa_pc, pa->pa_tag) * PCI_MSIX_TABLE_ENTRY_SIZE;
656 if (table_size == 0)
657 return NULL;
658
659 error = pci_mapreg_submap(pa, bar, pci_mapreg_type(pa->pa_pc, pa->pa_tag, bar),
660 BUS_SPACE_MAP_LINEAR, roundup(table_size, PAGE_SIZE), table_offset,
661 &bst, &bsh, NULL, &bsz);
662 if (error)
663 return NULL;
664
665 const uint32_t devid = gicv3_its_devid(pa->pa_pc, pa->pa_tag);
666
667 if (gicv3_its_device_map(its, devid, *count) != 0) {
668 bus_space_unmap(bst, bsh, bsz);
669 return NULL;
670 }
671
672 vectors = kmem_alloc(sizeof(*vectors) * *count, KM_SLEEP);
673 mutex_enter(its->its_lock);
674 for (n = 0; n < *count; n++) {
675 const int lpi = gicv3_its_msi_alloc_lpi(its, pa);
676 KASSERT(lpi >= 0);
677 const int msix_vec = table_indexes ? table_indexes[n] : n;
678 vectors[msix_vec] = ARM_PCI_INTR_MSIX |
679 __SHIFTIN(lpi, ARM_PCI_INTR_IRQ) |
680 __SHIFTIN(msix_vec, ARM_PCI_INTR_MSI_VEC) |
681 __SHIFTIN(msi->msi_id, ARM_PCI_INTR_FRAME);
682
683 gicv3_its_msix_enable(its, lpi, msix_vec, bst, bsh);
684
685 /*
686 * Record devid and target PE
687 */
688 its->its_devid[lpi - its->its_pic->pic_irqbase] = devid;
689 its->its_targets[lpi - its->its_pic->pic_irqbase] = ci;
690
691 /*
692 * Map event
693 */
694 gits_command_mapti(its, devid, lpi - its->its_pic->pic_irqbase, lpi, cpu_index(ci));
695 gits_command_sync(its, its->its_rdbase[cpu_index(ci)]);
696 }
697 gits_wait(its);
698 mutex_exit(its->its_lock);
699
700 bus_space_unmap(bst, bsh, bsz);
701
702 return vectors;
703 }
704
705 static void *
706 gicv3_its_msi_intr_establish(struct arm_pci_msi *msi,
707 pci_intr_handle_t ih, int ipl, int (*func)(void *), void *arg, const char *xname)
708 {
709 struct gicv3_its * const its = msi->msi_priv;
710 void *intrh;
711
712 const int lpi = __SHIFTOUT(ih, ARM_PCI_INTR_IRQ);
713 const int mpsafe = (ih & ARM_PCI_INTR_MPSAFE) ? IST_MPSAFE : 0;
714
715 intrh = pic_establish_intr(its->its_pic, lpi - its->its_pic->pic_irqbase, ipl,
716 IST_EDGE | mpsafe, func, arg, xname);
717 if (intrh == NULL)
718 return NULL;
719
720 /* Invalidate LPI configuration tables */
721 KASSERT(its->its_pa[lpi - its->its_pic->pic_irqbase] != NULL);
722 const uint32_t devid = its->its_devid[lpi - its->its_pic->pic_irqbase];
723 gits_command_inv(its, devid, lpi - its->its_pic->pic_irqbase);
724
725 return intrh;
726 }
727
728 static void
729 gicv3_its_msi_intr_release(struct arm_pci_msi *msi, pci_intr_handle_t *pih,
730 int count)
731 {
732 struct gicv3_its * const its = msi->msi_priv;
733 int n;
734
735 for (n = 0; n < count; n++) {
736 const int lpi = __SHIFTOUT(pih[n], ARM_PCI_INTR_IRQ);
737 KASSERT(lpi >= its->its_pic->pic_irqbase);
738 if (pih[n] & ARM_PCI_INTR_MSIX)
739 gicv3_its_msix_disable(its, lpi);
740 if (pih[n] & ARM_PCI_INTR_MSI)
741 gicv3_its_msi_disable(its, lpi);
742 gicv3_its_msi_free_lpi(its, lpi);
743 its->its_targets[lpi - its->its_pic->pic_irqbase] = NULL;
744 its->its_devid[lpi - its->its_pic->pic_irqbase] = 0;
745 struct intrsource * const is =
746 its->its_pic->pic_sources[lpi - its->its_pic->pic_irqbase];
747 if (is != NULL)
748 pic_disestablish_source(is);
749 }
750 }
751
752 static void
753 gicv3_its_command_init(struct gicv3_softc *sc, struct gicv3_its *its)
754 {
755 uint64_t cbaser, tmp;
756
757 gicv3_dma_alloc(sc, &its->its_cmd, GITS_COMMANDS_SIZE, GITS_COMMANDS_ALIGN);
758 if (its->its_cmd_flush) {
759 cpu_dcache_wb_range((vaddr_t)its->its_cmd.base, GITS_COMMANDS_SIZE);
760 }
761 dsb(sy);
762
763 KASSERT((gits_read_4(its, GITS_CTLR) & GITS_CTLR_Enabled) == 0);
764 KASSERT((gits_read_4(its, GITS_CTLR) & GITS_CTLR_Quiescent) != 0);
765
766 cbaser = its->its_cmd.segs[0].ds_addr;
767 cbaser |= __SHIFTIN((its->its_cmd.len / 4096) - 1, GITS_CBASER_Size);
768 cbaser |= GITS_CBASER_Valid;
769
770 cbaser |= __SHIFTIN(GITS_Cache_NORMAL_WA_WB, GITS_CBASER_InnerCache);
771 cbaser |= __SHIFTIN(GITS_Shareability_IS, GITS_CBASER_Shareability);
772 gits_write_8(its, GITS_CBASER, cbaser);
773
774 tmp = gits_read_8(its, GITS_CBASER);
775 if (__SHIFTOUT(tmp, GITS_CBASER_Shareability) != GITS_Shareability_IS) {
776 if (__SHIFTOUT(tmp, GITS_CBASER_InnerCache) == GITS_Shareability_NS) {
777 cbaser &= ~GITS_CBASER_InnerCache;
778 cbaser |= __SHIFTIN(GITS_Cache_NORMAL_NC, GITS_CBASER_InnerCache);
779 cbaser &= ~GITS_CBASER_Shareability;
780 cbaser |= __SHIFTIN(GITS_Shareability_NS, GITS_CBASER_Shareability);
781 gits_write_8(its, GITS_CBASER, cbaser);
782 }
783
784 its->its_cmd_flush = true;
785 }
786
787 gits_write_8(its, GITS_CWRITER, 0);
788 }
789
790 static void
791 gicv3_its_table_params(struct gicv3_softc *sc, struct gicv3_its *its,
792 u_int *devbits, u_int *innercache, u_int *share)
793 {
794
795 const uint64_t typer = gits_read_8(its, GITS_TYPER);
796 const uint32_t iidr = gits_read_4(its, GITS_IIDR);
797
798 /* Default values */
799 *devbits = __SHIFTOUT(typer, GITS_TYPER_Devbits) + 1;
800 *innercache = GITS_Cache_NORMAL_WA_WB;
801 *share = GITS_Shareability_IS;
802
803 /* Cavium ThunderX errata */
804 if ((iidr & GITS_IIDR_CAVIUM_ERRATA_MASK) == GITS_IIDR_CAVIUM_ERRATA_VALUE) {
805 *devbits = 20; /* 8Mb */
806 *innercache = GITS_Cache_DEVICE_nGnRnE;
807 aprint_normal_dev(sc->sc_dev, "Cavium ThunderX errata detected\n");
808 }
809 }
810
811 static u_int
812 gicv3_its_probe_page_size(struct gicv3_its *its, int tab)
813 {
814 uint64_t baser, tmp;
815 u_int page_size = 65536;
816
817 baser = gits_read_8(its, GITS_BASERn(tab));
818 for (;;) {
819 baser &= ~GITS_BASER_Page_Size;
820 switch (page_size) {
821 case 4096:
822 baser |= __SHIFTIN(GITS_Page_Size_4KB, GITS_BASER_Page_Size);
823 break;
824 case 16384:
825 baser |= __SHIFTIN(GITS_Page_Size_16KB, GITS_BASER_Page_Size);
826 break;
827 case 65536:
828 baser |= __SHIFTIN(GITS_Page_Size_64KB, GITS_BASER_Page_Size);
829 break;
830 }
831
832 gits_write_8(its, GITS_BASERn(tab), baser);
833 tmp = gits_read_8(its, GITS_BASERn(tab));
834 if ((baser & GITS_BASER_Page_Size) == (tmp & GITS_BASER_Page_Size)) {
835 return page_size;
836 }
837
838 if (page_size == 65536) {
839 page_size = 16384;
840 } else if (page_size == 16384) {
841 page_size = 4096;
842 } else {
843 aprint_error_dev(its->its_gic->sc_dev,
844 "WARNING: Couldn't determine ITS page size, "
845 "defaulting to 4KB\n");
846 return page_size;
847 }
848 }
849 }
850
851 static bool
852 gicv3_its_table_probe_indirect(struct gicv3_its *its, int tab)
853 {
854 uint64_t baser;
855
856 baser = gits_read_8(its, GITS_BASERn(tab));
857 baser |= GITS_BASER_Indirect;
858 gits_write_8(its, GITS_BASERn(tab), baser);
859
860 baser = gits_read_8(its, GITS_BASERn(tab));
861
862 return (baser & GITS_BASER_Indirect) != 0;
863 }
864
865 static void
866 gicv3_its_table_init(struct gicv3_softc *sc, struct gicv3_its *its)
867 {
868 u_int page_size, table_align;
869 u_int devbits, innercache, share;
870 const char *table_type;
871 uint64_t baser;
872 int tab;
873
874 gicv3_its_table_params(sc, its, &devbits, &innercache, &share);
875
876 DPRINTF(("ITS: devbits = %u\n", devbits));
877
878 for (tab = 0; tab < 8; tab++) {
879 struct gicv3_its_table *itstab;
880 bool indirect = false;
881 uint64_t l1_entry_size, l2_entry_size;
882 uint64_t l1_num_ids, l2_num_ids;
883 uint64_t table_size;
884
885 baser = gits_read_8(its, GITS_BASERn(tab));
886
887 l1_entry_size = __SHIFTOUT(baser, GITS_BASER_Entry_Size) + 1;
888 l2_entry_size = 0;
889 l2_num_ids = 0;
890
891 page_size = gicv3_its_probe_page_size(its, tab);
892 table_align = 65536; // why not
893
894 switch (__SHIFTOUT(baser, GITS_BASER_Type)) {
895 case GITS_Type_Devices:
896 /*
897 * Table size scales with the width of the DeviceID.
898 */
899 l1_num_ids = 1ULL << devbits;
900 DPRINTF(("ITS: l1_num_ids = %lu\n", l1_num_ids));
901 indirect =
902 gicv3_its_table_probe_indirect(its, tab);
903 if (indirect) {
904 DPRINTF(("ITS: indirect\n"));
905 l2_entry_size = l1_entry_size;
906 l2_num_ids = page_size / l2_entry_size;
907 l1_num_ids = l1_num_ids / l2_num_ids;
908 l1_entry_size = GITS_INDIRECT_ENTRY_SIZE;
909 }
910 table_size = roundup2(l1_entry_size * l1_num_ids, page_size);
911 if (howmany(table_size, page_size) > GITS_BASER_Size + 1) {
912 DPRINTF(("ITS: clamp table size 0x%lx -> ", table_size));
913 table_size = (GITS_BASER_Size + 1) * page_size;
914 DPRINTF(("0x%lx\n", table_size));
915 }
916 table_type = "Devices";
917
918 DPRINTF(("ITS: table_size is 0x%lx\n", table_size));
919
920 itstab = &its->its_tab_device;
921 itstab->tab_page_size = page_size;
922 itstab->tab_l1_entry_size = l1_entry_size;
923 itstab->tab_l1_num_ids = l1_num_ids;
924 itstab->tab_l2_entry_size = l2_entry_size;
925 itstab->tab_l2_num_ids = l2_num_ids;
926 itstab->tab_indirect = indirect;
927 LIST_INIT(&itstab->tab_pt);
928 break;
929 case GITS_Type_InterruptCollections:
930 /*
931 * Allocate space for one interrupt collection per CPU.
932 */
933 table_size = roundup(l1_entry_size * ncpu, page_size);
934 table_type = "Collections";
935 break;
936 default:
937 table_size = 0;
938 break;
939 }
940
941 if (table_size == 0)
942 continue;
943
944 gicv3_dma_alloc(sc, &its->its_tab[tab], table_size, table_align);
945 if (its->its_cmd_flush) {
946 cpu_dcache_wb_range((vaddr_t)its->its_tab[tab].base, table_size);
947 }
948 dsb(sy);
949
950 baser &= ~GITS_BASER_Size;
951 baser |= __SHIFTIN(howmany(table_size, page_size) - 1, GITS_BASER_Size);
952 baser &= ~GITS_BASER_Physical_Address;
953 baser |= its->its_tab[tab].segs[0].ds_addr;
954 baser &= ~GITS_BASER_InnerCache;
955 baser |= __SHIFTIN(innercache, GITS_BASER_InnerCache);
956 baser &= ~GITS_BASER_Shareability;
957 baser |= __SHIFTIN(share, GITS_BASER_Shareability);
958 baser |= GITS_BASER_Valid;
959 if (indirect) {
960 baser |= GITS_BASER_Indirect;
961 } else {
962 baser &= ~GITS_BASER_Indirect;
963 }
964
965 gits_write_8(its, GITS_BASERn(tab), baser);
966
967 baser = gits_read_8(its, GITS_BASERn(tab));
968 if (__SHIFTOUT(baser, GITS_BASER_Shareability) == GITS_Shareability_NS) {
969 baser &= ~GITS_BASER_InnerCache;
970 baser |= __SHIFTIN(GITS_Cache_NORMAL_NC, GITS_BASER_InnerCache);
971
972 gits_write_8(its, GITS_BASERn(tab), baser);
973 }
974
975 baser = gits_read_8(its, GITS_BASERn(tab));
976 aprint_normal_dev(sc->sc_dev, "ITS [#%d] %s table @ %#lx/%#lx, %s, %s%s\n",
977 tab, table_type, its->its_tab[tab].segs[0].ds_addr, table_size,
978 gits_cache_type[__SHIFTOUT(baser, GITS_BASER_InnerCache)],
979 gits_share_type[__SHIFTOUT(baser, GITS_BASER_Shareability)],
980 indirect ? ", indirect" : "");
981
982 if (__SHIFTOUT(baser, GITS_BASER_Type) == GITS_Type_Devices) {
983 its->its_tab_device.tab_l1 = its->its_tab[tab].base;
984 its->its_tab_device.tab_shareable =
985 __SHIFTOUT(baser, GITS_BASER_Shareability) != GITS_Shareability_NS;
986 }
987
988 }
989 }
990
991 static void
992 gicv3_its_enable(struct gicv3_softc *sc, struct gicv3_its *its)
993 {
994 uint32_t ctlr;
995
996 ctlr = gits_read_4(its, GITS_CTLR);
997 ctlr |= GITS_CTLR_Enabled;
998 gits_write_4(its, GITS_CTLR, ctlr);
999 }
1000
1001 static void
1002 gicv3_its_cpu_init(void *priv, struct cpu_info *ci)
1003 {
1004 struct gicv3_its * const its = priv;
1005 struct gicv3_softc * const sc = its->its_gic;
1006 uint64_t rdbase;
1007 size_t irq;
1008
1009 const uint64_t typer = bus_space_read_8(sc->sc_bst, its->its_bsh, GITS_TYPER);
1010 if (typer & GITS_TYPER_PTA) {
1011 void *va = bus_space_vaddr(sc->sc_bst, sc->sc_bsh_r[ci->ci_gic_redist]);
1012 rdbase = vtophys((vaddr_t)va);
1013 } else {
1014 rdbase = (uint64_t)sc->sc_processor_id[cpu_index(ci)] << 16;
1015 }
1016 its->its_rdbase[cpu_index(ci)] = rdbase;
1017
1018 /*
1019 * Map collection ID of this CPU's index to this CPU's redistributor.
1020 */
1021 mutex_enter(its->its_lock);
1022 gits_command_mapc(its, cpu_index(ci), rdbase, true);
1023 gits_command_invall(its, cpu_index(ci));
1024 gits_wait(its);
1025
1026 /*
1027 * Update routing for LPIs targetting this CPU
1028 */
1029 for (irq = 0; irq < its->its_pic->pic_maxsources; irq++) {
1030 if (its->its_targets[irq] != ci)
1031 continue;
1032 KASSERT(its->its_pa[irq] != NULL);
1033
1034 const uint32_t devid = its->its_devid[irq];
1035 gits_command_movi(its, devid, irq, cpu_index(ci));
1036 gits_command_sync(its, its->its_rdbase[cpu_index(ci)]);
1037 }
1038 gits_wait(its);
1039 mutex_exit(its->its_lock);
1040
1041 its->its_cpuonline[cpu_index(ci)] = true;
1042 }
1043
1044 static void
1045 gicv3_its_get_affinity(void *priv, size_t irq, kcpuset_t *affinity)
1046 {
1047 struct gicv3_its * const its = priv;
1048 struct cpu_info *ci;
1049
1050 ci = its->its_targets[irq];
1051 if (ci)
1052 kcpuset_set(affinity, cpu_index(ci));
1053 }
1054
1055 static int
1056 gicv3_its_set_affinity(void *priv, size_t irq, const kcpuset_t *affinity)
1057 {
1058 struct gicv3_its * const its = priv;
1059 const struct pci_attach_args *pa;
1060 struct cpu_info *ci;
1061
1062 const int set = kcpuset_countset(affinity);
1063 if (set != 1)
1064 return EINVAL;
1065
1066 pa = its->its_pa[irq];
1067 if (pa == NULL)
1068 return EPASSTHROUGH;
1069
1070 ci = cpu_lookup(kcpuset_ffs(affinity) - 1);
1071 its->its_targets[irq] = ci;
1072
1073 if (its->its_cpuonline[cpu_index(ci)] == true) {
1074 const uint32_t devid = gicv3_its_devid(pa->pa_pc, pa->pa_tag);
1075 mutex_enter(its->its_lock);
1076 gits_command_movi(its, devid, irq, cpu_index(ci));
1077 gits_command_sync(its, its->its_rdbase[cpu_index(ci)]);
1078 mutex_exit(its->its_lock);
1079 }
1080
1081 return 0;
1082 }
1083
1084 int
1085 gicv3_its_init(struct gicv3_softc *sc, bus_space_handle_t bsh,
1086 uint64_t its_base, uint32_t its_id)
1087 {
1088 struct gicv3_its *its;
1089 struct arm_pci_msi *msi;
1090
1091 const uint64_t typer = bus_space_read_8(sc->sc_bst, bsh, GITS_TYPER);
1092 if ((typer & GITS_TYPER_Physical) == 0)
1093 return ENXIO;
1094
1095 its = kmem_zalloc(sizeof(*its), KM_SLEEP);
1096 its->its_id = its_id;
1097 its->its_bst = sc->sc_bst;
1098 its->its_bsh = bsh;
1099 its->its_dmat = sc->sc_dmat;
1100 its->its_base = its_base;
1101 its->its_pic = &sc->sc_lpi;
1102 snprintf(its->its_pic->pic_name, sizeof(its->its_pic->pic_name), "gicv3-its");
1103 KASSERT(its->its_pic->pic_maxsources > 0);
1104 its->its_pa = kmem_zalloc(sizeof(struct pci_attach_args *) * its->its_pic->pic_maxsources, KM_SLEEP);
1105 its->its_targets = kmem_zalloc(sizeof(struct cpu_info *) * its->its_pic->pic_maxsources, KM_SLEEP);
1106 its->its_devid = kmem_zalloc(sizeof(uint32_t) * its->its_pic->pic_maxsources, KM_SLEEP);
1107 its->its_gic = sc;
1108 its->its_rdbase = kmem_zalloc(sizeof(*its->its_rdbase) * ncpu, KM_SLEEP);
1109 its->its_cpuonline = kmem_zalloc(sizeof(*its->its_cpuonline) * ncpu, KM_SLEEP);
1110 its->its_cb.cpu_init = gicv3_its_cpu_init;
1111 its->its_cb.get_affinity = gicv3_its_get_affinity;
1112 its->its_cb.set_affinity = gicv3_its_set_affinity;
1113 its->its_cb.priv = its;
1114 LIST_INIT(&its->its_devices);
1115 LIST_INSERT_HEAD(&sc->sc_lpi_callbacks, &its->its_cb, list);
1116 its->its_lock = mutex_obj_alloc(MUTEX_SPIN, IPL_NONE);
1117
1118 gicv3_its_command_init(sc, its);
1119 gicv3_its_table_init(sc, its);
1120
1121 gicv3_its_enable(sc, its);
1122
1123 gicv3_its_cpu_init(its, curcpu());
1124
1125 msi = &its->its_msi;
1126 msi->msi_id = its_id;
1127 msi->msi_dev = sc->sc_dev;
1128 msi->msi_priv = its;
1129 msi->msi_alloc = gicv3_its_msi_alloc;
1130 msi->msix_alloc = gicv3_its_msix_alloc;
1131 msi->msi_intr_establish = gicv3_its_msi_intr_establish;
1132 msi->msi_intr_release = gicv3_its_msi_intr_release;
1133
1134 return arm_pci_msi_add(msi);
1135 }
1136