acpi_machdep.c revision 1.27 1 /* $NetBSD: acpi_machdep.c,v 1.27 2024/12/09 21:56:19 jmcneill Exp $ */
2
3 /*-
4 * Copyright (c) 2018 The NetBSD Foundation, Inc.
5 * All rights reserved.
6 *
7 * This code is derived from software contributed to The NetBSD Foundation
8 * by Jared McNeill <jmcneill (at) invisible.ca>.
9 *
10 * Redistribution and use in source and binary forms, with or without
11 * modification, are permitted provided that the following conditions
12 * are met:
13 * 1. Redistributions of source code must retain the above copyright
14 * notice, this list of conditions and the following disclaimer.
15 * 2. Redistributions in binary form must reproduce the above copyright
16 * notice, this list of conditions and the following disclaimer in the
17 * documentation and/or other materials provided with the distribution.
18 *
19 * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
20 * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
21 * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
22 * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
23 * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
24 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
25 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
26 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
27 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
28 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
29 * POSSIBILITY OF SUCH DAMAGE.
30 */
31
32 #include "pci.h"
33
34 #include <sys/cdefs.h>
35 __KERNEL_RCSID(0, "$NetBSD: acpi_machdep.c,v 1.27 2024/12/09 21:56:19 jmcneill Exp $");
36
37 #include <sys/param.h>
38 #include <sys/systm.h>
39 #include <sys/bus.h>
40 #include <sys/cpu.h>
41 #include <sys/device.h>
42 #include <sys/kmem.h>
43
44 #include <uvm/uvm_extern.h>
45
46 #include <dev/fdt/fdtvar.h>
47
48 #include <dev/acpi/acpica.h>
49 #include <dev/acpi/acpivar.h>
50 #if NPCI > 0
51 #include <dev/acpi/acpi_mcfg.h>
52 #endif
53 #include <arm/acpi/acpi_iort.h>
54
55 #include <arm/arm/efi_runtime.h>
56
57 #include <arm/pic/picvar.h>
58
59 #include <arm/locore.h>
60
61 #include <machine/acpi_machdep.h>
62
63 extern struct bus_space arm_generic_bs_tag;
64 extern struct arm32_bus_dma_tag acpi_coherent_dma_tag;
65 extern struct arm32_bus_dma_tag arm_generic_dma_tag;
66
67 struct acpi_intrhandler {
68 int (*ah_fn)(void *);
69 void *ah_arg;
70 TAILQ_ENTRY(acpi_intrhandler) ah_list;
71 };
72
73 struct acpi_intrvec {
74 int ai_irq;
75 int ai_ipl;
76 int ai_type;
77 bool ai_mpsafe;
78 int ai_refcnt;
79 void *ai_arg;
80 void *ai_ih;
81 TAILQ_HEAD(, acpi_intrhandler) ai_handlers;
82 TAILQ_ENTRY(acpi_intrvec) ai_list;
83 };
84
85 static TAILQ_HEAD(, acpi_intrvec) acpi_intrvecs =
86 TAILQ_HEAD_INITIALIZER(acpi_intrvecs);
87
88 bus_dma_tag_t arm_acpi_dma32_tag(struct acpi_softc *, struct acpi_devnode *);
89 bus_dma_tag_t arm_acpi_dma64_tag(struct acpi_softc *, struct acpi_devnode *);
90
91 static int
92 acpi_md_pmapflags(paddr_t pa)
93 {
94 int len;
95
96 const int chosen = OF_finddevice("/chosen");
97 if (chosen == -1)
98 return 0;
99
100 const uint32_t *map = fdtbus_get_prop(chosen, "netbsd,uefi-memmap", &len);
101 if (map == NULL)
102 return 0;
103
104 while (len >= 28) {
105 const uint32_t type = be32dec(&map[0]);
106 const uint64_t phys_start = be64dec(&map[1]);
107 const uint64_t num_pages = be64dec(&map[3]);
108 const uint64_t attr = be64dec(&map[5]);
109
110 if (pa >= phys_start && pa < phys_start + (num_pages * EFI_PAGE_SIZE)) {
111 switch (type) {
112 case EFI_MD_TYPE_RECLAIM:
113 /* ACPI table memory */
114 return PMAP_WRITE_BACK;
115
116 case EFI_MD_TYPE_IOMEM:
117 case EFI_MD_TYPE_IOPORT:
118 return PMAP_DEV;
119
120 default:
121 if ((attr & EFI_MD_ATTR_WB) != 0)
122 return PMAP_WRITE_BACK;
123 else if ((attr & EFI_MD_ATTR_WC) != 0)
124 return PMAP_WRITE_COMBINE;
125 else if ((attr & EFI_MD_ATTR_WT) != 0)
126 return 0; /* XXX */
127
128 return PMAP_DEV;
129 }
130 }
131
132 map += 7;
133 len -= 28;
134 }
135
136 /* Not found; assume device memory */
137 return PMAP_DEV;
138 }
139
140 ACPI_STATUS
141 acpi_md_OsInitialize(void)
142 {
143 return AE_OK;
144 }
145
146 ACPI_PHYSICAL_ADDRESS
147 acpi_md_OsGetRootPointer(void)
148 {
149 uint64_t pa;
150
151 const int chosen = OF_finddevice("/chosen");
152 if (chosen == -1)
153 return 0;
154
155 if (of_getprop_uint64(chosen, "netbsd,acpi-root-table", &pa) != 0)
156 return 0;
157
158 return (ACPI_PHYSICAL_ADDRESS)pa;
159 }
160
161 ACPI_STATUS
162 acpi_md_OsInstallInterruptHandler(UINT32 irq, ACPI_OSD_HANDLER handler, void *context,
163 void **cookiep, const char *xname)
164 {
165 return AE_NOT_IMPLEMENTED;
166 }
167
168 void
169 acpi_md_OsRemoveInterruptHandler(void *cookie)
170 {
171 intr_disestablish(cookie);
172 }
173
174 ACPI_STATUS
175 acpi_md_OsMapMemory(ACPI_PHYSICAL_ADDRESS pa, UINT32 size, void **vap)
176 {
177 paddr_t spa, epa, curpa;
178 vaddr_t va, curva;
179
180 spa = trunc_page(pa);
181 epa = round_page(pa + size);
182
183 va = uvm_km_alloc(kernel_map, epa - spa, 0, UVM_KMF_VAONLY);
184 if (va == 0)
185 return AE_NO_MEMORY;
186
187 const int pmapflags = acpi_md_pmapflags(spa);
188
189 aprint_debug("%s: 0x%lx 0x%x flags = %#x\n", __func__, pa, size, pmapflags);
190
191 for (curpa = spa, curva = va; curpa < epa; curpa += PAGE_SIZE, curva += PAGE_SIZE)
192 pmap_kenter_pa(curva, curpa, VM_PROT_READ | VM_PROT_WRITE, pmapflags);
193 pmap_update(pmap_kernel());
194
195 *vap = (void *)(va + (pa - spa));
196
197 return AE_OK;
198 }
199
200 void
201 acpi_md_OsUnmapMemory(void *va, UINT32 size)
202 {
203 vaddr_t ova;
204 vsize_t osz;
205
206 ova = trunc_page((vaddr_t)va);
207 osz = round_page((vaddr_t)va + size) - ova;
208
209 pmap_kremove(ova, osz);
210 pmap_update(pmap_kernel());
211 uvm_km_free(kernel_map, ova, osz, UVM_KMF_VAONLY);
212 }
213
214 ACPI_STATUS
215 acpi_md_OsGetPhysicalAddress(void *va, ACPI_PHYSICAL_ADDRESS *pap)
216 {
217 paddr_t pa;
218
219 if (!pmap_extract(pmap_kernel(), (vaddr_t)va, &pa))
220 return AE_ERROR;
221
222 *pap = pa;
223
224 return AE_OK;
225 }
226
227 BOOLEAN
228 acpi_md_OsReadable(void *va, UINT32 len)
229 {
230 vaddr_t sva, eva;
231 pt_entry_t *pte;
232
233 sva = trunc_page((vaddr_t)va);
234 eva = round_page((vaddr_t)va + len);
235
236 if (sva < VM_MIN_KERNEL_ADDRESS)
237 return FALSE;
238
239 for (; sva < eva; sva += PAGE_SIZE) {
240 pte = kvtopte(sva);
241 if ((*pte & (LX_BLKPAG_AF|LX_BLKPAG_AP)) != (LX_BLKPAG_AF|LX_BLKPAG_AP_RO))
242 return FALSE;
243 }
244
245 return TRUE;
246 }
247
248 BOOLEAN
249 acpi_md_OsWritable(void *va, UINT32 len)
250 {
251 vaddr_t sva, eva;
252 pt_entry_t *pte;
253
254 sva = trunc_page((vaddr_t)va);
255 eva = round_page((vaddr_t)va + len);
256
257 if (sva < VM_MIN_KERNEL_ADDRESS)
258 return FALSE;
259
260 for (; sva < eva; sva += PAGE_SIZE) {
261 pte = kvtopte(sva);
262 if ((*pte & (LX_BLKPAG_AF|LX_BLKPAG_AP)) != (LX_BLKPAG_AF|LX_BLKPAG_AP_RW))
263 return FALSE;
264 }
265
266 return TRUE;
267 }
268
269 void
270 acpi_md_OsEnableInterrupt(void)
271 {
272 cpsie(I32_bit);
273 }
274
275 void
276 acpi_md_OsDisableInterrupt(void)
277 {
278 cpsid(I32_bit);
279 }
280
281 static struct acpi_intrvec *
282 acpi_md_intr_lookup(int irq)
283 {
284 struct acpi_intrvec *ai;
285
286 TAILQ_FOREACH(ai, &acpi_intrvecs, ai_list) {
287 if (ai->ai_irq == irq) {
288 return ai;
289 }
290 }
291
292 return NULL;
293 }
294
295 static int
296 acpi_md_intr(void *arg)
297 {
298 struct acpi_intrvec *ai = arg;
299 struct acpi_intrhandler *ah;
300 int rv = 0;
301
302 TAILQ_FOREACH(ah, &ai->ai_handlers, ah_list) {
303 rv += ah->ah_fn(ah->ah_arg);
304 }
305
306 return rv;
307 }
308
309 void *
310 acpi_md_intr_establish(uint32_t irq, int ipl, int type, int (*handler)(void *), void *arg, bool mpsafe, const char *xname)
311 {
312 struct acpi_intrvec *ai;
313 struct acpi_intrhandler *ah;
314
315 ai = acpi_md_intr_lookup(irq);
316 if (ai == NULL) {
317 ai = kmem_zalloc(sizeof(*ai), KM_SLEEP);
318 ai->ai_refcnt = 0;
319 ai->ai_irq = irq;
320 ai->ai_ipl = ipl;
321 ai->ai_type = type;
322 ai->ai_mpsafe = mpsafe;
323 ai->ai_arg = arg;
324 TAILQ_INIT(&ai->ai_handlers);
325 if (arg == NULL) {
326 ai->ai_ih = intr_establish_xname(irq, ipl,
327 type | (mpsafe ? IST_MPSAFE : 0), handler, NULL,
328 xname);
329 } else {
330 ai->ai_ih = intr_establish_xname(irq, ipl,
331 type | (mpsafe ? IST_MPSAFE : 0), acpi_md_intr, ai,
332 xname);
333 }
334 if (ai->ai_ih == NULL) {
335 kmem_free(ai, sizeof(*ai));
336 return NULL;
337 }
338 TAILQ_INSERT_TAIL(&acpi_intrvecs, ai, ai_list);
339 } else {
340 if (ai->ai_arg == NULL) {
341 printf("ACPI: cannot share irq with NULL arg\n");
342 return NULL;
343 }
344 if (ai->ai_ipl != ipl) {
345 printf("ACPI: cannot share irq with different ipl\n");
346 return NULL;
347 }
348 if (ai->ai_type != type) {
349 printf("ACPI: cannot share edge and level interrupts\n");
350 return NULL;
351 }
352 if (ai->ai_mpsafe != mpsafe) {
353 printf("ACPI: cannot share between mpsafe/non-mpsafe\n");
354 return NULL;
355 }
356 }
357
358 ai->ai_refcnt++;
359
360 ah = kmem_zalloc(sizeof(*ah), KM_SLEEP);
361 ah->ah_fn = handler;
362 ah->ah_arg = arg;
363 TAILQ_INSERT_TAIL(&ai->ai_handlers, ah, ah_list);
364
365 return ai->ai_ih;
366 }
367
368 void
369 acpi_md_intr_disestablish(void *ih)
370 {
371 struct acpi_intrvec *ai;
372 struct acpi_intrhandler *ah;
373
374 TAILQ_FOREACH(ai, &acpi_intrvecs, ai_list) {
375 if (ai->ai_ih == ih) {
376 KASSERT(ai->ai_refcnt > 0);
377 if (ai->ai_refcnt > 1) {
378 panic("%s: cannot disestablish shared irq", __func__);
379 }
380
381 TAILQ_REMOVE(&acpi_intrvecs, ai, ai_list);
382 ah = TAILQ_FIRST(&ai->ai_handlers);
383 kmem_free(ah, sizeof(*ah));
384 intr_disestablish(ai->ai_ih);
385 kmem_free(ai, sizeof(*ai));
386 return;
387 }
388 }
389
390 panic("%s: interrupt not established", __func__);
391 }
392
393 void
394 acpi_md_intr_mask(void *ih)
395 {
396 intr_mask(ih);
397 }
398
399 void
400 acpi_md_intr_unmask(void *ih)
401 {
402 intr_unmask(ih);
403 }
404
405 int
406 acpi_md_sleep(int state)
407 {
408 printf("ERROR: ACPI sleep not implemented on this platform\n");
409 return -1;
410 }
411
412 uint32_t
413 acpi_md_pdc(void)
414 {
415 return 0;
416 }
417
418 uint32_t
419 acpi_md_ncpus(void)
420 {
421 return kcpuset_countset(kcpuset_attached);
422 }
423
424 static ACPI_STATUS
425 acpi_md_madt_probe_cpu(ACPI_SUBTABLE_HEADER *hdrp, void *aux)
426 {
427 struct acpi_softc * const sc = aux;
428
429 if (hdrp->Type == ACPI_MADT_TYPE_GENERIC_INTERRUPT)
430 config_found(sc->sc_dev, hdrp, NULL,
431 CFARGS(.iattr = "acpimadtbus"));
432
433 return AE_OK;
434 }
435
436 static ACPI_STATUS
437 acpi_md_madt_probe_gic(ACPI_SUBTABLE_HEADER *hdrp, void *aux)
438 {
439 struct acpi_softc * const sc = aux;
440
441 if (hdrp->Type == ACPI_MADT_TYPE_GENERIC_DISTRIBUTOR)
442 config_found(sc->sc_dev, hdrp, NULL,
443 CFARGS(.iattr = "acpimadtbus"));
444
445 return AE_OK;
446 }
447
448 static ACPI_STATUS
449 acpi_md_gtdt_probe(ACPI_GTDT_HEADER *hdrp, void *aux)
450 {
451 struct acpi_softc * const sc = aux;
452
453 config_found(sc->sc_dev, hdrp, NULL,
454 CFARGS(.iattr = "acpigtdtbus"));
455
456 return AE_OK;
457 }
458
459 #if NPCI > 0
460 static struct bus_space acpi_md_mcfg_bs_tag;
461
462 static int
463 acpi_md_mcfg_bs_map(void *t, bus_addr_t bpa, bus_size_t size, int flag,
464 bus_space_handle_t *bshp)
465 {
466 return arm_generic_bs_tag.bs_map(t, bpa, size,
467 flag | BUS_SPACE_MAP_NONPOSTED, bshp);
468 }
469 #endif
470
471 void
472 acpi_md_callback(struct acpi_softc *sc)
473 {
474 #if NPCI > 0
475 acpi_md_mcfg_bs_tag = arm_generic_bs_tag;
476 acpi_md_mcfg_bs_tag.bs_map = acpi_md_mcfg_bs_map;
477 acpimcfg_init(&acpi_md_mcfg_bs_tag, NULL);
478 #endif
479
480 if (acpi_madt_map() != AE_OK)
481 panic("Failed to map MADT");
482 acpi_madt_walk(acpi_md_madt_probe_cpu, sc);
483 acpi_madt_walk(acpi_md_madt_probe_gic, sc);
484 acpi_madt_unmap();
485
486 if (acpi_gtdt_map() != AE_OK)
487 panic("Failed to map GTDT");
488 acpi_gtdt_walk(acpi_md_gtdt_probe, sc);
489 acpi_gtdt_unmap();
490 }
491
492 static const char * const module_hid[] = {
493 "ACPI0004", /* Module device */
494 NULL
495 };
496
497 static ACPI_HANDLE
498 arm_acpi_dma_module(struct acpi_softc *sc, struct acpi_devnode *ad)
499 {
500 ACPI_HANDLE tmp;
501 ACPI_STATUS rv;
502
503 /*
504 * Search up the tree for a module device with a _DMA method.
505 */
506 for (; ad != NULL; ad = ad->ad_parent) {
507 if (ad->ad_devinfo->Type != ACPI_TYPE_DEVICE)
508 continue;
509 if (!acpi_match_hid(ad->ad_devinfo, module_hid))
510 continue;
511 rv = AcpiGetHandle(ad->ad_handle, "_DMA", &tmp);
512 if (ACPI_SUCCESS(rv))
513 return ad->ad_handle;
514 }
515
516 return NULL;
517 }
518
519 static void
520 arm_acpi_dma_init_ranges(struct acpi_softc *sc, struct acpi_devnode *ad,
521 struct arm32_bus_dma_tag *dmat, uint32_t flags)
522 {
523 struct acpi_resources res;
524 struct acpi_mem *mem;
525 ACPI_HANDLE module;
526 ACPI_IORT_NAMED_COMPONENT *nc;
527 ACPI_STATUS rv;
528 uintptr_t dma_mask;
529 int n;
530
531 module = arm_acpi_dma_module(sc, ad->ad_parent);
532 if (module == NULL) {
533 default_tag:
534 rv = acpi_iort_named_component(ad, &nc);
535 if (ACPI_SUCCESS(rv) && nc->MemoryAddressLimit != 0) {
536 dma_mask = __BITS(nc->MemoryAddressLimit - 1, 0);
537 } else {
538 dma_mask = UINTPTR_MAX;
539 }
540
541 /* No translation required */
542 dmat->_nranges = 1;
543 dmat->_ranges = kmem_zalloc(sizeof(*dmat->_ranges), KM_SLEEP);
544 dmat->_ranges[0].dr_sysbase = 0;
545 dmat->_ranges[0].dr_busbase = 0;
546 dmat->_ranges[0].dr_len = dma_mask;
547 dmat->_ranges[0].dr_flags = flags;
548 return;
549 }
550
551 rv = acpi_resource_parse_any(sc->sc_dev, module, "_DMA", &res,
552 &acpi_resource_parse_ops_quiet);
553 if (ACPI_FAILURE(rv)) {
554 aprint_error_dev(sc->sc_dev,
555 "failed to parse _DMA on %s: %s\n",
556 acpi_name(module), AcpiFormatException(rv));
557 goto default_tag;
558 }
559 if (res.ar_nmem == 0) {
560 acpi_resource_cleanup(&res);
561 goto default_tag;
562 }
563
564 dmat->_nranges = res.ar_nmem;
565 dmat->_ranges = kmem_zalloc(sizeof(*dmat->_ranges) * res.ar_nmem,
566 KM_SLEEP);
567
568 for (n = 0; n < res.ar_nmem; n++) {
569 mem = acpi_res_mem(&res, n);
570 dmat->_ranges[n].dr_busbase = mem->ar_base;
571 dmat->_ranges[n].dr_sysbase = mem->ar_xbase;
572 dmat->_ranges[n].dr_len = mem->ar_length;
573 dmat->_ranges[n].dr_flags = flags;
574
575 aprint_debug_dev(sc->sc_dev,
576 "%s: DMA sys %#lx-%#lx bus %#lx-%#lx%s\n",
577 acpi_name(ad->ad_handle),
578 dmat->_ranges[n].dr_sysbase,
579 dmat->_ranges[n].dr_sysbase + dmat->_ranges[n].dr_len - 1,
580 dmat->_ranges[n].dr_busbase,
581 dmat->_ranges[n].dr_busbase + dmat->_ranges[n].dr_len - 1,
582 flags ? " (coherent)" : "");
583 }
584
585 acpi_resource_cleanup(&res);
586 }
587
588 static uint32_t
589 arm_acpi_dma_flags(struct acpi_softc *sc, struct acpi_devnode *ad)
590 {
591 ACPI_INTEGER cca = 1; /* default cache coherent */
592 ACPI_STATUS rv;
593
594 for (; ad != NULL; ad = ad->ad_parent) {
595 if (ad->ad_devinfo->Type != ACPI_TYPE_DEVICE)
596 continue;
597
598 rv = acpi_eval_integer(ad->ad_handle, "_CCA", &cca);
599 if (ACPI_SUCCESS(rv))
600 break;
601 }
602
603 return cca ? _BUS_DMAMAP_COHERENT : 0;
604 }
605
606 bus_dma_tag_t
607 arm_acpi_dma32_tag(struct acpi_softc *sc, struct acpi_devnode *ad)
608 {
609 bus_dma_tag_t dmat64, dmat32;
610 int error;
611
612 if (ad->ad_dmat != NULL)
613 return ad->ad_dmat;
614
615 dmat64 = arm_acpi_dma64_tag(sc, ad);
616
617 const uint32_t flags = arm_acpi_dma_flags(sc, ad);
618 error = bus_dmatag_subregion(dmat64, 0, UINT32_MAX, &dmat32, flags);
619 if (error != 0)
620 panic("arm_acpi_dma32_tag: bus_dmatag_subregion returned %d",
621 error);
622
623 return dmat32;
624 }
625 __strong_alias(acpi_get_dma_tag,arm_acpi_dma32_tag);
626
627 bus_dma_tag_t
628 arm_acpi_dma64_tag(struct acpi_softc *sc, struct acpi_devnode *ad)
629 {
630 struct arm32_bus_dma_tag *dmat;
631
632 if (ad->ad_dmat64 != NULL)
633 return ad->ad_dmat64;
634
635 dmat = kmem_alloc(sizeof(*dmat), KM_SLEEP);
636 *dmat = arm_generic_dma_tag;
637
638 const uint32_t flags = arm_acpi_dma_flags(sc, ad);
639 arm_acpi_dma_init_ranges(sc, ad, dmat, flags);
640
641 return dmat;
642 }
643 __strong_alias(acpi_get_dma64_tag,arm_acpi_dma64_tag);
644