rmixl_intr.c revision 1.1.2.30 1 1.1.2.30 matt /* $NetBSD: rmixl_intr.c,v 1.1.2.30 2011/04/29 08:26:32 matt Exp $ */
2 1.1.2.1 cliff
3 1.1.2.1 cliff /*-
4 1.1.2.1 cliff * Copyright (c) 2007 Ruslan Ermilov and Vsevolod Lobko.
5 1.1.2.1 cliff * All rights reserved.
6 1.1.2.1 cliff *
7 1.1.2.1 cliff * Redistribution and use in source and binary forms, with or
8 1.1.2.1 cliff * without modification, are permitted provided that the following
9 1.1.2.1 cliff * conditions are met:
10 1.1.2.1 cliff * 1. Redistributions of source code must retain the above copyright
11 1.1.2.1 cliff * notice, this list of conditions and the following disclaimer.
12 1.1.2.1 cliff * 2. Redistributions in binary form must reproduce the above
13 1.1.2.1 cliff * copyright notice, this list of conditions and the following
14 1.1.2.1 cliff * disclaimer in the documentation and/or other materials provided
15 1.1.2.1 cliff * with the distribution.
16 1.1.2.1 cliff * 3. The names of the authors may not be used to endorse or promote
17 1.1.2.1 cliff * products derived from this software without specific prior
18 1.1.2.1 cliff * written permission.
19 1.1.2.1 cliff *
20 1.1.2.1 cliff * THIS SOFTWARE IS PROVIDED BY THE AUTHORS ``AS IS'' AND ANY
21 1.1.2.1 cliff * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO,
22 1.1.2.1 cliff * THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A
23 1.1.2.1 cliff * PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHORS
24 1.1.2.1 cliff * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY,
25 1.1.2.1 cliff * OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
26 1.1.2.1 cliff * PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA,
27 1.1.2.1 cliff * OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
28 1.1.2.1 cliff * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR
29 1.1.2.1 cliff * TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT
30 1.1.2.1 cliff * OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY
31 1.1.2.1 cliff * OF SUCH DAMAGE.
32 1.1.2.1 cliff */
33 1.1.2.1 cliff /*-
34 1.1.2.1 cliff * Copyright (c) 2001 The NetBSD Foundation, Inc.
35 1.1.2.1 cliff * All rights reserved.
36 1.1.2.1 cliff *
37 1.1.2.1 cliff * This code is derived from software contributed to The NetBSD Foundation
38 1.1.2.1 cliff * by Jason R. Thorpe.
39 1.1.2.1 cliff *
40 1.1.2.1 cliff * Redistribution and use in source and binary forms, with or without
41 1.1.2.1 cliff * modification, are permitted provided that the following conditions
42 1.1.2.1 cliff * are met:
43 1.1.2.1 cliff * 1. Redistributions of source code must retain the above copyright
44 1.1.2.1 cliff * notice, this list of conditions and the following disclaimer.
45 1.1.2.1 cliff * 2. Redistributions in binary form must reproduce the above copyright
46 1.1.2.1 cliff * notice, this list of conditions and the following disclaimer in the
47 1.1.2.1 cliff * documentation and/or other materials provided with the distribution.
48 1.1.2.1 cliff *
49 1.1.2.1 cliff * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
50 1.1.2.1 cliff * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
51 1.1.2.1 cliff * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
52 1.1.2.1 cliff * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
53 1.1.2.1 cliff * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
54 1.1.2.1 cliff * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
55 1.1.2.1 cliff * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
56 1.1.2.1 cliff * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
57 1.1.2.1 cliff * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
58 1.1.2.1 cliff * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
59 1.1.2.1 cliff * POSSIBILITY OF SUCH DAMAGE.
60 1.1.2.1 cliff */
61 1.1.2.1 cliff
62 1.1.2.1 cliff /*
63 1.1.2.1 cliff * Platform-specific interrupt support for the RMI XLP, XLR, XLS
64 1.1.2.1 cliff */
65 1.1.2.1 cliff
66 1.1.2.1 cliff #include <sys/cdefs.h>
67 1.1.2.30 matt __KERNEL_RCSID(0, "$NetBSD: rmixl_intr.c,v 1.1.2.30 2011/04/29 08:26:32 matt Exp $");
68 1.1.2.1 cliff
69 1.1.2.1 cliff #include "opt_ddb.h"
70 1.1.2.30 matt #include "opt_multiprocessor.h"
71 1.1.2.14 matt #define __INTR_PRIVATE
72 1.1.2.1 cliff
73 1.1.2.1 cliff #include <sys/param.h>
74 1.1.2.1 cliff #include <sys/queue.h>
75 1.1.2.1 cliff #include <sys/malloc.h>
76 1.1.2.1 cliff #include <sys/systm.h>
77 1.1.2.1 cliff #include <sys/device.h>
78 1.1.2.1 cliff #include <sys/kernel.h>
79 1.1.2.15 cliff #include <sys/atomic.h>
80 1.1.2.25 cliff #include <sys/mutex.h>
81 1.1.2.15 cliff #include <sys/cpu.h>
82 1.1.2.1 cliff
83 1.1.2.1 cliff #include <machine/bus.h>
84 1.1.2.1 cliff #include <machine/intr.h>
85 1.1.2.1 cliff
86 1.1.2.5 cliff #include <mips/cpu.h>
87 1.1.2.30 matt #include <mips/cpuset.h>
88 1.1.2.1 cliff #include <mips/locore.h>
89 1.1.2.5 cliff
90 1.1.2.1 cliff #include <mips/rmi/rmixlreg.h>
91 1.1.2.1 cliff #include <mips/rmi/rmixlvar.h>
92 1.1.2.1 cliff
93 1.1.2.15 cliff #include <mips/rmi/rmixl_cpuvar.h>
94 1.1.2.15 cliff #include <mips/rmi/rmixl_intr.h>
95 1.1.2.15 cliff
96 1.1.2.1 cliff #include <dev/pci/pcireg.h>
97 1.1.2.1 cliff #include <dev/pci/pcivar.h>
98 1.1.2.1 cliff
99 1.1.2.30 matt //#define IOINTR_DEBUG 1
100 1.1.2.4 cliff #ifdef IOINTR_DEBUG
101 1.1.2.4 cliff int iointr_debug = IOINTR_DEBUG;
102 1.1.2.4 cliff # define DPRINTF(x) do { if (iointr_debug) printf x ; } while(0)
103 1.1.2.4 cliff #else
104 1.1.2.4 cliff # define DPRINTF(x)
105 1.1.2.4 cliff #endif
106 1.1.2.4 cliff
107 1.1.2.4 cliff #define RMIXL_PICREG_READ(off) \
108 1.1.2.4 cliff RMIXL_IOREG_READ(RMIXL_IO_DEV_PIC + (off))
109 1.1.2.4 cliff #define RMIXL_PICREG_WRITE(off, val) \
110 1.1.2.4 cliff RMIXL_IOREG_WRITE(RMIXL_IO_DEV_PIC + (off), (val))
111 1.1.2.15 cliff
112 1.1.2.1 cliff /*
113 1.1.2.15 cliff * do not clear these when acking EIRR
114 1.1.2.15 cliff * (otherwise they get lost)
115 1.1.2.15 cliff */
116 1.1.2.15 cliff #define RMIXL_EIRR_PRESERVE_MASK \
117 1.1.2.15 cliff ((MIPS_INT_MASK_5|MIPS_SOFT_INT_MASK) >> 8)
118 1.1.2.1 cliff
119 1.1.2.2 cliff /*
120 1.1.2.15 cliff * IRT assignments depends on the RMI chip family
121 1.1.2.15 cliff * (XLS1xx vs. XLS2xx vs. XLS3xx vs. XLS6xx)
122 1.1.2.20 cliff * use the right display string table for the CPU that's running.
123 1.1.2.4 cliff */
124 1.1.2.4 cliff
125 1.1.2.4 cliff /*
126 1.1.2.16 cliff * rmixl_irtnames_xlrxxx
127 1.1.2.16 cliff * - use for XLRxxx
128 1.1.2.16 cliff */
129 1.1.2.16 cliff static const char * const rmixl_irtnames_xlrxxx[NIRTS] = {
130 1.1.2.20 cliff "pic int 0 (watchdog)", /* 0 */
131 1.1.2.20 cliff "pic int 1 (timer0)", /* 1 */
132 1.1.2.20 cliff "pic int 2 (timer1)", /* 2 */
133 1.1.2.20 cliff "pic int 3 (timer2)", /* 3 */
134 1.1.2.20 cliff "pic int 4 (timer3)", /* 4 */
135 1.1.2.20 cliff "pic int 5 (timer4)", /* 5 */
136 1.1.2.20 cliff "pic int 6 (timer5)", /* 6 */
137 1.1.2.20 cliff "pic int 7 (timer6)", /* 7 */
138 1.1.2.20 cliff "pic int 8 (timer7)", /* 8 */
139 1.1.2.20 cliff "pic int 9 (uart0)", /* 9 */
140 1.1.2.20 cliff "pic int 10 (uart1)", /* 10 */
141 1.1.2.20 cliff "pic int 11 (i2c0)", /* 11 */
142 1.1.2.20 cliff "pic int 12 (i2c1)", /* 12 */
143 1.1.2.20 cliff "pic int 13 (pcmcia)", /* 13 */
144 1.1.2.20 cliff "pic int 14 (gpio)", /* 14 */
145 1.1.2.20 cliff "pic int 15 (hyper)", /* 15 */
146 1.1.2.20 cliff "pic int 16 (pcix)", /* 16 */
147 1.1.2.20 cliff "pic int 17 (gmac0)", /* 17 */
148 1.1.2.20 cliff "pic int 18 (gmac1)", /* 18 */
149 1.1.2.20 cliff "pic int 19 (gmac2)", /* 19 */
150 1.1.2.20 cliff "pic int 20 (gmac3)", /* 20 */
151 1.1.2.20 cliff "pic int 21 (xgs0)", /* 21 */
152 1.1.2.20 cliff "pic int 22 (xgs1)", /* 22 */
153 1.1.2.20 cliff "pic int 23 (irq23)", /* 23 */
154 1.1.2.20 cliff "pic int 24 (hyper_fatal)", /* 24 */
155 1.1.2.20 cliff "pic int 25 (bridge_aerr)", /* 25 */
156 1.1.2.20 cliff "pic int 26 (bridge_berr)", /* 26 */
157 1.1.2.20 cliff "pic int 27 (bridge_tb)", /* 27 */
158 1.1.2.20 cliff "pic int 28 (bridge_nmi)", /* 28 */
159 1.1.2.20 cliff "pic int 29 (bridge_sram_derr)",/* 29 */
160 1.1.2.20 cliff "pic int 30 (gpio_fatal)", /* 30 */
161 1.1.2.20 cliff "pic int 31 (reserved)", /* 31 */
162 1.1.2.16 cliff };
163 1.1.2.16 cliff
164 1.1.2.16 cliff /*
165 1.1.2.19 cliff * rmixl_irtnames_xls2xx
166 1.1.2.19 cliff * - use for XLS2xx
167 1.1.2.19 cliff */
168 1.1.2.19 cliff static const char * const rmixl_irtnames_xls2xx[NIRTS] = {
169 1.1.2.20 cliff "pic int 0 (watchdog)", /* 0 */
170 1.1.2.20 cliff "pic int 1 (timer0)", /* 1 */
171 1.1.2.20 cliff "pic int 2 (timer1)", /* 2 */
172 1.1.2.20 cliff "pic int 3 (timer2)", /* 3 */
173 1.1.2.20 cliff "pic int 4 (timer3)", /* 4 */
174 1.1.2.20 cliff "pic int 5 (timer4)", /* 5 */
175 1.1.2.20 cliff "pic int 6 (timer5)", /* 6 */
176 1.1.2.20 cliff "pic int 7 (timer6)", /* 7 */
177 1.1.2.20 cliff "pic int 8 (timer7)", /* 8 */
178 1.1.2.20 cliff "pic int 9 (uart0)", /* 9 */
179 1.1.2.20 cliff "pic int 10 (uart1)", /* 10 */
180 1.1.2.20 cliff "pic int 11 (i2c0)", /* 11 */
181 1.1.2.20 cliff "pic int 12 (i2c1)", /* 12 */
182 1.1.2.20 cliff "pic int 13 (pcmcia)", /* 13 */
183 1.1.2.20 cliff "pic int 14 (gpio_a)", /* 14 */
184 1.1.2.20 cliff "pic int 15 (irq15)", /* 15 */
185 1.1.2.20 cliff "pic int 16 (bridge_tb)", /* 16 */
186 1.1.2.20 cliff "pic int 17 (gmac0)", /* 17 */
187 1.1.2.20 cliff "pic int 18 (gmac1)", /* 18 */
188 1.1.2.20 cliff "pic int 19 (gmac2)", /* 19 */
189 1.1.2.20 cliff "pic int 20 (gmac3)", /* 20 */
190 1.1.2.20 cliff "pic int 21 (irq21)", /* 21 */
191 1.1.2.20 cliff "pic int 22 (irq22)", /* 22 */
192 1.1.2.20 cliff "pic int 23 (pcie_link2)", /* 23 */
193 1.1.2.20 cliff "pic int 24 (pcie_link3)", /* 24 */
194 1.1.2.20 cliff "pic int 25 (bridge_err)", /* 25 */
195 1.1.2.20 cliff "pic int 26 (pcie_link0)", /* 26 */
196 1.1.2.20 cliff "pic int 27 (pcie_link1)", /* 27 */
197 1.1.2.20 cliff "pic int 28 (irq28)", /* 28 */
198 1.1.2.20 cliff "pic int 29 (pcie_err)", /* 29 */
199 1.1.2.20 cliff "pic int 30 (gpio_b)", /* 30 */
200 1.1.2.20 cliff "pic int 31 (usb)", /* 31 */
201 1.1.2.19 cliff };
202 1.1.2.19 cliff
203 1.1.2.19 cliff /*
204 1.1.2.15 cliff * rmixl_irtnames_xls1xx
205 1.1.2.19 cliff * - use for XLS1xx, XLS4xx-Lite
206 1.1.2.2 cliff */
207 1.1.2.15 cliff static const char * const rmixl_irtnames_xls1xx[NIRTS] = {
208 1.1.2.20 cliff "pic int 0 (watchdog)", /* 0 */
209 1.1.2.20 cliff "pic int 1 (timer0)", /* 1 */
210 1.1.2.20 cliff "pic int 2 (timer1)", /* 2 */
211 1.1.2.20 cliff "pic int 3 (timer2)", /* 3 */
212 1.1.2.20 cliff "pic int 4 (timer3)", /* 4 */
213 1.1.2.20 cliff "pic int 5 (timer4)", /* 5 */
214 1.1.2.20 cliff "pic int 6 (timer5)", /* 6 */
215 1.1.2.20 cliff "pic int 7 (timer6)", /* 7 */
216 1.1.2.20 cliff "pic int 8 (timer7)", /* 8 */
217 1.1.2.20 cliff "pic int 9 (uart0)", /* 9 */
218 1.1.2.20 cliff "pic int 10 (uart1)", /* 10 */
219 1.1.2.20 cliff "pic int 11 (i2c0)", /* 11 */
220 1.1.2.20 cliff "pic int 12 (i2c1)", /* 12 */
221 1.1.2.20 cliff "pic int 13 (pcmcia)", /* 13 */
222 1.1.2.20 cliff "pic int 14 (gpio_a)", /* 14 */
223 1.1.2.20 cliff "pic int 15 (irq15)", /* 15 */
224 1.1.2.20 cliff "pic int 16 (bridge_tb)", /* 16 */
225 1.1.2.20 cliff "pic int 17 (gmac0)", /* 17 */
226 1.1.2.20 cliff "pic int 18 (gmac1)", /* 18 */
227 1.1.2.20 cliff "pic int 19 (gmac2)", /* 19 */
228 1.1.2.20 cliff "pic int 20 (gmac3)", /* 20 */
229 1.1.2.20 cliff "pic int 21 (irq21)", /* 21 */
230 1.1.2.20 cliff "pic int 22 (irq22)", /* 22 */
231 1.1.2.20 cliff "pic int 23 (irq23)", /* 23 */
232 1.1.2.20 cliff "pic int 24 (irq24)", /* 24 */
233 1.1.2.20 cliff "pic int 25 (bridge_err)", /* 25 */
234 1.1.2.20 cliff "pic int 26 (pcie_link0)", /* 26 */
235 1.1.2.20 cliff "pic int 27 (pcie_link1)", /* 27 */
236 1.1.2.20 cliff "pic int 28 (irq28)", /* 28 */
237 1.1.2.20 cliff "pic int 29 (pcie_err)", /* 29 */
238 1.1.2.20 cliff "pic int 30 (gpio_b)", /* 30 */
239 1.1.2.20 cliff "pic int 31 (usb)", /* 31 */
240 1.1.2.1 cliff };
241 1.1.2.1 cliff
242 1.1.2.2 cliff /*
243 1.1.2.15 cliff * rmixl_irtnames_xls4xx:
244 1.1.2.4 cliff * - use for XLS4xx, XLS6xx
245 1.1.2.4 cliff */
246 1.1.2.15 cliff static const char * const rmixl_irtnames_xls4xx[NIRTS] = {
247 1.1.2.20 cliff "pic int 0 (watchdog)", /* 0 */
248 1.1.2.20 cliff "pic int 1 (timer0)", /* 1 */
249 1.1.2.20 cliff "pic int 2 (timer1)", /* 2 */
250 1.1.2.20 cliff "pic int 3 (timer2)", /* 3 */
251 1.1.2.20 cliff "pic int 4 (timer3)", /* 4 */
252 1.1.2.20 cliff "pic int 5 (timer4)", /* 5 */
253 1.1.2.20 cliff "pic int 6 (timer5)", /* 6 */
254 1.1.2.20 cliff "pic int 7 (timer6)", /* 7 */
255 1.1.2.20 cliff "pic int 8 (timer7)", /* 8 */
256 1.1.2.20 cliff "pic int 9 (uart0)", /* 9 */
257 1.1.2.20 cliff "pic int 10 (uart1)", /* 10 */
258 1.1.2.20 cliff "pic int 11 (i2c0)", /* 11 */
259 1.1.2.20 cliff "pic int 12 (i2c1)", /* 12 */
260 1.1.2.20 cliff "pic int 13 (pcmcia)", /* 13 */
261 1.1.2.20 cliff "pic int 14 (gpio_a)", /* 14 */
262 1.1.2.20 cliff "pic int 15 (irq15)", /* 15 */
263 1.1.2.20 cliff "pic int 16 (bridge_tb)", /* 16 */
264 1.1.2.20 cliff "pic int 17 (gmac0)", /* 17 */
265 1.1.2.20 cliff "pic int 18 (gmac1)", /* 18 */
266 1.1.2.20 cliff "pic int 19 (gmac2)", /* 19 */
267 1.1.2.20 cliff "pic int 20 (gmac3)", /* 20 */
268 1.1.2.20 cliff "pic int 21 (irq21)", /* 21 */
269 1.1.2.20 cliff "pic int 22 (irq22)", /* 22 */
270 1.1.2.20 cliff "pic int 23 (irq23)", /* 23 */
271 1.1.2.20 cliff "pic int 24 (irq24)", /* 24 */
272 1.1.2.20 cliff "pic int 25 (bridge_err)", /* 25 */
273 1.1.2.20 cliff "pic int 26 (pcie_link0)", /* 26 */
274 1.1.2.20 cliff "pic int 27 (pcie_link1)", /* 27 */
275 1.1.2.20 cliff "pic int 28 (pcie_link2)", /* 28 */
276 1.1.2.20 cliff "pic int 29 (pcie_link3)", /* 29 */
277 1.1.2.20 cliff "pic int 30 (gpio_b)", /* 30 */
278 1.1.2.20 cliff "pic int 31 (usb)", /* 31 */
279 1.1.2.4 cliff };
280 1.1.2.4 cliff
281 1.1.2.4 cliff /*
282 1.1.2.15 cliff * rmixl_vecnames_common:
283 1.1.2.4 cliff * - use for unknown cpu implementation
284 1.1.2.15 cliff * - covers all vectors, not just IRT intrs
285 1.1.2.4 cliff */
286 1.1.2.15 cliff static const char * const rmixl_vecnames_common[NINTRVECS] = {
287 1.1.2.20 cliff "vec 0", /* 0 */
288 1.1.2.20 cliff "vec 1", /* 1 */
289 1.1.2.20 cliff "vec 2", /* 2 */
290 1.1.2.20 cliff "vec 3", /* 3 */
291 1.1.2.20 cliff "vec 4", /* 4 */
292 1.1.2.20 cliff "vec 5", /* 5 */
293 1.1.2.20 cliff "vec 6", /* 6 */
294 1.1.2.20 cliff "vec 7", /* 7 */
295 1.1.2.30 matt "vec 8 (ipi 0)", /* 8 */
296 1.1.2.30 matt "vec 9 (ipi 1)", /* 9 */
297 1.1.2.30 matt "vec 10 (ipi 2)", /* 10 */
298 1.1.2.30 matt "vec 11 (ipi 3)", /* 11 */
299 1.1.2.30 matt "vec 12 (ipi 4)", /* 12 */
300 1.1.2.30 matt "vec 13 (ipi 5)", /* 13 */
301 1.1.2.30 matt "vec 14 (ipi 6)", /* 14 */
302 1.1.2.30 matt "vec 15 (fmn)", /* 15 */
303 1.1.2.30 matt "vec 16", /* 16 */
304 1.1.2.20 cliff "vec 17", /* 17 */
305 1.1.2.20 cliff "vec 18", /* 18 */
306 1.1.2.20 cliff "vec 19", /* 19 */
307 1.1.2.20 cliff "vec 20", /* 20 */
308 1.1.2.20 cliff "vec 21", /* 21 */
309 1.1.2.20 cliff "vec 22", /* 22 */
310 1.1.2.20 cliff "vec 23", /* 23 */
311 1.1.2.20 cliff "vec 24", /* 24 */
312 1.1.2.20 cliff "vec 25", /* 25 */
313 1.1.2.20 cliff "vec 26", /* 26 */
314 1.1.2.20 cliff "vec 27", /* 27 */
315 1.1.2.20 cliff "vec 28", /* 28 */
316 1.1.2.20 cliff "vec 29", /* 29 */
317 1.1.2.20 cliff "vec 30", /* 30 */
318 1.1.2.20 cliff "vec 31", /* 31 */
319 1.1.2.20 cliff "vec 32", /* 32 */
320 1.1.2.20 cliff "vec 33", /* 33 */
321 1.1.2.20 cliff "vec 34", /* 34 */
322 1.1.2.20 cliff "vec 35", /* 35 */
323 1.1.2.20 cliff "vec 36", /* 36 */
324 1.1.2.20 cliff "vec 37", /* 37 */
325 1.1.2.20 cliff "vec 38", /* 38 */
326 1.1.2.20 cliff "vec 39", /* 39 */
327 1.1.2.20 cliff "vec 40", /* 40 */
328 1.1.2.20 cliff "vec 41", /* 41 */
329 1.1.2.20 cliff "vec 42", /* 42 */
330 1.1.2.20 cliff "vec 43", /* 43 */
331 1.1.2.20 cliff "vec 44", /* 44 */
332 1.1.2.20 cliff "vec 45", /* 45 */
333 1.1.2.20 cliff "vec 46", /* 46 */
334 1.1.2.20 cliff "vec 47", /* 47 */
335 1.1.2.20 cliff "vec 48", /* 48 */
336 1.1.2.20 cliff "vec 49", /* 49 */
337 1.1.2.20 cliff "vec 50", /* 50 */
338 1.1.2.20 cliff "vec 51", /* 51 */
339 1.1.2.20 cliff "vec 52", /* 52 */
340 1.1.2.20 cliff "vec 53", /* 53 */
341 1.1.2.20 cliff "vec 54", /* 54 */
342 1.1.2.20 cliff "vec 55", /* 55 */
343 1.1.2.20 cliff "vec 56", /* 56 */
344 1.1.2.20 cliff "vec 57", /* 57 */
345 1.1.2.20 cliff "vec 58", /* 58 */
346 1.1.2.20 cliff "vec 59", /* 59 */
347 1.1.2.20 cliff "vec 60", /* 60 */
348 1.1.2.20 cliff "vec 61", /* 61 */
349 1.1.2.20 cliff "vec 62", /* 63 */
350 1.1.2.20 cliff "vec 63", /* 63 */
351 1.1.2.4 cliff };
352 1.1.2.4 cliff
353 1.1.2.4 cliff /*
354 1.1.2.15 cliff * mask of CPUs attached
355 1.1.2.30 matt * once they are attached, this var is read-only so mp safe
356 1.1.2.2 cliff */
357 1.1.2.15 cliff static uint32_t cpu_present_mask;
358 1.1.2.1 cliff
359 1.1.2.30 matt kmutex_t *rmixl_ipi_lock; /* covers RMIXL_PIC_IPIBASE */
360 1.1.2.30 matt kmutex_t *rmixl_intr_lock; /* covers rest of PIC, and rmixl_intrhand[] */
361 1.1.2.30 matt rmixl_intrhand_t rmixl_intrhand[NINTRVECS];
362 1.1.2.1 cliff
363 1.1.2.15 cliff #ifdef DIAGNOSTIC
364 1.1.2.15 cliff static int rmixl_pic_init_done;
365 1.1.2.15 cliff #endif
366 1.1.2.2 cliff
367 1.1.2.1 cliff
368 1.1.2.16 cliff static const char *rmixl_intr_string_xlr(int);
369 1.1.2.16 cliff static const char *rmixl_intr_string_xls(int);
370 1.1.2.16 cliff static uint32_t rmixl_irt_thread_mask(int);
371 1.1.2.15 cliff static void rmixl_irt_init(int);
372 1.1.2.15 cliff static void rmixl_irt_disestablish(int);
373 1.1.2.20 cliff static void rmixl_irt_establish(int, int, int,
374 1.1.2.15 cliff rmixl_intr_trigger_t, rmixl_intr_polarity_t);
375 1.1.2.2 cliff
376 1.1.2.15 cliff #ifdef MULTIPROCESSOR
377 1.1.2.15 cliff static int rmixl_send_ipi(struct cpu_info *, int);
378 1.1.2.15 cliff static int rmixl_ipi_intr(void *);
379 1.1.2.15 cliff #endif
380 1.1.2.15 cliff
381 1.1.2.23 rmind #if defined(DIAGNOSTIC) || defined(IOINTR_DEBUG) || defined(DDB)
382 1.1.2.20 cliff int rmixl_intrhand_print_subr(int);
383 1.1.2.20 cliff int rmixl_intrhand_print(void);
384 1.1.2.20 cliff int rmixl_irt_print(void);
385 1.1.2.20 cliff void rmixl_ipl_eimr_map_print(void);
386 1.1.2.4 cliff #endif
387 1.1.2.2 cliff
388 1.1.2.6 cliff
389 1.1.2.15 cliff static inline u_int
390 1.1.2.15 cliff dclz(uint64_t val)
391 1.1.2.15 cliff {
392 1.1.2.15 cliff int nlz;
393 1.1.2.6 cliff
394 1.1.2.15 cliff asm volatile("dclz %0, %1;"
395 1.1.2.15 cliff : "=r"(nlz) : "r"(val));
396 1.1.2.15 cliff
397 1.1.2.15 cliff return nlz;
398 1.1.2.15 cliff }
399 1.1.2.6 cliff
400 1.1.2.1 cliff void
401 1.1.2.1 cliff evbmips_intr_init(void)
402 1.1.2.1 cliff {
403 1.1.2.2 cliff uint32_t r;
404 1.1.2.1 cliff
405 1.1.2.16 cliff KASSERT(cpu_rmixlr(mips_options.mips_cpu)
406 1.1.2.16 cliff || cpu_rmixls(mips_options.mips_cpu));
407 1.1.2.5 cliff
408 1.1.2.4 cliff
409 1.1.2.15 cliff #ifdef DIAGNOSTIC
410 1.1.2.15 cliff if (rmixl_pic_init_done != 0)
411 1.1.2.15 cliff panic("%s: rmixl_pic_init_done %d",
412 1.1.2.15 cliff __func__, rmixl_pic_init_done);
413 1.1.2.15 cliff #endif
414 1.1.2.1 cliff
415 1.1.2.28 cliff rmixl_ipi_lock = mutex_obj_alloc(MUTEX_DEFAULT, IPL_HIGH);
416 1.1.2.28 cliff rmixl_intr_lock = mutex_obj_alloc(MUTEX_DEFAULT, IPL_HIGH);
417 1.1.2.25 cliff
418 1.1.2.28 cliff mutex_enter(rmixl_intr_lock);
419 1.1.2.25 cliff
420 1.1.2.15 cliff /*
421 1.1.2.15 cliff * initialize (zero) all IRT Entries in the PIC
422 1.1.2.15 cliff */
423 1.1.2.30 matt for (u_int i = 0; i < NIRTS; i++) {
424 1.1.2.15 cliff rmixl_irt_init(i);
425 1.1.2.30 matt }
426 1.1.2.1 cliff
427 1.1.2.2 cliff /*
428 1.1.2.4 cliff * disable watchdog NMI, timers
429 1.1.2.4 cliff *
430 1.1.2.4 cliff * XXX
431 1.1.2.4 cliff * WATCHDOG_ENB is preserved because clearing it causes
432 1.1.2.4 cliff * hang on the XLS616 (but not on the XLS408)
433 1.1.2.4 cliff */
434 1.1.2.4 cliff r = RMIXL_PICREG_READ(RMIXL_PIC_CONTROL);
435 1.1.2.4 cliff r &= RMIXL_PIC_CONTROL_RESV|RMIXL_PIC_CONTROL_WATCHDOG_ENB;
436 1.1.2.4 cliff RMIXL_PICREG_WRITE(RMIXL_PIC_CONTROL, r);
437 1.1.2.2 cliff
438 1.1.2.4 cliff #ifdef DIAGNOSTIC
439 1.1.2.15 cliff rmixl_pic_init_done = 1;
440 1.1.2.4 cliff #endif
441 1.1.2.28 cliff mutex_exit(rmixl_intr_lock);
442 1.1.2.15 cliff
443 1.1.2.4 cliff }
444 1.1.2.4 cliff
445 1.1.2.15 cliff /*
446 1.1.2.15 cliff * establish vector for mips3 count/compare clock interrupt
447 1.1.2.15 cliff * this ensures we enable in EIRR,
448 1.1.2.15 cliff * even though cpu_intr() handles the interrupt
449 1.1.2.17 cliff * note the 'mpsafe' arg here is a placeholder only
450 1.1.2.15 cliff */
451 1.1.2.27 cliff void
452 1.1.2.15 cliff rmixl_intr_init_clk(void)
453 1.1.2.15 cliff {
454 1.1.2.30 matt const int vec = ffs(MIPS_INT_MASK_5 >> MIPS_INT_MASK_SHIFT) - 1;
455 1.1.2.25 cliff
456 1.1.2.28 cliff mutex_enter(rmixl_intr_lock);
457 1.1.2.25 cliff
458 1.1.2.17 cliff void *ih = rmixl_vec_establish(vec, 0, IPL_SCHED, NULL, NULL, false);
459 1.1.2.15 cliff if (ih == NULL)
460 1.1.2.15 cliff panic("%s: establish vec %d failed", __func__, vec);
461 1.1.2.25 cliff
462 1.1.2.28 cliff mutex_exit(rmixl_intr_lock);
463 1.1.2.15 cliff }
464 1.1.2.15 cliff
465 1.1.2.15 cliff #ifdef MULTIPROCESSOR
466 1.1.2.15 cliff /*
467 1.1.2.15 cliff * establish IPI interrupt and send function
468 1.1.2.15 cliff */
469 1.1.2.27 cliff void
470 1.1.2.15 cliff rmixl_intr_init_ipi(void)
471 1.1.2.15 cliff {
472 1.1.2.28 cliff mutex_enter(rmixl_intr_lock);
473 1.1.2.25 cliff
474 1.1.2.30 matt for (u_int ipi = 0; ipi < NIPIS; ipi++) {
475 1.1.2.30 matt const u_int vec = RMIXL_INTRVEC_IPI + ipi;
476 1.1.2.30 matt void * const ih = rmixl_vec_establish(vec, -1, IPL_SCHED,
477 1.1.2.25 cliff rmixl_ipi_intr, (void *)(uintptr_t)ipi, true);
478 1.1.2.25 cliff if (ih == NULL)
479 1.1.2.25 cliff panic("%s: establish ipi %d at vec %d failed",
480 1.1.2.25 cliff __func__, ipi, vec);
481 1.1.2.25 cliff }
482 1.1.2.15 cliff
483 1.1.2.15 cliff mips_locoresw.lsw_send_ipi = rmixl_send_ipi;
484 1.1.2.15 cliff
485 1.1.2.28 cliff mutex_exit(rmixl_intr_lock);
486 1.1.2.15 cliff }
487 1.1.2.15 cliff #endif /* MULTIPROCESSOR */
488 1.1.2.15 cliff
489 1.1.2.15 cliff /*
490 1.1.2.15 cliff * initialize per-cpu interrupt stuff in softc
491 1.1.2.15 cliff * accumulate per-cpu bits in 'cpu_present_mask'
492 1.1.2.15 cliff */
493 1.1.2.15 cliff void
494 1.1.2.15 cliff rmixl_intr_init_cpu(struct cpu_info *ci)
495 1.1.2.15 cliff {
496 1.1.2.15 cliff struct rmixl_cpu_softc *sc = (void *)ci->ci_softc;
497 1.1.2.21 cliff
498 1.1.2.15 cliff KASSERT(sc != NULL);
499 1.1.2.15 cliff
500 1.1.2.15 cliff for (int vec=0; vec < NINTRVECS; vec++)
501 1.1.2.15 cliff evcnt_attach_dynamic(&sc->sc_vec_evcnts[vec],
502 1.1.2.15 cliff EVCNT_TYPE_INTR, NULL,
503 1.1.2.15 cliff device_xname(sc->sc_dev),
504 1.1.2.15 cliff rmixl_intr_string(vec));
505 1.1.2.15 cliff
506 1.1.2.26 cliff KASSERT(cpu_index(ci) < (sizeof(cpu_present_mask) * 8));
507 1.1.2.26 cliff atomic_or_32((volatile uint32_t *)&cpu_present_mask, 1 << cpu_index(ci));
508 1.1.2.15 cliff }
509 1.1.2.15 cliff
510 1.1.2.15 cliff /*
511 1.1.2.15 cliff * rmixl_intr_string - return pointer to display name of a PIC-based interrupt
512 1.1.2.15 cliff */
513 1.1.2.4 cliff const char *
514 1.1.2.20 cliff rmixl_intr_string(int vec)
515 1.1.2.4 cliff {
516 1.1.2.20 cliff int irt;
517 1.1.2.20 cliff
518 1.1.2.20 cliff if (vec < 0 || vec >= NINTRVECS)
519 1.1.2.20 cliff panic("%s: vec index %d out of range, max %d",
520 1.1.2.20 cliff __func__, vec, NINTRVECS - 1);
521 1.1.2.15 cliff
522 1.1.2.20 cliff if (! RMIXL_VECTOR_IS_IRT(vec))
523 1.1.2.20 cliff return rmixl_vecnames_common[vec];
524 1.1.2.4 cliff
525 1.1.2.20 cliff irt = RMIXL_VECTOR_IRT(vec);
526 1.1.2.16 cliff switch(cpu_rmixl_chip_type(mips_options.mips_cpu)) {
527 1.1.2.16 cliff case CIDFL_RMI_TYPE_XLR:
528 1.1.2.20 cliff return rmixl_intr_string_xlr(irt);
529 1.1.2.16 cliff case CIDFL_RMI_TYPE_XLS:
530 1.1.2.20 cliff return rmixl_intr_string_xls(irt);
531 1.1.2.16 cliff case CIDFL_RMI_TYPE_XLP:
532 1.1.2.16 cliff panic("%s: RMI XLP not yet supported", __func__);
533 1.1.2.16 cliff }
534 1.1.2.16 cliff
535 1.1.2.16 cliff return "undefined"; /* appease gcc */
536 1.1.2.16 cliff }
537 1.1.2.16 cliff
538 1.1.2.16 cliff static const char *
539 1.1.2.20 cliff rmixl_intr_string_xlr(int irt)
540 1.1.2.16 cliff {
541 1.1.2.20 cliff return rmixl_irtnames_xlrxxx[irt];
542 1.1.2.16 cliff }
543 1.1.2.16 cliff
544 1.1.2.16 cliff static const char *
545 1.1.2.20 cliff rmixl_intr_string_xls(int irt)
546 1.1.2.16 cliff {
547 1.1.2.16 cliff const char *name;
548 1.1.2.16 cliff
549 1.1.2.7 matt switch (MIPS_PRID_IMPL(mips_options.mips_cpu_id)) {
550 1.1.2.4 cliff case MIPS_XLS104:
551 1.1.2.4 cliff case MIPS_XLS108:
552 1.1.2.4 cliff case MIPS_XLS404LITE:
553 1.1.2.4 cliff case MIPS_XLS408LITE:
554 1.1.2.20 cliff name = rmixl_irtnames_xls1xx[irt];
555 1.1.2.4 cliff break;
556 1.1.2.19 cliff case MIPS_XLS204:
557 1.1.2.19 cliff case MIPS_XLS208:
558 1.1.2.20 cliff name = rmixl_irtnames_xls2xx[irt];
559 1.1.2.19 cliff break;
560 1.1.2.8 cliff case MIPS_XLS404:
561 1.1.2.8 cliff case MIPS_XLS408:
562 1.1.2.8 cliff case MIPS_XLS416:
563 1.1.2.8 cliff case MIPS_XLS608:
564 1.1.2.8 cliff case MIPS_XLS616:
565 1.1.2.20 cliff name = rmixl_irtnames_xls4xx[irt];
566 1.1.2.4 cliff break;
567 1.1.2.4 cliff default:
568 1.1.2.20 cliff name = rmixl_vecnames_common[RMIXL_IRT_VECTOR(irt)];
569 1.1.2.4 cliff break;
570 1.1.2.4 cliff }
571 1.1.2.4 cliff
572 1.1.2.4 cliff return name;
573 1.1.2.1 cliff }
574 1.1.2.1 cliff
575 1.1.2.6 cliff /*
576 1.1.2.15 cliff * rmixl_irt_thread_mask
577 1.1.2.15 cliff *
578 1.1.2.15 cliff * given a bitmask of cpus, return a, IRT thread mask
579 1.1.2.6 cliff */
580 1.1.2.15 cliff static uint32_t
581 1.1.2.15 cliff rmixl_irt_thread_mask(int cpumask)
582 1.1.2.6 cliff {
583 1.1.2.15 cliff uint32_t irtc0;
584 1.1.2.15 cliff
585 1.1.2.15 cliff #if defined(MULTIPROCESSOR)
586 1.1.2.15 cliff #ifndef NOTYET
587 1.1.2.15 cliff if (cpumask == -1)
588 1.1.2.15 cliff return 1; /* XXX TMP FIXME */
589 1.1.2.15 cliff #endif
590 1.1.2.8 cliff
591 1.1.2.8 cliff /*
592 1.1.2.15 cliff * discount cpus not present
593 1.1.2.8 cliff */
594 1.1.2.15 cliff cpumask &= cpu_present_mask;
595 1.1.2.15 cliff
596 1.1.2.8 cliff switch (MIPS_PRID_IMPL(mips_options.mips_cpu_id)) {
597 1.1.2.8 cliff case MIPS_XLS104:
598 1.1.2.8 cliff case MIPS_XLS204:
599 1.1.2.8 cliff case MIPS_XLS404:
600 1.1.2.8 cliff case MIPS_XLS404LITE:
601 1.1.2.15 cliff irtc0 = ((cpumask >> 2) << 4) | (cpumask & __BITS(1,0));
602 1.1.2.15 cliff irtc0 &= (__BITS(5,4) | __BITS(1,0));
603 1.1.2.8 cliff break;
604 1.1.2.8 cliff case MIPS_XLS108:
605 1.1.2.8 cliff case MIPS_XLS208:
606 1.1.2.8 cliff case MIPS_XLS408:
607 1.1.2.8 cliff case MIPS_XLS408LITE:
608 1.1.2.8 cliff case MIPS_XLS608:
609 1.1.2.15 cliff irtc0 = cpumask & __BITS(7,0);
610 1.1.2.8 cliff break;
611 1.1.2.8 cliff case MIPS_XLS416:
612 1.1.2.8 cliff case MIPS_XLS616:
613 1.1.2.15 cliff irtc0 = cpumask & __BITS(15,0);
614 1.1.2.8 cliff break;
615 1.1.2.8 cliff default:
616 1.1.2.8 cliff panic("%s: unknown cpu ID %#x\n", __func__,
617 1.1.2.8 cliff mips_options.mips_cpu_id);
618 1.1.2.8 cliff }
619 1.1.2.8 cliff #else
620 1.1.2.15 cliff irtc0 = 1;
621 1.1.2.15 cliff #endif /* MULTIPROCESSOR */
622 1.1.2.15 cliff
623 1.1.2.15 cliff return irtc0;
624 1.1.2.15 cliff }
625 1.1.2.15 cliff
626 1.1.2.15 cliff /*
627 1.1.2.15 cliff * rmixl_irt_init
628 1.1.2.20 cliff * - initialize IRT Entry for given index
629 1.1.2.15 cliff * - unmask Thread#0 in low word (assume we only have 1 thread)
630 1.1.2.15 cliff */
631 1.1.2.15 cliff static void
632 1.1.2.20 cliff rmixl_irt_init(int irt)
633 1.1.2.15 cliff {
634 1.1.2.20 cliff KASSERT(irt < NIRTS);
635 1.1.2.20 cliff RMIXL_PICREG_WRITE(RMIXL_PIC_IRTENTRYC1(irt), 0); /* high word */
636 1.1.2.20 cliff RMIXL_PICREG_WRITE(RMIXL_PIC_IRTENTRYC0(irt), 0); /* low word */
637 1.1.2.6 cliff }
638 1.1.2.6 cliff
639 1.1.2.6 cliff /*
640 1.1.2.15 cliff * rmixl_irt_disestablish
641 1.1.2.20 cliff * - invalidate IRT Entry for given index
642 1.1.2.6 cliff */
643 1.1.2.6 cliff static void
644 1.1.2.20 cliff rmixl_irt_disestablish(int irt)
645 1.1.2.6 cliff {
646 1.1.2.28 cliff KASSERT(mutex_owned(rmixl_intr_lock));
647 1.1.2.20 cliff DPRINTF(("%s: irt %d, irtc1 %#x\n", __func__, irt, 0));
648 1.1.2.20 cliff rmixl_irt_init(irt);
649 1.1.2.6 cliff }
650 1.1.2.6 cliff
651 1.1.2.6 cliff /*
652 1.1.2.15 cliff * rmixl_irt_establish
653 1.1.2.20 cliff * - construct an IRT Entry for irt and write to PIC
654 1.1.2.6 cliff */
655 1.1.2.6 cliff static void
656 1.1.2.20 cliff rmixl_irt_establish(int irt, int vec, int cpumask, rmixl_intr_trigger_t trigger,
657 1.1.2.15 cliff rmixl_intr_polarity_t polarity)
658 1.1.2.6 cliff {
659 1.1.2.6 cliff uint32_t irtc1;
660 1.1.2.15 cliff uint32_t irtc0;
661 1.1.2.15 cliff
662 1.1.2.28 cliff KASSERT(mutex_owned(rmixl_intr_lock));
663 1.1.2.25 cliff
664 1.1.2.20 cliff if (irt >= NIRTS)
665 1.1.2.20 cliff panic("%s: bad irt %d\n", __func__, irt);
666 1.1.2.20 cliff
667 1.1.2.20 cliff if (! RMIXL_VECTOR_IS_IRT(vec))
668 1.1.2.20 cliff panic("%s: bad vec %d\n", __func__, vec);
669 1.1.2.20 cliff
670 1.1.2.15 cliff switch (trigger) {
671 1.1.2.15 cliff case RMIXL_TRIG_EDGE:
672 1.1.2.15 cliff case RMIXL_TRIG_LEVEL:
673 1.1.2.15 cliff break;
674 1.1.2.15 cliff default:
675 1.1.2.15 cliff panic("%s: bad trigger %d\n", __func__, trigger);
676 1.1.2.15 cliff }
677 1.1.2.15 cliff
678 1.1.2.15 cliff switch (polarity) {
679 1.1.2.15 cliff case RMIXL_POLR_RISING:
680 1.1.2.15 cliff case RMIXL_POLR_HIGH:
681 1.1.2.15 cliff case RMIXL_POLR_FALLING:
682 1.1.2.15 cliff case RMIXL_POLR_LOW:
683 1.1.2.15 cliff break;
684 1.1.2.15 cliff default:
685 1.1.2.15 cliff panic("%s: bad polarity %d\n", __func__, polarity);
686 1.1.2.15 cliff }
687 1.1.2.15 cliff
688 1.1.2.15 cliff /*
689 1.1.2.15 cliff * XXX IRT entries are not shared
690 1.1.2.15 cliff */
691 1.1.2.20 cliff KASSERT(RMIXL_PICREG_READ(RMIXL_PIC_IRTENTRYC0(irt)) == 0);
692 1.1.2.20 cliff KASSERT(RMIXL_PICREG_READ(RMIXL_PIC_IRTENTRYC1(irt)) == 0);
693 1.1.2.15 cliff
694 1.1.2.15 cliff irtc0 = rmixl_irt_thread_mask(cpumask);
695 1.1.2.6 cliff
696 1.1.2.6 cliff irtc1 = RMIXL_PIC_IRTENTRYC1_VALID;
697 1.1.2.6 cliff irtc1 |= RMIXL_PIC_IRTENTRYC1_GL; /* local */
698 1.1.2.30 matt KASSERT((irtc1 & RMIXL_PIC_IRTENTRYC1_NMI) == 0);
699 1.1.2.6 cliff
700 1.1.2.15 cliff if (trigger == RMIXL_TRIG_LEVEL)
701 1.1.2.6 cliff irtc1 |= RMIXL_PIC_IRTENTRYC1_TRG;
702 1.1.2.30 matt KASSERT((irtc1 & RMIXL_PIC_IRTENTRYC1_NMI) == 0);
703 1.1.2.6 cliff
704 1.1.2.15 cliff if ((polarity == RMIXL_POLR_FALLING) || (polarity == RMIXL_POLR_LOW))
705 1.1.2.6 cliff irtc1 |= RMIXL_PIC_IRTENTRYC1_P;
706 1.1.2.30 matt KASSERT((irtc1 & RMIXL_PIC_IRTENTRYC1_NMI) == 0);
707 1.1.2.6 cliff
708 1.1.2.20 cliff irtc1 |= vec; /* vector in EIRR */
709 1.1.2.30 matt KASSERT((irtc1 & RMIXL_PIC_IRTENTRYC1_NMI) == 0);
710 1.1.2.6 cliff
711 1.1.2.6 cliff /*
712 1.1.2.15 cliff * write IRT Entry to PIC
713 1.1.2.6 cliff */
714 1.1.2.30 matt DPRINTF(("%s: vec %d (%#x), irt %d, irtc0 %#x, irtc1 %#x\n",
715 1.1.2.30 matt __func__, vec, vec, irt, irtc0, irtc1));
716 1.1.2.20 cliff RMIXL_PICREG_WRITE(RMIXL_PIC_IRTENTRYC0(irt), irtc0); /* low word */
717 1.1.2.20 cliff RMIXL_PICREG_WRITE(RMIXL_PIC_IRTENTRYC1(irt), irtc1); /* high word */
718 1.1.2.6 cliff }
719 1.1.2.6 cliff
720 1.1.2.1 cliff void *
721 1.1.2.15 cliff rmixl_vec_establish(int vec, int cpumask, int ipl,
722 1.1.2.17 cliff int (*func)(void *), void *arg, bool mpsafe)
723 1.1.2.1 cliff {
724 1.1.2.15 cliff rmixl_intrhand_t *ih;
725 1.1.2.20 cliff uint64_t eimr_bit;
726 1.1.2.2 cliff int s;
727 1.1.2.1 cliff
728 1.1.2.28 cliff KASSERT(mutex_owned(rmixl_intr_lock));
729 1.1.2.25 cliff
730 1.1.2.30 matt DPRINTF(("%s: vec %d cpumask %#x ipl %d func %p arg %p mpsafe %d\n",
731 1.1.2.30 matt __func__, vec, cpumask, ipl, func, arg, mpsafe));
732 1.1.2.4 cliff #ifdef DIAGNOSTIC
733 1.1.2.15 cliff if (rmixl_pic_init_done == 0)
734 1.1.2.4 cliff panic("%s: called before evbmips_intr_init", __func__);
735 1.1.2.4 cliff #endif
736 1.1.2.4 cliff
737 1.1.2.2 cliff /*
738 1.1.2.15 cliff * check args
739 1.1.2.2 cliff */
740 1.1.2.15 cliff if (vec < 0 || vec >= NINTRVECS)
741 1.1.2.15 cliff panic("%s: vec %d out of range, max %d",
742 1.1.2.15 cliff __func__, vec, NINTRVECS - 1);
743 1.1.2.4 cliff if (ipl <= 0 || ipl >= _IPL_N)
744 1.1.2.4 cliff panic("%s: ipl %d out of range, min %d, max %d",
745 1.1.2.4 cliff __func__, ipl, 1, _IPL_N - 1);
746 1.1.2.2 cliff
747 1.1.2.15 cliff s = splhigh();
748 1.1.2.1 cliff
749 1.1.2.15 cliff ih = &rmixl_intrhand[vec];
750 1.1.2.20 cliff if (ih->ih_func != NULL) {
751 1.1.2.20 cliff #ifdef DIAGNOSTIC
752 1.1.2.20 cliff printf("%s: intrhand[%d] busy\n", __func__, vec);
753 1.1.2.20 cliff #endif
754 1.1.2.20 cliff splx(s);
755 1.1.2.20 cliff return NULL;
756 1.1.2.20 cliff }
757 1.1.2.2 cliff
758 1.1.2.15 cliff ih->ih_arg = arg;
759 1.1.2.17 cliff ih->ih_mpsafe = mpsafe;
760 1.1.2.20 cliff ih->ih_vec = vec;
761 1.1.2.15 cliff ih->ih_ipl = ipl;
762 1.1.2.15 cliff ih->ih_cpumask = cpumask;
763 1.1.2.2 cliff
764 1.1.2.20 cliff eimr_bit = (uint64_t)1 << vec;
765 1.1.2.20 cliff for (int i=ih->ih_ipl; --i >= 0; ) {
766 1.1.2.20 cliff KASSERT((ipl_eimr_map[i] & eimr_bit) == 0);
767 1.1.2.20 cliff ipl_eimr_map[i] |= eimr_bit;
768 1.1.2.20 cliff }
769 1.1.2.20 cliff
770 1.1.2.24 cliff ih->ih_func = func; /* do this last */
771 1.1.2.24 cliff
772 1.1.2.15 cliff splx(s);
773 1.1.2.15 cliff
774 1.1.2.15 cliff return ih;
775 1.1.2.15 cliff }
776 1.1.2.15 cliff
777 1.1.2.20 cliff /*
778 1.1.2.20 cliff * rmixl_intr_establish
779 1.1.2.20 cliff * - used to establish an IRT-based interrupt only
780 1.1.2.20 cliff */
781 1.1.2.15 cliff void *
782 1.1.2.20 cliff rmixl_intr_establish(int irt, int cpumask, int ipl,
783 1.1.2.17 cliff rmixl_intr_trigger_t trigger, rmixl_intr_polarity_t polarity,
784 1.1.2.17 cliff int (*func)(void *), void *arg, bool mpsafe)
785 1.1.2.15 cliff {
786 1.1.2.15 cliff rmixl_intrhand_t *ih;
787 1.1.2.20 cliff int vec;
788 1.1.2.4 cliff
789 1.1.2.4 cliff #ifdef DIAGNOSTIC
790 1.1.2.15 cliff if (rmixl_pic_init_done == 0)
791 1.1.2.15 cliff panic("%s: called before rmixl_pic_init_done", __func__);
792 1.1.2.15 cliff #endif
793 1.1.2.4 cliff
794 1.1.2.2 cliff /*
795 1.1.2.15 cliff * check args
796 1.1.2.2 cliff */
797 1.1.2.20 cliff if (irt < 0 || irt >= NIRTS)
798 1.1.2.20 cliff panic("%s: irt %d out of range, max %d",
799 1.1.2.20 cliff __func__, irt, NIRTS - 1);
800 1.1.2.15 cliff if (ipl <= 0 || ipl >= _IPL_N)
801 1.1.2.15 cliff panic("%s: ipl %d out of range, min %d, max %d",
802 1.1.2.15 cliff __func__, ipl, 1, _IPL_N - 1);
803 1.1.2.1 cliff
804 1.1.2.20 cliff vec = RMIXL_IRT_VECTOR(irt);
805 1.1.2.20 cliff
806 1.1.2.20 cliff DPRINTF(("%s: irt %d, vec %d, ipl %d\n", __func__, irt, vec, ipl));
807 1.1.2.1 cliff
808 1.1.2.28 cliff mutex_enter(rmixl_intr_lock);
809 1.1.2.1 cliff
810 1.1.2.2 cliff /*
811 1.1.2.15 cliff * establish vector
812 1.1.2.2 cliff */
813 1.1.2.20 cliff ih = rmixl_vec_establish(vec, cpumask, ipl, func, arg, mpsafe);
814 1.1.2.1 cliff
815 1.1.2.1 cliff /*
816 1.1.2.6 cliff * establish IRT Entry
817 1.1.2.1 cliff */
818 1.1.2.20 cliff rmixl_irt_establish(irt, vec, cpumask, trigger, polarity);
819 1.1.2.1 cliff
820 1.1.2.28 cliff mutex_exit(rmixl_intr_lock);
821 1.1.2.1 cliff
822 1.1.2.1 cliff return ih;
823 1.1.2.1 cliff }
824 1.1.2.1 cliff
825 1.1.2.1 cliff void
826 1.1.2.15 cliff rmixl_vec_disestablish(void *cookie)
827 1.1.2.15 cliff {
828 1.1.2.15 cliff rmixl_intrhand_t *ih = cookie;
829 1.1.2.20 cliff uint64_t eimr_bit;
830 1.1.2.15 cliff
831 1.1.2.28 cliff KASSERT(mutex_owned(rmixl_intr_lock));
832 1.1.2.20 cliff KASSERT(ih->ih_vec < NINTRVECS);
833 1.1.2.20 cliff KASSERT(ih == &rmixl_intrhand[ih->ih_vec]);
834 1.1.2.15 cliff
835 1.1.2.24 cliff ih->ih_func = NULL; /* do this first */
836 1.1.2.20 cliff
837 1.1.2.20 cliff eimr_bit = (uint64_t)1 << ih->ih_vec;
838 1.1.2.20 cliff for (int i=ih->ih_ipl; --i >= 0; ) {
839 1.1.2.20 cliff KASSERT((ipl_eimr_map[i] & eimr_bit) != 0);
840 1.1.2.20 cliff ipl_eimr_map[i] ^= eimr_bit;
841 1.1.2.20 cliff }
842 1.1.2.15 cliff }
843 1.1.2.15 cliff
844 1.1.2.15 cliff void
845 1.1.2.1 cliff rmixl_intr_disestablish(void *cookie)
846 1.1.2.1 cliff {
847 1.1.2.15 cliff rmixl_intrhand_t *ih = cookie;
848 1.1.2.30 matt const int vec = ih->ih_vec;
849 1.1.2.15 cliff
850 1.1.2.20 cliff KASSERT(vec < NINTRVECS);
851 1.1.2.20 cliff KASSERT(ih == &rmixl_intrhand[vec]);
852 1.1.2.1 cliff
853 1.1.2.28 cliff mutex_enter(rmixl_intr_lock);
854 1.1.2.1 cliff
855 1.1.2.1 cliff /*
856 1.1.2.15 cliff * disable/invalidate the IRT Entry if needed
857 1.1.2.1 cliff */
858 1.1.2.20 cliff if (RMIXL_VECTOR_IS_IRT(vec))
859 1.1.2.15 cliff rmixl_irt_disestablish(vec);
860 1.1.2.1 cliff
861 1.1.2.1 cliff /*
862 1.1.2.15 cliff * disasociate from vector and free the handle
863 1.1.2.1 cliff */
864 1.1.2.15 cliff rmixl_vec_disestablish(cookie);
865 1.1.2.1 cliff
866 1.1.2.28 cliff mutex_exit(rmixl_intr_lock);
867 1.1.2.1 cliff }
868 1.1.2.1 cliff
869 1.1.2.1 cliff void
870 1.1.2.15 cliff evbmips_iointr(int ipl, vaddr_t pc, uint32_t pending)
871 1.1.2.1 cliff {
872 1.1.2.15 cliff struct rmixl_cpu_softc *sc = (void *)curcpu()->ci_softc;
873 1.1.2.4 cliff
874 1.1.2.30 matt DPRINTF(("%s: cpu%u: ipl %d, pc %#"PRIxVADDR", pending %#x\n",
875 1.1.2.15 cliff __func__, cpu_number(), ipl, pc, pending));
876 1.1.2.2 cliff
877 1.1.2.15 cliff /*
878 1.1.2.15 cliff * 'pending' arg is a summary that there is something to do
879 1.1.2.15 cliff * the real pending status is obtained from EIRR
880 1.1.2.15 cliff */
881 1.1.2.15 cliff KASSERT(pending == MIPS_INT_MASK_1);
882 1.1.2.4 cliff
883 1.1.2.15 cliff for (;;) {
884 1.1.2.15 cliff rmixl_intrhand_t *ih;
885 1.1.2.15 cliff uint64_t eirr;
886 1.1.2.18 cliff uint64_t eimr;
887 1.1.2.15 cliff uint64_t vecbit;
888 1.1.2.15 cliff int vec;
889 1.1.2.1 cliff
890 1.1.2.15 cliff asm volatile("dmfc0 %0, $9, 6;" : "=r"(eirr));
891 1.1.2.22 cliff asm volatile("dmfc0 %0, $9, 7;" : "=r"(eimr));
892 1.1.2.4 cliff
893 1.1.2.15 cliff #ifdef IOINTR_DEBUG
894 1.1.2.30 matt printf("%s: cpu%u: eirr %#"PRIx64", eimr %#"PRIx64", mask %#"PRIx64"\n",
895 1.1.2.30 matt __func__, cpu_number(), eirr, eimr, ipl_eimr_map[ipl-1]);
896 1.1.2.15 cliff #endif /* IOINTR_DEBUG */
897 1.1.2.15 cliff
898 1.1.2.22 cliff /*
899 1.1.2.22 cliff * reduce eirr to
900 1.1.2.22 cliff * - ints that are enabled at or below this ipl
901 1.1.2.22 cliff * - exclude count/compare clock and soft ints
902 1.1.2.22 cliff * they are handled elsewhere
903 1.1.2.22 cliff */
904 1.1.2.15 cliff eirr &= ipl_eimr_map[ipl-1];
905 1.1.2.22 cliff eirr &= ~ipl_eimr_map[ipl];
906 1.1.2.22 cliff eirr &= ~((MIPS_INT_MASK_5 | MIPS_SOFT_INT_MASK) >> 8);
907 1.1.2.15 cliff if (eirr == 0)
908 1.1.2.15 cliff break;
909 1.1.2.15 cliff
910 1.1.2.15 cliff vec = 63 - dclz(eirr);
911 1.1.2.15 cliff ih = &rmixl_intrhand[vec];
912 1.1.2.15 cliff vecbit = 1ULL << vec;
913 1.1.2.22 cliff KASSERT (ih->ih_ipl == ipl);
914 1.1.2.18 cliff KASSERT ((vecbit & eimr) == 0);
915 1.1.2.15 cliff KASSERT ((vecbit & RMIXL_EIRR_PRESERVE_MASK) == 0);
916 1.1.2.22 cliff
917 1.1.2.22 cliff /*
918 1.1.2.30 matt * ack in EIRR, and in PIC if needed,
919 1.1.2.30 matt * the irq we are about to handle
920 1.1.2.22 cliff */
921 1.1.2.30 matt rmixl_eirr_ack(eimr, vecbit, RMIXL_EIRR_PRESERVE_MASK);
922 1.1.2.20 cliff if (RMIXL_VECTOR_IS_IRT(vec))
923 1.1.2.4 cliff RMIXL_PICREG_WRITE(RMIXL_PIC_INTRACK,
924 1.1.2.20 cliff 1 << RMIXL_VECTOR_IRT(vec));
925 1.1.2.15 cliff
926 1.1.2.17 cliff if (ih->ih_func != NULL) {
927 1.1.2.17 cliff #ifdef MULTIPROCESSOR
928 1.1.2.17 cliff if (ih->ih_mpsafe) {
929 1.1.2.17 cliff (void)(*ih->ih_func)(ih->ih_arg);
930 1.1.2.17 cliff } else {
931 1.1.2.30 matt KASSERTMSG(ipl == IPL_VM,
932 1.1.2.30 matt ("%s: %s: ipl (%d) != IPL_VM for KERNEL_LOCK",
933 1.1.2.30 matt __func__, sc->sc_vec_evcnts[vec].ev_name,
934 1.1.2.30 matt ipl));
935 1.1.2.17 cliff KERNEL_LOCK(1, NULL);
936 1.1.2.17 cliff (void)(*ih->ih_func)(ih->ih_arg);
937 1.1.2.17 cliff KERNEL_UNLOCK_ONE(NULL);
938 1.1.2.17 cliff }
939 1.1.2.17 cliff #else
940 1.1.2.15 cliff (void)(*ih->ih_func)(ih->ih_arg);
941 1.1.2.17 cliff #endif /* MULTIPROCESSOR */
942 1.1.2.17 cliff }
943 1.1.2.30 matt KASSERT(ipl == ih->ih_ipl);
944 1.1.2.30 matt KASSERTMSG(curcpu()->ci_cpl >= ipl,
945 1.1.2.30 matt ("%s: after %s: cpl (%d) < ipl %d",
946 1.1.2.30 matt __func__, sc->sc_vec_evcnts[vec].ev_name,
947 1.1.2.30 matt ipl, curcpu()->ci_cpl));
948 1.1.2.15 cliff sc->sc_vec_evcnts[vec].ev_count++;
949 1.1.2.1 cliff }
950 1.1.2.1 cliff }
951 1.1.2.4 cliff
952 1.1.2.15 cliff #ifdef MULTIPROCESSOR
953 1.1.2.15 cliff static int
954 1.1.2.15 cliff rmixl_send_ipi(struct cpu_info *ci, int tag)
955 1.1.2.4 cliff {
956 1.1.2.26 cliff const cpuid_t cpuid = ci->ci_cpuid;
957 1.1.2.26 cliff uint32_t core = (uint32_t)(cpuid >> 2);
958 1.1.2.26 cliff uint32_t thread = (uint32_t)(cpuid & __BITS(1,0));
959 1.1.2.15 cliff uint64_t req = 1 << tag;
960 1.1.2.15 cliff uint32_t r;
961 1.1.2.4 cliff
962 1.1.2.30 matt if (! CPUSET_HAS_P(cpus_running, cpu_index(ci)))
963 1.1.2.15 cliff return -1;
964 1.1.2.15 cliff
965 1.1.2.25 cliff KASSERT((tag >= 0) && (tag < NIPIS));
966 1.1.2.15 cliff
967 1.1.2.15 cliff r = (thread << RMIXL_PIC_IPIBASE_ID_THREAD_SHIFT)
968 1.1.2.15 cliff | (core << RMIXL_PIC_IPIBASE_ID_CORE_SHIFT)
969 1.1.2.25 cliff | (RMIXL_INTRVEC_IPI + tag);
970 1.1.2.15 cliff
971 1.1.2.28 cliff mutex_enter(rmixl_ipi_lock);
972 1.1.2.15 cliff atomic_or_64(&ci->ci_request_ipis, req);
973 1.1.2.15 cliff RMIXL_PICREG_WRITE(RMIXL_PIC_IPIBASE, r);
974 1.1.2.28 cliff mutex_exit(rmixl_ipi_lock);
975 1.1.2.15 cliff
976 1.1.2.15 cliff return 0;
977 1.1.2.15 cliff }
978 1.1.2.15 cliff
979 1.1.2.15 cliff static int
980 1.1.2.15 cliff rmixl_ipi_intr(void *arg)
981 1.1.2.15 cliff {
982 1.1.2.15 cliff struct cpu_info * const ci = curcpu();
983 1.1.2.30 matt const uint64_t ipi_mask = 1 << (uintptr_t)arg;
984 1.1.2.15 cliff
985 1.1.2.30 matt KASSERT(ci->ci_cpl >= IPL_SCHED);
986 1.1.2.25 cliff KASSERT((uintptr_t)arg < NIPIS);
987 1.1.2.30 matt
988 1.1.2.30 matt /* if the request is clear, it was previously processed */
989 1.1.2.30 matt if ((ci->ci_request_ipis & ipi_mask) == 0)
990 1.1.2.30 matt return 0;
991 1.1.2.25 cliff
992 1.1.2.25 cliff atomic_or_64(&ci->ci_active_ipis, ipi_mask);
993 1.1.2.25 cliff atomic_and_64(&ci->ci_request_ipis, ~ipi_mask);
994 1.1.2.15 cliff
995 1.1.2.15 cliff ipi_process(ci, ipi_mask);
996 1.1.2.15 cliff
997 1.1.2.25 cliff atomic_and_64(&ci->ci_active_ipis, ~ipi_mask);
998 1.1.2.25 cliff
999 1.1.2.15 cliff return 1;
1000 1.1.2.15 cliff }
1001 1.1.2.15 cliff #endif /* MULTIPROCESSOR */
1002 1.1.2.15 cliff
1003 1.1.2.20 cliff #if defined(DIAGNOSTIC) || defined(IOINTR_DEBUG) || defined(DDB)
1004 1.1.2.15 cliff int
1005 1.1.2.15 cliff rmixl_intrhand_print_subr(int vec)
1006 1.1.2.15 cliff {
1007 1.1.2.15 cliff rmixl_intrhand_t *ih = &rmixl_intrhand[vec];
1008 1.1.2.20 cliff printf("vec %d: func %p, arg %p, vec %d, ipl %d, mask %#x\n",
1009 1.1.2.20 cliff vec, ih->ih_func, ih->ih_arg, ih->ih_vec, ih->ih_ipl,
1010 1.1.2.15 cliff ih->ih_cpumask);
1011 1.1.2.15 cliff return 0;
1012 1.1.2.15 cliff }
1013 1.1.2.15 cliff int
1014 1.1.2.15 cliff rmixl_intrhand_print(void)
1015 1.1.2.15 cliff {
1016 1.1.2.15 cliff for (int vec=0; vec < NINTRVECS ; vec++)
1017 1.1.2.15 cliff rmixl_intrhand_print_subr(vec);
1018 1.1.2.15 cliff return 0;
1019 1.1.2.15 cliff }
1020 1.1.2.20 cliff
1021 1.1.2.20 cliff static inline void
1022 1.1.2.20 cliff rmixl_irt_entry_print(u_int irt)
1023 1.1.2.20 cliff {
1024 1.1.2.20 cliff uint32_t c0, c1;
1025 1.1.2.20 cliff
1026 1.1.2.20 cliff if ((irt < 0) || (irt > NIRTS))
1027 1.1.2.20 cliff return;
1028 1.1.2.20 cliff c0 = RMIXL_PICREG_READ(RMIXL_PIC_IRTENTRYC0(irt));
1029 1.1.2.20 cliff c1 = RMIXL_PICREG_READ(RMIXL_PIC_IRTENTRYC1(irt));
1030 1.1.2.20 cliff printf("irt[%d]: %#x, %#x\n", irt, c0, c1);
1031 1.1.2.20 cliff }
1032 1.1.2.20 cliff
1033 1.1.2.15 cliff int
1034 1.1.2.15 cliff rmixl_irt_print(void)
1035 1.1.2.15 cliff {
1036 1.1.2.15 cliff printf("%s:\n", __func__);
1037 1.1.2.15 cliff for (int irt=0; irt < NIRTS ; irt++)
1038 1.1.2.15 cliff rmixl_irt_entry_print(irt);
1039 1.1.2.4 cliff return 0;
1040 1.1.2.4 cliff }
1041 1.1.2.20 cliff
1042 1.1.2.20 cliff void
1043 1.1.2.20 cliff rmixl_ipl_eimr_map_print(void)
1044 1.1.2.20 cliff {
1045 1.1.2.20 cliff printf("IPL_NONE=%d, mask %#"PRIx64"\n",
1046 1.1.2.20 cliff IPL_NONE, ipl_eimr_map[IPL_NONE]);
1047 1.1.2.20 cliff printf("IPL_SOFTCLOCK=%d, mask %#"PRIx64"\n",
1048 1.1.2.20 cliff IPL_SOFTCLOCK, ipl_eimr_map[IPL_SOFTCLOCK]);
1049 1.1.2.20 cliff printf("IPL_SOFTNET=%d, mask %#"PRIx64"\n",
1050 1.1.2.20 cliff IPL_SOFTNET, ipl_eimr_map[IPL_SOFTNET]);
1051 1.1.2.20 cliff printf("IPL_VM=%d, mask %#"PRIx64"\n",
1052 1.1.2.20 cliff IPL_VM, ipl_eimr_map[IPL_VM]);
1053 1.1.2.20 cliff printf("IPL_SCHED=%d, mask %#"PRIx64"\n",
1054 1.1.2.20 cliff IPL_SCHED, ipl_eimr_map[IPL_SCHED]);
1055 1.1.2.20 cliff printf("IPL_DDB=%d, mask %#"PRIx64"\n",
1056 1.1.2.20 cliff IPL_DDB, ipl_eimr_map[IPL_DDB]);
1057 1.1.2.20 cliff printf("IPL_HIGH=%d, mask %#"PRIx64"\n",
1058 1.1.2.20 cliff IPL_HIGH, ipl_eimr_map[IPL_HIGH]);
1059 1.1.2.20 cliff }
1060 1.1.2.20 cliff
1061 1.1.2.4 cliff #endif
1062