rmixl_spl.S revision 1.6 1 /* $NetBSD: rmixl_spl.S,v 1.6 2020/12/01 08:39:39 skrll Exp $ */
2
3 /*-
4 * Copyright (c) 2009, 2010 The NetBSD Foundation, Inc.
5 * All rights reserved.
6 *
7 * This code is derived from software contributed to The NetBSD Foundation
8 * by Matt Thomas <matt (at) 3am-software.com>.
9 *
10 * Redistribution and use in source and binary forms, with or without
11 * modification, are permitted provided that the following conditions
12 * are met:
13 * 1. Redistributions of source code must retain the above copyright
14 * notice, this list of conditions and the following disclaimer.
15 * 2. Redistributions in binary form must reproduce the above copyright
16 * notice, this list of conditions and the following disclaimer in the
17 * documentation and/or other materials provided with the distribution.
18 *
19 * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
20 * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
21 * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
22 * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
23 * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
24 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
25 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
26 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
27 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
28 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
29 * POSSIBILITY OF SUCH DAMAGE.
30 */
31
32 #include "opt_cputype.h" /* which mips CPU levels do we support? */
33
34 #include <sys/cdefs.h>
35
36 #include <mips/asm.h>
37 #include <mips/cpuregs.h>
38
39 RCSID("$NetBSD: rmixl_spl.S,v 1.6 2020/12/01 08:39:39 skrll Exp $");
40
41 #include "assym.h"
42
43 #define MAP_SCALESHIFT 3
44 #define RMIXL_SOFT_INT_MASK_1 (MIPS_SOFT_INT_MASK_1 >> 8)
45 #define RMIXL_SOFT_INT_MASK (MIPS_SOFT_INT_MASK >> 8)
46 #define RMIXL_INT_MASK_5 (MIPS_INT_MASK_5 >> 8)
47 #define RMIXL_EIRR_PRESERVE (RMIXL_INT_MASK_5 | RMIXL_SOFT_INT_MASK)
48 #define RMIXL_INT_MASK_1 (MIPS_INT_MASK_1 >> 8)
49 #define RMIXL_INT_MASK_5 (MIPS_INT_MASK_5 >> 8)
50
51 .set noreorder
52
53 /*
54 * Array of mask of bits to set in the EIMR when we go to a
55 * given hardware interrupt priority level.
56 * The softint bits in [IPL_NONE] and [IPL_SOFTCLOCK] should stay constant
57 * Hard intr bits are managed by rmixl_vec_establish and rmixl_vec_disestablish.
58 */
59 .data
60 .globl _C_LABEL(ipl_eimr_map)
61 .type _C_LABEL(ipl_eimr_map),@object
62 .p2align MAP_SCALESHIFT
63 _C_LABEL(ipl_eimr_map):
64 .dword RMIXL_SOFT_INT_MASK /* IPL_NONE */
65 .dword RMIXL_SOFT_INT_MASK_1 /* IPL_SOFT{CLOCK,BIO} */
66 .dword 0 /* IPL_SOFT{NET,SERIAL} */
67 .dword 0 /* IPL_VM */
68 .dword 0 /* IPL_SCHED */
69 .dword 0 /* IPL_DDB */
70 .dword 0 /* IPL_HIGH */
71
72 .text
73
74 /*
75 * initialize cp0 interrupt control for this cpu
76 * - set STATUS[IE]
77 * - clear EIRR and EIMR
78 * on return, all interrupts are disabled by EIMR
79 *
80 * henceforth STATUS[IE] is expected to remain normally set
81 * but may be cleared and restored for temporary interrupt disablement
82 *
83 * call before the first call to spl0 on this cpu
84 */
85 LEAF_NOPROFILE(rmixl_spl_init_cpu)
86 mfc0 t0, MIPS_COP_0_STATUS # get STATUS
87 ori t0, MIPS_SR_INT_IE # set IE
88 mtc0 zero, MIPS_COP_0_STATUS ## disable all ints in STATUS
89 COP0_SYNC
90 dmtc0 zero, MIPS_COP_0_EIMR ## " " " " EIMR
91 COP0_SYNC
92 dmtc0 zero, MIPS_COP_0_EIRR ## clear EIRR
93 COP0_SYNC
94 mtc0 t0, MIPS_COP_0_STATUS ## set STATUS | IE
95 JR_HB_RA
96 END(rmixl_spl_init_cpu)
97
98 /*
99 * RMIXL processor interrupt control
100 *
101 * Used as building blocks for spl(9) kernel interface.
102 */
103
104 _splraise:
105 /*
106 * a0 = EIMR bits requested to be set for this IPL
107 * a1 = this IPL (IPL_*)
108 * Can only use a0-a3 and v0-v1
109 * old IPL is returned in v0
110 */
111 dmfc0 a2, MIPS_COP_0_EIMR # save EIMR
112 dmtc0 zero, MIPS_COP_0_EIMR ## disable all interrupts
113 COP0_SYNC
114 PTR_L a3, L_CPU(MIPS_CURLWP) ##
115 INT_L v0, CPU_INFO_CPL(a3) ## get current IPL from cpu_info
116 sltu v1, a1, v0 ## newipl < curipl
117 bnez v1, 1f ## yes, don't change.
118 nop
119 INT_S a1, CPU_INFO_CPL(a3) ## save IPL in cpu_info
120 dmtc0 a0, MIPS_COP_0_EIMR ## set new EIMR
121 JR_HB_RA
122 1:
123 dmtc0 a2, MIPS_COP_0_EIMR ## restore saved EIMR
124 JR_HB_RA
125
126 STATIC_LEAF(_splsw_splx)
127 STATIC_XLEAF(_splsw_splx_noprof) # does not get mcount hooks
128 PTR_LA v1, _C_LABEL(ipl_eimr_map) # get address of table
129 sll a2, a0, MAP_SCALESHIFT # convert IPL to array offset
130 PTR_ADDU v1, a2 # add to table addr
131 REG_L v1, (v1) # load EIMR bits for this IPL
132
133 dmtc0 zero, MIPS_COP_0_EIMR ## disable all interrupts
134 COP0_SYNC
135 PTR_L a3, L_CPU(MIPS_CURLWP) ## get cpu_info
136 INT_S a0, CPU_INFO_CPL(a3) ## save IPL in cpu_info
137 dmtc0 v1, MIPS_COP_0_EIMR ## set new EIMR
138 JR_HB_RA
139 END(_splsw_splx)
140
141 STATIC_LEAF(_splsw_spl0)
142 REG_L v1, _C_LABEL(ipl_eimr_map) + 8*IPL_NONE
143 dmtc0 zero, MIPS_COP_0_EIMR ## disable all interrupts
144 COP0_SYNC
145 PTR_L a3, L_CPU(MIPS_CURLWP) ## get cpu_info
146 #if IPL_NONE == 0
147 INT_S zero, CPU_INFO_CPL(a3) ## save IPL in cpu_info
148 #else
149 #error IPL_NONE != 0
150 #endif
151 dmtc0 v1, MIPS_COP_0_EIMR ## set new EIMR
152 JR_HB_RA
153 END(_splsw_spl0)
154
155 STATIC_LEAF(_splsw_setsoftintr)
156 dsrl a0, 8 # convert CAUSE bit to EIRR bit
157 and a0, RMIXL_SOFT_INT_MASK # restrict to softint bits
158 dmfc0 v1, MIPS_COP_0_EIMR # save EIMR register
159 dmtc0 zero, MIPS_COP_0_EIMR ## disable all interrupts
160 COP0_SYNC
161 dmfc0 v0, MIPS_COP_0_EIRR ## load EIRR
162 and v0, RMIXL_EIRR_PRESERVE ## preserve clock & softints
163 or v0, a0 ## set new softint bit
164 dmtc0 v0, MIPS_COP_0_EIRR ## store EIRR
165 COP0_SYNC
166 dmtc0 v1, MIPS_COP_0_EIMR ## restore EIMR
167 JR_HB_RA
168 END(_splsw_setsoftintr)
169
170 STATIC_LEAF(_splsw_clrsoftintr)
171 dsrl a0, 8 # convert CAUSE bit to EIRR bit
172 and a0, RMIXL_SOFT_INT_MASK # restrict to softint bits
173 xor a0, RMIXL_EIRR_PRESERVE # clear from preserve mask
174 dmfc0 v1, MIPS_COP_0_EIMR # save EIMR register
175 dmtc0 zero, MIPS_COP_0_EIMR ## disable all interrupts
176 COP0_SYNC
177 dmfc0 v0, MIPS_COP_0_EIRR ## load EIRR
178 and v0, a0 ## apply preserve mask
179 dmtc0 v0, MIPS_COP_0_EIRR ## store EIRR
180 COP0_SYNC
181 dmtc0 v1, MIPS_COP_0_EIMR ## restore EIMR
182 JR_HB_RA
183 END(_splsw_clrsoftintr)
184
185 STATIC_LEAF(_splsw_splraise)
186 move a1, a0
187 PTR_LA v1, _C_LABEL(ipl_eimr_map)
188 sll a2, a0, MAP_SCALESHIFT
189 PTR_ADDU v1, a2
190 REG_L a0, (v1)
191 b _splraise
192 nop
193 END(_splsw_splraise)
194
195 STATIC_LEAF(_splsw_splhigh)
196 STATIC_XLEAF(_splsw_splhigh_noprof)
197 dmtc0 zero, MIPS_COP_0_EIMR ## disable all interrupts
198 COP0_SYNC
199 PTR_L a3, L_CPU(MIPS_CURLWP) ## get cpu_info from curlwp
200 li a1, IPL_HIGH ##
201 INT_L v0, CPU_INFO_CPL(a3) ## old IPL for return value
202 INT_S a1, CPU_INFO_CPL(a3) ## save IPL in cpu_info
203 ## interrupts remain disabled!
204 j ra # return
205 nop
206 END(_splsw_splhigh)
207
208 STATIC_LEAF(_splsw_splddb)
209 REG_L a0, _C_LABEL(ipl_eimr_map) + 8*IPL_DDB
210 li a1, IPL_DDB
211 b _splraise
212 nop
213 END(_splsw_splddb)
214
215 STATIC_LEAF(_splsw_splsched)
216 REG_L a0, _C_LABEL(ipl_eimr_map) + 8*IPL_SCHED
217 li a1, IPL_SCHED
218 b _splraise
219 nop
220 END(_splsw_splsched)
221
222 STATIC_LEAF(_splsw_splvm)
223 REG_L a0, _C_LABEL(ipl_eimr_map) + 8*IPL_VM
224 li a1, IPL_VM
225 b _splraise
226 nop
227 END(_splsw_splvm)
228
229 STATIC_LEAF(_splsw_splsoftserial)
230 REG_L a0, _C_LABEL(ipl_eimr_map) + 8*IPL_SOFTSERIAL
231 li a1, IPL_SOFTSERIAL
232 b _splraise
233 nop
234 END(_splsw_splsoftserial)
235
236 STATIC_LEAF(_splsw_splsoftnet)
237 REG_L a0, _C_LABEL(ipl_eimr_map) + 8*IPL_SOFTNET
238 li a1, IPL_SOFTNET
239 b _splraise
240 nop
241 END(_splsw_splsoftnet)
242
243 STATIC_LEAF(_splsw_splsoftbio)
244 REG_L a0, _C_LABEL(ipl_eimr_map) + 8*IPL_SOFTBIO
245 li a1, IPL_SOFTBIO
246 b _splraise
247 nop
248 END(_splsw_splsoftbio)
249
250 STATIC_LEAF(_splsw_splsoftclock)
251 REG_L a0, _C_LABEL(ipl_eimr_map) + 8*IPL_SOFTCLOCK
252 li a1, IPL_SOFTCLOCK
253 b _splraise
254 nop
255 END(_splsw_splsoftclock)
256
257 STATIC_LEAF(_splsw_splintr)
258 dmfc0 ta1, MIPS_COP_0_EIRR # get active interrupts
259 # restrict to hard int bits:
260 and v1, ta1, RMIXL_SOFT_INT_MASK # v1 = ta1 & ~RMIXL_SOFT_INT_MASK
261 xor v1, ta1 # " "
262
263 li v0, IPL_NONE
264 PTR_LA ta3, _C_LABEL(ipl_eimr_map) + 8*IPL_VM
265 REG_L ta2, -8(ta3) # load 'enabled' bits for IPL_SOFTSERIAL
266 and v1, ta2 # apply to pending bits
267 beq v1, zero, 4f # if nothing pending...
268 nop # ... return IPL_NONE
269
270 li v0, IPL_VM # ipl=IPL_VM
271 1:
272 REG_L ta2, (ta3) # load 'enabled' bits for ipl
273 and ta2, v1 # any match to pending intrs?
274 beq ta2, zero, 2f # no, return ipl
275 PTR_ADDI ta3, 1 << MAP_SCALESHIFT # point to next entry
276 addiu v0, 1 # ipl++
277 move v1, ta2 # update highest pending
278 b 1b # loop
279 nop
280
281 2:
282 /*
283 * Emulate the CP0_SR 'IM' bits in 'pending'
284 * - if clock intr is requested, set MIPS_INT_MASK_5
285 * - if other HW intr is requested, set MIPS_INT_MASK_1 as summary bit
286 * the RMI evbmips_iointr function will sort through
287 * individial EIRR requests
288 */
289 li t2, RMIXL_INT_MASK_5 # load RMIXL_INT_MASK_5
290 and t1, v1, t2 # save count/compare intr request value
291 nor t0, zero, t2 # invert the mask
292 and v1, t0 # v1 &= ~RMIXL_INT_MASK_5
293 beq v1, zero, 3f # no non-clock intrs? skip ahead
294 li v1, RMIXL_INT_MASK_1 # use INT_MASK_1 as 'summary' bit
295 # for non-clock hw intrs
296 3:
297 or v1, t1 # combine clock and non-clock-summary
298 sll v1, MIPS_INT_MASK_SHIFT # shift to emulate COP0_SR 'IM' bits
299 4:
300 INT_S v1, (a0) # set a (fake) new pending mask
301 j ra # and return highest ipl pending
302 nop
303 END(_splsw_splintr)
304
305 STATIC_LEAF(_splsw_splcheck)
306 #ifdef PARANOIA
307 PTR_L t0, L_CPU(MIPS_CURLWP)
308 INT_L t1, CPU_INFO_CPL(t0) # get current priority level
309
310 dmfc0 t0, MIPS_COP_0_EIMR # get current EIMR
311
312 PTR_LA t2, _C_LABEL(ipl_eimr_map)
313 sll t1, MAP_SCALESHIFT # shift cpl to array index
314 PTR_ADDU t2, t1
315 REG_L t3, (t2) # load value
316 1: bne t0, t3, 1b # loop forever if not equal
317 nop
318 #endif /* PARANOIA */
319 j ra
320 nop
321 END(_splsw_splcheck)
322
323 .rdata
324 .globl _C_LABEL(rmixl_splsw)
325 _C_LABEL(rmixl_splsw):
326 PTR_WORD _C_LABEL(_splsw_splhigh)
327 PTR_WORD _C_LABEL(_splsw_splsched)
328 PTR_WORD _C_LABEL(_splsw_splvm)
329 PTR_WORD _C_LABEL(_splsw_splsoftserial)
330 PTR_WORD _C_LABEL(_splsw_splsoftnet)
331 PTR_WORD _C_LABEL(_splsw_splsoftbio)
332 PTR_WORD _C_LABEL(_splsw_splsoftclock)
333 PTR_WORD _C_LABEL(_splsw_splraise)
334 PTR_WORD _C_LABEL(_splsw_spl0)
335 PTR_WORD _C_LABEL(_splsw_splx)
336 PTR_WORD _C_LABEL(_splsw_splhigh_noprof)
337 PTR_WORD _C_LABEL(_splsw_splx_noprof)
338 PTR_WORD _C_LABEL(_splsw_setsoftintr)
339 PTR_WORD _C_LABEL(_splsw_clrsoftintr)
340 PTR_WORD _C_LABEL(_splsw_splintr)
341 PTR_WORD _C_LABEL(_splsw_splcheck)
342