locore.s revision 1.9 1 /* $NetBSD: locore.s,v 1.9 1997/02/20 00:54:21 jeremy Exp $ */
2
3 /*
4 * Copyright (c) 1988 University of Utah.
5 * Copyright (c) 1980, 1990, 1993
6 * The Regents of the University of California. All rights reserved.
7 *
8 * This code is derived from software contributed to Berkeley by
9 * the Systems Programming Group of the University of Utah Computer
10 * Science Department.
11 *
12 * Redistribution and use in source and binary forms, with or without
13 * modification, are permitted provided that the following conditions
14 * are met:
15 * 1. Redistributions of source code must retain the above copyright
16 * notice, this list of conditions and the following disclaimer.
17 * 2. Redistributions in binary form must reproduce the above copyright
18 * notice, this list of conditions and the following disclaimer in the
19 * documentation and/or other materials provided with the distribution.
20 * 3. All advertising materials mentioning features or use of this software
21 * must display the following acknowledgement:
22 * This product includes software developed by the University of
23 * California, Berkeley and its contributors.
24 * 4. Neither the name of the University nor the names of its contributors
25 * may be used to endorse or promote products derived from this software
26 * without specific prior written permission.
27 *
28 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
29 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
30 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
31 * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
32 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
33 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
34 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
35 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
36 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
37 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
38 * SUCH DAMAGE.
39 *
40 * from: Utah $Hdr: locore.s 1.66 92/12/22$
41 * @(#)locore.s 8.6 (Berkeley) 5/27/94
42 */
43
44 #include "assym.h"
45 #include <machine/trap.h>
46
47 | Remember this is a fun project!
48
49 .data
50 .globl _mon_crp
51 _mon_crp:
52 .long 0,0
53
54 | This is for kvm_mkdb, and should be the address of the beginning
55 | of the kernel text segment (not necessarily the same as kernbase).
56 .text
57 .globl _kernel_text
58 _kernel_text:
59
60 | This is the entry point, as well as the end of the temporary stack
61 | used during process switch (one 8K page ending at start)
62 .globl tmpstk
63 tmpstk:
64 .globl start
65 start:
66 | The first step, after disabling interrupts, is to map enough of the kernel
67 | into high virtual address space so that we can use position dependent code.
68 | This is a tricky task on the sun3x because the MMU is already enabled and
69 | the ROM monitor provides no indication of where the root MMU table is mapped.
70 | Therefore we must use one of the 68030's 'transparent translation' registers
71 | to define a range in the address space where the MMU translation is
72 | turned off. Once this is complete we can modify the MMU table directly
73 | without the need for it to be mapped into virtual memory.
74 | All code must be position independent until otherwise noted, as the
75 | boot loader has loaded us into low memory but all the symbols in this
76 | code have been linked high.
77 movw #PSL_HIGHIPL, sr | no interrupts
78 movl #KERNBASE, a5 | for vtop conversion
79 lea _mon_crp, a0 | where to store the CRP
80 subl a5, a0
81 | Note: borrowing mon_crp for tt0 setup...
82 movl #0x3F8107, a0@ | map the low 1GB v=p with the
83 pmove a0@, tt0 | transparent translation reg0
84
85 | In order to map the kernel into high memory we will copy the root table
86 | entry which maps the 16 megabytes of memory starting at 0x0 into the
87 | entry which maps the 16 megabytes starting at KERNBASE.
88 pmove crp, a0@ | Get monitor CPU root pointer
89 movl a0@(4), a1 | 2nd word is PA of level A table
90
91 movl a1, a0 | compute the descriptor address
92 addl #0x3e0, a1 | for VA starting at KERNBASE
93 movl a0@, a1@ | copy descriptor type
94 movl a0@(4), a1@(4) | copy physical address
95
96 | Kernel is now double mapped at zero and KERNBASE.
97 | Force a long jump to the relocated code (high VA).
98 movl #IC_CLEAR, d0 | Flush the I-cache
99 movc d0, cacr
100 jmp L_high_code:l | long jump
101
102 L_high_code:
103 | We are now running in the correctly relocated kernel, so
104 | we are no longer restricted to position-independent code.
105 | It is handy to leave transparent translation enabled while
106 | for the low 1GB while __bootstrap() is doing its thing.
107
108 | Do bootstrap stuff needed before main() gets called.
109 | Our boot loader leaves a copy of the kernel's exec header
110 | just before the start of the kernel text segment, so the
111 | kernel can sanity-check the DDB symbols at [end...esym].
112 | Pass the struct exec at tmpstk-32 to __bootstrap().
113 | Also, make sure the initial frame pointer is zero so that
114 | the backtrace algorithm used by KGDB terminates nicely.
115 lea tmpstk-32, sp
116 movl #0,a6
117 jsr __bootstrap | See _startup.c
118
119 | Now turn off the transparent translation of the low 1GB.
120 | (this also flushes the ATC)
121 clrl sp@-
122 pmove sp@,tt0
123 addql #4,sp
124
125 | Now that __bootstrap() is done using the PROM functions,
126 | we can safely set the sfc/dfc to something != FC_CONTROL
127 moveq #FC_USERD, d0 | make movs access "user data"
128 movc d0, sfc | space for copyin/copyout
129 movc d0, dfc
130
131 | Setup process zero user/kernel stacks.
132 movl _proc0paddr,a1 | get proc0 pcb addr
133 lea a1@(USPACE-4),sp | set SSP to last word
134 movl #USRSTACK-4,a2
135 movl a2,usp | init user SP
136
137 | Note curpcb was already set in __bootstrap().
138 | Will do fpu initialization during autoconfig (see fpu.c)
139 | The interrupt vector table and stack are now ready.
140 | Interrupts will be enabled later, AFTER autoconfiguration
141 | is finished, to avoid spurrious interrupts.
142
143 /*
144 * Final preparation for calling main.
145 *
146 * Create a fake exception frame that returns to user mode,
147 * and save its address in p->p_md.md_regs for cpu_fork().
148 * The new frames for process 1 and 2 will be adjusted by
149 * cpu_set_kpc() to arrange for a call to a kernel function
150 * before the new process does its rte out to user mode.
151 */
152 clrw sp@- | tf_format,tf_vector
153 clrl sp@- | tf_pc (filled in later)
154 movw #PSL_USER,sp@- | tf_sr for user mode
155 clrl sp@- | tf_stackadj
156 lea sp@(-64),sp | tf_regs[16]
157 movl sp,a1 | a1=trapframe
158 lea _proc0,a0 | proc0.p_md.md_regs =
159 movl a1,a0@(P_MDREGS) | trapframe
160 movl a2,a1@(FR_SP) | a2 == usp (from above)
161 pea a1@ | push &trapframe
162 jbsr _main | main(&trapframe)
163 addql #4,sp | help DDB backtrace
164 trap #15 | should not get here
165
166 | This is used by cpu_fork() to return to user mode.
167 | It is called with SP pointing to a struct trapframe.
168 .globl _proc_do_uret
169 _proc_do_uret:
170 movl sp@(FR_SP),a0 | grab and load
171 movl a0,usp | user SP
172 moveml sp@+,#0x7FFF | load most registers (all but SSP)
173 addql #8,sp | pop SSP and stack adjust count
174 rte
175
176 /*
177 * proc_trampoline:
178 * This is used by cpu_set_kpc() to "push" a function call onto the
179 * kernel stack of some process, very much like a signal delivery.
180 * When we get here, the stack has:
181 *
182 * SP+8: switchframe from before cpu_set_kpc
183 * SP+4: void *proc;
184 * SP: u_long func;
185 *
186 * On entry, the switchframe pushed by cpu_set_kpc has already been
187 * popped off the stack, so all this needs to do is pop the function
188 * pointer into a register, call it, then pop the arg, and finally
189 * return using the switchframe that remains on the stack.
190 */
191 .globl _proc_trampoline
192 _proc_trampoline:
193 movl sp@+,a0 | function pointer
194 jbsr a0@ | (*func)(procp)
195 addql #4,sp | toss the arg
196 rts | as cpu_switch would do
197
198 | That is all the assembly startup code we need on the sun3x!
199 | The rest of this is like the hp300/locore.s where possible.
200
201 /*
202 * Trap/interrupt vector routines
203 */
204
205 .globl _buserr, _addrerr, _illinst, _zerodiv, _chkinst
206 .globl _trapvinst, _privinst, _trace, _badtrap, _fmterr
207 .globl _trap0, _trap1, _trap2, _trap12, _trap15
208 .globl _coperr, _fpfline, _fpunsupp
209
210 .globl _trap, _nofault, _longjmp
211 _buserr:
212 tstl _nofault | device probe?
213 jeq _addrerr | no, handle as usual
214 movl _nofault,sp@- | yes,
215 jbsr _longjmp | longjmp(nofault)
216 _addrerr:
217 clrl sp@- | stack adjust count
218 moveml #0xFFFF,sp@- | save user registers
219 movl usp,a0 | save the user SP
220 movl a0,sp@(FR_SP) | in the savearea
221 lea sp@(FR_HW),a1 | grab base of HW berr frame
222 moveq #0,d0
223 movw a1@(10),d0 | grab SSW for fault processing
224 btst #12,d0 | RB set?
225 jeq LbeX0 | no, test RC
226 bset #14,d0 | yes, must set FB
227 movw d0,a1@(10) | for hardware too
228 LbeX0:
229 btst #13,d0 | RC set?
230 jeq LbeX1 | no, skip
231 bset #15,d0 | yes, must set FC
232 movw d0,a1@(10) | for hardware too
233 LbeX1:
234 btst #8,d0 | data fault?
235 jeq Lbe0 | no, check for hard cases
236 movl a1@(16),d1 | fault address is as given in frame
237 jra Lbe10 | thats it
238 Lbe0:
239 btst #4,a1@(6) | long (type B) stack frame?
240 jne Lbe4 | yes, go handle
241 movl a1@(2),d1 | no, can use save PC
242 btst #14,d0 | FB set?
243 jeq Lbe3 | no, try FC
244 addql #4,d1 | yes, adjust address
245 jra Lbe10 | done
246 Lbe3:
247 btst #15,d0 | FC set?
248 jeq Lbe10 | no, done
249 addql #2,d1 | yes, adjust address
250 jra Lbe10 | done
251 Lbe4:
252 movl a1@(36),d1 | long format, use stage B address
253 btst #15,d0 | FC set?
254 jeq Lbe10 | no, all done
255 subql #2,d1 | yes, adjust address
256 Lbe10:
257 movl d1,sp@- | push fault VA
258 movl d0,sp@- | and padded SSW
259 movw a1@(6),d0 | get frame format/vector offset
260 andw #0x0FFF,d0 | clear out frame format
261 cmpw #12,d0 | address error vector?
262 jeq Lisaerr | yes, go to it
263
264 /* MMU-specific code to determine reason for bus error. */
265 movl d1,a0 | fault address
266 movl sp@,d0 | function code from ssw
267 btst #8,d0 | data fault?
268 jne Lbe10a
269 movql #1,d0 | user program access FC
270 | (we dont separate data/program)
271 btst #5,a1@ | supervisor mode?
272 jeq Lbe10a | if no, done
273 movql #5,d0 | else supervisor program access
274 Lbe10a:
275 ptestr d0,a0@,#7 | do a table search
276 pmove psr,sp@ | save result
277 movb sp@,d1
278 btst #2,d1 | invalid? (incl. limit viol and berr)
279 jeq Lmightnotbemerr | no -> wp check
280 btst #7,d1 | is it MMU table berr?
281 jeq Lismerr | no, must be fast
282 jra Lisberr1 | real bus err needs not be fast
283 Lmightnotbemerr:
284 btst #3,d1 | write protect bit set?
285 jeq Lisberr1 | no, must be bus error
286 movl sp@,d0 | ssw into low word of d0
287 andw #0xc0,d0 | write protect is set on page:
288 cmpw #0x40,d0 | was it read cycle?
289 jeq Lisberr1 | yes, was not WPE, must be bus err
290 /* End of MMU-specific bus error code. */
291
292 Lismerr:
293 movl #T_MMUFLT,sp@- | show that we are an MMU fault
294 jra Ltrapnstkadj | and deal with it
295 Lisaerr:
296 movl #T_ADDRERR,sp@- | mark address error
297 jra Ltrapnstkadj | and deal with it
298 Lisberr1:
299 clrw sp@ | re-clear pad word
300 Lisberr:
301 movl #T_BUSERR,sp@- | mark bus error
302 Ltrapnstkadj:
303 jbsr _trap | handle the error
304 lea sp@(12),sp | pop value args
305 movl sp@(FR_SP),a0 | restore user SP
306 movl a0,usp | from save area
307 movw sp@(FR_ADJ),d0 | need to adjust stack?
308 jne Lstkadj | yes, go to it
309 moveml sp@+,#0x7FFF | no, restore most user regs
310 addql #8,sp | toss SSP and stkadj
311 jra rei | all done
312 Lstkadj:
313 lea sp@(FR_HW),a1 | pointer to HW frame
314 addql #8,a1 | source pointer
315 movl a1,a0 | source
316 addw d0,a0 | + hole size = dest pointer
317 movl a1@-,a0@- | copy
318 movl a1@-,a0@- | 8 bytes
319 movl a0,sp@(FR_SP) | new SSP
320 moveml sp@+,#0x7FFF | restore user registers
321 movl sp@,sp | and our SP
322 jra rei | all done
323
324 /*
325 * FP exceptions.
326 */
327 _fpfline:
328 clrl sp@- | stack adjust count
329 moveml #0xFFFF,sp@- | save registers
330 moveq #T_FPEMULI,d0 | denote as FP emulation trap
331 jra fault | do it
332
333 _fpunsupp:
334 clrl sp@- | stack adjust count
335 moveml #0xFFFF,sp@- | save registers
336 moveq #T_FPEMULD,d0 | denote as FP emulation trap
337 jra fault | do it
338
339 /*
340 * Handles all other FP coprocessor exceptions.
341 * Note that since some FP exceptions generate mid-instruction frames
342 * and may cause signal delivery, we need to test for stack adjustment
343 * after the trap call.
344 */
345 .globl _fpfault
346 _fpfault:
347 clrl sp@- | stack adjust count
348 moveml #0xFFFF,sp@- | save user registers
349 movl usp,a0 | and save
350 movl a0,sp@(FR_SP) | the user stack pointer
351 clrl sp@- | no VA arg
352 movl _curpcb,a0 | current pcb
353 lea a0@(PCB_FPCTX),a0 | address of FP savearea
354 fsave a0@ | save state
355 tstb a0@ | null state frame?
356 jeq Lfptnull | yes, safe
357 clrw d0 | no, need to tweak BIU
358 movb a0@(1),d0 | get frame size
359 bset #3,a0@(0,d0:w) | set exc_pend bit of BIU
360 Lfptnull:
361 fmovem fpsr,sp@- | push fpsr as code argument
362 frestore a0@ | restore state
363 movl #T_FPERR,sp@- | push type arg
364 jra Ltrapnstkadj | call trap and deal with stack cleanup
365
366 /*
367 * Coprocessor and format errors can generate mid-instruction stack
368 * frames and cause signal delivery hence we need to check for potential
369 * stack adjustment.
370 */
371 _coperr:
372 clrl sp@- | stack adjust count
373 moveml #0xFFFF,sp@-
374 movl usp,a0 | get and save
375 movl a0,sp@(FR_SP) | the user stack pointer
376 clrl sp@- | no VA arg
377 clrl sp@- | or code arg
378 movl #T_COPERR,sp@- | push trap type
379 jra Ltrapnstkadj | call trap and deal with stack adjustments
380
381 _fmterr:
382 clrl sp@- | stack adjust count
383 moveml #0xFFFF,sp@-
384 movl usp,a0 | get and save
385 movl a0,sp@(FR_SP) | the user stack pointer
386 clrl sp@- | no VA arg
387 clrl sp@- | or code arg
388 movl #T_FMTERR,sp@- | push trap type
389 jra Ltrapnstkadj | call trap and deal with stack adjustments
390
391 /*
392 * Other exceptions only cause four and six word stack frame and require
393 * no post-trap stack adjustment.
394 */
395 _illinst:
396 clrl sp@-
397 moveml #0xFFFF,sp@-
398 moveq #T_ILLINST,d0
399 jra fault
400
401 _zerodiv:
402 clrl sp@-
403 moveml #0xFFFF,sp@-
404 moveq #T_ZERODIV,d0
405 jra fault
406
407 _chkinst:
408 clrl sp@-
409 moveml #0xFFFF,sp@-
410 moveq #T_CHKINST,d0
411 jra fault
412
413 _trapvinst:
414 clrl sp@-
415 moveml #0xFFFF,sp@-
416 moveq #T_TRAPVINST,d0
417 jra fault
418
419 _privinst:
420 clrl sp@-
421 moveml #0xFFFF,sp@-
422 moveq #T_PRIVINST,d0
423 jra fault
424
425 .globl fault
426 fault:
427 movl usp,a0 | get and save
428 movl a0,sp@(FR_SP) | the user stack pointer
429 clrl sp@- | no VA arg
430 clrl sp@- | or code arg
431 movl d0,sp@- | push trap type
432 jbsr _trap | handle trap
433 lea sp@(12),sp | pop value args
434 movl sp@(FR_SP),a0 | restore
435 movl a0,usp | user SP
436 moveml sp@+,#0x7FFF | restore most user regs
437 addql #8,sp | pop SP and stack adjust
438 jra rei | all done
439
440 .globl _straytrap
441 _badtrap:
442 clrl sp@- | stack adjust count
443 moveml #0xFFFF,sp@- | save std frame regs
444 jbsr _straytrap | report
445 moveml sp@+,#0xFFFF | restore regs
446 addql #4, sp | stack adjust count
447 jra rei | all done
448
449 /*
450 * Trap 0 is for system calls
451 */
452 .globl _syscall
453 _trap0:
454 clrl sp@- | stack adjust count
455 moveml #0xFFFF,sp@- | save user registers
456 movl usp,a0 | save the user SP
457 movl a0,sp@(FR_SP) | in the savearea
458 movl d0,sp@- | push syscall number
459 jbsr _syscall | handle it
460 addql #4,sp | pop syscall arg
461 movl sp@(FR_SP),a0 | grab and restore
462 movl a0,usp | user SP
463 moveml sp@+,#0x7FFF | restore most registers
464 addql #8,sp | pop SP and stack adjust
465 jra rei | all done
466
467 /*
468 * Trap 1 is either:
469 * sigreturn (native NetBSD executable)
470 * breakpoint (HPUX executable)
471 */
472 _trap1:
473 #if 0 /* COMPAT_HPUX */
474 /* If process is HPUX, this is a user breakpoint. */
475 jne trap15 | breakpoint
476 #endif
477 /* fall into sigreturn */
478
479 /*
480 * The sigreturn() syscall comes here. It requires special handling
481 * because we must open a hole in the stack to fill in the (possibly much
482 * larger) original stack frame.
483 */
484 sigreturn:
485 lea sp@(-84),sp | leave enough space for largest frame
486 movl sp@(84),sp@ | move up current 8 byte frame
487 movl sp@(88),sp@(4)
488 movl #84,sp@- | default: adjust by 84 bytes
489 moveml #0xFFFF,sp@- | save user registers
490 movl usp,a0 | save the user SP
491 movl a0,sp@(FR_SP) | in the savearea
492 movl #SYS_sigreturn,sp@- | push syscall number
493 jbsr _syscall | handle it
494 addql #4,sp | pop syscall#
495 movl sp@(FR_SP),a0 | grab and restore
496 movl a0,usp | user SP
497 lea sp@(FR_HW),a1 | pointer to HW frame
498 movw sp@(FR_ADJ),d0 | do we need to adjust the stack?
499 jeq Lsigr1 | no, just continue
500 moveq #92,d1 | total size
501 subw d0,d1 | - hole size = frame size
502 lea a1@(92),a0 | destination
503 addw d1,a1 | source
504 lsrw #1,d1 | convert to word count
505 subqw #1,d1 | minus 1 for dbf
506 Lsigrlp:
507 movw a1@-,a0@- | copy a word
508 dbf d1,Lsigrlp | continue
509 movl a0,a1 | new HW frame base
510 Lsigr1:
511 movl a1,sp@(FR_SP) | new SP value
512 moveml sp@+,#0x7FFF | restore user registers
513 movl sp@,sp | and our SP
514 jra rei | all done
515
516 /*
517 * Trap 2 is one of:
518 * NetBSD: not used (ignore)
519 * SunOS: Some obscure FPU operation
520 * HPUX: sigreturn
521 */
522 _trap2:
523 #if 0 /* COMPAT_HPUX */
524 /* XXX: If HPUX, this is a user breakpoint. */
525 jne sigreturn
526 #endif
527 /* fall into trace (NetBSD or SunOS) */
528
529 /*
530 * Trace (single-step) trap. Kernel-mode is special.
531 * User mode traps are simply passed on to trap().
532 */
533 _trace:
534 clrl sp@- | stack adjust count
535 moveml #0xFFFF,sp@-
536 moveq #T_TRACE,d0
537 movw sp@(FR_HW),d1 | get PSW
538 andw #PSL_S,d1 | from system mode?
539 jne kbrkpt | yes, kernel breakpoint
540 jra fault | no, user-mode fault
541
542 /*
543 * Trap 15 is used for:
544 * - GDB breakpoints (in user programs)
545 * - KGDB breakpoints (in the kernel)
546 * - trace traps for SUN binaries (not fully supported yet)
547 * User mode traps are passed simply passed to trap()
548 */
549 _trap15:
550 clrl sp@- | stack adjust count
551 moveml #0xFFFF,sp@-
552 moveq #T_TRAP15,d0
553 movw sp@(FR_HW),d1 | get PSW
554 andw #PSL_S,d1 | from system mode?
555 jne kbrkpt | yes, kernel breakpoint
556 jra fault | no, user-mode fault
557
558 kbrkpt: | Kernel-mode breakpoint or trace trap. (d0=trap_type)
559 | Save the system sp rather than the user sp.
560 movw #PSL_HIGHIPL,sr | lock out interrupts
561 lea sp@(FR_SIZE),a6 | Save stack pointer
562 movl a6,sp@(FR_SP) | from before trap
563
564 | If we are not on tmpstk switch to it.
565 | (so debugger can change the stack pointer)
566 movl a6,d1
567 cmpl #tmpstk,d1
568 jls Lbrkpt2 | already on tmpstk
569 | Copy frame to the temporary stack
570 movl sp,a0 | a0=src
571 lea tmpstk-96,a1 | a1=dst
572 movl a1,sp | sp=new frame
573 moveq #FR_SIZE,d1
574 Lbrkpt1:
575 movl a0@+,a1@+
576 subql #4,d1
577 bgt Lbrkpt1
578
579 Lbrkpt2:
580 | Call the special kernel debugger trap handler.
581 | Do not call trap() to handle it, so that we can
582 | set breakpoints in trap() if we want. We know
583 | the trap type is either T_TRACE or T_BREAKPOINT.
584 movl d0,sp@- | push trap type
585 jbsr _trap_kdebug
586 addql #4,sp | pop args
587
588 | The stack pointer may have been modified, or
589 | data below it modified (by kgdb push call),
590 | so push the hardware frame at the current sp
591 | before restoring registers and returning.
592 movl sp@(FR_SP),a0 | modified sp
593 lea sp@(FR_SIZE),a1 | end of our frame
594 movl a1@-,a0@- | copy 2 longs with
595 movl a1@-,a0@- | ... predecrement
596 movl a0,sp@(FR_SP) | sp = h/w frame
597 moveml sp@+,#0x7FFF | restore all but sp
598 movl sp@,sp | ... and sp
599 rte | all done
600
601 /*
602 * Trap 12 is the entry point for the cachectl "syscall"
603 * cachectl(command, addr, length)
604 * command in d0, addr in a1, length in d1
605 */
606 .globl _cachectl
607 _trap12:
608 movl d1,sp@- | push length
609 movl a1,sp@- | push addr
610 movl d0,sp@- | push command
611 jbsr _cachectl | do it
612 lea sp@(12),sp | pop args
613 jra rei | all done
614
615 /*
616 * Interrupt handlers. Most are auto-vectored,
617 * and hard-wired the same way on all sun3 models.
618 * Format in the stack is:
619 * d0,d1,a0,a1, sr, pc, vo
620 */
621
622 #define INTERRUPT_SAVEREG \
623 moveml #0xC0C0,sp@-
624
625 #define INTERRUPT_RESTORE \
626 moveml sp@+,#0x0303
627
628 /*
629 * This is the common auto-vector interrupt handler,
630 * for which the CPU provides the vector=0x18+level.
631 * These are installed in the interrupt vector table.
632 */
633 .align 2
634 .globl __isr_autovec, _isr_autovec
635 __isr_autovec:
636 INTERRUPT_SAVEREG
637 jbsr _isr_autovec
638 INTERRUPT_RESTORE
639 jra rei
640
641 /* clock: see clock.c */
642 .align 2
643 .globl __isr_clock, _clock_intr
644 __isr_clock:
645 INTERRUPT_SAVEREG
646 jbsr _clock_intr
647 INTERRUPT_RESTORE
648 jra rei
649
650 | Handler for all vectored interrupts (i.e. VME interrupts)
651 .align 2
652 .globl __isr_vectored, _isr_vectored
653 __isr_vectored:
654 INTERRUPT_SAVEREG
655 jbsr _isr_vectored
656 INTERRUPT_RESTORE
657 jra rei
658
659 #undef INTERRUPT_SAVEREG
660 #undef INTERRUPT_RESTORE
661
662 /* interrupt counters (needed by vmstat) */
663 .globl _intrcnt,_eintrcnt,_intrnames,_eintrnames
664 _intrnames:
665 .asciz "spur" | 0
666 .asciz "lev1" | 1
667 .asciz "lev2" | 2
668 .asciz "lev3" | 3
669 .asciz "lev4" | 4
670 .asciz "clock" | 5
671 .asciz "lev6" | 6
672 .asciz "nmi" | 7
673 _eintrnames:
674
675 .data
676 .even
677 _intrcnt:
678 .long 0,0,0,0,0,0,0,0,0,0
679 _eintrcnt:
680 .text
681
682 /*
683 * Emulation of VAX REI instruction.
684 *
685 * This code is (mostly) un-altered from the hp300 code,
686 * except that sun machines do not need a simulated SIR
687 * because they have a real software interrupt register.
688 *
689 * This code deals with checking for and servicing ASTs
690 * (profiling, scheduling) and software interrupts (network, softclock).
691 * We check for ASTs first, just like the VAX. To avoid excess overhead
692 * the T_ASTFLT handling code will also check for software interrupts so we
693 * do not have to do it here. After identifying that we need an AST we
694 * drop the IPL to allow device interrupts.
695 *
696 * This code is complicated by the fact that sendsig may have been called
697 * necessitating a stack cleanup.
698 */
699
700 .globl _astpending
701 .globl rei
702 rei:
703 #ifdef DIAGNOSTIC
704 tstl _panicstr | have we paniced?
705 jne Ldorte | yes, do not make matters worse
706 #endif
707 tstl _astpending | AST pending?
708 jeq Ldorte | no, done
709 Lrei1:
710 btst #5,sp@ | yes, are we returning to user mode?
711 jne Ldorte | no, done
712 movw #PSL_LOWIPL,sr | lower SPL
713 clrl sp@- | stack adjust
714 moveml #0xFFFF,sp@- | save all registers
715 movl usp,a1 | including
716 movl a1,sp@(FR_SP) | the users SP
717 clrl sp@- | VA == none
718 clrl sp@- | code == none
719 movl #T_ASTFLT,sp@- | type == async system trap
720 jbsr _trap | go handle it
721 lea sp@(12),sp | pop value args
722 movl sp@(FR_SP),a0 | restore user SP
723 movl a0,usp | from save area
724 movw sp@(FR_ADJ),d0 | need to adjust stack?
725 jne Laststkadj | yes, go to it
726 moveml sp@+,#0x7FFF | no, restore most user regs
727 addql #8,sp | toss SP and stack adjust
728 rte | and do real RTE
729 Laststkadj:
730 lea sp@(FR_HW),a1 | pointer to HW frame
731 addql #8,a1 | source pointer
732 movl a1,a0 | source
733 addw d0,a0 | + hole size = dest pointer
734 movl a1@-,a0@- | copy
735 movl a1@-,a0@- | 8 bytes
736 movl a0,sp@(FR_SP) | new SSP
737 moveml sp@+,#0x7FFF | restore user registers
738 movl sp@,sp | and our SP
739 Ldorte:
740 rte | real return
741
742 /*
743 * Initialization is at the beginning of this file, because the
744 * kernel entry point needs to be at zero for compatibility with
745 * the Sun boot loader. This works on Sun machines because the
746 * interrupt vector table for reset is NOT at address zero.
747 * (The MMU has a "boot" bit that forces access to the PROM)
748 */
749
750 /*
751 * Signal "trampoline" code (18 bytes). Invoked from RTE setup by sendsig().
752 *
753 * Stack looks like:
754 *
755 * sp+0 -> signal number
756 * sp+4 signal specific code
757 * sp+8 pointer to signal context frame (scp)
758 * sp+12 address of handler
759 * sp+16 saved hardware state
760 * .
761 * .
762 * scp+0-> beginning of signal context frame
763 */
764 .globl _sigcode, _esigcode
765 .data
766 .align 2
767 _sigcode: /* Found at address: 0x0DFFffdc */
768 movl sp@(12),a0 | signal handler addr (4 bytes)
769 jsr a0@ | call signal handler (2 bytes)
770 addql #4,sp | pop signo (2 bytes)
771 trap #1 | special syscall entry (2 bytes)
772 movl d0,sp@(4) | save errno (4 bytes)
773 moveq #1,d0 | syscall == exit (2 bytes)
774 trap #0 | exit(errno) (2 bytes)
775 .align 2
776 _esigcode:
777 .text
778
779 /* XXX - hp300 still has icode here... */
780
781 /*
782 * Primitives
783 */
784 #include <machine/asm.h>
785
786 /*
787 * non-local gotos
788 */
789 ENTRY(setjmp)
790 movl sp@(4),a0 | savearea pointer
791 moveml #0xFCFC,a0@ | save d2-d7/a2-a7
792 movl sp@,a0@(48) | and return address
793 moveq #0,d0 | return 0
794 rts
795
796 ENTRY(longjmp)
797 movl sp@(4),a0
798 moveml a0@+,#0xFCFC
799 movl a0@,sp@
800 moveq #1,d0
801 rts
802
803 /*
804 * The following primitives manipulate the run queues.
805 * _whichqs tells which of the 32 queues _qs have processes in them.
806 * Setrunqueue puts processes into queues, Remrunqueue removes them
807 * from queues. The running process is on no queue, other processes
808 * are on a queue related to p->p_priority, divided by 4 actually to
809 * shrink the 0-127 range of priorities into the 32 available queues.
810 */
811
812 .globl _whichqs,_qs,_cnt,_panic
813 .globl _curproc
814 .comm _want_resched,4
815
816 /*
817 * setrunqueue(p)
818 *
819 * Call should be made at splclock(), and p->p_stat should be SRUN
820 */
821 ENTRY(setrunqueue)
822 movl sp@(4),a0
823 #ifdef DIAGNOSTIC
824 tstl a0@(P_BACK)
825 jne Lset1
826 tstl a0@(P_WCHAN)
827 jne Lset1
828 cmpb #SRUN,a0@(P_STAT)
829 jne Lset1
830 #endif
831 clrl d0
832 movb a0@(P_PRIORITY),d0
833 lsrb #2,d0
834 movl _whichqs,d1
835 bset d0,d1
836 movl d1,_whichqs
837 lslb #3,d0
838 addl #_qs,d0
839 movl d0,a0@(P_FORW)
840 movl d0,a1
841 movl a1@(P_BACK),a0@(P_BACK)
842 movl a0,a1@(P_BACK)
843 movl a0@(P_BACK),a1
844 movl a0,a1@(P_FORW)
845 rts
846 #ifdef DIAGNOSTIC
847 Lset1:
848 movl #Lset2,sp@-
849 jbsr _panic
850 Lset2:
851 .asciz "setrunqueue"
852 .even
853 #endif
854
855 /*
856 * remrunqueue(p)
857 *
858 * Call should be made at splclock().
859 */
860 ENTRY(remrunqueue)
861 movl sp@(4),a0 | proc *p
862 clrl d0
863 movb a0@(P_PRIORITY),d0
864 lsrb #2,d0
865 movl _whichqs,d1
866 bclr d0,d1 | if ((d1 & (1 << d0)) == 0)
867 jeq Lrem2 | panic (empty queue)
868 movl d1,_whichqs
869 movl a0@(P_FORW),a1
870 movl a0@(P_BACK),a1@(P_BACK)
871 movl a0@(P_BACK),a1
872 movl a0@(P_FORW),a1@(P_FORW)
873 movl #_qs,a1
874 movl d0,d1
875 lslb #3,d1
876 addl d1,a1
877 cmpl a1@(P_FORW),a1
878 jeq Lrem1
879 movl _whichqs,d1
880 bset d0,d1
881 movl d1,_whichqs
882 Lrem1:
883 clrl a0@(P_BACK)
884 rts
885 Lrem2:
886 movl #Lrem3,sp@-
887 jbsr _panic
888 Lrem3:
889 .asciz "remrunqueue"
890 .even
891
892 | Message for Lbadsw panic
893 Lsw0:
894 .asciz "cpu_switch"
895 .even
896
897 .globl _curpcb
898 .globl _masterpaddr | XXX compatibility (debuggers)
899 .data
900 _masterpaddr: | XXX compatibility (debuggers)
901 _curpcb:
902 .long 0
903 .comm nullpcb,SIZEOF_PCB
904 .text
905
906 /*
907 * At exit of a process, do a cpu_switch for the last time.
908 * Switch to a safe stack and PCB, and deallocate the process's resources.
909 * The ipl is high enough to prevent the memory from being reallocated.
910 */
911 ENTRY(switch_exit)
912 movl sp@(4),a0 | struct proc *p
913 movl #nullpcb,_curpcb | save state into garbage pcb
914 lea tmpstk,sp | goto a tmp stack
915 movl a0,sp@- | pass proc ptr down
916
917 /* Free old process's u-area. */
918 movl #USPACE,sp@- | size of u-area
919 movl a0@(P_ADDR),sp@- | address of process's u-area
920 movl _kernel_map,sp@- | map it was allocated in
921 jbsr _kmem_free | deallocate it
922 lea sp@(12),sp | pop args
923
924 jra _cpu_switch
925
926 /*
927 * When no processes are on the runq, cpu_switch() branches to idle
928 * to wait for something to come ready.
929 */
930 .data
931 .globl _Idle_count
932 _Idle_count:
933 .long 0
934 .text
935
936 .globl Idle
937 Lidle:
938 stop #PSL_LOWIPL
939 Idle:
940 movw #PSL_HIGHIPL,sr
941 addql #1, _Idle_count
942 tstl _whichqs
943 jeq Lidle
944 movw #PSL_LOWIPL,sr
945 jra Lsw1
946
947 Lbadsw:
948 movl #Lsw0,sp@-
949 jbsr _panic
950 /*NOTREACHED*/
951
952 /*
953 * cpu_switch()
954 * Hacked for sun3
955 * XXX - Arg 1 is a proc pointer (curproc) but this doesn't use it.
956 * XXX - Sould we use p->p_addr instead of curpcb? -gwr
957 */
958 ENTRY(cpu_switch)
959 movl _curpcb,a1 | current pcb
960 movw sr,a1@(PCB_PS) | save sr before changing ipl
961 #ifdef notyet
962 movl _curproc,sp@- | remember last proc running
963 #endif
964 clrl _curproc
965
966 Lsw1:
967 /*
968 * Find the highest-priority queue that isn't empty,
969 * then take the first proc from that queue.
970 */
971 clrl d0
972 lea _whichqs,a0
973 movl a0@,d1
974 Lswchk:
975 btst d0,d1
976 jne Lswfnd
977 addqb #1,d0
978 cmpb #32,d0
979 jne Lswchk
980 jra Idle
981 Lswfnd:
982 movw #PSL_HIGHIPL,sr | lock out interrupts
983 movl a0@,d1 | and check again...
984 bclr d0,d1
985 jeq Lsw1 | proc moved, rescan
986 movl d1,a0@ | update whichqs
987 moveq #1,d1 | double check for higher priority
988 lsll d0,d1 | process (which may have snuck in
989 subql #1,d1 | while we were finding this one)
990 andl a0@,d1
991 jeq Lswok | no one got in, continue
992 movl a0@,d1
993 bset d0,d1 | otherwise put this one back
994 movl d1,a0@
995 jra Lsw1 | and rescan
996 Lswok:
997 movl d0,d1
998 lslb #3,d1 | convert queue number to index
999 addl #_qs,d1 | locate queue (q)
1000 movl d1,a1
1001 cmpl a1@(P_FORW),a1 | anyone on queue?
1002 jeq Lbadsw | no, panic
1003 movl a1@(P_FORW),a0 | p = q->p_forw
1004 movl a0@(P_FORW),a1@(P_FORW) | q->p_forw = p->p_forw
1005 movl a0@(P_FORW),a1 | q = p->p_forw
1006 movl a0@(P_BACK),a1@(P_BACK) | q->p_back = p->p_back
1007 cmpl a0@(P_FORW),d1 | anyone left on queue?
1008 jeq Lsw2 | no, skip
1009 movl _whichqs,d1
1010 bset d0,d1 | yes, reset bit
1011 movl d1,_whichqs
1012 Lsw2:
1013 movl a0,_curproc
1014 clrl _want_resched
1015 #ifdef notyet
1016 movl sp@+,a1 | XXX - Make this work!
1017 cmpl a0,a1 | switching to same proc?
1018 jeq Lswdone | yes, skip save and restore
1019 #endif
1020 /*
1021 * Save state of previous process in its pcb.
1022 */
1023 movl _curpcb,a1
1024 moveml #0xFCFC,a1@(PCB_REGS) | save non-scratch registers
1025 movl usp,a2 | grab USP (a2 has been saved)
1026 movl a2,a1@(PCB_USP) | and save it
1027
1028 tstl _fpu_type | Do we have an fpu?
1029 jeq Lswnofpsave | No? Then don't try save.
1030 lea a1@(PCB_FPCTX),a2 | pointer to FP save area
1031 fsave a2@ | save FP state
1032 tstb a2@ | null state frame?
1033 jeq Lswnofpsave | yes, all done
1034 fmovem fp0-fp7,a2@(FPF_REGS) | save FP general regs
1035 fmovem fpcr/fpsr/fpi,a2@(FPF_FPCR) | save FP control regs
1036 Lswnofpsave:
1037
1038 /*
1039 * Now that we have saved all the registers that must be
1040 * preserved, we are free to use those registers until
1041 * we load the registers for the switched-to process.
1042 * In this section, keep: a0=curproc, a1=curpcb
1043 */
1044
1045 #ifdef DIAGNOSTIC
1046 tstl a0@(P_WCHAN)
1047 jne Lbadsw
1048 cmpb #SRUN,a0@(P_STAT)
1049 jne Lbadsw
1050 #endif
1051 clrl a0@(P_BACK) | clear back link
1052 movl a0@(P_ADDR),a1 | get p_addr
1053 movl a1,_curpcb
1054
1055 /*
1056 * Load the new VM context (new MMU root pointer)
1057 */
1058 movl a0@(P_VMSPACE),a2 | vm = p->p_vmspace
1059 #ifdef DIAGNOSTIC
1060 tstl a2 | map == VM_MAP_NULL?
1061 jeq Lbadsw | panic
1062 #endif
1063 #ifdef PMAP_DEBUG
1064 /*
1065 * Just call pmap_activate() for now. Later on,
1066 * use the in-line version below (for speed).
1067 */
1068 lea a2@(VM_PMAP),a2 | pmap = &vmspace.vm_pmap
1069 pea a2@ | push pmap
1070 jbsr _pmap_activate | pmap_activate(pmap)
1071 addql #4,sp
1072 movl _curpcb,a1 | restore p_addr
1073 #else
1074 /* XXX - Later, use this inline version. */
1075 /* Just load the new CPU Root Pointer (MMU) */
1076 lea _kernel_crp, a3 | our CPU Root Ptr. (CRP)
1077 lea a2@(VM_PMAP),a2 | pmap = &vmspace.vm_pmap
1078 movl a2@(PM_A_PHYS),d0 | phys = pmap->pm_a_phys
1079 cmpl a3@(4),d0 | == kernel_crp.rp_addr ?
1080 jeq Lsame_mmuctx | skip loadcrp/flush
1081 /* OK, it is a new MMU context. Load it up. */
1082 movl d0,a3@(4)
1083 movl #CACHE_CLR,d0
1084 movc d0,cacr | invalidate cache(s)
1085 pflusha | flush entire TLB
1086 pmove a3@,crp | load new user root pointer
1087 Lsame_mmuctx:
1088 #endif
1089
1090 /*
1091 * Reload the registers for the new process.
1092 * After this point we can only use d0,d1,a0,a1
1093 */
1094 moveml a1@(PCB_REGS),#0xFCFC | reload registers
1095 movl a1@(PCB_USP),a0
1096 movl a0,usp | and USP
1097
1098 tstl _fpu_type | If we don't have an fpu,
1099 jeq Lres_skip | don't try to restore it.
1100 lea a1@(PCB_FPCTX),a0 | pointer to FP save area
1101 tstb a0@ | null state frame?
1102 jeq Lresfprest | yes, easy
1103 fmovem a0@(FPF_FPCR),fpcr/fpsr/fpi | restore FP control regs
1104 fmovem a0@(FPF_REGS),fp0-fp7 | restore FP general regs
1105 Lresfprest:
1106 frestore a0@ | restore state
1107 Lres_skip:
1108 movw a1@(PCB_PS),d0 | no, restore PS
1109 #ifdef DIAGNOSTIC
1110 btst #13,d0 | supervisor mode?
1111 jeq Lbadsw | no? panic!
1112 #endif
1113 movw d0,sr | OK, restore PS
1114 moveq #1,d0 | return 1 (for alternate returns)
1115 rts
1116
1117 /*
1118 * savectx(pcb)
1119 * Update pcb, saving current processor state.
1120 */
1121 ENTRY(savectx)
1122 movl sp@(4),a1
1123 movw sr,a1@(PCB_PS)
1124 movl usp,a0 | grab USP
1125 movl a0,a1@(PCB_USP) | and save it
1126 moveml #0xFCFC,a1@(PCB_REGS) | save non-scratch registers
1127
1128 tstl _fpu_type | Do we have FPU?
1129 jeq Lsavedone | No? Then don't save state.
1130 lea a1@(PCB_FPCTX),a0 | pointer to FP save area
1131 fsave a0@ | save FP state
1132 tstb a0@ | null state frame?
1133 jeq Lsavedone | yes, all done
1134 fmovem fp0-fp7,a0@(FPF_REGS) | save FP general regs
1135 fmovem fpcr/fpsr/fpi,a0@(FPF_FPCR) | save FP control regs
1136 Lsavedone:
1137 moveq #0,d0 | return 0
1138 rts
1139
1140 /* suline() `040 only */
1141
1142 #ifdef DEBUG
1143 .data
1144 .globl fulltflush, fullcflush
1145 fulltflush:
1146 .long 0
1147 fullcflush:
1148 .long 0
1149 .text
1150 #endif
1151
1152 /*
1153 * Invalidate entire TLB.
1154 */
1155 ENTRY(TBIA)
1156 __TBIA:
1157 pflusha
1158 movl #DC_CLEAR,d0
1159 movc d0,cacr | invalidate on-chip d-cache
1160 rts
1161
1162 /*
1163 * Invalidate any TLB entry for given VA (TB Invalidate Single)
1164 */
1165 ENTRY(TBIS)
1166 #ifdef DEBUG
1167 tstl fulltflush | being conservative?
1168 jne __TBIA | yes, flush entire TLB
1169 #endif
1170 movl sp@(4),a0
1171 pflush #0,#0,a0@ | flush address from both sides
1172 movl #DC_CLEAR,d0
1173 movc d0,cacr | invalidate on-chip data cache
1174 rts
1175
1176 /*
1177 * Invalidate supervisor side of TLB
1178 */
1179 ENTRY(TBIAS)
1180 #ifdef DEBUG
1181 tstl fulltflush | being conservative?
1182 jne __TBIA | yes, flush everything
1183 #endif
1184 pflush #4,#4 | flush supervisor TLB entries
1185 movl #DC_CLEAR,d0
1186 movc d0,cacr | invalidate on-chip d-cache
1187 rts
1188
1189 /*
1190 * Invalidate user side of TLB
1191 */
1192 ENTRY(TBIAU)
1193 #ifdef DEBUG
1194 tstl fulltflush | being conservative?
1195 jne __TBIA | yes, flush everything
1196 #endif
1197 pflush #0,#4 | flush user TLB entries
1198 movl #DC_CLEAR,d0
1199 movc d0,cacr | invalidate on-chip d-cache
1200 rts
1201
1202 /*
1203 * Invalidate instruction cache
1204 */
1205 ENTRY(ICIA)
1206 movl #IC_CLEAR,d0
1207 movc d0,cacr | invalidate i-cache
1208 rts
1209
1210 /*
1211 * Invalidate data cache.
1212 * NOTE: we do not flush 68030 on-chip cache as there are no aliasing
1213 * problems with DC_WA. The only cases we have to worry about are context
1214 * switch and TLB changes, both of which are handled "in-line" in resume
1215 * and TBI*.
1216 */
1217 ENTRY(DCIA)
1218 __DCIA:
1219 rts
1220
1221 ENTRY(DCIS)
1222 __DCIS:
1223 rts
1224
1225 /*
1226 * Invalidate data cache.
1227 */
1228 ENTRY(DCIU)
1229 rts
1230
1231 /* ICPL, ICPP, DCPL, DCPP, DCPA, DCFL, DCFP */
1232
1233 ENTRY(PCIA)
1234 movl #DC_CLEAR,d0
1235 movc d0,cacr | invalidate on-chip d-cache
1236 rts
1237
1238 ENTRY(ecacheon)
1239 rts
1240
1241 ENTRY(ecacheoff)
1242 rts
1243
1244 /*
1245 * Get callers current SP value.
1246 * Note that simply taking the address of a local variable in a C function
1247 * doesn't work because callee saved registers may be outside the stack frame
1248 * defined by A6 (e.g. GCC generated code).
1249 *
1250 * [I don't think the ENTRY() macro will do the right thing with this -- glass]
1251 */
1252 .globl _getsp
1253 _getsp:
1254 movl sp,d0 | get current SP
1255 addql #4,d0 | compensate for return address
1256 rts
1257
1258 ENTRY(getsfc)
1259 movc sfc,d0
1260 rts
1261
1262 ENTRY(getdfc)
1263 movc dfc,d0
1264 rts
1265
1266 ENTRY(getvbr)
1267 movc vbr, d0
1268 rts
1269
1270 ENTRY(setvbr)
1271 movl sp@(4), d0
1272 movc d0, vbr
1273 rts
1274
1275 /*
1276 * Load a new CPU Root Pointer (CRP) into the MMU.
1277 * void loadcrp(struct mmu_rootptr *);
1278 */
1279 ENTRY(loadcrp)
1280 movl sp@(4),a0 | arg1: &CRP
1281 movl #CACHE_CLR,d0
1282 movc d0,cacr | invalidate cache(s)
1283 pflusha | flush entire TLB
1284 pmove a0@,crp | load new user root pointer
1285 rts
1286
1287 /*
1288 * Set processor priority level calls. Most are implemented with
1289 * inline asm expansions. However, we need one instantiation here
1290 * in case some non-optimized code makes external references.
1291 * Most places will use the inlined function param.h supplies.
1292 */
1293
1294 ENTRY(_spl)
1295 movl sp@(4),d1
1296 clrl d0
1297 movw sr,d0
1298 movw d1,sr
1299 rts
1300
1301 ENTRY(getsr)
1302 moveq #0, d0
1303 movw sr, d0
1304 rts
1305
1306 ENTRY(_insque)
1307 movw sr,d0
1308 movw #PSL_HIGHIPL,sr | atomic
1309 movl sp@(8),a0 | where to insert (after)
1310 movl sp@(4),a1 | element to insert (e)
1311 movl a0@,a1@ | e->next = after->next
1312 movl a0,a1@(4) | e->prev = after
1313 movl a1,a0@ | after->next = e
1314 movl a1@,a0
1315 movl a1,a0@(4) | e->next->prev = e
1316 movw d0,sr
1317 rts
1318
1319 ENTRY(_remque)
1320 movw sr,d0
1321 movw #PSL_HIGHIPL,sr | atomic
1322 movl sp@(4),a0 | element to remove (e)
1323 movl a0@,a1
1324 movl a0@(4),a0
1325 movl a0,a1@(4) | e->next->prev = e->prev
1326 movl a1,a0@ | e->prev->next = e->next
1327 movw d0,sr
1328 rts
1329
1330 /*
1331 * Save and restore 68881 state.
1332 */
1333 ENTRY(m68881_save)
1334 movl sp@(4),a0 | save area pointer
1335 fsave a0@ | save state
1336 tstb a0@ | null state frame?
1337 jeq Lm68881sdone | yes, all done
1338 fmovem fp0-fp7,a0@(FPF_REGS) | save FP general regs
1339 fmovem fpcr/fpsr/fpi,a0@(FPF_FPCR) | save FP control regs
1340 Lm68881sdone:
1341 rts
1342
1343 ENTRY(m68881_restore)
1344 movl sp@(4),a0 | save area pointer
1345 tstb a0@ | null state frame?
1346 jeq Lm68881rdone | yes, easy
1347 fmovem a0@(FPF_FPCR),fpcr/fpsr/fpi | restore FP control regs
1348 fmovem a0@(FPF_REGS),fp0-fp7 | restore FP general regs
1349 Lm68881rdone:
1350 frestore a0@ | restore state
1351 rts
1352
1353 /*
1354 * _delay(unsigned N)
1355 * Delay for at least (N/256) microseconds.
1356 * This routine depends on the variable: delay_divisor
1357 * which should be set based on the CPU clock rate.
1358 * XXX: Currently this is set in sun3_startup.c based on the
1359 * XXX: CPU model but this should be determined at run time...
1360 */
1361 .globl __delay
1362 __delay:
1363 | d0 = arg = (usecs << 8)
1364 movl sp@(4),d0
1365 | d1 = delay_divisor;
1366 movl _delay_divisor,d1
1367 L_delay:
1368 subl d1,d0
1369 jgt L_delay
1370 rts
1371
1372
1373 | Define some addresses, mostly so DDB can print useful info.
1374 .globl _kernbase
1375 .set _kernbase,KERNBASE
1376 .globl _dvma_base
1377 .set _dvma_base,DVMA_SPACE_START
1378 .globl _prom_start
1379 .set _prom_start,MONSTART
1380 .globl _prom_base
1381 .set _prom_base,PROM_BASE
1382
1383 |The end!
1384