Home | History | Annotate | Line # | Download | only in arm32
cpuswitch.S revision 1.94
      1  1.94     skrll /*	$NetBSD: cpuswitch.S,v 1.94 2019/09/13 06:30:10 skrll Exp $	*/
      2   1.1     chris 
      3   1.1     chris /*
      4  1.30       scw  * Copyright 2003 Wasabi Systems, Inc.
      5  1.30       scw  * All rights reserved.
      6  1.30       scw  *
      7  1.30       scw  * Written by Steve C. Woodford for Wasabi Systems, Inc.
      8  1.30       scw  *
      9  1.30       scw  * Redistribution and use in source and binary forms, with or without
     10  1.30       scw  * modification, are permitted provided that the following conditions
     11  1.30       scw  * are met:
     12  1.30       scw  * 1. Redistributions of source code must retain the above copyright
     13  1.30       scw  *    notice, this list of conditions and the following disclaimer.
     14  1.30       scw  * 2. Redistributions in binary form must reproduce the above copyright
     15  1.30       scw  *    notice, this list of conditions and the following disclaimer in the
     16  1.30       scw  *    documentation and/or other materials provided with the distribution.
     17  1.30       scw  * 3. All advertising materials mentioning features or use of this software
     18  1.30       scw  *    must display the following acknowledgement:
     19  1.30       scw  *      This product includes software developed for the NetBSD Project by
     20  1.30       scw  *      Wasabi Systems, Inc.
     21  1.30       scw  * 4. The name of Wasabi Systems, Inc. may not be used to endorse
     22  1.30       scw  *    or promote products derived from this software without specific prior
     23  1.30       scw  *    written permission.
     24  1.30       scw  *
     25  1.30       scw  * THIS SOFTWARE IS PROVIDED BY WASABI SYSTEMS, INC. ``AS IS'' AND
     26  1.30       scw  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
     27  1.30       scw  * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
     28  1.30       scw  * PURPOSE ARE DISCLAIMED.  IN NO EVENT SHALL WASABI SYSTEMS, INC
     29  1.30       scw  * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
     30  1.30       scw  * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
     31  1.30       scw  * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
     32  1.30       scw  * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
     33  1.30       scw  * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
     34  1.30       scw  * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
     35  1.30       scw  * POSSIBILITY OF SUCH DAMAGE.
     36  1.30       scw  */
     37  1.30       scw /*
     38   1.1     chris  * Copyright (c) 1994-1998 Mark Brinicombe.
     39   1.1     chris  * Copyright (c) 1994 Brini.
     40   1.1     chris  * All rights reserved.
     41   1.1     chris  *
     42   1.1     chris  * This code is derived from software written for Brini by Mark Brinicombe
     43   1.1     chris  *
     44   1.1     chris  * Redistribution and use in source and binary forms, with or without
     45   1.1     chris  * modification, are permitted provided that the following conditions
     46   1.1     chris  * are met:
     47   1.1     chris  * 1. Redistributions of source code must retain the above copyright
     48   1.1     chris  *    notice, this list of conditions and the following disclaimer.
     49   1.1     chris  * 2. Redistributions in binary form must reproduce the above copyright
     50   1.1     chris  *    notice, this list of conditions and the following disclaimer in the
     51   1.1     chris  *    documentation and/or other materials provided with the distribution.
     52   1.1     chris  * 3. All advertising materials mentioning features or use of this software
     53   1.1     chris  *    must display the following acknowledgement:
     54   1.1     chris  *	This product includes software developed by Brini.
     55   1.1     chris  * 4. The name of the company nor the name of the author may be used to
     56   1.1     chris  *    endorse or promote products derived from this software without specific
     57   1.1     chris  *    prior written permission.
     58   1.1     chris  *
     59   1.1     chris  * THIS SOFTWARE IS PROVIDED BY BRINI ``AS IS'' AND ANY EXPRESS OR IMPLIED
     60   1.1     chris  * WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF
     61   1.1     chris  * MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
     62   1.1     chris  * IN NO EVENT SHALL BRINI OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT,
     63   1.1     chris  * INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
     64   1.1     chris  * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
     65   1.1     chris  * SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
     66   1.1     chris  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
     67   1.1     chris  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
     68   1.1     chris  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
     69   1.1     chris  * SUCH DAMAGE.
     70   1.1     chris  *
     71   1.1     chris  * RiscBSD kernel project
     72   1.1     chris  *
     73   1.1     chris  * cpuswitch.S
     74   1.1     chris  *
     75   1.1     chris  * cpu switching functions
     76   1.1     chris  *
     77   1.1     chris  * Created      : 15/10/94
     78   1.1     chris  */
     79   1.1     chris 
     80   1.1     chris #include "opt_armfpe.h"
     81  1.30       scw #include "opt_arm32_pmap.h"
     82  1.19     bjh21 #include "opt_multiprocessor.h"
     83  1.58      matt #include "opt_cpuoptions.h"
     84  1.36    martin #include "opt_lockdebug.h"
     85   1.1     chris 
     86   1.1     chris #include "assym.h"
     87  1.79      matt #include <arm/asm.h>
     88  1.79      matt #include <arm/locore.h>
     89  1.58      matt 
     90  1.94     skrll 	RCSID("$NetBSD: cpuswitch.S,v 1.94 2019/09/13 06:30:10 skrll Exp $")
     91   1.1     chris 
     92  1.34  kristerw /* LINTSTUB: include <sys/param.h> */
     93  1.90      matt 
     94   1.1     chris #undef IRQdisable
     95   1.1     chris #undef IRQenable
     96   1.1     chris 
     97   1.1     chris /*
     98   1.1     chris  * New experimental definitions of IRQdisable and IRQenable
     99   1.1     chris  * These keep FIQ's enabled since FIQ's are special.
    100   1.1     chris  */
    101   1.1     chris 
    102  1.58      matt #ifdef _ARM_ARCH_6
    103  1.58      matt #define	IRQdisable	cpsid	i
    104  1.58      matt #define	IRQenable	cpsie	i
    105  1.58      matt #else
    106   1.1     chris #define IRQdisable \
    107  1.13   thorpej 	mrs	r14, cpsr ; \
    108   1.1     chris 	orr	r14, r14, #(I32_bit) ; \
    109  1.58      matt 	msr	cpsr_c, r14
    110   1.1     chris 
    111   1.1     chris #define IRQenable \
    112  1.13   thorpej 	mrs	r14, cpsr ; \
    113   1.1     chris 	bic	r14, r14, #(I32_bit) ; \
    114  1.58      matt 	msr	cpsr_c, r14
    115   1.1     chris 
    116  1.22     bjh21 #endif
    117   1.1     chris 
    118   1.1     chris 	.text
    119  1.30       scw 
    120   1.1     chris /*
    121  1.47      yamt  * struct lwp *
    122  1.47      yamt  * cpu_switchto(struct lwp *current, struct lwp *next)
    123  1.48     skrll  *
    124  1.47      yamt  * Switch to the specified next LWP
    125  1.47      yamt  * Arguments:
    126  1.16   thorpej  *
    127  1.58      matt  *	r0	'struct lwp *' of the current LWP (or NULL if exiting)
    128  1.47      yamt  *	r1	'struct lwp *' of the LWP to switch to
    129  1.58      matt  *	r2	returning
    130   1.1     chris  */
    131  1.47      yamt ENTRY(cpu_switchto)
    132  1.51     skrll 	mov	ip, sp
    133  1.78      matt 	push	{r4-r7, ip, lr}
    134   1.1     chris 
    135  1.58      matt 	/* move lwps into caller saved registers */
    136  1.55     chris 	mov	r6, r1
    137  1.55     chris 	mov	r4, r0
    138  1.58      matt 
    139  1.68      matt #ifdef TPIDRPRW_IS_CURCPU
    140  1.88      matt 	GET_CURCPU(r5)
    141  1.90      matt #else
    142  1.88      matt 	ldr	r5, [r6, #L_CPU]		/* get cpu from new lwp */
    143  1.59      matt #endif
    144   1.1     chris 
    145  1.47      yamt 	/* rem: r4 = old lwp */
    146  1.88      matt 	/* rem: r5 = curcpu() */
    147  1.43     skrll 	/* rem: r6 = new lwp */
    148   1.4     chris 	/* rem: interrupts are enabled */
    149   1.1     chris 
    150   1.1     chris 	/*
    151  1.47      yamt 	 * If the old lwp on entry to cpu_switchto was zero then the
    152   1.1     chris 	 * process that called it was exiting. This means that we do
    153   1.1     chris 	 * not need to save the current context. Instead we can jump
    154   1.1     chris 	 * straight to restoring the context for the new process.
    155   1.1     chris 	 */
    156  1.58      matt 	teq	r4, #0
    157  1.49       scw 	beq	.Ldo_switch
    158   1.1     chris 
    159  1.47      yamt 	/* rem: r4 = old lwp */
    160  1.88      matt 	/* rem: r5 = curcpu() */
    161  1.43     skrll 	/* rem: r6 = new lwp */
    162   1.4     chris 	/* rem: interrupts are enabled */
    163   1.1     chris 
    164  1.48     skrll 	/* Save old context */
    165   1.1     chris 
    166  1.29   thorpej 	/* Get the user structure for the old lwp. */
    167  1.88      matt 	ldr	r7, [r4, #(L_PCB)]
    168   1.1     chris 
    169  1.29   thorpej 	/* Save all the registers in the old lwp's pcb */
    170  1.76      matt #if defined(_ARM_ARCH_DWORD_OK)
    171  1.88      matt 	strd	r8, r9, [r7, #(PCB_R8)]
    172  1.88      matt 	strd	r10, r11, [r7, #(PCB_R10)]
    173  1.88      matt 	strd	r12, r13, [r7, #(PCB_R12)]
    174  1.58      matt #else
    175  1.88      matt 	add	r0, r7, #(PCB_R8)
    176  1.58      matt 	stmia	r0, {r8-r13}
    177  1.37       scw #endif
    178   1.1     chris 
    179  1.58      matt #ifdef _ARM_ARCH_6
    180  1.58      matt 	/*
    181  1.58      matt 	 * Save user read/write thread/process id register
    182  1.58      matt 	 */
    183  1.58      matt 	mrc	p15, 0, r0, c13, c0, 2
    184  1.88      matt 	str	r0, [r7, #(PCB_USER_PID_RW)]
    185  1.58      matt #endif
    186   1.1     chris 	/*
    187  1.29   thorpej 	 * NOTE: We can now use r8-r13 until it is time to restore
    188  1.29   thorpej 	 * them for the new process.
    189  1.29   thorpej 	 */
    190  1.29   thorpej 
    191  1.47      yamt 	/* rem: r4 = old lwp */
    192  1.88      matt 	/* rem: r5 = curcpu() */
    193  1.47      yamt 	/* rem: r6 = new lwp */
    194  1.47      yamt 	/* rem: interrupts are enabled */
    195  1.47      yamt 
    196  1.48     skrll 	/* Restore saved context */
    197   1.1     chris 
    198  1.49       scw .Ldo_switch:
    199  1.90      matt 	/* rem: r4 = old lwp */
    200  1.90      matt 	/* rem: r5 = curcpu() */
    201  1.90      matt 	/* rem: r6 = new lwp */
    202  1.90      matt 
    203  1.90      matt 	IRQdisable
    204  1.90      matt #if defined(TPIDRPRW_IS_CURLWP)
    205  1.90      matt 	mcr	p15, 0, r6, c13, c0, 4		/* set current lwp */
    206  1.90      matt #endif
    207  1.90      matt 
    208  1.90      matt 	/* We have a new curlwp now so make a note of it */
    209  1.90      matt 	str	r6, [r5, #(CI_CURLWP)]
    210  1.88      matt 	/* Get the new pcb */
    211  1.88      matt 	ldr	r7, [r6, #(L_PCB)]
    212  1.88      matt 
    213  1.90      matt 	/* make sure we are using the new lwp's stack */
    214  1.90      matt 	ldr	sp, [r7, #(PCB_KSP)]
    215  1.90      matt 
    216  1.90      matt 	/* At this point we can allow IRQ's again. */
    217  1.90      matt 	IRQenable
    218  1.90      matt 
    219  1.47      yamt 	/* rem: r4 = old lwp */
    220  1.88      matt 	/* rem: r5 = curcpu() */
    221  1.29   thorpej 	/* rem: r6 = new lwp */
    222  1.55     chris 	/* rem: r7 = new pcb */
    223  1.53     chris 	/* rem: interrupts are enabled */
    224  1.29   thorpej 
    225  1.88      matt 	/*
    226  1.88      matt 	 * If we are switching to a system lwp, don't bother restoring
    227  1.88      matt 	 * thread or vfp registers and skip the ras check.
    228  1.88      matt 	 */
    229  1.88      matt 	ldr	r0, [r6, #(L_FLAG)]
    230  1.88      matt 	tst	r0, #(LW_SYSTEM)
    231  1.88      matt 	bne	.Lswitch_do_restore
    232  1.88      matt 
    233  1.58      matt #ifdef _ARM_ARCH_6
    234  1.58      matt 	/*
    235  1.58      matt 	 * Restore user thread/process id registers
    236  1.58      matt 	 */
    237  1.58      matt 	ldr	r0, [r7, #(PCB_USER_PID_RW)]
    238  1.58      matt 	mcr	p15, 0, r0, c13, c0, 2
    239  1.63      matt 	ldr	r0, [r6, #(L_PRIVATE)]
    240  1.58      matt 	mcr	p15, 0, r0, c13, c0, 3
    241  1.58      matt #endif
    242  1.58      matt 
    243  1.76      matt #ifdef FPU_VFP
    244  1.76      matt 	/*
    245  1.76      matt 	 * If we have a VFP, we need to load FPEXC.
    246  1.76      matt 	 */
    247  1.88      matt 	ldr	r0, [r5, #(CI_VFP_ID)]
    248  1.76      matt 	cmp	r0, #0
    249  1.76      matt 	ldrne	r0, [r7, #(PCB_VFP_FPEXC)]
    250  1.81     joerg 	vmsrne	fpexc, r0
    251  1.76      matt #endif
    252  1.76      matt 
    253  1.91     skrll 	/*
    254  1.88      matt 	 * Check for restartable atomic sequences (RAS).
    255  1.88      matt 	 */
    256  1.55     chris 
    257  1.88      matt 	ldr	r0, [r6, #(L_PROC)]	/* fetch the proc for ras_lookup */
    258  1.88      matt 	ldr	r2, [r0, #(P_RASLIST)]
    259  1.88      matt 	cmp	r2, #0			/* p->p_nras == 0? */
    260  1.88      matt 	beq	.Lswitch_do_restore
    261  1.88      matt 
    262  1.88      matt 	/* we can use r8 since we haven't restored saved registers yet. */
    263  1.88      matt 	ldr	r8, [r6, #(L_MD_TF)]	/* r1 = trapframe (used below) */
    264  1.88      matt 	ldr	r1, [r8, #(TF_PC)]	/* second ras_lookup() arg */
    265  1.88      matt 	bl	_C_LABEL(ras_lookup)
    266  1.88      matt 	cmn	r0, #1			/* -1 means "not in a RAS" */
    267  1.88      matt 	strne	r0, [r8, #(TF_PC)]
    268  1.88      matt 
    269  1.88      matt 	/* rem: r4 = old lwp */
    270  1.88      matt 	/* rem: r5 = curcpu() */
    271  1.88      matt 	/* rem: r6 = new lwp */
    272  1.88      matt 	/* rem: r7 = new pcb */
    273  1.88      matt 
    274  1.88      matt .Lswitch_do_restore:
    275  1.52     skrll 	/* Restore all the saved registers */
    276  1.58      matt #ifdef __XSCALE__
    277  1.37       scw 	ldr	r8, [r7, #(PCB_R8)]
    278  1.37       scw 	ldr	r9, [r7, #(PCB_R9)]
    279  1.37       scw 	ldr	r10, [r7, #(PCB_R10)]
    280  1.37       scw 	ldr	r11, [r7, #(PCB_R11)]
    281  1.37       scw 	ldr	r12, [r7, #(PCB_R12)]
    282  1.76      matt #elif defined(_ARM_ARCH_DWORD_OK)
    283  1.80     joerg 	ldrd	r8, r9, [r7, #(PCB_R8)]
    284  1.80     joerg 	ldrd	r10, r11, [r7, #(PCB_R10)]
    285  1.90      matt 	ldr	r12, [r7, #(PCB_R12)]
    286  1.58      matt #else
    287  1.58      matt 	add	r0, r7, #PCB_R8
    288  1.90      matt 	ldmia	r0, {r8-r12}
    289  1.37       scw #endif
    290  1.29   thorpej 
    291  1.57       scw 	/* Record the old lwp for pmap_activate()'s benefit */
    292  1.83      matt #ifndef ARM_MMU_EXTENDED
    293  1.88      matt 	str	r4, [r5, #CI_LASTLWP]
    294  1.83      matt #endif
    295  1.57       scw 
    296  1.47      yamt 	/* cpu_switchto returns the old lwp */
    297  1.29   thorpej 	mov	r0, r4
    298  1.85       snj 	/* lwp_trampoline expects new lwp as its second argument */
    299  1.47      yamt 	mov	r1, r6
    300   1.1     chris 
    301  1.67      matt #ifdef _ARM_ARCH_7
    302  1.67      matt 	clrex				/* cause any subsequent STREX* to fail */
    303  1.67      matt #endif
    304  1.67      matt 
    305   1.1     chris 	/*
    306  1.51     skrll 	 * Pull the registers that got pushed when cpu_switchto() was called,
    307  1.51     skrll 	 * and return.
    308   1.1     chris 	 */
    309  1.78      matt 	pop	{r4-r7, ip, pc}
    310  1.18   thorpej 
    311  1.78      matt END(cpu_switchto)
    312   1.1     chris 
    313  1.73     skrll ENTRY_NP(lwp_trampoline)
    314  1.52     skrll 	/*
    315  1.52     skrll 	 * cpu_switchto gives us:
    316  1.67      matt 	 *	arg0(r0) = old lwp
    317  1.67      matt 	 *	arg1(r1) = new lwp
    318  1.67      matt 	 * setup by cpu_lwp_fork:
    319  1.67      matt 	 *	r4 = func to call
    320  1.67      matt 	 *	r5 = arg to func
    321  1.67      matt 	 *	r6 = <unused>
    322  1.67      matt 	 *	r7 = spsr mode
    323  1.52     skrll 	 */
    324  1.47      yamt 	bl	_C_LABEL(lwp_startup)
    325  1.38       scw 
    326  1.72      matt 	mov	fp, #0			/* top stack frame */
    327   1.1     chris 	mov	r0, r5
    328   1.1     chris 	mov	r1, sp
    329  1.70      matt #ifdef _ARM_ARCH_5
    330  1.67      matt 	blx	r4
    331  1.67      matt #else
    332  1.24     bjh21 	mov	lr, pc
    333   1.1     chris 	mov	pc, r4
    334  1.67      matt #endif
    335   1.1     chris 
    336  1.67      matt 	GET_CPSR(r0)
    337  1.67      matt 	CPSID_I(r0, r0)			/* Kill irq's */
    338   1.1     chris 
    339  1.71      matt 	GET_CURCPU(r4)			/* for DO_AST */
    340  1.67      matt 	DO_AST_AND_RESTORE_ALIGNMENT_FAULTS
    341   1.1     chris 	PULLFRAME
    342   1.1     chris 
    343   1.1     chris 	movs	pc, lr			/* Exit */
    344  1.78      matt END(lwp_trampoline)
    345  1.58      matt 
    346  1.69     skrll AST_ALIGNMENT_FAULT_LOCALS
    347  1.69     skrll 
    348  1.58      matt #ifdef __HAVE_FAST_SOFTINTS
    349  1.58      matt /*
    350  1.58      matt  *	Called at IPL_HIGH
    351  1.58      matt  *	r0 = new lwp
    352  1.58      matt  *	r1 = ipl for softint_dispatch
    353  1.58      matt  */
    354  1.58      matt ENTRY_NP(softint_switch)
    355  1.78      matt 	push	{r4, r6, r7, lr}
    356  1.58      matt 
    357  1.92     skrll 	ldr	r7, [r0, #L_CPU]	/* get curcpu */
    358  1.67      matt #if defined(TPIDRPRW_IS_CURLWP)
    359  1.92     skrll 	mrc	p15, 0, r4, c13, c0, 4	/* get old lwp */
    360  1.58      matt #else
    361  1.92     skrll 	ldr	r4, [r7, #(CI_CURLWP)]	/* get old lwp */
    362  1.58      matt #endif
    363  1.92     skrll 	mrs	r6, cpsr		/* we need to save this */
    364  1.58      matt 
    365  1.58      matt 	/*
    366  1.58      matt 	 * If the soft lwp blocks, it needs to return to softint_tramp
    367  1.58      matt 	 */
    368  1.92     skrll 	mov	r2, sp			/* think ip */
    369  1.92     skrll 	adr	r3, softint_tramp	/* think lr */
    370  1.78      matt 	push	{r2-r3}
    371  1.78      matt 	push	{r4-r7}
    372  1.58      matt 
    373  1.92     skrll 	mov	r5, r0			/* save new lwp */
    374  1.58      matt 
    375  1.92     skrll 	ldr	r2, [r4, #(L_PCB)]	/* get old lwp's pcb */
    376  1.58      matt 
    377  1.58      matt 	/* Save all the registers into the old lwp's pcb */
    378  1.58      matt #if defined(__XSCALE__) || defined(_ARM_ARCH_6)
    379  1.80     joerg 	strd	r8, r9, [r2, #(PCB_R8)]
    380  1.80     joerg 	strd	r10, r11, [r2, #(PCB_R10)]
    381  1.80     joerg 	strd	r12, r13, [r2, #(PCB_R12)]
    382  1.58      matt #else
    383  1.58      matt 	add	r3, r2, #(PCB_R8)
    384  1.58      matt 	stmia	r3, {r8-r13}
    385  1.58      matt #endif
    386  1.58      matt 
    387  1.86      matt #ifdef _ARM_ARCH_6
    388  1.86      matt 	/*
    389  1.93     skrll 	 * Save user read/write thread/process id register in case it was
    390  1.86      matt 	 * set in userland.
    391  1.86      matt 	 */
    392  1.86      matt 	mrc	p15, 0, r0, c13, c0, 2
    393  1.87      matt 	str	r0, [r2, #(PCB_USER_PID_RW)]
    394  1.86      matt #endif
    395  1.86      matt 
    396  1.58      matt 	/* this is an invariant so load before disabling intrs */
    397  1.60     rmind 	ldr	r2, [r5, #(L_PCB)]	/* get new lwp's pcb */
    398  1.58      matt 
    399  1.58      matt 	IRQdisable
    400  1.58      matt 	/*
    401  1.58      matt 	 * We're switching to a bound LWP so its l_cpu is already correct.
    402  1.58      matt 	 */
    403  1.67      matt #if defined(TPIDRPRW_IS_CURLWP)
    404  1.92     skrll 	mcr	p15, 0, r5, c13, c0, 4	/* save new lwp */
    405  1.58      matt #endif
    406  1.92     skrll 	str	r5, [r7, #(CI_CURLWP)]	/* save new lwp */
    407  1.58      matt 
    408  1.58      matt 	/*
    409  1.58      matt 	 * Normally, we'd get {r8-r13} but since this is a softint lwp
    410  1.85       snj 	 * its existing state doesn't matter.  We start the stack just
    411  1.58      matt 	 * below the trapframe.
    412  1.58      matt 	 */
    413  1.66      matt 	ldr	sp, [r5, #(L_MD_TF)]	/* get new lwp's stack ptr */
    414  1.58      matt 
    415  1.58      matt 	/* At this point we can allow IRQ's again. */
    416  1.58      matt 	IRQenable
    417  1.58      matt 					/* r1 still has ipl */
    418  1.58      matt 	mov	r0, r4			/* r0 has pinned (old) lwp */
    419  1.58      matt 	bl	_C_LABEL(softint_dispatch)
    420  1.58      matt 	/*
    421  1.58      matt 	 * If we've returned, we need to change everything back and return.
    422  1.58      matt 	 */
    423  1.60     rmind 	ldr	r2, [r4, #(L_PCB)]	/* get pinned lwp's pcb */
    424  1.58      matt 
    425  1.58      matt 	/*
    426  1.58      matt 	 * We don't need to restore all the registers since another lwp was
    427  1.58      matt 	 * never executed.  But we do need the SP from the formerly pinned lwp.
    428  1.58      matt 	 */
    429  1.58      matt 
    430  1.90      matt 	IRQdisable
    431  1.67      matt #if defined(TPIDRPRW_IS_CURLWP)
    432  1.92     skrll 	mcr	p15, 0, r4, c13, c0, 4	/* restore pinned lwp */
    433  1.58      matt #endif
    434  1.92     skrll 	str	r4, [r7, #(CI_CURLWP)]	/* restore pinned lwp */
    435  1.75      matt 	ldr	sp, [r2, #(PCB_KSP)]	/* now running on the old stack. */
    436  1.58      matt 
    437  1.58      matt 	/* At this point we can allow IRQ's again. */
    438  1.58      matt 	msr	cpsr_c, r6
    439  1.58      matt 
    440  1.58      matt 	/*
    441  1.58      matt 	 * Grab the registers that got pushed at the start and return.
    442  1.58      matt 	 */
    443  1.78      matt 	pop	{r4-r7, ip, lr}		/* eat switch frame */
    444  1.78      matt 	pop	{r4, r6, r7, pc}	/* pop stack and return */
    445  1.58      matt 
    446  1.58      matt END(softint_switch)
    447  1.58      matt 
    448  1.58      matt /*
    449  1.58      matt  * r0 = previous LWP (the soft lwp)
    450  1.58      matt  * r4 = original LWP (the current lwp)
    451  1.58      matt  * r6 = original CPSR
    452  1.58      matt  * r7 = curcpu()
    453  1.58      matt  */
    454  1.58      matt ENTRY_NP(softint_tramp)
    455  1.94     skrll 	ldr	r3, [r7, #(CI_MTX_COUNT)]	/* readjust after mi_switch */
    456  1.58      matt 	add	r3, r3, #1
    457  1.58      matt 	str	r3, [r7, #(CI_MTX_COUNT)]
    458  1.58      matt 
    459  1.58      matt 	mov	r3, #0				/* tell softint_dispatch */
    460  1.58      matt 	str	r3, [r0, #(L_CTXSWTCH)]		/*    the soft lwp blocked */
    461  1.58      matt 
    462  1.58      matt 	msr	cpsr_c, r6			/* restore interrupts */
    463  1.78      matt 	pop	{r4, r6, r7, pc}		/* pop stack and return */
    464  1.58      matt END(softint_tramp)
    465  1.58      matt #endif /* __HAVE_FAST_SOFTINTS */
    466