Home | History | Annotate | Line # | Download | only in include
locore.h revision 1.18.2.2
      1  1.18.2.2   skrll /*	$NetBSD: locore.h,v 1.18.2.2 2015/06/06 14:39:56 skrll Exp $	*/
      2       1.1    matt 
      3       1.1    matt /*
      4       1.1    matt  * Copyright (c) 1994-1996 Mark Brinicombe.
      5       1.1    matt  * Copyright (c) 1994 Brini.
      6       1.1    matt  * All rights reserved.
      7       1.1    matt  *
      8       1.1    matt  * This code is derived from software written for Brini by Mark Brinicombe
      9       1.1    matt  *
     10       1.1    matt  * Redistribution and use in source and binary forms, with or without
     11       1.1    matt  * modification, are permitted provided that the following conditions
     12       1.1    matt  * are met:
     13       1.1    matt  * 1. Redistributions of source code must retain the above copyright
     14       1.1    matt  *    notice, this list of conditions and the following disclaimer.
     15       1.1    matt  * 2. Redistributions in binary form must reproduce the above copyright
     16       1.1    matt  *    notice, this list of conditions and the following disclaimer in the
     17       1.1    matt  *    documentation and/or other materials provided with the distribution.
     18       1.1    matt  * 3. All advertising materials mentioning features or use of this software
     19       1.1    matt  *    must display the following acknowledgement:
     20       1.1    matt  *	This product includes software developed by Brini.
     21       1.1    matt  * 4. The name of the company nor the name of the author may be used to
     22       1.1    matt  *    endorse or promote products derived from this software without specific
     23       1.1    matt  *    prior written permission.
     24       1.1    matt  *
     25       1.1    matt  * THIS SOFTWARE IS PROVIDED BY BRINI ``AS IS'' AND ANY EXPRESS OR IMPLIED
     26       1.1    matt  * WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF
     27       1.1    matt  * MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
     28       1.1    matt  * IN NO EVENT SHALL BRINI OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT,
     29       1.1    matt  * INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
     30       1.1    matt  * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
     31       1.1    matt  * SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
     32       1.1    matt  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
     33       1.1    matt  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
     34       1.1    matt  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
     35       1.1    matt  * SUCH DAMAGE.
     36       1.1    matt  *
     37       1.1    matt  * RiscBSD kernel project
     38       1.1    matt  *
     39       1.1    matt  * cpu.h
     40       1.1    matt  *
     41       1.1    matt  * CPU specific symbols
     42       1.1    matt  *
     43       1.1    matt  * Created      : 18/09/94
     44       1.1    matt  *
     45       1.1    matt  * Based on kate/katelib/arm6.h
     46       1.1    matt  */
     47       1.1    matt 
     48       1.1    matt #ifndef _ARM_LOCORE_H_
     49       1.1    matt #define _ARM_LOCORE_H_
     50       1.1    matt 
     51       1.2    matt #ifdef _KERNEL_OPT
     52       1.2    matt #include "opt_cpuoptions.h"
     53       1.2    matt #include "opt_cputypes.h"
     54      1.15    matt #include "opt_arm_debug.h"
     55       1.2    matt #endif
     56       1.2    matt 
     57  1.18.2.1   skrll #include <sys/pcu.h>
     58  1.18.2.1   skrll 
     59       1.1    matt #include <arm/cpuconf.h>
     60       1.1    matt #include <arm/armreg.h>
     61       1.1    matt 
     62       1.1    matt #include <machine/frame.h>
     63       1.1    matt 
     64       1.1    matt #ifdef _LOCORE
     65       1.1    matt 
     66       1.1    matt #if defined(_ARM_ARCH_6)
     67       1.1    matt #define IRQdisable	cpsid	i
     68       1.1    matt #define IRQenable	cpsie	i
     69       1.1    matt #elif defined(__PROG32)
     70       1.1    matt #define IRQdisable \
     71       1.1    matt 	stmfd	sp!, {r0} ; \
     72       1.1    matt 	mrs	r0, cpsr ; \
     73       1.1    matt 	orr	r0, r0, #(I32_bit) ; \
     74       1.1    matt 	msr	cpsr_c, r0 ; \
     75       1.1    matt 	ldmfd	sp!, {r0}
     76       1.1    matt 
     77       1.1    matt #define IRQenable \
     78       1.1    matt 	stmfd	sp!, {r0} ; \
     79       1.1    matt 	mrs	r0, cpsr ; \
     80       1.1    matt 	bic	r0, r0, #(I32_bit) ; \
     81       1.1    matt 	msr	cpsr_c, r0 ; \
     82  1.18.2.2   skrll 	ldmfd	sp!, {r0}
     83       1.1    matt #else
     84       1.1    matt /* Not yet used in 26-bit code */
     85       1.1    matt #endif
     86       1.1    matt 
     87       1.1    matt #if defined (TPIDRPRW_IS_CURCPU)
     88       1.1    matt #define GET_CURCPU(rX)		mrc	p15, 0, rX, c13, c0, 4
     89       1.1    matt #define GET_CURLWP(rX)		GET_CURCPU(rX); ldr rX, [rX, #CI_CURLWP]
     90       1.1    matt #elif defined (TPIDRPRW_IS_CURLWP)
     91       1.1    matt #define GET_CURLWP(rX)		mrc	p15, 0, rX, c13, c0, 4
     92  1.18.2.1   skrll #if defined (MULTIPROCESSOR)
     93       1.1    matt #define GET_CURCPU(rX)		GET_CURLWP(rX); ldr rX, [rX, #L_CPU]
     94  1.18.2.1   skrll #elif defined(_ARM_ARCH_7)
     95  1.18.2.1   skrll #define GET_CURCPU(rX)		movw rX, #:lower16:cpu_info_store; movt rX, #:upper16:cpu_info_store
     96  1.18.2.2   skrll #else
     97  1.18.2.1   skrll #define GET_CURCPU(rX)		ldr rX, =_C_LABEL(cpu_info_store)
     98  1.18.2.1   skrll #endif
     99       1.1    matt #elif !defined(MULTIPROCESSOR)
    100       1.1    matt #define GET_CURCPU(rX)		ldr rX, =_C_LABEL(cpu_info_store)
    101       1.1    matt #define GET_CURLWP(rX)		GET_CURCPU(rX); ldr rX, [rX, #CI_CURLWP]
    102       1.1    matt #endif
    103       1.1    matt #define GET_CURPCB(rX)		GET_CURLWP(rX); ldr rX, [rX, #L_PCB]
    104       1.1    matt 
    105       1.1    matt #else /* !_LOCORE */
    106       1.1    matt 
    107       1.3    matt #include <arm/cpufunc.h>
    108       1.3    matt 
    109       1.1    matt #ifdef __PROG32
    110       1.1    matt #define IRQdisable __set_cpsr_c(I32_bit, I32_bit);
    111       1.1    matt #define IRQenable __set_cpsr_c(I32_bit, 0);
    112       1.1    matt #else
    113       1.1    matt #define IRQdisable set_r15(R15_IRQ_DISABLE, R15_IRQ_DISABLE);
    114       1.1    matt #define IRQenable set_r15(R15_IRQ_DISABLE, 0);
    115       1.1    matt #endif
    116       1.1    matt 
    117       1.1    matt /*
    118       1.1    matt  * Validate a PC or PSR for a user process.  Used by various system calls
    119       1.1    matt  * that take a context passed by the user and restore it.
    120       1.1    matt  */
    121       1.1    matt 
    122       1.1    matt #ifdef __PROG32
    123      1.17    matt #ifdef __NO_FIQ
    124       1.1    matt #define VALID_R15_PSR(r15,psr)						\
    125      1.17    matt 	(((psr) & PSR_MODE) == PSR_USR32_MODE && ((psr) & I32_bit) == 0)
    126      1.17    matt #else
    127      1.17    matt #define VALID_R15_PSR(r15,psr)						\
    128      1.17    matt 	(((psr) & PSR_MODE) == PSR_USR32_MODE && ((psr) & IF32_bits) == 0)
    129      1.17    matt #endif
    130       1.1    matt #else
    131       1.1    matt #define VALID_R15_PSR(r15,psr)						\
    132       1.1    matt 	(((r15) & R15_MODE) == R15_MODE_USR &&				\
    133       1.1    matt 		((r15) & (R15_IRQ_DISABLE | R15_FIQ_DISABLE)) == 0)
    134       1.1    matt #endif
    135       1.1    matt 
    136       1.1    matt 
    137       1.1    matt 
    138       1.1    matt /* The address of the vector page. */
    139       1.1    matt extern vaddr_t vector_page;
    140       1.1    matt #ifdef __PROG32
    141       1.1    matt void	arm32_vector_init(vaddr_t, int);
    142       1.1    matt 
    143       1.1    matt #define	ARM_VEC_RESET			(1 << 0)
    144       1.1    matt #define	ARM_VEC_UNDEFINED		(1 << 1)
    145       1.1    matt #define	ARM_VEC_SWI			(1 << 2)
    146       1.1    matt #define	ARM_VEC_PREFETCH_ABORT		(1 << 3)
    147       1.1    matt #define	ARM_VEC_DATA_ABORT		(1 << 4)
    148       1.1    matt #define	ARM_VEC_ADDRESS_EXCEPTION	(1 << 5)
    149       1.1    matt #define	ARM_VEC_IRQ			(1 << 6)
    150       1.1    matt #define	ARM_VEC_FIQ			(1 << 7)
    151       1.1    matt 
    152       1.1    matt #define	ARM_NVEC			8
    153       1.1    matt #define	ARM_VEC_ALL			0xffffffff
    154       1.1    matt #endif /* __PROG32 */
    155       1.1    matt 
    156       1.1    matt #ifndef acorn26
    157       1.1    matt /*
    158       1.1    matt  * cpu device glue (belongs in cpuvar.h)
    159       1.1    matt  */
    160       1.1    matt void	cpu_attach(device_t, cpuid_t);
    161       1.1    matt #endif
    162       1.1    matt 
    163       1.2    matt /* 1 == use cpu_sleep(), 0 == don't */
    164      1.14    matt extern int cpu_do_powersave;
    165       1.7    matt extern int cpu_printfataltraps;
    166       1.2    matt extern int cpu_fpu_present;
    167       1.5    matt extern int cpu_hwdiv_present;
    168      1.14    matt extern int cpu_neon_present;
    169      1.14    matt extern int cpu_simd_present;
    170      1.14    matt extern int cpu_simdex_present;
    171      1.14    matt extern int cpu_umull_present;
    172      1.14    matt extern int cpu_synchprim_present;
    173      1.14    matt 
    174      1.14    matt extern int cpu_instruction_set_attributes[6];
    175      1.14    matt extern int cpu_memory_model_features[4];
    176      1.14    matt extern int cpu_processor_features[2];
    177      1.14    matt extern int cpu_media_and_vfp_features[2];
    178       1.2    matt 
    179      1.16    matt extern bool arm_has_tlbiasid_p;
    180      1.16    matt #ifdef MULTIPROCESSOR
    181      1.13    matt extern u_int arm_cpu_max;
    182      1.16    matt extern volatile u_int arm_cpu_hatched;
    183      1.16    matt #endif
    184      1.13    matt 
    185       1.2    matt #if !defined(CPU_ARMV7)
    186       1.2    matt #define	CPU_IS_ARMV7_P()		false
    187       1.2    matt #elif defined(CPU_ARMV6) || defined(CPU_PRE_ARMV6)
    188       1.2    matt extern bool cpu_armv7_p;
    189       1.2    matt #define	CPU_IS_ARMV7_P()		(cpu_armv7_p)
    190       1.2    matt #else
    191       1.2    matt #define	CPU_IS_ARMV7_P()		true
    192       1.2    matt #endif
    193       1.6    matt #if !defined(CPU_ARMV6)
    194       1.6    matt #define	CPU_IS_ARMV6_P()		false
    195       1.6    matt #elif defined(CPU_ARMV7) || defined(CPU_PRE_ARMV6)
    196       1.6    matt extern bool cpu_armv6_p;
    197       1.6    matt #define	CPU_IS_ARMV6_P()		(cpu_armv6_p)
    198       1.6    matt #else
    199       1.6    matt #define	CPU_IS_ARMV6_P()		true
    200       1.6    matt #endif
    201       1.6    matt 
    202       1.8    matt /*
    203      1.12   joerg  * Used by the fault code to read the current instruction.
    204       1.8    matt  */
    205       1.8    matt static inline uint32_t
    206       1.8    matt read_insn(vaddr_t va, bool user_p)
    207       1.8    matt {
    208       1.8    matt 	uint32_t insn;
    209       1.8    matt 	if (user_p) {
    210       1.8    matt 		__asm __volatile("ldrt %0, [%1]" : "=&r"(insn) : "r"(va));
    211       1.8    matt 	} else {
    212       1.8    matt 		insn = *(const uint32_t *)va;
    213       1.8    matt 	}
    214       1.8    matt #if defined(__ARMEB__) && defined(_ARM_ARCH_7)
    215       1.8    matt 	insn = bswap32(insn);
    216       1.8    matt #endif
    217       1.8    matt 	return insn;
    218       1.8    matt }
    219       1.8    matt 
    220       1.8    matt /*
    221      1.12   joerg  * Used by the fault code to read the current thumb instruction.
    222       1.8    matt  */
    223       1.8    matt static inline uint32_t
    224       1.8    matt read_thumb_insn(vaddr_t va, bool user_p)
    225       1.8    matt {
    226       1.8    matt 	va &= ~1;
    227       1.8    matt 	uint32_t insn;
    228       1.8    matt 	if (user_p) {
    229  1.18.2.2   skrll #if defined(__thumb__) && defined(_ARM_ARCH_T2)
    230  1.18.2.2   skrll 		__asm __volatile("ldrht %0, [%1, #0]" : "=&r"(insn) : "r"(va));
    231  1.18.2.2   skrll #elif defined(_ARM_ARCH_7)
    232      1.10   joerg 		__asm __volatile("ldrht %0, [%1], #0" : "=&r"(insn) : "r"(va));
    233       1.9    matt #else
    234       1.9    matt 		__asm __volatile("ldrt %0, [%1]" : "=&r"(insn) : "r"(va & ~3));
    235       1.9    matt #ifdef __ARMEB__
    236       1.9    matt 		insn = (uint16_t) (insn >> (((va ^ 2) & 2) << 3));
    237       1.9    matt #else
    238       1.9    matt 		insn = (uint16_t) (insn >> ((va & 2) << 3));
    239       1.9    matt #endif
    240       1.9    matt #endif
    241       1.8    matt 	} else {
    242       1.8    matt 		insn = *(const uint16_t *)va;
    243       1.8    matt 	}
    244       1.8    matt #if defined(__ARMEB__) && defined(_ARM_ARCH_7)
    245       1.8    matt 	insn = bswap16(insn);
    246       1.8    matt #endif
    247       1.8    matt 	return insn;
    248       1.8    matt }
    249       1.2    matt 
    250      1.18  martin #ifndef _RUMPKERNEL
    251      1.16    matt static inline void
    252      1.16    matt arm_dmb(void)
    253      1.16    matt {
    254      1.16    matt 	if (CPU_IS_ARMV6_P())
    255      1.16    matt 		armreg_dmb_write(0);
    256      1.16    matt 	else if (CPU_IS_ARMV7_P())
    257  1.18.2.1   skrll 		__asm __volatile("dmb" ::: "memory");
    258      1.16    matt }
    259      1.16    matt 
    260      1.16    matt static inline void
    261      1.16    matt arm_dsb(void)
    262      1.16    matt {
    263      1.16    matt 	if (CPU_IS_ARMV6_P())
    264      1.16    matt 		armreg_dsb_write(0);
    265      1.16    matt 	else if (CPU_IS_ARMV7_P())
    266  1.18.2.1   skrll 		__asm __volatile("dsb" ::: "memory");
    267      1.16    matt }
    268      1.16    matt 
    269      1.16    matt static inline void
    270      1.16    matt arm_isb(void)
    271      1.16    matt {
    272      1.16    matt 	if (CPU_IS_ARMV6_P())
    273      1.16    matt 		armreg_isb_write(0);
    274      1.16    matt 	else if (CPU_IS_ARMV7_P())
    275  1.18.2.1   skrll 		__asm __volatile("isb" ::: "memory");
    276      1.16    matt }
    277      1.18  martin #endif
    278      1.16    matt 
    279       1.1    matt /*
    280       1.1    matt  * Random cruft
    281       1.1    matt  */
    282       1.1    matt 
    283       1.1    matt struct lwp;
    284       1.1    matt 
    285      1.16    matt /* cpu.c */
    286      1.16    matt void	identify_arm_cpu(device_t, struct cpu_info *);
    287       1.1    matt 
    288       1.1    matt /* cpuswitch.S */
    289       1.1    matt struct pcb;
    290       1.1    matt void	savectx(struct pcb *);
    291       1.1    matt 
    292       1.1    matt /* ast.c */
    293       1.1    matt void	userret(struct lwp *);
    294       1.1    matt 
    295       1.1    matt /* *_machdep.c */
    296       1.1    matt void	bootsync(void);
    297       1.1    matt 
    298       1.1    matt /* fault.c */
    299       1.1    matt int	badaddr_read(void *, size_t, void *);
    300       1.1    matt 
    301       1.1    matt /* syscall.c */
    302       1.1    matt void	swi_handler(trapframe_t *);
    303       1.1    matt 
    304       1.1    matt /* arm_machdep.c */
    305       1.1    matt void	ucas_ras_check(trapframe_t *);
    306       1.1    matt 
    307       1.1    matt /* vfp_init.c */
    308      1.16    matt void	vfp_attach(struct cpu_info *);
    309       1.4    matt void	vfp_discardcontext(bool);
    310       1.1    matt void	vfp_savecontext(void);
    311       1.1    matt void	vfp_kernel_acquire(void);
    312       1.1    matt void	vfp_kernel_release(void);
    313       1.4    matt bool	vfp_used_p(void);
    314       1.1    matt extern const pcu_ops_t arm_vfp_ops;
    315       1.1    matt 
    316       1.1    matt #endif	/* !_LOCORE */
    317       1.1    matt 
    318       1.1    matt #endif /* !_ARM_LOCORE_H_ */
    319