armreg.h revision 1.86 1 1.86 matt /* $NetBSD: armreg.h,v 1.86 2014/02/24 16:45:06 matt Exp $ */
2 1.1 bjh21
3 1.1 bjh21 /*
4 1.1 bjh21 * Copyright (c) 1998, 2001 Ben Harris
5 1.1 bjh21 * Copyright (c) 1994-1996 Mark Brinicombe.
6 1.1 bjh21 * Copyright (c) 1994 Brini.
7 1.1 bjh21 * All rights reserved.
8 1.1 bjh21 *
9 1.1 bjh21 * This code is derived from software written for Brini by Mark Brinicombe
10 1.1 bjh21 *
11 1.1 bjh21 * Redistribution and use in source and binary forms, with or without
12 1.1 bjh21 * modification, are permitted provided that the following conditions
13 1.1 bjh21 * are met:
14 1.1 bjh21 * 1. Redistributions of source code must retain the above copyright
15 1.1 bjh21 * notice, this list of conditions and the following disclaimer.
16 1.1 bjh21 * 2. Redistributions in binary form must reproduce the above copyright
17 1.1 bjh21 * notice, this list of conditions and the following disclaimer in the
18 1.1 bjh21 * documentation and/or other materials provided with the distribution.
19 1.1 bjh21 * 3. All advertising materials mentioning features or use of this software
20 1.1 bjh21 * must display the following acknowledgement:
21 1.1 bjh21 * This product includes software developed by Brini.
22 1.1 bjh21 * 4. The name of the company nor the name of the author may be used to
23 1.1 bjh21 * endorse or promote products derived from this software without specific
24 1.1 bjh21 * prior written permission.
25 1.1 bjh21 *
26 1.1 bjh21 * THIS SOFTWARE IS PROVIDED BY BRINI ``AS IS'' AND ANY EXPRESS OR IMPLIED
27 1.1 bjh21 * WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF
28 1.1 bjh21 * MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
29 1.1 bjh21 * IN NO EVENT SHALL BRINI OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT,
30 1.1 bjh21 * INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
31 1.1 bjh21 * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
32 1.1 bjh21 * SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
33 1.1 bjh21 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
34 1.1 bjh21 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
35 1.1 bjh21 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
36 1.1 bjh21 * SUCH DAMAGE.
37 1.1 bjh21 */
38 1.1 bjh21
39 1.1 bjh21 #ifndef _ARM_ARMREG_H
40 1.1 bjh21 #define _ARM_ARMREG_H
41 1.1 bjh21
42 1.1 bjh21 /*
43 1.1 bjh21 * ARM Process Status Register
44 1.1 bjh21 *
45 1.1 bjh21 * The picture in the ARM manuals looks like this:
46 1.1 bjh21 * 3 3 2 2 2 2
47 1.1 bjh21 * 1 0 9 8 7 6 8 7 6 5 4 0
48 1.1 bjh21 * +-+-+-+-+-+-------------------------------------+-+-+-+---------+
49 1.1 bjh21 * |N|Z|C|V|Q| reserved |I|F|T|M M M M M|
50 1.1 bjh21 * | | | | | | | | | |4 3 2 1 0|
51 1.1 bjh21 * +-+-+-+-+-+-------------------------------------+-+-+-+---------+
52 1.1 bjh21 */
53 1.1 bjh21
54 1.1 bjh21 #define PSR_FLAGS 0xf0000000 /* flags */
55 1.1 bjh21 #define PSR_N_bit (1 << 31) /* negative */
56 1.1 bjh21 #define PSR_Z_bit (1 << 30) /* zero */
57 1.1 bjh21 #define PSR_C_bit (1 << 29) /* carry */
58 1.1 bjh21 #define PSR_V_bit (1 << 28) /* overflow */
59 1.1 bjh21
60 1.1 bjh21 #define PSR_Q_bit (1 << 27) /* saturation */
61 1.85 matt #define PSR_IT1_bit (1 << 26)
62 1.85 matt #define PSR_IT0_bit (1 << 25)
63 1.85 matt #define PSR_J_bit (1 << 24) /* Jazelle mode */
64 1.85 matt #define PSR_GE_bits (15 << 16) /* SIMD GE bits */
65 1.85 matt #define PSR_IT7_bit (1 << 15)
66 1.85 matt #define PSR_IT6_bit (1 << 14)
67 1.85 matt #define PSR_IT5_bit (1 << 13)
68 1.85 matt #define PSR_IT4_bit (1 << 12)
69 1.85 matt #define PSR_IT3_bit (1 << 11)
70 1.85 matt #define PSR_IT2_bit (1 << 10)
71 1.85 matt #define PSR_E_BIT (1 << 9) /* Endian state */
72 1.85 matt #define PSR_A_BIT (1 << 8) /* Async abort disable */
73 1.1 bjh21
74 1.1 bjh21 #define I32_bit (1 << 7) /* IRQ disable */
75 1.1 bjh21 #define F32_bit (1 << 6) /* FIQ disable */
76 1.85 matt #define IF32_bits (3 << 6) /* IRQ/FIQ disable */
77 1.1 bjh21
78 1.1 bjh21 #define PSR_T_bit (1 << 5) /* Thumb state */
79 1.1 bjh21
80 1.1 bjh21 #define PSR_MODE 0x0000001f /* mode mask */
81 1.1 bjh21 #define PSR_USR26_MODE 0x00000000
82 1.1 bjh21 #define PSR_FIQ26_MODE 0x00000001
83 1.1 bjh21 #define PSR_IRQ26_MODE 0x00000002
84 1.1 bjh21 #define PSR_SVC26_MODE 0x00000003
85 1.1 bjh21 #define PSR_USR32_MODE 0x00000010
86 1.1 bjh21 #define PSR_FIQ32_MODE 0x00000011
87 1.1 bjh21 #define PSR_IRQ32_MODE 0x00000012
88 1.1 bjh21 #define PSR_SVC32_MODE 0x00000013
89 1.57 matt #define PSR_MON32_MODE 0x00000016
90 1.1 bjh21 #define PSR_ABT32_MODE 0x00000017
91 1.69 matt #define PSR_HYP32_MODE 0x0000001a
92 1.1 bjh21 #define PSR_UND32_MODE 0x0000001b
93 1.1 bjh21 #define PSR_SYS32_MODE 0x0000001f
94 1.1 bjh21 #define PSR_32_MODE 0x00000010
95 1.1 bjh21
96 1.1 bjh21 #define PSR_IN_USR_MODE(psr) (!((psr) & 3)) /* XXX */
97 1.1 bjh21 #define PSR_IN_32_MODE(psr) ((psr) & PSR_32_MODE)
98 1.1 bjh21
99 1.1 bjh21 /* In 26-bit modes, the PSR is stuffed into R15 along with the PC. */
100 1.1 bjh21
101 1.1 bjh21 #define R15_MODE 0x00000003
102 1.1 bjh21 #define R15_MODE_USR 0x00000000
103 1.1 bjh21 #define R15_MODE_FIQ 0x00000001
104 1.1 bjh21 #define R15_MODE_IRQ 0x00000002
105 1.1 bjh21 #define R15_MODE_SVC 0x00000003
106 1.1 bjh21
107 1.1 bjh21 #define R15_PC 0x03fffffc
108 1.1 bjh21
109 1.1 bjh21 #define R15_FIQ_DISABLE 0x04000000
110 1.1 bjh21 #define R15_IRQ_DISABLE 0x08000000
111 1.1 bjh21
112 1.1 bjh21 #define R15_FLAGS 0xf0000000
113 1.1 bjh21 #define R15_FLAG_N 0x80000000
114 1.1 bjh21 #define R15_FLAG_Z 0x40000000
115 1.1 bjh21 #define R15_FLAG_C 0x20000000
116 1.1 bjh21 #define R15_FLAG_V 0x10000000
117 1.1 bjh21
118 1.1 bjh21 /*
119 1.1 bjh21 * Co-processor 15: The system control co-processor.
120 1.1 bjh21 */
121 1.1 bjh21
122 1.1 bjh21 #define ARM_CP15_CPU_ID 0
123 1.1 bjh21
124 1.1 bjh21 /*
125 1.1 bjh21 * The CPU ID register is theoretically structured, but the definitions of
126 1.1 bjh21 * the fields keep changing.
127 1.1 bjh21 */
128 1.1 bjh21
129 1.1 bjh21 /* The high-order byte is always the implementor */
130 1.1 bjh21 #define CPU_ID_IMPLEMENTOR_MASK 0xff000000
131 1.1 bjh21 #define CPU_ID_ARM_LTD 0x41000000 /* 'A' */
132 1.1 bjh21 #define CPU_ID_DEC 0x44000000 /* 'D' */
133 1.1 bjh21 #define CPU_ID_INTEL 0x69000000 /* 'i' */
134 1.26 mycroft #define CPU_ID_TI 0x54000000 /* 'T' */
135 1.45 kiyohara #define CPU_ID_MARVELL 0x56000000 /* 'V' */
136 1.35 nonaka #define CPU_ID_FARADAY 0x66000000 /* 'f' */
137 1.1 bjh21
138 1.3 bjh21 /* How to decide what format the CPUID is in. */
139 1.7 bjh21 #define CPU_ID_ISOLD(x) (((x) & 0x0000f000) == 0x00000000)
140 1.7 bjh21 #define CPU_ID_IS7(x) (((x) & 0x0000f000) == 0x00007000)
141 1.3 bjh21 #define CPU_ID_ISNEW(x) (!CPU_ID_ISOLD(x) && !CPU_ID_IS7(x))
142 1.3 bjh21
143 1.1 bjh21 /* On ARM3 and ARM6, this byte holds the foundry ID. */
144 1.1 bjh21 #define CPU_ID_FOUNDRY_MASK 0x00ff0000
145 1.1 bjh21 #define CPU_ID_FOUNDRY_VLSI 0x00560000
146 1.1 bjh21
147 1.1 bjh21 /* On ARM7 it holds the architecture and variant (sub-model) */
148 1.1 bjh21 #define CPU_ID_7ARCH_MASK 0x00800000
149 1.1 bjh21 #define CPU_ID_7ARCH_V3 0x00000000
150 1.1 bjh21 #define CPU_ID_7ARCH_V4T 0x00800000
151 1.1 bjh21 #define CPU_ID_7VARIANT_MASK 0x007f0000
152 1.1 bjh21
153 1.1 bjh21 /* On more recent ARMs, it does the same, but in a different format */
154 1.1 bjh21 #define CPU_ID_ARCH_MASK 0x000f0000
155 1.1 bjh21 #define CPU_ID_ARCH_V3 0x00000000
156 1.1 bjh21 #define CPU_ID_ARCH_V4 0x00010000
157 1.1 bjh21 #define CPU_ID_ARCH_V4T 0x00020000
158 1.1 bjh21 #define CPU_ID_ARCH_V5 0x00030000
159 1.1 bjh21 #define CPU_ID_ARCH_V5T 0x00040000
160 1.1 bjh21 #define CPU_ID_ARCH_V5TE 0x00050000
161 1.32 rearnsha #define CPU_ID_ARCH_V5TEJ 0x00060000
162 1.32 rearnsha #define CPU_ID_ARCH_V6 0x00070000
163 1.1 bjh21 #define CPU_ID_VARIANT_MASK 0x00f00000
164 1.1 bjh21
165 1.1 bjh21 /* Next three nybbles are part number */
166 1.1 bjh21 #define CPU_ID_PARTNO_MASK 0x0000fff0
167 1.1 bjh21
168 1.23 bsh /* Intel XScale has sub fields in part number */
169 1.23 bsh #define CPU_ID_XSCALE_COREGEN_MASK 0x0000e000 /* core generation */
170 1.23 bsh #define CPU_ID_XSCALE_COREREV_MASK 0x00001c00 /* core revision */
171 1.23 bsh #define CPU_ID_XSCALE_PRODUCT_MASK 0x000003f0 /* product number */
172 1.10 rjs
173 1.1 bjh21 /* And finally, the revision number. */
174 1.1 bjh21 #define CPU_ID_REVISION_MASK 0x0000000f
175 1.2 bjh21
176 1.2 bjh21 /* Individual CPUs are probably best IDed by everything but the revision. */
177 1.2 bjh21 #define CPU_ID_CPU_MASK 0xfffffff0
178 1.1 bjh21
179 1.1 bjh21 /* Fake CPU IDs for ARMs without CP15 */
180 1.1 bjh21 #define CPU_ID_ARM2 0x41560200
181 1.1 bjh21 #define CPU_ID_ARM250 0x41560250
182 1.1 bjh21
183 1.1 bjh21 /* Pre-ARM7 CPUs -- [15:12] == 0 */
184 1.1 bjh21 #define CPU_ID_ARM3 0x41560300
185 1.1 bjh21 #define CPU_ID_ARM600 0x41560600
186 1.1 bjh21 #define CPU_ID_ARM610 0x41560610
187 1.1 bjh21 #define CPU_ID_ARM620 0x41560620
188 1.1 bjh21
189 1.1 bjh21 /* ARM7 CPUs -- [15:12] == 7 */
190 1.4 bjh21 #define CPU_ID_ARM700 0x41007000 /* XXX This is a guess. */
191 1.1 bjh21 #define CPU_ID_ARM710 0x41007100
192 1.36 bjh21 #define CPU_ID_ARM7500 0x41027100
193 1.75 kiyohara #define CPU_ID_ARM710A 0x41067100
194 1.6 bjh21 #define CPU_ID_ARM7500FE 0x41077100
195 1.1 bjh21 #define CPU_ID_ARM710T 0x41807100
196 1.1 bjh21 #define CPU_ID_ARM720T 0x41807200
197 1.1 bjh21 #define CPU_ID_ARM740T8K 0x41807400 /* XXX no MMU, 8KB cache */
198 1.1 bjh21 #define CPU_ID_ARM740T4K 0x41817400 /* XXX no MMU, 4KB cache */
199 1.1 bjh21
200 1.1 bjh21 /* Post-ARM7 CPUs */
201 1.1 bjh21 #define CPU_ID_ARM810 0x41018100
202 1.1 bjh21 #define CPU_ID_ARM920T 0x41129200
203 1.1 bjh21 #define CPU_ID_ARM922T 0x41029220
204 1.37 christos #define CPU_ID_ARM926EJS 0x41069260
205 1.1 bjh21 #define CPU_ID_ARM940T 0x41029400 /* XXX no MMU */
206 1.1 bjh21 #define CPU_ID_ARM946ES 0x41049460 /* XXX no MMU */
207 1.1 bjh21 #define CPU_ID_ARM966ES 0x41049660 /* XXX no MMU */
208 1.1 bjh21 #define CPU_ID_ARM966ESR1 0x41059660 /* XXX no MMU */
209 1.27 rearnsha #define CPU_ID_ARM1020E 0x4115a200 /* (AKA arm10 rev 1) */
210 1.11 bjh21 #define CPU_ID_ARM1022ES 0x4105a220
211 1.31 rearnsha #define CPU_ID_ARM1026EJS 0x4106a260
212 1.47 bsh #define CPU_ID_ARM11MPCORE 0x410fb020
213 1.32 rearnsha #define CPU_ID_ARM1136JS 0x4107b360
214 1.32 rearnsha #define CPU_ID_ARM1136JSR1 0x4117b360
215 1.51 skrll #define CPU_ID_ARM1156T2S 0x4107b560 /* MPU only */
216 1.49 skrll #define CPU_ID_ARM1176JZS 0x410fb760
217 1.58 matt #define CPU_ID_ARM11_P(n) ((n & 0xff07f000) == 0x4107b000)
218 1.52 matt #define CPU_ID_CORTEXA5R0 0x410fc050
219 1.83 matt #define CPU_ID_CORTEXA7R0 0x410fc070
220 1.38 matt #define CPU_ID_CORTEXA8R1 0x411fc080
221 1.38 matt #define CPU_ID_CORTEXA8R2 0x412fc080
222 1.42 jmcneill #define CPU_ID_CORTEXA8R3 0x413fc080
223 1.52 matt #define CPU_ID_CORTEXA9R2 0x411fc090
224 1.52 matt #define CPU_ID_CORTEXA9R3 0x412fc090
225 1.52 matt #define CPU_ID_CORTEXA9R4 0x413fc090
226 1.52 matt #define CPU_ID_CORTEXA15R2 0x412fc0f0
227 1.52 matt #define CPU_ID_CORTEXA15R3 0x413fc0f0
228 1.60 matt #define CPU_ID_CORTEX_P(n) ((n & 0xff0ff000) == 0x410fc000)
229 1.77 matt #define CPU_ID_CORTEX_A5_P(n) ((n & 0xff0ff0f0) == 0x410fc050)
230 1.77 matt #define CPU_ID_CORTEX_A7_P(n) ((n & 0xff0ff0f0) == 0x410fc070)
231 1.60 matt #define CPU_ID_CORTEX_A8_P(n) ((n & 0xff0ff0f0) == 0x410fc080)
232 1.60 matt #define CPU_ID_CORTEX_A9_P(n) ((n & 0xff0ff0f0) == 0x410fc090)
233 1.77 matt #define CPU_ID_CORTEX_A15_P(n) ((n & 0xff0ff0f0) == 0x410fc0f0)
234 1.1 bjh21 #define CPU_ID_SA110 0x4401a100
235 1.1 bjh21 #define CPU_ID_SA1100 0x4401a110
236 1.26 mycroft #define CPU_ID_TI925T 0x54029250
237 1.45 kiyohara #define CPU_ID_MV88FR571_VD 0x56155710
238 1.45 kiyohara #define CPU_ID_MV88SV131 0x56251310
239 1.35 nonaka #define CPU_ID_FA526 0x66015260
240 1.1 bjh21 #define CPU_ID_SA1110 0x6901b110
241 1.18 ichiro #define CPU_ID_IXP1200 0x6901c120
242 1.13 thorpej #define CPU_ID_80200 0x69052000
243 1.23 bsh #define CPU_ID_PXA250 0x69052100 /* sans core revision */
244 1.23 bsh #define CPU_ID_PXA210 0x69052120
245 1.20 ichiro #define CPU_ID_PXA250A 0x69052100 /* 1st version Core */
246 1.20 ichiro #define CPU_ID_PXA210A 0x69052120 /* 1st version Core */
247 1.20 ichiro #define CPU_ID_PXA250B 0x69052900 /* 3rd version Core */
248 1.20 ichiro #define CPU_ID_PXA210B 0x69052920 /* 3rd version Core */
249 1.22 rjs #define CPU_ID_PXA250C 0x69052d00 /* 4th version Core */
250 1.22 rjs #define CPU_ID_PXA210C 0x69052d20 /* 4th version Core */
251 1.29 bsh #define CPU_ID_PXA27X 0x69054110
252 1.19 thorpej #define CPU_ID_80321_400 0x69052420
253 1.19 thorpej #define CPU_ID_80321_600 0x69052430
254 1.21 briggs #define CPU_ID_80321_400_B0 0x69052c20
255 1.21 briggs #define CPU_ID_80321_600_B0 0x69052c30
256 1.33 nonaka #define CPU_ID_80219_400 0x69052e20
257 1.33 nonaka #define CPU_ID_80219_600 0x69052e30
258 1.25 ichiro #define CPU_ID_IXP425_533 0x690541c0
259 1.25 ichiro #define CPU_ID_IXP425_400 0x690541d0
260 1.25 ichiro #define CPU_ID_IXP425_266 0x690541f0
261 1.76 rkujawa #define CPU_ID_MV88SV58XX_P(n) ((n & 0xff0fff00) == 0x560f5800)
262 1.76 rkujawa #define CPU_ID_MV88SV581X_V6 0x560f5810 /* Marvell Sheeva 88SV581x v6 Core */
263 1.76 rkujawa #define CPU_ID_MV88SV581X_V7 0x561f5810 /* Marvell Sheeva 88SV581x v7 Core */
264 1.76 rkujawa #define CPU_ID_MV88SV584X_V6 0x561f5840 /* Marvell Sheeva 88SV584x v6 Core */
265 1.76 rkujawa #define CPU_ID_MV88SV584X_V7 0x562f5840 /* Marvell Sheeva 88SV584x v7 Core */
266 1.76 rkujawa /* Marvell's CPUIDs with ARM ID in implementor field */
267 1.76 rkujawa #define CPU_ID_ARM_88SV581X_V6 0x410fb760 /* Marvell Sheeva 88SV581x v6 Core */
268 1.76 rkujawa #define CPU_ID_ARM_88SV581X_V7 0x413fc080 /* Marvell Sheeva 88SV581x v7 Core */
269 1.76 rkujawa #define CPU_ID_ARM_88SV584X_V6 0x410fb020 /* Marvell Sheeva 88SV584x v6 Core */
270 1.76 rkujawa
271 1.76 rkujawa /* CPUID registers */
272 1.76 rkujawa #define ARM_PFR0_THUMBEE_MASK 0x0000f000
273 1.77 matt #define ARM_PFR1_GTIMER_MASK 0x000f0000
274 1.77 matt #define ARM_PFR1_VIRT_MASK 0x0000f000
275 1.78 matt #define ARM_PFR1_SEC_MASK 0x000000f0
276 1.1 bjh21
277 1.82 matt /* Media and VFP Feature registers */
278 1.82 matt #define ARM_MVFR0_ROUNDING_MASK 0xf0000000
279 1.82 matt #define ARM_MVFR0_SHORTVEC_MASK 0x0f000000
280 1.82 matt #define ARM_MVFR0_SQRT_MASK 0x00f00000
281 1.82 matt #define ARM_MVFR0_DIVIDE_MASK 0x000f0000
282 1.82 matt #define ARM_MVFR0_EXCEPT_MASK 0x0000f000
283 1.82 matt #define ARM_MVFR0_DFLOAT_MASK 0x00000f00
284 1.82 matt #define ARM_MVFR0_SFLOAT_MASK 0x000000f0
285 1.82 matt #define ARM_MVFR0_ASIMD_MASK 0x0000000f
286 1.82 matt #define ARM_MVFR1_ASIMD_FMACS_MASK 0xf0000000
287 1.82 matt #define ARM_MVFR1_VFP_HPFP_MASK 0x0f000000
288 1.82 matt #define ARM_MVFR1_ASIMD_HPFP_MASK 0x00f00000
289 1.82 matt #define ARM_MVFR1_ASIMD_SPFP_MASK 0x000f0000
290 1.82 matt #define ARM_MVFR1_ASIMD_INT_MASK 0x0000f000
291 1.82 matt #define ARM_MVFR1_ASIMD_LDST_MASK 0x00000f00
292 1.82 matt #define ARM_MVFR1_D_NAN_MASK 0x000000f0
293 1.82 matt #define ARM_MVFR1_FTZ_MASK 0x0000000f
294 1.82 matt
295 1.1 bjh21 /* ARM3-specific coprocessor 15 registers */
296 1.1 bjh21 #define ARM3_CP15_FLUSH 1
297 1.1 bjh21 #define ARM3_CP15_CONTROL 2
298 1.1 bjh21 #define ARM3_CP15_CACHEABLE 3
299 1.1 bjh21 #define ARM3_CP15_UPDATEABLE 4
300 1.1 bjh21 #define ARM3_CP15_DISRUPTIVE 5
301 1.1 bjh21
302 1.1 bjh21 /* ARM3 Control register bits */
303 1.1 bjh21 #define ARM3_CTL_CACHE_ON 0x00000001
304 1.1 bjh21 #define ARM3_CTL_SHARED 0x00000002
305 1.1 bjh21 #define ARM3_CTL_MONITOR 0x00000004
306 1.1 bjh21
307 1.1 bjh21 /*
308 1.1 bjh21 * Post-ARM3 CP15 registers:
309 1.14 thorpej *
310 1.14 thorpej * 1 Control register
311 1.14 thorpej *
312 1.14 thorpej * 2 Translation Table Base
313 1.14 thorpej *
314 1.14 thorpej * 3 Domain Access Control
315 1.14 thorpej *
316 1.14 thorpej * 4 Reserved
317 1.14 thorpej *
318 1.14 thorpej * 5 Fault Status
319 1.14 thorpej *
320 1.14 thorpej * 6 Fault Address
321 1.14 thorpej *
322 1.14 thorpej * 7 Cache/write-buffer Control
323 1.14 thorpej *
324 1.14 thorpej * 8 TLB Control
325 1.14 thorpej *
326 1.14 thorpej * 9 Cache Lockdown
327 1.14 thorpej *
328 1.14 thorpej * 10 TLB Lockdown
329 1.14 thorpej *
330 1.14 thorpej * 11 Reserved
331 1.14 thorpej *
332 1.14 thorpej * 12 Reserved
333 1.14 thorpej *
334 1.14 thorpej * 13 Process ID (for FCSE)
335 1.14 thorpej *
336 1.14 thorpej * 14 Reserved
337 1.14 thorpej *
338 1.14 thorpej * 15 Implementation Dependent
339 1.1 bjh21 */
340 1.14 thorpej
341 1.1 bjh21 /* Some of the definitions below need cleaning up for V3/V4 architectures */
342 1.1 bjh21
343 1.1 bjh21 /* CPU control register (CP15 register 1) */
344 1.1 bjh21 #define CPU_CONTROL_MMU_ENABLE 0x00000001 /* M: MMU/Protection unit enable */
345 1.1 bjh21 #define CPU_CONTROL_AFLT_ENABLE 0x00000002 /* A: Alignment fault enable */
346 1.1 bjh21 #define CPU_CONTROL_DC_ENABLE 0x00000004 /* C: IDC/DC enable */
347 1.1 bjh21 #define CPU_CONTROL_WBUF_ENABLE 0x00000008 /* W: Write buffer enable */
348 1.1 bjh21 #define CPU_CONTROL_32BP_ENABLE 0x00000010 /* P: 32-bit exception handlers */
349 1.1 bjh21 #define CPU_CONTROL_32BD_ENABLE 0x00000020 /* D: 32-bit addressing */
350 1.1 bjh21 #define CPU_CONTROL_LABT_ENABLE 0x00000040 /* L: Late abort enable */
351 1.1 bjh21 #define CPU_CONTROL_BEND_ENABLE 0x00000080 /* B: Big-endian mode */
352 1.1 bjh21 #define CPU_CONTROL_SYST_ENABLE 0x00000100 /* S: System protection bit */
353 1.1 bjh21 #define CPU_CONTROL_ROM_ENABLE 0x00000200 /* R: ROM protection bit */
354 1.1 bjh21 #define CPU_CONTROL_CPCLK 0x00000400 /* F: Implementation defined */
355 1.59 matt #define CPU_CONTROL_SWP_ENABLE 0x00000400 /* SW: SWP{B} perform normally. */
356 1.1 bjh21 #define CPU_CONTROL_BPRD_ENABLE 0x00000800 /* Z: Branch prediction enable */
357 1.1 bjh21 #define CPU_CONTROL_IC_ENABLE 0x00001000 /* I: IC enable */
358 1.1 bjh21 #define CPU_CONTROL_VECRELOC 0x00002000 /* V: Vector relocation */
359 1.1 bjh21 #define CPU_CONTROL_ROUNDROBIN 0x00004000 /* RR: Predictable replacement */
360 1.1 bjh21 #define CPU_CONTROL_V4COMPAT 0x00008000 /* L4: ARMv4 compat LDR R15 etc */
361 1.39 matt #define CPU_CONTROL_FI_ENABLE 0x00200000 /* FI: Low interrupt latency */
362 1.46 bsh #define CPU_CONTROL_UNAL_ENABLE 0x00400000 /* U: unaligned data access */
363 1.46 bsh #define CPU_CONTROL_XP_ENABLE 0x00800000 /* XP: extended page table */
364 1.47 bsh #define CPU_CONTROL_V_ENABLE 0x01000000 /* VE: Interrupt vectors enable */
365 1.47 bsh #define CPU_CONTROL_EX_BEND 0x02000000 /* EE: exception endianness */
366 1.47 bsh #define CPU_CONTROL_NMFI 0x08000000 /* NMFI: Non maskable FIQ */
367 1.47 bsh #define CPU_CONTROL_TR_ENABLE 0x10000000 /* TRE: */
368 1.47 bsh #define CPU_CONTROL_AF_ENABLE 0x20000000 /* AFE: Access flag enable */
369 1.47 bsh #define CPU_CONTROL_TE_ENABLE 0x40000000 /* TE: Thumb Exception enable */
370 1.1 bjh21
371 1.1 bjh21 #define CPU_CONTROL_IDC_ENABLE CPU_CONTROL_DC_ENABLE
372 1.16 thorpej
373 1.68 matt /* ARMv6/ARMv7 Co-Processor Access Control Register (CP15, 0, c1, c0, 2) */
374 1.68 matt #define CPACR_V7_ASEDIS 0x80000000 /* Disable Advanced SIMD Ext. */
375 1.68 matt #define CPACR_V7_D32DIS 0x40000000 /* Disable VFP regs 15-31 */
376 1.68 matt #define CPACR_CPn(n) (3 << (2*n))
377 1.68 matt #define CPACR_NOACCESS 0 /* reset value */
378 1.68 matt #define CPACR_PRIVED 1 /* Privileged mode access */
379 1.68 matt #define CPACR_RESERVED 2
380 1.68 matt #define CPACR_ALL 3 /* Privileged and User mode access */
381 1.68 matt
382 1.54 skrll /* ARM11x6 Auxiliary Control Register (CP15 register 1, opcode2 1) */
383 1.54 skrll #define ARM11X6_AUXCTL_RS 0x00000001 /* return stack */
384 1.54 skrll #define ARM11X6_AUXCTL_DB 0x00000002 /* dynamic branch prediction */
385 1.54 skrll #define ARM11X6_AUXCTL_SB 0x00000004 /* static branch prediction */
386 1.54 skrll #define ARM11X6_AUXCTL_TR 0x00000008 /* MicroTLB replacement strat. */
387 1.54 skrll #define ARM11X6_AUXCTL_EX 0x00000010 /* exclusive L1/L2 cache */
388 1.54 skrll #define ARM11X6_AUXCTL_RA 0x00000020 /* clean entire cache disable */
389 1.54 skrll #define ARM11X6_AUXCTL_RV 0x00000040 /* block transfer cache disable */
390 1.54 skrll #define ARM11X6_AUXCTL_CZ 0x00000080 /* restrict cache size */
391 1.54 skrll
392 1.56 skrll /* ARM1136 Auxiliary Control Register (CP15 register 1, opcode2 1) */
393 1.56 skrll #define ARM1136_AUXCTL_PFI 0x80000000 /* PFI: partial FI mode. */
394 1.56 skrll /* This is an undocumented flag
395 1.56 skrll * used to work around a cache bug
396 1.56 skrll * in r0 steppings. See errata
397 1.56 skrll * 364296.
398 1.56 skrll */
399 1.54 skrll /* ARM1176 Auxiliary Control Register (CP15 register 1, opcode2 1) */
400 1.54 skrll #define ARM1176_AUXCTL_PHD 0x10000000 /* inst. prefetch halting disable */
401 1.54 skrll #define ARM1176_AUXCTL_BFD 0x20000000 /* branch folding disable */
402 1.54 skrll #define ARM1176_AUXCTL_FSD 0x40000000 /* force speculative ops disable */
403 1.54 skrll #define ARM1176_AUXCTL_FIO 0x80000000 /* low intr latency override */
404 1.39 matt
405 1.60 matt /* Cortex-A9 Auxiliary Control Register (CP15 register 1, opcode2 1) */
406 1.60 matt #define CORTEXA9_AUXCTL_PARITY 0x00000200 /* Enable parity */
407 1.60 matt #define CORTEXA9_AUXCTL_1WAY 0x00000100 /* Alloc in one way only */
408 1.60 matt #define CORTEXA9_AUXCTL_EXCL 0x00000080 /* Exclusive cache */
409 1.60 matt #define CORTEXA9_AUXCTL_SMP 0x00000040 /* CPU is in SMP mode */
410 1.60 matt #define CORTEXA9_AUXCTL_WRZERO 0x00000008 /* Write full line of zeroes */
411 1.60 matt #define CORTEXA9_AUXCTL_L1PLD 0x00000004 /* L1 Dside prefetch */
412 1.60 matt #define CORTEXA9_AUXCTL_L2PLD 0x00000002 /* L2 Dside prefetch */
413 1.60 matt #define CORTEXA9_AUXCTL_FW 0x00000001 /* Forward Cache/TLB ops */
414 1.60 matt
415 1.55 skrll /* XScale Auxiliary Control Register (CP15 register 1, opcode2 1) */
416 1.16 thorpej #define XSCALE_AUXCTL_K 0x00000001 /* dis. write buffer coalescing */
417 1.16 thorpej #define XSCALE_AUXCTL_P 0x00000002 /* ECC protect page table access */
418 1.16 thorpej #define XSCALE_AUXCTL_MD_WB_RA 0x00000000 /* mini-D$ wb, read-allocate */
419 1.16 thorpej #define XSCALE_AUXCTL_MD_WB_RWA 0x00000010 /* mini-D$ wb, read/write-allocate */
420 1.17 thorpej #define XSCALE_AUXCTL_MD_WT 0x00000020 /* mini-D$ wt, read-allocate */
421 1.17 thorpej #define XSCALE_AUXCTL_MD_MASK 0x00000030
422 1.1 bjh21
423 1.55 skrll /* ARM11 MPCore Auxiliary Control Register (CP15 register 1, opcode2 1) */
424 1.47 bsh #define MPCORE_AUXCTL_RS 0x00000001 /* return stack */
425 1.47 bsh #define MPCORE_AUXCTL_DB 0x00000002 /* dynamic branch prediction */
426 1.47 bsh #define MPCORE_AUXCTL_SB 0x00000004 /* static branch prediction */
427 1.47 bsh #define MPCORE_AUXCTL_F 0x00000008 /* instruction folding enable */
428 1.47 bsh #define MPCORE_AUXCTL_EX 0x00000010 /* exclusive L1/L2 cache */
429 1.47 bsh #define MPCORE_AUXCTL_SA 0x00000020 /* SMP/AMP */
430 1.47 bsh
431 1.76 rkujawa /* Marvell PJ4B Auxillary Control Register */
432 1.76 rkujawa #define PJ4B_AUXCTL_SMPNAMP 0x00000040 /* SMP/AMP */
433 1.76 rkujawa
434 1.60 matt /* Cortex-A9 Auxiliary Control Register (CP15 register 1, opcode 1) */
435 1.60 matt #define CORTEXA9_AUXCTL_FW 0x00000001 /* Cache and TLB updates broadcast */
436 1.60 matt #define CORTEXA9_AUXCTL_L2_PLD 0x00000002 /* Prefetch hint enable */
437 1.60 matt #define CORTEXA9_AUXCTL_L1_PLD 0x00000004 /* Data prefetch hint enable */
438 1.60 matt #define CORTEXA9_AUXCTL_WR_ZERO 0x00000008 /* Ena. write full line of 0s mode */
439 1.60 matt #define CORTEXA9_AUXCTL_SMP 0x00000040 /* Coherency is active */
440 1.60 matt #define CORTEXA9_AUXCTL_EXCL 0x00000080 /* Exclusive cache bit */
441 1.60 matt #define CORTEXA9_AUXCTL_ONEWAY 0x00000100 /* Allocate in on cache way only */
442 1.60 matt #define CORTEXA9_AUXCTL_PARITY 0x00000200 /* Support parity checking */
443 1.60 matt
444 1.45 kiyohara /* Marvell Feroceon Extra Features Register (CP15 register 1, opcode2 0) */
445 1.45 kiyohara #define FC_DCACHE_REPL_LOCK 0x80000000 /* Replace DCache Lock */
446 1.45 kiyohara #define FC_DCACHE_STREAM_EN 0x20000000 /* DCache Streaming Switch */
447 1.45 kiyohara #define FC_WR_ALLOC_EN 0x10000000 /* Enable Write Allocate */
448 1.45 kiyohara #define FC_L2_PREF_DIS 0x01000000 /* L2 Cache Prefetch Disable */
449 1.45 kiyohara #define FC_L2_INV_EVICT_LINE 0x00800000 /* L2 Invalidates Uncorrectable Error Line Eviction */
450 1.45 kiyohara #define FC_L2CACHE_EN 0x00400000 /* L2 enable */
451 1.45 kiyohara #define FC_ICACHE_REPL_LOCK 0x00080000 /* Replace ICache Lock */
452 1.45 kiyohara #define FC_GLOB_HIST_REG_EN 0x00040000 /* Branch Global History Register Enable */
453 1.45 kiyohara #define FC_BRANCH_TARG_BUF_DIS 0x00020000 /* Branch Target Buffer Disable */
454 1.45 kiyohara #define FC_L1_PAR_ERR_EN 0x00010000 /* L1 Parity Error Enable */
455 1.45 kiyohara
456 1.41 matt /* Cache type register definitions 0 */
457 1.41 matt #define CPU_CT_FORMAT(x) (((x) >> 29) & 0x7) /* reg format */
458 1.9 thorpej #define CPU_CT_ISIZE(x) ((x) & 0xfff) /* I$ info */
459 1.9 thorpej #define CPU_CT_DSIZE(x) (((x) >> 12) & 0xfff) /* D$ info */
460 1.9 thorpej #define CPU_CT_S (1U << 24) /* split cache */
461 1.9 thorpej #define CPU_CT_CTYPE(x) (((x) >> 25) & 0xf) /* cache type */
462 1.9 thorpej
463 1.9 thorpej #define CPU_CT_CTYPE_WT 0 /* write-through */
464 1.9 thorpej #define CPU_CT_CTYPE_WB1 1 /* write-back, clean w/ read */
465 1.9 thorpej #define CPU_CT_CTYPE_WB2 2 /* w/b, clean w/ cp15,7 */
466 1.9 thorpej #define CPU_CT_CTYPE_WB6 6 /* w/b, cp15,7, lockdown fmt A */
467 1.9 thorpej #define CPU_CT_CTYPE_WB7 7 /* w/b, cp15,7, lockdown fmt B */
468 1.41 matt #define CPU_CT_CTYPE_WB14 14 /* w/b, cp15,7, lockdown fmt C */
469 1.9 thorpej
470 1.9 thorpej #define CPU_CT_xSIZE_LEN(x) ((x) & 0x3) /* line size */
471 1.9 thorpej #define CPU_CT_xSIZE_M (1U << 2) /* multiplier */
472 1.9 thorpej #define CPU_CT_xSIZE_ASSOC(x) (((x) >> 3) & 0x7) /* associativity */
473 1.9 thorpej #define CPU_CT_xSIZE_SIZE(x) (((x) >> 6) & 0x7) /* size */
474 1.38 matt #define CPU_CT_xSIZE_P (1U << 11) /* need to page-color */
475 1.1 bjh21
476 1.41 matt /* format 4 definitions */
477 1.41 matt #define CPU_CT4_ILINE(x) ((x) & 0xf) /* I$ line size */
478 1.41 matt #define CPU_CT4_DLINE(x) (((x) >> 16) & 0xf) /* D$ line size */
479 1.41 matt #define CPU_CT4_L1IPOLICY(x) (((x) >> 14) & 0x3) /* I$ policy */
480 1.65 matt #define CPU_CT4_L1_AIVIVT 1 /* ASID tagged VIVT */
481 1.41 matt #define CPU_CT4_L1_VIPT 2 /* VIPT */
482 1.65 matt #define CPU_CT4_L1_PIPT 3 /* PIPT */
483 1.65 matt #define CPU_CT4_ERG(x) (((x) >> 20) & 0xf) /* Cache WriteBack Granule */
484 1.65 matt #define CPU_CT4_CWG(x) (((x) >> 24) & 0xf) /* Exclusive Resv. Granule */
485 1.41 matt
486 1.41 matt /* Cache size identifaction register definitions 1, Rd, c0, c0, 0 */
487 1.41 matt #define CPU_CSID_CTYPE_WT 0x80000000 /* write-through avail */
488 1.41 matt #define CPU_CSID_CTYPE_WB 0x40000000 /* write-back avail */
489 1.41 matt #define CPU_CSID_CTYPE_RA 0x20000000 /* read-allocation avail */
490 1.41 matt #define CPU_CSID_CTYPE_WA 0x10000000 /* write-allocation avail */
491 1.44 matt #define CPU_CSID_NUMSETS(x) (((x) >> 13) & 0x7fff)
492 1.41 matt #define CPU_CSID_ASSOC(x) (((x) >> 3) & 0x1ff)
493 1.64 matt #define CPU_CSID_LEN(x) ((x) & 0x07)
494 1.41 matt
495 1.41 matt /* Cache size selection register definitions 2, Rd, c0, c0, 0 */
496 1.41 matt #define CPU_CSSR_L2 0x00000002
497 1.41 matt #define CPU_CSSR_L1 0x00000000
498 1.41 matt #define CPU_CSSR_InD 0x00000001
499 1.41 matt
500 1.71 matt /* ARMv7A CP15 Global Timer definitions */
501 1.71 matt #define CNTKCTL_PL0PTEN 0x00000200 /* PL0 Physical Timer Enable */
502 1.71 matt #define CNTKCTL_PL0VTEN 0x00000100 /* PL0 Virtual Timer Enable */
503 1.71 matt #define CNTKCTL_EVNTI 0x000000f0 /* CNTVCT Event Bit Select */
504 1.71 matt #define CNTKCTL_EVNTDIR 0x00000008 /* CNTVCT Event Dir (1->0) */
505 1.71 matt #define CNTKCTL_EVNTEN 0x00000004 /* CNTVCT Event Enable */
506 1.71 matt #define CNTKCTL_PL0PCTEN 0x00000200 /* PL0 Physical Counter Enable */
507 1.71 matt #define CNTKCTL_PL0VCTEN 0x00000100 /* PL0 Virtual Counter Enable */
508 1.71 matt
509 1.71 matt #define CNT_CTL_ISTATUS 0x00000004 /* Timer is asserted */
510 1.71 matt #define CNT_CTL_IMASK 0x00000002 /* Timer output is masked */
511 1.71 matt #define CNT_CTL_ENABLE 0x00000001 /* Timer is enabled */
512 1.71 matt
513 1.1 bjh21 /* Fault status register definitions */
514 1.1 bjh21
515 1.1 bjh21 #define FAULT_TYPE_MASK 0x0f
516 1.1 bjh21 #define FAULT_USER 0x10
517 1.1 bjh21
518 1.1 bjh21 #define FAULT_WRTBUF_0 0x00 /* Vector Exception */
519 1.1 bjh21 #define FAULT_WRTBUF_1 0x02 /* Terminal Exception */
520 1.1 bjh21 #define FAULT_BUSERR_0 0x04 /* External Abort on Linefetch -- Section */
521 1.1 bjh21 #define FAULT_BUSERR_1 0x06 /* External Abort on Linefetch -- Page */
522 1.1 bjh21 #define FAULT_BUSERR_2 0x08 /* External Abort on Non-linefetch -- Section */
523 1.1 bjh21 #define FAULT_BUSERR_3 0x0a /* External Abort on Non-linefetch -- Page */
524 1.1 bjh21 #define FAULT_BUSTRNL1 0x0c /* External abort on Translation -- Level 1 */
525 1.1 bjh21 #define FAULT_BUSTRNL2 0x0e /* External abort on Translation -- Level 2 */
526 1.1 bjh21 #define FAULT_ALIGN_0 0x01 /* Alignment */
527 1.1 bjh21 #define FAULT_ALIGN_1 0x03 /* Alignment */
528 1.1 bjh21 #define FAULT_TRANS_S 0x05 /* Translation -- Section */
529 1.1 bjh21 #define FAULT_TRANS_P 0x07 /* Translation -- Page */
530 1.1 bjh21 #define FAULT_DOMAIN_S 0x09 /* Domain -- Section */
531 1.1 bjh21 #define FAULT_DOMAIN_P 0x0b /* Domain -- Page */
532 1.1 bjh21 #define FAULT_PERM_S 0x0d /* Permission -- Section */
533 1.1 bjh21 #define FAULT_PERM_P 0x0f /* Permission -- Page */
534 1.28 scw
535 1.86 matt #define FAULT_LPAE 0x0200 /* (SW) used long descriptors */
536 1.86 matt #define FAULT_IMPRECISE 0x0400 /* Imprecise exception (XSCALE) */
537 1.86 matt #define FAULT_WRITE 0x0800 /* fault was due to write (ARMv6+) */
538 1.86 matt #define FAULT_EXT 0x1000 /* fault was due to external abort (ARMv6+) */
539 1.86 matt #define FAULT_CM 0x2000 /* fault was due to cache maintenance (ARMv7+) */
540 1.15 thorpej
541 1.15 thorpej /*
542 1.15 thorpej * Address of the vector page, low and high versions.
543 1.15 thorpej */
544 1.24 thorpej #define ARM_VECTORS_LOW 0x00000000U
545 1.24 thorpej #define ARM_VECTORS_HIGH 0xffff0000U
546 1.1 bjh21
547 1.1 bjh21 /*
548 1.1 bjh21 * ARM Instructions
549 1.1 bjh21 *
550 1.1 bjh21 * 3 3 2 2 2
551 1.1 bjh21 * 1 0 9 8 7 0
552 1.1 bjh21 * +-------+-------------------------------------------------------+
553 1.48 wiz * | cond | instruction dependent |
554 1.1 bjh21 * |c c c c| |
555 1.1 bjh21 * +-------+-------------------------------------------------------+
556 1.1 bjh21 */
557 1.1 bjh21
558 1.1 bjh21 #define INSN_SIZE 4 /* Always 4 bytes */
559 1.1 bjh21 #define INSN_COND_MASK 0xf0000000 /* Condition mask */
560 1.1 bjh21 #define INSN_COND_AL 0xe0000000 /* Always condition */
561 1.1 bjh21
562 1.30 rearnsha #define THUMB_INSN_SIZE 2 /* Some are 4 bytes. */
563 1.30 rearnsha
564 1.38 matt /*
565 1.38 matt * Defines and such for arm11 Performance Monitor Counters (p15, c15, c12, 0)
566 1.38 matt */
567 1.38 matt #define ARM11_PMCCTL_E __BIT(0) /* enable all three counters */
568 1.38 matt #define ARM11_PMCCTL_P __BIT(1) /* reset both Count Registers to zero */
569 1.38 matt #define ARM11_PMCCTL_C __BIT(2) /* reset the Cycle Counter Register to zero */
570 1.38 matt #define ARM11_PMCCTL_D __BIT(3) /* cycle count divide by 64 */
571 1.38 matt #define ARM11_PMCCTL_EC0 __BIT(4) /* Enable Counter Register 0 interrupt */
572 1.38 matt #define ARM11_PMCCTL_EC1 __BIT(5) /* Enable Counter Register 1 interrupt */
573 1.38 matt #define ARM11_PMCCTL_ECC __BIT(6) /* Enable Cycle Counter interrupt */
574 1.38 matt #define ARM11_PMCCTL_SBZa __BIT(7) /* UNP/SBZ */
575 1.38 matt #define ARM11_PMCCTL_CR0 __BIT(8) /* Count Register 0 overflow flag */
576 1.38 matt #define ARM11_PMCCTL_CR1 __BIT(9) /* Count Register 1 overflow flag */
577 1.38 matt #define ARM11_PMCCTL_CCR __BIT(10) /* Cycle Count Register overflow flag */
578 1.38 matt #define ARM11_PMCCTL_X __BIT(11) /* Enable Export of the events to the event bus */
579 1.38 matt #define ARM11_PMCCTL_EVT1 __BITS(19,12) /* source of events for Count Register 1 */
580 1.38 matt #define ARM11_PMCCTL_EVT0 __BITS(27,20) /* source of events for Count Register 0 */
581 1.38 matt #define ARM11_PMCCTL_SBZb __BITS(31,28) /* UNP/SBZ */
582 1.38 matt #define ARM11_PMCCTL_SBZ \
583 1.38 matt (ARM11_PMCCTL_SBZa | ARM11_PMCCTL_SBZb)
584 1.38 matt
585 1.38 matt #define ARM11_PMCEVT_ICACHE_MISS 0 /* Instruction Cache Miss */
586 1.38 matt #define ARM11_PMCEVT_ISTREAM_STALL 1 /* Instruction Stream Stall */
587 1.38 matt #define ARM11_PMCEVT_IUTLB_MISS 2 /* Instruction uTLB Miss */
588 1.38 matt #define ARM11_PMCEVT_DUTLB_MISS 3 /* Data uTLB Miss */
589 1.38 matt #define ARM11_PMCEVT_BRANCH 4 /* Branch Inst. Executed */
590 1.38 matt #define ARM11_PMCEVT_BRANCH_MISS 6 /* Branch mispredicted */
591 1.38 matt #define ARM11_PMCEVT_INST_EXEC 7 /* Instruction Executed */
592 1.38 matt #define ARM11_PMCEVT_DCACHE_ACCESS0 9 /* Data Cache Access */
593 1.38 matt #define ARM11_PMCEVT_DCACHE_ACCESS1 10 /* Data Cache Access */
594 1.38 matt #define ARM11_PMCEVT_DCACHE_MISS 11 /* Data Cache Miss */
595 1.38 matt #define ARM11_PMCEVT_DCACHE_WRITEBACK 12 /* Data Cache Writeback */
596 1.38 matt #define ARM11_PMCEVT_PC_CHANGE 13 /* Software PC change */
597 1.38 matt #define ARM11_PMCEVT_TLB_MISS 15 /* Main TLB Miss */
598 1.38 matt #define ARM11_PMCEVT_DATA_ACCESS 16 /* non-cached data access */
599 1.38 matt #define ARM11_PMCEVT_LSU_STALL 17 /* Load/Store Unit stall */
600 1.38 matt #define ARM11_PMCEVT_WBUF_DRAIN 18 /* Write buffer drained */
601 1.38 matt #define ARM11_PMCEVT_ETMEXTOUT0 32 /* ETMEXTOUT[0] asserted */
602 1.38 matt #define ARM11_PMCEVT_ETMEXTOUT1 33 /* ETMEXTOUT[1] asserted */
603 1.38 matt #define ARM11_PMCEVT_ETMEXTOUT 34 /* ETMEXTOUT[0 & 1] */
604 1.38 matt #define ARM11_PMCEVT_CALL_EXEC 35 /* Procedure call executed */
605 1.38 matt #define ARM11_PMCEVT_RETURN_EXEC 36 /* Return executed */
606 1.38 matt #define ARM11_PMCEVT_RETURN_HIT 37 /* return address predicted */
607 1.38 matt #define ARM11_PMCEVT_RETURN_MISS 38 /* return addr. mispredicted */
608 1.38 matt #define ARM11_PMCEVT_CYCLE 255 /* Increment each cycle */
609 1.38 matt
610 1.43 matt /* Defines for ARM CORTEX performance counters */
611 1.43 matt #define CORTEX_CNTENS_C __BIT(31) /* Enables the cycle counter */
612 1.43 matt #define CORTEX_CNTENC_C __BIT(31) /* Disables the cycle counter */
613 1.43 matt #define CORTEX_CNTOFL_C __BIT(31) /* Cycle counter overflow flag */
614 1.42 jmcneill
615 1.86 matt /* Defines for ARM Cortex A7/A15 L2CTRL */
616 1.86 matt #define L2CTRL_NUMCPU __BITS(25,24) // numcpus - 1
617 1.86 matt #define L2CTRL_ICPRES __BIT(23) // Interrupt Controller is present
618 1.86 matt
619 1.81 matt /* Translate Table Base Control Register */
620 1.81 matt #define TTBCR_S_EAE __BIT(31) // Extended Address Extension
621 1.81 matt #define TTBCR_S_PD1 __BIT(5) // Don't use TTBR1
622 1.81 matt #define TTBCR_S_PD0 __BIT(4) // Don't use TTBR0
623 1.81 matt #define TTBCR_S_N __BITS(2,0) // Width of base address in TTB0
624 1.81 matt
625 1.81 matt #define TTBCR_L_EAE __BIT(31) // Extended Address Extension
626 1.81 matt #define TTBCR_L_SH1 __BITS(29,28) // TTBR1 Shareability
627 1.81 matt #define TTBCR_L_ORGN1 __BITS(27,26) // TTBR1 Outer cacheability
628 1.81 matt #define TTBCR_L_IRGN1 __BITS(25,24) // TTBR1 inner cacheability
629 1.81 matt #define TTBCR_L_EPD1 __BIT(23) // Don't use TTBR1
630 1.81 matt #define TTBCR_L_A1 __BIT(22) // ASID is in TTBR1
631 1.81 matt #define TTBCR_L_T1SZ __BITS(18,16) // TTBR1 size offset
632 1.81 matt #define TTBCR_L_SH0 __BITS(13,12) // TTBR0 Shareability
633 1.81 matt #define TTBCR_L_ORGN0 __BITS(11,10) // TTBR0 Outer cacheability
634 1.81 matt #define TTBCR_L_IRGN0 __BITS(9,8) // TTBR0 inner cacheability
635 1.81 matt #define TTBCR_L_EPD0 __BIT(7) // Don't use TTBR0
636 1.81 matt #define TTBCR_L_T0SZ __BITS(2,0) // TTBR0 size offset
637 1.81 matt
638 1.77 matt /* Defines for ARM Generic Timer */
639 1.77 matt #define ARM_CNTCTL_ENABLE __BIT(0) // Timer Enabled
640 1.77 matt #define ARM_CNTCTL_IMASK __BIT(1) // Mask Interrupt
641 1.77 matt #define ARM_CNTCTL_ISTATUS __BIT(2) // Interrupt is pending
642 1.77 matt
643 1.77 matt #define ARM_CNTKCTL_PL0PTEN __BIT(9)
644 1.77 matt #define ARM_CNTKCTL_PL0VTEN __BIT(8)
645 1.77 matt #define ARM_CNTKCTL_EVNTI __BITS(7,4)
646 1.77 matt #define ARM_CNTKCTL_EVNTDIR __BIT(3)
647 1.77 matt #define ARM_CNTKCTL_EVNTEN __BIT(2)
648 1.77 matt #define ARM_CNTKCTL_PL0PCTEN __BIT(1)
649 1.77 matt #define ARM_CNTKCTL_PL0VCTEN __BIT(0)
650 1.77 matt
651 1.77 matt #define ARM_CNTHCTL_EVNTI __BITS(7,4)
652 1.77 matt #define ARM_CNTHCTL_EVNTDIR __BIT(3)
653 1.77 matt #define ARM_CNTHCTL_EVNTEN __BIT(2)
654 1.77 matt #define ARM_CNTHCTL_PL1PCTEN __BIT(1)
655 1.77 matt #define ARM_CNTHCTL_PL1VCTEN __BIT(0)
656 1.77 matt
657 1.72 matt #if !defined(__ASSEMBLER__) && !defined(_RUMPKERNEL)
658 1.60 matt #define ARMREG_READ_INLINE(name, __insnstring) \
659 1.60 matt static inline uint32_t armreg_##name##_read(void) \
660 1.60 matt { \
661 1.60 matt uint32_t __rv; \
662 1.60 matt __asm __volatile("mrc " __insnstring : "=r"(__rv)); \
663 1.60 matt return __rv; \
664 1.60 matt }
665 1.60 matt
666 1.60 matt #define ARMREG_WRITE_INLINE(name, __insnstring) \
667 1.60 matt static inline void armreg_##name##_write(uint32_t __val) \
668 1.60 matt { \
669 1.60 matt __asm __volatile("mcr " __insnstring :: "r"(__val)); \
670 1.60 matt }
671 1.60 matt
672 1.84 matt #define ARMREG_READ_INLINE2(name, __insnstring) \
673 1.84 matt static inline uint32_t armreg_##name##_read(void) \
674 1.84 matt { \
675 1.84 matt uint32_t __rv; \
676 1.84 matt __asm __volatile(__insnstring : "=r"(__rv)); \
677 1.84 matt return __rv; \
678 1.84 matt }
679 1.84 matt
680 1.84 matt #define ARMREG_WRITE_INLINE2(name, __insnstring) \
681 1.84 matt static inline void armreg_##name##_write(uint32_t __val) \
682 1.84 matt { \
683 1.84 matt __asm __volatile(__insnstring :: "r"(__val)); \
684 1.84 matt }
685 1.84 matt
686 1.71 matt #define ARMREG_READ64_INLINE(name, __insnstring) \
687 1.71 matt static inline uint64_t armreg_##name##_read(void) \
688 1.71 matt { \
689 1.71 matt uint64_t __rv; \
690 1.71 matt __asm __volatile("mrrc " __insnstring : "=r"(__rv)); \
691 1.71 matt return __rv; \
692 1.71 matt }
693 1.71 matt
694 1.71 matt #define ARMREG_WRITE64_INLINE(name, __insnstring) \
695 1.71 matt static inline void armreg_##name##_write(uint64_t __val) \
696 1.71 matt { \
697 1.71 matt __asm __volatile("mcrr " __insnstring :: "r"(__val)); \
698 1.71 matt }
699 1.71 matt
700 1.73 matt /* cp10 registers */
701 1.84 matt ARMREG_READ_INLINE2(fpsid, "vmrs\t%0, fpsid") /* VFP System ID */
702 1.84 matt ARMREG_READ_INLINE2(fpscr, "vmrs\t%0, fpscr") /* VFP Status/Control Register */
703 1.84 matt ARMREG_WRITE_INLINE2(fpscr, "vmsr\tfpscr, %0") /* VFP Status/Control Register */
704 1.84 matt ARMREG_READ_INLINE2(mvfr1, "vmrs\t%0, mvfr1") /* Media and VFP Feature Register 1 */
705 1.84 matt ARMREG_READ_INLINE2(mvfr0, "vmrs\t%0, mvfr0") /* Media and VFP Feature Register 0 */
706 1.84 matt ARMREG_READ_INLINE2(fpexc, "vmrs\t%0, fpexc") /* VFP Exception Register */
707 1.84 matt ARMREG_WRITE_INLINE2(fpexc, "vmsr\tfpexc, %0") /* VFP Exception Register */
708 1.84 matt ARMREG_READ_INLINE2(fpinst, "fmrx\t%0, fpinst") /* VFP Exception Instruction */
709 1.84 matt ARMREG_WRITE_INLINE2(fpinst, "fmxr\tfpinst, %0") /* VFP Exception Instruction */
710 1.84 matt ARMREG_READ_INLINE2(fpinst2, "fmrx\t%0, fpinst2") /* VFP Exception Instruction 2 */
711 1.84 matt ARMREG_WRITE_INLINE2(fpinst2, "fmxr\tfpinst2, %0") /* VFP Exception Instruction 2 */
712 1.73 matt
713 1.73 matt /* cp15 c0 registers */
714 1.60 matt ARMREG_READ_INLINE(midr, "p15,0,%0,c0,c0,0") /* Main ID Register */
715 1.60 matt ARMREG_READ_INLINE(ctr, "p15,0,%0,c0,c0,1") /* Cache Type Register */
716 1.60 matt ARMREG_READ_INLINE(mpidr, "p15,0,%0,c0,c0,5") /* Multiprocess Affinity Register */
717 1.60 matt ARMREG_READ_INLINE(pfr0, "p15,0,%0,c0,c1,0") /* Processor Feature Register 0 */
718 1.60 matt ARMREG_READ_INLINE(pfr1, "p15,0,%0,c0,c1,1") /* Processor Feature Register 1 */
719 1.60 matt ARMREG_READ_INLINE(mmfr0, "p15,0,%0,c0,c1,4") /* Memory Model Feature Register 0 */
720 1.60 matt ARMREG_READ_INLINE(mmfr1, "p15,0,%0,c0,c1,5") /* Memory Model Feature Register 1 */
721 1.60 matt ARMREG_READ_INLINE(mmfr2, "p15,0,%0,c0,c1,6") /* Memory Model Feature Register 2 */
722 1.60 matt ARMREG_READ_INLINE(mmfr3, "p15,0,%0,c0,c1,7") /* Memory Model Feature Register 3 */
723 1.60 matt ARMREG_READ_INLINE(isar0, "p15,0,%0,c0,c2,0") /* Instruction Set Attribute Register 0 */
724 1.60 matt ARMREG_READ_INLINE(isar1, "p15,0,%0,c0,c2,1") /* Instruction Set Attribute Register 1 */
725 1.60 matt ARMREG_READ_INLINE(isar2, "p15,0,%0,c0,c2,2") /* Instruction Set Attribute Register 2 */
726 1.60 matt ARMREG_READ_INLINE(isar3, "p15,0,%0,c0,c2,3") /* Instruction Set Attribute Register 3 */
727 1.60 matt ARMREG_READ_INLINE(isar4, "p15,0,%0,c0,c2,4") /* Instruction Set Attribute Register 4 */
728 1.60 matt ARMREG_READ_INLINE(isar5, "p15,0,%0,c0,c2,5") /* Instruction Set Attribute Register 5 */
729 1.60 matt ARMREG_READ_INLINE(ccsidr, "p15,1,%0,c0,c0,0") /* Cache Size ID Register */
730 1.60 matt ARMREG_READ_INLINE(clidr, "p15,1,%0,c0,c0,1") /* Cache Level ID Register */
731 1.60 matt ARMREG_READ_INLINE(csselr, "p15,2,%0,c0,c0,0") /* Cache Size Selection Register */
732 1.60 matt ARMREG_WRITE_INLINE(csselr, "p15,2,%0,c0,c0,0") /* Cache Size Selection Register */
733 1.73 matt /* cp15 c1 registers */
734 1.70 matt ARMREG_READ_INLINE(sctrl, "p15,0,%0,c1,c0,0") /* System Control Register */
735 1.70 matt ARMREG_WRITE_INLINE(sctrl, "p15,0,%0,c1,c0,0") /* System Control Register */
736 1.68 matt ARMREG_READ_INLINE(auxctl, "p15,0,%0,c1,c0,1") /* Auxiliary Control Register */
737 1.68 matt ARMREG_WRITE_INLINE(auxctl, "p15,0,%0,c1,c0,1") /* Auxiliary Control Register */
738 1.68 matt ARMREG_READ_INLINE(cpacr, "p15,0,%0,c1,c0,2") /* Co-Processor Access Control Register */
739 1.68 matt ARMREG_WRITE_INLINE(cpacr, "p15,0,%0,c1,c0,2") /* Co-Processor Access Control Register */
740 1.73 matt /* cp15 c2 registers */
741 1.63 matt ARMREG_READ_INLINE(ttbr, "p15,0,%0,c2,c0,0") /* Translation Table Base Register 0 */
742 1.63 matt ARMREG_WRITE_INLINE(ttbr, "p15,0,%0,c2,c0,0") /* Translation Table Base Register 0 */
743 1.63 matt ARMREG_READ_INLINE(ttbr1, "p15,0,%0,c2,c0,1") /* Translation Table Base Register 1 */
744 1.63 matt ARMREG_WRITE_INLINE(ttbr1, "p15,0,%0,c2,c0,1") /* Translation Table Base Register 1 */
745 1.63 matt ARMREG_READ_INLINE(ttbcr, "p15,0,%0,c2,c0,2") /* Translation Table Base Register */
746 1.63 matt ARMREG_WRITE_INLINE(ttbcr, "p15,0,%0,c2,c0,2") /* Translation Table Base Register */
747 1.86 matt /* cp15 c3 registers */
748 1.86 matt ARMREG_READ_INLINE(dacr, "p15,0,%0,c3,c0,0") /* Domain Access Control Register */
749 1.86 matt ARMREG_WRITE_INLINE(dacr, "p15,0,%0,c3,c0,0") /* Domain Access Control Register */
750 1.73 matt /* cp15 c5 registers */
751 1.66 matt ARMREG_READ_INLINE(dfsr, "p15,0,%0,c5,c0,0") /* Data Fault Status Register */
752 1.66 matt ARMREG_READ_INLINE(ifsr, "p15,0,%0,c5,c0,1") /* Instruction Fault Status Register */
753 1.73 matt /* cp15 c6 registers */
754 1.66 matt ARMREG_READ_INLINE(dfar, "p15,0,%0,c6,c0,0") /* Data Fault Address Register */
755 1.66 matt ARMREG_READ_INLINE(ifar, "p15,0,%0,c6,c0,2") /* Instruction Fault Address Register */
756 1.73 matt /* cp15 c7 registers */
757 1.65 matt ARMREG_WRITE_INLINE(icialluis, "p15,0,%0,c7,c1,0") /* Instruction Inv All (IS) */
758 1.65 matt ARMREG_WRITE_INLINE(bpiallis, "p15,0,%0,c7,c1,6") /* Branch Invalidate All (IS) */
759 1.65 matt ARMREG_READ_INLINE(par, "p15,0,%0,c7,c4,0") /* Physical Address Register */
760 1.65 matt ARMREG_WRITE_INLINE(iciallu, "p15,0,%0,c7,c5,0") /* Instruction Invalidate All */
761 1.65 matt ARMREG_WRITE_INLINE(icimvau, "p15,0,%0,c7,c5,1") /* Instruction Invalidate MVA */
762 1.65 matt ARMREG_WRITE_INLINE(isb, "p15,0,%0,c7,c5,4") /* Instruction Synchronization Barrier */
763 1.65 matt ARMREG_WRITE_INLINE(bpiall, "p15,0,%0,c5,c1,6") /* Breakpoint Invalidate All */
764 1.65 matt ARMREG_WRITE_INLINE(dcimvac, "p15,0,%0,c7,c6,1") /* Data Invalidate MVA to PoC */
765 1.65 matt ARMREG_WRITE_INLINE(dcisw, "p15,0,%0,c7,c6,2") /* Data Invalidate Set/Way */
766 1.65 matt ARMREG_WRITE_INLINE(ats1cpr, "p15,0,%0,c7,c8,0") /* AddrTrans CurState PL1 Read */
767 1.86 matt ARMREG_WRITE_INLINE(ats1cpw, "p15,0,%0,c7,c8,1") /* AddrTrans CurState PL1 Write */
768 1.86 matt ARMREG_WRITE_INLINE(ats1cur, "p15,0,%0,c7,c8,2") /* AddrTrans CurState PL0 Read */
769 1.86 matt ARMREG_WRITE_INLINE(ats1cuw, "p15,0,%0,c7,c8,3") /* AddrTrans CurState PL0 Write */
770 1.65 matt ARMREG_WRITE_INLINE(dccmvac, "p15,0,%0,c7,c10,1") /* Data Clean MVA to PoC */
771 1.65 matt ARMREG_WRITE_INLINE(dccsw, "p15,0,%0,c7,c10,2") /* Data Clean Set/Way */
772 1.65 matt ARMREG_WRITE_INLINE(dsb, "p15,0,%0,c7,c10,4") /* Data Synchronization Barrier */
773 1.65 matt ARMREG_WRITE_INLINE(dmb, "p15,0,%0,c7,c10,5") /* Data Memory Barrier */
774 1.65 matt ARMREG_WRITE_INLINE(dccmvau, "p15,0,%0,c7,c14,1") /* Data Clean MVA to PoU */
775 1.65 matt ARMREG_WRITE_INLINE(dccimvac, "p15,0,%0,c7,c14,1") /* Data Clean&Inv MVA to PoC */
776 1.65 matt ARMREG_WRITE_INLINE(dccisw, "p15,0,%0,c7,c14,2") /* Data Clean&Inv Set/Way */
777 1.74 matt /* cp15 c8 registers */
778 1.74 matt ARMREG_WRITE_INLINE(tlbiallis, "p15,0,%0,c8,c3,0") /* Invalidate entire unified TLB, inner shareable */
779 1.74 matt ARMREG_WRITE_INLINE(tlbimvais, "p15,0,%0,c8,c3,1") /* Invalidate unified TLB by MVA, inner shareable */
780 1.74 matt ARMREG_WRITE_INLINE(tlbiasidis, "p15,0,%0,c8,c3,2") /* Invalidate unified TLB by ASID, inner shareable */
781 1.74 matt ARMREG_WRITE_INLINE(tlbimvaais, "p15,0,%0,c8,c3,3") /* Invalidate unified TLB by MVA, all ASID, inner shareable */
782 1.74 matt ARMREG_WRITE_INLINE(itlbiall, "p15,0,%0,c8,c5,0") /* Invalidate entire instruction TLB */
783 1.74 matt ARMREG_WRITE_INLINE(itlbimva, "p15,0,%0,c8,c5,1") /* Invalidate instruction TLB by MVA */
784 1.74 matt ARMREG_WRITE_INLINE(itlbiasid, "p15,0,%0,c8,c5,2") /* Invalidate instruction TLB by ASID */
785 1.74 matt ARMREG_WRITE_INLINE(dtlbiall, "p15,0,%0,c8,c6,0") /* Invalidate entire data TLB */
786 1.74 matt ARMREG_WRITE_INLINE(dtlbimva, "p15,0,%0,c8,c6,1") /* Invalidate data TLB by MVA */
787 1.74 matt ARMREG_WRITE_INLINE(dtlbiasid, "p15,0,%0,c8,c6,2") /* Invalidate data TLB by ASID */
788 1.74 matt ARMREG_WRITE_INLINE(tlbiall, "p15,0,%0,c8,c7,0") /* Invalidate entire unified TLB */
789 1.74 matt ARMREG_WRITE_INLINE(tlbimva, "p15,0,%0,c8,c7,1") /* Invalidate unified TLB by MVA */
790 1.74 matt ARMREG_WRITE_INLINE(tlbiasid, "p15,0,%0,c8,c7,2") /* Invalidate unified TLB by ASID */
791 1.74 matt ARMREG_WRITE_INLINE(tlbimvaa, "p15,0,%0,c8,c7,3") /* Invalidate unified TLB by MVA, all ASID */
792 1.73 matt /* cp15 c9 registers */
793 1.60 matt ARMREG_READ_INLINE(pmcr, "p15,0,%0,c9,c12,0") /* PMC Control Register */
794 1.60 matt ARMREG_WRITE_INLINE(pmcr, "p15,0,%0,c9,c12,0") /* PMC Control Register */
795 1.60 matt ARMREG_READ_INLINE(pmcntenset, "p15,0,%0,c9,c12,1") /* PMC Count Enable Set */
796 1.60 matt ARMREG_WRITE_INLINE(pmcntenset, "p15,0,%0,c9,c12,1") /* PMC Count Enable Set */
797 1.60 matt ARMREG_READ_INLINE(pmcntenclr, "p15,0,%0,c9,c12,2") /* PMC Count Enable Clear */
798 1.60 matt ARMREG_WRITE_INLINE(pmcntenclr, "p15,0,%0,c9,c12,2") /* PMC Count Enable Clear */
799 1.60 matt ARMREG_READ_INLINE(pmovsr, "p15,0,%0,c9,c12,3") /* PMC Overflow Flag Status */
800 1.60 matt ARMREG_WRITE_INLINE(pmovsr, "p15,0,%0,c9,c12,3") /* PMC Overflow Flag Status */
801 1.60 matt ARMREG_READ_INLINE(pmccntr, "p15,0,%0,c9,c13,0") /* PMC Cycle Counter */
802 1.60 matt ARMREG_WRITE_INLINE(pmccntr, "p15,0,%0,c9,c13,0") /* PMC Cycle Counter */
803 1.71 matt ARMREG_READ_INLINE(pmuserenr, "p15,0,%0,c9,c14,0") /* PMC User Enable */
804 1.71 matt ARMREG_WRITE_INLINE(pmuserenr, "p15,0,%0,c9,c14,0") /* PMC User Enable */
805 1.86 matt ARMREG_READ_INLINE(l2ctrl, "p15,1,%0,c9,c0,2") /* A7/A15 L2 Control Register */
806 1.73 matt /* cp15 c13 registers */
807 1.71 matt ARMREG_READ_INLINE(contextidr, "p15,0,%0,c13,c0,1") /* Context ID Register */
808 1.71 matt ARMREG_WRITE_INLINE(contextidr, "p15,0,%0,c13,c0,1") /* Context ID Register */
809 1.60 matt ARMREG_READ_INLINE(tpidrprw, "p15,0,%0,c13,c0,4") /* PL1 only Thread ID Register */
810 1.60 matt ARMREG_WRITE_INLINE(tpidrprw, "p15,0,%0,c13,c0,4") /* PL1 only Thread ID Register */
811 1.77 matt /* cp14 c12 registers */
812 1.79 matt ARMREG_READ_INLINE(vbar, "p15,0,%0,c12,c0,0") /* Vector Base Address Register */
813 1.79 matt ARMREG_WRITE_INLINE(vbar, "p15,0,%0,c12,c0,0") /* Vector Base Address Register */
814 1.73 matt /* cp15 c14 registers */
815 1.73 matt /* cp15 Global Timer Registers */
816 1.80 matt ARMREG_READ_INLINE(cnt_frq, "p15,0,%0,c14,c0,0") /* Counter Frequency Register */
817 1.80 matt ARMREG_WRITE_INLINE(cnt_frq, "p15,0,%0,c14,c0,0") /* Counter Frequency Register */
818 1.80 matt ARMREG_READ_INLINE(cntk_ctl, "p15,0,%0,c14,c1,0") /* Timer PL1 Control Register */
819 1.80 matt ARMREG_WRITE_INLINE(cntk_ctl, "p15,0,%0,c14,c1,0") /* Timer PL1 Control Register */
820 1.71 matt ARMREG_READ_INLINE(cntp_tval, "p15,0,%0,c14,c2,0") /* PL1 Physical TimerValue Register */
821 1.71 matt ARMREG_WRITE_INLINE(cntp_tval, "p15,0,%0,c14,c2,0") /* PL1 Physical TimerValue Register */
822 1.71 matt ARMREG_READ_INLINE(cntp_ctl, "p15,0,%0,c14,c2,1") /* PL1 Physical Timer Control Register */
823 1.71 matt ARMREG_WRITE_INLINE(cntp_ctl, "p15,0,%0,c14,c2,1") /* PL1 Physical Timer Control Register */
824 1.71 matt ARMREG_READ_INLINE(cntv_tval, "p15,0,%0,c14,c3,0") /* Virtual TimerValue Register */
825 1.71 matt ARMREG_WRITE_INLINE(cntv_tval, "p15,0,%0,c14,c3,0") /* Virtual TimerValue Register */
826 1.71 matt ARMREG_READ_INLINE(cntv_ctl, "p15,0,%0,c14,c3,1") /* Virtual Timer Control Register */
827 1.71 matt ARMREG_WRITE_INLINE(cntv_ctl, "p15,0,%0,c14,c3,1") /* Virtual Timer Control Register */
828 1.80 matt ARMREG_READ64_INLINE(cntp_ct, "p15,0,%Q0,%R0,c14") /* Physical Count Register */
829 1.80 matt ARMREG_WRITE64_INLINE(cntp_ct, "p15,0,%Q0,%R0,c14") /* Physical Count Register */
830 1.80 matt ARMREG_READ64_INLINE(cntv_ct, "p15,1,%Q0,%R0,c14") /* Virtual Count Register */
831 1.80 matt ARMREG_WRITE64_INLINE(cntv_ct, "p15,1,%Q0,%R0,c14") /* Virtual Count Register */
832 1.71 matt ARMREG_READ64_INLINE(cntp_cval, "p15,2,%Q0,%R0,c14") /* PL1 Physical Timer CompareValue Register */
833 1.71 matt ARMREG_WRITE64_INLINE(cntp_cval, "p15,2,%Q0,%R0,c14") /* PL1 Physical Timer CompareValue Register */
834 1.71 matt ARMREG_READ64_INLINE(cntv_cval, "p15,3,%Q0,%R0,c14") /* PL1 Virtual Timer CompareValue Register */
835 1.71 matt ARMREG_WRITE64_INLINE(cntv_cval, "p15,3,%Q0,%R0,c14") /* PL1 Virtual Timer CompareValue Register */
836 1.73 matt /* cp15 c15 registers */
837 1.77 matt ARMREG_READ_INLINE(cbar, "p15,4,%0,c15,c0,0") /* Configuration Base Address Register */
838 1.60 matt ARMREG_READ_INLINE(pmcrv6, "p15,0,%0,c15,c12,0") /* PMC Control Register (armv6) */
839 1.60 matt ARMREG_WRITE_INLINE(pmcrv6, "p15,0,%0,c15,c12,0") /* PMC Control Register (armv6) */
840 1.60 matt ARMREG_READ_INLINE(pmccntrv6, "p15,0,%0,c15,c12,1") /* PMC Cycle Counter (armv6) */
841 1.60 matt ARMREG_WRITE_INLINE(pmccntrv6, "p15,0,%0,c15,c12,1") /* PMC Cycle Counter (armv6) */
842 1.60 matt
843 1.62 matt #endif /* !__ASSEMBLER__ */
844 1.61 matt
845 1.60 matt
846 1.60 matt #define MPIDR_31 0x80000000
847 1.60 matt #define MPIDR_U 0x40000000 // 1 = Uniprocessor
848 1.60 matt #define MPIDR_MT 0x01000000 // AFF0 for SMT
849 1.60 matt #define MPIDR_AFF2 0x00ff0000
850 1.60 matt #define MPIDR_AFF1 0x0000ff00
851 1.60 matt #define MPIDR_AFF0 0x000000ff
852 1.60 matt
853 1.38 matt #endif /* _ARM_ARMREG_H */
854