arm32_tlb.c revision 1.14 1 1.1 matt /*-
2 1.1 matt * Copyright (c) 2013 The NetBSD Foundation, Inc.
3 1.1 matt * All rights reserved.
4 1.1 matt *
5 1.1 matt * This code is derived from software contributed to The NetBSD Foundation
6 1.1 matt * by Matt Thomas of 3am Software Foundry.
7 1.1 matt *
8 1.1 matt * Redistribution and use in source and binary forms, with or without
9 1.1 matt * modification, are permitted provided that the following conditions
10 1.1 matt * are met:
11 1.1 matt * 1. Redistributions of source code must retain the above copyright
12 1.1 matt * notice, this list of conditions and the following disclaimer.
13 1.1 matt * 2. Redistributions in binary form must reproduce the above copyright
14 1.1 matt * notice, this list of conditions and the following disclaimer in the
15 1.1 matt * documentation and/or other materials provided with the distribution.
16 1.1 matt *
17 1.1 matt * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
18 1.1 matt * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
19 1.1 matt * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
20 1.1 matt * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
21 1.1 matt * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
22 1.1 matt * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
23 1.1 matt * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
24 1.1 matt * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
25 1.1 matt * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
26 1.1 matt * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
27 1.1 matt * POSSIBILITY OF SUCH DAMAGE.
28 1.1 matt */
29 1.5 skrll
30 1.12 skrll #include "opt_cputypes.h"
31 1.5 skrll #include "opt_multiprocessor.h"
32 1.5 skrll
33 1.1 matt #include <sys/cdefs.h>
34 1.14 skrll __KERNEL_RCSID(1, "$NetBSD: arm32_tlb.c,v 1.14 2020/10/30 18:54:36 skrll Exp $");
35 1.1 matt
36 1.1 matt #include <sys/param.h>
37 1.1 matt #include <sys/types.h>
38 1.1 matt
39 1.1 matt #include <uvm/uvm.h>
40 1.1 matt
41 1.1 matt #include <arm/locore.h>
42 1.1 matt
43 1.1 matt bool arm_has_tlbiasid_p; // CPU supports TLBIASID system coprocessor op
44 1.11 jmcneill bool arm_has_mpext_p; // CPU supports MP extensions
45 1.1 matt
46 1.1 matt tlb_asid_t
47 1.1 matt tlb_get_asid(void)
48 1.1 matt {
49 1.1 matt return armreg_contextidr_read() & 0xff;
50 1.1 matt }
51 1.1 matt
52 1.1 matt void
53 1.1 matt tlb_set_asid(tlb_asid_t asid)
54 1.1 matt {
55 1.14 skrll dsb(sy);
56 1.6 skrll if (asid == KERNEL_PID) {
57 1.1 matt armreg_ttbcr_write(armreg_ttbcr_read() | TTBCR_S_PD0);
58 1.14 skrll isb();
59 1.1 matt }
60 1.1 matt armreg_contextidr_write(asid);
61 1.14 skrll isb();
62 1.1 matt }
63 1.1 matt
64 1.1 matt void
65 1.1 matt tlb_invalidate_all(void)
66 1.1 matt {
67 1.2 matt const bool vivt_icache_p = arm_pcache.icache_type == CACHE_TYPE_VIVT;
68 1.14 skrll dsb(sy);
69 1.11 jmcneill if (arm_has_mpext_p) {
70 1.11 jmcneill armreg_tlbiallis_write(0);
71 1.11 jmcneill } else {
72 1.11 jmcneill armreg_tlbiall_write(0);
73 1.11 jmcneill }
74 1.14 skrll isb();
75 1.2 matt if (__predict_false(vivt_icache_p)) {
76 1.2 matt if (arm_has_tlbiasid_p) {
77 1.2 matt armreg_icialluis_write(0);
78 1.2 matt } else {
79 1.2 matt armreg_iciallu_write(0);
80 1.2 matt }
81 1.2 matt }
82 1.14 skrll dsb(sy);
83 1.14 skrll isb();
84 1.1 matt }
85 1.1 matt
86 1.1 matt void
87 1.1 matt tlb_invalidate_globals(void)
88 1.1 matt {
89 1.1 matt tlb_invalidate_all();
90 1.1 matt }
91 1.1 matt
92 1.1 matt void
93 1.1 matt tlb_invalidate_asids(tlb_asid_t lo, tlb_asid_t hi)
94 1.1 matt {
95 1.2 matt const bool vivt_icache_p = arm_pcache.icache_type == CACHE_TYPE_VIVT;
96 1.14 skrll dsb(sy);
97 1.1 matt if (arm_has_tlbiasid_p) {
98 1.1 matt for (; lo <= hi; lo++) {
99 1.11 jmcneill if (arm_has_mpext_p) {
100 1.11 jmcneill armreg_tlbiasidis_write(lo);
101 1.11 jmcneill } else {
102 1.11 jmcneill armreg_tlbiasid_write(lo);
103 1.11 jmcneill }
104 1.1 matt }
105 1.14 skrll dsb(sy);
106 1.14 skrll isb();
107 1.2 matt if (__predict_false(vivt_icache_p)) {
108 1.11 jmcneill if (arm_has_mpext_p) {
109 1.11 jmcneill armreg_icialluis_write(0);
110 1.11 jmcneill } else {
111 1.11 jmcneill armreg_iciallu_write(0);
112 1.11 jmcneill }
113 1.2 matt }
114 1.2 matt } else {
115 1.2 matt armreg_tlbiall_write(0);
116 1.14 skrll isb();
117 1.2 matt if (__predict_false(vivt_icache_p)) {
118 1.2 matt armreg_iciallu_write(0);
119 1.2 matt }
120 1.1 matt }
121 1.14 skrll isb();
122 1.1 matt }
123 1.1 matt
124 1.1 matt void
125 1.1 matt tlb_invalidate_addr(vaddr_t va, tlb_asid_t asid)
126 1.1 matt {
127 1.14 skrll dsb(sy);
128 1.1 matt va = trunc_page(va) | asid;
129 1.1 matt for (vaddr_t eva = va + PAGE_SIZE; va < eva; va += L2_S_SIZE) {
130 1.11 jmcneill if (arm_has_mpext_p) {
131 1.11 jmcneill armreg_tlbimvais_write(va);
132 1.11 jmcneill } else {
133 1.11 jmcneill armreg_tlbimva_write(va);
134 1.11 jmcneill }
135 1.1 matt }
136 1.14 skrll isb();
137 1.1 matt }
138 1.1 matt
139 1.1 matt bool
140 1.1 matt tlb_update_addr(vaddr_t va, tlb_asid_t asid, pt_entry_t pte, bool insert_p)
141 1.1 matt {
142 1.1 matt tlb_invalidate_addr(va, asid);
143 1.1 matt return true;
144 1.1 matt }
145 1.1 matt
146 1.13 jmcneill #if !defined(MULTIPROCESSOR)
147 1.1 matt static u_int
148 1.10 matt tlb_cortex_a5_record_asids(u_long *mapp, tlb_asid_t asid_max)
149 1.1 matt {
150 1.1 matt u_int nasids = 0;
151 1.1 matt for (size_t va_index = 0; va_index < 63; va_index++) {
152 1.1 matt for (size_t way = 0; way < 2; way++) {
153 1.1 matt armreg_tlbdataop_write(
154 1.1 matt __SHIFTIN(way, ARM_TLBDATAOP_WAY)
155 1.1 matt | __SHIFTIN(va_index, ARM_A5_TLBDATAOP_INDEX));
156 1.14 skrll isb();
157 1.1 matt const uint64_t d = ((uint64_t) armreg_tlbdata1_read())
158 1.1 matt | armreg_tlbdata0_read();
159 1.1 matt if (!(d & ARM_TLBDATA_VALID)
160 1.8 jmcneill || !(d & ARM_A5_TLBDATA_nG))
161 1.1 matt continue;
162 1.1 matt
163 1.1 matt const tlb_asid_t asid = __SHIFTOUT(d,
164 1.8 jmcneill ARM_A5_TLBDATA_ASID);
165 1.1 matt const u_long mask = 1L << (asid & 31);
166 1.1 matt const size_t idx = asid >> 5;
167 1.1 matt if (mapp[idx] & mask)
168 1.1 matt continue;
169 1.1 matt
170 1.1 matt mapp[idx] |= mask;
171 1.1 matt nasids++;
172 1.1 matt }
173 1.1 matt }
174 1.1 matt return nasids;
175 1.1 matt }
176 1.1 matt #endif
177 1.1 matt
178 1.13 jmcneill #if !defined(MULTIPROCESSOR)
179 1.1 matt static u_int
180 1.10 matt tlb_cortex_a7_record_asids(u_long *mapp, tlb_asid_t asid_max)
181 1.1 matt {
182 1.1 matt u_int nasids = 0;
183 1.1 matt for (size_t va_index = 0; va_index < 128; va_index++) {
184 1.1 matt for (size_t way = 0; way < 2; way++) {
185 1.1 matt armreg_tlbdataop_write(
186 1.1 matt __SHIFTIN(way, ARM_TLBDATAOP_WAY)
187 1.1 matt | __SHIFTIN(va_index, ARM_A7_TLBDATAOP_INDEX));
188 1.14 skrll isb();
189 1.1 matt const uint32_t d0 = armreg_tlbdata0_read();
190 1.1 matt const uint32_t d1 = armreg_tlbdata1_read();
191 1.1 matt if (!(d0 & ARM_TLBDATA_VALID)
192 1.1 matt || !(d1 & ARM_A7_TLBDATA1_nG))
193 1.1 matt continue;
194 1.1 matt
195 1.1 matt const uint64_t d01 = ((uint64_t) d1)|d0;
196 1.1 matt const tlb_asid_t asid = __SHIFTOUT(d01,
197 1.1 matt ARM_A7_TLBDATA01_ASID);
198 1.1 matt const u_long mask = 1L << (asid & 31);
199 1.1 matt const size_t idx = asid >> 5;
200 1.1 matt if (mapp[idx] & mask)
201 1.1 matt continue;
202 1.1 matt
203 1.1 matt mapp[idx] |= mask;
204 1.1 matt nasids++;
205 1.1 matt }
206 1.1 matt }
207 1.1 matt return nasids;
208 1.1 matt }
209 1.1 matt #endif
210 1.1 matt
211 1.1 matt u_int
212 1.10 matt tlb_record_asids(u_long *mapp, tlb_asid_t asid_max)
213 1.1 matt {
214 1.1 matt #ifndef MULTIPROCESSOR
215 1.1 matt if (CPU_ID_CORTEX_A5_P(curcpu()->ci_arm_cpuid))
216 1.10 matt return tlb_cortex_a5_record_asids(mapp, asid_max);
217 1.1 matt if (CPU_ID_CORTEX_A7_P(curcpu()->ci_arm_cpuid))
218 1.10 matt return tlb_cortex_a7_record_asids(mapp, asid_max);
219 1.1 matt #endif /* MULTIPROCESSOR */
220 1.1 matt #ifdef DIAGNOSTIC
221 1.1 matt mapp[0] = 0xfffffffe;
222 1.1 matt mapp[1] = 0xffffffff;
223 1.1 matt mapp[2] = 0xffffffff;
224 1.1 matt mapp[3] = 0xffffffff;
225 1.1 matt mapp[4] = 0xffffffff;
226 1.1 matt mapp[5] = 0xffffffff;
227 1.1 matt mapp[6] = 0xffffffff;
228 1.1 matt mapp[7] = 0xffffffff;
229 1.1 matt #endif
230 1.1 matt return 255;
231 1.1 matt }
232 1.1 matt
233 1.1 matt void
234 1.1 matt tlb_walk(void *ctx, bool (*func)(void *, vaddr_t, tlb_asid_t, pt_entry_t))
235 1.1 matt {
236 1.1 matt /* no way to view the TLB */
237 1.1 matt }
238