kvm_sparc.c revision 1.27 1 1.27 wiz /* $NetBSD: kvm_sparc.c,v 1.27 2003/05/16 10:24:56 wiz Exp $ */
2 1.8 thorpej
3 1.1 cgd /*-
4 1.1 cgd * Copyright (c) 1992, 1993
5 1.1 cgd * The Regents of the University of California. All rights reserved.
6 1.1 cgd *
7 1.1 cgd * This code is derived from software developed by the Computer Systems
8 1.1 cgd * Engineering group at Lawrence Berkeley Laboratory under DARPA contract
9 1.1 cgd * BG 91-66 and contributed to Berkeley.
10 1.1 cgd *
11 1.1 cgd * Redistribution and use in source and binary forms, with or without
12 1.1 cgd * modification, are permitted provided that the following conditions
13 1.1 cgd * are met:
14 1.1 cgd * 1. Redistributions of source code must retain the above copyright
15 1.1 cgd * notice, this list of conditions and the following disclaimer.
16 1.1 cgd * 2. Redistributions in binary form must reproduce the above copyright
17 1.1 cgd * notice, this list of conditions and the following disclaimer in the
18 1.1 cgd * documentation and/or other materials provided with the distribution.
19 1.1 cgd * 3. All advertising materials mentioning features or use of this software
20 1.1 cgd * must display the following acknowledgement:
21 1.1 cgd * This product includes software developed by the University of
22 1.1 cgd * California, Berkeley and its contributors.
23 1.1 cgd * 4. Neither the name of the University nor the names of its contributors
24 1.1 cgd * may be used to endorse or promote products derived from this software
25 1.1 cgd * without specific prior written permission.
26 1.1 cgd *
27 1.1 cgd * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
28 1.1 cgd * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
29 1.1 cgd * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
30 1.1 cgd * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
31 1.1 cgd * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
32 1.1 cgd * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
33 1.1 cgd * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
34 1.1 cgd * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
35 1.1 cgd * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
36 1.1 cgd * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
37 1.1 cgd * SUCH DAMAGE.
38 1.1 cgd */
39 1.1 cgd
40 1.13 mikel #include <sys/cdefs.h>
41 1.1 cgd #if defined(LIBC_SCCS) && !defined(lint)
42 1.8 thorpej #if 0
43 1.1 cgd static char sccsid[] = "@(#)kvm_sparc.c 8.1 (Berkeley) 6/4/93";
44 1.8 thorpej #else
45 1.27 wiz __RCSID("$NetBSD: kvm_sparc.c,v 1.27 2003/05/16 10:24:56 wiz Exp $");
46 1.8 thorpej #endif
47 1.1 cgd #endif /* LIBC_SCCS and not lint */
48 1.1 cgd
49 1.1 cgd /*
50 1.21 simonb * Sparc machine dependent routines for kvm. Hopefully, the forthcoming
51 1.1 cgd * vm code will one day obsolete this module.
52 1.1 cgd */
53 1.1 cgd
54 1.1 cgd #include <sys/param.h>
55 1.14 pk #include <sys/exec.h>
56 1.1 cgd #include <sys/user.h>
57 1.1 cgd #include <sys/proc.h>
58 1.1 cgd #include <sys/stat.h>
59 1.10 pk #include <sys/core.h>
60 1.10 pk #include <sys/kcore.h>
61 1.1 cgd #include <unistd.h>
62 1.1 cgd #include <nlist.h>
63 1.1 cgd #include <kvm.h>
64 1.1 cgd
65 1.23 mrg #include <uvm/uvm_extern.h>
66 1.22 mrg
67 1.25 matt #include <machine/pmap.h>
68 1.10 pk #include <machine/kcore.h>
69 1.1 cgd
70 1.1 cgd #include <limits.h>
71 1.1 cgd #include <db.h>
72 1.1 cgd
73 1.1 cgd #include "kvm_private.h"
74 1.1 cgd
75 1.1 cgd
76 1.4 deraadt static int cputyp = -1;
77 1.10 pk static int pgshift;
78 1.10 pk static int nptesg; /* [sun4/sun4c] only */
79 1.3 deraadt
80 1.26 mrg #undef VA_VPG
81 1.9 cgd #define VA_VPG(va) ((cputyp == CPU_SUN4C || cputyp == CPU_SUN4M) \
82 1.9 cgd ? VA_SUN4C_VPG(va) \
83 1.9 cgd : VA_SUN4_VPG(va))
84 1.4 deraadt
85 1.26 mrg #undef VA_OFF
86 1.10 pk #define VA_OFF(va) (va & (kd->nbpg - 1))
87 1.10 pk
88 1.15 mrg int _kvm_kvatop44c __P((kvm_t *, u_long, u_long *));
89 1.15 mrg int _kvm_kvatop4m __P((kvm_t *, u_long, u_long *));
90 1.20 mrg int _kvm_kvatop4u __P((kvm_t *, u_long, u_long *));
91 1.20 mrg
92 1.20 mrg /*
93 1.20 mrg * XXX
94 1.21 simonb * taken from /sys/arch/sparc64/include/kcore.h.
95 1.20 mrg * this is the same as the sparc one, except for the kphys addition,
96 1.20 mrg * so luckily we can use this here...
97 1.20 mrg */
98 1.20 mrg typedef struct sparc64_cpu_kcore_hdr {
99 1.20 mrg int cputype; /* CPU type associated with this dump */
100 1.20 mrg u_long kernbase; /* copy of KERNBASE goes here */
101 1.20 mrg int nmemseg; /* # of physical memory segments */
102 1.20 mrg u_long memsegoffset; /* start of memseg array (relative */
103 1.20 mrg /* to the start of this header) */
104 1.20 mrg int nsegmap; /* # of segmaps following */
105 1.20 mrg u_long segmapoffset; /* start of segmap array (relative */
106 1.20 mrg /* to the start of this header) */
107 1.20 mrg int npmeg; /* # of PMEGs; [sun4/sun4c] only */
108 1.20 mrg u_long pmegoffset; /* start of pmeg array (relative */
109 1.20 mrg /* to the start of this header) */
110 1.20 mrg /* SPARC64 stuff */
111 1.20 mrg paddr_t kphys; /* Physical address of 4MB locked TLB */
112 1.20 mrg } sparc64_cpu_kcore_hdr_t;
113 1.9 cgd
114 1.10 pk void
115 1.10 pk _kvm_freevtop(kd)
116 1.10 pk kvm_t *kd;
117 1.9 cgd {
118 1.10 pk if (kd->vmst != 0) {
119 1.10 pk _kvm_err(kd, kd->program, "_kvm_freevtop: internal error");
120 1.10 pk kd->vmst = 0;
121 1.10 pk }
122 1.9 cgd }
123 1.9 cgd
124 1.10 pk /*
125 1.10 pk * Prepare for translation of kernel virtual addresses into offsets
126 1.10 pk * into crash dump files. We use the MMU specific goop written at the
127 1.10 pk * front of the crash dump by pmap_dumpmmu().
128 1.10 pk */
129 1.10 pk int
130 1.10 pk _kvm_initvtop(kd)
131 1.3 deraadt kvm_t *kd;
132 1.3 deraadt {
133 1.20 mrg sparc64_cpu_kcore_hdr_t *cpup = kd->cpu_data;
134 1.9 cgd
135 1.10 pk switch (cputyp = cpup->cputype) {
136 1.10 pk case CPU_SUN4:
137 1.19 eeh case CPU_SUN4U:
138 1.10 pk kd->nbpg = 8196;
139 1.10 pk pgshift = 13;
140 1.10 pk break;
141 1.10 pk case CPU_SUN4C:
142 1.10 pk case CPU_SUN4M:
143 1.10 pk kd->nbpg = 4096;
144 1.10 pk pgshift = 12;
145 1.10 pk break;
146 1.10 pk default:
147 1.10 pk _kvm_err(kd, kd->program, "Unsupported CPU type");
148 1.9 cgd return (-1);
149 1.9 cgd }
150 1.10 pk nptesg = NBPSG / kd->nbpg;
151 1.9 cgd return (0);
152 1.3 deraadt }
153 1.3 deraadt
154 1.7 pk /*
155 1.9 cgd * Translate a kernel virtual address to a physical address using the
156 1.9 cgd * mapping information in kd->vm. Returns the result in pa, and returns
157 1.21 simonb * the number of bytes that are contiguously available from this
158 1.27 wiz * physical address. This routine is used only for crash dumps.
159 1.9 cgd */
160 1.9 cgd int
161 1.9 cgd _kvm_kvatop(kd, va, pa)
162 1.9 cgd kvm_t *kd;
163 1.9 cgd u_long va;
164 1.9 cgd u_long *pa;
165 1.9 cgd {
166 1.10 pk if (cputyp == -1)
167 1.10 pk if (_kvm_initvtop(kd) != 0)
168 1.10 pk return (-1);
169 1.9 cgd
170 1.19 eeh switch (cputyp) {
171 1.19 eeh case CPU_SUN4:
172 1.19 eeh case CPU_SUN4C:
173 1.19 eeh return _kvm_kvatop44c(kd, va, pa);
174 1.19 eeh break;
175 1.19 eeh case CPU_SUN4M:
176 1.19 eeh return _kvm_kvatop4m(kd, va, pa);
177 1.19 eeh break;
178 1.19 eeh case CPU_SUN4U:
179 1.19 eeh default:
180 1.19 eeh return _kvm_kvatop4u(kd, va, pa);
181 1.19 eeh }
182 1.9 cgd }
183 1.9 cgd
184 1.9 cgd /*
185 1.9 cgd * (note: sun4 3-level MMU not yet supported)
186 1.9 cgd */
187 1.9 cgd int
188 1.9 cgd _kvm_kvatop44c(kd, va, pa)
189 1.1 cgd kvm_t *kd;
190 1.1 cgd u_long va;
191 1.1 cgd u_long *pa;
192 1.1 cgd {
193 1.16 perry int vr, vs, pte;
194 1.20 mrg sparc64_cpu_kcore_hdr_t *cpup = kd->cpu_data;
195 1.14 pk struct segmap *sp, *segmaps;
196 1.10 pk int *ptes;
197 1.14 pk int nkreg, nureg;
198 1.14 pk u_long kernbase = cpup->kernbase;
199 1.1 cgd
200 1.14 pk if (va < kernbase)
201 1.7 pk goto err;
202 1.7 pk
203 1.10 pk /*
204 1.10 pk * Layout of CPU segment:
205 1.10 pk * cpu_kcore_hdr_t;
206 1.10 pk * [alignment]
207 1.10 pk * phys_ram_seg_t[cpup->nmemseg];
208 1.14 pk * segmap[cpup->nsegmap];
209 1.10 pk * ptes[cpup->npmegs];
210 1.10 pk */
211 1.14 pk segmaps = (struct segmap *)((long)kd->cpu_data + cpup->segmapoffset);
212 1.10 pk ptes = (int *)((int)kd->cpu_data + cpup->pmegoffset);
213 1.14 pk nkreg = ((int)((-(unsigned)kernbase) / NBPRG));
214 1.14 pk nureg = 256 - nkreg;
215 1.10 pk
216 1.7 pk vr = VA_VREG(va);
217 1.7 pk vs = VA_VSEG(va);
218 1.1 cgd
219 1.14 pk sp = &segmaps[(vr-nureg)*NSEGRG + vs];
220 1.7 pk if (sp->sg_npte == 0)
221 1.7 pk goto err;
222 1.11 pk if (sp->sg_pmeg == cpup->npmeg - 1) /* =seginval */
223 1.7 pk goto err;
224 1.10 pk pte = ptes[sp->sg_pmeg * nptesg + VA_VPG(va)];
225 1.7 pk if ((pte & PG_V) != 0) {
226 1.16 perry long p, off = VA_OFF(va);
227 1.7 pk
228 1.7 pk p = (pte & PG_PFNUM) << pgshift;
229 1.10 pk *pa = p + off;
230 1.7 pk return (kd->nbpg - off);
231 1.1 cgd }
232 1.7 pk err:
233 1.24 sommerfe _kvm_err(kd, 0, "invalid address (%lx)", va);
234 1.1 cgd return (0);
235 1.1 cgd }
236 1.4 deraadt
237 1.9 cgd int
238 1.9 cgd _kvm_kvatop4m(kd, va, pa)
239 1.9 cgd kvm_t *kd;
240 1.9 cgd u_long va;
241 1.9 cgd u_long *pa;
242 1.9 cgd {
243 1.20 mrg sparc64_cpu_kcore_hdr_t *cpup = kd->cpu_data;
244 1.16 perry int vr, vs;
245 1.9 cgd int pte;
246 1.9 cgd off_t foff;
247 1.14 pk struct segmap *sp, *segmaps;
248 1.14 pk int nkreg, nureg;
249 1.14 pk u_long kernbase = cpup->kernbase;
250 1.9 cgd
251 1.14 pk if (va < kernbase)
252 1.9 cgd goto err;
253 1.9 cgd
254 1.10 pk /*
255 1.10 pk * Layout of CPU segment:
256 1.10 pk * cpu_kcore_hdr_t;
257 1.10 pk * [alignment]
258 1.10 pk * phys_ram_seg_t[cpup->nmemseg];
259 1.14 pk * segmap[cpup->nsegmap];
260 1.10 pk */
261 1.14 pk segmaps = (struct segmap *)((long)kd->cpu_data + cpup->segmapoffset);
262 1.14 pk nkreg = ((int)((-(unsigned)kernbase) / NBPRG));
263 1.14 pk nureg = 256 - nkreg;
264 1.10 pk
265 1.9 cgd vr = VA_VREG(va);
266 1.9 cgd vs = VA_VSEG(va);
267 1.9 cgd
268 1.14 pk sp = &segmaps[(vr-nureg)*NSEGRG + vs];
269 1.9 cgd if (sp->sg_npte == 0)
270 1.9 cgd goto err;
271 1.9 cgd
272 1.10 pk /* XXX - assume page tables in initial kernel DATA or BSS. */
273 1.14 pk foff = _kvm_pa2off(kd, (u_long)&sp->sg_pte[VA_VPG(va)] - kernbase);
274 1.10 pk if (foff == (off_t)-1)
275 1.10 pk return (0);
276 1.10 pk
277 1.18 thorpej if (pread(kd->pmfd, &pte, sizeof(pte), foff) != sizeof(pte)) {
278 1.24 sommerfe _kvm_syserr(kd, kd->program, "cannot read pte for %lx", va);
279 1.10 pk return (0);
280 1.9 cgd }
281 1.9 cgd
282 1.9 cgd if ((pte & SRMMU_TETYPE) == SRMMU_TEPTE) {
283 1.16 perry long p, off = VA_OFF(va);
284 1.9 cgd
285 1.9 cgd p = (pte & SRMMU_PPNMASK) << SRMMU_PPNPASHIFT;
286 1.10 pk *pa = p + off;
287 1.9 cgd return (kd->nbpg - off);
288 1.9 cgd }
289 1.9 cgd err:
290 1.24 sommerfe _kvm_err(kd, 0, "invalid address (%lx)", va);
291 1.9 cgd return (0);
292 1.10 pk }
293 1.19 eeh
294 1.19 eeh /*
295 1.20 mrg * sparc64 pmap's 32-bit page table format
296 1.19 eeh */
297 1.19 eeh int
298 1.19 eeh _kvm_kvatop4u(kd, va, pa)
299 1.19 eeh kvm_t *kd;
300 1.19 eeh u_long va;
301 1.19 eeh u_long *pa;
302 1.19 eeh {
303 1.20 mrg sparc64_cpu_kcore_hdr_t *cpup = kd->cpu_data;
304 1.19 eeh int64_t **segmaps;
305 1.19 eeh int64_t *ptes;
306 1.19 eeh int64_t pte;
307 1.20 mrg int64_t kphys = cpup->kphys;
308 1.19 eeh u_long kernbase = cpup->kernbase;
309 1.19 eeh
310 1.19 eeh if (va < kernbase)
311 1.19 eeh goto err;
312 1.19 eeh
313 1.21 simonb /*
314 1.19 eeh * Kernel layout:
315 1.19 eeh *
316 1.19 eeh * kernbase:
317 1.19 eeh * 4MB locked TLB (text+data+BSS)
318 1.21 simonb * Random other stuff.
319 1.19 eeh */
320 1.20 mrg if (va >= kernbase && va < kernbase + 4*1024*1024)
321 1.19 eeh return (va - kernbase) + kphys;
322 1.19 eeh
323 1.20 mrg /* XXX: from sparc64/include/pmap.h */
324 1.20 mrg #define SPARC64_PTSZ (kd->nbpg/8)
325 1.20 mrg #define SPARC64_STSZ (SPARC64_PTSZ)
326 1.20 mrg #define SPARC64_PTMASK (SPARC64_PTSZ-1)
327 1.20 mrg #define SPARC64_PTSHIFT (13)
328 1.20 mrg #define SPARC64_PDSHIFT (10+SPARC64_PTSHIFT)
329 1.20 mrg #define SPARC64_STSHIFT (10+SPARC64_PDSHIFT)
330 1.20 mrg #define SPARC64_STMASK (SPARC64_STSZ-1)
331 1.20 mrg #define sparc64_va_to_seg(v) (int)((((int64_t)(v))>>SPARC64_STSHIFT)&SPARC64_STMASK)
332 1.20 mrg #define sparc64_va_to_pte(v) (int)((((int64_t)(v))>>SPARC64_PTSHIFT)&SPARC64_PTMASK)
333 1.20 mrg
334 1.20 mrg /* XXX: from sparc64/include/pte.h */
335 1.20 mrg #define SPARC64_TLB_V 0x8000000000000000LL
336 1.20 mrg #define SPARC64_TLB_PA_MASK 0x000001ffffffe000LL
337 1.20 mrg
338 1.19 eeh /*
339 1.19 eeh * Layout of CPU segment:
340 1.19 eeh * cpu_kcore_hdr_t;
341 1.19 eeh * [alignment]
342 1.19 eeh * phys_ram_seg_t[cpup->nmemseg];
343 1.19 eeh * segmap[cpup->nsegmap];
344 1.19 eeh */
345 1.20 mrg segmaps = (int64_t **)((long)kd->cpu_data + cpup->segmapoffset);
346 1.20 mrg /* XXX XXX XXX _kvm_pa2off takes u_long and returns off_t..
347 1.20 mrg should take off_t also!! */
348 1.20 mrg
349 1.20 mrg ptes = (int64_t *)(int)_kvm_pa2off(kd, (u_long)segmaps[sparc64_va_to_seg(va)]);
350 1.20 mrg pte = ptes[sparc64_va_to_pte(va)];
351 1.20 mrg if ((pte & SPARC64_TLB_V) != 0)
352 1.20 mrg return ((pte & SPARC64_TLB_PA_MASK) | (va & (kd->nbpg - 1)));
353 1.19 eeh err:
354 1.24 sommerfe _kvm_err(kd, 0, "invalid address (%lx)", va);
355 1.19 eeh return (0);
356 1.19 eeh }
357 1.19 eeh
358 1.10 pk
359 1.21 simonb /*
360 1.27 wiz * Translate a physical address to a file-offset in the crash dump.
361 1.21 simonb */
362 1.10 pk off_t
363 1.10 pk _kvm_pa2off(kd, pa)
364 1.10 pk kvm_t *kd;
365 1.10 pk u_long pa;
366 1.10 pk {
367 1.20 mrg sparc64_cpu_kcore_hdr_t *cpup = kd->cpu_data;
368 1.10 pk phys_ram_seg_t *mp;
369 1.10 pk off_t off;
370 1.10 pk int nmem;
371 1.10 pk
372 1.10 pk /*
373 1.10 pk * Layout of CPU segment:
374 1.10 pk * cpu_kcore_hdr_t;
375 1.10 pk * [alignment]
376 1.10 pk * phys_ram_seg_t[cpup->nmemseg];
377 1.10 pk */
378 1.10 pk mp = (phys_ram_seg_t *)((int)kd->cpu_data + cpup->memsegoffset);
379 1.10 pk off = 0;
380 1.10 pk
381 1.10 pk /* Translate (sparse) pfnum to (packed) dump offset */
382 1.10 pk for (nmem = cpup->nmemseg; --nmem >= 0; mp++) {
383 1.10 pk if (mp->start <= pa && pa < mp->start + mp->size)
384 1.10 pk break;
385 1.10 pk off += mp->size;
386 1.10 pk }
387 1.10 pk if (nmem < 0) {
388 1.24 sommerfe _kvm_err(kd, 0, "invalid address (%lx)", pa);
389 1.10 pk return (-1);
390 1.10 pk }
391 1.10 pk
392 1.10 pk return (kd->dump_off + off + pa - mp->start);
393 1.12 gwr }
394 1.12 gwr
395 1.12 gwr /*
396 1.12 gwr * Machine-dependent initialization for ALL open kvm descriptors,
397 1.12 gwr * not just those for a kernel crash dump. Some architectures
398 1.12 gwr * have to deal with these NOT being constants! (i.e. m68k)
399 1.12 gwr */
400 1.12 gwr int
401 1.12 gwr _kvm_mdopen(kd)
402 1.12 gwr kvm_t *kd;
403 1.12 gwr {
404 1.14 pk u_long max_uva;
405 1.14 pk extern struct ps_strings *__ps_strings;
406 1.12 gwr
407 1.14 pk max_uva = (u_long) (__ps_strings + 1);
408 1.14 pk kd->usrstack = max_uva;
409 1.14 pk kd->max_uva = max_uva;
410 1.14 pk kd->min_uva = 0;
411 1.12 gwr
412 1.12 gwr return (0);
413 1.4 deraadt }
414