cpu.h revision 1.112 1 /* $NetBSD: cpu.h,v 1.112 2014/09/04 18:48:29 palle Exp $ */
2
3 /*
4 * Copyright (c) 1992, 1993
5 * The Regents of the University of California. All rights reserved.
6 *
7 * This software was developed by the Computer Systems Engineering group
8 * at Lawrence Berkeley Laboratory under DARPA contract BG 91-66 and
9 * contributed to Berkeley.
10 *
11 * All advertising materials mentioning features or use of this software
12 * must display the following acknowledgement:
13 * This product includes software developed by the University of
14 * California, Lawrence Berkeley Laboratory.
15 *
16 * Redistribution and use in source and binary forms, with or without
17 * modification, are permitted provided that the following conditions
18 * are met:
19 * 1. Redistributions of source code must retain the above copyright
20 * notice, this list of conditions and the following disclaimer.
21 * 2. Redistributions in binary form must reproduce the above copyright
22 * notice, this list of conditions and the following disclaimer in the
23 * documentation and/or other materials provided with the distribution.
24 * 3. Neither the name of the University nor the names of its contributors
25 * may be used to endorse or promote products derived from this software
26 * without specific prior written permission.
27 *
28 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
29 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
30 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
31 * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
32 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
33 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
34 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
35 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
36 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
37 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
38 * SUCH DAMAGE.
39 *
40 * @(#)cpu.h 8.4 (Berkeley) 1/5/94
41 */
42
43 #ifndef _CPU_H_
44 #define _CPU_H_
45
46 /*
47 * CTL_MACHDEP definitions.
48 */
49 #define CPU_BOOTED_KERNEL 1 /* string: booted kernel name */
50 #define CPU_BOOTED_DEVICE 2 /* string: device booted from */
51 #define CPU_BOOT_ARGS 3 /* string: args booted with */
52 #define CPU_ARCH 4 /* integer: cpu architecture version */
53 #define CPU_VIS 5 /* 0 - no VIS, 1 - VIS 1.0, etc. */
54 #define CPU_MAXID 6 /* number of valid machdep ids */
55
56 #if defined(_KERNEL) || defined(_KMEMUSER)
57 /*
58 * Exported definitions unique to SPARC cpu support.
59 */
60
61 #if defined(_KERNEL_OPT)
62 #include "opt_multiprocessor.h"
63 #include "opt_lockdebug.h"
64 #endif
65
66 #include <machine/psl.h>
67 #include <machine/reg.h>
68 #include <machine/pte.h>
69 #include <machine/intr.h>
70 #if defined(_KERNEL)
71 #include <machine/cpuset.h>
72 #include <sparc64/sparc64/intreg.h>
73 #endif
74 #ifdef SUN4V
75 #include <sparc64/hypervisor.h>
76 #endif
77
78 #include <sys/cpu_data.h>
79 #include <sys/evcnt.h>
80
81 /*
82 * The cpu_info structure is part of a 64KB structure mapped both the kernel
83 * pmap and a single locked TTE a CPUINFO_VA for that particular processor.
84 * Each processor's cpu_info is accessible at CPUINFO_VA only for that
85 * processor. Other processors can access that through an additional mapping
86 * in the kernel pmap.
87 *
88 * The 64KB page contains:
89 *
90 * cpu_info
91 * interrupt stack (all remaining space)
92 * idle PCB
93 * idle stack (STACKSPACE - sizeof(PCB))
94 * 32KB TSB
95 */
96
97 struct cpu_info {
98 struct cpu_data ci_data; /* MI per-cpu data */
99
100
101 /*
102 * SPARC cpu_info structures live at two VAs: one global
103 * VA (so each CPU can access any other CPU's cpu_info)
104 * and an alias VA CPUINFO_VA which is the same on each
105 * CPU and maps to that CPU's cpu_info. Since the alias
106 * CPUINFO_VA is how we locate our cpu_info, we have to
107 * self-reference the global VA so that we can return it
108 * in the curcpu() macro.
109 */
110 struct cpu_info * volatile ci_self;
111
112 /* Most important fields first */
113 struct lwp *ci_curlwp;
114 struct pcb *ci_cpcb;
115 struct cpu_info *ci_next;
116
117 struct lwp *ci_fplwp;
118
119 void *ci_eintstack;
120
121 int ci_mtx_count;
122 int ci_mtx_oldspl;
123
124 /* Spinning up the CPU */
125 void (*ci_spinup)(void);
126 paddr_t ci_paddr;
127
128 int ci_cpuid;
129
130 /* CPU PROM information. */
131 u_int ci_node;
132
133 /* %tick and cpu frequency information */
134 u_long ci_tick_increment;
135 uint64_t ci_cpu_clockrate[2]; /* %tick */
136 uint64_t ci_system_clockrate[2]; /* %stick */
137
138 /* Interrupts */
139 struct intrhand *ci_intrpending[16];
140 struct intrhand *ci_tick_ih;
141
142 /* Event counters */
143 struct evcnt ci_tick_evcnt;
144
145 /* This could be under MULTIPROCESSOR, but there's no good reason */
146 struct evcnt ci_ipi_evcnt[IPI_EVCNT_NUM];
147
148 int ci_flags;
149 int ci_want_ast;
150 int ci_want_resched;
151 int ci_idepth;
152
153 /*
154 * A context is simply a small number that differentiates multiple mappings
155 * of the same address. Contexts on the spitfire are 13 bits, but could
156 * be as large as 17 bits.
157 *
158 * Each context is either free or attached to a pmap.
159 *
160 * The context table is an array of pointers to psegs. Just dereference
161 * the right pointer and you get to the pmap segment tables. These are
162 * physical addresses, of course.
163 *
164 * ci_ctx_lock protects this CPUs context allocation/free.
165 * These are all allocated almost with in the same cacheline.
166 */
167 kmutex_t ci_ctx_lock;
168 int ci_pmap_next_ctx;
169 int ci_numctx;
170 paddr_t *ci_ctxbusy;
171 LIST_HEAD(, pmap) ci_pmap_ctxlist;
172
173 /*
174 * The TSBs are per cpu too (since MMU context differs between
175 * cpus). These are just caches for the TLBs.
176 */
177 pte_t *ci_tsb_dmmu;
178 pte_t *ci_tsb_immu;
179
180 /* TSB description (sun4v). */
181 struct tsb_desc *ci_tsb_desc;
182
183 /* MMU Fault Status Area (sun4v).
184 * Will be initialized to the physical address of the bottom of
185 * the interrupt stack.
186 */
187 paddr_t ci_mmfsa;
188
189 /*
190 * sun4v mondo control fields
191 */
192 paddr_t ci_cpumq; /* cpu mondo queue address */
193 paddr_t ci_devmq; /* device mondo queue address */
194 paddr_t ci_cpuset; /* mondo recipient address */
195 paddr_t ci_mondo; /* mondo message address */
196
197 /* probe fault in PCI config space reads */
198 bool ci_pci_probe;
199 bool ci_pci_fault;
200
201 volatile void *ci_ddb_regs; /* DDB regs */
202 };
203
204 #endif /* _KERNEL || _KMEMUSER */
205
206 #ifdef _KERNEL
207
208 #define CPUF_PRIMARY 1
209
210 /*
211 * CPU boot arguments. Used by secondary CPUs at the bootstrap time.
212 */
213 struct cpu_bootargs {
214 u_int cb_node; /* PROM CPU node */
215 volatile int cb_flags;
216
217 vaddr_t cb_ktext;
218 paddr_t cb_ktextp;
219 vaddr_t cb_ektext;
220
221 vaddr_t cb_kdata;
222 paddr_t cb_kdatap;
223 vaddr_t cb_ekdata;
224
225 paddr_t cb_cpuinfo;
226 };
227
228 extern struct cpu_bootargs *cpu_args;
229
230 #if defined(MULTIPROCESSOR)
231 extern int sparc_ncpus;
232 #else
233 #define sparc_ncpus 1
234 #endif
235
236 extern struct cpu_info *cpus;
237 extern struct pool_cache *fpstate_cache;
238
239 #define curcpu() (((struct cpu_info *)CPUINFO_VA)->ci_self)
240 #define cpu_number() (curcpu()->ci_index)
241 #define CPU_IS_PRIMARY(ci) ((ci)->ci_flags & CPUF_PRIMARY)
242
243 #define CPU_INFO_ITERATOR int __unused
244 #define CPU_INFO_FOREACH(cii, ci) ci = cpus; ci != NULL; ci = ci->ci_next
245
246 #define curlwp curcpu()->ci_curlwp
247 #define fplwp curcpu()->ci_fplwp
248 #define curpcb curcpu()->ci_cpcb
249
250 #define want_ast curcpu()->ci_want_ast
251 #define want_resched curcpu()->ci_want_resched
252
253 /*
254 * definitions of cpu-dependent requirements
255 * referenced in generic code
256 */
257 #define cpu_wait(p) /* nothing */
258 void cpu_proc_fork(struct proc *, struct proc *);
259
260 /* run on the cpu itself */
261 void cpu_pmap_init(struct cpu_info *);
262 /* run upfront to prepare the cpu_info */
263 void cpu_pmap_prepare(struct cpu_info *, bool);
264
265 #if defined(MULTIPROCESSOR)
266 extern vaddr_t cpu_spinup_trampoline;
267
268 extern char *mp_tramp_code;
269 extern u_long mp_tramp_code_len;
270 extern u_long mp_tramp_tlb_slots;
271 extern u_long mp_tramp_func;
272 extern u_long mp_tramp_ci;
273
274 void cpu_hatch(void);
275 void cpu_boot_secondary_processors(void);
276
277 /*
278 * Call a function on other cpus:
279 * multicast - send to everyone in the sparc64_cpuset_t
280 * broadcast - send to to all cpus but ourselves
281 * send - send to just this cpu
282 * The called function do not follow the C ABI, so need to be coded in
283 * assembler.
284 */
285 typedef void (* ipifunc_t)(void *, void *);
286
287 void sparc64_multicast_ipi(sparc64_cpuset_t, ipifunc_t, uint64_t, uint64_t);
288 void sparc64_broadcast_ipi(ipifunc_t, uint64_t, uint64_t);
289 void sparc64_send_ipi(int, ipifunc_t, uint64_t, uint64_t);
290
291 /*
292 * Call an arbitrary C function on another cpu (or all others but ourself)
293 */
294 typedef void (*ipi_c_call_func_t)(void*);
295 void sparc64_generic_xcall(struct cpu_info*, ipi_c_call_func_t, void*);
296
297 #endif
298
299 /* Provide %pc of a lwp */
300 #define LWP_PC(l) ((l)->l_md.md_tf->tf_pc)
301
302 /*
303 * Arguments to hardclock, softclock and gatherstats encapsulate the
304 * previous machine state in an opaque clockframe. The ipl is here
305 * as well for strayintr (see locore.s:interrupt and intr.c:strayintr).
306 * Note that CLKF_INTR is valid only if CLKF_USERMODE is false.
307 */
308 struct clockframe {
309 struct trapframe64 t;
310 };
311
312 #define CLKF_USERMODE(framep) (((framep)->t.tf_tstate & TSTATE_PRIV) == 0)
313 #define CLKF_PC(framep) ((framep)->t.tf_pc)
314 /* Since some files in sys/kern do not know BIAS, I'm using 0x7ff here */
315 #define CLKF_INTR(framep) \
316 ((!CLKF_USERMODE(framep))&& \
317 (((framep)->t.tf_out[6] & 1 ) ? \
318 (((vaddr_t)(framep)->t.tf_out[6] < \
319 (vaddr_t)EINTSTACK-0x7ff) && \
320 ((vaddr_t)(framep)->t.tf_out[6] > \
321 (vaddr_t)INTSTACK-0x7ff)) : \
322 (((vaddr_t)(framep)->t.tf_out[6] < \
323 (vaddr_t)EINTSTACK) && \
324 ((vaddr_t)(framep)->t.tf_out[6] > \
325 (vaddr_t)INTSTACK))))
326
327 /*
328 * Give a profiling tick to the current process when the user profiling
329 * buffer pages are invalid. On the sparc, request an ast to send us
330 * through trap(), marking the proc as needing a profiling tick.
331 */
332 #define cpu_need_proftick(l) ((l)->l_pflag |= LP_OWEUPC, want_ast = 1)
333
334 /*
335 * Notify an LWP that it has a signal pending, process as soon as possible.
336 */
337 void cpu_signotify(struct lwp *);
338
339 /*
340 * Interrupt handler chains. Interrupt handlers should return 0 for
341 * ``not me'' or 1 (``I took care of it''). intr_establish() inserts a
342 * handler into the list. The handler is called with its (single)
343 * argument, or with a pointer to a clockframe if ih_arg is NULL.
344 */
345 struct intrhand {
346 int (*ih_fun)(void *);
347 void *ih_arg;
348 /* if we have to take the biglock, we interpose a wrapper
349 * and need to save the original function and arg */
350 int (*ih_realfun)(void *);
351 void *ih_realarg;
352 short ih_number; /* interrupt number */
353 /* the H/W provides */
354 char ih_pil; /* interrupt priority */
355 struct intrhand *ih_next; /* global list */
356 struct intrhand *ih_pending; /* interrupt queued */
357 volatile uint64_t *ih_map; /* Interrupt map reg */
358 volatile uint64_t *ih_clr; /* clear interrupt reg */
359 struct evcnt ih_cnt; /* counter for vmstat */
360 uint32_t ih_ivec;
361 char ih_name[32]; /* name for the above */
362 };
363 extern struct intrhand *intrhand[];
364 extern struct intrhand *intrlev[MAXINTNUM];
365
366 void intr_establish(int level, bool mpsafe, struct intrhand *);
367 void *sparc_softintr_establish(int, int (*)(void *), void *);
368 void sparc_softintr_schedule(void *);
369 void sparc_softintr_disestablish(void *);
370
371 /* cpu.c */
372 int cpu_myid(void);
373
374 /* disksubr.c */
375 struct dkbad;
376 int isbad(struct dkbad *bt, int, int, int);
377 /* machdep.c */
378 void * reserve_dumppages(void *);
379 /* clock.c */
380 struct timeval;
381 int tickintr(void *); /* level 10/14 (tick) interrupt code */
382 int stickintr(void *); /* system tick interrupt code */
383 int stick2eintr(void *); /* system tick interrupt code */
384 int clockintr(void *); /* level 10 (clock) interrupt code */
385 int statintr(void *); /* level 14 (statclock) interrupt code */
386 int schedintr(void *); /* level 10 (schedclock) interrupt code */
387 void tickintr_establish(int, int (*)(void *));
388 void stickintr_establish(int, int (*)(void *));
389 void stick2eintr_establish(int, int (*)(void *));
390
391 /* locore.s */
392 struct fpstate64;
393 void savefpstate(struct fpstate64 *);
394 void loadfpstate(struct fpstate64 *);
395 void clearfpstate(void);
396 uint64_t probeget(paddr_t, int, int);
397 int probeset(paddr_t, int, int, uint64_t);
398 void setcputyp(int);
399
400 #define write_all_windows() __asm volatile("flushw" : : )
401 #define write_user_windows() __asm volatile("flushw" : : )
402
403 struct pcb;
404 void snapshot(struct pcb *);
405 struct frame *getfp(void);
406 void switchtoctx_us(int);
407 void switchtoctx_usiii(int);
408 void next_tick(long);
409 void next_stick(long);
410 /* trap.c */
411 void cpu_vmspace_exec(struct lwp *, vaddr_t, vaddr_t);
412 int rwindow_save(struct lwp *);
413 /* cons.c */
414 int cnrom(void);
415 /* zs.c */
416 void zsconsole(struct tty *, int, int, void (**)(struct tty *, int));
417 /* fb.c */
418 void fb_unblank(void);
419 /* kgdb_stub.c */
420 #ifdef KGDB
421 void kgdb_attach(int (*)(void *), void (*)(void *, int), void *);
422 void kgdb_connect(int);
423 void kgdb_panic(void);
424 #endif
425 /* emul.c */
426 int fixalign(struct lwp *, struct trapframe64 *);
427 int emulinstr(vaddr_t, struct trapframe64 *);
428
429 #else /* _KERNEL */
430
431 /*
432 * XXX: provide some definitions for crash(8), probably can share
433 */
434 #if defined(_KMEMUSER)
435 #define curcpu() (((struct cpu_info *)CPUINFO_VA)->ci_self)
436 #define curlwp curcpu()->ci_curlwp
437 #endif
438
439 #endif /* _KERNEL */
440 #endif /* _CPU_H_ */
441