cpu.h revision 1.102 1 /* $NetBSD: cpu.h,v 1.102 2013/06/21 20:09:59 nakayama Exp $ */
2
3 /*
4 * Copyright (c) 1992, 1993
5 * The Regents of the University of California. All rights reserved.
6 *
7 * This software was developed by the Computer Systems Engineering group
8 * at Lawrence Berkeley Laboratory under DARPA contract BG 91-66 and
9 * contributed to Berkeley.
10 *
11 * All advertising materials mentioning features or use of this software
12 * must display the following acknowledgement:
13 * This product includes software developed by the University of
14 * California, Lawrence Berkeley Laboratory.
15 *
16 * Redistribution and use in source and binary forms, with or without
17 * modification, are permitted provided that the following conditions
18 * are met:
19 * 1. Redistributions of source code must retain the above copyright
20 * notice, this list of conditions and the following disclaimer.
21 * 2. Redistributions in binary form must reproduce the above copyright
22 * notice, this list of conditions and the following disclaimer in the
23 * documentation and/or other materials provided with the distribution.
24 * 3. Neither the name of the University nor the names of its contributors
25 * may be used to endorse or promote products derived from this software
26 * without specific prior written permission.
27 *
28 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
29 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
30 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
31 * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
32 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
33 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
34 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
35 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
36 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
37 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
38 * SUCH DAMAGE.
39 *
40 * @(#)cpu.h 8.4 (Berkeley) 1/5/94
41 */
42
43 #ifndef _CPU_H_
44 #define _CPU_H_
45
46 /*
47 * CTL_MACHDEP definitions.
48 */
49 #define CPU_BOOTED_KERNEL 1 /* string: booted kernel name */
50 #define CPU_BOOTED_DEVICE 2 /* string: device booted from */
51 #define CPU_BOOT_ARGS 3 /* string: args booted with */
52 #define CPU_ARCH 4 /* integer: cpu architecture version */
53 #define CPU_VIS 5 /* 0 - no VIS, 1 - VIS 1.0, etc. */
54 #define CPU_MAXID 6 /* number of valid machdep ids */
55
56 #if defined(_KERNEL) || defined(_KMEMUSER)
57 /*
58 * Exported definitions unique to SPARC cpu support.
59 */
60
61 #if defined(_KERNEL_OPT)
62 #include "opt_multiprocessor.h"
63 #include "opt_lockdebug.h"
64 #endif
65
66 #include <machine/psl.h>
67 #include <machine/reg.h>
68 #include <machine/pte.h>
69 #include <machine/intr.h>
70 #if defined(_KERNEL)
71 #include <machine/cpuset.h>
72 #include <sparc64/sparc64/intreg.h>
73 #endif
74
75 #include <sys/cpu_data.h>
76 #include <sys/evcnt.h>
77
78 /*
79 * The cpu_info structure is part of a 64KB structure mapped both the kernel
80 * pmap and a single locked TTE a CPUINFO_VA for that particular processor.
81 * Each processor's cpu_info is accessible at CPUINFO_VA only for that
82 * processor. Other processors can access that through an additional mapping
83 * in the kernel pmap.
84 *
85 * The 64KB page contains:
86 *
87 * cpu_info
88 * interrupt stack (all remaining space)
89 * idle PCB
90 * idle stack (STACKSPACE - sizeof(PCB))
91 * 32KB TSB
92 */
93
94 struct cpu_info {
95 struct cpu_data ci_data; /* MI per-cpu data */
96
97
98 /*
99 * SPARC cpu_info structures live at two VAs: one global
100 * VA (so each CPU can access any other CPU's cpu_info)
101 * and an alias VA CPUINFO_VA which is the same on each
102 * CPU and maps to that CPU's cpu_info. Since the alias
103 * CPUINFO_VA is how we locate our cpu_info, we have to
104 * self-reference the global VA so that we can return it
105 * in the curcpu() macro.
106 */
107 struct cpu_info * volatile ci_self;
108
109 /* Most important fields first */
110 struct lwp *ci_curlwp;
111 struct pcb *ci_cpcb;
112 struct cpu_info *ci_next;
113
114 struct lwp *ci_fplwp;
115
116 void *ci_eintstack;
117
118 int ci_mtx_count;
119 int ci_mtx_oldspl;
120
121 /* Spinning up the CPU */
122 void (*ci_spinup)(void);
123 paddr_t ci_paddr;
124
125 int ci_cpuid;
126
127 /* CPU PROM information. */
128 u_int ci_node;
129
130 /* %tick and cpu frequency information */
131 u_long ci_tick_increment;
132 uint64_t ci_cpu_clockrate[2]; /* %tick */
133 uint64_t ci_system_clockrate[2]; /* %stick */
134
135 /* Interrupts */
136 struct intrhand *ci_intrpending[16];
137 struct intrhand *ci_tick_ih;
138
139 /* Event counters */
140 struct evcnt ci_tick_evcnt;
141
142 /* This could be under MULTIPROCESSOR, but there's no good reason */
143 struct evcnt ci_ipi_evcnt[IPI_EVCNT_NUM];
144
145 int ci_flags;
146 int ci_want_ast;
147 int ci_want_resched;
148 int ci_idepth;
149
150 /*
151 * A context is simply a small number that differentiates multiple mappings
152 * of the same address. Contexts on the spitfire are 13 bits, but could
153 * be as large as 17 bits.
154 *
155 * Each context is either free or attached to a pmap.
156 *
157 * The context table is an array of pointers to psegs. Just dereference
158 * the right pointer and you get to the pmap segment tables. These are
159 * physical addresses, of course.
160 *
161 * ci_ctx_lock protects this CPUs context allocation/free.
162 * These are all allocated almost with in the same cacheline.
163 */
164 kmutex_t ci_ctx_lock;
165 int ci_pmap_next_ctx;
166 int ci_numctx;
167 paddr_t *ci_ctxbusy;
168 LIST_HEAD(, pmap) ci_pmap_ctxlist;
169
170 /*
171 * The TSBs are per cpu too (since MMU context differs between
172 * cpus). These are just caches for the TLBs.
173 */
174 pte_t *ci_tsb_dmmu;
175 pte_t *ci_tsb_immu;
176
177 /* probe fault in PCI config space reads */
178 bool ci_pci_probe;
179 bool ci_pci_fault;
180
181 volatile void *ci_ddb_regs; /* DDB regs */
182 };
183
184 #endif /* _KERNEL || _KMEMUSER */
185
186 #ifdef _KERNEL
187
188 #define CPUF_PRIMARY 1
189
190 /*
191 * CPU boot arguments. Used by secondary CPUs at the bootstrap time.
192 */
193 struct cpu_bootargs {
194 u_int cb_node; /* PROM CPU node */
195 volatile int cb_flags;
196
197 vaddr_t cb_ktext;
198 paddr_t cb_ktextp;
199 vaddr_t cb_ektext;
200
201 vaddr_t cb_kdata;
202 paddr_t cb_kdatap;
203 vaddr_t cb_ekdata;
204
205 paddr_t cb_cpuinfo;
206 };
207
208 extern struct cpu_bootargs *cpu_args;
209
210 #if defined(MULTIPROCESSOR)
211 extern int sparc_ncpus;
212 #else
213 #define sparc_ncpus 1
214 #endif
215
216 extern struct cpu_info *cpus;
217 extern struct pool_cache *fpstate_cache;
218
219 #define curcpu() (((struct cpu_info *)CPUINFO_VA)->ci_self)
220 #define cpu_number() (curcpu()->ci_index)
221 #define CPU_IS_PRIMARY(ci) ((ci)->ci_flags & CPUF_PRIMARY)
222
223 #define CPU_INFO_ITERATOR int
224 #define CPU_INFO_FOREACH(cii, ci) cii = 0, ci = cpus; ci != NULL; \
225 ci = ci->ci_next
226
227 #define curlwp curcpu()->ci_curlwp
228 #define fplwp curcpu()->ci_fplwp
229 #define curpcb curcpu()->ci_cpcb
230
231 #define want_ast curcpu()->ci_want_ast
232 #define want_resched curcpu()->ci_want_resched
233
234 /*
235 * definitions of cpu-dependent requirements
236 * referenced in generic code
237 */
238 #define cpu_wait(p) /* nothing */
239 void cpu_proc_fork(struct proc *, struct proc *);
240
241 /* run on the cpu itself */
242 void cpu_pmap_init(struct cpu_info *);
243 /* run upfront to prepare the cpu_info */
244 void cpu_pmap_prepare(struct cpu_info *, bool);
245
246 #if defined(MULTIPROCESSOR)
247 extern vaddr_t cpu_spinup_trampoline;
248
249 extern char *mp_tramp_code;
250 extern u_long mp_tramp_code_len;
251 extern u_long mp_tramp_tlb_slots;
252 extern u_long mp_tramp_func;
253 extern u_long mp_tramp_ci;
254
255 void cpu_hatch(void);
256 void cpu_boot_secondary_processors(void);
257
258 /*
259 * Call a function on other cpus:
260 * multicast - send to everyone in the sparc64_cpuset_t
261 * broadcast - send to to all cpus but ourselves
262 * send - send to just this cpu
263 * The called function do not follow the C ABI, so need to be coded in
264 * assembler.
265 */
266 typedef void (* ipifunc_t)(void *, void *);
267
268 void sparc64_multicast_ipi(sparc64_cpuset_t, ipifunc_t, uint64_t, uint64_t);
269 void sparc64_broadcast_ipi(ipifunc_t, uint64_t, uint64_t);
270 void sparc64_send_ipi(int, ipifunc_t, uint64_t, uint64_t);
271
272 /*
273 * Call an arbitrary C function on another cpu (or all others but ourself)
274 */
275 typedef void (*ipi_c_call_func_t)(void*);
276 void sparc64_generic_xcall(struct cpu_info*, ipi_c_call_func_t, void*);
277
278 #endif
279
280 /* Provide %pc of a lwp */
281 #define LWP_PC(l) ((l)->l_md.md_tf->tf_pc)
282
283 /*
284 * Arguments to hardclock, softclock and gatherstats encapsulate the
285 * previous machine state in an opaque clockframe. The ipl is here
286 * as well for strayintr (see locore.s:interrupt and intr.c:strayintr).
287 * Note that CLKF_INTR is valid only if CLKF_USERMODE is false.
288 */
289 struct clockframe {
290 struct trapframe64 t;
291 };
292
293 #define CLKF_USERMODE(framep) (((framep)->t.tf_tstate & TSTATE_PRIV) == 0)
294 #define CLKF_PC(framep) ((framep)->t.tf_pc)
295 /* Since some files in sys/kern do not know BIAS, I'm using 0x7ff here */
296 #define CLKF_INTR(framep) \
297 ((!CLKF_USERMODE(framep))&& \
298 (((framep)->t.tf_out[6] & 1 ) ? \
299 (((vaddr_t)(framep)->t.tf_out[6] < \
300 (vaddr_t)EINTSTACK-0x7ff) && \
301 ((vaddr_t)(framep)->t.tf_out[6] > \
302 (vaddr_t)INTSTACK-0x7ff)) : \
303 (((vaddr_t)(framep)->t.tf_out[6] < \
304 (vaddr_t)EINTSTACK) && \
305 ((vaddr_t)(framep)->t.tf_out[6] > \
306 (vaddr_t)INTSTACK))))
307
308 /*
309 * Give a profiling tick to the current process when the user profiling
310 * buffer pages are invalid. On the sparc, request an ast to send us
311 * through trap(), marking the proc as needing a profiling tick.
312 */
313 #define cpu_need_proftick(l) ((l)->l_pflag |= LP_OWEUPC, want_ast = 1)
314
315 /*
316 * Notify an LWP that it has a signal pending, process as soon as possible.
317 */
318 void cpu_signotify(struct lwp *);
319
320 /*
321 * Interrupt handler chains. Interrupt handlers should return 0 for
322 * ``not me'' or 1 (``I took care of it''). intr_establish() inserts a
323 * handler into the list. The handler is called with its (single)
324 * argument, or with a pointer to a clockframe if ih_arg is NULL.
325 */
326 struct intrhand {
327 int (*ih_fun)(void *);
328 void *ih_arg;
329 /* if we have to take the biglock, we interpose a wrapper
330 * and need to save the original function and arg */
331 int (*ih_realfun)(void *);
332 void *ih_realarg;
333 short ih_number; /* interrupt number */
334 /* the H/W provides */
335 char ih_pil; /* interrupt priority */
336 struct intrhand *ih_next; /* global list */
337 struct intrhand *ih_pending; /* interrupt queued */
338 volatile uint64_t *ih_map; /* Interrupt map reg */
339 volatile uint64_t *ih_clr; /* clear interrupt reg */
340 struct evcnt ih_cnt; /* counter for vmstat */
341 uint32_t ih_ivec;
342 char ih_name[32]; /* name for the above */
343 };
344 extern struct intrhand *intrhand[];
345 extern struct intrhand *intrlev[MAXINTNUM];
346
347 void intr_establish(int level, bool mpsafe, struct intrhand *);
348 void *sparc_softintr_establish(int, int (*)(void *), void *);
349 void sparc_softintr_schedule(void *);
350 void sparc_softintr_disestablish(void *);
351
352 /* disksubr.c */
353 struct dkbad;
354 int isbad(struct dkbad *bt, int, int, int);
355 /* machdep.c */
356 void * reserve_dumppages(void *);
357 /* clock.c */
358 struct timeval;
359 int tickintr(void *); /* level 10/14 (tick) interrupt code */
360 int stickintr(void *); /* system tick interrupt code */
361 int clockintr(void *); /* level 10 (clock) interrupt code */
362 int statintr(void *); /* level 14 (statclock) interrupt code */
363 int schedintr(void *); /* level 10 (schedclock) interrupt code */
364 void tickintr_establish(int, int (*)(void *));
365 void stickintr_establish(int, int (*)(void *));
366 /* locore.s */
367 struct fpstate64;
368 void savefpstate(struct fpstate64 *);
369 void loadfpstate(struct fpstate64 *);
370 void clearfpstate(void);
371 uint64_t probeget(paddr_t, int, int);
372 int probeset(paddr_t, int, int, uint64_t);
373
374 #define write_all_windows() __asm volatile("flushw" : : )
375 #define write_user_windows() __asm volatile("flushw" : : )
376
377 struct pcb;
378 void snapshot(struct pcb *);
379 struct frame *getfp(void);
380 void switchtoctx_us(int);
381 void switchtoctx_usiii(int);
382 void next_tick(long);
383 void next_stick(long);
384 /* trap.c */
385 void cpu_vmspace_exec(struct lwp *, vaddr_t, vaddr_t);
386 int rwindow_save(struct lwp *);
387 /* cons.c */
388 int cnrom(void);
389 /* zs.c */
390 void zsconsole(struct tty *, int, int, void (**)(struct tty *, int));
391 /* fb.c */
392 void fb_unblank(void);
393 /* kgdb_stub.c */
394 #ifdef KGDB
395 void kgdb_attach(int (*)(void *), void (*)(void *, int), void *);
396 void kgdb_connect(int);
397 void kgdb_panic(void);
398 #endif
399 /* emul.c */
400 int fixalign(struct lwp *, struct trapframe64 *);
401 int emulinstr(vaddr_t, struct trapframe64 *);
402
403 #else /* _KERNEL */
404
405 /*
406 * XXX: provide some definitions for crash(8), probably can share
407 */
408 #if defined(_KMEMUSER)
409 #define curcpu() (((struct cpu_info *)CPUINFO_VA)->ci_self)
410 #define curlwp curcpu()->ci_curlwp
411 #endif
412
413 #endif /* _KERNEL */
414 #endif /* _CPU_H_ */
415