cpu.h revision 1.98.12.2 1 /* $NetBSD: cpu.h,v 1.98.12.2 2013/02/25 00:28:59 tls Exp $ */
2
3 /*
4 * Copyright (c) 1992, 1993
5 * The Regents of the University of California. All rights reserved.
6 *
7 * This software was developed by the Computer Systems Engineering group
8 * at Lawrence Berkeley Laboratory under DARPA contract BG 91-66 and
9 * contributed to Berkeley.
10 *
11 * All advertising materials mentioning features or use of this software
12 * must display the following acknowledgement:
13 * This product includes software developed by the University of
14 * California, Lawrence Berkeley Laboratory.
15 *
16 * Redistribution and use in source and binary forms, with or without
17 * modification, are permitted provided that the following conditions
18 * are met:
19 * 1. Redistributions of source code must retain the above copyright
20 * notice, this list of conditions and the following disclaimer.
21 * 2. Redistributions in binary form must reproduce the above copyright
22 * notice, this list of conditions and the following disclaimer in the
23 * documentation and/or other materials provided with the distribution.
24 * 3. Neither the name of the University nor the names of its contributors
25 * may be used to endorse or promote products derived from this software
26 * without specific prior written permission.
27 *
28 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
29 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
30 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
31 * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
32 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
33 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
34 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
35 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
36 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
37 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
38 * SUCH DAMAGE.
39 *
40 * @(#)cpu.h 8.4 (Berkeley) 1/5/94
41 */
42
43 #ifndef _CPU_H_
44 #define _CPU_H_
45
46 /*
47 * CTL_MACHDEP definitions.
48 */
49 #define CPU_BOOTED_KERNEL 1 /* string: booted kernel name */
50 #define CPU_BOOTED_DEVICE 2 /* string: device booted from */
51 #define CPU_BOOT_ARGS 3 /* string: args booted with */
52 #define CPU_ARCH 4 /* integer: cpu architecture version */
53 #define CPU_VIS 5 /* 0 - no VIS, 1 - VIS 1.0, etc. */
54 #define CPU_MAXID 6 /* number of valid machdep ids */
55
56 #if defined(_KERNEL) || defined(_KMEMUSER)
57 /*
58 * Exported definitions unique to SPARC cpu support.
59 */
60
61 #if defined(_KERNEL_OPT)
62 #include "opt_multiprocessor.h"
63 #include "opt_lockdebug.h"
64 #endif
65
66 #include <machine/psl.h>
67 #include <machine/reg.h>
68 #include <machine/pte.h>
69 #include <machine/intr.h>
70 #if defined(_KERNEL)
71 #include <machine/cpuset.h>
72 #include <sparc64/sparc64/intreg.h>
73 #endif
74
75 #include <sys/cpu_data.h>
76 #include <sys/evcnt.h>
77
78 /*
79 * The cpu_info structure is part of a 64KB structure mapped both the kernel
80 * pmap and a single locked TTE a CPUINFO_VA for that particular processor.
81 * Each processor's cpu_info is accessible at CPUINFO_VA only for that
82 * processor. Other processors can access that through an additional mapping
83 * in the kernel pmap.
84 *
85 * The 64KB page contains:
86 *
87 * cpu_info
88 * interrupt stack (all remaining space)
89 * idle PCB
90 * idle stack (STACKSPACE - sizeof(PCB))
91 * 32KB TSB
92 */
93
94 struct cpu_info {
95 struct cpu_data ci_data; /* MI per-cpu data */
96
97
98 /*
99 * SPARC cpu_info structures live at two VAs: one global
100 * VA (so each CPU can access any other CPU's cpu_info)
101 * and an alias VA CPUINFO_VA which is the same on each
102 * CPU and maps to that CPU's cpu_info. Since the alias
103 * CPUINFO_VA is how we locate our cpu_info, we have to
104 * self-reference the global VA so that we can return it
105 * in the curcpu() macro.
106 */
107 struct cpu_info * volatile ci_self;
108
109 /* Most important fields first */
110 struct lwp *ci_curlwp;
111 struct pcb *ci_cpcb;
112 struct cpu_info *ci_next;
113
114 struct lwp *ci_fplwp;
115
116 void *ci_eintstack;
117
118 int ci_mtx_count;
119 int ci_mtx_oldspl;
120
121 /* Spinning up the CPU */
122 void (*ci_spinup)(void);
123 paddr_t ci_paddr;
124
125 int ci_cpuid;
126
127 /* CPU PROM information. */
128 u_int ci_node;
129
130 /* %tick and cpu frequency information */
131 u_long ci_tick_increment;
132 uint64_t ci_cpu_clockrate[2]; /* %tick */
133 uint64_t ci_system_clockrate[2]; /* %stick */
134
135 /* Interrupts */
136 struct intrhand *ci_intrpending[16];
137 struct intrhand *ci_tick_ih;
138
139 /* Event counters */
140 struct evcnt ci_tick_evcnt;
141
142 /* This could be under MULTIPROCESSOR, but there's no good reason */
143 struct evcnt ci_ipi_evcnt[IPI_EVCNT_NUM];
144
145 int ci_flags;
146 int ci_want_ast;
147 int ci_want_resched;
148 int ci_idepth;
149
150 /*
151 * A context is simply a small number that differentiates multiple mappings
152 * of the same address. Contexts on the spitfire are 13 bits, but could
153 * be as large as 17 bits.
154 *
155 * Each context is either free or attached to a pmap.
156 *
157 * The context table is an array of pointers to psegs. Just dereference
158 * the right pointer and you get to the pmap segment tables. These are
159 * physical addresses, of course.
160 *
161 * ci_ctx_lock protects this CPUs context allocation/free.
162 * These are all allocated almost with in the same cacheline.
163 */
164 kmutex_t ci_ctx_lock;
165 int ci_pmap_next_ctx;
166 int ci_numctx;
167 paddr_t *ci_ctxbusy;
168 LIST_HEAD(, pmap) ci_pmap_ctxlist;
169
170 /*
171 * The TSBs are per cpu too (since MMU context differs between
172 * cpus). These are just caches for the TLBs.
173 */
174 pte_t *ci_tsb_dmmu;
175 pte_t *ci_tsb_immu;
176
177 volatile void *ci_ddb_regs; /* DDB regs */
178 };
179
180 #endif /* _KERNEL || _KMEMUSER */
181
182 #ifdef _KERNEL
183
184 #define CPUF_PRIMARY 1
185
186 /*
187 * CPU boot arguments. Used by secondary CPUs at the bootstrap time.
188 */
189 struct cpu_bootargs {
190 u_int cb_node; /* PROM CPU node */
191 volatile int cb_flags;
192
193 vaddr_t cb_ktext;
194 paddr_t cb_ktextp;
195 vaddr_t cb_ektext;
196
197 vaddr_t cb_kdata;
198 paddr_t cb_kdatap;
199 vaddr_t cb_ekdata;
200
201 paddr_t cb_cpuinfo;
202 };
203
204 extern struct cpu_bootargs *cpu_args;
205
206 #if defined(MULTIPROCESSOR)
207 extern int sparc_ncpus;
208 #else
209 #define sparc_ncpus 1
210 #endif
211
212 extern struct cpu_info *cpus;
213 extern struct pool_cache *fpstate_cache;
214
215 #define curcpu() (((struct cpu_info *)CPUINFO_VA)->ci_self)
216 #define cpu_number() (curcpu()->ci_index)
217 #define CPU_IS_PRIMARY(ci) ((ci)->ci_flags & CPUF_PRIMARY)
218
219 #define CPU_INFO_ITERATOR int
220 #define CPU_INFO_FOREACH(cii, ci) cii = 0, ci = cpus; ci != NULL; \
221 ci = ci->ci_next
222
223 #define curlwp curcpu()->ci_curlwp
224 #define fplwp curcpu()->ci_fplwp
225 #define curpcb curcpu()->ci_cpcb
226
227 #define want_ast curcpu()->ci_want_ast
228 #define want_resched curcpu()->ci_want_resched
229
230 /*
231 * definitions of cpu-dependent requirements
232 * referenced in generic code
233 */
234 #define cpu_wait(p) /* nothing */
235 void cpu_proc_fork(struct proc *, struct proc *);
236
237 /* run on the cpu itself */
238 void cpu_pmap_init(struct cpu_info *);
239 /* run upfront to prepare the cpu_info */
240 void cpu_pmap_prepare(struct cpu_info *, bool);
241
242 #if defined(MULTIPROCESSOR)
243 extern vaddr_t cpu_spinup_trampoline;
244
245 extern char *mp_tramp_code;
246 extern u_long mp_tramp_code_len;
247 extern u_long mp_tramp_tlb_slots;
248 extern u_long mp_tramp_func;
249 extern u_long mp_tramp_ci;
250
251 void cpu_hatch(void);
252 void cpu_boot_secondary_processors(void);
253
254 /*
255 * Call a function on other cpus:
256 * multicast - send to everyone in the sparc64_cpuset_t
257 * broadcast - send to to all cpus but ourselves
258 * send - send to just this cpu
259 * The called function do not follow the C ABI, so need to be coded in
260 * assembler.
261 */
262 typedef void (* ipifunc_t)(void *, void *);
263
264 void sparc64_multicast_ipi(sparc64_cpuset_t, ipifunc_t, uint64_t, uint64_t);
265 void sparc64_broadcast_ipi(ipifunc_t, uint64_t, uint64_t);
266 void sparc64_send_ipi(int, ipifunc_t, uint64_t, uint64_t);
267
268 /*
269 * Call an arbitrary C function on another cpu (or all others but ourself)
270 */
271 typedef void (*ipi_c_call_func_t)(void*);
272 void sparc64_generic_xcall(struct cpu_info*, ipi_c_call_func_t, void*);
273
274 #endif
275
276 /* Provide %pc of a lwp */
277 #define LWP_PC(l) ((l)->l_md.md_tf->tf_pc)
278
279 /*
280 * Arguments to hardclock, softclock and gatherstats encapsulate the
281 * previous machine state in an opaque clockframe. The ipl is here
282 * as well for strayintr (see locore.s:interrupt and intr.c:strayintr).
283 * Note that CLKF_INTR is valid only if CLKF_USERMODE is false.
284 */
285 struct clockframe {
286 struct trapframe64 t;
287 };
288
289 #define CLKF_USERMODE(framep) (((framep)->t.tf_tstate & TSTATE_PRIV) == 0)
290 #define CLKF_PC(framep) ((framep)->t.tf_pc)
291 /* Since some files in sys/kern do not know BIAS, I'm using 0x7ff here */
292 #define CLKF_INTR(framep) \
293 ((!CLKF_USERMODE(framep))&& \
294 (((framep)->t.tf_out[6] & 1 ) ? \
295 (((vaddr_t)(framep)->t.tf_out[6] < \
296 (vaddr_t)EINTSTACK-0x7ff) && \
297 ((vaddr_t)(framep)->t.tf_out[6] > \
298 (vaddr_t)INTSTACK-0x7ff)) : \
299 (((vaddr_t)(framep)->t.tf_out[6] < \
300 (vaddr_t)EINTSTACK) && \
301 ((vaddr_t)(framep)->t.tf_out[6] > \
302 (vaddr_t)INTSTACK))))
303
304 /*
305 * Give a profiling tick to the current process when the user profiling
306 * buffer pages are invalid. On the sparc, request an ast to send us
307 * through trap(), marking the proc as needing a profiling tick.
308 */
309 #define cpu_need_proftick(l) ((l)->l_pflag |= LP_OWEUPC, want_ast = 1)
310
311 /*
312 * Notify an LWP that it has a signal pending, process as soon as possible.
313 */
314 void cpu_signotify(struct lwp *);
315
316 /*
317 * Interrupt handler chains. Interrupt handlers should return 0 for
318 * ``not me'' or 1 (``I took care of it''). intr_establish() inserts a
319 * handler into the list. The handler is called with its (single)
320 * argument, or with a pointer to a clockframe if ih_arg is NULL.
321 */
322 struct intrhand {
323 int (*ih_fun)(void *);
324 void *ih_arg;
325 /* if we have to take the biglock, we interpose a wrapper
326 * and need to save the original function and arg */
327 int (*ih_realfun)(void *);
328 void *ih_realarg;
329 short ih_number; /* interrupt number */
330 /* the H/W provides */
331 char ih_pil; /* interrupt priority */
332 struct intrhand *ih_next; /* global list */
333 struct intrhand *ih_pending; /* interrupt queued */
334 volatile uint64_t *ih_map; /* Interrupt map reg */
335 volatile uint64_t *ih_clr; /* clear interrupt reg */
336 struct evcnt ih_cnt; /* counter for vmstat */
337 uint32_t ih_ivec;
338 char ih_name[32]; /* name for the above */
339 };
340 extern struct intrhand *intrhand[];
341 extern struct intrhand *intrlev[MAXINTNUM];
342
343 void intr_establish(int level, bool mpsafe, struct intrhand *);
344 void *sparc_softintr_establish(int, int (*)(void *), void *);
345 void sparc_softintr_schedule(void *);
346 void sparc_softintr_disestablish(void *);
347
348 /* disksubr.c */
349 struct dkbad;
350 int isbad(struct dkbad *bt, int, int, int);
351 /* machdep.c */
352 void * reserve_dumppages(void *);
353 /* clock.c */
354 struct timeval;
355 int tickintr(void *); /* level 10/14 (tick) interrupt code */
356 int stickintr(void *); /* system tick interrupt code */
357 int clockintr(void *); /* level 10 (clock) interrupt code */
358 int statintr(void *); /* level 14 (statclock) interrupt code */
359 int schedintr(void *); /* level 10 (schedclock) interrupt code */
360 void tickintr_establish(int, int (*)(void *));
361 void stickintr_establish(int, int (*)(void *));
362 /* locore.s */
363 struct fpstate64;
364 void savefpstate(struct fpstate64 *);
365 void loadfpstate(struct fpstate64 *);
366 void clearfpstate(void);
367 uint64_t probeget(paddr_t, int, int);
368 int probeset(paddr_t, int, int, uint64_t);
369
370 #define write_all_windows() __asm volatile("flushw" : : )
371 #define write_user_windows() __asm volatile("flushw" : : )
372
373 struct pcb;
374 void snapshot(struct pcb *);
375 struct frame *getfp(void);
376 void switchtoctx_us(int);
377 void switchtoctx_usiii(int);
378 void next_tick(long);
379 void next_stick(long);
380 /* trap.c */
381 void cpu_vmspace_exec(struct lwp *, vaddr_t, vaddr_t);
382 int rwindow_save(struct lwp *);
383 /* cons.c */
384 int cnrom(void);
385 /* zs.c */
386 void zsconsole(struct tty *, int, int, void (**)(struct tty *, int));
387 /* fb.c */
388 void fb_unblank(void);
389 /* kgdb_stub.c */
390 #ifdef KGDB
391 void kgdb_attach(int (*)(void *), void (*)(void *, int), void *);
392 void kgdb_connect(int);
393 void kgdb_panic(void);
394 #endif
395 /* emul.c */
396 int fixalign(struct lwp *, struct trapframe64 *);
397 int emulinstr(vaddr_t, struct trapframe64 *);
398
399 #else /* _KERNEL */
400
401 /*
402 * XXX: provide some definitions for crash(8), probably can share
403 */
404 #if defined(_KMEMUSER)
405 #define curcpu() (((struct cpu_info *)CPUINFO_VA)->ci_self)
406 #define curlwp curcpu()->ci_curlwp
407 #endif
408
409 #endif /* _KERNEL */
410 #endif /* _CPU_H_ */
411