pthread_int.h revision 1.37 1 /* $NetBSD: pthread_int.h,v 1.37 2007/02/06 15:24:37 ad Exp $ */
2
3 /*-
4 * Copyright (c) 2001,2002,2003 The NetBSD Foundation, Inc.
5 * All rights reserved.
6 *
7 * This code is derived from software contributed to The NetBSD Foundation
8 * by Nathan J. Williams.
9 *
10 * Redistribution and use in source and binary forms, with or without
11 * modification, are permitted provided that the following conditions
12 * are met:
13 * 1. Redistributions of source code must retain the above copyright
14 * notice, this list of conditions and the following disclaimer.
15 * 2. Redistributions in binary form must reproduce the above copyright
16 * notice, this list of conditions and the following disclaimer in the
17 * documentation and/or other materials provided with the distribution.
18 * 3. All advertising materials mentioning features or use of this software
19 * must display the following acknowledgement:
20 * This product includes software developed by the NetBSD
21 * Foundation, Inc. and its contributors.
22 * 4. Neither the name of The NetBSD Foundation nor the names of its
23 * contributors may be used to endorse or promote products derived
24 * from this software without specific prior written permission.
25 *
26 * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
27 * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
28 * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
29 * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
30 * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
31 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
32 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
33 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
34 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
35 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
36 * POSSIBILITY OF SUCH DAMAGE.
37 */
38
39 #ifndef _LIB_PTHREAD_INT_H
40 #define _LIB_PTHREAD_INT_H
41
42 #define PTHREAD__DEBUG
43 #define ERRORCHECK
44
45 #include "pthread_types.h"
46 #include "pthread_queue.h"
47 #include "pthread_debug.h"
48 #include "pthread_md.h"
49
50 #ifndef PTHREAD_SA
51 #include <lwp.h>
52 #else
53 #include <sa.h>
54 #endif
55
56 #include <signal.h>
57
58 #define PTHREAD_KEYS_MAX 256
59 /*
60 * The size of this structure needs to be no larger than struct
61 * __pthread_cleanup_store, defined in pthread.h.
62 */
63 struct pt_clean_t {
64 PTQ_ENTRY(pt_clean_t) ptc_next;
65 void (*ptc_cleanup)(void *);
66 void *ptc_arg;
67 };
68
69 struct pt_alarm_t {
70 PTQ_ENTRY(pt_alarm_t) pta_next;
71 pthread_spin_t pta_lock;
72 const struct timespec *pta_time;
73 void (*pta_func)(void *);
74 void *pta_arg;
75 int pta_fired;
76 };
77
78 /* Private data for pthread_attr_t */
79 struct pthread_attr_private {
80 char ptap_name[PTHREAD_MAX_NAMELEN_NP];
81 void *ptap_namearg;
82 void *ptap_stackaddr;
83 size_t ptap_stacksize;
84 size_t ptap_guardsize;
85 };
86
87 struct __pthread_st {
88 unsigned int pt_magic;
89 /* Identifier, for debugging and for preventing recycling. */
90 int pt_num;
91
92 #ifdef PTHREAD_SA
93 int pt_type; /* normal, upcall, or idle */
94 #else
95 lwpid_t pt_lid; /* LWP ID */
96 #endif
97 int pt_state; /* running, blocked, etc. */
98 pthread_spin_t pt_statelock; /* lock on pt_state */
99 int pt_flags; /* see PT_FLAG_* below */
100 pthread_spin_t pt_flaglock; /* lock on pt_flag */
101 int pt_cancel; /* Deferred cancellation */
102 int pt_spinlocks; /* Number of spinlocks held. */
103 int pt_blockedlwp; /* LWP/SA number when blocked */
104 int pt_vpid; /* VP number */
105 int pt_blockgen; /* SA_UPCALL_BLOCKED counter */
106 #ifdef PTHREAD_SA
107 int pt_unblockgen; /* SA_UPCALL_UNBLOCKED counter */
108 #else
109 int pt_sleeponq; /* on a sleep queue */
110 #endif
111
112 int pt_errno; /* Thread-specific errno. */
113
114 /* Entry on the run queue */
115 PTQ_ENTRY(__pthread_st) pt_runq;
116 /* Entry on the list of all threads */
117 PTQ_ENTRY(__pthread_st) pt_allq;
118 /* Entry on the sleep queue (xxx should be same as run queue?) */
119 PTQ_ENTRY(__pthread_st) pt_sleep;
120 /*
121 * Object we're sleeping on. For 1:1 threads (!SA), this is
122 * protected by the interlock on the object that the thread is
123 * sleeping on.
124 */
125 void *pt_sleepobj;
126 /* Queue we're sleeping on */
127 struct pthread_queue_t *pt_sleepq;
128 /* Lock protecting that queue */
129 pthread_spin_t *pt_sleeplock;
130
131 stack_t pt_stack; /* Our stack */
132 ucontext_t *pt_uc; /* Saved context when we're stopped */
133 ucontext_t *pt_trapuc; /* Kernel-saved context */
134 ucontext_t *__pt_blockuc; /* Kernel-saved context when blocked */
135
136 sigset_t pt_sigmask; /* Signals we won't take. */
137 sigset_t pt_siglist; /* Signals pending for us. */
138 sigset_t pt_sigblocked; /* Signals delivered while blocked. */
139 sigset_t *pt_sigwait; /* Signals waited for in sigwait */
140 siginfo_t *pt_wsig;
141 pthread_spin_t pt_siglock; /* Lock on above */
142
143 void * pt_exitval; /* Read by pthread_join() */
144
145 /* Stack of cancellation cleanup handlers and their arguments */
146 PTQ_HEAD(, pt_clean_t) pt_cleanup_stack;
147
148 /* Thread's name, set by the application. */
149 char* pt_name;
150
151 /* Other threads trying to pthread_join() us. */
152 struct pthread_queue_t pt_joiners;
153 /* Lock for above, and for changing pt_state to ZOMBIE or DEAD,
154 * and for setting the DETACHED flag. Also protects pt_name.
155 */
156 pthread_spin_t pt_join_lock;
157
158 /* Thread we were going to switch to before we were preempted
159 * ourselves. Will be used by the upcall that's continuing us.
160 */
161 pthread_t pt_switchto;
162 ucontext_t* pt_switchtouc;
163
164 /* Threads that are preempted with spinlocks held will be
165 * continued until they unlock their spinlock. When they do
166 * so, they should jump ship to the thread pointed to by
167 * pt_next.
168 */
169 pthread_t pt_next;
170
171 /* The upcall that is continuing this thread */
172 pthread_t pt_parent;
173
174 /* A queue lock that this thread held while trying to
175 * context switch to another process.
176 */
177 pthread_spin_t* pt_heldlock;
178
179 #ifdef PTHREAD_SA
180 /* Upcall stack information shared between kernel and
181 * userland.
182 */
183 struct sa_stackinfo_t pt_stackinfo;
184 #endif
185
186 /* Thread-specific data */
187 void* pt_specific[PTHREAD_KEYS_MAX];
188
189 #ifdef PTHREAD__DEBUG
190 int blocks;
191 int preempts;
192 int rescheds;
193 #endif
194 };
195
196 struct pthread_lock_ops {
197 void (*plo_init)(__cpu_simple_lock_t *);
198 int (*plo_try)(__cpu_simple_lock_t *);
199 void (*plo_unlock)(__cpu_simple_lock_t *);
200 };
201
202 #ifdef PTHREAD_SA
203
204 /* Thread types */
205 #define PT_THREAD_NORMAL 1
206 #define PT_THREAD_UPCALL 2
207 #define PT_THREAD_IDLE 3
208
209 /* Thread states */
210 #define PT_STATE_RUNNING 1
211 #define PT_STATE_RUNNABLE 2
212 #define _PT_STATE_BLOCKED_SYS 3 /* Only used in libpthread_dbg */
213 #define PT_STATE_BLOCKED_QUEUE 4
214 #define PT_STATE_ZOMBIE 5
215 #define PT_STATE_DEAD 6
216 #define PT_STATE_SUSPENDED 7
217
218 #else /* PTHREAD_SA */
219
220 /* Thread states */
221 #define PT_STATE_RUNNING 1
222 #define PT_STATE_ZOMBIE 5
223 #define PT_STATE_DEAD 6
224
225 #endif /* PTHREAD_SA */
226
227 /* Flag values */
228
229 #define PT_FLAG_DETACHED 0x0001
230 #define PT_FLAG_IDLED 0x0002
231 #define PT_FLAG_CS_DISABLED 0x0004 /* Cancellation disabled */
232 #define PT_FLAG_CS_ASYNC 0x0008 /* Cancellation is async */
233 #define PT_FLAG_CS_PENDING 0x0010
234 #define PT_FLAG_SIGDEFERRED 0x0020 /* There are signals to take */
235 #define PT_FLAG_SCOPE_SYSTEM 0x0040
236 #define PT_FLAG_EXPLICIT_SCHED 0x0080
237 #define PT_FLAG_SUSPENDED 0x0100 /* In the suspended queue */
238 #define PT_FLAG_SIGNALED 0x0200
239
240 #define PT_MAGIC 0x11110001
241 #define PT_DEAD 0xDEAD0001
242
243 #define PT_ATTR_MAGIC 0x22220002
244 #define PT_ATTR_DEAD 0xDEAD0002
245
246 #ifdef PT_FIXEDSTACKSIZE_LG
247
248 #define PT_STACKSIZE_LG PT_FIXEDSTACKSIZE_LG
249 #define PT_STACKSIZE (1<<(PT_STACKSIZE_LG))
250 #define PT_STACKMASK (PT_STACKSIZE-1)
251
252 #else /* PT_FIXEDSTACKSIZE_LG */
253
254 extern int pthread_stacksize_lg;
255 extern size_t pthread_stacksize;
256 extern vaddr_t pthread_stackmask;
257
258 #define PT_STACKSIZE_LG pthread_stacksize_lg
259 #define PT_STACKSIZE pthread_stacksize
260 #define PT_STACKMASK pthread_stackmask
261
262 #endif /* PT_FIXEDSTACKSIZE_LG */
263
264
265 #define PT_UPCALLSTACKS 16
266
267 #define PT_ALARMTIMER_MAGIC 0x88880010
268 #define PT_RRTIMER_MAGIC 0x88880020
269 #define NIDLETHREADS 4
270
271 /* Flag to be used in a ucontext_t's uc_flags indicating that
272 * the saved register state is "user" state only, not full
273 * trap state.
274 */
275 #define _UC_USER_BIT 30
276 #define _UC_USER (1LU << _UC_USER_BIT)
277
278 void pthread_init(void) __attribute__ ((__constructor__));
279
280 /* Utility functions */
281
282 /* Set up/clean up a thread's basic state. */
283 void pthread__initthread(pthread_t self, pthread_t t);
284 /* Get offset from stack start to struct sa_stackinfo */
285 ssize_t pthread__stackinfo_offset(void);
286
287 /* Go do something else. Don't go back on the run queue */
288 void pthread__block(pthread_t self, pthread_spin_t* queuelock);
289 /* Put a thread back on the suspended queue */
290 void pthread__suspend(pthread_t self, pthread_t thread);
291 /* Put a thread back on the run queue */
292 void pthread__sched(pthread_t self, pthread_t thread);
293 void pthread__sched_sleepers(pthread_t self, struct pthread_queue_t *threadq);
294 void pthread__sched_idle(pthread_t self, pthread_t thread);
295 void pthread__sched_idle2(pthread_t self);
296
297 void pthread__sched_bulk(pthread_t self, pthread_t qhead);
298
299 void pthread__idle(void);
300
301 /* Get the next thread */
302 pthread_t pthread__next(pthread_t self);
303
304 #ifndef PTHREAD_SA
305 void pthread__unpark_all(pthread_t self, pthread_spin_t *lock,
306 void *obj, struct pthread_queue_t *threadq);
307 void pthread__unpark(pthread_t self, pthread_spin_t *lock,
308 void *obj, pthread_t target);
309 int pthread__park(pthread_t self, pthread_spin_t *lock,
310 void *obj, struct pthread_queue_t *threadq,
311 const struct timespec *abs_timeout, int tail,
312 int cancelpt);
313 #endif
314
315 int pthread__stackalloc(pthread_t *t);
316 void pthread__initmain(pthread_t *t);
317
318 void pthread__sa_start(void);
319 void pthread__sa_recycle(pthread_t old, pthread_t new);
320 void pthread__setconcurrency(int);
321
322 /* Alarm code */
323 void pthread__alarm_init(void);
324 void pthread__alarm_add(pthread_t, struct pt_alarm_t *,
325 const struct timespec *, void (*)(void *), void *);
326 void pthread__alarm_del(pthread_t, struct pt_alarm_t *);
327 int pthread__alarm_fired(struct pt_alarm_t *);
328 void pthread__alarm_process(pthread_t self, void *arg);
329
330 /* Internal locking primitives */
331 void pthread__lockprim_init(int ncpu);
332 void pthread_lockinit(pthread_spin_t *lock);
333 void pthread_spinlock(pthread_t thread, pthread_spin_t *lock);
334 int pthread_spintrylock(pthread_t thread, pthread_spin_t *lock);
335 void pthread_spinunlock(pthread_t thread, pthread_spin_t *lock);
336
337 extern const struct pthread_lock_ops *pthread__lock_ops;
338
339 #define pthread__simple_lock_init(alp) (*pthread__lock_ops->plo_init)(alp)
340 #define pthread__simple_lock_try(alp) (*pthread__lock_ops->plo_try)(alp)
341 #define pthread__simple_unlock(alp) (*pthread__lock_ops->plo_unlock)(alp)
342
343 #ifndef _getcontext_u
344 int _getcontext_u(ucontext_t *);
345 #endif
346 #ifndef _setcontext_u
347 int _setcontext_u(const ucontext_t *);
348 #endif
349 #ifndef _swapcontext_u
350 int _swapcontext_u(ucontext_t *, const ucontext_t *);
351 #endif
352
353 void pthread__testcancel(pthread_t self);
354 int pthread__find(pthread_t self, pthread_t target);
355
356 #ifndef PTHREAD_MD_INIT
357 #define PTHREAD_MD_INIT
358 #endif
359
360 #ifndef _INITCONTEXT_U_MD
361 #define _INITCONTEXT_U_MD(ucp)
362 #endif
363
364 #define _INITCONTEXT_U(ucp) do { \
365 (ucp)->uc_flags = _UC_CPU | _UC_STACK; \
366 _INITCONTEXT_U_MD(ucp) \
367 } while (/*CONSTCOND*/0)
368
369 #ifdef PTHREAD_MACHINE_HAS_ID_REGISTER
370 #define pthread__id(reg) (reg)
371 #else
372 /* Stack location of pointer to a particular thread */
373 #define pthread__id(sp) \
374 ((pthread_t) (((vaddr_t)(sp)) & ~PT_STACKMASK))
375
376 #define pthread__id_reg() pthread__sp()
377 #endif
378
379 #define pthread__self() (pthread__id(pthread__id_reg()))
380
381 #define pthread__abort() \
382 pthread__assertfunc(__FILE__, __LINE__, __func__, "unreachable")
383
384 #define pthread__assert(e) do { \
385 if (__predict_false(!(e))) \
386 pthread__assertfunc(__FILE__, __LINE__, __func__, #e); \
387 } while (/*CONSTCOND*/0)
388
389 #define pthread__error(err, msg, e) do { \
390 if (__predict_false(!(e))) { \
391 pthread__errorfunc(__FILE__, __LINE__, __func__, msg); \
392 return (err); \
393 } \
394 } while (/*CONSTCOND*/0)
395
396
397
398 /* These three routines are defined in processor-specific code. */
399 void pthread__upcall_switch(pthread_t self, pthread_t next);
400 void pthread__switch(pthread_t self, pthread_t next);
401 void pthread__locked_switch(pthread_t self, pthread_t next,
402 pthread_spin_t *lock);
403
404 void pthread__signal_init(void);
405 void pthread__signal_start(void);
406
407 void pthread__signal(pthread_t self, pthread_t t, siginfo_t *si);
408 void pthread__deliver_signal(pthread_t self, pthread_t t, siginfo_t *si);
409 void pthread__signal_deferred(pthread_t self, pthread_t t);
410
411 void pthread__destroy_tsd(pthread_t self);
412 void pthread__assertfunc(const char *file, int line, const char *function,
413 const char *expr);
414 void pthread__errorfunc(const char *file, int line, const char *function,
415 const char *msg);
416
417 #endif /* _LIB_PTHREAD_INT_H */
418