pthread_int.h revision 1.35 1 /* $NetBSD: pthread_int.h,v 1.35 2006/12/23 05:14:47 ad Exp $ */
2
3 /*-
4 * Copyright (c) 2001,2002,2003 The NetBSD Foundation, Inc.
5 * All rights reserved.
6 *
7 * This code is derived from software contributed to The NetBSD Foundation
8 * by Nathan J. Williams.
9 *
10 * Redistribution and use in source and binary forms, with or without
11 * modification, are permitted provided that the following conditions
12 * are met:
13 * 1. Redistributions of source code must retain the above copyright
14 * notice, this list of conditions and the following disclaimer.
15 * 2. Redistributions in binary form must reproduce the above copyright
16 * notice, this list of conditions and the following disclaimer in the
17 * documentation and/or other materials provided with the distribution.
18 * 3. All advertising materials mentioning features or use of this software
19 * must display the following acknowledgement:
20 * This product includes software developed by the NetBSD
21 * Foundation, Inc. and its contributors.
22 * 4. Neither the name of The NetBSD Foundation nor the names of its
23 * contributors may be used to endorse or promote products derived
24 * from this software without specific prior written permission.
25 *
26 * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
27 * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
28 * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
29 * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
30 * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
31 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
32 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
33 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
34 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
35 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
36 * POSSIBILITY OF SUCH DAMAGE.
37 */
38
39 #ifndef _LIB_PTHREAD_INT_H
40 #define _LIB_PTHREAD_INT_H
41
42 #define PTHREAD__DEBUG
43 #define ERRORCHECK
44
45 #include "pthread_types.h"
46 #include "pthread_queue.h"
47 #include "pthread_debug.h"
48 #include "pthread_md.h"
49
50 #ifndef PTHREAD_SA
51 #include <lwp.h>
52 #endif
53
54 #include <sa.h>
55 #include <signal.h>
56
57 #define PTHREAD_KEYS_MAX 256
58 /*
59 * The size of this structure needs to be no larger than struct
60 * __pthread_cleanup_store, defined in pthread.h.
61 */
62 struct pt_clean_t {
63 PTQ_ENTRY(pt_clean_t) ptc_next;
64 void (*ptc_cleanup)(void *);
65 void *ptc_arg;
66 };
67
68 struct pt_alarm_t {
69 PTQ_ENTRY(pt_alarm_t) pta_next;
70 pthread_spin_t pta_lock;
71 const struct timespec *pta_time;
72 void (*pta_func)(void *);
73 void *pta_arg;
74 int pta_fired;
75 };
76
77 /* Private data for pthread_attr_t */
78 struct pthread_attr_private {
79 char ptap_name[PTHREAD_MAX_NAMELEN_NP];
80 void *ptap_namearg;
81 void *ptap_stackaddr;
82 size_t ptap_stacksize;
83 size_t ptap_guardsize;
84 };
85
86 struct __pthread_st {
87 unsigned int pt_magic;
88 /* Identifier, for debugging and for preventing recycling. */
89 int pt_num;
90
91 #ifdef PTHREAD_SA
92 int pt_type; /* normal, upcall, or idle */
93 #else
94 lwpid_t pt_lid; /* LWP ID */
95 #endif
96 int pt_state; /* running, blocked, etc. */
97 pthread_spin_t pt_statelock; /* lock on pt_state */
98 int pt_flags; /* see PT_FLAG_* below */
99 pthread_spin_t pt_flaglock; /* lock on pt_flag */
100 int pt_cancel; /* Deferred cancellation */
101 int pt_spinlocks; /* Number of spinlocks held. */
102 int pt_blockedlwp; /* LWP/SA number when blocked */
103 int pt_vpid; /* VP number */
104 int pt_blockgen; /* SA_UPCALL_BLOCKED counter */
105 int pt_unblockgen; /* SA_UPCALL_UNBLOCKED counter */
106
107 int pt_errno; /* Thread-specific errno. */
108
109 /* Entry on the run queue */
110 PTQ_ENTRY(__pthread_st) pt_runq;
111 /* Entry on the list of all threads */
112 PTQ_ENTRY(__pthread_st) pt_allq;
113 /* Entry on the sleep queue (xxx should be same as run queue?) */
114 PTQ_ENTRY(__pthread_st) pt_sleep;
115 /*
116 * Object we're sleeping on. For 1:1 threads (!SA), this is
117 * protected by the interlock on the object that the thread is
118 * sleeping on.
119 */
120 void *pt_sleepobj;
121 /* Queue we're sleeping on */
122 struct pthread_queue_t *pt_sleepq;
123 /* Lock protecting that queue */
124 pthread_spin_t *pt_sleeplock;
125
126 stack_t pt_stack; /* Our stack */
127 ucontext_t *pt_uc; /* Saved context when we're stopped */
128 ucontext_t *pt_trapuc; /* Kernel-saved context */
129 ucontext_t *__pt_blockuc; /* Kernel-saved context when blocked */
130
131 sigset_t pt_sigmask; /* Signals we won't take. */
132 sigset_t pt_siglist; /* Signals pending for us. */
133 sigset_t pt_sigblocked; /* Signals delivered while blocked. */
134 sigset_t *pt_sigwait; /* Signals waited for in sigwait */
135 siginfo_t *pt_wsig;
136 pthread_spin_t pt_siglock; /* Lock on above */
137
138 void * pt_exitval; /* Read by pthread_join() */
139
140 /* Stack of cancellation cleanup handlers and their arguments */
141 PTQ_HEAD(, pt_clean_t) pt_cleanup_stack;
142
143 /* Thread's name, set by the application. */
144 char* pt_name;
145
146 /* Other threads trying to pthread_join() us. */
147 struct pthread_queue_t pt_joiners;
148 /* Lock for above, and for changing pt_state to ZOMBIE or DEAD,
149 * and for setting the DETACHED flag. Also protects pt_name.
150 */
151 pthread_spin_t pt_join_lock;
152
153 /* Thread we were going to switch to before we were preempted
154 * ourselves. Will be used by the upcall that's continuing us.
155 */
156 pthread_t pt_switchto;
157 ucontext_t* pt_switchtouc;
158
159 /* Threads that are preempted with spinlocks held will be
160 * continued until they unlock their spinlock. When they do
161 * so, they should jump ship to the thread pointed to by
162 * pt_next.
163 */
164 pthread_t pt_next;
165
166 /* The upcall that is continuing this thread */
167 pthread_t pt_parent;
168
169 /* A queue lock that this thread held while trying to
170 * context switch to another process.
171 */
172 pthread_spin_t* pt_heldlock;
173
174 /* Upcall stack information shared between kernel and
175 * userland.
176 */
177 struct sa_stackinfo_t pt_stackinfo;
178
179 /* Thread-specific data */
180 void* pt_specific[PTHREAD_KEYS_MAX];
181
182 #ifdef PTHREAD__DEBUG
183 int blocks;
184 int preempts;
185 int rescheds;
186 #endif
187 };
188
189 struct pthread_lock_ops {
190 void (*plo_init)(__cpu_simple_lock_t *);
191 int (*plo_try)(__cpu_simple_lock_t *);
192 void (*plo_unlock)(__cpu_simple_lock_t *);
193 };
194
195 #ifdef PTHREAD_SA
196
197 /* Thread types */
198 #define PT_THREAD_NORMAL 1
199 #define PT_THREAD_UPCALL 2
200 #define PT_THREAD_IDLE 3
201
202 /* Thread states */
203 #define PT_STATE_RUNNING 1
204 #define PT_STATE_RUNNABLE 2
205 #define _PT_STATE_BLOCKED_SYS 3 /* Only used in libpthread_dbg */
206 #define PT_STATE_BLOCKED_QUEUE 4
207 #define PT_STATE_ZOMBIE 5
208 #define PT_STATE_DEAD 6
209 #define PT_STATE_SUSPENDED 7
210
211 #else /* PTHREAD_SA */
212
213 /* Thread states */
214 #define PT_STATE_RUNNING 1
215 #define PT_STATE_ZOMBIE 5
216 #define PT_STATE_DEAD 6
217
218 #endif /* PTHREAD_SA */
219
220 /* Flag values */
221
222 #define PT_FLAG_DETACHED 0x0001
223 #define PT_FLAG_IDLED 0x0002
224 #define PT_FLAG_CS_DISABLED 0x0004 /* Cancellation disabled */
225 #define PT_FLAG_CS_ASYNC 0x0008 /* Cancellation is async */
226 #define PT_FLAG_CS_PENDING 0x0010
227 #define PT_FLAG_SIGDEFERRED 0x0020 /* There are signals to take */
228 #define PT_FLAG_SCOPE_SYSTEM 0x0040
229 #define PT_FLAG_EXPLICIT_SCHED 0x0080
230 #define PT_FLAG_SUSPENDED 0x0100 /* In the suspended queue */
231 #define PT_FLAG_SIGNALED 0x0200
232
233 #define PT_MAGIC 0x11110001
234 #define PT_DEAD 0xDEAD0001
235
236 #define PT_ATTR_MAGIC 0x22220002
237 #define PT_ATTR_DEAD 0xDEAD0002
238
239 #ifdef PT_FIXEDSTACKSIZE_LG
240
241 #define PT_STACKSIZE_LG PT_FIXEDSTACKSIZE_LG
242 #define PT_STACKSIZE (1<<(PT_STACKSIZE_LG))
243 #define PT_STACKMASK (PT_STACKSIZE-1)
244
245 #else /* PT_FIXEDSTACKSIZE_LG */
246
247 extern int pthread_stacksize_lg;
248 extern size_t pthread_stacksize;
249 extern vaddr_t pthread_stackmask;
250
251 #define PT_STACKSIZE_LG pthread_stacksize_lg
252 #define PT_STACKSIZE pthread_stacksize
253 #define PT_STACKMASK pthread_stackmask
254
255 #endif /* PT_FIXEDSTACKSIZE_LG */
256
257
258 #define PT_UPCALLSTACKS 16
259
260 #define PT_ALARMTIMER_MAGIC 0x88880010
261 #define PT_RRTIMER_MAGIC 0x88880020
262 #define NIDLETHREADS 4
263
264 /* Flag to be used in a ucontext_t's uc_flags indicating that
265 * the saved register state is "user" state only, not full
266 * trap state.
267 */
268 #define _UC_USER_BIT 30
269 #define _UC_USER (1LU << _UC_USER_BIT)
270
271 void pthread_init(void) __attribute__ ((__constructor__));
272
273 /* Utility functions */
274
275 /* Set up/clean up a thread's basic state. */
276 void pthread__initthread(pthread_t self, pthread_t t);
277 /* Get offset from stack start to struct sa_stackinfo */
278 ssize_t pthread__stackinfo_offset(void);
279
280 /* Go do something else. Don't go back on the run queue */
281 void pthread__block(pthread_t self, pthread_spin_t* queuelock);
282 /* Put a thread back on the suspended queue */
283 void pthread__suspend(pthread_t self, pthread_t thread);
284 /* Put a thread back on the run queue */
285 void pthread__sched(pthread_t self, pthread_t thread);
286 void pthread__sched_sleepers(pthread_t self, struct pthread_queue_t *threadq);
287 void pthread__sched_idle(pthread_t self, pthread_t thread);
288 void pthread__sched_idle2(pthread_t self);
289
290 void pthread__sched_bulk(pthread_t self, pthread_t qhead);
291
292 void pthread__idle(void);
293
294 /* Get the next thread */
295 pthread_t pthread__next(pthread_t self);
296
297 #ifndef PTHREAD_SA
298 void pthread__unpark_all(pthread_t self, pthread_spin_t *lock,
299 void *obj, struct pthread_queue_t *threadq);
300 void pthread__unpark(pthread_t self, pthread_spin_t *lock,
301 void *obj, pthread_t target);
302 int pthread__park(pthread_t self, pthread_spin_t *lock,
303 void *obj, struct pthread_queue_t *threadq,
304 const struct timespec *abs_timeout, int tail);
305 #endif
306
307 int pthread__stackalloc(pthread_t *t);
308 void pthread__initmain(pthread_t *t);
309
310 void pthread__sa_start(void);
311 void pthread__sa_recycle(pthread_t old, pthread_t new);
312 void pthread__setconcurrency(int);
313
314 /* Alarm code */
315 void pthread__alarm_init(void);
316 void pthread__alarm_add(pthread_t, struct pt_alarm_t *,
317 const struct timespec *, void (*)(void *), void *);
318 void pthread__alarm_del(pthread_t, struct pt_alarm_t *);
319 int pthread__alarm_fired(struct pt_alarm_t *);
320 void pthread__alarm_process(pthread_t self, void *arg);
321
322 /* Internal locking primitives */
323 void pthread__lockprim_init(int ncpu);
324 void pthread_lockinit(pthread_spin_t *lock);
325 void pthread_spinlock(pthread_t thread, pthread_spin_t *lock);
326 int pthread_spintrylock(pthread_t thread, pthread_spin_t *lock);
327 void pthread_spinunlock(pthread_t thread, pthread_spin_t *lock);
328
329 extern const struct pthread_lock_ops *pthread__lock_ops;
330
331 #define pthread__simple_lock_init(alp) (*pthread__lock_ops->plo_init)(alp)
332 #define pthread__simple_lock_try(alp) (*pthread__lock_ops->plo_try)(alp)
333 #define pthread__simple_unlock(alp) (*pthread__lock_ops->plo_unlock)(alp)
334
335 #ifndef _getcontext_u
336 int _getcontext_u(ucontext_t *);
337 #endif
338 #ifndef _setcontext_u
339 int _setcontext_u(const ucontext_t *);
340 #endif
341 #ifndef _swapcontext_u
342 int _swapcontext_u(ucontext_t *, const ucontext_t *);
343 #endif
344
345 void pthread__testcancel(pthread_t self);
346 int pthread__find(pthread_t self, pthread_t target);
347
348 #ifndef PTHREAD_MD_INIT
349 #define PTHREAD_MD_INIT
350 #endif
351
352 #ifndef _INITCONTEXT_U_MD
353 #define _INITCONTEXT_U_MD(ucp)
354 #endif
355
356 #define _INITCONTEXT_U(ucp) do { \
357 (ucp)->uc_flags = _UC_CPU | _UC_STACK; \
358 _INITCONTEXT_U_MD(ucp) \
359 } while (/*CONSTCOND*/0)
360
361 #ifdef PTHREAD_MACHINE_HAS_ID_REGISTER
362 #define pthread__id(reg) (reg)
363 #else
364 /* Stack location of pointer to a particular thread */
365 #define pthread__id(sp) \
366 ((pthread_t) (((vaddr_t)(sp)) & ~PT_STACKMASK))
367
368 #define pthread__id_reg() pthread__sp()
369 #endif
370
371 #define pthread__self() (pthread__id(pthread__id_reg()))
372
373 #define pthread__abort() \
374 pthread__assertfunc(__FILE__, __LINE__, __func__, "unreachable")
375
376 #define pthread__assert(e) do { \
377 if (__predict_false(!(e))) \
378 pthread__assertfunc(__FILE__, __LINE__, __func__, #e); \
379 } while (/*CONSTCOND*/0)
380
381 #define pthread__error(err, msg, e) do { \
382 if (__predict_false(!(e))) { \
383 pthread__errorfunc(__FILE__, __LINE__, __func__, msg); \
384 return (err); \
385 } \
386 } while (/*CONSTCOND*/0)
387
388
389
390 /* These three routines are defined in processor-specific code. */
391 void pthread__upcall_switch(pthread_t self, pthread_t next);
392 void pthread__switch(pthread_t self, pthread_t next);
393 void pthread__locked_switch(pthread_t self, pthread_t next,
394 pthread_spin_t *lock);
395
396 void pthread__signal_init(void);
397 void pthread__signal_start(void);
398
399 void pthread__signal(pthread_t self, pthread_t t, siginfo_t *si);
400 void pthread__deliver_signal(pthread_t self, pthread_t t, siginfo_t *si);
401 void pthread__signal_deferred(pthread_t self, pthread_t t);
402
403 void pthread__destroy_tsd(pthread_t self);
404 void pthread__assertfunc(const char *file, int line, const char *function,
405 const char *expr);
406 void pthread__errorfunc(const char *file, int line, const char *function,
407 const char *msg);
408
409 #endif /* _LIB_PTHREAD_INT_H */
410