pthread_int.h revision 1.34.4.3 1 /* $NetBSD: pthread_int.h,v 1.34.4.3 2007/10/19 05:35:38 wrstuden Exp $ */
2
3 /*-
4 * Copyright (c) 2001,2002,2003 The NetBSD Foundation, Inc.
5 * All rights reserved.
6 *
7 * This code is derived from software contributed to The NetBSD Foundation
8 * by Nathan J. Williams.
9 *
10 * Redistribution and use in source and binary forms, with or without
11 * modification, are permitted provided that the following conditions
12 * are met:
13 * 1. Redistributions of source code must retain the above copyright
14 * notice, this list of conditions and the following disclaimer.
15 * 2. Redistributions in binary form must reproduce the above copyright
16 * notice, this list of conditions and the following disclaimer in the
17 * documentation and/or other materials provided with the distribution.
18 * 3. All advertising materials mentioning features or use of this software
19 * must display the following acknowledgement:
20 * This product includes software developed by the NetBSD
21 * Foundation, Inc. and its contributors.
22 * 4. Neither the name of The NetBSD Foundation nor the names of its
23 * contributors may be used to endorse or promote products derived
24 * from this software without specific prior written permission.
25 *
26 * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
27 * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
28 * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
29 * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
30 * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
31 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
32 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
33 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
34 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
35 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
36 * POSSIBILITY OF SUCH DAMAGE.
37 */
38
39 #ifndef _LIB_PTHREAD_INT_H
40 #define _LIB_PTHREAD_INT_H
41
42 #define PTHREAD__DEBUG
43 #define ERRORCHECK
44
45 #include "pthread_types.h"
46 #include "pthread_queue.h"
47 #include "pthread_debug.h"
48 #include "pthread_md.h"
49
50 #include <sa.h>
51 #include <signal.h>
52
53 #define PTHREAD_KEYS_MAX 256
54 /*
55 * The size of this structure needs to be no larger than struct
56 * __pthread_cleanup_store, defined in pthread.h.
57 */
58 struct pt_clean_t {
59 PTQ_ENTRY(pt_clean_t) ptc_next;
60 void (*ptc_cleanup)(void *);
61 void *ptc_arg;
62 };
63
64 struct pt_alarm_t {
65 PTQ_ENTRY(pt_alarm_t) pta_next;
66 pthread_spin_t pta_lock;
67 const struct timespec *pta_time;
68 void (*pta_func)(void *);
69 void *pta_arg;
70 int pta_fired;
71 };
72
73 /* Private data for pthread_attr_t */
74 struct pthread_attr_private {
75 char ptap_name[PTHREAD_MAX_NAMELEN_NP];
76 void *ptap_namearg;
77 void *ptap_stackaddr;
78 size_t ptap_stacksize;
79 size_t ptap_guardsize;
80 };
81
82 struct __pthread_st {
83 unsigned int pt_magic;
84 /* Identifier, for debugging and for preventing recycling. */
85 int pt_num;
86
87 int pt_type; /* normal, upcall, or idle */
88 int pt_state; /* running, blocked, etc. */
89 pthread_spin_t pt_statelock; /* lock on pt_state */
90 int pt_flags; /* see PT_FLAG_* below */
91 pthread_spin_t pt_flaglock; /* lock on pt_flag */
92 int pt_cancel; /* Deferred cancellation */
93 int pt_spinlocks; /* Number of spinlocks held. */
94 int pt_lastlwp; /* LWP/SA number when running or blocked */
95 int pt_vpid; /* VP number */
96 int pt_blockgen; /* SA_UPCALL_BLOCKED counter */
97 int pt_unblockgen; /* SA_UPCALL_UNBLOCKED counter */
98
99 int pt_errno; /* Thread-specific errno. */
100
101 /* Entry on the run queue */
102 PTQ_ENTRY(__pthread_st) pt_runq;
103 /* Entry on the list of all threads */
104 PTQ_ENTRY(__pthread_st) pt_allq;
105 /* Entry on the sleep queue (xxx should be same as run queue?) */
106 PTQ_ENTRY(__pthread_st) pt_sleep;
107 /* Object we're sleeping on */
108 void *pt_sleepobj;
109 /* Queue we're sleeping on */
110 struct pthread_queue_t *pt_sleepq;
111 /* Lock protecting that queue */
112 pthread_spin_t *pt_sleeplock;
113
114 stack_t pt_stack; /* Our stack */
115 ucontext_t *pt_uc; /* Saved context when we're stopped */
116 ucontext_t *pt_trapuc; /* Kernel-saved context */
117 ucontext_t *__pt_blockuc; /* Kernel-saved context when blocked */
118
119 sigset_t pt_sigmask; /* Signals we won't take. */
120 sigset_t pt_siglist; /* Signals pending for us. */
121 sigset_t pt_sigblocked; /* Signals delivered while blocked. */
122 sigset_t *pt_sigwait; /* Signals waited for in sigwait */
123 siginfo_t *pt_wsig;
124 pthread_spin_t pt_siglock; /* Lock on above */
125
126 void * pt_exitval; /* Read by pthread_join() */
127
128 /* Stack of cancellation cleanup handlers and their arguments */
129 PTQ_HEAD(, pt_clean_t) pt_cleanup_stack;
130
131 /* Thread's name, set by the application. */
132 char* pt_name;
133
134 /* Other threads trying to pthread_join() us. */
135 struct pthread_queue_t pt_joiners;
136 /* Lock for above, and for changing pt_state to ZOMBIE or DEAD,
137 * and for setting the DETACHED flag. Also protects pt_name.
138 */
139 pthread_spin_t pt_join_lock;
140
141 /* Thread we were going to switch to before we were preempted
142 * ourselves. Will be used by the upcall that's continuing us.
143 */
144 pthread_t pt_switchto;
145 ucontext_t* pt_switchtouc;
146
147 /* Threads that are preempted with spinlocks held will be
148 * continued until they unlock their spinlock. When they do
149 * so, they should jump ship to the thread pointed to by
150 * pt_next.
151 */
152 pthread_t pt_next;
153
154 /* The upcall that is continuing this thread */
155 pthread_t pt_parent;
156
157 /* A queue lock that this thread held while trying to
158 * context switch to another process.
159 */
160 pthread_spin_t* pt_heldlock;
161
162 /* Upcall stack information shared between kernel and
163 * userland.
164 */
165 struct sa_stackinfo_t pt_stackinfo;
166
167 /* Thread-specific data */
168 void* pt_specific[PTHREAD_KEYS_MAX];
169
170 #ifdef PTHREAD__DEBUG
171 int blocks;
172 int preempts;
173 int rescheds;
174 #endif
175 };
176
177 /*
178 * Thread locking hierarcy. In order to avoid deadlocks when concurrency
179 * is enabled, locks mus be aquired in a consistent order. Locks are
180 * divided into groups, and no lock in a "higher" group may be taken while
181 * holding a lock in a "lower" group. Also, only one lock in a group may
182 * be taken at once, as otherwise the locks need to be listed separately
183 * to avoid an issue between them. Some locks listed are per-thread
184 * and some are global. A thread may take some per-thread locks on itself
185 * and per-thread locks on other threads, but any such locking must still
186 * respect the hierarcy. All per-thread locks of a given type are considered
187 * in the same group and thus only one may be taken at once by a given thread.
188 * It is believed that the current code never tries to take the same lock
189 * in more than one thread at once.
190 *
191 * "Top" group: pt_join_lock, pthread__deadqueue_lock, &barrier->ptb_lock,
192 * cond->ptc_lock, mutex->ptm_interlock, rwlock->ptr_interlock,
193 * pt_sigsuspended_lock, pt_sigwaiting_lock, pt_nanosleep_lock,
194 * pthread__allqueue_lock
195 *
196 * Group 2: pt_statelock
197 *
198 * Group 3: pt_siglock, pthread_alarmqlock
199 *
200 * Group 4: pt_flag_lock, pthread__runqueue_lock
201 *
202 * Group 5: alarm->pta_lock -- all locked in acending time order.
203 */
204
205 struct pthread_lock_ops {
206 void (*plo_init)(__cpu_simple_lock_t *);
207 int (*plo_try)(__cpu_simple_lock_t *);
208 void (*plo_unlock)(__cpu_simple_lock_t *);
209 };
210
211 /* Thread types */
212 #define PT_THREAD_NORMAL 1
213 #define PT_THREAD_UPCALL 2
214 #define PT_THREAD_IDLE 3
215
216 /* Thread states */
217 #define PT_STATE_RUNNING 1
218 #define PT_STATE_RUNNABLE 2
219 #define _PT_STATE_BLOCKED_SYS 3 /* Only used in libpthread_dbg */
220 #define PT_STATE_BLOCKED_QUEUE 4
221 #define PT_STATE_ZOMBIE 5
222 #define PT_STATE_DEAD 6
223 #define PT_STATE_SUSPENDED 7
224
225 /* Flag values */
226
227 #define PT_FLAG_DETACHED 0x0001
228 #define PT_FLAG_IDLED 0x0002
229 #define PT_FLAG_CS_DISABLED 0x0004 /* Cancellation disabled */
230 #define PT_FLAG_CS_ASYNC 0x0008 /* Cancellation is async */
231 #define PT_FLAG_CS_PENDING 0x0010
232 #define PT_FLAG_SIGDEFERRED 0x0020 /* There are signals to take */
233 #define PT_FLAG_SCOPE_SYSTEM 0x0040
234 #define PT_FLAG_EXPLICIT_SCHED 0x0080
235 #define PT_FLAG_SUSPENDED 0x0100 /* In the suspended queue */
236 #define PT_FLAG_SIGNALED 0x0200
237
238 #define PT_MAGIC 0x11110001
239 #define PT_DEAD 0xDEAD0001
240
241 #define PT_ATTR_MAGIC 0x22220002
242 #define PT_ATTR_DEAD 0xDEAD0002
243
244 #ifdef PT_FIXEDSTACKSIZE_LG
245
246 #define PT_STACKSIZE_LG PT_FIXEDSTACKSIZE_LG
247 #define PT_STACKSIZE (1<<(PT_STACKSIZE_LG))
248 #define PT_STACKMASK (PT_STACKSIZE-1)
249
250 #else /* PT_FIXEDSTACKSIZE_LG */
251
252 extern int pthread_stacksize_lg;
253 extern size_t pthread_stacksize;
254 extern vaddr_t pthread_stackmask;
255
256 #define PT_STACKSIZE_LG pthread_stacksize_lg
257 #define PT_STACKSIZE pthread_stacksize
258 #define PT_STACKMASK pthread_stackmask
259
260 #endif /* PT_FIXEDSTACKSIZE_LG */
261
262
263 #define PT_UPCALLSTACKS 16
264
265 #define PT_ALARMTIMER_MAGIC 0x88880010
266 #define PT_RRTIMER_MAGIC 0x88880020
267 #define NIDLETHREADS 4
268
269 /* Flag to be used in a ucontext_t's uc_flags indicating that
270 * the saved register state is "user" state only, not full
271 * trap state.
272 */
273 #define _UC_USER_BIT 30
274 #define _UC_USER (1LU << _UC_USER_BIT)
275
276 void pthread_init(void) __attribute__ ((__constructor__));
277
278 /* Utility functions */
279
280 /* Set up/clean up a thread's basic state. */
281 void pthread__initthread(pthread_t self, pthread_t t);
282 /* Get offset from stack start to struct sa_stackinfo */
283 ssize_t pthread__stackinfo_offset(void);
284
285 /* Go do something else. Don't go back on the run queue */
286 void pthread__block(pthread_t self, pthread_spin_t* queuelock);
287 /* Put a thread back on the suspended queue */
288 void pthread__suspend(pthread_t self, pthread_t thread);
289 /* Put a thread back on the run queue */
290 void pthread__sched(pthread_t self, pthread_t thread);
291 void pthread__sched_sleepers(pthread_t self, struct pthread_queue_t *threadq);
292 void pthread__sched_idle(pthread_t self, pthread_t thread);
293 void pthread__sched_idle2(pthread_t self);
294
295 void pthread__sched_bulk(pthread_t self, pthread_t qhead);
296
297 void pthread__idle(void);
298
299 /* Get the next thread */
300 pthread_t pthread__next(pthread_t self);
301
302 int pthread__stackalloc(pthread_t *t);
303 void pthread__initmain(pthread_t *t);
304
305 void pthread__sa_start(void);
306 void pthread__sa_recycle(pthread_t old, pthread_t new);
307 void pthread__setconcurrency(int);
308
309 /* Alarm code */
310 void pthread__alarm_init(void);
311 void pthread__alarm_add(pthread_t, struct pt_alarm_t *,
312 const struct timespec *, void (*)(void *), void *);
313 void pthread__alarm_del(pthread_t, struct pt_alarm_t *);
314 int pthread__alarm_fired(struct pt_alarm_t *);
315 void pthread__alarm_process(pthread_t self, void *arg);
316
317 /* Internal locking primitives */
318 void pthread__lockprim_init(int ncpu);
319 void pthread_lockinit(pthread_spin_t *lock);
320 void pthread_spinlock(pthread_t thread, pthread_spin_t *lock);
321 int pthread_spintrylock(pthread_t thread, pthread_spin_t *lock);
322 void pthread_spinunlock(pthread_t thread, pthread_spin_t *lock);
323
324 extern const struct pthread_lock_ops *pthread__lock_ops;
325
326 #define pthread__simple_lock_init(alp) (*pthread__lock_ops->plo_init)(alp)
327 #define pthread__simple_lock_try(alp) (*pthread__lock_ops->plo_try)(alp)
328 #define pthread__simple_unlock(alp) (*pthread__lock_ops->plo_unlock)(alp)
329
330 #ifndef _getcontext_u
331 int _getcontext_u(ucontext_t *);
332 #endif
333 #ifndef _setcontext_u
334 int _setcontext_u(const ucontext_t *);
335 #endif
336 #ifndef _swapcontext_u
337 int _swapcontext_u(ucontext_t *, const ucontext_t *);
338 #endif
339
340 void pthread__testcancel(pthread_t self);
341 int pthread__find(pthread_t self, pthread_t target);
342
343 #ifndef PTHREAD_MD_INIT
344 #define PTHREAD_MD_INIT
345 #endif
346
347 #ifndef _INITCONTEXT_U_MD
348 #define _INITCONTEXT_U_MD(ucp)
349 #endif
350
351 #define _INITCONTEXT_U(ucp) do { \
352 (ucp)->uc_flags = _UC_CPU | _UC_STACK; \
353 _INITCONTEXT_U_MD(ucp) \
354 } while (/*CONSTCOND*/0)
355
356 #ifdef PTHREAD_MACHINE_HAS_ID_REGISTER
357 #define pthread__id(reg) (reg)
358 #else
359 /* Stack location of pointer to a particular thread */
360 #define pthread__id(sp) \
361 ((pthread_t) (((vaddr_t)(sp)) & ~PT_STACKMASK))
362
363 #define pthread__id_reg() pthread__sp()
364 #endif
365
366 #define pthread__self() (pthread__id(pthread__id_reg()))
367
368 #define pthread__abort() \
369 pthread__assertfunc(__FILE__, __LINE__, __func__, "unreachable")
370
371 #define pthread__assert(e) do { \
372 if (__predict_false(!(e))) \
373 pthread__assertfunc(__FILE__, __LINE__, __func__, #e); \
374 } while (/*CONSTCOND*/0)
375
376 #define pthread__error(err, msg, e) do { \
377 if (__predict_false(!(e))) { \
378 pthread__errorfunc(__FILE__, __LINE__, __func__, msg); \
379 return (err); \
380 } \
381 } while (/*CONSTCOND*/0)
382
383 /*
384 * You must hold t->pt_statelock when making this check.
385 */
386 #define pthread_check_defsig(t) \
387 __predict_false((t)->pt_flags & PT_FLAG_SIGDEFERRED)
388
389
390
391 /* These three routines are defined in processor-specific code. */
392 void pthread__upcall_switch(pthread_t self, pthread_t next);
393 void pthread__switch(pthread_t self, pthread_t next);
394 void pthread__locked_switch(pthread_t self, pthread_t next,
395 pthread_spin_t *lock);
396
397 void pthread__signal_init(void);
398 void pthread__signal_start(void);
399
400 void pthread__signal(pthread_t self, pthread_t t, siginfo_t *si);
401 void pthread__deliver_signal(pthread_t self, pthread_t t, siginfo_t *si);
402 void pthread__signal_deferred(pthread_t self, pthread_t t);
403
404 void pthread__destroy_tsd(pthread_t self);
405 void pthread__assertfunc(const char *file, int line, const char *function,
406 const char *expr);
407 void pthread__errorfunc(const char *file, int line, const char *function,
408 const char *msg);
409
410 #endif /* _LIB_PTHREAD_INT_H */
411