pthread_int.h revision 1.18 1 /* $NetBSD: pthread_int.h,v 1.18 2003/09/07 14:47:44 cl Exp $ */
2
3 /*-
4 * Copyright (c) 2001,2002,2003 The NetBSD Foundation, Inc.
5 * All rights reserved.
6 *
7 * This code is derived from software contributed to The NetBSD Foundation
8 * by Nathan J. Williams.
9 *
10 * Redistribution and use in source and binary forms, with or without
11 * modification, are permitted provided that the following conditions
12 * are met:
13 * 1. Redistributions of source code must retain the above copyright
14 * notice, this list of conditions and the following disclaimer.
15 * 2. Redistributions in binary form must reproduce the above copyright
16 * notice, this list of conditions and the following disclaimer in the
17 * documentation and/or other materials provided with the distribution.
18 * 3. All advertising materials mentioning features or use of this software
19 * must display the following acknowledgement:
20 * This product includes software developed by the NetBSD
21 * Foundation, Inc. and its contributors.
22 * 4. Neither the name of The NetBSD Foundation nor the names of its
23 * contributors may be used to endorse or promote products derived
24 * from this software without specific prior written permission.
25 *
26 * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
27 * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
28 * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
29 * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
30 * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
31 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
32 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
33 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
34 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
35 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
36 * POSSIBILITY OF SUCH DAMAGE.
37 */
38
39 #ifndef _LIB_PTHREAD_INT_H
40 #define _LIB_PTHREAD_INT_H
41
42 #define PTHREAD__DEBUG
43 #define ERRORCHECK
44
45 #include "pthread_types.h"
46 #include "pthread_queue.h"
47 #include "pthread_debug.h"
48 #include "pthread_md.h"
49
50 #include <sa.h>
51 #include <signal.h>
52
53 #define PTHREAD_KEYS_MAX 256
54 /*
55 * The size of this structure needs to be no larger than struct
56 * __pthread_cleanup_store, defined in pthread.h.
57 */
58 struct pt_clean_t {
59 PTQ_ENTRY(pt_clean_t) ptc_next;
60 void (*ptc_cleanup)(void *);
61 void *ptc_arg;
62 };
63
64 struct pt_alarm_t {
65 PTQ_ENTRY(pt_alarm_t) pta_next;
66 pthread_spin_t pta_lock;
67 const struct timespec *pta_time;
68 void (*pta_func)(void *);
69 void *pta_arg;
70 int pta_fired;
71 };
72
73 /* Private data for pthread_attr_t */
74 struct pthread_attr_private {
75 char ptap_name[PTHREAD_MAX_NAMELEN_NP];
76 void *ptap_namearg;
77 void *ptap_stackaddr;
78 size_t ptap_stacksize;
79 size_t ptap_guardsize;
80 };
81
82 struct __pthread_st {
83 unsigned int pt_magic;
84 /* Identifier, for debugging and for preventing recycling. */
85 int pt_num;
86
87 int pt_type; /* normal, upcall, or idle */
88 int pt_state; /* running, blocked, etc. */
89 pthread_spin_t pt_statelock; /* lock on pt_state */
90 int pt_flags; /* see PT_FLAG_* below */
91 pthread_spin_t pt_flaglock; /* lock on pt_flag */
92 int pt_cancel; /* Deferred cancellation */
93 int pt_spinlocks; /* Number of spinlocks held. */
94 int pt_blockedlwp; /* LWP/SA number when blocked */
95
96 int pt_errno; /* Thread-specific errno. */
97
98 /* Entry on the run queue */
99 PTQ_ENTRY(__pthread_st) pt_runq;
100 /* Entry on the list of all threads */
101 PTQ_ENTRY(__pthread_st) pt_allq;
102 /* Entry on the sleep queue (xxx should be same as run queue?) */
103 PTQ_ENTRY(__pthread_st) pt_sleep;
104 /* Object we're sleeping on */
105 void *pt_sleepobj;
106 /* Queue we're sleeping on */
107 struct pthread_queue_t *pt_sleepq;
108 /* Lock protecting that queue */
109 pthread_spin_t *pt_sleeplock;
110
111 stack_t pt_stack; /* Our stack */
112 ucontext_t *pt_uc; /* Saved context when we're stopped */
113 ucontext_t *pt_trapuc; /* Kernel-saved context */
114
115 sigset_t pt_sigmask; /* Signals we won't take. */
116 sigset_t pt_siglist; /* Signals pending for us. */
117 sigset_t pt_sigblocked; /* Signals delivered while blocked. */
118 sigset_t *pt_sigwait; /* Signals waited for in sigwait */
119 siginfo_t *pt_wsig;
120 pthread_spin_t pt_siglock; /* Lock on above */
121
122 void * pt_exitval; /* Read by pthread_join() */
123
124 /* Stack of cancellation cleanup handlers and their arguments */
125 PTQ_HEAD(, pt_clean_t) pt_cleanup_stack;
126
127 /* Thread's name, set by the application. */
128 char* pt_name;
129
130 /* Other threads trying to pthread_join() us. */
131 struct pthread_queue_t pt_joiners;
132 /* Lock for above, and for changing pt_state to ZOMBIE or DEAD,
133 * and for setting the DETACHED flag. Also protects pt_name.
134 */
135 pthread_spin_t pt_join_lock;
136
137 /* Thread we were going to switch to before we were preempted
138 * ourselves. Will be used by the upcall that's continuing us.
139 */
140 pthread_t pt_switchto;
141 ucontext_t* pt_switchtouc;
142
143 /* Threads that are preempted with spinlocks held will be
144 * continued until they unlock their spinlock. When they do
145 * so, they should jump ship to the thread pointed to by
146 * pt_next.
147 */
148 pthread_t pt_next;
149
150 /* The upcall that is continuing this thread */
151 pthread_t pt_parent;
152
153 /* A queue lock that this thread held while trying to
154 * context switch to another process.
155 */
156 pthread_spin_t* pt_heldlock;
157
158 /* Thread-specific data */
159 void* pt_specific[PTHREAD_KEYS_MAX];
160
161 #ifdef PTHREAD__DEBUG
162 int blocks;
163 int preempts;
164 int rescheds;
165 #endif
166 };
167
168 struct pthread_lock_ops {
169 void (*plo_init)(__cpu_simple_lock_t *);
170 int (*plo_try)(__cpu_simple_lock_t *);
171 void (*plo_unlock)(__cpu_simple_lock_t *);
172 };
173
174 /* Thread types */
175 #define PT_THREAD_NORMAL 1
176 #define PT_THREAD_UPCALL 2
177 #define PT_THREAD_IDLE 3
178
179 /* Thread states */
180 #define PT_STATE_RUNNING 1
181 #define PT_STATE_RUNNABLE 2
182 #define PT_STATE_BLOCKED_SYS 3
183 #define PT_STATE_BLOCKED_QUEUE 4
184 #define PT_STATE_ZOMBIE 5
185 #define PT_STATE_DEAD 6
186
187 /* Flag values */
188
189 #define PT_FLAG_DETACHED 0x0001
190 #define PT_FLAG_IDLED 0x0002
191 #define PT_FLAG_CS_DISABLED 0x0004 /* Cancellation disabled */
192 #define PT_FLAG_CS_ASYNC 0x0008 /* Cancellation is async */
193 #define PT_FLAG_CS_PENDING 0x0010
194 #define PT_FLAG_SIGDEFERRED 0x0020 /* There are signals to take */
195 #define PT_FLAG_SCOPE_SYSTEM 0x0040
196 #define PT_FLAG_EXPLICIT_SCHED 0x0080
197
198 #define PT_MAGIC 0x11110001
199 #define PT_DEAD 0xDEAD0001
200
201 #define PT_ATTR_MAGIC 0x22220002
202 #define PT_ATTR_DEAD 0xDEAD0002
203
204 #define PT_STACKSIZE_LG 18
205 #define PT_STACKSIZE (1<<(PT_STACKSIZE_LG))
206 #define PT_STACKMASK (PT_STACKSIZE-1)
207
208 #define PT_UPCALLSTACKS 16
209
210 #define PT_ALARMTIMER_MAGIC 0x88880010
211 #define PT_RRTIMER_MAGIC 0x88880020
212 #define NIDLETHREADS 4
213 #define IDLESPINS 1000
214
215 /* Flag to be used in a ucontext_t's uc_flags indicating that
216 * the saved register state is "user" state only, not full
217 * trap state.
218 */
219 #define _UC_USER_BIT 30
220 #define _UC_USER (1LU << _UC_USER_BIT)
221
222 void pthread_init(void) __attribute__ ((__constructor__));
223
224 /* Utility functions */
225
226 /* Set up/clean up a thread's basic state. */
227 void pthread__initthread(pthread_t self, pthread_t t);
228
229 /* Go do something else. Don't go back on the run queue */
230 void pthread__block(pthread_t self, pthread_spin_t* queuelock);
231 /* Put a thread back on the run queue */
232 void pthread__sched(pthread_t self, pthread_t thread);
233 void pthread__sched_sleepers(pthread_t self, struct pthread_queue_t *threadq);
234 void pthread__sched_idle(pthread_t self, pthread_t thread);
235 void pthread__sched_idle2(pthread_t self);
236
237 void pthread__sched_bulk(pthread_t self, pthread_t qhead);
238
239 void pthread__idle(void);
240
241 /* Get the next thread */
242 pthread_t pthread__next(pthread_t self);
243
244 int pthread__stackalloc(pthread_t *t);
245 void pthread__initmain(pthread_t *t);
246
247 void pthread__sa_start(void);
248 void pthread__sa_recycle(pthread_t old, pthread_t new);
249
250 /* Alarm code */
251 void pthread__alarm_init(void);
252 void pthread__alarm_add(pthread_t, struct pt_alarm_t *,
253 const struct timespec *, void (*)(void *), void *);
254 void pthread__alarm_del(pthread_t, struct pt_alarm_t *);
255 int pthread__alarm_fired(struct pt_alarm_t *);
256 void pthread__alarm_process(pthread_t self, void *arg);
257
258 /* Internal locking primitives */
259 void pthread__lockprim_init(void);
260 void pthread_lockinit(pthread_spin_t *lock);
261 void pthread_spinlock(pthread_t thread, pthread_spin_t *lock);
262 int pthread_spintrylock(pthread_t thread, pthread_spin_t *lock);
263 void pthread_spinunlock(pthread_t thread, pthread_spin_t *lock);
264
265 extern const struct pthread_lock_ops *pthread__lock_ops;
266
267 #define pthread__simple_lock_init(alp) (*pthread__lock_ops->plo_init)(alp)
268 #define pthread__simple_lock_try(alp) (*pthread__lock_ops->plo_try)(alp)
269 #define pthread__simple_unlock(alp) (*pthread__lock_ops->plo_unlock)(alp)
270
271 #ifndef _getcontext_u
272 int _getcontext_u(ucontext_t *);
273 #endif
274 #ifndef _setcontext_u
275 int _setcontext_u(const ucontext_t *);
276 #endif
277 #ifndef _swapcontext_u
278 int _swapcontext_u(ucontext_t *, const ucontext_t *);
279 #endif
280
281 void pthread__testcancel(pthread_t self);
282 int pthread__find(pthread_t self, pthread_t target);
283
284 #ifndef PTHREAD_MD_INIT
285 #define PTHREAD_MD_INIT
286 #endif
287
288 #ifndef _INITCONTEXT_U_MD
289 #define _INITCONTEXT_U_MD(ucp)
290 #endif
291
292 #define _INITCONTEXT_U(ucp) do { \
293 (ucp)->uc_flags = _UC_CPU | _UC_STACK; \
294 _INITCONTEXT_U_MD(ucp) \
295 } while (/*CONSTCOND*/0)
296
297 #ifdef __PTHREAD_SIGNAL_PRIVATE
298
299 /*
300 * Macros for converting from ucontext to sigcontext and vice-versa.
301 * Note that going from sigcontext->ucontext is only safe for a
302 * sigcontext that was first created from a ucontext.
303 *
304 * Arch-specific code can override this, if necessary. It may also
305 * be necessary for arch-specific code to include extra info along with
306 * the sigcontext.
307 */
308 #ifndef PTHREAD_SIGCONTEXT_EXTRA
309 #define PTHREAD_SIGCONTEXT_EXTRA
310 #endif
311
312 struct pthread__sigcontext {
313 struct sigcontext psc_context;
314 PTHREAD_SIGCONTEXT_EXTRA
315 };
316
317 #ifndef PTHREAD_UCONTEXT_TO_SIGCONTEXT
318 #define PTHREAD_UCONTEXT_TO_SIGCONTEXT(mask, uc, psc) \
319 do { \
320 (uc)->uc_sigmask = *(mask); \
321 /* \
322 * XXX We may want to check for _UC_USER here and do a \
323 * XXX _INITCONTEXT_U_MD() and clearing _UC_USER on such \
324 * XXX contexts before converting to a signcontext, thus \
325 * XXX allowing signal handlers to modify the non-_UC_USER \
326 * XXX registers. Hazy territory; ignore it for now. \
327 */ \
328 _UCONTEXT_TO_SIGCONTEXT((uc), &(psc)->psc_context); \
329 } while (/*CONSTCOND*/0)
330
331 #define PTHREAD_SIGCONTEXT_TO_UCONTEXT(psc, uc) \
332 do { \
333 _SIGCONTEXT_TO_UCONTEXT(&(psc)->psc_context, (uc)); \
334 (uc)->uc_flags &= ~_UC_SIGMASK; \
335 } while (/*CONSTCOND*/0)
336 #else
337 void pthread__ucontext_to_sigcontext(const sigset_t *, ucontext_t *,
338 struct pthread__sigcontext *);
339 void pthread__sigcontext_to_ucontext(const struct pthread__sigcontext *,
340 ucontext_t *);
341 #endif /* PTHREAD_UCONTEXT_TO_SIGCONTEXT */
342
343 #endif /* __PTHREAD_SIGNAL_PRIVATE */
344
345 #ifdef PTHREAD_MACHINE_HAS_ID_REGISTER
346 #define pthread__id(reg) (reg)
347 #else
348 /* Stack location of pointer to a particular thread */
349 #define pthread__id(sp) \
350 ((pthread_t) (((vaddr_t)(sp)) & ~PT_STACKMASK))
351
352 #define pthread__id_reg() pthread__sp()
353 #endif
354
355 #define pthread__self() (pthread__id(pthread__id_reg()))
356
357 #define pthread__abort() \
358 pthread__assertfunc(__FILE__, __LINE__, __func__, "unreachable")
359
360 #define pthread__assert(e) do { \
361 if (__predict_false(!(e))) \
362 pthread__assertfunc(__FILE__, __LINE__, __func__, #e); \
363 } while (/*CONSTCOND*/0)
364
365 #define pthread__error(err, msg, e) do { \
366 if (__predict_false(!(e))) { \
367 pthread__errorfunc(__FILE__, __LINE__, __func__, msg); \
368 return (err); \
369 } \
370 } while (/*CONSTCOND*/0)
371
372
373
374 /* These three routines are defined in processor-specific code. */
375 void pthread__upcall_switch(pthread_t self, pthread_t next);
376 void pthread__switch(pthread_t self, pthread_t next);
377 void pthread__locked_switch(pthread_t self, pthread_t next,
378 pthread_spin_t *lock);
379
380 void pthread__signal_init(void);
381
382 void pthread__signal(pthread_t self, pthread_t t, int sig, int code);
383 void pthread__deliver_signal(pthread_t self, pthread_t t, int sig, int code);
384 void pthread__signal_deferred(pthread_t self, pthread_t t);
385
386 void pthread__destroy_tsd(pthread_t self);
387 void pthread__assertfunc(char *file, int line, char *function, char *expr);
388 void pthread__errorfunc(char *file, int line, char *function, char *msg);
389
390 #endif /* _LIB_PTHREAD_INT_H */
391