xref: /netbsd-src/lib/libpthread/pthread_int.h (revision d710132b4b8ce7f7cccaaf660cb16aa16b4077a0)
1 /*	$NetBSD: pthread_int.h,v 1.13 2003/06/06 21:06:07 nathanw Exp $	*/
2 
3 /*-
4  * Copyright (c) 2001 The NetBSD Foundation, Inc.
5  * All rights reserved.
6  *
7  * This code is derived from software contributed to The NetBSD Foundation
8  * by Nathan J. Williams.
9  *
10  * Redistribution and use in source and binary forms, with or without
11  * modification, are permitted provided that the following conditions
12  * are met:
13  * 1. Redistributions of source code must retain the above copyright
14  *    notice, this list of conditions and the following disclaimer.
15  * 2. Redistributions in binary form must reproduce the above copyright
16  *    notice, this list of conditions and the following disclaimer in the
17  *    documentation and/or other materials provided with the distribution.
18  * 3. All advertising materials mentioning features or use of this software
19  *    must display the following acknowledgement:
20  *        This product includes software developed by the NetBSD
21  *        Foundation, Inc. and its contributors.
22  * 4. Neither the name of The NetBSD Foundation nor the names of its
23  *    contributors may be used to endorse or promote products derived
24  *    from this software without specific prior written permission.
25  *
26  * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
27  * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
28  * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
29  * PURPOSE ARE DISCLAIMED.  IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
30  * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
31  * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
32  * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
33  * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
34  * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
35  * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
36  * POSSIBILITY OF SUCH DAMAGE.
37  */
38 
39 #ifndef _LIB_PTHREAD_INT_H
40 #define _LIB_PTHREAD_INT_H
41 
42 #define PTHREAD__DEBUG
43 #define ERRORCHECK
44 
45 #include "pthread_types.h"
46 #include "pthread_queue.h"
47 #include "pthread_debug.h"
48 #include "pthread_md.h"
49 
50 #include <sa.h>
51 #include <signal.h>
52 
53 /*
54  * The size of this structure needs to be no larger than struct
55  * __pthread_cleanup_store, defined in pthread.h.
56  */
57 struct pt_clean_t {
58 	PTQ_ENTRY(pt_clean_t)	ptc_next;
59 	void	(*ptc_cleanup)(void *);
60 	void	*ptc_arg;
61 };
62 
63 struct pt_alarm_t {
64 	PTQ_ENTRY(pt_alarm_t)	pta_next;
65 	pthread_spin_t	pta_lock;
66 	const struct timespec	*pta_time;
67 	void	(*pta_func)(void *);
68 	void	*pta_arg;
69 	int	pta_fired;
70 };
71 
72 struct	pthread_st {
73 	unsigned int	pt_magic;
74 	/* Identifier, for debugging and for preventing recycling. */
75 	int		pt_num;
76 
77 	int	pt_type;	/* normal, upcall, or idle */
78 	int	pt_state;	/* running, blocked, etc. */
79 	pthread_spin_t pt_statelock;	/* lock on pt_state */
80 	int	pt_flags;	/* see PT_FLAG_* below */
81 	int	pt_cancel;	/* Deferred cancellation */
82 	int	pt_spinlocks;	/* Number of spinlocks held. */
83 	int	pt_blockedlwp;	/* LWP/SA number when blocked */
84 
85 	int	pt_errno;	/* Thread-specific errno. */
86 
87 	/* Entry on the run queue */
88 	PTQ_ENTRY(pthread_st)	pt_runq;
89 	/* Entry on the list of all threads */
90 	PTQ_ENTRY(pthread_st)	pt_allq;
91 	/* Entry on the sleep queue (xxx should be same as run queue?) */
92 	PTQ_ENTRY(pthread_st)	pt_sleep;
93 	/* Object we're sleeping on */
94 	void			*pt_sleepobj;
95 	/* Queue we're sleeping on */
96 	struct pthread_queue_t	*pt_sleepq;
97 	/* Lock protecting that queue */
98 	pthread_spin_t		*pt_sleeplock;
99 
100 	stack_t		pt_stack;	/* Our stack */
101 	ucontext_t	*pt_uc;		/* Saved context when we're stopped */
102 
103 	sigset_t	pt_sigmask;	/* Signals we won't take. */
104 	sigset_t	pt_siglist;	/* Signals pending for us. */
105 	sigset_t	pt_sigblocked;	/* Signals delivered while blocked. */
106 	sigset_t	*pt_sigwait;	/* Signals waited for in sigwait */
107 	siginfo_t	*pt_wsig;
108 	pthread_spin_t	pt_siglock;	/* Lock on above */
109 
110 	void *		pt_exitval;	/* Read by pthread_join() */
111 
112 	/* Stack of cancellation cleanup handlers and their arguments */
113 	PTQ_HEAD(, pt_clean_t)	pt_cleanup_stack;
114 
115 	/* Thread's name, set by the application. */
116 	char*		pt_name;
117 
118 	/* Other threads trying to pthread_join() us. */
119 	struct pthread_queue_t	pt_joiners;
120 	/* Lock for above, and for changing pt_state to ZOMBIE or DEAD,
121 	 * and for setting the DETACHED flag.  Also protects pt_name.
122 	 */
123 	pthread_spin_t	pt_join_lock;
124 
125 	/* Thread we were going to switch to before we were preempted
126 	 * ourselves. Will be used by the upcall that's continuing us.
127 	 */
128 	pthread_t	pt_switchto;
129 	ucontext_t*	pt_switchtouc;
130 
131 	/* The context we saved in pthread__locked_switch but which
132 	 * was trashed when we were preempted before switching stacks.
133 	 */
134 	ucontext_t*	pt_sleepuc;
135 
136 	/* Threads that are preempted with spinlocks held will be
137 	 * continued until they unlock their spinlock. When they do
138 	 * so, they should jump ship to the thread pointed to by
139 	 * pt_next.
140 	 */
141 	pthread_t	pt_next;
142 
143 	/* The upcall that is continuing this thread */
144 	pthread_t	pt_parent;
145 
146 	/* A queue lock that this thread held while trying to
147 	 * context switch to another process.
148 	 */
149 	pthread_spin_t*	pt_heldlock;
150 
151 	/* Thread-specific data */
152 	void*		pt_specific[PTHREAD_KEYS_MAX];
153 
154 #ifdef PTHREAD__DEBUG
155 	int	blocks;
156 	int	preempts;
157 	int	rescheds;
158 #endif
159 };
160 
161 struct pthread_lock_ops {
162 	void	(*plo_init)(__cpu_simple_lock_t *);
163 	int	(*plo_try)(__cpu_simple_lock_t *);
164 	void	(*plo_unlock)(__cpu_simple_lock_t *);
165 };
166 
167 /* Thread types */
168 #define PT_THREAD_NORMAL	1
169 #define PT_THREAD_UPCALL	2
170 #define PT_THREAD_IDLE		3
171 
172 /* Thread states */
173 #define PT_STATE_RUNNING	1
174 #define PT_STATE_RUNNABLE	2
175 #define PT_STATE_BLOCKED_SYS	3
176 #define PT_STATE_BLOCKED_QUEUE	4
177 #define PT_STATE_ZOMBIE		5
178 #define PT_STATE_DEAD		6
179 #define PT_STATE_RECYCLABLE	7
180 
181 /* Flag values */
182 
183 #define PT_FLAG_DETACHED	0x0001
184 #define PT_FLAG_IDLED		0x0002
185 #define PT_FLAG_CS_DISABLED	0x0004	/* Cancellation disabled */
186 #define PT_FLAG_CS_ASYNC	0x0008  /* Cancellation is async */
187 #define PT_FLAG_CS_PENDING	0x0010
188 #define PT_FLAG_SIGDEFERRED     0x0020	/* There are signals to take */
189 
190 #define PT_MAGIC	0x11110001
191 #define PT_DEAD		0xDEAD0001
192 
193 #define PT_ATTR_MAGIC	0x22220002
194 #define PT_ATTR_DEAD	0xDEAD0002
195 
196 #define PT_STACKSIZE	(1<<18)
197 #define PT_STACKMASK	(PT_STACKSIZE-1)
198 
199 #define PT_UPCALLSTACKS	16
200 
201 #define PT_ALARMTIMER_MAGIC	0x88880010
202 #define PT_RRTIMER_MAGIC	0x88880020
203 #define NIDLETHREADS	4
204 #define IDLESPINS	1000
205 
206 /* Flag to be used in a ucontext_t's uc_flags indicating that
207  * the saved register state is "user" state only, not full
208  * trap state.
209  */
210 #define _UC_USER_BIT		30
211 #define _UC_USER		(1LU << _UC_USER_BIT)
212 
213 void	pthread_init(void)  __attribute__ ((__constructor__));
214 
215 /* Utility functions */
216 
217 /* Set up/clean up a thread's basic state. */
218 void	pthread__initthread(pthread_t self, pthread_t t);
219 
220 /* Go do something else. Don't go back on the run queue */
221 void	pthread__block(pthread_t self, pthread_spin_t* queuelock);
222 /* Put a thread back on the run queue */
223 void	pthread__sched(pthread_t self, pthread_t thread);
224 void	pthread__sched_sleepers(pthread_t self, struct pthread_queue_t *threadq);
225 void	pthread__sched_idle(pthread_t self, pthread_t thread);
226 void	pthread__sched_idle2(pthread_t self);
227 
228 void	pthread__sched_bulk(pthread_t self, pthread_t qhead);
229 
230 void	pthread__idle(void);
231 
232 /* Get the next thread */
233 pthread_t pthread__next(pthread_t self);
234 
235 int	pthread__stackalloc(pthread_t *t);
236 void	pthread__initmain(pthread_t *t);
237 
238 void	pthread__sa_start(void);
239 void	pthread__sa_recycle(pthread_t old, pthread_t new);
240 
241 /* Alarm code */
242 void	pthread__alarm_init(void);
243 void	pthread__alarm_add(pthread_t, struct pt_alarm_t *,
244     const struct timespec *, void (*)(void *), void *);
245 void	pthread__alarm_del(pthread_t, struct pt_alarm_t *);
246 int	pthread__alarm_fired(struct pt_alarm_t *);
247 void	pthread__alarm_process(pthread_t self, void *arg);
248 
249 /* Internal locking primitives */
250 void	pthread__lockprim_init(void);
251 void	pthread_lockinit(pthread_spin_t *lock);
252 void	pthread_spinlock(pthread_t thread, pthread_spin_t *lock);
253 int	pthread_spintrylock(pthread_t thread, pthread_spin_t *lock);
254 void	pthread_spinunlock(pthread_t thread, pthread_spin_t *lock);
255 
256 extern const struct pthread_lock_ops *pthread__lock_ops;
257 
258 #define	pthread__simple_lock_init(alp)	(*pthread__lock_ops->plo_init)(alp)
259 #define	pthread__simple_lock_try(alp)	(*pthread__lock_ops->plo_try)(alp)
260 #define	pthread__simple_unlock(alp)	(*pthread__lock_ops->plo_unlock)(alp)
261 
262 #ifndef _getcontext_u
263 int	_getcontext_u(ucontext_t *);
264 #endif
265 #ifndef _setcontext_u
266 int	_setcontext_u(const ucontext_t *);
267 #endif
268 #ifndef _swapcontext_u
269 int	_swapcontext_u(ucontext_t *, const ucontext_t *);
270 #endif
271 
272 void	pthread__testcancel(pthread_t self);
273 int	pthread__find(pthread_t self, pthread_t target);
274 
275 #ifndef PTHREAD_MD_INIT
276 #define PTHREAD_MD_INIT
277 #endif
278 
279 #ifndef _INITCONTEXT_U_MD
280 #define _INITCONTEXT_U_MD(ucp)
281 #endif
282 
283 #define _INITCONTEXT_U(ucp) do {					\
284 	(ucp)->uc_flags = _UC_CPU | _UC_STACK;				\
285 	_INITCONTEXT_U_MD(ucp)						\
286 	} while (/*CONSTCOND*/0)
287 
288 #ifdef __PTHREAD_SIGNAL_PRIVATE
289 
290 /*
291  * Macros for converting from ucontext to sigcontext and vice-versa.
292  * Note that going from sigcontext->ucontext is only safe for a
293  * sigcontext that was first created from a ucontext.
294  *
295  * Arch-specific code can override this, if necessary.  It may also
296  * be necessary for arch-specific code to include extra info along with
297  * the sigcontext.
298  */
299 #ifndef PTHREAD_SIGCONTEXT_EXTRA
300 #define	PTHREAD_SIGCONTEXT_EXTRA
301 #endif
302 
303 struct pthread__sigcontext {
304 	struct sigcontext	psc_context;
305 	PTHREAD_SIGCONTEXT_EXTRA
306 };
307 
308 #ifndef PTHREAD_UCONTEXT_TO_SIGCONTEXT
309 #define	PTHREAD_UCONTEXT_TO_SIGCONTEXT(mask, uc, psc)			\
310 do {									\
311 	(uc)->uc_sigmask = *(mask);					\
312 	/*								\
313 	 * XXX We may want to check for _UC_USER here and do a		\
314 	 * XXX _INITCONTEXT_U_MD() and clearing _UC_USER on such	\
315 	 * XXX contexts before converting to a signcontext, thus	\
316 	 * XXX allowing signal handlers to modify the non-_UC_USER	\
317 	 * XXX registers.  Hazy territory; ignore it for now.		\
318 	 */								\
319 	_UCONTEXT_TO_SIGCONTEXT((uc), &(psc)->psc_context);		\
320 } while (/*CONSTCOND*/0)
321 
322 #define	PTHREAD_SIGCONTEXT_TO_UCONTEXT(psc, uc)				\
323 do {									\
324 	_SIGCONTEXT_TO_UCONTEXT(&(psc)->psc_context, (uc));		\
325 	(uc)->uc_flags &= ~_UC_SIGMASK;					\
326 } while (/*CONSTCOND*/0)
327 #else
328 void	pthread__ucontext_to_sigcontext(const sigset_t *, ucontext_t *,
329 	    struct pthread__sigcontext *);
330 void	pthread__sigcontext_to_ucontext(const struct pthread__sigcontext *,
331 	    ucontext_t *);
332 #endif /* PTHREAD_UCONTEXT_TO_SIGCONTEXT */
333 
334 #endif /* __PTHREAD_SIGNAL_PRIVATE */
335 
336 #ifdef PTHREAD_MACHINE_HAS_ID_REGISTER
337 #define pthread__id(reg) (reg)
338 #else
339 /* Stack location of pointer to a particular thread */
340 #define pthread__id(sp) \
341 	((pthread_t) (((vaddr_t)(sp)) & ~PT_STACKMASK))
342 
343 #define pthread__id_reg() pthread__sp()
344 #endif
345 
346 #define pthread__self() (pthread__id(pthread__id_reg()))
347 
348 #define pthread__abort()						\
349 	pthread__assertfunc(__FILE__, __LINE__, __func__, "unreachable")
350 
351 #define pthread__assert(e) do {						\
352 	if (__predict_false(!(e)))					\
353        	       pthread__assertfunc(__FILE__, __LINE__, __func__, #e);	\
354         } while (/*CONSTCOND*/0)
355 
356 #define pthread__error(err, msg, e) do {				\
357 	if (__predict_false(!(e))) {					\
358        	       pthread__errorfunc(__FILE__, __LINE__, __func__, msg);	\
359 	       return (err);						\
360 	} 								\
361         } while (/*CONSTCOND*/0)
362 
363 
364 
365 /* These three routines are defined in processor-specific code. */
366 void	pthread__upcall_switch(pthread_t self, pthread_t next);
367 void	pthread__switch(pthread_t self, pthread_t next);
368 void	pthread__locked_switch(pthread_t self, pthread_t next,
369     pthread_spin_t *lock);
370 
371 void	pthread__signal_init(void);
372 
373 void	pthread__signal(pthread_t self, pthread_t t, int sig, int code);
374 void	pthread__deliver_signal(pthread_t self, pthread_t t, int sig, int code);
375 void	pthread__signal_deferred(pthread_t self, pthread_t t);
376 
377 void	pthread__destroy_tsd(pthread_t self);
378 void	pthread__assertfunc(char *file, int line, char *function, char *expr);
379 void	pthread__errorfunc(char *file, int line, char *function, char *msg);
380 
381 #endif /* _LIB_PTHREAD_INT_H */
382