Home | History | Annotate | Line # | Download | only in kern
sys_sig.c revision 1.47.4.4
      1 /*	$NetBSD: sys_sig.c,v 1.47.4.4 2024/08/07 10:11:45 martin Exp $	*/
      2 
      3 /*-
      4  * Copyright (c) 2006, 2007, 2008 The NetBSD Foundation, Inc.
      5  * All rights reserved.
      6  *
      7  * This code is derived from software contributed to The NetBSD Foundation
      8  * by Andrew Doran.
      9  *
     10  * Redistribution and use in source and binary forms, with or without
     11  * modification, are permitted provided that the following conditions
     12  * are met:
     13  * 1. Redistributions of source code must retain the above copyright
     14  *    notice, this list of conditions and the following disclaimer.
     15  * 2. Redistributions in binary form must reproduce the above copyright
     16  *    notice, this list of conditions and the following disclaimer in the
     17  *    documentation and/or other materials provided with the distribution.
     18  *
     19  * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
     20  * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
     21  * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
     22  * PURPOSE ARE DISCLAIMED.  IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
     23  * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
     24  * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
     25  * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
     26  * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
     27  * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
     28  * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
     29  * POSSIBILITY OF SUCH DAMAGE.
     30  */
     31 
     32 /*
     33  * Copyright (c) 1982, 1986, 1989, 1991, 1993
     34  *	The Regents of the University of California.  All rights reserved.
     35  * (c) UNIX System Laboratories, Inc.
     36  * All or some portions of this file are derived from material licensed
     37  * to the University of California by American Telephone and Telegraph
     38  * Co. or Unix System Laboratories, Inc. and are reproduced herein with
     39  * the permission of UNIX System Laboratories, Inc.
     40  *
     41  * Redistribution and use in source and binary forms, with or without
     42  * modification, are permitted provided that the following conditions
     43  * are met:
     44  * 1. Redistributions of source code must retain the above copyright
     45  *    notice, this list of conditions and the following disclaimer.
     46  * 2. Redistributions in binary form must reproduce the above copyright
     47  *    notice, this list of conditions and the following disclaimer in the
     48  *    documentation and/or other materials provided with the distribution.
     49  * 3. Neither the name of the University nor the names of its contributors
     50  *    may be used to endorse or promote products derived from this software
     51  *    without specific prior written permission.
     52  *
     53  * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
     54  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
     55  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
     56  * ARE DISCLAIMED.  IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
     57  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
     58  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
     59  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
     60  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
     61  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
     62  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
     63  * SUCH DAMAGE.
     64  *
     65  *	@(#)kern_sig.c	8.14 (Berkeley) 5/14/95
     66  */
     67 
     68 #include <sys/cdefs.h>
     69 __KERNEL_RCSID(0, "$NetBSD: sys_sig.c,v 1.47.4.4 2024/08/07 10:11:45 martin Exp $");
     70 
     71 #include "opt_dtrace.h"
     72 
     73 #include <sys/param.h>
     74 #include <sys/kernel.h>
     75 #include <sys/signalvar.h>
     76 #include <sys/proc.h>
     77 #include <sys/pool.h>
     78 #include <sys/syscallargs.h>
     79 #include <sys/kauth.h>
     80 #include <sys/wait.h>
     81 #include <sys/kmem.h>
     82 #include <sys/module.h>
     83 #include <sys/sdt.h>
     84 #include <sys/compat_stub.h>
     85 
     86 SDT_PROVIDER_DECLARE(proc);
     87 SDT_PROBE_DEFINE2(proc, kernel, , signal__clear,
     88     "int", 		/* signal */
     89     "ksiginfo_t *");	/* signal-info */
     90 
     91 int
     92 sys___sigaction_sigtramp(struct lwp *l,
     93     const struct sys___sigaction_sigtramp_args *uap, register_t *retval)
     94 {
     95 	/* {
     96 		syscallarg(int)				signum;
     97 		syscallarg(const struct sigaction *)	nsa;
     98 		syscallarg(struct sigaction *)		osa;
     99 		syscallarg(void *)			tramp;
    100 		syscallarg(int)				vers;
    101 	} */
    102 	struct sigaction nsa, osa;
    103 	int error;
    104 
    105 	if (SCARG(uap, nsa)) {
    106 		error = copyin(SCARG(uap, nsa), &nsa, sizeof(nsa));
    107 		if (error)
    108 			return (error);
    109 	}
    110 	error = sigaction1(l, SCARG(uap, signum),
    111 	    SCARG(uap, nsa) ? &nsa : 0, SCARG(uap, osa) ? &osa : 0,
    112 	    SCARG(uap, tramp), SCARG(uap, vers));
    113 	if (error)
    114 		return (error);
    115 	if (SCARG(uap, osa)) {
    116 		error = copyout(&osa, SCARG(uap, osa), sizeof(osa));
    117 		if (error)
    118 			return (error);
    119 	}
    120 	return 0;
    121 }
    122 
    123 /*
    124  * Manipulate signal mask.  Note that we receive new mask, not pointer, and
    125  * return old mask as return value; the library stub does the rest.
    126  */
    127 int
    128 sys___sigprocmask14(struct lwp *l, const struct sys___sigprocmask14_args *uap,
    129     register_t *retval)
    130 {
    131 	/* {
    132 		syscallarg(int)			how;
    133 		syscallarg(const sigset_t *)	set;
    134 		syscallarg(sigset_t *)		oset;
    135 	} */
    136 	struct proc	*p = l->l_proc;
    137 	sigset_t	nss, oss;
    138 	int		error;
    139 
    140 	if (SCARG(uap, set)) {
    141 		error = copyin(SCARG(uap, set), &nss, sizeof(nss));
    142 		if (error)
    143 			return error;
    144 	}
    145 	mutex_enter(p->p_lock);
    146 	error = sigprocmask1(l, SCARG(uap, how),
    147 	    SCARG(uap, set) ? &nss : 0, SCARG(uap, oset) ? &oss : 0);
    148 	mutex_exit(p->p_lock);
    149 	if (error)
    150 		return error;
    151 	if (SCARG(uap, oset)) {
    152 		error = copyout(&oss, SCARG(uap, oset), sizeof(oss));
    153 		if (error)
    154 			return error;
    155 	}
    156 	return 0;
    157 }
    158 
    159 int
    160 sys___sigpending14(struct lwp *l, const struct sys___sigpending14_args *uap,
    161     register_t *retval)
    162 {
    163 	/* {
    164 		syscallarg(sigset_t *)	set;
    165 	} */
    166 	sigset_t ss;
    167 
    168 	sigpending1(l, &ss);
    169 	return copyout(&ss, SCARG(uap, set), sizeof(ss));
    170 }
    171 
    172 /*
    173  * Suspend process until signal, providing mask to be set in the meantime.
    174  * Note nonstandard calling convention: libc stub passes mask, not pointer,
    175  * to save a copyin.
    176  */
    177 int
    178 sys___sigsuspend14(struct lwp *l, const struct sys___sigsuspend14_args *uap,
    179     register_t *retval)
    180 {
    181 	/* {
    182 		syscallarg(const sigset_t *)	set;
    183 	} */
    184 	sigset_t	ss;
    185 	int		error;
    186 
    187 	if (SCARG(uap, set)) {
    188 		error = copyin(SCARG(uap, set), &ss, sizeof(ss));
    189 		if (error)
    190 			return error;
    191 	}
    192 	return sigsuspend1(l, SCARG(uap, set) ? &ss : 0);
    193 }
    194 
    195 int
    196 sys___sigaltstack14(struct lwp *l, const struct sys___sigaltstack14_args *uap,
    197     register_t *retval)
    198 {
    199 	/* {
    200 		syscallarg(const struct sigaltstack *)	nss;
    201 		syscallarg(struct sigaltstack *)	oss;
    202 	} */
    203 	struct sigaltstack	nss, oss;
    204 	int			error;
    205 
    206 	if (SCARG(uap, nss)) {
    207 		error = copyin(SCARG(uap, nss), &nss, sizeof(nss));
    208 		if (error)
    209 			return error;
    210 	}
    211 	error = sigaltstack1(l,
    212 	    SCARG(uap, nss) ? &nss : 0, SCARG(uap, oss) ? &oss : 0);
    213 	if (error)
    214 		return error;
    215 	if (SCARG(uap, oss)) {
    216 		error = copyout(&oss, SCARG(uap, oss), sizeof(oss));
    217 		if (error)
    218 			return error;
    219 	}
    220 	return 0;
    221 }
    222 
    223 int
    224 kill1(struct lwp *l, pid_t pid, ksiginfo_t *ksi, register_t *retval)
    225 {
    226 	int error;
    227 	struct proc *p;
    228 
    229 	if ((u_int)ksi->ksi_signo >= NSIG)
    230 		return EINVAL;
    231 
    232 	if (pid != l->l_proc->p_pid) {
    233 		if (ksi->ksi_pid != l->l_proc->p_pid)
    234 			return EPERM;
    235 
    236 		if (ksi->ksi_uid != kauth_cred_geteuid(l->l_cred))
    237 			return EPERM;
    238 
    239 		switch (ksi->ksi_code) {
    240 		case SI_USER:
    241 		case SI_QUEUE:
    242 			break;
    243 		default:
    244 			return EPERM;
    245 		}
    246 	}
    247 
    248 	if (pid > 0) {
    249 		/* kill single process */
    250 		mutex_enter(proc_lock);
    251 		p = proc_find_raw(pid);
    252 		if (p == NULL || (p->p_stat != SACTIVE && p->p_stat != SSTOP)) {
    253 			mutex_exit(proc_lock);
    254 			/* IEEE Std 1003.1-2001: return success for zombies */
    255 			return p ? 0 : ESRCH;
    256 		}
    257 		mutex_enter(p->p_lock);
    258 		error = kauth_authorize_process(l->l_cred,
    259 		    KAUTH_PROCESS_SIGNAL, p, KAUTH_ARG(ksi->ksi_signo),
    260 		    NULL, NULL);
    261 		if (!error && ksi->ksi_signo) {
    262 			error = kpsignal2(p, ksi);
    263 		}
    264 		mutex_exit(p->p_lock);
    265 		mutex_exit(proc_lock);
    266 		return error;
    267 	}
    268 
    269 	switch (pid) {
    270 	case -1:		/* broadcast signal */
    271 		return killpg1(l, ksi, 0, 1);
    272 	case 0:			/* signal own process group */
    273 		return killpg1(l, ksi, 0, 0);
    274 	default:		/* negative explicit process group */
    275 		if (pid <= INT_MIN)
    276 			return ESRCH;
    277 		return killpg1(l, ksi, -pid, 0);
    278 	}
    279 	/* NOTREACHED */
    280 }
    281 
    282 int
    283 sys_sigqueueinfo(struct lwp *l, const struct sys_sigqueueinfo_args *uap,
    284     register_t *retval)
    285 {
    286 	/* {
    287 		syscallarg(pid_t int)	pid;
    288 		syscallarg(const siginfo_t *)	info;
    289 	} */
    290 	ksiginfo_t	ksi;
    291 	int error;
    292 
    293 	KSI_INIT(&ksi);
    294 
    295 	if ((error = copyin(&SCARG(uap, info)->_info, &ksi.ksi_info,
    296 	    sizeof(ksi.ksi_info))) != 0)
    297 		return error;
    298 
    299 	return kill1(l, SCARG(uap, pid), &ksi, retval);
    300 }
    301 
    302 int
    303 sys_kill(struct lwp *l, const struct sys_kill_args *uap, register_t *retval)
    304 {
    305 	/* {
    306 		syscallarg(pid_t)	pid;
    307 		syscallarg(int)	signum;
    308 	} */
    309 	ksiginfo_t	ksi;
    310 
    311 	KSI_INIT(&ksi);
    312 
    313 	ksi.ksi_signo = SCARG(uap, signum);
    314 	ksi.ksi_code = SI_USER;
    315 	ksi.ksi_pid = l->l_proc->p_pid;
    316 	ksi.ksi_uid = kauth_cred_geteuid(l->l_cred);
    317 
    318 	return kill1(l, SCARG(uap, pid), &ksi, retval);
    319 }
    320 
    321 int
    322 sys_getcontext(struct lwp *l, const struct sys_getcontext_args *uap,
    323     register_t *retval)
    324 {
    325 	/* {
    326 		syscallarg(struct __ucontext *) ucp;
    327 	} */
    328 	struct proc *p = l->l_proc;
    329 	ucontext_t uc;
    330 
    331 	memset(&uc, 0, sizeof(uc));
    332 
    333 	mutex_enter(p->p_lock);
    334 	getucontext(l, &uc);
    335 	mutex_exit(p->p_lock);
    336 
    337 	return copyout(&uc, SCARG(uap, ucp), sizeof (*SCARG(uap, ucp)));
    338 }
    339 
    340 int
    341 sys_setcontext(struct lwp *l, const struct sys_setcontext_args *uap,
    342     register_t *retval)
    343 {
    344 	/* {
    345 		syscallarg(const ucontext_t *) ucp;
    346 	} */
    347 	struct proc *p = l->l_proc;
    348 	ucontext_t uc;
    349 	int error;
    350 
    351 	error = copyin(SCARG(uap, ucp), &uc, sizeof (uc));
    352 	if (error)
    353 		return error;
    354 	if ((uc.uc_flags & _UC_CPU) == 0)
    355 		return EINVAL;
    356 	mutex_enter(p->p_lock);
    357 	error = setucontext(l, &uc);
    358 	mutex_exit(p->p_lock);
    359 	if (error)
    360  		return error;
    361 
    362 	return EJUSTRETURN;
    363 }
    364 
    365 /*
    366  * sigtimedwait(2) system call, used also for implementation
    367  * of sigwaitinfo() and sigwait().
    368  *
    369  * This only handles single LWP in signal wait. libpthread provides
    370  * its own sigtimedwait() wrapper to DTRT WRT individual threads.
    371  */
    372 int
    373 sys_____sigtimedwait50(struct lwp *l,
    374     const struct sys_____sigtimedwait50_args *uap, register_t *retval)
    375 {
    376 
    377 	return sigtimedwait1(l, uap, retval, copyin, copyout, copyin, copyout);
    378 }
    379 
    380 int
    381 sigaction1(struct lwp *l, int signum, const struct sigaction *nsa,
    382 	struct sigaction *osa, const void *tramp, int vers)
    383 {
    384 	struct proc *p;
    385 	struct sigacts *ps;
    386 	sigset_t tset;
    387 	int prop, error;
    388 	ksiginfoq_t kq;
    389 	static bool v0v1valid;
    390 
    391 	if (signum <= 0 || signum >= NSIG)
    392 		return EINVAL;
    393 
    394 	p = l->l_proc;
    395 	error = 0;
    396 	ksiginfo_queue_init(&kq);
    397 
    398 	/*
    399 	 * Trampoline ABI version 0 is reserved for the legacy kernel
    400 	 * provided on-stack trampoline.  Conversely, if we are using a
    401 	 * non-0 ABI version, we must have a trampoline.  Only validate the
    402 	 * vers if a new sigaction was supplied and there was an actual
    403 	 * handler specified (not SIG_IGN or SIG_DFL), which don't require
    404 	 * a trampoline. Emulations use legacy kernel trampolines with
    405 	 * version 0, alternatively check for that too.
    406 	 *
    407 	 * If version < 2, we try to autoload the compat module.  Note
    408 	 * that we interlock with the unload check in compat_modcmd()
    409 	 * using kernconfig_lock.  If the autoload fails, we don't try it
    410 	 * again for this process.
    411 	 */
    412 	if (nsa != NULL && nsa->sa_handler != SIG_IGN
    413 	    && nsa->sa_handler != SIG_DFL) {
    414 		if (__predict_false(vers < 2)) {
    415 			if (p->p_flag & PK_32)
    416 				v0v1valid = true;
    417 			else if ((p->p_lflag & PL_SIGCOMPAT) == 0) {
    418 				kernconfig_lock();
    419 				(void)module_autoload("compat_16",
    420 				    MODULE_CLASS_ANY);
    421 				if (sendsig_sigcontext_16_hook.hooked) {
    422 					/*
    423 					 * We need to remember if the
    424 					 * sigcontext method may be useable,
    425 					 * because libc may use it even
    426 					 * if siginfo is available.
    427 					 */
    428 					v0v1valid = true;
    429 				}
    430 				mutex_enter(proc_lock);
    431 				/*
    432 				 * Prevent unload of compat module while
    433 				 * this process remains.
    434 				 */
    435 				p->p_lflag |= PL_SIGCOMPAT;
    436 				mutex_exit(proc_lock);
    437 				kernconfig_unlock();
    438 			}
    439 		}
    440 
    441 		switch (vers) {
    442 		case 0:
    443 			/* sigcontext, kernel supplied trampoline. */
    444 			if (tramp != NULL || !v0v1valid) {
    445 				return EINVAL;
    446 			}
    447 			break;
    448 		case 1:
    449 			/* sigcontext, user supplied trampoline. */
    450 			if (tramp == NULL || !v0v1valid) {
    451 				return EINVAL;
    452 			}
    453 			break;
    454 		case 2:
    455 		case 3:
    456 			/* siginfo, user supplied trampoline. */
    457 			if (tramp == NULL) {
    458 				return EINVAL;
    459 			}
    460 			break;
    461 		default:
    462 			return EINVAL;
    463 		}
    464 	}
    465 
    466 	mutex_enter(p->p_lock);
    467 
    468 	ps = p->p_sigacts;
    469 	if (osa)
    470 		sigaction_copy(osa, &SIGACTION_PS(ps, signum));
    471 	if (!nsa)
    472 		goto out;
    473 
    474 	prop = sigprop[signum];
    475 	if ((nsa->sa_flags & ~SA_ALLBITS) || (prop & SA_CANTMASK)) {
    476 		error = EINVAL;
    477 		goto out;
    478 	}
    479 
    480 	sigaction_copy(&SIGACTION_PS(ps, signum), nsa);
    481 	ps->sa_sigdesc[signum].sd_tramp = tramp;
    482 	ps->sa_sigdesc[signum].sd_vers = vers;
    483 	sigminusset(&sigcantmask, &SIGACTION_PS(ps, signum).sa_mask);
    484 
    485 	if ((prop & SA_NORESET) != 0)
    486 		SIGACTION_PS(ps, signum).sa_flags &= ~SA_RESETHAND;
    487 
    488 	if (signum == SIGCHLD) {
    489 		if (nsa->sa_flags & SA_NOCLDSTOP)
    490 			p->p_sflag |= PS_NOCLDSTOP;
    491 		else
    492 			p->p_sflag &= ~PS_NOCLDSTOP;
    493 		if (nsa->sa_flags & SA_NOCLDWAIT) {
    494 			/*
    495 			 * Paranoia: since SA_NOCLDWAIT is implemented by
    496 			 * reparenting the dying child to PID 1 (and trust
    497 			 * it to reap the zombie), PID 1 itself is forbidden
    498 			 * to set SA_NOCLDWAIT.
    499 			 */
    500 			if (p->p_pid == 1)
    501 				p->p_flag &= ~PK_NOCLDWAIT;
    502 			else
    503 				p->p_flag |= PK_NOCLDWAIT;
    504 		} else
    505 			p->p_flag &= ~PK_NOCLDWAIT;
    506 
    507 		if (nsa->sa_handler == SIG_IGN) {
    508 			/*
    509 			 * Paranoia: same as above.
    510 			 */
    511 			if (p->p_pid == 1)
    512 				p->p_flag &= ~PK_CLDSIGIGN;
    513 			else
    514 				p->p_flag |= PK_CLDSIGIGN;
    515 		} else
    516 			p->p_flag &= ~PK_CLDSIGIGN;
    517 	}
    518 
    519 	if ((nsa->sa_flags & SA_NODEFER) == 0)
    520 		sigaddset(&SIGACTION_PS(ps, signum).sa_mask, signum);
    521 	else
    522 		sigdelset(&SIGACTION_PS(ps, signum).sa_mask, signum);
    523 
    524 	/*
    525 	 * Set bit in p_sigctx.ps_sigignore for signals that are set to
    526 	 * SIG_IGN, and for signals set to SIG_DFL where the default is to
    527 	 * ignore. However, don't put SIGCONT in p_sigctx.ps_sigignore, as
    528 	 * we have to restart the process.
    529 	 */
    530 	if (nsa->sa_handler == SIG_IGN ||
    531 	    (nsa->sa_handler == SIG_DFL && (prop & SA_IGNORE) != 0)) {
    532 		/* Never to be seen again. */
    533 		sigemptyset(&tset);
    534 		sigaddset(&tset, signum);
    535 		sigclearall(p, &tset, &kq);
    536 		if (signum != SIGCONT) {
    537 			/* Easier in psignal */
    538 			sigaddset(&p->p_sigctx.ps_sigignore, signum);
    539 		}
    540 		sigdelset(&p->p_sigctx.ps_sigcatch, signum);
    541 	} else {
    542 		sigdelset(&p->p_sigctx.ps_sigignore, signum);
    543 		if (nsa->sa_handler == SIG_DFL)
    544 			sigdelset(&p->p_sigctx.ps_sigcatch, signum);
    545 		else
    546 			sigaddset(&p->p_sigctx.ps_sigcatch, signum);
    547 	}
    548 
    549 	/*
    550 	 * Previously held signals may now have become visible.  Ensure that
    551 	 * we check for them before returning to userspace.
    552 	 */
    553 	if (sigispending(l, 0)) {
    554 		lwp_lock(l);
    555 		l->l_flag |= LW_PENDSIG;
    556 		lwp_unlock(l);
    557 	}
    558 out:
    559 	mutex_exit(p->p_lock);
    560 	ksiginfo_queue_drain(&kq);
    561 
    562 	return error;
    563 }
    564 
    565 int
    566 sigprocmask1(struct lwp *l, int how, const sigset_t *nss, sigset_t *oss)
    567 {
    568 	sigset_t *mask = &l->l_sigmask;
    569 	bool more;
    570 
    571 	KASSERT(mutex_owned(l->l_proc->p_lock));
    572 
    573 	if (oss) {
    574 		*oss = *mask;
    575 	}
    576 
    577 	if (nss == NULL) {
    578 		return 0;
    579 	}
    580 
    581 	switch (how) {
    582 	case SIG_BLOCK:
    583 		sigplusset(nss, mask);
    584 		more = false;
    585 		break;
    586 	case SIG_UNBLOCK:
    587 		sigminusset(nss, mask);
    588 		more = true;
    589 		break;
    590 	case SIG_SETMASK:
    591 		*mask = *nss;
    592 		more = true;
    593 		break;
    594 	default:
    595 		return EINVAL;
    596 	}
    597 	sigminusset(&sigcantmask, mask);
    598 	if (more && sigispending(l, 0)) {
    599 		/*
    600 		 * Check for pending signals on return to user.
    601 		 */
    602 		lwp_lock(l);
    603 		l->l_flag |= LW_PENDSIG;
    604 		lwp_unlock(l);
    605 	}
    606 	return 0;
    607 }
    608 
    609 void
    610 sigpending1(struct lwp *l, sigset_t *ss)
    611 {
    612 	struct proc *p = l->l_proc;
    613 
    614 	mutex_enter(p->p_lock);
    615 	*ss = l->l_sigpend.sp_set;
    616 	sigplusset(&p->p_sigpend.sp_set, ss);
    617 	mutex_exit(p->p_lock);
    618 }
    619 
    620 void
    621 sigsuspendsetup(struct lwp *l, const sigset_t *ss)
    622 {
    623 	struct proc *p = l->l_proc;
    624 
    625 	/*
    626 	 * When returning from sigsuspend/pselect/pollts, we want
    627 	 * the old mask to be restored after the
    628 	 * signal handler has finished.  Thus, we
    629 	 * save it here and mark the sigctx structure
    630 	 * to indicate this.
    631 	 */
    632 	mutex_enter(p->p_lock);
    633 	l->l_sigrestore = 1;
    634 	l->l_sigoldmask = l->l_sigmask;
    635 	l->l_sigmask = *ss;
    636 	sigminusset(&sigcantmask, &l->l_sigmask);
    637 
    638 	/* Check for pending signals when sleeping. */
    639 	if (sigispending(l, 0)) {
    640 		lwp_lock(l);
    641 		l->l_flag |= LW_PENDSIG;
    642 		lwp_unlock(l);
    643 	}
    644 	mutex_exit(p->p_lock);
    645 }
    646 
    647 void
    648 sigsuspendteardown(struct lwp *l)
    649 {
    650 	struct proc *p = l->l_proc;
    651 
    652 	mutex_enter(p->p_lock);
    653 	/* Check for pending signals when sleeping. */
    654 	if (l->l_sigrestore) {
    655 		if (sigispending(l, 0)) {
    656 			lwp_lock(l);
    657 			l->l_flag |= LW_PENDSIG;
    658 			lwp_unlock(l);
    659 		} else {
    660 			l->l_sigrestore = 0;
    661 			l->l_sigmask = l->l_sigoldmask;
    662 		}
    663 	}
    664 	mutex_exit(p->p_lock);
    665 }
    666 
    667 int
    668 sigsuspend1(struct lwp *l, const sigset_t *ss)
    669 {
    670 
    671 	if (ss)
    672 		sigsuspendsetup(l, ss);
    673 
    674 	while (kpause("pause", true, 0, NULL) == 0)
    675 		;
    676 
    677 	/* always return EINTR rather than ERESTART... */
    678 	return EINTR;
    679 }
    680 
    681 int
    682 sigaltstack1(struct lwp *l, const struct sigaltstack *nss,
    683     struct sigaltstack *oss)
    684 {
    685 	struct proc *p = l->l_proc;
    686 	int error = 0;
    687 
    688 	mutex_enter(p->p_lock);
    689 
    690 	if (oss)
    691 		*oss = l->l_sigstk;
    692 
    693 	if (nss) {
    694 		if (nss->ss_flags & ~SS_ALLBITS)
    695 			error = EINVAL;
    696 		else if (nss->ss_flags & SS_DISABLE) {
    697 			if (l->l_sigstk.ss_flags & SS_ONSTACK)
    698 				error = EINVAL;
    699 		} else if (nss->ss_size < MINSIGSTKSZ)
    700 			error = ENOMEM;
    701 
    702 		if (!error)
    703 			l->l_sigstk = *nss;
    704 	}
    705 
    706 	mutex_exit(p->p_lock);
    707 
    708 	return error;
    709 }
    710 
    711 int
    712 sigtimedwait1(struct lwp *l, const struct sys_____sigtimedwait50_args *uap,
    713     register_t *retval, copyin_t fetchss, copyout_t storeinf, copyin_t fetchts,
    714     copyout_t storets)
    715 {
    716 	/* {
    717 		syscallarg(const sigset_t *) set;
    718 		syscallarg(siginfo_t *) info;
    719 		syscallarg(struct timespec *) timeout;
    720 	} */
    721 	struct proc *p = l->l_proc;
    722 	int error, signum, timo;
    723 	struct timespec ts, tsstart, tsnow;
    724 	ksiginfo_t ksi;
    725 
    726 	/*
    727 	 * Calculate timeout, if it was specified.
    728 	 *
    729 	 * NULL pointer means an infinite timeout.
    730 	 * {.tv_sec = 0, .tv_nsec = 0} means do not block.
    731 	 */
    732 	if (SCARG(uap, timeout)) {
    733 		error = (*fetchts)(SCARG(uap, timeout), &ts, sizeof(ts));
    734 		if (error)
    735 			return error;
    736 
    737 		if ((error = itimespecfix(&ts)) != 0)
    738 			return error;
    739 
    740 		timo = tstohz(&ts);
    741 		if (timo == 0) {
    742 			if (ts.tv_sec == 0 && ts.tv_nsec == 0)
    743 				timo = -1; /* do not block */
    744 			else
    745 				timo = 1; /* the shortest possible timeout */
    746 		}
    747 
    748 		/*
    749 		 * Remember current uptime, it would be used in
    750 		 * ECANCELED/ERESTART case.
    751 		 */
    752 		getnanouptime(&tsstart);
    753 	} else {
    754 		memset(&tsstart, 0, sizeof(tsstart)); /* XXXgcc */
    755 		timo = 0; /* infinite timeout */
    756 	}
    757 
    758 	error = (*fetchss)(SCARG(uap, set), &l->l_sigwaitset,
    759 	    sizeof(l->l_sigwaitset));
    760 	if (error)
    761 		return error;
    762 
    763 	/*
    764 	 * Silently ignore SA_CANTMASK signals. psignal1() would ignore
    765 	 * SA_CANTMASK signals in waitset, we do this only for the below
    766 	 * siglist check.
    767 	 */
    768 	sigminusset(&sigcantmask, &l->l_sigwaitset);
    769 
    770 	memset(&ksi.ksi_info, 0, sizeof(ksi.ksi_info));
    771 
    772 	mutex_enter(p->p_lock);
    773 
    774 	/* Check for pending signals in the process, if no - then in LWP. */
    775 	if ((signum = sigget(&p->p_sigpend, &ksi, 0, &l->l_sigwaitset)) == 0)
    776 		signum = sigget(&l->l_sigpend, &ksi, 0, &l->l_sigwaitset);
    777 
    778 	if (signum != 0) {
    779 		/* If found a pending signal, just copy it out to the user. */
    780 		mutex_exit(p->p_lock);
    781 		goto out;
    782 	}
    783 
    784 	if (timo < 0) {
    785 		/* If not allowed to block, return an error */
    786 		mutex_exit(p->p_lock);
    787 		return EAGAIN;
    788 	}
    789 
    790 	/*
    791 	 * Set up the sigwait list and wait for signal to arrive.
    792 	 * We can either be woken up or time out.
    793 	 */
    794 	l->l_sigwaited = &ksi;
    795 	LIST_INSERT_HEAD(&p->p_sigwaiters, l, l_sigwaiter);
    796 	error = cv_timedwait_sig(&l->l_sigcv, p->p_lock, timo);
    797 
    798 	/*
    799 	 * Need to find out if we woke as a result of _lwp_wakeup() or a
    800 	 * signal outside our wait set.
    801 	 */
    802 	if (l->l_sigwaited != NULL) {
    803 		if (error == EINTR) {
    804 			/* Wakeup via _lwp_wakeup(). */
    805 			error = ECANCELED;
    806 		} else if (!error) {
    807 			/* Spurious wakeup - arrange for syscall restart. */
    808 			error = ERESTART;
    809 		}
    810 		l->l_sigwaited = NULL;
    811 		LIST_REMOVE(l, l_sigwaiter);
    812 	}
    813 	mutex_exit(p->p_lock);
    814 
    815 	/*
    816 	 * If the sleep was interrupted (either by signal or wakeup), update
    817 	 * the timeout and copyout new value back.  It would be used when
    818 	 * the syscall would be restarted or called again.
    819 	 */
    820 	if (timo && (error == ERESTART || error == ECANCELED)) {
    821 		getnanouptime(&tsnow);
    822 
    823 		/* Compute how much time has passed since start. */
    824 		timespecsub(&tsnow, &tsstart, &tsnow);
    825 
    826 		/* Substract passed time from timeout. */
    827 		timespecsub(&ts, &tsnow, &ts);
    828 
    829 		if (ts.tv_sec < 0)
    830 			error = EAGAIN;
    831 		else {
    832 			/* Copy updated timeout to userland. */
    833 			error = (*storets)(&ts, SCARG(uap, timeout),
    834 			    sizeof(ts));
    835 		}
    836 	}
    837 out:
    838 	/*
    839 	 * If a signal from the wait set arrived, copy it to userland.
    840 	 * Copy only the used part of siginfo, the padding part is
    841 	 * left unchanged (userland is not supposed to touch it anyway).
    842 	 */
    843 	if (error == 0 && SCARG(uap, info)) {
    844 		error = (*storeinf)(&ksi.ksi_info, SCARG(uap, info),
    845 		    sizeof(ksi.ksi_info));
    846 	}
    847 	if (error == 0) {
    848 		*retval = ksi.ksi_info._signo;
    849 		SDT_PROBE(proc, kernel, , signal__clear, *retval,
    850 		    &ksi, 0, 0, 0);
    851 	}
    852 	return error;
    853 }
    854