]> git.saurik.com Git - apple/xnu.git/blame_incremental - bsd/kern/kern_sig.c
xnu-7195.101.1.tar.gz
[apple/xnu.git] / bsd / kern / kern_sig.c
... / ...
CommitLineData
1/*
2 * Copyright (c) 1995-2016 Apple Inc. All rights reserved.
3 *
4 * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
5 *
6 * This file contains Original Code and/or Modifications of Original Code
7 * as defined in and that are subject to the Apple Public Source License
8 * Version 2.0 (the 'License'). You may not use this file except in
9 * compliance with the License. The rights granted to you under the License
10 * may not be used to create, or enable the creation or redistribution of,
11 * unlawful or unlicensed copies of an Apple operating system, or to
12 * circumvent, violate, or enable the circumvention or violation of, any
13 * terms of an Apple operating system software license agreement.
14 *
15 * Please obtain a copy of the License at
16 * http://www.opensource.apple.com/apsl/ and read it before using this file.
17 *
18 * The Original Code and all software distributed under the License are
19 * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
20 * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
21 * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
22 * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
23 * Please see the License for the specific language governing rights and
24 * limitations under the License.
25 *
26 * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
27 */
28/*
29 * Copyright (c) 1982, 1986, 1989, 1991, 1993
30 * The Regents of the University of California. All rights reserved.
31 * (c) UNIX System Laboratories, Inc.
32 * All or some portions of this file are derived from material licensed
33 * to the University of California by American Telephone and Telegraph
34 * Co. or Unix System Laboratories, Inc. and are reproduced herein with
35 * the permission of UNIX System Laboratories, Inc.
36 *
37 * Redistribution and use in source and binary forms, with or without
38 * modification, are permitted provided that the following conditions
39 * are met:
40 * 1. Redistributions of source code must retain the above copyright
41 * notice, this list of conditions and the following disclaimer.
42 * 2. Redistributions in binary form must reproduce the above copyright
43 * notice, this list of conditions and the following disclaimer in the
44 * documentation and/or other materials provided with the distribution.
45 * 3. All advertising materials mentioning features or use of this software
46 * must display the following acknowledgement:
47 * This product includes software developed by the University of
48 * California, Berkeley and its contributors.
49 * 4. Neither the name of the University nor the names of its contributors
50 * may be used to endorse or promote products derived from this software
51 * without specific prior written permission.
52 *
53 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
54 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
55 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
56 * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
57 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
58 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
59 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
60 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
61 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
62 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
63 * SUCH DAMAGE.
64 *
65 * @(#)kern_sig.c 8.7 (Berkeley) 4/18/94
66 */
67/*
68 * NOTICE: This file was modified by SPARTA, Inc. in 2005 to introduce
69 * support for mandatory and extensible security protections. This notice
70 * is included in support of clause 2.2 (b) of the Apple Public License,
71 * Version 2.0.
72 */
73
74#define SIGPROP /* include signal properties table */
75#include <sys/param.h>
76#include <sys/resourcevar.h>
77#include <sys/proc_internal.h>
78#include <sys/kauth.h>
79#include <sys/systm.h>
80#include <sys/timeb.h>
81#include <sys/times.h>
82#include <sys/acct.h>
83#include <sys/file_internal.h>
84#include <sys/kernel.h>
85#include <sys/wait.h>
86#include <sys/signalvar.h>
87#include <sys/syslog.h>
88#include <sys/stat.h>
89#include <sys/lock.h>
90#include <sys/kdebug.h>
91#include <sys/reason.h>
92
93#include <sys/mount.h>
94#include <sys/sysproto.h>
95
96#include <security/audit/audit.h>
97
98#include <kern/cpu_number.h>
99
100#include <sys/vm.h>
101#include <sys/user.h> /* for coredump */
102#include <kern/ast.h> /* for APC support */
103#include <kern/kalloc.h>
104#include <kern/task.h> /* extern void *get_bsdtask_info(task_t); */
105#include <kern/thread.h>
106#include <kern/sched_prim.h>
107#include <kern/thread_call.h>
108#include <kern/policy_internal.h>
109
110#include <mach/exception.h>
111#include <mach/task.h>
112#include <mach/thread_act.h>
113#include <libkern/OSAtomic.h>
114
115#include <sys/sdt.h>
116#include <sys/codesign.h>
117#include <sys/random.h>
118#include <libkern/section_keywords.h>
119
120#if CONFIG_MACF
121#include <security/mac_framework.h>
122#endif
123
124/*
125 * Missing prototypes that Mach should export
126 *
127 * +++
128 */
129extern int thread_enable_fpe(thread_t act, int onoff);
130extern kern_return_t get_signalact(task_t, thread_t *, int);
131extern unsigned int get_useraddr(void);
132extern boolean_t task_did_exec(task_t task);
133extern boolean_t task_is_exec_copy(task_t task);
134extern void vm_shared_region_reslide_stale(void);
135
136/*
137 * ---
138 */
139
140extern void doexception(int exc, mach_exception_code_t code,
141 mach_exception_subcode_t sub);
142
143static void stop(proc_t, proc_t);
144static int cansignal_nomac(proc_t, kauth_cred_t, proc_t, int);
145int cansignal(proc_t, kauth_cred_t, proc_t, int);
146int killpg1(proc_t, int, int, int, int);
147kern_return_t do_bsdexception(int, int, int);
148void __posix_sem_syscall_return(kern_return_t);
149char *proc_name_address(void *p);
150
151/* implementations in osfmk/kern/sync_sema.c. We do not want port.h in this scope, so void * them */
152kern_return_t semaphore_timedwait_signal_trap_internal(mach_port_name_t, mach_port_name_t, unsigned int, clock_res_t, void (*)(kern_return_t));
153kern_return_t semaphore_timedwait_trap_internal(mach_port_name_t, unsigned int, clock_res_t, void (*)(kern_return_t));
154kern_return_t semaphore_wait_signal_trap_internal(mach_port_name_t, mach_port_name_t, void (*)(kern_return_t));
155kern_return_t semaphore_wait_trap_internal(mach_port_name_t, void (*)(kern_return_t));
156
157static int filt_sigattach(struct knote *kn, struct kevent_qos_s *kev);
158static void filt_sigdetach(struct knote *kn);
159static int filt_signal(struct knote *kn, long hint);
160static int filt_signaltouch(struct knote *kn, struct kevent_qos_s *kev);
161static int filt_signalprocess(struct knote *kn, struct kevent_qos_s *kev);
162
163SECURITY_READ_ONLY_EARLY(struct filterops) sig_filtops = {
164 .f_attach = filt_sigattach,
165 .f_detach = filt_sigdetach,
166 .f_event = filt_signal,
167 .f_touch = filt_signaltouch,
168 .f_process = filt_signalprocess,
169};
170
171/* structures and fns for killpg1 iterartion callback and filters */
172struct killpg1_filtargs {
173 bool posix;
174 proc_t curproc;
175};
176
177struct killpg1_iterargs {
178 proc_t curproc;
179 kauth_cred_t uc;
180 int signum;
181 int nfound;
182};
183
184static int killpg1_allfilt(proc_t p, void * arg);
185static int killpg1_pgrpfilt(proc_t p, __unused void * arg);
186static int killpg1_callback(proc_t p, void * arg);
187
188static int pgsignal_filt(proc_t p, void * arg);
189static int pgsignal_callback(proc_t p, void * arg);
190static kern_return_t get_signalthread(proc_t, int, thread_t *);
191
192
193/* flags for psignal_internal */
194#define PSIG_LOCKED 0x1
195#define PSIG_VFORK 0x2
196#define PSIG_THREAD 0x4
197#define PSIG_TRY_THREAD 0x8
198
199static os_reason_t build_signal_reason(int signum, const char *procname);
200static void psignal_internal(proc_t p, task_t task, thread_t thread, int flavor, int signum, os_reason_t signal_reason);
201
202/*
203 * NOTE: Source and target may *NOT* overlap! (target is smaller)
204 */
205static void
206sigaltstack_kern_to_user32(struct kern_sigaltstack *in, struct user32_sigaltstack *out)
207{
208 out->ss_sp = CAST_DOWN_EXPLICIT(user32_addr_t, in->ss_sp);
209 out->ss_size = CAST_DOWN_EXPLICIT(user32_size_t, in->ss_size);
210 out->ss_flags = in->ss_flags;
211}
212
213static void
214sigaltstack_kern_to_user64(struct kern_sigaltstack *in, struct user64_sigaltstack *out)
215{
216 out->ss_sp = in->ss_sp;
217 out->ss_size = in->ss_size;
218 out->ss_flags = in->ss_flags;
219}
220
221/*
222 * NOTE: Source and target may are permitted to overlap! (source is smaller);
223 * this works because we copy fields in order from the end of the struct to
224 * the beginning.
225 */
226static void
227sigaltstack_user32_to_kern(struct user32_sigaltstack *in, struct kern_sigaltstack *out)
228{
229 out->ss_flags = in->ss_flags;
230 out->ss_size = in->ss_size;
231 out->ss_sp = CAST_USER_ADDR_T(in->ss_sp);
232}
233static void
234sigaltstack_user64_to_kern(struct user64_sigaltstack *in, struct kern_sigaltstack *out)
235{
236 out->ss_flags = in->ss_flags;
237 out->ss_size = (user_size_t)in->ss_size;
238 out->ss_sp = (user_addr_t)in->ss_sp;
239}
240
241static void
242sigaction_kern_to_user32(struct kern_sigaction *in, struct user32_sigaction *out)
243{
244 /* This assumes 32 bit __sa_handler is of type sig_t */
245 out->__sigaction_u.__sa_handler = CAST_DOWN_EXPLICIT(user32_addr_t, in->__sigaction_u.__sa_handler);
246 out->sa_mask = in->sa_mask;
247 out->sa_flags = in->sa_flags;
248}
249static void
250sigaction_kern_to_user64(struct kern_sigaction *in, struct user64_sigaction *out)
251{
252 /* This assumes 32 bit __sa_handler is of type sig_t */
253 out->__sigaction_u.__sa_handler = in->__sigaction_u.__sa_handler;
254 out->sa_mask = in->sa_mask;
255 out->sa_flags = in->sa_flags;
256}
257
258static void
259__sigaction_user32_to_kern(struct __user32_sigaction *in, struct __kern_sigaction *out)
260{
261 out->__sigaction_u.__sa_handler = CAST_USER_ADDR_T(in->__sigaction_u.__sa_handler);
262 out->sa_tramp = CAST_USER_ADDR_T(in->sa_tramp);
263 out->sa_mask = in->sa_mask;
264 out->sa_flags = in->sa_flags;
265
266 kern_return_t kr;
267 kr = machine_thread_function_pointers_convert_from_user(current_thread(),
268 &out->sa_tramp, 1);
269 assert(kr == KERN_SUCCESS);
270}
271
272static void
273__sigaction_user64_to_kern(struct __user64_sigaction *in, struct __kern_sigaction *out)
274{
275 out->__sigaction_u.__sa_handler = (user_addr_t)in->__sigaction_u.__sa_handler;
276 out->sa_tramp = (user_addr_t)in->sa_tramp;
277 out->sa_mask = in->sa_mask;
278 out->sa_flags = in->sa_flags;
279
280 kern_return_t kr;
281 kr = machine_thread_function_pointers_convert_from_user(current_thread(),
282 &out->sa_tramp, 1);
283 assert(kr == KERN_SUCCESS);
284}
285
286#if SIGNAL_DEBUG
287void ram_printf(int);
288int ram_debug = 0;
289unsigned int rdebug_proc = 0;
290void
291ram_printf(int x)
292{
293 printf("x is %d", x);
294}
295#endif /* SIGNAL_DEBUG */
296
297
298void
299signal_setast(thread_t sig_actthread)
300{
301 act_set_astbsd(sig_actthread);
302}
303
304static int
305cansignal_nomac(proc_t src, kauth_cred_t uc_src, proc_t dst, int signum)
306{
307 /* you can signal yourself */
308 if (src == dst) {
309 return 1;
310 }
311
312 /* you can't send the init proc SIGKILL, even if root */
313 if (signum == SIGKILL && dst == initproc) {
314 return 0;
315 }
316
317 /* otherwise, root can always signal */
318 if (kauth_cred_issuser(uc_src)) {
319 return 1;
320 }
321
322 /* processes in the same session can send SIGCONT to each other */
323 {
324 struct session *sess_src = SESSION_NULL;
325 struct session *sess_dst = SESSION_NULL;
326
327 /* The session field is protected by the list lock. */
328 proc_list_lock();
329 if (src->p_pgrp != PGRP_NULL) {
330 sess_src = src->p_pgrp->pg_session;
331 }
332 if (dst->p_pgrp != PGRP_NULL) {
333 sess_dst = dst->p_pgrp->pg_session;
334 }
335 proc_list_unlock();
336
337 /* allow SIGCONT within session and for processes without session */
338 if (signum == SIGCONT && sess_src == sess_dst) {
339 return 1;
340 }
341 }
342
343 /* the source process must be authorized to signal the target */
344 {
345 int allowed = 0;
346 kauth_cred_t uc_dst = NOCRED, uc_ref = NOCRED;
347
348 uc_dst = uc_ref = kauth_cred_proc_ref(dst);
349
350 /*
351 * If the real or effective UID of the sender matches the real or saved
352 * UID of the target, allow the signal to be sent.
353 */
354 if (kauth_cred_getruid(uc_src) == kauth_cred_getruid(uc_dst) ||
355 kauth_cred_getruid(uc_src) == kauth_cred_getsvuid(uc_dst) ||
356 kauth_cred_getuid(uc_src) == kauth_cred_getruid(uc_dst) ||
357 kauth_cred_getuid(uc_src) == kauth_cred_getsvuid(uc_dst)) {
358 allowed = 1;
359 }
360
361 if (uc_ref != NOCRED) {
362 kauth_cred_unref(&uc_ref);
363 uc_ref = NOCRED;
364 }
365
366 return allowed;
367 }
368}
369
370/*
371 * Can process `src`, with ucred `uc_src`, send the signal `signum` to process
372 * `dst`? The ucred is referenced by the caller so internal fileds can be used
373 * safely.
374 */
375int
376cansignal(proc_t src, kauth_cred_t uc_src, proc_t dst, int signum)
377{
378#if CONFIG_MACF
379 if (mac_proc_check_signal(src, dst, signum)) {
380 return 0;
381 }
382#endif
383
384 return cansignal_nomac(src, uc_src, dst, signum);
385}
386
387/*
388 * <rdar://problem/21952708> Some signals can be restricted from being handled,
389 * forcing the default action for that signal. This behavior applies only to
390 * non-root (EUID != 0) processes, and is configured with the "sigrestrict=x"
391 * bootarg:
392 *
393 * 0 (default): Disallow use of restricted signals. Trying to register a handler
394 * returns ENOTSUP, which userspace may use to take special action (e.g. abort).
395 * 1: As above, but return EINVAL. Restricted signals behave similarly to SIGKILL.
396 * 2: Usual POSIX semantics.
397 */
398unsigned sigrestrict_arg = 0;
399
400#if PLATFORM_WatchOS
401static int
402sigrestrictmask(void)
403{
404 if (kauth_getuid() != 0 && sigrestrict_arg != 2) {
405 return SIGRESTRICTMASK;
406 }
407 return 0;
408}
409
410static int
411signal_is_restricted(proc_t p, int signum)
412{
413 if (sigmask(signum) & sigrestrictmask()) {
414 if (sigrestrict_arg == 0 &&
415 task_get_apptype(p->task) == TASK_APPTYPE_APP_DEFAULT) {
416 return ENOTSUP;
417 } else {
418 return EINVAL;
419 }
420 }
421 return 0;
422}
423
424#else
425
426static inline int
427signal_is_restricted(proc_t p, int signum)
428{
429 (void)p;
430 (void)signum;
431 return 0;
432}
433#endif /* !PLATFORM_WatchOS */
434
435/*
436 * Returns: 0 Success
437 * EINVAL
438 * copyout:EFAULT
439 * copyin:EFAULT
440 *
441 * Notes: Uses current thread as a parameter to inform PPC to enable
442 * FPU exceptions via setsigvec(); this operation is not proxy
443 * safe!
444 */
445/* ARGSUSED */
446int
447sigaction(proc_t p, struct sigaction_args *uap, __unused int32_t *retval)
448{
449 struct kern_sigaction vec;
450 struct __kern_sigaction __vec;
451
452 struct kern_sigaction *sa = &vec;
453 struct sigacts *ps = p->p_sigacts;
454
455 int signum;
456 int bit, error = 0;
457 uint32_t sigreturn_validation = PS_SIGRETURN_VALIDATION_DEFAULT;
458
459 signum = uap->signum;
460 if (signum <= 0 || signum >= NSIG ||
461 signum == SIGKILL || signum == SIGSTOP) {
462 return EINVAL;
463 }
464
465 if (uap->nsa) {
466 if (IS_64BIT_PROCESS(p)) {
467 struct __user64_sigaction __vec64;
468 error = copyin(uap->nsa, &__vec64, sizeof(__vec64));
469 __sigaction_user64_to_kern(&__vec64, &__vec);
470 } else {
471 struct __user32_sigaction __vec32;
472 error = copyin(uap->nsa, &__vec32, sizeof(__vec32));
473 __sigaction_user32_to_kern(&__vec32, &__vec);
474 }
475 if (error) {
476 return error;
477 }
478
479 sigreturn_validation = (__vec.sa_flags & SA_VALIDATE_SIGRETURN_FROM_SIGTRAMP) ?
480 PS_SIGRETURN_VALIDATION_ENABLED : PS_SIGRETURN_VALIDATION_DISABLED;
481 __vec.sa_flags &= SA_USERSPACE_MASK; /* Only pass on valid sa_flags */
482
483 if ((__vec.sa_flags & SA_SIGINFO) || __vec.sa_handler != SIG_DFL) {
484 if ((error = signal_is_restricted(p, signum))) {
485 if (error == ENOTSUP) {
486 printf("%s(%d): denied attempt to register action for signal %d\n",
487 proc_name_address(p), proc_pid(p), signum);
488 }
489 return error;
490 }
491 }
492 }
493
494 if (uap->osa) {
495 sa->sa_handler = ps->ps_sigact[signum];
496 sa->sa_mask = ps->ps_catchmask[signum];
497 bit = sigmask(signum);
498 sa->sa_flags = 0;
499 if ((ps->ps_sigonstack & bit) != 0) {
500 sa->sa_flags |= SA_ONSTACK;
501 }
502 if ((ps->ps_sigintr & bit) == 0) {
503 sa->sa_flags |= SA_RESTART;
504 }
505 if (ps->ps_siginfo & bit) {
506 sa->sa_flags |= SA_SIGINFO;
507 }
508 if (ps->ps_signodefer & bit) {
509 sa->sa_flags |= SA_NODEFER;
510 }
511 if ((signum == SIGCHLD) && (p->p_flag & P_NOCLDSTOP)) {
512 sa->sa_flags |= SA_NOCLDSTOP;
513 }
514 if ((signum == SIGCHLD) && (p->p_flag & P_NOCLDWAIT)) {
515 sa->sa_flags |= SA_NOCLDWAIT;
516 }
517
518 if (IS_64BIT_PROCESS(p)) {
519 struct user64_sigaction vec64 = {};
520 sigaction_kern_to_user64(sa, &vec64);
521 error = copyout(&vec64, uap->osa, sizeof(vec64));
522 } else {
523 struct user32_sigaction vec32 = {};
524 sigaction_kern_to_user32(sa, &vec32);
525 error = copyout(&vec32, uap->osa, sizeof(vec32));
526 }
527 if (error) {
528 return error;
529 }
530 }
531
532 if (uap->nsa) {
533 uint32_t old_sigreturn_validation = atomic_load_explicit(
534 &ps->ps_sigreturn_validation, memory_order_relaxed);
535 if (old_sigreturn_validation == PS_SIGRETURN_VALIDATION_DEFAULT) {
536 atomic_compare_exchange_strong_explicit(&ps->ps_sigreturn_validation,
537 &old_sigreturn_validation, sigreturn_validation,
538 memory_order_relaxed, memory_order_relaxed);
539 }
540 error = setsigvec(p, current_thread(), signum, &__vec, FALSE);
541 }
542
543 return error;
544}
545
546/* Routines to manipulate bits on all threads */
547int
548clear_procsiglist(proc_t p, int bit, boolean_t in_signalstart)
549{
550 struct uthread * uth;
551 thread_t thact;
552
553 proc_lock(p);
554 if (!in_signalstart) {
555 proc_signalstart(p, 1);
556 }
557
558 if ((p->p_lflag & P_LINVFORK) && p->p_vforkact) {
559 thact = p->p_vforkact;
560 uth = (struct uthread *)get_bsdthread_info(thact);
561 if (uth) {
562 uth->uu_siglist &= ~bit;
563 }
564 if (!in_signalstart) {
565 proc_signalend(p, 1);
566 }
567 proc_unlock(p);
568 return 0;
569 }
570
571 TAILQ_FOREACH(uth, &p->p_uthlist, uu_list) {
572 uth->uu_siglist &= ~bit;
573 }
574 p->p_siglist &= ~bit;
575 if (!in_signalstart) {
576 proc_signalend(p, 1);
577 }
578 proc_unlock(p);
579
580 return 0;
581}
582
583
584static int
585unblock_procsigmask(proc_t p, int bit)
586{
587 struct uthread * uth;
588 thread_t thact;
589
590 proc_lock(p);
591 proc_signalstart(p, 1);
592
593 if ((p->p_lflag & P_LINVFORK) && p->p_vforkact) {
594 thact = p->p_vforkact;
595 uth = (struct uthread *)get_bsdthread_info(thact);
596 if (uth) {
597 uth->uu_sigmask &= ~bit;
598 }
599 p->p_sigmask &= ~bit;
600 proc_signalend(p, 1);
601 proc_unlock(p);
602 return 0;
603 }
604 TAILQ_FOREACH(uth, &p->p_uthlist, uu_list) {
605 uth->uu_sigmask &= ~bit;
606 }
607 p->p_sigmask &= ~bit;
608
609 proc_signalend(p, 1);
610 proc_unlock(p);
611 return 0;
612}
613
614static int
615block_procsigmask(proc_t p, int bit)
616{
617 struct uthread * uth;
618 thread_t thact;
619
620 proc_lock(p);
621 proc_signalstart(p, 1);
622
623 if ((p->p_lflag & P_LINVFORK) && p->p_vforkact) {
624 thact = p->p_vforkact;
625 uth = (struct uthread *)get_bsdthread_info(thact);
626 if (uth) {
627 uth->uu_sigmask |= bit;
628 }
629 p->p_sigmask |= bit;
630 proc_signalend(p, 1);
631 proc_unlock(p);
632 return 0;
633 }
634 TAILQ_FOREACH(uth, &p->p_uthlist, uu_list) {
635 uth->uu_sigmask |= bit;
636 }
637 p->p_sigmask |= bit;
638
639 proc_signalend(p, 1);
640 proc_unlock(p);
641 return 0;
642}
643
644int
645set_procsigmask(proc_t p, int bit)
646{
647 struct uthread * uth;
648 thread_t thact;
649
650 proc_lock(p);
651 proc_signalstart(p, 1);
652
653 if ((p->p_lflag & P_LINVFORK) && p->p_vforkact) {
654 thact = p->p_vforkact;
655 uth = (struct uthread *)get_bsdthread_info(thact);
656 if (uth) {
657 uth->uu_sigmask = bit;
658 }
659 p->p_sigmask = bit;
660 proc_signalend(p, 1);
661 proc_unlock(p);
662 return 0;
663 }
664 TAILQ_FOREACH(uth, &p->p_uthlist, uu_list) {
665 uth->uu_sigmask = bit;
666 }
667 p->p_sigmask = bit;
668 proc_signalend(p, 1);
669 proc_unlock(p);
670
671 return 0;
672}
673
674/* XXX should be static? */
675/*
676 * Notes: The thread parameter is used in the PPC case to select the
677 * thread on which the floating point exception will be enabled
678 * or disabled. We can't simply take current_thread(), since
679 * this is called from posix_spawn() on the not currently running
680 * process/thread pair.
681 *
682 * We mark thread as unused to alow compilation without warning
683 * on non-PPC platforms.
684 */
685int
686setsigvec(proc_t p, __unused thread_t thread, int signum, struct __kern_sigaction *sa, boolean_t in_sigstart)
687{
688 struct sigacts *ps = p->p_sigacts;
689 int bit;
690
691 assert(signum < NSIG);
692
693 if ((signum == SIGKILL || signum == SIGSTOP) &&
694 sa->sa_handler != SIG_DFL) {
695 return EINVAL;
696 }
697 bit = sigmask(signum);
698 /*
699 * Change setting atomically.
700 */
701 ps->ps_sigact[signum] = sa->sa_handler;
702 ps->ps_trampact[signum] = sa->sa_tramp;
703 ps->ps_catchmask[signum] = sa->sa_mask & ~sigcantmask;
704 if (sa->sa_flags & SA_SIGINFO) {
705 ps->ps_siginfo |= bit;
706 } else {
707 ps->ps_siginfo &= ~bit;
708 }
709 if ((sa->sa_flags & SA_RESTART) == 0) {
710 ps->ps_sigintr |= bit;
711 } else {
712 ps->ps_sigintr &= ~bit;
713 }
714 if (sa->sa_flags & SA_ONSTACK) {
715 ps->ps_sigonstack |= bit;
716 } else {
717 ps->ps_sigonstack &= ~bit;
718 }
719 if (sa->sa_flags & SA_RESETHAND) {
720 ps->ps_sigreset |= bit;
721 } else {
722 ps->ps_sigreset &= ~bit;
723 }
724 if (sa->sa_flags & SA_NODEFER) {
725 ps->ps_signodefer |= bit;
726 } else {
727 ps->ps_signodefer &= ~bit;
728 }
729 if (signum == SIGCHLD) {
730 if (sa->sa_flags & SA_NOCLDSTOP) {
731 OSBitOrAtomic(P_NOCLDSTOP, &p->p_flag);
732 } else {
733 OSBitAndAtomic(~((uint32_t)P_NOCLDSTOP), &p->p_flag);
734 }
735 if ((sa->sa_flags & SA_NOCLDWAIT) || (sa->sa_handler == SIG_IGN)) {
736 OSBitOrAtomic(P_NOCLDWAIT, &p->p_flag);
737 } else {
738 OSBitAndAtomic(~((uint32_t)P_NOCLDWAIT), &p->p_flag);
739 }
740 }
741
742 /*
743 * Set bit in p_sigignore for signals that are set to SIG_IGN,
744 * and for signals set to SIG_DFL where the default is to ignore.
745 * However, don't put SIGCONT in p_sigignore,
746 * as we have to restart the process.
747 */
748 if (sa->sa_handler == SIG_IGN ||
749 (sigprop[signum] & SA_IGNORE && sa->sa_handler == SIG_DFL)) {
750 clear_procsiglist(p, bit, in_sigstart);
751 if (signum != SIGCONT) {
752 p->p_sigignore |= bit; /* easier in psignal */
753 }
754 p->p_sigcatch &= ~bit;
755 } else {
756 p->p_sigignore &= ~bit;
757 if (sa->sa_handler == SIG_DFL) {
758 p->p_sigcatch &= ~bit;
759 } else {
760 p->p_sigcatch |= bit;
761 }
762 }
763 return 0;
764}
765
766/*
767 * Initialize signal state for process 0;
768 * set to ignore signals that are ignored by default.
769 */
770void
771siginit(proc_t p)
772{
773 int i;
774
775 for (i = 1; i < NSIG; i++) {
776 if (sigprop[i] & SA_IGNORE && i != SIGCONT) {
777 p->p_sigignore |= sigmask(i);
778 }
779 }
780}
781
782/*
783 * Reset signals for an exec of the specified process.
784 */
785void
786execsigs(proc_t p, thread_t thread)
787{
788 struct sigacts *ps = p->p_sigacts;
789 int nc, mask;
790 struct uthread *ut;
791
792 ut = (struct uthread *)get_bsdthread_info(thread);
793
794 /*
795 * transfer saved signal states from the process
796 * back to the current thread.
797 *
798 * NOTE: We do this without the process locked,
799 * because we are guaranteed to be single-threaded
800 * by this point in exec and the p_siglist is
801 * only accessed by threads inside the process.
802 */
803 ut->uu_siglist |= p->p_siglist;
804 p->p_siglist = 0;
805
806 /*
807 * Reset caught signals. Held signals remain held
808 * through p_sigmask (unless they were caught,
809 * and are now ignored by default).
810 */
811 while (p->p_sigcatch) {
812 nc = ffs((unsigned int)p->p_sigcatch);
813 mask = sigmask(nc);
814 p->p_sigcatch &= ~mask;
815 if (sigprop[nc] & SA_IGNORE) {
816 if (nc != SIGCONT) {
817 p->p_sigignore |= mask;
818 }
819 ut->uu_siglist &= ~mask;
820 }
821 ps->ps_sigact[nc] = SIG_DFL;
822 }
823
824 atomic_store_explicit(&ps->ps_sigreturn_validation,
825 PS_SIGRETURN_VALIDATION_DEFAULT, memory_order_relaxed);
826 /* Generate random token value used to validate sigreturn arguments */
827 read_random(&ps->ps_sigreturn_token, sizeof(ps->ps_sigreturn_token));
828
829 /*
830 * Reset stack state to the user stack.
831 * Clear set of signals caught on the signal stack.
832 */
833 /* thread */
834 ut->uu_sigstk.ss_flags = SA_DISABLE;
835 ut->uu_sigstk.ss_size = 0;
836 ut->uu_sigstk.ss_sp = USER_ADDR_NULL;
837 ut->uu_flag &= ~UT_ALTSTACK;
838 /* process */
839 ps->ps_sigonstack = 0;
840}
841
842/*
843 * Manipulate signal mask.
844 * Note that we receive new mask, not pointer,
845 * and return old mask as return value;
846 * the library stub does the rest.
847 */
848int
849sigprocmask(proc_t p, struct sigprocmask_args *uap, __unused int32_t *retval)
850{
851 int error = 0;
852 sigset_t oldmask, nmask;
853 user_addr_t omask = uap->omask;
854 struct uthread *ut;
855
856 ut = (struct uthread *)get_bsdthread_info(current_thread());
857 oldmask = ut->uu_sigmask;
858
859 if (uap->mask == USER_ADDR_NULL) {
860 /* just want old mask */
861 goto out;
862 }
863 error = copyin(uap->mask, &nmask, sizeof(sigset_t));
864 if (error) {
865 goto out;
866 }
867
868 switch (uap->how) {
869 case SIG_BLOCK:
870 block_procsigmask(p, (nmask & ~sigcantmask));
871 signal_setast(current_thread());
872 break;
873
874 case SIG_UNBLOCK:
875 unblock_procsigmask(p, (nmask & ~sigcantmask));
876 signal_setast(current_thread());
877 break;
878
879 case SIG_SETMASK:
880 set_procsigmask(p, (nmask & ~sigcantmask));
881 signal_setast(current_thread());
882 break;
883
884 default:
885 error = EINVAL;
886 break;
887 }
888out:
889 if (!error && omask != USER_ADDR_NULL) {
890 copyout(&oldmask, omask, sizeof(sigset_t));
891 }
892 return error;
893}
894
895int
896sigpending(__unused proc_t p, struct sigpending_args *uap, __unused int32_t *retval)
897{
898 struct uthread *ut;
899 sigset_t pendlist;
900
901 ut = (struct uthread *)get_bsdthread_info(current_thread());
902 pendlist = ut->uu_siglist;
903
904 if (uap->osv) {
905 copyout(&pendlist, uap->osv, sizeof(sigset_t));
906 }
907 return 0;
908}
909
910/*
911 * Suspend process until signal, providing mask to be set
912 * in the meantime. Note nonstandard calling convention:
913 * libc stub passes mask, not pointer, to save a copyin.
914 */
915
916static int
917sigcontinue(__unused int error)
918{
919// struct uthread *ut = get_bsdthread_info(current_thread());
920 unix_syscall_return(EINTR);
921}
922
923int
924sigsuspend(proc_t p, struct sigsuspend_args *uap, int32_t *retval)
925{
926 __pthread_testcancel(1);
927 return sigsuspend_nocancel(p, (struct sigsuspend_nocancel_args *)uap, retval);
928}
929
930int
931sigsuspend_nocancel(proc_t p, struct sigsuspend_nocancel_args *uap, __unused int32_t *retval)
932{
933 struct uthread *ut;
934
935 ut = (struct uthread *)get_bsdthread_info(current_thread());
936
937 /*
938 * When returning from sigpause, we want
939 * the old mask to be restored after the
940 * signal handler has finished. Thus, we
941 * save it here and mark the sigacts structure
942 * to indicate this.
943 */
944 ut->uu_oldmask = ut->uu_sigmask;
945 ut->uu_flag |= UT_SAS_OLDMASK;
946 ut->uu_sigmask = (uap->mask & ~sigcantmask);
947 (void) tsleep0((caddr_t) p, PPAUSE | PCATCH, "pause", 0, sigcontinue);
948 /* always return EINTR rather than ERESTART... */
949 return EINTR;
950}
951
952
953int
954__disable_threadsignal(__unused proc_t p,
955 __unused struct __disable_threadsignal_args *uap,
956 __unused int32_t *retval)
957{
958 struct uthread *uth;
959
960 uth = (struct uthread *)get_bsdthread_info(current_thread());
961
962 /* No longer valid to have any signal delivered */
963 uth->uu_flag |= (UT_NO_SIGMASK | UT_CANCELDISABLE);
964
965 return 0;
966}
967
968void
969__pthread_testcancel(int presyscall)
970{
971 thread_t self = current_thread();
972 struct uthread * uthread;
973
974 uthread = (struct uthread *)get_bsdthread_info(self);
975
976
977 uthread->uu_flag &= ~UT_NOTCANCELPT;
978
979 if ((uthread->uu_flag & (UT_CANCELDISABLE | UT_CANCEL | UT_CANCELED)) == UT_CANCEL) {
980 if (presyscall != 0) {
981 unix_syscall_return(EINTR);
982 /* NOTREACHED */
983 } else {
984 thread_abort_safely(self);
985 }
986 }
987}
988
989
990
991int
992__pthread_markcancel(__unused proc_t p,
993 struct __pthread_markcancel_args *uap, __unused int32_t *retval)
994{
995 thread_act_t target_act;
996 int error = 0;
997 struct uthread *uth;
998
999 target_act = (thread_act_t)port_name_to_thread(uap->thread_port,
1000 PORT_TO_THREAD_IN_CURRENT_TASK);
1001
1002 if (target_act == THR_ACT_NULL) {
1003 return ESRCH;
1004 }
1005
1006 uth = (struct uthread *)get_bsdthread_info(target_act);
1007
1008 /* if the thread is in vfork do not cancel */
1009 if ((uth->uu_flag & (UT_VFORK | UT_CANCEL | UT_CANCELED)) == 0) {
1010 uth->uu_flag |= (UT_CANCEL | UT_NO_SIGMASK);
1011 if (((uth->uu_flag & UT_NOTCANCELPT) == 0)
1012 && ((uth->uu_flag & UT_CANCELDISABLE) == 0)) {
1013 thread_abort_safely(target_act);
1014 }
1015 }
1016
1017 thread_deallocate(target_act);
1018 return error;
1019}
1020
1021/* if action =0 ; return the cancellation state ,
1022 * if marked for cancellation, make the thread canceled
1023 * if action = 1 ; Enable the cancel handling
1024 * if action = 2; Disable the cancel handling
1025 */
1026int
1027__pthread_canceled(__unused proc_t p,
1028 struct __pthread_canceled_args *uap, __unused int32_t *retval)
1029{
1030 thread_act_t thread;
1031 struct uthread *uth;
1032 int action = uap->action;
1033
1034 thread = current_thread();
1035 uth = (struct uthread *)get_bsdthread_info(thread);
1036
1037 switch (action) {
1038 case 1:
1039 uth->uu_flag &= ~UT_CANCELDISABLE;
1040 return 0;
1041 case 2:
1042 uth->uu_flag |= UT_CANCELDISABLE;
1043 return 0;
1044 case 0:
1045 default:
1046 /* if the thread is in vfork do not cancel */
1047 if ((uth->uu_flag & (UT_CANCELDISABLE | UT_CANCEL | UT_CANCELED)) == UT_CANCEL) {
1048 uth->uu_flag &= ~UT_CANCEL;
1049 uth->uu_flag |= (UT_CANCELED | UT_NO_SIGMASK);
1050 return 0;
1051 }
1052 return EINVAL;
1053 }
1054 return EINVAL;
1055}
1056
1057__attribute__((noreturn))
1058void
1059__posix_sem_syscall_return(kern_return_t kern_result)
1060{
1061 int error = 0;
1062
1063 if (kern_result == KERN_SUCCESS) {
1064 error = 0;
1065 } else if (kern_result == KERN_ABORTED) {
1066 error = EINTR;
1067 } else if (kern_result == KERN_OPERATION_TIMED_OUT) {
1068 error = ETIMEDOUT;
1069 } else {
1070 error = EINVAL;
1071 }
1072 unix_syscall_return(error);
1073 /* does not return */
1074}
1075
1076#if OLD_SEMWAIT_SIGNAL
1077/*
1078 * Returns: 0 Success
1079 * EINTR
1080 * ETIMEDOUT
1081 * EINVAL
1082 * EFAULT if timespec is NULL
1083 */
1084int
1085__old_semwait_signal(proc_t p, struct __old_semwait_signal_args *uap,
1086 int32_t *retval)
1087{
1088 __pthread_testcancel(0);
1089 return __old_semwait_signal_nocancel(p, (struct __old_semwait_signal_nocancel_args *)uap, retval);
1090}
1091
1092int
1093__old_semwait_signal_nocancel(proc_t p, struct __old_semwait_signal_nocancel_args *uap,
1094 __unused int32_t *retval)
1095{
1096 kern_return_t kern_result;
1097 int error;
1098 mach_timespec_t then;
1099 struct timespec now;
1100 struct user_timespec ts;
1101 boolean_t truncated_timeout = FALSE;
1102
1103 if (uap->timeout) {
1104 if (IS_64BIT_PROCESS(p)) {
1105 struct user64_timespec ts64;
1106 error = copyin(uap->ts, &ts64, sizeof(ts64));
1107 ts.tv_sec = (user_time_t)ts64.tv_sec;
1108 ts.tv_nsec = (user_long_t)ts64.tv_nsec;
1109 } else {
1110 struct user32_timespec ts32;
1111 error = copyin(uap->ts, &ts32, sizeof(ts32));
1112 ts.tv_sec = ts32.tv_sec;
1113 ts.tv_nsec = ts32.tv_nsec;
1114 }
1115
1116 if (error) {
1117 return error;
1118 }
1119
1120 if ((ts.tv_sec & 0xFFFFFFFF00000000ULL) != 0) {
1121 ts.tv_sec = 0xFFFFFFFF;
1122 ts.tv_nsec = 0;
1123 truncated_timeout = TRUE;
1124 }
1125
1126 if (uap->relative) {
1127 then.tv_sec = (unsigned int)ts.tv_sec;
1128 then.tv_nsec = (clock_res_t)ts.tv_nsec;
1129 } else {
1130 nanotime(&now);
1131
1132 /* if time has elapsed, set time to null timepsec to bailout rightaway */
1133 if (now.tv_sec == ts.tv_sec ?
1134 now.tv_nsec > ts.tv_nsec :
1135 now.tv_sec > ts.tv_sec) {
1136 then.tv_sec = 0;
1137 then.tv_nsec = 0;
1138 } else {
1139 then.tv_sec = (unsigned int)(ts.tv_sec - now.tv_sec);
1140 then.tv_nsec = (clock_res_t)(ts.tv_nsec - now.tv_nsec);
1141 if (then.tv_nsec < 0) {
1142 then.tv_nsec += NSEC_PER_SEC;
1143 then.tv_sec--;
1144 }
1145 }
1146 }
1147
1148 if (uap->mutex_sem == 0) {
1149 kern_result = semaphore_timedwait_trap_internal((mach_port_name_t)uap->cond_sem, then.tv_sec, then.tv_nsec, __posix_sem_syscall_return);
1150 } else {
1151 kern_result = semaphore_timedwait_signal_trap_internal(uap->cond_sem, uap->mutex_sem, then.tv_sec, then.tv_nsec, __posix_sem_syscall_return);
1152 }
1153 } else {
1154 if (uap->mutex_sem == 0) {
1155 kern_result = semaphore_wait_trap_internal(uap->cond_sem, __posix_sem_syscall_return);
1156 } else {
1157 kern_result = semaphore_wait_signal_trap_internal(uap->cond_sem, uap->mutex_sem, __posix_sem_syscall_return);
1158 }
1159 }
1160
1161 if (kern_result == KERN_SUCCESS && !truncated_timeout) {
1162 return 0;
1163 } else if (kern_result == KERN_SUCCESS && truncated_timeout) {
1164 return EINTR; /* simulate an exceptional condition because Mach doesn't support a longer timeout */
1165 } else if (kern_result == KERN_ABORTED) {
1166 return EINTR;
1167 } else if (kern_result == KERN_OPERATION_TIMED_OUT) {
1168 return ETIMEDOUT;
1169 } else {
1170 return EINVAL;
1171 }
1172}
1173#endif /* OLD_SEMWAIT_SIGNAL*/
1174
1175/*
1176 * Returns: 0 Success
1177 * EINTR
1178 * ETIMEDOUT
1179 * EINVAL
1180 * EFAULT if timespec is NULL
1181 */
1182int
1183__semwait_signal(proc_t p, struct __semwait_signal_args *uap,
1184 int32_t *retval)
1185{
1186 __pthread_testcancel(0);
1187 return __semwait_signal_nocancel(p, (struct __semwait_signal_nocancel_args *)uap, retval);
1188}
1189
1190int
1191__semwait_signal_nocancel(__unused proc_t p, struct __semwait_signal_nocancel_args *uap,
1192 __unused int32_t *retval)
1193{
1194 kern_return_t kern_result;
1195 mach_timespec_t then;
1196 struct timespec now;
1197 struct user_timespec ts;
1198 boolean_t truncated_timeout = FALSE;
1199
1200 if (uap->timeout) {
1201 ts.tv_sec = (user_time_t)uap->tv_sec;
1202 ts.tv_nsec = uap->tv_nsec;
1203
1204 if ((ts.tv_sec & 0xFFFFFFFF00000000ULL) != 0) {
1205 ts.tv_sec = 0xFFFFFFFF;
1206 ts.tv_nsec = 0;
1207 truncated_timeout = TRUE;
1208 }
1209
1210 if (uap->relative) {
1211 then.tv_sec = (unsigned int)ts.tv_sec;
1212 then.tv_nsec = (clock_res_t)ts.tv_nsec;
1213 } else {
1214 nanotime(&now);
1215
1216 /* if time has elapsed, set time to null timepsec to bailout rightaway */
1217 if (now.tv_sec == ts.tv_sec ?
1218 now.tv_nsec > ts.tv_nsec :
1219 now.tv_sec > ts.tv_sec) {
1220 then.tv_sec = 0;
1221 then.tv_nsec = 0;
1222 } else {
1223 then.tv_sec = (unsigned int)(ts.tv_sec - now.tv_sec);
1224 then.tv_nsec = (clock_res_t)(ts.tv_nsec - now.tv_nsec);
1225 if (then.tv_nsec < 0) {
1226 then.tv_nsec += NSEC_PER_SEC;
1227 then.tv_sec--;
1228 }
1229 }
1230 }
1231
1232 if (uap->mutex_sem == 0) {
1233 kern_result = semaphore_timedwait_trap_internal((mach_port_name_t)uap->cond_sem, then.tv_sec, then.tv_nsec, __posix_sem_syscall_return);
1234 } else {
1235 kern_result = semaphore_timedwait_signal_trap_internal(uap->cond_sem, uap->mutex_sem, then.tv_sec, then.tv_nsec, __posix_sem_syscall_return);
1236 }
1237 } else {
1238 if (uap->mutex_sem == 0) {
1239 kern_result = semaphore_wait_trap_internal(uap->cond_sem, __posix_sem_syscall_return);
1240 } else {
1241 kern_result = semaphore_wait_signal_trap_internal(uap->cond_sem, uap->mutex_sem, __posix_sem_syscall_return);
1242 }
1243 }
1244
1245 if (kern_result == KERN_SUCCESS && !truncated_timeout) {
1246 return 0;
1247 } else if (kern_result == KERN_SUCCESS && truncated_timeout) {
1248 return EINTR; /* simulate an exceptional condition because Mach doesn't support a longer timeout */
1249 } else if (kern_result == KERN_ABORTED) {
1250 return EINTR;
1251 } else if (kern_result == KERN_OPERATION_TIMED_OUT) {
1252 return ETIMEDOUT;
1253 } else {
1254 return EINVAL;
1255 }
1256}
1257
1258
1259int
1260__pthread_kill(__unused proc_t p, struct __pthread_kill_args *uap,
1261 __unused int32_t *retval)
1262{
1263 thread_t target_act;
1264 int error = 0;
1265 int signum = uap->sig;
1266 struct uthread *uth;
1267
1268 target_act = (thread_t)port_name_to_thread(uap->thread_port,
1269 PORT_TO_THREAD_NONE);
1270
1271 if (target_act == THREAD_NULL) {
1272 return ESRCH;
1273 }
1274 if ((u_int)signum >= NSIG) {
1275 error = EINVAL;
1276 goto out;
1277 }
1278
1279 uth = (struct uthread *)get_bsdthread_info(target_act);
1280
1281 if (uth->uu_flag & UT_NO_SIGMASK) {
1282 error = ESRCH;
1283 goto out;
1284 }
1285
1286 if ((thread_get_tag(target_act) & THREAD_TAG_WORKQUEUE) && !uth->uu_workq_pthread_kill_allowed) {
1287 error = ENOTSUP;
1288 goto out;
1289 }
1290
1291 if (signum) {
1292 psignal_uthread(target_act, signum);
1293 }
1294out:
1295 thread_deallocate(target_act);
1296 return error;
1297}
1298
1299
1300int
1301__pthread_sigmask(__unused proc_t p, struct __pthread_sigmask_args *uap,
1302 __unused int32_t *retval)
1303{
1304 user_addr_t set = uap->set;
1305 user_addr_t oset = uap->oset;
1306 sigset_t nset;
1307 int error = 0;
1308 struct uthread *ut;
1309 sigset_t oldset;
1310
1311 ut = (struct uthread *)get_bsdthread_info(current_thread());
1312 oldset = ut->uu_sigmask;
1313
1314 if (set == USER_ADDR_NULL) {
1315 /* need only old mask */
1316 goto out;
1317 }
1318
1319 error = copyin(set, &nset, sizeof(sigset_t));
1320 if (error) {
1321 goto out;
1322 }
1323
1324 switch (uap->how) {
1325 case SIG_BLOCK:
1326 ut->uu_sigmask |= (nset & ~sigcantmask);
1327 break;
1328
1329 case SIG_UNBLOCK:
1330 ut->uu_sigmask &= ~(nset);
1331 signal_setast(current_thread());
1332 break;
1333
1334 case SIG_SETMASK:
1335 ut->uu_sigmask = (nset & ~sigcantmask);
1336 signal_setast(current_thread());
1337 break;
1338
1339 default:
1340 error = EINVAL;
1341 }
1342out:
1343 if (!error && oset != USER_ADDR_NULL) {
1344 copyout(&oldset, oset, sizeof(sigset_t));
1345 }
1346
1347 return error;
1348}
1349
1350/*
1351 * Returns: 0 Success
1352 * EINVAL
1353 * copyin:EFAULT
1354 * copyout:EFAULT
1355 */
1356int
1357__sigwait(proc_t p, struct __sigwait_args *uap, int32_t *retval)
1358{
1359 __pthread_testcancel(1);
1360 return __sigwait_nocancel(p, (struct __sigwait_nocancel_args *)uap, retval);
1361}
1362
1363int
1364__sigwait_nocancel(proc_t p, struct __sigwait_nocancel_args *uap, __unused int32_t *retval)
1365{
1366 struct uthread *ut;
1367 struct uthread *uth;
1368 int error = 0;
1369 sigset_t mask;
1370 sigset_t siglist;
1371 sigset_t sigw = 0;
1372 int signum;
1373
1374 ut = (struct uthread *)get_bsdthread_info(current_thread());
1375
1376 if (uap->set == USER_ADDR_NULL) {
1377 return EINVAL;
1378 }
1379
1380 error = copyin(uap->set, &mask, sizeof(sigset_t));
1381 if (error) {
1382 return error;
1383 }
1384
1385 siglist = (mask & ~sigcantmask);
1386
1387 if (siglist == 0) {
1388 return EINVAL;
1389 }
1390
1391 proc_lock(p);
1392 if ((p->p_lflag & P_LINVFORK) && p->p_vforkact) {
1393 proc_unlock(p);
1394 return EINVAL;
1395 } else {
1396 proc_signalstart(p, 1);
1397 TAILQ_FOREACH(uth, &p->p_uthlist, uu_list) {
1398 if ((sigw = uth->uu_siglist & siglist)) {
1399 break;
1400 }
1401 }
1402 proc_signalend(p, 1);
1403 }
1404
1405 if (sigw) {
1406 /* The signal was pending on a thread */
1407 goto sigwait1;
1408 }
1409 /*
1410 * When returning from sigwait, we want
1411 * the old mask to be restored after the
1412 * signal handler has finished. Thus, we
1413 * save it here and mark the sigacts structure
1414 * to indicate this.
1415 */
1416 uth = ut; /* wait for it to be delivered to us */
1417 ut->uu_oldmask = ut->uu_sigmask;
1418 ut->uu_flag |= UT_SAS_OLDMASK;
1419 if (siglist == (sigset_t)0) {
1420 proc_unlock(p);
1421 return EINVAL;
1422 }
1423 /* SIGKILL and SIGSTOP are not maskable as well */
1424 ut->uu_sigmask = ~(siglist | sigcantmask);
1425 ut->uu_sigwait = siglist;
1426
1427 /* No Continuations for now */
1428 error = msleep((caddr_t)&ut->uu_sigwait, &p->p_mlock, PPAUSE | PCATCH, "pause", 0);
1429
1430 if (error == ERESTART) {
1431 error = 0;
1432 }
1433
1434 sigw = (ut->uu_sigwait & siglist);
1435 ut->uu_sigmask = ut->uu_oldmask;
1436 ut->uu_oldmask = 0;
1437 ut->uu_flag &= ~UT_SAS_OLDMASK;
1438sigwait1:
1439 ut->uu_sigwait = 0;
1440 if (!error) {
1441 signum = ffs((unsigned int)sigw);
1442 if (!signum) {
1443 panic("sigwait with no signal wakeup");
1444 }
1445 /* Clear the pending signal in the thread it was delivered */
1446 uth->uu_siglist &= ~(sigmask(signum));
1447
1448#if CONFIG_DTRACE
1449 DTRACE_PROC2(signal__clear, int, signum, siginfo_t *, &(ut->t_dtrace_siginfo));
1450#endif
1451
1452 proc_unlock(p);
1453 if (uap->sig != USER_ADDR_NULL) {
1454 error = copyout(&signum, uap->sig, sizeof(int));
1455 }
1456 } else {
1457 proc_unlock(p);
1458 }
1459
1460 return error;
1461}
1462
1463int
1464sigaltstack(__unused proc_t p, struct sigaltstack_args *uap, __unused int32_t *retval)
1465{
1466 struct kern_sigaltstack ss;
1467 struct kern_sigaltstack *pstk;
1468 int error;
1469 struct uthread *uth;
1470 int onstack;
1471
1472 uth = (struct uthread *)get_bsdthread_info(current_thread());
1473
1474 pstk = &uth->uu_sigstk;
1475 if ((uth->uu_flag & UT_ALTSTACK) == 0) {
1476 uth->uu_sigstk.ss_flags |= SA_DISABLE;
1477 }
1478 onstack = pstk->ss_flags & SA_ONSTACK;
1479 if (uap->oss) {
1480 if (IS_64BIT_PROCESS(p)) {
1481 struct user64_sigaltstack ss64 = {};
1482 sigaltstack_kern_to_user64(pstk, &ss64);
1483 error = copyout(&ss64, uap->oss, sizeof(ss64));
1484 } else {
1485 struct user32_sigaltstack ss32 = {};
1486 sigaltstack_kern_to_user32(pstk, &ss32);
1487 error = copyout(&ss32, uap->oss, sizeof(ss32));
1488 }
1489 if (error) {
1490 return error;
1491 }
1492 }
1493 if (uap->nss == USER_ADDR_NULL) {
1494 return 0;
1495 }
1496 if (IS_64BIT_PROCESS(p)) {
1497 struct user64_sigaltstack ss64;
1498 error = copyin(uap->nss, &ss64, sizeof(ss64));
1499 sigaltstack_user64_to_kern(&ss64, &ss);
1500 } else {
1501 struct user32_sigaltstack ss32;
1502 error = copyin(uap->nss, &ss32, sizeof(ss32));
1503 sigaltstack_user32_to_kern(&ss32, &ss);
1504 }
1505 if (error) {
1506 return error;
1507 }
1508 if ((ss.ss_flags & ~SA_DISABLE) != 0) {
1509 return EINVAL;
1510 }
1511
1512 if (ss.ss_flags & SA_DISABLE) {
1513 /* if we are here we are not in the signal handler ;so no need to check */
1514 if (uth->uu_sigstk.ss_flags & SA_ONSTACK) {
1515 return EINVAL;
1516 }
1517 uth->uu_flag &= ~UT_ALTSTACK;
1518 uth->uu_sigstk.ss_flags = ss.ss_flags;
1519 return 0;
1520 }
1521 if (onstack) {
1522 return EPERM;
1523 }
1524/* The older stacksize was 8K, enforce that one so no compat problems */
1525#define OLDMINSIGSTKSZ 8*1024
1526 if (ss.ss_size < OLDMINSIGSTKSZ) {
1527 return ENOMEM;
1528 }
1529 uth->uu_flag |= UT_ALTSTACK;
1530 uth->uu_sigstk = ss;
1531 return 0;
1532}
1533
1534int
1535kill(proc_t cp, struct kill_args *uap, __unused int32_t *retval)
1536{
1537 proc_t p;
1538 kauth_cred_t uc = kauth_cred_get();
1539 int posix = uap->posix; /* !0 if posix behaviour desired */
1540
1541 AUDIT_ARG(pid, uap->pid);
1542 AUDIT_ARG(signum, uap->signum);
1543
1544 if ((u_int)uap->signum >= NSIG) {
1545 return EINVAL;
1546 }
1547 if (uap->pid > 0) {
1548 /* kill single process */
1549 if ((p = proc_find(uap->pid)) == NULL) {
1550 if ((p = pzfind(uap->pid)) != NULL) {
1551 /*
1552 * POSIX 1003.1-2001 requires returning success when killing a
1553 * zombie; see Rationale for kill(2).
1554 */
1555 return 0;
1556 }
1557 return ESRCH;
1558 }
1559 AUDIT_ARG(process, p);
1560 if (!cansignal(cp, uc, p, uap->signum)) {
1561 proc_rele(p);
1562 return EPERM;
1563 }
1564 if (uap->signum) {
1565 psignal(p, uap->signum);
1566 }
1567 proc_rele(p);
1568 return 0;
1569 }
1570 switch (uap->pid) {
1571 case -1: /* broadcast signal */
1572 return killpg1(cp, uap->signum, 0, 1, posix);
1573 case 0: /* signal own process group */
1574 return killpg1(cp, uap->signum, 0, 0, posix);
1575 default: /* negative explicit process group */
1576 return killpg1(cp, uap->signum, -(uap->pid), 0, posix);
1577 }
1578 /* NOTREACHED */
1579}
1580
1581os_reason_t
1582build_userspace_exit_reason(uint32_t reason_namespace, uint64_t reason_code, user_addr_t payload, uint32_t payload_size,
1583 user_addr_t reason_string, uint64_t reason_flags)
1584{
1585 os_reason_t exit_reason = OS_REASON_NULL;
1586
1587 int error = 0;
1588 int num_items_to_copy = 0;
1589 uint32_t user_data_to_copy = 0;
1590 char *reason_user_desc = NULL;
1591 size_t reason_user_desc_len = 0;
1592
1593 exit_reason = os_reason_create(reason_namespace, reason_code);
1594 if (exit_reason == OS_REASON_NULL) {
1595 printf("build_userspace_exit_reason: failed to allocate exit reason\n");
1596 return exit_reason;
1597 }
1598
1599 exit_reason->osr_flags |= OS_REASON_FLAG_FROM_USERSPACE;
1600
1601 /*
1602 * Only apply flags that are allowed to be passed from userspace.
1603 */
1604 exit_reason->osr_flags |= (reason_flags & OS_REASON_FLAG_MASK_ALLOWED_FROM_USER);
1605 if ((reason_flags & OS_REASON_FLAG_MASK_ALLOWED_FROM_USER) != reason_flags) {
1606 printf("build_userspace_exit_reason: illegal flags passed from userspace (some masked off) 0x%llx, ns: %u, code 0x%llx\n",
1607 reason_flags, reason_namespace, reason_code);
1608 }
1609
1610 if (!(exit_reason->osr_flags & OS_REASON_FLAG_NO_CRASH_REPORT)) {
1611 exit_reason->osr_flags |= OS_REASON_FLAG_GENERATE_CRASH_REPORT;
1612 }
1613
1614 if (payload != USER_ADDR_NULL) {
1615 if (payload_size == 0) {
1616 printf("build_userspace_exit_reason: exit reason with namespace %u, nonzero payload but zero length\n",
1617 reason_namespace);
1618 exit_reason->osr_flags |= OS_REASON_FLAG_BAD_PARAMS;
1619 payload = USER_ADDR_NULL;
1620 } else {
1621 num_items_to_copy++;
1622
1623 if (payload_size > EXIT_REASON_PAYLOAD_MAX_LEN) {
1624 exit_reason->osr_flags |= OS_REASON_FLAG_PAYLOAD_TRUNCATED;
1625 payload_size = EXIT_REASON_PAYLOAD_MAX_LEN;
1626 }
1627
1628 user_data_to_copy += payload_size;
1629 }
1630 }
1631
1632 if (reason_string != USER_ADDR_NULL) {
1633 reason_user_desc = kheap_alloc(KHEAP_TEMP,
1634 EXIT_REASON_USER_DESC_MAX_LEN, Z_WAITOK);
1635
1636 if (reason_user_desc != NULL) {
1637 error = copyinstr(reason_string, (void *) reason_user_desc,
1638 EXIT_REASON_USER_DESC_MAX_LEN, &reason_user_desc_len);
1639
1640 if (error == 0) {
1641 num_items_to_copy++;
1642 user_data_to_copy += reason_user_desc_len;
1643 } else if (error == ENAMETOOLONG) {
1644 num_items_to_copy++;
1645 reason_user_desc[EXIT_REASON_USER_DESC_MAX_LEN - 1] = '\0';
1646 user_data_to_copy += reason_user_desc_len;
1647 } else {
1648 exit_reason->osr_flags |= OS_REASON_FLAG_FAILED_DATA_COPYIN;
1649 kheap_free(KHEAP_TEMP, reason_user_desc,
1650 EXIT_REASON_USER_DESC_MAX_LEN);
1651 reason_user_desc = NULL;
1652 reason_user_desc_len = 0;
1653 }
1654 }
1655 }
1656
1657 if (num_items_to_copy != 0) {
1658 uint32_t reason_buffer_size_estimate = 0;
1659 mach_vm_address_t data_addr = 0;
1660
1661 reason_buffer_size_estimate = kcdata_estimate_required_buffer_size(num_items_to_copy, user_data_to_copy);
1662
1663 error = os_reason_alloc_buffer(exit_reason, reason_buffer_size_estimate);
1664 if (error != 0) {
1665 printf("build_userspace_exit_reason: failed to allocate signal reason buffer\n");
1666 goto out_failed_copyin;
1667 }
1668
1669 if (reason_user_desc != NULL && reason_user_desc_len != 0) {
1670 if (KERN_SUCCESS == kcdata_get_memory_addr(&exit_reason->osr_kcd_descriptor,
1671 EXIT_REASON_USER_DESC,
1672 (uint32_t)reason_user_desc_len,
1673 &data_addr)) {
1674 kcdata_memcpy(&exit_reason->osr_kcd_descriptor, (mach_vm_address_t) data_addr,
1675 reason_user_desc, (uint32_t)reason_user_desc_len);
1676 } else {
1677 printf("build_userspace_exit_reason: failed to allocate space for reason string\n");
1678 goto out_failed_copyin;
1679 }
1680 }
1681
1682 if (payload != USER_ADDR_NULL) {
1683 if (KERN_SUCCESS ==
1684 kcdata_get_memory_addr(&exit_reason->osr_kcd_descriptor,
1685 EXIT_REASON_USER_PAYLOAD,
1686 payload_size,
1687 &data_addr)) {
1688 error = copyin(payload, (void *) data_addr, payload_size);
1689 if (error) {
1690 printf("build_userspace_exit_reason: failed to copy in payload data with error %d\n", error);
1691 goto out_failed_copyin;
1692 }
1693 } else {
1694 printf("build_userspace_exit_reason: failed to allocate space for payload data\n");
1695 goto out_failed_copyin;
1696 }
1697 }
1698 }
1699
1700 if (reason_user_desc != NULL) {
1701 kheap_free(KHEAP_TEMP, reason_user_desc, EXIT_REASON_USER_DESC_MAX_LEN);
1702 reason_user_desc = NULL;
1703 reason_user_desc_len = 0;
1704 }
1705
1706 return exit_reason;
1707
1708out_failed_copyin:
1709
1710 if (reason_user_desc != NULL) {
1711 kheap_free(KHEAP_TEMP, reason_user_desc, EXIT_REASON_USER_DESC_MAX_LEN);
1712 reason_user_desc = NULL;
1713 reason_user_desc_len = 0;
1714 }
1715
1716 exit_reason->osr_flags |= OS_REASON_FLAG_FAILED_DATA_COPYIN;
1717 os_reason_alloc_buffer(exit_reason, 0);
1718 return exit_reason;
1719}
1720
1721static int
1722terminate_with_payload_internal(struct proc *cur_proc, int target_pid, uint32_t reason_namespace,
1723 uint64_t reason_code, user_addr_t payload, uint32_t payload_size,
1724 user_addr_t reason_string, uint64_t reason_flags)
1725{
1726 proc_t target_proc = PROC_NULL;
1727 kauth_cred_t cur_cred = kauth_cred_get();
1728
1729 os_reason_t signal_reason = OS_REASON_NULL;
1730
1731 AUDIT_ARG(pid, target_pid);
1732 if ((target_pid <= 0)) {
1733 return EINVAL;
1734 }
1735
1736 target_proc = proc_find(target_pid);
1737 if (target_proc == PROC_NULL) {
1738 return ESRCH;
1739 }
1740
1741 AUDIT_ARG(process, target_proc);
1742
1743 if (!cansignal(cur_proc, cur_cred, target_proc, SIGKILL)) {
1744 proc_rele(target_proc);
1745 return EPERM;
1746 }
1747
1748 if (target_pid != cur_proc->p_pid) {
1749 /*
1750 * FLAG_ABORT should only be set on terminate_with_reason(getpid()) that
1751 * was a fallback from an unsuccessful abort_with_reason(). In that case
1752 * caller's pid matches the target one. Otherwise remove the flag.
1753 */
1754 reason_flags &= ~((typeof(reason_flags))OS_REASON_FLAG_ABORT);
1755 }
1756
1757 KERNEL_DEBUG_CONSTANT(BSDDBG_CODE(DBG_BSD_PROC, BSD_PROC_EXITREASON_CREATE) | DBG_FUNC_NONE,
1758 target_proc->p_pid, reason_namespace,
1759 reason_code, 0, 0);
1760
1761 signal_reason = build_userspace_exit_reason(reason_namespace, reason_code, payload, payload_size,
1762 reason_string, (reason_flags | OS_REASON_FLAG_NO_CRASHED_TID));
1763
1764 if (target_pid == cur_proc->p_pid) {
1765 /*
1766 * psignal_thread_with_reason() will pend a SIGKILL on the specified thread or
1767 * return if the thread and/or task are already terminating. Either way, the
1768 * current thread won't return to userspace.
1769 */
1770 psignal_thread_with_reason(target_proc, current_thread(), SIGKILL, signal_reason);
1771 } else {
1772 psignal_with_reason(target_proc, SIGKILL, signal_reason);
1773 }
1774
1775 proc_rele(target_proc);
1776
1777 return 0;
1778}
1779
1780int
1781terminate_with_payload(struct proc *cur_proc, struct terminate_with_payload_args *args,
1782 __unused int32_t *retval)
1783{
1784 return terminate_with_payload_internal(cur_proc, args->pid, args->reason_namespace, args->reason_code, args->payload,
1785 args->payload_size, args->reason_string, args->reason_flags);
1786}
1787
1788static int
1789killpg1_allfilt(proc_t p, void * arg)
1790{
1791 struct killpg1_filtargs * kfargp = (struct killpg1_filtargs *)arg;
1792
1793 /*
1794 * Don't signal initproc, a system process, or the current process if POSIX
1795 * isn't specified.
1796 */
1797 return p->p_pid > 1 && !(p->p_flag & P_SYSTEM) &&
1798 (kfargp->posix ? true : p != kfargp->curproc);
1799}
1800
1801static int
1802killpg1_pgrpfilt(proc_t p, __unused void * arg)
1803{
1804 /* XXX shouldn't this allow signalling zombies? */
1805 return p->p_pid > 1 && !(p->p_flag & P_SYSTEM) && p->p_stat != SZOMB;
1806}
1807
1808static int
1809killpg1_callback(proc_t p, void *arg)
1810{
1811 struct killpg1_iterargs *kargp = (struct killpg1_iterargs *)arg;
1812 int signum = kargp->signum;
1813
1814 if ((p->p_listflag & P_LIST_EXITED) == P_LIST_EXITED) {
1815 /*
1816 * Count zombies as found for the purposes of signalling, since POSIX
1817 * 1003.1-2001 sees signalling zombies as successful. If killpg(2) or
1818 * kill(2) with pid -1 only finds zombies that can be signalled, it
1819 * shouldn't return ESRCH. See the Rationale for kill(2).
1820 *
1821 * Don't call into MAC -- it's not expecting signal checks for exited
1822 * processes.
1823 */
1824 if (cansignal_nomac(kargp->curproc, kargp->uc, p, signum)) {
1825 kargp->nfound++;
1826 }
1827 } else if (cansignal(kargp->curproc, kargp->uc, p, signum)) {
1828 kargp->nfound++;
1829
1830 if (signum != 0) {
1831 psignal(p, signum);
1832 }
1833 }
1834
1835 return PROC_RETURNED;
1836}
1837
1838/*
1839 * Common code for kill process group/broadcast kill.
1840 */
1841int
1842killpg1(proc_t curproc, int signum, int pgid, int all, int posix)
1843{
1844 kauth_cred_t uc;
1845 struct pgrp *pgrp;
1846 int error = 0;
1847
1848 uc = kauth_cred_proc_ref(curproc);
1849 struct killpg1_iterargs karg = {
1850 .curproc = curproc, .uc = uc, .nfound = 0, .signum = signum
1851 };
1852
1853 if (all) {
1854 /*
1855 * Broadcast to all processes that the user can signal (pid was -1).
1856 */
1857 struct killpg1_filtargs kfarg = {
1858 .posix = posix, .curproc = curproc
1859 };
1860 proc_iterate(PROC_ALLPROCLIST | PROC_ZOMBPROCLIST, killpg1_callback,
1861 &karg, killpg1_allfilt, &kfarg);
1862 } else {
1863 if (pgid == 0) {
1864 /*
1865 * Send to current the current process' process group.
1866 */
1867 pgrp = proc_pgrp(curproc);
1868 } else {
1869 pgrp = pgfind(pgid);
1870 if (pgrp == NULL) {
1871 error = ESRCH;
1872 goto out;
1873 }
1874 }
1875
1876 /* PGRP_DROPREF drops the pgrp refernce */
1877 pgrp_iterate(pgrp, PGRP_DROPREF, killpg1_callback, &karg,
1878 killpg1_pgrpfilt, NULL);
1879 }
1880 error = (karg.nfound > 0 ? 0 : (posix ? EPERM : ESRCH));
1881out:
1882 kauth_cred_unref(&uc);
1883 return error;
1884}
1885
1886/*
1887 * Send a signal to a process group.
1888 */
1889void
1890gsignal(int pgid, int signum)
1891{
1892 struct pgrp *pgrp;
1893
1894 if (pgid && (pgrp = pgfind(pgid))) {
1895 pgsignal(pgrp, signum, 0);
1896 pg_rele(pgrp);
1897 }
1898}
1899
1900/*
1901 * Send a signal to a process group. If checkctty is 1,
1902 * limit to members which have a controlling terminal.
1903 */
1904
1905static int
1906pgsignal_filt(proc_t p, void * arg)
1907{
1908 int checkctty = *(int*)arg;
1909
1910 if ((checkctty == 0) || p->p_flag & P_CONTROLT) {
1911 return 1;
1912 } else {
1913 return 0;
1914 }
1915}
1916
1917
1918static int
1919pgsignal_callback(proc_t p, void * arg)
1920{
1921 int signum = *(int*)arg;
1922
1923 psignal(p, signum);
1924 return PROC_RETURNED;
1925}
1926
1927
1928void
1929pgsignal(struct pgrp *pgrp, int signum, int checkctty)
1930{
1931 if (pgrp != PGRP_NULL) {
1932 pgrp_iterate(pgrp, 0, pgsignal_callback, &signum, pgsignal_filt, &checkctty);
1933 }
1934}
1935
1936
1937void
1938tty_pgsignal(struct tty *tp, int signum, int checkctty)
1939{
1940 struct pgrp * pg;
1941
1942 pg = tty_pgrp(tp);
1943 if (pg != PGRP_NULL) {
1944 pgrp_iterate(pg, 0, pgsignal_callback, &signum, pgsignal_filt, &checkctty);
1945 pg_rele(pg);
1946 }
1947}
1948/*
1949 * Send a signal caused by a trap to a specific thread.
1950 */
1951void
1952threadsignal(thread_t sig_actthread, int signum, mach_exception_code_t code, boolean_t set_exitreason)
1953{
1954 struct uthread *uth;
1955 struct task * sig_task;
1956 proc_t p;
1957 int mask;
1958
1959 if ((u_int)signum >= NSIG || signum == 0) {
1960 return;
1961 }
1962
1963 mask = sigmask(signum);
1964 if ((mask & threadmask) == 0) {
1965 return;
1966 }
1967 sig_task = get_threadtask(sig_actthread);
1968 p = (proc_t)(get_bsdtask_info(sig_task));
1969
1970 uth = get_bsdthread_info(sig_actthread);
1971 if (uth->uu_flag & UT_VFORK) {
1972 p = uth->uu_proc;
1973 }
1974
1975 proc_lock(p);
1976 if (!(p->p_lflag & P_LTRACED) && (p->p_sigignore & mask)) {
1977 proc_unlock(p);
1978 return;
1979 }
1980
1981 uth->uu_siglist |= mask;
1982 uth->uu_code = code;
1983
1984 /* Attempt to establish whether the signal will be fatal (mirrors logic in psignal_internal()) */
1985 if (set_exitreason && ((p->p_lflag & P_LTRACED) || (!(uth->uu_sigwait & mask)
1986 && !(uth->uu_sigmask & mask) && !(p->p_sigcatch & mask))) &&
1987 !(mask & stopsigmask) && !(mask & contsigmask)) {
1988 if (uth->uu_exit_reason == OS_REASON_NULL) {
1989 KERNEL_DEBUG_CONSTANT(BSDDBG_CODE(DBG_BSD_PROC, BSD_PROC_EXITREASON_CREATE) | DBG_FUNC_NONE,
1990 p->p_pid, OS_REASON_SIGNAL, signum, 0, 0);
1991
1992 os_reason_t signal_reason = build_signal_reason(signum, "exc handler");
1993
1994 set_thread_exit_reason(sig_actthread, signal_reason, TRUE);
1995
1996 /* We dropped/consumed the reference in set_thread_exit_reason() */
1997 signal_reason = OS_REASON_NULL;
1998 }
1999 }
2000
2001 proc_unlock(p);
2002
2003 /* mark on process as well */
2004 signal_setast(sig_actthread);
2005}
2006
2007/* Called with proc locked */
2008static void
2009set_thread_extra_flags(struct uthread *uth, os_reason_t reason)
2010{
2011 extern int vm_shared_region_reslide_restrict;
2012 assert(uth != NULL);
2013 /*
2014 * Check whether the userland fault address falls within the shared
2015 * region and notify userland if so. This allows launchd to apply
2016 * special policies around this fault type.
2017 */
2018 if (reason->osr_namespace == OS_REASON_SIGNAL &&
2019 reason->osr_code == SIGSEGV) {
2020 mach_vm_address_t fault_address = uth->uu_subcode;
2021
2022#if defined(__arm64__)
2023 /* taken from osfmk/arm/misc_protos.h */
2024 #define TBI_MASK 0xff00000000000000
2025 #define tbi_clear(addr) ((addr) & ~(TBI_MASK))
2026 fault_address = tbi_clear(fault_address);
2027#endif /* __arm64__ */
2028
2029 if (fault_address >= SHARED_REGION_BASE &&
2030 fault_address <= SHARED_REGION_BASE + SHARED_REGION_SIZE) {
2031 /*
2032 * Always report whether the fault happened within the shared cache
2033 * region, but only stale the slide if the resliding is extended
2034 * to all processes or if the process faulting is a platform one.
2035 */
2036 reason->osr_flags |= OS_REASON_FLAG_SHAREDREGION_FAULT;
2037
2038#if __has_feature(ptrauth_calls)
2039 if (!vm_shared_region_reslide_restrict || csproc_get_platform_binary(current_proc())) {
2040 vm_shared_region_reslide_stale();
2041 }
2042#endif /* __has_feature(ptrauth_calls) */
2043 }
2044 }
2045}
2046
2047void
2048set_thread_exit_reason(void *th, void *reason, boolean_t proc_locked)
2049{
2050 struct uthread *targ_uth = get_bsdthread_info(th);
2051 struct task *targ_task = NULL;
2052 proc_t targ_proc = NULL;
2053
2054 os_reason_t exit_reason = (os_reason_t)reason;
2055
2056 if (exit_reason == OS_REASON_NULL) {
2057 return;
2058 }
2059
2060 if (!proc_locked) {
2061 targ_task = get_threadtask(th);
2062 targ_proc = (proc_t)(get_bsdtask_info(targ_task));
2063
2064 proc_lock(targ_proc);
2065 }
2066
2067 set_thread_extra_flags(targ_uth, exit_reason);
2068
2069 if (targ_uth->uu_exit_reason == OS_REASON_NULL) {
2070 targ_uth->uu_exit_reason = exit_reason;
2071 } else {
2072 /* The caller expects that we drop a reference on the exit reason */
2073 os_reason_free(exit_reason);
2074 }
2075
2076 if (!proc_locked) {
2077 assert(targ_proc != NULL);
2078 proc_unlock(targ_proc);
2079 }
2080}
2081
2082/*
2083 * get_signalthread
2084 *
2085 * Picks an appropriate thread from a process to target with a signal.
2086 *
2087 * Called with proc locked.
2088 * Returns thread with BSD ast set.
2089 *
2090 * We attempt to deliver a proc-wide signal to the first thread in the task.
2091 * This allows single threaded applications which use signals to
2092 * be able to be linked with multithreaded libraries.
2093 */
2094static kern_return_t
2095get_signalthread(proc_t p, int signum, thread_t * thr)
2096{
2097 struct uthread *uth;
2098 sigset_t mask = sigmask(signum);
2099 thread_t sig_thread;
2100 struct task * sig_task = p->task;
2101 kern_return_t kret;
2102 bool skip_wqthreads = true;
2103
2104 *thr = THREAD_NULL;
2105
2106 if ((p->p_lflag & P_LINVFORK) && p->p_vforkact) {
2107 sig_thread = p->p_vforkact;
2108 kret = check_actforsig(sig_task, sig_thread, 1);
2109 if (kret == KERN_SUCCESS) {
2110 *thr = sig_thread;
2111 return KERN_SUCCESS;
2112 } else {
2113 return KERN_FAILURE;
2114 }
2115 }
2116
2117again:
2118 TAILQ_FOREACH(uth, &p->p_uthlist, uu_list) {
2119 if (((uth->uu_flag & UT_NO_SIGMASK) == 0) &&
2120 (((uth->uu_sigmask & mask) == 0) || (uth->uu_sigwait & mask))) {
2121 thread_t th = uth->uu_context.vc_thread;
2122 if (skip_wqthreads && (thread_get_tag(th) & THREAD_TAG_WORKQUEUE)) {
2123 /* Workqueue threads may be parked in the kernel unable to
2124 * deliver signals for an extended period of time, so skip them
2125 * in favor of pthreads in a first pass. (rdar://50054475). */
2126 } else if (check_actforsig(p->task, th, 1) == KERN_SUCCESS) {
2127 *thr = th;
2128 return KERN_SUCCESS;
2129 }
2130 }
2131 }
2132 if (skip_wqthreads) {
2133 skip_wqthreads = false;
2134 goto again;
2135 }
2136 if (get_signalact(p->task, thr, 1) == KERN_SUCCESS) {
2137 return KERN_SUCCESS;
2138 }
2139
2140 return KERN_FAILURE;
2141}
2142
2143static os_reason_t
2144build_signal_reason(int signum, const char *procname)
2145{
2146 os_reason_t signal_reason = OS_REASON_NULL;
2147 proc_t sender_proc = current_proc();
2148 uint32_t reason_buffer_size_estimate = 0, proc_name_length = 0;
2149 const char *default_sender_procname = "unknown";
2150 mach_vm_address_t data_addr;
2151 int ret;
2152
2153 signal_reason = os_reason_create(OS_REASON_SIGNAL, signum);
2154 if (signal_reason == OS_REASON_NULL) {
2155 printf("build_signal_reason: unable to allocate signal reason structure.\n");
2156 return signal_reason;
2157 }
2158
2159 reason_buffer_size_estimate = kcdata_estimate_required_buffer_size(2, sizeof(sender_proc->p_name) +
2160 sizeof(sender_proc->p_pid));
2161
2162 ret = os_reason_alloc_buffer_noblock(signal_reason, reason_buffer_size_estimate);
2163 if (ret != 0) {
2164 printf("build_signal_reason: unable to allocate signal reason buffer.\n");
2165 return signal_reason;
2166 }
2167
2168 if (KERN_SUCCESS == kcdata_get_memory_addr(&signal_reason->osr_kcd_descriptor, KCDATA_TYPE_PID,
2169 sizeof(sender_proc->p_pid), &data_addr)) {
2170 kcdata_memcpy(&signal_reason->osr_kcd_descriptor, data_addr, &sender_proc->p_pid,
2171 sizeof(sender_proc->p_pid));
2172 } else {
2173 printf("build_signal_reason: exceeded space in signal reason buf, unable to log PID\n");
2174 }
2175
2176 proc_name_length = sizeof(sender_proc->p_name);
2177 if (KERN_SUCCESS == kcdata_get_memory_addr(&signal_reason->osr_kcd_descriptor, KCDATA_TYPE_PROCNAME,
2178 proc_name_length, &data_addr)) {
2179 if (procname) {
2180 char truncated_procname[proc_name_length];
2181 strncpy((char *) &truncated_procname, procname, proc_name_length);
2182 truncated_procname[proc_name_length - 1] = '\0';
2183
2184 kcdata_memcpy(&signal_reason->osr_kcd_descriptor, data_addr, truncated_procname,
2185 (uint32_t)strlen((char *) &truncated_procname));
2186 } else if (*sender_proc->p_name) {
2187 kcdata_memcpy(&signal_reason->osr_kcd_descriptor, data_addr, &sender_proc->p_name,
2188 sizeof(sender_proc->p_name));
2189 } else {
2190 kcdata_memcpy(&signal_reason->osr_kcd_descriptor, data_addr, &default_sender_procname,
2191 (uint32_t)strlen(default_sender_procname) + 1);
2192 }
2193 } else {
2194 printf("build_signal_reason: exceeded space in signal reason buf, unable to log procname\n");
2195 }
2196
2197 return signal_reason;
2198}
2199
2200/*
2201 * Send the signal to the process. If the signal has an action, the action
2202 * is usually performed by the target process rather than the caller; we add
2203 * the signal to the set of pending signals for the process.
2204 *
2205 * Always drops a reference on a signal_reason if one is provided, whether via
2206 * passing it to a thread or deallocating directly.
2207 *
2208 * Exceptions:
2209 * o When a stop signal is sent to a sleeping process that takes the
2210 * default action, the process is stopped without awakening it.
2211 * o SIGCONT restarts stopped processes (or puts them back to sleep)
2212 * regardless of the signal action (eg, blocked or ignored).
2213 *
2214 * Other ignored signals are discarded immediately.
2215 */
2216static void
2217psignal_internal(proc_t p, task_t task, thread_t thread, int flavor, int signum, os_reason_t signal_reason)
2218{
2219 int prop;
2220 user_addr_t action = USER_ADDR_NULL;
2221 proc_t sig_proc;
2222 thread_t sig_thread;
2223 task_t sig_task;
2224 int mask;
2225 struct uthread *uth;
2226 kern_return_t kret;
2227 uid_t r_uid;
2228 proc_t pp;
2229 kauth_cred_t my_cred;
2230 char *launchd_exit_reason_desc = NULL;
2231 boolean_t update_thread_policy = FALSE;
2232
2233 if ((u_int)signum >= NSIG || signum == 0) {
2234 panic("psignal: bad signal number %d", signum);
2235 }
2236
2237 mask = sigmask(signum);
2238 prop = sigprop[signum];
2239
2240#if SIGNAL_DEBUG
2241 if (rdebug_proc && (p != PROC_NULL) && (p == rdebug_proc)) {
2242 ram_printf(3);
2243 }
2244#endif /* SIGNAL_DEBUG */
2245
2246 /* catch unexpected initproc kills early for easier debuggging */
2247 if (signum == SIGKILL && p == initproc) {
2248 if (signal_reason == NULL) {
2249 panic_plain("unexpected SIGKILL of %s %s (no reason provided)",
2250 (p->p_name[0] != '\0' ? p->p_name : "initproc"),
2251 ((p->p_csflags & CS_KILLED) ? "(CS_KILLED)" : ""));
2252 } else {
2253 launchd_exit_reason_desc = launchd_exit_reason_get_string_desc(signal_reason);
2254 panic_plain("unexpected SIGKILL of %s %s with reason -- namespace %d code 0x%llx description %." LAUNCHD_PANIC_REASON_STRING_MAXLEN "s",
2255 (p->p_name[0] != '\0' ? p->p_name : "initproc"),
2256 ((p->p_csflags & CS_KILLED) ? "(CS_KILLED)" : ""),
2257 signal_reason->osr_namespace, signal_reason->osr_code,
2258 launchd_exit_reason_desc ? launchd_exit_reason_desc : "none");
2259 }
2260 }
2261
2262 /*
2263 * We will need the task pointer later. Grab it now to
2264 * check for a zombie process. Also don't send signals
2265 * to kernel internal tasks.
2266 */
2267 if (flavor & PSIG_VFORK) {
2268 sig_task = task;
2269 sig_thread = thread;
2270 sig_proc = p;
2271 } else if (flavor & PSIG_THREAD) {
2272 sig_task = get_threadtask(thread);
2273 sig_thread = thread;
2274 sig_proc = (proc_t)get_bsdtask_info(sig_task);
2275 } else if (flavor & PSIG_TRY_THREAD) {
2276 assert((thread == current_thread()) && (p == current_proc()));
2277 sig_task = p->task;
2278 sig_thread = thread;
2279 sig_proc = p;
2280 } else {
2281 sig_task = p->task;
2282 sig_thread = THREAD_NULL;
2283 sig_proc = p;
2284 }
2285
2286 if ((sig_task == TASK_NULL) || is_kerneltask(sig_task)) {
2287 os_reason_free(signal_reason);
2288 return;
2289 }
2290
2291 /*
2292 * do not send signals to the process that has the thread
2293 * doing a reboot(). Not doing so will mark that thread aborted
2294 * and can cause IO failures wich will cause data loss. There's
2295 * also no need to send a signal to a process that is in the middle
2296 * of being torn down.
2297 */
2298 if (ISSET(sig_proc->p_flag, P_REBOOT) || ISSET(sig_proc->p_lflag, P_LEXIT)) {
2299 DTRACE_PROC3(signal__discard, thread_t, sig_thread, proc_t, sig_proc, int, signum);
2300 os_reason_free(signal_reason);
2301 return;
2302 }
2303
2304 if ((flavor & (PSIG_VFORK | PSIG_THREAD)) == 0) {
2305 proc_knote(sig_proc, NOTE_SIGNAL | signum);
2306 }
2307
2308 if ((flavor & PSIG_LOCKED) == 0) {
2309 proc_signalstart(sig_proc, 0);
2310 }
2311
2312 /* Don't send signals to a process that has ignored them. */
2313 if (((flavor & PSIG_VFORK) == 0) && ((sig_proc->p_lflag & P_LTRACED) == 0) && (sig_proc->p_sigignore & mask)) {
2314 DTRACE_PROC3(signal__discard, thread_t, sig_thread, proc_t, sig_proc, int, signum);
2315 goto sigout_unlocked;
2316 }
2317
2318 /*
2319 * The proc_lock prevents the targeted thread from being deallocated
2320 * or handling the signal until we're done signaling it.
2321 *
2322 * Once the proc_lock is dropped, we have no guarantee the thread or uthread exists anymore.
2323 *
2324 * XXX: What if the thread goes inactive after the thread passes bsd ast point?
2325 */
2326 proc_lock(sig_proc);
2327
2328 if (flavor & PSIG_VFORK) {
2329 action = SIG_DFL;
2330 act_set_astbsd(sig_thread);
2331 kret = KERN_SUCCESS;
2332 } else if (flavor & PSIG_TRY_THREAD) {
2333 uth = get_bsdthread_info(sig_thread);
2334 if (((uth->uu_flag & UT_NO_SIGMASK) == 0) &&
2335 (((uth->uu_sigmask & mask) == 0) || (uth->uu_sigwait & mask)) &&
2336 ((kret = check_actforsig(sig_proc->task, sig_thread, 1)) == KERN_SUCCESS)) {
2337 /* deliver to specified thread */
2338 } else {
2339 /* deliver to any willing thread */
2340 kret = get_signalthread(sig_proc, signum, &sig_thread);
2341 }
2342 } else if (flavor & PSIG_THREAD) {
2343 /* If successful return with ast set */
2344 kret = check_actforsig(sig_task, sig_thread, 1);
2345 } else {
2346 /* If successful return with ast set */
2347 kret = get_signalthread(sig_proc, signum, &sig_thread);
2348 }
2349
2350 if (kret != KERN_SUCCESS) {
2351 DTRACE_PROC3(signal__discard, thread_t, sig_thread, proc_t, sig_proc, int, signum);
2352 proc_unlock(sig_proc);
2353 goto sigout_unlocked;
2354 }
2355
2356 uth = get_bsdthread_info(sig_thread);
2357
2358 /*
2359 * If proc is traced, always give parent a chance.
2360 */
2361
2362 if ((flavor & PSIG_VFORK) == 0) {
2363 if (sig_proc->p_lflag & P_LTRACED) {
2364 action = SIG_DFL;
2365 } else {
2366 /*
2367 * If the signal is being ignored,
2368 * then we forget about it immediately.
2369 * (Note: we don't set SIGCONT in p_sigignore,
2370 * and if it is set to SIG_IGN,
2371 * action will be SIG_DFL here.)
2372 */
2373 if (sig_proc->p_sigignore & mask) {
2374 goto sigout_locked;
2375 }
2376
2377 if (uth->uu_sigwait & mask) {
2378 action = KERN_SIG_WAIT;
2379 } else if (uth->uu_sigmask & mask) {
2380 action = KERN_SIG_HOLD;
2381 } else if (sig_proc->p_sigcatch & mask) {
2382 action = KERN_SIG_CATCH;
2383 } else {
2384 action = SIG_DFL;
2385 }
2386 }
2387 }
2388
2389 /* TODO: p_nice isn't hooked up to the scheduler... */
2390 if (sig_proc->p_nice > NZERO && action == SIG_DFL && (prop & SA_KILL) &&
2391 (sig_proc->p_lflag & P_LTRACED) == 0) {
2392 sig_proc->p_nice = NZERO;
2393 }
2394
2395 if (prop & SA_CONT) {
2396 uth->uu_siglist &= ~stopsigmask;
2397 }
2398
2399 if (prop & SA_STOP) {
2400 struct pgrp *pg;
2401 /*
2402 * If sending a tty stop signal to a member of an orphaned
2403 * process group, discard the signal here if the action
2404 * is default; don't stop the process below if sleeping,
2405 * and don't clear any pending SIGCONT.
2406 */
2407 pg = proc_pgrp(sig_proc);
2408 if (prop & SA_TTYSTOP && pg->pg_jobc == 0 &&
2409 action == SIG_DFL) {
2410 pg_rele(pg);
2411 goto sigout_locked;
2412 }
2413 pg_rele(pg);
2414 uth->uu_siglist &= ~contsigmask;
2415 }
2416
2417 uth->uu_siglist |= mask;
2418
2419 /*
2420 * Defer further processing for signals which are held,
2421 * except that stopped processes must be continued by SIGCONT.
2422 */
2423 /* vfork will not go thru as action is SIG_DFL */
2424 if ((action == KERN_SIG_HOLD) && ((prop & SA_CONT) == 0 || sig_proc->p_stat != SSTOP)) {
2425 goto sigout_locked;
2426 }
2427
2428 /*
2429 * SIGKILL priority twiddling moved here from above because
2430 * it needs sig_thread. Could merge it into large switch
2431 * below if we didn't care about priority for tracing
2432 * as SIGKILL's action is always SIG_DFL.
2433 *
2434 * TODO: p_nice isn't hooked up to the scheduler...
2435 */
2436 if ((signum == SIGKILL) && (sig_proc->p_nice > NZERO)) {
2437 sig_proc->p_nice = NZERO;
2438 }
2439
2440 /*
2441 * Process is traced - wake it up (if not already
2442 * stopped) so that it can discover the signal in
2443 * issig() and stop for the parent.
2444 */
2445 if (sig_proc->p_lflag & P_LTRACED) {
2446 if (sig_proc->p_stat != SSTOP) {
2447 goto runlocked;
2448 } else {
2449 goto sigout_locked;
2450 }
2451 }
2452
2453 if ((flavor & PSIG_VFORK) != 0) {
2454 goto runlocked;
2455 }
2456
2457 if (action == KERN_SIG_WAIT) {
2458#if CONFIG_DTRACE
2459 /*
2460 * DTrace proc signal-clear returns a siginfo_t. Collect the needed info.
2461 */
2462 r_uid = kauth_getruid(); /* per thread credential; protected by our thread context */
2463
2464 bzero((caddr_t)&(uth->t_dtrace_siginfo), sizeof(uth->t_dtrace_siginfo));
2465
2466 uth->t_dtrace_siginfo.si_signo = signum;
2467 uth->t_dtrace_siginfo.si_pid = current_proc()->p_pid;
2468 uth->t_dtrace_siginfo.si_status = W_EXITCODE(signum, 0);
2469 uth->t_dtrace_siginfo.si_uid = r_uid;
2470 uth->t_dtrace_siginfo.si_code = 0;
2471#endif
2472 uth->uu_sigwait = mask;
2473 uth->uu_siglist &= ~mask;
2474 wakeup(&uth->uu_sigwait);
2475 /* if it is SIGCONT resume whole process */
2476 if (prop & SA_CONT) {
2477 OSBitOrAtomic(P_CONTINUED, &sig_proc->p_flag);
2478 sig_proc->p_contproc = current_proc()->p_pid;
2479 (void) task_resume_internal(sig_task);
2480 }
2481 goto sigout_locked;
2482 }
2483
2484 if (action != SIG_DFL) {
2485 /*
2486 * User wants to catch the signal.
2487 * Wake up the thread, but don't un-suspend it
2488 * (except for SIGCONT).
2489 */
2490 if (prop & SA_CONT) {
2491 OSBitOrAtomic(P_CONTINUED, &sig_proc->p_flag);
2492 (void) task_resume_internal(sig_task);
2493 sig_proc->p_stat = SRUN;
2494 } else if (sig_proc->p_stat == SSTOP) {
2495 goto sigout_locked;
2496 }
2497 /*
2498 * Fill out siginfo structure information to pass to the
2499 * signalled process/thread sigaction handler, when it
2500 * wakes up. si_code is 0 because this is an ordinary
2501 * signal, not a SIGCHLD, and so si_status is the signal
2502 * number itself, instead of the child process exit status.
2503 * We shift this left because it will be shifted right before
2504 * it is passed to user space. kind of ugly to use W_EXITCODE
2505 * this way, but it beats defining a new macro.
2506 *
2507 * Note: Avoid the SIGCHLD recursion case!
2508 */
2509 if (signum != SIGCHLD) {
2510 r_uid = kauth_getruid();
2511
2512 sig_proc->si_pid = current_proc()->p_pid;
2513 sig_proc->si_status = W_EXITCODE(signum, 0);
2514 sig_proc->si_uid = r_uid;
2515 sig_proc->si_code = 0;
2516 }
2517
2518 goto runlocked;
2519 } else {
2520 /* Default action - varies */
2521 if (mask & stopsigmask) {
2522 assert(signal_reason == NULL);
2523 /*
2524 * These are the signals which by default
2525 * stop a process.
2526 *
2527 * Don't clog system with children of init
2528 * stopped from the keyboard.
2529 */
2530 if (!(prop & SA_STOP) && sig_proc->p_pptr == initproc) {
2531 uth->uu_siglist &= ~mask;
2532 proc_unlock(sig_proc);
2533 /* siglock still locked, proc_lock not locked */
2534 psignal_locked(sig_proc, SIGKILL);
2535 goto sigout_unlocked;
2536 }
2537
2538 /*
2539 * Stop the task
2540 * if task hasn't already been stopped by
2541 * a signal.
2542 */
2543 uth->uu_siglist &= ~mask;
2544 if (sig_proc->p_stat != SSTOP) {
2545 sig_proc->p_xstat = signum;
2546 sig_proc->p_stat = SSTOP;
2547 OSBitAndAtomic(~((uint32_t)P_CONTINUED), &sig_proc->p_flag);
2548 sig_proc->p_lflag &= ~P_LWAITED;
2549 proc_unlock(sig_proc);
2550
2551 pp = proc_parentholdref(sig_proc);
2552 stop(sig_proc, pp);
2553 if ((pp != PROC_NULL) && ((pp->p_flag & P_NOCLDSTOP) == 0)) {
2554 my_cred = kauth_cred_proc_ref(sig_proc);
2555 r_uid = kauth_cred_getruid(my_cred);
2556 kauth_cred_unref(&my_cred);
2557
2558 proc_lock(sig_proc);
2559 pp->si_pid = sig_proc->p_pid;
2560 /*
2561 * POSIX: sigaction for a stopped child
2562 * when sent to the parent must set the
2563 * child's signal number into si_status.
2564 */
2565 if (signum != SIGSTOP) {
2566 pp->si_status = WEXITSTATUS(sig_proc->p_xstat);
2567 } else {
2568 pp->si_status = W_EXITCODE(signum, signum);
2569 }
2570 pp->si_code = CLD_STOPPED;
2571 pp->si_uid = r_uid;
2572 proc_unlock(sig_proc);
2573
2574 psignal(pp, SIGCHLD);
2575 }
2576 if (pp != PROC_NULL) {
2577 proc_parentdropref(pp, 0);
2578 }
2579
2580 goto sigout_unlocked;
2581 }
2582
2583 goto sigout_locked;
2584 }
2585
2586 DTRACE_PROC3(signal__send, thread_t, sig_thread, proc_t, p, int, signum);
2587
2588 switch (signum) {
2589 /*
2590 * Signals ignored by default have been dealt
2591 * with already, since their bits are on in
2592 * p_sigignore.
2593 */
2594
2595 case SIGKILL:
2596 /*
2597 * Kill signal always sets process running and
2598 * unsuspends it.
2599 */
2600 /*
2601 * Process will be running after 'run'
2602 */
2603 sig_proc->p_stat = SRUN;
2604 /*
2605 * In scenarios where suspend/resume are racing
2606 * the signal we are missing AST_BSD by the time
2607 * we get here, set again to avoid races. This
2608 * was the scenario with spindump enabled shutdowns.
2609 * We would need to cover this approp down the line.
2610 */
2611 act_set_astbsd(sig_thread);
2612 kret = thread_abort(sig_thread);
2613 update_thread_policy = (kret == KERN_SUCCESS);
2614
2615 if (uth->uu_exit_reason == OS_REASON_NULL) {
2616 if (signal_reason == OS_REASON_NULL) {
2617 KERNEL_DEBUG_CONSTANT(BSDDBG_CODE(DBG_BSD_PROC, BSD_PROC_EXITREASON_CREATE) | DBG_FUNC_NONE,
2618 sig_proc->p_pid, OS_REASON_SIGNAL, signum, 0, 0);
2619
2620 signal_reason = build_signal_reason(signum, NULL);
2621 }
2622
2623 os_reason_ref(signal_reason);
2624 set_thread_exit_reason(sig_thread, signal_reason, TRUE);
2625 }
2626
2627 goto sigout_locked;
2628
2629 case SIGCONT:
2630 /*
2631 * Let the process run. If it's sleeping on an
2632 * event, it remains so.
2633 */
2634 assert(signal_reason == NULL);
2635 OSBitOrAtomic(P_CONTINUED, &sig_proc->p_flag);
2636 sig_proc->p_contproc = sig_proc->p_pid;
2637 sig_proc->p_xstat = signum;
2638
2639 (void) task_resume_internal(sig_task);
2640
2641 /*
2642 * When processing a SIGCONT, we need to check
2643 * to see if there are signals pending that
2644 * were not delivered because we had been
2645 * previously stopped. If that's the case,
2646 * we need to thread_abort_safely() to trigger
2647 * interruption of the current system call to
2648 * cause their handlers to fire. If it's only
2649 * the SIGCONT, then don't wake up.
2650 */
2651 if (((flavor & (PSIG_VFORK | PSIG_THREAD)) == 0) && (((uth->uu_siglist & ~uth->uu_sigmask) & ~sig_proc->p_sigignore) & ~mask)) {
2652 uth->uu_siglist &= ~mask;
2653 sig_proc->p_stat = SRUN;
2654 goto runlocked;
2655 }
2656
2657 uth->uu_siglist &= ~mask;
2658 sig_proc->p_stat = SRUN;
2659 goto sigout_locked;
2660
2661 default:
2662 /*
2663 * A signal which has a default action of killing
2664 * the process, and for which there is no handler,
2665 * needs to act like SIGKILL
2666 */
2667 if (((flavor & (PSIG_VFORK | PSIG_THREAD)) == 0) && (action == SIG_DFL) && (prop & SA_KILL)) {
2668 sig_proc->p_stat = SRUN;
2669 kret = thread_abort(sig_thread);
2670 update_thread_policy = (kret == KERN_SUCCESS);
2671
2672 if (uth->uu_exit_reason == OS_REASON_NULL) {
2673 if (signal_reason == OS_REASON_NULL) {
2674 KERNEL_DEBUG_CONSTANT(BSDDBG_CODE(DBG_BSD_PROC, BSD_PROC_EXITREASON_CREATE) | DBG_FUNC_NONE,
2675 sig_proc->p_pid, OS_REASON_SIGNAL, signum, 0, 0);
2676
2677 signal_reason = build_signal_reason(signum, NULL);
2678 }
2679
2680 os_reason_ref(signal_reason);
2681 set_thread_exit_reason(sig_thread, signal_reason, TRUE);
2682 }
2683
2684 goto sigout_locked;
2685 }
2686
2687 /*
2688 * All other signals wake up the process, but don't
2689 * resume it.
2690 */
2691 if (sig_proc->p_stat == SSTOP) {
2692 goto sigout_locked;
2693 }
2694 goto runlocked;
2695 }
2696 }
2697 /*NOTREACHED*/
2698
2699runlocked:
2700 /*
2701 * If we're being traced (possibly because someone attached us
2702 * while we were stopped), check for a signal from the debugger.
2703 */
2704 if (sig_proc->p_stat == SSTOP) {
2705 if ((sig_proc->p_lflag & P_LTRACED) != 0 && sig_proc->p_xstat != 0) {
2706 uth->uu_siglist |= sigmask(sig_proc->p_xstat);
2707 }
2708
2709 if ((flavor & PSIG_VFORK) != 0) {
2710 sig_proc->p_stat = SRUN;
2711 }
2712 } else {
2713 /*
2714 * setrunnable(p) in BSD and
2715 * Wake up the thread if it is interruptible.
2716 */
2717 sig_proc->p_stat = SRUN;
2718 if ((flavor & PSIG_VFORK) == 0) {
2719 thread_abort_safely(sig_thread);
2720 }
2721 }
2722
2723sigout_locked:
2724 if (update_thread_policy) {
2725 /*
2726 * Update the thread policy to heading to terminate, increase priority if
2727 * necessary. This needs to be done before we drop the proc lock because the
2728 * thread can take the fatal signal once it's dropped.
2729 */
2730 proc_set_thread_policy(sig_thread, TASK_POLICY_ATTRIBUTE, TASK_POLICY_TERMINATED, TASK_POLICY_ENABLE);
2731 }
2732
2733 proc_unlock(sig_proc);
2734
2735sigout_unlocked:
2736 os_reason_free(signal_reason);
2737 if ((flavor & PSIG_LOCKED) == 0) {
2738 proc_signalend(sig_proc, 0);
2739 }
2740}
2741
2742void
2743psignal(proc_t p, int signum)
2744{
2745 psignal_internal(p, NULL, NULL, 0, signum, NULL);
2746}
2747
2748void
2749psignal_with_reason(proc_t p, int signum, struct os_reason *signal_reason)
2750{
2751 psignal_internal(p, NULL, NULL, 0, signum, signal_reason);
2752}
2753
2754void
2755psignal_sigkill_with_reason(struct proc *p, struct os_reason *signal_reason)
2756{
2757 psignal_internal(p, NULL, NULL, 0, SIGKILL, signal_reason);
2758}
2759
2760void
2761psignal_locked(proc_t p, int signum)
2762{
2763 psignal_internal(p, NULL, NULL, PSIG_LOCKED, signum, NULL);
2764}
2765
2766void
2767psignal_vfork_with_reason(proc_t p, task_t new_task, thread_t thread, int signum, struct os_reason *signal_reason)
2768{
2769 psignal_internal(p, new_task, thread, PSIG_VFORK, signum, signal_reason);
2770}
2771
2772
2773void
2774psignal_vfork(proc_t p, task_t new_task, thread_t thread, int signum)
2775{
2776 psignal_internal(p, new_task, thread, PSIG_VFORK, signum, NULL);
2777}
2778
2779void
2780psignal_uthread(thread_t thread, int signum)
2781{
2782 psignal_internal(PROC_NULL, TASK_NULL, thread, PSIG_THREAD, signum, NULL);
2783}
2784
2785/* same as psignal(), but prefer delivery to 'thread' if possible */
2786void
2787psignal_try_thread(proc_t p, thread_t thread, int signum)
2788{
2789 psignal_internal(p, NULL, thread, PSIG_TRY_THREAD, signum, NULL);
2790}
2791
2792void
2793psignal_try_thread_with_reason(proc_t p, thread_t thread, int signum, struct os_reason *signal_reason)
2794{
2795 psignal_internal(p, TASK_NULL, thread, PSIG_TRY_THREAD, signum, signal_reason);
2796}
2797
2798void
2799psignal_thread_with_reason(proc_t p, thread_t thread, int signum, struct os_reason *signal_reason)
2800{
2801 psignal_internal(p, TASK_NULL, thread, PSIG_THREAD, signum, signal_reason);
2802}
2803
2804/*
2805 * If the current process has received a signal (should be caught or cause
2806 * termination, should interrupt current syscall), return the signal number.
2807 * Stop signals with default action are processed immediately, then cleared;
2808 * they aren't returned. This is checked after each entry to the system for
2809 * a syscall or trap (though this can usually be done without calling issignal
2810 * by checking the pending signal masks in the CURSIG macro.) The normal call
2811 * sequence is
2812 *
2813 * while (signum = CURSIG(curproc))
2814 * postsig(signum);
2815 */
2816int
2817issignal_locked(proc_t p)
2818{
2819 int signum, mask, prop, sigbits;
2820 thread_t cur_act;
2821 struct uthread * ut;
2822 proc_t pp;
2823 kauth_cred_t my_cred;
2824 int retval = 0;
2825 uid_t r_uid;
2826
2827 cur_act = current_thread();
2828
2829#if SIGNAL_DEBUG
2830 if (rdebug_proc && (p == rdebug_proc)) {
2831 ram_printf(3);
2832 }
2833#endif /* SIGNAL_DEBUG */
2834
2835 /*
2836 * Try to grab the signal lock.
2837 */
2838 if (sig_try_locked(p) <= 0) {
2839 return 0;
2840 }
2841
2842 proc_signalstart(p, 1);
2843
2844 ut = get_bsdthread_info(cur_act);
2845 for (;;) {
2846 sigbits = ut->uu_siglist & ~ut->uu_sigmask;
2847
2848 if (p->p_lflag & P_LPPWAIT) {
2849 sigbits &= ~stopsigmask;
2850 }
2851 if (sigbits == 0) { /* no signal to send */
2852 retval = 0;
2853 goto out;
2854 }
2855
2856 signum = ffs((unsigned int)sigbits);
2857 mask = sigmask(signum);
2858 prop = sigprop[signum];
2859
2860 /*
2861 * We should see pending but ignored signals
2862 * only if P_LTRACED was on when they were posted.
2863 */
2864 if (mask & p->p_sigignore && (p->p_lflag & P_LTRACED) == 0) {
2865 ut->uu_siglist &= ~mask;
2866 continue;
2867 }
2868
2869 if (p->p_lflag & P_LTRACED && (p->p_lflag & P_LPPWAIT) == 0) {
2870 /*
2871 * If traced, deliver the signal to the debugger, and wait to be
2872 * released.
2873 */
2874 task_t task;
2875 p->p_xstat = signum;
2876
2877 if (p->p_lflag & P_LSIGEXC) {
2878 p->sigwait = TRUE;
2879 p->sigwait_thread = cur_act;
2880 p->p_stat = SSTOP;
2881 OSBitAndAtomic(~((uint32_t)P_CONTINUED), &p->p_flag);
2882 p->p_lflag &= ~P_LWAITED;
2883 ut->uu_siglist &= ~mask; /* clear the current signal from the pending list */
2884 proc_signalend(p, 1);
2885 proc_unlock(p);
2886 do_bsdexception(EXC_SOFTWARE, EXC_SOFT_SIGNAL, signum);
2887 proc_lock(p);
2888 proc_signalstart(p, 1);
2889 } else {
2890 proc_unlock(p);
2891 my_cred = kauth_cred_proc_ref(p);
2892 r_uid = kauth_cred_getruid(my_cred);
2893 kauth_cred_unref(&my_cred);
2894
2895 pp = proc_parentholdref(p);
2896 if (pp != PROC_NULL) {
2897 proc_lock(pp);
2898
2899 pp->si_pid = p->p_pid;
2900 pp->p_xhighbits = p->p_xhighbits;
2901 p->p_xhighbits = 0;
2902 pp->si_status = p->p_xstat;
2903 pp->si_code = CLD_TRAPPED;
2904 pp->si_uid = r_uid;
2905
2906 proc_unlock(pp);
2907 }
2908
2909 /*
2910 * XXX Have to really stop for debuggers;
2911 * XXX stop() doesn't do the right thing.
2912 */
2913 task = p->task;
2914 task_suspend_internal(task);
2915
2916 proc_lock(p);
2917 p->sigwait = TRUE;
2918 p->sigwait_thread = cur_act;
2919 p->p_stat = SSTOP;
2920 OSBitAndAtomic(~((uint32_t)P_CONTINUED), &p->p_flag);
2921 p->p_lflag &= ~P_LWAITED;
2922 ut->uu_siglist &= ~mask;
2923
2924 proc_signalend(p, 1);
2925 proc_unlock(p);
2926
2927 if (pp != PROC_NULL) {
2928 psignal(pp, SIGCHLD);
2929 proc_list_lock();
2930 wakeup((caddr_t)pp);
2931 proc_parentdropref(pp, 1);
2932 proc_list_unlock();
2933 }
2934
2935 assert_wait((caddr_t)&p->sigwait, (THREAD_INTERRUPTIBLE));
2936 thread_block(THREAD_CONTINUE_NULL);
2937 proc_lock(p);
2938 proc_signalstart(p, 1);
2939 }
2940
2941 p->sigwait = FALSE;
2942 p->sigwait_thread = NULL;
2943 wakeup((caddr_t)&p->sigwait_thread);
2944
2945 if (signum == SIGKILL || ut->uu_siglist & sigmask(SIGKILL)) {
2946 /*
2947 * Deliver a pending sigkill even if it's not the current signal.
2948 * Necessary for PT_KILL, which should not be delivered to the
2949 * debugger, but we can't differentiate it from any other KILL.
2950 */
2951 signum = SIGKILL;
2952 goto deliver_sig;
2953 }
2954
2955 /* We may have to quit. */
2956 if (thread_should_abort(current_thread())) {
2957 retval = 0;
2958 goto out;
2959 }
2960
2961 /*
2962 * If parent wants us to take the signal,
2963 * then it will leave it in p->p_xstat;
2964 * otherwise we just look for signals again.
2965 */
2966 signum = p->p_xstat;
2967 if (signum == 0) {
2968 continue;
2969 }
2970
2971 /*
2972 * Put the new signal into p_siglist. If the
2973 * signal is being masked, look for other signals.
2974 */
2975 mask = sigmask(signum);
2976 ut->uu_siglist |= mask;
2977 if (ut->uu_sigmask & mask) {
2978 continue;
2979 }
2980 }
2981
2982 /*
2983 * Decide whether the signal should be returned.
2984 * Return the signal's number, or fall through
2985 * to clear it from the pending mask.
2986 */
2987
2988 switch ((long)p->p_sigacts->ps_sigact[signum]) {
2989 case (long)SIG_DFL:
2990 /*
2991 * If there is a pending stop signal to process
2992 * with default action, stop here,
2993 * then clear the signal. However,
2994 * if process is member of an orphaned
2995 * process group, ignore tty stop signals.
2996 */
2997 if (prop & SA_STOP) {
2998 struct pgrp * pg;
2999
3000 proc_unlock(p);
3001 pg = proc_pgrp(p);
3002 if (p->p_lflag & P_LTRACED ||
3003 (pg->pg_jobc == 0 &&
3004 prop & SA_TTYSTOP)) {
3005 proc_lock(p);
3006 pg_rele(pg);
3007 break; /* ignore signal */
3008 }
3009 pg_rele(pg);
3010 if (p->p_stat != SSTOP) {
3011 proc_lock(p);
3012 p->p_xstat = signum;
3013 p->p_stat = SSTOP;
3014 p->p_lflag &= ~P_LWAITED;
3015 proc_unlock(p);
3016
3017 pp = proc_parentholdref(p);
3018 stop(p, pp);
3019 if ((pp != PROC_NULL) && ((pp->p_flag & P_NOCLDSTOP) == 0)) {
3020 my_cred = kauth_cred_proc_ref(p);
3021 r_uid = kauth_cred_getruid(my_cred);
3022 kauth_cred_unref(&my_cred);
3023
3024 proc_lock(pp);
3025 pp->si_pid = p->p_pid;
3026 pp->si_status = WEXITSTATUS(p->p_xstat);
3027 pp->si_code = CLD_STOPPED;
3028 pp->si_uid = r_uid;
3029 proc_unlock(pp);
3030
3031 psignal(pp, SIGCHLD);
3032 }
3033 if (pp != PROC_NULL) {
3034 proc_parentdropref(pp, 0);
3035 }
3036 }
3037 proc_lock(p);
3038 break;
3039 } else if (prop & SA_IGNORE) {
3040 /*
3041 * Except for SIGCONT, shouldn't get here.
3042 * Default action is to ignore; drop it.
3043 */
3044 break; /* ignore signal */
3045 } else {
3046 goto deliver_sig;
3047 }
3048
3049 case (long)SIG_IGN:
3050 /*
3051 * Masking above should prevent us ever trying
3052 * to take action on an ignored signal other
3053 * than SIGCONT, unless process is traced.
3054 */
3055 if ((prop & SA_CONT) == 0 &&
3056 (p->p_lflag & P_LTRACED) == 0) {
3057 printf("issignal\n");
3058 }
3059 break; /* ignore signal */
3060
3061 default:
3062 /* This signal has an action - deliver it. */
3063 goto deliver_sig;
3064 }
3065
3066 /* If we dropped through, the signal was ignored - remove it from pending list. */
3067 ut->uu_siglist &= ~mask;
3068 } /* for(;;) */
3069
3070 /* NOTREACHED */
3071
3072deliver_sig:
3073 ut->uu_siglist &= ~mask;
3074 retval = signum;
3075
3076out:
3077 proc_signalend(p, 1);
3078 return retval;
3079}
3080
3081/* called from _sleep */
3082int
3083CURSIG(proc_t p)
3084{
3085 int signum, mask, prop, sigbits;
3086 thread_t cur_act;
3087 struct uthread * ut;
3088 int retnum = 0;
3089
3090
3091 cur_act = current_thread();
3092
3093 ut = get_bsdthread_info(cur_act);
3094
3095 if (ut->uu_siglist == 0) {
3096 return 0;
3097 }
3098
3099 if (((ut->uu_siglist & ~ut->uu_sigmask) == 0) && ((p->p_lflag & P_LTRACED) == 0)) {
3100 return 0;
3101 }
3102
3103 sigbits = ut->uu_siglist & ~ut->uu_sigmask;
3104
3105 for (;;) {
3106 if (p->p_lflag & P_LPPWAIT) {
3107 sigbits &= ~stopsigmask;
3108 }
3109 if (sigbits == 0) { /* no signal to send */
3110 return retnum;
3111 }
3112
3113 signum = ffs((unsigned int)sigbits);
3114 mask = sigmask(signum);
3115 prop = sigprop[signum];
3116 sigbits &= ~mask; /* take the signal out */
3117
3118 /*
3119 * We should see pending but ignored signals
3120 * only if P_LTRACED was on when they were posted.
3121 */
3122 if (mask & p->p_sigignore && (p->p_lflag & P_LTRACED) == 0) {
3123 continue;
3124 }
3125
3126 if (p->p_lflag & P_LTRACED && (p->p_lflag & P_LPPWAIT) == 0) {
3127 return signum;
3128 }
3129
3130 /*
3131 * Decide whether the signal should be returned.
3132 * Return the signal's number, or fall through
3133 * to clear it from the pending mask.
3134 */
3135
3136 switch ((long)p->p_sigacts->ps_sigact[signum]) {
3137 case (long)SIG_DFL:
3138 /*
3139 * If there is a pending stop signal to process
3140 * with default action, stop here,
3141 * then clear the signal. However,
3142 * if process is member of an orphaned
3143 * process group, ignore tty stop signals.
3144 */
3145 if (prop & SA_STOP) {
3146 struct pgrp *pg;
3147
3148 pg = proc_pgrp(p);
3149
3150 if (p->p_lflag & P_LTRACED ||
3151 (pg->pg_jobc == 0 &&
3152 prop & SA_TTYSTOP)) {
3153 pg_rele(pg);
3154 break; /* == ignore */
3155 }
3156 pg_rele(pg);
3157 retnum = signum;
3158 break;
3159 } else if (prop & SA_IGNORE) {
3160 /*
3161 * Except for SIGCONT, shouldn't get here.
3162 * Default action is to ignore; drop it.
3163 */
3164 break; /* == ignore */
3165 } else {
3166 return signum;
3167 }
3168 /*NOTREACHED*/
3169
3170 case (long)SIG_IGN:
3171 /*
3172 * Masking above should prevent us ever trying
3173 * to take action on an ignored signal other
3174 * than SIGCONT, unless process is traced.
3175 */
3176 if ((prop & SA_CONT) == 0 &&
3177 (p->p_lflag & P_LTRACED) == 0) {
3178 printf("issignal\n");
3179 }
3180 break; /* == ignore */
3181
3182 default:
3183 /*
3184 * This signal has an action, let
3185 * postsig() process it.
3186 */
3187 return signum;
3188 }
3189 }
3190 /* NOTREACHED */
3191}
3192
3193/*
3194 * Put the argument process into the stopped state and notify the parent
3195 * via wakeup. Signals are handled elsewhere. The process must not be
3196 * on the run queue.
3197 */
3198static void
3199stop(proc_t p, proc_t parent)
3200{
3201 OSBitAndAtomic(~((uint32_t)P_CONTINUED), &p->p_flag);
3202 if ((parent != PROC_NULL) && (parent->p_stat != SSTOP)) {
3203 proc_list_lock();
3204 wakeup((caddr_t)parent);
3205 proc_list_unlock();
3206 }
3207 (void) task_suspend_internal(p->task);
3208}
3209
3210/*
3211 * Take the action for the specified signal
3212 * from the current set of pending signals.
3213 */
3214void
3215postsig_locked(int signum)
3216{
3217 proc_t p = current_proc();
3218 struct sigacts *ps = p->p_sigacts;
3219 user_addr_t catcher;
3220 uint32_t code;
3221 int mask, returnmask;
3222 struct uthread * ut;
3223 os_reason_t ut_exit_reason = OS_REASON_NULL;
3224
3225#if DIAGNOSTIC
3226 if (signum == 0) {
3227 panic("postsig");
3228 }
3229 /*
3230 * This must be called on master cpu
3231 */
3232 if (cpu_number() != master_cpu) {
3233 panic("psig not on master");
3234 }
3235#endif
3236
3237 /*
3238 * Try to grab the signal lock.
3239 */
3240 if (sig_try_locked(p) <= 0) {
3241 return;
3242 }
3243
3244 proc_signalstart(p, 1);
3245
3246 ut = (struct uthread *)get_bsdthread_info(current_thread());
3247 mask = sigmask(signum);
3248 ut->uu_siglist &= ~mask;
3249 catcher = ps->ps_sigact[signum];
3250 if (catcher == SIG_DFL) {
3251 /*
3252 * Default catcher, where the default is to kill
3253 * the process. (Other cases were ignored above.)
3254 */
3255
3256 /*
3257 * exit_with_reason() below will consume a reference to the thread's exit reason, so we take another
3258 * reference so the thread still has one even after we call exit_with_reason(). The thread's reference will
3259 * ultimately be destroyed in uthread_cleanup().
3260 */
3261 ut_exit_reason = ut->uu_exit_reason;
3262 os_reason_ref(ut_exit_reason);
3263
3264 p->p_acflag |= AXSIG;
3265 if (sigprop[signum] & SA_CORE) {
3266 p->p_sigacts->ps_sig = signum;
3267 proc_signalend(p, 1);
3268 proc_unlock(p);
3269#if CONFIG_COREDUMP
3270 if (coredump(p, 0, 0) == 0) {
3271 signum |= WCOREFLAG;
3272 }
3273#endif
3274 } else {
3275 proc_signalend(p, 1);
3276 proc_unlock(p);
3277 }
3278
3279#if CONFIG_DTRACE
3280 bzero((caddr_t)&(ut->t_dtrace_siginfo), sizeof(ut->t_dtrace_siginfo));
3281
3282 ut->t_dtrace_siginfo.si_signo = signum;
3283 ut->t_dtrace_siginfo.si_pid = p->si_pid;
3284 ut->t_dtrace_siginfo.si_uid = p->si_uid;
3285 ut->t_dtrace_siginfo.si_status = WEXITSTATUS(p->si_status);
3286
3287 /* Fire DTrace proc:::fault probe when signal is generated by hardware. */
3288 switch (signum) {
3289 case SIGILL: case SIGBUS: case SIGSEGV: case SIGFPE: case SIGTRAP:
3290 DTRACE_PROC2(fault, int, (int)(ut->uu_code), siginfo_t *, &(ut->t_dtrace_siginfo));
3291 break;
3292 default:
3293 break;
3294 }
3295
3296
3297 DTRACE_PROC3(signal__handle, int, signum, siginfo_t *, &(ut->t_dtrace_siginfo),
3298 void (*)(void), SIG_DFL);
3299#endif
3300
3301 KERNEL_DEBUG_CONSTANT(BSDDBG_CODE(DBG_BSD_PROC, BSD_PROC_FRCEXIT) | DBG_FUNC_NONE,
3302 p->p_pid, W_EXITCODE(0, signum), 3, 0, 0);
3303
3304 exit_with_reason(p, W_EXITCODE(0, signum), (int *)NULL, TRUE, TRUE, 0, ut_exit_reason);
3305
3306 proc_lock(p);
3307 return;
3308 } else {
3309 /*
3310 * If we get here, the signal must be caught.
3311 */
3312#if DIAGNOSTIC
3313 if (catcher == SIG_IGN || (ut->uu_sigmask & mask)) {
3314 log(LOG_WARNING,
3315 "postsig: processing masked or ignored signal\n");
3316 }
3317#endif
3318
3319 /*
3320 * Set the new mask value and also defer further
3321 * occurences of this signal.
3322 *
3323 * Special case: user has done a sigpause. Here the
3324 * current mask is not of interest, but rather the
3325 * mask from before the sigpause is what we want
3326 * restored after the signal processing is completed.
3327 */
3328 if (ut->uu_flag & UT_SAS_OLDMASK) {
3329 returnmask = ut->uu_oldmask;
3330 ut->uu_flag &= ~UT_SAS_OLDMASK;
3331 ut->uu_oldmask = 0;
3332 } else {
3333 returnmask = ut->uu_sigmask;
3334 }
3335 ut->uu_sigmask |= ps->ps_catchmask[signum];
3336 if ((ps->ps_signodefer & mask) == 0) {
3337 ut->uu_sigmask |= mask;
3338 }
3339 sigset_t siginfo = ps->ps_siginfo;
3340 if ((signum != SIGILL) && (signum != SIGTRAP) && (ps->ps_sigreset & mask)) {
3341 if ((signum != SIGCONT) && (sigprop[signum] & SA_IGNORE)) {
3342 p->p_sigignore |= mask;
3343 }
3344 ps->ps_sigact[signum] = SIG_DFL;
3345 ps->ps_siginfo &= ~mask;
3346 ps->ps_signodefer &= ~mask;
3347 }
3348
3349 if (ps->ps_sig != signum) {
3350 code = 0;
3351 } else {
3352 code = ps->ps_code;
3353 ps->ps_code = 0;
3354 }
3355 OSIncrementAtomicLong(&p->p_stats->p_ru.ru_nsignals);
3356 sendsig(p, catcher, signum, returnmask, code, siginfo);
3357 }
3358 proc_signalend(p, 1);
3359}
3360
3361/*
3362 * Attach a signal knote to the list of knotes for this process.
3363 *
3364 * Signal knotes share the knote list with proc knotes. This
3365 * could be avoided by using a signal-specific knote list, but
3366 * probably isn't worth the trouble.
3367 */
3368
3369static int
3370filt_sigattach(struct knote *kn, __unused struct kevent_qos_s *kev)
3371{
3372 proc_t p = current_proc(); /* can attach only to oneself */
3373
3374 proc_klist_lock();
3375
3376 kn->kn_proc = p;
3377 kn->kn_flags |= EV_CLEAR; /* automatically set */
3378 kn->kn_sdata = 0; /* incoming data is ignored */
3379
3380 KNOTE_ATTACH(&p->p_klist, kn);
3381
3382 proc_klist_unlock();
3383
3384 /* edge-triggered events can't have fired before we attached */
3385 return 0;
3386}
3387
3388/*
3389 * remove the knote from the process list, if it hasn't already
3390 * been removed by exit processing.
3391 */
3392
3393static void
3394filt_sigdetach(struct knote *kn)
3395{
3396 proc_t p = kn->kn_proc;
3397
3398 proc_klist_lock();
3399 kn->kn_proc = NULL;
3400 KNOTE_DETACH(&p->p_klist, kn);
3401 proc_klist_unlock();
3402}
3403
3404/*
3405 * Post an event to the signal filter. Because we share the same list
3406 * as process knotes, we have to filter out and handle only signal events.
3407 *
3408 * We assume that we process fdfree() before we post the NOTE_EXIT for
3409 * a process during exit. Therefore, since signal filters can only be
3410 * set up "in-process", we should have already torn down the kqueue
3411 * hosting the EVFILT_SIGNAL knote and should never see NOTE_EXIT.
3412 */
3413static int
3414filt_signal(struct knote *kn, long hint)
3415{
3416 if (hint & NOTE_SIGNAL) {
3417 hint &= ~NOTE_SIGNAL;
3418
3419 if (kn->kn_id == (unsigned int)hint) {
3420 kn->kn_hook32++;
3421 }
3422 } else if (hint & NOTE_EXIT) {
3423 panic("filt_signal: detected NOTE_EXIT event");
3424 }
3425
3426 return kn->kn_hook32 != 0;
3427}
3428
3429static int
3430filt_signaltouch(struct knote *kn, struct kevent_qos_s *kev)
3431{
3432#pragma unused(kev)
3433
3434 int res;
3435
3436 proc_klist_lock();
3437
3438 /*
3439 * No data to save - just capture if it is already fired
3440 */
3441 res = (kn->kn_hook32 > 0);
3442
3443 proc_klist_unlock();
3444
3445 return res;
3446}
3447
3448static int
3449filt_signalprocess(struct knote *kn, struct kevent_qos_s *kev)
3450{
3451 int res = 0;
3452
3453 /*
3454 * Snapshot the event data.
3455 */
3456
3457 proc_klist_lock();
3458 if (kn->kn_hook32) {
3459 knote_fill_kevent(kn, kev, kn->kn_hook32);
3460 kn->kn_hook32 = 0;
3461 res = 1;
3462 }
3463 proc_klist_unlock();
3464 return res;
3465}
3466
3467void
3468bsd_ast(thread_t thread)
3469{
3470 proc_t p = current_proc();
3471 struct uthread *ut = get_bsdthread_info(thread);
3472 int signum;
3473 static int bsd_init_done = 0;
3474
3475 if (p == NULL) {
3476 return;
3477 }
3478
3479 /* don't run bsd ast on exec copy or exec'ed tasks */
3480 if (task_did_exec(current_task()) || task_is_exec_copy(current_task())) {
3481 return;
3482 }
3483
3484 if (timerisset(&p->p_vtimer_user.it_value)) {
3485 uint32_t microsecs;
3486
3487 task_vtimer_update(p->task, TASK_VTIMER_USER, &microsecs);
3488
3489 if (!itimerdecr(p, &p->p_vtimer_user, microsecs)) {
3490 if (timerisset(&p->p_vtimer_user.it_value)) {
3491 task_vtimer_set(p->task, TASK_VTIMER_USER);
3492 } else {
3493 task_vtimer_clear(p->task, TASK_VTIMER_USER);
3494 }
3495
3496 psignal_try_thread(p, thread, SIGVTALRM);
3497 }
3498 }
3499
3500 if (timerisset(&p->p_vtimer_prof.it_value)) {
3501 uint32_t microsecs;
3502
3503 task_vtimer_update(p->task, TASK_VTIMER_PROF, &microsecs);
3504
3505 if (!itimerdecr(p, &p->p_vtimer_prof, microsecs)) {
3506 if (timerisset(&p->p_vtimer_prof.it_value)) {
3507 task_vtimer_set(p->task, TASK_VTIMER_PROF);
3508 } else {
3509 task_vtimer_clear(p->task, TASK_VTIMER_PROF);
3510 }
3511
3512 psignal_try_thread(p, thread, SIGPROF);
3513 }
3514 }
3515
3516 if (timerisset(&p->p_rlim_cpu)) {
3517 struct timeval tv;
3518
3519 task_vtimer_update(p->task, TASK_VTIMER_RLIM, (uint32_t *) &tv.tv_usec);
3520
3521 proc_spinlock(p);
3522 if (p->p_rlim_cpu.tv_sec > 0 || p->p_rlim_cpu.tv_usec > tv.tv_usec) {
3523 tv.tv_sec = 0;
3524 timersub(&p->p_rlim_cpu, &tv, &p->p_rlim_cpu);
3525 proc_spinunlock(p);
3526 } else {
3527 timerclear(&p->p_rlim_cpu);
3528 proc_spinunlock(p);
3529
3530 task_vtimer_clear(p->task, TASK_VTIMER_RLIM);
3531
3532 psignal_try_thread(p, thread, SIGXCPU);
3533 }
3534 }
3535
3536#if CONFIG_DTRACE
3537 if (ut->t_dtrace_sig) {
3538 uint8_t dt_action_sig = ut->t_dtrace_sig;
3539 ut->t_dtrace_sig = 0;
3540 psignal(p, dt_action_sig);
3541 }
3542
3543 if (ut->t_dtrace_stop) {
3544 ut->t_dtrace_stop = 0;
3545 proc_lock(p);
3546 p->p_dtrace_stop = 1;
3547 proc_unlock(p);
3548 (void)task_suspend_internal(p->task);
3549 }
3550
3551 if (ut->t_dtrace_resumepid) {
3552 proc_t resumeproc = proc_find((int)ut->t_dtrace_resumepid);
3553 ut->t_dtrace_resumepid = 0;
3554 if (resumeproc != PROC_NULL) {
3555 proc_lock(resumeproc);
3556 /* We only act on processes stopped by dtrace */
3557 if (resumeproc->p_dtrace_stop) {
3558 resumeproc->p_dtrace_stop = 0;
3559 proc_unlock(resumeproc);
3560 task_resume_internal(resumeproc->task);
3561 } else {
3562 proc_unlock(resumeproc);
3563 }
3564 proc_rele(resumeproc);
3565 }
3566 }
3567
3568#endif /* CONFIG_DTRACE */
3569
3570 proc_lock(p);
3571 if (CHECK_SIGNALS(p, current_thread(), ut)) {
3572 while ((signum = issignal_locked(p))) {
3573 postsig_locked(signum);
3574 }
3575 }
3576 proc_unlock(p);
3577
3578#ifdef CONFIG_32BIT_TELEMETRY
3579 if (task_consume_32bit_log_flag(p->task)) {
3580 proc_log_32bit_telemetry(p);
3581 }
3582#endif /* CONFIG_32BIT_TELEMETRY */
3583
3584 if (!bsd_init_done) {
3585 bsd_init_done = 1;
3586 bsdinit_task();
3587 }
3588}
3589
3590/* ptrace set runnable */
3591void
3592pt_setrunnable(proc_t p)
3593{
3594 task_t task;
3595
3596 task = p->task;
3597
3598 if (p->p_lflag & P_LTRACED) {
3599 proc_lock(p);
3600 p->p_stat = SRUN;
3601 proc_unlock(p);
3602 if (p->sigwait) {
3603 wakeup((caddr_t)&(p->sigwait));
3604 if ((p->p_lflag & P_LSIGEXC) == 0) { // 5878479
3605 task_release(task);
3606 }
3607 }
3608 }
3609}
3610
3611kern_return_t
3612do_bsdexception(
3613 int exc,
3614 int code,
3615 int sub)
3616{
3617 mach_exception_data_type_t codes[EXCEPTION_CODE_MAX];
3618
3619 codes[0] = code;
3620 codes[1] = sub;
3621 return bsd_exception(exc, codes, 2);
3622}
3623
3624int
3625proc_pendingsignals(proc_t p, sigset_t mask)
3626{
3627 struct uthread * uth;
3628 thread_t th;
3629 sigset_t bits = 0;
3630
3631 proc_lock(p);
3632 /* If the process is in proc exit return no signal info */
3633 if (p->p_lflag & P_LPEXIT) {
3634 goto out;
3635 }
3636
3637 if ((p->p_lflag & P_LINVFORK) && p->p_vforkact) {
3638 th = p->p_vforkact;
3639 uth = (struct uthread *)get_bsdthread_info(th);
3640 if (uth) {
3641 bits = (((uth->uu_siglist & ~uth->uu_sigmask) & ~p->p_sigignore) & mask);
3642 }
3643 goto out;
3644 }
3645
3646 bits = 0;
3647 TAILQ_FOREACH(uth, &p->p_uthlist, uu_list) {
3648 bits |= (((uth->uu_siglist & ~uth->uu_sigmask) & ~p->p_sigignore) & mask);
3649 }
3650out:
3651 proc_unlock(p);
3652 return bits;
3653}
3654
3655int
3656thread_issignal(proc_t p, thread_t th, sigset_t mask)
3657{
3658 struct uthread * uth;
3659 sigset_t bits = 0;
3660
3661 proc_lock(p);
3662 uth = (struct uthread *)get_bsdthread_info(th);
3663 if (uth) {
3664 bits = (((uth->uu_siglist & ~uth->uu_sigmask) & ~p->p_sigignore) & mask);
3665 }
3666 proc_unlock(p);
3667 return bits;
3668}
3669
3670/*
3671 * Allow external reads of the sigprop array.
3672 */
3673int
3674hassigprop(int sig, int prop)
3675{
3676 return sigprop[sig] & prop;
3677}
3678
3679void
3680pgsigio(pid_t pgid, int sig)
3681{
3682 proc_t p = PROC_NULL;
3683
3684 if (pgid < 0) {
3685 gsignal(-(pgid), sig);
3686 } else if (pgid > 0 && (p = proc_find(pgid)) != 0) {
3687 psignal(p, sig);
3688 }
3689 if (p != PROC_NULL) {
3690 proc_rele(p);
3691 }
3692}
3693
3694void
3695proc_signalstart(proc_t p, int locked)
3696{
3697 if (!locked) {
3698 proc_lock(p);
3699 }
3700
3701 if (p->p_signalholder == current_thread()) {
3702 panic("proc_signalstart: thread attempting to signal a process for which it holds the signal lock");
3703 }
3704
3705 p->p_sigwaitcnt++;
3706 while ((p->p_lflag & P_LINSIGNAL) == P_LINSIGNAL) {
3707 msleep(&p->p_sigmask, &p->p_mlock, 0, "proc_signstart", NULL);
3708 }
3709 p->p_sigwaitcnt--;
3710
3711 p->p_lflag |= P_LINSIGNAL;
3712 p->p_signalholder = current_thread();
3713 if (!locked) {
3714 proc_unlock(p);
3715 }
3716}
3717
3718void
3719proc_signalend(proc_t p, int locked)
3720{
3721 if (!locked) {
3722 proc_lock(p);
3723 }
3724 p->p_lflag &= ~P_LINSIGNAL;
3725
3726 if (p->p_sigwaitcnt > 0) {
3727 wakeup(&p->p_sigmask);
3728 }
3729
3730 p->p_signalholder = NULL;
3731 if (!locked) {
3732 proc_unlock(p);
3733 }
3734}
3735
3736void
3737sig_lock_to_exit(proc_t p)
3738{
3739 thread_t self = current_thread();
3740
3741 p->exit_thread = self;
3742 proc_unlock(p);
3743
3744 task_hold(p->task);
3745 task_wait(p->task, FALSE);
3746
3747 proc_lock(p);
3748}
3749
3750int
3751sig_try_locked(proc_t p)
3752{
3753 thread_t self = current_thread();
3754
3755 while (p->sigwait || p->exit_thread) {
3756 if (p->exit_thread) {
3757 return 0;
3758 }
3759 msleep((caddr_t)&p->sigwait_thread, &p->p_mlock, PCATCH | PDROP, 0, 0);
3760 if (thread_should_abort(self)) {
3761 /*
3762 * Terminate request - clean up.
3763 */
3764 proc_lock(p);
3765 return -1;
3766 }
3767 proc_lock(p);
3768 }
3769 return 1;
3770}