X-Git-Url: https://git.saurik.com/apple/xnu.git/blobdiff_plain/527f99514973766e9c0382a4d8550dfb00f54939..c3c9b80d004dbbfdf763edeb97968c6997e3b45b:/bsd/kern/kern_sig.c diff --git a/bsd/kern/kern_sig.c b/bsd/kern/kern_sig.c index d390ded07..74ecaf338 100644 --- a/bsd/kern/kern_sig.c +++ b/bsd/kern/kern_sig.c @@ -2,7 +2,7 @@ * Copyright (c) 1995-2016 Apple Inc. All rights reserved. * * @APPLE_OSREFERENCE_LICENSE_HEADER_START@ - * + * * This file contains Original Code and/or Modifications of Original Code * as defined in and that are subject to the Apple Public Source License * Version 2.0 (the 'License'). You may not use this file except in @@ -11,10 +11,10 @@ * unlawful or unlicensed copies of an Apple operating system, or to * circumvent, violate, or enable the circumvention or violation of, any * terms of an Apple operating system software license agreement. - * + * * Please obtain a copy of the License at * http://www.opensource.apple.com/apsl/ and read it before using this file. - * + * * The Original Code and all software distributed under the License are * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES, @@ -22,7 +22,7 @@ * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT. * Please see the License for the specific language governing rights and * limitations under the License. - * + * * @APPLE_OSREFERENCE_LICENSE_HEADER_END@ */ /* @@ -71,7 +71,7 @@ * Version 2.0. */ -#define SIGPROP /* include signal properties table */ +#define SIGPROP /* include signal properties table */ #include #include #include @@ -98,10 +98,10 @@ #include #include -#include /* for coredump */ -#include /* for APC support */ +#include /* for coredump */ +#include /* for APC support */ #include -#include /* extern void *get_bsdtask_info(task_t); */ +#include /* extern void *get_bsdtask_info(task_t); */ #include #include #include @@ -114,6 +114,7 @@ #include #include +#include #include #if CONFIG_MACF @@ -126,21 +127,22 @@ * +++ */ extern int thread_enable_fpe(thread_t act, int onoff); -extern thread_t port_name_to_thread(mach_port_name_t port_name); -extern kern_return_t get_signalact(task_t , thread_t *, int); +extern kern_return_t get_signalact(task_t, thread_t *, int); extern unsigned int get_useraddr(void); extern boolean_t task_did_exec(task_t task); extern boolean_t task_is_exec_copy(task_t task); +extern void vm_shared_region_reslide_stale(void); /* * --- */ -extern void doexception(int exc, mach_exception_code_t code, - mach_exception_subcode_t sub); +extern void doexception(int exc, mach_exception_code_t code, + mach_exception_subcode_t sub); static void stop(proc_t, proc_t); -int cansignal(proc_t, kauth_cred_t, proc_t, int, int); +static int cansignal_nomac(proc_t, kauth_cred_t, proc_t, int); +int cansignal(proc_t, kauth_cred_t, proc_t, int); int killpg1(proc_t, int, int, int, int); kern_return_t do_bsdexception(int, int, int); void __posix_sem_syscall_return(kern_return_t); @@ -152,35 +154,34 @@ kern_return_t semaphore_timedwait_trap_internal(mach_port_name_t, unsigned int, kern_return_t semaphore_wait_signal_trap_internal(mach_port_name_t, mach_port_name_t, void (*)(kern_return_t)); kern_return_t semaphore_wait_trap_internal(mach_port_name_t, void (*)(kern_return_t)); -static int filt_sigattach(struct knote *kn, struct kevent_internal_s *kev); -static void filt_sigdetach(struct knote *kn); -static int filt_signal(struct knote *kn, long hint); -static int filt_signaltouch(struct knote *kn, struct kevent_internal_s *kev); -static int filt_signalprocess(struct knote *kn, struct filt_process_s *data, struct kevent_internal_s *kev); +static int filt_sigattach(struct knote *kn, struct kevent_qos_s *kev); +static void filt_sigdetach(struct knote *kn); +static int filt_signal(struct knote *kn, long hint); +static int filt_signaltouch(struct knote *kn, struct kevent_qos_s *kev); +static int filt_signalprocess(struct knote *kn, struct kevent_qos_s *kev); SECURITY_READ_ONLY_EARLY(struct filterops) sig_filtops = { - .f_attach = filt_sigattach, - .f_detach = filt_sigdetach, - .f_event = filt_signal, - .f_touch = filt_signaltouch, - .f_process = filt_signalprocess, + .f_attach = filt_sigattach, + .f_detach = filt_sigdetach, + .f_event = filt_signal, + .f_touch = filt_signaltouch, + .f_process = filt_signalprocess, }; /* structures and fns for killpg1 iterartion callback and filters */ struct killpg1_filtargs { - int posix; - proc_t cp; + bool posix; + proc_t curproc; }; struct killpg1_iterargs { - proc_t cp; + proc_t curproc; kauth_cred_t uc; int signum; - int * nfoundp; - int zombie; + int nfound; }; -static int killpg1_filt(proc_t p, void * arg); +static int killpg1_allfilt(proc_t p, void * arg); static int killpg1_pgrpfilt(proc_t p, __unused void * arg); static int killpg1_callback(proc_t p, void * arg); @@ -204,17 +205,17 @@ static void psignal_internal(proc_t p, task_t task, thread_t thread, int flavor, static void sigaltstack_kern_to_user32(struct kern_sigaltstack *in, struct user32_sigaltstack *out) { - out->ss_sp = CAST_DOWN_EXPLICIT(user32_addr_t, in->ss_sp); - out->ss_size = CAST_DOWN_EXPLICIT(user32_size_t, in->ss_size); - out->ss_flags = in->ss_flags; + out->ss_sp = CAST_DOWN_EXPLICIT(user32_addr_t, in->ss_sp); + out->ss_size = CAST_DOWN_EXPLICIT(user32_size_t, in->ss_size); + out->ss_flags = in->ss_flags; } static void sigaltstack_kern_to_user64(struct kern_sigaltstack *in, struct user64_sigaltstack *out) { - out->ss_sp = in->ss_sp; - out->ss_size = in->ss_size; - out->ss_flags = in->ss_flags; + out->ss_sp = in->ss_sp; + out->ss_size = in->ss_size; + out->ss_flags = in->ss_flags; } /* @@ -225,23 +226,23 @@ sigaltstack_kern_to_user64(struct kern_sigaltstack *in, struct user64_sigaltstac static void sigaltstack_user32_to_kern(struct user32_sigaltstack *in, struct kern_sigaltstack *out) { - out->ss_flags = in->ss_flags; - out->ss_size = in->ss_size; - out->ss_sp = CAST_USER_ADDR_T(in->ss_sp); + out->ss_flags = in->ss_flags; + out->ss_size = in->ss_size; + out->ss_sp = CAST_USER_ADDR_T(in->ss_sp); } static void sigaltstack_user64_to_kern(struct user64_sigaltstack *in, struct kern_sigaltstack *out) { - out->ss_flags = in->ss_flags; - out->ss_size = in->ss_size; - out->ss_sp = in->ss_sp; + out->ss_flags = in->ss_flags; + out->ss_size = (user_size_t)in->ss_size; + out->ss_sp = (user_addr_t)in->ss_sp; } static void sigaction_kern_to_user32(struct kern_sigaction *in, struct user32_sigaction *out) { /* This assumes 32 bit __sa_handler is of type sig_t */ - out->__sigaction_u.__sa_handler = CAST_DOWN_EXPLICIT(user32_addr_t,in->__sigaction_u.__sa_handler); + out->__sigaction_u.__sa_handler = CAST_DOWN_EXPLICIT(user32_addr_t, in->__sigaction_u.__sa_handler); out->sa_mask = in->sa_mask; out->sa_flags = in->sa_flags; } @@ -261,26 +262,35 @@ __sigaction_user32_to_kern(struct __user32_sigaction *in, struct __kern_sigactio out->sa_tramp = CAST_USER_ADDR_T(in->sa_tramp); out->sa_mask = in->sa_mask; out->sa_flags = in->sa_flags; + + kern_return_t kr; + kr = machine_thread_function_pointers_convert_from_user(current_thread(), + &out->sa_tramp, 1); + assert(kr == KERN_SUCCESS); } static void __sigaction_user64_to_kern(struct __user64_sigaction *in, struct __kern_sigaction *out) { - out->__sigaction_u.__sa_handler = in->__sigaction_u.__sa_handler; - out->sa_tramp = in->sa_tramp; + out->__sigaction_u.__sa_handler = (user_addr_t)in->__sigaction_u.__sa_handler; + out->sa_tramp = (user_addr_t)in->sa_tramp; out->sa_mask = in->sa_mask; out->sa_flags = in->sa_flags; + + kern_return_t kr; + kr = machine_thread_function_pointers_convert_from_user(current_thread(), + &out->sa_tramp, 1); + assert(kr == KERN_SUCCESS); } #if SIGNAL_DEBUG void ram_printf(int); -int ram_debug=0; -unsigned int rdebug_proc=0; +int ram_debug = 0; +unsigned int rdebug_proc = 0; void ram_printf(int x) { - printf("x is %d",x); - + printf("x is %d", x); } #endif /* SIGNAL_DEBUG */ @@ -291,75 +301,87 @@ signal_setast(thread_t sig_actthread) act_set_astbsd(sig_actthread); } -/* - * Can process p, with ucred uc, send the signal signum to process q? - * uc is refcounted by the caller so internal fileds can be used safely - * when called with zombie arg, list lock is held - */ -int -cansignal(proc_t p, kauth_cred_t uc, proc_t q, int signum, int zombie) +static int +cansignal_nomac(proc_t src, kauth_cred_t uc_src, proc_t dst, int signum) { - kauth_cred_t my_cred; - struct session * p_sessp = SESSION_NULL; - struct session * q_sessp = SESSION_NULL; -#if CONFIG_MACF - int error; - - error = mac_proc_check_signal(p, q, signum); - if (error) - return (0); -#endif - /* you can signal yourself */ - if (p == q) - return(1); + if (src == dst) { + return 1; + } + + /* you can't send the init proc SIGKILL, even if root */ + if (signum == SIGKILL && dst == initproc) { + return 0; + } - /* you can't send launchd SIGKILL, even if root */ - if (signum == SIGKILL && q == initproc) - return(0); + /* otherwise, root can always signal */ + if (kauth_cred_issuser(uc_src)) { + return 1; + } - if (!suser(uc, NULL)) - return (1); /* root can always signal */ + /* processes in the same session can send SIGCONT to each other */ + { + struct session *sess_src = SESSION_NULL; + struct session *sess_dst = SESSION_NULL; - if (zombie == 0) + /* The session field is protected by the list lock. */ proc_list_lock(); - if (p->p_pgrp != PGRP_NULL) - p_sessp = p->p_pgrp->pg_session; - if (q->p_pgrp != PGRP_NULL) - q_sessp = q->p_pgrp->pg_session; + if (src->p_pgrp != PGRP_NULL) { + sess_src = src->p_pgrp->pg_session; + } + if (dst->p_pgrp != PGRP_NULL) { + sess_dst = dst->p_pgrp->pg_session; + } + proc_list_unlock(); - if (signum == SIGCONT && q_sessp == p_sessp) { - if (zombie == 0) - proc_list_unlock(); - return (1); /* SIGCONT in session */ + /* allow SIGCONT within session and for processes without session */ + if (signum == SIGCONT && sess_src == sess_dst) { + return 1; + } } - if (zombie == 0) - proc_list_unlock(); + /* the source process must be authorized to signal the target */ + { + int allowed = 0; + kauth_cred_t uc_dst = NOCRED, uc_ref = NOCRED; - /* - * If the real or effective UID of the sender matches the real - * or saved UID of the target, permit the signal to - * be sent. - */ - if (zombie == 0) - my_cred = kauth_cred_proc_ref(q); - else - my_cred = proc_ucred(q); + uc_dst = uc_ref = kauth_cred_proc_ref(dst); + + /* + * If the real or effective UID of the sender matches the real or saved + * UID of the target, allow the signal to be sent. + */ + if (kauth_cred_getruid(uc_src) == kauth_cred_getruid(uc_dst) || + kauth_cred_getruid(uc_src) == kauth_cred_getsvuid(uc_dst) || + kauth_cred_getuid(uc_src) == kauth_cred_getruid(uc_dst) || + kauth_cred_getuid(uc_src) == kauth_cred_getsvuid(uc_dst)) { + allowed = 1; + } + + if (uc_ref != NOCRED) { + kauth_cred_unref(&uc_ref); + uc_ref = NOCRED; + } - if (kauth_cred_getruid(uc) == kauth_cred_getruid(my_cred) || - kauth_cred_getruid(uc) == kauth_cred_getsvuid(my_cred) || - kauth_cred_getuid(uc) == kauth_cred_getruid(my_cred) || - kauth_cred_getuid(uc) == kauth_cred_getsvuid(my_cred)) { - if (zombie == 0) - kauth_cred_unref(&my_cred); - return (1); + return allowed; } +} - if (zombie == 0) - kauth_cred_unref(&my_cred); +/* + * Can process `src`, with ucred `uc_src`, send the signal `signum` to process + * `dst`? The ucred is referenced by the caller so internal fileds can be used + * safely. + */ +int +cansignal(proc_t src, kauth_cred_t uc_src, proc_t dst, int signum) +{ +#if CONFIG_MACF + if (mac_proc_check_signal(src, dst, signum)) { + return 0; + } +#endif - return (0); + return cansignal_nomac(src, uc_src, dst, signum); } /* @@ -390,7 +412,7 @@ signal_is_restricted(proc_t p, int signum) { if (sigmask(signum) & sigrestrictmask()) { if (sigrestrict_arg == 0 && - task_get_apptype(p->task) == TASK_APPTYPE_APP_DEFAULT) { + task_get_apptype(p->task) == TASK_APPTYPE_APP_DEFAULT) { return ENOTSUP; } else { return EINVAL; @@ -431,32 +453,38 @@ sigaction(proc_t p, struct sigaction_args *uap, __unused int32_t *retval) struct sigacts *ps = p->p_sigacts; int signum; - int bit, error=0; + int bit, error = 0; + uint32_t sigreturn_validation = PS_SIGRETURN_VALIDATION_DEFAULT; signum = uap->signum; if (signum <= 0 || signum >= NSIG || - signum == SIGKILL || signum == SIGSTOP) - return (EINVAL); + signum == SIGKILL || signum == SIGSTOP) { + return EINVAL; + } if (uap->nsa) { if (IS_64BIT_PROCESS(p)) { - struct __user64_sigaction __vec64; + struct __user64_sigaction __vec64; error = copyin(uap->nsa, &__vec64, sizeof(__vec64)); __sigaction_user64_to_kern(&__vec64, &__vec); } else { - struct __user32_sigaction __vec32; + struct __user32_sigaction __vec32; error = copyin(uap->nsa, &__vec32, sizeof(__vec32)); __sigaction_user32_to_kern(&__vec32, &__vec); } - if (error) - return (error); + if (error) { + return error; + } + + sigreturn_validation = (__vec.sa_flags & SA_VALIDATE_SIGRETURN_FROM_SIGTRAMP) ? + PS_SIGRETURN_VALIDATION_ENABLED : PS_SIGRETURN_VALIDATION_DISABLED; __vec.sa_flags &= SA_USERSPACE_MASK; /* Only pass on valid sa_flags */ if ((__vec.sa_flags & SA_SIGINFO) || __vec.sa_handler != SIG_DFL) { if ((error = signal_is_restricted(p, signum))) { if (error == ENOTSUP) { printf("%s(%d): denied attempt to register action for signal %d\n", - proc_name_address(p), proc_pid(p), signum); + proc_name_address(p), proc_pid(p), signum); } return error; } @@ -468,78 +496,93 @@ sigaction(proc_t p, struct sigaction_args *uap, __unused int32_t *retval) sa->sa_mask = ps->ps_catchmask[signum]; bit = sigmask(signum); sa->sa_flags = 0; - if ((ps->ps_sigonstack & bit) != 0) + if ((ps->ps_sigonstack & bit) != 0) { sa->sa_flags |= SA_ONSTACK; - if ((ps->ps_sigintr & bit) == 0) + } + if ((ps->ps_sigintr & bit) == 0) { sa->sa_flags |= SA_RESTART; - if (ps->ps_siginfo & bit) + } + if (ps->ps_siginfo & bit) { sa->sa_flags |= SA_SIGINFO; - if (ps->ps_signodefer & bit) + } + if (ps->ps_signodefer & bit) { sa->sa_flags |= SA_NODEFER; - if (ps->ps_64regset & bit) - sa->sa_flags |= SA_64REGSET; - if ((signum == SIGCHLD) && (p->p_flag & P_NOCLDSTOP)) + } + if ((signum == SIGCHLD) && (p->p_flag & P_NOCLDSTOP)) { sa->sa_flags |= SA_NOCLDSTOP; - if ((signum == SIGCHLD) && (p->p_flag & P_NOCLDWAIT)) + } + if ((signum == SIGCHLD) && (p->p_flag & P_NOCLDWAIT)) { sa->sa_flags |= SA_NOCLDWAIT; + } if (IS_64BIT_PROCESS(p)) { - struct user64_sigaction vec64 = {}; + struct user64_sigaction vec64 = {}; sigaction_kern_to_user64(sa, &vec64); error = copyout(&vec64, uap->osa, sizeof(vec64)); } else { - struct user32_sigaction vec32 = {}; + struct user32_sigaction vec32 = {}; sigaction_kern_to_user32(sa, &vec32); error = copyout(&vec32, uap->osa, sizeof(vec32)); } - if (error) - return (error); + if (error) { + return error; + } } if (uap->nsa) { + uint32_t old_sigreturn_validation = atomic_load_explicit( + &ps->ps_sigreturn_validation, memory_order_relaxed); + if (old_sigreturn_validation == PS_SIGRETURN_VALIDATION_DEFAULT) { + atomic_compare_exchange_strong_explicit(&ps->ps_sigreturn_validation, + &old_sigreturn_validation, sigreturn_validation, + memory_order_relaxed, memory_order_relaxed); + } error = setsigvec(p, current_thread(), signum, &__vec, FALSE); } - return (error); + return error; } /* Routines to manipulate bits on all threads */ int -clear_procsiglist(proc_t p, int bit, boolean_t in_signalstart) +clear_procsiglist(proc_t p, int bit, boolean_t in_signalstart) { struct uthread * uth; thread_t thact; proc_lock(p); - if (!in_signalstart) + if (!in_signalstart) { proc_signalstart(p, 1); + } if ((p->p_lflag & P_LINVFORK) && p->p_vforkact) { - thact = p->p_vforkact; + thact = p->p_vforkact; uth = (struct uthread *)get_bsdthread_info(thact); if (uth) { uth->uu_siglist &= ~bit; } - if (!in_signalstart) + if (!in_signalstart) { proc_signalend(p, 1); + } proc_unlock(p); - return(0); - } + return 0; + } TAILQ_FOREACH(uth, &p->p_uthlist, uu_list) { uth->uu_siglist &= ~bit; } p->p_siglist &= ~bit; - if (!in_signalstart) + if (!in_signalstart) { proc_signalend(p, 1); + } proc_unlock(p); - return(0); + return 0; } static int -unblock_procsigmask(proc_t p, int bit) +unblock_procsigmask(proc_t p, int bit) { struct uthread * uth; thread_t thact; @@ -548,7 +591,7 @@ unblock_procsigmask(proc_t p, int bit) proc_signalstart(p, 1); if ((p->p_lflag & P_LINVFORK) && p->p_vforkact) { - thact = p->p_vforkact; + thact = p->p_vforkact; uth = (struct uthread *)get_bsdthread_info(thact); if (uth) { uth->uu_sigmask &= ~bit; @@ -556,8 +599,8 @@ unblock_procsigmask(proc_t p, int bit) p->p_sigmask &= ~bit; proc_signalend(p, 1); proc_unlock(p); - return(0); - } + return 0; + } TAILQ_FOREACH(uth, &p->p_uthlist, uu_list) { uth->uu_sigmask &= ~bit; } @@ -565,11 +608,11 @@ unblock_procsigmask(proc_t p, int bit) proc_signalend(p, 1); proc_unlock(p); - return(0); + return 0; } static int -block_procsigmask(proc_t p, int bit) +block_procsigmask(proc_t p, int bit) { struct uthread * uth; thread_t thact; @@ -578,7 +621,7 @@ block_procsigmask(proc_t p, int bit) proc_signalstart(p, 1); if ((p->p_lflag & P_LINVFORK) && p->p_vforkact) { - thact = p->p_vforkact; + thact = p->p_vforkact; uth = (struct uthread *)get_bsdthread_info(thact); if (uth) { uth->uu_sigmask |= bit; @@ -586,8 +629,8 @@ block_procsigmask(proc_t p, int bit) p->p_sigmask |= bit; proc_signalend(p, 1); proc_unlock(p); - return(0); - } + return 0; + } TAILQ_FOREACH(uth, &p->p_uthlist, uu_list) { uth->uu_sigmask |= bit; } @@ -595,11 +638,11 @@ block_procsigmask(proc_t p, int bit) proc_signalend(p, 1); proc_unlock(p); - return(0); + return 0; } int -set_procsigmask(proc_t p, int bit) +set_procsigmask(proc_t p, int bit) { struct uthread * uth; thread_t thact; @@ -608,7 +651,7 @@ set_procsigmask(proc_t p, int bit) proc_signalstart(p, 1); if ((p->p_lflag & P_LINVFORK) && p->p_vforkact) { - thact = p->p_vforkact; + thact = p->p_vforkact; uth = (struct uthread *)get_bsdthread_info(thact); if (uth) { uth->uu_sigmask = bit; @@ -616,8 +659,8 @@ set_procsigmask(proc_t p, int bit) p->p_sigmask = bit; proc_signalend(p, 1); proc_unlock(p); - return(0); - } + return 0; + } TAILQ_FOREACH(uth, &p->p_uthlist, uu_list) { uth->uu_sigmask = bit; } @@ -625,7 +668,7 @@ set_procsigmask(proc_t p, int bit) proc_signalend(p, 1); proc_unlock(p); - return(0); + return 0; } /* XXX should be static? */ @@ -648,52 +691,52 @@ setsigvec(proc_t p, __unused thread_t thread, int signum, struct __kern_sigactio assert(signum < NSIG); if ((signum == SIGKILL || signum == SIGSTOP) && - sa->sa_handler != SIG_DFL) - return(EINVAL); + sa->sa_handler != SIG_DFL) { + return EINVAL; + } bit = sigmask(signum); /* * Change setting atomically. */ ps->ps_sigact[signum] = sa->sa_handler; ps->ps_trampact[signum] = sa->sa_tramp; - ps->ps_catchmask[signum] = sa->sa_mask &~ sigcantmask; - if (sa->sa_flags & SA_SIGINFO) + ps->ps_catchmask[signum] = sa->sa_mask & ~sigcantmask; + if (sa->sa_flags & SA_SIGINFO) { ps->ps_siginfo |= bit; - else + } else { ps->ps_siginfo &= ~bit; - if (sa->sa_flags & SA_64REGSET) - ps->ps_64regset |= bit; - else - ps->ps_64regset &= ~bit; - if ((sa->sa_flags & SA_RESTART) == 0) + } + if ((sa->sa_flags & SA_RESTART) == 0) { ps->ps_sigintr |= bit; - else + } else { ps->ps_sigintr &= ~bit; - if (sa->sa_flags & SA_ONSTACK) + } + if (sa->sa_flags & SA_ONSTACK) { ps->ps_sigonstack |= bit; - else + } else { ps->ps_sigonstack &= ~bit; - if (sa->sa_flags & SA_USERTRAMP) - ps->ps_usertramp |= bit; - else - ps->ps_usertramp &= ~bit; - if (sa->sa_flags & SA_RESETHAND) + } + if (sa->sa_flags & SA_RESETHAND) { ps->ps_sigreset |= bit; - else + } else { ps->ps_sigreset &= ~bit; - if (sa->sa_flags & SA_NODEFER) + } + if (sa->sa_flags & SA_NODEFER) { ps->ps_signodefer |= bit; - else + } else { ps->ps_signodefer &= ~bit; + } if (signum == SIGCHLD) { - if (sa->sa_flags & SA_NOCLDSTOP) + if (sa->sa_flags & SA_NOCLDSTOP) { OSBitOrAtomic(P_NOCLDSTOP, &p->p_flag); - else + } else { OSBitAndAtomic(~((uint32_t)P_NOCLDSTOP), &p->p_flag); - if ((sa->sa_flags & SA_NOCLDWAIT) || (sa->sa_handler == SIG_IGN)) + } + if ((sa->sa_flags & SA_NOCLDWAIT) || (sa->sa_handler == SIG_IGN)) { OSBitOrAtomic(P_NOCLDWAIT, &p->p_flag); - else + } else { OSBitAndAtomic(~((uint32_t)P_NOCLDWAIT), &p->p_flag); + } } /* @@ -704,19 +747,20 @@ setsigvec(proc_t p, __unused thread_t thread, int signum, struct __kern_sigactio */ if (sa->sa_handler == SIG_IGN || (sigprop[signum] & SA_IGNORE && sa->sa_handler == SIG_DFL)) { - clear_procsiglist(p, bit, in_sigstart); - if (signum != SIGCONT) - p->p_sigignore |= bit; /* easier in psignal */ + if (signum != SIGCONT) { + p->p_sigignore |= bit; /* easier in psignal */ + } p->p_sigcatch &= ~bit; } else { p->p_sigignore &= ~bit; - if (sa->sa_handler == SIG_DFL) + if (sa->sa_handler == SIG_DFL) { p->p_sigcatch &= ~bit; - else + } else { p->p_sigcatch |= bit; + } } - return(0); + return 0; } /* @@ -728,9 +772,11 @@ siginit(proc_t p) { int i; - for (i = 1; i < NSIG; i++) - if (sigprop[i] & SA_IGNORE && i != SIGCONT) + for (i = 1; i < NSIG; i++) { + if (sigprop[i] & SA_IGNORE && i != SIGCONT) { p->p_sigignore |= sigmask(i); + } + } } /* @@ -763,17 +809,23 @@ execsigs(proc_t p, thread_t thread) * and are now ignored by default). */ while (p->p_sigcatch) { - nc = ffs((long)p->p_sigcatch); + nc = ffs((unsigned int)p->p_sigcatch); mask = sigmask(nc); p->p_sigcatch &= ~mask; if (sigprop[nc] & SA_IGNORE) { - if (nc != SIGCONT) + if (nc != SIGCONT) { p->p_sigignore |= mask; + } ut->uu_siglist &= ~mask; } ps->ps_sigact[nc] = SIG_DFL; } + atomic_store_explicit(&ps->ps_sigreturn_validation, + PS_SIGRETURN_VALIDATION_DEFAULT, memory_order_relaxed); + /* Generate random token value used to validate sigreturn arguments */ + read_random(&ps->ps_sigreturn_token, sizeof(ps->ps_sigreturn_token)); + /* * Reset stack state to the user stack. * Clear set of signals caught on the signal stack. @@ -809,8 +861,9 @@ sigprocmask(proc_t p, struct sigprocmask_args *uap, __unused int32_t *retval) goto out; } error = copyin(uap->mask, &nmask, sizeof(sigset_t)); - if (error) + if (error) { goto out; + } switch (uap->how) { case SIG_BLOCK: @@ -827,15 +880,16 @@ sigprocmask(proc_t p, struct sigprocmask_args *uap, __unused int32_t *retval) set_procsigmask(p, (nmask & ~sigcantmask)); signal_setast(current_thread()); break; - + default: error = EINVAL; break; } out: - if (!error && omask != USER_ADDR_NULL) + if (!error && omask != USER_ADDR_NULL) { copyout(&oldmask, omask, sizeof(sigset_t)); - return (error); + } + return error; } int @@ -847,9 +901,10 @@ sigpending(__unused proc_t p, struct sigpending_args *uap, __unused int32_t *ret ut = (struct uthread *)get_bsdthread_info(current_thread()); pendlist = ut->uu_siglist; - if (uap->osv) + if (uap->osv) { copyout(&pendlist, uap->osv, sizeof(sigset_t)); - return(0); + } + return 0; } /* @@ -869,7 +924,7 @@ int sigsuspend(proc_t p, struct sigsuspend_args *uap, int32_t *retval) { __pthread_testcancel(1); - return(sigsuspend_nocancel(p, (struct sigsuspend_nocancel_args *)uap, retval)); + return sigsuspend_nocancel(p, (struct sigsuspend_nocancel_args *)uap, retval); } int @@ -889,16 +944,16 @@ sigsuspend_nocancel(proc_t p, struct sigsuspend_nocancel_args *uap, __unused int ut->uu_oldmask = ut->uu_sigmask; ut->uu_flag |= UT_SAS_OLDMASK; ut->uu_sigmask = (uap->mask & ~sigcantmask); - (void) tsleep0((caddr_t) p, PPAUSE|PCATCH, "pause", 0, sigcontinue); + (void) tsleep0((caddr_t) p, PPAUSE | PCATCH, "pause", 0, sigcontinue); /* always return EINTR rather than ERESTART... */ - return (EINTR); + return EINTR; } int __disable_threadsignal(__unused proc_t p, - __unused struct __disable_threadsignal_args *uap, - __unused int32_t *retval) + __unused struct __disable_threadsignal_args *uap, + __unused int32_t *retval) { struct uthread *uth; @@ -907,28 +962,27 @@ __disable_threadsignal(__unused proc_t p, /* No longer valid to have any signal delivered */ uth->uu_flag |= (UT_NO_SIGMASK | UT_CANCELDISABLE); - return(0); - + return 0; } void __pthread_testcancel(int presyscall) { - thread_t self = current_thread(); struct uthread * uthread; uthread = (struct uthread *)get_bsdthread_info(self); - + uthread->uu_flag &= ~UT_NOTCANCELPT; if ((uthread->uu_flag & (UT_CANCELDISABLE | UT_CANCEL | UT_CANCELED)) == UT_CANCEL) { - if(presyscall != 0) { + if (presyscall != 0) { unix_syscall_return(EINTR); /* NOTREACHED */ - } else + } else { thread_abort_safely(self); + } } } @@ -936,39 +990,42 @@ __pthread_testcancel(int presyscall) int __pthread_markcancel(__unused proc_t p, - struct __pthread_markcancel_args *uap, __unused int32_t *retval) + struct __pthread_markcancel_args *uap, __unused int32_t *retval) { thread_act_t target_act; int error = 0; struct uthread *uth; - target_act = (thread_act_t)port_name_to_thread(uap->thread_port); + target_act = (thread_act_t)port_name_to_thread(uap->thread_port, + PORT_TO_THREAD_IN_CURRENT_TASK); - if (target_act == THR_ACT_NULL) - return (ESRCH); + if (target_act == THR_ACT_NULL) { + return ESRCH; + } uth = (struct uthread *)get_bsdthread_info(target_act); /* if the thread is in vfork do not cancel */ - if ((uth->uu_flag & (UT_VFORK | UT_CANCEL | UT_CANCELED )) == 0) { + if ((uth->uu_flag & (UT_VFORK | UT_CANCEL | UT_CANCELED)) == 0) { uth->uu_flag |= (UT_CANCEL | UT_NO_SIGMASK); - if (((uth->uu_flag & UT_NOTCANCELPT) == 0) - && ((uth->uu_flag & UT_CANCELDISABLE) == 0)) - thread_abort_safely(target_act); + if (((uth->uu_flag & UT_NOTCANCELPT) == 0) + && ((uth->uu_flag & UT_CANCELDISABLE) == 0)) { + thread_abort_safely(target_act); + } } thread_deallocate(target_act); - return (error); + return error; } -/* if action =0 ; return the cancellation state , +/* if action =0 ; return the cancellation state , * if marked for cancellation, make the thread canceled * if action = 1 ; Enable the cancel handling * if action = 2; Disable the cancel handling */ int __pthread_canceled(__unused proc_t p, - struct __pthread_canceled_args *uap, __unused int32_t *retval) + struct __pthread_canceled_args *uap, __unused int32_t *retval) { thread_act_t thread; struct uthread *uth; @@ -978,39 +1035,40 @@ __pthread_canceled(__unused proc_t p, uth = (struct uthread *)get_bsdthread_info(thread); switch (action) { - case 1: - uth->uu_flag &= ~UT_CANCELDISABLE; - return(0); - case 2: - uth->uu_flag |= UT_CANCELDISABLE; - return(0); - case 0: - default: - /* if the thread is in vfork do not cancel */ - if((uth->uu_flag & ( UT_CANCELDISABLE | UT_CANCEL | UT_CANCELED)) == UT_CANCEL) { - uth->uu_flag &= ~UT_CANCEL; - uth->uu_flag |= (UT_CANCELED | UT_NO_SIGMASK); - return(0); - } - return(EINVAL); - } - return(EINVAL); + case 1: + uth->uu_flag &= ~UT_CANCELDISABLE; + return 0; + case 2: + uth->uu_flag |= UT_CANCELDISABLE; + return 0; + case 0: + default: + /* if the thread is in vfork do not cancel */ + if ((uth->uu_flag & (UT_CANCELDISABLE | UT_CANCEL | UT_CANCELED)) == UT_CANCEL) { + uth->uu_flag &= ~UT_CANCEL; + uth->uu_flag |= (UT_CANCELED | UT_NO_SIGMASK); + return 0; + } + return EINVAL; + } + return EINVAL; } __attribute__((noreturn)) void -__posix_sem_syscall_return(kern_return_t kern_result) +__posix_sem_syscall_return(kern_return_t kern_result) { int error = 0; - if (kern_result == KERN_SUCCESS) + if (kern_result == KERN_SUCCESS) { error = 0; - else if (kern_result == KERN_ABORTED) + } else if (kern_result == KERN_ABORTED) { error = EINTR; - else if (kern_result == KERN_OPERATION_TIMED_OUT) + } else if (kern_result == KERN_OPERATION_TIMED_OUT) { error = ETIMEDOUT; - else + } else { error = EINVAL; + } unix_syscall_return(error); /* does not return */ } @@ -1025,94 +1083,92 @@ __posix_sem_syscall_return(kern_return_t kern_result) */ int __old_semwait_signal(proc_t p, struct __old_semwait_signal_args *uap, - int32_t *retval) + int32_t *retval) { __pthread_testcancel(0); - return(__old_semwait_signal_nocancel(p, (struct __old_semwait_signal_nocancel_args *)uap, retval)); + return __old_semwait_signal_nocancel(p, (struct __old_semwait_signal_nocancel_args *)uap, retval); } int __old_semwait_signal_nocancel(proc_t p, struct __old_semwait_signal_nocancel_args *uap, - __unused int32_t *retval) + __unused int32_t *retval) { - kern_return_t kern_result; int error; mach_timespec_t then; struct timespec now; struct user_timespec ts; boolean_t truncated_timeout = FALSE; - - if(uap->timeout) { - + + if (uap->timeout) { if (IS_64BIT_PROCESS(p)) { struct user64_timespec ts64; error = copyin(uap->ts, &ts64, sizeof(ts64)); - ts.tv_sec = ts64.tv_sec; - ts.tv_nsec = ts64.tv_nsec; + ts.tv_sec = (user_time_t)ts64.tv_sec; + ts.tv_nsec = (user_long_t)ts64.tv_nsec; } else { struct user32_timespec ts32; error = copyin(uap->ts, &ts32, sizeof(ts32)); ts.tv_sec = ts32.tv_sec; ts.tv_nsec = ts32.tv_nsec; } - + if (error) { return error; } - + if ((ts.tv_sec & 0xFFFFFFFF00000000ULL) != 0) { ts.tv_sec = 0xFFFFFFFF; ts.tv_nsec = 0; truncated_timeout = TRUE; } - + if (uap->relative) { - then.tv_sec = ts.tv_sec; - then.tv_nsec = ts.tv_nsec; + then.tv_sec = (unsigned int)ts.tv_sec; + then.tv_nsec = (clock_res_t)ts.tv_nsec; } else { nanotime(&now); - + /* if time has elapsed, set time to null timepsec to bailout rightaway */ if (now.tv_sec == ts.tv_sec ? - now.tv_nsec > ts.tv_nsec : - now.tv_sec > ts.tv_sec) { + now.tv_nsec > ts.tv_nsec : + now.tv_sec > ts.tv_sec) { then.tv_sec = 0; then.tv_nsec = 0; } else { - then.tv_sec = ts.tv_sec - now.tv_sec; - then.tv_nsec = ts.tv_nsec - now.tv_nsec; + then.tv_sec = (unsigned int)(ts.tv_sec - now.tv_sec); + then.tv_nsec = (clock_res_t)(ts.tv_nsec - now.tv_nsec); if (then.tv_nsec < 0) { then.tv_nsec += NSEC_PER_SEC; - then.tv_sec--; + then.tv_sec--; } } } - - if (uap->mutex_sem == 0) + + if (uap->mutex_sem == 0) { kern_result = semaphore_timedwait_trap_internal((mach_port_name_t)uap->cond_sem, then.tv_sec, then.tv_nsec, __posix_sem_syscall_return); - else + } else { kern_result = semaphore_timedwait_signal_trap_internal(uap->cond_sem, uap->mutex_sem, then.tv_sec, then.tv_nsec, __posix_sem_syscall_return); - + } } else { - - if (uap->mutex_sem == 0) + if (uap->mutex_sem == 0) { kern_result = semaphore_wait_trap_internal(uap->cond_sem, __posix_sem_syscall_return); - else - + } else { kern_result = semaphore_wait_signal_trap_internal(uap->cond_sem, uap->mutex_sem, __posix_sem_syscall_return); + } + } + + if (kern_result == KERN_SUCCESS && !truncated_timeout) { + return 0; + } else if (kern_result == KERN_SUCCESS && truncated_timeout) { + return EINTR; /* simulate an exceptional condition because Mach doesn't support a longer timeout */ + } else if (kern_result == KERN_ABORTED) { + return EINTR; + } else if (kern_result == KERN_OPERATION_TIMED_OUT) { + return ETIMEDOUT; + } else { + return EINVAL; } - - if (kern_result == KERN_SUCCESS && !truncated_timeout) - return(0); - else if (kern_result == KERN_SUCCESS && truncated_timeout) - return(EINTR); /* simulate an exceptional condition because Mach doesn't support a longer timeout */ - else if (kern_result == KERN_ABORTED) - return(EINTR); - else if (kern_result == KERN_OPERATION_TIMED_OUT) - return(ETIMEDOUT); - else - return(EINVAL); } #endif /* OLD_SEMWAIT_SIGNAL*/ @@ -1125,96 +1181,96 @@ __old_semwait_signal_nocancel(proc_t p, struct __old_semwait_signal_nocancel_arg */ int __semwait_signal(proc_t p, struct __semwait_signal_args *uap, - int32_t *retval) + int32_t *retval) { __pthread_testcancel(0); - return(__semwait_signal_nocancel(p, (struct __semwait_signal_nocancel_args *)uap, retval)); + return __semwait_signal_nocancel(p, (struct __semwait_signal_nocancel_args *)uap, retval); } int __semwait_signal_nocancel(__unused proc_t p, struct __semwait_signal_nocancel_args *uap, - __unused int32_t *retval) + __unused int32_t *retval) { - kern_return_t kern_result; mach_timespec_t then; struct timespec now; struct user_timespec ts; - boolean_t truncated_timeout = FALSE; - - if(uap->timeout) { - - ts.tv_sec = uap->tv_sec; - ts.tv_nsec = uap->tv_nsec; - - if ((ts.tv_sec & 0xFFFFFFFF00000000ULL) != 0) { - ts.tv_sec = 0xFFFFFFFF; - ts.tv_nsec = 0; - truncated_timeout = TRUE; - } - + boolean_t truncated_timeout = FALSE; + + if (uap->timeout) { + ts.tv_sec = (user_time_t)uap->tv_sec; + ts.tv_nsec = uap->tv_nsec; + + if ((ts.tv_sec & 0xFFFFFFFF00000000ULL) != 0) { + ts.tv_sec = 0xFFFFFFFF; + ts.tv_nsec = 0; + truncated_timeout = TRUE; + } + if (uap->relative) { - then.tv_sec = ts.tv_sec; - then.tv_nsec = ts.tv_nsec; + then.tv_sec = (unsigned int)ts.tv_sec; + then.tv_nsec = (clock_res_t)ts.tv_nsec; } else { nanotime(&now); - /* if time has elapsed, set time to null timepsec to bailout rightaway */ - if (now.tv_sec == ts.tv_sec ? - now.tv_nsec > ts.tv_nsec : - now.tv_sec > ts.tv_sec) { - then.tv_sec = 0; - then.tv_nsec = 0; - } else { - then.tv_sec = ts.tv_sec - now.tv_sec; - then.tv_nsec = ts.tv_nsec - now.tv_nsec; - if (then.tv_nsec < 0) { - then.tv_nsec += NSEC_PER_SEC; - then.tv_sec--; - } - } - } - - if (uap->mutex_sem == 0) + /* if time has elapsed, set time to null timepsec to bailout rightaway */ + if (now.tv_sec == ts.tv_sec ? + now.tv_nsec > ts.tv_nsec : + now.tv_sec > ts.tv_sec) { + then.tv_sec = 0; + then.tv_nsec = 0; + } else { + then.tv_sec = (unsigned int)(ts.tv_sec - now.tv_sec); + then.tv_nsec = (clock_res_t)(ts.tv_nsec - now.tv_nsec); + if (then.tv_nsec < 0) { + then.tv_nsec += NSEC_PER_SEC; + then.tv_sec--; + } + } + } + + if (uap->mutex_sem == 0) { kern_result = semaphore_timedwait_trap_internal((mach_port_name_t)uap->cond_sem, then.tv_sec, then.tv_nsec, __posix_sem_syscall_return); - else + } else { kern_result = semaphore_timedwait_signal_trap_internal(uap->cond_sem, uap->mutex_sem, then.tv_sec, then.tv_nsec, __posix_sem_syscall_return); - + } } else { - - if (uap->mutex_sem == 0) + if (uap->mutex_sem == 0) { kern_result = semaphore_wait_trap_internal(uap->cond_sem, __posix_sem_syscall_return); - else - + } else { kern_result = semaphore_wait_signal_trap_internal(uap->cond_sem, uap->mutex_sem, __posix_sem_syscall_return); + } + } + + if (kern_result == KERN_SUCCESS && !truncated_timeout) { + return 0; + } else if (kern_result == KERN_SUCCESS && truncated_timeout) { + return EINTR; /* simulate an exceptional condition because Mach doesn't support a longer timeout */ + } else if (kern_result == KERN_ABORTED) { + return EINTR; + } else if (kern_result == KERN_OPERATION_TIMED_OUT) { + return ETIMEDOUT; + } else { + return EINVAL; } - - if (kern_result == KERN_SUCCESS && !truncated_timeout) - return(0); - else if (kern_result == KERN_SUCCESS && truncated_timeout) - return(EINTR); /* simulate an exceptional condition because Mach doesn't support a longer timeout */ - else if (kern_result == KERN_ABORTED) - return(EINTR); - else if (kern_result == KERN_OPERATION_TIMED_OUT) - return(ETIMEDOUT); - else - return(EINVAL); } -int +int __pthread_kill(__unused proc_t p, struct __pthread_kill_args *uap, - __unused int32_t *retval) + __unused int32_t *retval) { thread_t target_act; int error = 0; int signum = uap->sig; struct uthread *uth; - target_act = (thread_t)port_name_to_thread(uap->thread_port); + target_act = (thread_t)port_name_to_thread(uap->thread_port, + PORT_TO_THREAD_NONE); - if (target_act == THREAD_NULL) - return (ESRCH); + if (target_act == THREAD_NULL) { + return ESRCH; + } if ((u_int)signum >= NSIG) { error = EINVAL; goto out; @@ -1227,17 +1283,23 @@ __pthread_kill(__unused proc_t p, struct __pthread_kill_args *uap, goto out; } - if (signum) + if ((thread_get_tag(target_act) & THREAD_TAG_WORKQUEUE) && !uth->uu_workq_pthread_kill_allowed) { + error = ENOTSUP; + goto out; + } + + if (signum) { psignal_uthread(target_act, signum); + } out: thread_deallocate(target_act); - return (error); + return error; } -int +int __pthread_sigmask(__unused proc_t p, struct __pthread_sigmask_args *uap, - __unused int32_t *retval) + __unused int32_t *retval) { user_addr_t set = uap->set; user_addr_t oset = uap->oset; @@ -1255,8 +1317,9 @@ __pthread_sigmask(__unused proc_t p, struct __pthread_sigmask_args *uap, } error = copyin(set, &nset, sizeof(sigset_t)); - if (error) + if (error) { goto out; + } switch (uap->how) { case SIG_BLOCK: @@ -1272,16 +1335,16 @@ __pthread_sigmask(__unused proc_t p, struct __pthread_sigmask_args *uap, ut->uu_sigmask = (nset & ~sigcantmask); signal_setast(current_thread()); break; - + default: error = EINVAL; - } out: - if (!error && oset != USER_ADDR_NULL) + if (!error && oset != USER_ADDR_NULL) { copyout(&oldset, oset, sizeof(sigset_t)); + } - return(error); + return error; } /* @@ -1290,14 +1353,14 @@ out: * copyin:EFAULT * copyout:EFAULT */ -int +int __sigwait(proc_t p, struct __sigwait_args *uap, int32_t *retval) { __pthread_testcancel(1); - return(__sigwait_nocancel(p, (struct __sigwait_nocancel_args *)uap, retval)); + return __sigwait_nocancel(p, (struct __sigwait_nocancel_args *)uap, retval); } -int +int __sigwait_nocancel(proc_t p, struct __sigwait_nocancel_args *uap, __unused int32_t *retval) { struct uthread *ut; @@ -1305,31 +1368,34 @@ __sigwait_nocancel(proc_t p, struct __sigwait_nocancel_args *uap, __unused int32 int error = 0; sigset_t mask; sigset_t siglist; - sigset_t sigw=0; + sigset_t sigw = 0; int signum; ut = (struct uthread *)get_bsdthread_info(current_thread()); - if (uap->set == USER_ADDR_NULL) - return(EINVAL); + if (uap->set == USER_ADDR_NULL) { + return EINVAL; + } error = copyin(uap->set, &mask, sizeof(sigset_t)); - if (error) - return(error); + if (error) { + return error; + } siglist = (mask & ~sigcantmask); - if (siglist == 0) - return(EINVAL); + if (siglist == 0) { + return EINVAL; + } proc_lock(p); if ((p->p_lflag & P_LINVFORK) && p->p_vforkact) { proc_unlock(p); - return(EINVAL); + return EINVAL; } else { proc_signalstart(p, 1); TAILQ_FOREACH(uth, &p->p_uthlist, uu_list) { - if ( (sigw = uth->uu_siglist & siglist) ) { + if ((sigw = uth->uu_siglist & siglist)) { break; } } @@ -1347,22 +1413,23 @@ __sigwait_nocancel(proc_t p, struct __sigwait_nocancel_args *uap, __unused int32 * save it here and mark the sigacts structure * to indicate this. */ - uth = ut; /* wait for it to be delivered to us */ + uth = ut; /* wait for it to be delivered to us */ ut->uu_oldmask = ut->uu_sigmask; ut->uu_flag |= UT_SAS_OLDMASK; if (siglist == (sigset_t)0) { proc_unlock(p); - return(EINVAL); + return EINVAL; } /* SIGKILL and SIGSTOP are not maskable as well */ - ut->uu_sigmask = ~(siglist|sigcantmask); - ut->uu_sigwait = siglist; + ut->uu_sigmask = ~(siglist | sigcantmask); + ut->uu_sigwait = siglist; /* No Continuations for now */ - error = msleep((caddr_t)&ut->uu_sigwait, &p->p_mlock, PPAUSE|PCATCH, "pause", 0); + error = msleep((caddr_t)&ut->uu_sigwait, &p->p_mlock, PPAUSE | PCATCH, "pause", 0); - if (error == ERESTART) + if (error == ERESTART) { error = 0; + } sigw = (ut->uu_sigwait & siglist); ut->uu_sigmask = ut->uu_oldmask; @@ -1372,8 +1439,9 @@ sigwait1: ut->uu_sigwait = 0; if (!error) { signum = ffs((unsigned int)sigw); - if (!signum) + if (!signum) { panic("sigwait with no signal wakeup"); + } /* Clear the pending signal in the thread it was delivered */ uth->uu_siglist &= ~(sigmask(signum)); @@ -1382,13 +1450,14 @@ sigwait1: #endif proc_unlock(p); - if (uap->sig != USER_ADDR_NULL) - error = copyout(&signum, uap->sig, sizeof(int)); - } else + if (uap->sig != USER_ADDR_NULL) { + error = copyout(&signum, uap->sig, sizeof(int)); + } + } else { proc_unlock(p); + } - return(error); - + return error; } int @@ -1403,24 +1472,27 @@ sigaltstack(__unused proc_t p, struct sigaltstack_args *uap, __unused int32_t *r uth = (struct uthread *)get_bsdthread_info(current_thread()); pstk = &uth->uu_sigstk; - if ((uth->uu_flag & UT_ALTSTACK) == 0) + if ((uth->uu_flag & UT_ALTSTACK) == 0) { uth->uu_sigstk.ss_flags |= SA_DISABLE; + } onstack = pstk->ss_flags & SA_ONSTACK; if (uap->oss) { if (IS_64BIT_PROCESS(p)) { struct user64_sigaltstack ss64 = {}; - sigaltstack_kern_to_user64(pstk, &ss64); + sigaltstack_kern_to_user64(pstk, &ss64); error = copyout(&ss64, uap->oss, sizeof(ss64)); } else { struct user32_sigaltstack ss32 = {}; - sigaltstack_kern_to_user32(pstk, &ss32); + sigaltstack_kern_to_user32(pstk, &ss32); error = copyout(&ss32, uap->oss, sizeof(ss32)); } - if (error) - return (error); + if (error) { + return error; + } + } + if (uap->nss == USER_ADDR_NULL) { + return 0; } - if (uap->nss == USER_ADDR_NULL) - return (0); if (IS_64BIT_PROCESS(p)) { struct user64_sigaltstack ss64; error = copyin(uap->nss, &ss64, sizeof(ss64)); @@ -1430,29 +1502,33 @@ sigaltstack(__unused proc_t p, struct sigaltstack_args *uap, __unused int32_t *r error = copyin(uap->nss, &ss32, sizeof(ss32)); sigaltstack_user32_to_kern(&ss32, &ss); } - if (error) - return (error); - if ((ss.ss_flags & ~SA_DISABLE) != 0) { - return(EINVAL); + if (error) { + return error; + } + if ((ss.ss_flags & ~SA_DISABLE) != 0) { + return EINVAL; } if (ss.ss_flags & SA_DISABLE) { /* if we are here we are not in the signal handler ;so no need to check */ - if (uth->uu_sigstk.ss_flags & SA_ONSTACK) - return (EINVAL); + if (uth->uu_sigstk.ss_flags & SA_ONSTACK) { + return EINVAL; + } uth->uu_flag &= ~UT_ALTSTACK; uth->uu_sigstk.ss_flags = ss.ss_flags; - return (0); + return 0; + } + if (onstack) { + return EPERM; } - if (onstack) - return (EPERM); /* The older stacksize was 8K, enforce that one so no compat problems */ #define OLDMINSIGSTKSZ 8*1024 - if (ss.ss_size < OLDMINSIGSTKSZ) - return (ENOMEM); + if (ss.ss_size < OLDMINSIGSTKSZ) { + return ENOMEM; + } uth->uu_flag |= UT_ALTSTACK; - uth->uu_sigstk= ss; - return (0); + uth->uu_sigstk = ss; + return 0; } int @@ -1460,49 +1536,51 @@ kill(proc_t cp, struct kill_args *uap, __unused int32_t *retval) { proc_t p; kauth_cred_t uc = kauth_cred_get(); - int posix = uap->posix; /* !0 if posix behaviour desired */ + int posix = uap->posix; /* !0 if posix behaviour desired */ - AUDIT_ARG(pid, uap->pid); - AUDIT_ARG(signum, uap->signum); + AUDIT_ARG(pid, uap->pid); + AUDIT_ARG(signum, uap->signum); - if ((u_int)uap->signum >= NSIG) - return (EINVAL); + if ((u_int)uap->signum >= NSIG) { + return EINVAL; + } if (uap->pid > 0) { /* kill single process */ if ((p = proc_find(uap->pid)) == NULL) { if ((p = pzfind(uap->pid)) != NULL) { /* - * IEEE Std 1003.1-2001: return success - * when killing a zombie. + * POSIX 1003.1-2001 requires returning success when killing a + * zombie; see Rationale for kill(2). */ - return (0); + return 0; } - return (ESRCH); + return ESRCH; } AUDIT_ARG(process, p); - if (!cansignal(cp, uc, p, uap->signum, 0)) { + if (!cansignal(cp, uc, p, uap->signum)) { proc_rele(p); - return(EPERM); + return EPERM; } - if (uap->signum) + if (uap->signum) { psignal(p, uap->signum); + } proc_rele(p); - return (0); + return 0; } switch (uap->pid) { - case -1: /* broadcast signal */ - return (killpg1(cp, uap->signum, 0, 1, posix)); - case 0: /* signal own process group */ - return (killpg1(cp, uap->signum, 0, 0, posix)); - default: /* negative explicit process group */ - return (killpg1(cp, uap->signum, -(uap->pid), 0, posix)); + case -1: /* broadcast signal */ + return killpg1(cp, uap->signum, 0, 1, posix); + case 0: /* signal own process group */ + return killpg1(cp, uap->signum, 0, 0, posix); + default: /* negative explicit process group */ + return killpg1(cp, uap->signum, -(uap->pid), 0, posix); } /* NOTREACHED */ } os_reason_t build_userspace_exit_reason(uint32_t reason_namespace, uint64_t reason_code, user_addr_t payload, uint32_t payload_size, - user_addr_t reason_string, uint64_t reason_flags) + user_addr_t reason_string, uint64_t reason_flags) { os_reason_t exit_reason = OS_REASON_NULL; @@ -1526,7 +1604,7 @@ build_userspace_exit_reason(uint32_t reason_namespace, uint64_t reason_code, use exit_reason->osr_flags |= (reason_flags & OS_REASON_FLAG_MASK_ALLOWED_FROM_USER); if ((reason_flags & OS_REASON_FLAG_MASK_ALLOWED_FROM_USER) != reason_flags) { printf("build_userspace_exit_reason: illegal flags passed from userspace (some masked off) 0x%llx, ns: %u, code 0x%llx\n", - reason_flags, reason_namespace, reason_code); + reason_flags, reason_namespace, reason_code); } if (!(exit_reason->osr_flags & OS_REASON_FLAG_NO_CRASH_REPORT)) { @@ -1536,7 +1614,7 @@ build_userspace_exit_reason(uint32_t reason_namespace, uint64_t reason_code, use if (payload != USER_ADDR_NULL) { if (payload_size == 0) { printf("build_userspace_exit_reason: exit reason with namespace %u, nonzero payload but zero length\n", - reason_namespace); + reason_namespace); exit_reason->osr_flags |= OS_REASON_FLAG_BAD_PARAMS; payload = USER_ADDR_NULL; } else { @@ -1552,11 +1630,12 @@ build_userspace_exit_reason(uint32_t reason_namespace, uint64_t reason_code, use } if (reason_string != USER_ADDR_NULL) { - reason_user_desc = (char *) kalloc(EXIT_REASON_USER_DESC_MAX_LEN); + reason_user_desc = kheap_alloc(KHEAP_TEMP, + EXIT_REASON_USER_DESC_MAX_LEN, Z_WAITOK); if (reason_user_desc != NULL) { error = copyinstr(reason_string, (void *) reason_user_desc, - EXIT_REASON_USER_DESC_MAX_LEN, &reason_user_desc_len); + EXIT_REASON_USER_DESC_MAX_LEN, &reason_user_desc_len); if (error == 0) { num_items_to_copy++; @@ -1567,7 +1646,8 @@ build_userspace_exit_reason(uint32_t reason_namespace, uint64_t reason_code, use user_data_to_copy += reason_user_desc_len; } else { exit_reason->osr_flags |= OS_REASON_FLAG_FAILED_DATA_COPYIN; - kfree(reason_user_desc, EXIT_REASON_USER_DESC_MAX_LEN); + kheap_free(KHEAP_TEMP, reason_user_desc, + EXIT_REASON_USER_DESC_MAX_LEN); reason_user_desc = NULL; reason_user_desc_len = 0; } @@ -1588,12 +1668,11 @@ build_userspace_exit_reason(uint32_t reason_namespace, uint64_t reason_code, use if (reason_user_desc != NULL && reason_user_desc_len != 0) { if (KERN_SUCCESS == kcdata_get_memory_addr(&exit_reason->osr_kcd_descriptor, - EXIT_REASON_USER_DESC, - reason_user_desc_len, - &data_addr)) { - + EXIT_REASON_USER_DESC, + (uint32_t)reason_user_desc_len, + &data_addr)) { kcdata_memcpy(&exit_reason->osr_kcd_descriptor, (mach_vm_address_t) data_addr, - reason_user_desc, reason_user_desc_len); + reason_user_desc, (uint32_t)reason_user_desc_len); } else { printf("build_userspace_exit_reason: failed to allocate space for reason string\n"); goto out_failed_copyin; @@ -1602,10 +1681,10 @@ build_userspace_exit_reason(uint32_t reason_namespace, uint64_t reason_code, use if (payload != USER_ADDR_NULL) { if (KERN_SUCCESS == - kcdata_get_memory_addr(&exit_reason->osr_kcd_descriptor, - EXIT_REASON_USER_PAYLOAD, - payload_size, - &data_addr)) { + kcdata_get_memory_addr(&exit_reason->osr_kcd_descriptor, + EXIT_REASON_USER_PAYLOAD, + payload_size, + &data_addr)) { error = copyin(payload, (void *) data_addr, payload_size); if (error) { printf("build_userspace_exit_reason: failed to copy in payload data with error %d\n", error); @@ -1619,7 +1698,7 @@ build_userspace_exit_reason(uint32_t reason_namespace, uint64_t reason_code, use } if (reason_user_desc != NULL) { - kfree(reason_user_desc, EXIT_REASON_USER_DESC_MAX_LEN); + kheap_free(KHEAP_TEMP, reason_user_desc, EXIT_REASON_USER_DESC_MAX_LEN); reason_user_desc = NULL; reason_user_desc_len = 0; } @@ -1629,7 +1708,7 @@ build_userspace_exit_reason(uint32_t reason_namespace, uint64_t reason_code, use out_failed_copyin: if (reason_user_desc != NULL) { - kfree(reason_user_desc, EXIT_REASON_USER_DESC_MAX_LEN); + kheap_free(KHEAP_TEMP, reason_user_desc, EXIT_REASON_USER_DESC_MAX_LEN); reason_user_desc = NULL; reason_user_desc_len = 0; } @@ -1641,8 +1720,8 @@ out_failed_copyin: static int terminate_with_payload_internal(struct proc *cur_proc, int target_pid, uint32_t reason_namespace, - uint64_t reason_code, user_addr_t payload, uint32_t payload_size, - user_addr_t reason_string, uint64_t reason_flags) + uint64_t reason_code, user_addr_t payload, uint32_t payload_size, + user_addr_t reason_string, uint64_t reason_flags) { proc_t target_proc = PROC_NULL; kauth_cred_t cur_cred = kauth_cred_get(); @@ -1661,17 +1740,26 @@ terminate_with_payload_internal(struct proc *cur_proc, int target_pid, uint32_t AUDIT_ARG(process, target_proc); - if (!cansignal(cur_proc, cur_cred, target_proc, SIGKILL, 0)) { + if (!cansignal(cur_proc, cur_cred, target_proc, SIGKILL)) { proc_rele(target_proc); return EPERM; } + if (target_pid != cur_proc->p_pid) { + /* + * FLAG_ABORT should only be set on terminate_with_reason(getpid()) that + * was a fallback from an unsuccessful abort_with_reason(). In that case + * caller's pid matches the target one. Otherwise remove the flag. + */ + reason_flags &= ~((typeof(reason_flags))OS_REASON_FLAG_ABORT); + } + KERNEL_DEBUG_CONSTANT(BSDDBG_CODE(DBG_BSD_PROC, BSD_PROC_EXITREASON_CREATE) | DBG_FUNC_NONE, - target_proc->p_pid, reason_namespace, - reason_code, 0, 0); + target_proc->p_pid, reason_namespace, + reason_code, 0, 0); signal_reason = build_userspace_exit_reason(reason_namespace, reason_code, payload, payload_size, - reason_string, (reason_flags | OS_REASON_FLAG_NO_CRASHED_TID)); + reason_string, (reason_flags | OS_REASON_FLAG_NO_CRASHED_TID)); if (target_pid == cur_proc->p_pid) { /* @@ -1691,116 +1779,93 @@ terminate_with_payload_internal(struct proc *cur_proc, int target_pid, uint32_t int terminate_with_payload(struct proc *cur_proc, struct terminate_with_payload_args *args, - __unused int32_t *retval) + __unused int32_t *retval) { return terminate_with_payload_internal(cur_proc, args->pid, args->reason_namespace, args->reason_code, args->payload, - args->payload_size, args->reason_string, args->reason_flags); + args->payload_size, args->reason_string, args->reason_flags); } static int -killpg1_filt(proc_t p, void * arg) +killpg1_allfilt(proc_t p, void * arg) { struct killpg1_filtargs * kfargp = (struct killpg1_filtargs *)arg; - proc_t cp = kfargp->cp; - int posix = kfargp->posix; - - if (p->p_pid <= 1 || p->p_flag & P_SYSTEM || - (!posix && p == cp)) - return(0); - else - return(1); + /* + * Don't signal initproc, a system process, or the current process if POSIX + * isn't specified. + */ + return p->p_pid > 1 && !(p->p_flag & P_SYSTEM) && + (kfargp->posix ? true : p != kfargp->curproc); } - static int killpg1_pgrpfilt(proc_t p, __unused void * arg) { - if (p->p_pid <= 1 || p->p_flag & P_SYSTEM || - (p->p_stat == SZOMB)) - return(0); - else - return(1); + /* XXX shouldn't this allow signalling zombies? */ + return p->p_pid > 1 && !(p->p_flag & P_SYSTEM) && p->p_stat != SZOMB; } - - static int -killpg1_callback(proc_t p, void * arg) -{ - struct killpg1_iterargs * kargp = (struct killpg1_iterargs *)arg; - proc_t cp = kargp->cp; - kauth_cred_t uc = kargp->uc; /* refcounted by the caller safe to use internal fields */ - int signum = kargp->signum; - int * nfoundp = kargp->nfoundp; - int n; - int zombie = 0; - int error = 0; - - if ((kargp->zombie != 0) && ((p->p_listflag & P_LIST_EXITED) == P_LIST_EXITED)) - zombie = 1; +killpg1_callback(proc_t p, void *arg) +{ + struct killpg1_iterargs *kargp = (struct killpg1_iterargs *)arg; + int signum = kargp->signum; - if (zombie != 0) { - proc_list_lock(); - error = cansignal(cp, uc, p, signum, zombie); - proc_list_unlock(); - - if (error != 0 && nfoundp != NULL) { - n = *nfoundp; - *nfoundp = n+1; + if ((p->p_listflag & P_LIST_EXITED) == P_LIST_EXITED) { + /* + * Count zombies as found for the purposes of signalling, since POSIX + * 1003.1-2001 sees signalling zombies as successful. If killpg(2) or + * kill(2) with pid -1 only finds zombies that can be signalled, it + * shouldn't return ESRCH. See the Rationale for kill(2). + * + * Don't call into MAC -- it's not expecting signal checks for exited + * processes. + */ + if (cansignal_nomac(kargp->curproc, kargp->uc, p, signum)) { + kargp->nfound++; } - } else { - if (cansignal(cp, uc, p, signum, 0) == 0) - return(PROC_RETURNED); + } else if (cansignal(kargp->curproc, kargp->uc, p, signum)) { + kargp->nfound++; - if (nfoundp != NULL) { - n = *nfoundp; - *nfoundp = n+1; - } - if (signum != 0) + if (signum != 0) { psignal(p, signum); + } } - return(PROC_RETURNED); + return PROC_RETURNED; } /* * Common code for kill process group/broadcast kill. - * cp is calling process. */ int -killpg1(proc_t cp, int signum, int pgid, int all, int posix) +killpg1(proc_t curproc, int signum, int pgid, int all, int posix) { kauth_cred_t uc; struct pgrp *pgrp; - int nfound = 0; - struct killpg1_iterargs karg; - struct killpg1_filtargs kfarg; int error = 0; - - uc = kauth_cred_proc_ref(cp); - if (all) { - /* - * broadcast - */ - kfarg.posix = posix; - kfarg.cp = cp; - - karg.cp = cp; - karg.uc = uc; - karg.nfoundp = &nfound; - karg.signum = signum; - karg.zombie = 1; - proc_iterate((PROC_ALLPROCLIST | PROC_ZOMBPROCLIST), killpg1_callback, &karg, killpg1_filt, (void *)&kfarg); + uc = kauth_cred_proc_ref(curproc); + struct killpg1_iterargs karg = { + .curproc = curproc, .uc = uc, .nfound = 0, .signum = signum + }; + if (all) { + /* + * Broadcast to all processes that the user can signal (pid was -1). + */ + struct killpg1_filtargs kfarg = { + .posix = posix, .curproc = curproc + }; + proc_iterate(PROC_ALLPROCLIST | PROC_ZOMBPROCLIST, killpg1_callback, + &karg, killpg1_allfilt, &kfarg); } else { if (pgid == 0) { - /* - * zero pgid means send to my process group. + /* + * Send to current the current process' process group. */ - pgrp = proc_pgrp(cp); - } else { + pgrp = proc_pgrp(curproc); + } else { pgrp = pgfind(pgid); if (pgrp == NULL) { error = ESRCH; @@ -1808,24 +1873,16 @@ killpg1(proc_t cp, int signum, int pgid, int all, int posix) } } - karg.nfoundp = &nfound; - karg.uc = uc; - karg.signum = signum; - karg.cp = cp; - karg.zombie = 0; - - /* PGRP_DROPREF drops the pgrp refernce */ pgrp_iterate(pgrp, PGRP_DROPREF, killpg1_callback, &karg, - killpg1_pgrpfilt, NULL); + killpg1_pgrpfilt, NULL); } - error = (nfound ? 0 : (posix ? EPERM : ESRCH)); + error = (karg.nfound > 0 ? 0 : (posix ? EPERM : ESRCH)); out: kauth_cred_unref(&uc); - return (error); + return error; } - /* * Send a signal to a process group. */ @@ -1850,20 +1907,21 @@ pgsignal_filt(proc_t p, void * arg) { int checkctty = *(int*)arg; - if ((checkctty == 0) || p->p_flag & P_CONTROLT) - return(1); - else - return(0); + if ((checkctty == 0) || p->p_flag & P_CONTROLT) { + return 1; + } else { + return 0; + } } static int pgsignal_callback(proc_t p, void * arg) { - int signum = *(int*)arg; + int signum = *(int*)arg; psignal(p, signum); - return(PROC_RETURNED); + return PROC_RETURNED; } @@ -1898,18 +1956,21 @@ threadsignal(thread_t sig_actthread, int signum, mach_exception_code_t code, boo proc_t p; int mask; - if ((u_int)signum >= NSIG || signum == 0) + if ((u_int)signum >= NSIG || signum == 0) { return; + } mask = sigmask(signum); - if ((mask & threadmask) == 0) + if ((mask & threadmask) == 0) { return; + } sig_task = get_threadtask(sig_actthread); p = (proc_t)(get_bsdtask_info(sig_task)); uth = get_bsdthread_info(sig_actthread); - if (uth->uu_flag & UT_VFORK) + if (uth->uu_flag & UT_VFORK) { p = uth->uu_proc; + } proc_lock(p); if (!(p->p_lflag & P_LTRACED) && (p->p_sigignore & mask)) { @@ -1922,12 +1983,11 @@ threadsignal(thread_t sig_actthread, int signum, mach_exception_code_t code, boo /* Attempt to establish whether the signal will be fatal (mirrors logic in psignal_internal()) */ if (set_exitreason && ((p->p_lflag & P_LTRACED) || (!(uth->uu_sigwait & mask) - && !(uth->uu_sigmask & mask) && !(p->p_sigcatch & mask))) && - !(mask & stopsigmask) && !(mask & contsigmask)) { - + && !(uth->uu_sigmask & mask) && !(p->p_sigcatch & mask))) && + !(mask & stopsigmask) && !(mask & contsigmask)) { if (uth->uu_exit_reason == OS_REASON_NULL) { KERNEL_DEBUG_CONSTANT(BSDDBG_CODE(DBG_BSD_PROC, BSD_PROC_EXITREASON_CREATE) | DBG_FUNC_NONE, - p->p_pid, OS_REASON_SIGNAL, signum, 0, 0); + p->p_pid, OS_REASON_SIGNAL, signum, 0, 0); os_reason_t signal_reason = build_signal_reason(signum, "exc handler"); @@ -1944,6 +2004,46 @@ threadsignal(thread_t sig_actthread, int signum, mach_exception_code_t code, boo signal_setast(sig_actthread); } +/* Called with proc locked */ +static void +set_thread_extra_flags(struct uthread *uth, os_reason_t reason) +{ + extern int vm_shared_region_reslide_restrict; + assert(uth != NULL); + /* + * Check whether the userland fault address falls within the shared + * region and notify userland if so. This allows launchd to apply + * special policies around this fault type. + */ + if (reason->osr_namespace == OS_REASON_SIGNAL && + reason->osr_code == SIGSEGV) { + mach_vm_address_t fault_address = uth->uu_subcode; + +#if defined(__arm64__) + /* taken from osfmk/arm/misc_protos.h */ + #define TBI_MASK 0xff00000000000000 + #define tbi_clear(addr) ((addr) & ~(TBI_MASK)) + fault_address = tbi_clear(fault_address); +#endif /* __arm64__ */ + + if (fault_address >= SHARED_REGION_BASE && + fault_address <= SHARED_REGION_BASE + SHARED_REGION_SIZE) { + /* + * Always report whether the fault happened within the shared cache + * region, but only stale the slide if the resliding is extended + * to all processes or if the process faulting is a platform one. + */ + reason->osr_flags |= OS_REASON_FLAG_SHAREDREGION_FAULT; + +#if __has_feature(ptrauth_calls) + if (!vm_shared_region_reslide_restrict || csproc_get_platform_binary(current_proc())) { + vm_shared_region_reslide_stale(); + } +#endif /* __has_feature(ptrauth_calls) */ + } + } +} + void set_thread_exit_reason(void *th, void *reason, boolean_t proc_locked) { @@ -1953,8 +2053,9 @@ set_thread_exit_reason(void *th, void *reason, boolean_t proc_locked) os_reason_t exit_reason = (os_reason_t)reason; - if (exit_reason == OS_REASON_NULL) + if (exit_reason == OS_REASON_NULL) { return; + } if (!proc_locked) { targ_task = get_threadtask(th); @@ -1963,6 +2064,8 @@ set_thread_exit_reason(void *th, void *reason, boolean_t proc_locked) proc_lock(targ_proc); } + set_thread_extra_flags(targ_uth, exit_reason); + if (targ_uth->uu_exit_reason == OS_REASON_NULL) { targ_uth->uu_exit_reason = exit_reason; } else { @@ -1996,33 +2099,45 @@ get_signalthread(proc_t p, int signum, thread_t * thr) thread_t sig_thread; struct task * sig_task = p->task; kern_return_t kret; + bool skip_wqthreads = true; *thr = THREAD_NULL; if ((p->p_lflag & P_LINVFORK) && p->p_vforkact) { - sig_thread = p->p_vforkact; + sig_thread = p->p_vforkact; kret = check_actforsig(sig_task, sig_thread, 1); - if (kret == KERN_SUCCESS) { + if (kret == KERN_SUCCESS) { *thr = sig_thread; - return(KERN_SUCCESS); - }else - return(KERN_FAILURE); + return KERN_SUCCESS; + } else { + return KERN_FAILURE; + } } +again: TAILQ_FOREACH(uth, &p->p_uthlist, uu_list) { - if(((uth->uu_flag & UT_NO_SIGMASK)== 0) && - (((uth->uu_sigmask & mask) == 0) || (uth->uu_sigwait & mask))) { - if (check_actforsig(p->task, uth->uu_context.vc_thread, 1) == KERN_SUCCESS) { - *thr = uth->uu_context.vc_thread; - return(KERN_SUCCESS); + if (((uth->uu_flag & UT_NO_SIGMASK) == 0) && + (((uth->uu_sigmask & mask) == 0) || (uth->uu_sigwait & mask))) { + thread_t th = uth->uu_context.vc_thread; + if (skip_wqthreads && (thread_get_tag(th) & THREAD_TAG_WORKQUEUE)) { + /* Workqueue threads may be parked in the kernel unable to + * deliver signals for an extended period of time, so skip them + * in favor of pthreads in a first pass. (rdar://50054475). */ + } else if (check_actforsig(p->task, th, 1) == KERN_SUCCESS) { + *thr = th; + return KERN_SUCCESS; } } } + if (skip_wqthreads) { + skip_wqthreads = false; + goto again; + } if (get_signalact(p->task, thr, 1) == KERN_SUCCESS) { - return(KERN_SUCCESS); + return KERN_SUCCESS; } - return(KERN_FAILURE); + return KERN_FAILURE; } static os_reason_t @@ -2042,7 +2157,7 @@ build_signal_reason(int signum, const char *procname) } reason_buffer_size_estimate = kcdata_estimate_required_buffer_size(2, sizeof(sender_proc->p_name) + - sizeof(sender_proc->p_pid)); + sizeof(sender_proc->p_pid)); ret = os_reason_alloc_buffer_noblock(signal_reason, reason_buffer_size_estimate); if (ret != 0) { @@ -2051,29 +2166,29 @@ build_signal_reason(int signum, const char *procname) } if (KERN_SUCCESS == kcdata_get_memory_addr(&signal_reason->osr_kcd_descriptor, KCDATA_TYPE_PID, - sizeof(sender_proc->p_pid), &data_addr)) { + sizeof(sender_proc->p_pid), &data_addr)) { kcdata_memcpy(&signal_reason->osr_kcd_descriptor, data_addr, &sender_proc->p_pid, - sizeof(sender_proc->p_pid)); + sizeof(sender_proc->p_pid)); } else { printf("build_signal_reason: exceeded space in signal reason buf, unable to log PID\n"); } proc_name_length = sizeof(sender_proc->p_name); if (KERN_SUCCESS == kcdata_get_memory_addr(&signal_reason->osr_kcd_descriptor, KCDATA_TYPE_PROCNAME, - proc_name_length, &data_addr)) { + proc_name_length, &data_addr)) { if (procname) { char truncated_procname[proc_name_length]; strncpy((char *) &truncated_procname, procname, proc_name_length); truncated_procname[proc_name_length - 1] = '\0'; kcdata_memcpy(&signal_reason->osr_kcd_descriptor, data_addr, truncated_procname, - strlen((char *) &truncated_procname)); + (uint32_t)strlen((char *) &truncated_procname)); } else if (*sender_proc->p_name) { kcdata_memcpy(&signal_reason->osr_kcd_descriptor, data_addr, &sender_proc->p_name, - sizeof(sender_proc->p_name)); + sizeof(sender_proc->p_name)); } else { kcdata_memcpy(&signal_reason->osr_kcd_descriptor, data_addr, &default_sender_procname, - strlen(default_sender_procname) + 1); + (uint32_t)strlen(default_sender_procname) + 1); } } else { printf("build_signal_reason: exceeded space in signal reason buf, unable to log procname\n"); @@ -2103,41 +2218,42 @@ psignal_internal(proc_t p, task_t task, thread_t thread, int flavor, int signum, { int prop; user_addr_t action = USER_ADDR_NULL; - proc_t sig_proc; - thread_t sig_thread; - task_t sig_task; - int mask; - struct uthread *uth; - kern_return_t kret; - uid_t r_uid; - proc_t pp; - kauth_cred_t my_cred; - char *launchd_exit_reason_desc = NULL; - boolean_t update_thread_policy = FALSE; - - if ((u_int)signum >= NSIG || signum == 0) + proc_t sig_proc; + thread_t sig_thread; + task_t sig_task; + int mask; + struct uthread *uth; + kern_return_t kret; + uid_t r_uid; + proc_t pp; + kauth_cred_t my_cred; + char *launchd_exit_reason_desc = NULL; + boolean_t update_thread_policy = FALSE; + + if ((u_int)signum >= NSIG || signum == 0) { panic("psignal: bad signal number %d", signum); + } mask = sigmask(signum); prop = sigprop[signum]; #if SIGNAL_DEBUG - if(rdebug_proc && (p != PROC_NULL) && (p == rdebug_proc)) { - ram_printf(3); - } + if (rdebug_proc && (p != PROC_NULL) && (p == rdebug_proc)) { + ram_printf(3); + } #endif /* SIGNAL_DEBUG */ /* catch unexpected initproc kills early for easier debuggging */ if (signum == SIGKILL && p == initproc) { if (signal_reason == NULL) { panic_plain("unexpected SIGKILL of %s %s (no reason provided)", - (p->p_name[0] != '\0' ? p->p_name : "initproc"), - ((p->p_csflags & CS_KILLED) ? "(CS_KILLED)" : "")); + (p->p_name[0] != '\0' ? p->p_name : "initproc"), + ((p->p_csflags & CS_KILLED) ? "(CS_KILLED)" : "")); } else { launchd_exit_reason_desc = launchd_exit_reason_get_string_desc(signal_reason); panic_plain("unexpected SIGKILL of %s %s with reason -- namespace %d code 0x%llx description %." LAUNCHD_PANIC_REASON_STRING_MAXLEN "s", - (p->p_name[0] != '\0' ? p->p_name : "initproc"), - ((p->p_csflags & CS_KILLED) ? "(CS_KILLED)" : ""), + (p->p_name[0] != '\0' ? p->p_name : "initproc"), + ((p->p_csflags & CS_KILLED) ? "(CS_KILLED)" : ""), signal_reason->osr_namespace, signal_reason->osr_code, launchd_exit_reason_desc ? launchd_exit_reason_desc : "none"); } @@ -2185,12 +2301,13 @@ psignal_internal(proc_t p, task_t task, thread_t thread, int flavor, int signum, return; } - if( (flavor & (PSIG_VFORK | PSIG_THREAD)) == 0) { + if ((flavor & (PSIG_VFORK | PSIG_THREAD)) == 0) { proc_knote(sig_proc, NOTE_SIGNAL | signum); } - if ((flavor & PSIG_LOCKED)== 0) + if ((flavor & PSIG_LOCKED) == 0) { proc_signalstart(sig_proc, 0); + } /* Don't send signals to a process that has ignored them. */ if (((flavor & PSIG_VFORK) == 0) && ((sig_proc->p_lflag & P_LTRACED) == 0) && (sig_proc->p_sigignore & mask)) { @@ -2215,8 +2332,8 @@ psignal_internal(proc_t p, task_t task, thread_t thread, int flavor, int signum, } else if (flavor & PSIG_TRY_THREAD) { uth = get_bsdthread_info(sig_thread); if (((uth->uu_flag & UT_NO_SIGMASK) == 0) && - (((uth->uu_sigmask & mask) == 0) || (uth->uu_sigwait & mask)) && - ((kret = check_actforsig(sig_proc->task, sig_thread, 1)) == KERN_SUCCESS)) { + (((uth->uu_sigmask & mask) == 0) || (uth->uu_sigwait & mask)) && + ((kret = check_actforsig(sig_proc->task, sig_thread, 1)) == KERN_SUCCESS)) { /* deliver to specified thread */ } else { /* deliver to any willing thread */ @@ -2243,9 +2360,9 @@ psignal_internal(proc_t p, task_t task, thread_t thread, int flavor, int signum, */ if ((flavor & PSIG_VFORK) == 0) { - if (sig_proc->p_lflag & P_LTRACED) + if (sig_proc->p_lflag & P_LTRACED) { action = SIG_DFL; - else { + } else { /* * If the signal is being ignored, * then we forget about it immediately. @@ -2253,27 +2370,31 @@ psignal_internal(proc_t p, task_t task, thread_t thread, int flavor, int signum, * and if it is set to SIG_IGN, * action will be SIG_DFL here.) */ - if (sig_proc->p_sigignore & mask) + if (sig_proc->p_sigignore & mask) { goto sigout_locked; + } - if (uth->uu_sigwait & mask) + if (uth->uu_sigwait & mask) { action = KERN_SIG_WAIT; - else if (uth->uu_sigmask & mask) + } else if (uth->uu_sigmask & mask) { action = KERN_SIG_HOLD; - else if (sig_proc->p_sigcatch & mask) + } else if (sig_proc->p_sigcatch & mask) { action = KERN_SIG_CATCH; - else + } else { action = SIG_DFL; + } } } /* TODO: p_nice isn't hooked up to the scheduler... */ if (sig_proc->p_nice > NZERO && action == SIG_DFL && (prop & SA_KILL) && - (sig_proc->p_lflag & P_LTRACED) == 0) - sig_proc->p_nice = NZERO; + (sig_proc->p_lflag & P_LTRACED) == 0) { + sig_proc->p_nice = NZERO; + } - if (prop & SA_CONT) + if (prop & SA_CONT) { uth->uu_siglist &= ~stopsigmask; + } if (prop & SA_STOP) { struct pgrp *pg; @@ -2285,7 +2406,7 @@ psignal_internal(proc_t p, task_t task, thread_t thread, int flavor, int signum, */ pg = proc_pgrp(sig_proc); if (prop & SA_TTYSTOP && pg->pg_jobc == 0 && - action == SIG_DFL) { + action == SIG_DFL) { pg_rele(pg); goto sigout_locked; } @@ -2300,8 +2421,9 @@ psignal_internal(proc_t p, task_t task, thread_t thread, int flavor, int signum, * except that stopped processes must be continued by SIGCONT. */ /* vfork will not go thru as action is SIG_DFL */ - if ((action == KERN_SIG_HOLD) && ((prop & SA_CONT) == 0 || sig_proc->p_stat != SSTOP)) + if ((action == KERN_SIG_HOLD) && ((prop & SA_CONT) == 0 || sig_proc->p_stat != SSTOP)) { goto sigout_locked; + } /* * SIGKILL priority twiddling moved here from above because @@ -2321,14 +2443,16 @@ psignal_internal(proc_t p, task_t task, thread_t thread, int flavor, int signum, * issig() and stop for the parent. */ if (sig_proc->p_lflag & P_LTRACED) { - if (sig_proc->p_stat != SSTOP) + if (sig_proc->p_stat != SSTOP) { goto runlocked; - else + } else { goto sigout_locked; + } } - if ((flavor & PSIG_VFORK) != 0) + if ((flavor & PSIG_VFORK) != 0) { goto runlocked; + } if (action == KERN_SIG_WAIT) { #if CONFIG_DTRACE @@ -2367,7 +2491,7 @@ psignal_internal(proc_t p, task_t task, thread_t thread, int flavor, int signum, OSBitOrAtomic(P_CONTINUED, &sig_proc->p_flag); (void) task_resume_internal(sig_task); sig_proc->p_stat = SRUN; - } else if (sig_proc->p_stat == SSTOP) { + } else if (sig_proc->p_stat == SSTOP) { goto sigout_locked; } /* @@ -2426,8 +2550,7 @@ psignal_internal(proc_t p, task_t task, thread_t thread, int flavor, int signum, pp = proc_parentholdref(sig_proc); stop(sig_proc, pp); - if (( pp != PROC_NULL) && ((pp->p_flag & P_NOCLDSTOP) == 0)) { - + if ((pp != PROC_NULL) && ((pp->p_flag & P_NOCLDSTOP) == 0)) { my_cred = kauth_cred_proc_ref(sig_proc); r_uid = kauth_cred_getruid(my_cred); kauth_cred_unref(&my_cred); @@ -2439,10 +2562,11 @@ psignal_internal(proc_t p, task_t task, thread_t thread, int flavor, int signum, * when sent to the parent must set the * child's signal number into si_status. */ - if (signum != SIGSTOP) + if (signum != SIGSTOP) { pp->si_status = WEXITSTATUS(sig_proc->p_xstat); - else + } else { pp->si_status = W_EXITCODE(signum, signum); + } pp->si_code = CLD_STOPPED; pp->si_uid = r_uid; proc_unlock(sig_proc); @@ -2462,11 +2586,11 @@ psignal_internal(proc_t p, task_t task, thread_t thread, int flavor, int signum, DTRACE_PROC3(signal__send, thread_t, sig_thread, proc_t, p, int, signum); switch (signum) { - /* - * Signals ignored by default have been dealt - * with already, since their bits are on in - * p_sigignore. - */ + /* + * Signals ignored by default have been dealt + * with already, since their bits are on in + * p_sigignore. + */ case SIGKILL: /* @@ -2491,7 +2615,7 @@ psignal_internal(proc_t p, task_t task, thread_t thread, int flavor, int signum, if (uth->uu_exit_reason == OS_REASON_NULL) { if (signal_reason == OS_REASON_NULL) { KERNEL_DEBUG_CONSTANT(BSDDBG_CODE(DBG_BSD_PROC, BSD_PROC_EXITREASON_CREATE) | DBG_FUNC_NONE, - sig_proc->p_pid, OS_REASON_SIGNAL, signum, 0, 0); + sig_proc->p_pid, OS_REASON_SIGNAL, signum, 0, 0); signal_reason = build_signal_reason(signum, NULL); } @@ -2524,7 +2648,7 @@ psignal_internal(proc_t p, task_t task, thread_t thread, int flavor, int signum, * cause their handlers to fire. If it's only * the SIGCONT, then don't wake up. */ - if (((flavor & (PSIG_VFORK|PSIG_THREAD)) == 0) && (((uth->uu_siglist & ~uth->uu_sigmask) & ~sig_proc->p_sigignore) & ~mask)) { + if (((flavor & (PSIG_VFORK | PSIG_THREAD)) == 0) && (((uth->uu_siglist & ~uth->uu_sigmask) & ~sig_proc->p_sigignore) & ~mask)) { uth->uu_siglist &= ~mask; sig_proc->p_stat = SRUN; goto runlocked; @@ -2540,7 +2664,7 @@ psignal_internal(proc_t p, task_t task, thread_t thread, int flavor, int signum, * the process, and for which there is no handler, * needs to act like SIGKILL */ - if (((flavor & (PSIG_VFORK|PSIG_THREAD)) == 0) && (action == SIG_DFL) && (prop & SA_KILL)) { + if (((flavor & (PSIG_VFORK | PSIG_THREAD)) == 0) && (action == SIG_DFL) && (prop & SA_KILL)) { sig_proc->p_stat = SRUN; kret = thread_abort(sig_thread); update_thread_policy = (kret == KERN_SUCCESS); @@ -2548,7 +2672,7 @@ psignal_internal(proc_t p, task_t task, thread_t thread, int flavor, int signum, if (uth->uu_exit_reason == OS_REASON_NULL) { if (signal_reason == OS_REASON_NULL) { KERNEL_DEBUG_CONSTANT(BSDDBG_CODE(DBG_BSD_PROC, BSD_PROC_EXITREASON_CREATE) | DBG_FUNC_NONE, - sig_proc->p_pid, OS_REASON_SIGNAL, signum, 0, 0); + sig_proc->p_pid, OS_REASON_SIGNAL, signum, 0, 0); signal_reason = build_signal_reason(signum, NULL); } @@ -2578,20 +2702,22 @@ runlocked: * while we were stopped), check for a signal from the debugger. */ if (sig_proc->p_stat == SSTOP) { - if ((sig_proc->p_lflag & P_LTRACED) != 0 && sig_proc->p_xstat != 0) + if ((sig_proc->p_lflag & P_LTRACED) != 0 && sig_proc->p_xstat != 0) { uth->uu_siglist |= sigmask(sig_proc->p_xstat); + } if ((flavor & PSIG_VFORK) != 0) { sig_proc->p_stat = SRUN; } } else { /* - * setrunnable(p) in BSD and - * Wake up the thread if it is interruptible. - */ + * setrunnable(p) in BSD and + * Wake up the thread if it is interruptible. + */ sig_proc->p_stat = SRUN; - if ((flavor & PSIG_VFORK) == 0) + if ((flavor & PSIG_VFORK) == 0) { thread_abort_safely(sig_thread); + } } sigout_locked: @@ -2608,7 +2734,7 @@ sigout_locked: sigout_unlocked: os_reason_free(signal_reason); - if ((flavor & PSIG_LOCKED)== 0) { + if ((flavor & PSIG_LOCKED) == 0) { proc_signalend(sig_proc, 0); } } @@ -2625,6 +2751,12 @@ psignal_with_reason(proc_t p, int signum, struct os_reason *signal_reason) psignal_internal(p, NULL, NULL, 0, signum, signal_reason); } +void +psignal_sigkill_with_reason(struct proc *p, struct os_reason *signal_reason) +{ + psignal_internal(p, NULL, NULL, 0, SIGKILL, signal_reason); +} + void psignal_locked(proc_t p, int signum) { @@ -2695,7 +2827,7 @@ issignal_locked(proc_t p) cur_act = current_thread(); #if SIGNAL_DEBUG - if(rdebug_proc && (p == rdebug_proc)) { + if (rdebug_proc && (p == rdebug_proc)) { ram_printf(3); } #endif /* SIGNAL_DEBUG */ @@ -2713,14 +2845,15 @@ issignal_locked(proc_t p) for (;;) { sigbits = ut->uu_siglist & ~ut->uu_sigmask; - if (p->p_lflag & P_LPPWAIT) + if (p->p_lflag & P_LPPWAIT) { sigbits &= ~stopsigmask; - if (sigbits == 0) { /* no signal to send */ + } + if (sigbits == 0) { /* no signal to send */ retval = 0; goto out; } - signum = ffs((long)sigbits); + signum = ffs((unsigned int)sigbits); mask = sigmask(signum); prop = sigprop[signum]; @@ -2733,12 +2866,12 @@ issignal_locked(proc_t p) continue; } - if (p->p_lflag & P_LTRACED && (p->p_lflag & P_LPPWAIT) == 0) { + if (p->p_lflag & P_LTRACED && (p->p_lflag & P_LPPWAIT) == 0) { /* * If traced, deliver the signal to the debugger, and wait to be * released. */ - task_t task; + task_t task; p->p_xstat = signum; if (p->p_lflag & P_LSIGEXC) { @@ -2774,9 +2907,9 @@ issignal_locked(proc_t p) } /* - * XXX Have to really stop for debuggers; - * XXX stop() doesn't do the right thing. - */ + * XXX Have to really stop for debuggers; + * XXX stop() doesn't do the right thing. + */ task = p->task; task_suspend_internal(task); @@ -2831,8 +2964,9 @@ issignal_locked(proc_t p) * otherwise we just look for signals again. */ signum = p->p_xstat; - if (signum == 0) + if (signum == 0) { continue; + } /* * Put the new signal into p_siglist. If the @@ -2840,8 +2974,9 @@ issignal_locked(proc_t p) */ mask = sigmask(signum); ut->uu_siglist |= mask; - if (ut->uu_sigmask & mask) + if (ut->uu_sigmask & mask) { continue; + } } /* @@ -2851,7 +2986,6 @@ issignal_locked(proc_t p) */ switch ((long)p->p_sigacts->ps_sigact[signum]) { - case (long)SIG_DFL: /* * If there is a pending stop signal to process @@ -2866,8 +3000,8 @@ issignal_locked(proc_t p) proc_unlock(p); pg = proc_pgrp(p); if (p->p_lflag & P_LTRACED || - (pg->pg_jobc == 0 && - prop & SA_TTYSTOP)) { + (pg->pg_jobc == 0 && + prop & SA_TTYSTOP)) { proc_lock(p); pg_rele(pg); break; /* ignore signal */ @@ -2896,8 +3030,9 @@ issignal_locked(proc_t p) psignal(pp, SIGCHLD); } - if (pp != PROC_NULL) + if (pp != PROC_NULL) { proc_parentdropref(pp, 0); + } } proc_lock(p); break; @@ -2918,8 +3053,9 @@ issignal_locked(proc_t p) * than SIGCONT, unless process is traced. */ if ((prop & SA_CONT) == 0 && - (p->p_lflag & P_LTRACED) == 0) + (p->p_lflag & P_LTRACED) == 0) { printf("issignal\n"); + } break; /* ignore signal */ default: @@ -2929,7 +3065,6 @@ issignal_locked(proc_t p) /* If we dropped through, the signal was ignored - remove it from pending list. */ ut->uu_siglist &= ~mask; - } /* for(;;) */ /* NOTREACHED */ @@ -2951,31 +3086,34 @@ CURSIG(proc_t p) thread_t cur_act; struct uthread * ut; int retnum = 0; - + cur_act = current_thread(); ut = get_bsdthread_info(cur_act); - if (ut->uu_siglist == 0) - return (0); + if (ut->uu_siglist == 0) { + return 0; + } - if (((ut->uu_siglist & ~ut->uu_sigmask) == 0) && ((p->p_lflag & P_LTRACED) == 0)) - return (0); + if (((ut->uu_siglist & ~ut->uu_sigmask) == 0) && ((p->p_lflag & P_LTRACED) == 0)) { + return 0; + } sigbits = ut->uu_siglist & ~ut->uu_sigmask; - for(;;) { - if (p->p_lflag & P_LPPWAIT) + for (;;) { + if (p->p_lflag & P_LPPWAIT) { sigbits &= ~stopsigmask; - if (sigbits == 0) { /* no signal to send */ - return (retnum); + } + if (sigbits == 0) { /* no signal to send */ + return retnum; } - signum = ffs((long)sigbits); + signum = ffs((unsigned int)sigbits); mask = sigmask(signum); prop = sigprop[signum]; - sigbits &= ~mask; /* take the signal out */ + sigbits &= ~mask; /* take the signal out */ /* * We should see pending but ignored signals @@ -2986,7 +3124,7 @@ CURSIG(proc_t p) } if (p->p_lflag & P_LTRACED && (p->p_lflag & P_LPPWAIT) == 0) { - return(signum); + return signum; } /* @@ -2996,7 +3134,6 @@ CURSIG(proc_t p) */ switch ((long)p->p_sigacts->ps_sigact[signum]) { - case (long)SIG_DFL: /* * If there is a pending stop signal to process @@ -3011,10 +3148,10 @@ CURSIG(proc_t p) pg = proc_pgrp(p); if (p->p_lflag & P_LTRACED || - (pg->pg_jobc == 0 && - prop & SA_TTYSTOP)) { + (pg->pg_jobc == 0 && + prop & SA_TTYSTOP)) { pg_rele(pg); - break; /* == ignore */ + break; /* == ignore */ } pg_rele(pg); retnum = signum; @@ -3024,11 +3161,11 @@ CURSIG(proc_t p) * Except for SIGCONT, shouldn't get here. * Default action is to ignore; drop it. */ - break; /* == ignore */ + break; /* == ignore */ } else { - return (signum); + return signum; } - /*NOTREACHED*/ + /*NOTREACHED*/ case (long)SIG_IGN: /* @@ -3037,16 +3174,17 @@ CURSIG(proc_t p) * than SIGCONT, unless process is traced. */ if ((prop & SA_CONT) == 0 && - (p->p_lflag & P_LTRACED) == 0) + (p->p_lflag & P_LTRACED) == 0) { printf("issignal\n"); - break; /* == ignore */ + } + break; /* == ignore */ default: /* * This signal has an action, let * postsig() process it. */ - return (signum); + return signum; } } /* NOTREACHED */ @@ -3082,15 +3220,18 @@ postsig_locked(int signum) uint32_t code; int mask, returnmask; struct uthread * ut; + os_reason_t ut_exit_reason = OS_REASON_NULL; #if DIAGNOSTIC - if (signum == 0) + if (signum == 0) { panic("postsig"); + } /* * This must be called on master cpu */ - if (cpu_number() != master_cpu) + if (cpu_number() != master_cpu) { panic("psig not on master"); + } #endif /* @@ -3111,21 +3252,30 @@ postsig_locked(int signum) * Default catcher, where the default is to kill * the process. (Other cases were ignored above.) */ - sig_lock_to_exit(p); + + /* + * exit_with_reason() below will consume a reference to the thread's exit reason, so we take another + * reference so the thread still has one even after we call exit_with_reason(). The thread's reference will + * ultimately be destroyed in uthread_cleanup(). + */ + ut_exit_reason = ut->uu_exit_reason; + os_reason_ref(ut_exit_reason); + p->p_acflag |= AXSIG; if (sigprop[signum] & SA_CORE) { p->p_sigacts->ps_sig = signum; proc_signalend(p, 1); proc_unlock(p); #if CONFIG_COREDUMP - if (coredump(p, 0, 0) == 0) + if (coredump(p, 0, 0) == 0) { signum |= WCOREFLAG; + } #endif - } else { + } else { proc_signalend(p, 1); proc_unlock(p); } - + #if CONFIG_DTRACE bzero((caddr_t)&(ut->t_dtrace_siginfo), sizeof(ut->t_dtrace_siginfo)); @@ -3142,21 +3292,16 @@ postsig_locked(int signum) default: break; } - + DTRACE_PROC3(signal__handle, int, signum, siginfo_t *, &(ut->t_dtrace_siginfo), - void (*)(void), SIG_DFL); + void (*)(void), SIG_DFL); #endif KERNEL_DEBUG_CONSTANT(BSDDBG_CODE(DBG_BSD_PROC, BSD_PROC_FRCEXIT) | DBG_FUNC_NONE, - p->p_pid, W_EXITCODE(0, signum), 3, 0, 0); + p->p_pid, W_EXITCODE(0, signum), 3, 0, 0); - /* - * exit_with_reason() will consume a reference to the thread's exit reason, so we take another - * reference for the thread. This reference will be destroyed in uthread_cleanup(). - */ - os_reason_ref(ut->uu_exit_reason); - exit_with_reason(p, W_EXITCODE(0, signum), (int *)NULL, TRUE, TRUE, 0, ut->uu_exit_reason); + exit_with_reason(p, W_EXITCODE(0, signum), (int *)NULL, TRUE, TRUE, 0, ut_exit_reason); proc_lock(p); return; @@ -3165,9 +3310,10 @@ postsig_locked(int signum) * If we get here, the signal must be caught. */ #if DIAGNOSTIC - if (catcher == SIG_IGN || (ut->uu_sigmask & mask)) + if (catcher == SIG_IGN || (ut->uu_sigmask & mask)) { log(LOG_WARNING, - "postsig: processing masked or ignored signal\n"); + "postsig: processing masked or ignored signal\n"); + } #endif /* @@ -3183,14 +3329,18 @@ postsig_locked(int signum) returnmask = ut->uu_oldmask; ut->uu_flag &= ~UT_SAS_OLDMASK; ut->uu_oldmask = 0; - } else + } else { returnmask = ut->uu_sigmask; + } ut->uu_sigmask |= ps->ps_catchmask[signum]; - if ((ps->ps_signodefer & mask) == 0) + if ((ps->ps_signodefer & mask) == 0) { ut->uu_sigmask |= mask; + } + sigset_t siginfo = ps->ps_siginfo; if ((signum != SIGILL) && (signum != SIGTRAP) && (ps->ps_sigreset & mask)) { - if ((signum != SIGCONT) && (sigprop[signum] & SA_IGNORE)) + if ((signum != SIGCONT) && (sigprop[signum] & SA_IGNORE)) { p->p_sigignore |= mask; + } ps->ps_sigact[signum] = SIG_DFL; ps->ps_siginfo &= ~mask; ps->ps_signodefer &= ~mask; @@ -3203,7 +3353,7 @@ postsig_locked(int signum) ps->ps_code = 0; } OSIncrementAtomicLong(&p->p_stats->p_ru.ru_nsignals); - sendsig(p, catcher, signum, returnmask, code); + sendsig(p, catcher, signum, returnmask, code, siginfo); } proc_signalend(p, 1); } @@ -3217,34 +3367,36 @@ postsig_locked(int signum) */ static int -filt_sigattach(struct knote *kn, __unused struct kevent_internal_s *kev) +filt_sigattach(struct knote *kn, __unused struct kevent_qos_s *kev) { proc_t p = current_proc(); /* can attach only to oneself */ proc_klist_lock(); - kn->kn_ptr.p_proc = p; + kn->kn_proc = p; + kn->kn_flags |= EV_CLEAR; /* automatically set */ + kn->kn_sdata = 0; /* incoming data is ignored */ KNOTE_ATTACH(&p->p_klist, kn); proc_klist_unlock(); /* edge-triggered events can't have fired before we attached */ - return (0); + return 0; } /* * remove the knote from the process list, if it hasn't already - * been removed by exit processing. + * been removed by exit processing. */ - + static void filt_sigdetach(struct knote *kn) { - proc_t p = kn->kn_ptr.p_proc; + proc_t p = kn->kn_proc; proc_klist_lock(); - kn->kn_ptr.p_proc = NULL; + kn->kn_proc = NULL; KNOTE_DETACH(&p->p_klist, kn); proc_klist_unlock(); } @@ -3261,23 +3413,21 @@ filt_sigdetach(struct knote *kn) static int filt_signal(struct knote *kn, long hint) { - if (hint & NOTE_SIGNAL) { hint &= ~NOTE_SIGNAL; - if (kn->kn_id == (unsigned int)hint) - kn->kn_data++; + if (kn->kn_id == (unsigned int)hint) { + kn->kn_hook32++; + } } else if (hint & NOTE_EXIT) { panic("filt_signal: detected NOTE_EXIT event"); } - return (kn->kn_data != 0); + return kn->kn_hook32 != 0; } static int -filt_signaltouch( - struct knote *kn, - struct kevent_internal_s *kev) +filt_signaltouch(struct knote *kn, struct kevent_qos_s *kev) { #pragma unused(kev) @@ -3285,13 +3435,10 @@ filt_signaltouch( proc_klist_lock(); - if ((kn->kn_status & KN_UDATA_SPECIFIC) == 0) - kn->kn_udata = kev->udata; - /* - * No data to save - - * just capture if it is already fired + /* + * No data to save - just capture if it is already fired */ - res = (kn->kn_data > 0); + res = (kn->kn_hook32 > 0); proc_klist_unlock(); @@ -3299,29 +3446,22 @@ filt_signaltouch( } static int -filt_signalprocess( - struct knote *kn, - __unused struct filt_process_s *data, - struct kevent_internal_s *kev) +filt_signalprocess(struct knote *kn, struct kevent_qos_s *kev) { - proc_klist_lock(); - - if (kn->kn_data == 0) { - proc_klist_unlock(); - return 0; - } + int res = 0; /* * Snapshot the event data. - * All signal events are EV_CLEAR, so - * add that and clear out the data field. */ - *kev = kn->kn_kevent; - kev->flags |= EV_CLEAR; - kn->kn_data = 0; + proc_klist_lock(); + if (kn->kn_hook32) { + knote_fill_kevent(kn, kev, kn->kn_hook32); + kn->kn_hook32 = 0; + res = 1; + } proc_klist_unlock(); - return 1; + return res; } void @@ -3329,56 +3469,52 @@ bsd_ast(thread_t thread) { proc_t p = current_proc(); struct uthread *ut = get_bsdthread_info(thread); - int signum; - user_addr_t pc; + int signum; static int bsd_init_done = 0; - if (p == NULL) + if (p == NULL) { return; + } /* don't run bsd ast on exec copy or exec'ed tasks */ if (task_did_exec(current_task()) || task_is_exec_copy(current_task())) { return; } - if ((p->p_flag & P_OWEUPC) && (p->p_flag & P_PROFIL)) { - pc = get_useraddr(); - addupc_task(p, pc, 1); - OSBitAndAtomic(~((uint32_t)P_OWEUPC), &p->p_flag); - } - if (timerisset(&p->p_vtimer_user.it_value)) { - uint32_t microsecs; + uint32_t microsecs; task_vtimer_update(p->task, TASK_VTIMER_USER, µsecs); if (!itimerdecr(p, &p->p_vtimer_user, microsecs)) { - if (timerisset(&p->p_vtimer_user.it_value)) + if (timerisset(&p->p_vtimer_user.it_value)) { task_vtimer_set(p->task, TASK_VTIMER_USER); - else + } else { task_vtimer_clear(p->task, TASK_VTIMER_USER); + } psignal_try_thread(p, thread, SIGVTALRM); } } if (timerisset(&p->p_vtimer_prof.it_value)) { - uint32_t microsecs; + uint32_t microsecs; task_vtimer_update(p->task, TASK_VTIMER_PROF, µsecs); if (!itimerdecr(p, &p->p_vtimer_prof, microsecs)) { - if (timerisset(&p->p_vtimer_prof.it_value)) + if (timerisset(&p->p_vtimer_prof.it_value)) { task_vtimer_set(p->task, TASK_VTIMER_PROF); - else + } else { task_vtimer_clear(p->task, TASK_VTIMER_PROF); + } psignal_try_thread(p, thread, SIGPROF); } } if (timerisset(&p->p_rlim_cpu)) { - struct timeval tv; + struct timeval tv; task_vtimer_update(p->task, TASK_VTIMER_RLIM, (uint32_t *) &tv.tv_usec); @@ -3388,7 +3524,6 @@ bsd_ast(thread_t thread) timersub(&p->p_rlim_cpu, &tv, &p->p_rlim_cpu); proc_spinunlock(p); } else { - timerclear(&p->p_rlim_cpu); proc_spinunlock(p); @@ -3400,9 +3535,9 @@ bsd_ast(thread_t thread) #if CONFIG_DTRACE if (ut->t_dtrace_sig) { - uint8_t dt_action_sig = ut->t_dtrace_sig; - ut->t_dtrace_sig = 0; - psignal(p, dt_action_sig); + uint8_t dt_action_sig = ut->t_dtrace_sig; + ut->t_dtrace_sig = 0; + psignal(p, dt_action_sig); } if (ut->t_dtrace_stop) { @@ -3414,7 +3549,7 @@ bsd_ast(thread_t thread) } if (ut->t_dtrace_resumepid) { - proc_t resumeproc = proc_find(ut->t_dtrace_resumepid); + proc_t resumeproc = proc_find((int)ut->t_dtrace_resumepid); ut->t_dtrace_resumepid = 0; if (resumeproc != PROC_NULL) { proc_lock(resumeproc); @@ -3423,20 +3558,20 @@ bsd_ast(thread_t thread) resumeproc->p_dtrace_stop = 0; proc_unlock(resumeproc); task_resume_internal(resumeproc->task); - } - else { + } else { proc_unlock(resumeproc); } proc_rele(resumeproc); } } - + #endif /* CONFIG_DTRACE */ proc_lock(p); if (CHECK_SIGNALS(p, current_thread(), ut)) { - while ( (signum = issignal_locked(p)) ) + while ((signum = issignal_locked(p))) { postsig_locked(signum); + } } proc_unlock(p); @@ -3466,7 +3601,7 @@ pt_setrunnable(proc_t p) proc_unlock(p); if (p->sigwait) { wakeup((caddr_t)&(p->sigwait)); - if ((p->p_lflag & P_LSIGEXC) == 0) { // 5878479 + if ((p->p_lflag & P_LSIGEXC) == 0) { // 5878479 task_release(task); } } @@ -3475,15 +3610,15 @@ pt_setrunnable(proc_t p) kern_return_t do_bsdexception( - int exc, - int code, - int sub) + int exc, + int code, + int sub) { mach_exception_data_type_t codes[EXCEPTION_CODE_MAX]; - codes[0] = code; + codes[0] = code; codes[1] = sub; - return(bsd_exception(exc, codes, 2)); + return bsd_exception(exc, codes, 2); } int @@ -3495,18 +3630,18 @@ proc_pendingsignals(proc_t p, sigset_t mask) proc_lock(p); /* If the process is in proc exit return no signal info */ - if (p->p_lflag & P_LPEXIT) { + if (p->p_lflag & P_LPEXIT) { goto out; } if ((p->p_lflag & P_LINVFORK) && p->p_vforkact) { - th = p->p_vforkact; + th = p->p_vforkact; uth = (struct uthread *)get_bsdthread_info(th); if (uth) { bits = (((uth->uu_siglist & ~uth->uu_sigmask) & ~p->p_sigignore) & mask); } goto out; - } + } bits = 0; TAILQ_FOREACH(uth, &p->p_uthlist, uu_list) { @@ -3514,14 +3649,14 @@ proc_pendingsignals(proc_t p, sigset_t mask) } out: proc_unlock(p); - return(bits); + return bits; } int thread_issignal(proc_t p, thread_t th, sigset_t mask) { struct uthread * uth; - sigset_t bits=0; + sigset_t bits = 0; proc_lock(p); uth = (struct uthread *)get_bsdthread_info(th); @@ -3529,7 +3664,7 @@ thread_issignal(proc_t p, thread_t th, sigset_t mask) bits = (((uth->uu_siglist & ~uth->uu_sigmask) & ~p->p_sigignore) & mask); } proc_unlock(p); - return(bits); + return bits; } /* @@ -3538,62 +3673,70 @@ thread_issignal(proc_t p, thread_t th, sigset_t mask) int hassigprop(int sig, int prop) { - return (sigprop[sig] & prop); + return sigprop[sig] & prop; } void pgsigio(pid_t pgid, int sig) -{ +{ proc_t p = PROC_NULL; - if (pgid < 0) + if (pgid < 0) { gsignal(-(pgid), sig); - - else if (pgid > 0 && (p = proc_find(pgid)) != 0) + } else if (pgid > 0 && (p = proc_find(pgid)) != 0) { psignal(p, sig); - if (p != PROC_NULL) + } + if (p != PROC_NULL) { proc_rele(p); + } } void proc_signalstart(proc_t p, int locked) { - if (!locked) + if (!locked) { proc_lock(p); - - if(p->p_signalholder == current_thread()) - panic("proc_signalstart: thread attempting to signal a process for which it holds the signal lock"); - + } + + if (p->p_signalholder == current_thread()) { + panic("proc_signalstart: thread attempting to signal a process for which it holds the signal lock"); + } + p->p_sigwaitcnt++; - while ((p->p_lflag & P_LINSIGNAL) == P_LINSIGNAL) + while ((p->p_lflag & P_LINSIGNAL) == P_LINSIGNAL) { msleep(&p->p_sigmask, &p->p_mlock, 0, "proc_signstart", NULL); + } p->p_sigwaitcnt--; p->p_lflag |= P_LINSIGNAL; p->p_signalholder = current_thread(); - if (!locked) + if (!locked) { proc_unlock(p); + } } void proc_signalend(proc_t p, int locked) { - if (!locked) + if (!locked) { proc_lock(p); + } p->p_lflag &= ~P_LINSIGNAL; - if (p->p_sigwaitcnt > 0) + if (p->p_sigwaitcnt > 0) { wakeup(&p->p_sigmask); + } p->p_signalholder = NULL; - if (!locked) + if (!locked) { proc_unlock(p); + } } void sig_lock_to_exit(proc_t p) { - thread_t self = current_thread(); + thread_t self = current_thread(); p->exit_thread = self; proc_unlock(p); @@ -3607,11 +3750,11 @@ sig_lock_to_exit(proc_t p) int sig_try_locked(proc_t p) { - thread_t self = current_thread(); + thread_t self = current_thread(); while (p->sigwait || p->exit_thread) { if (p->exit_thread) { - return(0); + return 0; } msleep((caddr_t)&p->sigwait_thread, &p->p_mlock, PCATCH | PDROP, 0, 0); if (thread_should_abort(self)) {