]> git.saurik.com Git - apple/xnu.git/blob - bsd/kern/kern_exit.c
xnu-792.tar.gz
[apple/xnu.git] / bsd / kern / kern_exit.c
1 /*
2 * Copyright (c) 2000-2004 Apple Computer, Inc. All rights reserved.
3 *
4 * @APPLE_LICENSE_HEADER_START@
5 *
6 * The contents of this file constitute Original Code as defined in and
7 * are subject to the Apple Public Source License Version 1.1 (the
8 * "License"). You may not use this file except in compliance with the
9 * License. Please obtain a copy of the License at
10 * http://www.apple.com/publicsource and read it before using this file.
11 *
12 * This Original Code and all software distributed under the License are
13 * distributed on an "AS IS" basis, WITHOUT WARRANTY OF ANY KIND, EITHER
14 * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
15 * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
16 * FITNESS FOR A PARTICULAR PURPOSE OR NON-INFRINGEMENT. Please see the
17 * License for the specific language governing rights and limitations
18 * under the License.
19 *
20 * @APPLE_LICENSE_HEADER_END@
21 */
22 /* Copyright (c) 1995, 1997 Apple Computer, Inc. All Rights Reserved */
23 /*
24 * Copyright (c) 1982, 1986, 1989, 1991, 1993
25 * The Regents of the University of California. All rights reserved.
26 * (c) UNIX System Laboratories, Inc.
27 * All or some portions of this file are derived from material licensed
28 * to the University of California by American Telephone and Telegraph
29 * Co. or Unix System Laboratories, Inc. and are reproduced herein with
30 * the permission of UNIX System Laboratories, Inc.
31 *
32 * Redistribution and use in source and binary forms, with or without
33 * modification, are permitted provided that the following conditions
34 * are met:
35 * 1. Redistributions of source code must retain the above copyright
36 * notice, this list of conditions and the following disclaimer.
37 * 2. Redistributions in binary form must reproduce the above copyright
38 * notice, this list of conditions and the following disclaimer in the
39 * documentation and/or other materials provided with the distribution.
40 * 3. All advertising materials mentioning features or use of this software
41 * must display the following acknowledgement:
42 * This product includes software developed by the University of
43 * California, Berkeley and its contributors.
44 * 4. Neither the name of the University nor the names of its contributors
45 * may be used to endorse or promote products derived from this software
46 * without specific prior written permission.
47 *
48 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
49 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
50 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
51 * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
52 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
53 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
54 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
55 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
56 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
57 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
58 * SUCH DAMAGE.
59 *
60 * @(#)kern_exit.c 8.7 (Berkeley) 2/12/94
61 */
62
63 #include <machine/reg.h>
64 #include <machine/psl.h>
65
66 #include "compat_43.h"
67
68 #include <sys/param.h>
69 #include <sys/systm.h>
70 #include <sys/ioctl.h>
71 #include <sys/proc_internal.h>
72 #include <sys/kauth.h>
73 #include <sys/tty.h>
74 #include <sys/time.h>
75 #include <sys/resource.h>
76 #include <sys/kernel.h>
77 #include <sys/wait.h>
78 #include <sys/file_internal.h>
79 #include <sys/vnode_internal.h>
80 #include <sys/syslog.h>
81 #include <sys/malloc.h>
82 #include <sys/resourcevar.h>
83 #include <sys/ptrace.h>
84 #include <sys/user.h>
85 #include <sys/aio_kern.h>
86 #include <sys/sysproto.h>
87 #include <sys/signalvar.h>
88 #include <sys/filedesc.h> /* fdfree */
89 #include <sys/shm_internal.h> /* shmexit */
90 #include <sys/acct.h> /* acct_process */
91 #include <machine/spl.h>
92
93 #include <bsm/audit_kernel.h>
94 #include <bsm/audit_kevents.h>
95
96 #include <mach/mach_types.h>
97
98 #include <kern/kern_types.h>
99 #include <kern/kalloc.h>
100 #include <kern/task.h>
101 #include <kern/thread.h>
102 #include <kern/sched_prim.h>
103 #include <kern/assert.h>
104 #if KTRACE
105 #include <sys/ktrace.h>
106 #endif
107
108 #include <mach/mach_types.h>
109 #include <mach/task.h>
110 #include <mach/thread_act.h>
111 #include <mach/mach_traps.h> /* init_process */
112
113 extern char init_task_failure_data[];
114 int exit1(struct proc *, int, int *);
115 void proc_prepareexit(struct proc *p);
116 void vfork_exit(struct proc *p, int rv);
117 void vproc_exit(struct proc *p);
118 __private_extern__ void munge_rusage(struct rusage *a_rusage_p, struct user_rusage *a_user_rusage_p);
119
120 /*
121 * Things which should have prototypes in headers, but don't
122 */
123 void unix_syscall_return(int);
124 void *get_bsduthreadarg(thread_t);
125 void proc_exit(struct proc *p);
126 int wait1continue(int result);
127 int waitidcontinue(int result);
128 int *get_bsduthreadrval(thread_t);
129 kern_return_t sys_perf_notify(struct task *task, exception_data_t code,
130 mach_msg_type_number_t codeCnt);
131
132 /*
133 * NOTE: Source and target may *NOT* overlap!
134 * XXX Should share code with bsd/dev/ppc/unix_signal.c
135 */
136 static void
137 siginfo_64to32(user_siginfo_t *in, siginfo_t *out)
138 {
139 out->si_signo = in->si_signo;
140 out->si_errno = in->si_errno;
141 out->si_code = in->si_code;
142 out->si_pid = in->si_pid;
143 out->si_uid = in->si_uid;
144 out->si_status = in->si_status;
145 out->si_addr = CAST_DOWN(void *,in->si_addr);
146 /* following cast works for sival_int because of padding */
147 out->si_value.sival_ptr = CAST_DOWN(void *,in->si_value.sival_ptr);
148 out->si_band = in->si_band; /* range reduction */
149 out->pad[0] = in->pad[0]; /* mcontext.ss.r1 */
150 }
151
152 /*
153 * exit --
154 * Death of process.
155 */
156 void
157 exit(struct proc *p, struct exit_args *uap, int *retval)
158 {
159 exit1(p, W_EXITCODE(uap->rval, 0), retval);
160
161 /* drop funnel before we return */
162 thread_funnel_set(kernel_flock, FALSE);
163 thread_exception_return();
164 /* NOTREACHED */
165 while (TRUE)
166 thread_block(THREAD_CONTINUE_NULL);
167 /* NOTREACHED */
168 }
169
170 /*
171 * Exit: deallocate address space and other resources, change proc state
172 * to zombie, and unlink proc from allproc and parent's lists. Save exit
173 * status and rusage for wait(). Check for child processes and orphan them.
174 */
175 int
176 exit1(struct proc *p, int rv, int *retval)
177 {
178 thread_t self = current_thread();
179 struct task *task = p->task;
180 register int s;
181 struct uthread *ut;
182
183 /*
184 * If a thread in this task has already
185 * called exit(), then halt any others
186 * right here.
187 */
188
189 ut = get_bsdthread_info(self);
190 if (ut->uu_flag & UT_VFORK) {
191 vfork_exit(p, rv);
192 vfork_return(self, p->p_pptr, p , retval);
193 unix_syscall_return(0);
194 /* NOT REACHED */
195 }
196 AUDIT_SYSCALL_EXIT(0, p, ut); /* Exit is always successfull */
197 signal_lock(p);
198 while (p->exit_thread != self) {
199 if (sig_try_locked(p) <= 0) {
200 if (get_threadtask(self) != task) {
201 signal_unlock(p);
202 return(0);
203 }
204 signal_unlock(p);
205 thread_terminate(self);
206 thread_funnel_set(kernel_flock, FALSE);
207 thread_exception_return();
208 /* NOTREACHED */
209 }
210 sig_lock_to_exit(p);
211 }
212 signal_unlock(p);
213 if (p->p_pid == 1) {
214 printf("pid 1 exited (signal %d, exit %d)",
215 WTERMSIG(rv), WEXITSTATUS(rv));
216 panic("init died\nState at Last Exception:\n\n%s",
217 init_task_failure_data);
218 }
219
220 s = splsched();
221 p->p_flag |= P_WEXIT;
222 splx(s);
223 proc_prepareexit(p);
224 p->p_xstat = rv;
225
226 /* task terminate will call proc_terminate and that cleans it up */
227 task_terminate_internal(task);
228
229 return(0);
230 }
231
232 void
233 proc_prepareexit(struct proc *p)
234 {
235 struct uthread *ut;
236 exception_data_t code[EXCEPTION_CODE_MAX];
237 thread_t self = current_thread();
238
239 code[0] = (exception_data_t)0xFF000001; /* Set terminate code */
240 code[1] = (exception_data_t)p->p_pid; /* Pass out the pid */
241 /* Notify the perf server */
242 (void)sys_perf_notify(p->task, (exception_data_t)&code, 2);
243
244 /*
245 * Remove proc from allproc queue and from pidhash chain.
246 * Need to do this before we do anything that can block.
247 * Not doing causes things like mount() find this on allproc
248 * in partially cleaned state.
249 */
250 LIST_REMOVE(p, p_list);
251 LIST_INSERT_HEAD(&zombproc, p, p_list); /* Place onto zombproc. */
252 LIST_REMOVE(p, p_hash);
253
254 #ifdef PGINPROF
255 vmsizmon();
256 #endif
257 /*
258 * If parent is waiting for us to exit or exec,
259 * P_PPWAIT is set; we will wakeup the parent below.
260 */
261 p->p_flag &= ~(P_TRACED | P_PPWAIT);
262 p->p_sigignore = ~(sigcantmask);
263 p->p_siglist = 0;
264 ut = get_bsdthread_info(self);
265 ut->uu_siglist = 0;
266 untimeout(realitexpire, (caddr_t)p->p_pid);
267 }
268
269 void
270 proc_exit(struct proc *p)
271 {
272 register struct proc *q, *nq, *pp;
273 struct task *task = p->task;
274 register int s;
275 boolean_t funnel_state;
276
277 /* This can happen if thread_terminate of the single thread
278 * process
279 */
280
281 funnel_state = thread_funnel_set(kernel_flock, TRUE);
282 if( !(p->p_flag & P_WEXIT)) {
283 s = splsched();
284 p->p_flag |= P_WEXIT;
285 splx(s);
286 proc_prepareexit(p);
287 }
288
289 p->p_lflag |= P_LPEXIT;
290 /* XXX Zombie allocation may fail, in which case stats get lost */
291 MALLOC_ZONE(p->p_ru, struct rusage *,
292 sizeof (*p->p_ru), M_ZOMBIE, M_WAITOK);
293
294 /*
295 * need to cancel async IO requests that can be cancelled and wait for those
296 * already active. MAY BLOCK!
297 */
298 _aio_exit( p );
299
300 /*
301 * Close open files and release open-file table.
302 * This may block!
303 */
304 fdfree(p);
305
306 /* Close ref SYSV Shared memory*/
307 if (p->vm_shm)
308 shmexit(p);
309 /* Release SYSV semaphores */
310 semexit(p);
311
312 if (SESS_LEADER(p)) {
313 register struct session *sp = p->p_session;
314
315 if (sp->s_ttyvp) {
316 struct vnode *ttyvp;
317 struct vfs_context context;
318
319 /*
320 * Controlling process.
321 * Signal foreground pgrp,
322 * drain controlling terminal
323 * and revoke access to controlling terminal.
324 */
325 if (sp->s_ttyp->t_session == sp) {
326 if (sp->s_ttyp->t_pgrp)
327 pgsignal(sp->s_ttyp->t_pgrp, SIGHUP, 1);
328 (void) ttywait(sp->s_ttyp);
329 /*
330 * The tty could have been revoked
331 * if we blocked.
332 */
333 context.vc_proc = p;
334 context.vc_ucred = p->p_ucred;
335 if (sp->s_ttyvp)
336 VNOP_REVOKE(sp->s_ttyvp, REVOKEALL, &context);
337 }
338 ttyvp = sp->s_ttyvp;
339 sp->s_ttyvp = NULL;
340 if (ttyvp) {
341 vnode_rele(ttyvp);
342 }
343 /*
344 * s_ttyp is not zero'd; we use this to indicate
345 * that the session once had a controlling terminal.
346 * (for logging and informational purposes)
347 */
348 }
349 sp->s_leader = NULL;
350 }
351
352 fixjobc(p, p->p_pgrp, 0);
353 p->p_rlimit[RLIMIT_FSIZE].rlim_cur = RLIM_INFINITY;
354 (void)acct_process(p);
355
356 #if KTRACE
357 /*
358 * release trace file
359 */
360 p->p_traceflag = 0; /* don't trace the vnode_put() */
361 if (p->p_tracep) {
362 struct vnode *tvp = p->p_tracep;
363 p->p_tracep = NULL;
364 vnode_rele(tvp);
365 }
366 #endif
367
368 while (q = p->p_children.lh_first) {
369 proc_reparent(q, initproc);
370 /*
371 * Traced processes are killed
372 * since their existence means someone is messing up.
373 */
374 if (q->p_flag & P_TRACED) {
375 q->p_flag &= ~P_TRACED;
376 if (q->sigwait_thread) {
377 /*
378 * The sigwait_thread could be stopped at a
379 * breakpoint. Wake it up to kill.
380 * Need to do this as it could be a thread which is not
381 * the first thread in the task. So any attempts to kill
382 * the process would result into a deadlock on q->sigwait.
383 */
384 thread_resume((thread_t)q->sigwait_thread);
385 clear_wait(q->sigwait_thread, THREAD_INTERRUPTED);
386 threadsignal((thread_t)q->sigwait_thread, SIGKILL, 0);
387 }
388 psignal(q, SIGKILL);
389 }
390 }
391
392 /*
393 * Save exit status and final rusage info, adding in child rusage
394 * info and self times. If we were unable to allocate a zombie
395 * structure, this information is lost.
396 */
397 if (p->p_ru != NULL) {
398 *p->p_ru = p->p_stats->p_ru;
399
400 timerclear(&p->p_ru->ru_utime);
401 timerclear(&p->p_ru->ru_stime);
402
403 if (task) {
404 task_basic_info_data_t tinfo;
405 task_thread_times_info_data_t ttimesinfo;
406 int task_info_stuff, task_ttimes_stuff;
407 struct timeval ut,st;
408
409 task_info_stuff = TASK_BASIC_INFO_COUNT;
410 task_info(task, TASK_BASIC_INFO,
411 (task_info_t)&tinfo, &task_info_stuff);
412 p->p_ru->ru_utime.tv_sec = tinfo.user_time.seconds;
413 p->p_ru->ru_utime.tv_usec = tinfo.user_time.microseconds;
414 p->p_ru->ru_stime.tv_sec = tinfo.system_time.seconds;
415 p->p_ru->ru_stime.tv_usec = tinfo.system_time.microseconds;
416
417 task_ttimes_stuff = TASK_THREAD_TIMES_INFO_COUNT;
418 task_info(task, TASK_THREAD_TIMES_INFO,
419 (task_info_t)&ttimesinfo, &task_ttimes_stuff);
420
421 ut.tv_sec = ttimesinfo.user_time.seconds;
422 ut.tv_usec = ttimesinfo.user_time.microseconds;
423 st.tv_sec = ttimesinfo.system_time.seconds;
424 st.tv_usec = ttimesinfo.system_time.microseconds;
425 timeradd(&ut,&p->p_ru->ru_utime,&p->p_ru->ru_utime);
426 timeradd(&st,&p->p_ru->ru_stime,&p->p_ru->ru_stime);
427 }
428
429 ruadd(p->p_ru, &p->p_stats->p_cru);
430 }
431
432 /*
433 * Free up profiling buffers.
434 */
435 {
436 struct uprof *p0 = &p->p_stats->p_prof, *p1, *pn;
437
438 p1 = p0->pr_next;
439 p0->pr_next = NULL;
440 p0->pr_scale = 0;
441
442 for (; p1 != NULL; p1 = pn) {
443 pn = p1->pr_next;
444 kfree(p1, sizeof *p1);
445 }
446 }
447
448 /*
449 * Other substructures are freed from wait().
450 */
451 FREE_ZONE(p->p_stats, sizeof *p->p_stats, M_SUBPROC);
452 p->p_stats = NULL;
453
454 FREE_ZONE(p->p_sigacts, sizeof *p->p_sigacts, M_SUBPROC);
455 p->p_sigacts = NULL;
456
457 if (--p->p_limit->p_refcnt == 0)
458 FREE_ZONE(p->p_limit, sizeof *p->p_limit, M_SUBPROC);
459 p->p_limit = NULL;
460
461 /*
462 * Finish up by terminating the task
463 * and halt this thread (only if a
464 * member of the task exiting).
465 */
466 p->task = TASK_NULL;
467 //task->proc = NULL;
468 set_bsdtask_info(task, NULL);
469
470 KNOTE(&p->p_klist, NOTE_EXIT);
471
472 /*
473 * Notify parent that we're gone.
474 */
475 if (p->p_pptr->p_flag & P_NOCLDWAIT) {
476 struct proc *opp = p->p_pptr;
477
478 /*
479 * Add child resource usage to parent before giving
480 * zombie to init. If we were unable to allocate a
481 * zombie structure, this information is lost.
482 */
483 if (p->p_ru != NULL)
484 ruadd(&p->p_pptr->p_stats->p_cru, p->p_ru);
485
486 proc_reparent(p, initproc);
487 /* If there are no more children wakeup parent */
488 if (LIST_EMPTY(&opp->p_children))
489 wakeup((caddr_t)opp);
490 }
491 /* should be fine as parent proc would be initproc */
492 pp = p->p_pptr;
493 if (pp != initproc) {
494 pp->si_pid = p->p_pid;
495 pp->si_status = p->p_xstat;
496 pp->si_code = CLD_EXITED;
497 pp->si_uid = p->p_ucred->cr_ruid;
498 }
499 /* mark as a zombie */
500 p->p_stat = SZOMB;
501
502 psignal(pp, SIGCHLD);
503
504 /* and now wakeup the parent */
505 wakeup((caddr_t)p->p_pptr);
506
507 (void) thread_funnel_set(kernel_flock, funnel_state);
508 }
509
510
511 /*
512 * reap_child_process
513 *
514 * Description: Given a process from which all status information needed
515 * has already been extracted, if the process is a ptrace
516 * attach process, detach it and give it back to its real
517 * parent, else recover all resources remaining associated
518 * with it.
519 *
520 * Parameters: struct proc *parent Parent of process being reaped
521 * struct proc *child Process to reap
522 *
523 * Returns: 0 Process was not reaped because it
524 * came from an attach
525 * 1 Process was reaped
526 */
527 static int
528 reap_child_process(struct proc *parent, struct proc *child)
529 {
530 struct proc *trace_parent; /* Traced parent process, if tracing */
531 struct vnode *tvp; /* Traced vnode pointer, if used */
532
533 /*
534 * If we got the child via a ptrace 'attach',
535 * we need to give it back to the old parent.
536 */
537 if (child->p_oppid && (trace_parent = pfind(child->p_oppid))) {
538 child->p_oppid = 0;
539 proc_reparent(child, trace_parent);
540 if (trace_parent != initproc) {
541 trace_parent->si_pid = child->p_pid;
542 trace_parent->si_status = child->p_xstat;
543 trace_parent->si_code = CLD_CONTINUED;
544 trace_parent->si_uid = child->p_ucred->cr_ruid;
545 }
546 psignal(trace_parent, SIGCHLD);
547 wakeup((caddr_t)trace_parent);
548 return (0);
549 }
550 child->p_xstat = 0;
551 if (child->p_ru) {
552 ruadd(&parent->p_stats->p_cru, child->p_ru);
553 FREE_ZONE(child->p_ru, sizeof *child->p_ru, M_ZOMBIE);
554 child->p_ru = NULL;
555 } else {
556 printf("Warning : lost p_ru for %s\n", child->p_comm);
557 }
558
559 /*
560 * Decrement the count of procs running with this uid.
561 */
562 (void)chgproccnt(child->p_ucred->cr_ruid, -1);
563
564 /*
565 * Free up credentials.
566 */
567 if (child->p_ucred != NOCRED) {
568 kauth_cred_t ucr = child->p_ucred;
569 child->p_ucred = NOCRED;
570 kauth_cred_rele(ucr);
571 }
572
573 /*
574 * Release reference to text vnode
575 */
576 tvp = child->p_textvp;
577 child->p_textvp = NULL;
578 if (tvp) {
579 vnode_rele(tvp);
580 }
581 /*
582 * Finally finished with old proc entry.
583 * Unlink it from its process group and free it.
584 */
585 leavepgrp(child);
586 LIST_REMOVE(child, p_list); /* off zombproc */
587 LIST_REMOVE(child, p_sibling);
588 child->p_flag &= ~P_WAITING;
589
590 lck_mtx_destroy(&child->p_mlock, proc_lck_grp);
591 lck_mtx_destroy(&child->p_fdmlock, proc_lck_grp);
592 FREE_ZONE(child, sizeof *child, M_PROC);
593 nprocs--;
594 return (1);
595 }
596
597
598 int
599 wait1continue(int result)
600 {
601 void *vt;
602 thread_t thread;
603 int *retval;
604 struct proc *p;
605
606 if (result)
607 return(result);
608
609 p = current_proc();
610 thread = current_thread();
611 vt = get_bsduthreadarg(thread);
612 retval = get_bsduthreadrval(thread);
613 return(wait4((struct proc *)p, (struct wait4_args *)vt, retval));
614 }
615
616 int
617 wait4(struct proc *q, struct wait4_args *uap, register_t *retval)
618 {
619 register int nfound;
620 register struct proc *p;
621 int status, error;
622
623 if (uap->pid == 0)
624 uap->pid = -q->p_pgid;
625
626 loop:
627 nfound = 0;
628 for (p = q->p_children.lh_first; p != 0; p = p->p_sibling.le_next) {
629 if (uap->pid != WAIT_ANY &&
630 p->p_pid != uap->pid &&
631 p->p_pgid != -(uap->pid))
632 continue;
633 nfound++;
634
635 /* XXX This is racy because we don't get the lock!!!! */
636
637 if (p->p_flag & P_WAITING) {
638 (void)tsleep(&p->p_stat, PWAIT, "waitcoll", 0);
639 goto loop;
640 }
641 p->p_flag |= P_WAITING; /* only allow single thread to wait() */
642
643 if (p->p_stat == SZOMB) {
644 retval[0] = p->p_pid;
645 if (uap->status) {
646 status = p->p_xstat; /* convert to int */
647 error = copyout((caddr_t)&status,
648 uap->status,
649 sizeof(status));
650 if (error) {
651 p->p_flag &= ~P_WAITING;
652 wakeup(&p->p_stat);
653 return (error);
654 }
655 }
656 if (uap->rusage) {
657 if (p->p_ru == NULL) {
658 error = ENOMEM;
659 } else {
660 if (IS_64BIT_PROCESS(q)) {
661 struct user_rusage my_rusage;
662 munge_rusage(p->p_ru, &my_rusage);
663 error = copyout((caddr_t)&my_rusage,
664 uap->rusage,
665 sizeof (my_rusage));
666 }
667 else {
668 error = copyout((caddr_t)p->p_ru,
669 uap->rusage,
670 sizeof (struct rusage));
671 }
672 }
673 /* information unavailable? */
674 if (error) {
675 p->p_flag &= ~P_WAITING;
676 wakeup(&p->p_stat);
677 return (error);
678 }
679 }
680
681 /* Clean up */
682 if (!reap_child_process(q, p))
683 p->p_flag &= ~P_WAITING;
684
685 /* Wake other wait'ers, if any */
686 wakeup(&p->p_stat);
687
688 return (0);
689 }
690 if (p->p_stat == SSTOP && (p->p_flag & P_WAITED) == 0 &&
691 (p->p_flag & P_TRACED || uap->options & WUNTRACED)) {
692 p->p_flag |= P_WAITED;
693 retval[0] = p->p_pid;
694 if (uap->status) {
695 status = W_STOPCODE(p->p_xstat);
696 error = copyout((caddr_t)&status,
697 uap->status,
698 sizeof(status));
699 } else
700 error = 0;
701 p->p_flag &= ~P_WAITING;
702 wakeup(&p->p_stat);
703 return (error);
704 }
705 p->p_flag &= ~P_WAITING;
706 wakeup(&p->p_stat);
707 }
708 if (nfound == 0)
709 return (ECHILD);
710
711 if (uap->options & WNOHANG) {
712 retval[0] = 0;
713 return (0);
714 }
715
716 if ((error = tsleep0((caddr_t)q, PWAIT | PCATCH, "wait", 0, wait1continue)))
717 return (error);
718
719 goto loop;
720 }
721
722
723 int
724 waitidcontinue(int result)
725 {
726 void *vt;
727 thread_t thread;
728 int *retval;
729 struct proc *p;
730
731 if (result)
732 return(result);
733
734 p = current_proc();
735 thread = current_thread();
736 vt = get_bsduthreadarg(thread);
737 retval = get_bsduthreadrval(thread);
738 return(waitid((struct proc *)p, (struct waitid_args *)vt, retval));
739 }
740
741 /*
742 * Description: Suspend the calling thread until one child of the process
743 * containing the calling thread changes state.
744 *
745 * Parameters: uap->idtype one of P_PID, P_PGID, P_ALL
746 * uap->id pid_t or gid_t or ignored
747 * uap->infop Address of signinfo_t struct in
748 * user space into which to return status
749 * uap->options flag values
750 *
751 * Returns: 0 Success
752 * !0 Error returning status to user space
753 */
754 int
755 waitid(struct proc *q, struct waitid_args *uap, register_t *retval)
756 {
757 user_siginfo_t collect64; /* siginfo data to return to caller */
758
759 register int nfound;
760 register struct proc *p;
761 int error;
762
763 loop:
764 nfound = 0;
765 for (p = q->p_children.lh_first; p != 0; p = p->p_sibling.le_next) {
766 switch(uap->idtype) {
767 case P_PID: /* child with process ID equal to... */
768 if (p->p_pid != (pid_t)uap->id)
769 continue;
770 break;
771 case P_PGID: /* child with process group ID equal to... */
772 if (p->p_pgid != (pid_t)uap->id)
773 continue;
774 break;
775 case P_ALL: /* any child */
776 break;
777 }
778
779 /* XXX This is racy because we don't get the lock!!!! */
780
781 /*
782 * Wait collision; go to sleep and restart; used to maintain
783 * the single return for waited process guarantee.
784 */
785 if (p->p_flag & P_WAITING) {
786 (void)tsleep(&p->p_stat, PWAIT, "waitidcoll", 0);
787 goto loop;
788 }
789 p->p_flag |= P_WAITING; /* mark busy */
790
791 nfound++;
792
793 /*
794 * Types of processes we are interested in
795 *
796 * XXX Don't know what to do for WCONTINUED?!?
797 */
798 switch(p->p_stat) {
799 case SZOMB: /* Exited */
800 if (!(uap->options & WEXITED))
801 break;
802
803 /* Collect "siginfo" information for caller */
804 collect64.si_signo = 0;
805 collect64.si_code = 0;
806 collect64.si_errno = 0;
807 collect64.si_pid = 0;
808 collect64.si_uid = 0;
809 collect64.si_addr = 0;
810 collect64.si_status = p->p_xstat;
811 collect64.si_band = 0;
812
813 if (IS_64BIT_PROCESS(p)) {
814 error = copyout((caddr_t)&collect64,
815 uap->infop,
816 sizeof(collect64));
817 } else {
818 siginfo_t collect;
819 siginfo_64to32(&collect64,&collect);
820 error = copyout((caddr_t)&collect,
821 uap->infop,
822 sizeof(collect));
823 }
824 /* information unavailable? */
825 if (error) {
826 p->p_flag &= ~P_WAITING;
827 wakeup(&p->p_stat);
828 return (error);
829 }
830
831 /* Prevent other process for waiting for this event? */
832 if (!(uap->options & WNOWAIT)) {
833 /* Clean up */
834 if (!reap_child_process(q, p))
835 p->p_flag &= ~P_WAITING;
836
837 /* Wake other wait'ers, if any */
838 wakeup(&p->p_stat);
839 }
840
841 return (0);
842
843 case SSTOP: /* Stopped */
844 /*
845 * If we are not interested in stopped processes, then
846 * ignore this one.
847 */
848 if (!(uap->options & WSTOPPED))
849 break;
850
851 /*
852 * If someone has already waited it, we lost a race
853 * to be the one to return status.
854 */
855 if ((p->p_flag & P_WAITED) != 0)
856 break;
857
858 /*
859 * If this is not a traced process, and they haven't
860 * indicated an interest in untraced processes, then
861 * ignore this one.
862 */
863 if (!(p->p_flag & P_TRACED) && !(uap->options & WUNTRACED))
864 break;
865
866 /* Collect "siginfo" information for caller */
867 collect64.si_signo = 0;
868 collect64.si_code = 0;
869 collect64.si_errno = 0;
870 collect64.si_pid = 0;
871 collect64.si_uid = 0;
872 collect64.si_addr = 0;
873 collect64.si_status = p->p_xstat;
874 collect64.si_band = 0;
875
876 if (IS_64BIT_PROCESS(p)) {
877 error = copyout((caddr_t)&collect64,
878 uap->infop,
879 sizeof(collect64));
880 } else {
881 siginfo_t collect;
882 siginfo_64to32(&collect64,&collect);
883 error = copyout((caddr_t)&collect,
884 uap->infop,
885 sizeof(collect));
886 }
887 /* information unavailable? */
888 if (error) {
889 p->p_flag &= ~P_WAITING;
890 wakeup(&p->p_stat);
891 return (error);
892 }
893
894 /* Prevent other process for waiting for this event? */
895 if (!(uap->options & WNOWAIT)) {
896 p->p_flag |= P_WAITED;
897 }
898
899 p->p_flag &= ~P_WAITING;
900 wakeup(&p->p_stat);
901 return (0);
902
903 default: /* All others */
904 /* ...meaning Continued */
905 if (!(uap->options & WCONTINUED))
906 break;
907
908 /*
909 * If the flag isn't set, then this process has not
910 * been stopped and continued, or the status has
911 * already been reaped by another caller of waitid().
912 */
913 if ((p->p_flag & P_CONTINUED) == 0)
914 break;
915
916 /* Collect "siginfo" information for caller */
917 collect64.si_signo = 0;
918 collect64.si_code = 0;
919 collect64.si_errno = 0;
920 collect64.si_pid = 0;
921 collect64.si_uid = 0;
922 collect64.si_addr = 0;
923 collect64.si_status = p->p_xstat;
924 collect64.si_band = 0;
925
926 if (IS_64BIT_PROCESS(p)) {
927 error = copyout((caddr_t)&collect64,
928 uap->infop,
929 sizeof(collect64));
930 } else {
931 siginfo_t collect;
932 siginfo_64to32(&collect64,&collect);
933 error = copyout((caddr_t)&collect,
934 uap->infop,
935 sizeof(collect));
936 }
937 /* information unavailable? */
938 if (error) {
939 p->p_flag &= ~P_WAITING;
940 wakeup(&p->p_stat);
941 return (error);
942 }
943
944 /* Prevent other process for waiting for this event? */
945 if (!(uap->options & WNOWAIT)) {
946 p->p_flag &= ~P_CONTINUED;
947 }
948
949 p->p_flag &= ~P_WAITING;
950 wakeup(&p->p_stat);
951 return (0);
952
953 break;
954 }
955
956
957 /* Not a process we are interested in; go on to next child */
958 p->p_flag &= ~P_WAITING;
959 wakeup(&p->p_stat);
960 }
961
962 /* No child processes that could possibly satisfy the request? */
963 if (nfound == 0)
964 return (ECHILD);
965
966 if (uap->options & WNOHANG) {
967 retval[0] = 0;
968 return (0);
969 }
970
971 if ((error = tsleep0((caddr_t)q, PWAIT | PCATCH, "waitid", 0, waitidcontinue)))
972 return (error);
973
974 goto loop;
975 }
976
977 /*
978 * make process 'parent' the new parent of process 'child'.
979 */
980 void
981 proc_reparent(struct proc *child, struct proc *parent)
982 {
983
984 if (child->p_pptr == parent)
985 return;
986
987 LIST_REMOVE(child, p_sibling);
988 LIST_INSERT_HEAD(&parent->p_children, child, p_sibling);
989 child->p_pptr = parent;
990
991 if (initproc == parent && child->p_stat == SZOMB)
992 psignal(initproc, SIGCHLD);
993 }
994
995 /*
996 * Make the current process an "init" process, meaning
997 * that it doesn't have a parent, and that it won't be
998 * gunned down by kill(-1, 0).
999 */
1000 kern_return_t
1001 init_process(__unused struct init_process_args *args)
1002 {
1003 register struct proc *p = current_proc();
1004
1005 AUDIT_MACH_SYSCALL_ENTER(AUE_INITPROCESS);
1006 if (suser(kauth_cred_get(), &p->p_acflag)) {
1007 AUDIT_MACH_SYSCALL_EXIT(KERN_NO_ACCESS);
1008 return(KERN_NO_ACCESS);
1009 }
1010
1011 if (p->p_pid != 1 && p->p_pgid != p->p_pid)
1012 enterpgrp(p, p->p_pid, 0);
1013 p->p_flag |= P_SYSTEM;
1014
1015 /*
1016 * Take us out of the sibling chain, and
1017 * out of our parent's child chain.
1018 */
1019 LIST_REMOVE(p, p_sibling);
1020 p->p_sibling.le_prev = NULL;
1021 p->p_sibling.le_next = NULL;
1022 p->p_pptr = kernproc;
1023
1024 AUDIT_MACH_SYSCALL_EXIT(KERN_SUCCESS);
1025 return(KERN_SUCCESS);
1026 }
1027
1028
1029 /*
1030 * Exit: deallocate address space and other resources, change proc state
1031 * to zombie, and unlink proc from allproc and parent's lists. Save exit
1032 * status and rusage for wait(). Check for child processes and orphan them.
1033 */
1034
1035 void
1036 vfork_exit(struct proc *p, int rv)
1037 {
1038 thread_t self = current_thread();
1039 #ifdef FIXME
1040 struct task *task = p->task;
1041 #endif
1042 register int s;
1043 struct uthread *ut;
1044 exception_data_t code[EXCEPTION_CODE_MAX];
1045
1046 /*
1047 * If a thread in this task has already
1048 * called exit(), then halt any others
1049 * right here.
1050 */
1051
1052 ut = get_bsdthread_info(self);
1053 #ifdef FIXME
1054 signal_lock(p);
1055 while (p->exit_thread != self) {
1056 if (sig_try_locked(p) <= 0) {
1057 if (get_threadtask(self) != task) {
1058 signal_unlock(p);
1059 return;
1060 }
1061 signal_unlock(p);
1062 thread_terminate(self);
1063 thread_funnel_set(kernel_flock, FALSE);
1064 thread_exception_return();
1065 /* NOTREACHED */
1066 }
1067 sig_lock_to_exit(p);
1068 }
1069 signal_unlock(p);
1070 if (p->p_pid == 1) {
1071 printf("pid 1 exited (signal %d, exit %d)",
1072 WTERMSIG(rv), WEXITSTATUS(rv));
1073 panic("init died\nState at Last Exception:\n\n%s", init_task_failure_data);
1074 }
1075 #endif /* FIXME */
1076
1077 s = splsched();
1078 p->p_flag |= P_WEXIT;
1079 p->p_lflag |= P_LPEXIT;
1080 splx(s);
1081
1082 code[0] = (exception_data_t)0xFF000001; /* Set terminate code */
1083 code[1] = (exception_data_t)p->p_pid; /* Pass out the pid */
1084 /* Notify the perf server */
1085 (void)sys_perf_notify(p->task, (exception_data_t)&code, 2);
1086
1087 /*
1088 * Remove proc from allproc queue and from pidhash chain.
1089 * Need to do this before we do anything that can block.
1090 * Not doing causes things like mount() find this on allproc
1091 * in partially cleaned state.
1092 */
1093 LIST_REMOVE(p, p_list);
1094 LIST_INSERT_HEAD(&zombproc, p, p_list); /* Place onto zombproc. */
1095 LIST_REMOVE(p, p_hash);
1096 /*
1097 * If parent is waiting for us to exit or exec,
1098 * P_PPWAIT is set; we will wakeup the parent below.
1099 */
1100 p->p_flag &= ~(P_TRACED | P_PPWAIT);
1101 p->p_sigignore = ~0;
1102 p->p_siglist = 0;
1103
1104 ut->uu_siglist = 0;
1105 untimeout(realitexpire, (caddr_t)p->p_pid);
1106
1107 p->p_xstat = rv;
1108
1109 vproc_exit(p);
1110 }
1111
1112 void
1113 vproc_exit(struct proc *p)
1114 {
1115 register struct proc *q, *nq, *pp;
1116 #ifdef FIXME
1117 struct task *task = p->task;
1118 #endif
1119
1120 /* XXX Zombie allocation may fail, in which case stats get lost */
1121 MALLOC_ZONE(p->p_ru, struct rusage *,
1122 sizeof (*p->p_ru), M_ZOMBIE, M_WAITOK);
1123
1124 /*
1125 * Close open files and release open-file table.
1126 * This may block!
1127 */
1128 fdfree(p);
1129
1130 if (SESS_LEADER(p)) {
1131 register struct session *sp = p->p_session;
1132
1133 if (sp->s_ttyvp) {
1134 struct vnode *ttyvp;
1135 struct vfs_context context;
1136
1137 /*
1138 * Controlling process.
1139 * Signal foreground pgrp,
1140 * drain controlling terminal
1141 * and revoke access to controlling terminal.
1142 */
1143 if (sp->s_ttyp->t_session == sp) {
1144 if (sp->s_ttyp->t_pgrp)
1145 pgsignal(sp->s_ttyp->t_pgrp, SIGHUP, 1);
1146 (void) ttywait(sp->s_ttyp);
1147 /*
1148 * The tty could have been revoked
1149 * if we blocked.
1150 */
1151 context.vc_proc = p;
1152 context.vc_ucred = p->p_ucred;
1153 if (sp->s_ttyvp)
1154 VNOP_REVOKE(sp->s_ttyvp, REVOKEALL, &context);
1155 }
1156 ttyvp = sp->s_ttyvp;
1157 sp->s_ttyvp = NULL;
1158 if (ttyvp) {
1159 vnode_rele(ttyvp);
1160 }
1161 /*
1162 * s_ttyp is not zero'd; we use this to indicate
1163 * that the session once had a controlling terminal.
1164 * (for logging and informational purposes)
1165 */
1166 }
1167 sp->s_leader = NULL;
1168 }
1169
1170 fixjobc(p, p->p_pgrp, 0);
1171 p->p_rlimit[RLIMIT_FSIZE].rlim_cur = RLIM_INFINITY;
1172
1173 #if KTRACE
1174 /*
1175 * release trace file
1176 */
1177 p->p_traceflag = 0; /* don't trace the vnode_rele() */
1178 if (p->p_tracep) {
1179 struct vnode *tvp = p->p_tracep;
1180 p->p_tracep = NULL;
1181 vnode_rele(tvp);
1182 }
1183 #endif
1184
1185 while (q = p->p_children.lh_first) {
1186 proc_reparent(q, initproc);
1187 /*
1188 * Traced processes are killed
1189 * since their existence means someone is messing up.
1190 */
1191 if (q->p_flag & P_TRACED) {
1192 q->p_flag &= ~P_TRACED;
1193 if (q->sigwait_thread) {
1194 /*
1195 * The sigwait_thread could be stopped at a
1196 * breakpoint. Wake it up to kill.
1197 * Need to do this as it could be a thread which is not
1198 * the first thread in the task. So any attempts to kill
1199 * the process would result into a deadlock on q->sigwait.
1200 */
1201 thread_resume((thread_t)q->sigwait_thread);
1202 clear_wait(q->sigwait_thread, THREAD_INTERRUPTED);
1203 threadsignal((thread_t)q->sigwait_thread, SIGKILL, 0);
1204 }
1205 psignal(q, SIGKILL);
1206 }
1207 }
1208
1209 /*
1210 * Save exit status and final rusage info, adding in child rusage
1211 * info and self times. If we were unable to allocate a zombie
1212 * structure, this information is lost.
1213 */
1214 if (p->p_ru != NULL) {
1215 *p->p_ru = p->p_stats->p_ru;
1216 timerclear(&p->p_ru->ru_utime);
1217 timerclear(&p->p_ru->ru_stime);
1218
1219 #ifdef FIXME
1220 if (task) {
1221 task_basic_info_data_t tinfo;
1222 task_thread_times_info_data_t ttimesinfo;
1223 int task_info_stuff, task_ttimes_stuff;
1224 struct timeval ut,st;
1225
1226 task_info_stuff = TASK_BASIC_INFO_COUNT;
1227 task_info(task, TASK_BASIC_INFO,
1228 &tinfo, &task_info_stuff);
1229 p->p_ru->ru_utime.tv_sec = tinfo.user_time.seconds;
1230 p->p_ru->ru_utime.tv_usec = tinfo.user_time.microseconds;
1231 p->p_ru->ru_stime.tv_sec = tinfo.system_time.seconds;
1232 p->p_ru->ru_stime.tv_usec = tinfo.system_time.microseconds;
1233
1234 task_ttimes_stuff = TASK_THREAD_TIMES_INFO_COUNT;
1235 task_info(task, TASK_THREAD_TIMES_INFO,
1236 &ttimesinfo, &task_ttimes_stuff);
1237
1238 ut.tv_sec = ttimesinfo.user_time.seconds;
1239 ut.tv_usec = ttimesinfo.user_time.microseconds;
1240 st.tv_sec = ttimesinfo.system_time.seconds;
1241 st.tv_usec = ttimesinfo.system_time.microseconds;
1242 timeradd(&ut,&p->p_ru->ru_utime,&p->p_ru->ru_utime);
1243 timeradd(&st,&p->p_ru->ru_stime,&p->p_ru->ru_stime);
1244 }
1245 #endif /* FIXME */
1246
1247 ruadd(p->p_ru, &p->p_stats->p_cru);
1248 }
1249
1250 /*
1251 * Free up profiling buffers.
1252 */
1253 {
1254 struct uprof *p0 = &p->p_stats->p_prof, *p1, *pn;
1255
1256 p1 = p0->pr_next;
1257 p0->pr_next = NULL;
1258 p0->pr_scale = 0;
1259
1260 for (; p1 != NULL; p1 = pn) {
1261 pn = p1->pr_next;
1262 kfree(p1, sizeof *p1);
1263 }
1264 }
1265
1266 /*
1267 * Other substructures are freed from wait().
1268 */
1269 FREE_ZONE(p->p_stats, sizeof *p->p_stats, M_SUBPROC);
1270 p->p_stats = NULL;
1271
1272 FREE_ZONE(p->p_sigacts, sizeof *p->p_sigacts, M_SUBPROC);
1273 p->p_sigacts = NULL;
1274
1275 if (--p->p_limit->p_refcnt == 0)
1276 FREE_ZONE(p->p_limit, sizeof *p->p_limit, M_SUBPROC);
1277 p->p_limit = NULL;
1278
1279 /*
1280 * Finish up by terminating the task
1281 * and halt this thread (only if a
1282 * member of the task exiting).
1283 */
1284 p->task = TASK_NULL;
1285
1286 /*
1287 * Notify parent that we're gone.
1288 */
1289 pp = p->p_pptr;
1290 if (pp != initproc) {
1291 pp->si_pid = p->p_pid;
1292 pp->si_status = p->p_xstat;
1293 pp->si_code = CLD_EXITED;
1294 pp->si_uid = p->p_ucred->cr_ruid;
1295 }
1296 /* mark as a zombie */
1297 p->p_stat = SZOMB;
1298
1299 psignal(p->p_pptr, SIGCHLD);
1300
1301 /* and now wakeup the parent */
1302 wakeup((caddr_t)p->p_pptr);
1303 }
1304
1305
1306 /*
1307 * munge_rusage
1308 * LP64 support - long is 64 bits if we are dealing with a 64 bit user
1309 * process. We munge the kernel (32 bit) version of rusage into the
1310 * 64 bit version.
1311 */
1312 __private_extern__ void
1313 munge_rusage(struct rusage *a_rusage_p, struct user_rusage *a_user_rusage_p)
1314 {
1315 /* timeval changes size, so utime and stime need special handling */
1316 a_user_rusage_p->ru_utime.tv_sec = a_rusage_p->ru_utime.tv_sec;
1317 a_user_rusage_p->ru_utime.tv_usec = a_rusage_p->ru_utime.tv_usec;
1318 a_user_rusage_p->ru_stime.tv_sec = a_rusage_p->ru_stime.tv_sec;
1319 a_user_rusage_p->ru_stime.tv_usec = a_rusage_p->ru_stime.tv_usec;
1320 /*
1321 * everything else can be a direct assign, since there is no loss
1322 * of precision implied boing 32->64.
1323 */
1324 a_user_rusage_p->ru_maxrss = a_rusage_p->ru_maxrss;
1325 a_user_rusage_p->ru_ixrss = a_rusage_p->ru_ixrss;
1326 a_user_rusage_p->ru_idrss = a_rusage_p->ru_idrss;
1327 a_user_rusage_p->ru_isrss = a_rusage_p->ru_isrss;
1328 a_user_rusage_p->ru_minflt = a_rusage_p->ru_minflt;
1329 a_user_rusage_p->ru_majflt = a_rusage_p->ru_majflt;
1330 a_user_rusage_p->ru_nswap = a_rusage_p->ru_nswap;
1331 a_user_rusage_p->ru_inblock = a_rusage_p->ru_inblock;
1332 a_user_rusage_p->ru_oublock = a_rusage_p->ru_oublock;
1333 a_user_rusage_p->ru_msgsnd = a_rusage_p->ru_msgsnd;
1334 a_user_rusage_p->ru_msgrcv = a_rusage_p->ru_msgrcv;
1335 a_user_rusage_p->ru_nsignals = a_rusage_p->ru_nsignals;
1336 a_user_rusage_p->ru_nvcsw = a_rusage_p->ru_nvcsw;
1337 a_user_rusage_p->ru_nivcsw = a_rusage_p->ru_nivcsw;
1338 }