X-Git-Url: https://git.saurik.com/apple/xnu.git/blobdiff_plain/55e303ae13a4cf49d70f2294092726f2fffb9ef2..bb59bff194111743b33cc36712410b5656329d3c:/bsd/kern/kern_synch.c?ds=sidebyside diff --git a/bsd/kern/kern_synch.c b/bsd/kern/kern_synch.c index 8bb748231..b1c4eda1c 100644 --- a/bsd/kern/kern_synch.c +++ b/bsd/kern/kern_synch.c @@ -1,16 +1,19 @@ /* - * Copyright (c) 2000-2001 Apple Computer, Inc. All rights reserved. + * Copyright (c) 2000-2006 Apple Computer, Inc. All rights reserved. * - * @APPLE_LICENSE_HEADER_START@ - * - * Copyright (c) 1999-2003 Apple Computer, Inc. All Rights Reserved. + * @APPLE_OSREFERENCE_LICENSE_HEADER_START@ * * This file contains Original Code and/or Modifications of Original Code * as defined in and that are subject to the Apple Public Source License * Version 2.0 (the 'License'). You may not use this file except in - * compliance with the License. Please obtain a copy of the License at - * http://www.opensource.apple.com/apsl/ and read it before using this - * file. + * compliance with the License. The rights granted to you under the License + * may not be used to create, or enable the creation or redistribution of, + * unlawful or unlicensed copies of an Apple operating system, or to + * circumvent, violate, or enable the circumvention or violation of, any + * terms of an Apple operating system software license agreement. + * + * Please obtain a copy of the License at + * http://www.opensource.apple.com/apsl/ and read it before using this file. * * The Original Code and all software distributed under the License are * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER @@ -20,7 +23,7 @@ * Please see the License for the specific language governing rights and * limitations under the License. * - * @APPLE_LICENSE_HEADER_END@ + * @APPLE_OSREFERENCE_LICENSE_HEADER_END@ */ /* * Mach Operating System @@ -31,12 +34,11 @@ #include #include -#include +#include #include -#include +#include #include #include -#include #include @@ -51,26 +53,31 @@ #include #include +#include + +#include /* for unix_syscall_return() */ +#include + +extern void compute_averunnable(void *); /* XXX */ + -#if KTRACE -#include -#include -#endif static void -_sleep_continue(void) +_sleep_continue( __unused void *parameter, wait_result_t wresult) { - register struct proc *p; - register thread_t self = current_act(); + struct proc *p = current_proc(); + thread_t self = current_thread(); struct uthread * ut; int sig, catch; int error = 0; + int dropmutex, spinmutex; ut = get_bsdthread_info(self); - catch = ut->uu_pri & PCATCH; - p = current_proc(); + catch = ut->uu_pri & PCATCH; + dropmutex = ut->uu_pri & PDROP; + spinmutex = ut->uu_pri & PSPIN; - switch (get_thread_waitresult(self)) { + switch (wresult) { case THREAD_TIMED_OUT: error = EWOULDBLOCK; break; @@ -88,7 +95,7 @@ _sleep_continue(void) if (thread_should_abort(self)) { error = EINTR; } else if (SHOULDissignal(p,ut)) { - if (sig = CURSIG(p)) { + if ((sig = CURSIG(p)) != 0) { if (p->p_sigacts->ps_sigintr & sigmask(sig)) error = EINTR; else @@ -97,7 +104,10 @@ _sleep_continue(void) if (thread_should_abort(self)) { error = EINTR; } - } + } else if( (ut->uu_flag & ( UT_CANCELDISABLE | UT_CANCEL | UT_CANCELED)) == UT_CANCEL) { + /* due to thread cancel */ + error = EINTR; + } } else error = EINTR; break; @@ -106,13 +116,14 @@ _sleep_continue(void) if (error == EINTR || error == ERESTART) act_set_astbsd(self); - if (ut->uu_timo) - thread_cancel_timer(); - -#if KTRACE - if (KTRPOINT(p, KTR_CSW)) - ktrcsw(p->p_tracep, 0, 0, -1); -#endif + if (ut->uu_mtx && !dropmutex) { + if (spinmutex) + lck_mtx_lock_spin(ut->uu_mtx); + else + lck_mtx_lock(ut->uu_mtx); + } + ut->uu_wchan = NULL; + ut->uu_wmesg = NULL; unix_syscall_return((*ut->uu_continuation)(error)); } @@ -129,104 +140,126 @@ _sleep_continue(void) * Callers of this routine must be prepared for * premature return, and check that the reason for * sleeping has gone away. + * + * if msleep was the entry point, than we have a mutex to deal with + * + * The mutex is unlocked before the caller is blocked, and + * relocked before msleep returns unless the priority includes the PDROP + * flag... if PDROP is specified, _sleep returns with the mutex unlocked + * regardless of whether it actually blocked or not. */ static int _sleep( caddr_t chan, - int pri, - char *wmsg, + int pri, + const char *wmsg, u_int64_t abstime, - int (*continuation)(int)) + int (*continuation)(int), + lck_mtx_t *mtx) { - register struct proc *p; - register thread_t self = current_act(); + struct proc *p; + thread_t self = current_thread(); struct uthread * ut; - int sig, catch = pri & PCATCH; - int sigttblock = pri & PTTYBLOCK; + int sig, catch; + int dropmutex = pri & PDROP; + int spinmutex = pri & PSPIN; int wait_result; int error = 0; - spl_t s; - - s = splhigh(); ut = get_bsdthread_info(self); - + p = current_proc(); -#if KTRACE - if (KTRPOINT(p, KTR_CSW)) - ktrcsw(p->p_tracep, 1, 0, -1); -#endif p->p_priority = pri & PRIMASK; - - if (chan != NULL) - assert_wait_prim(chan, NULL, abstime, - (catch) ? THREAD_ABORTSAFE : THREAD_UNINT); + /* It can still block in proc_exit() after the teardown. */ + if (p->p_stats != NULL) + OSIncrementAtomicLong(&p->p_stats->p_ru.ru_nvcsw); + + if (pri & PCATCH) + catch = THREAD_ABORTSAFE; else - if (abstime != 0) - thread_set_timer_deadline(abstime); - - /* - * We start our timeout - * before calling CURSIG, as we could stop there, and a wakeup - * or a SIGCONT (or both) could occur while we were stopped. - * A SIGCONT would cause us to be marked as SSLEEP - * without resuming us, thus we must be ready for sleep - * when CURSIG is called. If the wakeup happens while we're - * stopped, p->p_wchan will be 0 upon return from CURSIG. - */ - if (catch) { - if (SHOULDissignal(p,ut)) { - if (sig = CURSIG(p)) { + catch = THREAD_UNINT; + + /* set wait message & channel */ + ut->uu_wchan = chan; + ut->uu_wmesg = wmsg ? wmsg : "unknown"; + + if (mtx != NULL && chan != NULL && (thread_continue_t)continuation == THREAD_CONTINUE_NULL) { + int flags; + + if (dropmutex) + flags = LCK_SLEEP_UNLOCK; + else + flags = LCK_SLEEP_DEFAULT; + + if (spinmutex) + flags |= LCK_SLEEP_SPIN; + + if (abstime) + wait_result = lck_mtx_sleep_deadline(mtx, flags, chan, catch, abstime); + else + wait_result = lck_mtx_sleep(mtx, flags, chan, catch); + } + else { + if (chan != NULL) + assert_wait_deadline(chan, catch, abstime); + if (mtx) + lck_mtx_unlock(mtx); + + if (catch == THREAD_ABORTSAFE) { + if (SHOULDissignal(p,ut)) { + if ((sig = CURSIG(p)) != 0) { + if (clear_wait(self, THREAD_INTERRUPTED) == KERN_FAILURE) + goto block; + if (p->p_sigacts->ps_sigintr & sigmask(sig)) + error = EINTR; + else + error = ERESTART; + if (mtx && !dropmutex) { + if (spinmutex) + lck_mtx_lock_spin(mtx); + else + lck_mtx_lock(mtx); + } + goto out; + } + } + if (thread_should_abort(self)) { if (clear_wait(self, THREAD_INTERRUPTED) == KERN_FAILURE) goto block; - /* if SIGTTOU or SIGTTIN then block till SIGCONT */ - if (sigttblock && ((sig == SIGTTOU) || (sig == SIGTTIN))) { - p->p_flag |= P_TTYSLEEP; - /* reset signal bits */ - clear_procsiglist(p, sig); - assert_wait(&p->p_siglist, THREAD_ABORTSAFE); - /* assert wait can block and SIGCONT should be checked */ - if (p->p_flag & P_TTYSLEEP) - thread_block(THREAD_CONTINUE_NULL); - /* return with success */ - error = 0; - goto out; + error = EINTR; + + if (mtx && !dropmutex) { + if (spinmutex) + lck_mtx_lock_spin(mtx); + else + lck_mtx_lock(mtx); } - if (p->p_sigacts->ps_sigintr & sigmask(sig)) - error = EINTR; - else - error = ERESTART; goto out; } - } - if (thread_should_abort(self)) { - if (clear_wait(self, THREAD_INTERRUPTED) == KERN_FAILURE) - goto block; - error = EINTR; - goto out; - } - if (get_thread_waitresult(self) != THREAD_WAITING) { - /*already happened */ - goto out; - } - } + } -block: - splx(s); - p->p_stats->p_ru.ru_nvcsw++; +block: + if ((thread_continue_t)continuation != THREAD_CONTINUE_NULL) { + ut->uu_continuation = continuation; + ut->uu_pri = pri; + ut->uu_timo = abstime? 1: 0; + ut->uu_mtx = mtx; + (void) thread_block(_sleep_continue); + /* NOTREACHED */ + } + + wait_result = thread_block(THREAD_CONTINUE_NULL); - if ((thread_continue_t)continuation != THREAD_CONTINUE_NULL ) { - ut->uu_continuation = continuation; - ut->uu_pri = pri; - ut->uu_timo = abstime? 1: 0; - (void) thread_block(_sleep_continue); - /* NOTREACHED */ + if (mtx && !dropmutex) { + if (spinmutex) + lck_mtx_lock_spin(mtx); + else + lck_mtx_lock(mtx); + } } - wait_result = thread_block(THREAD_CONTINUE_NULL); - switch (wait_result) { case THREAD_TIMED_OUT: error = EWOULDBLOCK; @@ -237,15 +270,15 @@ block: * first, regardless of whether awakened due * to receiving event. */ - if (!catch) + if (catch != THREAD_ABORTSAFE) break; /* else fall through */ case THREAD_INTERRUPTED: - if (catch) { + if (catch == THREAD_ABORTSAFE) { if (thread_should_abort(self)) { error = EINTR; - } else if (SHOULDissignal(p,ut)) { - if (sig = CURSIG(p)) { + } else if (SHOULDissignal(p, ut)) { + if ((sig = CURSIG(p)) != 0) { if (p->p_sigacts->ps_sigintr & sigmask(sig)) error = EINTR; else @@ -254,7 +287,10 @@ block: if (thread_should_abort(self)) { error = EINTR; } - } + } else if( (ut->uu_flag & ( UT_CANCELDISABLE | UT_CANCEL | UT_CANCELED)) == UT_CANCEL) { + /* due to thread cancel */ + error = EINTR; + } } else error = EINTR; break; @@ -262,13 +298,9 @@ block: out: if (error == EINTR || error == ERESTART) act_set_astbsd(self); - if (abstime) - thread_cancel_timer(); - (void) splx(s); -#if KTRACE - if (KTRPOINT(p, KTR_CSW)) - ktrcsw(p->p_tracep, 0, 0, -1); -#endif + ut->uu_wchan = NULL; + ut->uu_wmesg = NULL; + return (error); } @@ -277,28 +309,74 @@ sleep( void *chan, int pri) { - return _sleep((caddr_t)chan, pri, (char *)NULL, 0, (int (*)(int))0); + return _sleep((caddr_t)chan, pri, (char *)NULL, 0, (int (*)(int))0, (lck_mtx_t *)0); +} + +int +msleep0( + void *chan, + lck_mtx_t *mtx, + int pri, + const char *wmsg, + int timo, + int (*continuation)(int)) +{ + u_int64_t abstime = 0; + + if (timo) + clock_interval_to_deadline(timo, NSEC_PER_SEC / hz, &abstime); + + return _sleep((caddr_t)chan, pri, wmsg, abstime, continuation, mtx); +} + +int +msleep( + void *chan, + lck_mtx_t *mtx, + int pri, + const char *wmsg, + struct timespec *ts) +{ + u_int64_t abstime = 0; + + if (ts && (ts->tv_sec || ts->tv_nsec)) { + nanoseconds_to_absolutetime((uint64_t)ts->tv_sec * NSEC_PER_SEC + ts->tv_nsec, &abstime ); + clock_absolutetime_interval_to_deadline( abstime, &abstime ); + } + + return _sleep((caddr_t)chan, pri, wmsg, abstime, (int (*)(int))0, mtx); +} + +int +msleep1( + void *chan, + lck_mtx_t *mtx, + int pri, + const char *wmsg, + u_int64_t abstime) +{ + return _sleep((caddr_t)chan, pri, wmsg, abstime, (int (*)(int))0, mtx); } int tsleep( - void *chan, + void *chan, int pri, - char *wmsg, + const char *wmsg, int timo) { u_int64_t abstime = 0; if (timo) clock_interval_to_deadline(timo, NSEC_PER_SEC / hz, &abstime); - return _sleep((caddr_t)chan, pri, wmsg, abstime, (int (*)(int))0); + return _sleep((caddr_t)chan, pri, wmsg, abstime, (int (*)(int))0, (lck_mtx_t *)0); } int tsleep0( - void *chan, + void *chan, int pri, - char *wmsg, + const char *wmsg, int timo, int (*continuation)(int)) { @@ -306,28 +384,27 @@ tsleep0( if (timo) clock_interval_to_deadline(timo, NSEC_PER_SEC / hz, &abstime); - return _sleep((caddr_t)chan, pri, wmsg, abstime, continuation); + return _sleep((caddr_t)chan, pri, wmsg, abstime, continuation, (lck_mtx_t *)0); } int tsleep1( void *chan, - int pri, - char *wmsg, + int pri, + const char *wmsg, u_int64_t abstime, - int (*continuation)(int)) + int (*continuation)(int)) { - return _sleep((caddr_t)chan, pri, wmsg, abstime, continuation); + return _sleep((caddr_t)chan, pri, wmsg, abstime, continuation, (lck_mtx_t *)0); } /* * Wake up all processes sleeping on chan. */ void -wakeup(chan) - register void *chan; +wakeup(void *chan) { - thread_wakeup_prim((caddr_t)chan, FALSE, THREAD_AWAKENED); + thread_wakeup((caddr_t)chan); } /* @@ -337,10 +414,9 @@ wakeup(chan) * the right one to wakeup. */ void -wakeup_one(chan) - register caddr_t chan; +wakeup_one(caddr_t chan) { - thread_wakeup_prim((caddr_t)chan, TRUE, THREAD_AWAKENED); + thread_wakeup_one((caddr_t)chan); } /* @@ -349,8 +425,7 @@ wakeup_one(chan) * than that of the current process. */ void -resetpriority(p) - register struct proc *p; +resetpriority(struct proc *p) { (void)task_importance(p->task, -p->p_nice); } @@ -368,11 +443,11 @@ static fixpt_t cexp[3] = { }; void -compute_averunnable( - register int nrun) +compute_averunnable(void *arg) { - register int i; + unsigned int nrun = *(unsigned int *)arg; struct loadavg *avg = &averunnable; + int i; for (i = 0; i < 3; i++) avg->ldavg[i] = (cexp[i] * avg->ldavg[i] +