]>
git.saurik.com Git - apple/xnu.git/blob - bsd/kern/kern_synch.c
c6b4888c35692b0fd16ae0cb6ebde657ea0ce1bc
2 * Copyright (c) 2000-2006 Apple Computer, Inc. All rights reserved.
4 * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
6 * This file contains Original Code and/or Modifications of Original Code
7 * as defined in and that are subject to the Apple Public Source License
8 * Version 2.0 (the 'License'). You may not use this file except in
9 * compliance with the License. The rights granted to you under the License
10 * may not be used to create, or enable the creation or redistribution of,
11 * unlawful or unlicensed copies of an Apple operating system, or to
12 * circumvent, violate, or enable the circumvention or violation of, any
13 * terms of an Apple operating system software license agreement.
15 * Please obtain a copy of the License at
16 * http://www.opensource.apple.com/apsl/ and read it before using this file.
18 * The Original Code and all software distributed under the License are
19 * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
20 * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
21 * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
22 * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
23 * Please see the License for the specific language governing rights and
24 * limitations under the License.
26 * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
29 * Mach Operating System
30 * Copyright (c) 1987 Carnegie-Mellon University
31 * All rights reserved. The CMU software License Agreement specifies
32 * the terms and conditions for use and redistribution.
35 #include <sys/param.h>
36 #include <sys/systm.h>
37 #include <sys/proc_internal.h>
39 #include <sys/file_internal.h>
40 #include <sys/vnode.h>
41 #include <sys/kernel.h>
43 #include <machine/spl.h>
45 #include <kern/queue.h>
47 #include <kern/thread.h>
48 #include <kern/sched_prim.h>
51 #include <kern/cpu_number.h>
52 #include <vm/vm_kern.h>
54 #include <kern/task.h>
55 #include <mach/time_value.h>
56 #include <kern/lock.h>
58 #include <sys/systm.h> /* for unix_syscall_return() */
59 #include <libkern/OSAtomic.h>
61 extern boolean_t
thread_should_abort(thread_t
); /* XXX */
62 extern void compute_averunnable(void *); /* XXX */
67 _sleep_continue( __unused
void *parameter
, wait_result_t wresult
)
69 struct proc
*p
= current_proc();
70 thread_t self
= current_thread();
74 int dropmutex
, spinmutex
;
76 ut
= get_bsdthread_info(self
);
77 catch = ut
->uu_pri
& PCATCH
;
78 dropmutex
= ut
->uu_pri
& PDROP
;
79 spinmutex
= ut
->uu_pri
& PSPIN
;
82 case THREAD_TIMED_OUT
:
87 * Posix implies any signal should be delivered
88 * first, regardless of whether awakened due
93 /* else fall through */
94 case THREAD_INTERRUPTED
:
96 if (thread_should_abort(self
)) {
98 } else if (SHOULDissignal(p
,ut
)) {
99 if ((sig
= CURSIG(p
)) != 0) {
100 if (p
->p_sigacts
->ps_sigintr
& sigmask(sig
))
105 if (thread_should_abort(self
)) {
108 } else if( (ut
->uu_flag
& ( UT_CANCELDISABLE
| UT_CANCEL
| UT_CANCELED
)) == UT_CANCEL
) {
109 /* due to thread cancel */
117 if (error
== EINTR
|| error
== ERESTART
)
118 act_set_astbsd(self
);
120 if (ut
->uu_mtx
&& !dropmutex
) {
122 lck_mtx_lock_spin(ut
->uu_mtx
);
124 lck_mtx_lock(ut
->uu_mtx
);
129 unix_syscall_return((*ut
->uu_continuation
)(error
));
133 * Give up the processor till a wakeup occurs
134 * on chan, at which time the process
135 * enters the scheduling queue at priority pri.
136 * The most important effect of pri is that when
137 * pri<=PZERO a signal cannot disturb the sleep;
138 * if pri>PZERO signals will be processed.
139 * If pri&PCATCH is set, signals will cause sleep
140 * to return 1, rather than longjmp.
141 * Callers of this routine must be prepared for
142 * premature return, and check that the reason for
143 * sleeping has gone away.
145 * if msleep was the entry point, than we have a mutex to deal with
147 * The mutex is unlocked before the caller is blocked, and
148 * relocked before msleep returns unless the priority includes the PDROP
149 * flag... if PDROP is specified, _sleep returns with the mutex unlocked
150 * regardless of whether it actually blocked or not.
159 int (*continuation
)(int),
163 thread_t self
= current_thread();
166 int dropmutex
= pri
& PDROP
;
167 int spinmutex
= pri
& PSPIN
;
171 ut
= get_bsdthread_info(self
);
174 p
->p_priority
= pri
& PRIMASK
;
175 /* It can still block in proc_exit() after the teardown. */
176 if (p
->p_stats
!= NULL
)
177 OSIncrementAtomicLong(&p
->p_stats
->p_ru
.ru_nvcsw
);
180 catch = THREAD_ABORTSAFE
;
182 catch = THREAD_UNINT
;
184 /* set wait message & channel */
186 ut
->uu_wmesg
= wmsg
? wmsg
: "unknown";
188 if (mtx
!= NULL
&& chan
!= NULL
&& (thread_continue_t
)continuation
== THREAD_CONTINUE_NULL
) {
192 flags
= LCK_SLEEP_UNLOCK
;
194 flags
= LCK_SLEEP_DEFAULT
;
197 flags
|= LCK_SLEEP_SPIN
;
200 wait_result
= lck_mtx_sleep_deadline(mtx
, flags
, chan
, catch, abstime
);
202 wait_result
= lck_mtx_sleep(mtx
, flags
, chan
, catch);
206 assert_wait_deadline(chan
, catch, abstime
);
210 if (catch == THREAD_ABORTSAFE
) {
211 if (SHOULDissignal(p
,ut
)) {
212 if ((sig
= CURSIG(p
)) != 0) {
213 if (clear_wait(self
, THREAD_INTERRUPTED
) == KERN_FAILURE
)
215 if (p
->p_sigacts
->ps_sigintr
& sigmask(sig
))
219 if (mtx
&& !dropmutex
) {
221 lck_mtx_lock_spin(mtx
);
228 if (thread_should_abort(self
)) {
229 if (clear_wait(self
, THREAD_INTERRUPTED
) == KERN_FAILURE
)
233 if (mtx
&& !dropmutex
) {
235 lck_mtx_lock_spin(mtx
);
245 if ((thread_continue_t
)continuation
!= THREAD_CONTINUE_NULL
) {
246 ut
->uu_continuation
= continuation
;
248 ut
->uu_timo
= abstime
? 1: 0;
250 (void) thread_block(_sleep_continue
);
254 wait_result
= thread_block(THREAD_CONTINUE_NULL
);
256 if (mtx
&& !dropmutex
) {
258 lck_mtx_lock_spin(mtx
);
264 switch (wait_result
) {
265 case THREAD_TIMED_OUT
:
268 case THREAD_AWAKENED
:
270 * Posix implies any signal should be delivered
271 * first, regardless of whether awakened due
272 * to receiving event.
274 if (catch != THREAD_ABORTSAFE
)
276 /* else fall through */
277 case THREAD_INTERRUPTED
:
278 if (catch == THREAD_ABORTSAFE
) {
279 if (thread_should_abort(self
)) {
281 } else if (SHOULDissignal(p
, ut
)) {
282 if ((sig
= CURSIG(p
)) != 0) {
283 if (p
->p_sigacts
->ps_sigintr
& sigmask(sig
))
288 if (thread_should_abort(self
)) {
291 } else if( (ut
->uu_flag
& ( UT_CANCELDISABLE
| UT_CANCEL
| UT_CANCELED
)) == UT_CANCEL
) {
292 /* due to thread cancel */
300 if (error
== EINTR
|| error
== ERESTART
)
301 act_set_astbsd(self
);
313 return _sleep((caddr_t
)chan
, pri
, (char *)NULL
, 0, (int (*)(int))0, (lck_mtx_t
*)0);
323 int (*continuation
)(int))
325 u_int64_t abstime
= 0;
328 clock_interval_to_deadline(timo
, NSEC_PER_SEC
/ hz
, &abstime
);
330 return _sleep((caddr_t
)chan
, pri
, wmsg
, abstime
, continuation
, mtx
);
341 u_int64_t abstime
= 0;
343 if (ts
&& (ts
->tv_sec
|| ts
->tv_nsec
)) {
344 nanoseconds_to_absolutetime((uint64_t)ts
->tv_sec
* NSEC_PER_SEC
+ ts
->tv_nsec
, &abstime
);
345 clock_absolutetime_interval_to_deadline( abstime
, &abstime
);
348 return _sleep((caddr_t
)chan
, pri
, wmsg
, abstime
, (int (*)(int))0, mtx
);
359 return _sleep((caddr_t
)chan
, pri
, wmsg
, abstime
, (int (*)(int))0, mtx
);
369 u_int64_t abstime
= 0;
372 clock_interval_to_deadline(timo
, NSEC_PER_SEC
/ hz
, &abstime
);
373 return _sleep((caddr_t
)chan
, pri
, wmsg
, abstime
, (int (*)(int))0, (lck_mtx_t
*)0);
382 int (*continuation
)(int))
384 u_int64_t abstime
= 0;
387 clock_interval_to_deadline(timo
, NSEC_PER_SEC
/ hz
, &abstime
);
388 return _sleep((caddr_t
)chan
, pri
, wmsg
, abstime
, continuation
, (lck_mtx_t
*)0);
397 int (*continuation
)(int))
399 return _sleep((caddr_t
)chan
, pri
, wmsg
, abstime
, continuation
, (lck_mtx_t
*)0);
403 * Wake up all processes sleeping on chan.
408 thread_wakeup((caddr_t
)chan
);
412 * Wake up the first process sleeping on chan.
414 * Be very sure that the first process is really
415 * the right one to wakeup.
418 wakeup_one(caddr_t chan
)
420 thread_wakeup_one((caddr_t
)chan
);
424 * Compute the priority of a process when running in user mode.
425 * Arrange to reschedule if the resulting priority is better
426 * than that of the current process.
429 resetpriority(struct proc
*p
)
431 (void)task_importance(p
->task
, -p
->p_nice
);
434 struct loadavg averunnable
=
435 { {0, 0, 0}, FSCALE
}; /* load average, of runnable procs */
437 * Constants for averages over 1, 5, and 15 minutes
438 * when sampling at 5 second intervals.
440 static fixpt_t cexp
[3] = {
441 (fixpt_t
)(0.9200444146293232 * FSCALE
), /* exp(-1/12) */
442 (fixpt_t
)(0.9834714538216174 * FSCALE
), /* exp(-1/60) */
443 (fixpt_t
)(0.9944598480048967 * FSCALE
), /* exp(-1/180) */
447 compute_averunnable(void *arg
)
449 unsigned int nrun
= *(unsigned int *)arg
;
450 struct loadavg
*avg
= &averunnable
;
453 for (i
= 0; i
< 3; i
++)
454 avg
->ldavg
[i
] = (cexp
[i
] * avg
->ldavg
[i
] +
455 nrun
* FSCALE
* (FSCALE
- cexp
[i
])) >> FSHIFT
;