]> git.saurik.com Git - apple/xnu.git/blobdiff - osfmk/kern/priority.c
xnu-1228.tar.gz
[apple/xnu.git] / osfmk / kern / priority.c
index 23f5e43f0c72dc1f180b8f5cd4482ba362274fd5..4f08fd378bf2d1e017a7febbb7d0fe4cc4e60367 100644 (file)
@@ -1,16 +1,19 @@
 /*
 /*
- * Copyright (c) 2000 Apple Computer, Inc. All rights reserved.
+ * Copyright (c) 2000-2007 Apple Inc. All rights reserved.
  *
  *
- * @APPLE_LICENSE_HEADER_START@
- * 
- * Copyright (c) 1999-2003 Apple Computer, Inc.  All Rights Reserved.
+ * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
  * 
  * This file contains Original Code and/or Modifications of Original Code
  * as defined in and that are subject to the Apple Public Source License
  * Version 2.0 (the 'License'). You may not use this file except in
  * 
  * This file contains Original Code and/or Modifications of Original Code
  * as defined in and that are subject to the Apple Public Source License
  * Version 2.0 (the 'License'). You may not use this file except in
- * compliance with the License. Please obtain a copy of the License at
- * http://www.opensource.apple.com/apsl/ and read it before using this
- * file.
+ * compliance with the License. The rights granted to you under the License
+ * may not be used to create, or enable the creation or redistribution of,
+ * unlawful or unlicensed copies of an Apple operating system, or to
+ * circumvent, violate, or enable the circumvention or violation of, any
+ * terms of an Apple operating system software license agreement.
+ * 
+ * Please obtain a copy of the License at
+ * http://www.opensource.apple.com/apsl/ and read it before using this file.
  * 
  * The Original Code and all software distributed under the License are
  * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
  * 
  * The Original Code and all software distributed under the License are
  * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
@@ -20,7 +23,7 @@
  * Please see the License for the specific language governing rights and
  * limitations under the License.
  * 
  * Please see the License for the specific language governing rights and
  * limitations under the License.
  * 
- * @APPLE_LICENSE_HEADER_END@
+ * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
  */
 /*
  * @OSF_COPYRIGHT@
  */
 /*
  * @OSF_COPYRIGHT@
@@ -60,8 +63,6 @@
  *     Clock primitives.
  */
 
  *     Clock primitives.
  */
 
-#include <cpus.h>
-
 #include <mach/boolean.h>
 #include <mach/kern_return.h>
 #include <mach/machine.h>
 #include <mach/boolean.h>
 #include <mach/kern_return.h>
 #include <mach/machine.h>
@@ -77,6 +78,8 @@
  *     thread_quantum_expire:
  *
  *     Recalculate the quantum and priority for a thread.
  *     thread_quantum_expire:
  *
  *     Recalculate the quantum and priority for a thread.
+ *
+ *     Called at splsched.
  */
 
 void
  */
 
 void
@@ -84,25 +87,22 @@ thread_quantum_expire(
        timer_call_param_t      p0,
        timer_call_param_t      p1)
 {
        timer_call_param_t      p0,
        timer_call_param_t      p1)
 {
-       register processor_t            myprocessor = p0;
-       register thread_t                       thread = p1;
-       spl_t                                           s;
+       processor_t                     processor = p0;
+       thread_t                        thread = p1;
+       ast_t                           preempt;
 
 
-       s = splsched();
        thread_lock(thread);
 
        /*
         *      Check for fail-safe trip.
         */
        if (!(thread->sched_mode & TH_MODE_TIMESHARE)) {
        thread_lock(thread);
 
        /*
         *      Check for fail-safe trip.
         */
        if (!(thread->sched_mode & TH_MODE_TIMESHARE)) {
-               extern uint64_t         max_unsafe_computation;
                uint64_t                        new_computation;
 
                uint64_t                        new_computation;
 
-               new_computation = myprocessor->quantum_end;
+               new_computation = processor->quantum_end;
                new_computation -= thread->computation_epoch;
                if (new_computation + thread->computation_metered >
                                                                                        max_unsafe_computation) {
                new_computation -= thread->computation_epoch;
                if (new_computation + thread->computation_metered >
                                                                                        max_unsafe_computation) {
-                       extern uint32_t         sched_safe_duration;
 
                        if (thread->sched_mode & TH_MODE_REALTIME) {
                                thread->priority = DEPRESSPRI;
 
                        if (thread->sched_mode & TH_MODE_REALTIME) {
                                thread->priority = DEPRESSPRI;
@@ -111,6 +111,8 @@ thread_quantum_expire(
                                thread->sched_mode &= ~TH_MODE_REALTIME;
                        }
 
                                thread->sched_mode &= ~TH_MODE_REALTIME;
                        }
 
+                       sched_share_incr();
+
                        thread->safe_release = sched_tick + sched_safe_duration;
                        thread->sched_mode |= (TH_MODE_FAILSAFE|TH_MODE_TIMESHARE);
                        thread->sched_mode &= ~TH_MODE_PREEMPT;
                        thread->safe_release = sched_tick + sched_safe_duration;
                        thread->sched_mode |= (TH_MODE_FAILSAFE|TH_MODE_TIMESHARE);
                        thread->sched_mode &= ~TH_MODE_PREEMPT;
@@ -124,9 +126,19 @@ thread_quantum_expire(
                update_priority(thread);
        else
        if (thread->sched_mode & TH_MODE_TIMESHARE) {
                update_priority(thread);
        else
        if (thread->sched_mode & TH_MODE_TIMESHARE) {
-               thread_timer_delta(thread);
-               thread->sched_usage += thread->sched_delta;
-               thread->sched_delta = 0;
+               register uint32_t       delta;
+
+               thread_timer_delta(thread, delta);
+
+               /*
+                *      Accumulate timesharing usage only
+                *      during contention for processor
+                *      resources.
+                */
+               if (thread->pri_shift < INT8_MAX)
+                       thread->sched_usage += delta;
+
+               thread->cpu_delta += delta;
 
                /*
                 * Adjust the scheduled priority if
 
                /*
                 * Adjust the scheduled priority if
@@ -138,24 +150,241 @@ thread_quantum_expire(
                        compute_my_priority(thread);
        }
 
                        compute_my_priority(thread);
        }
 
+       processor->current_pri = thread->sched_pri;
+
        /*
         *      This quantum is up, give this thread another.
         */
        /*
         *      This quantum is up, give this thread another.
         */
-       if (first_quantum(myprocessor))
-               myprocessor->slice_quanta--;
+       if (first_timeslice(processor))
+               processor->timeslice--;
 
 
-       thread->current_quantum = (thread->sched_mode & TH_MODE_REALTIME)?
-                                                                       thread->realtime.computation: std_quantum;
-       myprocessor->quantum_end += thread->current_quantum;
-       timer_call_enter1(&myprocessor->quantum_timer,
-                                                       thread, myprocessor->quantum_end);
+       thread_quantum_init(thread);
+       processor->quantum_end += thread->current_quantum;
+       timer_call_enter1(&processor->quantum_timer,
+                                                       thread, processor->quantum_end);
+
+       /*
+        *      Context switch check.
+        */
+       if ((preempt = csw_check(thread, processor)) != AST_NONE)
+               ast_on(preempt);
+       else {
+               processor_set_t         pset = processor->processor_set;
+
+               pset_lock(pset);
+
+               pset_hint_low(pset, processor);
+               pset_hint_high(pset, processor);
+
+               pset_unlock(pset);
+       }
 
        thread_unlock(thread);
 
        thread_unlock(thread);
+}
+
+/*
+ *     Define shifts for simulating (5/8) ** n
+ *
+ *     Shift structures for holding update shifts.  Actual computation
+ *     is  usage = (usage >> shift1) +/- (usage >> abs(shift2))  where the
+ *     +/- is determined by the sign of shift 2.
+ */
+struct shift_data {
+       int     shift1;
+       int     shift2;
+};
+
+#define SCHED_DECAY_TICKS      32
+static struct shift_data       sched_decay_shifts[SCHED_DECAY_TICKS] = {
+       {1,1},{1,3},{1,-3},{2,-7},{3,5},{3,-5},{4,-8},{5,7},
+       {5,-7},{6,-10},{7,10},{7,-9},{8,-11},{9,12},{9,-11},{10,-13},
+       {11,14},{11,-13},{12,-15},{13,17},{13,-15},{14,-17},{15,19},{16,18},
+       {16,-19},{17,22},{18,20},{18,-20},{19,26},{20,22},{20,-22},{21,-27}
+};
+
+/*
+ *     do_priority_computation:
+ *
+ *     Calculate the timesharing priority based upon usage and load.
+ */
+#define do_priority_computation(thread, pri)                                                   \
+       MACRO_BEGIN                                                                                                                     \
+       (pri) = (thread)->priority              /* start with base priority */          \
+           - ((thread)->sched_usage >> (thread)->pri_shift);                           \
+       if ((pri) < MINPRI_USER)                                                                                        \
+               (pri) = MINPRI_USER;                                                                                    \
+       else                                                                                                                            \
+       if ((pri) > MAXPRI_KERNEL)                                                                                      \
+               (pri) = MAXPRI_KERNEL;                                                                                  \
+       MACRO_END
+
+/*
+ *     set_priority:
+ *
+ *     Set the base priority of the thread
+ *     and reset its scheduled priority.
+ *
+ *     Called with the thread locked.
+ */
+void
+set_priority(
+       register thread_t       thread,
+       register int            priority)
+{
+       thread->priority = priority;
+       compute_priority(thread, FALSE);
+}
+
+/*
+ *     compute_priority:
+ *
+ *     Reset the scheduled priority of the thread
+ *     according to its base priority if the
+ *     thread has not been promoted or depressed.
+ *
+ *     Called with the thread locked.
+ */
+void
+compute_priority(
+       register thread_t       thread,
+       boolean_t                       override_depress)
+{
+       register int            priority;
+
+       if (    !(thread->sched_mode & TH_MODE_PROMOTED)                        &&
+                       (!(thread->sched_mode & TH_MODE_ISDEPRESSED)    ||
+                                override_depress                                                       )               ) {
+               if (thread->sched_mode & TH_MODE_TIMESHARE)
+                       do_priority_computation(thread, priority);
+               else
+                       priority = thread->priority;
+
+               set_sched_pri(thread, priority);
+       }
+}
+
+/*
+ *     compute_my_priority:
+ *
+ *     Reset the scheduled priority for
+ *     a timesharing thread.
+ *
+ *     Only for use on the current thread
+ *     if timesharing and not depressed.
+ *
+ *     Called with the thread locked.
+ */
+void
+compute_my_priority(
+       register thread_t       thread)
+{
+       register int            priority;
+
+       do_priority_computation(thread, priority);
+       assert(thread->runq == PROCESSOR_NULL);
+       thread->sched_pri = priority;
+}
+
+/*
+ *     update_priority
+ *
+ *     Perform housekeeping operations driven by scheduler tick.
+ *
+ *     Called with the thread locked.
+ */
+void
+update_priority(
+       register thread_t       thread)
+{
+       register unsigned       ticks;
+       register uint32_t       delta;
+
+       ticks = sched_tick - thread->sched_stamp;
+       assert(ticks != 0);
+       thread->sched_stamp += ticks;
+       thread->pri_shift = sched_pri_shift;
+
+       /*
+        *      Gather cpu usage data.
+        */
+       thread_timer_delta(thread, delta);
+       if (ticks < SCHED_DECAY_TICKS) {
+               register struct shift_data      *shiftp;
+
+               /*
+                *      Accumulate timesharing usage only
+                *      during contention for processor
+                *      resources.
+                */
+               if (thread->pri_shift < INT8_MAX)
+                       thread->sched_usage += delta;
+
+               thread->cpu_usage += delta + thread->cpu_delta;
+               thread->cpu_delta = 0;
+
+               shiftp = &sched_decay_shifts[ticks];
+               if (shiftp->shift2 > 0) {
+                   thread->cpu_usage =
+                                               (thread->cpu_usage >> shiftp->shift1) +
+                                               (thread->cpu_usage >> shiftp->shift2);
+                   thread->sched_usage =
+                                               (thread->sched_usage >> shiftp->shift1) +
+                                               (thread->sched_usage >> shiftp->shift2);
+               }
+               else {
+                   thread->cpu_usage =
+                                               (thread->cpu_usage >> shiftp->shift1) -
+                                               (thread->cpu_usage >> -(shiftp->shift2));
+                   thread->sched_usage =
+                                               (thread->sched_usage >> shiftp->shift1) -
+                                               (thread->sched_usage >> -(shiftp->shift2));
+               }
+       }
+       else {
+               thread->cpu_usage = thread->cpu_delta = 0;
+               thread->sched_usage = 0;
+       }
 
        /*
 
        /*
-        * Check for and schedule ast if needed.
+        *      Check for fail-safe release.
         */
         */
-       ast_check(myprocessor);
+       if (    (thread->sched_mode & TH_MODE_FAILSAFE)         &&
+                       thread->sched_stamp >= thread->safe_release             ) {
+               if (!(thread->safe_mode & TH_MODE_TIMESHARE)) {
+                       if (thread->safe_mode & TH_MODE_REALTIME) {
+                               thread->priority = BASEPRI_RTQUEUES;
+
+                               thread->sched_mode |= TH_MODE_REALTIME;
+                       }
+
+                       thread->sched_mode &= ~TH_MODE_TIMESHARE;
+
+                       if ((thread->state & (TH_RUN|TH_IDLE)) == TH_RUN)
+                               sched_share_decr();
 
 
-       splx(s);
+                       if (!(thread->sched_mode & TH_MODE_ISDEPRESSED))
+                               set_sched_pri(thread, thread->priority);
+               }
+
+               thread->safe_mode = 0;
+               thread->sched_mode &= ~TH_MODE_FAILSAFE;
+       }
+
+       /*
+        *      Recompute scheduled priority if appropriate.
+        */
+       if (    (thread->sched_mode & TH_MODE_TIMESHARE)        &&
+                       !(thread->sched_mode & TH_MODE_PROMOTED)        &&
+                       !(thread->sched_mode & TH_MODE_ISDEPRESSED)             ) {
+               register int            new_pri;
+
+               do_priority_computation(thread, new_pri);
+               if (new_pri != thread->sched_pri) {
+                       boolean_t               removed = run_queue_remove(thread);
+
+                       thread->sched_pri = new_pri;
+                       if (removed)
+                               thread_setrun(thread, SCHED_TAILQ);
+               }
+       }
 }
 }