+ thread_recompute_sched_pri(thread, FALSE);
+}
+
+/*
+ * thread_recompute_sched_pri:
+ *
+ * Reset the scheduled priority of the thread
+ * according to its base priority if the
+ * thread has not been promoted or depressed.
+ *
+ * This is the standard way to push base_pri changes into sched_pri,
+ * or to recalculate the appropriate sched_pri after clearing
+ * a promotion or depression.
+ *
+ * Called at splsched with the thread locked.
+ */
+void
+thread_recompute_sched_pri(
+ thread_t thread,
+ boolean_t override_depress)
+{
+ int priority;
+
+ if (thread->sched_mode == TH_MODE_TIMESHARE)
+ priority = SCHED(compute_timeshare_priority)(thread);
+ else
+ priority = thread->base_pri;
+
+ if ((!(thread->sched_flags & TH_SFLAG_PROMOTED_MASK) || (priority > thread->sched_pri)) &&
+ (!(thread->sched_flags & TH_SFLAG_DEPRESSED_MASK) || override_depress)) {
+ set_sched_pri(thread, priority);
+ }
+}
+
+void
+sched_default_quantum_expire(thread_t thread __unused)
+{
+ /*
+ * No special behavior when a timeshare, fixed, or realtime thread
+ * uses up its entire quantum
+ */
+}
+
+#if defined(CONFIG_SCHED_TIMESHARE_CORE)
+
+/*
+ * lightweight_update_priority:
+ *
+ * Update the scheduled priority for
+ * a timesharing thread.
+ *
+ * Only for use on the current thread.
+ *
+ * Called with the thread locked.
+ */
+void
+lightweight_update_priority(thread_t thread)
+{
+ assert(thread->runq == PROCESSOR_NULL);
+ assert(thread == current_thread());
+
+ if (thread->sched_mode == TH_MODE_TIMESHARE) {
+ int priority;
+ uint32_t delta;
+
+ thread_timer_delta(thread, delta);
+
+ /*
+ * Accumulate timesharing usage only
+ * during contention for processor
+ * resources.
+ */
+ if (thread->pri_shift < INT8_MAX)
+ thread->sched_usage += delta;
+
+ thread->cpu_delta += delta;
+
+ priority = sched_compute_timeshare_priority(thread);
+
+ /*
+ * Adjust the scheduled priority like thread_recompute_sched_pri,
+ * except with the benefit of knowing the thread is on this core.
+ */
+ if ((!(thread->sched_flags & TH_SFLAG_PROMOTED_MASK) || (priority > thread->sched_pri)) &&
+ (!(thread->sched_flags & TH_SFLAG_DEPRESSED_MASK)) &&
+ priority != thread->sched_pri) {
+
+ thread->sched_pri = priority;
+
+ KERNEL_DEBUG_CONSTANT(MACHDBG_CODE(DBG_MACH_SCHED, MACH_SCHED_CHANGE_PRIORITY),
+ (uintptr_t)thread_tid(thread),
+ thread->base_pri,
+ thread->sched_pri,
+ 0, /* eventually, 'reason' */
+ 0);
+ }
+ }