X-Git-Url: https://git.saurik.com/apple/xnu.git/blobdiff_plain/316670eb35587141e969394ae8537d66b9211e80..008676633c2ad2c325837c2b64915f7ded690a8f:/osfmk/kern/sched.h diff --git a/osfmk/kern/sched.h b/osfmk/kern/sched.h index 8f87afad2..f1225c9d2 100644 --- a/osfmk/kern/sched.h +++ b/osfmk/kern/sched.h @@ -70,17 +70,18 @@ #include #include #include -#include #include #include #include +#include +#include #define NRQS 128 /* 128 levels per run queue */ -#define NRQBM (NRQS / 32) /* number of words per bit map */ #define MAXPRI (NRQS-1) -#define MINPRI IDLEPRI /* lowest legal priority schedulable */ -#define IDLEPRI 0 /* idle thread priority */ +#define MINPRI 0 /* lowest legal priority schedulable */ +#define IDLEPRI MINPRI /* idle thread priority */ +#define NOPRI -1 /* * High-level priority assignments @@ -143,26 +144,27 @@ #define BASEPRI_REALTIME (MAXPRI - (NRQS / 4) + 1) /* 96 */ #define MAXPRI_KERNEL (BASEPRI_REALTIME - 1) /* 95 */ -#define BASEPRI_PREEMPT (MAXPRI_KERNEL - 2) /* 93 */ -#define BASEPRI_KERNEL (MINPRI_KERNEL + 1) /* 81 */ -#define MINPRI_KERNEL (MAXPRI_KERNEL - (NRQS / 8) + 1) /* 80 */ +#define BASEPRI_PREEMPT (MAXPRI_KERNEL - 2) /* 93 */ +#define BASEPRI_KERNEL (MINPRI_KERNEL + 1) /* 81 */ +#define MINPRI_KERNEL (MAXPRI_KERNEL - (NRQS / 8) + 1) /* 80 */ -#define MAXPRI_RESERVED (MINPRI_KERNEL - 1) /* 79 */ -#define MINPRI_RESERVED (MAXPRI_RESERVED - (NRQS / 8) + 1) /* 64 */ +#define MAXPRI_RESERVED (MINPRI_KERNEL - 1) /* 79 */ +#define BASEPRI_GRAPHICS (MAXPRI_RESERVED - 3) /* 76 */ +#define MINPRI_RESERVED (MAXPRI_RESERVED - (NRQS / 8) + 1) /* 64 */ -#define MAXPRI_USER (MINPRI_RESERVED - 1) /* 63 */ +#define MAXPRI_USER (MINPRI_RESERVED - 1) /* 63 */ #define BASEPRI_CONTROL (BASEPRI_DEFAULT + 17) /* 48 */ #define BASEPRI_FOREGROUND (BASEPRI_DEFAULT + 16) /* 47 */ #define BASEPRI_BACKGROUND (BASEPRI_DEFAULT + 15) /* 46 */ +#define BASEPRI_USER_INITIATED (BASEPRI_DEFAULT + 6) /* 37 */ #define BASEPRI_DEFAULT (MAXPRI_USER - (NRQS / 4)) /* 31 */ -#define MAXPRI_THROTTLE (MINPRI + 4) /* 4 */ -#define MINPRI_USER MINPRI /* 0 */ +#define MAXPRI_SUPPRESSED (BASEPRI_DEFAULT - 3) /* 28 */ +#define BASEPRI_UTILITY (BASEPRI_DEFAULT - 11) /* 20 */ +#define MAXPRI_THROTTLE (MINPRI + 4) /* 4 */ +#define MINPRI_USER MINPRI /* 0 */ -#ifdef CONFIG_EMBEDDED -#define DEPRESSPRI MAXPRI_THROTTLE -#else -#define DEPRESSPRI MINPRI /* depress priority */ -#endif +#define DEPRESSPRI MINPRI /* depress priority */ +#define MAXPRI_PROMOTE (MAXPRI_KERNEL) /* ceiling for mutex promotion */ /* Type used for thread->sched_mode and saved_mode */ typedef enum { @@ -170,9 +172,18 @@ typedef enum { TH_MODE_REALTIME, /* time constraints supplied */ TH_MODE_FIXED, /* use fixed priorities, no decay */ TH_MODE_TIMESHARE, /* use timesharing algorithm */ - TH_MODE_FAIRSHARE /* use fair-share scheduling */ } sched_mode_t; +/* Buckets used for load calculation */ +typedef enum { + TH_BUCKET_RUN = 0, /* All runnable threads */ + TH_BUCKET_FIXPRI, /* Fixed-priority */ + TH_BUCKET_SHARE_FG, /* Timeshare thread above BASEPRI_UTILITY */ + TH_BUCKET_SHARE_UT, /* Timeshare thread between BASEPRI_UTILITY and MAXPRI_THROTTLE */ + TH_BUCKET_SHARE_BG, /* Timeshare thread between MAXPRI_THROTTLE and MINPRI */ + TH_BUCKET_MAX, +} sched_bucket_t; + /* * Macro to check for invalid priorities. */ @@ -183,11 +194,11 @@ struct runq_stats { uint64_t last_change_timestamp; }; -#if defined(CONFIG_SCHED_TRADITIONAL) || defined(CONFIG_SCHED_PROTO) || defined(CONFIG_SCHED_FIXEDPRIORITY) +#if defined(CONFIG_SCHED_TIMESHARE_CORE) || defined(CONFIG_SCHED_PROTO) struct run_queue { int highq; /* highest runnable queue */ - int bitmap[NRQBM]; /* run queue bitmap array */ + bitmap_t bitmap[BITMAP_LEN(NRQS)]; /* run queue bitmap array */ int count; /* # of threads total */ int urgency; /* level of preemption urgency */ queue_head_t queues[NRQS]; /* one for each priority */ @@ -195,7 +206,21 @@ struct run_queue { struct runq_stats runq_stats; }; -#endif /* defined(CONFIG_SCHED_TRADITIONAL) || defined(CONFIG_SCHED_PROTO) || defined(CONFIG_SCHED_FIXEDPRIORITY) */ +inline static void +rq_bitmap_set(bitmap_t *map, u_int n) +{ + assert(n < NRQS); + bitmap_set(map, n); +} + +inline static void +rq_bitmap_clear(bitmap_t *map, u_int n) +{ + assert(n < NRQS); + bitmap_clear(map, n); +} + +#endif /* defined(CONFIG_SCHED_TIMESHARE_CORE) || defined(CONFIG_SCHED_PROTO) */ struct rt_queue { int count; /* # of threads total */ @@ -204,15 +229,6 @@ struct rt_queue { struct runq_stats runq_stats; }; -#if defined(CONFIG_SCHED_TRADITIONAL) || defined(CONFIG_SCHED_PROTO) || defined(CONFIG_SCHED_FIXEDPRIORITY) -struct fairshare_queue { - int count; /* # of threads total */ - queue_head_t queue; /* all runnable threads demoted to fairshare scheduling */ - - struct runq_stats runq_stats; -}; -#endif - #if defined(CONFIG_SCHED_GRRR_CORE) /* @@ -259,10 +275,15 @@ struct grrr_run_queue { #endif /* defined(CONFIG_SCHED_GRRR_CORE) */ -#define first_timeslice(processor) ((processor)->timeslice > 0) - extern struct rt_queue rt_runq; +#if defined(CONFIG_SCHED_MULTIQ) +sched_group_t sched_group_create(void); +void sched_group_destroy(sched_group_t sched_group); +#endif /* defined(CONFIG_SCHED_MULTIQ) */ + + + /* * Scheduler routines. */ @@ -273,12 +294,13 @@ extern void thread_quantum_expire( timer_call_param_t thread); /* Context switch check for current processor */ -extern ast_t csw_check(processor_t processor); +extern ast_t csw_check(processor_t processor, + ast_t check_reason); -#if defined(CONFIG_SCHED_TRADITIONAL) +#if defined(CONFIG_SCHED_TIMESHARE_CORE) extern uint32_t std_quantum, min_std_quantum; extern uint32_t std_quantum_us; -#endif +#endif /* CONFIG_SCHED_TIMESHARE_CORE */ extern uint32_t thread_depress_time; extern uint32_t default_timeshare_computation; @@ -289,22 +311,25 @@ extern uint32_t max_rt_quantum, min_rt_quantum; extern int default_preemption_rate; extern int default_bg_preemption_rate; -#if defined(CONFIG_SCHED_TRADITIONAL) +#if defined(CONFIG_SCHED_TIMESHARE_CORE) /* - * Age usage (1 << SCHED_TICK_SHIFT) times per second. + * Age usage at approximately (1 << SCHED_TICK_SHIFT) times per second + * Aging may be deferred during periods where all processors are idle + * and cumulatively applied during periods of activity. */ #define SCHED_TICK_SHIFT 3 +#define SCHED_TICK_MAX_DELTA (8) extern unsigned sched_tick; extern uint32_t sched_tick_interval; -#endif /* CONFIG_SCHED_TRADITIONAL */ +#endif /* CONFIG_SCHED_TIMESHARE_CORE */ extern uint64_t sched_one_second_interval; /* Periodic computation of various averages */ -extern void compute_averages(void); +extern void compute_averages(uint64_t); extern void compute_averunnable( void *nrun); @@ -315,9 +340,6 @@ extern void compute_stack_target( extern void compute_memory_pressure( void *arg); -extern void compute_zone_gc_throttle( - void *arg); - extern void compute_pageout_gc_throttle( void *arg); @@ -328,16 +350,21 @@ extern void compute_pmap_gc_throttle( * Conversion factor from usage * to priority. */ -#if defined(CONFIG_SCHED_TRADITIONAL) -extern uint32_t sched_pri_shift; +#if defined(CONFIG_SCHED_TIMESHARE_CORE) + +#define MAX_LOAD (NRQS - 1) +extern uint32_t sched_pri_shifts[TH_BUCKET_MAX]; extern uint32_t sched_fixed_shift; extern int8_t sched_load_shifts[NRQS]; -#endif +extern uint32_t sched_decay_usage_age_factor; +void sched_timeshare_consider_maintenance(uint64_t ctime); +#endif /* CONFIG_SCHED_TIMESHARE_CORE */ + +void sched_consider_recommended_cores(uint64_t ctime, thread_t thread); extern int32_t sched_poll_yield_shift; extern uint64_t sched_safe_duration; -extern uint32_t sched_run_count, sched_share_count; extern uint32_t sched_load_average, sched_mach_factor; extern uint32_t avenrun[3], mach_factor[3]; @@ -345,25 +372,10 @@ extern uint32_t avenrun[3], mach_factor[3]; extern uint64_t max_unsafe_computation; extern uint64_t max_poll_computation; -#define sched_run_incr() \ -MACRO_BEGIN \ - hw_atomic_add(&sched_run_count, 1); \ -MACRO_END - -#define sched_run_decr() \ -MACRO_BEGIN \ - hw_atomic_sub(&sched_run_count, 1); \ -MACRO_END +extern volatile uint32_t sched_run_buckets[TH_BUCKET_MAX]; -#define sched_share_incr() \ -MACRO_BEGIN \ - (void)hw_atomic_add(&sched_share_count, 1); \ -MACRO_END - -#define sched_share_decr() \ -MACRO_BEGIN \ - (void)hw_atomic_sub(&sched_share_count, 1); \ -MACRO_END +extern uint32_t sched_run_incr(thread_t thread); +extern uint32_t sched_run_decr(thread_t thread); /* * thread_timer_delta macro takes care of both thread timers.