]> git.saurik.com Git - apple/xnu.git/blobdiff - osfmk/kern/sched.h
xnu-2782.40.9.tar.gz
[apple/xnu.git] / osfmk / kern / sched.h
index 10f7b46805dd50baf9225165dd3963d2084d4bd0..980891362d80efa735df85951ddc12011ea2d059 100644 (file)
 #ifndef        _KERN_SCHED_H_
 #define _KERN_SCHED_H_
 
-#include <stat_time.h>
-
 #include <mach/policy.h>
 #include <kern/kern_types.h>
 #include <kern/queue.h>
-#include <kern/lock.h>
 #include <kern/macro_help.h>
 #include <kern/timer_call.h>
 #include <kern/ast.h>
 #define MINPRI_KERNEL          (MAXPRI_KERNEL - (NRQS / 8) + 1)        /* 80 */
 
 #define MAXPRI_RESERVED                (MINPRI_KERNEL - 1)                                     /* 79 */
+#define BASEPRI_GRAPHICS       (MAXPRI_RESERVED - 3)                           /* 76 */
 #define MINPRI_RESERVED                (MAXPRI_RESERVED - (NRQS / 8) + 1)      /* 64 */
 
 #define MAXPRI_USER                    (MINPRI_RESERVED - 1)                           /* 63 */
 #define BASEPRI_CONTROL                (BASEPRI_DEFAULT + 17)                          /* 48 */
 #define BASEPRI_FOREGROUND     (BASEPRI_DEFAULT + 16)                          /* 47 */
 #define BASEPRI_BACKGROUND     (BASEPRI_DEFAULT + 15)                          /* 46 */
+#define BASEPRI_USER_INITIATED (BASEPRI_DEFAULT +  6)                          /* 37 */
 #define BASEPRI_DEFAULT                (MAXPRI_USER - (NRQS / 4))                      /* 31 */
+#define MAXPRI_SUPPRESSED      (BASEPRI_DEFAULT - 3)                           /* 28 */
+#define BASEPRI_UTILITY                (BASEPRI_DEFAULT - 11)                          /* 20 */
 #define MAXPRI_THROTTLE                (MINPRI + 4)                                            /*  4 */
 #define MINPRI_USER                    MINPRI                                                          /*  0 */
 
-#ifdef CONFIG_EMBEDDED
-#define DEPRESSPRI     MAXPRI_THROTTLE
-#else
 #define DEPRESSPRI     MINPRI                  /* depress priority */
-#endif
+#define MAXPRI_PROMOTE         (MAXPRI_KERNEL)         /* ceiling for mutex promotion */
+
+/* Type used for thread->sched_mode and saved_mode */
+typedef enum {
+       TH_MODE_NONE = 0,                                       /* unassigned, usually for saved_mode only */
+       TH_MODE_REALTIME,                                       /* time constraints supplied */
+       TH_MODE_FIXED,                                          /* use fixed priorities, no decay */
+       TH_MODE_TIMESHARE,                                      /* use timesharing algorithm */
+       TH_MODE_FAIRSHARE                                       /* use fair-share scheduling */         
+} sched_mode_t;
 
 /*
  *     Macro to check for invalid priorities.
  */
 #define invalid_pri(pri) ((pri) < MINPRI || (pri) > MAXPRI)
 
+struct runq_stats {
+       uint64_t                                count_sum;
+       uint64_t                                last_change_timestamp;
+};
+
+#if defined(CONFIG_SCHED_TIMESHARE_CORE) || defined(CONFIG_SCHED_PROTO)
+
 struct run_queue {
        int                                     highq;                          /* highest runnable queue */
        int                                     bitmap[NRQBM];          /* run queue bitmap array */
        int                                     count;                          /* # of threads total */
        int                                     urgency;                        /* level of preemption urgency */
        queue_head_t            queues[NRQS];           /* one for each priority */
+
+       struct runq_stats       runq_stats;
 };
 
-typedef struct run_queue       *run_queue_t;
-#define RUN_QUEUE_NULL         ((run_queue_t) 0)
+#endif /* defined(CONFIG_SCHED_TIMESHARE_CORE) || defined(CONFIG_SCHED_PROTO) */
+
+struct rt_queue {
+       int                                     count;                          /* # of threads total */
+       queue_head_t            queue;                          /* all runnable RT threads */
+
+       struct runq_stats       runq_stats;
+};
+
+#if defined(CONFIG_SCHED_FAIRSHARE_CORE)
+struct fairshare_queue {
+       int                                     count;                          /* # of threads total */
+       queue_head_t            queue;                          /* all runnable threads demoted to fairshare scheduling */
+       
+       struct runq_stats       runq_stats;
+};
+#endif /* CONFIG_SCHED_FAIRSHARE_CORE */
+
+#if defined(CONFIG_SCHED_GRRR_CORE)
+
+/*
+ * We map standard Mach priorities to an abstract scale that more properly
+ * indicates how we want processor time allocated under contention.
+ */
+typedef uint8_t        grrr_proportional_priority_t;
+typedef uint8_t grrr_group_index_t;
+
+#define NUM_GRRR_PROPORTIONAL_PRIORITIES       256
+#define MAX_GRRR_PROPORTIONAL_PRIORITY ((grrr_proportional_priority_t)255)
+
+#if 0
+#define NUM_GRRR_GROUPS 8                                      /* log(256) */
+#endif
+
+#define NUM_GRRR_GROUPS 64                                     /* 256/4 */
+
+struct grrr_group {
+       queue_chain_t                   priority_order;                         /* next greatest weight group */
+       grrr_proportional_priority_t            minpriority;
+       grrr_group_index_t              index;
+
+       queue_head_t                    clients;
+       int                                             count;
+       uint32_t                                weight;
+#if 0
+       uint32_t                                deferred_removal_weight;
+#endif
+       uint32_t                                work;
+       thread_t                                current_client;
+};
+
+struct grrr_run_queue {
+       int                                     count;
+       uint32_t                        last_rescale_tick;
+       struct grrr_group       groups[NUM_GRRR_GROUPS];
+       queue_head_t            sorted_group_list;
+       uint32_t                        weight;
+       grrr_group_t            current_group;
+       
+       struct runq_stats   runq_stats;
+};
+
+#endif /* defined(CONFIG_SCHED_GRRR_CORE) */
 
 #define first_timeslice(processor)             ((processor)->timeslice > 0)
 
-#define thread_quantum_init(thread)                                                    \
-MACRO_BEGIN                                                                                                    \
-       (thread)->current_quantum =                                                     \
-               ((thread)->sched_mode & TH_MODE_REALTIME)?                      \
-                       (thread)->realtime.computation: std_quantum;    \
-MACRO_END
+extern struct rt_queue         rt_runq;
 
-extern struct run_queue                rt_runq;
+#if defined(CONFIG_SCHED_MULTIQ)
+sched_group_t   sched_group_create(void);
+void            sched_group_destroy(sched_group_t sched_group);
+
+extern boolean_t sched_groups_enabled;
+
+#endif /* defined(CONFIG_SCHED_MULTIQ) */
 
 /*
  *     Scheduler routines.
  */
 
-/* Remove thread from its run queue */
-extern boolean_t       run_queue_remove(
-                                               thread_t        thread);
-
 /* Handle quantum expiration for an executing thread */
 extern void            thread_quantum_expire(
                                        timer_call_param_t      processor,
                                        timer_call_param_t      thread);
 
 /* Context switch check for current processor */
-extern ast_t   csw_check(processor_t           processor);
+extern ast_t   csw_check(processor_t           processor,
+                                               ast_t                   check_reason);
 
+#if defined(CONFIG_SCHED_TIMESHARE_CORE)
 extern uint32_t        std_quantum, min_std_quantum;
 extern uint32_t        std_quantum_us;
+#endif /* CONFIG_SCHED_TIMESHARE_CORE */
+
+extern uint32_t thread_depress_time;
+extern uint32_t default_timeshare_computation;
+extern uint32_t default_timeshare_constraint;
 
 extern uint32_t        max_rt_quantum, min_rt_quantum;
 
-extern uint32_t        sched_cswtime;
+extern int default_preemption_rate;
+extern int default_bg_preemption_rate;
+
+#if defined(CONFIG_SCHED_TIMESHARE_CORE)
 
 /*
- *     Age usage (1 << SCHED_TICK_SHIFT) times per second.
+ *     Age usage  at approximately (1 << SCHED_TICK_SHIFT) times per second
+ *     Aging may be deferred during periods where all processors are idle
+ *     and cumulatively applied during periods of activity.
  */
 #define SCHED_TICK_SHIFT       3
+#define SCHED_TICK_MAX_DELTA   (8)
 
 extern unsigned                sched_tick;
 extern uint32_t                sched_tick_interval;
 
+#endif /* CONFIG_SCHED_TIMESHARE_CORE */
+
+extern uint64_t                sched_one_second_interval;
+
 /* Periodic computation of various averages */
-extern void            compute_averages(void);
+extern void            compute_averages(uint64_t);
 
 extern void            compute_averunnable(
                                        void                    *nrun);
@@ -236,18 +327,34 @@ extern void               compute_stack_target(
 extern void            compute_memory_pressure(
                                        void                    *arg);
 
+extern void            compute_zone_gc_throttle(
+                                       void                    *arg);
+
+extern void            compute_pageout_gc_throttle(
+                                       void                    *arg);
+
+extern void            compute_pmap_gc_throttle(
+                                       void                    *arg);
+
 /*
  *     Conversion factor from usage
  *     to priority.
  */
+#if defined(CONFIG_SCHED_TIMESHARE_CORE)
 extern uint32_t                sched_pri_shift;
+extern uint32_t                sched_background_pri_shift;
+extern uint32_t                sched_combined_fgbg_pri_shift;
 extern uint32_t                sched_fixed_shift;
 extern int8_t          sched_load_shifts[NRQS];
+extern uint32_t                sched_decay_usage_age_factor;
+extern uint32_t                sched_use_combined_fgbg_decay;
+void sched_traditional_consider_maintenance(uint64_t);
+#endif /* CONFIG_SCHED_TIMESHARE_CORE */
 
 extern int32_t         sched_poll_yield_shift;
-extern uint32_t                sched_safe_duration;
+extern uint64_t                sched_safe_duration;
 
-extern uint32_t                sched_run_count, sched_share_count;
+extern uint32_t                sched_run_count, sched_share_count, sched_background_count;
 extern uint32_t                sched_load_average, sched_mach_factor;
 
 extern uint32_t                avenrun[3], mach_factor[3];
@@ -255,26 +362,50 @@ extern uint32_t           avenrun[3], mach_factor[3];
 extern uint64_t                max_unsafe_computation;
 extern uint64_t                max_poll_computation;
 
-#define sched_run_incr()                       \
-MACRO_BEGIN                                                                                                    \
-       machine_run_count(hw_atomic_add(&sched_run_count, 1));  \
+/* TH_RUN & !TH_IDLE controls whether a thread has a run count */
+#define sched_run_incr(th)                                      \
+       hw_atomic_add(&sched_run_count, 1)                      \
+
+#define sched_run_decr(th)                                      \
+       hw_atomic_sub(&sched_run_count, 1)                      \
+
+#if MACH_ASSERT
+extern void sched_share_incr(thread_t thread);
+extern void sched_share_decr(thread_t thread);
+extern void sched_background_incr(thread_t thread);
+extern void sched_background_decr(thread_t thread);
+
+extern void assert_thread_sched_count(thread_t thread);
+
+#else /* MACH_ASSERT */
+/* sched_mode == TH_MODE_TIMESHARE controls whether a thread has a timeshare count when it has a run count */
+#define sched_share_incr(th)                            \
+MACRO_BEGIN                                             \
+       (void)hw_atomic_add(&sched_share_count, 1);     \
 MACRO_END
 
-#define sched_run_decr()                       \
-MACRO_BEGIN                                                                                                    \
-       machine_run_count(hw_atomic_sub(&sched_run_count, 1));  \
+#define sched_share_decr(th)                            \
+MACRO_BEGIN                                             \
+       (void)hw_atomic_sub(&sched_share_count, 1);     \
 MACRO_END
 
-#define sched_share_incr()                     \
-MACRO_BEGIN                                                                                    \
-       (void)hw_atomic_add(&sched_share_count, 1);             \
+/* TH_SFLAG_THROTTLED controls whether a thread has a background count when it has a run count and a share count */
+#define sched_background_incr(th)                       \
+MACRO_BEGIN                                             \
+       hw_atomic_add(&sched_background_count, 1);      \
 MACRO_END
 
-#define sched_share_decr()                     \
-MACRO_BEGIN                                                                                    \
-       (void)hw_atomic_sub(&sched_share_count, 1);             \
+#define sched_background_decr(th)                       \
+MACRO_BEGIN                                             \
+       hw_atomic_sub(&sched_background_count, 1);      \
 MACRO_END
 
+#define assert_thread_sched_count(th)                   \
+MACRO_BEGIN                                             \
+MACRO_END
+
+#endif /* !MACH_ASSERT */
+
 /*
  *     thread_timer_delta macro takes care of both thread timers.
  */