]> git.saurik.com Git - apple/xnu.git/blobdiff - bsd/kern/kern_memorystatus.c
xnu-2050.48.11.tar.gz
[apple/xnu.git] / bsd / kern / kern_memorystatus.c
index 35e9a43a685b85dd66234d4bade9f89cb9a31172..1bf8dd616c128713e27601a50046bec5ba7e5916 100644 (file)
  *
  */
 
-#include <sys/kern_event.h>
-#include <sys/kern_memorystatus.h>
-
 #include <kern/sched_prim.h>
+#include <kern/kalloc.h>
+#include <kern/assert.h>
+#include <kern/debug.h>
 #include <kern/lock.h>
 #include <kern/task.h>
 #include <kern/thread.h>
+#include <kern/host.h>
 #include <libkern/libkern.h>
+#include <mach/mach_time.h>
 #include <mach/task.h>
 #include <mach/task_info.h>
+#include <mach/host_priv.h>
+#include <sys/kern_event.h>
 #include <sys/proc.h>
 #include <sys/signal.h>
 #include <sys/signalvar.h>
 #include <sys/sysctl.h>
+#include <sys/sysproto.h>
 #include <sys/wait.h>
+#include <sys/tree.h>
+#include <sys/priv.h>
+#include <pexpert/pexpert.h>
+
+#if CONFIG_FREEZE
+#include <vm/vm_protos.h>
+#include <vm/vm_map.h>
+#endif
+
+#include <sys/kern_memorystatus.h> 
+
+/* These are very verbose printfs(), enable with
+ * MEMORYSTATUS_DEBUG_LOG
+ */
+#if MEMORYSTATUS_DEBUG_LOG
+#define MEMORYSTATUS_DEBUG(cond, format, ...)      \
+do {                                              \
+       if (cond) { printf(format, ##__VA_ARGS__); } \
+} while(0)
+#else
+#define MEMORYSTATUS_DEBUG(cond, format, ...)
+#endif
+
+/* General memorystatus stuff */
+
+static void memorystatus_add_node(memorystatus_node *node);
+static void memorystatus_remove_node(memorystatus_node *node);
+static memorystatus_node *memorystatus_get_node(pid_t pid);
+static void memorystatus_release_node(memorystatus_node *node);
+
+int memorystatus_wakeup = 0;
+
+static void memorystatus_thread(void *param __unused, wait_result_t wr __unused);
+
+static memorystatus_node *next_memorystatus_node = NULL;
+
+static int memorystatus_list_count = 0;
+
+static lck_mtx_t * memorystatus_list_mlock;
+static lck_attr_t * memorystatus_lck_attr;
+static lck_grp_t * memorystatus_lck_grp;
+static lck_grp_attr_t * memorystatus_lck_grp_attr;
+
+static TAILQ_HEAD(memorystatus_list_head, memorystatus_node) memorystatus_list;
+
+static uint64_t memorystatus_idle_delay_time = 0;
+
+static unsigned int memorystatus_dirty_count = 0;
+
+extern void proc_dirty_start(struct proc *p);
+extern void proc_dirty_end(struct proc *p);
+
+/* Jetsam */
+
+#if CONFIG_JETSAM
+
+extern unsigned int    vm_page_free_count;
+extern unsigned int    vm_page_active_count;
+extern unsigned int    vm_page_inactive_count;
+extern unsigned int    vm_page_throttled_count;
+extern unsigned int    vm_page_purgeable_count;
+extern unsigned int    vm_page_wire_count;
+
+static lck_mtx_t * exit_list_mlock;
+
+static TAILQ_HEAD(exit_list_head, memorystatus_node) exit_list;
+
+static unsigned int memorystatus_kev_failure_count = 0;
+
+/* Counted in pages... */
+unsigned int memorystatus_delta = 0;
+
+unsigned int memorystatus_available_pages = (unsigned int)-1;
+unsigned int memorystatus_available_pages_critical = 0;
+unsigned int memorystatus_available_pages_highwater = 0;
+
+/* ...with the exception of the legacy level in percent. */
+unsigned int memorystatus_level = 0;
+
+SYSCTL_UINT(_kern, OID_AUTO, memorystatus_kev_failure_count, CTLFLAG_RD, &memorystatus_kev_failure_count, 0, "");
+SYSCTL_INT(_kern, OID_AUTO, memorystatus_level, CTLFLAG_RD, &memorystatus_level, 0, "");
+
+unsigned int memorystatus_jetsam_policy = kPolicyDefault;
+
+unsigned int memorystatus_jetsam_policy_offset_pages_more_free = 0;
+#if DEVELOPMENT || DEBUG
+unsigned int memorystatus_jetsam_policy_offset_pages_diagnostic = 0;
+#endif
+
+static memorystatus_jetsam_snapshot_t memorystatus_jetsam_snapshot;
+#define memorystatus_jetsam_snapshot_list memorystatus_jetsam_snapshot.entries
+
+static int memorystatus_jetsam_snapshot_list_count = 0;
+
+int memorystatus_jetsam_wakeup = 0;
+unsigned int memorystatus_jetsam_running = 1;
+
+static uint32_t memorystatus_task_page_count(task_t task);
+
+static void memorystatus_move_node_to_exit_list(memorystatus_node *node);
+
+static void memorystatus_update_levels_locked(void);
+
+static void memorystatus_jetsam_thread_block(void);
+static void memorystatus_jetsam_thread(void *param __unused, wait_result_t wr __unused);
+
+static int memorystatus_send_note(int event_code, void *data, size_t data_length);
+
+static uint32_t memorystatus_build_flags_from_state(uint32_t state);
+
+/* VM pressure */
+
+#if VM_PRESSURE_EVENTS
+
+typedef enum vm_pressure_level {
+        kVMPressureNormal   = 0,
+        kVMPressureWarning  = 1,
+        kVMPressureUrgent   = 2,
+        kVMPressureCritical = 3,
+} vm_pressure_level_t;
+
+static vm_pressure_level_t memorystatus_vm_pressure_level = kVMPressureNormal;
+
+unsigned int memorystatus_available_pages_pressure = 0;
+
+static inline boolean_t memorystatus_get_pressure_locked(void);
+static void memorystatus_check_pressure_reset(void);
+
+#endif /* VM_PRESSURE_EVENTS */
+
+#endif /* CONFIG_JETSAM */
+
+/* Freeze */
+
+#if CONFIG_FREEZE
+
+static unsigned int memorystatus_suspended_resident_count = 0;
+static unsigned int memorystatus_suspended_count = 0;
+
+boolean_t memorystatus_freeze_enabled = FALSE;
+int memorystatus_freeze_wakeup = 0;
+
+static inline boolean_t memorystatus_can_freeze_processes(void);
+static boolean_t memorystatus_can_freeze(boolean_t *memorystatus_freeze_swap_low);
+
+static void memorystatus_freeze_thread(void *param __unused, wait_result_t wr __unused);
+
+/* Thresholds */
+static unsigned int memorystatus_freeze_threshold = 0;
+
+static unsigned int memorystatus_freeze_pages_min = FREEZE_PAGES_MIN;
+static unsigned int memorystatus_freeze_pages_max = FREEZE_PAGES_MAX;
+
+static unsigned int memorystatus_frozen_count = 0;
+
+static unsigned int memorystatus_freeze_suspended_threshold = FREEZE_SUSPENDED_THRESHOLD_DEFAULT;
+
+/* Stats */
+static uint64_t memorystatus_freeze_count = 0;
+static uint64_t memorystatus_freeze_pageouts = 0;
+
+/* Throttling */
+static throttle_interval_t throttle_intervals[] = {
+       {      60,  8, 0, 0, { 0, 0 }, FALSE }, /* 1 hour intermediate interval, 8x burst */
+       { 24 * 60,  1, 0, 0, { 0, 0 }, FALSE }, /* 24 hour long interval, no burst */
+};
+
+static uint64_t memorystatus_freeze_throttle_count = 0;
+
+#endif /* CONFIG_FREEZE */
+
+#if CONFIG_JETSAM
+
+/* Debug */
+
+#if DEVELOPMENT || DEBUG
+
+SYSCTL_UINT(_kern, OID_AUTO, memorystatus_available_pages, CTLFLAG_RD, &memorystatus_available_pages, 0, "");
+SYSCTL_UINT(_kern, OID_AUTO, memorystatus_available_pages_critical, CTLFLAG_RW, &memorystatus_available_pages_critical, 0, "");
+SYSCTL_UINT(_kern, OID_AUTO, memorystatus_available_pages_highwater, CTLFLAG_RW, &memorystatus_available_pages_highwater, 0, "");
+#if VM_PRESSURE_EVENTS
+SYSCTL_UINT(_kern, OID_AUTO, memorystatus_available_pages_pressure, CTLFLAG_RW, &memorystatus_available_pages_pressure, 0, "");
+#endif /* VM_PRESSURE_EVENTS */
+
+/* Diagnostic code */
+enum {
+       kJetsamDiagnosticModeNone =              0, 
+       kJetsamDiagnosticModeAll  =              1,
+       kJetsamDiagnosticModeStopAtFirstActive = 2,
+       kJetsamDiagnosticModeCount
+} jetsam_diagnostic_mode = kJetsamDiagnosticModeNone;
+
+static int jetsam_diagnostic_suspended_one_active_proc = 0;
+
+static int
+sysctl_jetsam_diagnostic_mode SYSCTL_HANDLER_ARGS
+{
+#pragma unused(arg1, arg2)
+
+       const char *diagnosticStrings[] = {
+               "jetsam: diagnostic mode: resetting critical level.",
+               "jetsam: diagnostic mode: will examine all processes",
+               "jetsam: diagnostic mode: will stop at first active process"                
+       };
+        
+       int error, val = jetsam_diagnostic_mode;
+       boolean_t changed = FALSE;
+
+       error = sysctl_handle_int(oidp, &val, 0, req);
+       if (error || !req->newptr)
+               return (error);
+       if ((val < 0) || (val >= kJetsamDiagnosticModeCount)) {
+               printf("jetsam: diagnostic mode: invalid value - %d\n", val);
+               return EINVAL;
+       }
+       
+       lck_mtx_lock(memorystatus_list_mlock);
+       
+       if ((unsigned int) val != jetsam_diagnostic_mode) {
+               jetsam_diagnostic_mode = val;
+
+               memorystatus_jetsam_policy &= ~kPolicyDiagnoseActive;
+                
+               switch (jetsam_diagnostic_mode) {
+               case kJetsamDiagnosticModeNone:
+                       /* Already cleared */
+                       break;
+               case kJetsamDiagnosticModeAll:
+                       memorystatus_jetsam_policy |= kPolicyDiagnoseAll;
+                       break;
+               case kJetsamDiagnosticModeStopAtFirstActive:
+                       memorystatus_jetsam_policy |= kPolicyDiagnoseFirst;
+                       break;
+               default:
+                       /* Already validated */
+                       break;
+               }
+               
+               memorystatus_update_levels_locked();
+               changed = TRUE;
+       }
+        
+       lck_mtx_unlock(memorystatus_list_mlock);
+       
+       if (changed) {
+               printf("%s\n", diagnosticStrings[val]);
+       }
+       
+       return (0);
+}
+
+SYSCTL_PROC(_debug, OID_AUTO, jetsam_diagnostic_mode, CTLTYPE_INT|CTLFLAG_RW|CTLFLAG_ANYBODY,
+               &jetsam_diagnostic_mode, 0, sysctl_jetsam_diagnostic_mode, "I", "Jetsam Diagnostic Mode");
+
+SYSCTL_UINT(_kern, OID_AUTO, memorystatus_jetsam_policy_offset_pages_more_free, CTLFLAG_RW, &memorystatus_jetsam_policy_offset_pages_more_free, 0, "");
+SYSCTL_UINT(_kern, OID_AUTO, memorystatus_jetsam_policy_offset_pages_diagnostic, CTLFLAG_RW, &memorystatus_jetsam_policy_offset_pages_diagnostic, 0, "");
+
+#if VM_PRESSURE_EVENTS
+
+#include "vm_pressure.h"
+
+static int
+sysctl_memorystatus_vm_pressure_level SYSCTL_HANDLER_ARGS
+{
+#pragma unused(arg1, arg2, oidp)
+       int error = 0;
+
+       error = priv_check_cred(kauth_cred_get(), PRIV_VM_PRESSURE, 0);
+       if (error)
+               return (error);
+
+       return SYSCTL_OUT(req, &memorystatus_vm_pressure_level, sizeof(memorystatus_vm_pressure_level));
+}
+
+SYSCTL_PROC(_kern, OID_AUTO, memorystatus_vm_pressure_level, CTLTYPE_INT|CTLFLAG_RD|CTLFLAG_LOCKED|CTLFLAG_MASKED,
+    0, 0, &sysctl_memorystatus_vm_pressure_level, "I", "");
+
+static int
+sysctl_memorystatus_vm_pressure_send SYSCTL_HANDLER_ARGS
+{
+#pragma unused(arg1, arg2)
+
+       int error, pid = 0;
+
+       error = sysctl_handle_int(oidp, &pid, 0, req);
+       if (error || !req->newptr)
+               return (error);
+
+       if (vm_dispatch_pressure_note_to_pid(pid)) {
+               return 0;
+       }
+
+       return EINVAL;
+}
+
+SYSCTL_PROC(_kern, OID_AUTO, memorystatus_vm_pressure_send, CTLTYPE_INT|CTLFLAG_WR|CTLFLAG_LOCKED|CTLFLAG_MASKED,
+    0, 0, &sysctl_memorystatus_vm_pressure_send, "I", "");
+
+#endif /* VM_PRESSURE_EVENTS */
+
+#endif /* CONFIG_JETSAM */
+
+#if CONFIG_FREEZE
+
+SYSCTL_UINT(_kern, OID_AUTO, memorystatus_freeze_threshold, CTLFLAG_RW, &memorystatus_freeze_threshold, 0, "");
+
+SYSCTL_UINT(_kern, OID_AUTO, memorystatus_freeze_pages_min, CTLFLAG_RW, &memorystatus_freeze_pages_min, 0, "");
+SYSCTL_UINT(_kern, OID_AUTO, memorystatus_freeze_pages_max, CTLFLAG_RW, &memorystatus_freeze_pages_max, 0, "");
+
+SYSCTL_QUAD(_kern, OID_AUTO, memorystatus_freeze_count, CTLFLAG_RD, &memorystatus_freeze_count, "");
+SYSCTL_QUAD(_kern, OID_AUTO, memorystatus_freeze_pageouts, CTLFLAG_RD, &memorystatus_freeze_pageouts, "");
+SYSCTL_QUAD(_kern, OID_AUTO, memorystatus_freeze_throttle_count, CTLFLAG_RD, &memorystatus_freeze_throttle_count, "");
+SYSCTL_UINT(_kern, OID_AUTO, memorystatus_freeze_min_processes, CTLFLAG_RW, &memorystatus_freeze_suspended_threshold, 0, "");
+
+boolean_t memorystatus_freeze_throttle_enabled = TRUE;
+SYSCTL_UINT(_kern, OID_AUTO, memorystatus_freeze_throttle_enabled, CTLFLAG_RW, &memorystatus_freeze_throttle_enabled, 0, "");
+
+/* 
+ * Manual trigger of freeze and thaw for dev / debug kernels only.
+ */
+static int
+sysctl_memorystatus_freeze SYSCTL_HANDLER_ARGS
+{
+#pragma unused(arg1, arg2)
+
+       int error, pid = 0;
+       proc_t p;
+
+       error = sysctl_handle_int(oidp, &pid, 0, req);
+       if (error || !req->newptr)
+               return (error);
+
+       p = proc_find(pid);
+       if (p != NULL) {
+               uint32_t purgeable, wired, clean, dirty;
+               boolean_t shared;
+               uint32_t max_pages = MIN(default_pager_swap_pages_free(), memorystatus_freeze_pages_max);
+               task_freeze(p->task, &purgeable, &wired, &clean, &dirty, max_pages, &shared, FALSE);
+               proc_rele(p);
+        return 0;
+       }
+
+       return EINVAL;
+}
+
+SYSCTL_PROC(_kern, OID_AUTO, memorystatus_freeze, CTLTYPE_INT|CTLFLAG_WR|CTLFLAG_LOCKED|CTLFLAG_MASKED,
+    0, 0, &sysctl_memorystatus_freeze, "I", "");
+
+static int
+sysctl_memorystatus_available_pages_thaw SYSCTL_HANDLER_ARGS
+{
+#pragma unused(arg1, arg2)
+
+       int error, pid = 0;
+       proc_t p;
+
+       error = sysctl_handle_int(oidp, &pid, 0, req);
+       if (error || !req->newptr)
+               return (error);
+
+       p = proc_find(pid);
+       if (p != NULL) {
+               task_thaw(p->task);
+               proc_rele(p);
+               return 0;
+       }
+
+       return EINVAL;
+}
+
+SYSCTL_PROC(_kern, OID_AUTO, memorystatus_thaw, CTLTYPE_INT|CTLFLAG_WR|CTLFLAG_LOCKED|CTLFLAG_MASKED,
+    0, 0, &sysctl_memorystatus_available_pages_thaw, "I", "");
+
+#endif /* CONFIG_FREEZE */
+
+#endif /* DEVELOPMENT || DEBUG */
+
+__private_extern__ void
+memorystatus_init(void)
+{
+       thread_t thread = THREAD_NULL;
+       kern_return_t result;
+       
+       memorystatus_lck_attr = lck_attr_alloc_init();
+       memorystatus_lck_grp_attr = lck_grp_attr_alloc_init();
+       memorystatus_lck_grp = lck_grp_alloc_init("memorystatus",  memorystatus_lck_grp_attr);
+       memorystatus_list_mlock = lck_mtx_alloc_init(memorystatus_lck_grp, memorystatus_lck_attr);
+       TAILQ_INIT(&memorystatus_list);
+
+#if CONFIG_JETSAM
+       exit_list_mlock = lck_mtx_alloc_init(memorystatus_lck_grp, memorystatus_lck_attr);
+       TAILQ_INIT(&exit_list);
+       
+       memorystatus_delta = DELTA_PERCENT * atop_64(max_mem) / 100;
+#endif
+
+#if CONFIG_FREEZE
+       memorystatus_freeze_threshold = (FREEZE_PERCENT / DELTA_PERCENT) * memorystatus_delta;
+#endif
+
+       nanoseconds_to_absolutetime((uint64_t)IDLE_EXIT_TIME_SECS * NSEC_PER_SEC, &memorystatus_idle_delay_time);
+
+       result = kernel_thread_start(memorystatus_thread, NULL, &thread);
+       if (result == KERN_SUCCESS) {
+               thread_deallocate(thread);
+       } else {
+               panic("Could not create memorystatus_thread");
+       }
+
+#if CONFIG_JETSAM
+       memorystatus_jetsam_policy_offset_pages_more_free = (POLICY_MORE_FREE_OFFSET_PERCENT / DELTA_PERCENT) * memorystatus_delta;
+#if DEVELOPMENT || DEBUG
+       memorystatus_jetsam_policy_offset_pages_diagnostic = (POLICY_DIAGNOSTIC_OFFSET_PERCENT / DELTA_PERCENT) * memorystatus_delta;
+#endif
+
+       /* No contention at this point */
+       memorystatus_update_levels_locked();
+       
+       result = kernel_thread_start(memorystatus_jetsam_thread, NULL, &thread);
+       if (result == KERN_SUCCESS) {
+               thread_deallocate(thread);
+       } else {
+               panic("Could not create memorystatus_jetsam_thread");
+       }
+#endif
+}
+
+/*
+ * Node manipulation
+ */
+
+static void
+memorystatus_add_node(memorystatus_node *new_node)
+{
+       memorystatus_node *node;
+
+       /* Make sure we're called with the list lock held */
+       lck_mtx_assert(memorystatus_list_mlock, LCK_MTX_ASSERT_OWNED);
+
+       TAILQ_FOREACH(node, &memorystatus_list, link) {
+               if (node->priority <= new_node->priority) {
+                       break;
+               }
+       }
+
+       if (node) {
+               TAILQ_INSERT_BEFORE(node, new_node, link);
+       } else {
+               TAILQ_INSERT_TAIL(&memorystatus_list, new_node, link);
+       }
+
+       next_memorystatus_node = TAILQ_FIRST(&memorystatus_list);
+
+       memorystatus_list_count++;
+}
+
+static void
+memorystatus_remove_node(memorystatus_node *node) 
+{
+       /* Make sure we're called with the list lock held */
+       lck_mtx_assert(memorystatus_list_mlock, LCK_MTX_ASSERT_OWNED);
+
+       TAILQ_REMOVE(&memorystatus_list, node, link);
+       next_memorystatus_node = TAILQ_FIRST(&memorystatus_list);
+
+#if CONFIG_FREEZE    
+       if (node->state & (kProcessFrozen)) {
+               memorystatus_frozen_count--;
+       }
+
+       if (node->state & kProcessSuspended) {
+               memorystatus_suspended_resident_count -= node->resident_pages;
+               memorystatus_suspended_count--;
+       }
+#endif
+
+       memorystatus_list_count--;
+}
+
+/* Returns with the lock taken if found */
+static memorystatus_node *
+memorystatus_get_node(pid_t pid) 
+{
+       memorystatus_node *node;
+
+       lck_mtx_lock(memorystatus_list_mlock);
+
+       TAILQ_FOREACH(node, &memorystatus_list, link) {
+               if (node->pid == pid) {
+                       break;
+               }
+       }
+
+       if (!node) {
+               lck_mtx_unlock(memorystatus_list_mlock);                
+       }
+
+       return node;
+}
+
+static void
+memorystatus_release_node(memorystatus_node *node) 
+{
+#pragma unused(node)
+       lck_mtx_unlock(memorystatus_list_mlock);        
+}
+
+/* 
+ * List manipulation
+ */
+kern_return_t 
+memorystatus_list_add(pid_t pid, int priority, int high_water_mark)
+{
+
+#if !CONFIG_JETSAM
+#pragma unused(high_water_mark)
+#endif
+
+       memorystatus_node *new_node;
+
+       new_node = (memorystatus_node*)kalloc(sizeof(memorystatus_node));
+       if (!new_node) {
+               assert(FALSE);
+       }
+       memset(new_node, 0, sizeof(memorystatus_node));
+    
+       MEMORYSTATUS_DEBUG(1, "memorystatus_list_add: adding process %d with priority %d, high water mark %d.\n", pid, priority, high_water_mark);
+    
+       new_node->pid = pid;
+       new_node->priority = priority;
+#if CONFIG_JETSAM
+       new_node->hiwat_pages = high_water_mark;
+#endif    
+
+       lck_mtx_lock(memorystatus_list_mlock);
+    
+       memorystatus_add_node(new_node);
+        
+       lck_mtx_unlock(memorystatus_list_mlock);
+       
+       return KERN_SUCCESS;
+}
+
+kern_return_t
+memorystatus_list_change(boolean_t effective, pid_t pid, int priority, int state_flags, int high_water_mark)
+{
+
+#if !CONFIG_JETSAM
+#pragma unused(high_water_mark)
+#endif
+       
+       kern_return_t ret;
+       memorystatus_node *node, *search;
+
+       MEMORYSTATUS_DEBUG(1, "memorystatus_list_change: changing process %d to priority %d with flags %d\n", pid, priority, state_flags);
+
+       lck_mtx_lock(memorystatus_list_mlock);
+
+       TAILQ_FOREACH(node, &memorystatus_list, link) {
+               if (node->pid == pid) {
+                       break;
+               }
+       }
+    
+       if (!node) {
+               ret = KERN_FAILURE;
+               goto out;             
+       }
+
+       if (effective && (node->state & kProcessPriorityUpdated)) {
+               MEMORYSTATUS_DEBUG(1, "memorystatus_list_change: effective change specified for pid %d, but change already occurred.\n", pid);
+               ret = KERN_FAILURE;
+               goto out;             
+       }
+
+       node->state |= kProcessPriorityUpdated;
+       if (state_flags != -1) {
+               node->state &= ~(kProcessActive|kProcessForeground);
+               if (state_flags & kMemorystatusFlagsFrontmost) {
+                       node->state |= kProcessForeground;
+               }
+               if (state_flags & kMemorystatusFlagsActive) {
+                       node->state |= kProcessActive;
+               }
+       }
+
+#if CONFIG_JETSAM        
+       if (high_water_mark != -1) {
+               node->hiwat_pages = high_water_mark;
+       }
+#endif
+
+       if (node->priority == priority) {
+               /* Priority unchanged */
+               MEMORYSTATUS_DEBUG(1, "memorystatus_list_change: same priority set for pid %d\n", pid);
+               ret = KERN_SUCCESS;
+               goto out;
+       }
+
+       if (node->priority < priority) {
+               /* Higher priority value (ie less important) - search backwards */
+               search = TAILQ_PREV(node, memorystatus_list_head, link);
+               TAILQ_REMOVE(&memorystatus_list, node, link);
+
+               node->priority = priority;
+               while (search && (search->priority <= node->priority)) {
+                       search = TAILQ_PREV(search, memorystatus_list_head, link);
+               }
+               if (search) {
+                       TAILQ_INSERT_AFTER(&memorystatus_list, search, node, link);
+               } else {
+                       TAILQ_INSERT_HEAD(&memorystatus_list, node, link);
+               }
+       } else {
+               /* Lower priority value (ie more important) - search forwards */
+               search = TAILQ_NEXT(node, link);
+               TAILQ_REMOVE(&memorystatus_list, node, link);
+
+               node->priority = priority;
+               while (search && (search->priority >= node->priority)) {
+                       search = TAILQ_NEXT(search, link);
+               }
+               if (search) {
+                       TAILQ_INSERT_BEFORE(search, node, link);
+               } else {
+                       TAILQ_INSERT_TAIL(&memorystatus_list, node, link);
+               }
+       }
+
+       next_memorystatus_node = TAILQ_FIRST(&memorystatus_list);
+       ret = KERN_SUCCESS;
+
+out:
+       lck_mtx_unlock(memorystatus_list_mlock);
+       return ret;
+}
+
+kern_return_t memorystatus_list_remove(pid_t pid)
+{
+       kern_return_t ret;
+       memorystatus_node *node = NULL;
+
+       MEMORYSTATUS_DEBUG(1, "memorystatus_list_remove: removing process %d\n", pid);
+
+#if CONFIG_JETSAM
+       /* Did we mark this as a exited process? */
+       lck_mtx_lock(exit_list_mlock);
+
+       TAILQ_FOREACH(node, &exit_list, link) {
+               if (node->pid == pid) {
+                       /* We did, so remove it from the list. The stats were updated when the queues were shifted. */
+                       TAILQ_REMOVE(&exit_list, node, link);
+                       break;
+               }
+       }
+
+       lck_mtx_unlock(exit_list_mlock);
+#endif
+
+       /* If not, search the main list */
+       if (!node) {
+               lck_mtx_lock(memorystatus_list_mlock);
+
+               TAILQ_FOREACH(node, &memorystatus_list, link) {
+                       if (node->pid == pid) {
+                               /* Remove from the list, and update accounting accordingly */
+                               memorystatus_remove_node(node);
+                               break;
+                       }
+               }
+
+               lck_mtx_unlock(memorystatus_list_mlock);
+       }
+
+       if (node) {
+               kfree(node, sizeof(memorystatus_node));
+               ret = KERN_SUCCESS; 
+       } else {
+               ret = KERN_FAILURE;
+       }
+
+       return ret;
+}
+
+kern_return_t 
+memorystatus_on_track_dirty(int pid, boolean_t track)
+{
+       kern_return_t ret = KERN_FAILURE;
+       memorystatus_node *node;
+       
+       node = memorystatus_get_node((pid_t)pid);
+       if (!node) {
+               return KERN_FAILURE;
+       }
+       
+       if (track & !(node->state & kProcessSupportsIdleExit)) {
+               node->state |= kProcessSupportsIdleExit;
+               node->clean_time = mach_absolute_time() + memorystatus_idle_delay_time;
+               ret = KERN_SUCCESS;
+       } else  if (!track & (node->state & kProcessSupportsIdleExit)) {
+               node->state &= ~kProcessSupportsIdleExit;
+               node->clean_time = 0;
+               ret = KERN_SUCCESS;             
+       }
+       
+       memorystatus_release_node(node);
+               
+       return ret;     
+}
+
+kern_return_t 
+memorystatus_on_dirty(int pid, boolean_t dirty)
+{
+       kern_return_t ret = KERN_FAILURE;
+       memorystatus_node *node;
+       
+       node = memorystatus_get_node((pid_t)pid);
+       if (!node) {
+               return KERN_FAILURE;
+       }
+       
+       if (dirty) {
+               if (!(node->state & kProcessDirty)) {
+                       node->state |= kProcessDirty;
+                       node->clean_time = 0;
+                       memorystatus_dirty_count++;
+                       ret = KERN_SUCCESS;
+               }
+       } else {
+               if (node->state & kProcessDirty) {
+                       node->state &= ~kProcessDirty;
+                       node->clean_time = mach_absolute_time() + memorystatus_idle_delay_time;
+                       memorystatus_dirty_count--;
+                       ret = KERN_SUCCESS;
+               }
+       }
+       
+       memorystatus_release_node(node);
+       
+       return ret;
+}
+
+void 
+memorystatus_on_suspend(int pid)
+{      
+       memorystatus_node *node = memorystatus_get_node((pid_t)pid);
+
+       if (node) {
+#if CONFIG_FREEZE
+               proc_t p;
+
+               p = proc_find(pid);
+               if (p != NULL) {
+                       uint32_t pages = memorystatus_task_page_count(p->task);
+                       proc_rele(p);
+                       node->resident_pages = pages;
+                       memorystatus_suspended_resident_count += pages;
+               }
+               memorystatus_suspended_count++;
+#endif
 
-extern unsigned int    vm_page_free_count;
-extern unsigned int    vm_page_active_count;
-extern unsigned int    vm_page_inactive_count;
-extern unsigned int    vm_page_purgeable_count;
-extern unsigned int    vm_page_wire_count;
+               node->state |= kProcessSuspended;
 
-static void kern_memorystatus_thread(void);
+               memorystatus_release_node(node);
+       }
+}
 
-int kern_memorystatus_wakeup = 0;
-int kern_memorystatus_level = 0;
-int kern_memorystatus_last_level = 0;
-unsigned int kern_memorystatus_kev_failure_count = 0;
-int kern_memorystatus_level_critical = 5;
+void
+memorystatus_on_resume(int pid)
+{      
+       memorystatus_node *node = memorystatus_get_node((pid_t)pid);
 
-static struct {
-       jetsam_kernel_stats_t stats;
-       size_t entry_count;
-       jetsam_snapshot_entry_t entries[kMaxSnapshotEntries];
-} jetsam_snapshot;
+       if (node) {
+#if CONFIG_FREEZE
+               boolean_t frozen = (node->state & kProcessFrozen);
+               if (node->state & (kProcessFrozen)) {
+                       memorystatus_frozen_count--;
+               }
+               memorystatus_suspended_resident_count -= node->resident_pages;
+               memorystatus_suspended_count--;
+#endif
 
-static jetsam_priority_entry_t jetsam_priority_list[kMaxPriorityEntries];
-#define jetsam_snapshot_list jetsam_snapshot.entries
+               node->state &= ~(kProcessSuspended | kProcessFrozen | kProcessIgnored);
 
-static int jetsam_priority_list_index = 0;
-static int jetsam_priority_list_count = 0;
-static int jetsam_snapshot_list_count = 0;
+               memorystatus_release_node(node);
 
-static lck_mtx_t * jetsam_list_mlock;
-static lck_attr_t * jetsam_lck_attr;
-static lck_grp_t * jetsam_lck_grp;
-static lck_grp_attr_t * jetsam_lck_grp_attr;
+#if CONFIG_FREEZE
+               if (frozen) {
+                       memorystatus_freeze_entry_t data = { pid, kMemorystatusFlagsThawed, 0 };
+                       memorystatus_send_note(kMemorystatusFreezeNote, &data, sizeof(data));
+               }
+#endif
+       }
+}
 
-SYSCTL_INT(_kern, OID_AUTO, memorystatus_level, CTLFLAG_RD, &kern_memorystatus_level, 0, "");
-SYSCTL_UINT(_kern, OID_AUTO, memorystatus_kev_failure_count, CTLFLAG_RD, &kern_memorystatus_kev_failure_count, 0, "");
+void
+memorystatus_on_inactivity(int pid)
+{
+#pragma unused(pid)
+#if CONFIG_FREEZE
+       /* Wake the freeze thread */
+       thread_wakeup((event_t)&memorystatus_freeze_wakeup);
+#endif 
+}
 
-__private_extern__ void
-kern_memorystatus_init(void)
+static void
+memorystatus_thread(void *param __unused, wait_result_t wr __unused)
 {
-    jetsam_lck_attr = lck_attr_alloc_init();
-    jetsam_lck_grp_attr= lck_grp_attr_alloc_init();
-    jetsam_lck_grp = lck_grp_alloc_init("jetsam",  jetsam_lck_grp_attr);
-    jetsam_list_mlock = lck_mtx_alloc_init(jetsam_lck_grp, jetsam_lck_attr);
+       static boolean_t initialized = FALSE;
+       memorystatus_node *node;
+       uint64_t current_time;
+       pid_t victim_pid = -1;
+
+       if (initialized == FALSE) {
+               initialized = TRUE;
+               assert_wait(&memorystatus_wakeup, THREAD_UNINT);
+               (void)thread_block((thread_continue_t)memorystatus_thread);
+       }
+
+       /*  Pick next idle exit victim. For now, just iterate through; ideally, this would be be more intelligent. */
+       current_time = mach_absolute_time();
+       
+       /* Set a cutoff so that we don't idle exit processes that went recently clean */
+       
+       lck_mtx_lock(memorystatus_list_mlock);
+       
+       if (memorystatus_dirty_count) {
+               TAILQ_FOREACH(node, &memorystatus_list, link) {
+                       if ((node->state & kProcessSupportsIdleExit) && !(node->state & (kProcessDirty|kProcessIgnoreIdleExit))) {                              
+                               if (current_time >= node->clean_time) {
+                                       victim_pid = node->pid;
+                                       break;
+                               }
+                       }
+               }
+       }
+
+       lck_mtx_unlock(memorystatus_list_mlock);
+       
+       if (-1 != victim_pid) {         
+               proc_t p = proc_find(victim_pid);
+               if (p != NULL) {
+                       boolean_t kill = FALSE;
+                       proc_dirty_start(p);
+                       /* Ensure process is still marked for idle exit and is clean */
+                       if ((p->p_dirty & (P_DIRTY_ALLOW_IDLE_EXIT|P_DIRTY_IS_DIRTY|P_DIRTY_TERMINATED)) == (P_DIRTY_ALLOW_IDLE_EXIT)) {
+                               /* Clean; issue SIGKILL */
+                               p->p_dirty |= P_DIRTY_TERMINATED;
+                               kill = TRUE;
+                       }
+                       proc_dirty_end(p);
+                       if (TRUE == kill) {
+                               printf("memorystatus_thread: idle exiting pid %d [%s]\n", victim_pid, (p->p_comm ? p->p_comm : "(unknown)"));
+                               psignal(p, SIGKILL);
+                       }
+                       proc_rele(p);
+               }
+       }
 
-       (void)kernel_thread(kernel_task, kern_memorystatus_thread);
+       assert_wait(&memorystatus_wakeup, THREAD_UNINT);
+       (void)thread_block((thread_continue_t)memorystatus_thread);
 }
 
+#if CONFIG_JETSAM
+
 static uint32_t
-jetsam_task_page_count(task_t task)
+memorystatus_task_page_count(task_t task)
 {
        kern_return_t ret;
        static task_info_data_t data;
@@ -104,271 +929,964 @@ jetsam_task_page_count(task_t task)
        return 0;
 }
 
+static int
+memorystatus_send_note(int event_code, void *data, size_t data_length) {
+       int ret;
+       struct kev_msg ev_msg;
+       
+       ev_msg.vendor_code    = KEV_VENDOR_APPLE;
+       ev_msg.kev_class      = KEV_SYSTEM_CLASS;
+       ev_msg.kev_subclass   = KEV_MEMORYSTATUS_SUBCLASS;
+
+       ev_msg.event_code     = event_code;
+
+       ev_msg.dv[0].data_length = data_length;
+       ev_msg.dv[0].data_ptr = data;
+       ev_msg.dv[1].data_length = 0;
+
+       ret = kev_post_msg(&ev_msg);
+       if (ret) {
+               memorystatus_kev_failure_count++;
+               printf("%s: kev_post_msg() failed, err %d\n", __func__, ret);
+       }
+       
+    return ret;
+}
+
 static uint32_t
-jetsam_flags_for_pid(pid_t pid)
+memorystatus_build_flags_from_state(uint32_t state) {
+    uint32_t flags = 0;
+    
+    if (state & kProcessForeground) {
+        flags |= kMemorystatusFlagsFrontmost;
+    }
+    if (state & kProcessActive) {
+        flags |= kMemorystatusFlagsActive;
+    }
+    if (state & kProcessSupportsIdleExit) {
+        flags |= kMemorystatusFlagsSupportsIdleExit;
+    }
+    if (state & kProcessDirty) {
+        flags |= kMemorystatusFlagsDirty;
+    }
+    
+    return flags;
+}
+
+static void 
+memorystatus_move_node_to_exit_list(memorystatus_node *node) 
 {
-       int i;
+       /* Make sure we're called with the list lock held */
+       lck_mtx_assert(memorystatus_list_mlock, LCK_MTX_ASSERT_OWNED);
+    
+       /* Now, acquire the exit list lock... */
+       lck_mtx_lock(exit_list_mlock);
+       
+       /* Remove from list + update accounting... */
+       memorystatus_remove_node(node);
+       
+       /* ...then insert at the end of the exit queue */
+       TAILQ_INSERT_TAIL(&exit_list, node, link);
+       
+       /* And relax */
+       lck_mtx_unlock(exit_list_mlock);
+}
 
-       for (i = 0; i < jetsam_priority_list_count; i++) {
-               if (pid == jetsam_priority_list[i].pid) {
-                       return jetsam_priority_list[i].flags;
+void memorystatus_update(unsigned int pages_avail)
+{        
+       if (!memorystatus_delta) {
+           return;
+       }
+                     
+       if ((pages_avail < memorystatus_available_pages_critical) ||
+            (pages_avail >= (memorystatus_available_pages + memorystatus_delta)) ||
+            (memorystatus_available_pages >= (pages_avail + memorystatus_delta))) {
+               memorystatus_available_pages = pages_avail;
+               memorystatus_level = memorystatus_available_pages * 100 / atop_64(max_mem);
+               /* Only wake the thread if currently blocked */
+               if (OSCompareAndSwap(0, 1, &memorystatus_jetsam_running)) {
+                       thread_wakeup((event_t)&memorystatus_jetsam_wakeup);
                }
        }
-       return 0;
+}
+
+static boolean_t
+memorystatus_get_snapshot_properties_for_proc_locked(proc_t p, memorystatus_jetsam_snapshot_entry_t *entry)
+{      
+       memorystatus_node *node;
+    
+       TAILQ_FOREACH(node, &memorystatus_list, link) {
+               if (node->pid == p->p_pid) {
+                       break;
+               }
+       }
+       
+       if (!node) {
+               return FALSE;
+       }
+       
+       entry->pid = p->p_pid;
+       strlcpy(&entry->name[0], p->p_comm, MAXCOMLEN+1);
+       entry->priority = node->priority;
+       entry->pages = memorystatus_task_page_count(p->task);
+       entry->flags = memorystatus_build_flags_from_state(node->state);
+       memcpy(&entry->uuid[0], &p->p_uuid[0], sizeof(p->p_uuid));
+
+       return TRUE;    
 }
 
 static void
-jetsam_snapshot_procs(void)
+memorystatus_jetsam_snapshot_procs_locked(void)
 {
        proc_t p;
        int i = 0;
 
-       jetsam_snapshot.stats.free_pages = vm_page_free_count;
-       jetsam_snapshot.stats.active_pages = vm_page_active_count;
-       jetsam_snapshot.stats.inactive_pages = vm_page_inactive_count;
-       jetsam_snapshot.stats.purgeable_pages = vm_page_purgeable_count;
-       jetsam_snapshot.stats.wired_pages = vm_page_wire_count;
+       memorystatus_jetsam_snapshot.stats.free_pages = vm_page_free_count;
+       memorystatus_jetsam_snapshot.stats.active_pages = vm_page_active_count;
+       memorystatus_jetsam_snapshot.stats.inactive_pages = vm_page_inactive_count;
+       memorystatus_jetsam_snapshot.stats.throttled_pages = vm_page_throttled_count;
+       memorystatus_jetsam_snapshot.stats.purgeable_pages = vm_page_purgeable_count;
+       memorystatus_jetsam_snapshot.stats.wired_pages = vm_page_wire_count;
        proc_list_lock();
        LIST_FOREACH(p, &allproc, p_list) {
-               task_t task = p->task;
-               jetsam_snapshot_list[i].pid = p->p_pid;
-               jetsam_snapshot_list[i].pages = jetsam_task_page_count(task);
-               jetsam_snapshot_list[i].flags = jetsam_flags_for_pid(p->p_pid);
-               strlcpy(&jetsam_snapshot_list[i].name[0], p->p_comm, MAXCOMLEN+1);
-#ifdef DEBUG
-               printf("jetsam snapshot pid = %d, uuid = %02x%02x%02x%02x%02x%02x%02x%02x%02x%02x%02x%02x%02x%02x%02x%02x\n",
+               if (FALSE == memorystatus_get_snapshot_properties_for_proc_locked(p, &memorystatus_jetsam_snapshot_list[i])) {
+                       continue;
+               }
+               
+               MEMORYSTATUS_DEBUG(0, "jetsam snapshot pid = %d, uuid = %02x%02x%02x%02x%02x%02x%02x%02x%02x%02x%02x%02x%02x%02x%02x%02x\n",
                        p->p_pid, 
                        p->p_uuid[0], p->p_uuid[1], p->p_uuid[2], p->p_uuid[3], p->p_uuid[4], p->p_uuid[5], p->p_uuid[6], p->p_uuid[7],
                        p->p_uuid[8], p->p_uuid[9], p->p_uuid[10], p->p_uuid[11], p->p_uuid[12], p->p_uuid[13], p->p_uuid[14], p->p_uuid[15]);
-#endif
-               memcpy(&jetsam_snapshot_list[i].uuid[0], &p->p_uuid[0], sizeof(p->p_uuid));
-               i++;
-               if (i == kMaxSnapshotEntries) {
+
+               if (++i == kMaxSnapshotEntries) {
                        break;
                }       
        }
        proc_list_unlock();     
-       jetsam_snapshot.entry_count = jetsam_snapshot_list_count = i - 1;
+       memorystatus_jetsam_snapshot.snapshot_time = mach_absolute_time();
+       memorystatus_jetsam_snapshot.entry_count = memorystatus_jetsam_snapshot_list_count = i - 1;
 }
 
 static void
-jetsam_mark_pid_in_snapshot(pid_t pid)
+memorystatus_mark_pid_in_snapshot(pid_t pid, int flags)
 {
-
        int i = 0;
 
-       for (i = 0; i < jetsam_snapshot_list_count; i++) {
-               if (jetsam_snapshot_list[i].pid == pid) {
-                       jetsam_snapshot_list[i].flags |= kJetsamFlagsKilled;
+       for (i = 0; i < memorystatus_jetsam_snapshot_list_count; i++) {
+               if (memorystatus_jetsam_snapshot_list[i].pid == pid) {
+                       memorystatus_jetsam_snapshot_list[i].flags |= flags;
                        return;
                }
        }
 }
 
-static int
-jetsam_kill_top_proc(void)
+int
+memorystatus_kill_top_proc(boolean_t any, uint32_t cause)
 {
        proc_t p;
+       int pending_snapshot = 0;
 
-       if (jetsam_snapshot_list_count == 0) {
-               jetsam_snapshot_procs();
+#ifndef CONFIG_FREEZE
+#pragma unused(any)
+#endif
+       
+       lck_mtx_lock(memorystatus_list_mlock);
+
+       if (memorystatus_jetsam_snapshot_list_count == 0) {
+               memorystatus_jetsam_snapshot_procs_locked();
+       } else {
+               pending_snapshot = 1;
        }
-       lck_mtx_lock(jetsam_list_mlock);
-       while (jetsam_priority_list_index < jetsam_priority_list_count) {
+
+       while (next_memorystatus_node) {
+               memorystatus_node *node;
                pid_t aPid;
-               aPid = jetsam_priority_list[jetsam_priority_list_index].pid;
-               jetsam_priority_list_index++;
+#if DEVELOPMENT || DEBUG
+               int activeProcess;
+               int procSuspendedForDiagnosis;
+#endif /* DEVELOPMENT || DEBUG */
+
+               node = next_memorystatus_node;
+               next_memorystatus_node = TAILQ_NEXT(next_memorystatus_node, link);
+
+#if DEVELOPMENT || DEBUG
+               activeProcess = node->state & kProcessForeground;
+               procSuspendedForDiagnosis = node->state & kProcessSuspendedForDiag;
+#endif /* DEVELOPMENT || DEBUG */
+               
+               aPid = node->pid;
+
                /* skip empty slots in the list */
-               if (aPid == 0) {
+               if (aPid == 0  || (node->state & kProcessKilled)) {
                        continue; // with lock held
                }
-               lck_mtx_unlock(jetsam_list_mlock);
-               jetsam_mark_pid_in_snapshot(aPid);
+
                p = proc_find(aPid);
                if (p != NULL) {
-#if DEBUG
-                       printf("jetsam: killing pid %d [%s] - memory_status_level: %d - ", aPid, p->p_comm, kern_memorystatus_level);
-#endif /* DEBUG */
-                       exit1(p, W_EXITCODE(0, SIGKILL), (int *)NULL);
-                       proc_rele(p);
-#if DEBUG
-                       printf("jetsam: pid %d killed - memory_status_level: %d\n", aPid, kern_memorystatus_level);
-#endif /* DEBUG */
-                       return 0;
+                       int flags = cause;
+                       
+#if DEVELOPMENT || DEBUG
+                       if ((memorystatus_jetsam_policy & kPolicyDiagnoseActive) && procSuspendedForDiagnosis) {
+                               printf("jetsam: continuing after ignoring proc suspended already for diagnosis - %d\n", aPid);
+                               proc_rele(p);
+                               continue;
+                       }
+#endif /* DEVELOPMENT || DEBUG */
+
+#if CONFIG_FREEZE
+                       boolean_t skip;
+                       boolean_t reclaim_proc = !(node->state & (kProcessLocked | kProcessNoReclaimWorth));
+                       if (any || reclaim_proc) {
+                               if (node->state & kProcessFrozen) {
+                                       flags |= kMemorystatusFlagsFrozen;
+                               }
+                               skip = FALSE;
+                       } else {
+                               skip = TRUE;
+                       }
+                       
+                       if (skip) {
+                               proc_rele(p);                   
+                       } else
+#endif
+                       {
+#if DEVELOPMENT || DEBUG
+                               if ((memorystatus_jetsam_policy & kPolicyDiagnoseActive) && activeProcess) {
+                                       MEMORYSTATUS_DEBUG(1, "jetsam: suspending pid %d [%s] (active) for diagnosis - memory_status_level: %d\n",
+                                               aPid, (p->p_comm ? p->p_comm: "(unknown)"), memorystatus_level);
+                                       memorystatus_mark_pid_in_snapshot(aPid, kMemorystatusFlagsSuspForDiagnosis);
+                                       node->state |= kProcessSuspendedForDiag;
+                                       if (memorystatus_jetsam_policy & kPolicyDiagnoseFirst) {
+                                               jetsam_diagnostic_suspended_one_active_proc = 1;
+                                               printf("jetsam: returning after suspending first active proc - %d\n", aPid);
+                                       }
+                                       lck_mtx_unlock(memorystatus_list_mlock);
+                                       task_suspend(p->task);
+                                       proc_rele(p);
+                                       return 0;
+                               } else
+#endif /* DEVELOPMENT || DEBUG */
+                               {
+                                       printf("memorystatus: jetsam killing pid %d [%s] - memorystatus_available_pages: %d\n", 
+                                               aPid, (p->p_comm ? p->p_comm : "(unknown)"), memorystatus_available_pages);
+                                       /* Shift queue, update stats */
+                                       memorystatus_move_node_to_exit_list(node);
+                                       memorystatus_mark_pid_in_snapshot(aPid, flags);
+                                       lck_mtx_unlock(memorystatus_list_mlock);
+                                       exit1_internal(p, W_EXITCODE(0, SIGKILL), (int *)NULL, FALSE, FALSE);
+                                       proc_rele(p);
+                                       return 0;
+                               }
+                       }
+               }
+       }
+       
+       lck_mtx_unlock(memorystatus_list_mlock);
+       
+       // If we didn't kill anything, toss any newly-created snapshot
+       if (!pending_snapshot) {
+           memorystatus_jetsam_snapshot.entry_count = memorystatus_jetsam_snapshot_list_count = 0;
+       }
+       
+       return -1;
+}
+
+int memorystatus_kill_top_proc_from_VM(void) {
+       return memorystatus_kill_top_proc(TRUE, kMemorystatusFlagsKilledVM);
+}
+
+static int
+memorystatus_kill_hiwat_proc(void)
+{
+       proc_t p;
+       int pending_snapshot = 0;
+       memorystatus_node *next_hiwat_node;
+       
+       lck_mtx_lock(memorystatus_list_mlock);
+       
+       if (memorystatus_jetsam_snapshot_list_count == 0) {
+               memorystatus_jetsam_snapshot_procs_locked();
+       } else {
+               pending_snapshot = 1;
+       }
+       
+       next_hiwat_node = next_memorystatus_node;
+       
+       while (next_hiwat_node) {
+               pid_t aPid;
+               int32_t hiwat;
+               memorystatus_node *node;
+        
+               node = next_hiwat_node;
+               next_hiwat_node = TAILQ_NEXT(next_hiwat_node, link);
+               
+               aPid = node->pid;
+               hiwat = node->hiwat_pages;
+               
+               /* skip empty or non-hiwat slots in the list */
+               if (aPid == 0 || (hiwat < 0) || (node->state & kProcessKilled)) {
+                       continue; // with lock held
+               }
+               
+               p = proc_find(aPid);
+               if (p != NULL) {
+                       int32_t pages = (int32_t)memorystatus_task_page_count(p->task);
+                       boolean_t skip = (pages <= hiwat);
+#if DEVELOPMENT || DEBUG
+                       if (!skip && (memorystatus_jetsam_policy & kPolicyDiagnoseActive)) {
+                               if (node->state & kProcessSuspendedForDiag) {
+                                       proc_rele(p);
+                                       continue;
+                               }
+                       }
+#endif /* DEVELOPMENT || DEBUG */
+
+#if CONFIG_FREEZE
+                       if (!skip) {
+                               if (node->state & kProcessLocked) {
+                                       skip = TRUE;
+                               } else {
+                                       skip = FALSE;
+                               }                               
+                       }
+#endif
+
+                       if (!skip) {
+                               MEMORYSTATUS_DEBUG(1, "jetsam: %s pid %d [%s] - %d pages > 1 (%d)\n",
+                                       (memorystatus_jetsam_policy & kPolicyDiagnoseActive) ? "suspending": "killing", aPid, p->p_comm, pages, hiwat);
+#if DEVELOPMENT || DEBUG
+                               if (memorystatus_jetsam_policy & kPolicyDiagnoseActive) {
+                                   memorystatus_mark_pid_in_snapshot(aPid, kMemorystatusFlagsSuspForDiagnosis);
+                                       node->state |= kProcessSuspendedForDiag;
+                                       lck_mtx_unlock(memorystatus_list_mlock);
+                                       task_suspend(p->task);
+                                       proc_rele(p);
+                                       MEMORYSTATUS_DEBUG(1, "jetsam: pid %d suspended for diagnosis - memorystatus_available_pages: %d\n", aPid, memorystatus_available_pages);
+                               } else
+#endif /* DEVELOPMENT || DEBUG */
+                               {       
+                                       printf("memorystatus: jetsam killing pid %d [%s] (highwater) - memorystatus_available_pages: %d\n", 
+                                               aPid, (p->p_comm ? p->p_comm : "(unknown)"), memorystatus_available_pages);
+                                       /* Shift queue, update stats */
+                                       memorystatus_move_node_to_exit_list(node);
+                                       memorystatus_mark_pid_in_snapshot(aPid, kMemorystatusFlagsKilledHiwat);
+                                       lck_mtx_unlock(memorystatus_list_mlock);                    
+                                       exit1(p, W_EXITCODE(0, SIGKILL), (int *)NULL);
+                                       proc_rele(p);
+                               }
+                               return 0;
+                       } else {
+                               proc_rele(p);
+                       }
+
                }
-           lck_mtx_lock(jetsam_list_mlock);
        }
-       lck_mtx_unlock(jetsam_list_mlock);
+       
+       lck_mtx_unlock(memorystatus_list_mlock);
+       
+       // If we didn't kill anything, toss any newly-created snapshot
+       if (!pending_snapshot) {
+               memorystatus_jetsam_snapshot.entry_count = memorystatus_jetsam_snapshot_list_count = 0;
+       }
+       
        return -1;
 }
 
 static void
-kern_memorystatus_thread(void)
+memorystatus_jetsam_thread_block(void)
 {
-       struct kev_msg ev_msg;
-       jetsam_kernel_stats_t data;
-       int ret;
+       assert_wait(&memorystatus_jetsam_wakeup, THREAD_UNINT);
+       assert(memorystatus_jetsam_running == 1);
+       OSDecrementAtomic(&memorystatus_jetsam_running);
+       (void)thread_block((thread_continue_t)memorystatus_jetsam_thread);   
+}
 
+static void
+memorystatus_jetsam_thread(void *param __unused, wait_result_t wr __unused)
+{
+       boolean_t post_snapshot = FALSE; 
+       static boolean_t is_vm_privileged = FALSE;
+
+       if (is_vm_privileged == FALSE) {
+               /* 
+                * It's the first time the thread has run, so just mark the thread as privileged and block.
+                * This avoids a spurious pass with unset variables, as set out in <rdar://problem/9609402>.
+                */
+               thread_wire(host_priv_self(), current_thread(), TRUE);
+               is_vm_privileged = TRUE;
+               memorystatus_jetsam_thread_block();
+       }
+       
+       assert(memorystatus_available_pages != (unsigned)-1);
+       
        while(1) {
+               unsigned int last_available_pages;
 
-               while (kern_memorystatus_level <= kern_memorystatus_level_critical) {
-                       if (jetsam_kill_top_proc() < 0) {
+#if DEVELOPMENT || DEBUG
+               jetsam_diagnostic_suspended_one_active_proc = 0;
+#endif /* DEVELOPMENT || DEBUG */
+           
+               while (memorystatus_available_pages <= memorystatus_available_pages_highwater) {
+                       if (memorystatus_kill_hiwat_proc() < 0) {
                                break;
                        }
+                       post_snapshot = TRUE;
+               }
+
+               while (memorystatus_available_pages <= memorystatus_available_pages_critical) {
+                       if (memorystatus_kill_top_proc(FALSE, kMemorystatusFlagsKilled) < 0) {
+                               /* No victim was found - panic */
+                               panic("memorystatus_jetsam_thread: no victim! available pages:%d, critical page level: %d\n",
+                                        memorystatus_available_pages, memorystatus_available_pages_critical);
+                       }
+                       post_snapshot = TRUE;
+#if DEVELOPMENT || DEBUG
+                       if ((memorystatus_jetsam_policy & kPolicyDiagnoseFirst) && jetsam_diagnostic_suspended_one_active_proc) {
+                               printf("jetsam: stopping killing since 1 active proc suspended already for diagnosis\n");
+                               break; // we found first active proc, let's not kill any more
+                       }
+#endif /* DEVELOPMENT || DEBUG */
                }
+               
+               last_available_pages = memorystatus_available_pages;
 
-               kern_memorystatus_last_level = kern_memorystatus_level;
+               if (post_snapshot) {
+                       size_t snapshot_size = sizeof(memorystatus_jetsam_snapshot_t) + sizeof(memorystatus_jetsam_snapshot_entry_t) * (memorystatus_jetsam_snapshot_list_count - 1);
+                       memorystatus_jetsam_snapshot.notification_time = mach_absolute_time();
+                       memorystatus_send_note(kMemorystatusSnapshotNote, &snapshot_size, sizeof(snapshot_size));
+               }
+
+               if (memorystatus_available_pages >= (last_available_pages + memorystatus_delta) ||
+                   last_available_pages >= (memorystatus_available_pages + memorystatus_delta)) {
+                       continue;
+               }
+
+#if VM_PRESSURE_EVENTS
+               memorystatus_check_pressure_reset();
+#endif
+
+               memorystatus_jetsam_thread_block();
+       }
+}
+
+#endif /* CONFIG_JETSAM */
+
+#if CONFIG_FREEZE
+
+__private_extern__ void
+memorystatus_freeze_init(void)
+{
+       kern_return_t result;
+       thread_t thread;
+       
+       result = kernel_thread_start(memorystatus_freeze_thread, NULL, &thread);
+       if (result == KERN_SUCCESS) {
+               thread_deallocate(thread);
+       } else {
+               panic("Could not create memorystatus_freeze_thread");
+       }
+}
 
-               ev_msg.vendor_code    = KEV_VENDOR_APPLE;
-               ev_msg.kev_class      = KEV_SYSTEM_CLASS;
-               ev_msg.kev_subclass   = KEV_MEMORYSTATUS_SUBCLASS;
+static int
+memorystatus_freeze_top_proc(boolean_t *memorystatus_freeze_swap_low)
+{
+       proc_t p;
+       uint32_t i;
+       memorystatus_node *next_freeze_node;
 
-               /* pass the memory status level (percent free) */
-               ev_msg.event_code     = kMemoryStatusLevelNote;
+       lck_mtx_lock(memorystatus_list_mlock);
+       
+       next_freeze_node = next_memorystatus_node;
+       
+       while (next_freeze_node) {
+               memorystatus_node *node;
+               pid_t aPid;
+               uint32_t state;
+               
+               node = next_freeze_node;
+               next_freeze_node = TAILQ_NEXT(next_freeze_node, link);
 
-               ev_msg.dv[0].data_length = sizeof kern_memorystatus_last_level;
-               ev_msg.dv[0].data_ptr = &kern_memorystatus_last_level;
-               ev_msg.dv[1].data_length = sizeof data;
-               ev_msg.dv[1].data_ptr = &data;
-               ev_msg.dv[2].data_length = 0;
+               aPid = node->pid;
+               state = node->state;
 
-               data.free_pages = vm_page_free_count;
-               data.active_pages = vm_page_active_count;
-               data.inactive_pages = vm_page_inactive_count;
-               data.purgeable_pages = vm_page_purgeable_count;
-               data.wired_pages = vm_page_wire_count;
+               /* skip empty slots in the list */
+               if (aPid == 0) {
+                       continue; // with lock held
+               }
 
-               ret = kev_post_msg(&ev_msg);
-               if (ret) {
-                       kern_memorystatus_kev_failure_count++;
-                       printf("%s: kev_post_msg() failed, err %d\n", __func__, ret);
+               /* Ensure the process is eligible for freezing */
+               if ((state & (kProcessKilled | kProcessLocked | kProcessFrozen)) || !(state & kProcessSuspended)) {
+                       continue; // with lock held
                }
 
-               if (jetsam_snapshot_list_count) {
-                       size_t snapshot_size =  sizeof(jetsam_kernel_stats_t) + sizeof(size_t) + sizeof(jetsam_snapshot_entry_t) * jetsam_snapshot_list_count;
-                       ev_msg.event_code = kMemoryStatusSnapshotNote;
-                       ev_msg.dv[0].data_length = sizeof snapshot_size;
-                       ev_msg.dv[0].data_ptr = &snapshot_size;
-                       ev_msg.dv[1].data_length = 0;
+               p = proc_find(aPid);
+               if (p != NULL) {
+                       kern_return_t kr;
+                       uint32_t purgeable, wired, clean, dirty;
+                       boolean_t shared;
+                       uint32_t max_pages = 0;
+                                       
+                       /* Only freeze processes meeting our minimum resident page criteria */
+                       if (memorystatus_task_page_count(p->task) < memorystatus_freeze_pages_min) {
+                               proc_rele(p);
+                               continue;
+                       } 
 
-                       ret = kev_post_msg(&ev_msg);
-                       if (ret) {
-                               kern_memorystatus_kev_failure_count++;
-                               printf("%s: kev_post_msg() failed, err %d\n", __func__, ret);
+                       /* Ensure there's enough free space to freeze this process. */                  
+                       max_pages = MIN(default_pager_swap_pages_free(), memorystatus_freeze_pages_max);
+                       if (max_pages < memorystatus_freeze_pages_min) {
+                               *memorystatus_freeze_swap_low = TRUE;
+                               proc_rele(p);
+                               lck_mtx_unlock(memorystatus_list_mlock);
+                               return 0;
                        }
-               }
+                       
+                       /* Mark as locked temporarily to avoid kill */
+                       node->state |= kProcessLocked;
+                       
+                       kr = task_freeze(p->task, &purgeable, &wired, &clean, &dirty, max_pages, &shared, FALSE);
+                       
+                       MEMORYSTATUS_DEBUG(1, "memorystatus_freeze_top_proc: task_freeze %s for pid %d [%s] - "
+                       "memorystatus_pages: %d, purgeable: %d, wired: %d, clean: %d, dirty: %d, shared %d, free swap: %d\n", 
+                       (kr == KERN_SUCCESS) ? "SUCCEEDED" : "FAILED", aPid, (p->p_comm ? p->p_comm : "(unknown)"), 
+                       memorystatus_available_pages, purgeable, wired, clean, dirty, shared, default_pager_swap_pages_free());
+                       
+                       proc_rele(p);
+               
+                       node->state &= ~kProcessLocked;
+                       
+                       if (KERN_SUCCESS == kr) {
+                               memorystatus_freeze_entry_t data = { aPid, kMemorystatusFlagsFrozen, dirty };
+                               
+                               memorystatus_frozen_count++;
+                               
+                               node->state |= (kProcessFrozen | (shared ? 0: kProcessNoReclaimWorth));
+                       
+                               /* Update stats */
+                               for (i = 0; i < sizeof(throttle_intervals) / sizeof(struct throttle_interval_t); i++) {
+                               throttle_intervals[i].pageouts += dirty;
+                               }
+                       
+                               memorystatus_freeze_pageouts += dirty;
+                               memorystatus_freeze_count++;
 
-               if (kern_memorystatus_level >= kern_memorystatus_last_level + 5 ||
-                   kern_memorystatus_level <= kern_memorystatus_last_level - 5)
-                       continue;
+                               lck_mtx_unlock(memorystatus_list_mlock);
+
+                               memorystatus_send_note(kMemorystatusFreezeNote, &data, sizeof(data));
 
-               assert_wait(&kern_memorystatus_wakeup, THREAD_UNINT);
-               (void)thread_block((thread_continue_t)kern_memorystatus_thread);
+                               return dirty;
+                       }
+                       
+                       /* Failed; go round again */
+               }
        }
+       
+       lck_mtx_unlock(memorystatus_list_mlock);
+       
+       return -1;
 }
 
-static int
-sysctl_io_variable(struct sysctl_req *req, void *pValue, size_t currentsize, size_t maxsize, size_t *newsize)
+static inline boolean_t 
+memorystatus_can_freeze_processes(void) 
 {
-    int error;
+       boolean_t ret;
+       
+       lck_mtx_lock(memorystatus_list_mlock);
+       
+       if (memorystatus_suspended_count) {
+               uint32_t average_resident_pages, estimated_processes;
+        
+               /* Estimate the number of suspended processes we can fit */
+               average_resident_pages = memorystatus_suspended_resident_count / memorystatus_suspended_count;
+               estimated_processes = memorystatus_suspended_count +
+                       ((memorystatus_available_pages - memorystatus_available_pages_critical) / average_resident_pages);
 
-    /* Copy blob out */
-    error = SYSCTL_OUT(req, pValue, currentsize);
+               /* If it's predicted that no freeze will occur, lower the threshold temporarily */
+               if (estimated_processes <= FREEZE_SUSPENDED_THRESHOLD_DEFAULT) {
+                       memorystatus_freeze_suspended_threshold = FREEZE_SUSPENDED_THRESHOLD_LOW;
+               } else {
+                       memorystatus_freeze_suspended_threshold = FREEZE_SUSPENDED_THRESHOLD_DEFAULT;        
+               }
 
-    /* error or nothing to set */
-    if (error || !req->newptr)
-        return(error);
+               MEMORYSTATUS_DEBUG(1, "memorystatus_can_freeze_processes: %d suspended processes, %d average resident pages / process, %d suspended processes estimated\n", 
+                       memorystatus_suspended_count, average_resident_pages, estimated_processes);
+       
+               if ((memorystatus_suspended_count - memorystatus_frozen_count) > memorystatus_freeze_suspended_threshold) {
+                       ret = TRUE;
+               } else {
+                       ret = FALSE;
+               }
+       } else {
+               ret = FALSE;
+       }
+                               
+       lck_mtx_unlock(memorystatus_list_mlock);
+       
+       return ret;
+}
 
-    if (req->newlen > maxsize) {
-               return EINVAL;
+static boolean_t 
+memorystatus_can_freeze(boolean_t *memorystatus_freeze_swap_low)
+{
+       /* Only freeze if we're sufficiently low on memory; this holds off freeze right
+          after boot,  and is generally is a no-op once we've reached steady state. */
+       if (memorystatus_available_pages > memorystatus_freeze_threshold) {
+               return FALSE;
+       }
+       
+       /* Check minimum suspended process threshold. */
+       if (!memorystatus_can_freeze_processes()) {
+               return FALSE;
        }
-       error = SYSCTL_IN(req, pValue, req->newlen);
 
-       if (!error) {
-               *newsize = req->newlen;
+       /* Is swap running low? */
+       if (*memorystatus_freeze_swap_low) {
+               /* If there's been no movement in free swap pages since we last attempted freeze, return. */
+               if (default_pager_swap_pages_free() < memorystatus_freeze_pages_min) {
+                       return FALSE;
+               }
+               
+               /* Pages have been freed - we can retry. */
+               *memorystatus_freeze_swap_low = FALSE;  
        }
+       
+       /* OK */
+       return TRUE;
+}
+
+static void
+memorystatus_freeze_update_throttle_interval(mach_timespec_t *ts, struct throttle_interval_t *interval)
+{
+       if (CMP_MACH_TIMESPEC(ts, &interval->ts) >= 0) {
+               if (!interval->max_pageouts) {
+                       interval->max_pageouts = (interval->burst_multiple * (((uint64_t)interval->mins * FREEZE_DAILY_PAGEOUTS_MAX) / (24 * 60)));
+               } else {
+                       printf("memorystatus_freeze_update_throttle_interval: %d minute throttle timeout, resetting\n", interval->mins);
+               }
+               interval->ts.tv_sec = interval->mins * 60;
+               interval->ts.tv_nsec = 0;
+               ADD_MACH_TIMESPEC(&interval->ts, ts);
+               /* Since we update the throttle stats pre-freeze, adjust for overshoot here */
+               if (interval->pageouts > interval->max_pageouts) {
+                       interval->pageouts -= interval->max_pageouts;
+               } else {
+                       interval->pageouts = 0;
+               }
+               interval->throttle = FALSE;
+       } else if (!interval->throttle && interval->pageouts >= interval->max_pageouts) {
+               printf("memorystatus_freeze_update_throttle_interval: %d minute pageout limit exceeded; enabling throttle\n", interval->mins);
+               interval->throttle = TRUE;
+       }       
 
-    return(error);
+       MEMORYSTATUS_DEBUG(1, "memorystatus_freeze_update_throttle_interval: throttle updated - %d frozen (%d max) within %dm; %dm remaining; throttle %s\n", 
+               interval->pageouts, interval->max_pageouts, interval->mins, (interval->ts.tv_sec - ts->tv_sec) / 60, 
+               interval->throttle ? "on" : "off");
 }
 
-static int
-sysctl_handle_kern_memorystatus_priority_list(__unused struct sysctl_oid *oid, __unused void *arg1, __unused int arg2, struct sysctl_req *req)
+static boolean_t
+memorystatus_freeze_update_throttle(void) 
+{
+       clock_sec_t sec;
+       clock_nsec_t nsec;
+       mach_timespec_t ts;
+       uint32_t i;
+       boolean_t throttled = FALSE;
+
+#if DEVELOPMENT || DEBUG
+       if (!memorystatus_freeze_throttle_enabled)
+               return FALSE;
+#endif
+
+       clock_get_system_nanotime(&sec, &nsec);
+       ts.tv_sec = sec;
+       ts.tv_nsec = nsec;
+       
+       /* Check freeze pageouts over multiple intervals and throttle if we've exceeded our budget.
+        *
+        * This ensures that periods of inactivity can't be used as 'credit' towards freeze if the device has
+        * remained dormant for a long period. We do, however, allow increased thresholds for shorter intervals in
+        * order to allow for bursts of activity.
+        */
+       for (i = 0; i < sizeof(throttle_intervals) / sizeof(struct throttle_interval_t); i++) {
+               memorystatus_freeze_update_throttle_interval(&ts, &throttle_intervals[i]);
+               if (throttle_intervals[i].throttle == TRUE)
+                       throttled = TRUE;
+       }                                                               
+
+       return throttled;
+}
+
+static void
+memorystatus_freeze_thread(void *param __unused, wait_result_t wr __unused)
 {
-       int i, ret;
-       jetsam_priority_entry_t temp_list[kMaxPriorityEntries];
-       size_t newsize, currentsize;
+       static boolean_t memorystatus_freeze_swap_low = FALSE;
+       
+       if (memorystatus_freeze_enabled) {
+               if (memorystatus_can_freeze(&memorystatus_freeze_swap_low)) {
+                       /* Only freeze if we've not exceeded our pageout budgets */
+                       if (!memorystatus_freeze_update_throttle()) {
+                               memorystatus_freeze_top_proc(&memorystatus_freeze_swap_low);
+                       } else {
+                               printf("memorystatus_freeze_thread: in throttle, ignoring freeze\n");
+                               memorystatus_freeze_throttle_count++; /* Throttled, update stats */
+                       }
+               }
+       }
+
+       assert_wait((event_t) &memorystatus_freeze_wakeup, THREAD_UNINT);
+       thread_block((thread_continue_t) memorystatus_freeze_thread);   
+}
+
+#endif /* CONFIG_FREEZE */
+
+#if CONFIG_JETSAM
+
+#if VM_PRESSURE_EVENTS
+
+static inline boolean_t
+memorystatus_get_pressure_locked(void) {
+       if (memorystatus_available_pages > memorystatus_available_pages_pressure) {
+                /* Too many free pages */
+                return kVMPressureNormal;
+       }
+       
+#if CONFIG_FREEZE
+       if (memorystatus_frozen_count > 0) {
+                /* Frozen processes exist */
+                return kVMPressureNormal;              
+       }
+#endif
+
+       if (memorystatus_suspended_count > MEMORYSTATUS_SUSPENDED_THRESHOLD) {
+               /* Too many supended processes */
+               return kVMPressureNormal;
+       }
+       
+       if (memorystatus_suspended_count > 0) {
+               /* Some suspended processes - warn */
+               return kVMPressureWarning;
+       }
+    
+       /* Otherwise, pressure level is urgent */
+       return kVMPressureUrgent;
+}
+
+pid_t
+memorystatus_request_vm_pressure_candidate(void) {
+       memorystatus_node *node;
+       pid_t pid = -1;
+
+       lck_mtx_lock(memorystatus_list_mlock);
 
-       if (req->oldptr) {
-               lck_mtx_lock(jetsam_list_mlock);
-               for (i = 0; i < jetsam_priority_list_count; i++) {
-                       temp_list[i] = jetsam_priority_list[i];
+       /* Are we in a low memory state? */
+       memorystatus_vm_pressure_level = memorystatus_get_pressure_locked();
+       if (kVMPressureNormal != memorystatus_vm_pressure_level) {
+               TAILQ_FOREACH(node, &memorystatus_list, link) {
+                       /* Skip ineligible processes */
+                       if (node->state & (kProcessKilled | kProcessLocked | kProcessSuspended | kProcessFrozen | kProcessNotifiedForPressure)) {
+                               continue;
+                       }
+                       node->state |= kProcessNotifiedForPressure;
+                       pid = node->pid;
+                       break;
                }
-               lck_mtx_unlock(jetsam_list_mlock);
        }
+    
+       lck_mtx_unlock(memorystatus_list_mlock);
 
-       currentsize = sizeof(jetsam_priority_list[0]) * jetsam_priority_list_count;
+       return pid;
+}
 
-       ret = sysctl_io_variable(req, &temp_list[0], currentsize, sizeof(temp_list), &newsize);
+void
+memorystatus_send_pressure_note(pid_t pid) {
+    memorystatus_send_note(kMemorystatusPressureNote, &pid, sizeof(pid));
+}
 
-       if (!ret && req->newptr) {
-               jetsam_priority_list_count = newsize / sizeof(jetsam_priority_list[0]);
-#if DEBUG 
-               printf("set jetsam priority pids = { ");
-               for (i = 0; i < jetsam_priority_list_count; i++) {
-                       printf("(%d, 0x%08x, %d) ", temp_list[i].pid, temp_list[i].flags, temp_list[i].hiwat_pages);
+static void
+memorystatus_check_pressure_reset() {        
+       lck_mtx_lock(memorystatus_list_mlock);
+       
+       if (kVMPressureNormal != memorystatus_vm_pressure_level) {
+               memorystatus_vm_pressure_level = memorystatus_get_pressure_locked();
+               if (kVMPressureNormal == memorystatus_vm_pressure_level) {
+                       memorystatus_node *node;
+                       TAILQ_FOREACH(node, &memorystatus_list, link) {
+                               node->state &= ~kProcessNotifiedForPressure;
+                       }
                }
-               printf("}\n");
-#endif /* DEBUG */
-               lck_mtx_lock(jetsam_list_mlock);
-               for (i = 0; i < jetsam_priority_list_count; i++) {
-                       jetsam_priority_list[i] = temp_list[i];
+       }
+    
+       lck_mtx_unlock(memorystatus_list_mlock);
+}
+
+#endif /* VM_PRESSURE_EVENTS */
+
+/* Sysctls... */
+
+static int
+sysctl_memorystatus_list_change SYSCTL_HANDLER_ARGS
+{
+       int ret;
+       memorystatus_priority_entry_t entry;
+
+#pragma unused(oidp, arg1, arg2)
+
+       if (!req->newptr || req->newlen > sizeof(entry)) {
+               return EINVAL;
+       }
+
+       ret = SYSCTL_IN(req, &entry, req->newlen);
+       if (ret) {
+               return ret;
+       }
+
+       memorystatus_list_change(FALSE, entry.pid, entry.priority, entry.flags, -1);
+
+       return ret;
+}
+
+SYSCTL_PROC(_kern, OID_AUTO, memorystatus_jetsam_change, CTLTYPE_INT|CTLFLAG_WR|CTLFLAG_LOCKED|CTLFLAG_MASKED,
+    0, 0, &sysctl_memorystatus_list_change, "I", "");
+    
+static int
+sysctl_memorystatus_priority_list(__unused struct sysctl_oid *oid, __unused void *arg1, __unused int arg2, struct sysctl_req *req)
+{
+       int ret;
+       size_t allocated_size, list_size = 0;
+       memorystatus_priority_entry_t *list;
+       uint32_t list_count, i = 0;
+       memorystatus_node *node;
+        
+       /* Races, but this is only for diagnostic purposes */
+       list_count = memorystatus_list_count;
+       allocated_size = sizeof(memorystatus_priority_entry_t) * list_count;
+       list = kalloc(allocated_size);
+       if (!list) {
+               return ENOMEM;
+       }
+
+       memset(list, 0, allocated_size);
+        
+       lck_mtx_lock(memorystatus_list_mlock);
+
+       TAILQ_FOREACH(node, &memorystatus_list, link) {
+               list[i].pid = node->pid;
+               list[i].priority = node->priority; 
+               list[i].flags = memorystatus_build_flags_from_state(node->state);
+               list[i].hiwat_pages = node->hiwat_pages;
+               list_size += sizeof(memorystatus_priority_entry_t);
+               if (++i >= list_count) {
+                       break;
+               }       
+       }
+       
+       lck_mtx_unlock(memorystatus_list_mlock);
+       
+       if (!list_size) {
+               if (req->oldptr) {
+                       MEMORYSTATUS_DEBUG(1, "kern.memorystatus_priority_list returning EINVAL\n");
+                       return EINVAL;
                }
-               for (i = jetsam_priority_list_count; i < kMaxPriorityEntries; i++) {
-                       jetsam_priority_list[i].pid = 0;
-                       jetsam_priority_list[i].flags = 0;
-                       jetsam_priority_list[i].hiwat_pages = -1;
-                       jetsam_priority_list[i].hiwat_reserved1 = -1;
-                       jetsam_priority_list[i].hiwat_reserved2 = -1;
-                       jetsam_priority_list[i].hiwat_reserved3 = -1;
+               else {
+                       MEMORYSTATUS_DEBUG(1, "kern.memorystatus_priority_list returning 0 for size\n");
                }
-               jetsam_priority_list_index = 0;
-               lck_mtx_unlock(jetsam_list_mlock);
-       }       
+       } else {
+               MEMORYSTATUS_DEBUG(1, "kern.memorystatus_priority_list returning %ld for size\n", (long)list_size);
+       }
+       
+       ret = SYSCTL_OUT(req, list, list_size);
+
+       kfree(list, allocated_size);
+       
        return ret;
 }
 
+SYSCTL_PROC(_kern, OID_AUTO, memorystatus_priority_list, CTLTYPE_OPAQUE|CTLFLAG_RD | CTLFLAG_LOCKED, 0, 0, sysctl_memorystatus_priority_list, "S,jetsam_priorities", "");
+
+static void
+memorystatus_update_levels_locked(void) {
+       /* Set the baseline levels in pages */
+       memorystatus_available_pages_critical = (CRITICAL_PERCENT / DELTA_PERCENT) * memorystatus_delta;
+       memorystatus_available_pages_highwater = (HIGHWATER_PERCENT / DELTA_PERCENT) * memorystatus_delta;
+#if VM_PRESSURE_EVENTS
+       memorystatus_available_pages_pressure = (PRESSURE_PERCENT / DELTA_PERCENT) * memorystatus_delta;
+#endif
+       
+#if DEBUG || DEVELOPMENT
+       if (memorystatus_jetsam_policy & kPolicyDiagnoseActive) {
+               memorystatus_available_pages_critical += memorystatus_jetsam_policy_offset_pages_diagnostic;
+               memorystatus_available_pages_highwater += memorystatus_jetsam_policy_offset_pages_diagnostic;
+#if VM_PRESSURE_EVENTS
+               memorystatus_available_pages_pressure += memorystatus_jetsam_policy_offset_pages_diagnostic;
+#endif
+       }
+#endif
+       
+       /* Only boost the critical level - it's more important to kill right away than issue warnings */
+       if (memorystatus_jetsam_policy & kPolicyMoreFree) {
+               memorystatus_available_pages_critical += memorystatus_jetsam_policy_offset_pages_more_free;
+       }
+}
+
+static int
+sysctl_memorystatus_jetsam_policy_more_free SYSCTL_HANDLER_ARGS
+{
+#pragma unused(arg1, arg2, oidp)
+       int error, more_free = 0;
+
+       error = priv_check_cred(kauth_cred_get(), PRIV_VM_JETSAM, 0);
+       if (error)
+               return (error);
+
+       error = sysctl_handle_int(oidp, &more_free, 0, req);
+       if (error || !req->newptr)
+               return (error);
+
+       lck_mtx_lock(memorystatus_list_mlock);
+       
+       if (more_free) {
+               memorystatus_jetsam_policy |= kPolicyMoreFree;
+       } else {
+               memorystatus_jetsam_policy &= ~kPolicyMoreFree;               
+       }
+        
+       memorystatus_update_levels_locked();
+               
+       lck_mtx_unlock(memorystatus_list_mlock);
+       
+       return 0;
+}
+
+SYSCTL_PROC(_kern, OID_AUTO, memorystatus_jetsam_policy_more_free, CTLTYPE_INT|CTLFLAG_WR|CTLFLAG_LOCKED|CTLFLAG_MASKED|CTLFLAG_ANYBODY,
+    0, 0, &sysctl_memorystatus_jetsam_policy_more_free, "I", "");
+
 static int
-sysctl_handle_kern_memorystatus_snapshot(__unused struct sysctl_oid *oid, __unused void *arg1, __unused int arg2, struct sysctl_req *req)
+sysctl_handle_memorystatus_snapshot(__unused struct sysctl_oid *oid, __unused void *arg1, __unused int arg2, struct sysctl_req *req)
 {
        int ret;
        size_t currentsize = 0;
 
-       if (jetsam_snapshot_list_count > 0) {
-               currentsize = sizeof(jetsam_kernel_stats_t) + sizeof(size_t) + sizeof(jetsam_snapshot_entry_t) * jetsam_snapshot_list_count;
+       if (memorystatus_jetsam_snapshot_list_count > 0) {
+               currentsize = sizeof(memorystatus_jetsam_snapshot_t) + sizeof(memorystatus_jetsam_snapshot_entry_t) * (memorystatus_jetsam_snapshot_list_count - 1);
        }
        if (!currentsize) {
                if (req->oldptr) {
-#ifdef DEBUG
-                       printf("kern.memorystatus_snapshot returning EINVAL\n");
-#endif
+                       MEMORYSTATUS_DEBUG(1, "kern.memorystatus_snapshot returning EINVAL\n");
                        return EINVAL;
                }
                else {
-#ifdef DEBUG
-                       printf("kern.memorystatus_snapshot returning 0 for size\n");
-#endif
+                       MEMORYSTATUS_DEBUG(1, "kern.memorystatus_snapshot returning 0 for size\n");
                }
        } else {
-#ifdef DEBUG
-                       printf("kern.memorystatus_snapshot returning %ld for size\n", (long)currentsize);
-#endif
+               MEMORYSTATUS_DEBUG(1, "kern.memorystatus_snapshot returning %ld for size\n", (long)currentsize);
        }       
-       ret = sysctl_io_variable(req, &jetsam_snapshot, currentsize, 0, NULL);
+       ret = SYSCTL_OUT(req, &memorystatus_jetsam_snapshot, currentsize);
        if (!ret && req->oldptr) {
-               jetsam_snapshot.entry_count = jetsam_snapshot_list_count = 0;
+               memorystatus_jetsam_snapshot.entry_count = memorystatus_jetsam_snapshot_list_count = 0;
        }
        return ret;
 }
 
-SYSCTL_PROC(_kern, OID_AUTO, memorystatus_priority_list, CTLTYPE_OPAQUE|CTLFLAG_RW, 0, 0, sysctl_handle_kern_memorystatus_priority_list, "S,jetsam_priorities", "");
-SYSCTL_PROC(_kern, OID_AUTO, memorystatus_snapshot, CTLTYPE_OPAQUE|CTLFLAG_RD, 0, 0, sysctl_handle_kern_memorystatus_snapshot, "S,jetsam_snapshot", "");
+SYSCTL_PROC(_kern, OID_AUTO, memorystatus_snapshot, CTLTYPE_OPAQUE|CTLFLAG_RD, 0, 0, sysctl_handle_memorystatus_snapshot, "S,memorystatus_snapshot", "");
+
+#endif /* CONFIG_JETSAM */