#include <mach/boolean.h>
#include <kern/kern_types.h>
+#include <kern/locks.h>
#include <kern/host.h>
#include <kern/task.h>
#include <kern/zalloc.h>
#include <security/mac_framework.h>
#endif
+#include <pexpert/pexpert.h>
+
+void macx_init(void);
+
+static lck_grp_t *macx_lock_group;
+static lck_mtx_t *macx_lock;
+
/*
* temporary support for delayed instantiation
* of default_pager
/* ###################################################### */
+/*
+ * Routine: macx_init
+ * Function:
+ * Initialize locks so that only one caller can change
+ * state at a time.
+ */
+void
+macx_init(void)
+{
+ macx_lock_group = lck_grp_alloc_init("macx", NULL);
+ macx_lock = lck_mtx_alloc_init(macx_lock_group, NULL);
+}
/*
* Routine: macx_backing_store_recovery
int pid = args->pid;
int error;
struct proc *p = current_proc();
- boolean_t funnel_state;
- funnel_state = thread_funnel_set(kernel_flock, TRUE);
if ((error = suser(kauth_cred_get(), 0)))
goto backing_store_recovery_return;
task_backing_store_privileged(p->task);
backing_store_recovery_return:
- (void) thread_funnel_set(kernel_flock, FALSE);
return(error);
}
{
boolean_t suspend = args->suspend;
int error;
- boolean_t funnel_state;
- funnel_state = thread_funnel_set(kernel_flock, TRUE);
+ lck_mtx_lock(macx_lock);
if ((error = suser(kauth_cred_get(), 0)))
goto backing_store_suspend_return;
+ /* Multiple writers protected by macx_lock */
vm_backing_store_disable(suspend);
backing_store_suspend_return:
- (void) thread_funnel_set(kernel_flock, FALSE);
+ lck_mtx_unlock(macx_lock);
return(error);
}
extern boolean_t backing_store_stop_compaction;
+extern boolean_t compressor_store_stop_compaction;
/*
* Routine: macx_backing_store_compaction
* on by default when the system comes up and is turned
* off when a shutdown/restart is requested. It is
* re-enabled if the shutdown/restart is aborted for any reason.
+ *
+ * This routine assumes macx_lock has been locked by macx_triggers ->
+ * mach_macx_triggers -> macx_backing_store_compaction
*/
int
{
int error;
+ lck_mtx_assert(macx_lock, LCK_MTX_ASSERT_OWNED);
if ((error = suser(kauth_cred_get(), 0)))
return error;
if (flags & SWAP_COMPACT_DISABLE) {
backing_store_stop_compaction = TRUE;
+ compressor_store_stop_compaction = TRUE;
+
+ kprintf("backing_store_stop_compaction = TRUE\n");
} else if (flags & SWAP_COMPACT_ENABLE) {
backing_store_stop_compaction = FALSE;
+ compressor_store_stop_compaction = FALSE;
+
+ kprintf("backing_store_stop_compaction = FALSE\n");
}
return 0;
{
int error;
+ lck_mtx_lock(macx_lock);
error = suser(kauth_cred_get(), 0);
if (error)
return error;
- return mach_macx_triggers(args);
+ error = mach_macx_triggers(args);
+
+ lck_mtx_unlock(macx_lock);
+ return error;
}
extern boolean_t dp_isssd;
+/*
+ * In the compressed pager world, the swapfiles are created by the kernel.
+ * Well, all except the first one. That swapfile is absorbed by the kernel at
+ * the end of the macx_swapon function (if swap is enabled). That's why
+ * we allow the first invocation of macx_swapon to succeed.
+ *
+ * If the compressor pool is running low, the kernel messages the dynamic pager
+ * on the port it has registered with the kernel. That port can transport 1 of 2
+ * pieces of information to dynamic pager: create a swapfile or delete a swapfile.
+ *
+ * We choose to transmit the former. So, that message tells dynamic pager
+ * to create a swapfile and activate it by calling macx_swapon.
+ *
+ * We deny this new macx_swapon request. That leads dynamic pager to interpret the
+ * failure as a serious error and notify all it's clients that swap is running low.
+ * That's how we get the loginwindow "Resume / Force Quit Applications" dialog to appear.
+ *
+ * NOTE:
+ * If the kernel has already created multiple swapfiles by the time the compressor
+ * pool is running low (and it has to play this trick), dynamic pager won't be able to
+ * create a file in user-space and, that too will lead to a similar notification blast
+ * to all of it's clients. So, that behaves as desired too.
+ */
+boolean_t macx_swapon_allowed = TRUE;
+
/*
* Routine: macx_swapon
* Function:
mach_port_t backing_store;
memory_object_default_t default_pager;
int i;
- boolean_t funnel_state;
off_t file_size;
vfs_context_t ctx = vfs_context_current();
struct proc *p = current_proc();
int dp_cluster_size;
-
AUDIT_MACH_SYSCALL_ENTER(AUE_SWAPON);
AUDIT_ARG(value32, args->priority);
+
+ lck_mtx_lock(macx_lock);
+
+ if (COMPRESSED_PAGER_IS_ACTIVE) {
+ if (macx_swapon_allowed == FALSE) {
+ error = EINVAL;
+ goto swapon_bailout;
+ } else {
+ macx_swapon_allowed = FALSE;
+ error = 0;
+ goto swapon_bailout;
+ }
+ }
- funnel_state = thread_funnel_set(kernel_flock, TRUE);
ndp = &nd;
if ((error = suser(kauth_cred_get(), 0)))
#if CONFIG_PROTECT
{
- void *cnode = NULL;
/* initialize content protection keys manually */
- if ((cnode = cp_get_protected_cnode(vp)) != 0) {
- if ((error = cp_handle_vnop(cnode, CP_WRITE_ACCESS)) != 0)
- goto swapon_bailout;
- }
+ if ((error = cp_handle_vnop(vp, CP_WRITE_ACCESS, 0)) != 0) {
+ goto swapon_bailout;
+ }
}
#endif
goto swapon_bailout;
}
-#if CONFIG_EMBEDDED
- dp_cluster_size = 1 * PAGE_SIZE;
-#else
if ((dp_isssd = vnode_pager_isSSD(vp)) == TRUE) {
/*
* keep the cluster size small since the
*/
dp_cluster_size = 0;
}
-#endif
kr = default_pager_backing_store_create(default_pager,
-1, /* default priority */
dp_cluster_size,
if (vp) {
vnode_put(vp);
}
- (void) thread_funnel_set(kernel_flock, FALSE);
+ lck_mtx_unlock(macx_lock);
AUDIT_MACH_SYSCALL_EXIT(error);
if (error)
struct proc *p = current_proc();
int i;
int error;
- boolean_t funnel_state;
vfs_context_t ctx = vfs_context_current();
- struct uthread *ut;
int orig_iopol_disk;
AUDIT_MACH_SYSCALL_ENTER(AUE_SWAPOFF);
- funnel_state = thread_funnel_set(kernel_flock, TRUE);
+ lck_mtx_lock(macx_lock);
+
backing_store = NULL;
ndp = &nd;
}
backing_store = (mach_port_t)bs_port_table[i].bs;
- ut = get_bsdthread_info(current_thread());
+ orig_iopol_disk = proc_get_task_policy(current_task(), current_thread(),
+ TASK_POLICY_INTERNAL, TASK_POLICY_IOPOL);
-#if !CONFIG_EMBEDDED
- orig_iopol_disk = proc_get_thread_selfdiskacc();
- proc_apply_thread_selfdiskacc(IOPOL_THROTTLE);
-#else /* !CONFIG_EMBEDDED */
- orig_iopol_disk = ut->uu_iopol_disk;
- ut->uu_iopol_disk = IOPOL_THROTTLE;
-#endif /* !CONFIG_EMBEDDED */
+ proc_set_task_policy(current_task(), current_thread(), TASK_POLICY_INTERNAL,
+ TASK_POLICY_IOPOL, IOPOL_THROTTLE);
kr = default_pager_backing_store_delete(backing_store);
-#if !CONFIG_EMBEDDED
- proc_apply_thread_selfdiskacc(orig_iopol_disk);
-#else /* !CONFIG_EMBEDDED */
- ut->uu_iopol_disk = orig_iopol_disk;
-#endif /* !CONFIG_EMBEDDED */
+ proc_set_task_policy(current_task(), current_thread(), TASK_POLICY_INTERNAL,
+ TASK_POLICY_IOPOL, orig_iopol_disk);
switch (kr) {
case KERN_SUCCESS:
/* get rid of macx_swapoff() namei() reference */
if (vp)
vnode_put(vp);
-
- (void) thread_funnel_set(kernel_flock, FALSE);
+ lck_mtx_unlock(macx_lock);
AUDIT_MACH_SYSCALL_EXIT(error);
if (error)
* Function:
* Syscall interface to get general swap statistics
*/
+extern uint64_t vm_swap_get_total_space(void);
+extern uint64_t vm_swap_get_used_space(void);
+extern uint64_t vm_swap_get_free_space(void);
+extern boolean_t vm_swap_up;
+
int
macx_swapinfo(
memory_object_size_t *total_p,
kern_return_t kr;
error = 0;
+ if (COMPRESSED_PAGER_IS_ACTIVE || DEFAULT_FREEZER_COMPRESSED_PAGER_IS_ACTIVE) {
+
+ if (vm_swap_up == TRUE) {
+
+ *total_p = vm_swap_get_total_space();
+ *avail_p = vm_swap_get_free_space();
+ *pagesize_p = (vm_size_t)PAGE_SIZE_64;
+ *encrypted_p = TRUE;
+
+ } else {
+
+ *total_p = 0;
+ *avail_p = 0;
+ *pagesize_p = 0;
+ *encrypted_p = FALSE;
+ }
+ } else {
- /*
- * Get a handle on the default pager.
- */
- default_pager = MEMORY_OBJECT_DEFAULT_NULL;
- kr = host_default_memory_manager(host_priv_self(), &default_pager, 0);
- if (kr != KERN_SUCCESS) {
- error = EAGAIN; /* XXX why EAGAIN ? */
- goto done;
- }
- if (default_pager == MEMORY_OBJECT_DEFAULT_NULL) {
/*
- * The default pager has not initialized yet,
- * so it can't be using any swap space at all.
+ * Get a handle on the default pager.
*/
- *total_p = 0;
- *avail_p = 0;
- *pagesize_p = 0;
- *encrypted_p = FALSE;
- goto done;
- }
-
- /*
- * Get swap usage data from default pager.
- */
- kr = default_pager_info_64(default_pager, &dpi64);
- if (kr != KERN_SUCCESS) {
- error = ENOTSUP;
- goto done;
- }
+ default_pager = MEMORY_OBJECT_DEFAULT_NULL;
+ kr = host_default_memory_manager(host_priv_self(), &default_pager, 0);
+ if (kr != KERN_SUCCESS) {
+ error = EAGAIN; /* XXX why EAGAIN ? */
+ goto done;
+ }
+ if (default_pager == MEMORY_OBJECT_DEFAULT_NULL) {
+ /*
+ * The default pager has not initialized yet,
+ * so it can't be using any swap space at all.
+ */
+ *total_p = 0;
+ *avail_p = 0;
+ *pagesize_p = 0;
+ *encrypted_p = FALSE;
+ goto done;
+ }
+
+ /*
+ * Get swap usage data from default pager.
+ */
+ kr = default_pager_info_64(default_pager, &dpi64);
+ if (kr != KERN_SUCCESS) {
+ error = ENOTSUP;
+ goto done;
+ }
- /*
- * Provide default pager info to caller.
- */
- *total_p = dpi64.dpi_total_space;
- *avail_p = dpi64.dpi_free_space;
- *pagesize_p = dpi64.dpi_page_size;
- if (dpi64.dpi_flags & DPI_ENCRYPTED) {
- *encrypted_p = TRUE;
- } else {
- *encrypted_p = FALSE;
- }
+ /*
+ * Provide default pager info to caller.
+ */
+ *total_p = dpi64.dpi_total_space;
+ *avail_p = dpi64.dpi_free_space;
+ *pagesize_p = dpi64.dpi_page_size;
+ if (dpi64.dpi_flags & DPI_ENCRYPTED) {
+ *encrypted_p = TRUE;
+ } else {
+ *encrypted_p = FALSE;
+ }
done:
- if (default_pager != MEMORY_OBJECT_DEFAULT_NULL) {
- /* release our handle on default pager */
- memory_object_default_deallocate(default_pager);
+ if (default_pager != MEMORY_OBJECT_DEFAULT_NULL) {
+ /* release our handle on default pager */
+ memory_object_default_deallocate(default_pager);
+ }
}
return error;
}