2 * Copyright (c) 2000-2010 Apple Inc. All rights reserved.
4 * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
6 * This file contains Original Code and/or Modifications of Original Code
7 * as defined in and that are subject to the Apple Public Source License
8 * Version 2.0 (the 'License'). You may not use this file except in
9 * compliance with the License. The rights granted to you under the License
10 * may not be used to create, or enable the creation or redistribution of,
11 * unlawful or unlicensed copies of an Apple operating system, or to
12 * circumvent, violate, or enable the circumvention or violation of, any
13 * terms of an Apple operating system software license agreement.
15 * Please obtain a copy of the License at
16 * http://www.opensource.apple.com/apsl/ and read it before using this file.
18 * The Original Code and all software distributed under the License are
19 * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
20 * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
21 * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
22 * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
23 * Please see the License for the specific language governing rights and
24 * limitations under the License.
26 * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
29 * NOTICE: This file was modified by SPARTA, Inc. in 2005 to introduce
30 * support for mandatory and extensible security protections. This notice
31 * is included in support of clause 2.2 (b) of the Apple Public License,
35 #include <sys/param.h>
36 #include <sys/systm.h>
38 #include <sys/proc_internal.h>
39 #include <sys/kauth.h>
42 #include <sys/vnode_internal.h>
43 #include <sys/namei.h>
44 #include <sys/ubc_internal.h>
45 #include <sys/malloc.h>
48 #include <sys/cprotect.h>
51 #include <default_pager/default_pager_types.h>
52 #include <default_pager/default_pager_object.h>
54 #include <security/audit/audit.h>
55 #include <bsm/audit_kevents.h>
57 #include <mach/mach_types.h>
58 #include <mach/host_priv.h>
59 #include <mach/mach_traps.h>
60 #include <mach/boolean.h>
62 #include <kern/kern_types.h>
63 #include <kern/locks.h>
64 #include <kern/host.h>
65 #include <kern/task.h>
66 #include <kern/zalloc.h>
67 #include <kern/kalloc.h>
68 #include <kern/assert.h>
70 #include <libkern/libkern.h>
72 #include <vm/vm_pageout.h>
73 #include <vm/vm_map.h>
74 #include <vm/vm_kern.h>
75 #include <vm/vnode_pager.h>
76 #include <vm/vm_protos.h>
78 #include <security/mac_framework.h>
81 #include <pexpert/pexpert.h>
85 static lck_grp_t
*macx_lock_group
;
86 static lck_mtx_t
*macx_lock
;
89 * temporary support for delayed instantiation
92 int default_pager_init_flag
= 0;
94 struct bs_map bs_port_table
[MAX_BACKING_STORE
] = {
95 {0,0},{0,0},{0,0},{0,0},{0,0},{0,0},{0,0},{0,0},{0,0},{0,0},
96 {0,0},{0,0},{0,0},{0,0},{0,0},{0,0},{0,0},{0,0},{0,0},{0,0},
97 {0,0},{0,0},{0,0},{0,0},{0,0},{0,0},{0,0},{0,0},{0,0},{0,0},
98 {0,0},{0,0},{0,0},{0,0},{0,0},{0,0},{0,0},{0,0},{0,0},{0,0},
99 {0,0},{0,0},{0,0},{0,0},{0,0},{0,0},{0,0},{0,0},{0,0},{0,0},
100 {0,0},{0,0},{0,0},{0,0},{0,0},{0,0},{0,0},{0,0},{0,0},{0,0},
101 {0,0},{0,0},{0,0},{0,0},{0,0},{0,0},{0,0},{0,0},{0,0},{0,0},
102 {0,0},{0,0},{0,0},{0,0},{0,0},{0,0},{0,0},{0,0},{0,0},{0,0},
103 {0,0},{0,0},{0,0},{0,0},{0,0},{0,0},{0,0},{0,0},{0,0},{0,0},
104 {0,0},{0,0},{0,0},{0,0},{0,0},{0,0},{0,0},{0,0},{0,0},{0,0}};
106 /* ###################################################### */
111 * Initialize locks so that only one caller can change
117 macx_lock_group
= lck_grp_alloc_init("macx", NULL
);
118 macx_lock
= lck_mtx_alloc_init(macx_lock_group
, NULL
);
122 * Routine: macx_backing_store_recovery
124 * Syscall interface to set a tasks privilege
125 * level so that it is not subject to
126 * macx_backing_store_suspend
129 macx_backing_store_recovery(
130 struct macx_backing_store_recovery_args
*args
)
134 struct proc
*p
= current_proc();
136 if ((error
= suser(kauth_cred_get(), 0)))
137 goto backing_store_recovery_return
;
139 /* for now restrict backing_store_recovery */
140 /* usage to only present task */
141 if(pid
!= proc_selfpid()) {
143 goto backing_store_recovery_return
;
146 task_backing_store_privileged(p
->task
);
148 backing_store_recovery_return
:
153 * Routine: macx_backing_store_suspend
155 * Syscall interface to stop new demand for
156 * backing store when backing store is low
160 macx_backing_store_suspend(
161 struct macx_backing_store_suspend_args
*args
)
163 boolean_t suspend
= args
->suspend
;
166 lck_mtx_lock(macx_lock
);
167 if ((error
= suser(kauth_cred_get(), 0)))
168 goto backing_store_suspend_return
;
170 /* Multiple writers protected by macx_lock */
171 vm_backing_store_disable(suspend
);
173 backing_store_suspend_return
:
174 lck_mtx_unlock(macx_lock
);
178 extern boolean_t backing_store_stop_compaction
;
179 extern boolean_t compressor_store_stop_compaction
;
182 * Routine: macx_backing_store_compaction
184 * Turn compaction of swap space on or off. This is
185 * used during shutdown/restart so that the kernel
186 * doesn't waste time compacting swap files that are
187 * about to be deleted anyway. Compaction is always
188 * on by default when the system comes up and is turned
189 * off when a shutdown/restart is requested. It is
190 * re-enabled if the shutdown/restart is aborted for any reason.
192 * This routine assumes macx_lock has been locked by macx_triggers ->
193 * mach_macx_triggers -> macx_backing_store_compaction
197 macx_backing_store_compaction(int flags
)
201 lck_mtx_assert(macx_lock
, LCK_MTX_ASSERT_OWNED
);
202 if ((error
= suser(kauth_cred_get(), 0)))
205 if (flags
& SWAP_COMPACT_DISABLE
) {
206 backing_store_stop_compaction
= TRUE
;
207 compressor_store_stop_compaction
= TRUE
;
209 kprintf("backing_store_stop_compaction = TRUE\n");
211 } else if (flags
& SWAP_COMPACT_ENABLE
) {
212 backing_store_stop_compaction
= FALSE
;
213 compressor_store_stop_compaction
= FALSE
;
215 kprintf("backing_store_stop_compaction = FALSE\n");
222 * Routine: macx_triggers
224 * Syscall interface to set the call backs for low and
229 struct macx_triggers_args
*args
)
233 lck_mtx_lock(macx_lock
);
234 error
= suser(kauth_cred_get(), 0);
238 error
= mach_macx_triggers(args
);
240 lck_mtx_unlock(macx_lock
);
245 extern boolean_t dp_isssd
;
248 * In the compressed pager world, the swapfiles are created by the kernel.
249 * Well, all except the first one. That swapfile is absorbed by the kernel at
250 * the end of the macx_swapon function (if swap is enabled). That's why
251 * we allow the first invocation of macx_swapon to succeed.
253 * If the compressor pool is running low, the kernel messages the dynamic pager
254 * on the port it has registered with the kernel. That port can transport 1 of 2
255 * pieces of information to dynamic pager: create a swapfile or delete a swapfile.
257 * We choose to transmit the former. So, that message tells dynamic pager
258 * to create a swapfile and activate it by calling macx_swapon.
260 * We deny this new macx_swapon request. That leads dynamic pager to interpret the
261 * failure as a serious error and notify all it's clients that swap is running low.
262 * That's how we get the loginwindow "Resume / Force Quit Applications" dialog to appear.
265 * If the kernel has already created multiple swapfiles by the time the compressor
266 * pool is running low (and it has to play this trick), dynamic pager won't be able to
267 * create a file in user-space and, that too will lead to a similar notification blast
268 * to all of it's clients. So, that behaves as desired too.
270 boolean_t macx_swapon_allowed
= TRUE
;
273 * Routine: macx_swapon
275 * Syscall interface to add a file to backing store
279 struct macx_swapon_args
*args
)
281 int size
= args
->size
;
282 vnode_t vp
= (vnode_t
)NULL
;
283 struct nameidata nd
, *ndp
;
286 mach_port_t backing_store
;
287 memory_object_default_t default_pager
;
290 vfs_context_t ctx
= vfs_context_current();
291 struct proc
*p
= current_proc();
294 AUDIT_MACH_SYSCALL_ENTER(AUE_SWAPON
);
295 AUDIT_ARG(value32
, args
->priority
);
297 lck_mtx_lock(macx_lock
);
299 if (COMPRESSED_PAGER_IS_ACTIVE
) {
300 if (macx_swapon_allowed
== FALSE
) {
304 macx_swapon_allowed
= FALSE
;
312 if ((error
= suser(kauth_cred_get(), 0)))
316 * Get a vnode for the paging area.
318 NDINIT(ndp
, LOOKUP
, OP_LOOKUP
, FOLLOW
| LOCKLEAF
| AUDITVNPATH1
,
319 ((IS_64BIT_PROCESS(p
)) ? UIO_USERSPACE64
: UIO_USERSPACE32
),
320 (user_addr_t
) args
->filename
, ctx
);
322 if ((error
= namei(ndp
)))
327 if (vp
->v_type
!= VREG
) {
333 if ((error
= vnode_size(vp
, &file_size
, ctx
)) != 0)
337 error
= mac_system_check_swapon(vfs_context_ucred(ctx
), vp
);
343 /* resize to desired size if it's too small */
344 if ((file_size
< (off_t
)size
) && ((error
= vnode_setsize(vp
, (off_t
)size
, 0, ctx
)) != 0))
349 /* initialize content protection keys manually */
350 if ((error
= cp_handle_vnop(vp
, CP_WRITE_ACCESS
, 0)) != 0) {
357 if (default_pager_init_flag
== 0) {
358 start_def_pager(NULL
);
359 default_pager_init_flag
= 1;
362 /* add new backing store to list */
364 while(bs_port_table
[i
].vp
!= 0) {
365 if(i
== MAX_BACKING_STORE
)
369 if(i
== MAX_BACKING_STORE
) {
374 /* remember the vnode. This vnode has namei() reference */
375 bs_port_table
[i
].vp
= vp
;
378 * Look to see if we are already paging to this file.
380 /* make certain the copy send of kernel call will work */
381 default_pager
= MEMORY_OBJECT_DEFAULT_NULL
;
382 kr
= host_default_memory_manager(host_priv_self(), &default_pager
, 0);
383 if(kr
!= KERN_SUCCESS
) {
385 bs_port_table
[i
].vp
= 0;
389 if ((dp_isssd
= vnode_pager_isSSD(vp
)) == TRUE
) {
391 * keep the cluster size small since the
392 * seek cost is effectively 0 which means
393 * we don't care much about fragmentation
395 dp_cluster_size
= 2 * PAGE_SIZE
;
398 * use the default cluster size
402 kr
= default_pager_backing_store_create(default_pager
,
403 -1, /* default priority */
406 memory_object_default_deallocate(default_pager
);
408 if(kr
!= KERN_SUCCESS
) {
410 bs_port_table
[i
].vp
= 0;
414 /* Mark this vnode as being used for swapfile */
416 SET(vp
->v_flag
, VSWAP
);
420 * NOTE: we are able to supply PAGE_SIZE here instead of
421 * an actual record size or block number because:
422 * a: we do not support offsets from the beginning of the
423 * file (allowing for non page size/record modulo offsets.
424 * b: because allow paging will be done modulo page size
427 kr
= default_pager_add_file(backing_store
, (vnode_ptr_t
) vp
,
428 PAGE_SIZE
, (int)(file_size
/PAGE_SIZE
));
429 if(kr
!= KERN_SUCCESS
) {
430 bs_port_table
[i
].vp
= 0;
431 if(kr
== KERN_INVALID_ARGUMENT
)
436 /* This vnode is not to be used for swapfile */
438 CLR(vp
->v_flag
, VSWAP
);
443 bs_port_table
[i
].bs
= (void *)backing_store
;
446 ubc_setthreadcred(vp
, p
, current_thread());
449 * take a long term reference on the vnode to keep
450 * vnreclaim() away from this vnode.
458 lck_mtx_unlock(macx_lock
);
459 AUDIT_MACH_SYSCALL_EXIT(error
);
462 printf("macx_swapon FAILED - %d\n", error
);
464 printf("macx_swapon SUCCESS\n");
470 * Routine: macx_swapoff
472 * Syscall interface to remove a file from backing store
476 struct macx_swapoff_args
*args
)
478 __unused
int flags
= args
->flags
;
480 mach_port_t backing_store
;
482 struct vnode
*vp
= 0;
483 struct nameidata nd
, *ndp
;
484 struct proc
*p
= current_proc();
487 vfs_context_t ctx
= vfs_context_current();
490 AUDIT_MACH_SYSCALL_ENTER(AUE_SWAPOFF
);
492 lck_mtx_lock(macx_lock
);
494 backing_store
= NULL
;
497 if ((error
= suser(kauth_cred_get(), 0)))
498 goto swapoff_bailout
;
501 * Get the vnode for the paging area.
503 NDINIT(ndp
, LOOKUP
, OP_LOOKUP
, FOLLOW
| LOCKLEAF
| AUDITVNPATH1
,
504 ((IS_64BIT_PROCESS(p
)) ? UIO_USERSPACE64
: UIO_USERSPACE32
),
505 (user_addr_t
) args
->filename
, ctx
);
507 if ((error
= namei(ndp
)))
508 goto swapoff_bailout
;
512 if (vp
->v_type
!= VREG
) {
514 goto swapoff_bailout
;
518 error
= mac_system_check_swapoff(vfs_context_ucred(ctx
), vp
);
521 goto swapoff_bailout
;
524 for(i
= 0; i
< MAX_BACKING_STORE
; i
++) {
525 if(bs_port_table
[i
].vp
== vp
) {
529 if (i
== MAX_BACKING_STORE
) {
531 goto swapoff_bailout
;
533 backing_store
= (mach_port_t
)bs_port_table
[i
].bs
;
535 orig_iopol_disk
= proc_get_task_policy(current_task(), current_thread(),
536 TASK_POLICY_INTERNAL
, TASK_POLICY_IOPOL
);
538 proc_set_task_policy(current_task(), current_thread(), TASK_POLICY_INTERNAL
,
539 TASK_POLICY_IOPOL
, IOPOL_THROTTLE
);
541 kr
= default_pager_backing_store_delete(backing_store
);
543 proc_set_task_policy(current_task(), current_thread(), TASK_POLICY_INTERNAL
,
544 TASK_POLICY_IOPOL
, orig_iopol_disk
);
549 bs_port_table
[i
].vp
= 0;
550 /* This vnode is no longer used for swapfile */
552 CLR(vp
->v_flag
, VSWAP
);
555 /* get rid of macx_swapon() "long term" reference */
568 /* get rid of macx_swapoff() namei() reference */
571 lck_mtx_unlock(macx_lock
);
572 AUDIT_MACH_SYSCALL_EXIT(error
);
575 printf("macx_swapoff FAILED - %d\n", error
);
577 printf("macx_swapoff SUCCESS\n");
583 * Routine: macx_swapinfo
585 * Syscall interface to get general swap statistics
587 extern uint64_t vm_swap_get_total_space(void);
588 extern uint64_t vm_swap_get_used_space(void);
589 extern uint64_t vm_swap_get_free_space(void);
590 extern boolean_t vm_swap_up
;
594 memory_object_size_t
*total_p
,
595 memory_object_size_t
*avail_p
,
596 vm_size_t
*pagesize_p
,
597 boolean_t
*encrypted_p
)
600 memory_object_default_t default_pager
;
601 default_pager_info_64_t dpi64
;
605 if (COMPRESSED_PAGER_IS_ACTIVE
|| DEFAULT_FREEZER_COMPRESSED_PAGER_IS_ACTIVE
) {
607 if (vm_swap_up
== TRUE
) {
609 *total_p
= vm_swap_get_total_space();
610 *avail_p
= vm_swap_get_free_space();
611 *pagesize_p
= (vm_size_t
)PAGE_SIZE_64
;
619 *encrypted_p
= FALSE
;
624 * Get a handle on the default pager.
626 default_pager
= MEMORY_OBJECT_DEFAULT_NULL
;
627 kr
= host_default_memory_manager(host_priv_self(), &default_pager
, 0);
628 if (kr
!= KERN_SUCCESS
) {
629 error
= EAGAIN
; /* XXX why EAGAIN ? */
632 if (default_pager
== MEMORY_OBJECT_DEFAULT_NULL
) {
634 * The default pager has not initialized yet,
635 * so it can't be using any swap space at all.
640 *encrypted_p
= FALSE
;
645 * Get swap usage data from default pager.
647 kr
= default_pager_info_64(default_pager
, &dpi64
);
648 if (kr
!= KERN_SUCCESS
) {
654 * Provide default pager info to caller.
656 *total_p
= dpi64
.dpi_total_space
;
657 *avail_p
= dpi64
.dpi_free_space
;
658 *pagesize_p
= dpi64
.dpi_page_size
;
659 if (dpi64
.dpi_flags
& DPI_ENCRYPTED
) {
662 *encrypted_p
= FALSE
;
666 if (default_pager
!= MEMORY_OBJECT_DEFAULT_NULL
) {
667 /* release our handle on default pager */
668 memory_object_default_deallocate(default_pager
);