]> git.saurik.com Git - apple/xnu.git/blob - osfmk/vm/vm_compressor.h
xnu-3248.30.4.tar.gz
[apple/xnu.git] / osfmk / vm / vm_compressor.h
1 /*
2 * Copyright (c) 2000-2013 Apple Inc. All rights reserved.
3 *
4 * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
5 *
6 * This file contains Original Code and/or Modifications of Original Code
7 * as defined in and that are subject to the Apple Public Source License
8 * Version 2.0 (the 'License'). You may not use this file except in
9 * compliance with the License. The rights granted to you under the License
10 * may not be used to create, or enable the creation or redistribution of,
11 * unlawful or unlicensed copies of an Apple operating system, or to
12 * circumvent, violate, or enable the circumvention or violation of, any
13 * terms of an Apple operating system software license agreement.
14 *
15 * Please obtain a copy of the License at
16 * http://www.opensource.apple.com/apsl/ and read it before using this file.
17 *
18 * The Original Code and all software distributed under the License are
19 * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
20 * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
21 * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
22 * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
23 * Please see the License for the specific language governing rights and
24 * limitations under the License.
25 *
26 * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
27 */
28
29 #include <kern/kalloc.h>
30 #include <vm/vm_compressor_pager.h>
31 #include <vm/vm_kern.h>
32 #include <vm/vm_page.h>
33 #include <vm/vm_protos.h>
34 #include <vm/WKdm_new.h>
35 #include <vm/vm_object.h>
36 #include <machine/pmap.h>
37 #include <kern/locks.h>
38
39 #include <sys/kdebug.h>
40
41
42 #define C_SEG_OFFSET_BITS 16
43 #define C_SEG_BUFSIZE (1024 * 256)
44 #define C_SEG_MAX_PAGES (C_SEG_BUFSIZE / PAGE_SIZE)
45
46 #define C_SEG_OFF_LIMIT (C_SEG_BYTES_TO_OFFSET((C_SEG_BUFSIZE - 128)))
47 #define C_SEG_ALLOCSIZE (C_SEG_BUFSIZE)
48 #define C_SEG_MAX_POPULATE_SIZE (4 * PAGE_SIZE)
49
50
51 #define CHECKSUM_THE_SWAP 0 /* Debug swap data */
52 #define CHECKSUM_THE_DATA 0 /* Debug compressor/decompressor data */
53 #define CHECKSUM_THE_COMPRESSED_DATA 0 /* Debug compressor/decompressor compressed data */
54 #define VALIDATE_C_SEGMENTS 0 /* Debug compaction */
55
56 #define RECORD_THE_COMPRESSED_DATA 0
57
58
59
60 struct c_slot {
61 uint64_t c_offset:C_SEG_OFFSET_BITS,
62 c_size:12,
63 c_packed_ptr:36;
64 #if CHECKSUM_THE_DATA
65 unsigned int c_hash_data;
66 #endif
67 #if CHECKSUM_THE_COMPRESSED_DATA
68 unsigned int c_hash_compressed_data;
69 #endif
70
71 };
72
73 #define C_IS_EMPTY 0
74 #define C_IS_FREE 1
75 #define C_IS_FILLING 2
76 #define C_ON_AGE_Q 3
77 #define C_ON_SWAPOUT_Q 4
78 #define C_ON_SWAPPEDOUT_Q 5
79 #define C_ON_SWAPPEDOUTSPARSE_Q 6
80 #define C_ON_SWAPPEDIN_Q 7
81 #define C_ON_MAJORCOMPACT_Q 8
82 #define C_ON_BAD_Q 9
83
84
85 struct c_segment {
86 #if __i386__ || __x86_64__
87 lck_mtx_t c_lock;
88 #else /* __i386__ || __x86_64__ */
89 lck_spin_t c_lock;
90 #endif /* __i386__ || __x86_64__ */
91 queue_chain_t c_age_list;
92 queue_chain_t c_list;
93
94 uint64_t c_generation_id;
95 int32_t c_bytes_used;
96 int32_t c_bytes_unused;
97
98 #define C_SEG_MAX_LIMIT (1 << 19) /* this needs to track the size of c_mysegno */
99 uint32_t c_mysegno:19,
100 c_busy:1,
101 c_busy_swapping:1,
102 c_wanted:1,
103 c_on_minorcompact_q:1, /* can also be on the age_q, the majorcompact_q or the swappedin_q */
104
105 c_state:4, /* what state is the segment in which dictates which q to find it on */
106 c_overage_swap:1,
107 c_reserved:4;
108
109 uint16_t c_firstemptyslot;
110 uint16_t c_nextslot;
111 uint32_t c_nextoffset;
112 uint32_t c_populated_offset;
113
114 uint32_t c_creation_ts;
115 uint32_t c_swappedin_ts;
116
117 union {
118 int32_t *c_buffer;
119 uint64_t c_swap_handle;
120 } c_store;
121
122 #if VALIDATE_C_SEGMENTS
123 uint32_t c_was_minor_compacted;
124 uint32_t c_was_major_compacted;
125 uint32_t c_was_major_donor;
126 #endif
127 #if CHECKSUM_THE_SWAP
128 unsigned int cseg_hash;
129 unsigned int cseg_swap_size;
130 #endif /* CHECKSUM_THE_SWAP */
131
132 #if MACH_ASSERT
133 thread_t c_busy_for_thread;
134 #endif /* MACH_ASSERT */
135
136 int c_slot_var_array_len;
137 struct c_slot *c_slot_var_array;
138 struct c_slot c_slot_fixed_array[0];
139 };
140
141 #define C_SEG_SLOT_VAR_ARRAY_MIN_LEN C_SEG_MAX_PAGES
142
143 extern int c_seg_fixed_array_len;
144 extern vm_offset_t c_buffers;
145 #define C_SEG_BUFFER_ADDRESS(c_segno) ((c_buffers + ((uint64_t)c_segno * (uint64_t)C_SEG_ALLOCSIZE)))
146
147 #define C_SEG_SLOT_FROM_INDEX(cseg, index) (index < c_seg_fixed_array_len ? &(cseg->c_slot_fixed_array[index]) : &(cseg->c_slot_var_array[index - c_seg_fixed_array_len]))
148
149 #define C_SEG_OFFSET_TO_BYTES(off) ((off) * (int) sizeof(int32_t))
150 #define C_SEG_BYTES_TO_OFFSET(bytes) ((bytes) / (int) sizeof(int32_t))
151
152 #define C_SEG_UNUSED_BYTES(cseg) (cseg->c_bytes_unused + (C_SEG_OFFSET_TO_BYTES(cseg->c_populated_offset - cseg->c_nextoffset)))
153
154 #define C_SEG_OFFSET_ALIGNMENT_MASK 0x3
155
156 #define C_SEG_ONDISK_IS_SPARSE(cseg) ((cseg->c_bytes_used < (C_SEG_BUFSIZE / 2)) ? 1 : 0)
157 #define C_SEG_SHOULD_MINORCOMPACT(cseg) ((C_SEG_UNUSED_BYTES(cseg) >= (C_SEG_BUFSIZE / 3)) ? 1 : 0)
158 #define C_SEG_SHOULD_MAJORCOMPACT(cseg) (((cseg->c_bytes_unused + (C_SEG_BUFSIZE - C_SEG_OFFSET_TO_BYTES(c_seg->c_nextoffset))) >= (C_SEG_BUFSIZE / 8)) ? 1 : 0)
159
160 #define C_SEG_IS_ONDISK(cseg) ((cseg->c_state == C_ON_SWAPPEDOUT_Q || cseg->c_state == C_ON_SWAPPEDOUTSPARSE_Q))
161
162
163 #define C_SEG_WAKEUP_DONE(cseg) \
164 MACRO_BEGIN \
165 assert((cseg)->c_busy); \
166 (cseg)->c_busy = 0; \
167 assert((cseg)->c_busy_for_thread != NULL); \
168 assert((((cseg)->c_busy_for_thread = NULL), TRUE)); \
169 if ((cseg)->c_wanted) { \
170 (cseg)->c_wanted = 0; \
171 thread_wakeup((event_t) (cseg)); \
172 } \
173 MACRO_END
174
175 #define C_SEG_BUSY(cseg) \
176 MACRO_BEGIN \
177 assert((cseg)->c_busy == 0); \
178 (cseg)->c_busy = 1; \
179 assert((cseg)->c_busy_for_thread == NULL); \
180 assert((((cseg)->c_busy_for_thread = current_thread()), TRUE)); \
181 MACRO_END
182
183
184
185 typedef struct c_segment *c_segment_t;
186 typedef struct c_slot *c_slot_t;
187
188 uint64_t vm_compressor_total_compressions(void);
189 void vm_wake_compactor_swapper(void);
190 void vm_thrashing_jetsam_done(void);
191 void vm_consider_waking_compactor_swapper(void);
192 void vm_consider_swapping(void);
193 void vm_compressor_flush(void);
194 void c_seg_free(c_segment_t);
195 void c_seg_free_locked(c_segment_t);
196 void c_seg_insert_into_age_q(c_segment_t);
197
198 void vm_decompressor_lock(void);
199 void vm_decompressor_unlock(void);
200
201 void vm_compressor_delay_trim(void);
202 void vm_compressor_do_warmup(void);
203 void vm_compressor_record_warmup_start(void);
204 void vm_compressor_record_warmup_end(void);
205
206 int vm_wants_task_throttled(task_t);
207 boolean_t vm_compression_available(void);
208
209 extern void vm_compressor_swap_init(void);
210 extern void vm_compressor_init_locks(void);
211 extern lck_rw_t c_master_lock;
212
213 #if ENCRYPTED_SWAP
214 extern void vm_swap_decrypt(c_segment_t);
215 #endif /* ENCRYPTED_SWAP */
216
217 extern int vm_swap_low_on_space(void);
218 extern kern_return_t vm_swap_get(vm_offset_t, uint64_t, uint64_t);
219 extern void vm_swap_free(uint64_t);
220 extern void vm_swap_consider_defragmenting(void);
221
222 extern void c_seg_swapin_requeue(c_segment_t, boolean_t);
223 extern void c_seg_swapin(c_segment_t, boolean_t);
224 extern void c_seg_wait_on_busy(c_segment_t);
225 extern void c_seg_trim_tail(c_segment_t);
226 extern void c_seg_switch_state(c_segment_t, int, boolean_t);
227
228 extern boolean_t fastwake_recording_in_progress;
229 extern int compaction_swapper_running;
230 extern uint64_t vm_swap_put_failures;
231
232 extern int c_overage_swapped_count;
233 extern int c_overage_swapped_limit;
234
235 extern queue_head_t c_minor_list_head;
236 extern queue_head_t c_age_list_head;
237 extern queue_head_t c_swapout_list_head;
238 extern queue_head_t c_swappedout_list_head;
239 extern queue_head_t c_swappedout_sparse_list_head;
240
241 extern uint32_t c_age_count;
242 extern uint32_t c_swapout_count;
243 extern uint32_t c_swappedout_count;
244 extern uint32_t c_swappedout_sparse_count;
245
246 extern int64_t compressor_bytes_used;
247 extern uint64_t first_c_segment_to_warm_generation_id;
248 extern uint64_t last_c_segment_to_warm_generation_id;
249 extern boolean_t hibernate_flushing;
250 extern boolean_t hibernate_no_swapspace;
251 extern uint32_t swapout_target_age;
252
253 extern void c_seg_insert_into_q(queue_head_t *, c_segment_t);
254
255 extern uint32_t vm_compressor_minorcompact_threshold_divisor;
256 extern uint32_t vm_compressor_majorcompact_threshold_divisor;
257 extern uint32_t vm_compressor_unthrottle_threshold_divisor;
258 extern uint32_t vm_compressor_catchup_threshold_divisor;
259 extern uint64_t vm_compressor_compute_elapsed_msecs(clock_sec_t, clock_nsec_t, clock_sec_t, clock_nsec_t);
260
261 #define PAGE_REPLACEMENT_DISALLOWED(enable) (enable == TRUE ? lck_rw_lock_shared(&c_master_lock) : lck_rw_done(&c_master_lock))
262 #define PAGE_REPLACEMENT_ALLOWED(enable) (enable == TRUE ? lck_rw_lock_exclusive(&c_master_lock) : lck_rw_done(&c_master_lock))
263
264
265 #define AVAILABLE_NON_COMPRESSED_MEMORY (vm_page_active_count + vm_page_inactive_count + vm_page_free_count + vm_page_speculative_count)
266 #define AVAILABLE_MEMORY (AVAILABLE_NON_COMPRESSED_MEMORY + VM_PAGE_COMPRESSOR_COUNT)
267
268 #define VM_PAGE_COMPRESSOR_COMPACT_THRESHOLD (((AVAILABLE_MEMORY) * 10) / (vm_compressor_minorcompact_threshold_divisor ? vm_compressor_minorcompact_threshold_divisor : 1))
269 #define VM_PAGE_COMPRESSOR_SWAP_THRESHOLD (((AVAILABLE_MEMORY) * 10) / (vm_compressor_majorcompact_threshold_divisor ? vm_compressor_majorcompact_threshold_divisor : 1))
270 #define VM_PAGE_COMPRESSOR_SWAP_UNTHROTTLE_THRESHOLD (((AVAILABLE_MEMORY) * 10) / (vm_compressor_unthrottle_threshold_divisor ? vm_compressor_unthrottle_threshold_divisor : 1))
271 #define VM_PAGE_COMPRESSOR_SWAP_CATCHUP_THRESHOLD (((AVAILABLE_MEMORY) * 10) / (vm_compressor_catchup_threshold_divisor ? vm_compressor_catchup_threshold_divisor : 1))
272
273 #define COMPRESSOR_NEEDS_TO_SWAP() ((AVAILABLE_NON_COMPRESSED_MEMORY < VM_PAGE_COMPRESSOR_SWAP_THRESHOLD) ? 1 : 0)
274
275 #define VM_PAGEOUT_SCAN_NEEDS_TO_THROTTLE() \
276 (vm_compressor_mode == VM_PAGER_COMPRESSOR_WITH_SWAP && \
277 ((AVAILABLE_NON_COMPRESSED_MEMORY < VM_PAGE_COMPRESSOR_SWAP_CATCHUP_THRESHOLD) ? 1 : 0))
278 #define HARD_THROTTLE_LIMIT_REACHED() ((AVAILABLE_NON_COMPRESSED_MEMORY < (VM_PAGE_COMPRESSOR_SWAP_UNTHROTTLE_THRESHOLD) / 2) ? 1 : 0)
279 #define SWAPPER_NEEDS_TO_UNTHROTTLE() ((AVAILABLE_NON_COMPRESSED_MEMORY < VM_PAGE_COMPRESSOR_SWAP_UNTHROTTLE_THRESHOLD) ? 1 : 0)
280 #define COMPRESSOR_NEEDS_TO_MINOR_COMPACT() ((AVAILABLE_NON_COMPRESSED_MEMORY < VM_PAGE_COMPRESSOR_COMPACT_THRESHOLD) ? 1 : 0)
281
282 /*
283 * indicate the need to do a major compaction if
284 * the overall set of in-use compression segments
285 * becomes sparse... on systems that support pressure
286 * driven swapping, this will also cause swapouts to
287 * be initiated.
288 */
289 #define COMPRESSOR_NEEDS_TO_MAJOR_COMPACT() (((c_segment_count >= (c_segments_nearing_limit / 8)) && \
290 ((c_segment_count * C_SEG_MAX_PAGES) - VM_PAGE_COMPRESSOR_COUNT) > \
291 ((c_segment_count / 8) * C_SEG_MAX_PAGES)) \
292 ? 1 : 0)
293
294 #define COMPRESSOR_FREE_RESERVED_LIMIT 128
295
296 #define COMPRESSOR_SCRATCH_BUF_SIZE WKdm_SCRATCH_BUF_SIZE
297
298
299 #if RECORD_THE_COMPRESSED_DATA
300 extern void c_compressed_record_init(void);
301 extern void c_compressed_record_write(char *, int);
302 #endif
303
304
305 #if __i386__ || __x86_64__
306 extern lck_mtx_t *c_list_lock;
307 #else /* __i386__ || __x86_64__ */
308 extern lck_spin_t *c_list_lock;
309 #endif /* __i386__ || __x86_64__ */