2 * Copyright (c) 2013 Apple Inc. All rights reserved.
4 * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
6 * This file contains Original Code and/or Modifications of Original Code
7 * as defined in and that are subject to the Apple Public Source License
8 * Version 2.0 (the 'License'). You may not use this file except in
9 * compliance with the License. The rights granted to you under the License
10 * may not be used to create, or enable the creation or redistribution of,
11 * unlawful or unlicensed copies of an Apple operating system, or to
12 * circumvent, violate, or enable the circumvention or violation of, any
13 * terms of an Apple operating system software license agreement.
15 * Please obtain a copy of the License at
16 * http://www.opensource.apple.com/apsl/ and read it before using this file.
18 * The Original Code and all software distributed under the License are
19 * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
20 * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
21 * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
22 * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
23 * Please see the License for the specific language governing rights and
24 * limitations under the License.
26 * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
29 #include <kern/locks.h>
30 #include <kern/task.h>
31 #include <kern/thread.h>
32 #include <libkern/OSAtomic.h>
33 #include <vm/vm_pageout.h>
35 #if defined(__x86_64__) && CONFIG_VMX
36 #include <i386/vmx/vmx_cpu.h>
39 #include <kern/hv_support.h>
41 int hv_support_available
= 0;
43 /* callbacks for tasks/threads with associated hv objects */
44 hv_callbacks_t hv_callbacks
= {
45 .dispatch
= NULL
, /* thread is being dispatched for execution */
46 .preempt
= NULL
, /* thread is being preempted */
47 .thread_destroy
= NULL
, /* thread is being destroyed */
48 .task_destroy
= NULL
, /* task is being destroyed */
49 .volatile_state
= NULL
, /* thread state is becoming volatile */
50 .memory_pressure
= NULL
/* memory pressure notification */
53 /* trap tables for hv_*_trap syscalls */
54 static hv_trap_table_t hv_trap_table
[] = {
65 static int hv_callbacks_enabled
= 0;
66 static int hv_mp_notify_enabled
= 0;
67 static int hv_mp_notify_destroy
= 0;
68 static lck_grp_t
*hv_support_lck_grp
= NULL
;
69 static lck_mtx_t
*hv_support_lck_mtx
= NULL
;
70 static thread_t hv_mp_notify_thread
= THREAD_NULL
;
71 static void hv_mp_notify(void);
73 /* hv_support boot initialization */
75 hv_support_init(void) {
76 #if defined(__x86_64__) && CONFIG_VMX
77 hv_support_available
= vmx_hv_support();
80 hv_support_lck_grp
= lck_grp_alloc_init("hv_support", LCK_GRP_ATTR_NULL
);
81 assert(hv_support_lck_grp
);
83 hv_support_lck_mtx
= lck_mtx_alloc_init(hv_support_lck_grp
, LCK_ATTR_NULL
);
84 assert(hv_support_lck_mtx
);
87 /* returns true if hv_support is available on this machine */
89 hv_get_support(void) {
90 return hv_support_available
;
93 /* associate an hv object with the current task */
95 hv_set_task_target(void *target
) {
96 current_task()->hv_task_target
= target
;
99 /* associate an hv object with the current thread */
101 hv_set_thread_target(void *target
) {
102 current_thread()->hv_thread_target
= target
;
105 /* get hv object associated with the current task */
107 hv_get_task_target(void) {
108 return current_task()->hv_task_target
;
111 /* get hv object associated with the current thread */
113 hv_get_thread_target(void) {
114 return current_thread()->hv_thread_target
;
117 /* test if a given thread state may be volatile between dispatch
120 hv_get_volatile_state(hv_volatile_state_t state
) {
123 #if (defined(__x86_64__))
124 if (state
== HV_DEBUG_STATE
) {
125 is_volatile
= (current_thread()->machine
.ids
!= NULL
);
132 /* memory pressure monitor thread */
136 mach_vm_pressure_monitor(TRUE
, 0, NULL
, NULL
);
138 lck_mtx_lock(hv_support_lck_mtx
);
139 if (hv_mp_notify_destroy
== 1) {
140 hv_mp_notify_destroy
= 0;
141 hv_mp_notify_enabled
= 0;
142 lck_mtx_unlock(hv_support_lck_mtx
);
145 hv_callbacks
.memory_pressure(NULL
);
147 lck_mtx_unlock(hv_support_lck_mtx
);
150 thread_deallocate(current_thread());
153 /* subscribe to memory pressure notifications */
155 hv_set_mp_notify(void) {
158 lck_mtx_lock(hv_support_lck_mtx
);
159 if (hv_callbacks_enabled
== 0) {
160 lck_mtx_unlock(hv_support_lck_mtx
);
164 if (hv_mp_notify_enabled
== 1) {
165 hv_mp_notify_destroy
= 0;
166 lck_mtx_unlock(hv_support_lck_mtx
);
170 kr
= kernel_thread_start((thread_continue_t
) &hv_mp_notify
, NULL
,
171 &hv_mp_notify_thread
);
173 if (kr
== KERN_SUCCESS
) {
174 hv_mp_notify_enabled
= 1;
176 lck_mtx_unlock(hv_support_lck_mtx
);
181 /* unsubscribe from memory pressure notifications */
183 hv_release_mp_notify(void) {
184 lck_mtx_lock(hv_support_lck_mtx
);
185 if (hv_mp_notify_enabled
== 1) {
186 hv_mp_notify_destroy
= 1;
188 lck_mtx_unlock(hv_support_lck_mtx
);
191 /* register a list of trap handlers for the hv_*_trap syscalls */
193 hv_set_traps(hv_trap_type_t trap_type
, const hv_trap_t
*traps
,
196 hv_trap_table_t
*trap_table
= &hv_trap_table
[trap_type
];
197 kern_return_t kr
= KERN_FAILURE
;
199 lck_mtx_lock(hv_support_lck_mtx
);
200 if (trap_table
->trap_count
== 0) {
201 trap_table
->traps
= traps
;
203 trap_table
->trap_count
= trap_count
;
206 lck_mtx_unlock(hv_support_lck_mtx
);
211 /* release hv_*_trap traps */
213 hv_release_traps(hv_trap_type_t trap_type
) {
214 hv_trap_table_t
*trap_table
= &hv_trap_table
[trap_type
];
216 lck_mtx_lock(hv_support_lck_mtx
);
217 trap_table
->trap_count
= 0;
219 trap_table
->traps
= NULL
;
220 lck_mtx_unlock(hv_support_lck_mtx
);
223 /* register callbacks for certain task/thread events for tasks/threads with
224 associated hv objects */
226 hv_set_callbacks(hv_callbacks_t callbacks
) {
227 kern_return_t kr
= KERN_FAILURE
;
229 lck_mtx_lock(hv_support_lck_mtx
);
230 if (hv_callbacks_enabled
== 0) {
231 hv_callbacks
= callbacks
;
232 hv_callbacks_enabled
= 1;
235 lck_mtx_unlock(hv_support_lck_mtx
);
240 /* release callbacks for task/thread events */
242 hv_release_callbacks(void) {
243 lck_mtx_lock(hv_support_lck_mtx
);
244 hv_callbacks
= (hv_callbacks_t
) {
247 .thread_destroy
= NULL
,
248 .task_destroy
= NULL
,
249 .volatile_state
= NULL
,
250 .memory_pressure
= NULL
253 hv_callbacks_enabled
= 0;
254 lck_mtx_unlock(hv_support_lck_mtx
);
257 /* dispatch hv_task_trap/hv_thread_trap syscalls to trap handlers,
258 fail for invalid index or absence of trap handlers, trap handler is
259 responsible for validating targets */
260 #define HV_TRAP_DISPATCH(type, index, target, argument)\
261 ((__probable(index < hv_trap_table[type].trap_count)) ? \
262 hv_trap_table[type].traps[index](target, argument) \
263 : KERN_INVALID_ARGUMENT)
265 kern_return_t
hv_task_trap(uint64_t index
, uint64_t arg
) {
266 return HV_TRAP_DISPATCH(HV_TASK_TRAP
, index
, hv_get_task_target(), arg
);
269 kern_return_t
hv_thread_trap(uint64_t index
, uint64_t arg
) {
270 return HV_TRAP_DISPATCH(HV_THREAD_TRAP
, index
, hv_get_thread_target(), arg
);