]>
Commit | Line | Data |
---|---|---|
1c79356b | 1 | /* |
b0d623f7 | 2 | * Copyright (c) 2000-2009 Apple Inc. All rights reserved. |
1c79356b | 3 | * |
2d21ac55 | 4 | * @APPLE_OSREFERENCE_LICENSE_HEADER_START@ |
1c79356b | 5 | * |
2d21ac55 A |
6 | * This file contains Original Code and/or Modifications of Original Code |
7 | * as defined in and that are subject to the Apple Public Source License | |
8 | * Version 2.0 (the 'License'). You may not use this file except in | |
9 | * compliance with the License. The rights granted to you under the License | |
10 | * may not be used to create, or enable the creation or redistribution of, | |
11 | * unlawful or unlicensed copies of an Apple operating system, or to | |
12 | * circumvent, violate, or enable the circumvention or violation of, any | |
13 | * terms of an Apple operating system software license agreement. | |
8f6c56a5 | 14 | * |
2d21ac55 A |
15 | * Please obtain a copy of the License at |
16 | * http://www.opensource.apple.com/apsl/ and read it before using this file. | |
17 | * | |
18 | * The Original Code and all software distributed under the License are | |
19 | * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER | |
8f6c56a5 A |
20 | * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES, |
21 | * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY, | |
2d21ac55 A |
22 | * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT. |
23 | * Please see the License for the specific language governing rights and | |
24 | * limitations under the License. | |
8f6c56a5 | 25 | * |
2d21ac55 | 26 | * @APPLE_OSREFERENCE_LICENSE_HEADER_END@ |
1c79356b A |
27 | */ |
28 | /* | |
29 | * @OSF_FREE_COPYRIGHT@ | |
30 | */ | |
31 | /* | |
32 | * Mach Operating System | |
33 | * Copyright (c) 1991,1990,1989,1988 Carnegie Mellon University | |
34 | * All Rights Reserved. | |
35 | * | |
36 | * Permission to use, copy, modify and distribute this software and its | |
37 | * documentation is hereby granted, provided that both the copyright | |
38 | * notice and this permission notice appear in all copies of the | |
39 | * software, derivative works or modified versions, and any portions | |
40 | * thereof, and that both notices appear in supporting documentation. | |
41 | * | |
42 | * CARNEGIE MELLON ALLOWS FREE USE OF THIS SOFTWARE IN ITS "AS IS" | |
43 | * CONDITION. CARNEGIE MELLON DISCLAIMS ANY LIABILITY OF ANY KIND FOR | |
44 | * ANY DAMAGES WHATSOEVER RESULTING FROM THE USE OF THIS SOFTWARE. | |
45 | * | |
46 | * Carnegie Mellon requests users of this software to return to | |
47 | * | |
48 | * Software Distribution Coordinator or Software.Distribution@CS.CMU.EDU | |
49 | * School of Computer Science | |
50 | * Carnegie Mellon University | |
51 | * Pittsburgh PA 15213-3890 | |
52 | * | |
53 | * any improvements or extensions that they make and grant Carnegie Mellon | |
54 | * the rights to redistribute these changes. | |
55 | */ | |
56 | /* | |
57 | * File: kern/task.c | |
58 | * Author: Avadis Tevanian, Jr., Michael Wayne Young, David Golub, | |
59 | * David Black | |
60 | * | |
61 | * Task management primitives implementation. | |
62 | */ | |
63 | /* | |
64 | * Copyright (c) 1993 The University of Utah and | |
65 | * the Computer Systems Laboratory (CSL). All rights reserved. | |
66 | * | |
67 | * Permission to use, copy, modify and distribute this software and its | |
68 | * documentation is hereby granted, provided that both the copyright | |
69 | * notice and this permission notice appear in all copies of the | |
70 | * software, derivative works or modified versions, and any portions | |
71 | * thereof, and that both notices appear in supporting documentation. | |
72 | * | |
73 | * THE UNIVERSITY OF UTAH AND CSL ALLOW FREE USE OF THIS SOFTWARE IN ITS "AS | |
74 | * IS" CONDITION. THE UNIVERSITY OF UTAH AND CSL DISCLAIM ANY LIABILITY OF | |
75 | * ANY KIND FOR ANY DAMAGES WHATSOEVER RESULTING FROM THE USE OF THIS SOFTWARE. | |
76 | * | |
77 | * CSL requests users of this software to return to csl-dist@cs.utah.edu any | |
78 | * improvements that they make and grant CSL redistribution rights. | |
79 | * | |
80 | */ | |
2d21ac55 A |
81 | /* |
82 | * NOTICE: This file was modified by McAfee Research in 2004 to introduce | |
83 | * support for mandatory and extensible security protections. This notice | |
84 | * is included in support of clause 2.2 (b) of the Apple Public License, | |
85 | * Version 2.0. | |
86 | * Copyright (c) 2005 SPARTA, Inc. | |
87 | */ | |
1c79356b A |
88 | |
89 | #include <mach_kdb.h> | |
1c79356b | 90 | #include <fast_tas.h> |
1c79356b A |
91 | #include <platforms.h> |
92 | ||
91447636 | 93 | #include <mach/mach_types.h> |
1c79356b | 94 | #include <mach/boolean.h> |
91447636 | 95 | #include <mach/host_priv.h> |
1c79356b A |
96 | #include <mach/machine/vm_types.h> |
97 | #include <mach/vm_param.h> | |
98 | #include <mach/semaphore.h> | |
99 | #include <mach/task_info.h> | |
100 | #include <mach/task_special_ports.h> | |
91447636 A |
101 | |
102 | #include <ipc/ipc_types.h> | |
1c79356b A |
103 | #include <ipc/ipc_space.h> |
104 | #include <ipc/ipc_entry.h> | |
91447636 A |
105 | |
106 | #include <kern/kern_types.h> | |
1c79356b A |
107 | #include <kern/mach_param.h> |
108 | #include <kern/misc_protos.h> | |
109 | #include <kern/task.h> | |
110 | #include <kern/thread.h> | |
111 | #include <kern/zalloc.h> | |
112 | #include <kern/kalloc.h> | |
113 | #include <kern/processor.h> | |
114 | #include <kern/sched_prim.h> /* for thread_wakeup */ | |
1c79356b A |
115 | #include <kern/ipc_tt.h> |
116 | #include <kern/ledger.h> | |
117 | #include <kern/host.h> | |
91447636 A |
118 | #include <kern/clock.h> |
119 | #include <kern/timer.h> | |
1c79356b A |
120 | #include <kern/assert.h> |
121 | #include <kern/sync_lock.h> | |
2d21ac55 | 122 | #include <kern/affinity.h> |
91447636 A |
123 | |
124 | #include <vm/pmap.h> | |
125 | #include <vm/vm_map.h> | |
126 | #include <vm/vm_kern.h> /* for kernel_map, ipc_kernel_map */ | |
127 | #include <vm/vm_pageout.h> | |
2d21ac55 | 128 | #include <vm/vm_protos.h> |
91447636 | 129 | |
1c79356b A |
130 | #if MACH_KDB |
131 | #include <ddb/db_sym.h> | |
132 | #endif /* MACH_KDB */ | |
133 | ||
1c79356b A |
134 | /* |
135 | * Exported interfaces | |
136 | */ | |
137 | ||
138 | #include <mach/task_server.h> | |
139 | #include <mach/mach_host_server.h> | |
140 | #include <mach/host_security_server.h> | |
91447636 | 141 | #include <mach/mach_port_server.h> |
2d21ac55 | 142 | #include <mach/security_server.h> |
91447636 | 143 | |
2d21ac55 A |
144 | #include <vm/vm_shared_region.h> |
145 | ||
146 | #if CONFIG_MACF_MACH | |
147 | #include <security/mac_mach_internal.h> | |
148 | #endif | |
1c79356b | 149 | |
b0d623f7 A |
150 | #if CONFIG_COUNTERS |
151 | #include <pmc/pmc.h> | |
152 | #endif /* CONFIG_COUNTERS */ | |
153 | ||
154 | task_t kernel_task; | |
155 | zone_t task_zone; | |
156 | lck_attr_t task_lck_attr; | |
157 | lck_grp_t task_lck_grp; | |
158 | lck_grp_attr_t task_lck_grp_attr; | |
159 | ||
6d2010ae A |
160 | zinfo_usage_store_t tasks_tkm_private; |
161 | zinfo_usage_store_t tasks_tkm_shared; | |
162 | ||
b0d623f7 | 163 | int task_max = CONFIG_TASK_MAX; /* Max number of tasks */ |
1c79356b | 164 | |
6d2010ae A |
165 | /* externs for BSD kernel */ |
166 | extern void proc_getexecutableuuid(void *, unsigned char *, unsigned long); | |
167 | ||
1c79356b A |
168 | /* Forwards */ |
169 | ||
170 | void task_hold_locked( | |
171 | task_t task); | |
172 | void task_wait_locked( | |
173 | task_t task); | |
174 | void task_release_locked( | |
175 | task_t task); | |
1c79356b A |
176 | void task_free( |
177 | task_t task ); | |
178 | void task_synchronizer_destroy_all( | |
179 | task_t task); | |
1c79356b A |
180 | |
181 | kern_return_t task_set_ledger( | |
182 | task_t task, | |
183 | ledger_t wired, | |
184 | ledger_t paged); | |
185 | ||
b0d623f7 A |
186 | int check_for_tasksuspend( |
187 | task_t task); | |
188 | ||
55e303ae A |
189 | void |
190 | task_backing_store_privileged( | |
191 | task_t task) | |
192 | { | |
193 | task_lock(task); | |
194 | task->priv_flags |= VM_BACKING_STORE_PRIV; | |
195 | task_unlock(task); | |
196 | return; | |
197 | } | |
198 | ||
91447636 A |
199 | |
200 | void | |
201 | task_set_64bit( | |
202 | task_t task, | |
203 | boolean_t is64bit) | |
204 | { | |
b0d623f7 | 205 | #if defined(__i386__) || defined(__x86_64__) |
2d21ac55 A |
206 | thread_t thread; |
207 | #endif /* __i386__ */ | |
208 | int vm_flags = 0; | |
0c530ab8 A |
209 | |
210 | if (is64bit) { | |
2d21ac55 A |
211 | if (task_has_64BitAddr(task)) |
212 | return; | |
0c530ab8 | 213 | |
91447636 | 214 | task_set_64BitAddr(task); |
91447636 | 215 | } else { |
2d21ac55 A |
216 | if ( !task_has_64BitAddr(task)) |
217 | return; | |
0c530ab8 | 218 | |
91447636 A |
219 | /* |
220 | * Deallocate all memory previously allocated | |
221 | * above the 32-bit address space, since it won't | |
222 | * be accessible anymore. | |
223 | */ | |
2d21ac55 A |
224 | /* remove regular VM map entries & pmap mappings */ |
225 | (void) vm_map_remove(task->map, | |
226 | (vm_map_offset_t) VM_MAX_ADDRESS, | |
227 | MACH_VM_MAX_ADDRESS, | |
228 | 0); | |
2d21ac55 | 229 | /* remove the higher VM mappings */ |
91447636 | 230 | (void) vm_map_remove(task->map, |
91447636 | 231 | MACH_VM_MAX_ADDRESS, |
2d21ac55 A |
232 | 0xFFFFFFFFFFFFF000ULL, |
233 | vm_flags); | |
91447636 | 234 | task_clear_64BitAddr(task); |
91447636 | 235 | } |
0c530ab8 A |
236 | /* FIXME: On x86, the thread save state flavor can diverge from the |
237 | * task's 64-bit feature flag due to the 32-bit/64-bit register save | |
238 | * state dichotomy. Since we can be pre-empted in this interval, | |
239 | * certain routines may observe the thread as being in an inconsistent | |
240 | * state with respect to its task's 64-bitness. | |
241 | */ | |
b0d623f7 A |
242 | #if defined(__i386__) || defined(__x86_64__) |
243 | task_lock(task); | |
0c530ab8 | 244 | queue_iterate(&task->threads, thread, thread_t, task_threads) { |
b0d623f7 | 245 | thread_mtx_lock(thread); |
2d21ac55 | 246 | machine_thread_switch_addrmode(thread); |
b0d623f7 | 247 | thread_mtx_unlock(thread); |
0c530ab8 | 248 | } |
b0d623f7 | 249 | task_unlock(task); |
2d21ac55 | 250 | #endif /* __i386__ */ |
91447636 A |
251 | } |
252 | ||
b0d623f7 A |
253 | |
254 | void | |
255 | task_set_dyld_info(task_t task, mach_vm_address_t addr, mach_vm_size_t size) | |
256 | { | |
257 | task_lock(task); | |
258 | task->all_image_info_addr = addr; | |
259 | task->all_image_info_size = size; | |
260 | task_unlock(task); | |
261 | } | |
262 | ||
1c79356b A |
263 | void |
264 | task_init(void) | |
265 | { | |
b0d623f7 A |
266 | |
267 | lck_grp_attr_setdefault(&task_lck_grp_attr); | |
268 | lck_grp_init(&task_lck_grp, "task", &task_lck_grp_attr); | |
269 | lck_attr_setdefault(&task_lck_attr); | |
270 | lck_mtx_init(&tasks_threads_lock, &task_lck_grp, &task_lck_attr); | |
271 | ||
1c79356b A |
272 | task_zone = zinit( |
273 | sizeof(struct task), | |
b0d623f7 | 274 | task_max * sizeof(struct task), |
1c79356b A |
275 | TASK_CHUNK * sizeof(struct task), |
276 | "tasks"); | |
6d2010ae | 277 | |
0b4c1975 | 278 | zone_change(task_zone, Z_NOENCRYPT, TRUE); |
1c79356b | 279 | |
1c79356b A |
280 | /* |
281 | * Create the kernel task as the first task. | |
1c79356b | 282 | */ |
b0d623f7 A |
283 | #ifdef __LP64__ |
284 | if (task_create_internal(TASK_NULL, FALSE, TRUE, &kernel_task) != KERN_SUCCESS) | |
285 | #else | |
0c530ab8 | 286 | if (task_create_internal(TASK_NULL, FALSE, FALSE, &kernel_task) != KERN_SUCCESS) |
b0d623f7 | 287 | #endif |
1c79356b | 288 | panic("task_init\n"); |
55e303ae | 289 | |
1c79356b A |
290 | vm_map_deallocate(kernel_task->map); |
291 | kernel_task->map = kernel_map; | |
1c79356b A |
292 | } |
293 | ||
1c79356b A |
294 | /* |
295 | * Create a task running in the kernel address space. It may | |
296 | * have its own map of size mem_size and may have ipc privileges. | |
297 | */ | |
298 | kern_return_t | |
299 | kernel_task_create( | |
91447636 A |
300 | __unused task_t parent_task, |
301 | __unused vm_offset_t map_base, | |
302 | __unused vm_size_t map_size, | |
303 | __unused task_t *child_task) | |
1c79356b | 304 | { |
55e303ae | 305 | return (KERN_INVALID_ARGUMENT); |
1c79356b A |
306 | } |
307 | ||
308 | kern_return_t | |
309 | task_create( | |
2d21ac55 | 310 | task_t parent_task, |
91447636 | 311 | __unused ledger_port_array_t ledger_ports, |
2d21ac55 A |
312 | __unused mach_msg_type_number_t num_ledger_ports, |
313 | __unused boolean_t inherit_memory, | |
314 | __unused task_t *child_task) /* OUT */ | |
1c79356b A |
315 | { |
316 | if (parent_task == TASK_NULL) | |
317 | return(KERN_INVALID_ARGUMENT); | |
318 | ||
2d21ac55 A |
319 | /* |
320 | * No longer supported: too many calls assume that a task has a valid | |
321 | * process attached. | |
322 | */ | |
323 | return(KERN_FAILURE); | |
1c79356b A |
324 | } |
325 | ||
326 | kern_return_t | |
327 | host_security_create_task_token( | |
91447636 | 328 | host_security_t host_security, |
2d21ac55 A |
329 | task_t parent_task, |
330 | __unused security_token_t sec_token, | |
331 | __unused audit_token_t audit_token, | |
332 | __unused host_priv_t host_priv, | |
91447636 A |
333 | __unused ledger_port_array_t ledger_ports, |
334 | __unused mach_msg_type_number_t num_ledger_ports, | |
2d21ac55 A |
335 | __unused boolean_t inherit_memory, |
336 | __unused task_t *child_task) /* OUT */ | |
1c79356b | 337 | { |
1c79356b A |
338 | if (parent_task == TASK_NULL) |
339 | return(KERN_INVALID_ARGUMENT); | |
340 | ||
341 | if (host_security == HOST_NULL) | |
342 | return(KERN_INVALID_SECURITY); | |
343 | ||
2d21ac55 A |
344 | /* |
345 | * No longer supported. | |
346 | */ | |
347 | return(KERN_FAILURE); | |
1c79356b A |
348 | } |
349 | ||
350 | kern_return_t | |
55e303ae | 351 | task_create_internal( |
1c79356b A |
352 | task_t parent_task, |
353 | boolean_t inherit_memory, | |
0c530ab8 | 354 | boolean_t is_64bit, |
1c79356b A |
355 | task_t *child_task) /* OUT */ |
356 | { | |
2d21ac55 A |
357 | task_t new_task; |
358 | vm_shared_region_t shared_region; | |
1c79356b A |
359 | |
360 | new_task = (task_t) zalloc(task_zone); | |
361 | ||
362 | if (new_task == TASK_NULL) | |
363 | return(KERN_RESOURCE_SHORTAGE); | |
364 | ||
365 | /* one ref for just being alive; one for our caller */ | |
366 | new_task->ref_count = 2; | |
367 | ||
b0d623f7 | 368 | /* if inherit_memory is true, parent_task MUST not be NULL */ |
1c79356b A |
369 | if (inherit_memory) |
370 | new_task->map = vm_map_fork(parent_task->map); | |
371 | else | |
0c530ab8 | 372 | new_task->map = vm_map_create(pmap_create(0, is_64bit), |
91447636 A |
373 | (vm_map_offset_t)(VM_MIN_ADDRESS), |
374 | (vm_map_offset_t)(VM_MAX_ADDRESS), TRUE); | |
1c79356b | 375 | |
2d21ac55 A |
376 | /* Inherit memlock limit from parent */ |
377 | if (parent_task) | |
b0d623f7 | 378 | vm_map_set_user_wire_limit(new_task->map, (vm_size_t)parent_task->map->user_wire_limit); |
2d21ac55 | 379 | |
b0d623f7 | 380 | lck_mtx_init(&new_task->lock, &task_lck_grp, &task_lck_attr); |
55e303ae | 381 | queue_init(&new_task->threads); |
1c79356b | 382 | new_task->suspend_count = 0; |
55e303ae | 383 | new_task->thread_count = 0; |
55e303ae | 384 | new_task->active_thread_count = 0; |
1c79356b | 385 | new_task->user_stop_count = 0; |
0b4e3aa0 | 386 | new_task->role = TASK_UNSPECIFIED; |
1c79356b | 387 | new_task->active = TRUE; |
b0d623f7 | 388 | new_task->halting = FALSE; |
2d21ac55 | 389 | new_task->user_data = NULL; |
1c79356b A |
390 | new_task->faults = 0; |
391 | new_task->cow_faults = 0; | |
392 | new_task->pageins = 0; | |
393 | new_task->messages_sent = 0; | |
394 | new_task->messages_received = 0; | |
395 | new_task->syscalls_mach = 0; | |
55e303ae | 396 | new_task->priv_flags = 0; |
1c79356b | 397 | new_task->syscalls_unix=0; |
2d21ac55 | 398 | new_task->c_switch = new_task->p_switch = new_task->ps_switch = 0; |
55e303ae A |
399 | new_task->taskFeatures[0] = 0; /* Init task features */ |
400 | new_task->taskFeatures[1] = 0; /* Init task features */ | |
1c79356b | 401 | |
6d2010ae A |
402 | new_task->tkm_private.alloc = 0; |
403 | new_task->tkm_private.free = 0; | |
404 | new_task->tkm_shared.alloc = 0; | |
405 | new_task->tkm_shared.free = 0; | |
406 | ||
407 | zinfo_task_init(new_task); | |
408 | ||
1c79356b | 409 | #ifdef MACH_BSD |
2d21ac55 | 410 | new_task->bsd_info = NULL; |
1c79356b A |
411 | #endif /* MACH_BSD */ |
412 | ||
b0d623f7 | 413 | #if defined(__i386__) || defined(__x86_64__) |
0c530ab8 | 414 | new_task->i386_ldt = 0; |
b0d623f7 | 415 | new_task->task_debug = NULL; |
0c530ab8 A |
416 | #endif |
417 | ||
55e303ae | 418 | |
1c79356b A |
419 | queue_init(&new_task->semaphore_list); |
420 | queue_init(&new_task->lock_set_list); | |
421 | new_task->semaphores_owned = 0; | |
422 | new_task->lock_sets_owned = 0; | |
423 | ||
2d21ac55 | 424 | #if CONFIG_MACF_MACH |
2d21ac55 A |
425 | new_task->label = labelh_new(1); |
426 | mac_task_label_init (&new_task->maclabel); | |
427 | #endif | |
1c79356b A |
428 | |
429 | ipc_task_init(new_task, parent_task); | |
430 | ||
91447636 A |
431 | new_task->total_user_time = 0; |
432 | new_task->total_system_time = 0; | |
1c79356b | 433 | |
2d21ac55 | 434 | new_task->vtimers = 0; |
1c79356b | 435 | |
2d21ac55 A |
436 | new_task->shared_region = NULL; |
437 | ||
438 | new_task->affinity_space = NULL; | |
1c79356b | 439 | |
b0d623f7 A |
440 | #if CONFIG_COUNTERS |
441 | new_task->t_chud = 0U; | |
442 | #endif | |
443 | ||
2d21ac55 | 444 | if (parent_task != TASK_NULL) { |
1c79356b | 445 | new_task->sec_token = parent_task->sec_token; |
55e303ae | 446 | new_task->audit_token = parent_task->audit_token; |
1c79356b | 447 | |
2d21ac55 A |
448 | /* inherit the parent's shared region */ |
449 | shared_region = vm_shared_region_get(parent_task); | |
450 | vm_shared_region_set(new_task, shared_region); | |
1c79356b A |
451 | |
452 | new_task->wired_ledger_port = ledger_copy( | |
453 | convert_port_to_ledger(parent_task->wired_ledger_port)); | |
454 | new_task->paged_ledger_port = ledger_copy( | |
455 | convert_port_to_ledger(parent_task->paged_ledger_port)); | |
91447636 A |
456 | if(task_has_64BitAddr(parent_task)) |
457 | task_set_64BitAddr(new_task); | |
b0d623f7 A |
458 | new_task->all_image_info_addr = parent_task->all_image_info_addr; |
459 | new_task->all_image_info_size = parent_task->all_image_info_size; | |
0c530ab8 | 460 | |
b0d623f7 | 461 | #if defined(__i386__) || defined(__x86_64__) |
0c530ab8 A |
462 | if (inherit_memory && parent_task->i386_ldt) |
463 | new_task->i386_ldt = user_ldt_copy(parent_task->i386_ldt); | |
464 | #endif | |
2d21ac55 A |
465 | if (inherit_memory && parent_task->affinity_space) |
466 | task_affinity_create(parent_task, new_task); | |
b0d623f7 A |
467 | |
468 | new_task->pset_hint = parent_task->pset_hint = task_choose_pset(parent_task); | |
6d2010ae A |
469 | new_task->policystate = parent_task->policystate; |
470 | /* inherit the self action state */ | |
471 | new_task->actionstate = parent_task->actionstate; | |
472 | new_task->ext_policystate = parent_task->ext_policystate; | |
473 | #if NOTYET | |
474 | /* till the child lifecycle is cleared do not inherit external action */ | |
475 | new_task->ext_actionstate = parent_task->ext_actionstate; | |
476 | #else | |
477 | new_task->ext_actionstate = default_task_null_policy; | |
478 | #endif | |
1c79356b A |
479 | } |
480 | else { | |
1c79356b | 481 | new_task->sec_token = KERNEL_SECURITY_TOKEN; |
55e303ae | 482 | new_task->audit_token = KERNEL_AUDIT_TOKEN; |
1c79356b A |
483 | new_task->wired_ledger_port = ledger_copy(root_wired_ledger); |
484 | new_task->paged_ledger_port = ledger_copy(root_paged_ledger); | |
b0d623f7 A |
485 | #ifdef __LP64__ |
486 | if(is_64bit) | |
487 | task_set_64BitAddr(new_task); | |
488 | #endif | |
6d2010ae A |
489 | new_task->all_image_info_addr = (mach_vm_address_t)0; |
490 | new_task->all_image_info_size = (mach_vm_size_t)0; | |
b0d623f7 A |
491 | |
492 | new_task->pset_hint = PROCESSOR_SET_NULL; | |
6d2010ae A |
493 | new_task->policystate = default_task_proc_policy; |
494 | new_task->ext_policystate = default_task_proc_policy; | |
495 | new_task->actionstate = default_task_null_policy; | |
496 | new_task->ext_actionstate = default_task_null_policy; | |
1c79356b A |
497 | } |
498 | ||
0b4e3aa0 | 499 | if (kernel_task == TASK_NULL) { |
55e303ae | 500 | new_task->priority = BASEPRI_KERNEL; |
0b4e3aa0 A |
501 | new_task->max_priority = MAXPRI_KERNEL; |
502 | } | |
503 | else { | |
504 | new_task->priority = BASEPRI_DEFAULT; | |
505 | new_task->max_priority = MAXPRI_USER; | |
506 | } | |
6d2010ae A |
507 | |
508 | bzero(&new_task->extmod_statistics, sizeof(new_task->extmod_statistics)); | |
2d21ac55 | 509 | |
b0d623f7 | 510 | lck_mtx_lock(&tasks_threads_lock); |
2d21ac55 A |
511 | queue_enter(&tasks, new_task, task_t, tasks); |
512 | tasks_count++; | |
b0d623f7 | 513 | lck_mtx_unlock(&tasks_threads_lock); |
1c79356b | 514 | |
55e303ae A |
515 | if (vm_backing_store_low && parent_task != NULL) |
516 | new_task->priv_flags |= (parent_task->priv_flags&VM_BACKING_STORE_PRIV); | |
1c79356b A |
517 | |
518 | ipc_task_enable(new_task); | |
519 | ||
1c79356b A |
520 | *child_task = new_task; |
521 | return(KERN_SUCCESS); | |
522 | } | |
523 | ||
524 | /* | |
91447636 | 525 | * task_deallocate: |
1c79356b | 526 | * |
91447636 | 527 | * Drop a reference on a task. |
1c79356b A |
528 | */ |
529 | void | |
9bccf70c | 530 | task_deallocate( |
1c79356b A |
531 | task_t task) |
532 | { | |
9bccf70c A |
533 | if (task == TASK_NULL) |
534 | return; | |
535 | ||
91447636 | 536 | if (task_deallocate_internal(task) > 0) |
9bccf70c | 537 | return; |
1c79356b | 538 | |
6d2010ae A |
539 | lck_mtx_lock(&tasks_threads_lock); |
540 | queue_remove(&terminated_tasks, task, task_t, tasks); | |
541 | lck_mtx_unlock(&tasks_threads_lock); | |
542 | ||
9bccf70c A |
543 | ipc_task_terminate(task); |
544 | ||
2d21ac55 A |
545 | if (task->affinity_space) |
546 | task_affinity_deallocate(task); | |
547 | ||
1c79356b A |
548 | vm_map_deallocate(task->map); |
549 | is_release(task->itk_space); | |
1c79356b | 550 | |
b0d623f7 A |
551 | lck_mtx_destroy(&task->lock, &task_lck_grp); |
552 | ||
2d21ac55 A |
553 | #if CONFIG_MACF_MACH |
554 | labelh_release(task->label); | |
555 | #endif | |
6d2010ae A |
556 | OSAddAtomic64(task->tkm_private.alloc, (int64_t *)&tasks_tkm_private.alloc); |
557 | OSAddAtomic64(task->tkm_private.free, (int64_t *)&tasks_tkm_private.free); | |
558 | OSAddAtomic64(task->tkm_shared.alloc, (int64_t *)&tasks_tkm_shared.alloc); | |
559 | OSAddAtomic64(task->tkm_shared.free, (int64_t *)&tasks_tkm_shared.free); | |
560 | zinfo_task_free(task); | |
91447636 | 561 | zfree(task_zone, task); |
1c79356b A |
562 | } |
563 | ||
0c530ab8 A |
564 | /* |
565 | * task_name_deallocate: | |
566 | * | |
567 | * Drop a reference on a task name. | |
568 | */ | |
569 | void | |
570 | task_name_deallocate( | |
571 | task_name_t task_name) | |
572 | { | |
573 | return(task_deallocate((task_t)task_name)); | |
574 | } | |
575 | ||
576 | ||
1c79356b A |
577 | /* |
578 | * task_terminate: | |
579 | * | |
580 | * Terminate the specified task. See comments on thread_terminate | |
581 | * (kern/thread.c) about problems with terminating the "current task." | |
582 | */ | |
583 | ||
584 | kern_return_t | |
585 | task_terminate( | |
586 | task_t task) | |
587 | { | |
588 | if (task == TASK_NULL) | |
91447636 A |
589 | return (KERN_INVALID_ARGUMENT); |
590 | ||
1c79356b | 591 | if (task->bsd_info) |
91447636 A |
592 | return (KERN_FAILURE); |
593 | ||
1c79356b A |
594 | return (task_terminate_internal(task)); |
595 | } | |
596 | ||
597 | kern_return_t | |
598 | task_terminate_internal( | |
91447636 | 599 | task_t task) |
1c79356b | 600 | { |
91447636 A |
601 | thread_t thread, self; |
602 | task_t self_task; | |
603 | boolean_t interrupt_save; | |
1c79356b A |
604 | |
605 | assert(task != kernel_task); | |
606 | ||
91447636 A |
607 | self = current_thread(); |
608 | self_task = self->task; | |
1c79356b A |
609 | |
610 | /* | |
611 | * Get the task locked and make sure that we are not racing | |
612 | * with someone else trying to terminate us. | |
613 | */ | |
91447636 | 614 | if (task == self_task) |
1c79356b | 615 | task_lock(task); |
91447636 A |
616 | else |
617 | if (task < self_task) { | |
1c79356b | 618 | task_lock(task); |
91447636 A |
619 | task_lock(self_task); |
620 | } | |
621 | else { | |
622 | task_lock(self_task); | |
1c79356b A |
623 | task_lock(task); |
624 | } | |
625 | ||
6d2010ae | 626 | if (!task->active) { |
1c79356b | 627 | /* |
6d2010ae | 628 | * Task is already being terminated. |
1c79356b A |
629 | * Just return an error. If we are dying, this will |
630 | * just get us to our AST special handler and that | |
631 | * will get us to finalize the termination of ourselves. | |
632 | */ | |
633 | task_unlock(task); | |
91447636 A |
634 | if (self_task != task) |
635 | task_unlock(self_task); | |
636 | ||
637 | return (KERN_FAILURE); | |
1c79356b | 638 | } |
91447636 A |
639 | |
640 | if (self_task != task) | |
641 | task_unlock(self_task); | |
1c79356b | 642 | |
e7c99d92 A |
643 | /* |
644 | * Make sure the current thread does not get aborted out of | |
645 | * the waits inside these operations. | |
646 | */ | |
9bccf70c | 647 | interrupt_save = thread_interrupt_level(THREAD_UNINT); |
e7c99d92 | 648 | |
1c79356b A |
649 | /* |
650 | * Indicate that we want all the threads to stop executing | |
651 | * at user space by holding the task (we would have held | |
652 | * each thread independently in thread_terminate_internal - | |
653 | * but this way we may be more likely to already find it | |
654 | * held there). Mark the task inactive, and prevent | |
655 | * further task operations via the task port. | |
656 | */ | |
657 | task_hold_locked(task); | |
658 | task->active = FALSE; | |
659 | ipc_task_disable(task); | |
660 | ||
661 | /* | |
91447636 A |
662 | * Terminate each thread in the task. |
663 | */ | |
664 | queue_iterate(&task->threads, thread, thread_t, task_threads) { | |
665 | thread_terminate_internal(thread); | |
1c79356b | 666 | } |
e7c99d92 A |
667 | |
668 | /* | |
55e303ae A |
669 | * Give the machine dependent code a chance |
670 | * to perform cleanup before ripping apart | |
671 | * the task. | |
e7c99d92 | 672 | */ |
91447636 | 673 | if (self_task == task) |
55e303ae | 674 | machine_thread_terminate_self(); |
e7c99d92 | 675 | |
1c79356b A |
676 | task_unlock(task); |
677 | ||
678 | /* | |
679 | * Destroy all synchronizers owned by the task. | |
680 | */ | |
681 | task_synchronizer_destroy_all(task); | |
682 | ||
1c79356b A |
683 | /* |
684 | * Destroy the IPC space, leaving just a reference for it. | |
685 | */ | |
55e303ae | 686 | ipc_space_destroy(task->itk_space); |
1c79356b | 687 | |
0c530ab8 A |
688 | if (vm_map_has_4GB_pagezero(task->map)) |
689 | vm_map_clear_4GB_pagezero(task->map); | |
91447636 | 690 | |
1c79356b A |
691 | /* |
692 | * If the current thread is a member of the task | |
693 | * being terminated, then the last reference to | |
694 | * the task will not be dropped until the thread | |
695 | * is finally reaped. To avoid incurring the | |
696 | * expense of removing the address space regions | |
697 | * at reap time, we do it explictly here. | |
698 | */ | |
2d21ac55 A |
699 | vm_map_remove(task->map, |
700 | task->map->min_offset, | |
701 | task->map->max_offset, | |
702 | VM_MAP_NO_FLAGS); | |
1c79356b | 703 | |
2d21ac55 A |
704 | /* release our shared region */ |
705 | vm_shared_region_set(task, NULL); | |
9bccf70c | 706 | |
b0d623f7 | 707 | lck_mtx_lock(&tasks_threads_lock); |
2d21ac55 | 708 | queue_remove(&tasks, task, task_t, tasks); |
6d2010ae | 709 | queue_enter(&terminated_tasks, task, task_t, tasks); |
2d21ac55 | 710 | tasks_count--; |
b0d623f7 | 711 | lck_mtx_unlock(&tasks_threads_lock); |
9bccf70c | 712 | |
1c79356b | 713 | /* |
e7c99d92 A |
714 | * We no longer need to guard against being aborted, so restore |
715 | * the previous interruptible state. | |
716 | */ | |
9bccf70c | 717 | thread_interrupt_level(interrupt_save); |
e7c99d92 A |
718 | |
719 | /* | |
720 | * Get rid of the task active reference on itself. | |
1c79356b | 721 | */ |
1c79356b A |
722 | task_deallocate(task); |
723 | ||
91447636 | 724 | return (KERN_SUCCESS); |
1c79356b A |
725 | } |
726 | ||
727 | /* | |
b0d623f7 | 728 | * task_start_halt: |
91447636 A |
729 | * |
730 | * Shut the current task down (except for the current thread) in | |
731 | * preparation for dramatic changes to the task (probably exec). | |
b0d623f7 A |
732 | * We hold the task and mark all other threads in the task for |
733 | * termination. | |
1c79356b A |
734 | */ |
735 | kern_return_t | |
b0d623f7 | 736 | task_start_halt( |
1c79356b A |
737 | task_t task) |
738 | { | |
91447636 | 739 | thread_t thread, self; |
1c79356b A |
740 | |
741 | assert(task != kernel_task); | |
742 | ||
91447636 | 743 | self = current_thread(); |
1c79356b | 744 | |
91447636 A |
745 | if (task != self->task) |
746 | return (KERN_INVALID_ARGUMENT); | |
1c79356b A |
747 | |
748 | task_lock(task); | |
749 | ||
b0d623f7 | 750 | if (task->halting || !task->active || !self->active) { |
1c79356b A |
751 | /* |
752 | * Task or current thread is already being terminated. | |
753 | * Hurry up and return out of the current kernel context | |
754 | * so that we run our AST special handler to terminate | |
755 | * ourselves. | |
756 | */ | |
757 | task_unlock(task); | |
91447636 A |
758 | |
759 | return (KERN_FAILURE); | |
1c79356b A |
760 | } |
761 | ||
b0d623f7 A |
762 | task->halting = TRUE; |
763 | ||
55e303ae | 764 | if (task->thread_count > 1) { |
b0d623f7 | 765 | |
1c79356b A |
766 | /* |
767 | * Mark all the threads to keep them from starting any more | |
768 | * user-level execution. The thread_terminate_internal code | |
769 | * would do this on a thread by thread basis anyway, but this | |
770 | * gives us a better chance of not having to wait there. | |
771 | */ | |
772 | task_hold_locked(task); | |
773 | ||
774 | /* | |
91447636 | 775 | * Terminate all the other threads in the task. |
1c79356b | 776 | */ |
91447636 A |
777 | queue_iterate(&task->threads, thread, thread_t, task_threads) { |
778 | if (thread != self) | |
779 | thread_terminate_internal(thread); | |
1c79356b | 780 | } |
91447636 | 781 | |
1c79356b A |
782 | task_release_locked(task); |
783 | } | |
b0d623f7 A |
784 | task_unlock(task); |
785 | return KERN_SUCCESS; | |
786 | } | |
787 | ||
788 | ||
789 | /* | |
790 | * task_complete_halt: | |
791 | * | |
792 | * Complete task halt by waiting for threads to terminate, then clean | |
793 | * up task resources (VM, port namespace, etc...) and then let the | |
794 | * current thread go in the (practically empty) task context. | |
795 | */ | |
796 | void | |
797 | task_complete_halt(task_t task) | |
798 | { | |
799 | task_lock(task); | |
800 | assert(task->halting); | |
801 | assert(task == current_task()); | |
e7c99d92 A |
802 | |
803 | /* | |
55e303ae | 804 | * Give the machine dependent code a chance |
b0d623f7 A |
805 | * to perform cleanup of task-level resources |
806 | * associated with the current thread before | |
807 | * ripping apart the task. | |
808 | * | |
809 | * This must be done with the task locked. | |
e7c99d92 | 810 | */ |
55e303ae | 811 | machine_thread_terminate_self(); |
e7c99d92 | 812 | |
b0d623f7 A |
813 | /* |
814 | * Wait for the other threads to get shut down. | |
815 | * When the last other thread is reaped, we'll be | |
816 | * worken up. | |
817 | */ | |
818 | if (task->thread_count > 1) { | |
819 | assert_wait((event_t)&task->halting, THREAD_UNINT); | |
820 | task_unlock(task); | |
821 | thread_block(THREAD_CONTINUE_NULL); | |
822 | } else { | |
823 | task_unlock(task); | |
824 | } | |
1c79356b A |
825 | |
826 | /* | |
827 | * Destroy all synchronizers owned by the task. | |
828 | */ | |
829 | task_synchronizer_destroy_all(task); | |
830 | ||
831 | /* | |
9bccf70c A |
832 | * Destroy the contents of the IPC space, leaving just |
833 | * a reference for it. | |
e7c99d92 | 834 | */ |
55e303ae | 835 | ipc_space_clean(task->itk_space); |
1c79356b A |
836 | |
837 | /* | |
838 | * Clean out the address space, as we are going to be | |
839 | * getting a new one. | |
840 | */ | |
91447636 A |
841 | vm_map_remove(task->map, task->map->min_offset, |
842 | task->map->max_offset, VM_MAP_NO_FLAGS); | |
1c79356b | 843 | |
b0d623f7 | 844 | task->halting = FALSE; |
1c79356b A |
845 | } |
846 | ||
847 | /* | |
848 | * task_hold_locked: | |
849 | * | |
850 | * Suspend execution of the specified task. | |
851 | * This is a recursive-style suspension of the task, a count of | |
852 | * suspends is maintained. | |
853 | * | |
854 | * CONDITIONS: the task is locked and active. | |
855 | */ | |
856 | void | |
857 | task_hold_locked( | |
91447636 | 858 | register task_t task) |
1c79356b | 859 | { |
91447636 | 860 | register thread_t thread; |
1c79356b A |
861 | |
862 | assert(task->active); | |
863 | ||
9bccf70c A |
864 | if (task->suspend_count++ > 0) |
865 | return; | |
1c79356b A |
866 | |
867 | /* | |
91447636 | 868 | * Iterate through all the threads and hold them. |
1c79356b | 869 | */ |
91447636 A |
870 | queue_iterate(&task->threads, thread, thread_t, task_threads) { |
871 | thread_mtx_lock(thread); | |
872 | thread_hold(thread); | |
873 | thread_mtx_unlock(thread); | |
1c79356b A |
874 | } |
875 | } | |
876 | ||
877 | /* | |
878 | * task_hold: | |
879 | * | |
880 | * Same as the internal routine above, except that is must lock | |
881 | * and verify that the task is active. This differs from task_suspend | |
882 | * in that it places a kernel hold on the task rather than just a | |
883 | * user-level hold. This keeps users from over resuming and setting | |
884 | * it running out from under the kernel. | |
885 | * | |
886 | * CONDITIONS: the caller holds a reference on the task | |
887 | */ | |
888 | kern_return_t | |
91447636 A |
889 | task_hold( |
890 | register task_t task) | |
1c79356b | 891 | { |
1c79356b A |
892 | if (task == TASK_NULL) |
893 | return (KERN_INVALID_ARGUMENT); | |
91447636 | 894 | |
1c79356b | 895 | task_lock(task); |
91447636 | 896 | |
1c79356b A |
897 | if (!task->active) { |
898 | task_unlock(task); | |
91447636 | 899 | |
1c79356b A |
900 | return (KERN_FAILURE); |
901 | } | |
1c79356b | 902 | |
91447636 A |
903 | task_hold_locked(task); |
904 | task_unlock(task); | |
905 | ||
906 | return (KERN_SUCCESS); | |
1c79356b A |
907 | } |
908 | ||
909 | /* | |
91447636 A |
910 | * task_wait_locked: |
911 | * | |
1c79356b A |
912 | * Wait for all threads in task to stop. |
913 | * | |
914 | * Conditions: | |
915 | * Called with task locked, active, and held. | |
916 | */ | |
917 | void | |
918 | task_wait_locked( | |
919 | register task_t task) | |
920 | { | |
91447636 | 921 | register thread_t thread, self; |
1c79356b A |
922 | |
923 | assert(task->active); | |
924 | assert(task->suspend_count > 0); | |
925 | ||
91447636 A |
926 | self = current_thread(); |
927 | ||
1c79356b | 928 | /* |
91447636 | 929 | * Iterate through all the threads and wait for them to |
1c79356b A |
930 | * stop. Do not wait for the current thread if it is within |
931 | * the task. | |
932 | */ | |
91447636 A |
933 | queue_iterate(&task->threads, thread, thread_t, task_threads) { |
934 | if (thread != self) | |
55e303ae | 935 | thread_wait(thread); |
1c79356b A |
936 | } |
937 | } | |
938 | ||
939 | /* | |
940 | * task_release_locked: | |
941 | * | |
942 | * Release a kernel hold on a task. | |
943 | * | |
944 | * CONDITIONS: the task is locked and active | |
945 | */ | |
946 | void | |
947 | task_release_locked( | |
91447636 | 948 | register task_t task) |
1c79356b | 949 | { |
91447636 | 950 | register thread_t thread; |
1c79356b A |
951 | |
952 | assert(task->active); | |
9bccf70c | 953 | assert(task->suspend_count > 0); |
1c79356b | 954 | |
9bccf70c A |
955 | if (--task->suspend_count > 0) |
956 | return; | |
1c79356b | 957 | |
91447636 A |
958 | queue_iterate(&task->threads, thread, thread_t, task_threads) { |
959 | thread_mtx_lock(thread); | |
960 | thread_release(thread); | |
961 | thread_mtx_unlock(thread); | |
1c79356b A |
962 | } |
963 | } | |
964 | ||
965 | /* | |
966 | * task_release: | |
967 | * | |
968 | * Same as the internal routine above, except that it must lock | |
969 | * and verify that the task is active. | |
970 | * | |
971 | * CONDITIONS: The caller holds a reference to the task | |
972 | */ | |
973 | kern_return_t | |
91447636 A |
974 | task_release( |
975 | task_t task) | |
1c79356b | 976 | { |
1c79356b A |
977 | if (task == TASK_NULL) |
978 | return (KERN_INVALID_ARGUMENT); | |
91447636 | 979 | |
1c79356b | 980 | task_lock(task); |
91447636 | 981 | |
1c79356b A |
982 | if (!task->active) { |
983 | task_unlock(task); | |
91447636 | 984 | |
1c79356b A |
985 | return (KERN_FAILURE); |
986 | } | |
1c79356b | 987 | |
91447636 A |
988 | task_release_locked(task); |
989 | task_unlock(task); | |
990 | ||
991 | return (KERN_SUCCESS); | |
1c79356b A |
992 | } |
993 | ||
994 | kern_return_t | |
995 | task_threads( | |
91447636 A |
996 | task_t task, |
997 | thread_act_array_t *threads_out, | |
1c79356b A |
998 | mach_msg_type_number_t *count) |
999 | { | |
91447636 | 1000 | mach_msg_type_number_t actual; |
2d21ac55 | 1001 | thread_t *thread_list; |
91447636 A |
1002 | thread_t thread; |
1003 | vm_size_t size, size_needed; | |
1004 | void *addr; | |
1005 | unsigned int i, j; | |
1c79356b A |
1006 | |
1007 | if (task == TASK_NULL) | |
91447636 | 1008 | return (KERN_INVALID_ARGUMENT); |
1c79356b | 1009 | |
2d21ac55 | 1010 | size = 0; addr = NULL; |
1c79356b A |
1011 | |
1012 | for (;;) { | |
1013 | task_lock(task); | |
1014 | if (!task->active) { | |
1015 | task_unlock(task); | |
91447636 | 1016 | |
1c79356b A |
1017 | if (size != 0) |
1018 | kfree(addr, size); | |
91447636 A |
1019 | |
1020 | return (KERN_FAILURE); | |
1c79356b A |
1021 | } |
1022 | ||
55e303ae | 1023 | actual = task->thread_count; |
1c79356b A |
1024 | |
1025 | /* do we have the memory we need? */ | |
91447636 | 1026 | size_needed = actual * sizeof (mach_port_t); |
1c79356b A |
1027 | if (size_needed <= size) |
1028 | break; | |
1029 | ||
1030 | /* unlock the task and allocate more memory */ | |
1031 | task_unlock(task); | |
1032 | ||
1033 | if (size != 0) | |
1034 | kfree(addr, size); | |
1035 | ||
1036 | assert(size_needed > 0); | |
1037 | size = size_needed; | |
1038 | ||
1039 | addr = kalloc(size); | |
1040 | if (addr == 0) | |
91447636 | 1041 | return (KERN_RESOURCE_SHORTAGE); |
1c79356b A |
1042 | } |
1043 | ||
1044 | /* OK, have memory and the task is locked & active */ | |
2d21ac55 | 1045 | thread_list = (thread_t *)addr; |
91447636 A |
1046 | |
1047 | i = j = 0; | |
1048 | ||
1049 | for (thread = (thread_t)queue_first(&task->threads); i < actual; | |
1050 | ++i, thread = (thread_t)queue_next(&thread->task_threads)) { | |
1051 | thread_reference_internal(thread); | |
2d21ac55 | 1052 | thread_list[j++] = thread; |
1c79356b | 1053 | } |
91447636 A |
1054 | |
1055 | assert(queue_end(&task->threads, (queue_entry_t)thread)); | |
1c79356b A |
1056 | |
1057 | actual = j; | |
91447636 | 1058 | size_needed = actual * sizeof (mach_port_t); |
1c79356b | 1059 | |
91447636 | 1060 | /* can unlock task now that we've got the thread refs */ |
1c79356b A |
1061 | task_unlock(task); |
1062 | ||
1063 | if (actual == 0) { | |
91447636 | 1064 | /* no threads, so return null pointer and deallocate memory */ |
1c79356b | 1065 | |
2d21ac55 | 1066 | *threads_out = NULL; |
1c79356b A |
1067 | *count = 0; |
1068 | ||
1069 | if (size != 0) | |
1070 | kfree(addr, size); | |
91447636 A |
1071 | } |
1072 | else { | |
1c79356b A |
1073 | /* if we allocated too much, must copy */ |
1074 | ||
1075 | if (size_needed < size) { | |
91447636 | 1076 | void *newaddr; |
1c79356b A |
1077 | |
1078 | newaddr = kalloc(size_needed); | |
1079 | if (newaddr == 0) { | |
91447636 | 1080 | for (i = 0; i < actual; ++i) |
2d21ac55 | 1081 | thread_deallocate(thread_list[i]); |
1c79356b | 1082 | kfree(addr, size); |
91447636 | 1083 | return (KERN_RESOURCE_SHORTAGE); |
1c79356b A |
1084 | } |
1085 | ||
91447636 | 1086 | bcopy(addr, newaddr, size_needed); |
1c79356b | 1087 | kfree(addr, size); |
2d21ac55 | 1088 | thread_list = (thread_t *)newaddr; |
1c79356b A |
1089 | } |
1090 | ||
2d21ac55 | 1091 | *threads_out = thread_list; |
1c79356b A |
1092 | *count = actual; |
1093 | ||
1094 | /* do the conversion that Mig should handle */ | |
1095 | ||
91447636 | 1096 | for (i = 0; i < actual; ++i) |
2d21ac55 | 1097 | ((ipc_port_t *) thread_list)[i] = convert_thread_to_port(thread_list[i]); |
1c79356b A |
1098 | } |
1099 | ||
91447636 | 1100 | return (KERN_SUCCESS); |
1c79356b A |
1101 | } |
1102 | ||
1103 | /* | |
91447636 A |
1104 | * task_suspend: |
1105 | * | |
1c79356b A |
1106 | * Implement a user-level suspension on a task. |
1107 | * | |
1108 | * Conditions: | |
1109 | * The caller holds a reference to the task | |
1110 | */ | |
1111 | kern_return_t | |
1112 | task_suspend( | |
1113 | register task_t task) | |
1114 | { | |
91447636 | 1115 | if (task == TASK_NULL || task == kernel_task) |
1c79356b A |
1116 | return (KERN_INVALID_ARGUMENT); |
1117 | ||
1118 | task_lock(task); | |
91447636 | 1119 | |
1c79356b A |
1120 | if (!task->active) { |
1121 | task_unlock(task); | |
91447636 | 1122 | |
1c79356b A |
1123 | return (KERN_FAILURE); |
1124 | } | |
91447636 A |
1125 | |
1126 | if (task->user_stop_count++ > 0) { | |
1c79356b A |
1127 | /* |
1128 | * If the stop count was positive, the task is | |
1129 | * already stopped and we can exit. | |
1130 | */ | |
1131 | task_unlock(task); | |
91447636 | 1132 | |
1c79356b A |
1133 | return (KERN_SUCCESS); |
1134 | } | |
1135 | ||
1136 | /* | |
1137 | * Put a kernel-level hold on the threads in the task (all | |
1138 | * user-level task suspensions added together represent a | |
1139 | * single kernel-level hold). We then wait for the threads | |
1140 | * to stop executing user code. | |
1141 | */ | |
1142 | task_hold_locked(task); | |
1143 | task_wait_locked(task); | |
91447636 | 1144 | |
1c79356b | 1145 | task_unlock(task); |
91447636 | 1146 | |
1c79356b A |
1147 | return (KERN_SUCCESS); |
1148 | } | |
1149 | ||
1150 | /* | |
91447636 | 1151 | * task_resume: |
1c79356b A |
1152 | * Release a kernel hold on a task. |
1153 | * | |
1154 | * Conditions: | |
1155 | * The caller holds a reference to the task | |
1156 | */ | |
1157 | kern_return_t | |
91447636 A |
1158 | task_resume( |
1159 | register task_t task) | |
1c79356b | 1160 | { |
91447636 | 1161 | register boolean_t release = FALSE; |
1c79356b | 1162 | |
91447636 A |
1163 | if (task == TASK_NULL || task == kernel_task) |
1164 | return (KERN_INVALID_ARGUMENT); | |
1c79356b | 1165 | |
1c79356b | 1166 | task_lock(task); |
91447636 | 1167 | |
1c79356b A |
1168 | if (!task->active) { |
1169 | task_unlock(task); | |
91447636 A |
1170 | |
1171 | return (KERN_FAILURE); | |
1c79356b | 1172 | } |
91447636 | 1173 | |
1c79356b | 1174 | if (task->user_stop_count > 0) { |
6d2010ae | 1175 | if (--task->user_stop_count == 0) { |
91447636 | 1176 | release = TRUE; |
6d2010ae | 1177 | } |
1c79356b A |
1178 | } |
1179 | else { | |
1180 | task_unlock(task); | |
91447636 A |
1181 | |
1182 | return (KERN_FAILURE); | |
1c79356b A |
1183 | } |
1184 | ||
1185 | /* | |
1186 | * Release the task if necessary. | |
1187 | */ | |
1188 | if (release) | |
1189 | task_release_locked(task); | |
1190 | ||
1191 | task_unlock(task); | |
91447636 A |
1192 | |
1193 | return (KERN_SUCCESS); | |
1c79356b A |
1194 | } |
1195 | ||
6d2010ae A |
1196 | #if CONFIG_FREEZE |
1197 | ||
1198 | /* | |
1199 | * task_freeze: | |
1200 | * | |
1201 | * Freeze a currently suspended task. | |
1202 | * | |
1203 | * Conditions: | |
1204 | * The caller holds a reference to the task | |
1205 | */ | |
1206 | kern_return_t | |
1207 | task_freeze( | |
1208 | register task_t task, | |
1209 | uint32_t *purgeable_count, | |
1210 | uint32_t *wired_count, | |
1211 | uint32_t *clean_count, | |
1212 | uint32_t *dirty_count, | |
1213 | boolean_t *shared, | |
1214 | boolean_t walk_only) | |
1215 | { | |
1216 | if (task == TASK_NULL || task == kernel_task) | |
1217 | return (KERN_INVALID_ARGUMENT); | |
1218 | ||
1219 | if (walk_only) { | |
1220 | vm_map_freeze_walk(task->map, purgeable_count, wired_count, clean_count, dirty_count, shared); | |
1221 | } else { | |
1222 | vm_map_freeze(task->map, purgeable_count, wired_count, clean_count, dirty_count, shared); | |
1223 | } | |
1224 | ||
1225 | return (KERN_SUCCESS); | |
1226 | } | |
1227 | ||
1228 | /* | |
1229 | * task_thaw: | |
1230 | * | |
1231 | * Thaw a currently frozen task. | |
1232 | * | |
1233 | * Conditions: | |
1234 | * The caller holds a reference to the task | |
1235 | */ | |
1236 | kern_return_t | |
1237 | task_thaw( | |
1238 | register task_t task) | |
1239 | { | |
1240 | if (task == TASK_NULL || task == kernel_task) | |
1241 | return (KERN_INVALID_ARGUMENT); | |
1242 | ||
1243 | vm_map_thaw(task->map); | |
1244 | ||
1245 | return (KERN_SUCCESS); | |
1246 | } | |
1247 | ||
1248 | #endif /* CONFIG_FREEZE */ | |
1249 | ||
1c79356b A |
1250 | kern_return_t |
1251 | host_security_set_task_token( | |
1252 | host_security_t host_security, | |
1253 | task_t task, | |
1254 | security_token_t sec_token, | |
55e303ae | 1255 | audit_token_t audit_token, |
1c79356b A |
1256 | host_priv_t host_priv) |
1257 | { | |
55e303ae | 1258 | ipc_port_t host_port; |
1c79356b A |
1259 | kern_return_t kr; |
1260 | ||
1261 | if (task == TASK_NULL) | |
1262 | return(KERN_INVALID_ARGUMENT); | |
1263 | ||
1264 | if (host_security == HOST_NULL) | |
1265 | return(KERN_INVALID_SECURITY); | |
1266 | ||
1267 | task_lock(task); | |
1268 | task->sec_token = sec_token; | |
55e303ae | 1269 | task->audit_token = audit_token; |
1c79356b A |
1270 | task_unlock(task); |
1271 | ||
1272 | if (host_priv != HOST_PRIV_NULL) { | |
55e303ae | 1273 | kr = host_get_host_priv_port(host_priv, &host_port); |
1c79356b | 1274 | } else { |
55e303ae | 1275 | kr = host_get_host_port(host_priv_self(), &host_port); |
1c79356b | 1276 | } |
55e303ae A |
1277 | assert(kr == KERN_SUCCESS); |
1278 | kr = task_set_special_port(task, TASK_HOST_PORT, host_port); | |
1c79356b A |
1279 | return(kr); |
1280 | } | |
1281 | ||
1282 | /* | |
1283 | * Utility routine to set a ledger | |
1284 | */ | |
1285 | kern_return_t | |
1286 | task_set_ledger( | |
1287 | task_t task, | |
1288 | ledger_t wired, | |
1289 | ledger_t paged) | |
1290 | { | |
1291 | if (task == TASK_NULL) | |
1292 | return(KERN_INVALID_ARGUMENT); | |
1293 | ||
1294 | task_lock(task); | |
1295 | if (wired) { | |
1296 | ipc_port_release_send(task->wired_ledger_port); | |
1297 | task->wired_ledger_port = ledger_copy(wired); | |
1298 | } | |
1299 | if (paged) { | |
1300 | ipc_port_release_send(task->paged_ledger_port); | |
1301 | task->paged_ledger_port = ledger_copy(paged); | |
1302 | } | |
1303 | task_unlock(task); | |
1304 | ||
1305 | return(KERN_SUCCESS); | |
1306 | } | |
1307 | ||
1308 | /* | |
1309 | * This routine was added, pretty much exclusively, for registering the | |
1310 | * RPC glue vector for in-kernel short circuited tasks. Rather than | |
1311 | * removing it completely, I have only disabled that feature (which was | |
1312 | * the only feature at the time). It just appears that we are going to | |
1313 | * want to add some user data to tasks in the future (i.e. bsd info, | |
1314 | * task names, etc...), so I left it in the formal task interface. | |
1315 | */ | |
1316 | kern_return_t | |
1317 | task_set_info( | |
1318 | task_t task, | |
1319 | task_flavor_t flavor, | |
91447636 A |
1320 | __unused task_info_t task_info_in, /* pointer to IN array */ |
1321 | __unused mach_msg_type_number_t task_info_count) | |
1c79356b | 1322 | { |
1c79356b A |
1323 | if (task == TASK_NULL) |
1324 | return(KERN_INVALID_ARGUMENT); | |
1325 | ||
1326 | switch (flavor) { | |
1327 | default: | |
1328 | return (KERN_INVALID_ARGUMENT); | |
1329 | } | |
1330 | return (KERN_SUCCESS); | |
1331 | } | |
1332 | ||
1333 | kern_return_t | |
1334 | task_info( | |
91447636 A |
1335 | task_t task, |
1336 | task_flavor_t flavor, | |
1337 | task_info_t task_info_out, | |
1c79356b A |
1338 | mach_msg_type_number_t *task_info_count) |
1339 | { | |
b0d623f7 A |
1340 | kern_return_t error = KERN_SUCCESS; |
1341 | ||
1c79356b | 1342 | if (task == TASK_NULL) |
91447636 | 1343 | return (KERN_INVALID_ARGUMENT); |
1c79356b | 1344 | |
b0d623f7 A |
1345 | task_lock(task); |
1346 | ||
1347 | if ((task != current_task()) && (!task->active)) { | |
1348 | task_unlock(task); | |
1349 | return (KERN_INVALID_ARGUMENT); | |
1350 | } | |
1351 | ||
1c79356b A |
1352 | switch (flavor) { |
1353 | ||
91447636 | 1354 | case TASK_BASIC_INFO_32: |
2d21ac55 | 1355 | case TASK_BASIC2_INFO_32: |
91447636 A |
1356 | { |
1357 | task_basic_info_32_t basic_info; | |
b0d623f7 A |
1358 | vm_map_t map; |
1359 | clock_sec_t secs; | |
1360 | clock_usec_t usecs; | |
1c79356b | 1361 | |
b0d623f7 A |
1362 | if (*task_info_count < TASK_BASIC_INFO_32_COUNT) { |
1363 | error = KERN_INVALID_ARGUMENT; | |
1364 | break; | |
1365 | } | |
1c79356b | 1366 | |
91447636 | 1367 | basic_info = (task_basic_info_32_t)task_info_out; |
1c79356b | 1368 | |
91447636 | 1369 | map = (task == kernel_task)? kernel_map: task->map; |
b0d623f7 | 1370 | basic_info->virtual_size = (typeof(basic_info->virtual_size))map->size; |
2d21ac55 A |
1371 | if (flavor == TASK_BASIC2_INFO_32) { |
1372 | /* | |
1373 | * The "BASIC2" flavor gets the maximum resident | |
1374 | * size instead of the current resident size... | |
1375 | */ | |
1376 | basic_info->resident_size = pmap_resident_max(map->pmap); | |
1377 | } else { | |
1378 | basic_info->resident_size = pmap_resident_count(map->pmap); | |
1379 | } | |
1380 | basic_info->resident_size *= PAGE_SIZE; | |
1c79356b | 1381 | |
0b4e3aa0 A |
1382 | basic_info->policy = ((task != kernel_task)? |
1383 | POLICY_TIMESHARE: POLICY_RR); | |
1c79356b | 1384 | basic_info->suspend_count = task->user_stop_count; |
91447636 | 1385 | |
b0d623f7 A |
1386 | absolutetime_to_microtime(task->total_user_time, &secs, &usecs); |
1387 | basic_info->user_time.seconds = | |
1388 | (typeof(basic_info->user_time.seconds))secs; | |
1389 | basic_info->user_time.microseconds = usecs; | |
1390 | ||
1391 | absolutetime_to_microtime(task->total_system_time, &secs, &usecs); | |
1392 | basic_info->system_time.seconds = | |
1393 | (typeof(basic_info->system_time.seconds))secs; | |
1394 | basic_info->system_time.microseconds = usecs; | |
1c79356b | 1395 | |
91447636 | 1396 | *task_info_count = TASK_BASIC_INFO_32_COUNT; |
1c79356b | 1397 | break; |
91447636 | 1398 | } |
1c79356b | 1399 | |
91447636 A |
1400 | case TASK_BASIC_INFO_64: |
1401 | { | |
1402 | task_basic_info_64_t basic_info; | |
b0d623f7 A |
1403 | vm_map_t map; |
1404 | clock_sec_t secs; | |
1405 | clock_usec_t usecs; | |
1c79356b | 1406 | |
b0d623f7 A |
1407 | if (*task_info_count < TASK_BASIC_INFO_64_COUNT) { |
1408 | error = KERN_INVALID_ARGUMENT; | |
1409 | break; | |
1410 | } | |
91447636 A |
1411 | |
1412 | basic_info = (task_basic_info_64_t)task_info_out; | |
1413 | ||
1414 | map = (task == kernel_task)? kernel_map: task->map; | |
1415 | basic_info->virtual_size = map->size; | |
2d21ac55 A |
1416 | basic_info->resident_size = |
1417 | (mach_vm_size_t)(pmap_resident_count(map->pmap)) | |
1418 | * PAGE_SIZE_64; | |
91447636 | 1419 | |
91447636 A |
1420 | basic_info->policy = ((task != kernel_task)? |
1421 | POLICY_TIMESHARE: POLICY_RR); | |
1422 | basic_info->suspend_count = task->user_stop_count; | |
1423 | ||
b0d623f7 A |
1424 | absolutetime_to_microtime(task->total_user_time, &secs, &usecs); |
1425 | basic_info->user_time.seconds = | |
1426 | (typeof(basic_info->user_time.seconds))secs; | |
1427 | basic_info->user_time.microseconds = usecs; | |
1428 | ||
1429 | absolutetime_to_microtime(task->total_system_time, &secs, &usecs); | |
1430 | basic_info->system_time.seconds = | |
1431 | (typeof(basic_info->system_time.seconds))secs; | |
1432 | basic_info->system_time.microseconds = usecs; | |
91447636 A |
1433 | |
1434 | *task_info_count = TASK_BASIC_INFO_64_COUNT; | |
1435 | break; | |
1436 | } | |
1437 | ||
1438 | case TASK_THREAD_TIMES_INFO: | |
1439 | { | |
1440 | register task_thread_times_info_t times_info; | |
1441 | register thread_t thread; | |
1442 | ||
b0d623f7 A |
1443 | if (*task_info_count < TASK_THREAD_TIMES_INFO_COUNT) { |
1444 | error = KERN_INVALID_ARGUMENT; | |
1445 | break; | |
1446 | } | |
1c79356b A |
1447 | |
1448 | times_info = (task_thread_times_info_t) task_info_out; | |
1449 | times_info->user_time.seconds = 0; | |
1450 | times_info->user_time.microseconds = 0; | |
1451 | times_info->system_time.seconds = 0; | |
1452 | times_info->system_time.microseconds = 0; | |
1453 | ||
1c79356b | 1454 | |
91447636 A |
1455 | queue_iterate(&task->threads, thread, thread_t, task_threads) { |
1456 | time_value_t user_time, system_time; | |
1c79356b A |
1457 | |
1458 | thread_read_times(thread, &user_time, &system_time); | |
1459 | ||
1c79356b A |
1460 | time_value_add(×_info->user_time, &user_time); |
1461 | time_value_add(×_info->system_time, &system_time); | |
1462 | } | |
91447636 | 1463 | |
1c79356b A |
1464 | |
1465 | *task_info_count = TASK_THREAD_TIMES_INFO_COUNT; | |
1466 | break; | |
91447636 A |
1467 | } |
1468 | ||
1469 | case TASK_ABSOLUTETIME_INFO: | |
1470 | { | |
1471 | task_absolutetime_info_t info; | |
1472 | register thread_t thread; | |
1473 | ||
b0d623f7 A |
1474 | if (*task_info_count < TASK_ABSOLUTETIME_INFO_COUNT) { |
1475 | error = KERN_INVALID_ARGUMENT; | |
1476 | break; | |
1477 | } | |
91447636 A |
1478 | |
1479 | info = (task_absolutetime_info_t)task_info_out; | |
1480 | info->threads_user = info->threads_system = 0; | |
1481 | ||
91447636 A |
1482 | |
1483 | info->total_user = task->total_user_time; | |
1484 | info->total_system = task->total_system_time; | |
1485 | ||
1486 | queue_iterate(&task->threads, thread, thread_t, task_threads) { | |
1487 | uint64_t tval; | |
1488 | ||
1489 | tval = timer_grab(&thread->user_timer); | |
1490 | info->threads_user += tval; | |
1491 | info->total_user += tval; | |
1492 | ||
1493 | tval = timer_grab(&thread->system_timer); | |
1494 | info->threads_system += tval; | |
1495 | info->total_system += tval; | |
1496 | } | |
1497 | ||
91447636 A |
1498 | |
1499 | *task_info_count = TASK_ABSOLUTETIME_INFO_COUNT; | |
1500 | break; | |
1501 | } | |
1c79356b | 1502 | |
b0d623f7 A |
1503 | case TASK_DYLD_INFO: |
1504 | { | |
1505 | task_dyld_info_t info; | |
1506 | ||
6d2010ae A |
1507 | /* |
1508 | * We added the format field to TASK_DYLD_INFO output. For | |
1509 | * temporary backward compatibility, accept the fact that | |
1510 | * clients may ask for the old version - distinquished by the | |
1511 | * size of the expected result structure. | |
1512 | */ | |
1513 | #define TASK_LEGACY_DYLD_INFO_COUNT \ | |
1514 | offsetof(struct task_dyld_info, all_image_info_format)/sizeof(natural_t) | |
1515 | ||
1516 | if (*task_info_count < TASK_LEGACY_DYLD_INFO_COUNT) { | |
b0d623f7 A |
1517 | error = KERN_INVALID_ARGUMENT; |
1518 | break; | |
1519 | } | |
6d2010ae | 1520 | |
b0d623f7 A |
1521 | info = (task_dyld_info_t)task_info_out; |
1522 | info->all_image_info_addr = task->all_image_info_addr; | |
1523 | info->all_image_info_size = task->all_image_info_size; | |
6d2010ae A |
1524 | |
1525 | /* only set format on output for those expecting it */ | |
1526 | if (*task_info_count >= TASK_DYLD_INFO_COUNT) { | |
1527 | info->all_image_info_format = task_has_64BitAddr(task) ? | |
1528 | TASK_DYLD_ALL_IMAGE_INFO_64 : | |
1529 | TASK_DYLD_ALL_IMAGE_INFO_32 ; | |
1530 | *task_info_count = TASK_DYLD_INFO_COUNT; | |
1531 | } else { | |
1532 | *task_info_count = TASK_LEGACY_DYLD_INFO_COUNT; | |
1533 | } | |
b0d623f7 A |
1534 | break; |
1535 | } | |
1536 | ||
6d2010ae A |
1537 | case TASK_EXTMOD_INFO: |
1538 | { | |
1539 | task_extmod_info_t info; | |
1540 | void *p; | |
1541 | ||
1542 | if (*task_info_count < TASK_EXTMOD_INFO_COUNT) { | |
1543 | error = KERN_INVALID_ARGUMENT; | |
1544 | break; | |
1545 | } | |
1546 | ||
1547 | info = (task_extmod_info_t)task_info_out; | |
1548 | ||
1549 | p = get_bsdtask_info(task); | |
1550 | if (p) { | |
1551 | proc_getexecutableuuid(p, info->task_uuid, sizeof(info->task_uuid)); | |
1552 | } else { | |
1553 | bzero(info->task_uuid, sizeof(info->task_uuid)); | |
1554 | } | |
1555 | info->extmod_statistics = task->extmod_statistics; | |
1556 | *task_info_count = TASK_EXTMOD_INFO_COUNT; | |
1557 | ||
1558 | break; | |
1559 | } | |
1560 | ||
1561 | case TASK_KERNELMEMORY_INFO: | |
1562 | { | |
1563 | task_kernelmemory_info_t tkm_info; | |
1564 | thread_t thread; | |
1565 | ||
1566 | if (*task_info_count < TASK_KERNELMEMORY_INFO_COUNT) { | |
1567 | error = KERN_INVALID_ARGUMENT; | |
1568 | break; | |
1569 | } | |
1570 | ||
1571 | tkm_info = (task_kernelmemory_info_t) task_info_out; | |
1572 | ||
1573 | if (task == kernel_task) { | |
1574 | /* | |
1575 | * All shared allocs/frees from other tasks count against | |
1576 | * the kernel private memory usage. If we are looking up | |
1577 | * info for the kernel task, gather from everywhere. | |
1578 | */ | |
1579 | task_unlock(task); | |
1580 | ||
1581 | /* start by accounting for all the terminated tasks against the kernel */ | |
1582 | tkm_info->total_palloc = tasks_tkm_private.alloc + tasks_tkm_shared.alloc; | |
1583 | tkm_info->total_pfree = tasks_tkm_private.free + tasks_tkm_shared.free; | |
1584 | tkm_info->total_salloc = 0; | |
1585 | tkm_info->total_sfree = 0; | |
1586 | ||
1587 | /* count all other task/thread shared alloc/free against the kernel */ | |
1588 | lck_mtx_lock(&tasks_threads_lock); | |
1589 | queue_iterate(&tasks, task, task_t, tasks) { | |
1590 | if (task == kernel_task) { | |
1591 | tkm_info->total_palloc += task->tkm_private.alloc; | |
1592 | tkm_info->total_pfree += task->tkm_private.free; | |
1593 | } | |
1594 | tkm_info->total_palloc += task->tkm_shared.alloc; | |
1595 | tkm_info->total_pfree += task->tkm_shared.free; | |
1596 | } | |
1597 | queue_iterate(&threads, thread, thread_t, threads) { | |
1598 | if (thread->task == kernel_task) { | |
1599 | tkm_info->total_palloc += thread->tkm_private.alloc; | |
1600 | tkm_info->total_pfree += thread->tkm_private.free; | |
1601 | } | |
1602 | tkm_info->total_palloc += thread->tkm_shared.alloc; | |
1603 | tkm_info->total_pfree += thread->tkm_shared.free; | |
1604 | } | |
1605 | lck_mtx_unlock(&tasks_threads_lock); | |
1606 | } else { | |
1607 | /* account for all the terminated threads in the process */ | |
1608 | tkm_info->total_palloc = task->tkm_private.alloc; | |
1609 | tkm_info->total_pfree = task->tkm_private.free; | |
1610 | tkm_info->total_salloc = task->tkm_shared.alloc; | |
1611 | tkm_info->total_sfree = task->tkm_shared.free; | |
1612 | ||
1613 | /* then add in all the running threads */ | |
1614 | queue_iterate(&task->threads, thread, thread_t, task_threads) { | |
1615 | tkm_info->total_palloc += thread->tkm_private.alloc; | |
1616 | tkm_info->total_pfree += thread->tkm_private.free; | |
1617 | tkm_info->total_salloc += thread->tkm_shared.alloc; | |
1618 | tkm_info->total_sfree += thread->tkm_shared.free; | |
1619 | } | |
1620 | task_unlock(task); | |
1621 | } | |
1622 | ||
1623 | *task_info_count = TASK_KERNELMEMORY_INFO_COUNT; | |
1624 | return KERN_SUCCESS; | |
1625 | } | |
1626 | ||
91447636 A |
1627 | /* OBSOLETE */ |
1628 | case TASK_SCHED_FIFO_INFO: | |
1629 | { | |
1c79356b | 1630 | |
b0d623f7 A |
1631 | if (*task_info_count < POLICY_FIFO_BASE_COUNT) { |
1632 | error = KERN_INVALID_ARGUMENT; | |
1633 | break; | |
1634 | } | |
1c79356b | 1635 | |
b0d623f7 | 1636 | error = KERN_INVALID_POLICY; |
6d2010ae | 1637 | break; |
91447636 | 1638 | } |
1c79356b | 1639 | |
91447636 A |
1640 | /* OBSOLETE */ |
1641 | case TASK_SCHED_RR_INFO: | |
1642 | { | |
1c79356b | 1643 | register policy_rr_base_t rr_base; |
6d2010ae A |
1644 | uint32_t quantum_time; |
1645 | uint64_t quantum_ns; | |
1c79356b | 1646 | |
b0d623f7 A |
1647 | if (*task_info_count < POLICY_RR_BASE_COUNT) { |
1648 | error = KERN_INVALID_ARGUMENT; | |
1649 | break; | |
1650 | } | |
1c79356b A |
1651 | |
1652 | rr_base = (policy_rr_base_t) task_info_out; | |
1653 | ||
0b4e3aa0 | 1654 | if (task != kernel_task) { |
b0d623f7 A |
1655 | error = KERN_INVALID_POLICY; |
1656 | break; | |
1c79356b A |
1657 | } |
1658 | ||
1659 | rr_base->base_priority = task->priority; | |
1c79356b | 1660 | |
6d2010ae A |
1661 | quantum_time = SCHED(initial_quantum_size)(THREAD_NULL); |
1662 | absolutetime_to_nanoseconds(quantum_time, &quantum_ns); | |
1663 | ||
1664 | rr_base->quantum = (uint32_t)(quantum_ns / 1000 / 1000); | |
1c79356b A |
1665 | |
1666 | *task_info_count = POLICY_RR_BASE_COUNT; | |
1667 | break; | |
91447636 | 1668 | } |
1c79356b | 1669 | |
91447636 A |
1670 | /* OBSOLETE */ |
1671 | case TASK_SCHED_TIMESHARE_INFO: | |
1672 | { | |
1c79356b A |
1673 | register policy_timeshare_base_t ts_base; |
1674 | ||
b0d623f7 A |
1675 | if (*task_info_count < POLICY_TIMESHARE_BASE_COUNT) { |
1676 | error = KERN_INVALID_ARGUMENT; | |
1677 | break; | |
1678 | } | |
1c79356b A |
1679 | |
1680 | ts_base = (policy_timeshare_base_t) task_info_out; | |
1681 | ||
0b4e3aa0 | 1682 | if (task == kernel_task) { |
b0d623f7 A |
1683 | error = KERN_INVALID_POLICY; |
1684 | break; | |
1c79356b A |
1685 | } |
1686 | ||
1687 | ts_base->base_priority = task->priority; | |
1c79356b A |
1688 | |
1689 | *task_info_count = POLICY_TIMESHARE_BASE_COUNT; | |
1690 | break; | |
91447636 | 1691 | } |
1c79356b | 1692 | |
91447636 A |
1693 | case TASK_SECURITY_TOKEN: |
1694 | { | |
1695 | register security_token_t *sec_token_p; | |
1c79356b | 1696 | |
b0d623f7 A |
1697 | if (*task_info_count < TASK_SECURITY_TOKEN_COUNT) { |
1698 | error = KERN_INVALID_ARGUMENT; | |
1699 | break; | |
1700 | } | |
1c79356b A |
1701 | |
1702 | sec_token_p = (security_token_t *) task_info_out; | |
1703 | ||
1c79356b | 1704 | *sec_token_p = task->sec_token; |
1c79356b A |
1705 | |
1706 | *task_info_count = TASK_SECURITY_TOKEN_COUNT; | |
91447636 A |
1707 | break; |
1708 | } | |
1c79356b | 1709 | |
91447636 A |
1710 | case TASK_AUDIT_TOKEN: |
1711 | { | |
1712 | register audit_token_t *audit_token_p; | |
55e303ae | 1713 | |
b0d623f7 A |
1714 | if (*task_info_count < TASK_AUDIT_TOKEN_COUNT) { |
1715 | error = KERN_INVALID_ARGUMENT; | |
1716 | break; | |
1717 | } | |
55e303ae A |
1718 | |
1719 | audit_token_p = (audit_token_t *) task_info_out; | |
1720 | ||
55e303ae | 1721 | *audit_token_p = task->audit_token; |
55e303ae A |
1722 | |
1723 | *task_info_count = TASK_AUDIT_TOKEN_COUNT; | |
91447636 A |
1724 | break; |
1725 | } | |
55e303ae | 1726 | |
91447636 | 1727 | case TASK_SCHED_INFO: |
b0d623f7 | 1728 | error = KERN_INVALID_ARGUMENT; |
6d2010ae | 1729 | break; |
1c79356b | 1730 | |
91447636 A |
1731 | case TASK_EVENTS_INFO: |
1732 | { | |
1c79356b | 1733 | register task_events_info_t events_info; |
2d21ac55 | 1734 | register thread_t thread; |
1c79356b | 1735 | |
b0d623f7 A |
1736 | if (*task_info_count < TASK_EVENTS_INFO_COUNT) { |
1737 | error = KERN_INVALID_ARGUMENT; | |
1738 | break; | |
1739 | } | |
1c79356b A |
1740 | |
1741 | events_info = (task_events_info_t) task_info_out; | |
1742 | ||
2d21ac55 | 1743 | |
1c79356b A |
1744 | events_info->faults = task->faults; |
1745 | events_info->pageins = task->pageins; | |
1746 | events_info->cow_faults = task->cow_faults; | |
1747 | events_info->messages_sent = task->messages_sent; | |
1748 | events_info->messages_received = task->messages_received; | |
1749 | events_info->syscalls_mach = task->syscalls_mach; | |
1750 | events_info->syscalls_unix = task->syscalls_unix; | |
2d21ac55 A |
1751 | |
1752 | events_info->csw = task->c_switch; | |
1753 | ||
1754 | queue_iterate(&task->threads, thread, thread_t, task_threads) { | |
6d2010ae A |
1755 | events_info->csw += thread->c_switch; |
1756 | events_info->syscalls_mach += thread->syscalls_mach; | |
1757 | events_info->syscalls_unix += thread->syscalls_unix; | |
2d21ac55 A |
1758 | } |
1759 | ||
1c79356b A |
1760 | |
1761 | *task_info_count = TASK_EVENTS_INFO_COUNT; | |
1762 | break; | |
91447636 | 1763 | } |
2d21ac55 A |
1764 | case TASK_AFFINITY_TAG_INFO: |
1765 | { | |
b0d623f7 A |
1766 | if (*task_info_count < TASK_AFFINITY_TAG_INFO_COUNT) { |
1767 | error = KERN_INVALID_ARGUMENT; | |
1768 | break; | |
1769 | } | |
2d21ac55 | 1770 | |
b0d623f7 | 1771 | error = task_affinity_info(task, task_info_out, task_info_count); |
6d2010ae | 1772 | break; |
2d21ac55 | 1773 | } |
91447636 | 1774 | default: |
b0d623f7 | 1775 | error = KERN_INVALID_ARGUMENT; |
1c79356b A |
1776 | } |
1777 | ||
b0d623f7 A |
1778 | task_unlock(task); |
1779 | return (error); | |
1c79356b A |
1780 | } |
1781 | ||
2d21ac55 A |
1782 | void |
1783 | task_vtimer_set( | |
1784 | task_t task, | |
1785 | integer_t which) | |
1786 | { | |
1787 | thread_t thread; | |
1788 | ||
1789 | /* assert(task == current_task()); */ /* bogus assert 4803227 4807483 */ | |
1790 | ||
1791 | task_lock(task); | |
1792 | ||
1793 | task->vtimers |= which; | |
1794 | ||
1795 | switch (which) { | |
1796 | ||
1797 | case TASK_VTIMER_USER: | |
1798 | queue_iterate(&task->threads, thread, thread_t, task_threads) { | |
1799 | thread->vtimer_user_save = timer_grab(&thread->user_timer); | |
1800 | } | |
1801 | break; | |
1802 | ||
1803 | case TASK_VTIMER_PROF: | |
1804 | queue_iterate(&task->threads, thread, thread_t, task_threads) { | |
1805 | thread->vtimer_prof_save = timer_grab(&thread->user_timer); | |
1806 | thread->vtimer_prof_save += timer_grab(&thread->system_timer); | |
1807 | } | |
1808 | break; | |
1809 | ||
1810 | case TASK_VTIMER_RLIM: | |
1811 | queue_iterate(&task->threads, thread, thread_t, task_threads) { | |
1812 | thread->vtimer_rlim_save = timer_grab(&thread->user_timer); | |
1813 | thread->vtimer_rlim_save += timer_grab(&thread->system_timer); | |
1814 | } | |
1815 | break; | |
1816 | } | |
1817 | ||
1818 | task_unlock(task); | |
1819 | } | |
1820 | ||
1821 | void | |
1822 | task_vtimer_clear( | |
1823 | task_t task, | |
1824 | integer_t which) | |
1825 | { | |
1826 | assert(task == current_task()); | |
1827 | ||
1828 | task_lock(task); | |
1829 | ||
1830 | task->vtimers &= ~which; | |
1831 | ||
1832 | task_unlock(task); | |
1833 | } | |
1834 | ||
1835 | void | |
1836 | task_vtimer_update( | |
1837 | __unused | |
1838 | task_t task, | |
1839 | integer_t which, | |
1840 | uint32_t *microsecs) | |
1841 | { | |
1842 | thread_t thread = current_thread(); | |
b0d623f7 A |
1843 | uint32_t tdelt; |
1844 | clock_sec_t secs; | |
2d21ac55 A |
1845 | uint64_t tsum; |
1846 | ||
1847 | assert(task == current_task()); | |
1848 | ||
1849 | assert(task->vtimers & which); | |
1850 | ||
b0d623f7 | 1851 | secs = tdelt = 0; |
2d21ac55 A |
1852 | |
1853 | switch (which) { | |
1854 | ||
1855 | case TASK_VTIMER_USER: | |
b0d623f7 | 1856 | tdelt = (uint32_t)timer_delta(&thread->user_timer, |
2d21ac55 | 1857 | &thread->vtimer_user_save); |
b0d623f7 | 1858 | absolutetime_to_microtime(tdelt, &secs, microsecs); |
2d21ac55 A |
1859 | break; |
1860 | ||
1861 | case TASK_VTIMER_PROF: | |
1862 | tsum = timer_grab(&thread->user_timer); | |
1863 | tsum += timer_grab(&thread->system_timer); | |
b0d623f7 A |
1864 | tdelt = (uint32_t)(tsum - thread->vtimer_prof_save); |
1865 | absolutetime_to_microtime(tdelt, &secs, microsecs); | |
1866 | /* if the time delta is smaller than a usec, ignore */ | |
1867 | if (*microsecs != 0) | |
1868 | thread->vtimer_prof_save = tsum; | |
2d21ac55 A |
1869 | break; |
1870 | ||
1871 | case TASK_VTIMER_RLIM: | |
1872 | tsum = timer_grab(&thread->user_timer); | |
1873 | tsum += timer_grab(&thread->system_timer); | |
b0d623f7 | 1874 | tdelt = (uint32_t)(tsum - thread->vtimer_rlim_save); |
2d21ac55 | 1875 | thread->vtimer_rlim_save = tsum; |
b0d623f7 | 1876 | absolutetime_to_microtime(tdelt, &secs, microsecs); |
2d21ac55 A |
1877 | break; |
1878 | } | |
1879 | ||
2d21ac55 A |
1880 | } |
1881 | ||
1c79356b A |
1882 | /* |
1883 | * task_assign: | |
1884 | * | |
1885 | * Change the assigned processor set for the task | |
1886 | */ | |
1887 | kern_return_t | |
1888 | task_assign( | |
91447636 A |
1889 | __unused task_t task, |
1890 | __unused processor_set_t new_pset, | |
1891 | __unused boolean_t assign_threads) | |
1c79356b | 1892 | { |
1c79356b A |
1893 | return(KERN_FAILURE); |
1894 | } | |
1895 | ||
1896 | /* | |
1897 | * task_assign_default: | |
1898 | * | |
1899 | * Version of task_assign to assign to default processor set. | |
1900 | */ | |
1901 | kern_return_t | |
1902 | task_assign_default( | |
1903 | task_t task, | |
1904 | boolean_t assign_threads) | |
1905 | { | |
2d21ac55 | 1906 | return (task_assign(task, &pset0, assign_threads)); |
1c79356b A |
1907 | } |
1908 | ||
1909 | /* | |
1910 | * task_get_assignment | |
1911 | * | |
1912 | * Return name of processor set that task is assigned to. | |
1913 | */ | |
1914 | kern_return_t | |
1915 | task_get_assignment( | |
1916 | task_t task, | |
1917 | processor_set_t *pset) | |
1918 | { | |
1919 | if (!task->active) | |
1920 | return(KERN_FAILURE); | |
1921 | ||
2d21ac55 A |
1922 | *pset = &pset0; |
1923 | ||
1924 | return (KERN_SUCCESS); | |
1c79356b A |
1925 | } |
1926 | ||
1927 | ||
1928 | /* | |
1929 | * task_policy | |
1930 | * | |
1931 | * Set scheduling policy and parameters, both base and limit, for | |
1932 | * the given task. Policy must be a policy which is enabled for the | |
1933 | * processor set. Change contained threads if requested. | |
1934 | */ | |
1935 | kern_return_t | |
1936 | task_policy( | |
91447636 A |
1937 | __unused task_t task, |
1938 | __unused policy_t policy_id, | |
1939 | __unused policy_base_t base, | |
1940 | __unused mach_msg_type_number_t count, | |
1941 | __unused boolean_t set_limit, | |
1942 | __unused boolean_t change) | |
1c79356b A |
1943 | { |
1944 | return(KERN_FAILURE); | |
1945 | } | |
1946 | ||
1947 | /* | |
1948 | * task_set_policy | |
1949 | * | |
1950 | * Set scheduling policy and parameters, both base and limit, for | |
1951 | * the given task. Policy can be any policy implemented by the | |
1952 | * processor set, whether enabled or not. Change contained threads | |
1953 | * if requested. | |
1954 | */ | |
1955 | kern_return_t | |
1956 | task_set_policy( | |
91447636 A |
1957 | __unused task_t task, |
1958 | __unused processor_set_t pset, | |
1959 | __unused policy_t policy_id, | |
1960 | __unused policy_base_t base, | |
1961 | __unused mach_msg_type_number_t base_count, | |
1962 | __unused policy_limit_t limit, | |
1963 | __unused mach_msg_type_number_t limit_count, | |
1964 | __unused boolean_t change) | |
1c79356b A |
1965 | { |
1966 | return(KERN_FAILURE); | |
1967 | } | |
1968 | ||
91447636 | 1969 | #if FAST_TAS |
1c79356b A |
1970 | kern_return_t |
1971 | task_set_ras_pc( | |
1972 | task_t task, | |
1973 | vm_offset_t pc, | |
1974 | vm_offset_t endpc) | |
1975 | { | |
1c79356b A |
1976 | extern int fast_tas_debug; |
1977 | ||
1978 | if (fast_tas_debug) { | |
1979 | printf("task 0x%x: setting fast_tas to [0x%x, 0x%x]\n", | |
1980 | task, pc, endpc); | |
1981 | } | |
1982 | task_lock(task); | |
1983 | task->fast_tas_base = pc; | |
1984 | task->fast_tas_end = endpc; | |
1985 | task_unlock(task); | |
1986 | return KERN_SUCCESS; | |
91447636 | 1987 | } |
1c79356b | 1988 | #else /* FAST_TAS */ |
91447636 A |
1989 | kern_return_t |
1990 | task_set_ras_pc( | |
1991 | __unused task_t task, | |
1992 | __unused vm_offset_t pc, | |
1993 | __unused vm_offset_t endpc) | |
1994 | { | |
1c79356b | 1995 | return KERN_FAILURE; |
1c79356b | 1996 | } |
91447636 | 1997 | #endif /* FAST_TAS */ |
1c79356b A |
1998 | |
1999 | void | |
2000 | task_synchronizer_destroy_all(task_t task) | |
2001 | { | |
2002 | semaphore_t semaphore; | |
2003 | lock_set_t lock_set; | |
2004 | ||
2005 | /* | |
2006 | * Destroy owned semaphores | |
2007 | */ | |
2008 | ||
2009 | while (!queue_empty(&task->semaphore_list)) { | |
2010 | semaphore = (semaphore_t) queue_first(&task->semaphore_list); | |
2011 | (void) semaphore_destroy(task, semaphore); | |
2012 | } | |
2013 | ||
2014 | /* | |
2015 | * Destroy owned lock sets | |
2016 | */ | |
2017 | ||
2018 | while (!queue_empty(&task->lock_set_list)) { | |
2019 | lock_set = (lock_set_t) queue_first(&task->lock_set_list); | |
2020 | (void) lock_set_destroy(task, lock_set); | |
2021 | } | |
2022 | } | |
2023 | ||
b0d623f7 A |
2024 | /* |
2025 | * Install default (machine-dependent) initial thread state | |
2026 | * on the task. Subsequent thread creation will have this initial | |
2027 | * state set on the thread by machine_thread_inherit_taskwide(). | |
2028 | * Flavors and structures are exactly the same as those to thread_set_state() | |
2029 | */ | |
2030 | kern_return_t | |
2031 | task_set_state( | |
2032 | task_t task, | |
2033 | int flavor, | |
2034 | thread_state_t state, | |
2035 | mach_msg_type_number_t state_count) | |
2036 | { | |
2037 | kern_return_t ret; | |
2038 | ||
2039 | if (task == TASK_NULL) { | |
2040 | return (KERN_INVALID_ARGUMENT); | |
2041 | } | |
2042 | ||
2043 | task_lock(task); | |
2044 | ||
2045 | if (!task->active) { | |
2046 | task_unlock(task); | |
2047 | return (KERN_FAILURE); | |
2048 | } | |
2049 | ||
2050 | ret = machine_task_set_state(task, flavor, state, state_count); | |
2051 | ||
2052 | task_unlock(task); | |
2053 | return ret; | |
2054 | } | |
2055 | ||
2056 | /* | |
2057 | * Examine the default (machine-dependent) initial thread state | |
2058 | * on the task, as set by task_set_state(). Flavors and structures | |
2059 | * are exactly the same as those passed to thread_get_state(). | |
2060 | */ | |
2061 | kern_return_t | |
2062 | task_get_state( | |
2063 | task_t task, | |
2064 | int flavor, | |
2065 | thread_state_t state, | |
2066 | mach_msg_type_number_t *state_count) | |
2067 | { | |
2068 | kern_return_t ret; | |
2069 | ||
2070 | if (task == TASK_NULL) { | |
2071 | return (KERN_INVALID_ARGUMENT); | |
2072 | } | |
2073 | ||
2074 | task_lock(task); | |
2075 | ||
2076 | if (!task->active) { | |
2077 | task_unlock(task); | |
2078 | return (KERN_FAILURE); | |
2079 | } | |
2080 | ||
2081 | ret = machine_task_get_state(task, flavor, state, state_count); | |
2082 | ||
2083 | task_unlock(task); | |
2084 | return ret; | |
2085 | } | |
2086 | ||
2087 | ||
1c79356b A |
2088 | /* |
2089 | * We need to export some functions to other components that | |
2090 | * are currently implemented in macros within the osfmk | |
2091 | * component. Just export them as functions of the same name. | |
2092 | */ | |
2093 | boolean_t is_kerneltask(task_t t) | |
2094 | { | |
2095 | if (t == kernel_task) | |
55e303ae A |
2096 | return (TRUE); |
2097 | ||
2098 | return (FALSE); | |
1c79356b A |
2099 | } |
2100 | ||
b0d623f7 A |
2101 | int |
2102 | check_for_tasksuspend(task_t task) | |
2103 | { | |
2104 | ||
2105 | if (task == TASK_NULL) | |
2106 | return (0); | |
2107 | ||
2108 | return (task->suspend_count > 0); | |
2109 | } | |
2110 | ||
1c79356b | 2111 | #undef current_task |
91447636 A |
2112 | task_t current_task(void); |
2113 | task_t current_task(void) | |
1c79356b A |
2114 | { |
2115 | return (current_task_fast()); | |
2116 | } | |
91447636 A |
2117 | |
2118 | #undef task_reference | |
2119 | void task_reference(task_t task); | |
2120 | void | |
2121 | task_reference( | |
2122 | task_t task) | |
2123 | { | |
2124 | if (task != TASK_NULL) | |
2125 | task_reference_internal(task); | |
2126 | } | |
2d21ac55 | 2127 | |
6d2010ae A |
2128 | /* |
2129 | * This routine is called always with task lock held. | |
2130 | * And it returns a thread handle without reference as the caller | |
2131 | * operates on it under the task lock held. | |
2132 | */ | |
2133 | thread_t | |
2134 | task_findtid(task_t task, uint64_t tid) | |
2135 | { | |
2136 | thread_t thread= THREAD_NULL; | |
2137 | ||
2138 | queue_iterate(&task->threads, thread, thread_t, task_threads) { | |
2139 | if (thread->thread_id == tid) | |
2140 | break; | |
2141 | } | |
2142 | return(thread); | |
2143 | } | |
2144 | ||
2145 | ||
2d21ac55 A |
2146 | #if CONFIG_MACF_MACH |
2147 | /* | |
2148 | * Protect 2 task labels against modification by adding a reference on | |
2149 | * both label handles. The locks do not actually have to be held while | |
2150 | * using the labels as only labels with one reference can be modified | |
2151 | * in place. | |
2152 | */ | |
2153 | ||
2154 | void | |
2155 | tasklabel_lock2( | |
2156 | task_t a, | |
2157 | task_t b) | |
2158 | { | |
2159 | labelh_reference(a->label); | |
2160 | labelh_reference(b->label); | |
2161 | } | |
2162 | ||
2163 | void | |
2164 | tasklabel_unlock2( | |
2165 | task_t a, | |
2166 | task_t b) | |
2167 | { | |
2168 | labelh_release(a->label); | |
2169 | labelh_release(b->label); | |
2170 | } | |
2171 | ||
2172 | void | |
2173 | mac_task_label_update_internal( | |
2174 | struct label *pl, | |
2175 | struct task *task) | |
2176 | { | |
2177 | ||
2178 | tasklabel_lock(task); | |
2179 | task->label = labelh_modify(task->label); | |
2180 | mac_task_label_update(pl, &task->maclabel); | |
2181 | tasklabel_unlock(task); | |
2182 | ip_lock(task->itk_self); | |
2183 | mac_port_label_update_cred(pl, &task->itk_self->ip_label); | |
2184 | ip_unlock(task->itk_self); | |
2185 | } | |
2186 | ||
2187 | void | |
2188 | mac_task_label_modify( | |
2189 | struct task *task, | |
2190 | void *arg, | |
2191 | void (*f) (struct label *l, void *arg)) | |
2192 | { | |
2193 | ||
2194 | tasklabel_lock(task); | |
2195 | task->label = labelh_modify(task->label); | |
2196 | (*f)(&task->maclabel, arg); | |
2197 | tasklabel_unlock(task); | |
2198 | } | |
2199 | ||
2200 | struct label * | |
2201 | mac_task_get_label(struct task *task) | |
2202 | { | |
2203 | return (&task->maclabel); | |
2204 | } | |
2205 | #endif |