]>
Commit | Line | Data |
---|---|---|
2d21ac55 A |
1 | /* |
2 | * Copyright (c) 2007 Apple Inc. All rights reserved. | |
3 | * | |
4 | * @APPLE_OSREFERENCE_LICENSE_HEADER_START@ | |
0a7de745 | 5 | * |
2d21ac55 A |
6 | * This file contains Original Code and/or Modifications of Original Code |
7 | * as defined in and that are subject to the Apple Public Source License | |
8 | * Version 2.0 (the 'License'). You may not use this file except in | |
9 | * compliance with the License. The rights granted to you under the License | |
10 | * may not be used to create, or enable the creation or redistribution of, | |
11 | * unlawful or unlicensed copies of an Apple operating system, or to | |
12 | * circumvent, violate, or enable the circumvention or violation of, any | |
13 | * terms of an Apple operating system software license agreement. | |
0a7de745 | 14 | * |
2d21ac55 A |
15 | * Please obtain a copy of the License at |
16 | * http://www.opensource.apple.com/apsl/ and read it before using this file. | |
0a7de745 | 17 | * |
2d21ac55 A |
18 | * The Original Code and all software distributed under the License are |
19 | * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER | |
20 | * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES, | |
21 | * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY, | |
22 | * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT. | |
23 | * Please see the License for the specific language governing rights and | |
24 | * limitations under the License. | |
0a7de745 | 25 | * |
2d21ac55 A |
26 | * @APPLE_OSREFERENCE_LICENSE_HEADER_END@ |
27 | */ | |
28 | ||
29 | #include <kern/affinity.h> | |
30 | #include <kern/task.h> | |
31 | #include <kern/kalloc.h> | |
32 | #include <machine/cpu_affinity.h> | |
33 | ||
34 | /* | |
35 | * Affinity involves 2 objects: | |
36 | * - affinity namespace: | |
37 | * shared by a task family, this controls affinity tag lookup and | |
38 | * allocation; it anchors all affinity sets in one namespace | |
39 | * - affinity set: | |
0a7de745 | 40 | * anchors all threads with membership of this affinity set |
2d21ac55 | 41 | * and which share an affinity tag in the owning namespace. |
0a7de745 | 42 | * |
2d21ac55 A |
43 | * Locking: |
44 | * - The task lock protects the creation of an affinity namespace. | |
45 | * - The affinity namespace mutex protects the inheritance of a namespace | |
46 | * and its thread membership. This includes its destruction when the task | |
47 | * reference count goes to zero. | |
48 | * - The thread mutex protects a thread's affinity set membership, but in | |
49 | * addition, the thread_lock is taken to write thread->affinity_set since this | |
50 | * field (representng the active affinity set) is read by the scheduler. | |
0a7de745 | 51 | * |
2d21ac55 A |
52 | * The lock ordering is: task lock, thread mutex, namespace mutex, thread lock. |
53 | */ | |
54 | ||
55 | #if AFFINITY_DEBUG | |
0a7de745 | 56 | #define DBG(x...) kprintf("DBG: " x) |
2d21ac55 A |
57 | #else |
58 | #define DBG(x...) | |
59 | #endif | |
60 | ||
61 | struct affinity_space { | |
0a7de745 A |
62 | lck_mtx_t aspc_lock; |
63 | uint32_t aspc_task_count; | |
64 | queue_head_t aspc_affinities; | |
2d21ac55 A |
65 | }; |
66 | typedef struct affinity_space *affinity_space_t; | |
67 | ||
68 | static affinity_space_t affinity_space_alloc(void); | |
69 | static void affinity_space_free(affinity_space_t aspc); | |
70 | static affinity_set_t affinity_set_alloc(void); | |
71 | static void affinity_set_free(affinity_set_t aset); | |
72 | static affinity_set_t affinity_set_find(affinity_space_t aspc, uint32_t tag); | |
73 | static void affinity_set_place(affinity_space_t aspc, affinity_set_t aset); | |
74 | static void affinity_set_add(affinity_set_t aset, thread_t thread); | |
75 | static affinity_set_t affinity_set_remove(affinity_set_t aset, thread_t thread); | |
76 | ||
77 | /* | |
78 | * The following globals may be modified by the sysctls | |
79 | * kern.affinity_sets_enabled - disables hinting if cleared | |
80 | * kern.affinity_sets_mapping - controls cache distribution policy | |
81 | * See bsd/kern_sysctl.c | |
fe8ab488 A |
82 | * |
83 | * Affinity sets are not used on embedded, which typically only | |
84 | * has a single pset, and last-processor affinity is | |
85 | * more important than pset affinity. | |
2d21ac55 | 86 | */ |
5ba3f43e | 87 | #if CONFIG_EMBEDDED |
0a7de745 A |
88 | boolean_t affinity_sets_enabled = FALSE; |
89 | int affinity_sets_mapping = 0; | |
5ba3f43e | 90 | #else /* !CONFIG_EMBEDDED */ |
0a7de745 A |
91 | boolean_t affinity_sets_enabled = TRUE; |
92 | int affinity_sets_mapping = 1; | |
5ba3f43e | 93 | #endif /* !CONFIG_EMBEDDED */ |
2d21ac55 A |
94 | |
95 | boolean_t | |
96 | thread_affinity_is_supported(void) | |
97 | { | |
0a7de745 | 98 | return ml_get_max_affinity_sets() != 0; |
2d21ac55 A |
99 | } |
100 | ||
101 | ||
102 | /* | |
0a7de745 A |
103 | * thread_affinity_get() |
104 | * Return the affinity tag for a thread. | |
2d21ac55 A |
105 | * Called with the thread mutex held. |
106 | */ | |
107 | uint32_t | |
108 | thread_affinity_get(thread_t thread) | |
109 | { | |
110 | uint32_t tag; | |
111 | ||
0a7de745 | 112 | if (thread->affinity_set != NULL) { |
2d21ac55 | 113 | tag = thread->affinity_set->aset_tag; |
0a7de745 | 114 | } else { |
2d21ac55 | 115 | tag = THREAD_AFFINITY_TAG_NULL; |
0a7de745 | 116 | } |
2d21ac55 A |
117 | |
118 | return tag; | |
119 | } | |
120 | ||
121 | ||
122 | /* | |
0a7de745 | 123 | * thread_affinity_set() |
2d21ac55 A |
124 | * Place a thread in an affinity set identified by a tag. |
125 | * Called with thread referenced but not locked. | |
126 | */ | |
127 | kern_return_t | |
128 | thread_affinity_set(thread_t thread, uint32_t tag) | |
129 | { | |
0a7de745 A |
130 | affinity_set_t aset; |
131 | affinity_set_t empty_aset = NULL; | |
132 | affinity_space_t aspc; | |
133 | affinity_space_t new_aspc = NULL; | |
2d21ac55 A |
134 | |
135 | DBG("thread_affinity_set(%p,%u)\n", thread, tag); | |
136 | ||
137 | task_lock(thread->task); | |
138 | aspc = thread->task->affinity_space; | |
139 | if (aspc == NULL) { | |
140 | task_unlock(thread->task); | |
141 | new_aspc = affinity_space_alloc(); | |
0a7de745 | 142 | if (new_aspc == NULL) { |
2d21ac55 | 143 | return KERN_RESOURCE_SHORTAGE; |
0a7de745 | 144 | } |
2d21ac55 A |
145 | task_lock(thread->task); |
146 | if (thread->task->affinity_space == NULL) { | |
147 | thread->task->affinity_space = new_aspc; | |
148 | new_aspc = NULL; | |
149 | } | |
150 | aspc = thread->task->affinity_space; | |
151 | } | |
152 | task_unlock(thread->task); | |
0a7de745 | 153 | if (new_aspc) { |
2d21ac55 | 154 | affinity_space_free(new_aspc); |
0a7de745 | 155 | } |
2d21ac55 A |
156 | |
157 | thread_mtx_lock(thread); | |
158 | if (!thread->active) { | |
159 | /* Beaten to lock and the thread is dead */ | |
160 | thread_mtx_unlock(thread); | |
161 | return KERN_TERMINATED; | |
162 | } | |
163 | ||
b0d623f7 | 164 | lck_mtx_lock(&aspc->aspc_lock); |
2d21ac55 A |
165 | aset = thread->affinity_set; |
166 | if (aset != NULL) { | |
167 | /* | |
168 | * Remove thread from current affinity set | |
169 | */ | |
170 | DBG("thread_affinity_set(%p,%u) removing from aset %p\n", | |
0a7de745 | 171 | thread, tag, aset); |
2d21ac55 A |
172 | empty_aset = affinity_set_remove(aset, thread); |
173 | } | |
174 | ||
175 | if (tag != THREAD_AFFINITY_TAG_NULL) { | |
176 | aset = affinity_set_find(aspc, tag); | |
177 | if (aset != NULL) { | |
178 | /* | |
179 | * Add thread to existing affinity set | |
180 | */ | |
181 | DBG("thread_affinity_set(%p,%u) found aset %p\n", | |
0a7de745 | 182 | thread, tag, aset); |
2d21ac55 A |
183 | } else { |
184 | /* | |
185 | * Use the new affinity set, add this thread | |
186 | * and place it in a suitable processor set. | |
187 | */ | |
188 | if (empty_aset != NULL) { | |
189 | aset = empty_aset; | |
190 | empty_aset = NULL; | |
191 | } else { | |
192 | aset = affinity_set_alloc(); | |
193 | if (aset == NULL) { | |
b0d623f7 | 194 | lck_mtx_unlock(&aspc->aspc_lock); |
2d21ac55 A |
195 | thread_mtx_unlock(thread); |
196 | return KERN_RESOURCE_SHORTAGE; | |
197 | } | |
198 | } | |
199 | DBG("thread_affinity_set(%p,%u) (re-)using aset %p\n", | |
0a7de745 | 200 | thread, tag, aset); |
2d21ac55 A |
201 | aset->aset_tag = tag; |
202 | affinity_set_place(aspc, aset); | |
203 | } | |
204 | affinity_set_add(aset, thread); | |
205 | } | |
206 | ||
b0d623f7 | 207 | lck_mtx_unlock(&aspc->aspc_lock); |
2d21ac55 A |
208 | thread_mtx_unlock(thread); |
209 | ||
210 | /* | |
211 | * If we wound up not using an empty aset we created, | |
212 | * free it here. | |
213 | */ | |
0a7de745 | 214 | if (empty_aset != NULL) { |
2d21ac55 | 215 | affinity_set_free(empty_aset); |
0a7de745 | 216 | } |
2d21ac55 | 217 | |
0a7de745 A |
218 | if (thread == current_thread()) { |
219 | thread_block(THREAD_CONTINUE_NULL); | |
220 | } | |
2d21ac55 A |
221 | |
222 | return KERN_SUCCESS; | |
223 | } | |
224 | ||
225 | /* | |
226 | * task_affinity_create() | |
227 | * Called from task create. | |
228 | */ | |
229 | void | |
230 | task_affinity_create(task_t parent_task, task_t child_task) | |
231 | { | |
0a7de745 | 232 | affinity_space_t aspc = parent_task->affinity_space; |
2d21ac55 A |
233 | |
234 | DBG("task_affinity_create(%p,%p)\n", parent_task, child_task); | |
235 | ||
236 | assert(aspc); | |
237 | ||
238 | /* | |
239 | * Bump the task reference count on the shared namespace and | |
240 | * give it to the child. | |
241 | */ | |
b0d623f7 | 242 | lck_mtx_lock(&aspc->aspc_lock); |
2d21ac55 A |
243 | aspc->aspc_task_count++; |
244 | child_task->affinity_space = aspc; | |
b0d623f7 | 245 | lck_mtx_unlock(&aspc->aspc_lock); |
2d21ac55 A |
246 | } |
247 | ||
248 | /* | |
249 | * task_affinity_deallocate() | |
250 | * Called from task_deallocate() when there's a namespace to dereference. | |
251 | */ | |
252 | void | |
0a7de745 | 253 | task_affinity_deallocate(task_t task) |
2d21ac55 | 254 | { |
0a7de745 | 255 | affinity_space_t aspc = task->affinity_space; |
2d21ac55 A |
256 | |
257 | DBG("task_affinity_deallocate(%p) aspc %p task_count %d\n", | |
0a7de745 | 258 | task, aspc, aspc->aspc_task_count); |
2d21ac55 | 259 | |
b0d623f7 | 260 | lck_mtx_lock(&aspc->aspc_lock); |
2d21ac55 A |
261 | if (--(aspc->aspc_task_count) == 0) { |
262 | assert(queue_empty(&aspc->aspc_affinities)); | |
b0d623f7 | 263 | lck_mtx_unlock(&aspc->aspc_lock); |
2d21ac55 A |
264 | affinity_space_free(aspc); |
265 | } else { | |
b0d623f7 | 266 | lck_mtx_unlock(&aspc->aspc_lock); |
2d21ac55 A |
267 | } |
268 | } | |
269 | ||
270 | /* | |
271 | * task_affinity_info() | |
272 | * Return affinity tag info (number, min, max) for the task. | |
b0d623f7 A |
273 | * |
274 | * Conditions: task is locked. | |
2d21ac55 A |
275 | */ |
276 | kern_return_t | |
277 | task_affinity_info( | |
0a7de745 A |
278 | task_t task, |
279 | task_info_t task_info_out, | |
280 | mach_msg_type_number_t *task_info_count) | |
2d21ac55 | 281 | { |
0a7de745 A |
282 | affinity_set_t aset; |
283 | affinity_space_t aspc; | |
284 | task_affinity_tag_info_t info; | |
2d21ac55 A |
285 | |
286 | *task_info_count = TASK_AFFINITY_TAG_INFO_COUNT; | |
287 | info = (task_affinity_tag_info_t) task_info_out; | |
288 | info->set_count = 0; | |
289 | info->task_count = 0; | |
290 | info->min = THREAD_AFFINITY_TAG_NULL; | |
291 | info->max = THREAD_AFFINITY_TAG_NULL; | |
292 | ||
2d21ac55 A |
293 | aspc = task->affinity_space; |
294 | if (aspc) { | |
b0d623f7 | 295 | lck_mtx_lock(&aspc->aspc_lock); |
2d21ac55 | 296 | queue_iterate(&aspc->aspc_affinities, |
0a7de745 | 297 | aset, affinity_set_t, aset_affinities) { |
2d21ac55 A |
298 | info->set_count++; |
299 | if (info->min == THREAD_AFFINITY_TAG_NULL || | |
0a7de745 | 300 | aset->aset_tag < (uint32_t) info->min) { |
2d21ac55 | 301 | info->min = aset->aset_tag; |
0a7de745 | 302 | } |
2d21ac55 | 303 | if (info->max == THREAD_AFFINITY_TAG_NULL || |
0a7de745 | 304 | aset->aset_tag > (uint32_t) info->max) { |
2d21ac55 | 305 | info->max = aset->aset_tag; |
0a7de745 | 306 | } |
2d21ac55 A |
307 | } |
308 | info->task_count = aspc->aspc_task_count; | |
b0d623f7 | 309 | lck_mtx_unlock(&aspc->aspc_lock); |
2d21ac55 | 310 | } |
2d21ac55 A |
311 | return KERN_SUCCESS; |
312 | } | |
313 | ||
314 | /* | |
315 | * Called from thread_dup() during fork() with child's mutex held. | |
316 | * Set the child into the parent's affinity set. | |
317 | * Note the affinity space is shared. | |
318 | */ | |
319 | void | |
320 | thread_affinity_dup(thread_t parent, thread_t child) | |
321 | { | |
0a7de745 A |
322 | affinity_set_t aset; |
323 | affinity_space_t aspc; | |
2d21ac55 A |
324 | |
325 | thread_mtx_lock(parent); | |
326 | aset = parent->affinity_set; | |
327 | DBG("thread_affinity_dup(%p,%p) aset %p\n", parent, child, aset); | |
328 | if (aset == NULL) { | |
329 | thread_mtx_unlock(parent); | |
330 | return; | |
331 | } | |
332 | ||
333 | aspc = aset->aset_space; | |
334 | assert(aspc == parent->task->affinity_space); | |
335 | assert(aspc == child->task->affinity_space); | |
336 | ||
b0d623f7 | 337 | lck_mtx_lock(&aspc->aspc_lock); |
2d21ac55 | 338 | affinity_set_add(aset, child); |
b0d623f7 | 339 | lck_mtx_unlock(&aspc->aspc_lock); |
2d21ac55 A |
340 | |
341 | thread_mtx_unlock(parent); | |
342 | } | |
343 | ||
344 | /* | |
0a7de745 | 345 | * thread_affinity_terminate() |
2d21ac55 A |
346 | * Remove thread from any affinity set. |
347 | * Called with the thread mutex locked. | |
348 | */ | |
349 | void | |
350 | thread_affinity_terminate(thread_t thread) | |
351 | { | |
0a7de745 A |
352 | affinity_set_t aset = thread->affinity_set; |
353 | affinity_space_t aspc; | |
2d21ac55 A |
354 | |
355 | DBG("thread_affinity_terminate(%p)\n", thread); | |
356 | ||
357 | aspc = aset->aset_space; | |
b0d623f7 | 358 | lck_mtx_lock(&aspc->aspc_lock); |
2d21ac55 A |
359 | if (affinity_set_remove(aset, thread)) { |
360 | affinity_set_free(aset); | |
361 | } | |
b0d623f7 A |
362 | lck_mtx_unlock(&aspc->aspc_lock); |
363 | } | |
364 | ||
365 | /* | |
366 | * thread_affinity_exec() | |
367 | * Called from execve() to cancel any current affinity - a new image implies | |
368 | * the calling thread terminates any expressed or inherited affinity. | |
369 | */ | |
370 | void | |
371 | thread_affinity_exec(thread_t thread) | |
372 | { | |
0a7de745 | 373 | if (thread->affinity_set != AFFINITY_SET_NULL) { |
b0d623f7 | 374 | thread_affinity_terminate(thread); |
0a7de745 | 375 | } |
2d21ac55 A |
376 | } |
377 | ||
378 | /* | |
379 | * Create an empty affinity namespace data structure. | |
380 | */ | |
381 | static affinity_space_t | |
0a7de745 | 382 | affinity_space_alloc(void) |
2d21ac55 | 383 | { |
0a7de745 | 384 | affinity_space_t aspc; |
2d21ac55 A |
385 | |
386 | aspc = (affinity_space_t) kalloc(sizeof(struct affinity_space)); | |
0a7de745 | 387 | if (aspc == NULL) { |
2d21ac55 | 388 | return NULL; |
0a7de745 | 389 | } |
2d21ac55 | 390 | |
b0d623f7 | 391 | lck_mtx_init(&aspc->aspc_lock, &task_lck_grp, &task_lck_attr); |
2d21ac55 A |
392 | queue_init(&aspc->aspc_affinities); |
393 | aspc->aspc_task_count = 1; | |
394 | ||
395 | DBG("affinity_space_create() returns %p\n", aspc); | |
396 | return aspc; | |
397 | } | |
398 | ||
399 | /* | |
400 | * Destroy the given empty affinity namespace data structure. | |
401 | */ | |
402 | static void | |
403 | affinity_space_free(affinity_space_t aspc) | |
404 | { | |
405 | assert(queue_empty(&aspc->aspc_affinities)); | |
406 | ||
39236c6e | 407 | lck_mtx_destroy(&aspc->aspc_lock, &task_lck_grp); |
2d21ac55 A |
408 | DBG("affinity_space_free(%p)\n", aspc); |
409 | kfree(aspc, sizeof(struct affinity_space)); | |
410 | } | |
411 | ||
412 | ||
413 | /* | |
414 | * Create an empty affinity set data structure | |
415 | * entering it into a list anchored by the owning task. | |
416 | */ | |
417 | static affinity_set_t | |
0a7de745 | 418 | affinity_set_alloc(void) |
2d21ac55 | 419 | { |
0a7de745 | 420 | affinity_set_t aset; |
2d21ac55 A |
421 | |
422 | aset = (affinity_set_t) kalloc(sizeof(struct affinity_set)); | |
0a7de745 | 423 | if (aset == NULL) { |
2d21ac55 | 424 | return NULL; |
0a7de745 | 425 | } |
2d21ac55 A |
426 | |
427 | aset->aset_thread_count = 0; | |
428 | queue_init(&aset->aset_affinities); | |
429 | queue_init(&aset->aset_threads); | |
430 | aset->aset_num = 0; | |
431 | aset->aset_pset = PROCESSOR_SET_NULL; | |
432 | aset->aset_space = NULL; | |
433 | ||
434 | DBG("affinity_set_create() returns %p\n", aset); | |
435 | return aset; | |
436 | } | |
437 | ||
438 | /* | |
439 | * Destroy the given empty affinity set data structure | |
440 | * after removing it from the parent task. | |
441 | */ | |
442 | static void | |
443 | affinity_set_free(affinity_set_t aset) | |
444 | { | |
445 | assert(queue_empty(&aset->aset_threads)); | |
446 | ||
447 | DBG("affinity_set_free(%p)\n", aset); | |
448 | kfree(aset, sizeof(struct affinity_set)); | |
449 | } | |
450 | ||
451 | /* | |
452 | * Add a thread to an affinity set. | |
453 | * The caller must have the thread mutex and space locked. | |
454 | */ | |
455 | static void | |
456 | affinity_set_add(affinity_set_t aset, thread_t thread) | |
457 | { | |
0a7de745 | 458 | spl_t s; |
2d21ac55 A |
459 | |
460 | DBG("affinity_set_add(%p,%p)\n", aset, thread); | |
461 | queue_enter(&aset->aset_threads, | |
0a7de745 | 462 | thread, thread_t, affinity_threads); |
2d21ac55 A |
463 | aset->aset_thread_count++; |
464 | s = splsched(); | |
465 | thread_lock(thread); | |
466 | thread->affinity_set = affinity_sets_enabled ? aset : NULL; | |
467 | thread_unlock(thread); | |
468 | splx(s); | |
469 | } | |
470 | ||
471 | /* | |
472 | * Remove a thread from an affinity set returning the set if now empty. | |
473 | * The caller must have the thread mutex and space locked. | |
474 | */ | |
475 | static affinity_set_t | |
476 | affinity_set_remove(affinity_set_t aset, thread_t thread) | |
477 | { | |
0a7de745 | 478 | spl_t s; |
2d21ac55 A |
479 | |
480 | s = splsched(); | |
481 | thread_lock(thread); | |
482 | thread->affinity_set = NULL; | |
483 | thread_unlock(thread); | |
484 | splx(s); | |
485 | ||
486 | aset->aset_thread_count--; | |
487 | queue_remove(&aset->aset_threads, | |
0a7de745 | 488 | thread, thread_t, affinity_threads); |
2d21ac55 A |
489 | if (queue_empty(&aset->aset_threads)) { |
490 | queue_remove(&aset->aset_space->aspc_affinities, | |
0a7de745 | 491 | aset, affinity_set_t, aset_affinities); |
2d21ac55 A |
492 | assert(aset->aset_thread_count == 0); |
493 | aset->aset_tag = THREAD_AFFINITY_TAG_NULL; | |
494 | aset->aset_num = 0; | |
495 | aset->aset_pset = PROCESSOR_SET_NULL; | |
496 | aset->aset_space = NULL; | |
497 | DBG("affinity_set_remove(%p,%p) set now empty\n", aset, thread); | |
498 | return aset; | |
499 | } else { | |
500 | DBG("affinity_set_remove(%p,%p)\n", aset, thread); | |
501 | return NULL; | |
502 | } | |
503 | } | |
504 | ||
505 | /* | |
506 | * Find an affinity set in the parent task with the given affinity tag. | |
507 | * The caller must have the space locked. | |
508 | */ | |
509 | static affinity_set_t | |
510 | affinity_set_find(affinity_space_t space, uint32_t tag) | |
511 | { | |
0a7de745 | 512 | affinity_set_t aset; |
2d21ac55 A |
513 | |
514 | queue_iterate(&space->aspc_affinities, | |
0a7de745 | 515 | aset, affinity_set_t, aset_affinities) { |
2d21ac55 A |
516 | if (aset->aset_tag == tag) { |
517 | DBG("affinity_set_find(%p,%u) finds %p\n", | |
0a7de745 | 518 | space, tag, aset); |
2d21ac55 A |
519 | return aset; |
520 | } | |
521 | } | |
522 | DBG("affinity_set_find(%p,%u) not found\n", space, tag); | |
523 | return NULL; | |
524 | } | |
525 | ||
526 | /* | |
527 | * affinity_set_place() assigns an affinity set to a suitable processor_set. | |
528 | * The selection criteria is: | |
529 | * - the set currently occupied by the least number of affinities | |
530 | * belonging to the owning the task. | |
531 | * The caller must have the space locked. | |
532 | */ | |
533 | static void | |
534 | affinity_set_place(affinity_space_t aspc, affinity_set_t new_aset) | |
535 | { | |
0a7de745 A |
536 | unsigned int num_cpu_asets = ml_get_max_affinity_sets(); |
537 | unsigned int set_occupancy[num_cpu_asets]; | |
538 | unsigned int i; | |
539 | unsigned int i_least_occupied; | |
540 | affinity_set_t aset; | |
2d21ac55 | 541 | |
0a7de745 | 542 | for (i = 0; i < num_cpu_asets; i++) { |
2d21ac55 | 543 | set_occupancy[i] = 0; |
0a7de745 | 544 | } |
2d21ac55 A |
545 | |
546 | /* | |
547 | * Scan the affinity sets calculating the number of sets | |
548 | * occupy the available physical affinities. | |
549 | */ | |
550 | queue_iterate(&aspc->aspc_affinities, | |
0a7de745 A |
551 | aset, affinity_set_t, aset_affinities) { |
552 | if (aset->aset_num < num_cpu_asets) { | |
316670eb | 553 | set_occupancy[aset->aset_num]++; |
0a7de745 | 554 | } else { |
316670eb | 555 | panic("aset_num = %d in %s\n", aset->aset_num, __FUNCTION__); |
0a7de745 | 556 | } |
2d21ac55 A |
557 | } |
558 | ||
559 | /* | |
560 | * Find the least occupied set (or the first empty set). | |
561 | * To distribute placements somewhat, start searching from | |
562 | * a cpu affinity chosen randomly per namespace: | |
563 | * [(unsigned int)aspc % 127] % num_cpu_asets | |
564 | * unless this mapping policy is overridden. | |
565 | */ | |
0a7de745 | 566 | if (affinity_sets_mapping == 0) { |
2d21ac55 | 567 | i_least_occupied = 0; |
0a7de745 | 568 | } else { |
b0d623f7 | 569 | i_least_occupied = (unsigned int)(((uintptr_t)aspc % 127) % num_cpu_asets); |
0a7de745 | 570 | } |
2d21ac55 | 571 | for (i = 0; i < num_cpu_asets; i++) { |
0a7de745 | 572 | unsigned int j = (i_least_occupied + i) % num_cpu_asets; |
2d21ac55 A |
573 | if (set_occupancy[j] == 0) { |
574 | i_least_occupied = j; | |
575 | break; | |
576 | } | |
0a7de745 | 577 | if (set_occupancy[j] < set_occupancy[i_least_occupied]) { |
2d21ac55 | 578 | i_least_occupied = j; |
0a7de745 | 579 | } |
2d21ac55 A |
580 | } |
581 | new_aset->aset_num = i_least_occupied; | |
582 | new_aset->aset_pset = ml_affinity_to_pset(i_least_occupied); | |
583 | ||
584 | /* Add the new affinity set to the group */ | |
585 | new_aset->aset_space = aspc; | |
586 | queue_enter(&aspc->aspc_affinities, | |
0a7de745 | 587 | new_aset, affinity_set_t, aset_affinities); |
2d21ac55 A |
588 | |
589 | DBG("affinity_set_place(%p,%p) selected affinity %u pset %p\n", | |
590 | aspc, new_aset, new_aset->aset_num, new_aset->aset_pset); | |
591 | } |