#include <kern/thread.h>
#include <kern/zalloc.h>
#include <kern/kalloc.h>
+#include <kern/ledger.h>
#include <vm/vm_map.h>
#include <vm/vm_kern.h>
task_t task;
zinfo_usage_t zinfo;
- thread->tkm_private.alloc += kernel_stack_size;
+ ledger_credit(thread->t_ledger, task_ledgers.tkm_private, kernel_stack_size);
+
if (stack_fake_zone_index != -1 &&
(task = thread->task) != NULL && (zinfo = task->tkm_zinfo) != NULL)
OSAddAtomic64(kernel_stack_size,
task_t task;
zinfo_usage_t zinfo;
- thread->tkm_private.free += kernel_stack_size;
+ ledger_debit(thread->t_ledger, task_ledgers.tkm_private, kernel_stack_size);
+
if (stack_fake_zone_index != -1 &&
(task = thread->task) != NULL && (zinfo = task->tkm_zinfo) != NULL)
OSAddAtomic64(kernel_stack_size,
static inline void
STACK_ZINFO_HANDOFF(thread_t from, thread_t to)
{
- from->tkm_private.free += kernel_stack_size;
- to->tkm_private.alloc += kernel_stack_size;
+ ledger_debit(from->t_ledger, task_ledgers.tkm_private, kernel_stack_size);
+ ledger_credit(to->t_ledger, task_ledgers.tkm_private, kernel_stack_size);
+
if (stack_fake_zone_index != -1) {
task_t task;
zinfo_usage_t zinfo;
if (kernel_memory_allocate(kernel_map, &stack,
kernel_stack_size + (2*PAGE_SIZE),
stack_addr_mask,
- KMA_KOBJECT | guard_flags)
+ KMA_KSTACK | KMA_KOBJECT | guard_flags)
!= KERN_SUCCESS)
panic("stack_alloc: kernel_memory_allocate");
* back in stack_alloc().
*/
- stack = (vm_offset_t)vm_map_trunc_page(stack);
+ stack = (vm_offset_t)vm_map_trunc_page(
+ stack,
+ VM_MAP_PAGE_MASK(kernel_map));
stack -= PAGE_SIZE;
if (vm_map_remove(
kernel_map,
/* OK, have memory and list is locked */
thread_list = (thread_t *) addr;
- for (i = 0, thread = (thread_t) queue_first(&threads);
+ for (i = 0, thread = (thread_t)(void *) queue_first(&threads);
!queue_end(&threads, (queue_entry_t) thread);
- thread = (thread_t) queue_next(&thread->threads)) {
+ thread = (thread_t)(void *) queue_next(&thread->threads)) {
thread_reference_internal(thread);
thread_list[i++] = thread;
}