]> git.saurik.com Git - apple/xnu.git/blobdiff - osfmk/kern/processor.c
xnu-2050.22.13.tar.gz
[apple/xnu.git] / osfmk / kern / processor.c
index a37913b1502e401447c35ffb1ac9162535bf04af..23a5496119f7ea63cce73147b87f3c0cd5b8bdc6 100644 (file)
@@ -1,23 +1,29 @@
 /*
- * Copyright (c) 2000 Apple Computer, Inc. All rights reserved.
+ * Copyright (c) 2000-2009 Apple Inc. All rights reserved.
  *
- * @APPLE_LICENSE_HEADER_START@
+ * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
  * 
- * The contents of this file constitute Original Code as defined in and
- * are subject to the Apple Public Source License Version 1.1 (the
- * "License").  You may not use this file except in compliance with the
- * License.  Please obtain a copy of the License at
- * http://www.apple.com/publicsource and read it before using this file.
+ * This file contains Original Code and/or Modifications of Original Code
+ * as defined in and that are subject to the Apple Public Source License
+ * Version 2.0 (the 'License'). You may not use this file except in
+ * compliance with the License. The rights granted to you under the License
+ * may not be used to create, or enable the creation or redistribution of,
+ * unlawful or unlicensed copies of an Apple operating system, or to
+ * circumvent, violate, or enable the circumvention or violation of, any
+ * terms of an Apple operating system software license agreement.
  * 
- * This Original Code and all software distributed under the License are
- * distributed on an "AS IS" basis, WITHOUT WARRANTY OF ANY KIND, EITHER
+ * Please obtain a copy of the License at
+ * http://www.opensource.apple.com/apsl/ and read it before using this file.
+ * 
+ * The Original Code and all software distributed under the License are
+ * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
  * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
  * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
- * FITNESS FOR A PARTICULAR PURPOSE OR NON-INFRINGEMENT.  Please see the
- * License for the specific language governing rights and limitations
- * under the License.
+ * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
+ * Please see the License for the specific language governing rights and
+ * limitations under the License.
  * 
- * @APPLE_LICENSE_HEADER_END@
+ * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
  */
 /*
  * @OSF_COPYRIGHT@
  *     processor.c: processor and processor_set manipulation routines.
  */
 
-#include <cpus.h>
-
 #include <mach/boolean.h>
 #include <mach/policy.h>
+#include <mach/processor.h>
 #include <mach/processor_info.h>
 #include <mach/vm_param.h>
 #include <kern/cpu_number.h>
  * Exported interface
  */
 #include <mach/mach_host_server.h>
+#include <mach/processor_set_server.h>
 
-/*
- *     Exported variables.
- */
-struct processor_set default_pset;
-struct processor processor_array[NCPUS];
-
-processor_t    master_processor;
-processor_t    processor_ptr[NCPUS];
-
-/* Forwards */
-void   pset_init(
-               processor_set_t pset);
+struct processor_set   pset0;
+struct pset_node               pset_node0;
+decl_simple_lock_data(static,pset_node_lock)
 
-void   processor_init(
-               register processor_t    pr,
-               int                     slot_num);
+queue_head_t                   tasks;
+queue_head_t                   terminated_tasks;       /* To be used ONLY for stackshot. */
+int                                            tasks_count;
+queue_head_t                   threads;
+int                                            threads_count;
+decl_lck_mtx_data(,tasks_threads_lock)
 
-void   pset_quanta_set(
-               processor_set_t         pset);
+processor_t                            processor_list;
+unsigned int                   processor_count;
+static processor_t             processor_list_tail;
+decl_simple_lock_data(,processor_list_lock)
 
-kern_return_t  processor_set_base(
-               processor_set_t         pset,
-               policy_t                policy,
-               policy_base_t           base,
-               boolean_t               change);
+uint32_t                               processor_avail_count;
 
-kern_return_t  processor_set_limit(
-               processor_set_t         pset,
-               policy_t                policy,
-               policy_limit_t          limit,
-               boolean_t               change);
+processor_t            master_processor;
+int                    master_cpu = 0;
+boolean_t              sched_stats_active = FALSE;
 
+/* Forwards */
 kern_return_t  processor_set_things(
                processor_set_t         pset,
                mach_port_t             **thing_list,
                mach_msg_type_number_t  *count,
                int                     type);
 
-
-/*
- *     Bootstrap the processor/pset system so the scheduler can run.
- */
 void
-pset_sys_bootstrap(void)
+processor_bootstrap(void)
 {
-       register int    i;
-
-       pset_init(&default_pset);
-       for (i = 0; i < NCPUS; i++) {
-               /*
-                *      Initialize processor data structures.
-                *      Note that cpu_to_processor(i) is processor_ptr[i].
-                */
-               processor_ptr[i] = &processor_array[i];
-               processor_init(processor_ptr[i], i);
-       }
-       master_processor = cpu_to_processor(master_cpu);
-       default_pset.active = TRUE;
-}
+       pset_init(&pset0, &pset_node0);
+       pset_node0.psets = &pset0;
 
-/*
- *     Initialize the given processor_set structure.
- */
+       simple_lock_init(&pset_node_lock, 0);
 
-void pset_init(
-       register processor_set_t        pset)
-{
-       int     i;
+       queue_init(&tasks);
+       queue_init(&terminated_tasks);
+       queue_init(&threads);
 
-       /* setup run-queues */
-       simple_lock_init(&pset->runq.lock, ETAP_THREAD_PSET_RUNQ);
-       pset->runq.count = 0;
-       for (i = 0; i < NRQBM; i++) {
-           pset->runq.bitmap[i] = 0;
-       }
-       setbit(MAXPRI - IDLEPRI, pset->runq.bitmap); 
-       pset->runq.highq = IDLEPRI;
-       for (i = 0; i < NRQS; i++) {
-           queue_init(&(pset->runq.queues[i]));
-       }
+       simple_lock_init(&processor_list_lock, 0);
 
-       queue_init(&pset->idle_queue);
-       pset->idle_count = 0;
-       simple_lock_init(&pset->idle_lock, ETAP_THREAD_PSET_IDLE);
-       pset->mach_factor = pset->load_average = 0;
-       pset->sched_load = 0;
-       queue_init(&pset->processors);
-       pset->processor_count = 0;
-       simple_lock_init(&pset->processors_lock, ETAP_THREAD_PSET);
-       queue_init(&pset->tasks);
-       pset->task_count = 0;
-       queue_init(&pset->threads);
-       pset->thread_count = 0;
-       pset->ref_count = 1;
-       pset->active = FALSE;
-       mutex_init(&pset->lock, ETAP_THREAD_PSET);
-       pset->pset_self = IP_NULL;
-       pset->pset_name_self = IP_NULL;
-       pset->set_quanta = 1;
+       master_processor = cpu_to_processor(master_cpu);
 
-       for (i = 0; i <= NCPUS; i++)
-           pset->machine_quanta[i] = 1;
+       processor_init(master_processor, master_cpu, &pset0);
 }
 
 /*
- *     Initialize the given processor structure for the processor in
- *     the slot specified by slot_num.
+ *     Initialize the given processor for the cpu
+ *     indicated by cpu_id, and assign to the
+ *     specified processor set.
  */
 void
 processor_init(
-       register processor_t    pr,
-       int                     slot_num)
+       processor_t                     processor,
+       int                                     cpu_id,
+       processor_set_t         pset)
 {
-       int     i;
-
-       /* setup run-queues */
-       simple_lock_init(&pr->runq.lock, ETAP_THREAD_PROC_RUNQ);
-       pr->runq.count = 0;
-       for (i = 0; i < NRQBM; i++) {
-           pr->runq.bitmap[i] = 0;
+       if (processor != master_processor) {
+               /* Scheduler state deferred until sched_init() */
+               SCHED(processor_init)(processor);
        }
-       setbit(MAXPRI - IDLEPRI, pr->runq.bitmap); 
-       pr->runq.highq = IDLEPRI;
-       for (i = 0; i < NRQS; i++) {
-           queue_init(&(pr->runq.queues[i]));
+
+       processor->state = PROCESSOR_OFF_LINE;
+       processor->active_thread = processor->next_thread = processor->idle_thread = THREAD_NULL;
+       processor->processor_set = pset;
+       processor->current_pri = MINPRI;
+       processor->current_thmode = TH_MODE_NONE;
+       processor->cpu_id = cpu_id;
+       timer_call_setup(&processor->quantum_timer, thread_quantum_expire, processor);
+       processor->deadline = UINT64_MAX;
+       processor->timeslice = 0;
+       processor->processor_meta = PROCESSOR_META_NULL;
+       processor->processor_self = IP_NULL;
+       processor_data_init(processor);
+       processor->processor_list = NULL;
+
+       pset_lock(pset);
+       if (pset->cpu_set_count++ == 0)
+               pset->cpu_set_low = pset->cpu_set_hi = cpu_id;
+       else {
+               pset->cpu_set_low = (cpu_id < pset->cpu_set_low)? cpu_id: pset->cpu_set_low;
+               pset->cpu_set_hi = (cpu_id > pset->cpu_set_hi)? cpu_id: pset->cpu_set_hi;
        }
+       pset_unlock(pset);
 
-       queue_init(&pr->processor_queue);
-       pr->state = PROCESSOR_OFF_LINE;
-       pr->next_thread = THREAD_NULL;
-       pr->idle_thread = THREAD_NULL;
-       timer_call_setup(&pr->quantum_timer, thread_quantum_expire, pr);
-       pr->slice_quanta = 0;
-       pr->processor_set = PROCESSOR_SET_NULL;
-       pr->processor_set_next = PROCESSOR_SET_NULL;
-       queue_init(&pr->processors);
-       simple_lock_init(&pr->lock, ETAP_THREAD_PROC);
-       pr->processor_self = IP_NULL;
-       pr->slot_num = slot_num;
+       simple_lock(&processor_list_lock);
+       if (processor_list == NULL)
+               processor_list = processor;
+       else
+               processor_list_tail->processor_list = processor;
+       processor_list_tail = processor;
+       processor_count++;
+       simple_unlock(&processor_list_lock);
 }
 
-/*
- *     pset_remove_processor() removes a processor from a processor_set.
- *     It can only be called on the current processor.  Caller must
- *     hold lock on current processor and processor set.
- */
 void
-pset_remove_processor(
-       processor_set_t pset,
-       processor_t     processor)
+processor_meta_init(
+       processor_t             processor,
+       processor_t             primary)
 {
-       if (pset != processor->processor_set)
-               panic("pset_remove_processor: wrong pset");
+       processor_meta_t        pmeta = primary->processor_meta;
 
-       queue_remove(&pset->processors, processor, processor_t, processors);
-       processor->processor_set = PROCESSOR_SET_NULL;
-       pset->processor_count--;
-       pset_quanta_set(pset);
-}
+       if (pmeta == PROCESSOR_META_NULL) {
+               pmeta = kalloc(sizeof (*pmeta));
 
-/*
- *     pset_add_processor() adds a  processor to a processor_set.
- *     It can only be called on the current processor.  Caller must
- *     hold lock on curent processor and on pset.  No reference counting on
- *     processors.  Processor reference to pset is implicit.
- */
-void
-pset_add_processor(
-       processor_set_t pset,
-       processor_t     processor)
-{
-       queue_enter(&pset->processors, processor, processor_t, processors);
-       processor->processor_set = pset;
-       pset->processor_count++;
-       pset_quanta_set(pset);
-}
+               queue_init(&pmeta->idle_queue);
 
-/*
- *     pset_remove_task() removes a task from a processor_set.
- *     Caller must hold locks on pset and task.  Pset reference count
- *     is not decremented; caller must explicitly pset_deallocate.
- */
-void
-pset_remove_task(
-       processor_set_t pset,
-       task_t          task)
-{
-       if (pset != task->processor_set)
-               return;
+               pmeta->primary = primary;
+       }
 
-       queue_remove(&pset->tasks, task, task_t, pset_tasks);
-       task->processor_set = PROCESSOR_SET_NULL;
-       pset->task_count--;
+       processor->processor_meta = pmeta;
 }
 
-/*
- *     pset_add_task() adds a  task to a processor_set.
- *     Caller must hold locks on pset and task.  Pset references to
- *     tasks are implicit.
- */
-void
-pset_add_task(
-       processor_set_t pset,
-       task_t          task)
+processor_set_t
+processor_pset(
+       processor_t     processor)
 {
-       queue_enter(&pset->tasks, task, task_t, pset_tasks);
-       task->processor_set = pset;
-       pset->task_count++;
-       pset->ref_count++;
+       return (processor->processor_set);
 }
 
-/*
- *     pset_remove_thread() removes a thread from a processor_set.
- *     Caller must hold locks on pset and thread.  Pset reference count
- *     is not decremented; caller must explicitly pset_deallocate.
- */
-void
-pset_remove_thread(
-       processor_set_t pset,
-       thread_t        thread)
+pset_node_t
+pset_node_root(void)
 {
-       queue_remove(&pset->threads, thread, thread_t, pset_threads);
-       thread->processor_set = PROCESSOR_SET_NULL;
-       pset->thread_count--;
+       return &pset_node0;
 }
 
-/*
- *     pset_add_thread() adds a  thread to a processor_set.
- *     Caller must hold locks on pset and thread.  Pset references to
- *     threads are implicit.
- */
-void
-pset_add_thread(
-       processor_set_t pset,
-       thread_t        thread)
+processor_set_t
+pset_create(
+       pset_node_t                     node)
 {
-       queue_enter(&pset->threads, thread, thread_t, pset_threads);
-       thread->processor_set = pset;
-       pset->thread_count++;
-       pset->ref_count++;
-}
+       processor_set_t         *prev, pset = kalloc(sizeof (*pset));
 
-/*
- *     thread_change_psets() changes the pset of a thread.  Caller must
- *     hold locks on both psets and thread.  The old pset must be
- *     explicitly pset_deallocat()'ed by caller.
- */
-void
-thread_change_psets(
-       thread_t        thread,
-       processor_set_t old_pset,
-       processor_set_t new_pset)
-{
-       queue_remove(&old_pset->threads, thread, thread_t, pset_threads);
-       old_pset->thread_count--;
-       queue_enter(&new_pset->threads, thread, thread_t, pset_threads);
-       thread->processor_set = new_pset;
-       new_pset->thread_count++;
-       new_pset->ref_count++;
-}      
+       if (pset != PROCESSOR_SET_NULL) {
+               pset_init(pset, node);
 
-/*
- *     pset_deallocate:
- *
- *     Remove one reference to the processor set.  Destroy processor_set
- *     if this was the last reference.
- */
-void
-pset_deallocate(
-       processor_set_t pset)
-{
-       if (pset == PROCESSOR_SET_NULL)
-               return;
+               simple_lock(&pset_node_lock);
 
-       pset_lock(pset);
-       if (--pset->ref_count > 0) {
-               pset_unlock(pset);
-               return;
+               prev = &node->psets;
+               while (*prev != PROCESSOR_SET_NULL)
+                       prev = &(*prev)->pset_list;
+
+               *prev = pset;
+
+               simple_unlock(&pset_node_lock);
        }
 
-       panic("pset_deallocate: default_pset destroyed");
+       return (pset);
 }
 
 /*
- *     pset_reference:
- *
- *     Add one reference to the processor set.
+ *     Initialize the given processor_set structure.
  */
 void
-pset_reference(
-       processor_set_t pset)
+pset_init(
+       processor_set_t         pset,
+       pset_node_t                     node)
 {
-       pset_lock(pset);
-       pset->ref_count++;
-       pset_unlock(pset);
-}
+       if (pset != &pset0) {
+               /* Scheduler state deferred until sched_init() */
+               SCHED(pset_init)(pset);
+       }
 
+       queue_init(&pset->active_queue);
+       queue_init(&pset->idle_queue);
+       pset->online_processor_count = 0;
+       pset_pri_init_hint(pset, PROCESSOR_NULL);
+       pset_count_init_hint(pset, PROCESSOR_NULL);
+       pset->cpu_set_low = pset->cpu_set_hi = 0;
+       pset->cpu_set_count = 0;
+       pset_lock_init(pset);
+       pset->pset_self = IP_NULL;
+       pset->pset_name_self = IP_NULL;
+       pset->pset_list = PROCESSOR_SET_NULL;
+       pset->node = node;
+}
 
 kern_return_t
 processor_info_count(
-       processor_flavor_t      flavor,
+       processor_flavor_t              flavor,
        mach_msg_type_number_t  *count)
 {
-       kern_return_t           kr;
-
        switch (flavor) {
+
        case PROCESSOR_BASIC_INFO:
                *count = PROCESSOR_BASIC_INFO_COUNT;
-               return KERN_SUCCESS;
+               break;
+
        case PROCESSOR_CPU_LOAD_INFO:
                *count = PROCESSOR_CPU_LOAD_INFO_COUNT;
-               return KERN_SUCCESS;
+               break;
+
        default:
-               kr = cpu_info_count(flavor, count);
-               return kr;
+               return (cpu_info_count(flavor, count));
        }
+
+       return (KERN_SUCCESS);
 }
 
 
 kern_return_t
 processor_info(
        register processor_t    processor,
-       processor_flavor_t      flavor,
-       host_t                  *host,
-       processor_info_t        info,
+       processor_flavor_t              flavor,
+       host_t                                  *host,
+       processor_info_t                info,
        mach_msg_type_number_t  *count)
 {
-       register int    i, slot_num, state;
-       register processor_basic_info_t         basic_info;
-       register processor_cpu_load_info_t      cpu_load_info;
-       kern_return_t   kr;
+       register int    cpu_id, state;
+       kern_return_t   result;
 
        if (processor == PROCESSOR_NULL)
-               return(KERN_INVALID_ARGUMENT);
+               return (KERN_INVALID_ARGUMENT);
 
-       slot_num = processor->slot_num;
+       cpu_id = processor->cpu_id;
 
        switch (flavor) {
 
        case PROCESSOR_BASIC_INFO:
-         {
-           if (*count < PROCESSOR_BASIC_INFO_COUNT)
-             return(KERN_FAILURE);
-
-           basic_info = (processor_basic_info_t) info;
-           basic_info->cpu_type = machine_slot[slot_num].cpu_type;
-           basic_info->cpu_subtype = machine_slot[slot_num].cpu_subtype;
-           state = processor->state;
-           if (state == PROCESSOR_OFF_LINE)
-             basic_info->running = FALSE;
-           else
-             basic_info->running = TRUE;
-           basic_info->slot_num = slot_num;
-           if (processor == master_processor) 
-             basic_info->is_master = TRUE;
-           else
-             basic_info->is_master = FALSE;
-
-           *count = PROCESSOR_BASIC_INFO_COUNT;
-           *host = &realhost;
-           return(KERN_SUCCESS);
-         }
+       {
+               register processor_basic_info_t         basic_info;
+
+               if (*count < PROCESSOR_BASIC_INFO_COUNT)
+                       return (KERN_FAILURE);
+
+               basic_info = (processor_basic_info_t) info;
+               basic_info->cpu_type = slot_type(cpu_id);
+               basic_info->cpu_subtype = slot_subtype(cpu_id);
+               state = processor->state;
+               if (state == PROCESSOR_OFF_LINE)
+                       basic_info->running = FALSE;
+               else
+                       basic_info->running = TRUE;
+               basic_info->slot_num = cpu_id;
+               if (processor == master_processor) 
+                       basic_info->is_master = TRUE;
+               else
+                       basic_info->is_master = FALSE;
+
+               *count = PROCESSOR_BASIC_INFO_COUNT;
+               *host = &realhost;
+
+           return (KERN_SUCCESS);
+       }
+
        case PROCESSOR_CPU_LOAD_INFO:
-         {
-           if (*count < PROCESSOR_CPU_LOAD_INFO_COUNT)
-             return(KERN_FAILURE);
+       {
+               processor_cpu_load_info_t       cpu_load_info;
+               timer_data_t    idle_temp;
+               timer_t         idle_state;
+
+               if (*count < PROCESSOR_CPU_LOAD_INFO_COUNT)
+                       return (KERN_FAILURE);
+
+               cpu_load_info = (processor_cpu_load_info_t) info;
+               if (precise_user_kernel_time) {
+                       cpu_load_info->cpu_ticks[CPU_STATE_USER] =
+                                                       (uint32_t)(timer_grab(&PROCESSOR_DATA(processor, user_state)) / hz_tick_interval);
+                       cpu_load_info->cpu_ticks[CPU_STATE_SYSTEM] =
+                                                       (uint32_t)(timer_grab(&PROCESSOR_DATA(processor, system_state)) / hz_tick_interval);
+               } else {
+                       uint64_t tval = timer_grab(&PROCESSOR_DATA(processor, user_state)) +
+                               timer_grab(&PROCESSOR_DATA(processor, system_state));
+
+                       cpu_load_info->cpu_ticks[CPU_STATE_USER] = (uint32_t)(tval / hz_tick_interval);
+                       cpu_load_info->cpu_ticks[CPU_STATE_SYSTEM] = 0;
+               }
+
+               idle_state = &PROCESSOR_DATA(processor, idle_state);
+               idle_temp = *idle_state;
+
+               if (PROCESSOR_DATA(processor, current_state) != idle_state ||
+                   timer_grab(&idle_temp) != timer_grab(idle_state)) {
+                       cpu_load_info->cpu_ticks[CPU_STATE_IDLE] =
+                                                       (uint32_t)(timer_grab(&PROCESSOR_DATA(processor, idle_state)) / hz_tick_interval);
+               } else {
+                       timer_advance(&idle_temp, mach_absolute_time() - idle_temp.tstamp);
+                               
+                       cpu_load_info->cpu_ticks[CPU_STATE_IDLE] =
+                               (uint32_t)(timer_grab(&idle_temp) / hz_tick_interval);
+               }
 
-           cpu_load_info = (processor_cpu_load_info_t) info;
-           for (i=0;i<CPU_STATE_MAX;i++)
-             cpu_load_info->cpu_ticks[i] = machine_slot[slot_num].cpu_ticks[i];
+               cpu_load_info->cpu_ticks[CPU_STATE_NICE] = 0;
 
            *count = PROCESSOR_CPU_LOAD_INFO_COUNT;
            *host = &realhost;
-           return(KERN_SUCCESS);
-         }
+
+           return (KERN_SUCCESS);
+       }
+
        default:
-         {
-           kr=cpu_info(flavor, slot_num, info, count);
-           if (kr == KERN_SUCCESS)
-               *host = &realhost;                 
-           return(kr);
-         }
+           result = cpu_info(flavor, cpu_id, info, count);
+           if (result == KERN_SUCCESS)
+                       *host = &realhost;                 
+
+           return (result);
        }
 }
 
 kern_return_t
 processor_start(
-       processor_t     processor)
+       processor_t                     processor)
 {
-       int     state;
-       spl_t   s;
-       kern_return_t   kr;
+       processor_set_t         pset;
+       thread_t                        thread;   
+       kern_return_t           result;
+       spl_t                           s;
 
-       if (processor == PROCESSOR_NULL)
-               return(KERN_INVALID_ARGUMENT);
+       if (processor == PROCESSOR_NULL || processor->processor_set == PROCESSOR_SET_NULL)
+               return (KERN_INVALID_ARGUMENT);
 
        if (processor == master_processor) {
-               thread_bind(current_thread(), processor);
-               thread_block((void (*)(void)) 0);
-               kr = cpu_start(processor->slot_num);
-               thread_bind(current_thread(), PROCESSOR_NULL);
+               processor_t             prev;
+
+               prev = thread_bind(processor);
+               thread_block(THREAD_CONTINUE_NULL);
+
+               result = cpu_start(processor->cpu_id);
 
-               return(kr);
+               thread_bind(prev);
+
+               return (result);
        }
 
        s = splsched();
-       processor_lock(processor);
-
-       state = processor->state;
-       if (state != PROCESSOR_OFF_LINE) {
-               processor_unlock(processor);
+       pset = processor->processor_set;
+       pset_lock(pset);
+       if (processor->state != PROCESSOR_OFF_LINE) {
+               pset_unlock(pset);
                splx(s);
-               return(KERN_FAILURE);
+
+               return (KERN_FAILURE);
        }
+
        processor->state = PROCESSOR_START;
-       processor_unlock(processor);
+       pset_unlock(pset);
        splx(s);
 
-       if (processor->next_thread == THREAD_NULL) {
-               thread_t                thread;   
-               extern void             start_cpu_thread(void);
-       
-               thread = kernel_thread_with_priority(
-                                                                       kernel_task, MAXPRI_KERNEL,
-                                                                               start_cpu_thread, TRUE, FALSE);
+       /*
+        *      Create the idle processor thread.
+        */
+       if (processor->idle_thread == THREAD_NULL) {
+               result = idle_thread_create(processor);
+               if (result != KERN_SUCCESS) {
+                       s = splsched();
+                       pset_lock(pset);
+                       processor->state = PROCESSOR_OFF_LINE;
+                       pset_unlock(pset);
+                       splx(s);
+
+                       return (result);
+               }
+       }
+
+       /*
+        *      If there is no active thread, the processor
+        *      has never been started.  Create a dedicated
+        *      start up thread.
+        */
+       if (    processor->active_thread == THREAD_NULL         &&
+                       processor->next_thread == THREAD_NULL           ) {
+               result = kernel_thread_create((thread_continue_t)processor_start_thread, NULL, MAXPRI_KERNEL, &thread);
+               if (result != KERN_SUCCESS) {
+                       s = splsched();
+                       pset_lock(pset);
+                       processor->state = PROCESSOR_OFF_LINE;
+                       pset_unlock(pset);
+                       splx(s);
+
+                       return (result);
+               }
 
                s = splsched();
                thread_lock(thread);
-               thread_bind_locked(thread, processor);
-               thread_go_locked(thread, THREAD_AWAKENED);
-               (void)rem_runq(thread);
+               thread->bound_processor = processor;
                processor->next_thread = thread;
+               thread->state = TH_RUN;
                thread_unlock(thread);
                splx(s);
+
+               thread_deallocate(thread);
        }
 
-       kr = cpu_start(processor->slot_num);
+       if (processor->processor_self == IP_NULL)
+               ipc_processor_init(processor);
 
-       if (kr != KERN_SUCCESS) {
+       result = cpu_start(processor->cpu_id);
+       if (result != KERN_SUCCESS) {
                s = splsched();
-               processor_lock(processor);
+               pset_lock(pset);
                processor->state = PROCESSOR_OFF_LINE;
-               processor_unlock(processor);
+               pset_unlock(pset);
                splx(s);
+
+               return (result);
        }
 
-       return(kr);
+       ipc_processor_enable(processor);
+
+       return (KERN_SUCCESS);
 }
 
 kern_return_t
@@ -546,52 +506,22 @@ processor_control(
        if (processor == PROCESSOR_NULL)
                return(KERN_INVALID_ARGUMENT);
 
-       return(cpu_control(processor->slot_num, info, count));
-}
-
-/*
- *     Precalculate the appropriate timesharing quanta based on load.  The
- *     index into machine_quanta is the number of threads on the
- *     processor set queue.  It is limited to the number of processors in
- *     the set.
- */
-
-void
-pset_quanta_set(
-       processor_set_t         pset)
-{
-       register int    i, ncpus;
-
-       ncpus = pset->processor_count;
-
-       for (i=1; i <= ncpus; i++)
-               pset->machine_quanta[i] = (ncpus + (i / 2)) / i;
-
-       pset->machine_quanta[0] = pset->machine_quanta[1];
-
-       i = (pset->runq.count > ncpus)? ncpus: pset->runq.count;
-       pset->set_quanta = pset->machine_quanta[i];
+       return(cpu_control(processor->cpu_id, info, count));
 }
            
 kern_return_t
 processor_set_create(
-       host_t  host,
-       processor_set_t *new_set,
-       processor_set_t *new_name)
+       __unused host_t         host,
+       __unused processor_set_t        *new_set,
+       __unused processor_set_t        *new_name)
 {
-#ifdef lint
-       host++; new_set++; new_name++;
-#endif /* lint */
        return(KERN_FAILURE);
 }
 
 kern_return_t
 processor_set_destroy(
-       processor_set_t pset)
+       __unused processor_set_t        pset)
 {
-#ifdef lint
-       pset++;
-#endif /* lint */
        return(KERN_FAILURE);
 }
 
@@ -600,14 +530,17 @@ processor_get_assignment(
        processor_t     processor,
        processor_set_t *pset)
 {
-       int state;
+       int state;
+
+       if (processor == PROCESSOR_NULL)
+               return(KERN_INVALID_ARGUMENT);
 
        state = processor->state;
        if (state == PROCESSOR_SHUTDOWN || state == PROCESSOR_OFF_LINE)
                return(KERN_FAILURE);
 
-       *pset = processor->processor_set;
-       pset_reference(*pset);
+       *pset = &pset0;
+
        return(KERN_SUCCESS);
 }
 
@@ -629,7 +562,7 @@ processor_set_info(
                        return(KERN_FAILURE);
 
                basic_info = (processor_set_basic_info_t) info;
-               basic_info->processor_count = pset->processor_count;
+               basic_info->processor_count = processor_avail_count;
                basic_info->default_policy = POLICY_TIMESHARE;
 
                *count = PROCESSOR_SET_BASIC_INFO_COUNT;
@@ -683,7 +616,7 @@ processor_set_info(
                        return(KERN_FAILURE);
 
                ts_limit = (policy_timeshare_limit_t) info;
-               ts_limit->max_priority = MAXPRI_STANDARD;
+               ts_limit->max_priority = MAXPRI_KERNEL;
 
                *count = POLICY_TIMESHARE_LIMIT_COUNT;
                *host = &realhost;
@@ -696,7 +629,7 @@ processor_set_info(
                        return(KERN_FAILURE);
 
                fifo_limit = (policy_fifo_limit_t) info;
-               fifo_limit->max_priority = MAXPRI_STANDARD;
+               fifo_limit->max_priority = MAXPRI_KERNEL;
 
                *count = POLICY_FIFO_LIMIT_COUNT;
                *host = &realhost;
@@ -709,7 +642,7 @@ processor_set_info(
                        return(KERN_FAILURE);
 
                rr_limit = (policy_rr_limit_t) info;
-               rr_limit->max_priority = MAXPRI_STANDARD;
+               rr_limit->max_priority = MAXPRI_KERNEL;
 
                *count = POLICY_RR_LIMIT_COUNT;
                *host = &realhost;
@@ -746,31 +679,28 @@ processor_set_statistics(
        processor_set_info_t    info,
        mach_msg_type_number_t  *count)
 {
-        if (pset == PROCESSOR_SET_NULL)
-                return (KERN_INVALID_PROCESSOR_SET);
+       if (pset == PROCESSOR_SET_NULL || pset != &pset0)
+               return (KERN_INVALID_PROCESSOR_SET);
 
-        if (flavor == PROCESSOR_SET_LOAD_INFO) {
-                register processor_set_load_info_t     load_info;
+       if (flavor == PROCESSOR_SET_LOAD_INFO) {
+               register processor_set_load_info_t     load_info;
 
-                if (*count < PROCESSOR_SET_LOAD_INFO_COUNT)
-                        return(KERN_FAILURE);
+               if (*count < PROCESSOR_SET_LOAD_INFO_COUNT)
+                       return(KERN_FAILURE);
 
-                load_info = (processor_set_load_info_t) info;
+               load_info = (processor_set_load_info_t) info;
 
-                pset_lock(pset);
-                load_info->task_count = pset->task_count;
-                load_info->thread_count = pset->thread_count;
-                               simple_lock(&pset->processors_lock);
-                load_info->mach_factor = pset->mach_factor;
-                load_info->load_average = pset->load_average;
-                               simple_unlock(&pset->processors_lock);
-                pset_unlock(pset);
+               load_info->mach_factor = sched_mach_factor;
+               load_info->load_average = sched_load_average;
 
-                *count = PROCESSOR_SET_LOAD_INFO_COUNT;
-                return(KERN_SUCCESS);
-        }
+               load_info->task_count = tasks_count;
+               load_info->thread_count = threads_count;
+
+               *count = PROCESSOR_SET_LOAD_INFO_COUNT;
+               return(KERN_SUCCESS);
+       }
 
-        return(KERN_INVALID_ARGUMENT);
+       return(KERN_INVALID_ARGUMENT);
 }
 
 /*
@@ -782,9 +712,9 @@ processor_set_statistics(
  */
 kern_return_t
 processor_set_max_priority(
-       processor_set_t pset,
-       int             max_priority,
-       boolean_t       change_threads)
+       __unused processor_set_t        pset,
+       __unused int                    max_priority,
+       __unused boolean_t              change_threads)
 {
        return (KERN_INVALID_ARGUMENT);
 }
@@ -797,8 +727,8 @@ processor_set_max_priority(
 
 kern_return_t
 processor_set_policy_enable(
-       processor_set_t pset,
-       int             policy)
+       __unused processor_set_t        pset,
+       __unused int                    policy)
 {
        return (KERN_INVALID_ARGUMENT);
 }
@@ -811,9 +741,9 @@ processor_set_policy_enable(
  */
 kern_return_t
 processor_set_policy_disable(
-       processor_set_t pset,
-       int             policy,
-       boolean_t       change_threads)
+       __unused processor_set_t        pset,
+       __unused int                    policy,
+       __unused boolean_t              change_threads)
 {
        return (KERN_INVALID_ARGUMENT);
 }
@@ -828,42 +758,40 @@ processor_set_policy_disable(
  */
 kern_return_t
 processor_set_things(
-       processor_set_t         pset,
-       mach_port_t             **thing_list,
+       processor_set_t                 pset,
+       mach_port_t                             **thing_list,
        mach_msg_type_number_t  *count,
-       int                     type)
+       int                                             type)
 {
        unsigned int actual;    /* this many things */
-       int i;
+       unsigned int maxthings;
+       unsigned int i;
 
        vm_size_t size, size_needed;
-       vm_offset_t addr;
+       void  *addr;
 
-       if (pset == PROCESSOR_SET_NULL)
-               return KERN_INVALID_ARGUMENT;
+       if (pset == PROCESSOR_SET_NULL || pset != &pset0)
+               return (KERN_INVALID_ARGUMENT);
 
-       size = 0; addr = 0;
+       size = 0;
+       addr = NULL;
 
        for (;;) {
-               pset_lock(pset);
-               if (!pset->active) {
-                       pset_unlock(pset);
-                       return KERN_FAILURE;
-               }
+               lck_mtx_lock(&tasks_threads_lock);
 
                if (type == THING_TASK)
-                       actual = pset->task_count;
+                       maxthings = tasks_count;
                else
-                       actual = pset->thread_count;
+                       maxthings = threads_count;
 
                /* do we have the memory we need? */
 
-               size_needed = actual * sizeof(mach_port_t);
+               size_needed = maxthings * sizeof (mach_port_t);
                if (size_needed <= size)
                        break;
 
-               /* unlock the pset and allocate more memory */
-               pset_unlock(pset);
+               /* unlock and allocate more memory */
+               lck_mtx_unlock(&tasks_threads_lock);
 
                if (size != 0)
                        kfree(addr, size);
@@ -873,124 +801,125 @@ processor_set_things(
 
                addr = kalloc(size);
                if (addr == 0)
-                       return KERN_RESOURCE_SHORTAGE;
+                       return (KERN_RESOURCE_SHORTAGE);
        }
 
-       /* OK, have memory and the processor_set is locked & active */
+       /* OK, have memory and the list locked */
 
+       actual = 0;
        switch (type) {
-           case THING_TASK: {
-               task_t *tasks = (task_t *) addr;
-               task_t task;
-
-               for (i = 0, task = (task_t) queue_first(&pset->tasks);
-                    i < actual;
-                    i++, task = (task_t) queue_next(&task->pset_tasks)) {
-                       /* take ref for convert_task_to_port */
-                       task_reference(task);
-                       tasks[i] = task;
-               }
-               assert(queue_end(&pset->tasks, (queue_entry_t) task));
-               break;
-           }
-
-           case THING_THREAD: {
-               thread_act_t *thr_acts = (thread_act_t *) addr;
-               thread_t thread;
-               thread_act_t thr_act;
-               queue_head_t *list;
-
-               list = &pset->threads;
-               thread = (thread_t) queue_first(list);
-               i = 0;
-               while (i < actual && !queue_end(list, (queue_entry_t)thread)) {
-                       thr_act = thread_lock_act(thread);
-                       if (thr_act && thr_act->ref_count > 0) {
-                               /* take ref for convert_act_to_port */
-                               act_locked_act_reference(thr_act);
-                               thr_acts[i] = thr_act;
-                               i++;
+
+       case THING_TASK: {
+               task_t          task, *task_list = (task_t *)addr;
+
+               for (task = (task_t)queue_first(&tasks);
+                                               !queue_end(&tasks, (queue_entry_t)task);
+                                                               task = (task_t)queue_next(&task->tasks)) {
+#if defined(SECURE_KERNEL)
+                       if (task != kernel_task) {
+#endif
+                               task_reference_internal(task);
+                               task_list[actual++] = task;
+#if defined(SECURE_KERNEL)
                        }
-                       thread_unlock_act(thread);
-                       thread = (thread_t) queue_next(&thread->pset_threads);
+#endif
                }
-               if (i < actual) {
-                       actual = i;
-                       size_needed = actual * sizeof(mach_port_t);
+
+               break;
+       }
+
+       case THING_THREAD: {
+               thread_t        thread, *thread_list = (thread_t *)addr;
+
+               for (thread = (thread_t)queue_first(&threads);
+                                               !queue_end(&threads, (queue_entry_t)thread);
+                                                               thread = (thread_t)queue_next(&thread->threads)) {
+                       thread_reference_internal(thread);
+                       thread_list[actual++] = thread;
                }
+
                break;
-           }
        }
 
-       /* can unlock processor set now that we have the task/thread refs */
-       pset_unlock(pset);
+       }
+               
+       lck_mtx_unlock(&tasks_threads_lock);
+
+       if (actual < maxthings)
+               size_needed = actual * sizeof (mach_port_t);
 
        if (actual == 0) {
                /* no things, so return null pointer and deallocate memory */
-               *thing_list = 0;
+               *thing_list = NULL;
                *count = 0;
 
                if (size != 0)
                        kfree(addr, size);
-       } else {
+       }
+       else {
                /* if we allocated too much, must copy */
 
                if (size_needed < size) {
-                       vm_offset_t newaddr;
+                       void *newaddr;
 
                        newaddr = kalloc(size_needed);
                        if (newaddr == 0) {
                                switch (type) {
-                                   case THING_TASK: {
-                                       task_t *tasks = (task_t *) addr;
+
+                               case THING_TASK: {
+                                       task_t          *task_list = (task_t *)addr;
 
                                        for (i = 0; i < actual; i++)
-                                               task_deallocate(tasks[i]);
+                                               task_deallocate(task_list[i]);
                                        break;
-                                   }
+                               }
 
-                                   case THING_THREAD: {
-                                       thread_t *threads = (thread_t *) addr;
+                               case THING_THREAD: {
+                                       thread_t        *thread_list = (thread_t *)addr;
 
                                        for (i = 0; i < actual; i++)
-                                               thread_deallocate(threads[i]);
+                                               thread_deallocate(thread_list[i]);
                                        break;
-                                   }
                                }
+
+                               }
+
                                kfree(addr, size);
-                               return KERN_RESOURCE_SHORTAGE;
+                               return (KERN_RESOURCE_SHORTAGE);
                        }
 
-                       bcopy((char *) addr, (char *) newaddr, size_needed);
+                       bcopy((void *) addr, (void *) newaddr, size_needed);
                        kfree(addr, size);
                        addr = newaddr;
                }
 
-               *thing_list = (mach_port_t *) addr;
+               *thing_list = (mach_port_t *)addr;
                *count = actual;
 
                /* do the conversion that Mig should handle */
 
                switch (type) {
-                   case THING_TASK: {
-                       task_t *tasks = (task_t *) addr;
+
+               case THING_TASK: {
+                       task_t          *task_list = (task_t *)addr;
 
                        for (i = 0; i < actual; i++)
-                               (*thing_list)[i] = convert_task_to_port(tasks[i]);
+                               (*thing_list)[i] = convert_task_to_port(task_list[i]);
                        break;
-                   }
+               }
 
-                   case THING_THREAD: {
-                       thread_act_t *thr_acts = (thread_act_t *) addr;
+               case THING_THREAD: {
+                       thread_t        *thread_list = (thread_t *)addr;
 
                        for (i = 0; i < actual; i++)
-                               (*thing_list)[i] = convert_act_to_port(thr_acts[i]);
+                               (*thing_list)[i] = convert_thread_to_port(thread_list[i]);
                        break;
-                   }
+               }
+
                }
        }
 
-       return(KERN_SUCCESS);
+       return (KERN_SUCCESS);
 }
 
 
@@ -1013,48 +942,34 @@ processor_set_tasks(
  *
  *     List all threads in the processor set.
  */
+#if defined(SECURE_KERNEL)
 kern_return_t
 processor_set_threads(
-       processor_set_t         pset,
-       thread_array_t          *thread_list,
-       mach_msg_type_number_t  *count)
+       __unused processor_set_t                pset,
+       __unused thread_array_t         *thread_list,
+       __unused mach_msg_type_number_t *count)
 {
-    return(processor_set_things(pset, (mach_port_t **)thread_list, count, THING_THREAD));
+    return KERN_FAILURE;
 }
-
-/*
- *      processor_set_base:
- *
- *      Specify per-policy base priority for a processor set.  Set processor
- *     set default policy to the given policy. This affects newly created
- *      and assigned threads.  Optionally change existing ones.
- */
+#elif defined(CONFIG_EMBEDDED)
 kern_return_t
-processor_set_base(
-       processor_set_t         pset,
-       policy_t                policy,
-        policy_base_t           base,
-       boolean_t               change)
+processor_set_threads(
+       __unused processor_set_t                pset,
+       __unused thread_array_t         *thread_list,
+       __unused mach_msg_type_number_t *count)
 {
-       return (KERN_INVALID_ARGUMENT);
+    return KERN_NOT_SUPPORTED;
 }
-
-/*
- *      processor_set_limit:
- *
- *      Specify per-policy limits for a processor set.  This affects
- *      newly created and assigned threads.  Optionally change existing
- *      ones.
- */
+#else
 kern_return_t
-processor_set_limit(
-       processor_set_t         pset,
-       policy_t                policy,
-        policy_limit_t         limit,
-       boolean_t               change)
+processor_set_threads(
+       processor_set_t         pset,
+       thread_array_t          *thread_list,
+       mach_msg_type_number_t  *count)
 {
-       return (KERN_POLICY_LIMIT);
+    return(processor_set_things(pset, (mach_port_t **)thread_list, count, THING_THREAD));
 }
+#endif
 
 /*
  *     processor_set_policy_control
@@ -1065,11 +980,29 @@ processor_set_limit(
  */
 kern_return_t
 processor_set_policy_control(
-       processor_set_t         pset,
-       int                     flavor,
-       processor_set_info_t    policy_info,
-       mach_msg_type_number_t  count,
-       boolean_t               change)
+       __unused processor_set_t                pset,
+       __unused int                            flavor,
+       __unused processor_set_info_t   policy_info,
+       __unused mach_msg_type_number_t count,
+       __unused boolean_t                      change)
 {
        return (KERN_INVALID_ARGUMENT);
 }
+
+#undef pset_deallocate
+void pset_deallocate(processor_set_t pset);
+void
+pset_deallocate(
+__unused processor_set_t       pset)
+{
+       return;
+}
+
+#undef pset_reference
+void pset_reference(processor_set_t pset);
+void
+pset_reference(
+__unused processor_set_t       pset)
+{
+       return;
+}