X-Git-Url: https://git.saurik.com/apple/xnu.git/blobdiff_plain/0b4e3aa066abc0728aacb4bbeb86f53f9737156e..c7d2c2c6ee645e10cbccdd01c6191873ec77239d:/osfmk/kern/machine.c?ds=inline diff --git a/osfmk/kern/machine.c b/osfmk/kern/machine.c index b1a1ee73a..dfe33564d 100644 --- a/osfmk/kern/machine.c +++ b/osfmk/kern/machine.c @@ -1,23 +1,29 @@ /* - * Copyright (c) 2000 Apple Computer, Inc. All rights reserved. + * Copyright (c) 2000-2009 Apple Inc. All rights reserved. * - * @APPLE_LICENSE_HEADER_START@ + * @APPLE_OSREFERENCE_LICENSE_HEADER_START@ * - * The contents of this file constitute Original Code as defined in and - * are subject to the Apple Public Source License Version 1.1 (the - * "License"). You may not use this file except in compliance with the - * License. Please obtain a copy of the License at - * http://www.apple.com/publicsource and read it before using this file. + * This file contains Original Code and/or Modifications of Original Code + * as defined in and that are subject to the Apple Public Source License + * Version 2.0 (the 'License'). You may not use this file except in + * compliance with the License. The rights granted to you under the License + * may not be used to create, or enable the creation or redistribution of, + * unlawful or unlicensed copies of an Apple operating system, or to + * circumvent, violate, or enable the circumvention or violation of, any + * terms of an Apple operating system software license agreement. * - * This Original Code and all software distributed under the License are - * distributed on an "AS IS" basis, WITHOUT WARRANTY OF ANY KIND, EITHER + * Please obtain a copy of the License at + * http://www.opensource.apple.com/apsl/ and read it before using this file. + * + * The Original Code and all software distributed under the License are + * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES, * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY, - * FITNESS FOR A PARTICULAR PURPOSE OR NON-INFRINGEMENT. Please see the - * License for the specific language governing rights and limitations - * under the License. + * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT. + * Please see the License for the specific language governing rights and + * limitations under the License. * - * @APPLE_LICENSE_HEADER_END@ + * @APPLE_OSREFERENCE_LICENSE_HEADER_END@ */ /* * @OSF_COPYRIGHT@ @@ -57,125 +63,87 @@ * Support for machine independent machine abstraction. */ -#include - #include + +#include #include #include -#include #include #include #include +#include +#include + +#include #include #include #include #include -#include #include +#include #include #include #include #include #include -#include -#include -#include +#include + +#if HIBERNATION +#include +#endif +#include + +#if CONFIG_DTRACE +extern void (*dtrace_cpu_state_changed_hook)(int, boolean_t); +#endif /* * Exported variables: */ struct machine_info machine_info; -struct machine_slot machine_slot[NCPUS]; - -static queue_head_t processor_action_queue; -static boolean_t processor_action_active; -static thread_call_t processor_action_call; -static thread_call_data_t processor_action_call_data; -decl_simple_lock_data(static,processor_action_lock) - -thread_t machine_wake_thread; /* Forwards */ -processor_set_t processor_request_action( - processor_t processor, - processor_set_t new_pset); - -void processor_doaction( - processor_t processor); - void processor_doshutdown( processor_t processor); /* - * cpu_up: + * processor_up: * - * Flag specified cpu as up and running. Called when a processor comes - * online. + * Flag processor as up and running, and available + * for scheduling. */ void -cpu_up( - int cpu) +processor_up( + processor_t processor) { - processor_t processor = cpu_to_processor(cpu); - struct machine_slot *ms; - spl_t s; - - /* - * Just twiddle our thumbs; we've got nothing better to do - * yet, anyway. - */ - while (!simple_lock_try(&default_pset.processors_lock)) - continue; + processor_set_t pset; + spl_t s; s = splsched(); - processor_lock(processor); init_ast_check(processor); - ms = &machine_slot[cpu]; - ms->running = TRUE; - machine_info.avail_cpus++; - pset_add_processor(&default_pset, processor); + pset = processor->processor_set; + pset_lock(pset); + ++pset->online_processor_count; + enqueue_tail(&pset->active_queue, (queue_entry_t)processor); processor->state = PROCESSOR_RUNNING; - processor_unlock(processor); + (void)hw_atomic_add(&processor_avail_count, 1); + commpage_update_active_cpus(); + pset_unlock(pset); + ml_cpu_up(); splx(s); - simple_unlock(&default_pset.processors_lock); -} - -/* - * cpu_down: - * - * Flag specified cpu as down. Called when a processor is about to - * go offline. - */ -void -cpu_down( - int cpu) -{ - processor_t processor; - struct machine_slot *ms; - spl_t s; - - processor = cpu_to_processor(cpu); - - s = splsched(); - processor_lock(processor); - ms = &machine_slot[cpu]; - ms->running = FALSE; - machine_info.avail_cpus--; - /* - * processor has already been removed from pset. - */ - processor->processor_set_next = PROCESSOR_SET_NULL; - processor->state = PROCESSOR_OFF_LINE; - processor_unlock(processor); - splx(s); +#if CONFIG_DTRACE + if (dtrace_cpu_state_changed_hook) + (*dtrace_cpu_state_changed_hook)(processor->cpu_id, TRUE); +#endif } +#include kern_return_t host_reboot( - host_priv_t host_priv, + host_priv_t host_priv, int options) { if (host_priv == HOST_PRIV_NULL) @@ -183,296 +151,179 @@ host_reboot( assert(host_priv == &realhost); +#if DEVELOPMENT || DEBUG if (options & HOST_REBOOT_DEBUGGER) { Debugger("Debugger"); + return (KERN_SUCCESS); } - else - halt_all_cpus(!(options & HOST_REBOOT_HALT)); - - return (KERN_SUCCESS); -} - -/* - * processor_request_action: - * - * Common internals of processor_assign and processor_shutdown. - * If new_pset is null, this is a shutdown, else it's an assign - * and caller must donate a reference. - * For assign operations, it returns an old pset that must be deallocated - * if it's not NULL. - * For shutdown operations, it always returns PROCESSOR_SET_NULL. - */ -processor_set_t -processor_request_action( - processor_t processor, - processor_set_t new_pset) -{ - processor_set_t pset, old_next_pset; - - /* - * Processor must be in a processor set. Must lock its idle lock to - * get at processor state. - */ - pset = processor->processor_set; - simple_lock(&pset->idle_lock); - - /* - * If the processor is dispatching, let it finish - it will set its - * state to running very soon. - */ - while (*(volatile int *)&processor->state == PROCESSOR_DISPATCHING) { - simple_unlock(&pset->idle_lock); - simple_lock(&pset->idle_lock); - } - - /* - * Now lock the action queue and do the dirty work. - */ - simple_lock(&processor_action_lock); - - switch (processor->state) { - - case PROCESSOR_IDLE: - /* - * Remove from idle queue. - */ - queue_remove(&pset->idle_queue, processor, - processor_t, processor_queue); - pset->idle_count--; - - /* fall through ... */ - case PROCESSOR_RUNNING: - /* - * Put it on the action queue. - */ - queue_enter(&processor_action_queue, processor, - processor_t,processor_queue); - - /* Fall through ... */ - case PROCESSOR_ASSIGN: - /* - * And ask the action_thread to do the work. - */ - - if (new_pset == PROCESSOR_SET_NULL) { - processor->state = PROCESSOR_SHUTDOWN; - old_next_pset = PROCESSOR_SET_NULL; - } else { - processor->state = PROCESSOR_ASSIGN; - old_next_pset = processor->processor_set_next; - processor->processor_set_next = new_pset; - } - break; - - default: - printf("state: %d\n", processor->state); - panic("processor_request_action: bad state"); - } +#endif - if (processor_action_active == FALSE) { - processor_action_active = TRUE; - simple_unlock(&processor_action_lock); - simple_unlock(&pset->idle_lock); - processor_unlock(processor); - thread_call_enter(processor_action_call); - processor_lock(processor); - } else { - simple_unlock(&processor_action_lock); - simple_unlock(&pset->idle_lock); - } + if (options & HOST_REBOOT_UPSDELAY) { + // UPS power cutoff path + PEHaltRestart( kPEUPSDelayHaltCPU ); + } else { + halt_all_cpus(!(options & HOST_REBOOT_HALT)); + } - return (old_next_pset); + return (KERN_SUCCESS); } kern_return_t processor_assign( - processor_t processor, - processor_set_t new_pset, - boolean_t wait) + __unused processor_t processor, + __unused processor_set_t new_pset, + __unused boolean_t wait) { -#ifdef lint - processor++; new_pset++; wait++; -#endif /* lint */ return (KERN_FAILURE); } -/* - * processor_shutdown() queues a processor up for shutdown. - * Any assignment in progress is overriden. - */ kern_return_t processor_shutdown( - processor_t processor) + processor_t processor) { - spl_t s; + processor_set_t pset; + spl_t s; s = splsched(); - processor_lock(processor); - if ((processor->state == PROCESSOR_OFF_LINE) || - (processor->state == PROCESSOR_SHUTDOWN)) { + pset = processor->processor_set; + pset_lock(pset); + if (processor->state == PROCESSOR_OFF_LINE) { /* - * Already shutdown or being shutdown -- nothing to do. + * Success if already shutdown. */ - processor_unlock(processor); + pset_unlock(pset); splx(s); return (KERN_SUCCESS); } - (void) processor_request_action(processor, PROCESSOR_SET_NULL); - - assert_wait((event_t)processor, THREAD_UNINT); - - processor_unlock(processor); - splx(s); + if (processor->state == PROCESSOR_START) { + /* + * Failure if currently being started. + */ + pset_unlock(pset); + splx(s); - thread_block((void (*)(void)) 0); + return (KERN_FAILURE); + } - return (KERN_SUCCESS); -} + /* + * If the processor is dispatching, let it finish. + */ + while (processor->state == PROCESSOR_DISPATCHING) { + pset_unlock(pset); + splx(s); + delay(1); + s = splsched(); + pset_lock(pset); + } -/* - * processor_action() shuts down processors or changes their assignment. - */ -static void -_processor_action( - thread_call_param_t p0, - thread_call_param_t p1) -{ - register processor_t processor; - spl_t s; + /* + * Success if already being shutdown. + */ + if (processor->state == PROCESSOR_SHUTDOWN) { + pset_unlock(pset); + splx(s); - s = splsched(); - simple_lock(&processor_action_lock); + return (KERN_SUCCESS); + } - while (!queue_empty(&processor_action_queue)) { - processor = (processor_t) queue_first(&processor_action_queue); - queue_remove(&processor_action_queue, processor, - processor_t, processor_queue); - simple_unlock(&processor_action_lock); - splx(s); + if (processor->state == PROCESSOR_IDLE) + remqueue((queue_entry_t)processor); + else + if (processor->state == PROCESSOR_RUNNING) + remqueue((queue_entry_t)processor); - processor_doaction(processor); + processor->state = PROCESSOR_SHUTDOWN; - s = splsched(); - simple_lock(&processor_action_lock); - } + pset_unlock(pset); - processor_action_active = FALSE; - simple_unlock(&processor_action_lock); + processor_doshutdown(processor); splx(s); -} -void -processor_action(void) -{ - queue_init(&processor_action_queue); - simple_lock_init(&processor_action_lock, ETAP_THREAD_ACTION); - processor_action_active = FALSE; + cpu_exit_wait(processor->cpu_id); - thread_call_setup(&processor_action_call_data, _processor_action, NULL); - processor_action_call = &processor_action_call_data; + return (KERN_SUCCESS); } /* - * processor_doaction actually does the shutdown. The trick here - * is to schedule ourselves onto a cpu and then save our - * context back into the runqs before taking out the cpu. + * Called with interrupts disabled. */ void -processor_doaction( - processor_t processor) +processor_doshutdown( + processor_t processor) { - thread_t self = current_thread(); - processor_set_t pset; - thread_t old_thread; - spl_t s; + thread_t old_thread, self = current_thread(); + processor_t prev; + processor_set_t pset; /* * Get onto the processor to shutdown */ - thread_bind(self, processor); - thread_block((void (*)(void)) 0); + prev = thread_bind(processor); + thread_block(THREAD_CONTINUE_NULL); - pset = processor->processor_set; - simple_lock(&pset->processors_lock); + assert(processor->state == PROCESSOR_SHUTDOWN); - if (pset->processor_count == 1) { - thread_t thread; - extern void start_cpu_thread(void); +#if CONFIG_DTRACE + if (dtrace_cpu_state_changed_hook) + (*dtrace_cpu_state_changed_hook)(processor->cpu_id, FALSE); +#endif - simple_unlock(&pset->processors_lock); + ml_cpu_down(); - /* - * Create the thread, and point it at the routine. - */ - thread = kernel_thread_with_priority( - kernel_task, MAXPRI_KERNEL, - start_cpu_thread, TRUE, FALSE); - - disable_preemption(); - - s = splsched(); - thread_lock(thread); - thread->state |= TH_RUN; - _mk_sp_thread_unblock(thread); - (void)rem_runq(thread); - machine_wake_thread = thread; - thread_unlock(thread); - splx(s); - - simple_lock(&pset->processors_lock); - enable_preemption(); +#if HIBERNATION + if (processor_avail_count < 2) { + hibernate_vm_lock(); + hibernate_vm_unlock(); } +#endif - s = splsched(); - processor_lock(processor); + pset = processor->processor_set; + pset_lock(pset); + processor->state = PROCESSOR_OFF_LINE; + --pset->online_processor_count; + (void)hw_atomic_sub(&processor_avail_count, 1); + commpage_update_active_cpus(); + SCHED(processor_queue_shutdown)(processor); + /* pset lock dropped */ /* - * Do shutdown, make sure we live when processor dies. + * Continue processor shutdown in shutdown context. */ - if (processor->state != PROCESSOR_SHUTDOWN) { - panic("action_thread -- bad processor state"); - } + thread_bind(prev); + old_thread = machine_processor_shutdown(self, processor_offline, processor); - pset_remove_processor(pset, processor); - processor_unlock(processor); - simple_unlock(&pset->processors_lock); - - /* - * Clean up. - */ - thread_bind(self, PROCESSOR_NULL); - self->continuation = 0; - old_thread = switch_to_shutdown_context(self, - processor_doshutdown, processor); - thread_dispatch(old_thread); - thread_wakeup((event_t)processor); - splx(s); + thread_dispatch(old_thread, self); } /* - * Actually do the processor shutdown. This is called at splsched, - * running on the processor's shutdown stack. + *Complete the shutdown and place the processor offline. + * + * Called at splsched in the shutdown context. */ - void -processor_doshutdown( - processor_t processor) +processor_offline( + processor_t processor) { - register int cpu = processor->slot_num; + thread_t new_thread, old_thread = processor->active_thread; - timer_call_cancel(&processor->quantum_timer); - thread_dispatch(current_thread()); - timer_switch(&kernel_timer[cpu]); + new_thread = processor->idle_thread; + processor->active_thread = new_thread; + processor->current_pri = IDLEPRI; + processor->current_thmode = TH_MODE_NONE; + processor->deadline = UINT64_MAX; + new_thread->last_processor = processor; + + processor->last_dispatch = mach_absolute_time(); + timer_stop(PROCESSOR_DATA(processor, thread_timer), processor->last_dispatch); + + machine_set_current_thread(new_thread); + + thread_dispatch(old_thread, new_thread); + + PMAP_DEACTIVATE_KERNEL(processor->cpu_id); - /* - * OK, now exit this cpu. - */ - PMAP_DEACTIVATE_KERNEL(cpu); - cpu_data[cpu].active_thread = THREAD_NULL; - active_kloaded[cpu] = THR_ACT_NULL; - cpu_down(cpu); cpu_sleep(); panic("zombie processor"); /*NOTREACHED*/ @@ -483,11 +334,7 @@ host_get_boot_info( host_priv_t host_priv, kernel_boot_info_t boot_info) { - char *src = ""; - extern char *machine_boot_info( - kernel_boot_info_t boot_info, - vm_size_t buf_len); - + const char *src = ""; if (host_priv == HOST_PRIV_NULL) return (KERN_INVALID_HOST);