X-Git-Url: https://git.saurik.com/apple/xnu.git/blobdiff_plain/2d21ac55c334faf3a56e5634905ed6987fc787d4..527f99514973766e9c0382a4d8550dfb00f54939:/osfmk/i386/cpu.c diff --git a/osfmk/i386/cpu.c b/osfmk/i386/cpu.c index b0f87d7da..84cf06615 100644 --- a/osfmk/i386/cpu.c +++ b/osfmk/i386/cpu.c @@ -1,5 +1,5 @@ /* - * Copyright (c) 2000-2007 Apple Inc. All rights reserved. + * Copyright (c) 2000-2009 Apple Inc. All rights reserved. * * @APPLE_OSREFERENCE_LICENSE_HEADER_START@ * @@ -35,14 +35,18 @@ #include #include #include +#include #include #include -#include #include #include -#include +#include +#include +#if CONFIG_VMX +#include +#endif #include -#include "cpuid.h" +#include struct processor processor_master; @@ -86,8 +90,6 @@ cpu_sleep(void) { cpu_data_t *cdp = current_cpu_datap(); - i386_deactivate_cpu(); - PE_cpu_machine_quiesce(cdp->cpu_id); cpu_thread_halt(); @@ -98,6 +100,9 @@ cpu_init(void) { cpu_data_t *cdp = current_cpu_datap(); + timer_call_queue_init(&cdp->rtclock_timer.queue); + cdp->rtclock_timer.deadline = EndOfAllTime; + cdp->cpu_type = cpuid_cputype(); cdp->cpu_subtype = cpuid_cpusubtype(); @@ -113,14 +118,26 @@ cpu_start( if (cpu == cpu_number()) { cpu_machine_init(); return KERN_SUCCESS; - } else { + } + + /* + * Try to bring the CPU back online without a reset. + * If the fast restart doesn't succeed, fall back to + * the slow way. + */ + ret = intel_startCPU_fast(cpu); + if (ret != KERN_SUCCESS) { /* * Should call out through PE. * But take the shortcut here. */ ret = intel_startCPU(cpu); - return(ret); } + + if (ret != KERN_SUCCESS) + kprintf("cpu: cpu_start(%d) returning failure!\n", cpu); + + return(ret); } void @@ -128,14 +145,32 @@ cpu_exit_wait( int cpu) { cpu_data_t *cdp = cpu_datap(cpu); - - simple_lock(&x86_topo_lock); - while (!cdp->lcpu.halted) { + boolean_t intrs_enabled; + uint64_t tsc_timeout; + + /* + * Wait until the CPU indicates that it has stopped. + * Disable interrupts while the topo lock is held -- arguably + * this should always be done but in this instance it can lead to + * a timeout if long-running interrupt were to occur here. + */ + intrs_enabled = ml_set_interrupts_enabled(FALSE); + mp_safe_spin_lock(&x86_topo_lock); + /* Set a generous timeout of several seconds (in TSC ticks) */ + tsc_timeout = rdtsc64() + (10ULL * 1000 * 1000 * 1000); + while ((cdp->lcpu.state != LCPU_HALT) + && (cdp->lcpu.state != LCPU_OFF) + && !cdp->lcpu.stopped) { simple_unlock(&x86_topo_lock); + ml_set_interrupts_enabled(intrs_enabled); cpu_pause(); - simple_lock(&x86_topo_lock); + if (rdtsc64() > tsc_timeout) + panic("cpu_exit_wait(%d) timeout", cpu); + ml_set_interrupts_enabled(FALSE); + mp_safe_spin_lock(&x86_topo_lock); } simple_unlock(&x86_topo_lock); + ml_set_interrupts_enabled(intrs_enabled); } void @@ -147,17 +182,12 @@ cpu_machine_init( PE_cpu_machine_init(cdp->cpu_id, !cdp->cpu_boot_complete); cdp->cpu_boot_complete = TRUE; cdp->cpu_running = TRUE; -#if 0 - if (cpu_datap(cpu)->hibernate) - { - cpu_datap(cpu)->hibernate = 0; - hibernate_machine_init(); - } -#endif ml_init_interrupt(); - /* for every CPU, get the VT specs */ - vmx_get_specs(); +#if CONFIG_VMX + /* initialize VMX for every CPU */ + vmx_cpu_init(); +#endif } processor_t @@ -169,7 +199,7 @@ cpu_processor_alloc(boolean_t is_boot_cpu) if (is_boot_cpu) return &processor_master; - ret = kmem_alloc(kernel_map, (vm_offset_t *) &proc, sizeof(*proc)); + ret = kmem_alloc(kernel_map, (vm_offset_t *) &proc, sizeof(*proc), VM_KERN_MEMORY_OSFMK); if (ret != KERN_SUCCESS) return NULL;