X-Git-Url: https://git.saurik.com/apple/xnu.git/blobdiff_plain/4452a7af2eac33dbad800bcc91f2399d62c18f53..d9a64523371fa019c4575bb400cbbc3a50ac9903:/osfmk/i386/cpu.c diff --git a/osfmk/i386/cpu.c b/osfmk/i386/cpu.c index 5ad19a057..84cf06615 100644 --- a/osfmk/i386/cpu.c +++ b/osfmk/i386/cpu.c @@ -1,5 +1,5 @@ /* - * Copyright (c) 2000 Apple Computer, Inc. All rights reserved. + * Copyright (c) 2000-2009 Apple Inc. All rights reserved. * * @APPLE_OSREFERENCE_LICENSE_HEADER_START@ * @@ -35,15 +35,18 @@ #include #include #include -#include +#include #include #include -#include #include #include -#include +#include +#include +#if CONFIG_VMX +#include +#endif #include -#include "cpuid.h" +#include struct processor processor_master; @@ -54,7 +57,7 @@ cpu_control( processor_info_t info, unsigned int count) { - printf("cpu_control(%d,0x%x,%d) not implemented\n", + printf("cpu_control(%d,%p,%d) not implemented\n", slot_num, info, count); return (KERN_FAILURE); } @@ -77,7 +80,7 @@ cpu_info( processor_info_t info, unsigned int *count) { - printf("cpu_info(%d,%d,0x%x,0x%x) not implemented\n", + printf("cpu_info(%d,%d,%p,%p) not implemented\n", flavor, slot_num, info, count); return (KERN_FAILURE); } @@ -85,11 +88,9 @@ cpu_info( void cpu_sleep(void) { - cpu_data_t *proc_info = current_cpu_datap(); - - proc_info->cpu_running = FALSE; + cpu_data_t *cdp = current_cpu_datap(); - PE_cpu_machine_quiesce(proc_info->cpu_id); + PE_cpu_machine_quiesce(cdp->cpu_id); cpu_thread_halt(); } @@ -99,10 +100,13 @@ cpu_init(void) { cpu_data_t *cdp = current_cpu_datap(); + timer_call_queue_init(&cdp->rtclock_timer.queue); + cdp->rtclock_timer.deadline = EndOfAllTime; + cdp->cpu_type = cpuid_cputype(); cdp->cpu_subtype = cpuid_cpusubtype(); - cdp->cpu_running = TRUE; + i386_activate_cpu(); } kern_return_t @@ -114,20 +118,59 @@ cpu_start( if (cpu == cpu_number()) { cpu_machine_init(); return KERN_SUCCESS; - } else { + } + + /* + * Try to bring the CPU back online without a reset. + * If the fast restart doesn't succeed, fall back to + * the slow way. + */ + ret = intel_startCPU_fast(cpu); + if (ret != KERN_SUCCESS) { /* * Should call out through PE. * But take the shortcut here. */ ret = intel_startCPU(cpu); - return(ret); } + + if (ret != KERN_SUCCESS) + kprintf("cpu: cpu_start(%d) returning failure!\n", cpu); + + return(ret); } void cpu_exit_wait( - __unused int cpu) + int cpu) { + cpu_data_t *cdp = cpu_datap(cpu); + boolean_t intrs_enabled; + uint64_t tsc_timeout; + + /* + * Wait until the CPU indicates that it has stopped. + * Disable interrupts while the topo lock is held -- arguably + * this should always be done but in this instance it can lead to + * a timeout if long-running interrupt were to occur here. + */ + intrs_enabled = ml_set_interrupts_enabled(FALSE); + mp_safe_spin_lock(&x86_topo_lock); + /* Set a generous timeout of several seconds (in TSC ticks) */ + tsc_timeout = rdtsc64() + (10ULL * 1000 * 1000 * 1000); + while ((cdp->lcpu.state != LCPU_HALT) + && (cdp->lcpu.state != LCPU_OFF) + && !cdp->lcpu.stopped) { + simple_unlock(&x86_topo_lock); + ml_set_interrupts_enabled(intrs_enabled); + cpu_pause(); + if (rdtsc64() > tsc_timeout) + panic("cpu_exit_wait(%d) timeout", cpu); + ml_set_interrupts_enabled(FALSE); + mp_safe_spin_lock(&x86_topo_lock); + } + simple_unlock(&x86_topo_lock); + ml_set_interrupts_enabled(intrs_enabled); } void @@ -139,14 +182,12 @@ cpu_machine_init( PE_cpu_machine_init(cdp->cpu_id, !cdp->cpu_boot_complete); cdp->cpu_boot_complete = TRUE; cdp->cpu_running = TRUE; -#if 0 - if (cpu_datap(cpu)->hibernate) - { - cpu_datap(cpu)->hibernate = 0; - hibernate_machine_init(); - } -#endif ml_init_interrupt(); + +#if CONFIG_VMX + /* initialize VMX for every CPU */ + vmx_cpu_init(); +#endif } processor_t @@ -158,7 +199,7 @@ cpu_processor_alloc(boolean_t is_boot_cpu) if (is_boot_cpu) return &processor_master; - ret = kmem_alloc(kernel_map, (vm_offset_t *) &proc, sizeof(*proc)); + ret = kmem_alloc(kernel_map, (vm_offset_t *) &proc, sizeof(*proc), VM_KERN_MEMORY_OSFMK); if (ret != KERN_SUCCESS) return NULL;