]> git.saurik.com Git - apple/xnu.git/blobdiff - osfmk/i386/cpu_topology.c
xnu-1228.tar.gz
[apple/xnu.git] / osfmk / i386 / cpu_topology.c
diff --git a/osfmk/i386/cpu_topology.c b/osfmk/i386/cpu_topology.c
new file mode 100644 (file)
index 0000000..56b3b43
--- /dev/null
@@ -0,0 +1,258 @@
+/*
+ * Copyright (c) 2007 Apple Inc. All rights reserved.
+ *
+ * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
+ * 
+ * This file contains Original Code and/or Modifications of Original Code
+ * as defined in and that are subject to the Apple Public Source License
+ * Version 2.0 (the 'License'). You may not use this file except in
+ * compliance with the License. The rights granted to you under the License
+ * may not be used to create, or enable the creation or redistribution of,
+ * unlawful or unlicensed copies of an Apple operating system, or to
+ * circumvent, violate, or enable the circumvention or violation of, any
+ * terms of an Apple operating system software license agreement.
+ * 
+ * Please obtain a copy of the License at
+ * http://www.opensource.apple.com/apsl/ and read it before using this file.
+ * 
+ * The Original Code and all software distributed under the License are
+ * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
+ * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
+ * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
+ * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
+ * Please see the License for the specific language governing rights and
+ * limitations under the License.
+ * 
+ * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
+ */
+
+#include <mach/machine.h>
+#include <mach/processor.h>
+#include <kern/kalloc.h>
+#include <i386/cpu_affinity.h>
+#include <i386/cpu_topology.h>
+#include <i386/cpu_data.h>
+#include <i386/cpu_threads.h>
+#include <i386/machine_cpu.h>
+#include <i386/machine_routines.h>
+#include <i386/lock.h>
+#include <i386/mp.h>
+
+//#define TOPO_DEBUG 1
+#if TOPO_DEBUG
+#define DBG(x...)      kprintf("DBG: " x)
+#else
+#define DBG(x...)
+#endif
+
+__private_extern__ void qsort(
+    void * array,
+    size_t nmembers,
+    size_t member_size,
+    int (*)(const void *, const void *));
+
+static int lapicid_cmp(const void *x, const void *y);
+static x86_affinity_set_t *find_cache_affinity(x86_cpu_cache_t *L2_cachep);
+
+x86_affinity_set_t     *x86_affinities = NULL;
+static int             x86_affinity_count = 0;
+
+/*
+ * cpu_topology_start() is called after all processors have been registered
+ * but before any non-boot processor id started.
+ * We establish canonical logical processor numbering - logical cpus must be
+ * contiguous, zero-based and assigned in physical (local apic id) order.
+ * This step is required because the discovery/registration order is
+ * non-deterministic - cores are registered in differing orders over boots.
+ * Enforcing canonical numbering simplifies identification
+ * of processors - in particular, for stopping/starting from CHUD.
+ */ 
+void
+cpu_topology_start(void)
+{
+       int             ncpus = machine_info.max_cpus;
+       int             i;
+       boolean_t       istate;
+
+       assert(machine_info.physical_cpu == 1);
+       assert(machine_info.logical_cpu == 1);
+       assert(master_cpu == 0);
+       assert(cpu_number() == 0);
+       assert(cpu_datap(0)->cpu_number == 0);
+       
+       /* Lights out for this */
+       istate = ml_set_interrupts_enabled(FALSE);
+
+#ifdef TOPO_DEBUG
+       DBG("cpu_topology_start() %d cpu%s registered\n",
+               ncpus, (ncpus > 1) ? "s" : "");
+       for (i = 0; i < ncpus; i++) {
+               cpu_data_t      *cpup = cpu_datap(i);
+               DBG("\tcpu_data[%d]:0x%08x local apic 0x%x\n",
+                       i, (unsigned) cpup, cpup->cpu_phys_number);
+       }
+#endif
+       /*
+        * Re-order the cpu_data_ptr vector sorting by physical id.
+        * Skip the boot processor, it's required to be correct.
+        */
+       if (ncpus > 1) {
+               qsort((void *) &cpu_data_ptr[1],
+                       ncpus - 1,
+                       sizeof(cpu_data_t *),
+                       lapicid_cmp);
+       }
+#ifdef TOPO_DEBUG
+       DBG("cpu_topology_start() after sorting:\n");
+       for (i = 0; i < ncpus; i++) {
+               cpu_data_t      *cpup = cpu_datap(i);
+               DBG("\tcpu_data[%d]:0x%08x local apic 0x%x\n",
+                       i, (unsigned) cpup, cpup->cpu_phys_number);
+       }
+#endif
+
+       /*
+        * Fix up logical numbers and reset the map kept by the lapic code.
+        */
+       for (i = 1; i < ncpus; i++) {
+               cpu_data_t      *cpup = cpu_datap(i);
+
+               if (cpup->cpu_number != i) {
+                       kprintf("cpu_datap(%d):0x%08x local apic id 0x%x "
+                               "remapped from %d\n",
+                               i, (unsigned) cpup, cpup->cpu_phys_number,
+                               cpup->cpu_number);
+               }
+               cpup->cpu_number = i;
+               cpup->lcpu.lnum = i;
+               lapic_cpu_map(cpup->cpu_phys_number, i);
+       }
+
+       ml_set_interrupts_enabled(istate);
+
+       /*
+        * Iterate over all logical cpus finding or creating the affinity set
+        * for their L2 cache. Each affinity set possesses a processor set
+        * into which each logical processor is added.
+        */
+       DBG("cpu_topology_start() creating affinity sets:\n");
+       for (i = 0; i < ncpus; i++) {
+               cpu_data_t              *cpup = cpu_datap(i);
+               x86_lcpu_t              *lcpup = cpu_to_lcpu(i);
+               x86_cpu_cache_t         *L2_cachep;
+               x86_affinity_set_t      *aset;
+
+               L2_cachep = lcpup->caches[CPU_CACHE_DEPTH_L2];
+               assert(L2_cachep->type == CPU_CACHE_TYPE_UNIF);
+               aset = find_cache_affinity(L2_cachep); 
+               if (aset == NULL) {
+                       aset = (x86_affinity_set_t *) kalloc(sizeof(*aset));
+                       if (aset == NULL)
+                               panic("cpu_topology_start() failed aset alloc");
+                       aset->next = x86_affinities;
+                       x86_affinities = aset;
+                       aset->num = x86_affinity_count++;
+                       aset->cache = L2_cachep;
+                       aset->pset = (i == master_cpu) ?
+                                       processor_pset(master_processor) :
+                                       pset_create(pset_node_root());
+                       if (aset->pset == PROCESSOR_SET_NULL)
+                               panic("cpu_topology_start: pset_create");
+                       DBG("\tnew set %p(%d) pset %p for cache %p\n",
+                               aset, aset->num, aset->pset, aset->cache);
+               }
+
+               DBG("\tprocessor_init set %p(%d) lcpup %p(%d) cpu %p processor %p\n",
+                       aset, aset->num, lcpup, lcpup->lnum, cpup, cpup->cpu_processor);
+
+               if (i != master_cpu)
+                       processor_init(cpup->cpu_processor, i, aset->pset);
+       }
+
+       /*
+        * Finally we start all processors (including the boot cpu we're
+        * running on).
+        */
+       DBG("cpu_topology_start() processor_start():\n");
+       for (i = 0; i < ncpus; i++) {
+               DBG("\tlcpu %d\n", cpu_datap(i)->cpu_number);
+               processor_start(cpu_datap(i)->cpu_processor); 
+       }
+}
+
+static int
+lapicid_cmp(const void *x, const void *y)
+{
+       cpu_data_t      *cpu_x = *((cpu_data_t **)(uintptr_t)x);
+       cpu_data_t      *cpu_y = *((cpu_data_t **)(uintptr_t)y);
+
+       DBG("lapicid_cmp(%p,%p) (%d,%d)\n",
+               x, y, cpu_x->cpu_phys_number, cpu_y->cpu_phys_number);
+       if (cpu_x->cpu_phys_number < cpu_y->cpu_phys_number)
+               return -1;
+       if (cpu_x->cpu_phys_number == cpu_y->cpu_phys_number)
+               return 0;
+       return 1;
+}
+
+static x86_affinity_set_t *
+find_cache_affinity(x86_cpu_cache_t *l2_cachep)
+{
+       x86_affinity_set_t      *aset;
+
+       for (aset = x86_affinities; aset != NULL; aset = aset->next) {
+               if (l2_cachep == aset->cache)
+                       break;
+       }
+       return aset;                    
+}
+
+int
+ml_get_max_affinity_sets(void)
+{
+       return x86_affinity_count;
+}
+
+processor_set_t
+ml_affinity_to_pset(uint32_t affinity_num) 
+{
+       x86_affinity_set_t      *aset;
+
+       for (aset = x86_affinities; aset != NULL; aset = aset->next) {
+               if (affinity_num == aset->num)
+                       break;
+       }
+       return (aset == NULL) ? PROCESSOR_SET_NULL : aset->pset;                        
+       
+}
+
+uint64_t
+ml_cpu_cache_size(unsigned int level)
+{
+       x86_cpu_cache_t *cachep;
+
+       if (level == 0) {
+               return machine_info.max_mem;
+       } else if ( 1 <= level && level <= 3) {
+               cachep = current_cpu_datap()->lcpu.caches[level-1];
+               return cachep ? cachep->cache_size : 0;
+       } else {
+               return 0;
+       }
+}
+
+uint64_t
+ml_cpu_cache_sharing(unsigned int level)
+{
+       x86_cpu_cache_t *cachep;
+
+       if (level == 0) {
+               return machine_info.max_cpus;
+       } else if ( 1 <= level && level <= 3) {
+               cachep = current_cpu_datap()->lcpu.caches[level-1];
+               return cachep ? cachep->nlcpus : 0;
+       } else {
+               return 0;
+       }
+}
+