2 * Copyright (c) 2000-2012 Apple Inc. All rights reserved.
4 * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
6 * This file contains Original Code and/or Modifications of Original Code
7 * as defined in and that are subject to the Apple Public Source License
8 * Version 2.0 (the 'License'). You may not use this file except in
9 * compliance with the License. The rights granted to you under the License
10 * may not be used to create, or enable the creation or redistribution of,
11 * unlawful or unlicensed copies of an Apple operating system, or to
12 * circumvent, violate, or enable the circumvention or violation of, any
13 * terms of an Apple operating system software license agreement.
15 * Please obtain a copy of the License at
16 * http://www.opensource.apple.com/apsl/ and read it before using this file.
18 * The Original Code and all software distributed under the License are
19 * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
20 * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
21 * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
22 * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
23 * Please see the License for the specific language governing rights and
24 * limitations under the License.
26 * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
32 * Mach Operating System
33 * Copyright (c) 1991,1990 Carnegie Mellon University
34 * All Rights Reserved.
36 * Permission to use, copy, modify and distribute this software and its
37 * documentation is hereby granted, provided that both the copyright
38 * notice and this permission notice appear in all copies of the
39 * software, derivative works or modified versions, and any portions
40 * thereof, and that both notices appear in supporting documentation.
42 * CARNEGIE MELLON ALLOWS FREE USE OF THIS SOFTWARE IN ITS "AS IS"
43 * CONDITION. CARNEGIE MELLON DISCLAIMS ANY LIABILITY OF ANY KIND FOR
44 * ANY DAMAGES WHATSOEVER RESULTING FROM THE USE OF THIS SOFTWARE.
46 * Carnegie Mellon requests users of this software to return to
48 * Software Distribution Coordinator or Software.Distribution@CS.CMU.EDU
49 * School of Computer Science
50 * Carnegie Mellon University
51 * Pittsburgh PA 15213-3890
53 * any improvements or extensions that they make and grant Carnegie Mellon
54 * the rights to redistribute these changes.
66 #include <i386/apic.h>
67 #include <i386/mp_events.h>
69 #define MAX_CPUS 64 /* 8 * sizeof(cpumask_t) */
73 #include <sys/cdefs.h>
74 #include <mach/boolean.h>
75 #include <mach/kern_return.h>
76 #include <mach/i386/thread_status.h>
77 #include <mach/vm_types.h>
78 #include <kern/simple_lock.h>
82 extern kern_return_t
intel_startCPU(int slot_num
);
83 extern kern_return_t
intel_startCPU_fast(int slot_num
);
84 extern void i386_init_slave(void);
85 extern void i386_init_slave_fast(void);
86 extern void smp_init(void);
88 extern void cpu_interrupt(int cpu
);
91 extern unsigned int real_ncpus
; /* real number of cpus */
92 extern unsigned int max_ncpus
; /* max number of cpus */
93 decl_simple_lock_data(extern,kdb_lock
) /* kdb lock */
97 extern void console_init(void);
98 extern void *console_cpu_alloc(boolean_t boot_cpu
);
99 extern void console_cpu_free(void *console_buf
);
101 extern int kdb_cpu
; /* current cpu running kdb */
102 extern int kdb_debug
;
103 extern int kdb_active
[];
105 extern volatile boolean_t mp_kdp_trap
;
106 extern volatile boolean_t mp_kdp_is_NMI
;
107 extern volatile boolean_t force_immediate_debugger_NMI
;
108 extern volatile boolean_t pmap_tlb_flush_timeout
;
109 extern volatile usimple_lock_t spinlock_timed_out
;
110 extern volatile uint32_t spinlock_owner_cpu
;
111 extern uint32_t spinlock_timeout_NMI(uintptr_t thread_addr
);
113 extern uint64_t LastDebuggerEntryAllowance
;
115 extern void mp_kdp_enter(boolean_t proceed_on_failure
);
116 extern void mp_kdp_exit(void);
118 extern boolean_t
mp_recent_debugger_activity(void);
119 extern void kernel_spin(uint64_t spin_ns
);
122 * All cpu rendezvous:
124 extern void mp_rendezvous(
125 void (*setup_func
)(void *),
126 void (*action_func
)(void *),
127 void (*teardown_func
)(void *),
129 extern void mp_rendezvous_no_intrs(
130 void (*action_func
)(void *),
132 extern void mp_rendezvous_break_lock(void);
136 * Called from thread context, this blocks until all active cpus have
139 extern void mp_broadcast(
140 void (*action_func
)(void *),
143 typedef long (*kdp_x86_xcpu_func_t
) (void *arg0
, void *arg1
, uint16_t lcpu
);
145 extern long kdp_x86_xcpu_invoke(const uint16_t lcpu
,
146 kdp_x86_xcpu_func_t func
,
147 void *arg0
, void *arg1
);
148 typedef enum {KDP_XCPU_NONE
= 0xffff, KDP_CURRENT_LCPU
= 0xfffe} kdp_cpu_t
;
151 typedef uint32_t cpu_t
;
152 typedef volatile uint64_t cpumask_t
;
153 static inline cpumask_t
154 cpu_to_cpumask(cpu_t cpu
)
156 return (cpu
< MAX_CPUS
) ? (1ULL << cpu
) : 0;
158 #define CPUMASK_ALL 0xffffffffffffffffULL
159 #define CPUMASK_SELF cpu_to_cpumask(cpu_number())
160 #define CPUMASK_OTHERS (CPUMASK_ALL & ~CPUMASK_SELF)
162 /* Initialation routing called at processor registration */
163 extern void mp_cpus_call_cpu_init(int cpu
);
166 * Invoke a function (possibly NULL) on a set of cpus specified by a mask.
167 * The mask may include the local cpu.
169 * - ASYNC: other cpus make their calls in parallel
170 * - SYNC: the calls are performed serially in logical cpu order
171 * - NOSYNC: the calls are queued
172 * Unless the mode is NOSYNC, mp_cpus_call() returns when the function has been
173 * called on all specified cpus.
174 * The return value is the number of cpus where the call was made or queued.
175 * The action function is called with interrupts disabled.
177 extern cpu_t
mp_cpus_call(
180 void (*action_func
)(void *),
182 extern cpu_t
mp_cpus_call1(
185 void (*action_func
)(void *, void*),
188 cpumask_t
*cpus_calledp
);
197 extern void NMIPI_panic(cpumask_t cpus
, NMI_reason_t reason
);
199 /* Interrupt a set of cpus, forcing an exit out of non-root mode */
200 extern void mp_cpus_kick(cpumask_t cpus
);
202 * Power-management-specific SPI to:
203 * - register a callout function, and
204 * - request the callout (if registered) on a given cpu.
206 extern void PM_interrupt_register(void (*fn
)(void));
207 extern void cpu_PM_interrupt(int cpu
);
216 } cpu_signal_event_t
;
218 #define LOG_NENTRIES 100
220 uint64_t count
[MP_LAST
];
222 cpu_signal_event_t entry
[LOG_NENTRIES
];
223 } cpu_signal_event_log_t
;
225 extern cpu_signal_event_log_t
*cpu_signal
[];
226 extern cpu_signal_event_log_t
*cpu_handle
[];
228 #define DBGLOG(log,_cpu,_event) { \
229 boolean_t spl = ml_set_interrupts_enabled(FALSE); \
230 cpu_signal_event_log_t *logp = log[cpu_number()]; \
231 int next = logp->next_entry; \
232 cpu_signal_event_t *eventp = &logp->entry[next]; \
234 logp->count[_event]++; \
236 eventp->time = rdtsc64(); \
237 eventp->cpu = _cpu; \
238 eventp->event = _event; \
239 if (next == (LOG_NENTRIES - 1)) \
240 logp->next_entry = 0; \
242 logp->next_entry++; \
244 (void) ml_set_interrupts_enabled(spl); \
247 #define DBGLOG_CPU_INIT(cpu) { \
248 cpu_signal_event_log_t **sig_logpp = &cpu_signal[cpu]; \
249 cpu_signal_event_log_t **hdl_logpp = &cpu_handle[cpu]; \
251 if (*sig_logpp == NULL && \
252 kmem_alloc(kernel_map, \
253 (vm_offset_t *) sig_logpp, \
254 sizeof(cpu_signal_event_log_t)) != KERN_SUCCESS)\
255 panic("DBGLOG_CPU_INIT cpu_signal allocation failed\n");\
256 bzero(*sig_logpp, sizeof(cpu_signal_event_log_t)); \
257 if (*hdl_logpp == NULL && \
258 kmem_alloc(kernel_map, \
259 (vm_offset_t *) hdl_logpp, \
260 sizeof(cpu_signal_event_log_t)) != KERN_SUCCESS)\
261 panic("DBGLOG_CPU_INIT cpu_handle allocation failed\n");\
262 bzero(*hdl_logpp, sizeof(cpu_signal_event_log_t)); \
265 #define DBGLOG(log,_cpu,_event)
266 #define DBGLOG_CPU_INIT(cpu)
267 #endif /* MP_DEBUG */
269 #endif /* ASSEMBLER */
272 #define i_bit(bit, word) ((long)(*(word)) & (1L << (bit)))
274 __attribute__((always_inline
)) static inline long
275 i_bit_impl(long word
, long bit
) {
276 long bitmask
= 1L << bit
;
277 return word
& bitmask
;
279 #define i_bit(bit, word) i_bit_impl((long)(*(word)), bit)
283 #endif /* _I386_MP_H_ */
285 #endif /* KERNEL_PRIVATE */