+}
+#define MP_DEBUG_KPRINTF(x...) _kprintf(x)
+#else /* MP_DEBUG */
+#define MP_DEBUG_KPRINTF(x...)
+#endif /* MP_DEBUG */
+
+static int cpu_last_locked = 0;
+
+__attribute__((noinline,not_tail_called))
+void kprintf(const char *fmt, ...)
+{
+ va_list listp;
+ va_list listp2;
+ boolean_t state;
+ void *caller = __builtin_return_address(0);
+
+ if (!disable_serial_output) {
+ boolean_t early = FALSE;
+ if (rdmsr64(MSR_IA32_GS_BASE) == 0) {
+ early = TRUE;
+ }
+ /* If PE_kputc has not yet been initialized, don't
+ * take any locks, just dump to serial */
+ if (!PE_kputc || early) {
+ va_start(listp, fmt);
+ va_copy(listp2, listp);
+
+ _doprnt_log(fmt, &listp, pal_serial_putc, 16);
+ va_end(listp);
+
+ // If interrupts are enabled
+ if (ml_get_interrupts_enabled()) {
+ os_log_with_args(OS_LOG_DEFAULT, OS_LOG_TYPE_DEFAULT, fmt, listp2, caller);
+ }
+ va_end(listp2);
+ return;
+ }
+
+ /*
+ * Spin to get kprintf lock but poll for incoming signals
+ * while interrupts are masked.
+ */
+ state = ml_set_interrupts_enabled(FALSE);
+
+ pal_preemption_assert();
+
+ while (!simple_lock_try(&kprintf_lock)) {
+ (void) cpu_signal_handler(NULL);
+ }
+
+ if (cpu_number() != cpu_last_locked) {
+ MP_DEBUG_KPRINTF("[cpu%d...]\n", cpu_number());
+ cpu_last_locked = cpu_number();
+ }
+
+ va_start(listp, fmt);
+ va_copy(listp2, listp);
+ _doprnt(fmt, &listp, PE_kputc, 16);
+ va_end(listp);
+
+ simple_unlock(&kprintf_lock);
+ ml_set_interrupts_enabled(state);
+
+ // If interrupts are enabled
+ if (ml_get_interrupts_enabled()) {
+ os_log_with_args(OS_LOG_DEFAULT, OS_LOG_TYPE_DEFAULT, fmt, listp2, caller);
+ }
+ va_end(listp2);
+
+ }
+ else {
+ if (ml_get_interrupts_enabled()) {
+ va_start(listp, fmt);
+ os_log_with_args(OS_LOG_DEFAULT, OS_LOG_TYPE_DEFAULT, fmt, listp, caller);
+ va_end(listp);
+ }
+ }
+}
+
+
+
+extern void kprintf_break_lock(void);
+void
+kprintf_break_lock(void)
+{
+ simple_lock_init(&kprintf_lock, 0);