- va_list listp;
- boolean_t state;
-
- if (!disableSerialOuput) {
-
- /*
- * Spin to get kprintf lock but re-enable interrupts while failing.
- * This allows interrupts to be handled while waiting but
- * interrupts are disabled once we have the lock.
- */
- state = ml_set_interrupts_enabled(FALSE);
- while (!simple_lock_try(&kprintf_lock)) {
- ml_set_interrupts_enabled(state);
- ml_set_interrupts_enabled(FALSE);
- }
+ va_list listp;
+ va_list listp2;
+ boolean_t state;
+ void *caller = __builtin_return_address(0);
+
+ if (!disable_serial_output) {
+ boolean_t early = FALSE;
+ if (rdmsr64(MSR_IA32_GS_BASE) == 0) {
+ early = TRUE;
+ }
+ /* If PE_kputc has not yet been initialized, don't
+ * take any locks, just dump to serial */
+ if (!PE_kputc || early) {
+ va_start(listp, fmt);
+ va_copy(listp2, listp);
+
+ _doprnt_log(fmt, &listp, pal_serial_putc, 16);
+ va_end(listp);
+
+ // If interrupts are enabled
+ if (ml_get_interrupts_enabled()) {
+ os_log_with_args(OS_LOG_DEFAULT, OS_LOG_TYPE_DEFAULT, fmt, listp2, caller);
+ }
+ va_end(listp2);
+ return;
+ }
+
+ /*
+ * Spin to get kprintf lock but poll for incoming signals
+ * while interrupts are masked.
+ */
+ state = ml_set_interrupts_enabled(FALSE);
+
+ pal_preemption_assert();
+
+ while (!simple_lock_try(&kprintf_lock)) {
+ (void) cpu_signal_handler(NULL);
+ }
+
+ if (cpu_number() != cpu_last_locked) {
+ MP_DEBUG_KPRINTF("[cpu%d...]\n", cpu_number());
+ cpu_last_locked = cpu_number();
+ }
+
+ va_start(listp, fmt);
+ va_copy(listp2, listp);
+ _doprnt(fmt, &listp, PE_kputc, 16);
+ va_end(listp);
+
+ simple_unlock(&kprintf_lock);
+ ml_set_interrupts_enabled(state);
+
+ // If interrupts are enabled
+ if (ml_get_interrupts_enabled()) {
+ os_log_with_args(OS_LOG_DEFAULT, OS_LOG_TYPE_DEFAULT, fmt, listp2, caller);
+ }
+ va_end(listp2);