]> git.saurik.com Git - apple/xnu.git/blobdiff - osfmk/arm/model_dep.c
xnu-7195.101.1.tar.gz
[apple/xnu.git] / osfmk / arm / model_dep.c
index ab930d549eabf449726ef891c5a679bd308df194..053fe3a94633506f9ba1cd8782833e069cc86fc7 100644 (file)
@@ -1,5 +1,5 @@
 /*
- * Copyright (c) 2007-2013 Apple Inc. All rights reserved.
+ * Copyright (c) 2007-2020 Apple Inc. All rights reserved.
  *
  * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
  *
@@ -40,6 +40,8 @@
 #include <pexpert/boot.h>
 #include <pexpert/pexpert.h>
 
+#include <ptrauth.h>
+
 #include <kern/misc_protos.h>
 #include <kern/startup.h>
 #include <kern/clock.h>
@@ -57,6 +59,7 @@
 #include <kdp/kdp_dyld.h>
 #include <kdp/kdp_internal.h>
 #include <uuid/uuid.h>
+#include <sys/codesign.h>
 #include <sys/time.h>
 
 #include <IOKit/IOPlatformExpert.h>
 #include <vm/pmap.h>
 #include <vm/vm_shared_region.h>
 #include <mach/time_value.h>
-#include <machine/machparam.h> /* for btop */
+#include <machine/machparam.h>  /* for btop */
 
 #include <console/video_console.h>
+#include <console/serial_protos.h>
 #include <arm/cpu_data.h>
 #include <arm/cpu_data_internal.h>
 #include <arm/cpu_internal.h>
 #include <kern/kern_cdata.h>
 
 #if     MACH_KDP
-void   kdp_trap(unsigned int, struct arm_saved_state *);
+void    kdp_trap(unsigned int, struct arm_saved_state *);
 #endif
 
-extern kern_return_t   do_stackshot(void *);
-extern void            kdp_snapshot_preflight(int pid, void *tracebuf,
-                                              uint32_t tracebuf_size, uint32_t flags,
-                                              kcdata_descriptor_t data_p,
-                                               boolean_t enable_faulting);
-extern int             kdp_stack_snapshot_bytes_traced(void);
+extern kern_return_t    do_stackshot(void *);
+extern void                    kdp_snapshot_preflight(int pid, void * tracebuf,
+    uint32_t tracebuf_size, uint64_t flags,
+    kcdata_descriptor_t data_p,
+    uint64_t since_timestamp, uint32_t pagetable_mask);
+extern int              kdp_stack_snapshot_bytes_traced(void);
+extern int              kdp_stack_snapshot_bytes_uncompressed(void);
+
+#if INTERRUPT_MASKED_DEBUG
+extern boolean_t interrupt_masked_debug;
+#endif
 
 /*
  * Increment the PANICLOG_VERSION if you change the format of the panic
  * log in any way.
  */
-#define PANICLOG_VERSION 8
+#define PANICLOG_VERSION 13
 static struct kcdata_descriptor kc_panic_data;
 
 extern char                 firmware_version[];
-extern volatile uint32_t       debug_enabled;
+extern volatile uint32_t        debug_enabled;
 extern unsigned int         not_in_kdp;
 
-extern int                             copyinframe(vm_address_t fp, uint32_t * frame);
-extern void                            kdp_callouts(kdp_event_t event);
+extern int                              copyinframe(vm_address_t fp, uint32_t * frame);
+extern void                             kdp_callouts(kdp_event_t event);
 
 /* #include <sys/proc.h> */
 #define MAXCOMLEN 16
-extern int                             proc_pid(void *p);
-extern void                            proc_name_kdp(task_t, char *, int);
+struct proc;
+extern int                              proc_pid(struct proc *p);
+extern void                     proc_name_kdp(task_t, char *, int);
+
+/*
+ * Make sure there's enough space to include the relevant bits in the format required
+ * within the space allocated for the panic version string in the panic header.
+ * The format required by OSAnalytics/DumpPanic is 'Product Version (OS Version)'.
+ */
+#define PANIC_HEADER_VERSION_FMT_STR "%.14s (%.14s)"
+
+extern const char version[];
+extern char       osversion[];
+extern char       osproductversion[];
+extern char       osreleasetype[];
+
+#if defined(XNU_TARGET_OS_BRIDGE)
+extern char     macosproductversion[];
+extern char     macosversion[];
+#endif
 
-extern const char              version[];
-extern char                            osversion[];
 extern uint8_t          gPlatformECID[8];
 extern uint32_t         gPlatformMemoryID;
 
-extern uint64_t                last_hwaccess_thread;
+extern uint64_t         last_hwaccess_thread;
 
-/*Choosing the size for gTargetTypeBuffer as 8 and size for gModelTypeBuffer as 32
-  since the target name and model name typically  doesn't exceed this size */
-extern char  gTargetTypeBuffer[8];
+/*Choosing the size for gTargetTypeBuffer as 16 and size for gModelTypeBuffer as 32
 since the target name and model name typically  doesn't exceed this size */
+extern char  gTargetTypeBuffer[16];
 extern char  gModelTypeBuffer[32];
 
-decl_simple_lock_data(extern,clock_lock)
-extern struct timeval   gIOLastSleepTime;
-extern struct timeval   gIOLastWakeTime;
-extern boolean_t                is_clock_configured;
+decl_simple_lock_data(extern, clock_lock);
+extern struct timeval    gIOLastSleepTime;
+extern struct timeval    gIOLastWakeTime;
+extern boolean_t                 is_clock_configured;
+extern boolean_t kernelcache_uuid_valid;
 extern uuid_t kernelcache_uuid;
 
+extern void stackshot_memcpy(void *dst, const void *src, size_t len);
+
 /* Definitions for frame pointers */
 #define FP_ALIGNMENT_MASK      ((uint32_t)(0x3))
 #define FP_LR_OFFSET           ((uint32_t)4)
@@ -135,7 +163,7 @@ extern uuid_t kernelcache_uuid;
 #define DEBUG_ACK_TIMEOUT ((uint64_t) 10000000)
 
 /* Forward functions definitions */
-void panic_display_times(void) ;
+void panic_display_times(void);
 void panic_print_symbol_name(vm_address_t search);
 
 
@@ -146,23 +174,34 @@ boolean_t             force_immediate_debug_halt = FALSE;
 unsigned int          debug_ack_timeout_count = 0;
 volatile unsigned int debugger_sync = 0;
 volatile unsigned int mp_kdp_trap = 0; /* CPUs signalled by the debug CPU will spin on this */
+volatile unsigned int debug_cpus_spinning = 0; /* Number of signalled CPUs still spinning on mp_kdp_trap (in DebuggerXCall). */
 unsigned int          DebugContextCount = 0;
-// Convenient macros to easily validate one or more pointers if 
+
+#if defined(__arm64__)
+uint8_t PE_smc_stashed_x86_system_state = 0xFF;
+uint8_t PE_smc_stashed_x86_power_state = 0xFF;
+uint8_t PE_smc_stashed_x86_efi_boot_state = 0xFF;
+uint8_t PE_smc_stashed_x86_shutdown_cause = 0xFF;
+uint64_t PE_smc_stashed_x86_prev_power_transitions = UINT64_MAX;
+uint32_t PE_pcie_stashed_link_state = UINT32_MAX;
+#endif
+
+
+// Convenient macros to easily validate one or more pointers if
 // they have defined types
 #define VALIDATE_PTR(ptr) \
        validate_ptr((vm_offset_t)(ptr), sizeof(*(ptr)), #ptr)
 
 #define VALIDATE_PTR_2(ptr0, ptr1) \
-       VALIDATE_PTR(ptr0) && VALIDATE_PTR(ptr1) 
-       
+       VALIDATE_PTR(ptr0) && VALIDATE_PTR(ptr1)
+
 #define VALIDATE_PTR_3(ptr0, ptr1, ptr2) \
        VALIDATE_PTR_2(ptr0, ptr1) && VALIDATE_PTR(ptr2)
 
 #define VALIDATE_PTR_4(ptr0, ptr1, ptr2, ptr3) \
        VALIDATE_PTR_2(ptr0, ptr1) && VALIDATE_PTR_2(ptr2, ptr3)
 
-#define GET_MACRO(_1,_2,_3,_4,NAME,...) NAME
+#define GET_MACRO(_1, _2, _3, _4, NAME, ...) NAME
 
 #define VALIDATE_PTR_LIST(...) GET_MACRO(__VA_ARGS__, VALIDATE_PTR_4, VALIDATE_PTR_3, VALIDATE_PTR_2, VALIDATE_PTR)(__VA_ARGS__)
 
@@ -170,7 +209,8 @@ unsigned int          DebugContextCount = 0;
  * Evaluate if a pointer is valid
  * Print a message if pointer is invalid
  */
-static boolean_t validate_ptr(
+static boolean_t
+validate_ptr(
        vm_offset_t ptr, vm_size_t size, const char * ptr_name)
 {
        if (ptr) {
@@ -178,7 +218,7 @@ static boolean_t validate_ptr(
                        return TRUE;
                } else {
                        paniclog_append_noflush("Invalid %s pointer: %p size: %d\n",
-                               ptr_name, (void *)ptr, (int)size);
+                           ptr_name, (void *)ptr, (int)size);
                        return FALSE;
                }
        } else {
@@ -192,32 +232,37 @@ static boolean_t validate_ptr(
  */
 static void
 print_one_backtrace(pmap_t pmap, vm_offset_t topfp, const char *cur_marker,
-       boolean_t is_64_bit)
+    boolean_t is_64_bit, boolean_t print_kexts_in_backtrace)
 {
-       int                 i = 0;
-       addr64_t        lr;
-       addr64_t        fp;
-       addr64_t        fp_for_ppn;
-       ppnum_t         ppn;
-       boolean_t       dump_kernel_stack;
+       int                 i = 0;
+       addr64_t        lr;
+       addr64_t        fp;
+       addr64_t        fp_for_ppn;
+       ppnum_t         ppn;
+       boolean_t       dump_kernel_stack;
+       vm_offset_t     raddrs[FP_MAX_NUM_TO_EVALUATE];
 
        fp = topfp;
        fp_for_ppn = 0;
        ppn = (ppnum_t)NULL;
 
-       if (fp >= VM_MIN_KERNEL_ADDRESS)
+       if (fp >= VM_MIN_KERNEL_ADDRESS) {
                dump_kernel_stack = TRUE;
-       else
+       } else {
                dump_kernel_stack = FALSE;
+       }
 
        do {
-               if ((fp == 0) || ((fp & FP_ALIGNMENT_MASK) != 0))
+               if ((fp == 0) || ((fp & FP_ALIGNMENT_MASK) != 0)) {
                        break;
-               if (dump_kernel_stack && ((fp < VM_MIN_KERNEL_ADDRESS) || (fp > VM_MAX_KERNEL_ADDRESS)))
+               }
+               if (dump_kernel_stack && ((fp < VM_MIN_KERNEL_ADDRESS) || (fp > VM_MAX_KERNEL_ADDRESS))) {
                        break;
-               if ((!dump_kernel_stack) && (fp >=VM_MIN_KERNEL_ADDRESS))
+               }
+               if ((!dump_kernel_stack) && (fp >= VM_MIN_KERNEL_ADDRESS)) {
                        break;
-                       
+               }
+
                /*
                 * Check to see if current address will result in a different
                 * ppn than previously computed (to avoid recomputation) via
@@ -230,6 +275,10 @@ print_one_backtrace(pmap_t pmap, vm_offset_t topfp, const char *cur_marker,
                if (ppn != (ppnum_t)NULL) {
                        if (is_64_bit) {
                                lr = ml_phys_read_double_64(((((vm_offset_t)ppn) << PAGE_SHIFT)) | ((fp + FP_LR_OFFSET64) & PAGE_MASK));
+#if defined(HAS_APPLE_PAC)
+                               /* return addresses on stack will be signed by arm64e ABI */
+                               lr = (addr64_t) ptrauth_strip((void *)lr, ptrauth_key_return_address);
+#endif
                        } else {
                                lr = ml_phys_read_word(((((vm_offset_t)ppn) << PAGE_SHIFT)) | ((fp + FP_LR_OFFSET) & PAGE_MASK));
                        }
@@ -266,8 +315,13 @@ print_one_backtrace(pmap_t pmap, vm_offset_t topfp, const char *cur_marker,
                        } else {
                                paniclog_append_noflush("%s\t  lr: 0x%08x  fp: 0x%08x\n", cur_marker, (uint32_t)lr, (uint32_t)fp);
                        }
+                       raddrs[i] = lr;
                }
        } while ((++i < FP_MAX_NUM_TO_EVALUATE) && (fp != topfp));
+
+       if (print_kexts_in_backtrace && i != 0) {
+               kmod_panic_dump(&raddrs[0], i);
+       }
 }
 
 #define SANE_TASK_LIMIT 256
@@ -277,66 +331,171 @@ print_one_backtrace(pmap_t pmap, vm_offset_t topfp, const char *cur_marker,
 extern void panic_print_vnodes(void);
 
 static void
-do_print_all_backtraces(
-       const char      *message)
+panic_display_hung_cpus_help(void)
+{
+#if defined(__arm64__)
+       const uint32_t pcsr_offset = 0x90;
+
+       /*
+        * Print some info that might help in cases where nothing
+        * else does
+        */
+       const ml_topology_info_t *info = ml_get_topology_info();
+       if (info) {
+               unsigned i, retry;
+
+               for (i = 0; i < info->num_cpus; i++) {
+                       if (info->cpus[i].cpu_UTTDBG_regs) {
+                               volatile uint64_t *pcsr = (volatile uint64_t*)(info->cpus[i].cpu_UTTDBG_regs + pcsr_offset);
+                               volatile uint32_t *pcsrTrigger = (volatile uint32_t*)pcsr;
+                               uint64_t pc = 0;
+
+                               // a number of retries are needed till this works
+                               for (retry = 1024; retry && !pc; retry--) {
+                                       //a 32-bit read is required to make a PC sample be produced, else we'll only get a zero
+                                       (void)*pcsrTrigger;
+                                       pc = *pcsr;
+                               }
+
+                               //postprocessing (same as astris does)
+                               if (pc >> 48) {
+                                       pc |= 0xffff000000000000ull;
+                               }
+                               paniclog_append_noflush("CORE %u recently retired instr at 0x%016llx\n", i, pc);
+                       }
+               }
+       }
+#endif //defined(__arm64__)
+}
+
+static void
+do_print_all_backtraces(const char *message, uint64_t panic_options)
 {
-       int             logversion = PANICLOG_VERSION;
+       int             logversion = PANICLOG_VERSION;
        thread_t        cur_thread = current_thread();
-       uintptr_t       cur_fp;
+       uintptr_t       cur_fp;
        task_t          task;
-       int             i;
-       size_t          index;
        int             print_vnodes = 0;
-       const char *nohilite_thread_marker="\t";
+       const char *nohilite_thread_marker = "\t";
 
        /* end_marker_bytes set to 200 for printing END marker + stackshot summary info always */
        int bytes_traced = 0, bytes_remaining = 0, end_marker_bytes = 200;
+       int bytes_uncompressed = 0;
        uint64_t bytes_used = 0ULL;
        int err = 0;
        char *stackshot_begin_loc = NULL;
+       kc_format_t kc_format;
+       bool filesetKC = false;
 
 #if defined(__arm__)
-       __asm__         volatile("mov %0, r7":"=r"(cur_fp));
+       __asm__         volatile ("mov %0, r7":"=r"(cur_fp));
 #elif defined(__arm64__)
-       __asm__         volatile("add %0, xzr, fp":"=r"(cur_fp));
+       __asm__         volatile ("add %0, xzr, fp":"=r"(cur_fp));
 #else
 #error Unknown architecture.
 #endif
-       if (panic_bt_depth != 0)
+       if (panic_bt_depth != 0) {
                return;
+       }
        panic_bt_depth++;
 
-       /* Truncate panic string to 1200 bytes -- WDT log can be ~1100 bytes */
+       __unused bool result = PE_get_primary_kc_format(&kc_format);
+       assert(result == true);
+       filesetKC = kc_format == KCFormatFileset;
+
+       /* Truncate panic string to 1200 bytes */
        paniclog_append_noflush("Debugger message: %.1200s\n", message);
        if (debug_enabled) {
                paniclog_append_noflush("Device: %s\n",
-                       ('\0' != gTargetTypeBuffer[0]) ? gTargetTypeBuffer : "Not set yet");
+                   ('\0' != gTargetTypeBuffer[0]) ? gTargetTypeBuffer : "Not set yet");
                paniclog_append_noflush("Hardware Model: %s\n",
-                       ('\0' != gModelTypeBuffer[0]) ? gModelTypeBuffer:"Not set yet");
+                   ('\0' != gModelTypeBuffer[0]) ? gModelTypeBuffer:"Not set yet");
                paniclog_append_noflush("ECID: %02X%02X%02X%02X%02X%02X%02X%02X\n", gPlatformECID[7],
-                       gPlatformECID[6], gPlatformECID[5], gPlatformECID[4], gPlatformECID[3],
-                       gPlatformECID[2], gPlatformECID[1], gPlatformECID[0]);
+                   gPlatformECID[6], gPlatformECID[5], gPlatformECID[4], gPlatformECID[3],
+                   gPlatformECID[2], gPlatformECID[1], gPlatformECID[0]);
                if (last_hwaccess_thread) {
                        paniclog_append_noflush("AppleHWAccess Thread: 0x%llx\n", last_hwaccess_thread);
                }
+               paniclog_append_noflush("Boot args: %s\n", PE_boot_args());
        }
        paniclog_append_noflush("Memory ID: 0x%x\n", gPlatformMemoryID);
+       paniclog_append_noflush("OS release type: %.256s\n",
+           ('\0' != osreleasetype[0]) ? osreleasetype : "Not set yet");
        paniclog_append_noflush("OS version: %.256s\n",
-                       ('\0' != osversion[0]) ? osversion : "Not set yet");
+           ('\0' != osversion[0]) ? osversion : "Not set yet");
+#if defined(XNU_TARGET_OS_BRIDGE)
+       paniclog_append_noflush("macOS version: %.256s\n",
+           ('\0' != macosversion[0]) ? macosversion : "Not set");
+#endif
        paniclog_append_noflush("Kernel version: %.512s\n", version);
-       paniclog_append_noflush("KernelCache UUID: ");
-       for (index = 0; index < sizeof(uuid_t); index++) {
-               paniclog_append_noflush("%02X", kernelcache_uuid[index]);
+
+       if (kernelcache_uuid_valid) {
+               if (filesetKC) {
+                       paniclog_append_noflush("Fileset Kernelcache UUID: ");
+               } else {
+                       paniclog_append_noflush("KernelCache UUID: ");
+               }
+               for (size_t index = 0; index < sizeof(uuid_t); index++) {
+                       paniclog_append_noflush("%02X", kernelcache_uuid[index]);
+               }
+               paniclog_append_noflush("\n");
        }
-       paniclog_append_noflush("\n");
+       panic_display_kernel_uuid();
 
        paniclog_append_noflush("iBoot version: %.128s\n", firmware_version);
        paniclog_append_noflush("secure boot?: %s\n", debug_enabled ? "NO": "YES");
+#if defined(XNU_TARGET_OS_BRIDGE)
+       paniclog_append_noflush("x86 EFI Boot State: ");
+       if (PE_smc_stashed_x86_efi_boot_state != 0xFF) {
+               paniclog_append_noflush("0x%x\n", PE_smc_stashed_x86_efi_boot_state);
+       } else {
+               paniclog_append_noflush("not available\n");
+       }
+       paniclog_append_noflush("x86 System State: ");
+       if (PE_smc_stashed_x86_system_state != 0xFF) {
+               paniclog_append_noflush("0x%x\n", PE_smc_stashed_x86_system_state);
+       } else {
+               paniclog_append_noflush("not available\n");
+       }
+       paniclog_append_noflush("x86 Power State: ");
+       if (PE_smc_stashed_x86_power_state != 0xFF) {
+               paniclog_append_noflush("0x%x\n", PE_smc_stashed_x86_power_state);
+       } else {
+               paniclog_append_noflush("not available\n");
+       }
+       paniclog_append_noflush("x86 Shutdown Cause: ");
+       if (PE_smc_stashed_x86_shutdown_cause != 0xFF) {
+               paniclog_append_noflush("0x%x\n", PE_smc_stashed_x86_shutdown_cause);
+       } else {
+               paniclog_append_noflush("not available\n");
+       }
+       paniclog_append_noflush("x86 Previous Power Transitions: ");
+       if (PE_smc_stashed_x86_prev_power_transitions != UINT64_MAX) {
+               paniclog_append_noflush("0x%llx\n", PE_smc_stashed_x86_prev_power_transitions);
+       } else {
+               paniclog_append_noflush("not available\n");
+       }
+       paniclog_append_noflush("PCIeUp link state: ");
+       if (PE_pcie_stashed_link_state != UINT32_MAX) {
+               paniclog_append_noflush("0x%x\n", PE_pcie_stashed_link_state);
+       } else {
+               paniclog_append_noflush("not available\n");
+       }
+#endif
+       if (panic_data_buffers != NULL) {
+               paniclog_append_noflush("%s data: ", panic_data_buffers->producer_name);
+               uint8_t *panic_buffer_data = (uint8_t *) panic_data_buffers->buf;
+               for (int i = 0; i < panic_data_buffers->len; i++) {
+                       paniclog_append_noflush("%02X", panic_buffer_data[i]);
+               }
+               paniclog_append_noflush("\n");
+       }
        paniclog_append_noflush("Paniclog version: %d\n", logversion);
 
        panic_display_kernel_aslr();
        panic_display_times();
        panic_display_zprint();
+       panic_display_hung_cpus_help();
 #if CONFIG_ZLEAKS
        panic_display_ztrace();
 #endif /* CONFIG_ZLEAKS */
@@ -344,26 +503,32 @@ do_print_all_backtraces(
        panic_display_ecc_errors();
 #endif /* CONFIG_ECC_LOGGING */
 
-       // Just print threads with high CPU usage for WDT timeouts
-       if (strncmp(message, "WDT timeout", 11) == 0) {
-               thread_t        top_runnable[5] = {0};
-               thread_t        thread;
-               int                     total_cpu_usage = 0;
+#if DEVELOPMENT || DEBUG
+       if (cs_debug_unsigned_exec_failures != 0 || cs_debug_unsigned_mmap_failures != 0) {
+               paniclog_append_noflush("Unsigned code exec failures: %u\n", cs_debug_unsigned_exec_failures);
+               paniclog_append_noflush("Unsigned code mmap failures: %u\n", cs_debug_unsigned_mmap_failures);
+       }
+#endif
+
+       // Highlight threads that used high amounts of CPU in the panic log if requested (historically requested for watchdog panics)
+       if (panic_options & DEBUGGER_OPTION_PRINT_CPU_USAGE_PANICLOG) {
+               thread_t        top_runnable[5] = {0};
+               thread_t        thread;
+               int                     total_cpu_usage = 0;
 
                print_vnodes = 1;
 
-       
+
                for (thread = (thread_t)queue_first(&threads);
-             VALIDATE_PTR(thread) && !queue_end(&threads, (queue_entry_t)thread);
-             thread = (thread_t)queue_next(&thread->threads)) {
-                                                        
+                   VALIDATE_PTR(thread) && !queue_end(&threads, (queue_entry_t)thread);
+                   thread = (thread_t)queue_next(&thread->threads)) {
                        total_cpu_usage += thread->cpu_usage;
-                                               
+
                        // Look for the 5 runnable threads with highest priority
                        if (thread->state & TH_RUN) {
-                               int                     k;
-                               thread_t        comparison_thread = thread;
-                               
+                               int                     k;
+                               thread_t        comparison_thread = thread;
+
                                for (k = 0; k < TOP_RUNNABLE_LIMIT; k++) {
                                        if (top_runnable[k] == 0) {
                                                top_runnable[k] = comparison_thread;
@@ -376,36 +541,33 @@ do_print_all_backtraces(
                                } // loop through highest priority runnable threads
                        } // Check if thread is runnable
                } // Loop through all threads
-               
+
                // Print the relevant info for each thread identified
                paniclog_append_noflush("Total cpu_usage: %d\n", total_cpu_usage);
                paniclog_append_noflush("Thread task pri cpu_usage\n");
 
-               for (i = 0; i < TOP_RUNNABLE_LIMIT; i++) {                      
-
+               for (int i = 0; i < TOP_RUNNABLE_LIMIT; i++) {
                        if (top_runnable[i] && VALIDATE_PTR(top_runnable[i]->task) &&
-                               validate_ptr((vm_offset_t)top_runnable[i]->task->bsd_info, 1, "bsd_info")) {
-                               
+                           validate_ptr((vm_offset_t)top_runnable[i]->task->bsd_info, 1, "bsd_info")) {
                                char            name[MAXCOMLEN + 1];
                                proc_name_kdp(top_runnable[i]->task, name, sizeof(name));
                                paniclog_append_noflush("%p %s %d %d\n",
-                                       top_runnable[i], name, top_runnable[i]->sched_pri, top_runnable[i]->cpu_usage);
-                       } 
+                                   top_runnable[i], name, top_runnable[i]->sched_pri, top_runnable[i]->cpu_usage);
+                       }
                } // Loop through highest priority runnable threads
                paniclog_append_noflush("\n");
-       } // Check if message is "WDT timeout"
+       }
 
-    // print current task info
+       // print current task info
        if (VALIDATE_PTR_LIST(cur_thread, cur_thread->task)) {
-
                task = cur_thread->task;
 
                if (VALIDATE_PTR_LIST(task->map, task->map->pmap)) {
                        paniclog_append_noflush("Panicked task %p: %d pages, %d threads: ",
-                               task, task->map->pmap->stats.resident_count, task->thread_count);
+                           task, task->map->pmap->stats.resident_count, task->thread_count);
                } else {
                        paniclog_append_noflush("Panicked task %p: %d threads: ",
-                               task, task->thread_count);
+                           task, task->thread_count);
                }
 
                if (validate_ptr((vm_offset_t)task->bsd_info, 1, "bsd_info")) {
@@ -422,19 +584,34 @@ do_print_all_backtraces(
 
        if (cur_fp < VM_MAX_KERNEL_ADDRESS) {
                paniclog_append_noflush("Panicked thread: %p, backtrace: 0x%llx, tid: %llu\n",
-                       cur_thread, (addr64_t)cur_fp, thread_tid(cur_thread));
+                   cur_thread, (addr64_t)cur_fp, thread_tid(cur_thread));
 #if __LP64__
-               print_one_backtrace(kernel_pmap, cur_fp, nohilite_thread_marker, TRUE);
+               print_one_backtrace(kernel_pmap, cur_fp, nohilite_thread_marker, TRUE, filesetKC);
 #else
-               print_one_backtrace(kernel_pmap, cur_fp, nohilite_thread_marker, FALSE);
+               print_one_backtrace(kernel_pmap, cur_fp, nohilite_thread_marker, FALSE, filesetKC);
 #endif
        } else {
                paniclog_append_noflush("Could not print panicked thread backtrace:"
-                          "frame pointer outside kernel vm.\n");
+                   "frame pointer outside kernel vm.\n");
        }
 
        paniclog_append_noflush("\n");
+       if (filesetKC) {
+               kext_dump_panic_lists(&paniclog_append_noflush);
+               paniclog_append_noflush("\n");
+       }
        panic_info->eph_panic_log_len = PE_get_offset_into_panic_region(debug_buf_ptr) - panic_info->eph_panic_log_offset;
+       /* set the os version data in the panic header in the format 'Product Version (OS Version)' (only if they have been set) */
+       if ((osversion[0] != '\0') && (osproductversion[0] != '\0')) {
+               snprintf((char *)&panic_info->eph_os_version, sizeof(panic_info->eph_os_version), PANIC_HEADER_VERSION_FMT_STR,
+                   osproductversion, osversion);
+       }
+#if defined(XNU_TARGET_OS_BRIDGE)
+       if ((macosversion[0] != '\0') && (macosproductversion[0] != '\0')) {
+               snprintf((char *)&panic_info->eph_macos_version, sizeof(panic_info->eph_macos_version), PANIC_HEADER_VERSION_FMT_STR,
+                   macosproductversion, macosversion);
+       }
+#endif
 
        if (debug_ack_timeout_count) {
                panic_info->eph_panic_flags |= EMBEDDED_PANIC_HEADER_FLAG_STACKSHOT_FAILED_DEBUGGERSYNC;
@@ -451,13 +628,25 @@ do_print_all_backtraces(
 
                bytes_remaining = debug_buf_size - (unsigned int)((uintptr_t)stackshot_begin_loc - (uintptr_t)debug_buf_base);
                err = kcdata_memory_static_init(&kc_panic_data, (mach_vm_address_t)debug_buf_ptr,
-                                                                               KCDATA_BUFFER_BEGIN_STACKSHOT, bytes_remaining - end_marker_bytes,
-                                                                               KCFLAG_USE_MEMCOPY);
+                   KCDATA_BUFFER_BEGIN_COMPRESSED, bytes_remaining - end_marker_bytes,
+                   KCFLAG_USE_MEMCOPY);
                if (err == KERN_SUCCESS) {
+                       uint64_t stackshot_flags = (STACKSHOT_GET_GLOBAL_MEM_STATS | STACKSHOT_SAVE_LOADINFO | STACKSHOT_KCDATA_FORMAT |
+                           STACKSHOT_ENABLE_BT_FAULTING | STACKSHOT_ENABLE_UUID_FAULTING | STACKSHOT_FROM_PANIC | STACKSHOT_DO_COMPRESS |
+                           STACKSHOT_DISABLE_LATENCY_INFO | STACKSHOT_NO_IO_STATS | STACKSHOT_THREAD_WAITINFO | STACKSHOT_GET_DQ |
+                           STACKSHOT_COLLECT_SHAREDCACHE_LAYOUT);
+
+                       err = kcdata_init_compress(&kc_panic_data, KCDATA_BUFFER_BEGIN_STACKSHOT, stackshot_memcpy, KCDCT_ZLIB);
+                       if (err != KERN_SUCCESS) {
+                               panic_info->eph_panic_flags |= EMBEDDED_PANIC_HEADER_FLAG_COMPRESS_FAILED;
+                               stackshot_flags &= ~STACKSHOT_DO_COMPRESS;
+                       }
+                       if (filesetKC) {
+                               stackshot_flags |= STACKSHOT_SAVE_KEXT_LOADINFO;
+                       }
+
                        kdp_snapshot_preflight(-1, stackshot_begin_loc, bytes_remaining - end_marker_bytes,
-                                                                  (STACKSHOT_GET_GLOBAL_MEM_STATS | STACKSHOT_SAVE_LOADINFO | STACKSHOT_KCDATA_FORMAT |
-                                                                       STACKSHOT_ENABLE_BT_FAULTING | STACKSHOT_ENABLE_UUID_FAULTING | STACKSHOT_FROM_PANIC |
-                                                                       STACKSHOT_NO_IO_STATS | STACKSHOT_THREAD_WAITINFO), &kc_panic_data, 0);
+                           stackshot_flags, &kc_panic_data, 0, 0);
                        err = do_stackshot(NULL);
                        bytes_traced = kdp_stack_snapshot_bytes_traced();
                        if (bytes_traced > 0 && !err) {
@@ -467,7 +656,13 @@ do_print_all_backtraces(
                                panic_info->eph_stackshot_len = bytes_traced;
 
                                panic_info->eph_other_log_offset = PE_get_offset_into_panic_region(debug_buf_ptr);
-                               paniclog_append_noflush("\n** Stackshot Succeeded ** Bytes Traced %d **\n", bytes_traced);
+                               if (stackshot_flags & STACKSHOT_DO_COMPRESS) {
+                                       panic_info->eph_panic_flags |= EMBEDDED_PANIC_HEADER_FLAG_STACKSHOT_DATA_COMPRESSED;
+                                       bytes_uncompressed = kdp_stack_snapshot_bytes_uncompressed();
+                                       paniclog_append_noflush("\n** Stackshot Succeeded ** Bytes Traced %d (Uncompressed %d) **\n", bytes_traced, bytes_uncompressed);
+                               } else {
+                                       paniclog_append_noflush("\n** Stackshot Succeeded ** Bytes Traced %d **\n", bytes_traced);
+                               }
                        } else {
                                bytes_used = kcdata_memory_get_used_bytes(&kc_panic_data);
                                if (bytes_used > 0) {
@@ -494,8 +689,9 @@ do_print_all_backtraces(
 
        assert(panic_info->eph_other_log_offset != 0);
 
-       if (print_vnodes != 0)
+       if (print_vnodes != 0) {
                panic_print_vnodes();
+       }
 
        panic_bt_depth--;
 }
@@ -504,7 +700,7 @@ do_print_all_backtraces(
  * Entry to print_all_backtraces is serialized by the debugger lock
  */
 static void
-print_all_backtraces(const char        *message)
+print_all_backtraces(const char *message, uint64_t panic_options)
 {
        unsigned int initial_not_in_kdp = not_in_kdp;
 
@@ -519,7 +715,7 @@ print_all_backtraces(const char     *message)
         * not_in_kdp.
         */
        not_in_kdp = 0;
-       do_print_all_backtraces(message);
+       do_print_all_backtraces(message, panic_options);
 
        not_in_kdp = initial_not_in_kdp;
 
@@ -534,15 +730,16 @@ panic_display_times()
                return;
        }
 
-       if ((is_clock_configured) && (simple_lock_try(&clock_lock))) {
-               clock_sec_t     secs, boot_secs;
-               clock_usec_t    usecs, boot_usecs;
+       if ((is_clock_configured) && (simple_lock_try(&clock_lock, LCK_GRP_NULL))) {
+               clock_sec_t     secs, boot_secs;
+               clock_usec_t    usecs, boot_usecs;
 
                simple_unlock(&clock_lock);
 
                clock_get_calendar_microtime(&secs, &usecs);
                clock_get_boottime_microtime(&boot_secs, &boot_usecs);
 
+               paniclog_append_noflush("mach_absolute_time: 0x%llx\n", mach_absolute_time());
                paniclog_append_noflush("Epoch Time:        sec       usec\n");
                paniclog_append_noflush("  Boot    : 0x%08x 0x%08x\n", (unsigned int)boot_secs, (unsigned int)boot_usecs);
                paniclog_append_noflush("  Sleep   : 0x%08x 0x%08x\n", (unsigned int)gIOLastSleepTime.tv_sec, (unsigned int)gIOLastSleepTime.tv_usec);
@@ -551,7 +748,8 @@ panic_display_times()
        }
 }
 
-void panic_print_symbol_name(vm_address_t search)
+void
+panic_print_symbol_name(vm_address_t search)
 {
 #pragma unused(search)
        // empty stub. Really only used on x86_64.
@@ -560,11 +758,20 @@ void panic_print_symbol_name(vm_address_t search)
 
 void
 SavePanicInfo(
-       const char *message, __unused uint64_t panic_options)
+       const char *message, __unused void *panic_data, uint64_t panic_options)
 {
-
-       /* This should be initialized by the time we get here */
-       assert(panic_info->eph_panic_log_offset != 0);
+       /*
+        * This should be initialized by the time we get here, but
+        * if it is not, asserting about it will be of no use (it will
+        * come right back to here), so just loop right here and now.
+        * This prevents early-boot panics from becoming recursive and
+        * thus makes them easier to debug. If you attached to a device
+        * and see your PC here, look down a few frames to see your
+        * early-boot panic there.
+        */
+       while (!panic_info || panic_info->eph_panic_log_offset == 0) {
+               ;
+       }
 
        if (panic_options & DEBUGGER_OPTION_PANICLOGANDREBOOT) {
                panic_info->eph_panic_flags  |= EMBEDDED_PANIC_HEADER_FLAG_BUTTON_RESET_PANIC;
@@ -574,6 +781,12 @@ SavePanicInfo(
                panic_info->eph_panic_flags |= EMBEDDED_PANIC_HEADER_FLAG_COPROC_INITIATED_PANIC;
        }
 
+#if defined(XNU_TARGET_OS_BRIDGE)
+       panic_info->eph_x86_power_state = PE_smc_stashed_x86_power_state;
+       panic_info->eph_x86_efi_boot_state = PE_smc_stashed_x86_efi_boot_state;
+       panic_info->eph_x86_system_state = PE_smc_stashed_x86_system_state;
+#endif
+
        /*
         * On newer targets, panic data is stored directly into the iBoot panic region.
         * If we re-enter SavePanicInfo (e.g. on a double panic) on such a target, update the
@@ -585,12 +798,13 @@ SavePanicInfo(
                PE_sync_panic_buffers(); // extra precaution; panic path likely isn't reliable if we're here
        }
 
-       if (PanicInfoSaved || (debug_buf_size == 0))
+       if (PanicInfoSaved || (debug_buf_size == 0)) {
                return;
+       }
 
        PanicInfoSaved = TRUE;
 
-       print_all_backtraces(message);
+       print_all_backtraces(message, panic_options);
 
        assert(panic_info->eph_panic_log_len != 0);
        panic_info->eph_other_log_len = PE_get_offset_into_panic_region(debug_buf_ptr) - panic_info->eph_other_log_offset;
@@ -611,8 +825,9 @@ paniclog_flush()
        unsigned int panicbuf_length = 0;
 
        panicbuf_length = (unsigned int)(debug_buf_ptr - gPanicBase);
-       if (!panicbuf_length)
+       if (!panicbuf_length) {
                return;
+       }
 
        /*
         * Updates the log length of the last part of the panic log.
@@ -634,6 +849,20 @@ paniclog_flush()
        PE_sync_panic_buffers();
 }
 
+/*
+ * @function _was_in_userspace
+ *
+ * @abstract Unused function used to indicate that a CPU was in userspace
+ * before it was IPI'd to enter the Debugger context.
+ *
+ * @discussion This function should never actually be called.
+ */
+static void __attribute__((__noreturn__))
+_was_in_userspace(void)
+{
+       panic("%s: should not have been invoked.", __FUNCTION__);
+}
+
 /*
  * @function DebuggerXCallEnter
  *
@@ -657,13 +886,14 @@ DebuggerXCallEnter(
        uint64_t max_mabs_time, current_mabs_time;
        int cpu;
        int max_cpu;
-       cpu_data_t      *target_cpu_datap;
-       cpu_data_t      *cpu_data_ptr = getCpuDatap();
+       cpu_data_t      *target_cpu_datap;
+       cpu_data_t      *cpu_data_ptr = getCpuDatap();
 
        /* Check for nested debugger entry. */
        cpu_data_ptr->debugger_active++;
-       if (cpu_data_ptr->debugger_active != 1)
+       if (cpu_data_ptr->debugger_active != 1) {
                return KERN_SUCCESS;
+       }
 
        /*
         * If debugger_sync is not 0, someone responded excessively late to the last
@@ -674,6 +904,7 @@ DebuggerXCallEnter(
 
        debugger_sync = 0;
        mp_kdp_trap = 1;
+       debug_cpus_spinning = 0;
 
        /*
         * We need a barrier here to ensure CPUs see mp_kdp_trap and spin when responding
@@ -690,18 +921,21 @@ DebuggerXCallEnter(
        max_cpu = ml_get_max_cpu_number();
 
        boolean_t immediate_halt = FALSE;
-       if (proceed_on_sync_failure && force_immediate_debug_halt)
-               immediate_halt = TRUE; 
+       if (proceed_on_sync_failure && force_immediate_debug_halt) {
+               immediate_halt = TRUE;
+       }
 
        if (!immediate_halt) {
-               for (cpu=0; cpu <= max_cpu; cpu++) {
+               for (cpu = 0; cpu <= max_cpu; cpu++) {
                        target_cpu_datap = (cpu_data_t *)CpuDataEntries[cpu].cpu_data_vaddr;
 
-                       if ((target_cpu_datap == NULL) || (target_cpu_datap == cpu_data_ptr))
+                       if ((target_cpu_datap == NULL) || (target_cpu_datap == cpu_data_ptr)) {
                                continue;
+                       }
 
-                       if(KERN_SUCCESS == cpu_signal(target_cpu_datap, SIGPdebug, (void *)NULL, NULL)) {
-                               (void)hw_atomic_add(&debugger_sync, 1);
+                       if (KERN_SUCCESS == cpu_signal(target_cpu_datap, SIGPdebug, (void *)NULL, NULL)) {
+                               os_atomic_inc(&debugger_sync, relaxed);
+                               os_atomic_inc(&debug_cpus_spinning, relaxed);
                        } else {
                                cpu_signal_failed = true;
                                kprintf("cpu_signal failed in DebuggerXCallEnter\n");
@@ -720,9 +954,9 @@ DebuggerXCallEnter(
                 * all other CPUs have either responded or are spinning in a context that is
                 * debugger safe.
                 */
-               while ((debugger_sync != 0) && (current_mabs_time < max_mabs_time))
+               while ((debugger_sync != 0) && (current_mabs_time < max_mabs_time)) {
                        current_mabs_time = mach_absolute_time();
-
+               }
        }
 
        if (cpu_signal_failed && !proceed_on_sync_failure) {
@@ -734,44 +968,52 @@ DebuggerXCallEnter(
                 * but will be sufficient to let the other core respond.
                 */
                __builtin_arm_dmb(DMB_ISH);
-               for (cpu=0; cpu <= max_cpu; cpu++) {
+               for (cpu = 0; cpu <= max_cpu; cpu++) {
                        target_cpu_datap = (cpu_data_t *)CpuDataEntries[cpu].cpu_data_vaddr;
 
-                       if ((target_cpu_datap == NULL) || (target_cpu_datap == cpu_data_ptr))
+                       if ((target_cpu_datap == NULL) || (target_cpu_datap == cpu_data_ptr)) {
                                continue;
-                       if (!(target_cpu_datap->cpu_signal & SIGPdebug) && !immediate_halt)
+                       }
+                       if (!(target_cpu_datap->cpu_signal & SIGPdebug) && !immediate_halt) {
                                continue;
+                       }
                        if (proceed_on_sync_failure) {
                                paniclog_append_noflush("Attempting to forcibly halt cpu %d\n", cpu);
                                dbgwrap_status_t halt_status = ml_dbgwrap_halt_cpu(cpu, 0);
-                               if (halt_status < 0)
-                                       paniclog_append_noflush("Unable to halt cpu %d: %d\n", cpu, halt_status);
-                               else {
-                                       if (halt_status > 0)
-                                               paniclog_append_noflush("cpu %d halted with warning %d\n", cpu, halt_status);
+                               if (halt_status < 0) {
+                                       paniclog_append_noflush("cpu %d failed to halt with error %d: %s\n", cpu, halt_status, ml_dbgwrap_strerror(halt_status));
+                               } else {
+                                       if (halt_status > 0) {
+                                               paniclog_append_noflush("cpu %d halted with warning %d: %s\n", cpu, halt_status, ml_dbgwrap_strerror(halt_status));
+                                       }
                                        target_cpu_datap->halt_status = CPU_HALTED;
                                }
-                       } else
+                       } else {
                                kprintf("Debugger synch pending on cpu %d\n", cpu);
+                       }
                }
                if (proceed_on_sync_failure) {
                        for (cpu = 0; cpu <= max_cpu; cpu++) {
                                target_cpu_datap = (cpu_data_t *)CpuDataEntries[cpu].cpu_data_vaddr;
 
                                if ((target_cpu_datap == NULL) || (target_cpu_datap == cpu_data_ptr) ||
-                                   (target_cpu_datap->halt_status == CPU_NOT_HALTED))
+                                   (target_cpu_datap->halt_status == CPU_NOT_HALTED)) {
                                        continue;
+                               }
                                dbgwrap_status_t halt_status = ml_dbgwrap_halt_cpu_with_state(cpu,
                                    NSEC_PER_SEC, &target_cpu_datap->halt_state);
-                               if ((halt_status < 0) || (halt_status == DBGWRAP_WARN_CPU_OFFLINE))
-                                       paniclog_append_noflush("Unable to obtain state for cpu %d: %d\n", cpu, halt_status);
-                               else
+                               if ((halt_status < 0) || (halt_status == DBGWRAP_WARN_CPU_OFFLINE)) {
+                                       paniclog_append_noflush("Unable to obtain state for cpu %d with status %d: %s\n", cpu, halt_status, ml_dbgwrap_strerror(halt_status));
+                               } else {
+                                       paniclog_append_noflush("cpu %d successfully halted\n", cpu);
                                        target_cpu_datap->halt_status = CPU_HALTED_WITH_STATE;
+                               }
                        }
-                       if (immediate_halt)
+                       if (immediate_halt) {
                                paniclog_append_noflush("Immediate halt requested on all cores\n");
-                       else
+                       } else {
                                paniclog_append_noflush("Debugger synchronization timed out; waited %llu nanoseconds\n", DEBUG_ACK_TIMEOUT);
+                       }
                        debug_ack_timeout_count++;
                        return KERN_SUCCESS;
                } else {
@@ -794,26 +1036,47 @@ void
 DebuggerXCallReturn(
        void)
 {
-       cpu_data_t      *cpu_data_ptr = getCpuDatap();
+       cpu_data_t      *cpu_data_ptr = getCpuDatap();
+       uint64_t max_mabs_time, current_mabs_time;
 
        cpu_data_ptr->debugger_active--;
-       if (cpu_data_ptr->debugger_active != 0)
+       if (cpu_data_ptr->debugger_active != 0) {
                return;
+       }
 
        mp_kdp_trap = 0;
        debugger_sync = 0;
 
+       nanoseconds_to_absolutetime(DEBUG_ACK_TIMEOUT, &max_mabs_time);
+       current_mabs_time = mach_absolute_time();
+       max_mabs_time += current_mabs_time;
+       assert(max_mabs_time > current_mabs_time);
+
+       /*
+        * Wait for other CPUs to stop spinning on mp_kdp_trap (see DebuggerXCall).
+        * It's possible for one or more CPUs to not decrement debug_cpus_spinning,
+        * since they may be stuck somewhere else with interrupts disabled.
+        * Wait for DEBUG_ACK_TIMEOUT ns for a response and move on if we don't get it.
+        *
+        * Note that the same is done in DebuggerXCallEnter, when we wait for other
+        * CPUS to update debugger_sync. If we time out, let's hope for all CPUs to be
+        * spinning in a debugger-safe context
+        */
+       while ((debug_cpus_spinning != 0) && (current_mabs_time < max_mabs_time)) {
+               current_mabs_time = mach_absolute_time();
+       }
+
        /* Do we need a barrier here? */
        __builtin_arm_dmb(DMB_ISH);
 }
 
 void
 DebuggerXCall(
-       void            *ctx)
+       void            *ctx)
 {
-       boolean_t               save_context = FALSE;
-       vm_offset_t             kstackptr = 0;
-       arm_saved_state_t       *regs = (arm_saved_state_t *) ctx;
+       boolean_t               save_context = FALSE;
+       vm_offset_t             kstackptr = 0;
+       arm_saved_state_t       *regs = (arm_saved_state_t *) ctx;
 
        if (regs != NULL) {
 #if defined(__arm64__)
@@ -824,33 +1087,80 @@ DebuggerXCall(
        }
 
        kstackptr = current_thread()->machine.kstackptr;
+
+#if defined(__arm64__)
+       arm_kernel_saved_state_t *state = (arm_kernel_saved_state_t *)kstackptr;
+
+       if (save_context) {
+               /* Save the interrupted context before acknowledging the signal */
+               current_thread()->machine.kpcb = regs;
+       } else if (regs) {
+               /* zero old state so machine_trace_thread knows not to backtrace it */
+               register_t pc = (register_t)ptrauth_strip((void *)&_was_in_userspace, ptrauth_key_function_pointer);
+               state->fp = 0;
+               state->pc = pc;
+               state->lr = 0;
+               state->sp = 0;
+       }
+#else
        arm_saved_state_t *state = (arm_saved_state_t *)kstackptr;
 
        if (save_context) {
                /* Save the interrupted context before acknowledging the signal */
-               *state = *regs;
+               copy_signed_thread_state(state, regs);
        } else if (regs) {
                /* zero old state so machine_trace_thread knows not to backtrace it */
+               register_t pc = (register_t)ptrauth_strip((void *)&_was_in_userspace, ptrauth_key_function_pointer);
                set_saved_state_fp(state, 0);
-               set_saved_state_pc(state, 0);
+               set_saved_state_pc(state, pc);
                set_saved_state_lr(state, 0);
                set_saved_state_sp(state, 0);
        }
+#endif
 
-       (void)hw_atomic_sub(&debugger_sync, 1);
+       /*
+        * When running in serial mode, the core capturing the dump may hold interrupts disabled
+        * for a time longer than the timeout. That path includes logic to reset the timestamp
+        * so that we do not eventually trigger the interrupt timeout assert().
+        *
+        * Here we check whether other cores have already gone over the timeout at this point
+        * before spinning, so we at least cover the IPI reception path. After spinning, however,
+        * we reset the timestamp so as to avoid hitting the interrupt timeout assert().
+        */
+       if ((serialmode & SERIALMODE_OUTPUT) || stackshot_active()) {
+               INTERRUPT_MASKED_DEBUG_END();
+       }
+
+       os_atomic_dec(&debugger_sync, relaxed);
        __builtin_arm_dmb(DMB_ISH);
-       while (mp_kdp_trap);
+       while (mp_kdp_trap) {
+               ;
+       }
+
+       /**
+        * Alert the triggering CPU that this CPU is done spinning. The CPU that
+        * signalled all of the other CPUs will wait (in DebuggerXCallReturn) for
+        * all of the CPUs to exit the above loop before continuing.
+        */
+       os_atomic_dec(&debug_cpus_spinning, relaxed);
+
+       if ((serialmode & SERIALMODE_OUTPUT) || stackshot_active()) {
+               INTERRUPT_MASKED_DEBUG_START(current_thread()->machine.int_handler_addr, current_thread()->machine.int_type);
+       }
+
+#if defined(__arm64__)
+       current_thread()->machine.kpcb = NULL;
+#endif /* defined(__arm64__) */
 
        /* Any cleanup for our pushed context should go here */
 }
 
-
 void
 DebuggerCall(
-       unsigned int    reason,
-       void            *ctx)
+       unsigned int    reason,
+       void            *ctx)
 {
-#if    !MACH_KDP
+#if     !MACH_KDP
 #pragma unused(reason,ctx)
 #endif /* !MACH_KDP */
 
@@ -858,11 +1168,15 @@ DebuggerCall(
        alternate_debugger_enter();
 #endif
 
-#if    MACH_KDP
+#if     MACH_KDP
        kdp_trap(reason, (struct arm_saved_state *)ctx);
 #else
        /* TODO: decide what to do if no debugger config */
 #endif
 }
 
-
+boolean_t
+bootloader_valid_page(ppnum_t ppn)
+{
+       return pmap_bootloader_page(ppn);
+}