]> git.saurik.com Git - apple/xnu.git/blobdiff - osfmk/arm64/cswitch.s
xnu-7195.101.1.tar.gz
[apple/xnu.git] / osfmk / arm64 / cswitch.s
index e3a0cb317dc8e819e7279c2f21cd7fcca72965e8..1e0e214e66ec3bec6f95fe6a9783262a24d1c85b 100644 (file)
@@ -27,6 +27,7 @@
  */
 #include <machine/asm.h>
 #include <arm64/machine_machdep.h>
+#include <arm64/machine_routines_asm.h>
 #include <arm64/proc_reg.h>
 #include "assym.s"
 
  * A subroutine invocation must preserve the contents of the registers r19-r29
  * and SP. We also save IP0 and IP1, as machine_idle uses IP0 for saving the LR.
  */
-       stp             x16, x17, [$0, SS64_X16]
-       stp             x19, x20, [$0, SS64_X19]
-       stp             x21, x22, [$0, SS64_X21]
-       stp             x23, x24, [$0, SS64_X23]
-       stp             x25, x26, [$0, SS64_X25]
-       stp             x27, x28, [$0, SS64_X27]
-       stp             fp, lr, [$0, SS64_FP]
-       mov             $1, sp
-       str             $1, [$0, SS64_SP]
+       stp             x16, x17, [$0, SS64_KERNEL_X16]
+       stp             x19, x20, [$0, SS64_KERNEL_X19]
+       stp             x21, x22, [$0, SS64_KERNEL_X21]
+       stp             x23, x24, [$0, SS64_KERNEL_X23]
+       stp             x25, x26, [$0, SS64_KERNEL_X25]
+       stp             x27, x28, [$0, SS64_KERNEL_X27]
+       stp             fp, lr, [$0, SS64_KERNEL_FP]
+       str             xzr, [$0, SS64_KERNEL_PC]
+       MOV32   w$1, PSR64_KERNEL_POISON
+       str             w$1, [$0, SS64_KERNEL_CPSR]     
+#ifdef HAS_APPLE_PAC
+       stp             x0, x1, [sp, #-16]!
+       stp             x2, x3, [sp, #-16]!
+       stp             x4, x5, [sp, #-16]!
+
+       /*
+        * Arg0: The ARM context pointer
+        * Arg1: PC value to sign
+        * Arg2: CPSR value to sign
+        * Arg3: LR to sign
+        */
+       mov             x0, $0
+       mov             x1, #0
+       mov             w2, w$1
+       mov             x3, lr
+       mov             x4, x16
+       mov             x5, x17
+       bl              EXT(ml_sign_kernel_thread_state)
+
+       ldp             x4, x5, [sp], #16
+       ldp             x2, x3, [sp], #16
+       ldp             x0, x1, [sp], #16
+       ldp             fp, lr, [$0, SS64_KERNEL_FP]
+#endif /* defined(HAS_APPLE_PAC) */
+       mov             x$1, sp
+       str             x$1, [$0, SS64_KERNEL_SP]
 
 /* AAPCS-64 Page 14
  *
  * calls; the remaining registers (v0-v7, v16-v31) do not need to be preserved
  * (or should be preserved by the caller).
  */
-       str             d8,     [$0, NS64_D8]
-       str             d9,     [$0, NS64_D9]
-       str             d10,[$0, NS64_D10]
-       str             d11,[$0, NS64_D11]
-       str             d12,[$0, NS64_D12]
-       str             d13,[$0, NS64_D13]
-       str             d14,[$0, NS64_D14]
-       str             d15,[$0, NS64_D15]
+       str             d8,     [$0, NS64_KERNEL_D8]
+       str             d9,     [$0, NS64_KERNEL_D9]
+       str             d10,[$0, NS64_KERNEL_D10]
+       str             d11,[$0, NS64_KERNEL_D11]
+       str             d12,[$0, NS64_KERNEL_D12]
+       str             d13,[$0, NS64_KERNEL_D13]
+       str             d14,[$0, NS64_KERNEL_D14]
+       str             d15,[$0, NS64_KERNEL_D15]
+
+       mrs             x$1, FPCR
+       str             w$1, [$0, NS64_KERNEL_FPCR]
 .endmacro
 
 /*
  *   arg1 - Scratch register
  */
 .macro load_general_registers
-       ldp             x16, x17, [$0, SS64_X16]
-       ldp             x19, x20, [$0, SS64_X19]
-       ldp             x21, x22, [$0, SS64_X21]
-       ldp             x23, x24, [$0, SS64_X23]
-       ldp             x25, x26, [$0, SS64_X25]
-       ldp             x27, x28, [$0, SS64_X27]
-       ldp             fp, lr, [$0, SS64_FP]
-       ldr             $1, [$0, SS64_SP]
-       mov             sp, $1
-
-       ldr             d8,     [$0, NS64_D8]
-       ldr             d9,     [$0, NS64_D9]
-       ldr             d10,[$0, NS64_D10]
-       ldr             d11,[$0, NS64_D11]
-       ldr             d12,[$0, NS64_D12]
-       ldr             d13,[$0, NS64_D13]
-       ldr             d14,[$0, NS64_D14]
-       ldr             d15,[$0, NS64_D15]
+       mov             x20, x0
+       mov             x21, x1
+       mov             x22, x2
+
+       mov             x0, $0
+       AUTH_KERNEL_THREAD_STATE_IN_X0  x23, x24, x25, x26, x27
+
+       mov             x0, x20
+       mov             x1, x21
+       mov             x2, x22
+
+       ldr             w$1, [$0, NS64_KERNEL_FPCR]
+       mrs             x19, FPCR
+       CMSR FPCR, x19, x$1, 1
+1:
+
+       // Skip x16, x17 - already loaded + authed by AUTH_THREAD_STATE_IN_X0
+       ldp             x19, x20, [$0, SS64_KERNEL_X19]
+       ldp             x21, x22, [$0, SS64_KERNEL_X21]
+       ldp             x23, x24, [$0, SS64_KERNEL_X23]
+       ldp             x25, x26, [$0, SS64_KERNEL_X25]
+       ldp             x27, x28, [$0, SS64_KERNEL_X27]
+       ldr             fp, [$0, SS64_KERNEL_FP]
+       // Skip lr - already loaded + authed by AUTH_THREAD_STATE_IN_X0
+       ldr             x$1, [$0, SS64_KERNEL_SP]
+       mov             sp, x$1
+
+       ldr             d8,     [$0, NS64_KERNEL_D8]
+       ldr             d9,     [$0, NS64_KERNEL_D9]
+       ldr             d10,[$0, NS64_KERNEL_D10]
+       ldr             d11,[$0, NS64_KERNEL_D11]
+       ldr             d12,[$0, NS64_KERNEL_D12]
+       ldr             d13,[$0, NS64_KERNEL_D13]
+       ldr             d14,[$0, NS64_KERNEL_D14]
+       ldr             d15,[$0, NS64_KERNEL_D15]
 .endmacro
 
+
 /*
  * set_thread_registers
  *
  */
 .macro set_thread_registers
        msr             TPIDR_EL1, $0                                           // Write new thread pointer to TPIDR_EL1
+       ldr             $1, [$0, ACT_CPUDATAP]
+       str             $0, [$1, CPU_ACTIVE_THREAD]
        ldr             $1, [$0, TH_CTH_SELF]                           // Get cthread pointer
        mrs             $2, TPIDRRO_EL0                                         // Extract cpu number from TPIDRRO_EL0
        and             $2, $2, #(MACHDEP_CPUNUM_MASK)
        orr             $2, $1, $2                                                      // Save new cthread/cpu to TPIDRRO_EL0
        msr             TPIDRRO_EL0, $2
-       ldr             $1, [$0, TH_CTH_DATA]                           // Get new cthread data pointer
-       msr             TPIDR_EL0, $1                                           // Save data pointer to TPIDRRW_EL0
-       /* ARM64_TODO Reserve x18 until we decide what to do with it */
-       mov             x18, $1                                                         // ... and trash reserved x18
+       msr             TPIDR_EL0, xzr
+#if DEBUG || DEVELOPMENT
+       ldr             $1, [$0, TH_THREAD_ID]                          // Save the bottom 32-bits of the thread ID into
+       msr             CONTEXTIDR_EL1, $1                                      // CONTEXTIDR_EL1 (top 32-bits are RES0).
+#endif /* DEBUG || DEVELOPMENT */
 .endmacro
 
+/*
+ * set_process_dependent_keys_and_sync_context
+ *
+ * Updates process dependent keys and issues explicit context sync during context switch if necessary
+ *  Per CPU Data rop_key is initialized in arm_init() for bootstrap processor
+ *  and in cpu_data_init for slave processors
+ *
+ *  thread - New thread pointer
+ *  new_key - Scratch register: New Thread Key
+ *  tmp_key - Scratch register: Current CPU Key
+ *  cpudatap - Scratch register: Current CPU Data pointer
+ *  wsync - Half-width scratch register: CPU sync required flag
+ *
+ *  to save on ISBs, for ARMv8.5 we use the CPU_SYNC_ON_CSWITCH field, cached in wsync, for pre-ARMv8.5,
+ *  we just use wsync to keep track of needing an ISB
+ */
+.macro set_process_dependent_keys_and_sync_context     thread, new_key, tmp_key, cpudatap, wsync
+
+
+#if defined(__ARM_ARCH_8_5__) || defined(HAS_APPLE_PAC)
+       ldr             \cpudatap, [\thread, ACT_CPUDATAP]
+#endif /* defined(__ARM_ARCH_8_5__) || defined(HAS_APPLE_PAC) */
+
+#if defined(__ARM_ARCH_8_5__)
+       ldrb    \wsync, [\cpudatap, CPU_SYNC_ON_CSWITCH]
+#else /* defined(__ARM_ARCH_8_5__) */
+       mov             \wsync, #0
+#endif
+
+
+#if CSWITCH_ROP_KEYS
+       ldr             \new_key, [\thread, TH_ROP_PID]
+       REPROGRAM_ROP_KEYS      Lskip_rop_keys_\@, \new_key, \cpudatap, \tmp_key
+       mov             \wsync, #1
+Lskip_rop_keys_\@:
+#endif /* CSWITCH_ROP_KEYS */
+
+#if CSWITCH_JOP_KEYS
+       ldr             \new_key, [\thread, TH_JOP_PID]
+       REPROGRAM_JOP_KEYS      Lskip_jop_keys_\@, \new_key, \cpudatap, \tmp_key
+       mov             \wsync, #1
+Lskip_jop_keys_\@:
+#endif /* CSWITCH_JOP_KEYS */
+
+       cbz             \wsync, 1f
+       isb     sy
+
+#if defined(__ARM_ARCH_8_5__)
+       strb    wzr, [\cpudatap, CPU_SYNC_ON_CSWITCH]
+#endif
+1:
+.endmacro
 
 /*
  * void     machine_load_context(thread_t        thread)
 LEXT(machine_load_context)
        set_thread_registers    x0, x1, x2
        ldr             x1, [x0, TH_KSTACKPTR]                          // Get top of kernel stack
-       load_general_registers  x1, x2
-       mov             x0, xzr                                                         // Clear argument to thread_continue
+       load_general_registers  x1, 2
+       set_process_dependent_keys_and_sync_context     x0, x1, x2, x3, w4
+       mov             x0, #0                                                          // Clear argument to thread_continue
        ret
 
 /*
- *     void Call_continuation( void (*continuation)(void), 
- *                             void *param, 
- *                             wait_result_t wresult, 
- *                             vm_offset_t stack_ptr)
+ *  typedef void (*thread_continue_t)(void *param, wait_result_t)
+ *
+ *     void Call_continuation( thread_continue_t continuation,
+ *                                     void *param,
+ *                                         wait_result_t wresult,
+ *                          bool enable interrupts)
  */
        .text
        .align  5
@@ -153,12 +260,27 @@ LEXT(Call_continuation)
        /* ARM64_TODO arm loads the kstack top instead of arg4. What should we use? */
        ldr             x5, [x4, TH_KSTACKPTR]                          // Get the top of the kernel stack
        mov             sp, x5                                                          // Set stack pointer
+       mov             fp, #0                                                          // Clear the frame pointer
+
+       set_process_dependent_keys_and_sync_context     x4, x5, x6, x7, w20
+
+       mov x20, x0  //continuation
+       mov x21, x1  //continuation parameter
+       mov x22, x2  //wait result
+
+       cbz x3, 1f
+       mov x0, #1
+       bl EXT(ml_set_interrupts_enabled)
+1:
 
-       mov             fp, xzr                                                         // Clear the frame pointer
-       mov             x4, x0                                                          // Load the continuation
-       mov             x0, x1                                                          // Set the first parameter
-       mov             x1, x2                                                          // Set the wait result arg
-       blr             x4                                                                      // Branch to the continuation
+       mov             x0, x21                                                         // Set the first parameter
+       mov             x1, x22                                                         // Set the wait result arg
+#ifdef HAS_APPLE_PAC
+       mov             x21, THREAD_CONTINUE_T_DISC
+       blraa   x20, x21                                                        // Branch to the continuation
+#else
+       blr             x20                                                                     // Branch to the continuation
+#endif
        mrs             x0, TPIDR_EL1                                           // Get the current thread pointer
        b               EXT(thread_terminate)                           // Kill the thread
 
@@ -175,11 +297,12 @@ LEXT(Call_continuation)
 LEXT(Switch_context)
        cbnz    x1, Lswitch_threads                                     // Skip saving old state if blocking on continuation
        ldr             x3, [x0, TH_KSTACKPTR]                          // Get the old kernel stack top
-       save_general_registers  x3, x4
+       save_general_registers  x3, 4
 Lswitch_threads:
        set_thread_registers    x2, x3, x4
        ldr             x3, [x2, TH_KSTACKPTR]
-       load_general_registers  x3, x4
+       load_general_registers  x3, 4
+       set_process_dependent_keys_and_sync_context     x2, x3, x4, x5, w6
        ret
 
 /*
@@ -193,14 +316,13 @@ Lswitch_threads:
 LEXT(Shutdown_context)
        mrs             x10, TPIDR_EL1                                                  // Get thread pointer
        ldr             x11, [x10, TH_KSTACKPTR]                                // Get the top of the kernel stack
-       save_general_registers  x11, x12
+       save_general_registers  x11, 12
        msr             DAIFSet, #(DAIFSC_FIQF | DAIFSC_IRQF)   // Disable interrupts
        ldr             x11, [x10, ACT_CPUDATAP]                                // Get current cpu
        ldr             x12, [x11, CPU_ISTACKPTR]                               // Switch to interrupt stack
        mov             sp, x12
        b               EXT(cpu_doshutdown)
 
-
 /*
  *     thread_t Idle_context(void)
  *
@@ -212,7 +334,7 @@ LEXT(Shutdown_context)
 LEXT(Idle_context)
        mrs             x0, TPIDR_EL1                                           // Get thread pointer
        ldr             x1, [x0, TH_KSTACKPTR]                          // Get the top of the kernel stack
-       save_general_registers  x1, x2
+       save_general_registers  x1, 2
        ldr             x1, [x0, ACT_CPUDATAP]                          // Get current cpu
        ldr             x2, [x1, CPU_ISTACKPTR]                         // Switch to interrupt stack
        mov             sp, x2
@@ -229,7 +351,8 @@ LEXT(Idle_context)
 LEXT(Idle_load_context)
        mrs             x0, TPIDR_EL1                                           // Get thread pointer
        ldr             x1, [x0, TH_KSTACKPTR]                          // Get the top of the kernel stack
-       load_general_registers  x1, x2
+       load_general_registers  x1, 2
+       set_process_dependent_keys_and_sync_context     x0, x1, x2, x3, w4
        ret
 
        .align  2
@@ -237,3 +360,6 @@ LEXT(Idle_load_context)
 LEXT(machine_set_current_thread)
        set_thread_registers x0, x1, x2
        ret
+
+
+/* vim: set ts=4: */