]> git.saurik.com Git - apple/xnu.git/blobdiff - osfmk/i386/i386_lock.s
xnu-792.25.20.tar.gz
[apple/xnu.git] / osfmk / i386 / i386_lock.s
index 8f3900f318fb5929439a1922688048fdc22d63bc..838fd0ed871d51f7444a54a580bd18d520efe795 100644 (file)
 /*
  *  Routines for general lock debugging.
  */
-#define        S_TYPE          SLOCK_TYPE(%edx)
-#define        S_PC            SLOCK_PC(%edx)
-#define        S_THREAD        SLOCK_THREAD(%edx)
-#define        S_DURATIONH     SLOCK_DURATIONH(%edx)
-#define        S_DURATIONL     SLOCK_DURATIONL(%edx)
 
 /* 
  * Checks for expected lock types and calls "panic" on
        .text                                           ;       \
 1:
 
-#define        CHECK_SIMPLE_LOCK_TYPE()                                \
-       cmpl    $ USLOCK_TAG,S_TYPE                     ;       \
-       je      1f                                      ;       \
-       pushl   $2f                                     ;       \
-       call    EXT(panic)                              ;       \
-       hlt                                             ;       \
-       .data                                           ;       \
-2:     String  "not a simple lock!"                    ;       \
-       .text                                           ;       \
-1:
-
 /*
  * If one or more simplelocks are currently held by a thread,
  * an attempt to acquire a mutex will cause this check to fail
@@ -261,8 +245,9 @@ LEAF_ENTRY(hw_lock_init)
 LEAF_ENTRY(hw_lock_lock)
        movl    L_ARG0,%edx             /* fetch lock pointer */
 
-       movl    L_PC,%ecx
-1:     DISABLE_PREEMPTION
+       movl    %gs:CPU_ACTIVE_THREAD,%ecx
+       DISABLE_PREEMPTION
+1:
        movl    0(%edx), %eax
        testl   %eax,%eax               /* lock locked? */
        jne     3f                      /* branch if so */
@@ -270,8 +255,7 @@ LEAF_ENTRY(hw_lock_lock)
        jne     3f
        movl    $1,%eax                 /* In case this was a timeout call */
        LEAF_RET                        /* if yes, then nothing left to do */
-
-3:     ENABLE_PREEMPTION               /* no reason we can't be preemptable */
+3:
        PAUSE                           /* pause for hyper-threading */
        jmp     1b                      /* try again */
 
@@ -284,7 +268,7 @@ LEAF_ENTRY(hw_lock_lock)
 LEAF_ENTRY(hw_lock_to)
 1:
        movl    L_ARG0,%edx             /* fetch lock pointer */
-       movl    L_PC,%ecx
+       movl    %gs:CPU_ACTIVE_THREAD,%ecx
        /*
         * Attempt to grab the lock immediately
         * - fastpath without timeout nonsense.
@@ -315,8 +299,6 @@ LEAF_ENTRY(hw_lock_to)
        adcl    $0,%edx                 /* add carry */
        mov     %edx,%ecx
        mov     %eax,%ebx               /* %ecx:%ebx is the timeout expiry */
-3:
-       ENABLE_PREEMPTION               /* no reason not to be preempted now */
 4:
        /*
         * The inner-loop spin to look for the lock being freed.
@@ -337,7 +319,7 @@ LEAF_ENTRY(hw_lock_to)
        cmpl    %ecx,%edx               /* compare high-order 32-bits */
        jb      4b                      /* continue spinning if less, or */
        cmpl    %ebx,%eax               /* compare low-order 32-bits */ 
-       jb      5b                      /* continue if less, else bail */
+       jb      4b                      /* continue if less, else bail */
        xor     %eax,%eax               /* with 0 return value */
        pop     %ebx
        pop     %edi
@@ -348,10 +330,9 @@ LEAF_ENTRY(hw_lock_to)
         * Here to try to grab the lock that now appears to be free
         * after contention.
         */
-       movl    8+L_PC,%edx             /* calling pc (8+ for pushed regs) */
-       DISABLE_PREEMPTION
+       movl    %gs:CPU_ACTIVE_THREAD,%edx
        lock; cmpxchgl  %edx,0(%edi)    /* try to acquire the HW lock */
-       jne     3b                      /* no - spin again */
+       jne     4b                      /* no - spin again */
        movl    $1,%eax                 /* yes */
        pop     %ebx
        pop     %edi
@@ -369,6 +350,26 @@ LEAF_ENTRY(hw_lock_unlock)
        ENABLE_PREEMPTION
        LEAF_RET
 
+/*
+ *     void i386_lock_unlock_with_flush(hw_lock_t)
+ *
+ *     Unconditionally release lock, followed by a cacheline flush of
+ *     the line corresponding to the lock dword. This routine is currently
+ *     used with certain locks which are susceptible to lock starvation,
+ *     minimizing cache affinity for lock acquisitions. A queued spinlock
+ *     or other mechanism that ensures fairness would obviate the need
+ *     for this routine, but ideally few or no spinlocks should exhibit
+ *     enough contention to require such measures.
+ *     MACH_RT:  release preemption level.
+ */
+LEAF_ENTRY(i386_lock_unlock_with_flush)
+       movl    L_ARG0,%edx             /* Fetch lock pointer */
+       movl    $0,0(%edx)              /* Clear the lock */
+       mfence                          /* Serialize prior stores */
+       clflush 0(%edx)                 /* Write back and invalidate line */
+       ENABLE_PREEMPTION
+       LEAF_RET
+
 /*
  *     unsigned int hw_lock_try(hw_lock_t)
  *     MACH_RT:  returns with preemption disabled on success.
@@ -376,7 +377,7 @@ LEAF_ENTRY(hw_lock_unlock)
 LEAF_ENTRY(hw_lock_try)
        movl    L_ARG0,%edx             /* fetch lock pointer */
 
-       movl    L_PC,%ecx
+       movl    %gs:CPU_ACTIVE_THREAD,%ecx
        DISABLE_PREEMPTION
        movl    0(%edx),%eax
        testl   %eax,%eax
@@ -387,7 +388,8 @@ LEAF_ENTRY(hw_lock_try)
        movl    $1,%eax                 /* success */
        LEAF_RET
 
-1:     ENABLE_PREEMPTION               /* failure:  release preemption... */
+1:
+       ENABLE_PREEMPTION               /* failure:  release preemption... */
        xorl    %eax,%eax               /* ...and return failure */
        LEAF_RET
 
@@ -431,23 +433,21 @@ NONLEAF_ENTRY2(mutex_lock,_mutex_lock)
 
        pushf                           /* save interrupt state */
        cli                             /* disable interrupts */
+Lml_retry:
+       movl    %gs:CPU_ACTIVE_THREAD,%ecx
 
-ml_retry:
-       movl    B_PC,%ecx
-
-ml_get_hw:
+Lml_get_hw:
        movl    M_ILK,%eax              /* read interlock */
        testl   %eax,%eax               /* unlocked? */
-       je      1f                      /* yes - attempt to lock it */
-       PAUSE                           /* no  - pause */
-       jmp     ml_get_hw               /* try again */
-1:
+       jne     Lml_ilk_fail            /* no - take the slow path */
+
        lock; cmpxchgl  %ecx,M_ILK      /* atomic compare and exchange */
-       jne     ml_get_hw               /* branch on failure to retry */
+       jne     Lml_get_hw              /* branch on failure to retry */
 
        movl    M_LOCKED,%ecx           /* get lock owner */
        testl   %ecx,%ecx               /* is the mutex locked? */
-       jne     ml_fail                 /* yes, we lose */
+       jne     Lml_fail                /* yes, we lose */
+Lml_acquire:
        movl    %gs:CPU_ACTIVE_THREAD,%ecx
        movl    %ecx,M_LOCKED
 
@@ -457,28 +457,94 @@ ml_get_hw:
        movl    %ecx,M_PC
 #endif
 
+       cmpw    $0,M_WAITERS            /* are there any waiters? */
+       jne     Lml_waiters             /* yes, more work to do */
+Lml_return:
+       xorl    %eax,%eax
+       movl    %eax,M_ILK
+
+       popf                            /* restore interrupt state */
+
+       NONLEAF_RET
+
+Lml_waiters:
        pushl   %edx                    /* save mutex address */
        pushl   %edx
        call    EXT(lck_mtx_lock_acquire)
        addl    $4,%esp
        popl    %edx                    /* restore mutex address */
+       jmp     Lml_return
+
+Lml_ilk_fail:
+       /*
+        * Slow path: call out to do the spinning.
+        */
+       pushl   %edx                    /* lock address */
+       call    EXT(lck_mtx_interlock_spin)
+       popl    %edx                    /* lock pointer */
+       jmp     Lml_retry               /* try again */
 
+Lml_fail:
+       /*
+        n Check if the owner is on another processor and therefore
+        * we should try to spin before blocking.
+        */
+       testl   $(OnProc),ACT_SPF(%ecx)
+       jz      Lml_block
+
+       /*
+        * Here if owner is on another processor:
+        *  - release the interlock
+        *  - spin on the holder until release or timeout
+        *  - in either case re-acquire the interlock
+        *  - if released, acquire it
+        *  - otherwise drop thru to block.
+        */
        xorl    %eax,%eax
-       movl    %eax,M_ILK
+       movl    %eax,M_ILK              /* zero interlock */
+       popf
+       pushf                           /* restore interrupt state */
 
-       popf                            /* restore interrupt state */
+       push    %edx                    /* lock address */
+       call    EXT(lck_mtx_lock_spin)  /* call out to do spinning */
+       addl    $4,%esp
+       movl    B_ARG0,%edx             /* refetch mutex address */
 
-       NONLEAF_RET
+       /* Re-acquire interlock */
+       cli                             /* disable interrupts */
+Lml_reget_retry:
+       movl    %gs:CPU_ACTIVE_THREAD,%ecx
+
+Lml_reget_hw:
+       movl    M_ILK,%eax              /* read interlock */
+       testl   %eax,%eax               /* unlocked? */
+       jne     Lml_ilk_refail          /* no - slow path */
+
+       lock; cmpxchgl  %ecx,M_ILK      /* atomic compare and exchange */
+       jne     Lml_reget_hw            /* branch on failure to retry */
 
-ml_fail:
-ml_block:
+       movl    M_LOCKED,%ecx           /* get lock owner */
+       testl   %ecx,%ecx               /* is the mutex free? */
+       je      Lml_acquire             /* yes, acquire */
+       
+Lml_block:
        CHECK_MYLOCK(M_THREAD)
        pushl   M_LOCKED
        pushl   %edx                    /* push mutex address */
        call    EXT(lck_mtx_lock_wait)  /* wait for the lock */
        addl    $8,%esp
        movl    B_ARG0,%edx             /* refetch mutex address */
-       jmp     ml_retry                /* and try again */
+       cli                             /* ensure interrupts disabled */
+       jmp     Lml_retry               /* and try again */
+
+Lml_ilk_refail:
+       /*
+        * Slow path: call out to do the spinning.
+        */
+       pushl   %edx                    /* lock address */
+       call    EXT(lck_mtx_interlock_spin)
+       popl    %edx                    /* lock pointer */
+       jmp     Lml_reget_retry         /* try again */
 
 NONLEAF_ENTRY2(mutex_try,_mutex_try)   
 
@@ -487,24 +553,22 @@ NONLEAF_ENTRY2(mutex_try,_mutex_try)
        CHECK_MUTEX_TYPE()
        CHECK_NO_SIMPLELOCKS()
 
-       movl    B_PC,%ecx
-
        pushf                           /* save interrupt state */
        cli                             /* disable interrupts */
+Lmt_retry:
+       movl    %gs:CPU_ACTIVE_THREAD,%ecx
 
-mt_get_hw:
+Lmt_get_hw:
        movl    M_ILK,%eax              /* read interlock */
        testl   %eax,%eax               /* unlocked? */
-       je      1f                      /* yes - attempt to lock it */
-       PAUSE                           /* no  - pause */
-       jmp     mt_get_hw               /* try again */
-1:
+       jne     Lmt_ilk_fail            /* no - slow path */
+
        lock; cmpxchgl  %ecx,M_ILK      /* atomic compare and exchange */
-       jne     mt_get_hw               /* branch on failure to retry */
+       jne     Lmt_get_hw              /* branch on failure to retry */
 
        movl    M_LOCKED,%ecx           /* get lock owner */
        testl   %ecx,%ecx               /* is the mutex locked? */
-       jne     mt_fail                 /* yes, we lose */
+       jne     Lmt_fail                /* yes, we lose */
        movl    %gs:CPU_ACTIVE_THREAD,%ecx
        movl    %ecx,M_LOCKED
 
@@ -514,22 +578,35 @@ mt_get_hw:
        movl    %ecx,M_PC
 #endif
 
-       pushl   %edx                    /* save mutex address */
-       pushl   %edx
-       call    EXT(lck_mtx_lock_acquire)
-       addl    $4,%esp
-       popl    %edx                    /* restore mutex address */
-
+       cmpl    $0,M_WAITERS            /* are there any waiters? */
+       jne     Lmt_waiters             /* yes, more work to do */
+Lmt_return:
        xorl    %eax,%eax
        movl    %eax,M_ILK
-
        popf                            /* restore interrupt state */
 
        movl    $1,%eax
 
        NONLEAF_RET
 
-mt_fail:
+Lmt_waiters:
+       pushl   %edx                    /* save mutex address */
+       pushl   %edx
+       call    EXT(lck_mtx_lock_acquire)
+       addl    $4,%esp
+       popl    %edx                    /* restore mutex address */
+       jmp     Lmt_return
+
+Lmt_ilk_fail:
+       /*
+        * Slow path: call out to do the spinning.
+        */
+       pushl   %edx                    /* lock address */
+       call    EXT(lck_mtx_interlock_spin)
+       popl    %edx                    /* lock pointer */
+       jmp     Lmt_retry               /* try again */
+
+Lmt_fail:
        xorl    %eax,%eax
        movl    %eax,M_ILK
 
@@ -545,25 +622,23 @@ NONLEAF_ENTRY(mutex_unlock)
        CHECK_MUTEX_TYPE()
        CHECK_THREAD(M_THREAD)
 
-       movl    B_PC,%ecx
-
        pushf                           /* save interrupt state */
        cli                             /* disable interrupts */
+Lmu_retry:
+       movl    %gs:CPU_ACTIVE_THREAD,%ecx
 
-mu_get_hw:
+Lmu_get_hw:
        movl    M_ILK,%eax              /* read interlock */
        testl   %eax,%eax               /* unlocked? */
-       je      1f                      /* yes - attempt to lock it */
-       PAUSE                           /* no  - pause */
-       jmp     mu_get_hw               /* try again */
-1:
+       jne     Lmu_ilk_fail            /* no - slow path */
+
        lock; cmpxchgl  %ecx,M_ILK      /* atomic compare and exchange */
-       jne     mu_get_hw               /* branch on failure to retry */
+       jne     Lmu_get_hw              /* branch on failure to retry */
 
        cmpw    $0,M_WAITERS            /* are there any waiters? */
-       jne     mu_wakeup               /* yes, more work to do */
+       jne     Lmu_wakeup              /* yes, more work to do */
 
-mu_doit:
+Lmu_doit:
 
 #if    MACH_LDEBUG
        movl    $0,M_THREAD             /* disown thread */
@@ -578,13 +653,22 @@ mu_doit:
 
        NONLEAF_RET
 
-mu_wakeup:
+Lmu_ilk_fail:
+       /*
+        * Slow path: call out to do the spinning.
+        */
+       pushl   %edx                    /* lock address */
+       call    EXT(lck_mtx_interlock_spin)
+       popl    %edx                    /* lock pointer */
+       jmp     Lmu_retry               /* try again */
+
+Lmu_wakeup:
        pushl   M_LOCKED
        pushl   %edx                    /* push mutex address */
        call    EXT(lck_mtx_unlock_wakeup)/* yes, wake a thread */
        addl    $8,%esp
        movl    B_ARG0,%edx             /* restore lock pointer */
-       jmp     mu_doit
+       jmp     Lmu_doit
 
 /*
  * lck_mtx_lock()
@@ -605,40 +689,96 @@ NONLEAF_ENTRY(lck_mtx_lock)
 
        pushf                           /* save interrupt state */
        cli                             /* disable interrupts */
+Llml_retry:
+       movl    %gs:CPU_ACTIVE_THREAD,%ecx
 
-lml_retry:
-       movl    B_PC,%ecx
-
-lml_get_hw:
+Llml_get_hw:
        movl    M_ILK,%eax              /* read interlock */
        testl   %eax,%eax               /* unlocked? */
-       je      1f                      /* yes - attempt to lock it */
-       PAUSE                           /* no  - pause */
-       jmp     lml_get_hw              /* try again */
-1:
+       jne     Llml_ilk_fail           /* no - slow path */
+
        lock; cmpxchgl  %ecx,M_ILK      /* atomic compare and exchange */
-       jne     lml_get_hw              /* branch on failure to retry */
+       jne     Llml_get_hw             /* branch on failure to retry */
 
        movl    M_LOCKED,%ecx           /* get lock owner */
        testl   %ecx,%ecx               /* is the mutex locked? */
-       jne     lml_fail                /* yes, we lose */
+       jne     Llml_fail               /* yes, we lose */
+Llml_acquire:
        movl    %gs:CPU_ACTIVE_THREAD,%ecx
        movl    %ecx,M_LOCKED
 
+       cmpl    $0,M_WAITERS            /* are there any waiters? */
+       jne     Llml_waiters            /* yes, more work to do */
+Llml_return:
+       xorl    %eax,%eax
+       movl    %eax,M_ILK
+
+       popf                            /* restore interrupt state */
+
+       NONLEAF_RET
+
+Llml_waiters:
        pushl   %edx                    /* save mutex address */
        pushl   %edx
        call    EXT(lck_mtx_lock_acquire)
        addl    $4,%esp
        popl    %edx                    /* restore mutex address */
+       jmp     Llml_return
+
+Llml_ilk_fail:
+       /*
+        * Slow path: call out to do the spinning.
+        */
+       pushl   %edx                    /* lock address */
+       call    EXT(lck_mtx_interlock_spin)
+       popl    %edx                    /* lock pointer */
+       jmp     Llml_retry              /* try again */
+
+Llml_fail:
+       /*
+        * Check if the owner is on another processor and therefore
+        * we should try to spin before blocking.
+        */
+       testl   $(OnProc),ACT_SPF(%ecx)
+       jz      Llml_block
 
+       /*
+        * Here if owner is on another processor:
+        *  - release the interlock
+        *  - spin on the holder until release or timeout
+        *  - in either case re-acquire the interlock
+        *  - if released, acquire it
+        *  - otherwise drop thru to block.
+        */
        xorl    %eax,%eax
-       movl    %eax,M_ILK
+       movl    %eax,M_ILK              /* zero interlock */
+       popf
+       pushf                           /* restore interrupt state */
 
-       popf                            /* restore interrupt state */
+       pushl   %edx                    /* save mutex address */
+       pushl   %edx
+       call    EXT(lck_mtx_lock_spin)
+       addl    $4,%esp
+       popl    %edx                    /* restore mutex address */
 
-       NONLEAF_RET
+       /* Re-acquire interlock */
+       cli                             /* disable interrupts */
+Llml_reget_retry:
+       movl    %gs:CPU_ACTIVE_THREAD,%ecx
+
+Llml_reget_hw:
+       movl    M_ILK,%eax              /* read interlock */
+       testl   %eax,%eax               /* unlocked? */
+       jne     Llml_ilk_refail         /* no - slow path */
+
+       lock; cmpxchgl  %ecx,M_ILK      /* atomic compare and exchange */
+       jne     Llml_reget_hw           /* branch on failure to retry */
 
-lml_fail:
+       movl    M_LOCKED,%ecx           /* get lock owner */
+       testl   %ecx,%ecx               /* is the mutex free? */
+       je      Llml_acquire            /* yes, acquire */
+       
+Llml_block:
        CHECK_MYLOCK(M_THREAD)
        pushl   %edx                    /* save mutex address */
        pushl   M_LOCKED
@@ -646,7 +786,17 @@ lml_fail:
        call    EXT(lck_mtx_lock_wait)  /* wait for the lock */
        addl    $8,%esp
        popl    %edx                    /* restore mutex address */
-       jmp     lml_retry               /* and try again */
+       cli                             /* ensure interrupts disabled */
+       jmp     Llml_retry              /* and try again */
+
+Llml_ilk_refail:
+       /*
+        * Slow path: call out to do the spinning.
+        */
+       pushl   %edx                    /* lock address */
+       call    EXT(lck_mtx_interlock_spin)
+       popl    %edx                    /* lock pointer */
+       jmp     Llml_reget_retry        /* try again */
 
 NONLEAF_ENTRY(lck_mtx_try_lock)
 
@@ -657,33 +807,28 @@ NONLEAF_ENTRY(lck_mtx_try_lock)
        CHECK_NO_SIMPLELOCKS()
        CHECK_PREEMPTION_LEVEL()
 
-       movl    B_PC,%ecx
-
        pushf                           /* save interrupt state */
        cli                             /* disable interrupts */
+Llmt_retry:
+       movl    %gs:CPU_ACTIVE_THREAD,%ecx
 
-lmt_get_hw:
+Llmt_get_hw:
        movl    M_ILK,%eax              /* read interlock */
        testl   %eax,%eax               /* unlocked? */
-       je      1f                      /* yes - attempt to lock it */
-       PAUSE                           /* no  - pause */
-       jmp     lmt_get_hw              /* try again */
-1:
+       jne     Llmt_ilk_fail           /* no - slow path */
+
        lock; cmpxchgl  %ecx,M_ILK      /* atomic compare and exchange */
-       jne     lmt_get_hw              /* branch on failure to retry */
+       jne     Llmt_get_hw             /* branch on failure to retry */
 
        movl    M_LOCKED,%ecx           /* get lock owner */
        testl   %ecx,%ecx               /* is the mutex locked? */
-       jne     lmt_fail                /* yes, we lose */
+       jne     Llmt_fail               /* yes, we lose */
        movl    %gs:CPU_ACTIVE_THREAD,%ecx
        movl    %ecx,M_LOCKED
 
-       pushl   %edx                    /* save mutex address */
-       pushl   %edx
-       call    EXT(lck_mtx_lock_acquire)
-       addl    $4,%esp
-       popl    %edx                    /* restore mutex address */
-
+       cmpl    $0,M_WAITERS            /* are there any waiters? */
+       jne     Llmt_waiters            /* yes, more work to do */
+Llmt_return:
        xorl    %eax,%eax
        movl    %eax,M_ILK
 
@@ -692,7 +837,24 @@ lmt_get_hw:
        movl    $1,%eax                 /* return success */
        NONLEAF_RET
 
-lmt_fail:
+Llmt_waiters:
+       pushl   %edx                    /* save mutex address */
+       pushl   %edx
+       call    EXT(lck_mtx_lock_acquire)
+       addl    $4,%esp
+       popl    %edx                    /* restore mutex address */
+       jmp     Llmt_return
+
+Llmt_ilk_fail:
+       /*
+        * Slow path: call out to do the spinning.
+        */
+       pushl   %edx                    /* lock address */
+       call    EXT(lck_mtx_interlock_spin)
+       popl    %edx                    /* lock pointer */
+       jmp     Llmt_retry              /* try again */
+
+Llmt_fail:
        xorl    %eax,%eax
        movl    %eax,M_ILK
 
@@ -707,25 +869,23 @@ NONLEAF_ENTRY(lck_mtx_unlock)
        cmpl    $(MUTEX_IND),M_ITAG     /* is this indirect? */
        cmove   M_PTR,%edx              /* yes - take indirection */
 
-       movl    B_PC,%ecx
-
        pushf                           /* save interrupt state */
        cli                             /* disable interrupts */
+Llmu_retry:
+       movl    %gs:CPU_ACTIVE_THREAD,%ecx
 
-lmu_get_hw:
+Llmu_get_hw:
        movl    M_ILK,%eax              /* read interlock */
        testl   %eax,%eax               /* unlocked? */
-       je      1f                      /* yes - attempt to lock it */
-       PAUSE                           /* no  - pause */
-       jmp     lmu_get_hw              /* try again */
-1:
+       jne     Llmu_ilk_fail           /* no - slow path */
+
        lock; cmpxchgl  %ecx,M_ILK      /* atomic compare and exchange */
-       jne     lmu_get_hw              /* branch on failure to retry */
+       jne     Llmu_get_hw             /* branch on failure to retry */
 
        cmpw    $0,M_WAITERS            /* are there any waiters? */
-       jne     lmu_wakeup              /* yes, more work to do */
+       jne     Llmu_wakeup             /* yes, more work to do */
 
-lmu_doit:
+Llmu_doit:
        xorl    %ecx,%ecx
        movl    %ecx,M_LOCKED           /* unlock the mutex */
 
@@ -735,14 +895,23 @@ lmu_doit:
 
        NONLEAF_RET
 
-lmu_wakeup:
+Llmu_ilk_fail:
+       /*
+        * Slow path: call out to do the spinning.
+        */
+       pushl   %edx                    /* lock address */
+       call    EXT(lck_mtx_interlock_spin)
+       popl    %edx                    /* lock pointer */
+       jmp     Llmu_retry              /* try again */
+
+Llmu_wakeup:
        pushl   %edx                    /* save mutex address */
        pushl   M_LOCKED
        pushl   %edx                    /* push mutex address */
        call    EXT(lck_mtx_unlock_wakeup)/* yes, wake a thread */
        addl    $8,%esp
        popl    %edx                    /* restore mutex pointer */
-       jmp     lmu_doit
+       jmp     Llmu_doit
 
 LEAF_ENTRY(lck_mtx_ilk_unlock)
        movl    L_ARG0,%edx             /* no indirection here */