2 * Copyright (c) 2000-2008 Apple Inc. All rights reserved.
4 * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
6 * This file contains Original Code and/or Modifications of Original Code
7 * as defined in and that are subject to the Apple Public Source License
8 * Version 2.0 (the 'License'). You may not use this file except in
9 * compliance with the License. The rights granted to you under the License
10 * may not be used to create, or enable the creation or redistribution of,
11 * unlawful or unlicensed copies of an Apple operating system, or to
12 * circumvent, violate, or enable the circumvention or violation of, any
13 * terms of an Apple operating system software license agreement.
15 * Please obtain a copy of the License at
16 * http://www.opensource.apple.com/apsl/ and read it before using this file.
18 * The Original Code and all software distributed under the License are
19 * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
20 * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
21 * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
22 * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
23 * Please see the License for the specific language governing rights and
24 * limitations under the License.
26 * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
32 * Mach Operating System
33 * Copyright (c) 1989 Carnegie-Mellon University
34 * All rights reserved. The CMU software License Agreement specifies
35 * the terms and conditions for use and redistribution.
39 #include <platforms.h>
40 #include <mach_ldebug.h>
42 #include <i386/eflags.h>
43 #include <i386/trap.h>
44 #include <config_dtrace.h>
49 #define PAUSE rep; nop
58 * When performance isn't the only concern, it's
59 * nice to build stack frames...
61 #define BUILD_STACK_FRAMES (GPROF || \
62 ((MACH_LDEBUG) && MACH_KDB))
64 #if BUILD_STACK_FRAMES
66 /* Stack-frame-relative: */
71 #define LEAF_ENTRY(name) \
76 #define LEAF_ENTRY2(n1,n2) \
86 #else /* BUILD_STACK_FRAMES */
88 /* Stack-pointer-relative: */
93 #define LEAF_ENTRY(name) \
96 #define LEAF_ENTRY2(n1,n2) \
103 #endif /* BUILD_STACK_FRAMES */
106 /* Non-leaf routines always have a stack frame: */
108 #define NONLEAF_ENTRY(name) \
113 #define NONLEAF_ENTRY2(n1,n2) \
119 #define NONLEAF_RET \
124 /* For x86_64, the varargs ABI requires that %al indicate
125 * how many SSE register contain arguments. In our case, 0 */
127 #define LOAD_STRING_ARG0(label) pushl $##label ;
128 #define LOAD_ARG1(x) pushl x ;
129 #define CALL_PANIC() call EXT(panic) ;
131 #define LOAD_STRING_ARG0(label) leaq label(%rip), %rdi ;
132 #define LOAD_ARG1(x) movq x, %rsi ;
133 #define CALL_PANIC() xorb %al,%al ; call EXT(panic) ;
136 #define CHECK_UNLOCK(current, owner) \
137 cmp current, owner ; \
139 LOAD_STRING_ARG0(2f) ; \
143 2: String "Mutex unlock attempted from non-owner thread"; \
149 * Routines for general lock debugging.
153 * Checks for expected lock types and calls "panic" on
154 * mismatch. Detects calls to Mutex functions with
155 * type simplelock and vice versa.
157 #define CHECK_MUTEX_TYPE() \
158 cmpl $ MUTEX_TAG,M_TYPE ; \
160 LOAD_STRING_ARG0(2f) ; \
164 2: String "not a mutex!" ; \
169 * If one or more simplelocks are currently held by a thread,
170 * an attempt to acquire a mutex will cause this check to fail
171 * (since a mutex lock may context switch, holding a simplelock
172 * is not a good thing).
175 #define CHECK_PREEMPTION_LEVEL() \
176 cmpl $0,%gs:CPU_HIBERNATE ; \
178 cmpl $0,%gs:CPU_PREEMPTION_LEVEL ; \
180 LOAD_ARG1(%gs:CPU_PREEMPTION_LEVEL) ; \
181 LOAD_STRING_ARG0(2f) ; \
185 2: String "preemption_level(%d) != 0!" ; \
189 #define CHECK_PREEMPTION_LEVEL()
192 #define CHECK_MYLOCK(current, owner) \
193 cmp current, owner ; \
195 LOAD_STRING_ARG0(2f) ; \
199 2: String "Attempt to recursively lock a non-recursive lock"; \
203 #else /* MACH_LDEBUG */
204 #define CHECK_MUTEX_TYPE()
205 #define CHECK_PREEMPTION_LEVEL()
206 #define CHECK_MYLOCK(thd)
207 #endif /* MACH_LDEBUG */
210 #define PREEMPTION_DISABLE \
211 incl %gs:CPU_PREEMPTION_LEVEL
214 #define PREEMPTION_ENABLE \
215 decl %gs:CPU_PREEMPTION_LEVEL ; \
218 testl $ EFL_IF,S_PC ; \
221 movl %gs:CPU_PENDING_AST,%eax ; \
222 testl $ AST_URGENT,%eax ; \
224 movl %gs:CPU_INTERRUPT_LEVEL,%eax ; \
238 .globl _lockstat_probe
239 .globl _lockstat_probemap
242 * LOCKSTAT_LABEL creates a dtrace symbol which contains
243 * a pointer into the lock code function body. At that
244 * point is a "ret" instruction that can be patched into
248 #if defined(__i386__)
250 #define LOCKSTAT_LABEL(lab) \
258 #define LOCKSTAT_RECORD(id, lck) \
261 sub $0x38,%esp /* size of dtrace_probe args */ ; \
262 movl _lockstat_probemap + (id * 4),%eax ; \
273 movl lck,4(%esp) /* copy lock pointer to arg 1 */ ; \
276 call *_lockstat_probe ; \
278 /* ret - left to subsequent code, e.g. return values */
280 #elif defined(__x86_64__)
281 #define LOCKSTAT_LABEL(lab) \
289 #define LOCKSTAT_RECORD(id, lck) \
292 movl _lockstat_probemap + (id * 4)(%rip),%eax ; \
301 call *_lockstat_probe(%rip) ; \
303 /* ret - left to subsequent code, e.g. return values */
305 #error Unsupported architecture
307 #endif /* CONFIG_DTRACE */
310 * For most routines, the hw_lock_t pointer is loaded into a
311 * register initially, and then either a byte or register-sized
312 * word is loaded/stored to the pointer
315 #if defined(__i386__)
316 #define HW_LOCK_REGISTER %edx
317 #define LOAD_HW_LOCK_REGISTER mov L_ARG0, HW_LOCK_REGISTER
318 #define HW_LOCK_THREAD_REGISTER %ecx
319 #define LOAD_HW_LOCK_THREAD_REGISTER mov %gs:CPU_ACTIVE_THREAD, HW_LOCK_THREAD_REGISTER
320 #define HW_LOCK_MOV_WORD movl
321 #define HW_LOCK_EXAM_REGISTER %eax
322 #elif defined(__x86_64__)
323 #define HW_LOCK_REGISTER %rdi
324 #define LOAD_HW_LOCK_REGISTER
325 #define HW_LOCK_THREAD_REGISTER %rcx
326 #define LOAD_HW_LOCK_THREAD_REGISTER mov %gs:CPU_ACTIVE_THREAD, HW_LOCK_THREAD_REGISTER
327 #define HW_LOCK_MOV_WORD movq
328 #define HW_LOCK_EXAM_REGISTER %rax
330 #error Unsupported architecture
334 * void hw_lock_init(hw_lock_t)
336 * Initialize a hardware lock.
338 LEAF_ENTRY(hw_lock_init)
339 LOAD_HW_LOCK_REGISTER /* fetch lock pointer */
340 HW_LOCK_MOV_WORD $0, (HW_LOCK_REGISTER) /* clear the lock */
345 * void hw_lock_byte_init(uint8_t *)
347 * Initialize a hardware byte lock.
349 LEAF_ENTRY(hw_lock_byte_init)
350 LOAD_HW_LOCK_REGISTER /* fetch lock pointer */
351 movb $0, (HW_LOCK_REGISTER) /* clear the lock */
355 * void hw_lock_lock(hw_lock_t)
357 * Acquire lock, spinning until it becomes available.
358 * MACH_RT: also return with preemption disabled.
360 LEAF_ENTRY(hw_lock_lock)
361 LOAD_HW_LOCK_REGISTER /* fetch lock pointer */
362 LOAD_HW_LOCK_THREAD_REGISTER /* get thread pointer */
366 mov (HW_LOCK_REGISTER), HW_LOCK_EXAM_REGISTER
367 test HW_LOCK_EXAM_REGISTER,HW_LOCK_EXAM_REGISTER /* lock locked? */
368 jne 3f /* branch if so */
369 lock; cmpxchg HW_LOCK_THREAD_REGISTER,(HW_LOCK_REGISTER) /* try to acquire the HW lock */
371 movl $1,%eax /* In case this was a timeout call */
372 LEAF_RET /* if yes, then nothing left to do */
374 PAUSE /* pause for hyper-threading */
375 jmp 1b /* try again */
378 * void hw_lock_byte_lock(uint8_t *lock_byte)
380 * Acquire byte sized lock operand, spinning until it becomes available.
381 * MACH_RT: also return with preemption disabled.
384 LEAF_ENTRY(hw_lock_byte_lock)
385 LOAD_HW_LOCK_REGISTER /* Load lock pointer */
387 movl $1, %ecx /* Set lock value */
389 movb (HW_LOCK_REGISTER), %al /* Load byte at address */
390 testb %al,%al /* lock locked? */
391 jne 3f /* branch if so */
392 lock; cmpxchg %cl,(HW_LOCK_REGISTER) /* attempt atomic compare exchange */
394 LEAF_RET /* if yes, then nothing left to do */
396 PAUSE /* pause for hyper-threading */
397 jmp 1b /* try again */
400 * unsigned int hw_lock_to(hw_lock_t, unsigned int)
402 * Acquire lock, spinning until it becomes available or timeout.
403 * MACH_RT: also return with preemption disabled.
405 LEAF_ENTRY(hw_lock_to)
407 LOAD_HW_LOCK_REGISTER /* fetch lock pointer */
408 LOAD_HW_LOCK_THREAD_REGISTER
411 * Attempt to grab the lock immediately
412 * - fastpath without timeout nonsense.
416 mov (HW_LOCK_REGISTER), HW_LOCK_EXAM_REGISTER
417 test HW_LOCK_EXAM_REGISTER,HW_LOCK_EXAM_REGISTER /* lock locked? */
418 jne 2f /* branch if so */
419 lock; cmpxchg HW_LOCK_THREAD_REGISTER,(HW_LOCK_REGISTER) /* try to acquire the HW lock */
420 jne 2f /* branch on failure */
425 #define INNER_LOOP_COUNT 1000
427 * Failed to get the lock so set the timeout
428 * and then spin re-checking the lock but pausing
429 * every so many (INNER_LOOP_COUNT) spins to check for timeout.
432 movl L_ARG1,%ecx /* fetch timeout */
438 rdtsc /* read cyclecount into %edx:%eax */
440 addl %ecx,%eax /* fetch and timeout */
441 adcl $0,%edx /* add carry */
443 mov %eax,%ebx /* %ecx:%ebx is the timeout expiry */
444 mov %edi, %edx /* load lock back into %edx */
448 rdtsc /* read cyclecount into %edx:%eax */
451 orq %rdx, %rax /* load 64-bit quantity into %rax */
452 addq %rax, %rsi /* %rsi is the timeout expiry */
457 * The inner-loop spin to look for the lock being freed.
460 mov $(INNER_LOOP_COUNT),%edi
462 mov $(INNER_LOOP_COUNT),%r9
465 PAUSE /* pause for hyper-threading */
466 mov (HW_LOCK_REGISTER),HW_LOCK_EXAM_REGISTER /* spin checking lock value in cache */
467 test HW_LOCK_EXAM_REGISTER,HW_LOCK_EXAM_REGISTER
468 je 6f /* zero => unlocked, try to grab it */
470 decl %edi /* decrement inner loop count */
472 decq %r9 /* decrement inner loop count */
474 jnz 5b /* time to check for timeout? */
477 * Here after spinning INNER_LOOP_COUNT times, check for timeout
480 mov %edx,%edi /* Save %edx */
482 rdtsc /* cyclecount into %edx:%eax */
484 xchg %edx,%edi /* cyclecount into %edi:%eax */
485 cmpl %ecx,%edi /* compare high-order 32-bits */
486 jb 4b /* continue spinning if less, or */
487 cmpl %ebx,%eax /* compare low-order 32-bits */
488 jb 4b /* continue if less, else bail */
489 xor %eax,%eax /* with 0 return value */
494 rdtsc /* cyclecount into %edx:%eax */
497 orq %rdx, %rax /* load 64-bit quantity into %rax */
498 cmpq %rsi, %rax /* compare to timeout */
499 jb 4b /* continue spinning if less, or */
500 xor %rax,%rax /* with 0 return value */
507 * Here to try to grab the lock that now appears to be free
510 LOAD_HW_LOCK_THREAD_REGISTER
511 lock; cmpxchg HW_LOCK_THREAD_REGISTER,(HW_LOCK_REGISTER) /* try to acquire the HW lock */
512 jne 4b /* no - spin again */
513 movl $1,%eax /* yes */
523 * void hw_lock_unlock(hw_lock_t)
525 * Unconditionally release lock.
526 * MACH_RT: release preemption level.
528 LEAF_ENTRY(hw_lock_unlock)
529 LOAD_HW_LOCK_REGISTER /* fetch lock pointer */
530 HW_LOCK_MOV_WORD $0, (HW_LOCK_REGISTER) /* clear the lock */
535 * void hw_lock_byte_unlock(uint8_t *lock_byte)
537 * Unconditionally release byte sized lock operand.
538 * MACH_RT: release preemption level.
541 LEAF_ENTRY(hw_lock_byte_unlock)
542 LOAD_HW_LOCK_REGISTER /* Load lock pointer */
543 movb $0, (HW_LOCK_REGISTER) /* Clear the lock byte */
548 * unsigned int hw_lock_try(hw_lock_t)
549 * MACH_RT: returns with preemption disabled on success.
551 LEAF_ENTRY(hw_lock_try)
552 LOAD_HW_LOCK_REGISTER /* fetch lock pointer */
553 LOAD_HW_LOCK_THREAD_REGISTER
556 mov (HW_LOCK_REGISTER),HW_LOCK_EXAM_REGISTER
557 test HW_LOCK_EXAM_REGISTER,HW_LOCK_EXAM_REGISTER
559 lock; cmpxchg HW_LOCK_THREAD_REGISTER,(HW_LOCK_REGISTER) /* try to acquire the HW lock */
562 movl $1,%eax /* success */
566 PREEMPTION_ENABLE /* failure: release preemption... */
567 xorl %eax,%eax /* ...and return failure */
571 * unsigned int hw_lock_held(hw_lock_t)
572 * MACH_RT: doesn't change preemption state.
573 * N.B. Racy, of course.
575 LEAF_ENTRY(hw_lock_held)
576 LOAD_HW_LOCK_REGISTER /* fetch lock pointer */
577 mov (HW_LOCK_REGISTER),HW_LOCK_EXAM_REGISTER /* check lock value */
578 test HW_LOCK_EXAM_REGISTER,HW_LOCK_EXAM_REGISTER
580 cmovne %ecx,%eax /* 0 => unlocked, 1 => locked */
585 * Reader-writer lock fastpaths. These currently exist for the
586 * shared lock acquire, the exclusive lock acquire, the shared to
587 * exclusive upgrade and the release paths (where they reduce overhead
588 * considerably) -- these are by far the most frequently used routines
590 * The following should reflect the layout of the bitfield embedded within
591 * the lck_rw_t structure (see i386/locks.h).
593 #define LCK_RW_INTERLOCK (0x1 << 16)
595 #define LCK_RW_PRIV_EXCL (0x1 << 24)
596 #define LCK_RW_WANT_UPGRADE (0x2 << 24)
597 #define LCK_RW_WANT_WRITE (0x4 << 24)
598 #define LCK_R_WAITING (0x8 << 24)
599 #define LCK_W_WAITING (0x10 << 24)
601 #define LCK_RW_SHARED_MASK (0xffff)
604 * For most routines, the lck_rw_t pointer is loaded into a
605 * register initially, and the flags bitfield loaded into another
606 * register and examined
609 #if defined(__i386__)
610 #define LCK_RW_REGISTER %edx
611 #define LOAD_LCK_RW_REGISTER mov S_ARG0, LCK_RW_REGISTER
612 #define LCK_RW_FLAGS_REGISTER %eax
613 #define LOAD_LCK_RW_FLAGS_REGISTER mov (LCK_RW_REGISTER), LCK_RW_FLAGS_REGISTER
614 #elif defined(__x86_64__)
615 #define LCK_RW_REGISTER %rdi
616 #define LOAD_LCK_RW_REGISTER
617 #define LCK_RW_FLAGS_REGISTER %eax
618 #define LOAD_LCK_RW_FLAGS_REGISTER mov (LCK_RW_REGISTER), LCK_RW_FLAGS_REGISTER
620 #error Unsupported architecture
623 #define RW_LOCK_SHARED_MASK (LCK_RW_INTERLOCK | LCK_RW_WANT_UPGRADE | LCK_RW_WANT_WRITE)
625 * void lck_rw_lock_shared(lck_rw_t *)
628 Entry(lck_rw_lock_shared)
631 LOAD_LCK_RW_FLAGS_REGISTER /* Load state bitfield and interlock */
632 testl $(RW_LOCK_SHARED_MASK), %eax /* Eligible for fastpath? */
635 movl %eax, %ecx /* original value in %eax for cmpxchgl */
636 incl %ecx /* Increment reader refcount */
638 cmpxchgl %ecx, (LCK_RW_REGISTER) /* Attempt atomic exchange */
643 * Dtrace lockstat event: LS_LCK_RW_LOCK_SHARED_ACQUIRE
644 * Implemented by swapping between return and no-op instructions.
645 * See bsd/dev/dtrace/lockstat.c.
647 LOCKSTAT_LABEL(_lck_rw_lock_shared_lockstat_patch_point)
649 /* Fall thru when patched, counting on lock pointer in LCK_RW_REGISTER */
650 LOCKSTAT_RECORD(LS_LCK_RW_LOCK_SHARED_ACQUIRE, LCK_RW_REGISTER)
657 jmp EXT(lck_rw_lock_shared_gen)
661 #define RW_TRY_LOCK_SHARED_MASK (LCK_RW_WANT_UPGRADE | LCK_RW_WANT_WRITE)
663 * void lck_rw_try_lock_shared(lck_rw_t *)
666 Entry(lck_rw_try_lock_shared)
669 LOAD_LCK_RW_FLAGS_REGISTER /* Load state bitfield and interlock */
670 testl $(LCK_RW_INTERLOCK), %eax
672 testl $(RW_TRY_LOCK_SHARED_MASK), %eax
673 jne 3f /* lock is busy */
675 movl %eax, %ecx /* original value in %eax for cmpxchgl */
676 incl %ecx /* Increment reader refcount */
678 cmpxchgl %ecx, (LCK_RW_REGISTER) /* Attempt atomic exchange */
684 * Dtrace lockstat event: LS_LCK_RW_TRY_LOCK_SHARED_ACQUIRE
685 * Implemented by swapping between return and no-op instructions.
686 * See bsd/dev/dtrace/lockstat.c.
688 LOCKSTAT_LABEL(_lck_rw_try_lock_shared_lockstat_patch_point)
690 /* Fall thru when patched, counting on lock pointer in LCK_RW_REGISTER */
691 LOCKSTAT_RECORD(LS_LCK_RW_LOCK_SHARED_ACQUIRE, LCK_RW_REGISTER)
693 movl $1, %eax /* return TRUE */
703 #define RW_LOCK_EXCLUSIVE_HELD (LCK_RW_WANT_WRITE | LCK_RW_WANT_UPGRADE)
705 * int lck_rw_grab_shared(lck_rw_t *)
708 Entry(lck_rw_grab_shared)
711 LOAD_LCK_RW_FLAGS_REGISTER /* Load state bitfield and interlock */
712 testl $(LCK_RW_INTERLOCK), %eax
714 testl $(RW_LOCK_EXCLUSIVE_HELD), %eax
717 movl %eax, %ecx /* original value in %eax for cmpxchgl */
718 incl %ecx /* Increment reader refcount */
720 cmpxchgl %ecx, (LCK_RW_REGISTER) /* Attempt atomic exchange */
723 movl $1, %eax /* return success */
726 testl $(LCK_RW_SHARED_MASK), %eax
728 testl $(LCK_RW_PRIV_EXCL), %eax
731 xorl %eax, %eax /* return failure */
739 #define RW_LOCK_EXCLUSIVE_MASK (LCK_RW_SHARED_MASK | LCK_RW_INTERLOCK | \
740 LCK_RW_WANT_UPGRADE | LCK_RW_WANT_WRITE)
742 * void lck_rw_lock_exclusive(lck_rw_t*)
745 Entry(lck_rw_lock_exclusive)
748 LOAD_LCK_RW_FLAGS_REGISTER /* Load state bitfield, interlock and shared count */
749 testl $(RW_LOCK_EXCLUSIVE_MASK), %eax /* Eligible for fastpath? */
750 jne 3f /* no, go slow */
752 movl %eax, %ecx /* original value in %eax for cmpxchgl */
753 orl $(LCK_RW_WANT_WRITE), %ecx
755 cmpxchgl %ecx, (LCK_RW_REGISTER) /* Attempt atomic exchange */
760 * Dtrace lockstat event: LS_LCK_RW_LOCK_EXCL_ACQUIRE
761 * Implemented by swapping between return and no-op instructions.
762 * See bsd/dev/dtrace/lockstat.c.
764 LOCKSTAT_LABEL(_lck_rw_lock_exclusive_lockstat_patch_point)
766 /* Fall thru when patched, counting on lock pointer in LCK_RW_REGISTER */
767 LOCKSTAT_RECORD(LS_LCK_RW_LOCK_SHARED_ACQUIRE, LCK_RW_REGISTER)
774 jmp EXT(lck_rw_lock_exclusive_gen)
778 #define RW_TRY_LOCK_EXCLUSIVE_MASK (LCK_RW_SHARED_MASK | LCK_RW_WANT_UPGRADE | LCK_RW_WANT_WRITE)
780 * void lck_rw_try_lock_exclusive(lck_rw_t *)
782 * Tries to get a write lock.
784 * Returns FALSE if the lock is not held on return.
786 Entry(lck_rw_try_lock_exclusive)
789 LOAD_LCK_RW_FLAGS_REGISTER /* Load state bitfield, interlock and shared count */
790 testl $(LCK_RW_INTERLOCK), %eax
792 testl $(RW_TRY_LOCK_EXCLUSIVE_MASK), %eax
793 jne 3f /* can't get it */
795 movl %eax, %ecx /* original value in %eax for cmpxchgl */
796 orl $(LCK_RW_WANT_WRITE), %ecx
798 cmpxchgl %ecx, (LCK_RW_REGISTER) /* Attempt atomic exchange */
804 * Dtrace lockstat event: LS_LCK_RW_TRY_LOCK_EXCL_ACQUIRE
805 * Implemented by swapping between return and no-op instructions.
806 * See bsd/dev/dtrace/lockstat.c.
808 LOCKSTAT_LABEL(_lck_rw_try_lock_exclusive_lockstat_patch_point)
810 /* Fall thru when patched, counting on lock pointer in LCK_RW_REGISTER */
811 LOCKSTAT_RECORD(LS_LCK_RW_LOCK_SHARED_ACQUIRE, LCK_RW_REGISTER)
813 movl $1, %eax /* return TRUE */
819 xorl %eax, %eax /* return FALSE */
825 * void lck_rw_lock_shared_to_exclusive(lck_rw_t*)
827 * fastpath can be taken if
828 * the current rw_shared_count == 1
829 * AND the interlock is clear
830 * AND RW_WANT_UPGRADE is not set
832 * note that RW_WANT_WRITE could be set, but will not
833 * be indicative of an exclusive hold since we have
834 * a read count on the lock that we have not yet released
835 * we can blow by that state since the lck_rw_lock_exclusive
836 * function will block until rw_shared_count == 0 and
837 * RW_WANT_UPGRADE is clear... it does this check behind
838 * the interlock which we are also checking for
840 * to make the transition we must be able to atomically
841 * set RW_WANT_UPGRADE and get rid of the read count we hold
843 Entry(lck_rw_lock_shared_to_exclusive)
846 LOAD_LCK_RW_FLAGS_REGISTER /* Load state bitfield, interlock and shared count */
847 testl $(LCK_RW_INTERLOCK), %eax
849 testl $(LCK_RW_WANT_UPGRADE), %eax
852 movl %eax, %ecx /* original value in %eax for cmpxchgl */
853 orl $(LCK_RW_WANT_UPGRADE), %ecx /* ask for WANT_UPGRADE */
854 decl %ecx /* and shed our read count */
856 cmpxchgl %ecx, (LCK_RW_REGISTER) /* Attempt atomic exchange */
858 /* we now own the WANT_UPGRADE */
859 testl $(LCK_RW_SHARED_MASK), %ecx /* check to see if all of the readers are drained */
860 jne 8f /* if not, we need to go wait */
865 * Dtrace lockstat event: LS_LCK_RW_LOCK_SHARED_TO_EXCL_UPGRADE
866 * Implemented by swapping between return and no-op instructions.
867 * See bsd/dev/dtrace/lockstat.c.
869 LOCKSTAT_LABEL(_lck_rw_lock_shared_to_exclusive_lockstat_patch_point)
871 /* Fall thru when patched, counting on lock pointer in LCK_RW_REGISTER */
872 LOCKSTAT_RECORD(LS_LCK_RW_LOCK_SHARED_ACQUIRE, LCK_RW_REGISTER)
874 movl $1, %eax /* return success */
877 2: /* someone else already holds WANT_UPGRADE */
878 movl %eax, %ecx /* original value in %eax for cmpxchgl */
879 decl %ecx /* shed our read count */
880 testl $(LCK_RW_SHARED_MASK), %ecx
881 jne 3f /* we were the last reader */
882 andl $(~LCK_W_WAITING), %ecx /* so clear the wait indicator */
885 cmpxchgl %ecx, (LCK_RW_REGISTER) /* Attempt atomic exchange */
889 pushl %eax /* go check to see if we need to */
890 push %edx /* wakeup anyone */
891 call EXT(lck_rw_lock_shared_to_exclusive_failure)
894 mov %eax, %esi /* put old flags as second arg */
895 /* lock is alread in %rdi */
896 call EXT(lck_rw_lock_shared_to_exclusive_failure)
898 ret /* and pass the failure return along */
903 jmp EXT(lck_rw_lock_shared_to_exclusive_success)
908 rwl_release_error_str:
909 .asciz "Releasing non-exclusive RW lock without a reader refcount!"
913 * lck_rw_type_t lck_rw_done(lck_rw_t *)
919 LOAD_LCK_RW_FLAGS_REGISTER /* Load state bitfield, interlock and reader count */
920 testl $(LCK_RW_INTERLOCK), %eax
921 jne 7f /* wait for interlock to clear */
923 movl %eax, %ecx /* keep original value in %eax for cmpxchgl */
924 testl $(LCK_RW_SHARED_MASK), %ecx /* if reader count == 0, must be exclusive lock */
926 decl %ecx /* Decrement reader count */
927 testl $(LCK_RW_SHARED_MASK), %ecx /* if reader count has now gone to 0, check for waiters */
931 testl $(LCK_RW_WANT_UPGRADE), %ecx
933 andl $(~LCK_RW_WANT_UPGRADE), %ecx
936 testl $(LCK_RW_WANT_WRITE), %ecx
937 je 8f /* lock is not 'owned', go panic */
938 andl $(~LCK_RW_WANT_WRITE), %ecx
941 * test the original values to match what
942 * lck_rw_done_gen is going to do to determine
943 * which wakeups need to happen...
945 * if !(fake_lck->lck_rw_priv_excl && fake_lck->lck_w_waiting)
947 testl $(LCK_W_WAITING), %eax
949 andl $(~LCK_W_WAITING), %ecx
951 testl $(LCK_RW_PRIV_EXCL), %eax
954 andl $(~LCK_R_WAITING), %ecx
957 cmpxchgl %ecx, (LCK_RW_REGISTER) /* Attempt atomic exchange */
963 call EXT(lck_rw_done_gen)
966 mov %eax,%esi /* old flags in %rsi */
967 /* lock is in %rdi already */
968 call EXT(lck_rw_done_gen)
975 LOAD_STRING_ARG0(rwl_release_error_str)
981 * lck_rw_type_t lck_rw_lock_exclusive_to_shared(lck_rw_t *)
984 Entry(lck_rw_lock_exclusive_to_shared)
987 LOAD_LCK_RW_FLAGS_REGISTER /* Load state bitfield, interlock and reader count */
988 testl $(LCK_RW_INTERLOCK), %eax
989 jne 6f /* wait for interlock to clear */
991 movl %eax, %ecx /* keep original value in %eax for cmpxchgl */
992 incl %ecx /* Increment reader count */
994 testl $(LCK_RW_WANT_UPGRADE), %ecx
996 andl $(~LCK_RW_WANT_UPGRADE), %ecx
999 andl $(~LCK_RW_WANT_WRITE), %ecx
1002 * test the original values to match what
1003 * lck_rw_lock_exclusive_to_shared_gen is going to do to determine
1004 * which wakeups need to happen...
1006 * if !(fake_lck->lck_rw_priv_excl && fake_lck->lck_w_waiting)
1008 testl $(LCK_W_WAITING), %eax
1010 testl $(LCK_RW_PRIV_EXCL), %eax
1013 andl $(~LCK_R_WAITING), %ecx
1016 cmpxchgl %ecx, (LCK_RW_REGISTER) /* Attempt atomic exchange */
1022 call EXT(lck_rw_lock_exclusive_to_shared_gen)
1026 call EXT(lck_rw_lock_exclusive_to_shared_gen)
1036 * int lck_rw_grab_want(lck_rw_t *)
1039 Entry(lck_rw_grab_want)
1040 LOAD_LCK_RW_REGISTER
1042 LOAD_LCK_RW_FLAGS_REGISTER /* Load state bitfield, interlock and reader count */
1043 testl $(LCK_RW_INTERLOCK), %eax
1044 jne 3f /* wait for interlock to clear */
1045 testl $(LCK_RW_WANT_WRITE), %eax /* want_write has been grabbed by someone else */
1046 jne 2f /* go return failure */
1048 movl %eax, %ecx /* original value in %eax for cmpxchgl */
1049 orl $(LCK_RW_WANT_WRITE), %ecx
1051 cmpxchgl %ecx, (LCK_RW_REGISTER) /* Attempt atomic exchange */
1053 /* we now own want_write */
1054 movl $1, %eax /* return success */
1057 xorl %eax, %eax /* return failure */
1064 #define RW_LOCK_SHARED_OR_UPGRADE_MASK (LCK_RW_SHARED_MASK | LCK_RW_INTERLOCK | LCK_RW_WANT_UPGRADE)
1066 * int lck_rw_held_read_or_upgrade(lck_rw_t *)
1069 Entry(lck_rw_held_read_or_upgrade)
1070 LOAD_LCK_RW_REGISTER
1071 LOAD_LCK_RW_FLAGS_REGISTER /* Load state bitfield, interlock and reader count */
1072 andl $(RW_LOCK_SHARED_OR_UPGRADE_MASK), %eax
1078 * N.B.: On x86, statistics are currently recorded for all indirect mutexes.
1079 * Also, only the acquire attempt count (GRP_MTX_STAT_UTIL) is maintained
1080 * as a 64-bit quantity (this matches the existing PowerPC implementation,
1081 * and the new x86 specific statistics are also maintained as 32-bit
1085 * Enable this preprocessor define to record the first miss alone
1086 * By default, we count every miss, hence multiple misses may be
1087 * recorded for a single lock acquire attempt via lck_mtx_lock
1089 #undef LOG_FIRST_MISS_ALONE
1092 * This preprocessor define controls whether the R-M-W update of the
1093 * per-group statistics elements are atomic (LOCK-prefixed)
1094 * Enabled by default.
1096 #define ATOMIC_STAT_UPDATES 1
1098 #if defined(ATOMIC_STAT_UPDATES)
1099 #define LOCK_IF_ATOMIC_STAT_UPDATES lock
1101 #define LOCK_IF_ATOMIC_STAT_UPDATES
1102 #endif /* ATOMIC_STAT_UPDATES */
1106 * For most routines, the lck_mtx_t pointer is loaded into a
1107 * register initially, and the owner field checked for indirection.
1108 * Eventually the lock owner is loaded into a register and examined.
1111 #define M_OWNER MUTEX_OWNER
1112 #define M_PTR MUTEX_PTR
1113 #define M_STATE MUTEX_STATE
1115 #if defined(__i386__)
1117 #define LMTX_ARG0 B_ARG0
1118 #define LMTX_ARG1 B_ARG1
1119 #define LMTX_REG %edx
1120 #define LMTX_A_REG %eax
1121 #define LMTX_A_REG32 %eax
1122 #define LMTX_C_REG %ecx
1123 #define LMTX_C_REG32 %ecx
1124 #define LMTX_D_REG %edx
1125 #define LMTX_RET_REG %eax
1126 #define LMTX_LGROUP_REG %esi
1127 #define LMTX_SSTATE_REG %edi
1128 #define LOAD_LMTX_REG(arg) mov arg, LMTX_REG
1129 #define LOAD_REG_ARG0(reg) push reg
1130 #define LOAD_REG_ARG1(reg) push reg
1131 #define LMTX_CHK_EXTENDED cmp LMTX_REG, LMTX_ARG0
1132 #define LMTX_ASSERT_OWNED cmpl $(MUTEX_ASSERT_OWNED), LMTX_ARG1
1134 #define LMTX_ENTER_EXTENDED \
1135 mov M_PTR(LMTX_REG), LMTX_REG ; \
1136 push LMTX_LGROUP_REG ; \
1137 push LMTX_SSTATE_REG ; \
1138 xor LMTX_SSTATE_REG, LMTX_SSTATE_REG ; \
1139 mov MUTEX_GRP(LMTX_REG), LMTX_LGROUP_REG ; \
1140 LOCK_IF_ATOMIC_STAT_UPDATES ; \
1141 addl $1, GRP_MTX_STAT_UTIL(LMTX_LGROUP_REG) ; \
1143 incl GRP_MTX_STAT_UTIL+4(LMTX_LGROUP_REG) ; \
1146 #define LMTX_EXIT_EXTENDED \
1147 pop LMTX_SSTATE_REG ; \
1151 #define LMTX_CHK_EXTENDED_EXIT \
1152 cmp LMTX_REG, LMTX_ARG0 ; \
1154 pop LMTX_SSTATE_REG ; \
1155 pop LMTX_LGROUP_REG ; \
1159 #if LOG_FIRST_MISS_ALONE
1160 #define LMTX_UPDATE_MISS \
1161 test $1, LMTX_SSTATE_REG ; \
1163 LOCK_IF_ATOMIC_STAT_UPDATES ; \
1164 incl GRP_MTX_STAT_MISS(LMTX_LGROUP_REG) ; \
1165 or $1, LMTX_SSTATE_REG ; \
1168 #define LMTX_UPDATE_MISS \
1169 LOCK_IF_ATOMIC_STAT_UPDATES ; \
1170 incl GRP_MTX_STAT_MISS(LMTX_LGROUP_REG)
1174 #if LOG_FIRST_MISS_ALONE
1175 #define LMTX_UPDATE_WAIT \
1176 test $2, LMTX_SSTATE_REG ; \
1178 LOCK_IF_ATOMIC_STAT_UPDATES ; \
1179 incl GRP_MTX_STAT_WAIT(LMTX_LGROUP_REG) ; \
1180 or $2, LMTX_SSTATE_REG ; \
1183 #define LMTX_UPDATE_WAIT \
1184 LOCK_IF_ATOMIC_STAT_UPDATES ; \
1185 incl GRP_MTX_STAT_WAIT(LMTX_LGROUP_REG)
1190 * Record the "direct wait" statistic, which indicates if a
1191 * miss proceeded to block directly without spinning--occurs
1192 * if the owner of the mutex isn't running on another processor
1193 * at the time of the check.
1195 #define LMTX_UPDATE_DIRECT_WAIT \
1196 LOCK_IF_ATOMIC_STAT_UPDATES ; \
1197 incl GRP_MTX_STAT_DIRECT_WAIT(LMTX_LGROUP_REG)
1200 #define LMTX_CALLEXT1(func_name) \
1203 call EXT(func_name) ; \
1207 #define LMTX_CALLEXT2(func_name, reg) \
1211 call EXT(func_name) ; \
1215 #elif defined(__x86_64__)
1217 #define LMTX_ARG0 %rdi
1218 #define LMTX_ARG1 %rsi
1219 #define LMTX_REG_ORIG %rdi
1220 #define LMTX_REG %rdx
1221 #define LMTX_A_REG %rax
1222 #define LMTX_A_REG32 %eax
1223 #define LMTX_C_REG %rcx
1224 #define LMTX_C_REG32 %ecx
1225 #define LMTX_D_REG %rdx
1226 #define LMTX_RET_REG %rax
1227 #define LMTX_LGROUP_REG %r10
1228 #define LMTX_SSTATE_REG %r11
1229 #define LOAD_LMTX_REG(arg) mov %rdi, %rdx
1230 #define LOAD_REG_ARG0(reg) mov reg, %rdi
1231 #define LOAD_REG_ARG1(reg) mov reg, %rsi
1232 #define LMTX_CHK_EXTENDED cmp LMTX_REG, LMTX_REG_ORIG
1233 #define LMTX_ASSERT_OWNED cmp $(MUTEX_ASSERT_OWNED), LMTX_ARG1
1235 #define LMTX_ENTER_EXTENDED \
1236 mov M_PTR(LMTX_REG), LMTX_REG ; \
1237 xor LMTX_SSTATE_REG, LMTX_SSTATE_REG ; \
1238 mov MUTEX_GRP(LMTX_REG), LMTX_LGROUP_REG ; \
1239 LOCK_IF_ATOMIC_STAT_UPDATES ; \
1240 incq GRP_MTX_STAT_UTIL(LMTX_LGROUP_REG)
1242 #define LMTX_EXIT_EXTENDED
1244 #define LMTX_CHK_EXTENDED_EXIT
1247 #if LOG_FIRST_MISS_ALONE
1248 #define LMTX_UPDATE_MISS \
1249 test $1, LMTX_SSTATE_REG ; \
1251 LOCK_IF_ATOMIC_STAT_UPDATES ; \
1252 incl GRP_MTX_STAT_MISS(LMTX_LGROUP_REG) ; \
1253 or $1, LMTX_SSTATE_REG ; \
1256 #define LMTX_UPDATE_MISS \
1257 LOCK_IF_ATOMIC_STAT_UPDATES ; \
1258 incl GRP_MTX_STAT_MISS(LMTX_LGROUP_REG)
1262 #if LOG_FIRST_MISS_ALONE
1263 #define LMTX_UPDATE_WAIT \
1264 test $2, LMTX_SSTATE_REG ; \
1266 LOCK_IF_ATOMIC_STAT_UPDATES ; \
1267 incl GRP_MTX_STAT_WAIT(LMTX_LGROUP_REG) ; \
1268 or $2, LMTX_SSTATE_REG ; \
1271 #define LMTX_UPDATE_WAIT \
1272 LOCK_IF_ATOMIC_STAT_UPDATES ; \
1273 incl GRP_MTX_STAT_WAIT(LMTX_LGROUP_REG)
1278 * Record the "direct wait" statistic, which indicates if a
1279 * miss proceeded to block directly without spinning--occurs
1280 * if the owner of the mutex isn't running on another processor
1281 * at the time of the check.
1283 #define LMTX_UPDATE_DIRECT_WAIT \
1284 LOCK_IF_ATOMIC_STAT_UPDATES ; \
1285 incl GRP_MTX_STAT_DIRECT_WAIT(LMTX_LGROUP_REG)
1288 #define LMTX_CALLEXT1(func_name) \
1289 LMTX_CHK_EXTENDED ; \
1291 push LMTX_LGROUP_REG ; \
1292 push LMTX_SSTATE_REG ; \
1293 12: push LMTX_REG_ORIG ; \
1295 mov LMTX_REG, LMTX_ARG0 ; \
1296 call EXT(func_name) ; \
1298 pop LMTX_REG_ORIG ; \
1299 LMTX_CHK_EXTENDED ; \
1301 pop LMTX_SSTATE_REG ; \
1302 pop LMTX_LGROUP_REG ; \
1305 #define LMTX_CALLEXT2(func_name, reg) \
1306 LMTX_CHK_EXTENDED ; \
1308 push LMTX_LGROUP_REG ; \
1309 push LMTX_SSTATE_REG ; \
1310 12: push LMTX_REG_ORIG ; \
1312 mov reg, LMTX_ARG1 ; \
1313 mov LMTX_REG, LMTX_ARG0 ; \
1314 call EXT(func_name) ; \
1316 pop LMTX_REG_ORIG ; \
1317 LMTX_CHK_EXTENDED ; \
1319 pop LMTX_SSTATE_REG ; \
1320 pop LMTX_LGROUP_REG ; \
1324 #error Unsupported architecture
1328 #define M_WAITERS_MSK 0x0000ffff
1329 #define M_PRIORITY_MSK 0x00ff0000
1330 #define M_ILOCKED_MSK 0x01000000
1331 #define M_MLOCKED_MSK 0x02000000
1332 #define M_PROMOTED_MSK 0x04000000
1333 #define M_SPIN_MSK 0x08000000
1338 * void lck_mtx_assert(lck_mtx_t* l, unsigned int)
1339 * Takes the address of a lock, and an assertion type as parameters.
1340 * The assertion can take one of two forms determine by the type
1341 * parameter: either the lock is held by the current thread, and the
1342 * type is LCK_MTX_ASSERT_OWNED, or it isn't and the type is
1343 * LCK_MTX_ASSERT_NOTOWNED. Calls panic on assertion failure.
1347 NONLEAF_ENTRY(lck_mtx_assert)
1348 LOAD_LMTX_REG(B_ARG0) /* Load lock address */
1349 mov %gs:CPU_ACTIVE_THREAD, LMTX_A_REG /* Load current thread */
1351 mov M_OWNER(LMTX_REG), LMTX_C_REG
1352 cmp $(MUTEX_IND), LMTX_C_REG /* Is this an indirect mutex? */
1353 cmove M_PTR(LMTX_REG), LMTX_REG /* If so, take indirection */
1355 mov M_OWNER(LMTX_REG), LMTX_C_REG /* Load owner */
1357 jne 2f /* Assert ownership? */
1358 cmp LMTX_A_REG, LMTX_C_REG /* Current thread match? */
1359 jne 3f /* no, go panic */
1360 testl $(M_ILOCKED_MSK | M_MLOCKED_MSK), M_STATE(LMTX_REG)
1362 1: /* yes, we own it */
1365 cmp LMTX_A_REG, LMTX_C_REG /* Current thread match? */
1366 jne 1b /* No, return */
1367 LOAD_REG_ARG1(LMTX_REG)
1368 LOAD_STRING_ARG0(mutex_assert_owned_str)
1371 LOAD_REG_ARG1(LMTX_REG)
1372 LOAD_STRING_ARG0(mutex_assert_not_owned_str)
1378 LOAD_REG_ARG1(LMTX_REG)
1379 LOAD_STRING_ARG0(mutex_interlock_destroyed_str)
1384 mutex_assert_not_owned_str:
1385 .asciz "mutex (%p) not owned\n"
1386 mutex_assert_owned_str:
1387 .asciz "mutex (%p) owned\n"
1388 mutex_interlock_destroyed_str:
1389 .asciz "trying to interlock destroyed mutex (%p)"
1396 * lck_mtx_try_lock()
1398 * lck_mtx_lock_spin()
1399 * lck_mtx_convert_spin()
1402 NONLEAF_ENTRY(lck_mtx_lock_spin)
1403 LOAD_LMTX_REG(B_ARG0) /* fetch lock pointer */
1405 CHECK_PREEMPTION_LEVEL()
1407 mov M_STATE(LMTX_REG), LMTX_C_REG32
1408 test $(M_ILOCKED_MSK), LMTX_C_REG /* is the interlock held */
1409 je Llmls_enter /* no - can't be INDIRECT or DESTROYED */
1411 mov M_OWNER(LMTX_REG), LMTX_A_REG
1412 cmp $(MUTEX_DESTROYED), LMTX_A_REG /* check to see if its marked destroyed */
1413 je lck_mtx_destroyed
1414 cmp $(MUTEX_IND), LMTX_A_REG /* Is this an indirect mutex */
1419 mov M_STATE(LMTX_REG), LMTX_C_REG32
1420 test $(M_SPIN_MSK), LMTX_C_REG
1426 mov M_STATE(LMTX_REG), LMTX_C_REG32
1428 test $(M_ILOCKED_MSK), LMTX_C_REG /* is the interlock held */
1431 test $(M_MLOCKED_MSK), LMTX_C_REG /* is the mutex locked */
1432 jne Llml_contended /* fall back to normal mutex handling */
1434 PUSHF /* save interrupt state */
1435 mov LMTX_C_REG, LMTX_A_REG /* eax contains snapshot for cmpxchgl */
1436 or $(M_ILOCKED_MSK | M_SPIN_MSK), LMTX_C_REG
1437 CLI /* disable interrupts */
1439 cmpxchg LMTX_C_REG32, M_STATE(LMTX_REG) /* atomic compare and exchange */
1442 mov %gs:CPU_ACTIVE_THREAD, LMTX_A_REG
1443 mov LMTX_A_REG, M_OWNER(LMTX_REG) /* record owner of interlock */
1446 POPF /* restore interrupt state */
1448 LMTX_CHK_EXTENDED_EXIT
1449 /* return with the interlock held and preemption disabled */
1452 LOCKSTAT_LABEL(_lck_mtx_lock_spin_lockstat_patch_point)
1454 /* inherit lock pointer in LMTX_REG above */
1455 LOCKSTAT_RECORD(LS_LCK_MTX_LOCK_SPIN_ACQUIRE, LMTX_REG)
1460 POPF /* restore interrupt state */
1465 NONLEAF_ENTRY(lck_mtx_lock)
1466 LOAD_LMTX_REG(B_ARG0) /* fetch lock pointer */
1468 CHECK_PREEMPTION_LEVEL()
1470 mov M_STATE(LMTX_REG), LMTX_C_REG32
1471 test $(M_ILOCKED_MSK), LMTX_C_REG /* is the interlock held */
1472 je Llml_enter /* no - can't be INDIRECT or DESTROYED */
1474 mov M_OWNER(LMTX_REG), LMTX_A_REG
1475 cmp $(MUTEX_DESTROYED), LMTX_A_REG /* check to see if its marked destroyed */
1476 je lck_mtx_destroyed
1477 cmp $(MUTEX_IND), LMTX_A_REG /* Is this an indirect mutex? */
1482 mov M_STATE(LMTX_REG), LMTX_C_REG32
1483 test $(M_SPIN_MSK), LMTX_C_REG
1489 mov M_STATE(LMTX_REG), LMTX_C_REG32
1491 test $(M_ILOCKED_MSK), LMTX_C_REG
1494 test $(M_MLOCKED_MSK), LMTX_C_REG
1495 jne Llml_contended /* mutex owned by someone else, go contend for it */
1497 mov LMTX_C_REG, LMTX_A_REG /* eax contains snapshot for cmpxchgl */
1498 or $(M_MLOCKED_MSK), LMTX_C_REG
1500 cmpxchg LMTX_C_REG32, M_STATE(LMTX_REG) /* atomic compare and exchange */
1503 mov %gs:CPU_ACTIVE_THREAD, LMTX_A_REG
1504 mov LMTX_A_REG, M_OWNER(LMTX_REG) /* record owner of mutex */
1507 testl $(M_WAITERS_MSK), M_STATE(LMTX_REG)
1510 LMTX_CALLEXT1(lck_mtx_lock_acquire_x86)
1512 LMTX_CHK_EXTENDED /* is this an extended mutex */
1517 LOCKSTAT_LABEL(_lck_mtx_lock_lockstat_patch_point)
1519 /* inherit lock pointer in LMTX_REG above */
1520 LOCKSTAT_RECORD(LS_LCK_MTX_LOCK_ACQUIRE, LMTX_REG)
1527 LOCKSTAT_LABEL(_lck_mtx_lock_ext_lockstat_patch_point)
1529 /* inherit lock pointer in LMTX_REG above */
1530 LOCKSTAT_RECORD(LS_LCK_MTX_EXT_LOCK_ACQUIRE, LMTX_REG)
1536 LMTX_CHK_EXTENDED /* is this an extended mutex */
1540 LMTX_CALLEXT1(lck_mtx_lock_spinwait_x86)
1542 test LMTX_RET_REG, LMTX_RET_REG
1543 je Llml_acquired /* acquired mutex */
1544 cmp $1, LMTX_RET_REG /* check for direct wait status */
1546 LMTX_CHK_EXTENDED /* is this an extended mutex */
1548 LMTX_UPDATE_DIRECT_WAIT
1550 mov M_STATE(LMTX_REG), LMTX_C_REG32
1551 test $(M_ILOCKED_MSK), LMTX_C_REG
1554 PUSHF /* save state of interrupt mask */
1555 mov LMTX_C_REG, LMTX_A_REG /* eax contains snapshot for cmpxchgl */
1556 or $(M_ILOCKED_MSK), LMTX_C_REG /* try to take the interlock */
1557 CLI /* disable interrupts */
1559 cmpxchg LMTX_C_REG32, M_STATE(LMTX_REG) /* atomic compare and exchange */
1562 test $(M_MLOCKED_MSK), LMTX_C_REG /* we've got the interlock and */
1564 or $(M_MLOCKED_MSK), LMTX_C_REG /* the mutex is free... grab it directly */
1565 and $(~M_ILOCKED_MSK), LMTX_C_REG
1567 mov %gs:CPU_ACTIVE_THREAD, LMTX_A_REG
1568 mov LMTX_A_REG, M_OWNER(LMTX_REG) /* record owner of mutex */
1569 mov LMTX_C_REG32, M_STATE(LMTX_REG) /* now drop the interlock */
1571 POPF /* restore interrupt state */
1573 3: /* interlock held, mutex busy */
1575 POPF /* restore interrupt state */
1577 LMTX_CHK_EXTENDED /* is this an extended mutex */
1581 LMTX_CALLEXT1(lck_mtx_lock_wait_x86)
1584 POPF /* restore interrupt state */
1591 NONLEAF_ENTRY(lck_mtx_try_lock_spin)
1592 LOAD_LMTX_REG(B_ARG0) /* fetch lock pointer */
1594 mov M_STATE(LMTX_REG), LMTX_C_REG32
1595 test $(M_ILOCKED_MSK), LMTX_C_REG /* is the interlock held */
1596 je Llmts_enter /* no - can't be INDIRECT or DESTROYED */
1598 mov M_OWNER(LMTX_REG), LMTX_A_REG
1599 cmp $(MUTEX_DESTROYED), LMTX_A_REG /* check to see if its marked destroyed */
1600 je lck_mtx_destroyed
1601 cmp $(MUTEX_IND), LMTX_A_REG /* Is this an indirect mutex? */
1607 mov M_STATE(LMTX_REG), LMTX_C_REG32
1609 test $(M_MLOCKED_MSK | M_SPIN_MSK), LMTX_C_REG
1611 test $(M_ILOCKED_MSK), LMTX_C_REG
1614 PUSHF /* save interrupt state */
1615 mov LMTX_C_REG, LMTX_A_REG /* eax contains snapshot for cmpxchgl */
1616 or $(M_ILOCKED_MSK | M_SPIN_MSK), LMTX_C_REG
1617 CLI /* disable interrupts */
1619 cmpxchg LMTX_C_REG32, M_STATE(LMTX_REG) /* atomic compare and exchange */
1622 mov %gs:CPU_ACTIVE_THREAD, LMTX_A_REG
1623 mov LMTX_A_REG, M_OWNER(LMTX_REG) /* record owner of mutex */
1626 POPF /* restore interrupt state */
1628 LMTX_CHK_EXTENDED_EXIT
1632 mov $1, LMTX_RET_REG /* return success */
1633 LOCKSTAT_LABEL(_lck_mtx_try_lock_spin_lockstat_patch_point)
1635 /* inherit lock pointer in LMTX_REG above */
1636 LOCKSTAT_RECORD(LS_LCK_MTX_TRY_SPIN_LOCK_ACQUIRE, LMTX_REG)
1638 mov $1, LMTX_RET_REG /* return success */
1641 POPF /* restore interrupt state */
1646 NONLEAF_ENTRY(lck_mtx_try_lock)
1647 LOAD_LMTX_REG(B_ARG0) /* fetch lock pointer */
1649 mov M_STATE(LMTX_REG), LMTX_C_REG32
1650 test $(M_ILOCKED_MSK), LMTX_C_REG /* is the interlock held */
1651 je Llmt_enter /* no - can't be INDIRECT or DESTROYED */
1653 mov M_OWNER(LMTX_REG), LMTX_A_REG
1654 cmp $(MUTEX_DESTROYED), LMTX_A_REG /* check to see if its marked destroyed */
1655 je lck_mtx_destroyed
1656 cmp $(MUTEX_IND), LMTX_A_REG /* Is this an indirect mutex? */
1662 mov M_STATE(LMTX_REG), LMTX_C_REG32
1664 test $(M_MLOCKED_MSK | M_SPIN_MSK), LMTX_C_REG
1666 test $(M_ILOCKED_MSK), LMTX_C_REG
1669 mov LMTX_C_REG, LMTX_A_REG /* eax contains snapshot for cmpxchgl */
1670 or $(M_MLOCKED_MSK), LMTX_C_REG
1672 cmpxchg LMTX_C_REG32, M_STATE(LMTX_REG) /* atomic compare and exchange */
1675 mov %gs:CPU_ACTIVE_THREAD, LMTX_A_REG
1676 mov LMTX_A_REG, M_OWNER(LMTX_REG) /* record owner of mutex */
1678 LMTX_CHK_EXTENDED_EXIT
1680 test $(M_WAITERS_MSK), LMTX_C_REG
1682 LMTX_CALLEXT1(lck_mtx_lock_acquire_x86)
1687 mov $1, LMTX_RET_REG /* return success */
1688 /* Dtrace probe: LS_LCK_MTX_TRY_LOCK_ACQUIRE */
1689 LOCKSTAT_LABEL(_lck_mtx_try_lock_lockstat_patch_point)
1691 /* inherit lock pointer in LMTX_REG from above */
1692 LOCKSTAT_RECORD(LS_LCK_MTX_TRY_LOCK_ACQUIRE, LMTX_REG)
1694 mov $1, LMTX_RET_REG /* return success */
1700 LMTX_CHK_EXTENDED /* is this an extended mutex */
1705 xor LMTX_RET_REG, LMTX_RET_REG
1710 NONLEAF_ENTRY(lck_mtx_convert_spin)
1711 LOAD_LMTX_REG(B_ARG0) /* fetch lock pointer */
1713 mov M_OWNER(LMTX_REG), LMTX_A_REG
1714 cmp $(MUTEX_IND), LMTX_A_REG /* Is this an indirect mutex? */
1715 cmove M_PTR(LMTX_REG), LMTX_REG /* If so, take indirection */
1717 mov M_STATE(LMTX_REG), LMTX_C_REG32
1718 test $(M_MLOCKED_MSK), LMTX_C_REG /* already owned as a mutex, just return */
1721 and $(~(M_ILOCKED_MSK | M_SPIN_MSK)), LMTX_C_REG /* convert from spin version to mutex */
1722 or $(M_MLOCKED_MSK), LMTX_C_REG
1723 mov LMTX_C_REG32, M_STATE(LMTX_REG) /* since I own the interlock, I don't need an atomic update */
1725 PREEMPTION_ENABLE /* only %eax is consumed */
1727 test $(M_WAITERS_MSK), LMTX_C_REG /* are there any waiters? */
1730 LMTX_CALLEXT1(lck_mtx_lock_acquire_x86)
1735 #if defined(__i386__)
1736 NONLEAF_ENTRY(lck_mtx_unlock)
1737 LOAD_LMTX_REG(B_ARG0) /* fetch lock pointer */
1738 mov M_OWNER(LMTX_REG), LMTX_A_REG
1739 test LMTX_A_REG, LMTX_A_REG
1743 NONLEAF_ENTRY(lck_mtx_unlock_darwin10)
1745 NONLEAF_ENTRY(lck_mtx_unlock)
1747 LOAD_LMTX_REG(B_ARG0) /* fetch lock pointer */
1748 mov M_OWNER(LMTX_REG), LMTX_A_REG
1750 cmp $(MUTEX_IND), LMTX_A_REG /* Is this an indirect mutex? */
1753 mov M_STATE(LMTX_REG), LMTX_C_REG32
1754 test $(M_MLOCKED_MSK), LMTX_C_REG /* check for full mutex */
1757 xor LMTX_A_REG, LMTX_A_REG
1758 mov LMTX_A_REG, M_OWNER(LMTX_REG)
1759 mov LMTX_C_REG, LMTX_A_REG /* keep original state in %ecx for later evaluation */
1760 and $(~(M_ILOCKED_MSK | M_SPIN_MSK | M_PROMOTED_MSK)), LMTX_A_REG
1761 mov LMTX_A_REG32, M_STATE(LMTX_REG) /* since I own the interlock, I don't need an atomic update */
1763 PREEMPTION_ENABLE /* need to re-enable preemption - clobbers eax */
1766 test $(M_ILOCKED_MSK), LMTX_C_REG /* have to wait for interlock to clear */
1769 PUSHF /* save interrupt state */
1770 mov LMTX_C_REG, LMTX_A_REG /* eax contains snapshot for cmpxchgl */
1771 and $(~M_MLOCKED_MSK), LMTX_C_REG /* drop mutex */
1772 or $(M_ILOCKED_MSK), LMTX_C_REG /* pick up interlock */
1775 cmpxchg LMTX_C_REG32, M_STATE(LMTX_REG) /* atomic compare and exchange */
1776 jne 6f /* branch on failure to spin loop */
1778 xor LMTX_A_REG, LMTX_A_REG
1779 mov LMTX_A_REG, M_OWNER(LMTX_REG)
1780 mov LMTX_C_REG, LMTX_A_REG /* keep original state in %ecx for later evaluation */
1781 and $(~(M_ILOCKED_MSK | M_PROMOTED_MSK)), LMTX_A_REG
1782 mov LMTX_A_REG32, M_STATE(LMTX_REG) /* since I own the interlock, I don't need an atomic update */
1783 POPF /* restore interrupt state */
1785 test $(M_PROMOTED_MSK | M_WAITERS_MSK), LMTX_C_REG
1787 and $(M_PROMOTED_MSK), LMTX_C_REG
1789 LMTX_CALLEXT2(lck_mtx_unlock_wakeup_x86, LMTX_C_REG)
1796 /* Dtrace: LS_LCK_MTX_UNLOCK_RELEASE */
1797 LOCKSTAT_LABEL(_lck_mtx_unlock_lockstat_patch_point)
1799 /* inherit lock pointer in LMTX_REG from above */
1800 LOCKSTAT_RECORD(LS_LCK_MTX_UNLOCK_RELEASE, LMTX_REG)
1806 /* Dtrace: LS_LCK_MTX_EXT_UNLOCK_RELEASE */
1807 LOCKSTAT_LABEL(_lck_mtx_ext_unlock_lockstat_patch_point)
1809 /* inherit lock pointer in LMTX_REG from above */
1810 LOCKSTAT_RECORD(LS_LCK_MTX_EXT_UNLOCK_RELEASE, LMTX_REG)
1814 POPF /* restore interrupt state */
1817 mov M_STATE(LMTX_REG), LMTX_C_REG32
1820 mov M_PTR(LMTX_REG), LMTX_REG
1821 mov M_OWNER(LMTX_REG), LMTX_A_REG
1822 mov %gs:CPU_ACTIVE_THREAD, LMTX_C_REG
1823 CHECK_UNLOCK(LMTX_C_REG, LMTX_A_REG)
1827 LEAF_ENTRY(lck_mtx_lock_decr_waiter)
1828 LOAD_LMTX_REG(L_ARG0) /* fetch lock pointer - no indirection here */
1830 mov M_STATE(LMTX_REG), LMTX_C_REG32
1832 test $(M_WAITERS_MSK), LMTX_C_REG
1834 test $(M_ILOCKED_MSK), LMTX_C_REG /* have to wait for interlock to clear */
1837 mov LMTX_C_REG, LMTX_A_REG /* eax contains snapshot for cmpxchgl */
1838 dec LMTX_C_REG /* decrement waiter count */
1840 cmpxchg LMTX_C_REG32, M_STATE(LMTX_REG) /* atomic compare and exchange */
1841 jne 3f /* branch on failure to spin loop */
1843 mov $1, LMTX_RET_REG
1846 xor LMTX_RET_REG, LMTX_RET_REG
1854 LEAF_ENTRY(lck_mtx_lock_get_pri)
1855 LOAD_LMTX_REG(L_ARG0) /* fetch lock pointer - no indirection here */
1857 mov M_STATE(LMTX_REG), LMTX_C_REG32
1859 test $(M_WAITERS_MSK), LMTX_C_REG
1861 test $(M_ILOCKED_MSK), LMTX_C_REG /* have to wait for interlock to clear */
1864 mov LMTX_C_REG, LMTX_A_REG /* eax contains snapshot for cmpxchgl */
1865 and $(~M_PRIORITY_MSK), LMTX_C_REG /* no waiters, reset mutex priority to 0 */
1867 cmpxchg LMTX_C_REG32, M_STATE(LMTX_REG) /* atomic compare and exchange */
1868 jne 3f /* branch on failure to spin loop */
1870 xor LMTX_RET_REG, LMTX_RET_REG /* return mutex priority == 0 */
1873 mov LMTX_C_REG, LMTX_RET_REG
1874 and $(M_PRIORITY_MSK), LMTX_RET_REG
1875 shr $16, LMTX_RET_REG /* return current mutex priority */
1884 LEAF_ENTRY(lck_mtx_ilk_unlock)
1885 LOAD_LMTX_REG(L_ARG0) /* fetch lock pointer - no indirection here */
1887 andl $(~M_ILOCKED_MSK), M_STATE(LMTX_REG)
1889 PREEMPTION_ENABLE /* need to re-enable preemption */
1895 LEAF_ENTRY(lck_mtx_lock_grab_mutex)
1896 LOAD_LMTX_REG(L_ARG0) /* fetch lock pointer - no indirection here */
1898 mov M_STATE(LMTX_REG), LMTX_C_REG32
1900 test $(M_ILOCKED_MSK | M_MLOCKED_MSK), LMTX_C_REG /* can't have the mutex yet */
1903 mov LMTX_C_REG, LMTX_A_REG /* eax contains snapshot for cmpxchgl */
1904 or $(M_MLOCKED_MSK), LMTX_C_REG
1906 cmpxchg LMTX_C_REG32, M_STATE(LMTX_REG) /* atomic compare and exchange */
1907 jne 2f /* branch on failure to spin loop */
1909 mov %gs:CPU_ACTIVE_THREAD, LMTX_A_REG
1910 mov LMTX_A_REG, M_OWNER(LMTX_REG) /* record owner of mutex */
1912 mov $1, LMTX_RET_REG /* return success */
1915 xor LMTX_RET_REG, LMTX_RET_REG /* return failure */
1920 LEAF_ENTRY(lck_mtx_lock_mark_promoted)
1921 LOAD_LMTX_REG(L_ARG0) /* fetch lock pointer - no indirection here */
1923 mov M_STATE(LMTX_REG), LMTX_C_REG32
1925 test $(M_PROMOTED_MSK), LMTX_C_REG
1927 test $(M_ILOCKED_MSK), LMTX_C_REG /* have to wait for interlock to clear */
1930 mov LMTX_C_REG, LMTX_A_REG /* eax contains snapshot for cmpxchgl */
1931 or $(M_PROMOTED_MSK), LMTX_C_REG
1933 cmpxchg LMTX_C_REG32, M_STATE(LMTX_REG) /* atomic compare and exchange */
1934 jne 2f /* branch on failure to spin loop */
1936 mov $1, LMTX_RET_REG
1942 xor LMTX_RET_REG, LMTX_RET_REG
1947 LEAF_ENTRY(lck_mtx_lock_mark_destroyed)
1948 LOAD_LMTX_REG(L_ARG0)
1950 mov M_OWNER(LMTX_REG), LMTX_A_REG
1952 cmp $(MUTEX_DESTROYED), LMTX_A_REG /* check to see if its marked destroyed */
1954 cmp $(MUTEX_IND), LMTX_A_REG /* Is this an indirect mutex? */
1957 movl $(MUTEX_DESTROYED), M_OWNER(LMTX_REG) /* convert to destroyed state */
1960 mov M_STATE(LMTX_REG), LMTX_C_REG32
1962 test $(M_ILOCKED_MSK), LMTX_C_REG /* have to wait for interlock to clear */
1965 PUSHF /* save interrupt state */
1966 mov LMTX_C_REG, LMTX_A_REG /* eax contains snapshot for cmpxchgl */
1967 or $(M_ILOCKED_MSK), LMTX_C_REG
1970 cmpxchg LMTX_C_REG32, M_STATE(LMTX_REG) /* atomic compare and exchange */
1971 jne 4f /* branch on failure to spin loop */
1972 movl $(MUTEX_DESTROYED), M_OWNER(LMTX_REG) /* convert to destroyed state */
1973 POPF /* restore interrupt state */
1975 LEAF_RET /* return with M_ILOCKED set */
1977 POPF /* restore interrupt state */
1984 LEAF_ENTRY(_disable_preemption)
1987 #endif /* MACH_RT */
1990 LEAF_ENTRY(_enable_preemption)
1993 cmpl $0,%gs:CPU_PREEMPTION_LEVEL
1996 pushl %gs:CPU_PREEMPTION_LEVEL
1998 movl %gs:CPU_PREEMPTION_LEVEL,%esi
2000 LOAD_STRING_ARG0(_enable_preemption_less_than_zero)
2004 _enable_preemption_less_than_zero:
2005 .asciz "_enable_preemption: preemption_level(%d) < 0!"
2008 #endif /* MACH_ASSERT */
2010 #endif /* MACH_RT */
2013 LEAF_ENTRY(_enable_preemption_no_check)
2016 cmpl $0,%gs:CPU_PREEMPTION_LEVEL
2018 LOAD_STRING_ARG0(_enable_preemption_no_check_less_than_zero)
2022 _enable_preemption_no_check_less_than_zero:
2023 .asciz "_enable_preemption_no_check: preemption_level <= 0!"
2026 #endif /* MACH_ASSERT */
2027 _ENABLE_PREEMPTION_NO_CHECK
2028 #endif /* MACH_RT */
2032 LEAF_ENTRY(_mp_disable_preemption)
2035 #endif /* MACH_RT */
2038 LEAF_ENTRY(_mp_enable_preemption)
2041 cmpl $0,%gs:CPU_PREEMPTION_LEVEL
2044 pushl %gs:CPU_PREEMPTION_LEVEL
2046 movl %gs:CPU_PREEMPTION_LEVEL,%esi
2048 LOAD_STRING_ARG0(_mp_enable_preemption_less_than_zero)
2052 _mp_enable_preemption_less_than_zero:
2053 .asciz "_mp_enable_preemption: preemption_level (%d) <= 0!"
2056 #endif /* MACH_ASSERT */
2058 #endif /* MACH_RT */
2061 LEAF_ENTRY(_mp_enable_preemption_no_check)
2064 cmpl $0,%gs:CPU_PREEMPTION_LEVEL
2066 LOAD_STRING_ARG0(_mp_enable_preemption_no_check_less_than_zero)
2070 _mp_enable_preemption_no_check_less_than_zero:
2071 .asciz "_mp_enable_preemption_no_check: preemption_level <= 0!"
2074 #endif /* MACH_ASSERT */
2075 _ENABLE_PREEMPTION_NO_CHECK
2076 #endif /* MACH_RT */
2081 LEAF_ENTRY(i_bit_set)
2088 LEAF_ENTRY(i_bit_clear)
2096 LEAF_ENTRY(bit_lock)
2106 LEAF_ENTRY(bit_lock_try)
2112 LEAF_RET /* %eax better not be null ! */
2117 LEAF_ENTRY(bit_unlock)
2125 * Atomic primitives, prototyped in kern/simple_lock.h
2127 LEAF_ENTRY(hw_atomic_add)
2128 movl L_ARG0, %ecx /* Load address of operand */
2129 movl L_ARG1, %eax /* Load addend */
2132 xaddl %eax, (%ecx) /* Atomic exchange and add */
2133 addl %edx, %eax /* Calculate result */
2136 LEAF_ENTRY(hw_atomic_sub)
2137 movl L_ARG0, %ecx /* Load address of operand */
2138 movl L_ARG1, %eax /* Load subtrahend */
2142 xaddl %eax, (%ecx) /* Atomic exchange and add */
2143 addl %edx, %eax /* Calculate result */
2146 LEAF_ENTRY(hw_atomic_or)
2147 movl L_ARG0, %ecx /* Load address of operand */
2150 movl L_ARG1, %edx /* Load mask */
2153 cmpxchgl %edx, (%ecx) /* Atomic CAS */
2155 movl %edx, %eax /* Result */
2158 * A variant of hw_atomic_or which doesn't return a value.
2159 * The implementation is thus comparatively more efficient.
2162 LEAF_ENTRY(hw_atomic_or_noret)
2163 movl L_ARG0, %ecx /* Load address of operand */
2164 movl L_ARG1, %edx /* Load mask */
2166 orl %edx, (%ecx) /* Atomic OR */
2169 LEAF_ENTRY(hw_atomic_and)
2170 movl L_ARG0, %ecx /* Load address of operand */
2173 movl L_ARG1, %edx /* Load mask */
2176 cmpxchgl %edx, (%ecx) /* Atomic CAS */
2178 movl %edx, %eax /* Result */
2181 * A variant of hw_atomic_and which doesn't return a value.
2182 * The implementation is thus comparatively more efficient.
2185 LEAF_ENTRY(hw_atomic_and_noret)
2186 movl L_ARG0, %ecx /* Load address of operand */
2187 movl L_ARG1, %edx /* Load mask */
2189 andl %edx, (%ecx) /* Atomic AND */
2192 #else /* !__i386__ */
2194 LEAF_ENTRY(i_bit_set)
2199 LEAF_ENTRY(i_bit_clear)
2205 LEAF_ENTRY(bit_lock)
2213 LEAF_ENTRY(bit_lock_try)
2223 LEAF_ENTRY(bit_unlock)
2230 * Atomic primitives, prototyped in kern/simple_lock.h
2232 LEAF_ENTRY(hw_atomic_add)
2233 movl %esi, %eax /* Load addend */
2235 xaddl %eax, (%rdi) /* Atomic exchange and add */
2236 addl %esi, %eax /* Calculate result */
2239 LEAF_ENTRY(hw_atomic_sub)
2243 xaddl %eax, (%rdi) /* Atomic exchange and add */
2244 addl %esi, %eax /* Calculate result */
2247 LEAF_ENTRY(hw_atomic_or)
2250 movl %esi, %edx /* Load mask */
2253 cmpxchgl %edx, (%rdi) /* Atomic CAS */
2255 movl %edx, %eax /* Result */
2258 * A variant of hw_atomic_or which doesn't return a value.
2259 * The implementation is thus comparatively more efficient.
2262 LEAF_ENTRY(hw_atomic_or_noret)
2264 orl %esi, (%rdi) /* Atomic OR */
2268 LEAF_ENTRY(hw_atomic_and)
2271 movl %esi, %edx /* Load mask */
2274 cmpxchgl %edx, (%rdi) /* Atomic CAS */
2276 movl %edx, %eax /* Result */
2279 * A variant of hw_atomic_and which doesn't return a value.
2280 * The implementation is thus comparatively more efficient.
2283 LEAF_ENTRY(hw_atomic_and_noret)
2285 andl %esi, (%rdi) /* Atomic OR */
2288 #endif /* !__i386 __ */