]> git.saurik.com Git - apple/xnu.git/blobdiff - iokit/Kernel/IOLocks.cpp
xnu-3789.31.2.tar.gz
[apple/xnu.git] / iokit / Kernel / IOLocks.cpp
index a61f6a326ac7d3bd1509fcb4d02d1060fff514f7..c2ece6b9c79a9d41ea97c0f6a16a52cccb9a61a9 100644 (file)
 extern "C" {
 #include <kern/locks.h>
 
+#if defined(__x86_64__)
+/* Synthetic event if none is specified, for backwards compatibility only. */
+static bool IOLockSleep_NO_EVENT __attribute__((used)) = 0;
+#endif
+
 void   IOLockInitWithState( IOLock * lock, IOLockState state)
 {
     if( state == kIOLockStateLocked)
@@ -60,13 +65,13 @@ lck_mtx_t * IOLockGetMachLock( IOLock * lock)
 
 int    IOLockSleep( IOLock * lock, void *event, UInt32 interType)
 {
-    return (int) lck_mtx_sleep(lock, LCK_SLEEP_DEFAULT, (event_t) event, (wait_interrupt_t) interType);
+    return (int) lck_mtx_sleep(lock, LCK_SLEEP_PROMOTED_PRI, (event_t) event, (wait_interrupt_t) interType);
 }
 
 int    IOLockSleepDeadline( IOLock * lock, void *event,
                                 AbsoluteTime deadline, UInt32 interType)
 {
-    return (int) lck_mtx_sleep_deadline(lock, LCK_SLEEP_DEFAULT, (event_t) event,
+    return (int) lck_mtx_sleep_deadline(lock, LCK_SLEEP_PROMOTED_PRI, (event_t) event,
                                        (wait_interrupt_t) interType, __OSAbsoluteTime(deadline));
 }
 
@@ -75,9 +80,45 @@ void IOLockWakeup(IOLock * lock, void *event, bool oneThread)
        thread_wakeup_prim((event_t) event, oneThread, THREAD_AWAKENED);
 }   
 
+#if defined(__x86_64__)
+/*
+ * For backwards compatibility, kexts built against pre-Darwin 14 headers will bind at runtime to this function,
+ * which supports a NULL event,
+ */
+int    IOLockSleep_legacy_x86_64( IOLock * lock, void *event, UInt32 interType) __asm("_IOLockSleep");
+int    IOLockSleepDeadline_legacy_x86_64( IOLock * lock, void *event,
+                                          AbsoluteTime deadline, UInt32 interType) __asm("_IOLockSleepDeadline");
+void   IOLockWakeup_legacy_x86_64(IOLock * lock, void *event, bool oneThread) __asm("_IOLockWakeup");
+
+int    IOLockSleep_legacy_x86_64( IOLock * lock, void *event, UInt32 interType)
+{
+    if (event == NULL)
+        event = (void *)&IOLockSleep_NO_EVENT;
+
+    return IOLockSleep(lock, event, interType);
+}
+
+int    IOLockSleepDeadline_legacy_x86_64( IOLock * lock, void *event,
+                            AbsoluteTime deadline, UInt32 interType)
+{
+    if (event == NULL)
+        event = (void *)&IOLockSleep_NO_EVENT;
+
+    return IOLockSleepDeadline(lock, event, deadline, interType);
+}
+
+void   IOLockWakeup_legacy_x86_64(IOLock * lock, void *event, bool oneThread)
+{   
+    if (event == NULL)
+        event = (void *)&IOLockSleep_NO_EVENT;
+
+    IOLockWakeup(lock, event, oneThread);
+}   
+#endif /* defined(__x86_64__) */
+
 
 struct _IORecursiveLock {
-       lck_mtx_t       *mutex;
+       lck_mtx_t       mutex;
        lck_grp_t       *group;
        thread_t        thread;
        UInt32          count;
@@ -94,15 +135,10 @@ IORecursiveLock * IORecursiveLockAllocWithLockGroup( lck_grp_t * lockGroup )
     if( !lock )
         return( 0 );
 
-    lock->mutex = lck_mtx_alloc_init( lockGroup, LCK_ATTR_NULL );
-    if( lock->mutex ) {
-               lock->group = lockGroup;
-        lock->thread = 0;
-        lock->count  = 0;
-    } else {
-        IODelete( lock, _IORecursiveLock, 1 );
-        lock = 0;
-    }
+    lck_mtx_init( &lock->mutex, lockGroup, LCK_ATTR_NULL );
+    lock->group = lockGroup;
+    lock->thread = 0;
+    lock->count  = 0;
 
     return( (IORecursiveLock *) lock );
 }
@@ -117,13 +153,13 @@ void IORecursiveLockFree( IORecursiveLock * _lock )
 {
     _IORecursiveLock * lock = (_IORecursiveLock *)_lock;
        
-    lck_mtx_free( lock->mutex, lock->group );
+    lck_mtx_destroy(&lock->mutex, lock->group);
     IODelete( lock, _IORecursiveLock, 1 );
 }
 
 lck_mtx_t * IORecursiveLockGetMachLock( IORecursiveLock * lock )
 {
-    return( lock->mutex );
+    return( &lock->mutex );
 }
 
 void IORecursiveLockLock( IORecursiveLock * _lock)
@@ -133,7 +169,7 @@ void IORecursiveLockLock( IORecursiveLock * _lock)
     if( lock->thread == IOThreadSelf())
         lock->count++;
     else {
-        lck_mtx_lock( lock->mutex );
+        lck_mtx_lock( &lock->mutex );
         assert( lock->thread == 0 );
         assert( lock->count == 0 );
         lock->thread = IOThreadSelf();
@@ -149,7 +185,7 @@ boolean_t IORecursiveLockTryLock( IORecursiveLock * _lock)
         lock->count++;
        return( true );
     } else {
-        if( lck_mtx_try_lock( lock->mutex )) {
+        if( lck_mtx_try_lock( &lock->mutex )) {
             assert( lock->thread == 0 );
             assert( lock->count == 0 );
             lock->thread = IOThreadSelf();
@@ -168,7 +204,7 @@ void IORecursiveLockUnlock( IORecursiveLock * _lock)
 
     if( 0 == (--lock->count)) {
         lock->thread = 0;
-        lck_mtx_unlock( lock->mutex );
+        lck_mtx_unlock( &lock->mutex );
     }
 }
 
@@ -189,7 +225,7 @@ int IORecursiveLockSleep(IORecursiveLock *_lock, void *event, UInt32 interType)
     
     lock->count = 0;
     lock->thread = 0;
-    res = lck_mtx_sleep(lock->mutex, LCK_SLEEP_DEFAULT, (event_t) event, (wait_interrupt_t) interType);
+    res = lck_mtx_sleep(&lock->mutex, LCK_SLEEP_PROMOTED_PRI, (event_t) event, (wait_interrupt_t) interType);
 
     // Must re-establish the recursive lock no matter why we woke up
     // otherwise we would potentially leave the return path corrupted.
@@ -211,8 +247,8 @@ int IORecursiveLockSleepDeadline( IORecursiveLock * _lock, void *event,
     
     lock->count = 0;
     lock->thread = 0;
-    res = lck_mtx_sleep_deadline(lock->mutex, LCK_SLEEP_DEFAULT, (event_t) event, 
-                                                                (wait_interrupt_t) interType, __OSAbsoluteTime(deadline));
+    res = lck_mtx_sleep_deadline(&lock->mutex, LCK_SLEEP_PROMOTED_PRI, (event_t) event,
+                                                                     (wait_interrupt_t) interType, __OSAbsoluteTime(deadline));
 
     // Must re-establish the recursive lock no matter why we woke up
     // otherwise we would potentially leave the return path corrupted.