]> git.saurik.com Git - apple/xnu.git/blob - iokit/Kernel/IOLocks.cpp
xnu-4570.51.1.tar.gz
[apple/xnu.git] / iokit / Kernel / IOLocks.cpp
1 /*
2 * Copyright (c) 1998-2007 Apple Inc. All rights reserved.
3 *
4 * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
5 *
6 * This file contains Original Code and/or Modifications of Original Code
7 * as defined in and that are subject to the Apple Public Source License
8 * Version 2.0 (the 'License'). You may not use this file except in
9 * compliance with the License. The rights granted to you under the License
10 * may not be used to create, or enable the creation or redistribution of,
11 * unlawful or unlicensed copies of an Apple operating system, or to
12 * circumvent, violate, or enable the circumvention or violation of, any
13 * terms of an Apple operating system software license agreement.
14 *
15 * Please obtain a copy of the License at
16 * http://www.opensource.apple.com/apsl/ and read it before using this file.
17 *
18 * The Original Code and all software distributed under the License are
19 * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
20 * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
21 * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
22 * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
23 * Please see the License for the specific language governing rights and
24 * limitations under the License.
25 *
26 * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
27 */
28
29 #include <IOKit/system.h>
30
31 #include <IOKit/IOReturn.h>
32 #include <IOKit/IOLib.h>
33 #include <IOKit/assert.h>
34
35 #include <IOKit/IOLocksPrivate.h>
36
37 extern "C" {
38 #include <kern/locks.h>
39
40 #if defined(__x86_64__)
41 /* Synthetic event if none is specified, for backwards compatibility only. */
42 static bool IOLockSleep_NO_EVENT __attribute__((used)) = 0;
43 #endif
44
45 void IOLockInitWithState( IOLock * lock, IOLockState state)
46 {
47 if( state == kIOLockStateLocked)
48 lck_mtx_lock( lock);
49 }
50
51 IOLock * IOLockAlloc( void )
52 {
53 return( lck_mtx_alloc_init(IOLockGroup, LCK_ATTR_NULL) );
54 }
55
56 void IOLockFree( IOLock * lock)
57 {
58 lck_mtx_free( lock, IOLockGroup);
59 }
60
61 lck_mtx_t * IOLockGetMachLock( IOLock * lock)
62 {
63 return( (lck_mtx_t *)lock);
64 }
65
66 int IOLockSleep( IOLock * lock, void *event, UInt32 interType)
67 {
68 return (int) lck_mtx_sleep(lock, LCK_SLEEP_PROMOTED_PRI, (event_t) event, (wait_interrupt_t) interType);
69 }
70
71 int IOLockSleepDeadline( IOLock * lock, void *event,
72 AbsoluteTime deadline, UInt32 interType)
73 {
74 return (int) lck_mtx_sleep_deadline(lock, LCK_SLEEP_PROMOTED_PRI, (event_t) event,
75 (wait_interrupt_t) interType, __OSAbsoluteTime(deadline));
76 }
77
78 void IOLockWakeup(IOLock * lock, void *event, bool oneThread)
79 {
80 thread_wakeup_prim((event_t) event, oneThread, THREAD_AWAKENED);
81 }
82
83
84 #if defined(__x86_64__)
85 /*
86 * For backwards compatibility, kexts built against pre-Darwin 14 headers will bind at runtime to this function,
87 * which supports a NULL event,
88 */
89 int IOLockSleep_legacy_x86_64( IOLock * lock, void *event, UInt32 interType) __asm("_IOLockSleep");
90 int IOLockSleepDeadline_legacy_x86_64( IOLock * lock, void *event,
91 AbsoluteTime deadline, UInt32 interType) __asm("_IOLockSleepDeadline");
92 void IOLockWakeup_legacy_x86_64(IOLock * lock, void *event, bool oneThread) __asm("_IOLockWakeup");
93
94 int IOLockSleep_legacy_x86_64( IOLock * lock, void *event, UInt32 interType)
95 {
96 if (event == NULL)
97 event = (void *)&IOLockSleep_NO_EVENT;
98
99 return IOLockSleep(lock, event, interType);
100 }
101
102 int IOLockSleepDeadline_legacy_x86_64( IOLock * lock, void *event,
103 AbsoluteTime deadline, UInt32 interType)
104 {
105 if (event == NULL)
106 event = (void *)&IOLockSleep_NO_EVENT;
107
108 return IOLockSleepDeadline(lock, event, deadline, interType);
109 }
110
111 void IOLockWakeup_legacy_x86_64(IOLock * lock, void *event, bool oneThread)
112 {
113 if (event == NULL)
114 event = (void *)&IOLockSleep_NO_EVENT;
115
116 IOLockWakeup(lock, event, oneThread);
117 }
118 #endif /* defined(__x86_64__) */
119
120
121 struct _IORecursiveLock {
122 lck_mtx_t mutex;
123 lck_grp_t *group;
124 thread_t thread;
125 UInt32 count;
126 };
127
128 IORecursiveLock * IORecursiveLockAllocWithLockGroup( lck_grp_t * lockGroup )
129 {
130 _IORecursiveLock * lock;
131
132 if( lockGroup == 0 )
133 return( 0 );
134
135 lock = IONew( _IORecursiveLock, 1 );
136 if( !lock )
137 return( 0 );
138
139 lck_mtx_init( &lock->mutex, lockGroup, LCK_ATTR_NULL );
140 lock->group = lockGroup;
141 lock->thread = 0;
142 lock->count = 0;
143
144 return( (IORecursiveLock *) lock );
145 }
146
147
148 IORecursiveLock * IORecursiveLockAlloc( void )
149 {
150 return IORecursiveLockAllocWithLockGroup( IOLockGroup );
151 }
152
153 void IORecursiveLockFree( IORecursiveLock * _lock )
154 {
155 _IORecursiveLock * lock = (_IORecursiveLock *)_lock;
156
157 lck_mtx_destroy(&lock->mutex, lock->group);
158 IODelete( lock, _IORecursiveLock, 1 );
159 }
160
161 lck_mtx_t * IORecursiveLockGetMachLock( IORecursiveLock * lock )
162 {
163 return( &lock->mutex );
164 }
165
166 void IORecursiveLockLock( IORecursiveLock * _lock)
167 {
168 _IORecursiveLock * lock = (_IORecursiveLock *)_lock;
169
170 if( lock->thread == IOThreadSelf())
171 lock->count++;
172 else {
173 lck_mtx_lock( &lock->mutex );
174 assert( lock->thread == 0 );
175 assert( lock->count == 0 );
176 lock->thread = IOThreadSelf();
177 lock->count = 1;
178 }
179 }
180
181 boolean_t IORecursiveLockTryLock( IORecursiveLock * _lock)
182 {
183 _IORecursiveLock * lock = (_IORecursiveLock *)_lock;
184
185 if( lock->thread == IOThreadSelf()) {
186 lock->count++;
187 return( true );
188 } else {
189 if( lck_mtx_try_lock( &lock->mutex )) {
190 assert( lock->thread == 0 );
191 assert( lock->count == 0 );
192 lock->thread = IOThreadSelf();
193 lock->count = 1;
194 return( true );
195 }
196 }
197 return( false );
198 }
199
200 void IORecursiveLockUnlock( IORecursiveLock * _lock)
201 {
202 _IORecursiveLock * lock = (_IORecursiveLock *)_lock;
203
204 assert( lock->thread == IOThreadSelf() );
205
206 if( 0 == (--lock->count)) {
207 lock->thread = 0;
208 lck_mtx_unlock( &lock->mutex );
209 }
210 }
211
212 boolean_t IORecursiveLockHaveLock( const IORecursiveLock * _lock)
213 {
214 _IORecursiveLock * lock = (_IORecursiveLock *)_lock;
215
216 return( lock->thread == IOThreadSelf());
217 }
218
219 int IORecursiveLockSleep(IORecursiveLock *_lock, void *event, UInt32 interType)
220 {
221 _IORecursiveLock * lock = (_IORecursiveLock *)_lock;
222 UInt32 count = lock->count;
223 int res;
224
225 assert(lock->thread == IOThreadSelf());
226
227 lock->count = 0;
228 lock->thread = 0;
229 res = lck_mtx_sleep(&lock->mutex, LCK_SLEEP_PROMOTED_PRI, (event_t) event, (wait_interrupt_t) interType);
230
231 // Must re-establish the recursive lock no matter why we woke up
232 // otherwise we would potentially leave the return path corrupted.
233 assert(lock->thread == 0);
234 assert(lock->count == 0);
235 lock->thread = IOThreadSelf();
236 lock->count = count;
237 return res;
238 }
239
240 int IORecursiveLockSleepDeadline( IORecursiveLock * _lock, void *event,
241 AbsoluteTime deadline, UInt32 interType)
242 {
243 _IORecursiveLock * lock = (_IORecursiveLock *)_lock;
244 UInt32 count = lock->count;
245 int res;
246
247 assert(lock->thread == IOThreadSelf());
248
249 lock->count = 0;
250 lock->thread = 0;
251 res = lck_mtx_sleep_deadline(&lock->mutex, LCK_SLEEP_PROMOTED_PRI, (event_t) event,
252 (wait_interrupt_t) interType, __OSAbsoluteTime(deadline));
253
254 // Must re-establish the recursive lock no matter why we woke up
255 // otherwise we would potentially leave the return path corrupted.
256 assert(lock->thread == 0);
257 assert(lock->count == 0);
258 lock->thread = IOThreadSelf();
259 lock->count = count;
260 return res;
261 }
262
263 void IORecursiveLockWakeup(IORecursiveLock *, void *event, bool oneThread)
264 {
265 thread_wakeup_prim((event_t) event, oneThread, THREAD_AWAKENED);
266 }
267
268 /*
269 * Complex (read/write) lock operations
270 */
271
272 IORWLock * IORWLockAlloc( void )
273 {
274 return( lck_rw_alloc_init(IOLockGroup, LCK_ATTR_NULL) );
275 }
276
277 void IORWLockFree( IORWLock * lock)
278 {
279 lck_rw_free( lock, IOLockGroup);
280 }
281
282 lck_rw_t * IORWLockGetMachLock( IORWLock * lock)
283 {
284 return( (lck_rw_t *)lock);
285 }
286
287
288 /*
289 * Spin locks
290 */
291
292 IOSimpleLock * IOSimpleLockAlloc( void )
293 {
294 return( lck_spin_alloc_init( IOLockGroup, LCK_ATTR_NULL) );
295 }
296
297 void IOSimpleLockInit( IOSimpleLock * lock)
298 {
299 lck_spin_init( lock, IOLockGroup, LCK_ATTR_NULL);
300 }
301
302 void IOSimpleLockFree( IOSimpleLock * lock )
303 {
304 lck_spin_free( lock, IOLockGroup);
305 }
306
307 lck_spin_t * IOSimpleLockGetMachLock( IOSimpleLock * lock)
308 {
309 return( (lck_spin_t *)lock);
310 }
311
312 #ifndef IOLOCKS_INLINE
313 /*
314 * Lock assertions
315 */
316
317 void
318 IOLockAssert(IOLock * lock, IOLockAssertState type)
319 {
320 LCK_MTX_ASSERT(lock, type);
321 }
322
323 void
324 IORWLockAssert(IORWLock * lock, IORWLockAssertState type)
325 {
326 LCK_RW_ASSERT(lock, type);
327 }
328
329 void
330 IOSimpleLockAssert(IOSimpleLock *lock, IOSimpleLockAssertState type)
331 {
332 LCK_SPIN_ASSERT(l, type);
333 }
334 #endif /* !IOLOCKS_INLINE */
335
336 } /* extern "C" */
337
338