* Copyright (c) 2000-2005 Apple Computer, Inc. All rights reserved.
*
* @APPLE_OSREFERENCE_LICENSE_HEADER_START@
- *
+ *
* This file contains Original Code and/or Modifications of Original Code
* as defined in and that are subject to the Apple Public Source License
* Version 2.0 (the 'License'). You may not use this file except in
* unlawful or unlicensed copies of an Apple operating system, or to
* circumvent, violate, or enable the circumvention or violation of, any
* terms of an Apple operating system software license agreement.
- *
+ *
* Please obtain a copy of the License at
* http://www.opensource.apple.com/apsl/ and read it before using this file.
- *
+ *
* The Original Code and all software distributed under the License are
* distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
* EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
* FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
* Please see the License for the specific language governing rights and
* limitations under the License.
- *
+ *
* @APPLE_OSREFERENCE_LICENSE_HEADER_END@
*/
-/*
+/*
* Copyright (C) 1998 Apple Computer
* All Rights Reserved
*/
* Mach Operating System
* Copyright (c) 1991,1990,1989,1988,1987 Carnegie Mellon University
* All Rights Reserved.
- *
+ *
* Permission to use, copy, modify and distribute this software and its
* documentation is hereby granted, provided that both the copyright
* notice and this permission notice appear in all copies of the
* software, derivative works or modified versions, and any portions
* thereof, and that both notices appear in supporting documentation.
- *
+ *
* CARNEGIE MELLON ALLOWS FREE USE OF THIS SOFTWARE IN ITS "AS IS"
* CONDITION. CARNEGIE MELLON DISCLAIMS ANY LIABILITY OF ANY KIND FOR
* ANY DAMAGES WHATSOEVER RESULTING FROM THE USE OF THIS SOFTWARE.
- *
+ *
* Carnegie Mellon requests users of this software to return to
- *
+ *
* Software Distribution Coordinator or Software.Distribution@CS.CMU.EDU
* School of Computer Science
* Carnegie Mellon University
* Pittsburgh PA 15213-3890
- *
+ *
* any improvements or extensions that they make and grant Carnegie Mellon
* the rights to redistribute these changes.
*/
* Atomic primitives and Simple Locking primitives definitions
*/
-#ifdef KERNEL_PRIVATE
+#ifdef KERNEL_PRIVATE
-#ifndef _KERN_SIMPLE_LOCK_H_
-#define _KERN_SIMPLE_LOCK_H_
+#ifndef _KERN_SIMPLE_LOCK_H_
+#define _KERN_SIMPLE_LOCK_H_
#include <sys/cdefs.h>
#include <mach/boolean.h>
#include <kern/kern_types.h>
+#include <kern/lock_group.h>
#include <machine/simple_lock.h>
-#ifdef MACH_KERNEL_PRIVATE
+#ifdef MACH_KERNEL_PRIVATE
#include <mach_ldebug.h>
-extern void hw_lock_init(
- hw_lock_t);
+extern void hw_lock_init(
+ hw_lock_t);
+
+#if LOCK_STATS
+extern void hw_lock_lock(
+ hw_lock_t,
+ lck_grp_t*);
+
+extern void hw_lock_lock_nopreempt(
+ hw_lock_t,
+ lck_grp_t*);
+
+extern unsigned int hw_lock_to(
+ hw_lock_t,
+ uint64_t,
+ lck_grp_t*);
-extern void hw_lock_lock(
- hw_lock_t);
+extern unsigned int hw_lock_try(
+ hw_lock_t,
+ lck_grp_t*);
-extern void hw_lock_lock_nopreempt(
- hw_lock_t);
+extern unsigned int hw_lock_try_nopreempt(
+ hw_lock_t,
+ lck_grp_t*);
-extern void hw_lock_unlock(
- hw_lock_t);
+#else
-extern void hw_lock_unlock_nopreempt(
- hw_lock_t);
+extern void hw_lock_lock(
+ hw_lock_t);
-extern unsigned int hw_lock_to(
- hw_lock_t,
- uint64_t);
+#define hw_lock_lock(lck, grp) hw_lock_lock(lck)
-extern unsigned int hw_lock_try(
- hw_lock_t);
+extern void hw_lock_lock_nopreempt(
+ hw_lock_t);
+#define hw_lock_lock_nopreempt(lck, grp) hw_lock_lock_nopreempt(lck)
-extern unsigned int hw_lock_try_nopreempt(
- hw_lock_t);
+extern unsigned int hw_lock_to(
+ hw_lock_t,
+ uint64_t);
+#define hw_lock_to(lck, timeout, grp) hw_lock_to(lck, timeout)
-extern unsigned int hw_lock_held(
- hw_lock_t);
-#endif /* MACH_KERNEL_PRIVATE */
+extern unsigned int hw_lock_try(
+ hw_lock_t);
+#define hw_lock_try(lck, grp) hw_lock_try(lck)
+
+extern unsigned int hw_lock_try_nopreempt(
+ hw_lock_t);
+#define hw_lock_try_nopreempt(lck, grp) hw_lock_try_nopreempt(lck)
+
+
+#endif /* LOCK_STATS */
+
+extern void hw_lock_unlock(
+ hw_lock_t);
+
+extern void hw_lock_unlock_nopreempt(
+ hw_lock_t);
+
+extern unsigned int hw_lock_held(
+ hw_lock_t);
+
+#endif /* MACH_KERNEL_PRIVATE */
__BEGIN_DECLS
-extern uint32_t hw_atomic_add(
- volatile uint32_t *dest,
- uint32_t delt);
+extern uint32_t hw_atomic_add(
+ volatile uint32_t *dest,
+ uint32_t delt);
-extern uint32_t hw_atomic_sub(
- volatile uint32_t *dest,
- uint32_t delt);
+extern uint32_t hw_atomic_sub(
+ volatile uint32_t *dest,
+ uint32_t delt);
-extern uint32_t hw_atomic_or(
- volatile uint32_t *dest,
- uint32_t mask);
+extern uint32_t hw_atomic_or(
+ volatile uint32_t *dest,
+ uint32_t mask);
-extern uint32_t hw_atomic_and(
- volatile uint32_t *dest,
- uint32_t mask);
+extern uint32_t hw_atomic_and(
+ volatile uint32_t *dest,
+ uint32_t mask);
/*
* Variant of hw_atomic_or which doesn't return a value; potentially
* more efficient on some platforms.
*/
-extern void hw_atomic_or_noret(
- volatile uint32_t *dest,
- uint32_t mask);
+extern void hw_atomic_or_noret(
+ volatile uint32_t *dest,
+ uint32_t mask);
/*
* Variant of hw_atomic_and which doesn't return a value; potentially
* more efficient on some platforms.
*/
-extern void hw_atomic_and_noret(
- volatile uint32_t *dest,
- uint32_t mask);
+extern void hw_atomic_and_noret(
+ volatile uint32_t *dest,
+ uint32_t mask);
+
+extern uint32_t hw_compare_and_store(
+ uint32_t oldval,
+ uint32_t newval,
+ volatile uint32_t *dest);
+
+extern void hw_queue_atomic(
+ unsigned int *anchor,
+ unsigned int *elem,
+ unsigned int disp);
+
+extern void hw_queue_atomic_list(
+ unsigned int *anchor,
+ unsigned int *first,
+ unsigned int *last,
+ unsigned int disp);
+
+extern unsigned int *hw_dequeue_atomic(
+ unsigned int *anchor,
+ unsigned int disp);
+
+extern void usimple_lock_init(
+ usimple_lock_t,
+ unsigned short);
+
+#if LOCK_STATS
+extern void usimple_lock(
+ usimple_lock_t,
+ lck_grp_t*);
-extern uint32_t hw_compare_and_store(
- uint32_t oldval,
- uint32_t newval,
- volatile uint32_t *dest);
+extern unsigned int usimple_lock_try(
+ usimple_lock_t,
+ lck_grp_t*);
-extern void hw_queue_atomic(
- unsigned int *anchor,
- unsigned int *elem,
- unsigned int disp);
+extern void usimple_lock_try_lock_loop(
+ usimple_lock_t,
+ lck_grp_t*);
+#else
+extern void usimple_lock(
+ usimple_lock_t);
+#define usimple_lock(lck, grp) usimple_lock(lck)
-extern void hw_queue_atomic_list(
- unsigned int *anchor,
- unsigned int *first,
- unsigned int *last,
- unsigned int disp);
-extern unsigned int *hw_dequeue_atomic(
- unsigned int *anchor,
- unsigned int disp);
+extern unsigned int usimple_lock_try(
+ usimple_lock_t);
-extern void usimple_lock_init(
- usimple_lock_t,
- unsigned short);
+#define usimple_lock_try(lck, grp) usimple_lock_try(lck)
-extern void usimple_lock(
- usimple_lock_t);
+extern void usimple_lock_try_lock_loop(
+ usimple_lock_t);
+#define usimple_lock_try_lock_loop(lck, grp) usimple_lock_try_lock_loop(lck)
-extern void usimple_unlock(
- usimple_lock_t);
+#endif /* LOCK_STATS */
-extern unsigned int usimple_lock_try(
- usimple_lock_t);
+extern void usimple_unlock(
+ usimple_lock_t);
-extern void usimple_lock_try_lock_loop(
- usimple_lock_t);
__END_DECLS
-#define ETAP_NO_TRACE 0
-#define ETAP_IO_AHA 0
+#define ETAP_NO_TRACE 0
+#define ETAP_IO_AHA 0
/*
* If we got to here and we still don't have simple_lock_init
* running on a true SMP, or need debug.
*/
#if !defined(simple_lock_init)
-#define simple_lock_init(l,t) usimple_lock_init(l,t)
-#define simple_lock(l) usimple_lock(l)
-#define simple_unlock(l) usimple_unlock(l)
-#define simple_lock_try(l) usimple_lock_try(l)
-#define simple_lock_try_lock_loop(l) usimple_lock_try_lock_loop(l)
-#define simple_lock_addr(l) (&(l))
+#define simple_lock_init(l, t) usimple_lock_init(l,t)
+#define simple_lock(l, grp) usimple_lock(l, grp)
+#define simple_unlock(l) usimple_unlock(l)
+#define simple_lock_try(l, grp) usimple_lock_try(l, grp)
+#define simple_lock_try_lock_loop(l, grp) usimple_lock_try_lock_loop(l, grp)
+#define simple_lock_addr(l) (&(l))
#endif /* !defined(simple_lock_init) */
#endif /*!_KERN_SIMPLE_LOCK_H_*/
-#endif /* KERNEL_PRIVATE */
+#endif /* KERNEL_PRIVATE */