/*
- * Copyright (c) 2000 Apple Computer, Inc. All rights reserved.
+ * Copyright (c) 2000-2004 Apple Computer, Inc. All rights reserved.
*
- * @APPLE_LICENSE_HEADER_START@
+ * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
*
- * The contents of this file constitute Original Code as defined in and
- * are subject to the Apple Public Source License Version 1.1 (the
- * "License"). You may not use this file except in compliance with the
- * License. Please obtain a copy of the License at
- * http://www.apple.com/publicsource and read it before using this file.
+ * This file contains Original Code and/or Modifications of Original Code
+ * as defined in and that are subject to the Apple Public Source License
+ * Version 2.0 (the 'License'). You may not use this file except in
+ * compliance with the License. The rights granted to you under the License
+ * may not be used to create, or enable the creation or redistribution of,
+ * unlawful or unlicensed copies of an Apple operating system, or to
+ * circumvent, violate, or enable the circumvention or violation of, any
+ * terms of an Apple operating system software license agreement.
*
- * This Original Code and all software distributed under the License are
- * distributed on an "AS IS" basis, WITHOUT WARRANTY OF ANY KIND, EITHER
+ * Please obtain a copy of the License at
+ * http://www.opensource.apple.com/apsl/ and read it before using this file.
+ *
+ * The Original Code and all software distributed under the License are
+ * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
* EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
* INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
- * FITNESS FOR A PARTICULAR PURPOSE OR NON-INFRINGEMENT. Please see the
- * License for the specific language governing rights and limitations
- * under the License.
+ * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
+ * Please see the License for the specific language governing rights and
+ * limitations under the License.
*
- * @APPLE_LICENSE_HEADER_END@
+ * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
*/
/*
* Copyright (C) 1998 Apple Computer
/*
* Machine-dependent simple locks for the i386.
*/
+#ifdef KERNEL_PRIVATE
#ifndef _I386_LOCK_H_
#define _I386_LOCK_H_
-#include <kern/macro_help.h>
-#include <kern/assert.h>
-#include <i386/hw_lock_types.h>
-
-#ifdef MACH_KERNEL_PRIVATE
-
-#include <mach_rt.h>
-#include <mach_ldebug.h>
-#include <cpus.h>
-
-
-#if defined(__GNUC__)
-
-/*
- * General bit-lock routines.
- */
-
-#define bit_lock(bit,l) \
- __asm__ volatile(" jmp 1f \n \
- 0: btl %0, %1 \n \
- jb 0b \n \
- 1: lock \n \
- btsl %0,%1 \n \
- jb 0b" : \
- : \
- "r" (bit), "m" (*(volatile int *)(l)) : \
- "memory");
-
-#define bit_unlock(bit,l) \
- __asm__ volatile(" lock \n \
- btrl %0,%1" : \
- : \
- "r" (bit), "m" (*(volatile int *)(l)));
-
-/*
- * Set or clear individual bits in a long word.
- * The locked access is needed only to lock access
- * to the word, not to individual bits.
- */
-
-#define i_bit_set(bit,l) \
- __asm__ volatile(" lock \n \
- btsl %0,%1" : \
- : \
- "r" (bit), "m" (*(volatile int *)(l)));
-
-#define i_bit_clear(bit,l) \
- __asm__ volatile(" lock \n \
- btrl %0,%1" : \
- : \
- "r" (bit), "m" (*(volatile int *)(l)));
-
-extern __inline__ unsigned long i_bit_isset(unsigned int testbit, volatile unsigned long *word)
-{
- int bit;
-
- __asm__ volatile("btl %2,%1\n\tsbbl %0,%0" : "=r" (bit)
- : "m" (word), "ir" (testbit));
- return bit;
-}
-
-extern __inline__ char xchgb(volatile char * cp, char new);
-
-extern __inline__ void atomic_incl(long * p, long delta);
-extern __inline__ void atomic_incs(short * p, short delta);
-extern __inline__ void atomic_incb(char * p, char delta);
-
-extern __inline__ void atomic_decl(long * p, long delta);
-extern __inline__ void atomic_decs(short * p, short delta);
-extern __inline__ void atomic_decb(char * p, char delta);
-
-extern __inline__ long atomic_getl(long * p);
-extern __inline__ short atomic_gets(short * p);
-extern __inline__ char atomic_getb(char * p);
-
-extern __inline__ void atomic_setl(long * p, long value);
-extern __inline__ void atomic_sets(short * p, short value);
-extern __inline__ void atomic_setb(char * p, char value);
-
-extern __inline__ char xchgb(volatile char * cp, char new)
-{
- register char old = new;
-
- __asm__ volatile (" xchgb %0,%2" :
- "=q" (old) :
- "0" (new), "m" (*(volatile char *)cp) : "memory");
- return (old);
-}
-
-extern __inline__ void atomic_incl(long * p, long delta)
-{
-#if NEED_ATOMIC
- __asm__ volatile (" lock \n \
- addl %0,%1" : \
- : \
- "r" (delta), "m" (*(volatile long *)p));
-#else /* NEED_ATOMIC */
- *p += delta;
-#endif /* NEED_ATOMIC */
-}
-
-extern __inline__ void atomic_incs(short * p, short delta)
-{
-#if NEED_ATOMIC
- __asm__ volatile (" lock \n \
- addw %0,%1" : \
- : \
- "q" (delta), "m" (*(volatile short *)p));
-#else /* NEED_ATOMIC */
- *p += delta;
-#endif /* NEED_ATOMIC */
-}
-
-extern __inline__ void atomic_incb(char * p, char delta)
-{
-#if NEED_ATOMIC
- __asm__ volatile (" lock \n \
- addb %0,%1" : \
- : \
- "q" (delta), "m" (*(volatile char *)p));
-#else /* NEED_ATOMIC */
- *p += delta;
-#endif /* NEED_ATOMIC */
-}
-
-extern __inline__ void atomic_decl(long * p, long delta)
-{
-#if NCPUS > 1
- __asm__ volatile (" lock \n \
- subl %0,%1" : \
- : \
- "r" (delta), "m" (*(volatile long *)p));
-#else /* NCPUS > 1 */
- *p -= delta;
-#endif /* NCPUS > 1 */
-}
-
-extern __inline__ void atomic_decs(short * p, short delta)
-{
-#if NEED_ATOMIC
- __asm__ volatile (" lock \n \
- subw %0,%1" : \
- : \
- "q" (delta), "m" (*(volatile short *)p));
-#else /* NEED_ATOMIC */
- *p -= delta;
-#endif /* NEED_ATOMIC */
-}
-
-extern __inline__ void atomic_decb(char * p, char delta)
-{
-#if NEED_ATOMIC
- __asm__ volatile (" lock \n \
- subb %0,%1" : \
- : \
- "q" (delta), "m" (*(volatile char *)p));
-#else /* NEED_ATOMIC */
- *p -= delta;
-#endif /* NEED_ATOMIC */
-}
-
-extern __inline__ long atomic_getl(long * p)
-{
- return (*p);
-}
-
-extern __inline__ short atomic_gets(short * p)
-{
- return (*p);
-}
-
-extern __inline__ char atomic_getb(char * p)
-{
- return (*p);
-}
-
-extern __inline__ void atomic_setl(long * p, long value)
-{
- *p = value;
-}
-
-extern __inline__ void atomic_sets(short * p, short value)
-{
- *p = value;
-}
-
-extern __inline__ void atomic_setb(char * p, char value)
-{
- *p = value;
-}
-
-
-#else /* !defined(__GNUC__) */
-
-extern void i_bit_set(
- int index,
- void *addr);
-
-extern void i_bit_clear(
- int index,
- void *addr);
-
-extern void bit_lock(
- int index,
- void *addr);
-
-extern void bit_unlock(
- int index,
- void *addr);
-
-/*
- * All other routines defined in __GNUC__ case lack
- * definitions otherwise. - XXX
- */
-
-#endif /* !defined(__GNUC__) */
-
-
-#if !(USLOCK_DEBUG || USLOCK_STATS)
-/*
- * Take responsibility for production-quality usimple_locks.
- * Let the portable lock package build simple_locks in terms
- * of usimple_locks, which is done efficiently with macros.
- * Currently, these aren't inlined although they probably
- * should be. The portable lock package is used for the
- * usimple_lock prototypes and data declarations.
- *
- * For non-production configurations, punt entirely to the
- * portable lock package.
- *
- * N.B. I've left in the hooks for ETAP, so we can
- * compare the performance of stats-gathering on top
- * of "production" locks v. stats-gathering on top
- * of portable, C-based locks.
- */
-#define USIMPLE_LOCK_CALLS
-#endif /* !(USLOCK_DEBUG || USLOCK_STATS) */
-
-
-#if MACH_RT || (NCPUS > 1) || MACH_LDEBUG
-#if MACH_LDEBUG || !MACH_RT
-#define mutex_try(m) (!(m)->interlock && _mutex_try(m))
-#define mutex_lock(m) \
-MACRO_BEGIN \
- assert(assert_wait_possible()); \
- _mutex_lock((m)); \
-MACRO_END
-
-#else /* MACH_LDEBUG || !MACH_RT */
-#define mutex_try(m) (!(m)->interlock && \
- !xchgb ((volatile char *)&((m)->locked), 1))
-#define mutex_lock(m) \
-MACRO_BEGIN \
- assert(assert_wait_possible()); \
- _mutex_lock (m); \
-MACRO_END
-
-#endif /* MACH_LDEBUG || !MACH_RT */
-#else /* MACH_RT || (NCPUS > 1) || MACH_LDEBUG */
-#define mutex_try _mutex_try
-#define mutex_lock _mutex_lock
-#endif /* MACH_RT || (NCPUS > 1) || MACH_LDEBUG */
-
-#else /* !MACH_KERNEL_PRIVATE */
-
-#define mutex_try _mutex_try
-#define mutex_lock(m) \
-MACRO_BEGIN \
- assert(assert_wait_possible()); \
- _mutex_lock((m)); \
-MACRO_END
-
-#endif /* !MACH_KERNEL_PRIVATE */
-
-extern void kernel_preempt_check (void);
+#warning This header is deprecated. Use <kern/locks.h> instead.
#endif /* _I386_LOCK_H_ */
+#endif /* KERNEL_PRIVATE */