/*
- * Copyright (c) 2000-2003 Apple Computer, Inc. All rights reserved.
+ * Copyright (c) 2000-2003, 2007, 2008 Apple Inc. All rights reserved.
*
* @APPLE_LICENSE_HEADER_START@
*
- * Copyright (c) 1999-2003 Apple Computer, Inc. All Rights Reserved.
- *
* This file contains Original Code and/or Modifications of Original Code
* as defined in and that are subject to the Apple Public Source License
* Version 2.0 (the 'License'). You may not use this file except in
#include "pthread_internals.h"
#include <sys/time.h> /* For struct timespec and getclock(). */
#include <stdio.h>
-
-extern void _pthread_mutex_remove(pthread_mutex_t *, pthread_t);
-/*
- * Destroy a condition variable.
- */
+#ifdef PLOCKSTAT
+#include "plockstat.h"
+#else /* !PLOCKSTAT */
+#define PLOCKSTAT_MUTEX_RELEASE(x, y)
+#endif /* PLOCKSTAT */
+
+extern int __semwait_signal(int, int, int, int, int64_t, int32_t);
+extern int _pthread_cond_init(pthread_cond_t *, const pthread_condattr_t *, int);
+extern int __unix_conforming;
+extern int usenew_mtximpl;
+
+#ifdef PR_5243343
+/* 5243343 - temporary hack to detect if we are running the conformance test */
+extern int PR_5243343_flag;
+#endif /* PR_5243343 */
+
+__private_extern__ int _pthread_cond_wait(pthread_cond_t *cond, pthread_mutex_t *mutex, const struct timespec *abstime, int isRelative, int isconforming);
+#ifndef BUILDING_VARIANT
+static void cond_cleanup(void *arg);
+static void cond_dropwait(npthread_cond_t * cond, int error, uint32_t updateval);
+static void __pthread_cond_set_signature(npthread_cond_t * cond);
+static int _pthread_cond_destroy_locked(pthread_cond_t *cond);
+#endif
+
+#if defined(__LP64__)
+#define COND_GETSEQ_ADDR(cond, c_lseqcnt, c_useqcnt, c_sseqcnt) \
+{ \
+ if (cond->misalign != 0) { \
+ c_lseqcnt = &cond->c_seq[1]; \
+ c_sseqcnt = &cond->c_seq[2]; \
+ c_useqcnt = &cond->c_seq[0]; \
+ } else { \
+ /* aligned */ \
+ c_lseqcnt = &cond->c_seq[0]; \
+ c_sseqcnt = &cond->c_seq[1]; \
+ c_useqcnt = &cond->c_seq[2]; \
+ } \
+}
+#else /* __LP64__ */
+#define COND_GETSEQ_ADDR(cond, c_lseqcnt, c_useqcnt, c_sseqcnt) \
+{ \
+ if (cond->misalign != 0) { \
+ c_lseqcnt = &cond->c_seq[1]; \
+ c_sseqcnt = &cond->c_seq[2]; \
+ c_useqcnt = &cond->c_seq[0]; \
+ } else { \
+ /* aligned */ \
+ c_lseqcnt = &cond->c_seq[0]; \
+ c_sseqcnt = &cond->c_seq[1]; \
+ c_useqcnt = &cond->c_seq[2]; \
+ } \
+}
+#endif /* __LP64__ */
+
+
+#define _KSYN_TRACE_ 0
+
+#if _KSYN_TRACE_
+/* The Function qualifiers */
+#define DBG_FUNC_START 1
+#define DBG_FUNC_END 2
+#define DBG_FUNC_NONE 0
+
+int __kdebug_trace(uint32_t, uint32_t, uint32_t, uint32_t, uint32_t, uint32_t);
+
+#define _KSYN_TRACE_UM_LOCK 0x9000060
+#define _KSYN_TRACE_UM_UNLOCK 0x9000064
+#define _KSYN_TRACE_UM_MHOLD 0x9000068
+#define _KSYN_TRACE_UM_MDROP 0x900006c
+#define _KSYN_TRACE_UM_CVWAIT 0x9000070
+#define _KSYN_TRACE_UM_CVSIG 0x9000074
+#define _KSYN_TRACE_UM_CVBRD 0x9000078
+#define _KSYN_TRACE_UM_CDROPWT 0x90000a0
+#define _KSYN_TRACE_UM_CVCLRPRE 0x90000a4
+
+#endif /* _KSYN_TRACE_ */
+
+
+#ifndef BUILDING_VARIANT /* [ */
+
+int
+pthread_condattr_init(pthread_condattr_t *attr)
+{
+ attr->sig = _PTHREAD_COND_ATTR_SIG;
+ attr->pshared = _PTHREAD_DEFAULT_PSHARED;
+ return (0);
+}
+
int
-pthread_cond_destroy(pthread_cond_t *cond)
+pthread_condattr_destroy(pthread_condattr_t *attr)
+{
+ attr->sig = _PTHREAD_NO_SIG; /* Uninitialized */
+ return (0);
+}
+
+int
+pthread_condattr_getpshared(const pthread_condattr_t *attr,
+ int *pshared)
+{
+ if (attr->sig == _PTHREAD_COND_ATTR_SIG)
+ {
+ *pshared = (int)attr->pshared;
+ return (0);
+ } else
+ {
+ return (EINVAL); /* Not an initialized 'attribute' structure */
+ }
+}
+
+
+
+
+/* temp home till pshared is fixed correctly */
+int
+pthread_condattr_setpshared(pthread_condattr_t * attr, int pshared)
+{
+
+ if (attr->sig == _PTHREAD_COND_ATTR_SIG)
+ {
+#if __DARWIN_UNIX03
+ if (( pshared == PTHREAD_PROCESS_PRIVATE) || (pshared == PTHREAD_PROCESS_SHARED))
+#else /* __DARWIN_UNIX03 */
+ if ( pshared == PTHREAD_PROCESS_PRIVATE)
+#endif /* __DARWIN_UNIX03 */
+ {
+ attr->pshared = pshared;
+ return (0);
+ } else {
+ return (EINVAL); /* Invalid parameter */
+ }
+ } else
+ {
+ return (EINVAL); /* Not an initialized 'attribute' structure */
+ }
+
+}
+
+__private_extern__ int
+_pthread_cond_init(pthread_cond_t *ocond,
+ const pthread_condattr_t *attr,
+ int conforming)
{
+ npthread_cond_t * cond = (npthread_cond_t *)ocond;
+
+ cond->busy = (npthread_mutex_t *)NULL;
+ cond->c_seq[0] = 0;
+ cond->c_seq[1] = 0;
+ cond->c_seq[2] = 0;
+ cond->rfu = 0;
+
+ if (((uintptr_t)cond & 0x07) != 0) {
+ cond->misalign = 1;
+ cond->c_seq[2] = PTH_RWS_CV_CBIT;
+ } else {
+ cond->misalign = 0;
+ cond->c_seq[1] = PTH_RWS_CV_CBIT; /* set Sword to 0c */
+ }
+ if (conforming) {
+ if (attr)
+ cond->pshared = attr->pshared;
+ else
+ cond->pshared = _PTHREAD_DEFAULT_PSHARED;
+ } else
+ cond->pshared = _PTHREAD_DEFAULT_PSHARED;
+ /*
+ * For the new style mutex, interlocks are not held all the time.
+ * We needed the signature to be set in the end. And we need
+ * to protect against the code getting reorganized by compiler.
+ * cond->sig = _PTHREAD_COND_SIG;
+ */
+ __pthread_cond_set_signature(cond);
+ return (0);
+}
+
+int
+pthread_cond_destroy(pthread_cond_t * ocond)
+{
+ npthread_cond_t *cond = (npthread_cond_t *)ocond;
int ret;
- int sig = cond->sig;
/* to provide backwards compat for apps using united condtn vars */
- if((sig != _PTHREAD_COND_SIG) && (sig !=_PTHREAD_COND_SIG_init))
+ if((cond->sig != _PTHREAD_COND_SIG) && (cond->sig != _PTHREAD_COND_SIG_init))
return(EINVAL);
LOCK(cond->lock);
+ ret = _pthread_cond_destroy_locked(ocond);
+ UNLOCK(cond->lock);
+
+ return(ret);
+}
+
+static int
+_pthread_cond_destroy_locked(pthread_cond_t * ocond)
+{
+ npthread_cond_t *cond = (npthread_cond_t *)ocond;
+ int ret;
+ volatile uint32_t * c_lseqcnt, *c_useqcnt, *c_sseqcnt;
+ uint32_t lcntval , ucntval, scntval;
+ uint64_t oldval64, newval64;
+
+retry:
if (cond->sig == _PTHREAD_COND_SIG)
{
- if (cond->busy == (pthread_mutex_t *)NULL)
- {
+ COND_GETSEQ_ADDR(cond, c_lseqcnt, c_useqcnt, c_sseqcnt);
+ lcntval = *c_lseqcnt;
+ ucntval = *c_useqcnt;
+ scntval = *c_sseqcnt;
+
+ if ((lcntval & PTHRW_COUNT_MASK) == (scntval & PTHRW_COUNT_MASK)) {
+ /* validate it is not busy */
+ oldval64 = (((uint64_t)scntval) << 32);
+ oldval64 |= lcntval;
+ newval64 = oldval64;
+
+ if (OSAtomicCompareAndSwap64(oldval64, newval64, (volatile int64_t *)c_lseqcnt) != TRUE)
+ goto retry;
cond->sig = _PTHREAD_NO_SIG;
- ret = ESUCCESS;
+ ret = 0;
} else
ret = EBUSY;
+ } else if (cond->sig == _PTHREAD_COND_SIG_init) {
+ cond->sig = _PTHREAD_NO_SIG;
+ ret = 0;
} else
ret = EINVAL; /* Not an initialized condition variable structure */
- UNLOCK(cond->lock);
return (ret);
}
-/*
- * Initialize a condition variable. Note: 'attr' is ignored.
- */
-static int
-_pthread_cond_init(pthread_cond_t *cond,
- const pthread_condattr_t *attr)
-{
- cond->next = (pthread_cond_t *)NULL;
- cond->prev = (pthread_cond_t *)NULL;
- cond->busy = (pthread_mutex_t *)NULL;
- cond->waiters = 0;
- cond->sigspending = 0;
- cond->sem = SEMAPHORE_NULL;
- cond->sig = _PTHREAD_COND_SIG;
- return (ESUCCESS);
-}
-
-/*
- * Initialize a condition variable. This is the public interface.
- * We can't trust the lock, so initialize it first before taking
- * it.
- */
-int
-pthread_cond_init(pthread_cond_t *cond,
- const pthread_condattr_t *attr)
-{
- LOCK_INIT(cond->lock);
- return (_pthread_cond_init(cond, attr));
-}
-
/*
* Signal a condition variable, waking up all threads waiting for it.
*/
int
-pthread_cond_broadcast(pthread_cond_t *cond)
+pthread_cond_broadcast(pthread_cond_t *ocond)
{
- kern_return_t kern_res;
- semaphore_t sem;
+ npthread_cond_t * cond = (npthread_cond_t *)ocond;
int sig = cond->sig;
+ uint32_t flags, updateval;
+ uint32_t lcntval , ucntval, scntval;
+ uint64_t oldval64, newval64, mugen, cvlsgen, cvudgen, mtid=0;
+ int diffgen, error = 0;
+ volatile uint32_t * c_lseqcnt, *c_useqcnt, *c_sseqcnt;
+ uint32_t * pmtx = NULL;
+ uint32_t nlval, ulval;
+ int needclearpre = 0, retry_count = 0;
/* to provide backwards compat for apps using united condtn vars */
- if((sig != _PTHREAD_COND_SIG) && (sig !=_PTHREAD_COND_SIG_init))
+ if((sig != _PTHREAD_COND_SIG) && (sig != _PTHREAD_COND_SIG_init))
return(EINVAL);
- LOCK(cond->lock);
- if (cond->sig != _PTHREAD_COND_SIG)
+ if (sig != _PTHREAD_COND_SIG)
{
- int res;
-
+ LOCK(cond->lock);
if (cond->sig == _PTHREAD_COND_SIG_init)
{
- _pthread_cond_init(cond, NULL);
- res = ESUCCESS;
- } else
- res = EINVAL; /* Not a condition variable */
+ _pthread_cond_init(ocond, NULL, 0);
+ /* just inited nothing to post */
+ UNLOCK(cond->lock);
+ return (0);
+ } else if (cond->sig != _PTHREAD_COND_SIG) {
+ /* Not a condition variable */
+ UNLOCK(cond->lock);
+ return (EINVAL);
+ }
UNLOCK(cond->lock);
- return (res);
}
- else if ((sem = cond->sem) == SEMAPHORE_NULL)
- {
- /* Avoid kernel call since there are no waiters... */
- UNLOCK(cond->lock);
- return (ESUCCESS);
+
+#if _KSYN_TRACE_
+ (void)__kdebug_trace(_KSYN_TRACE_UM_CVBRD | DBG_FUNC_START, (uint32_t)cond, 0, 0, 0, 0);
+#endif
+
+ COND_GETSEQ_ADDR(cond, c_lseqcnt, c_useqcnt, c_sseqcnt);
+retry:
+ lcntval = *c_lseqcnt;
+ ucntval = *c_useqcnt;
+ scntval = *c_sseqcnt;
+
+#if _KSYN_TRACE_
+ (void)__kdebug_trace(_KSYN_TRACE_UM_CVBRD | DBG_FUNC_NONE, (uint32_t)cond, lcntval, ucntval, scntval, 0);
+#endif
+
+ if (((lcntval & PTHRW_COUNT_MASK) == (scntval & PTHRW_COUNT_MASK)) ||
+ ((lcntval & PTHRW_COUNT_MASK) == (ucntval & PTHRW_COUNT_MASK))) {
+ /* validate it is spurious and return */
+ oldval64 = (((uint64_t)scntval) << 32);
+ oldval64 |= lcntval;
+ newval64 = oldval64;
+
+ if (OSAtomicCompareAndSwap64(oldval64, newval64, (volatile int64_t *)c_lseqcnt) != TRUE)
+ goto retry;
+#if _KSYN_TRACE_
+ (void)__kdebug_trace(_KSYN_TRACE_UM_CVBRD | DBG_FUNC_NONE, (uint32_t)cond, lcntval, ucntval, 0xf1f1f1f1, 0);
+ (void)__kdebug_trace(_KSYN_TRACE_UM_CVBRD | DBG_FUNC_END, (uint32_t)cond, scntval, 0, 0xf1f1f1f1, 0);
+#endif
+ return(0);
}
- cond->sigspending++;
- UNLOCK(cond->lock);
- PTHREAD_MACH_CALL(semaphore_signal_all(sem), kern_res);
+ if (is_seqhigher((ucntval & PTHRW_COUNT_MASK), (lcntval & PTHRW_COUNT_MASK)) || is_seqhigher((scntval & PTHRW_COUNT_MASK), (lcntval & PTHRW_COUNT_MASK))) {
+ /* since ucntval may be newer, just redo */
+ retry_count++;
+ if (retry_count > 8192) {
+ return(EAGAIN);
+ } else {
+ sched_yield();
+ goto retry;
+ }
+ }
+
+ if (is_seqlower(ucntval & PTHRW_COUNT_MASK, scntval & PTHRW_COUNT_MASK) != 0) {
+ /* If U < S, set U = S+diff due to intr's TO, etc */
+ ulval = (scntval & PTHRW_COUNT_MASK);
+ } else {
+ /* If U >= S, set U = U+diff due to intr's TO, etc */
+ ulval = (ucntval & PTHRW_COUNT_MASK);
+ }
+#if _KSYN_TRACE_
+ (void)__kdebug_trace(_KSYN_TRACE_UM_CVBRD | DBG_FUNC_NONE, lcntval, ucntval, scntval, diffgen, 0);
+#endif
+
+ diffgen = diff_genseq((lcntval & PTHRW_COUNT_MASK), (ulval & PTHRW_COUNT_MASK));
- LOCK(cond->lock);
- cond->sigspending--;
- if (cond->waiters == 0 && cond->sigspending == 0)
- {
- cond->sem = SEMAPHORE_NULL;
- restore_sem_to_pool(sem);
+ /* set U = L */
+ ulval = (lcntval & PTHRW_COUNT_MASK);
+ if (OSAtomicCompareAndSwap32(ucntval, ulval, (volatile int32_t *)c_useqcnt) != TRUE) {
+ goto retry;
}
- UNLOCK(cond->lock);
- if (kern_res != KERN_SUCCESS)
- return (EINVAL);
- return (ESUCCESS);
+
+ flags = 0;
+ if (cond->pshared == PTHREAD_PROCESS_SHARED)
+ flags |= _PTHREAD_MTX_OPT_PSHARED;
+ pmtx = NULL;
+
+#if _KSYN_TRACE_
+ (void)__kdebug_trace(_KSYN_TRACE_UM_CVBRD | DBG_FUNC_NONE, (uint32_t)cond, 3, diffgen, flags, 0);
+#endif
+ nlval = lcntval;
+
+ /* pass old u val so kernel will know the diffgen */
+ mugen = 0;
+ cvlsgen = ((uint64_t)scntval << 32) | nlval;
+ cvudgen = ((uint64_t)ucntval << 32) | diffgen;
+
+ updateval = __psynch_cvbroad(ocond, cvlsgen, cvudgen, flags, (pthread_mutex_t *)pmtx, mugen, mtid);
+
+ if (updateval != (uint32_t)-1) {
+
+ /* if kernel granted woke some threads, updatwe S for them as they will not access cv on their way out */
+ /* Were any threads woken or bits to be set? */
+ if (updateval != 0) {
+retry2:
+ needclearpre = 0;
+ lcntval = *c_lseqcnt;
+ ucntval = *c_useqcnt;
+ scntval = *c_sseqcnt;
+ /* update scntval with number of expected returns and bits */
+ nlval = (scntval & PTHRW_COUNT_MASK) + (updateval & PTHRW_COUNT_MASK);
+ /* set bits */
+ nlval |= ((scntval & PTH_RWS_CV_BITSALL) | (updateval & PTH_RWS_CV_BITSALL));
+
+#if _KSYN_TRACE_
+ (void)__kdebug_trace(_KSYN_TRACE_UM_CVBRD | DBG_FUNC_NONE, 0x25, lcntval, scntval, updateval, 0);
+#endif
+ /* if L==S and c&p bits are set, needs clearpre */
+ if (((nlval & PTHRW_COUNT_MASK) == (lcntval & PTHRW_COUNT_MASK))
+ && ((nlval & PTH_RWS_CV_BITSALL) == PTH_RWS_CV_BITSALL)) {
+ /* reset p bit but retain c bit on the sword */
+ nlval &= PTH_RWS_CV_RESET_PBIT;
+ needclearpre = 1;
+ }
+
+ oldval64 = (((uint64_t)scntval) << 32);
+ oldval64 |= lcntval;
+ newval64 = (((uint64_t)nlval) << 32);
+ newval64 |= lcntval;
+
+#if _KSYN_TRACE_
+ (void)__kdebug_trace(_KSYN_TRACE_UM_CVBRD | DBG_FUNC_NONE, 0x25, nlval, scntval, updateval, 0);
+#endif
+
+ if (OSAtomicCompareAndSwap64(oldval64, newval64, (volatile int64_t *)c_lseqcnt) != TRUE)
+ goto retry2;
+
+ /* if L == S, then reset associated mutex */
+ if ((nlval & PTHRW_COUNT_MASK) == (lcntval & PTHRW_COUNT_MASK)) {
+ cond->busy = (npthread_mutex_t *)NULL;
+ }
+
+ if (needclearpre != 0) {
+ (void)__psynch_cvclrprepost(ocond, lcntval, ucntval, nlval, 0, lcntval, flags);
+ }
+ }
+
+ }
+ error = 0;
+
+#if _KSYN_TRACE_
+ (void)__kdebug_trace(_KSYN_TRACE_UM_CVBRD | DBG_FUNC_END, (uint32_t)cond, 0, error, 0, 0);
+#endif
+ return(error);
}
+
/*
* Signal a condition variable, waking a specified thread.
*/
+
int
-pthread_cond_signal_thread_np(pthread_cond_t *cond, pthread_t thread)
+pthread_cond_signal_thread_np(pthread_cond_t *ocond, pthread_t thread)
{
- kern_return_t kern_res;
- semaphore_t sem;
+ npthread_cond_t * cond = (npthread_cond_t *)ocond;
int sig = cond->sig;
+ uint32_t flags, updateval;
+ uint32_t lcntval , ucntval, scntval;
+ uint32_t nlval, ulval=0;
+ volatile uint32_t * c_lseqcnt, *c_useqcnt, *c_sseqcnt;
+ uint64_t oldval64, newval64, mugen, cvlsgen, mtid = 0;
+ int needclearpre = 0, retry_count = 0;
+ int error;
/* to provide backwards compat for apps using united condtn vars */
- if((sig != _PTHREAD_COND_SIG) && (sig !=_PTHREAD_COND_SIG_init))
- return(EINVAL);
- LOCK(cond->lock);
- if (cond->sig != _PTHREAD_COND_SIG)
- {
- int ret;
+ if((sig != _PTHREAD_COND_SIG) && (sig != _PTHREAD_COND_SIG_init))
+ return(EINVAL);
- if (cond->sig == _PTHREAD_COND_SIG_init)
- {
- _pthread_cond_init(cond, NULL);
- ret = ESUCCESS;
+ if (cond->sig != _PTHREAD_COND_SIG) {
+ LOCK(cond->lock);
+ if (cond->sig != _PTHREAD_COND_SIG) {
+ if (cond->sig == _PTHREAD_COND_SIG_init) {
+ _pthread_cond_init(ocond, NULL, 0);
+ /* just inited, nothing to post yet */
+ UNLOCK(cond->lock);
+ return(0);
+ } else {
+ UNLOCK(cond->lock);
+ return(EINVAL);
+ }
}
- else
- ret = EINVAL; /* Not a condition variable */
- UNLOCK(cond->lock);
- return (ret);
- }
- else if ((sem = cond->sem) == SEMAPHORE_NULL)
- {
- /* Avoid kernel call since there are not enough waiters... */
UNLOCK(cond->lock);
- return (ESUCCESS);
}
- cond->sigspending++;
- UNLOCK(cond->lock);
- if (thread == (pthread_t)NULL)
- {
- kern_res = semaphore_signal_thread(sem, THREAD_NULL);
- if (kern_res == KERN_NOT_WAITING)
- kern_res = KERN_SUCCESS;
- }
- else if (thread->sig == _PTHREAD_SIG)
- {
- PTHREAD_MACH_CALL(semaphore_signal_thread(
- sem, pthread_mach_thread_np(thread)), kern_res);
+#if _KSYN_TRACE_
+ (void)__kdebug_trace(_KSYN_TRACE_UM_CVSIG | DBG_FUNC_START, (uint32_t)cond, 0, 0, 0, 0);
+#endif
+ COND_GETSEQ_ADDR(cond, c_lseqcnt, c_useqcnt, c_sseqcnt);
+retry:
+ lcntval = *c_lseqcnt;
+ ucntval = *c_useqcnt;
+ scntval = *c_sseqcnt;
+#if _KSYN_TRACE_
+ (void)__kdebug_trace(_KSYN_TRACE_UM_CVSIG | DBG_FUNC_NONE, (uint32_t)cond, lcntval, ucntval, scntval, 0);
+#endif
+
+ if (((lcntval & PTHRW_COUNT_MASK) == (scntval & PTHRW_COUNT_MASK)) ||
+ ((thread == 0) && ((lcntval & PTHRW_COUNT_MASK) == (ucntval & PTHRW_COUNT_MASK)))) {
+ /* If L <= S+U, it is spurious broadcasr */
+ /* validate it is spurious and return */
+ oldval64 = (((uint64_t)scntval) << 32);
+ oldval64 |= lcntval;
+ newval64 = oldval64;
+
+ if (OSAtomicCompareAndSwap64(oldval64, newval64, (volatile int64_t *)c_lseqcnt) != TRUE)
+ goto retry;
+#if _KSYN_TRACE_
+ (void)__kdebug_trace(_KSYN_TRACE_UM_CVSIG | DBG_FUNC_NONE, (uint32_t)cond, lcntval, ucntval, 0xf1f1f1f1, 0);
+ (void)__kdebug_trace(_KSYN_TRACE_UM_CVSIG | DBG_FUNC_END, (uint32_t)cond, scntval, 0, 0xf1f1f1f1, 0);
+#endif
+ return(0);
}
- else
- kern_res = KERN_FAILURE;
- LOCK(cond->lock);
- cond->sigspending--;
- if (cond->waiters == 0 && cond->sigspending == 0)
- {
- cond->sem = SEMAPHORE_NULL;
- restore_sem_to_pool(sem);
+ if (((thread == 0) && (is_seqhigher((ucntval & PTHRW_COUNT_MASK), (lcntval & PTHRW_COUNT_MASK)))) || is_seqhigher((scntval & PTHRW_COUNT_MASK), (lcntval & PTHRW_COUNT_MASK))) {
+ /* since ucntval may be newer, just redo */
+ retry_count++;
+ if (retry_count > 8192) {
+ return(EAGAIN);
+ } else {
+ sched_yield();
+ goto retry;
+ }
+ }
+
+ if (thread == 0) {
+ /*
+ * skip manipulating U count as ESRCH from kernel cannot be handled properly.
+ * S count will cover the imbalance and next signal without thread or broadcast
+ * will correct it. But we need to send the right U to kernel so it will use
+ * that to look for the appropriate sequenc. So the ulval is computed anyway.
+ */
+
+ if (is_seqlower(ucntval & PTHRW_COUNT_MASK, scntval & PTHRW_COUNT_MASK) != 0) {
+ /* If U < S, set U = S+1 due to intr's TO, etc */
+ ulval = (scntval & PTHRW_COUNT_MASK) + PTHRW_INC;
+ } else {
+ /* If U >= S, set U = U+1 due to intr's TO, etc */
+ ulval = (ucntval & PTHRW_COUNT_MASK) + PTHRW_INC;
+ }
+
+ if (OSAtomicCompareAndSwap32(ucntval, ulval, (volatile int32_t *)c_useqcnt) != TRUE) {
+ goto retry;
+ }
+ }
+
+ flags = 0;
+ if (cond->pshared == PTHREAD_PROCESS_SHARED)
+ flags |= _PTHREAD_MTX_OPT_PSHARED;
+
+#if _KSYN_TRACE_
+ (void)__kdebug_trace(_KSYN_TRACE_UM_CVSIG | DBG_FUNC_NONE, (uint32_t)cond, 3, nlval, ulval, 0);
+#endif
+ nlval = lcntval;
+ /* pass old u val so kernel will know the diffgen */
+ mugen = 0;
+ cvlsgen = ((uint64_t)scntval << 32) | nlval;
+
+ updateval = __psynch_cvsignal(ocond, cvlsgen, ucntval, pthread_mach_thread_np(thread), (pthread_mutex_t *)0, mugen, mtid, flags);
+
+
+ if (updateval != (uint32_t)-1) {
+
+ /* if kernel granted woke some threads, updatwe S for them as they will not access cv on their way out */
+ /* Were any threads woken or bits to be set? */
+ if (updateval != 0) {
+retry2:
+ lcntval = *c_lseqcnt;
+ ucntval = *c_useqcnt;
+ scntval = *c_sseqcnt;
+ /* update scntval with number of expected returns and bits */
+ nlval = (scntval & PTHRW_COUNT_MASK) + (updateval & PTHRW_COUNT_MASK);
+ /* set bits */
+ nlval |= ((scntval & PTH_RWS_CV_BITSALL) | (updateval & PTH_RWS_CV_BITSALL));
+
+#if _KSYN_TRACE_
+ (void)__kdebug_trace(_KSYN_TRACE_UM_CVSIG | DBG_FUNC_NONE, 0x25, 0, 0, updateval, 0);
+#endif
+ /* if L==S and c&p bits are set, needs clearpre */
+ if (((nlval & PTHRW_COUNT_MASK) == (lcntval & PTHRW_COUNT_MASK))
+ && ((nlval & PTH_RWS_CV_BITSALL) == PTH_RWS_CV_BITSALL)) {
+ /* reset p bit but retain c bit on the sword */
+ nlval &= PTH_RWS_CV_RESET_PBIT;
+ needclearpre = 1;
+ } else
+ needclearpre = 0;
+
+ oldval64 = (((uint64_t)scntval) << 32);
+ oldval64 |= lcntval;
+ newval64 = (((uint64_t)nlval) << 32);
+ newval64 |= lcntval;
+
+#if _KSYN_TRACE_
+(void)__kdebug_trace(_KSYN_TRACE_UM_CVSIG | DBG_FUNC_NONE, 0x25, nlval, ulval, updateval, 0);
+#endif
+
+ if (OSAtomicCompareAndSwap64(oldval64, newval64, (volatile int64_t *)c_lseqcnt) != TRUE)
+ goto retry2;
+
+ /* if L == S, then reset associated mutex */
+ if ((nlval & PTHRW_COUNT_MASK) == (lcntval & PTHRW_COUNT_MASK)) {
+ cond->busy = (npthread_mutex_t *)NULL;
+ }
+
+ if (needclearpre != 0) {
+ (void)__psynch_cvclrprepost(ocond, lcntval, ucntval, nlval, 0, lcntval, flags);
+ }
+ }
}
- UNLOCK(cond->lock);
- if (kern_res != KERN_SUCCESS)
- return (EINVAL);
- return (ESUCCESS);
+
+ error = 0;
+
+#if _KSYN_TRACE_
+ (void)__kdebug_trace(_KSYN_TRACE_UM_CVSIG | DBG_FUNC_END, (uint32_t)cond, 0, 0, 0, 0);
+#endif
+ return (error);
}
/*
* Manage a list of condition variables associated with a mutex
*/
-static void
-_pthread_cond_add(pthread_cond_t *cond, pthread_mutex_t *mutex)
-{
- pthread_cond_t *c;
- LOCK(mutex->lock);
- if ((c = mutex->busy) != (pthread_cond_t *)NULL)
- {
- c->prev = cond;
- }
- cond->next = c;
- cond->prev = (pthread_cond_t *)NULL;
- mutex->busy = cond;
- UNLOCK(mutex->lock);
- if (cond->sem == SEMAPHORE_NULL)
- cond->sem = new_sem_from_pool();
-}
-
-static void
-_pthread_cond_remove(pthread_cond_t *cond, pthread_mutex_t *mutex)
-{
- pthread_cond_t *n, *p;
-
- LOCK(mutex->lock);
- if ((n = cond->next) != (pthread_cond_t *)NULL)
- {
- n->prev = cond->prev;
- }
- if ((p = cond->prev) != (pthread_cond_t *)NULL)
- {
- p->next = cond->next;
- }
- else
- { /* This is the first in the list */
- mutex->busy = n;
- }
- UNLOCK(mutex->lock);
- if (cond->sigspending == 0)
- {
- restore_sem_to_pool(cond->sem);
- cond->sem = SEMAPHORE_NULL;
- }
-}
/*
* Suspend waiting for a condition variable.
* Note: we have to keep a list of condition variables which are using
* this same mutex variable so we can detect invalid 'destroy' sequences.
+ * If isconforming < 0, we skip the _pthread_testcancel(), but keep the
+ * remaining conforming behavior..
*/
-static int
-_pthread_cond_wait(pthread_cond_t *cond,
- pthread_mutex_t *mutex,
+__private_extern__ int
+_pthread_cond_wait(pthread_cond_t *ocond,
+ pthread_mutex_t *omutex,
const struct timespec *abstime,
- int isRelative)
+ int isRelative,
+ int isconforming)
{
- int res;
- kern_return_t kern_res;
- pthread_mutex_t *busy;
- mach_timespec_t then;
+ int retval;
+ npthread_cond_t * cond = (npthread_cond_t *)ocond;
+ npthread_mutex_t * mutex = (npthread_mutex_t * )omutex;
+ mach_timespec_t then = {0,0};
+ struct timespec cthen = {0,0};
int sig = cond->sig;
+ int msig = mutex->sig;
+ npthread_mutex_t * pmtx;
+ uint32_t mtxgen, mtxugen, flags=0, updateval;
+ uint32_t lcntval , ucntval, scntval;
+ uint32_t nlval, ulval, savebits;
+ volatile uint32_t * c_lseqcnt, *c_useqcnt, *c_sseqcnt;
+ uint64_t oldval64, newval64, mugen, cvlsgen;
+ uint32_t * npmtx = NULL;
+ int error, local_error;
+
+extern void _pthread_testcancel(pthread_t thread, int isconforming);
/* to provide backwards compat for apps using united condtn vars */
- if((sig != _PTHREAD_COND_SIG) && (sig !=_PTHREAD_COND_SIG_init))
+ if((sig != _PTHREAD_COND_SIG) && (sig != _PTHREAD_COND_SIG_init))
return(EINVAL);
- LOCK(cond->lock);
+
+ if (isconforming) {
+ if((msig != _PTHREAD_MUTEX_SIG) && ((msig & _PTHREAD_MUTEX_SIG_init_MASK) != _PTHREAD_MUTEX_SIG_CMP))
+ return(EINVAL);
+ if (isconforming > 0)
+ _pthread_testcancel(pthread_self(), 1);
+ }
+
if (cond->sig != _PTHREAD_COND_SIG)
{
- if (cond->sig != _PTHREAD_COND_SIG_init)
- {
- UNLOCK(cond->lock);
- return (EINVAL); /* Not a condition variable */
+ LOCK(cond->lock);
+ if (cond->sig != _PTHREAD_COND_SIG) {
+ if (cond->sig == _PTHREAD_COND_SIG_init) {
+ _pthread_cond_init(ocond, NULL, 0);
+ } else {
+ UNLOCK(cond->lock);
+ return(EINVAL);
+ }
}
- _pthread_cond_init(cond, NULL);
+ UNLOCK(cond->lock);
}
- if (abstime)
- {
- if (isRelative == 0)
- {
+#if _KSYN_TRACE_
+ (void)__kdebug_trace(_KSYN_TRACE_UM_CVWAIT | DBG_FUNC_START, (uint32_t)cond, isRelative, 0, (uint32_t)abstime, 0);
+#endif
+ COND_GETSEQ_ADDR(cond, c_lseqcnt, c_useqcnt, c_sseqcnt);
+
+ /* send relative time to kernel */
+ if (abstime) {
+#if _KSYN_TRACE_
+ (void)__kdebug_trace(_KSYN_TRACE_UM_CVWAIT | DBG_FUNC_START, 0x11111111, abstime->tv_nsec, abstime->tv_sec, 0, 0);
+#endif
+ if (isRelative == 0) {
struct timespec now;
struct timeval tv;
gettimeofday(&tv, NULL);
if (((int)then.tv_sec < 0) ||
((then.tv_sec == 0) && (then.tv_nsec == 0)))
{
- UNLOCK(cond->lock);
return ETIMEDOUT;
}
- }
- else
- {
+ if (isconforming != 0) {
+ cthen.tv_sec = abstime->tv_sec;
+ cthen.tv_nsec = abstime->tv_nsec;
+ if ((cthen.tv_sec < 0) || (cthen.tv_nsec < 0)) {
+ return EINVAL;
+ }
+ if (cthen.tv_nsec >= NSEC_PER_SEC) {
+ return EINVAL;
+ }
+ }
+ } else {
then.tv_sec = abstime->tv_sec;
then.tv_nsec = abstime->tv_nsec;
+ if ((then.tv_sec == 0) && (then.tv_nsec == 0)) {
+ return ETIMEDOUT;
+ }
}
- if (then.tv_nsec >= NSEC_PER_SEC)
- {
- UNLOCK(cond->lock);
+ if(isconforming && ((then.tv_sec < 0) || (then.tv_nsec < 0))) {
+ return EINVAL;
+ }
+ if (then.tv_nsec >= NSEC_PER_SEC) {
return EINVAL;
}
}
- if (++cond->waiters == 1)
- {
- _pthread_cond_add(cond, mutex);
- cond->busy = mutex;
- }
- else if ((busy = cond->busy) != mutex)
- {
- /* Must always specify the same mutex! */
- cond->waiters--;
- UNLOCK(cond->lock);
+ if ((cond->busy != (npthread_mutex_t *)NULL) && (cond->busy != mutex))
return (EINVAL);
- }
- UNLOCK(cond->lock);
+
+ pmtx = mutex;
+retry:
+ lcntval = *c_lseqcnt;
+ ucntval = *c_useqcnt;
+ scntval = *c_sseqcnt;
+
+ oldval64 = (((uint64_t)scntval) << 32);
+ oldval64 |= lcntval;
+
+ /* remove c and p bits on S word */
+ savebits = scntval & PTH_RWS_CV_BITSALL;
+ ulval = (scntval & PTHRW_COUNT_MASK);
+ nlval = lcntval + PTHRW_INC;
+ newval64 = (((uint64_t)ulval) << 32);
+ newval64 |= nlval;
+
+ if (OSAtomicCompareAndSwap64(oldval64, newval64, (volatile int64_t *)c_lseqcnt) != TRUE)
+ goto retry;
+
+ cond->busy = mutex;
+
+#if _KSYN_TRACE_
+ (void)__kdebug_trace(_KSYN_TRACE_UM_CVWAIT | DBG_FUNC_NONE, (uint32_t)cond, lcntval, ucntval, scntval, 0);
+#endif
+ retval = __mtx_droplock(pmtx, PTHRW_INC, &flags, &npmtx, &mtxgen, &mtxugen);
+
+ /* TBD: cases are for normal (non owner for recursive mutex; error checking)*/
+ if (retval != 0)
+ return(EINVAL);
+ if ((flags & _PTHREAD_MTX_OPT_NOTIFY) == 0) {
+ npmtx = NULL;
+ mugen = 0;
+ } else
+ mugen = ((uint64_t)mtxugen << 32) | mtxgen;
+ flags &= ~_PTHREAD_MTX_OPT_MUTEX; /* reset the mutex bit as this is cvar */
-#if defined(DEBUG)
- _pthread_mutex_remove(mutex, pthread_self());
+#if _KSYN_TRACE_
+ (void)__kdebug_trace(_KSYN_TRACE_UM_CVWAIT | DBG_FUNC_NONE, (uint32_t)cond, 3, (uint32_t)mutex, flags, 0);
#endif
- LOCK(mutex->lock);
- if (--mutex->lock_count == 0)
- {
- if (mutex->sem == SEMAPHORE_NULL)
- mutex->sem = new_sem_from_pool();
- mutex->owner = _PTHREAD_MUTEX_OWNER_SWITCHING;
- UNLOCK(mutex->lock);
+
+
+ cvlsgen = ((uint64_t)(ulval | savebits)<< 32) | nlval;
+
+ if (isconforming) {
+ pthread_cleanup_push(cond_cleanup, (void *)cond);
+ updateval = __psynch_cvwait(ocond, cvlsgen, ucntval, (pthread_mutex_t *)npmtx, mugen, flags, (int64_t)then.tv_sec, (int32_t)then.tv_nsec);
+ pthread_cleanup_pop(0);
+ } else {
+ updateval = __psynch_cvwait(ocond, cvlsgen, ucntval, (pthread_mutex_t *)npmtx, mugen, flags, (int64_t)then.tv_sec, (int32_t)then.tv_nsec);
- if (abstime) {
- kern_res = semaphore_timedwait_signal(cond->sem, mutex->sem, then);
- } else {
- PTHREAD_MACH_CALL(semaphore_wait_signal(cond->sem, mutex->sem), kern_res);
- }
}
- else
- {
- UNLOCK(mutex->lock);
- if (abstime) {
- kern_res = semaphore_timedwait(cond->sem, then);
- } else {
- PTHREAD_MACH_CALL(semaphore_wait(cond->sem), kern_res);
+
+ retval = 0;
+
+ if (updateval == (uint32_t)-1) {
+ local_error = errno;
+ error = local_error & 0xff;
+ if (error == ETIMEDOUT) {
+ retval = ETIMEDOUT;
+ } else if (error == EINTR) {
+ /*
+ ** EINTR can be treated as a spurious wakeup unless we were canceled.
+ */
+ retval = 0;
+ } else
+ retval = EINVAL;
+//#if _KSYN_TRACE_
+// (void)__kdebug_trace(0x9000070 | 0, (uint32_t)cond, 0xf1f1f2f2, local_error, error, 0);
+//#endif
+
+ /* add unlock ref to show one less waiter */
+ cond_dropwait(cond, local_error, 0);
+ } else {
+//#if _KSYN_TRACE_
+// (void)__kdebug_trace(0x9000070 | 0, (uint32_t)cond, 0xf3f3f4f4, updateval, 0, 0);
+//#endif
+ /* succesful wait */
+ if (updateval != 0) {
+ /* the return due to prepost and might have bit states */
+ /* update S and return for prepo if needed */
+ cond_dropwait(cond, 0, updateval);
}
+ retval = 0;
}
+#if _KSYN_TRACE_
+ (void)__kdebug_trace(_KSYN_TRACE_UM_CVWAIT | DBG_FUNC_NONE, (uint32_t)cond, 4, retval, 0, 0);
+#endif
+ pthread_mutex_lock(omutex);
- LOCK(cond->lock);
- cond->waiters--;
- if (cond->waiters == 0)
- {
- _pthread_cond_remove(cond, mutex);
- cond->busy = (pthread_mutex_t *)NULL;
- }
- UNLOCK(cond->lock);
- if ((res = pthread_mutex_lock(mutex)) != ESUCCESS)
- return (res);
-
- /* KERN_ABORTED can be treated as a spurious wakeup */
- if ((kern_res == KERN_SUCCESS) || (kern_res == KERN_ABORTED))
- return (ESUCCESS);
- else if (kern_res == KERN_OPERATION_TIMED_OUT)
- return (ETIMEDOUT);
- return (EINVAL);
+#if _KSYN_TRACE_
+ (void)__kdebug_trace(_KSYN_TRACE_UM_CVWAIT | DBG_FUNC_END, (uint32_t)cond, 0, 0, retval, 0);
+#endif
+ return(retval);
}
-int
-pthread_cond_wait(pthread_cond_t *cond,
- pthread_mutex_t *mutex)
+/*
+ * For the new style mutex, interlocks are not held all the time.
+ * We needed the signature to be set in the end. And we need
+ * to protect against the code getting reorganized by compiler.
+ */
+static void
+__pthread_cond_set_signature(npthread_cond_t * cond)
{
- return (_pthread_cond_wait(cond, mutex, (struct timespec *)NULL, 0));
+ cond->sig = _PTHREAD_COND_SIG;
}
-int
-pthread_cond_timedwait(pthread_cond_t *cond,
- pthread_mutex_t *mutex,
- const struct timespec *abstime)
+
+static void
+cond_cleanup(void *arg)
{
- return (_pthread_cond_wait(cond, mutex, abstime, 0));
+ npthread_cond_t *cond = (npthread_cond_t *)arg;
+ pthread_mutex_t *mutex;
+
+// 4597450: begin
+ pthread_t thread = pthread_self();
+ int thcanceled = 0;
+
+ LOCK(thread->lock);
+ thcanceled = (thread->detached & _PTHREAD_WASCANCEL);
+ UNLOCK(thread->lock);
+
+ if (thcanceled == 0)
+ return;
+
+// 4597450: end
+ mutex = (pthread_mutex_t *) cond->busy;
+
+ /* add unlock ref to show one less waiter */
+ cond_dropwait(cond, thread->cancel_error, 0);
+
+ /*
+ ** Can't do anything if this fails -- we're on the way out
+ */
+ if (mutex != NULL)
+ (void)pthread_mutex_lock(mutex);
+}
+
+#define ECVCERORR 256
+#define ECVPERORR 512
+
+void
+cond_dropwait(npthread_cond_t * cond, int error, uint32_t updateval)
+{
+ int sig = cond->sig;
+ pthread_cond_t * ocond = (pthread_cond_t *)cond;
+ int needclearpre = 0;
+ uint32_t diffgen, nlval, ulval, flags;
+ uint32_t lcntval , ucntval, scntval, lval;
+ volatile uint32_t * c_lseqcnt, *c_useqcnt, *c_sseqcnt;
+ uint64_t oldval64, newval64;
+
+ /* to provide backwards compat for apps using united condtn vars */
+
+ if (sig != _PTHREAD_COND_SIG)
+ return;
+
+ COND_GETSEQ_ADDR(cond, c_lseqcnt, c_useqcnt, c_sseqcnt);
+
+ if (error != 0) {
+ lval = PTHRW_INC;
+ if ((error & ECVCERORR) != 0)
+ lval |= PTH_RWS_CV_CBIT;
+ if ((error & ECVPERORR) != 0)
+ lval |= PTH_RWS_CV_PBIT;
+ } else {
+ lval = updateval;
+ }
+#if _KSYN_TRACE_
+ (void)__kdebug_trace(_KSYN_TRACE_UM_CDROPWT | DBG_FUNC_START, (uint32_t)cond, error, updateval, 0xee, 0);
+#endif
+retry:
+ lcntval = *c_lseqcnt;
+ ucntval = *c_useqcnt;
+ scntval = *c_sseqcnt;
+
+ diffgen = diff_genseq((lcntval & PTHRW_COUNT_MASK), (scntval & PTHRW_COUNT_MASK)); /* pendig waiters */
+#if _KSYN_TRACE_
+ (void)__kdebug_trace(_KSYN_TRACE_UM_CDROPWT | DBG_FUNC_NONE, (uint32_t)cond, lcntval, scntval, diffgen, 0);
+#endif
+ if (diffgen <= 0) {
+ /* TBD: Assert, should not be the case */
+ /* validate it is spurious and return */
+ oldval64 = (((uint64_t)scntval) << 32);
+ oldval64 |= lcntval;
+ newval64 = oldval64;
+ if (OSAtomicCompareAndSwap64(oldval64, newval64, (volatile int64_t *)c_lseqcnt) != TRUE)
+ goto retry;
+#if _KSYN_TRACE_
+ (void)__kdebug_trace(_KSYN_TRACE_UM_CDROPWT | DBG_FUNC_END, (uint32_t)cond, 0, 0, 0, 0);
+#endif
+ return;
+ }
+
+ /* update S by one */
+ oldval64 = (((uint64_t)scntval) << 32);
+ oldval64 |= lcntval;
+
+ /* update scntval with number of expected returns and bits */
+ ulval = (scntval & PTHRW_COUNT_MASK) + (lval & PTHRW_COUNT_MASK);
+ /* set bits */
+ ulval |= ((scntval & PTH_RWS_CV_BITSALL) | (lval & PTH_RWS_CV_BITSALL));
+
+ nlval = lcntval;
+
+ needclearpre = 0;
+
+ /* If L==S, need to return to kernel */
+ if ((nlval & PTHRW_COUNT_MASK) == (ulval & PTHRW_COUNT_MASK)) {
+ if ((ulval & PTH_RWS_CV_BITSALL) == PTH_RWS_CV_BITSALL) {
+ /* reset p bit but retain c bit on the sword */
+ needclearpre = 1;
+ ulval &= PTH_RWS_CV_RESET_PBIT;
+ }
+ }
+
+ newval64 = (((uint64_t)ulval) << 32);
+ newval64 |= nlval;
+
+#if _KSYN_TRACE_
+ (void)__kdebug_trace(_KSYN_TRACE_UM_CDROPWT | DBG_FUNC_NONE, (uint32_t)cond, 0xffff, nlval, ulval, 0);
+#endif
+ if (OSAtomicCompareAndSwap64(oldval64, newval64, (volatile int64_t *)c_lseqcnt) != TRUE)
+ goto retry;
+
+#if _KSYN_TRACE_
+ (void)__kdebug_trace(_KSYN_TRACE_UM_CDROPWT | DBG_FUNC_NONE, (uint32_t)cond, 2, 0, 0xee, 0);
+#endif
+ if ((nlval & PTHRW_COUNT_MASK) == (ulval & PTHRW_COUNT_MASK)) {
+ /* last usage remove the mutex */
+ cond->busy = NULL;
+ }
+
+#if _KSYN_TRACE_
+ (void)__kdebug_trace(_KSYN_TRACE_UM_CDROPWT | DBG_FUNC_NONE, nlval, ucntval, ulval, PTHRW_INC, 0);
+#endif
+ if (needclearpre != 0) {
+ flags = 0;
+ if (cond->pshared == PTHREAD_PROCESS_SHARED)
+ flags |= _PTHREAD_MTX_OPT_PSHARED;
+ /* reset prepost */
+ (void)__psynch_cvclrprepost(ocond, nlval, ucntval, ulval, 0, nlval, flags);
+ }
+#if _KSYN_TRACE_
+ (void)__kdebug_trace(_KSYN_TRACE_UM_CDROPWT | DBG_FUNC_END, nlval, ucntval, ulval, PTHRW_INC, 0);
+#endif
+ return;
}
+
int
pthread_cond_timedwait_relative_np(pthread_cond_t *cond,
pthread_mutex_t *mutex,
const struct timespec *abstime)
{
- return (_pthread_cond_wait(cond, mutex, abstime, 1));
+ return (_pthread_cond_wait(cond, mutex, abstime, 1, 0));
}
-int
-pthread_condattr_init(pthread_condattr_t *attr)
-{
- attr->sig = _PTHREAD_COND_ATTR_SIG;
- return (ESUCCESS);
-}
-int
-pthread_condattr_destroy(pthread_condattr_t *attr)
-{
- attr->sig = _PTHREAD_NO_SIG; /* Uninitialized */
- return (ESUCCESS);
-}
-int
-pthread_condattr_getpshared(const pthread_condattr_t *attr,
- int *pshared)
-{
- if (attr->sig == _PTHREAD_COND_ATTR_SIG)
- {
- *pshared = (int)PTHREAD_PROCESS_PRIVATE;
- return (ESUCCESS);
- } else
- {
- return (EINVAL); /* Not an initialized 'attribute' structure */
- }
-}
+#else /* !BUILDING_VARIANT */
+extern int _pthread_cond_wait(pthread_cond_t *cond,
+ pthread_mutex_t *mutex,
+ const struct timespec *abstime,
+ int isRelative,
+ int isconforming);
-int
-pthread_condattr_setpshared(pthread_condattr_t * attr, int pshared)
+#endif /* !BUILDING_VARIANT ] */
+/*
+ * Initialize a condition variable. Note: 'attr' is ignored.
+ */
+
+/*
+ * Initialize a condition variable. This is the public interface.
+ * We can't trust the lock, so initialize it first before taking
+ * it.
+ */
+int
+pthread_cond_init(pthread_cond_t *cond,
+ const pthread_condattr_t *attr)
{
- if (attr->sig == _PTHREAD_COND_ATTR_SIG)
- {
- if ( pshared == PTHREAD_PROCESS_PRIVATE)
- {
- /* attr->pshared = pshared */
- return (ESUCCESS);
- } else
- {
- return (EINVAL); /* Invalid parameter */
- }
- } else
- {
- return (EINVAL); /* Not an initialized 'attribute' structure */
- }
+ int conforming;
+
+#if __DARWIN_UNIX03
+ conforming = 1;
+#else /* __DARWIN_UNIX03 */
+ conforming = 0;
+#endif /* __DARWIN_UNIX03 */
+ /* lock is same offset in both structures */
+ LOCK_INIT(cond->lock);
+
+ return (_pthread_cond_init(cond, attr, conforming));
}
+/*
+int
+pthread_cond_wait(pthread_cond_t *cond,
+ pthread_mutex_t *mutex)
+
+int
+pthread_cond_timedwait(pthread_cond_t *cond,
+ pthread_mutex_t *mutex,
+ const struct timespec *abstime)
+
+moved to pthread_cancelable.c */