]> git.saurik.com Git - apple/libc.git/blobdiff - pthreads/pthread.c
Libc-262.tar.gz
[apple/libc.git] / pthreads / pthread.c
diff --git a/pthreads/pthread.c b/pthreads/pthread.c
new file mode 100644 (file)
index 0000000..14d7446
--- /dev/null
@@ -0,0 +1,1431 @@
+/*
+ * Copyright 1996 1995 by Open Software Foundation, Inc. 1997 1996 1995 1994 1993 1992 1991  
+ *              All Rights Reserved 
+ *  
+ * Permission to use, copy, modify, and distribute this software and 
+ * its documentation for any purpose and without fee is hereby granted, 
+ * provided that the above copyright notice appears in all copies and 
+ * that both the copyright notice and this permission notice appear in 
+ * supporting documentation. 
+ *  
+ * OSF DISCLAIMS ALL WARRANTIES WITH REGARD TO THIS SOFTWARE 
+ * INCLUDING ALL IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS 
+ * FOR A PARTICULAR PURPOSE. 
+ *  
+ * IN NO EVENT SHALL OSF BE LIABLE FOR ANY SPECIAL, INDIRECT, OR 
+ * CONSEQUENTIAL DAMAGES OR ANY DAMAGES WHATSOEVER RESULTING FROM 
+ * LOSS OF USE, DATA OR PROFITS, WHETHER IN ACTION OF CONTRACT, 
+ * NEGLIGENCE, OR OTHER TORTIOUS ACTION, ARISING OUT OF OR IN CONNECTION 
+ * WITH THE USE OR PERFORMANCE OF THIS SOFTWARE. 
+ * 
+ */
+/*
+ * MkLinux
+ */
+
+/*
+ * POSIX Pthread Library
+ */
+
+#define __POSIX_LIB__
+#include <assert.h>
+#include <stdio.h>     /* For printf(). */
+#include <stdlib.h>
+#include <errno.h>     /* For __mach_errno_addr() prototype. */
+#include <sys/time.h>
+#include <sys/resource.h>
+#include <sys/sysctl.h>
+#include <sys/syscall.h>
+#include <machine/vmparam.h>
+#include <mach/vm_statistics.h>
+
+#include "pthread_internals.h"
+
+/* Per-thread kernel support */
+extern void _pthread_set_self(pthread_t);
+extern void mig_init(int);
+
+/* Needed to tell the malloc subsystem we're going multithreaded */
+extern void set_malloc_singlethreaded(int);
+
+/* Used when we need to call into the kernel with no reply port */
+extern pthread_lock_t reply_port_lock;
+
+/* We'll implement this when the main thread is a pthread */
+/* Use the local _pthread struct to avoid malloc before our MiG reply port is set */
+static struct _pthread _thread = {0};
+
+/* This global should be used (carefully) by anyone needing to know if a 
+** pthread has been created.
+*/
+int __is_threaded = 0;
+static int _pthread_count = 1;
+
+static pthread_lock_t _pthread_count_lock = LOCK_INITIALIZER;
+
+/* Same implementation as LOCK, but without the __is_threaded check */
+int _spin_tries = 0;
+__private_extern__ void _spin_lock_retry(pthread_lock_t *lock)
+{ 
+       int tries = _spin_tries;
+       do {
+               if (tries-- > 0)
+                       continue;
+               syscall_thread_switch(THREAD_NULL, SWITCH_OPTION_DEPRESS, 1);
+               tries = _spin_tries;
+       } while(!_spin_lock_try(lock));
+}
+
+/* Apparently, bcopy doesn't declare _cpu_has_altivec anymore */
+int _cpu_has_altivec = 0;
+
+extern mach_port_t thread_recycle_port;
+
+/* These are used to keep track of a semaphore pool shared by mutexes and condition
+** variables.
+*/
+
+static semaphore_t *sem_pool = NULL;
+static int sem_pool_count = 0;
+static int sem_pool_current = 0;
+static pthread_lock_t sem_pool_lock = LOCK_INITIALIZER;
+
+static int default_priority;
+static int max_priority;
+static int min_priority;
+static int pthread_concurrency;
+
+/*
+ * [Internal] stack support
+ */
+size_t _pthread_stack_size = 0;
+#define STACK_LOWEST(sp)       ((sp) & ~__pthread_stack_mask)
+#define STACK_RESERVED         (sizeof (struct _pthread))
+
+#ifdef STACK_GROWS_UP
+
+/* The stack grows towards higher addresses:
+   |struct _pthread|user stack---------------->|
+   ^STACK_BASE     ^STACK_START
+   ^STACK_SELF
+   ^STACK_LOWEST  */
+#define STACK_BASE(sp)         STACK_LOWEST(sp)
+#define STACK_START(stack_low) (STACK_BASE(stack_low) + STACK_RESERVED)
+#define STACK_SELF(sp)         STACK_BASE(sp)
+
+#else
+
+/* The stack grows towards lower addresses:
+   |<----------------user stack|struct _pthread|
+   ^STACK_LOWEST               ^STACK_START    ^STACK_BASE
+                              ^STACK_SELF  */
+
+#define STACK_BASE(sp)         (((sp) | __pthread_stack_mask) + 1)
+#define STACK_START(stack_low) (STACK_BASE(stack_low) - STACK_RESERVED)
+#define STACK_SELF(sp)         STACK_START(sp)
+
+#endif
+
+#if defined(__ppc__)
+static const vm_address_t PTHREAD_STACK_HINT = 0xF0000000;
+#elif defined(__i386__)
+static const vm_address_t PTHREAD_STACK_HINT = 0xB0000000;
+#else
+#error Need to define a stack address hint for this architecture
+#endif
+
+/* Set the base address to use as the stack pointer, before adjusting due to the ABI */
+
+static int
+_pthread_allocate_stack(pthread_attr_t *attrs, void **stack)
+{
+    kern_return_t kr;
+#if 1
+    assert(attrs->stacksize >= PTHREAD_STACK_MIN);
+    if (attrs->stackaddr != NULL) {
+        assert(((vm_address_t)(attrs->stackaddr) & (vm_page_size - 1)) == 0);
+        *stack = attrs->stackaddr;
+         return 0;
+    }
+
+    *((vm_address_t *)stack) = PTHREAD_STACK_HINT;
+    kr = vm_map(mach_task_self(), (vm_address_t *)stack,
+                       attrs->stacksize + vm_page_size,
+                       vm_page_size-1,
+                       VM_MAKE_TAG(VM_MEMORY_STACK)| VM_FLAGS_ANYWHERE , MEMORY_OBJECT_NULL,
+                       0, FALSE, VM_PROT_DEFAULT, VM_PROT_ALL,
+                       VM_INHERIT_DEFAULT);
+    if (kr != KERN_SUCCESS)
+       kr = vm_allocate(mach_task_self(),
+                                       (vm_address_t *)stack, attrs->stacksize + vm_page_size,
+                                       VM_MAKE_TAG(VM_MEMORY_STACK)| VM_FLAGS_ANYWHERE);
+    if (kr != KERN_SUCCESS) {
+        return EAGAIN;
+    }
+ #ifdef STACK_GROWS_UP
+     /* The guard page is the page one higher than the stack */
+     /* The stack base is at the lowest address */
+    kr = vm_protect(mach_task_self(), *stack + attrs->stacksize, vm_page_size, FALSE, VM_PROT_NONE);
+ #else
+     /* The guard page is at the lowest address */
+     /* The stack base is the highest address */
+    kr = vm_protect(mach_task_self(), (vm_address_t)*stack, vm_page_size, FALSE, VM_PROT_NONE);
+    *stack += attrs->stacksize + vm_page_size;
+ #endif
+
+#else
+    vm_address_t cur_stack = (vm_address_t)0;
+       if (free_stacks == 0)
+       {
+           /* Allocating guard pages is done by doubling
+            * the actual stack size, since STACK_BASE() needs
+            * to have stacks aligned on stack_size. Allocating just 
+            * one page takes as much memory as allocating more pages
+            * since it will remain one entry in the vm map.
+            * Besides, allocating more than one page allows tracking the
+            * overflow pattern when the overflow is bigger than one page.
+            */
+#ifndef        NO_GUARD_PAGES
+# define       GUARD_SIZE(a)   (2*(a))
+# define       GUARD_MASK(a)   (((a)<<1) | 1)
+#else
+# define       GUARD_SIZE(a)   (a)
+# define       GUARD_MASK(a)   (a)
+#endif
+               while (lowest_stack > GUARD_SIZE(__pthread_stack_size))
+               {
+                       lowest_stack -= GUARD_SIZE(__pthread_stack_size);
+                       /* Ensure stack is there */
+                       kr = vm_allocate(mach_task_self(),
+                                        &lowest_stack,
+                                        GUARD_SIZE(__pthread_stack_size),
+                                        FALSE);
+#ifndef        NO_GUARD_PAGES
+                       if (kr == KERN_SUCCESS) {
+# ifdef        STACK_GROWS_UP
+                           kr = vm_protect(mach_task_self(),
+                                           lowest_stack+__pthread_stack_size,
+                                           __pthread_stack_size,
+                                           FALSE, VM_PROT_NONE);
+# else /* STACK_GROWS_UP */
+                           kr = vm_protect(mach_task_self(),
+                                           lowest_stack,
+                                           __pthread_stack_size,
+                                           FALSE, VM_PROT_NONE);
+                           lowest_stack += __pthread_stack_size;
+# endif        /* STACK_GROWS_UP */
+                           if (kr == KERN_SUCCESS)
+                               break;
+                       }
+#else
+                       if (kr == KERN_SUCCESS)
+                           break;
+#endif
+               }
+               if (lowest_stack > 0)
+                       free_stacks = (vm_address_t *)lowest_stack;
+               else
+               {
+                       /* Too bad.  We'll just have to take what comes.
+                          Use vm_map instead of vm_allocate so we can
+                          specify alignment.  */
+                       kr = vm_map(mach_task_self(), &lowest_stack,
+                                   GUARD_SIZE(__pthread_stack_size),
+                                   GUARD_MASK(__pthread_stack_mask),
+                                   TRUE /* anywhere */, MEMORY_OBJECT_NULL,
+                                   0, FALSE, VM_PROT_DEFAULT, VM_PROT_ALL,
+                                   VM_INHERIT_DEFAULT);
+                       /* This really shouldn't fail and if it does I don't
+                          know what to do.  */
+#ifndef        NO_GUARD_PAGES
+                       if (kr == KERN_SUCCESS) {
+# ifdef        STACK_GROWS_UP
+                           kr = vm_protect(mach_task_self(),
+                                           lowest_stack+__pthread_stack_size,
+                                           __pthread_stack_size,
+                                           FALSE, VM_PROT_NONE);
+# else /* STACK_GROWS_UP */
+                           kr = vm_protect(mach_task_self(),
+                                           lowest_stack,
+                                           __pthread_stack_size,
+                                           FALSE, VM_PROT_NONE);
+                           lowest_stack += __pthread_stack_size;
+# endif        /* STACK_GROWS_UP */
+                       }
+#endif
+                       free_stacks = (vm_address_t *)lowest_stack;
+                       lowest_stack = 0;
+               }
+               *free_stacks = 0; /* No other free stacks */
+       }
+       cur_stack = STACK_START((vm_address_t) free_stacks);
+       free_stacks = (vm_address_t *)*free_stacks;
+       cur_stack = _adjust_sp(cur_stack); /* Machine dependent stack fudging */
+#endif
+        return 0;
+}
+
+static pthread_attr_t _pthread_attr_default = {0};
+
+/*
+ * Destroy a thread attribute structure
+ */
+int       
+pthread_attr_destroy(pthread_attr_t *attr)
+{
+       if (attr->sig == _PTHREAD_ATTR_SIG)
+       {
+               return (ESUCCESS);
+       } else
+       {
+               return (EINVAL); /* Not an attribute structure! */
+       }
+}
+
+/*
+ * Get the 'detach' state from a thread attribute structure.
+ * Note: written as a helper function for info hiding
+ */
+int       
+pthread_attr_getdetachstate(const pthread_attr_t *attr, 
+                           int *detachstate)
+{
+       if (attr->sig == _PTHREAD_ATTR_SIG)
+       {
+               *detachstate = attr->detached;
+               return (ESUCCESS);
+       } else
+       {
+               return (EINVAL); /* Not an attribute structure! */
+       }
+}
+
+/*
+ * Get the 'inherit scheduling' info from a thread attribute structure.
+ * Note: written as a helper function for info hiding
+ */
+int       
+pthread_attr_getinheritsched(const pthread_attr_t *attr, 
+                            int *inheritsched)
+{
+       if (attr->sig == _PTHREAD_ATTR_SIG)
+       {
+               *inheritsched = attr->inherit;
+               return (ESUCCESS);
+       } else
+       {
+               return (EINVAL); /* Not an attribute structure! */
+       }
+}
+
+/*
+ * Get the scheduling parameters from a thread attribute structure.
+ * Note: written as a helper function for info hiding
+ */
+int       
+pthread_attr_getschedparam(const pthread_attr_t *attr, 
+                          struct sched_param *param)
+{
+       if (attr->sig == _PTHREAD_ATTR_SIG)
+       {
+               *param = attr->param;
+               return (ESUCCESS);
+       } else
+       {
+               return (EINVAL); /* Not an attribute structure! */
+       }
+}
+
+/*
+ * Get the scheduling policy from a thread attribute structure.
+ * Note: written as a helper function for info hiding
+ */
+int       
+pthread_attr_getschedpolicy(const pthread_attr_t *attr, 
+                           int *policy)
+{
+       if (attr->sig == _PTHREAD_ATTR_SIG)
+       {
+               *policy = attr->policy;
+               return (ESUCCESS);
+       } else
+       {
+               return (EINVAL); /* Not an attribute structure! */
+       }
+}
+
+static const size_t DEFAULT_STACK_SIZE = DFLSSIZ;
+/*
+ * Initialize a thread attribute structure to default values.
+ */
+int       
+pthread_attr_init(pthread_attr_t *attr)
+{
+        attr->stacksize = DEFAULT_STACK_SIZE;
+        attr->stackaddr = NULL;
+       attr->sig = _PTHREAD_ATTR_SIG;
+       attr->policy = _PTHREAD_DEFAULT_POLICY;
+       attr->param.sched_priority = default_priority;
+       attr->param.quantum = 10; /* quantum isn't public yet */
+       attr->inherit = _PTHREAD_DEFAULT_INHERITSCHED;
+       attr->detached = PTHREAD_CREATE_JOINABLE;
+        attr->freeStackOnExit = TRUE;
+       return (ESUCCESS);
+}
+
+/*
+ * Set the 'detach' state in a thread attribute structure.
+ * Note: written as a helper function for info hiding
+ */
+int       
+pthread_attr_setdetachstate(pthread_attr_t *attr, 
+                           int detachstate)
+{
+       if (attr->sig == _PTHREAD_ATTR_SIG)
+       {
+               if ((detachstate == PTHREAD_CREATE_JOINABLE) ||
+                   (detachstate == PTHREAD_CREATE_DETACHED))
+               {
+                       attr->detached = detachstate;
+                       return (ESUCCESS);
+               } else
+               {
+                       return (EINVAL);
+               }
+       } else
+       {
+               return (EINVAL); /* Not an attribute structure! */
+       }
+}
+
+/*
+ * Set the 'inherit scheduling' state in a thread attribute structure.
+ * Note: written as a helper function for info hiding
+ */
+int       
+pthread_attr_setinheritsched(pthread_attr_t *attr, 
+                            int inheritsched)
+{
+       if (attr->sig == _PTHREAD_ATTR_SIG)
+       {
+               if ((inheritsched == PTHREAD_INHERIT_SCHED) ||
+                   (inheritsched == PTHREAD_EXPLICIT_SCHED))
+               {
+                       attr->inherit = inheritsched;
+                       return (ESUCCESS);
+               } else
+               {
+                       return (EINVAL);
+               }
+       } else
+       {
+               return (EINVAL); /* Not an attribute structure! */
+       }
+}
+
+/*
+ * Set the scheduling paramters in a thread attribute structure.
+ * Note: written as a helper function for info hiding
+ */
+int       
+pthread_attr_setschedparam(pthread_attr_t *attr, 
+                          const struct sched_param *param)
+{
+       if (attr->sig == _PTHREAD_ATTR_SIG)
+       {
+               /* TODO: Validate sched_param fields */
+               attr->param = *param;
+               return (ESUCCESS);
+       } else
+       {
+               return (EINVAL); /* Not an attribute structure! */
+       }
+}
+
+/*
+ * Set the scheduling policy in a thread attribute structure.
+ * Note: written as a helper function for info hiding
+ */
+int       
+pthread_attr_setschedpolicy(pthread_attr_t *attr, 
+                           int policy)
+{
+       if (attr->sig == _PTHREAD_ATTR_SIG)
+       {
+               if ((policy == SCHED_OTHER) ||
+                   (policy == SCHED_RR) ||
+                   (policy == SCHED_FIFO))
+               {
+                       attr->policy = policy;
+                       return (ESUCCESS);
+               } else
+               {
+                       return (EINVAL);
+               }
+       } else
+       {
+               return (EINVAL); /* Not an attribute structure! */
+       }
+}
+
+/*
+ * Set the scope for the thread.
+ * We currently only provide PTHREAD_SCOPE_SYSTEM
+ */
+int
+pthread_attr_setscope(pthread_attr_t *attr,
+                            int scope)
+{
+    if (attr->sig == _PTHREAD_ATTR_SIG) {
+        if (scope == PTHREAD_SCOPE_SYSTEM) {
+            /* No attribute yet for the scope */
+            return (ESUCCESS);
+        } else if (scope == PTHREAD_SCOPE_PROCESS) {
+            return (ENOTSUP);
+        }
+    }
+    return (EINVAL); /* Not an attribute structure! */
+}
+
+/*
+ * Get the scope for the thread.
+ * We currently only provide PTHREAD_SCOPE_SYSTEM
+ */
+int
+pthread_attr_getscope(pthread_attr_t *attr,
+                            int *scope)
+{
+    if (attr->sig == _PTHREAD_ATTR_SIG) {
+        *scope = PTHREAD_SCOPE_SYSTEM;
+        return (ESUCCESS);
+    }
+    return (EINVAL); /* Not an attribute structure! */
+}
+
+/* Get the base stack address of the given thread */
+int
+pthread_attr_getstackaddr(const pthread_attr_t *attr, void **stackaddr)
+{
+    if (attr->sig == _PTHREAD_ATTR_SIG) {
+        *stackaddr = attr->stackaddr;
+        return (ESUCCESS);
+    } else {
+        return (EINVAL); /* Not an attribute structure! */
+    }
+}
+
+int
+pthread_attr_setstackaddr(pthread_attr_t *attr, void *stackaddr)
+{
+    if ((attr->sig == _PTHREAD_ATTR_SIG) && (((vm_offset_t)stackaddr & (vm_page_size - 1)) == 0)) {
+        attr->stackaddr = stackaddr;
+        attr->freeStackOnExit = FALSE;
+        return (ESUCCESS);
+    } else {
+        return (EINVAL); /* Not an attribute structure! */
+    }
+}
+
+int
+pthread_attr_getstacksize(const pthread_attr_t *attr, size_t *stacksize)
+{
+    if (attr->sig == _PTHREAD_ATTR_SIG) {
+        *stacksize = attr->stacksize;
+        return (ESUCCESS);
+    } else {
+        return (EINVAL); /* Not an attribute structure! */
+    }
+}
+
+int
+pthread_attr_setstacksize(pthread_attr_t *attr, size_t stacksize)
+{
+    if ((attr->sig == _PTHREAD_ATTR_SIG) && ((stacksize % vm_page_size) == 0) && (stacksize >= PTHREAD_STACK_MIN)) {
+        attr->stacksize = stacksize;
+        return (ESUCCESS);
+    } else {
+        return (EINVAL); /* Not an attribute structure! */
+    }
+}
+
+int
+pthread_attr_getstack(const pthread_attr_t *attr, void **stackaddr, size_t * stacksize)
+{
+    if (attr->sig == _PTHREAD_ATTR_SIG) {
+        *stackaddr = attr->stackaddr;
+        *stacksize = attr->stacksize;
+        return (ESUCCESS);
+    } else {
+        return (EINVAL); /* Not an attribute structure! */
+    }
+}
+
+int
+pthread_attr_setstack(pthread_attr_t *attr, void *stackaddr, size_t stacksize)
+{
+    if ((attr->sig == _PTHREAD_ATTR_SIG) && 
+       (((vm_offset_t)stackaddr & (vm_page_size - 1)) == 0) && 
+       ((stacksize % vm_page_size) == 0) && (stacksize >= PTHREAD_STACK_MIN)) {
+               attr->stackaddr = stackaddr;
+               attr->freeStackOnExit = FALSE;
+               attr->stacksize = stacksize;
+               return (ESUCCESS);
+    } else {
+        return (EINVAL); /* Not an attribute structure! */
+    }
+}
+
+/*
+ * Create and start execution of a new thread.
+ */
+
+static void
+_pthread_body(pthread_t self)
+{
+    _pthread_set_self(self);
+    pthread_exit((self->fun)(self->arg));
+}
+
+int
+_pthread_create(pthread_t t,
+               const pthread_attr_t *attrs,
+                void *stack,
+               const mach_port_t kernel_thread)
+{
+       int res;
+       res = ESUCCESS;
+       
+       do
+       {
+               memset(t, 0, sizeof(*t));
+               t->stacksize = attrs->stacksize;
+                t->stackaddr = (void *)stack;
+                t->kernel_thread = kernel_thread;
+               t->detached = attrs->detached;
+               t->inherit = attrs->inherit;
+               t->policy = attrs->policy;
+               t->param = attrs->param;
+                t->freeStackOnExit = attrs->freeStackOnExit;
+               t->mutexes = (struct _pthread_mutex *)NULL;
+               t->sig = _PTHREAD_SIG;
+                t->reply_port = MACH_PORT_NULL;
+                t->cthread_self = NULL;
+               LOCK_INIT(t->lock);
+               t->cancel_state = PTHREAD_CANCEL_ENABLE | PTHREAD_CANCEL_DEFERRED;
+               t->cleanup_stack = (struct _pthread_handler_rec *)NULL;
+               t->death = SEMAPHORE_NULL;
+
+               if (kernel_thread != MACH_PORT_NULL)
+                       pthread_setschedparam(t, t->policy, &t->param);
+       } while (0);
+       return (res);
+}
+
+/* Need to deprecate this in future */
+int
+_pthread_is_threaded(void)
+{
+    return __is_threaded;
+}
+
+/* Non portable public api to know whether this process has(had) atleast one thread 
+ * apart from main thread. There could be race if there is a thread in the process of
+ * creation at the time of call . It does not tell whether there are more than one thread
+ * at this point of time.
+ */
+int
+pthread_is_threaded_np(void)
+{
+    return (__is_threaded);
+}
+
+mach_port_t
+pthread_mach_thread_np(pthread_t t)
+{
+    thread_t kernel_thread;
+
+    /* Wait for the creator to initialize it */
+    while ((kernel_thread = t->kernel_thread) == MACH_PORT_NULL)
+       sched_yield();
+
+    return kernel_thread;
+}
+
+size_t
+pthread_get_stacksize_np(pthread_t t)
+{
+    return t->stacksize;
+}
+
+void *
+pthread_get_stackaddr_np(pthread_t t)
+{
+    return t->stackaddr;
+}
+
+mach_port_t
+_pthread_reply_port(pthread_t t)
+{
+    return t->reply_port;
+}
+
+
+/* returns non-zero if the current thread is the main thread */
+int
+pthread_main_np(void)
+{
+       pthread_t self = pthread_self();
+
+       return ((self->detached & _PTHREAD_CREATE_PARENT) == _PTHREAD_CREATE_PARENT);
+}
+
+static int       
+_pthread_create_suspended(pthread_t *thread, 
+              const pthread_attr_t *attr,
+              void *(*start_routine)(void *), 
+              void *arg,
+           int suspended)
+{
+       pthread_attr_t *attrs;
+       void *stack;
+       int res;
+       pthread_t t;
+       kern_return_t kern_res;
+       mach_port_t kernel_thread = MACH_PORT_NULL;
+       int needresume;
+
+       if ((attrs = (pthread_attr_t *)attr) == (pthread_attr_t *)NULL)
+       {                       /* Set up default paramters */
+               attrs = &_pthread_attr_default;
+       } else if (attrs->sig != _PTHREAD_ATTR_SIG) {
+            return EINVAL;
+       }
+       res = ESUCCESS;
+
+       /* In default policy (ie SCHED_OTHER) only sched_priority is used. Check for
+        * any change in priority or policy is needed here.
+        */
+       if (((attrs->policy != _PTHREAD_DEFAULT_POLICY)  || 
+               (attrs->param.sched_priority != default_priority)) && (suspended == 0)) {
+               needresume = 1; 
+               suspended = 1;  
+       } else 
+               needresume = 0;
+
+       do
+       {
+               /* Allocate a stack for the thread */
+                if ((res = _pthread_allocate_stack(attrs, &stack)) != 0) {
+                    break;
+                }
+               t = (pthread_t)malloc(sizeof(struct _pthread));
+               *thread = t;
+               if (suspended) {
+                       /* Create the Mach thread for this thread */
+                       PTHREAD_MACH_CALL(thread_create(mach_task_self(), &kernel_thread), kern_res);
+                       if (kern_res != KERN_SUCCESS)
+                       {
+                               printf("Can't create thread: %d\n", kern_res);
+                               res = EINVAL; /* Need better error here? */
+                               break;
+                       }
+               }
+                if ((res = _pthread_create(t, attrs, stack, kernel_thread)) != 0)
+               {
+                       break;
+               }
+               set_malloc_singlethreaded(0);
+               __is_threaded = 1;
+               LOCK(_pthread_count_lock);
+               _pthread_count++;
+               UNLOCK(_pthread_count_lock);
+
+               /* Send it on it's way */
+               t->arg = arg;
+               t->fun = start_routine;
+               /* Now set it up to execute */
+               _pthread_setup(t, _pthread_body, stack, suspended, needresume);
+       } while (0);
+       return (res);
+}
+
+int
+pthread_create(pthread_t *thread,
+           const pthread_attr_t *attr,
+           void *(*start_routine)(void *),
+           void *arg)
+{
+    return _pthread_create_suspended(thread, attr, start_routine, arg, 0);
+}
+
+int
+pthread_create_suspended_np(pthread_t *thread,
+           const pthread_attr_t *attr,
+           void *(*start_routine)(void *),
+           void *arg)
+{
+    return _pthread_create_suspended(thread, attr, start_routine, arg, 1);
+}
+
+/*
+ * Make a thread 'undetached' - no longer 'joinable' with other threads.
+ */
+int       
+pthread_detach(pthread_t thread)
+{
+       if (thread->sig == _PTHREAD_SIG)
+       {
+               LOCK(thread->lock);
+               if (thread->detached & PTHREAD_CREATE_JOINABLE)
+               {
+                       if (thread->detached & _PTHREAD_EXITED) {
+                               UNLOCK(thread->lock);
+                               pthread_join(thread, NULL);
+                               return ESUCCESS;
+                       } else {
+                               semaphore_t death = thread->death;
+
+                               thread->detached &= ~PTHREAD_CREATE_JOINABLE;
+                               thread->detached |= PTHREAD_CREATE_DETACHED;
+                               UNLOCK(thread->lock);
+                               if (death)
+                                       (void) semaphore_signal(death);
+                               return (ESUCCESS);
+                       }
+               } else {
+                       UNLOCK(thread->lock);
+                       return (EINVAL);
+               }
+       } else {
+               return (ESRCH); /* Not a valid thread */
+       }
+}
+
+
+/* 
+ * pthread_kill call to system call
+ */
+
+
+int   
+pthread_kill (
+        pthread_t th,
+        int sig)
+{
+       int error = 0;
+       
+       if ((sig < 0) || (sig > NSIG))
+               return(EINVAL);
+
+       if (th && (th->sig == _PTHREAD_SIG)) {
+               error = __pthread_kill(pthread_mach_thread_np(th), sig);
+               if (error == -1)
+                       error = errno;
+               return(error);
+       }
+       else
+               return(ESRCH);
+}
+
+/* Announce that there are pthread resources ready to be reclaimed in a */
+/* subsequent pthread_exit or reaped by pthread_join. In either case, the Mach */
+/* thread underneath is terminated right away. */
+static
+void _pthread_become_available(pthread_t thread, mach_port_t kernel_thread) {
+       mach_msg_empty_rcv_t msg;
+       kern_return_t ret;
+
+       msg.header.msgh_bits = MACH_MSGH_BITS(MACH_MSG_TYPE_MAKE_SEND,
+                                             MACH_MSG_TYPE_MOVE_SEND);
+       msg.header.msgh_size = sizeof msg - sizeof msg.trailer;
+       msg.header.msgh_remote_port = thread_recycle_port;
+       msg.header.msgh_local_port = kernel_thread; 
+       msg.header.msgh_id = (int)thread;
+       ret = mach_msg_send(&msg.header);
+       assert(ret == MACH_MSG_SUCCESS);
+}
+
+/* Reap the resources for available threads */
+static
+int _pthread_reap_thread(pthread_t th, mach_port_t kernel_thread, void **value_ptr) {
+       mach_port_type_t ptype;
+       kern_return_t ret;
+       task_t self;
+
+       self = mach_task_self();
+       if (kernel_thread != MACH_PORT_DEAD) {
+               ret = mach_port_type(self, kernel_thread, &ptype);
+               if (ret == KERN_SUCCESS && ptype != MACH_PORT_TYPE_DEAD_NAME) {
+                       /* not quite dead yet... */
+                       return EAGAIN;
+               }
+               ret = mach_port_deallocate(self, kernel_thread);
+               if (ret != KERN_SUCCESS) {
+                       fprintf(stderr,
+                                 "mach_port_deallocate(kernel_thread) failed: %s\n",
+                                 mach_error_string(ret));
+               }
+       }
+
+       if (th->reply_port != MACH_PORT_NULL) {
+               ret = mach_port_mod_refs(self, th->reply_port,
+                                                MACH_PORT_RIGHT_RECEIVE, -1);
+               if (ret != KERN_SUCCESS) {
+                       fprintf(stderr,
+                                 "mach_port_mod_refs(reply_port) failed: %s\n",
+                                 mach_error_string(ret));
+               }
+       }
+
+       if (th->freeStackOnExit) {
+               vm_address_t addr = (vm_address_t)th->stackaddr;
+               vm_size_t size;
+
+               size = (vm_size_t)th->stacksize + vm_page_size;
+               
+#if !defined(STACK_GROWS_UP)
+               addr -= size;
+#endif
+               ret = vm_deallocate(self, addr, size);
+               if (ret != KERN_SUCCESS) {
+                 fprintf(stderr,
+                                   "vm_deallocate(stack) failed: %s\n",
+                                   mach_error_string(ret));
+               }
+       }
+
+       if (value_ptr)
+               *value_ptr = th->exit_value;
+
+       if (th != &_thread)
+               free(th);
+
+       return ESUCCESS;
+}
+
+static
+void _pthread_reap_threads(void)
+{
+       mach_msg_empty_rcv_t msg;
+       kern_return_t ret;
+
+       ret = mach_msg(&msg.header, MACH_RCV_MSG|MACH_RCV_TIMEOUT, 0,
+                       sizeof(mach_msg_empty_rcv_t), thread_recycle_port, 
+                       MACH_MSG_TIMEOUT_NONE, MACH_PORT_NULL);
+       while (ret == MACH_MSG_SUCCESS) {
+               mach_port_t kernel_thread = msg.header.msgh_remote_port;
+               pthread_t thread = (pthread_t)msg.header.msgh_id;
+
+               if (_pthread_reap_thread(thread, kernel_thread, (void **)0) == EAGAIN)
+               {
+                       /* not dead yet, put it back for someone else to reap, stop here */
+                       _pthread_become_available(thread, kernel_thread);
+                       return;
+               }
+               ret = mach_msg(&msg.header, MACH_RCV_MSG|MACH_RCV_TIMEOUT, 0,
+                               sizeof(mach_msg_empty_rcv_t), thread_recycle_port, 
+                               MACH_MSG_TIMEOUT_NONE, MACH_PORT_NULL);
+       }
+}
+
+/* For compatibility... */
+
+pthread_t
+_pthread_self() {
+       return pthread_self();
+}
+
+/*
+ * Terminate a thread.
+ */
+void 
+pthread_exit(void *value_ptr)
+{
+       struct _pthread_handler_rec *handler;
+       pthread_t self = pthread_self();
+       kern_return_t kern_res;
+       int thread_count;
+
+       /* Make this thread not to receive any signals */
+       syscall(331,1);
+
+       while ((handler = self->cleanup_stack) != 0)
+       {
+               (handler->routine)(handler->arg);
+               self->cleanup_stack = handler->next;
+       }
+       _pthread_tsd_cleanup(self);
+
+       _pthread_reap_threads();
+
+       LOCK(self->lock);
+       self->detached |= _PTHREAD_EXITED;
+
+       if (self->detached & PTHREAD_CREATE_JOINABLE) {
+               mach_port_t death = self->death;
+               self->exit_value = value_ptr;
+               UNLOCK(self->lock);
+               /* the joiner will need a kernel thread reference, leave ours for it */
+               if (death) {
+                       PTHREAD_MACH_CALL(semaphore_signal(death), kern_res);
+                       if (kern_res != KERN_SUCCESS)
+                               fprintf(stderr,
+                                   "semaphore_signal(death) failed: %s\n",
+                                       mach_error_string(kern_res));
+               }
+       } else {
+               UNLOCK(self->lock);
+               /* with no joiner, we let become available consume our cached ref */
+               _pthread_become_available(self, pthread_mach_thread_np(self));
+       }
+
+       LOCK(_pthread_count_lock);
+       thread_count = --_pthread_count;
+       UNLOCK(_pthread_count_lock);
+       if (thread_count <= 0)
+               exit(0);
+
+       /* Use a new reference to terminate ourselves. Should never return. */
+       PTHREAD_MACH_CALL(thread_terminate(mach_thread_self()), kern_res);
+       fprintf(stderr, "thread_terminate(mach_thread_self()) failed: %s\n",
+                       mach_error_string(kern_res));
+       abort();
+}
+
+/*
+ * Wait for a thread to terminate and obtain its exit value.
+ */
+int       
+pthread_join(pthread_t thread, 
+            void **value_ptr)
+{
+       kern_return_t kern_res;
+       int res = ESUCCESS;
+
+       if (thread->sig == _PTHREAD_SIG)
+       {
+               semaphore_t death = new_sem_from_pool(); /* in case we need it */
+
+               LOCK(thread->lock);
+               if ((thread->detached & PTHREAD_CREATE_JOINABLE) &&
+                       thread->death == SEMAPHORE_NULL)
+               {
+                       pthread_t self = pthread_self();
+
+                       assert(thread->joiner == NULL);
+                       if (thread != self && (self == NULL || self->joiner != thread))
+                       {
+                               int already_exited = (thread->detached & _PTHREAD_EXITED);
+
+                               thread->death = death;
+                               thread->joiner = self;
+                               UNLOCK(thread->lock);
+
+                               if (!already_exited)
+                               {
+                                       /* Wait for it to signal... */ 
+                                       do {
+                                               PTHREAD_MACH_CALL(semaphore_wait(death), kern_res);
+                                       } while (kern_res != KERN_SUCCESS);
+                               }
+
+                               /* ... and wait for it to really be dead */
+                               while ((res = _pthread_reap_thread(thread,
+                                                       thread->kernel_thread,
+                                                       value_ptr)) == EAGAIN)
+                               {
+                                       sched_yield();
+                               }
+                       } else {
+                               UNLOCK(thread->lock);
+                               res = EDEADLK;
+                       }
+               } else {
+                       UNLOCK(thread->lock);
+                       res = EINVAL;
+               }
+               restore_sem_to_pool(death);
+               return res;
+       }
+       return ESRCH;
+}
+
+/*
+ * Get the scheduling policy and scheduling paramters for a thread.
+ */
+int       
+pthread_getschedparam(pthread_t thread, 
+                     int *policy,
+                     struct sched_param *param)
+{
+       if (thread->sig == _PTHREAD_SIG)
+       {
+            *policy = thread->policy;
+            *param = thread->param;
+            return (ESUCCESS);
+       } else
+       {
+               return (ESRCH);  /* Not a valid thread structure */
+       }
+}
+
+/*
+ * Set the scheduling policy and scheduling paramters for a thread.
+ */
+int       
+pthread_setschedparam(pthread_t thread, 
+                     int policy,
+                     const struct sched_param *param)
+{
+       policy_base_data_t bases;
+       policy_base_t base;
+       mach_msg_type_number_t count;
+       kern_return_t ret;
+
+       if (thread->sig == _PTHREAD_SIG)
+       {
+               switch (policy)
+               {
+               case SCHED_OTHER:
+                       bases.ts.base_priority = param->sched_priority;
+                       base = (policy_base_t)&bases.ts;
+                       count = POLICY_TIMESHARE_BASE_COUNT;
+                       break;
+               case SCHED_FIFO:
+                       bases.fifo.base_priority = param->sched_priority;
+                       base = (policy_base_t)&bases.fifo;
+                       count = POLICY_FIFO_BASE_COUNT;
+                       break;
+               case SCHED_RR:
+                       bases.rr.base_priority = param->sched_priority;
+                       /* quantum isn't public yet */
+                       bases.rr.quantum = param->quantum;
+                       base = (policy_base_t)&bases.rr;
+                       count = POLICY_RR_BASE_COUNT;
+                       break;
+               default:
+                       return (EINVAL);
+               }
+            thread->policy = policy;
+            thread->param = *param;
+               ret = thread_policy(pthread_mach_thread_np(thread), policy, base, count, TRUE);
+               if (ret != KERN_SUCCESS)
+               {
+                       return (EINVAL);
+               }
+               return (ESUCCESS);
+       } else
+       {
+               return (ESRCH);  /* Not a valid thread structure */
+       }
+}
+
+/*
+ * Get the minimum priority for the given policy
+ */
+int
+sched_get_priority_min(int policy)
+{
+    return default_priority - 16;
+}
+
+/*
+ * Get the maximum priority for the given policy
+ */
+int
+sched_get_priority_max(int policy)
+{
+    return default_priority + 16;
+}
+
+/*
+ * Determine if two thread identifiers represent the same thread.
+ */
+int       
+pthread_equal(pthread_t t1, 
+             pthread_t t2)
+{
+       return (t1 == t2);
+}
+
+void 
+cthread_set_self(void *cself)
+{
+    pthread_t self = pthread_self();
+    if ((self == (pthread_t)NULL) || (self->sig != _PTHREAD_SIG)) {
+        _pthread_set_self(cself);
+        return;
+    }
+    self->cthread_self = cself;
+}
+
+void *
+ur_cthread_self(void) {
+    pthread_t self = pthread_self();
+    if ((self == (pthread_t)NULL) || (self->sig != _PTHREAD_SIG)) {
+        return (void *)self;
+    }
+    return self->cthread_self;
+}
+
+/*
+ * Execute a function exactly one time in a thread-safe fashion.
+ */
+int       
+pthread_once(pthread_once_t *once_control, 
+            void (*init_routine)(void))
+{
+       LOCK(once_control->lock);
+       if (once_control->sig == _PTHREAD_ONCE_SIG_init)
+       {
+               (*init_routine)();
+               once_control->sig = _PTHREAD_ONCE_SIG;
+       }
+       UNLOCK(once_control->lock);
+       return (ESUCCESS);  /* Spec defines no possible errors! */
+}
+
+/*
+ * Cancel a thread
+ */
+int
+pthread_cancel(pthread_t thread)
+{
+       if (thread->sig == _PTHREAD_SIG)
+       {
+               thread->cancel_state |= _PTHREAD_CANCEL_PENDING;
+               return (ESUCCESS);
+       } else
+       {
+               return (ESRCH);
+       }
+}
+
+/*
+ * Insert a cancellation point in a thread.
+ */
+static void
+_pthread_testcancel(pthread_t thread)
+{
+       LOCK(thread->lock);
+       if ((thread->cancel_state & (PTHREAD_CANCEL_ENABLE|_PTHREAD_CANCEL_PENDING)) == 
+           (PTHREAD_CANCEL_ENABLE|_PTHREAD_CANCEL_PENDING))
+       {
+               UNLOCK(thread->lock);
+               pthread_exit(0);
+       }
+       UNLOCK(thread->lock);
+}
+
+void
+pthread_testcancel(void)
+{
+       pthread_t self = pthread_self();
+       _pthread_testcancel(self);
+}
+
+/*
+ * Query/update the cancelability 'state' of a thread
+ */
+int
+pthread_setcancelstate(int state, int *oldstate)
+{
+       pthread_t self = pthread_self();
+       int err = ESUCCESS;
+       LOCK(self->lock);
+       if (oldstate)
+               *oldstate = self->cancel_state & ~_PTHREAD_CANCEL_STATE_MASK;
+       if ((state == PTHREAD_CANCEL_ENABLE) || (state == PTHREAD_CANCEL_DISABLE))
+       {
+               self->cancel_state = (self->cancel_state & _PTHREAD_CANCEL_STATE_MASK) | state;
+       } else
+       {
+               err = EINVAL;
+       }
+       UNLOCK(self->lock);
+       _pthread_testcancel(self);  /* See if we need to 'die' now... */
+       return (err);
+}
+
+/*
+ * Query/update the cancelability 'type' of a thread
+ */
+int
+pthread_setcanceltype(int type, int *oldtype)
+{
+       pthread_t self = pthread_self();
+       int err = ESUCCESS;
+       LOCK(self->lock);
+       if (oldtype)
+               *oldtype = self->cancel_state & ~_PTHREAD_CANCEL_TYPE_MASK;
+       if ((type == PTHREAD_CANCEL_DEFERRED) || (type == PTHREAD_CANCEL_ASYNCHRONOUS))
+       {
+               self->cancel_state = (self->cancel_state & _PTHREAD_CANCEL_TYPE_MASK) | type;
+       } else
+       {
+               err = EINVAL;
+       }
+       UNLOCK(self->lock);
+       _pthread_testcancel(self);  /* See if we need to 'die' now... */
+       return (err);
+}
+
+int
+pthread_getconcurrency(void)
+{
+       return(pthread_concurrency);
+}
+
+int
+pthread_setconcurrency(int new_level)
+{
+       pthread_concurrency = new_level;
+       return(ESUCCESS);
+}
+
+/*
+ * Perform package initialization - called automatically when application starts
+ */
+
+extern int _cpu_capabilities;
+
+#define kHasAltivec     0x01                                            
+#define kCache32                0x04                                    
+#define kUseDcba                0x20                                    
+#define kNoDcba         0x40                                            
+
+static int
+pthread_init(void)
+{
+       pthread_attr_t *attrs;
+        pthread_t thread;
+       kern_return_t kr;
+       host_basic_info_data_t basic_info;
+       host_priority_info_data_t priority_info;
+       host_info_t info;
+       host_flavor_t flavor;
+       host_t host;
+       mach_msg_type_number_t count;
+       int mib[2];
+       size_t len;
+       int hasvectorunit, numcpus;
+                                                                                
+       extern  int     _bcopy_initialize(void);
+       int     dynamic_choice;
+                                                                                
+
+        count = HOST_PRIORITY_INFO_COUNT;
+       info = (host_info_t)&priority_info;
+       flavor = HOST_PRIORITY_INFO;
+       host = mach_host_self();
+       kr = host_info(host, flavor, info, &count);
+        if (kr != KERN_SUCCESS)
+                printf("host_info failed (%d); probably need privilege.\n", kr);
+        else {
+               default_priority = priority_info.user_priority;
+            min_priority = priority_info.minimum_priority;
+            max_priority = priority_info.maximum_priority;
+       }
+       attrs = &_pthread_attr_default;
+       pthread_attr_init(attrs);
+
+       thread = &_thread;
+       _pthread_set_self(thread);
+       _pthread_create(thread, attrs, (void *)USRSTACK, mach_thread_self());
+       thread->detached = PTHREAD_CREATE_JOINABLE|_PTHREAD_CREATE_PARENT;
+
+        /* See if we're on a multiprocessor and set _spin_tries if so.  */
+       mib[0] = CTL_HW;
+       mib[1] = HW_NCPU;
+       len = sizeof(numcpus);
+       if (sysctl(mib, 2, &numcpus, &len, NULL, 0) == 0) {
+               if (numcpus > 1) {
+                       _spin_tries = MP_SPIN_TRIES;
+               }
+       } else {
+               count = HOST_BASIC_INFO_COUNT;
+               info = (host_info_t)&basic_info;
+               flavor = HOST_BASIC_INFO;
+               kr = host_info(host, flavor, info, &count);
+               if (kr != KERN_SUCCESS)
+                       printf("host_info failed (%d)\n", kr);
+               else {
+                       if (basic_info.avail_cpus > 1)
+                               _spin_tries = MP_SPIN_TRIES;
+                       /* This is a crude test */
+                       if (basic_info.cpu_subtype >= CPU_SUBTYPE_POWERPC_7400) 
+                               _cpu_has_altivec = 1;
+               }
+       }
+       mach_port_deallocate(mach_task_self(), host);
+
+       mib[0] = CTL_HW;
+       mib[1] = HW_VECTORUNIT;
+       len = sizeof(hasvectorunit);
+       if (sysctl(mib, 2, &hasvectorunit, &len, NULL, 0) == 0) {
+               _cpu_has_altivec = hasvectorunit;
+       }
+       if (_cpu_has_altivec) {                                 // G4, let bcopy decide whether to use dcba                                        
+               _cpu_capabilities = kCache32 + kHasAltivec;                     
+       } else {                                                         // G3, no altivec and no dcba                                              
+               _cpu_capabilities = kCache32 + kNoDcba;                         
+       }                                                                       
+                                                                                
+       dynamic_choice = _bcopy_initialize();           // returns 0, kUseDcba, or kNoDcba                                                    
+       _cpu_capabilities |= dynamic_choice;            // remember dynamic choice, if any                                                                  
+       mig_init(1);            /* enable multi-threaded mig interfaces */
+       return 0;
+}
+
+int sched_yield(void)
+{
+    swtch_pri(0);
+    return 0;
+}
+
+/* This is the "magic" that gets the initialization routine called when the application starts */
+int (*_cthread_init_routine)(void) = pthread_init;
+
+/* Get a semaphore from the pool, growing it if necessary */
+
+__private_extern__ semaphore_t new_sem_from_pool(void) {
+       kern_return_t res;
+       semaphore_t sem;
+        int i;
+        
+       LOCK(sem_pool_lock);
+       if (sem_pool_current == sem_pool_count) {
+               sem_pool_count += 16;
+               sem_pool = realloc(sem_pool, sem_pool_count * sizeof(semaphore_t));
+               for (i = sem_pool_current; i < sem_pool_count; i++) {
+                       PTHREAD_MACH_CALL(semaphore_create(mach_task_self(), &sem_pool[i], SYNC_POLICY_FIFO, 0), res);
+               }
+       }
+       sem = sem_pool[sem_pool_current++];
+       UNLOCK(sem_pool_lock);
+       return sem;
+}
+
+/* Put a semaphore back into the pool */
+__private_extern__ void restore_sem_to_pool(semaphore_t sem) {
+       LOCK(sem_pool_lock);
+       sem_pool[--sem_pool_current] = sem;
+       UNLOCK(sem_pool_lock);
+}
+
+static void sem_pool_reset(void) {
+       LOCK(sem_pool_lock);
+       sem_pool_count = 0;
+       sem_pool_current = 0;
+       sem_pool = NULL; 
+       UNLOCK(sem_pool_lock);
+}
+
+__private_extern__ void _pthread_fork_child(void) {
+       /* Just in case somebody had it locked... */
+       UNLOCK(sem_pool_lock);
+       sem_pool_reset();
+       UNLOCK(_pthread_count_lock);
+       _pthread_count = 1;
+}
+