]> git.saurik.com Git - apple/xnu.git/blobdiff - osfmk/kern/sync_lock.c
xnu-1504.7.4.tar.gz
[apple/xnu.git] / osfmk / kern / sync_lock.c
index 027754519d31837b96699274b2eb6410b979377d..174381f5faeccfae197be468285080d195f0bbee 100644 (file)
@@ -1,23 +1,29 @@
 /*
  * Copyright (c) 2000-2004 Apple Computer, Inc. All rights reserved.
  *
- * @APPLE_LICENSE_HEADER_START@
+ * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
  * 
- * The contents of this file constitute Original Code as defined in and
- * are subject to the Apple Public Source License Version 1.1 (the
- * "License").  You may not use this file except in compliance with the
- * License.  Please obtain a copy of the License at
- * http://www.apple.com/publicsource and read it before using this file.
+ * This file contains Original Code and/or Modifications of Original Code
+ * as defined in and that are subject to the Apple Public Source License
+ * Version 2.0 (the 'License'). You may not use this file except in
+ * compliance with the License. The rights granted to you under the License
+ * may not be used to create, or enable the creation or redistribution of,
+ * unlawful or unlicensed copies of an Apple operating system, or to
+ * circumvent, violate, or enable the circumvention or violation of, any
+ * terms of an Apple operating system software license agreement.
  * 
- * This Original Code and all software distributed under the License are
- * distributed on an "AS IS" basis, WITHOUT WARRANTY OF ANY KIND, EITHER
+ * Please obtain a copy of the License at
+ * http://www.opensource.apple.com/apsl/ and read it before using this file.
+ * 
+ * The Original Code and all software distributed under the License are
+ * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
  * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
  * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
- * FITNESS FOR A PARTICULAR PURPOSE OR NON-INFRINGEMENT.  Please see the
- * License for the specific language governing rights and limitations
- * under the License.
+ * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
+ * Please see the License for the specific language governing rights and
+ * limitations under the License.
  * 
- * @APPLE_LICENSE_HEADER_END@
+ * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
  */
 /*
  * @OSF_COPYRIGHT@
@@ -62,9 +68,9 @@
 
 #define ulock_ownership_clear(ul)                              \
        MACRO_BEGIN                                             \
-       thread_t th;                                    \
+       thread_t th;                                            \
        th = (ul)->holder;                                      \
-        if (th->active) {                                      \
+        if ((th)->active) {                                    \
                thread_mtx_lock(th);                            \
                remqueue(&th->held_ulocks,                      \
                         (queue_entry_t) (ul));                 \
        MACRO_END
 
 unsigned int lock_set_event;
-#define LOCK_SET_EVENT ((event64_t)&lock_set_event)
+#define LOCK_SET_EVENT CAST_EVENT64_T(&lock_set_event)
 
 unsigned int lock_set_handoff;
-#define LOCK_SET_HANDOFF ((event64_t)&lock_set_handoff)
+#define LOCK_SET_HANDOFF CAST_EVENT64_T(&lock_set_handoff)
+
+
+lck_attr_t                             lock_set_attr;
+lck_grp_t                              lock_set_grp;
+static lck_grp_attr_t  lock_set_grp_attr;
+
+
 
 /*
  *     ROUTINE:        lock_set_init           [private]
  *
  *     Initialize the lock_set subsystem.
- *
- *     For now, we don't have anything to do here.
  */
 void
 lock_set_init(void)
 {
-       return;
+       lck_grp_attr_setdefault(&lock_set_grp_attr);
+       lck_grp_init(&lock_set_grp, "lock_set", &lock_set_grp_attr);
+       lck_attr_setdefault(&lock_set_attr);
 }
 
 
@@ -132,7 +145,7 @@ lock_set_create (
 {
        lock_set_t      lock_set = LOCK_SET_NULL;
        ulock_t         ulock;
-       int             size;
+       vm_size_t       size;
        int             x;
 
        *new_lock_set = LOCK_SET_NULL;
@@ -140,6 +153,9 @@ lock_set_create (
        if (task == TASK_NULL || n_ulocks <= 0 || policy > SYNC_POLICY_MAX)
                return KERN_INVALID_ARGUMENT;
 
+       if ((VM_MAX_ADDRESS - sizeof(struct lock_set))/sizeof(struct ulock) < (unsigned)n_ulocks)
+               return KERN_RESOURCE_SHORTAGE;
+
        size = sizeof(struct lock_set) + (sizeof(struct ulock) * (n_ulocks-1));
        lock_set = (lock_set_t) kalloc (size);
 
@@ -149,15 +165,14 @@ lock_set_create (
 
        lock_set_lock_init(lock_set);
        lock_set->n_ulocks = n_ulocks;
-       lock_set->ref_count = 1;
+       lock_set->ref_count = (task == kernel_task) ? 1 : 2; /* one for kernel, one for port */
 
        /*
         *  Create and initialize the lock set port
         */
        lock_set->port = ipc_port_alloc_kernel();
        if (lock_set->port == IP_NULL) {        
-               /* This will deallocate the lock set */
-               lock_set_dereference(lock_set);
+               kfree(lock_set, size);
                return KERN_RESOURCE_SHORTAGE; 
        }
 
@@ -177,6 +192,7 @@ lock_set_create (
                ulock->blocked   = FALSE;
                ulock->unstable  = FALSE;
                ulock->ho_wait   = FALSE;
+               ulock->accept_wait = FALSE;
                wait_queue_init(&ulock->wait_queue, policy);
        }
 
@@ -268,13 +284,10 @@ lock_set_destroy (task_t task, lock_set_t lock_set)
        lock_set_ownership_clear(lock_set, task);
 
        /*
-        *  Deallocate  
-        *
-        *  Drop the lock set reference, which inturn destroys the
-        *  lock set structure if the reference count goes to zero.
+        *  Drop the lock set reference given to the containing task,
+        *  which inturn destroys the lock set structure if the reference
+        *  count goes to zero.
         */
-
-       ipc_port_dealloc_kernel(lock_set->port);
        lock_set_dereference(lock_set);
 
        return KERN_SUCCESS;
@@ -543,16 +556,6 @@ ulock_release_internal (ulock_t ulock, thread_t thread)
                /* wait_queue now unlocked, thread locked */
 
                if (wqthread != THREAD_NULL) {
-                       /*
-                        * JMM - These ownership transfer macros have a
-                        * locking/race problem.  To keep the thread from
-                        * changing states on us (nullifying the ownership
-                        * assignment) we need to keep the thread locked
-                        * during the assignment.  But we can't because the
-                        * macros take an activation lock, which is a mutex.
-                        * Since this code was already broken before I got
-                        * here, I will leave it for now.
-                        */
                        thread_unlock(wqthread);
                        splx(s);
 
@@ -637,16 +640,11 @@ lock_handoff (lock_set_t lock_set, int lock_id)
                 *  Transfer lock ownership
                 */
                if (thread != THREAD_NULL) {
-                       /*
-                        * JMM - These ownership transfer macros have a
-                        * locking/race problem.  To keep the thread from
-                        * changing states on us (nullifying the ownership
-                        * assignment) we need to keep the thread locked
-                        * during the assignment.  But we can't because the
-                        * macros take a thread mutex lock.
-                        *
-                        * Since this code was already broken before I got
-                        * here, I will leave it for now.
+                       /* 
+                        * The thread we are transferring to will try
+                        * to take the lock on the ulock, and therefore
+                        * will wait for us complete the handoff even
+                        * through we set the thread running.
                         */
                        thread_unlock(thread);
                        splx(s);
@@ -690,7 +688,15 @@ lock_handoff (lock_set_t lock_set, int lock_id)
         */
        switch (wait_result) {
 
+
        case THREAD_AWAKENED:
+               /*
+                * we take the ulock lock to syncronize with the
+                * thread that is accepting ownership.
+                */
+               ulock_lock(ulock);
+               assert(ulock->holder != current_thread());
+               ulock_unlock(ulock);
                return KERN_SUCCESS;
 
        case THREAD_INTERRUPTED:
@@ -800,6 +806,15 @@ lock_handoff_accept (lock_set_t lock_set, int lock_id)
        switch (wait_result) {
 
        case THREAD_AWAKENED:
+               /*
+                * Take the lock to synchronize with the thread handing
+                * off the lock to us.  We don't want to continue until
+                * they complete the handoff.
+                */
+               ulock_lock(ulock);
+               assert(ulock->accept_wait == FALSE);
+               assert(ulock->holder == current_thread());
+               ulock_unlock(ulock);
                return KERN_SUCCESS;
 
        case THREAD_INTERRUPTED:
@@ -847,8 +862,9 @@ lock_set_dereference(lock_set_t lock_set)
        lock_set_unlock(lock_set);
 
        if (ref_count == 0) {
-               size =  sizeof(struct lock_set) +
-                       (sizeof(struct ulock) * (lock_set->n_ulocks - 1));
+               ipc_port_dealloc_kernel(lock_set->port);
+               size = (int)(sizeof(struct lock_set) +
+                       (sizeof(struct ulock) * (lock_set->n_ulocks - 1)));
                kfree(lock_set, size);
        }
 }