2 * Copyright (c) 2000-2004 Apple Computer, Inc. All rights reserved.
4 * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
6 * This file contains Original Code and/or Modifications of Original Code
7 * as defined in and that are subject to the Apple Public Source License
8 * Version 2.0 (the 'License'). You may not use this file except in
9 * compliance with the License. The rights granted to you under the License
10 * may not be used to create, or enable the creation or redistribution of,
11 * unlawful or unlicensed copies of an Apple operating system, or to
12 * circumvent, violate, or enable the circumvention or violation of, any
13 * terms of an Apple operating system software license agreement.
15 * Please obtain a copy of the License at
16 * http://www.opensource.apple.com/apsl/ and read it before using this file.
18 * The Original Code and all software distributed under the License are
19 * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
20 * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
21 * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
22 * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
23 * Please see the License for the specific language governing rights and
24 * limitations under the License.
26 * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
32 * Mach Operating System
33 * Copyright (c) 1991,1990,1989 Carnegie Mellon University
34 * All Rights Reserved.
36 * Permission to use, copy, modify and distribute this software and its
37 * documentation is hereby granted, provided that both the copyright
38 * notice and this permission notice appear in all copies of the
39 * software, derivative works or modified versions, and any portions
40 * thereof, and that both notices appear in supporting documentation.
42 * CARNEGIE MELLON ALLOWS FREE USE OF THIS SOFTWARE IN ITS "AS IS"
43 * CONDITION. CARNEGIE MELLON DISCLAIMS ANY LIABILITY OF ANY KIND FOR
44 * ANY DAMAGES WHATSOEVER RESULTING FROM THE USE OF THIS SOFTWARE.
46 * Carnegie Mellon requests users of this software to return to
48 * Software Distribution Coordinator or Software.Distribution@CS.CMU.EDU
49 * School of Computer Science
50 * Carnegie Mellon University
51 * Pittsburgh PA 15213-3890
53 * any improvements or extensions that they make and grant Carnegie Mellon
54 * the rights to redistribute these changes.
59 * File: ipc/ipc_entry.c
63 * Primitive functions to manipulate translation entries.
66 #include <mach_debug.h>
68 #include <mach/kern_return.h>
69 #include <mach/port.h>
70 #include <kern/assert.h>
71 #include <kern/sched_prim.h>
72 #include <kern/zalloc.h>
73 #include <kern/misc_protos.h>
75 #include <ipc/ipc_entry.h>
76 #include <ipc/ipc_space.h>
77 #include <ipc/ipc_object.h>
78 #include <ipc/ipc_hash.h>
79 #include <ipc/ipc_table.h>
80 #include <ipc/ipc_port.h>
82 #include <sys/kdebug.h>
85 * Routine: ipc_entry_lookup
87 * Searches for an entry, given its name.
89 * The space must be read or write locked throughout.
90 * The space must be active.
96 mach_port_name_t name
)
98 mach_port_index_t index
;
101 assert(is_active(space
));
103 index
= MACH_PORT_INDEX(name
);
104 if (index
< space
->is_table_size
) {
105 entry
= &space
->is_table
[index
];
106 if (IE_BITS_GEN(entry
->ie_bits
) != MACH_PORT_GEN(name
) ||
107 IE_BITS_TYPE(entry
->ie_bits
) == MACH_PORT_TYPE_NONE
) {
114 assert((entry
== IE_NULL
) || IE_BITS_TYPE(entry
->ie_bits
));
120 * Routine: ipc_entries_hold
122 * Verifies that there are at least 'entries_needed'
125 * The space is write-locked and active throughout.
126 * An object may be locked. Will not allocate memory.
128 * KERN_SUCCESS Free entries were found.
129 * KERN_NO_SPACE No entry allocated.
135 uint32_t entries_needed
)
138 mach_port_index_t next_free
= 0;
142 * Assume that all new entries will need hashing.
143 * If the table is more than 87.5% full pretend we didn't have space.
145 if (space
->is_table_hashed
+ entries_needed
>
146 space
->is_table_size
* 7 / 8) {
147 return KERN_NO_SPACE
;
150 assert(is_active(space
));
152 table
= &space
->is_table
[0];
154 for (i
= 0; i
< entries_needed
; i
++) {
155 next_free
= table
[next_free
].ie_next
;
156 if (next_free
== 0) {
157 return KERN_NO_SPACE
;
159 assert(next_free
< space
->is_table_size
);
160 assert(table
[next_free
].ie_object
== IO_NULL
);
166 * Routine: ipc_entry_claim
168 * Take formal ownership of a held entry.
170 * The space is write-locked and active throughout.
171 * An object may be locked. Will not allocate memory.
173 * Note: The returned entry must be marked as modified before
174 * releasing the space lock
180 mach_port_name_t
*namep
,
185 mach_port_index_t first_free
;
187 mach_port_name_t new_name
;
189 table
= &space
->is_table
[0];
191 first_free
= table
->ie_next
;
192 assert(first_free
!= 0);
194 entry
= &table
[first_free
];
195 table
->ie_next
= entry
->ie_next
;
196 space
->is_table_free
--;
198 assert(table
->ie_next
< space
->is_table_size
);
201 * Initialize the new entry: increment gencount and reset
202 * rollover point if it rolled over, and clear ie_request.
204 gen
= ipc_entry_new_gen(entry
->ie_bits
);
205 if (__improbable(ipc_entry_gen_rolled(entry
->ie_bits
, gen
))) {
206 ipc_entry_bits_t roll
= ipc_space_get_rollpoint(space
);
207 gen
= ipc_entry_new_rollpoint(roll
);
209 entry
->ie_bits
= gen
;
210 entry
->ie_request
= IE_REQ_NONE
;
213 * The new name can't be MACH_PORT_NULL because index
214 * is non-zero. It can't be MACH_PORT_DEAD because
215 * the table isn't allowed to grow big enough.
216 * (See comment in ipc/ipc_table.h.)
218 new_name
= MACH_PORT_MAKE(first_free
, gen
);
219 assert(MACH_PORT_VALID(new_name
));
227 * Routine: ipc_entry_get
229 * Tries to allocate an entry out of the space.
231 * The space is write-locked and active throughout.
232 * An object may be locked. Will not allocate memory.
234 * KERN_SUCCESS A free entry was found.
235 * KERN_NO_SPACE No entry allocated.
241 mach_port_name_t
*namep
,
246 kr
= ipc_entries_hold(space
, 1);
247 if (KERN_SUCCESS
!= kr
) {
251 return ipc_entry_claim(space
, namep
, entryp
);
255 * Routine: ipc_entry_alloc
257 * Allocate an entry out of the space.
259 * The space is not locked before, but it is write-locked after
260 * if the call is successful. May allocate memory.
262 * KERN_SUCCESS An entry was allocated.
263 * KERN_INVALID_TASK The space is dead.
264 * KERN_NO_SPACE No room for an entry in the space.
265 * KERN_RESOURCE_SHORTAGE Couldn't allocate memory for an entry.
271 mach_port_name_t
*namep
,
276 is_write_lock(space
);
279 if (!is_active(space
)) {
280 is_write_unlock(space
);
281 return KERN_INVALID_TASK
;
284 kr
= ipc_entry_get(space
, namep
, entryp
);
285 if (kr
== KERN_SUCCESS
) {
289 kr
= ipc_entry_grow_table(space
, ITS_SIZE_NONE
);
290 if (kr
!= KERN_SUCCESS
) {
291 return kr
; /* space is unlocked */
297 * Routine: ipc_entry_alloc_name
299 * Allocates/finds an entry with a specific name.
300 * If an existing entry is returned, its type will be nonzero.
302 * The space is not locked before, but it is write-locked after
303 * if the call is successful. May allocate memory.
305 * KERN_SUCCESS Found existing entry with same name.
306 * KERN_SUCCESS Allocated a new entry.
307 * KERN_INVALID_TASK The space is dead.
308 * KERN_RESOURCE_SHORTAGE Couldn't allocate memory.
309 * KERN_FAILURE Couldn't allocate requested name.
313 ipc_entry_alloc_name(
315 mach_port_name_t name
,
318 mach_port_index_t index
= MACH_PORT_INDEX(name
);
319 mach_port_gen_t gen
= MACH_PORT_GEN(name
);
321 if (index
> ipc_table_max_entries()) {
322 return KERN_NO_SPACE
;
325 assert(MACH_PORT_VALID(name
));
328 is_write_lock(space
);
333 if (!is_active(space
)) {
334 is_write_unlock(space
);
335 return KERN_INVALID_TASK
;
339 * If we are under the table cutoff,
340 * there are usually four cases:
341 * 1) The entry is reserved (index 0)
342 * 2) The entry is inuse, for the same name
343 * 3) The entry is inuse, for a different name
344 * 4) The entry is free
345 * For a task with a "fast" IPC space, we disallow
346 * cases 1) and 3), because ports cannot be renamed.
348 if (index
< space
->is_table_size
) {
349 ipc_entry_t table
= space
->is_table
;
351 entry
= &table
[index
];
354 /* case #1 - the entry is reserved */
355 assert(!IE_BITS_TYPE(entry
->ie_bits
));
356 assert(!IE_BITS_GEN(entry
->ie_bits
));
357 is_write_unlock(space
);
359 } else if (IE_BITS_TYPE(entry
->ie_bits
)) {
360 if (IE_BITS_GEN(entry
->ie_bits
) == gen
) {
361 /* case #2 -- the entry is inuse, for the same name */
365 /* case #3 -- the entry is inuse, for a different name. */
366 /* Collisions are not allowed */
367 is_write_unlock(space
);
371 mach_port_index_t free_index
, next_index
;
374 * case #4 -- the entry is free
375 * Rip the entry out of the free list.
379 (next_index
= table
[free_index
].ie_next
)
381 free_index
= next_index
) {
385 table
[free_index
].ie_next
=
386 table
[next_index
].ie_next
;
387 space
->is_table_free
--;
389 /* mark the previous entry modified - reconstructing the name */
390 ipc_entry_modified(space
,
391 MACH_PORT_MAKE(free_index
,
392 IE_BITS_GEN(table
[free_index
].ie_bits
)),
395 entry
->ie_bits
= gen
;
396 entry
->ie_request
= IE_REQ_NONE
;
399 assert(entry
->ie_object
== IO_NULL
);
405 * We grow the table so that the name
406 * index fits in the array space.
407 * Because the space will be unlocked,
411 kr
= ipc_entry_grow_table(space
, index
+ 1);
412 assert(kr
!= KERN_NO_SPACE
);
413 if (kr
!= KERN_SUCCESS
) {
414 /* space is unlocked */
422 * Routine: ipc_entry_dealloc
424 * Deallocates an entry from a space.
426 * The space must be write-locked throughout.
427 * The space must be active.
433 mach_port_name_t name
,
437 ipc_entry_num_t size
;
438 mach_port_index_t index
;
440 assert(is_active(space
));
441 assert(entry
->ie_object
== IO_NULL
);
442 assert(entry
->ie_request
== IE_REQ_NONE
);
445 if (entry
->ie_request
!= IE_REQ_NONE
) {
446 panic("ipc_entry_dealloc()\n");
450 index
= MACH_PORT_INDEX(name
);
451 table
= space
->is_table
;
452 size
= space
->is_table_size
;
454 if ((index
< size
) && (entry
== &table
[index
])) {
455 assert(IE_BITS_GEN(entry
->ie_bits
) == MACH_PORT_GEN(name
));
456 entry
->ie_bits
&= (IE_BITS_GEN_MASK
| IE_BITS_ROLL_MASK
);
457 entry
->ie_next
= table
->ie_next
;
458 table
->ie_next
= index
;
459 space
->is_table_free
++;
462 * Nothing to do. The entry does not match
463 * so there is nothing to deallocate.
465 assert(index
< size
);
466 assert(entry
== &table
[index
]);
467 assert(IE_BITS_GEN(entry
->ie_bits
) == MACH_PORT_GEN(name
));
469 ipc_entry_modified(space
, name
, entry
);
473 * Routine: ipc_entry_modified
475 * Note that an entry was modified in a space.
477 * Assumes exclusive write access to the space,
478 * either through a write lock or being the cleaner
479 * on an inactive space.
485 mach_port_name_t name
,
486 __assert_only ipc_entry_t entry
)
489 ipc_entry_num_t size
;
490 mach_port_index_t index
;
492 index
= MACH_PORT_INDEX(name
);
493 table
= space
->is_table
;
494 size
= space
->is_table_size
;
496 assert(index
< size
);
497 assert(entry
== &table
[index
]);
499 assert(space
->is_low_mod
<= size
);
500 assert(space
->is_high_mod
< size
);
502 if (index
< space
->is_low_mod
) {
503 space
->is_low_mod
= index
;
505 if (index
> space
->is_high_mod
) {
506 space
->is_high_mod
= index
;
509 KERNEL_DEBUG_CONSTANT(
510 MACHDBG_CODE(DBG_MACH_IPC
, MACH_IPC_PORT_ENTRY_MODIFY
) | DBG_FUNC_NONE
,
511 space
->is_task
? task_pid(space
->is_task
) : 0,
518 #define IPC_ENTRY_GROW_STATS 1
519 #if IPC_ENTRY_GROW_STATS
520 static uint64_t ipc_entry_grow_count
= 0;
521 static uint64_t ipc_entry_grow_rescan
= 0;
522 static uint64_t ipc_entry_grow_rescan_max
= 0;
523 static uint64_t ipc_entry_grow_rescan_entries
= 0;
524 static uint64_t ipc_entry_grow_rescan_entries_max
= 0;
525 static uint64_t ipc_entry_grow_freelist_entries
= 0;
526 static uint64_t ipc_entry_grow_freelist_entries_max
= 0;
530 * Routine: ipc_entry_grow_table
532 * Grows the table in a space.
534 * The space must be write-locked and active before.
535 * If successful, the space is also returned locked.
536 * On failure, the space is returned unlocked.
539 * KERN_SUCCESS Grew the table.
540 * KERN_SUCCESS Somebody else grew the table.
541 * KERN_SUCCESS The space died.
542 * KERN_NO_SPACE Table has maximum size already.
543 * KERN_RESOURCE_SHORTAGE Couldn't allocate a new table.
547 ipc_entry_grow_table(
549 ipc_table_elems_t target_size
)
551 ipc_entry_num_t osize
, size
, nsize
, psize
;
553 ipc_entry_t otable
, table
;
554 ipc_table_size_t oits
, its
, nits
;
555 mach_port_index_t i
, free_index
;
556 mach_port_index_t low_mod
, hi_mod
;
557 ipc_table_index_t sanity
;
558 #if IPC_ENTRY_GROW_STATS
559 uint64_t rescan_count
= 0;
561 assert(is_active(space
));
563 if (is_growing(space
)) {
565 * Somebody else is growing the table.
566 * We just wait for them to finish.
569 is_write_sleep(space
);
573 otable
= space
->is_table
;
575 its
= space
->is_table_next
;
576 size
= its
->its_size
;
579 * Since is_table_next points to the next natural size
580 * we can identify the current size entry.
583 osize
= oits
->its_size
;
586 * If there is no target size, then the new size is simply
587 * specified by is_table_next. If there is a target
588 * size, then search for the next entry.
590 if (target_size
!= ITS_SIZE_NONE
) {
591 if (target_size
<= osize
) {
592 /* the space is locked */
597 while ((psize
!= size
) && (target_size
> size
)) {
600 size
= its
->its_size
;
603 is_write_unlock(space
);
604 return KERN_NO_SPACE
;
609 is_write_unlock(space
);
610 return KERN_NO_SPACE
;
614 nsize
= nits
->its_size
;
615 assert((osize
< size
) && (size
<= nsize
));
618 * We'll attempt to grow the table.
620 * Because we will be copying without the space lock, reset
621 * the lowest_mod index to just beyond the end of the current
622 * table. Modification of entries (other than hashes) will
623 * bump this downward, and we only have to reprocess entries
624 * above that mark. Eventually, we'll get done.
626 is_start_growing(space
);
627 space
->is_low_mod
= osize
;
628 space
->is_high_mod
= 0;
629 #if IPC_ENTRY_GROW_STATS
630 ipc_entry_grow_count
++;
632 is_write_unlock(space
);
634 table
= it_entries_alloc(its
);
635 if (table
== IE_NULL
) {
636 is_write_lock(space
);
637 is_done_growing(space
);
638 is_write_unlock(space
);
639 thread_wakeup((event_t
) space
);
640 return KERN_RESOURCE_SHORTAGE
;
643 ipc_space_rand_freelist(space
, table
, osize
, size
);
645 /* clear out old entries in new table */
646 memset((void *)table
, 0, osize
* sizeof(*table
));
652 * Within the range of the table that changed, determine what we
653 * have to take action on. For each entry, take a snapshot of the
654 * corresponding entry in the old table (so it won't change
655 * during this iteration). The snapshot may not be self-consistent
656 * (if we caught it in the middle of being changed), so be very
657 * cautious with the values.
659 for (i
= low_mod
; i
<= hi_mod
; i
++) {
660 ipc_entry_t entry
= &table
[i
];
661 struct ipc_entry osnap
= otable
[i
];
663 if (entry
->ie_object
!= osnap
.ie_object
||
664 IE_BITS_TYPE(entry
->ie_bits
) != IE_BITS_TYPE(osnap
.ie_bits
)) {
665 if (entry
->ie_object
!= IO_NULL
&&
666 IE_BITS_TYPE(entry
->ie_bits
) == MACH_PORT_TYPE_SEND
) {
667 ipc_hash_table_delete(table
, size
, entry
->ie_object
, i
, entry
);
670 entry
->ie_object
= osnap
.ie_object
;
671 entry
->ie_bits
= osnap
.ie_bits
;
672 entry
->ie_request
= osnap
.ie_request
; /* or ie_next */
674 if (entry
->ie_object
!= IO_NULL
&&
675 IE_BITS_TYPE(entry
->ie_bits
) == MACH_PORT_TYPE_SEND
) {
676 ipc_hash_table_insert(table
, size
, entry
->ie_object
, i
, entry
);
679 assert(entry
->ie_object
== osnap
.ie_object
);
680 entry
->ie_bits
= osnap
.ie_bits
;
681 entry
->ie_request
= osnap
.ie_request
; /* or ie_next */
684 table
[0].ie_next
= otable
[0].ie_next
; /* always rebase the freelist */
687 * find the end of the freelist (should be short). But be careful,
688 * the list items can change so only follow through truly free entries
689 * (no problem stopping short in those cases, because we'll rescan).
692 for (sanity
= 0; sanity
< osize
; sanity
++) {
693 if (table
[free_index
].ie_object
!= IPC_OBJECT_NULL
) {
696 i
= table
[free_index
].ie_next
;
697 if (i
== 0 || i
>= osize
) {
702 #if IPC_ENTRY_GROW_STATS
703 ipc_entry_grow_freelist_entries
+= sanity
;
704 if (sanity
> ipc_entry_grow_freelist_entries_max
) {
705 ipc_entry_grow_freelist_entries_max
= sanity
;
709 is_write_lock(space
);
712 * We need to do a wakeup on the space,
713 * to rouse waiting threads. We defer
714 * this until the space is unlocked,
715 * because we don't want them to spin.
718 if (!is_active(space
)) {
720 * The space died while it was unlocked.
723 is_done_growing(space
);
724 is_write_unlock(space
);
725 thread_wakeup((event_t
) space
);
726 it_entries_free(its
, table
);
727 is_write_lock(space
);
731 /* If the space changed while unlocked, go back and process the changes */
732 if (space
->is_low_mod
< osize
) {
733 assert(space
->is_high_mod
> 0);
734 low_mod
= space
->is_low_mod
;
735 space
->is_low_mod
= osize
;
736 hi_mod
= space
->is_high_mod
;
737 space
->is_high_mod
= 0;
738 is_write_unlock(space
);
739 #if IPC_ENTRY_GROW_STATS
741 if (rescan_count
> ipc_entry_grow_rescan_max
) {
742 ipc_entry_grow_rescan_max
= rescan_count
;
745 ipc_entry_grow_rescan
++;
746 ipc_entry_grow_rescan_entries
+= hi_mod
- low_mod
+ 1;
747 if (hi_mod
- low_mod
+ 1 > ipc_entry_grow_rescan_entries_max
) {
748 ipc_entry_grow_rescan_entries_max
= hi_mod
- low_mod
+ 1;
754 /* link new free entries onto the rest of the freelist */
755 assert(table
[free_index
].ie_next
== 0 &&
756 table
[free_index
].ie_object
== IO_NULL
);
757 table
[free_index
].ie_next
= osize
;
759 assert(space
->is_table
== otable
);
760 assert((space
->is_table_next
== its
) ||
761 (target_size
!= ITS_SIZE_NONE
));
762 assert(space
->is_table_size
== osize
);
764 space
->is_table
= table
;
765 space
->is_table_size
= size
;
766 space
->is_table_next
= nits
;
767 space
->is_table_free
+= size
- osize
;
769 is_done_growing(space
);
770 is_write_unlock(space
);
772 thread_wakeup((event_t
) space
);
775 * Now we need to free the old table.
777 it_entries_free(oits
, otable
);
778 is_write_lock(space
);
785 * Routine: ipc_entry_name_mask
787 * Ensure a mach port name has the default ipc entry
788 * generation bits set. This can be used to ensure that
789 * a name passed in by user space matches names generated
794 * 'name' input with default generation bits masked or added
798 ipc_entry_name_mask(mach_port_name_t name
)
801 static mach_port_name_t null_name
= MACH_PORT_MAKE(0, IE_BITS_GEN_MASK
+ IE_BITS_GEN_ONE
);
802 return name
| null_name
;
804 static mach_port_name_t null_name
= MACH_PORT_MAKE(0, ~(IE_BITS_GEN_MASK
+ IE_BITS_GEN_ONE
));
805 return name
& ~null_name
;