2 * Copyright (c) 2002-2005 Apple Computer, Inc. All rights reserved.
4 * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
6 * This file contains Original Code and/or Modifications of Original Code
7 * as defined in and that are subject to the Apple Public Source License
8 * Version 2.0 (the 'License'). You may not use this file except in
9 * compliance with the License. The rights granted to you under the License
10 * may not be used to create, or enable the creation or redistribution of,
11 * unlawful or unlicensed copies of an Apple operating system, or to
12 * circumvent, violate, or enable the circumvention or violation of, any
13 * terms of an Apple operating system software license agreement.
15 * Please obtain a copy of the License at
16 * http://www.opensource.apple.com/apsl/ and read it before using this file.
18 * The Original Code and all software distributed under the License are
19 * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
20 * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
21 * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
22 * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
23 * Please see the License for the specific language governing rights and
24 * limitations under the License.
26 * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
30 * Copyright (c) 1982, 1986, 1989, 1991, 1993, 1995
31 * The Regents of the University of California. All rights reserved.
33 * Redistribution and use in source and binary forms, with or without
34 * modification, are permitted provided that the following conditions
36 * 1. Redistributions of source code must retain the above copyright
37 * notice, this list of conditions and the following disclaimer.
38 * 2. Redistributions in binary form must reproduce the above copyright
39 * notice, this list of conditions and the following disclaimer in the
40 * documentation and/or other materials provided with the distribution.
41 * 3. All advertising materials mentioning features or use of this software
42 * must display the following acknowledgement:
43 * This product includes software developed by the University of
44 * California, Berkeley and its contributors.
45 * 4. Neither the name of the University nor the names of its contributors
46 * may be used to endorse or promote products derived from this software
47 * without specific prior written permission.
49 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
50 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
51 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
52 * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
53 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
54 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
55 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
56 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
57 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
58 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
62 * derived from @(#)ufs_ihash.c 8.7 (Berkeley) 5/17/95
65 #include <sys/param.h>
66 #include <sys/systm.h>
67 #include <sys/vnode.h>
68 #include <sys/kernel.h>
69 #include <sys/malloc.h>
71 #include <sys/queue.h>
74 #include "hfs.h" /* XXX bringup */
75 #include "hfs_cnode.h"
77 extern lck_attr_t
* hfs_lock_attr
;
78 extern lck_grp_t
* hfs_mutex_group
;
79 extern lck_grp_t
* hfs_rwlock_group
;
81 lck_grp_t
* chash_lck_grp
;
82 lck_grp_attr_t
* chash_lck_grp_attr
;
83 lck_attr_t
* chash_lck_attr
;
86 * Structures associated with cnode caching.
88 LIST_HEAD(cnodehashhead
, cnode
) *cnodehashtbl
;
89 u_long cnodehash
; /* size of hash table - 1 */
90 #define CNODEHASH(device, inum) (&cnodehashtbl[((device) + (inum)) & cnodehash])
92 lck_mtx_t hfs_chash_mutex
;
96 * Initialize cnode hash table.
102 cnodehashtbl
= hashinit(desiredvnodes
, M_HFSMNT
, &cnodehash
);
104 chash_lck_grp_attr
= lck_grp_attr_alloc_init();
105 chash_lck_grp
= lck_grp_alloc_init("cnode_hash", chash_lck_grp_attr
);
106 chash_lck_attr
= lck_attr_alloc_init();
108 lck_mtx_init(&hfs_chash_mutex
, chash_lck_grp
, chash_lck_attr
);
113 * Use the device, inum pair to find the incore cnode.
115 * If it is in core, but locked, wait for it.
119 hfs_chash_getvnode(dev_t dev
, ino_t inum
, int wantrsrc
, int skiplock
)
127 * Go through the hash list
128 * If a cnode is in the process of being cleaned out or being
129 * allocated, wait for it to be finished and then try again.
132 lck_mtx_lock(&hfs_chash_mutex
);
133 for (cp
= CNODEHASH(dev
, inum
)->lh_first
; cp
; cp
= cp
->c_hash
.le_next
) {
134 if ((cp
->c_fileid
!= inum
) || (cp
->c_dev
!= dev
))
136 /* Wait if cnode is being created or reclaimed. */
137 if (ISSET(cp
->c_hflag
, H_ALLOC
| H_TRANSIT
| H_ATTACH
)) {
138 SET(cp
->c_hflag
, H_WAITING
);
140 (void) msleep(cp
, &hfs_chash_mutex
, PDROP
| PINOD
,
141 "hfs_chash_getvnode", 0);
145 * Skip cnodes that are not in the name space anymore
146 * note that this check is done outside of the proper
147 * lock to catch nodes already in this state... this
148 * state must be rechecked after we acquire the cnode lock
150 if (cp
->c_flag
& (C_NOEXISTS
| C_DELETED
)) {
153 /* Obtain the desired vnode. */
154 vp
= wantrsrc
? cp
->c_rsrc_vp
: cp
->c_vp
;
159 lck_mtx_unlock(&hfs_chash_mutex
);
161 if ((error
= vnode_getwithvid(vp
, vid
))) {
163 * If vnode is being reclaimed, or has
164 * already changed identity, no need to wait
168 if (!skiplock
&& hfs_lock(cp
, HFS_EXCLUSIVE_LOCK
) != 0) {
174 * Skip cnodes that are not in the name space anymore
175 * we need to check again with the cnode lock held
176 * because we may have blocked acquiring the vnode ref
177 * or the lock on the cnode which would allow the node
180 if (cp
->c_flag
& (C_NOEXISTS
| C_DELETED
)) {
190 lck_mtx_unlock(&hfs_chash_mutex
);
196 * Use the device, fileid pair to find the incore cnode.
197 * If no cnode if found one is created
199 * If it is in core, but locked, wait for it.
203 hfs_chash_snoop(dev_t dev
, ino_t inum
, int (*callout
)(const struct cat_desc
*,
204 const struct cat_attr
*, void *), void * arg
)
210 * Go through the hash list
211 * If a cnode is in the process of being cleaned out or being
212 * allocated, wait for it to be finished and then try again.
214 lck_mtx_lock(&hfs_chash_mutex
);
215 for (cp
= CNODEHASH(dev
, inum
)->lh_first
; cp
; cp
= cp
->c_hash
.le_next
) {
216 if ((cp
->c_fileid
!= inum
) || (cp
->c_dev
!= dev
))
218 /* Skip cnodes being created or reclaimed. */
219 if (!ISSET(cp
->c_hflag
, H_ALLOC
| H_TRANSIT
| H_ATTACH
)) {
220 result
= callout(&cp
->c_desc
, &cp
->c_attr
, arg
);
224 lck_mtx_unlock(&hfs_chash_mutex
);
230 * Use the device, fileid pair to find the incore cnode.
231 * If no cnode if found one is created
233 * If it is in core, but locked, wait for it.
237 hfs_chash_getcnode(dev_t dev
, ino_t inum
, struct vnode
**vpp
, int wantrsrc
, int skiplock
)
240 struct cnode
*ncp
= NULL
;
245 * Go through the hash list
246 * If a cnode is in the process of being cleaned out or being
247 * allocated, wait for it to be finished and then try again.
250 lck_mtx_lock(&hfs_chash_mutex
);
253 for (cp
= CNODEHASH(dev
, inum
)->lh_first
; cp
; cp
= cp
->c_hash
.le_next
) {
254 if ((cp
->c_fileid
!= inum
) || (cp
->c_dev
!= dev
))
257 * Wait if cnode is being created, attached to or reclaimed.
259 if (ISSET(cp
->c_hflag
, H_ALLOC
| H_ATTACH
| H_TRANSIT
)) {
260 SET(cp
->c_hflag
, H_WAITING
);
262 (void) msleep(cp
, &hfs_chash_mutex
, PINOD
,
263 "hfs_chash_getcnode", 0);
267 * Skip cnodes that are not in the name space anymore
268 * note that this check is done outside of the proper
269 * lock to catch nodes already in this state... this
270 * state must be rechecked after we acquire the cnode lock
272 if (cp
->c_flag
& (C_NOEXISTS
| C_DELETED
)) {
275 vp
= wantrsrc
? cp
->c_rsrc_vp
: cp
->c_vp
;
278 * The desired vnode isn't there so tag the cnode.
280 SET(cp
->c_hflag
, H_ATTACH
);
282 lck_mtx_unlock(&hfs_chash_mutex
);
286 lck_mtx_unlock(&hfs_chash_mutex
);
288 if (vnode_getwithvid(vp
, vid
))
293 * someone else won the race to create
294 * this cnode and add it to the hash
295 * just dump our allocation
297 FREE_ZONE(ncp
, sizeof(struct cnode
), M_HFSNODE
);
300 if (!skiplock
&& hfs_lock(cp
, HFS_EXCLUSIVE_LOCK
) != 0) {
303 lck_mtx_lock(&hfs_chash_mutex
);
306 CLR(cp
->c_hflag
, H_ATTACH
);
310 * Skip cnodes that are not in the name space anymore
311 * we need to check again with the cnode lock held
312 * because we may have blocked acquiring the vnode ref
313 * or the lock on the cnode which would allow the node
316 if (cp
->c_flag
& (C_NOEXISTS
| C_DELETED
)) {
321 lck_mtx_lock(&hfs_chash_mutex
);
324 CLR(cp
->c_hflag
, H_ATTACH
);
332 * Allocate a new cnode
335 panic("%s - should never get here when skiplock is set \n", __FUNCTION__
);
338 lck_mtx_unlock(&hfs_chash_mutex
);
340 MALLOC_ZONE(ncp
, struct cnode
*, sizeof(struct cnode
), M_HFSNODE
, M_WAITOK
);
342 * since we dropped the chash lock,
343 * we need to go back and re-verify
344 * that this node hasn't come into
349 bzero(ncp
, sizeof(struct cnode
));
350 SET(ncp
->c_hflag
, H_ALLOC
);
351 ncp
->c_fileid
= inum
;
353 TAILQ_INIT(&ncp
->c_hintlist
); /* make the list empty */
355 lck_rw_init(&ncp
->c_rwlock
, hfs_rwlock_group
, hfs_lock_attr
);
357 (void) hfs_lock(ncp
, HFS_EXCLUSIVE_LOCK
);
359 /* Insert the new cnode with it's H_ALLOC flag set */
360 LIST_INSERT_HEAD(CNODEHASH(dev
, inum
), ncp
, c_hash
);
361 lck_mtx_unlock(&hfs_chash_mutex
);
370 hfs_chashwakeup(struct cnode
*cp
, int hflags
)
372 lck_mtx_lock(&hfs_chash_mutex
);
374 CLR(cp
->c_hflag
, hflags
);
376 if (ISSET(cp
->c_hflag
, H_WAITING
)) {
377 CLR(cp
->c_hflag
, H_WAITING
);
380 lck_mtx_unlock(&hfs_chash_mutex
);
385 * Re-hash two cnodes in the hash table.
389 hfs_chash_rehash(struct cnode
*cp1
, struct cnode
*cp2
)
391 lck_mtx_lock(&hfs_chash_mutex
);
393 LIST_REMOVE(cp1
, c_hash
);
394 LIST_REMOVE(cp2
, c_hash
);
395 LIST_INSERT_HEAD(CNODEHASH(cp1
->c_dev
, cp1
->c_fileid
), cp1
, c_hash
);
396 LIST_INSERT_HEAD(CNODEHASH(cp2
->c_dev
, cp2
->c_fileid
), cp2
, c_hash
);
398 lck_mtx_unlock(&hfs_chash_mutex
);
403 * Remove a cnode from the hash table.
407 hfs_chashremove(struct cnode
*cp
)
409 lck_mtx_lock(&hfs_chash_mutex
);
411 /* Check if a vnode is getting attached */
412 if (ISSET(cp
->c_hflag
, H_ATTACH
)) {
413 lck_mtx_unlock(&hfs_chash_mutex
);
416 LIST_REMOVE(cp
, c_hash
);
417 cp
->c_hash
.le_next
= NULL
;
418 cp
->c_hash
.le_prev
= NULL
;
420 lck_mtx_unlock(&hfs_chash_mutex
);
425 * Remove a cnode from the hash table and wakeup any waiters.
429 hfs_chash_abort(struct cnode
*cp
)
431 lck_mtx_lock(&hfs_chash_mutex
);
433 LIST_REMOVE(cp
, c_hash
);
434 cp
->c_hash
.le_next
= NULL
;
435 cp
->c_hash
.le_prev
= NULL
;
437 CLR(cp
->c_hflag
, H_ATTACH
| H_ALLOC
);
438 if (ISSET(cp
->c_hflag
, H_WAITING
)) {
439 CLR(cp
->c_hflag
, H_WAITING
);
442 lck_mtx_unlock(&hfs_chash_mutex
);
447 * mark a cnode as in transistion
451 hfs_chash_mark_in_transit(struct cnode
*cp
)
453 lck_mtx_lock(&hfs_chash_mutex
);
455 SET(cp
->c_hflag
, H_TRANSIT
);
457 lck_mtx_unlock(&hfs_chash_mutex
);