2 * Copyright (c) 2002-2005 Apple Computer, Inc. All rights reserved.
4 * @APPLE_LICENSE_HEADER_START@
6 * The contents of this file constitute Original Code as defined in and
7 * are subject to the Apple Public Source License Version 1.1 (the
8 * "License"). You may not use this file except in compliance with the
9 * License. Please obtain a copy of the License at
10 * http://www.apple.com/publicsource and read it before using this file.
12 * This Original Code and all software distributed under the License are
13 * distributed on an "AS IS" basis, WITHOUT WARRANTY OF ANY KIND, EITHER
14 * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
15 * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
16 * FITNESS FOR A PARTICULAR PURPOSE OR NON-INFRINGEMENT. Please see the
17 * License for the specific language governing rights and limitations
20 * @APPLE_LICENSE_HEADER_END@
24 * Copyright (c) 1982, 1986, 1989, 1991, 1993, 1995
25 * The Regents of the University of California. All rights reserved.
27 * Redistribution and use in source and binary forms, with or without
28 * modification, are permitted provided that the following conditions
30 * 1. Redistributions of source code must retain the above copyright
31 * notice, this list of conditions and the following disclaimer.
32 * 2. Redistributions in binary form must reproduce the above copyright
33 * notice, this list of conditions and the following disclaimer in the
34 * documentation and/or other materials provided with the distribution.
35 * 3. All advertising materials mentioning features or use of this software
36 * must display the following acknowledgement:
37 * This product includes software developed by the University of
38 * California, Berkeley and its contributors.
39 * 4. Neither the name of the University nor the names of its contributors
40 * may be used to endorse or promote products derived from this software
41 * without specific prior written permission.
43 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
44 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
45 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
46 * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
47 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
48 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
49 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
50 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
51 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
52 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
56 * derived from @(#)ufs_ihash.c 8.7 (Berkeley) 5/17/95
59 #include <sys/param.h>
60 #include <sys/systm.h>
61 #include <sys/vnode.h>
62 #include <sys/kernel.h>
63 #include <sys/malloc.h>
65 #include <sys/queue.h>
68 #include "hfs.h" /* XXX bringup */
69 #include "hfs_cnode.h"
71 extern lck_attr_t
* hfs_lock_attr
;
72 extern lck_grp_t
* hfs_mutex_group
;
73 extern lck_grp_t
* hfs_rwlock_group
;
75 lck_grp_t
* chash_lck_grp
;
76 lck_grp_attr_t
* chash_lck_grp_attr
;
77 lck_attr_t
* chash_lck_attr
;
80 * Structures associated with cnode caching.
82 LIST_HEAD(cnodehashhead
, cnode
) *cnodehashtbl
;
83 u_long cnodehash
; /* size of hash table - 1 */
84 #define CNODEHASH(device, inum) (&cnodehashtbl[((device) + (inum)) & cnodehash])
86 lck_mtx_t hfs_chash_mutex
;
90 * Initialize cnode hash table.
96 cnodehashtbl
= hashinit(desiredvnodes
, M_HFSMNT
, &cnodehash
);
98 chash_lck_grp_attr
= lck_grp_attr_alloc_init();
99 chash_lck_grp
= lck_grp_alloc_init("cnode_hash", chash_lck_grp_attr
);
100 chash_lck_attr
= lck_attr_alloc_init();
102 lck_mtx_init(&hfs_chash_mutex
, chash_lck_grp
, chash_lck_attr
);
107 * Use the device, inum pair to find the incore cnode.
109 * If it is in core, but locked, wait for it.
113 hfs_chash_getvnode(dev_t dev
, ino_t inum
, int wantrsrc
, int skiplock
)
121 * Go through the hash list
122 * If a cnode is in the process of being cleaned out or being
123 * allocated, wait for it to be finished and then try again.
126 lck_mtx_lock(&hfs_chash_mutex
);
127 for (cp
= CNODEHASH(dev
, inum
)->lh_first
; cp
; cp
= cp
->c_hash
.le_next
) {
128 if ((cp
->c_fileid
!= inum
) || (cp
->c_dev
!= dev
))
130 /* Wait if cnode is being created or reclaimed. */
131 if (ISSET(cp
->c_hflag
, H_ALLOC
| H_TRANSIT
| H_ATTACH
)) {
132 SET(cp
->c_hflag
, H_WAITING
);
134 (void) msleep(cp
, &hfs_chash_mutex
, PDROP
| PINOD
,
135 "hfs_chash_getvnode", 0);
139 * Skip cnodes that are not in the name space anymore
140 * note that this check is done outside of the proper
141 * lock to catch nodes already in this state... this
142 * state must be rechecked after we acquire the cnode lock
144 if (cp
->c_flag
& (C_NOEXISTS
| C_DELETED
)) {
147 /* Obtain the desired vnode. */
148 vp
= wantrsrc
? cp
->c_rsrc_vp
: cp
->c_vp
;
153 lck_mtx_unlock(&hfs_chash_mutex
);
155 if ((error
= vnode_getwithvid(vp
, vid
))) {
157 * If vnode is being reclaimed, or has
158 * already changed identity, no need to wait
162 if (!skiplock
&& hfs_lock(cp
, HFS_EXCLUSIVE_LOCK
) != 0) {
168 * Skip cnodes that are not in the name space anymore
169 * we need to check again with the cnode lock held
170 * because we may have blocked acquiring the vnode ref
171 * or the lock on the cnode which would allow the node
174 if (cp
->c_flag
& (C_NOEXISTS
| C_DELETED
)) {
184 lck_mtx_unlock(&hfs_chash_mutex
);
190 * Use the device, fileid pair to find the incore cnode.
191 * If no cnode if found one is created
193 * If it is in core, but locked, wait for it.
197 hfs_chash_snoop(dev_t dev
, ino_t inum
, int (*callout
)(const struct cat_desc
*,
198 const struct cat_attr
*, void *), void * arg
)
204 * Go through the hash list
205 * If a cnode is in the process of being cleaned out or being
206 * allocated, wait for it to be finished and then try again.
208 lck_mtx_lock(&hfs_chash_mutex
);
209 for (cp
= CNODEHASH(dev
, inum
)->lh_first
; cp
; cp
= cp
->c_hash
.le_next
) {
210 if ((cp
->c_fileid
!= inum
) || (cp
->c_dev
!= dev
))
212 /* Skip cnodes being created or reclaimed. */
213 if (!ISSET(cp
->c_hflag
, H_ALLOC
| H_TRANSIT
| H_ATTACH
)) {
214 result
= callout(&cp
->c_desc
, &cp
->c_attr
, arg
);
218 lck_mtx_unlock(&hfs_chash_mutex
);
224 * Use the device, fileid pair to find the incore cnode.
225 * If no cnode if found one is created
227 * If it is in core, but locked, wait for it.
231 hfs_chash_getcnode(dev_t dev
, ino_t inum
, struct vnode
**vpp
, int wantrsrc
, int skiplock
)
234 struct cnode
*ncp
= NULL
;
239 * Go through the hash list
240 * If a cnode is in the process of being cleaned out or being
241 * allocated, wait for it to be finished and then try again.
244 lck_mtx_lock(&hfs_chash_mutex
);
247 for (cp
= CNODEHASH(dev
, inum
)->lh_first
; cp
; cp
= cp
->c_hash
.le_next
) {
248 if ((cp
->c_fileid
!= inum
) || (cp
->c_dev
!= dev
))
251 * Wait if cnode is being created, attached to or reclaimed.
253 if (ISSET(cp
->c_hflag
, H_ALLOC
| H_ATTACH
| H_TRANSIT
)) {
254 SET(cp
->c_hflag
, H_WAITING
);
256 (void) msleep(cp
, &hfs_chash_mutex
, PINOD
,
257 "hfs_chash_getcnode", 0);
261 * Skip cnodes that are not in the name space anymore
262 * note that this check is done outside of the proper
263 * lock to catch nodes already in this state... this
264 * state must be rechecked after we acquire the cnode lock
266 if (cp
->c_flag
& (C_NOEXISTS
| C_DELETED
)) {
269 vp
= wantrsrc
? cp
->c_rsrc_vp
: cp
->c_vp
;
272 * The desired vnode isn't there so tag the cnode.
274 SET(cp
->c_hflag
, H_ATTACH
);
276 lck_mtx_unlock(&hfs_chash_mutex
);
280 lck_mtx_unlock(&hfs_chash_mutex
);
282 if (vnode_getwithvid(vp
, vid
))
287 * someone else won the race to create
288 * this cnode and add it to the hash
289 * just dump our allocation
291 FREE_ZONE(ncp
, sizeof(struct cnode
), M_HFSNODE
);
294 if (!skiplock
&& hfs_lock(cp
, HFS_EXCLUSIVE_LOCK
) != 0) {
297 lck_mtx_lock(&hfs_chash_mutex
);
300 CLR(cp
->c_hflag
, H_ATTACH
);
304 * Skip cnodes that are not in the name space anymore
305 * we need to check again with the cnode lock held
306 * because we may have blocked acquiring the vnode ref
307 * or the lock on the cnode which would allow the node
310 if (cp
->c_flag
& (C_NOEXISTS
| C_DELETED
)) {
315 lck_mtx_lock(&hfs_chash_mutex
);
318 CLR(cp
->c_hflag
, H_ATTACH
);
326 * Allocate a new cnode
329 panic("%s - should never get here when skiplock is set \n", __FUNCTION__
);
332 lck_mtx_unlock(&hfs_chash_mutex
);
334 MALLOC_ZONE(ncp
, struct cnode
*, sizeof(struct cnode
), M_HFSNODE
, M_WAITOK
);
336 * since we dropped the chash lock,
337 * we need to go back and re-verify
338 * that this node hasn't come into
343 bzero(ncp
, sizeof(struct cnode
));
344 SET(ncp
->c_hflag
, H_ALLOC
);
345 ncp
->c_fileid
= inum
;
347 TAILQ_INIT(&ncp
->c_hintlist
); /* make the list empty */
349 lck_rw_init(&ncp
->c_rwlock
, hfs_rwlock_group
, hfs_lock_attr
);
351 (void) hfs_lock(ncp
, HFS_EXCLUSIVE_LOCK
);
353 /* Insert the new cnode with it's H_ALLOC flag set */
354 LIST_INSERT_HEAD(CNODEHASH(dev
, inum
), ncp
, c_hash
);
355 lck_mtx_unlock(&hfs_chash_mutex
);
364 hfs_chashwakeup(struct cnode
*cp
, int hflags
)
366 lck_mtx_lock(&hfs_chash_mutex
);
368 CLR(cp
->c_hflag
, hflags
);
370 if (ISSET(cp
->c_hflag
, H_WAITING
)) {
371 CLR(cp
->c_hflag
, H_WAITING
);
374 lck_mtx_unlock(&hfs_chash_mutex
);
379 * Re-hash two cnodes in the hash table.
383 hfs_chash_rehash(struct cnode
*cp1
, struct cnode
*cp2
)
385 lck_mtx_lock(&hfs_chash_mutex
);
387 LIST_REMOVE(cp1
, c_hash
);
388 LIST_REMOVE(cp2
, c_hash
);
389 LIST_INSERT_HEAD(CNODEHASH(cp1
->c_dev
, cp1
->c_fileid
), cp1
, c_hash
);
390 LIST_INSERT_HEAD(CNODEHASH(cp2
->c_dev
, cp2
->c_fileid
), cp2
, c_hash
);
392 lck_mtx_unlock(&hfs_chash_mutex
);
397 * Remove a cnode from the hash table.
401 hfs_chashremove(struct cnode
*cp
)
403 lck_mtx_lock(&hfs_chash_mutex
);
405 /* Check if a vnode is getting attached */
406 if (ISSET(cp
->c_hflag
, H_ATTACH
)) {
407 lck_mtx_unlock(&hfs_chash_mutex
);
410 LIST_REMOVE(cp
, c_hash
);
411 cp
->c_hash
.le_next
= NULL
;
412 cp
->c_hash
.le_prev
= NULL
;
414 lck_mtx_unlock(&hfs_chash_mutex
);
419 * Remove a cnode from the hash table and wakeup any waiters.
423 hfs_chash_abort(struct cnode
*cp
)
425 lck_mtx_lock(&hfs_chash_mutex
);
427 LIST_REMOVE(cp
, c_hash
);
428 cp
->c_hash
.le_next
= NULL
;
429 cp
->c_hash
.le_prev
= NULL
;
431 CLR(cp
->c_hflag
, H_ATTACH
| H_ALLOC
);
432 if (ISSET(cp
->c_hflag
, H_WAITING
)) {
433 CLR(cp
->c_hflag
, H_WAITING
);
436 lck_mtx_unlock(&hfs_chash_mutex
);
441 * mark a cnode as in transistion
445 hfs_chash_mark_in_transit(struct cnode
*cp
)
447 lck_mtx_lock(&hfs_chash_mutex
);
449 SET(cp
->c_hflag
, H_TRANSIT
);
451 lck_mtx_unlock(&hfs_chash_mutex
);