]> git.saurik.com Git - apple/xnu.git/blob - bsd/nfs/nfs_node.c
xnu-6153.61.1.tar.gz
[apple/xnu.git] / bsd / nfs / nfs_node.c
1 /*
2 * Copyright (c) 2000-2019 Apple Inc. All rights reserved.
3 *
4 * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
5 *
6 * This file contains Original Code and/or Modifications of Original Code
7 * as defined in and that are subject to the Apple Public Source License
8 * Version 2.0 (the 'License'). You may not use this file except in
9 * compliance with the License. The rights granted to you under the License
10 * may not be used to create, or enable the creation or redistribution of,
11 * unlawful or unlicensed copies of an Apple operating system, or to
12 * circumvent, violate, or enable the circumvention or violation of, any
13 * terms of an Apple operating system software license agreement.
14 *
15 * Please obtain a copy of the License at
16 * http://www.opensource.apple.com/apsl/ and read it before using this file.
17 *
18 * The Original Code and all software distributed under the License are
19 * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
20 * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
21 * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
22 * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
23 * Please see the License for the specific language governing rights and
24 * limitations under the License.
25 *
26 * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
27 */
28 /* Copyright (c) 1995 NeXT Computer, Inc. All Rights Reserved */
29 /*
30 * Copyright (c) 1989, 1993
31 * The Regents of the University of California. All rights reserved.
32 *
33 * This code is derived from software contributed to Berkeley by
34 * Rick Macklem at The University of Guelph.
35 *
36 * Redistribution and use in source and binary forms, with or without
37 * modification, are permitted provided that the following conditions
38 * are met:
39 * 1. Redistributions of source code must retain the above copyright
40 * notice, this list of conditions and the following disclaimer.
41 * 2. Redistributions in binary form must reproduce the above copyright
42 * notice, this list of conditions and the following disclaimer in the
43 * documentation and/or other materials provided with the distribution.
44 * 3. All advertising materials mentioning features or use of this software
45 * must display the following acknowledgement:
46 * This product includes software developed by the University of
47 * California, Berkeley and its contributors.
48 * 4. Neither the name of the University nor the names of its contributors
49 * may be used to endorse or promote products derived from this software
50 * without specific prior written permission.
51 *
52 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
53 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
54 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
55 * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
56 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
57 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
58 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
59 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
60 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
61 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
62 * SUCH DAMAGE.
63 *
64 * @(#)nfs_node.c 8.6 (Berkeley) 5/22/95
65 * FreeBSD-Id: nfs_node.c,v 1.22 1997/10/28 14:06:20 bde Exp $
66 */
67
68
69 #include <sys/param.h>
70 #include <sys/kernel.h>
71 #include <sys/systm.h>
72 #include <sys/proc.h>
73 #include <sys/kauth.h>
74 #include <sys/mount_internal.h>
75 #include <sys/vnode_internal.h>
76 #include <sys/vnode.h>
77 #include <sys/ubc.h>
78 #include <sys/malloc.h>
79 #include <sys/fcntl.h>
80 #include <sys/time.h>
81
82 #include <nfs/rpcv2.h>
83 #include <nfs/nfsproto.h>
84 #include <nfs/nfs.h>
85 #include <nfs/nfsnode.h>
86 #include <nfs/nfs_gss.h>
87 #include <nfs/nfsmount.h>
88
89 #define NFSNOHASH(fhsum) \
90 (&nfsnodehashtbl[(fhsum) & nfsnodehash])
91 static LIST_HEAD(nfsnodehashhead, nfsnode) * nfsnodehashtbl;
92 static u_long nfsnodehash;
93
94 static lck_grp_t *nfs_node_hash_lck_grp;
95 static lck_grp_t *nfs_node_lck_grp;
96 static lck_grp_t *nfs_data_lck_grp;
97 lck_mtx_t *nfs_node_hash_mutex;
98
99 #define NFS_NODE_DBG(...) NFS_DBG(NFS_FAC_NODE, 7, ## __VA_ARGS__)
100
101 /*
102 * Initialize hash links for nfsnodes
103 * and build nfsnode free list.
104 */
105 void
106 nfs_nhinit(void)
107 {
108 nfs_node_hash_lck_grp = lck_grp_alloc_init("nfs_node_hash", LCK_GRP_ATTR_NULL);
109 nfs_node_hash_mutex = lck_mtx_alloc_init(nfs_node_hash_lck_grp, LCK_ATTR_NULL);
110 nfs_node_lck_grp = lck_grp_alloc_init("nfs_node", LCK_GRP_ATTR_NULL);
111 nfs_data_lck_grp = lck_grp_alloc_init("nfs_data", LCK_GRP_ATTR_NULL);
112 }
113
114 void
115 nfs_nhinit_finish(void)
116 {
117 lck_mtx_lock(nfs_node_hash_mutex);
118 if (!nfsnodehashtbl) {
119 nfsnodehashtbl = hashinit(desiredvnodes, M_NFSNODE, &nfsnodehash);
120 }
121 lck_mtx_unlock(nfs_node_hash_mutex);
122 }
123
124 /*
125 * Compute an entry in the NFS hash table structure
126 */
127 u_long
128 nfs_hash(u_char *fhp, int fhsize)
129 {
130 u_long fhsum;
131 int i;
132
133 fhsum = 0;
134 for (i = 0; i < fhsize; i++) {
135 fhsum += *fhp++;
136 }
137 return fhsum;
138 }
139
140
141 int nfs_case_insensitive(mount_t);
142
143 int
144 nfs_case_insensitive(mount_t mp)
145 {
146 struct nfsmount *nmp = VFSTONFS(mp);
147 int answer = 0;
148 int skip = 0;
149
150 if (nfs_mount_gone(nmp)) {
151 return 0;
152 }
153
154 if (nmp->nm_vers == NFS_VER2) {
155 /* V2 has no way to know */
156 return 0;
157 }
158
159 lck_mtx_lock(&nmp->nm_lock);
160 if (nmp->nm_vers == NFS_VER3) {
161 if (!(nmp->nm_state & NFSSTA_GOTPATHCONF)) {
162 /* We're holding the node lock so we just return
163 * with answer as case sensitive. Is very rare
164 * for file systems not to be homogenous w.r.t. pathconf
165 */
166 skip = 1;
167 }
168 } else if (!(nmp->nm_fsattr.nfsa_flags & NFS_FSFLAG_HOMOGENEOUS)) {
169 /* no pathconf info cached */
170 skip = 1;
171 }
172
173 if (!skip && (nmp->nm_fsattr.nfsa_flags & NFS_FSFLAG_CASE_INSENSITIVE)) {
174 answer = 1;
175 }
176
177 lck_mtx_unlock(&nmp->nm_lock);
178
179 return answer;
180 }
181
182
183 /*
184 * Look up a vnode/nfsnode by file handle.
185 * Callers must check for mount points!!
186 * In all cases, a pointer to a
187 * nfsnode structure is returned.
188 */
189 int
190 nfs_nget(
191 mount_t mp,
192 nfsnode_t dnp,
193 struct componentname *cnp,
194 u_char *fhp,
195 int fhsize,
196 struct nfs_vattr *nvap,
197 u_int64_t *xidp,
198 uint32_t auth,
199 int flags,
200 nfsnode_t *npp)
201 {
202 nfsnode_t np;
203 struct nfsnodehashhead *nhpp;
204 vnode_t vp;
205 int error, nfsvers;
206 mount_t mp2;
207 struct vnode_fsparam vfsp;
208 uint32_t vid;
209
210 FSDBG_TOP(263, mp, dnp, flags, npp);
211
212 /* Check for unmount in progress */
213 if (!mp || vfs_isforce(mp)) {
214 *npp = NULL;
215 error = ENXIO;
216 FSDBG_BOT(263, mp, dnp, 0xd1e, error);
217 return error;
218 }
219 nfsvers = VFSTONFS(mp)->nm_vers;
220
221 nhpp = NFSNOHASH(nfs_hash(fhp, fhsize));
222 loop:
223 lck_mtx_lock(nfs_node_hash_mutex);
224 for (np = nhpp->lh_first; np != 0; np = np->n_hash.le_next) {
225 mp2 = (np->n_hflag & NHINIT) ? np->n_mount : NFSTOMP(np);
226 if (mp != mp2 || np->n_fhsize != fhsize ||
227 bcmp(fhp, np->n_fhp, fhsize)) {
228 continue;
229 }
230 if (nvap && (nvap->nva_flags & NFS_FFLAG_TRIGGER_REFERRAL) &&
231 cnp && (cnp->cn_namelen > (fhsize - (int)sizeof(dnp)))) {
232 /* The name was too long to fit in the file handle. Check it against the node's name. */
233 int namecmp = 0;
234 const char *vname = vnode_getname(NFSTOV(np));
235 if (vname) {
236 if (cnp->cn_namelen != (int)strlen(vname)) {
237 namecmp = 1;
238 } else {
239 namecmp = strncmp(vname, cnp->cn_nameptr, cnp->cn_namelen);
240 }
241 vnode_putname(vname);
242 }
243 if (namecmp) { /* full name didn't match */
244 continue;
245 }
246 }
247 FSDBG(263, dnp, np, np->n_flag, 0xcace0000);
248 /* if the node is locked, sleep on it */
249 if ((np->n_hflag & NHLOCKED) && !(flags & NG_NOCREATE)) {
250 np->n_hflag |= NHLOCKWANT;
251 FSDBG(263, dnp, np, np->n_flag, 0xcace2222);
252 msleep(np, nfs_node_hash_mutex, PDROP | PINOD, "nfs_nget", NULL);
253 FSDBG(263, dnp, np, np->n_flag, 0xcace3333);
254 goto loop;
255 }
256 vp = NFSTOV(np);
257 vid = vnode_vid(vp);
258 lck_mtx_unlock(nfs_node_hash_mutex);
259 if ((error = vnode_getwithvid(vp, vid))) {
260 /*
261 * If vnode is being reclaimed or has already
262 * changed identity, no need to wait.
263 */
264 FSDBG_BOT(263, dnp, *npp, 0xcace0d1e, error);
265 return error;
266 }
267 if ((error = nfs_node_lock(np))) {
268 /* this only fails if the node is now unhashed */
269 /* so let's see if we can find/create it again */
270 FSDBG(263, dnp, *npp, 0xcaced1e2, error);
271 vnode_put(vp);
272 if (flags & NG_NOCREATE) {
273 *npp = 0;
274 FSDBG_BOT(263, dnp, *npp, 0xcaced1e0, ENOENT);
275 return ENOENT;
276 }
277 goto loop;
278 }
279 /* update attributes */
280 if (nvap) {
281 error = nfs_loadattrcache(np, nvap, xidp, 0);
282 }
283 if (error) {
284 nfs_node_unlock(np);
285 vnode_put(vp);
286 } else {
287 if (dnp && cnp && (flags & NG_MAKEENTRY)) {
288 cache_enter(NFSTOV(dnp), vp, cnp);
289 }
290 /*
291 * Update the vnode if the name/and or the parent has
292 * changed. We need to do this so that if getattrlist is
293 * called asking for ATTR_CMN_NAME, that the "most"
294 * correct name is being returned. In addition for
295 * monitored vnodes we need to kick the vnode out of the
296 * name cache. We do this so that if there are hard
297 * links in the same directory the link will not be
298 * found and a lookup will get us here to return the
299 * name of the current link. In addition by removing the
300 * name from the name cache the old name will not be
301 * found after a rename done on another client or the
302 * server. The principle reason to do this is because
303 * Finder is asking for notifications on a directory.
304 * The directory changes, Finder gets notified, reads
305 * the directory (which we have purged) and for each
306 * entry returned calls getattrlist with the name
307 * returned from readdir. gettattrlist has to call
308 * namei/lookup to resolve the name, because its not in
309 * the cache we end up here. We need to update the name
310 * so Finder will get the name it called us with.
311 *
312 * We had an imperfect solution with respect to case
313 * sensitivity. There is a test that is run in
314 * FileBuster that does renames from some name to
315 * another name differing only in case. It then reads
316 * the directory looking for the new name, after it
317 * finds that new name, it ask gettattrlist to verify
318 * that the name is the new name. Usually that works,
319 * but renames generate fsevents and fseventsd will do a
320 * lookup on the name via lstat. Since that test renames
321 * old name to new name back and forth there is a race
322 * that an fsevent will be behind and will access the
323 * file by the old name, on a case insensitive file
324 * system that will work. Problem is if we do a case
325 * sensitive compare, we're going to change the name,
326 * which the test's getattrlist verification step is
327 * going to fail. So we will check the case sensitivity
328 * of the file system and do the appropriate compare. In
329 * a rare instance for non homogeneous file systems
330 * w.r.t. pathconf we will use case sensitive compares.
331 * That could break if the file system is actually case
332 * insensitive.
333 *
334 * Note that V2 does not know the case, so we just
335 * assume case sensitivity.
336 *
337 * This is clearly not perfect due to races, but this is
338 * as good as its going to get. You can defeat the
339 * handling of hard links simply by doing:
340 *
341 * while :; do ls -l > /dev/null; done
342 *
343 * in a terminal window. Even a single ls -l can cause a
344 * race.
345 *
346 * <rant>What we really need is for the caller, that
347 * knows the name being used is valid since it got it
348 * from a readdir to use that name and not ask for the
349 * ATTR_CMN_NAME</rant>
350 */
351 if (dnp && cnp && (vp != NFSTOV(dnp))) {
352 int update_flags = (vnode_ismonitored((NFSTOV(dnp)))) ? VNODE_UPDATE_CACHE : 0;
353 int (*cmp)(const char *s1, const char *s2, size_t n);
354
355 cmp = nfs_case_insensitive(mp) ? strncasecmp : strncmp;
356
357 if (vp->v_name && (size_t)cnp->cn_namelen != strnlen(vp->v_name, MAXPATHLEN)) {
358 update_flags |= VNODE_UPDATE_NAME;
359 }
360 if (vp->v_name && cnp->cn_namelen && (*cmp)(cnp->cn_nameptr, vp->v_name, cnp->cn_namelen)) {
361 update_flags |= VNODE_UPDATE_NAME;
362 }
363 if ((vp->v_name == NULL && cnp->cn_namelen != 0) || (vp->v_name != NULL && cnp->cn_namelen == 0)) {
364 update_flags |= VNODE_UPDATE_NAME;
365 }
366 if (vnode_parent(vp) != NFSTOV(dnp)) {
367 update_flags |= VNODE_UPDATE_PARENT;
368 }
369 if (update_flags) {
370 NFS_NODE_DBG("vnode_update_identity old name %s new name %.*s update flags = %x\n",
371 vp->v_name, cnp->cn_namelen, cnp->cn_nameptr ? cnp->cn_nameptr : "", update_flags);
372 vnode_update_identity(vp, NFSTOV(dnp), cnp->cn_nameptr, cnp->cn_namelen, 0, update_flags);
373 }
374 }
375
376 *npp = np;
377 }
378 FSDBG_BOT(263, dnp, *npp, 0xcace0000, error);
379 return error;
380 }
381
382 FSDBG(263, mp, dnp, npp, 0xaaaaaaaa);
383
384 if (flags & NG_NOCREATE) {
385 lck_mtx_unlock(nfs_node_hash_mutex);
386 *npp = 0;
387 FSDBG_BOT(263, dnp, *npp, 0x80000001, ENOENT);
388 return ENOENT;
389 }
390
391 /*
392 * allocate and initialize nfsnode and stick it in the hash
393 * before calling getnewvnode(). Anyone finding it in the
394 * hash before initialization is complete will wait for it.
395 */
396 MALLOC_ZONE(np, nfsnode_t, sizeof *np, M_NFSNODE, M_WAITOK);
397 if (!np) {
398 lck_mtx_unlock(nfs_node_hash_mutex);
399 *npp = 0;
400 FSDBG_BOT(263, dnp, *npp, 0x80000001, ENOMEM);
401 return ENOMEM;
402 }
403 bzero(np, sizeof *np);
404 np->n_hflag |= (NHINIT | NHLOCKED);
405 np->n_mount = mp;
406 np->n_auth = auth;
407 TAILQ_INIT(&np->n_opens);
408 TAILQ_INIT(&np->n_lock_owners);
409 TAILQ_INIT(&np->n_locks);
410 np->n_dlink.tqe_next = NFSNOLIST;
411 np->n_dreturn.tqe_next = NFSNOLIST;
412 np->n_monlink.le_next = NFSNOLIST;
413
414 /* ugh... need to keep track of ".zfs" directories to workaround server bugs */
415 if ((nvap->nva_type == VDIR) && cnp && (cnp->cn_namelen == 4) &&
416 (cnp->cn_nameptr[0] == '.') && (cnp->cn_nameptr[1] == 'z') &&
417 (cnp->cn_nameptr[2] == 'f') && (cnp->cn_nameptr[3] == 's')) {
418 np->n_flag |= NISDOTZFS;
419 }
420 if (dnp && (dnp->n_flag & NISDOTZFS)) {
421 np->n_flag |= NISDOTZFSCHILD;
422 }
423
424 if (dnp && cnp && ((cnp->cn_namelen != 2) ||
425 (cnp->cn_nameptr[0] != '.') || (cnp->cn_nameptr[1] != '.'))) {
426 vnode_t dvp = NFSTOV(dnp);
427 if (!vnode_get(dvp)) {
428 if (!vnode_ref(dvp)) {
429 np->n_parent = dvp;
430 }
431 vnode_put(dvp);
432 }
433 }
434
435 /* setup node's file handle */
436 if (fhsize > NFS_SMALLFH) {
437 MALLOC_ZONE(np->n_fhp, u_char *,
438 fhsize, M_NFSBIGFH, M_WAITOK);
439 if (!np->n_fhp) {
440 lck_mtx_unlock(nfs_node_hash_mutex);
441 FREE_ZONE(np, sizeof *np, M_NFSNODE);
442 *npp = 0;
443 FSDBG_BOT(263, dnp, *npp, 0x80000002, ENOMEM);
444 return ENOMEM;
445 }
446 } else {
447 np->n_fhp = &np->n_fh[0];
448 }
449 bcopy(fhp, np->n_fhp, fhsize);
450 np->n_fhsize = fhsize;
451
452 /* Insert the nfsnode in the hash queue for its new file handle */
453 LIST_INSERT_HEAD(nhpp, np, n_hash);
454 np->n_hflag |= NHHASHED;
455 FSDBG(266, 0, np, np->n_flag, np->n_hflag);
456
457 /* lock the new nfsnode */
458 lck_mtx_init(&np->n_lock, nfs_node_lck_grp, LCK_ATTR_NULL);
459 lck_rw_init(&np->n_datalock, nfs_data_lck_grp, LCK_ATTR_NULL);
460 lck_mtx_init(&np->n_openlock, nfs_open_grp, LCK_ATTR_NULL);
461 lck_mtx_lock(&np->n_lock);
462
463 /* release lock on hash table */
464 lck_mtx_unlock(nfs_node_hash_mutex);
465
466 /* do initial loading of attributes */
467 NACLINVALIDATE(np);
468 NACCESSINVALIDATE(np);
469 error = nfs_loadattrcache(np, nvap, xidp, 1);
470 if (error) {
471 FSDBG(266, 0, np, np->n_flag, 0xb1eb1e);
472 nfs_node_unlock(np);
473 lck_mtx_lock(nfs_node_hash_mutex);
474 LIST_REMOVE(np, n_hash);
475 np->n_hflag &= ~(NHHASHED | NHINIT | NHLOCKED);
476 if (np->n_hflag & NHLOCKWANT) {
477 np->n_hflag &= ~NHLOCKWANT;
478 wakeup(np);
479 }
480 lck_mtx_unlock(nfs_node_hash_mutex);
481 if (np->n_parent) {
482 if (!vnode_get(np->n_parent)) {
483 vnode_rele(np->n_parent);
484 vnode_put(np->n_parent);
485 }
486 np->n_parent = NULL;
487 }
488 lck_mtx_destroy(&np->n_lock, nfs_node_lck_grp);
489 lck_rw_destroy(&np->n_datalock, nfs_data_lck_grp);
490 lck_mtx_destroy(&np->n_openlock, nfs_open_grp);
491 if (np->n_fhsize > NFS_SMALLFH) {
492 FREE_ZONE(np->n_fhp, np->n_fhsize, M_NFSBIGFH);
493 }
494 FREE_ZONE(np, sizeof *np, M_NFSNODE);
495 *npp = 0;
496 FSDBG_BOT(263, dnp, *npp, 0x80000003, error);
497 return error;
498 }
499 NFS_CHANGED_UPDATE(nfsvers, np, nvap);
500 if (nvap->nva_type == VDIR) {
501 NFS_CHANGED_UPDATE_NC(nfsvers, np, nvap);
502 }
503
504 /* now, attempt to get a new vnode */
505 vfsp.vnfs_mp = mp;
506 vfsp.vnfs_vtype = nvap->nva_type;
507 vfsp.vnfs_str = "nfs";
508 vfsp.vnfs_dvp = dnp ? NFSTOV(dnp) : NULL;
509 vfsp.vnfs_fsnode = np;
510 #if CONFIG_NFS4
511 if (nfsvers == NFS_VER4) {
512 #if FIFO
513 if (nvap->nva_type == VFIFO) {
514 vfsp.vnfs_vops = fifo_nfsv4nodeop_p;
515 } else
516 #endif /* FIFO */
517 if (nvap->nva_type == VBLK || nvap->nva_type == VCHR) {
518 vfsp.vnfs_vops = spec_nfsv4nodeop_p;
519 } else {
520 vfsp.vnfs_vops = nfsv4_vnodeop_p;
521 }
522 } else
523 #endif /* CONFIG_NFS4 */
524 {
525 #if FIFO
526 if (nvap->nva_type == VFIFO) {
527 vfsp.vnfs_vops = fifo_nfsv2nodeop_p;
528 } else
529 #endif /* FIFO */
530 if (nvap->nva_type == VBLK || nvap->nva_type == VCHR) {
531 vfsp.vnfs_vops = spec_nfsv2nodeop_p;
532 } else {
533 vfsp.vnfs_vops = nfsv2_vnodeop_p;
534 }
535 }
536 vfsp.vnfs_markroot = (flags & NG_MARKROOT) ? 1 : 0;
537 vfsp.vnfs_marksystem = 0;
538 vfsp.vnfs_rdev = 0;
539 vfsp.vnfs_filesize = nvap->nva_size;
540 vfsp.vnfs_cnp = cnp;
541 vfsp.vnfs_flags = VNFS_ADDFSREF;
542 if (!dnp || !cnp || !(flags & NG_MAKEENTRY)) {
543 vfsp.vnfs_flags |= VNFS_NOCACHE;
544 }
545
546 #if CONFIG_TRIGGERS
547 if (((nfsvers >= NFS_VER4)
548 )
549 && (nvap->nva_type == VDIR) && (np->n_vattr.nva_flags & NFS_FFLAG_TRIGGER)
550 && !(flags & NG_MARKROOT)) {
551 struct vnode_trigger_param vtp;
552 bzero(&vtp, sizeof(vtp));
553 bcopy(&vfsp, &vtp.vnt_params, sizeof(vfsp));
554 vtp.vnt_resolve_func = nfs_mirror_mount_trigger_resolve;
555 vtp.vnt_unresolve_func = nfs_mirror_mount_trigger_unresolve;
556 vtp.vnt_rearm_func = nfs_mirror_mount_trigger_rearm;
557 vtp.vnt_flags = VNT_AUTO_REARM | VNT_KERN_RESOLVE;
558 error = vnode_create(VNCREATE_TRIGGER, VNCREATE_TRIGGER_SIZE, &vtp, &np->n_vnode);
559 } else
560 #endif
561 {
562 error = vnode_create(VNCREATE_FLAVOR, VCREATESIZE, &vfsp, &np->n_vnode);
563 }
564 notsup:
565 if (error) {
566 FSDBG(266, 0, np, np->n_flag, 0xb1eb1e);
567 nfs_node_unlock(np);
568 lck_mtx_lock(nfs_node_hash_mutex);
569 LIST_REMOVE(np, n_hash);
570 np->n_hflag &= ~(NHHASHED | NHINIT | NHLOCKED);
571 if (np->n_hflag & NHLOCKWANT) {
572 np->n_hflag &= ~NHLOCKWANT;
573 wakeup(np);
574 }
575 lck_mtx_unlock(nfs_node_hash_mutex);
576 if (np->n_parent) {
577 if (!vnode_get(np->n_parent)) {
578 vnode_rele(np->n_parent);
579 vnode_put(np->n_parent);
580 }
581 np->n_parent = NULL;
582 }
583 lck_mtx_destroy(&np->n_lock, nfs_node_lck_grp);
584 lck_rw_destroy(&np->n_datalock, nfs_data_lck_grp);
585 lck_mtx_destroy(&np->n_openlock, nfs_open_grp);
586 if (np->n_fhsize > NFS_SMALLFH) {
587 FREE_ZONE(np->n_fhp, np->n_fhsize, M_NFSBIGFH);
588 }
589 FREE_ZONE(np, sizeof *np, M_NFSNODE);
590 *npp = 0;
591 FSDBG_BOT(263, dnp, *npp, 0x80000004, error);
592 return error;
593 }
594 vp = np->n_vnode;
595 vnode_settag(vp, VT_NFS);
596 /* node is now initialized */
597
598 /* check if anyone's waiting on this node */
599 lck_mtx_lock(nfs_node_hash_mutex);
600 np->n_hflag &= ~(NHINIT | NHLOCKED);
601 if (np->n_hflag & NHLOCKWANT) {
602 np->n_hflag &= ~NHLOCKWANT;
603 wakeup(np);
604 }
605 lck_mtx_unlock(nfs_node_hash_mutex);
606
607 *npp = np;
608
609 FSDBG_BOT(263, dnp, vp, *npp, error);
610 return error;
611 }
612
613
614 int
615 nfs_vnop_inactive(
616 struct vnop_inactive_args /* {
617 * struct vnodeop_desc *a_desc;
618 * vnode_t a_vp;
619 * vfs_context_t a_context;
620 * } */*ap)
621 {
622 vnode_t vp = ap->a_vp;
623 vfs_context_t ctx = ap->a_context;
624 nfsnode_t np;
625 struct nfs_sillyrename *nsp;
626 struct nfs_vattr nvattr;
627 int unhash, attrerr, busyerror, error, inuse, busied, force;
628 struct nfs_open_file *nofp;
629 struct componentname cn;
630 struct nfsmount *nmp;
631 mount_t mp;
632
633 if (vp == NULL) {
634 panic("nfs_vnop_inactive: vp == NULL");
635 }
636 np = VTONFS(vp);
637 if (np == NULL) {
638 panic("nfs_vnop_inactive: np == NULL");
639 }
640
641 nmp = NFSTONMP(np);
642 mp = vnode_mount(vp);
643
644 restart:
645 force = (!mp || vfs_isforce(mp));
646 error = 0;
647 inuse = (nfs_mount_state_in_use_start(nmp, NULL) == 0);
648
649 /* There shouldn't be any open or lock state at this point */
650 lck_mtx_lock(&np->n_openlock);
651 if (np->n_openrefcnt && !force) {
652 /*
653 * vnode_rele and vnode_put drop the vnode lock before
654 * calling VNOP_INACTIVE, so there is a race were the
655 * vnode could become active again. Perhaps there are
656 * other places where this can happen, so if we've got
657 * here we need to get out.
658 */
659 #ifdef NFS_NODE_DEBUG
660 NP(np, "nfs_vnop_inactive: still open: %d", np->n_openrefcnt);
661 #endif
662 lck_mtx_unlock(&np->n_openlock);
663 return 0;
664 }
665
666 TAILQ_FOREACH(nofp, &np->n_opens, nof_link) {
667 lck_mtx_lock(&nofp->nof_lock);
668 if (nofp->nof_flags & NFS_OPEN_FILE_BUSY) {
669 if (!force) {
670 NP(np, "nfs_vnop_inactive: open file busy");
671 }
672 busied = 0;
673 } else {
674 nofp->nof_flags |= NFS_OPEN_FILE_BUSY;
675 busied = 1;
676 }
677 lck_mtx_unlock(&nofp->nof_lock);
678 if ((np->n_flag & NREVOKE) || (nofp->nof_flags & NFS_OPEN_FILE_LOST)) {
679 if (busied) {
680 nfs_open_file_clear_busy(nofp);
681 }
682 continue;
683 }
684 /*
685 * If we just created the file, we already had it open in
686 * anticipation of getting a subsequent open call. If the
687 * node has gone inactive without being open, we need to
688 * clean up (close) the open done in the create.
689 */
690 #if CONFIG_NFS4
691 if ((nofp->nof_flags & NFS_OPEN_FILE_CREATE) && nofp->nof_creator && !force) {
692 if (nofp->nof_flags & NFS_OPEN_FILE_REOPEN) {
693 lck_mtx_unlock(&np->n_openlock);
694 if (busied) {
695 nfs_open_file_clear_busy(nofp);
696 }
697 if (inuse) {
698 nfs_mount_state_in_use_end(nmp, 0);
699 }
700 if (!nfs4_reopen(nofp, NULL)) {
701 goto restart;
702 }
703 }
704 nofp->nof_flags &= ~NFS_OPEN_FILE_CREATE;
705 lck_mtx_unlock(&np->n_openlock);
706 error = nfs_close(np, nofp, NFS_OPEN_SHARE_ACCESS_BOTH, NFS_OPEN_SHARE_DENY_NONE, ctx);
707 if (error) {
708 NP(np, "nfs_vnop_inactive: create close error: %d", error);
709 nofp->nof_flags |= NFS_OPEN_FILE_CREATE;
710 }
711 if (busied) {
712 nfs_open_file_clear_busy(nofp);
713 }
714 if (inuse) {
715 nfs_mount_state_in_use_end(nmp, error);
716 }
717 goto restart;
718 }
719 #endif
720 if (nofp->nof_flags & NFS_OPEN_FILE_NEEDCLOSE) {
721 /*
722 * If the file is marked as needing reopen, but this was the only
723 * open on the file, just drop the open.
724 */
725 nofp->nof_flags &= ~NFS_OPEN_FILE_NEEDCLOSE;
726 if ((nofp->nof_flags & NFS_OPEN_FILE_REOPEN) && (nofp->nof_opencnt == 1)) {
727 nofp->nof_flags &= ~NFS_OPEN_FILE_REOPEN;
728 nofp->nof_r--;
729 nofp->nof_opencnt--;
730 nofp->nof_access = 0;
731 } else if (!force) {
732 lck_mtx_unlock(&np->n_openlock);
733 if (nofp->nof_flags & NFS_OPEN_FILE_REOPEN) {
734 if (busied) {
735 nfs_open_file_clear_busy(nofp);
736 }
737 if (inuse) {
738 nfs_mount_state_in_use_end(nmp, 0);
739 }
740 #if CONFIG_NFS4
741 if (!nfs4_reopen(nofp, NULL)) {
742 goto restart;
743 }
744 #endif
745 }
746 error = nfs_close(np, nofp, NFS_OPEN_SHARE_ACCESS_READ, NFS_OPEN_SHARE_DENY_NONE, ctx);
747 if (error) {
748 NP(np, "nfs_vnop_inactive: need close error: %d", error);
749 nofp->nof_flags |= NFS_OPEN_FILE_NEEDCLOSE;
750 }
751 if (busied) {
752 nfs_open_file_clear_busy(nofp);
753 }
754 if (inuse) {
755 nfs_mount_state_in_use_end(nmp, error);
756 }
757 goto restart;
758 }
759 }
760 if (nofp->nof_opencnt && !force) {
761 NP(np, "nfs_vnop_inactive: file still open: %d", nofp->nof_opencnt);
762 }
763 if (!force && (nofp->nof_access || nofp->nof_deny ||
764 nofp->nof_mmap_access || nofp->nof_mmap_deny ||
765 nofp->nof_r || nofp->nof_w || nofp->nof_rw ||
766 nofp->nof_r_dw || nofp->nof_w_dw || nofp->nof_rw_dw ||
767 nofp->nof_r_drw || nofp->nof_w_drw || nofp->nof_rw_drw ||
768 nofp->nof_d_r || nofp->nof_d_w || nofp->nof_d_rw ||
769 nofp->nof_d_r_dw || nofp->nof_d_w_dw || nofp->nof_d_rw_dw ||
770 nofp->nof_d_r_drw || nofp->nof_d_w_drw || nofp->nof_d_rw_drw)) {
771 NP(np, "nfs_vnop_inactive: non-zero access: %d %d %d %d # %u.%u %u.%u %u.%u dw %u.%u %u.%u %u.%u drw %u.%u %u.%u %u.%u",
772 nofp->nof_access, nofp->nof_deny,
773 nofp->nof_mmap_access, nofp->nof_mmap_deny,
774 nofp->nof_r, nofp->nof_d_r,
775 nofp->nof_w, nofp->nof_d_w,
776 nofp->nof_rw, nofp->nof_d_rw,
777 nofp->nof_r_dw, nofp->nof_d_r_dw,
778 nofp->nof_w_dw, nofp->nof_d_w_dw,
779 nofp->nof_rw_dw, nofp->nof_d_rw_dw,
780 nofp->nof_r_drw, nofp->nof_d_r_drw,
781 nofp->nof_w_drw, nofp->nof_d_w_drw,
782 nofp->nof_rw_drw, nofp->nof_d_rw_drw);
783 }
784 if (busied) {
785 nfs_open_file_clear_busy(nofp);
786 }
787 }
788 lck_mtx_unlock(&np->n_openlock);
789
790 if (inuse && nfs_mount_state_in_use_end(nmp, error)) {
791 goto restart;
792 }
793
794 nfs_node_lock_force(np);
795
796 if (vnode_vtype(vp) != VDIR) {
797 nsp = np->n_sillyrename;
798 np->n_sillyrename = NULL;
799 } else {
800 nsp = NULL;
801 }
802
803 FSDBG_TOP(264, vp, np, np->n_flag, nsp);
804
805 if (!nsp) {
806 /* no silly file to clean up... */
807 /* clear all flags other than these */
808 np->n_flag &= (NMODIFIED);
809 nfs_node_unlock(np);
810 FSDBG_BOT(264, vp, np, np->n_flag, 0);
811 return 0;
812 }
813 nfs_node_unlock(np);
814
815 /* Remove the silly file that was rename'd earlier */
816
817 /* flush all the buffers */
818 nfs_vinvalbuf2(vp, V_SAVE, vfs_context_thread(ctx), nsp->nsr_cred, 1);
819
820 /* try to get the latest attributes */
821 attrerr = nfs_getattr(np, &nvattr, ctx, NGA_UNCACHED);
822
823 /* Check if we should remove it from the node hash. */
824 /* Leave it if inuse or it has multiple hard links. */
825 if (vnode_isinuse(vp, 0) || (!attrerr && (nvattr.nva_nlink > 1))) {
826 unhash = 0;
827 } else {
828 unhash = 1;
829 ubc_setsize(vp, 0);
830 }
831
832 /* mark this node and the directory busy while we do the remove */
833 busyerror = nfs_node_set_busy2(nsp->nsr_dnp, np, vfs_context_thread(ctx));
834
835 /* lock the node while we remove the silly file */
836 lck_mtx_lock(nfs_node_hash_mutex);
837 while (np->n_hflag & NHLOCKED) {
838 np->n_hflag |= NHLOCKWANT;
839 msleep(np, nfs_node_hash_mutex, PINOD, "nfs_inactive", NULL);
840 }
841 np->n_hflag |= NHLOCKED;
842 lck_mtx_unlock(nfs_node_hash_mutex);
843
844 /* purge the name cache to deter others from finding it */
845 bzero(&cn, sizeof(cn));
846 cn.cn_nameptr = nsp->nsr_name;
847 cn.cn_namelen = nsp->nsr_namlen;
848 nfs_name_cache_purge(nsp->nsr_dnp, np, &cn, ctx);
849
850 FSDBG(264, np, np->n_size, np->n_vattr.nva_size, 0xf00d00f1);
851
852 /* now remove the silly file */
853 nfs_removeit(nsp);
854
855 /* clear all flags other than these */
856 nfs_node_lock_force(np);
857 np->n_flag &= (NMODIFIED);
858 nfs_node_unlock(np);
859
860 if (!busyerror) {
861 nfs_node_clear_busy2(nsp->nsr_dnp, np);
862 }
863
864 if (unhash && vnode_isinuse(vp, 0)) {
865 /* vnode now inuse after silly remove? */
866 unhash = 0;
867 ubc_setsize(vp, np->n_size);
868 }
869
870 lck_mtx_lock(nfs_node_hash_mutex);
871 if (unhash) {
872 /*
873 * remove nfsnode from hash now so we can't accidentally find it
874 * again if another object gets created with the same filehandle
875 * before this vnode gets reclaimed
876 */
877 if (np->n_hflag & NHHASHED) {
878 LIST_REMOVE(np, n_hash);
879 np->n_hflag &= ~NHHASHED;
880 FSDBG(266, 0, np, np->n_flag, 0xb1eb1e);
881 }
882 vnode_recycle(vp);
883 }
884 /* unlock the node */
885 np->n_hflag &= ~NHLOCKED;
886 if (np->n_hflag & NHLOCKWANT) {
887 np->n_hflag &= ~NHLOCKWANT;
888 wakeup(np);
889 }
890 lck_mtx_unlock(nfs_node_hash_mutex);
891
892 /* cleanup sillyrename info */
893 if (nsp->nsr_cred != NOCRED) {
894 kauth_cred_unref(&nsp->nsr_cred);
895 }
896 vnode_rele(NFSTOV(nsp->nsr_dnp));
897 FREE_ZONE(nsp, sizeof(*nsp), M_NFSREQ);
898
899 FSDBG_BOT(264, vp, np, np->n_flag, 0);
900 return 0;
901 }
902
903 /*
904 * Reclaim an nfsnode so that it can be used for other purposes.
905 */
906 int
907 nfs_vnop_reclaim(
908 struct vnop_reclaim_args /* {
909 * struct vnodeop_desc *a_desc;
910 * vnode_t a_vp;
911 * vfs_context_t a_context;
912 * } */*ap)
913 {
914 vnode_t vp = ap->a_vp;
915 nfsnode_t np = VTONFS(vp);
916 vfs_context_t ctx = ap->a_context;
917 struct nfs_open_file *nofp, *nextnofp;
918 struct nfs_file_lock *nflp, *nextnflp;
919 struct nfs_lock_owner *nlop, *nextnlop;
920 struct nfsmount *nmp = np->n_mount ? VFSTONFS(np->n_mount) : NFSTONMP(np);
921 mount_t mp = vnode_mount(vp);
922 int force;
923
924 FSDBG_TOP(265, vp, np, np->n_flag, 0);
925 force = (!mp || vfs_isforce(mp) || nfs_mount_gone(nmp));
926
927
928 /* There shouldn't be any open or lock state at this point */
929 lck_mtx_lock(&np->n_openlock);
930
931 #if CONFIG_NFS4
932 if (nmp && (nmp->nm_vers >= NFS_VER4)) {
933 /* need to drop a delegation */
934 if (np->n_dreturn.tqe_next != NFSNOLIST) {
935 /* remove this node from the delegation return list */
936 lck_mtx_lock(&nmp->nm_lock);
937 if (np->n_dreturn.tqe_next != NFSNOLIST) {
938 TAILQ_REMOVE(&nmp->nm_dreturnq, np, n_dreturn);
939 np->n_dreturn.tqe_next = NFSNOLIST;
940 }
941 lck_mtx_unlock(&nmp->nm_lock);
942 }
943 if (np->n_dlink.tqe_next != NFSNOLIST) {
944 /* remove this node from the delegation list */
945 lck_mtx_lock(&nmp->nm_lock);
946 if (np->n_dlink.tqe_next != NFSNOLIST) {
947 TAILQ_REMOVE(&nmp->nm_delegations, np, n_dlink);
948 np->n_dlink.tqe_next = NFSNOLIST;
949 }
950 lck_mtx_unlock(&nmp->nm_lock);
951 }
952 if ((np->n_openflags & N_DELEG_MASK) && !force) {
953 /* try to return the delegation */
954 np->n_openflags &= ~N_DELEG_MASK;
955 nfs4_delegreturn_rpc(nmp, np->n_fhp, np->n_fhsize, &np->n_dstateid,
956 R_RECOVER, vfs_context_thread(ctx), vfs_context_ucred(ctx));
957 }
958 if (np->n_attrdirfh) {
959 FREE(np->n_attrdirfh, M_TEMP);
960 np->n_attrdirfh = NULL;
961 }
962 }
963 #endif
964
965 /* clean up file locks */
966 TAILQ_FOREACH_SAFE(nflp, &np->n_locks, nfl_link, nextnflp) {
967 if (!(nflp->nfl_flags & NFS_FILE_LOCK_DEAD) && !force) {
968 NP(np, "nfs_vnop_reclaim: lock 0x%llx 0x%llx 0x%x (bc %d)",
969 nflp->nfl_start, nflp->nfl_end, nflp->nfl_flags, nflp->nfl_blockcnt);
970 }
971 if (!(nflp->nfl_flags & (NFS_FILE_LOCK_BLOCKED | NFS_FILE_LOCK_DEAD))) {
972 /* try sending an unlock RPC if it wasn't delegated */
973 if (!(nflp->nfl_flags & NFS_FILE_LOCK_DELEGATED) && !force) {
974 nmp->nm_funcs->nf_unlock_rpc(np, nflp->nfl_owner, F_WRLCK, nflp->nfl_start, nflp->nfl_end, R_RECOVER,
975 NULL, nflp->nfl_owner->nlo_open_owner->noo_cred);
976 }
977 lck_mtx_lock(&nflp->nfl_owner->nlo_lock);
978 TAILQ_REMOVE(&nflp->nfl_owner->nlo_locks, nflp, nfl_lolink);
979 lck_mtx_unlock(&nflp->nfl_owner->nlo_lock);
980 }
981 TAILQ_REMOVE(&np->n_locks, nflp, nfl_link);
982 nfs_file_lock_destroy(nflp);
983 }
984 /* clean up lock owners */
985 TAILQ_FOREACH_SAFE(nlop, &np->n_lock_owners, nlo_link, nextnlop) {
986 if (!TAILQ_EMPTY(&nlop->nlo_locks) && !force) {
987 NP(np, "nfs_vnop_reclaim: lock owner with locks");
988 }
989 TAILQ_REMOVE(&np->n_lock_owners, nlop, nlo_link);
990 nfs_lock_owner_destroy(nlop);
991 }
992 /* clean up open state */
993 if (np->n_openrefcnt && !force) {
994 NP(np, "nfs_vnop_reclaim: still open: %d", np->n_openrefcnt);
995 }
996 TAILQ_FOREACH_SAFE(nofp, &np->n_opens, nof_link, nextnofp) {
997 if (nofp->nof_flags & NFS_OPEN_FILE_BUSY) {
998 NP(np, "nfs_vnop_reclaim: open file busy");
999 }
1000 if (!(np->n_flag & NREVOKE) && !(nofp->nof_flags & NFS_OPEN_FILE_LOST)) {
1001 if (nofp->nof_opencnt && !force) {
1002 NP(np, "nfs_vnop_reclaim: file still open: %d", nofp->nof_opencnt);
1003 }
1004 if (!force && (nofp->nof_access || nofp->nof_deny ||
1005 nofp->nof_mmap_access || nofp->nof_mmap_deny ||
1006 nofp->nof_r || nofp->nof_w || nofp->nof_rw ||
1007 nofp->nof_r_dw || nofp->nof_w_dw || nofp->nof_rw_dw ||
1008 nofp->nof_r_drw || nofp->nof_w_drw || nofp->nof_rw_drw ||
1009 nofp->nof_d_r || nofp->nof_d_w || nofp->nof_d_rw ||
1010 nofp->nof_d_r_dw || nofp->nof_d_w_dw || nofp->nof_d_rw_dw ||
1011 nofp->nof_d_r_drw || nofp->nof_d_w_drw || nofp->nof_d_rw_drw)) {
1012 NP(np, "nfs_vnop_reclaim: non-zero access: %d %d %d %d # %u.%u %u.%u %u.%u dw %u.%u %u.%u %u.%u drw %u.%u %u.%u %u.%u",
1013 nofp->nof_access, nofp->nof_deny,
1014 nofp->nof_mmap_access, nofp->nof_mmap_deny,
1015 nofp->nof_r, nofp->nof_d_r,
1016 nofp->nof_w, nofp->nof_d_w,
1017 nofp->nof_rw, nofp->nof_d_rw,
1018 nofp->nof_r_dw, nofp->nof_d_r_dw,
1019 nofp->nof_w_dw, nofp->nof_d_w_dw,
1020 nofp->nof_rw_dw, nofp->nof_d_rw_dw,
1021 nofp->nof_r_drw, nofp->nof_d_r_drw,
1022 nofp->nof_w_drw, nofp->nof_d_w_drw,
1023 nofp->nof_rw_drw, nofp->nof_d_rw_drw);
1024 #if CONFIG_NFS4
1025 /* try sending a close RPC if it wasn't delegated */
1026 if (nofp->nof_r || nofp->nof_w || nofp->nof_rw ||
1027 nofp->nof_r_dw || nofp->nof_w_dw || nofp->nof_rw_dw ||
1028 nofp->nof_r_drw || nofp->nof_w_drw || nofp->nof_rw_drw) {
1029 nfs4_close_rpc(np, nofp, NULL, nofp->nof_owner->noo_cred, R_RECOVER);
1030 }
1031 #endif
1032 }
1033 }
1034 TAILQ_REMOVE(&np->n_opens, nofp, nof_link);
1035 nfs_open_file_destroy(nofp);
1036 }
1037 lck_mtx_unlock(&np->n_openlock);
1038
1039 if (np->n_monlink.le_next != NFSNOLIST) {
1040 /* Wait for any in-progress getattr to complete, */
1041 /* then remove this node from the monitored node list. */
1042 lck_mtx_lock(&nmp->nm_lock);
1043 while (np->n_mflag & NMMONSCANINPROG) {
1044 struct timespec ts = { .tv_sec = 1, .tv_nsec = 0 };
1045 np->n_mflag |= NMMONSCANWANT;
1046 msleep(&np->n_mflag, &nmp->nm_lock, PZERO - 1, "nfswaitmonscan", &ts);
1047 }
1048 if (np->n_monlink.le_next != NFSNOLIST) {
1049 LIST_REMOVE(np, n_monlink);
1050 np->n_monlink.le_next = NFSNOLIST;
1051 }
1052 lck_mtx_unlock(&nmp->nm_lock);
1053 }
1054
1055 lck_mtx_lock(nfs_buf_mutex);
1056 if (!force && (!LIST_EMPTY(&np->n_dirtyblkhd) || !LIST_EMPTY(&np->n_cleanblkhd))) {
1057 NP(np, "nfs_reclaim: dropping %s buffers", (!LIST_EMPTY(&np->n_dirtyblkhd) ? "dirty" : "clean"));
1058 }
1059 lck_mtx_unlock(nfs_buf_mutex);
1060 nfs_vinvalbuf(vp, V_IGNORE_WRITEERR, ap->a_context, 0);
1061
1062 lck_mtx_lock(nfs_node_hash_mutex);
1063
1064 if ((vnode_vtype(vp) != VDIR) && np->n_sillyrename) {
1065 if (!force) {
1066 NP(np, "nfs_reclaim: leaving unlinked file %s", np->n_sillyrename->nsr_name);
1067 }
1068 if (np->n_sillyrename->nsr_cred != NOCRED) {
1069 kauth_cred_unref(&np->n_sillyrename->nsr_cred);
1070 }
1071 vnode_rele(NFSTOV(np->n_sillyrename->nsr_dnp));
1072 FREE_ZONE(np->n_sillyrename, sizeof(*np->n_sillyrename), M_NFSREQ);
1073 }
1074
1075 vnode_removefsref(vp);
1076
1077 if (np->n_hflag & NHHASHED) {
1078 LIST_REMOVE(np, n_hash);
1079 np->n_hflag &= ~NHHASHED;
1080 FSDBG(266, 0, np, np->n_flag, 0xb1eb1e);
1081 }
1082 lck_mtx_unlock(nfs_node_hash_mutex);
1083
1084 /*
1085 * Free up any directory cookie structures and large file handle
1086 * structures that might be associated with this nfs node.
1087 */
1088 nfs_node_lock_force(np);
1089 if ((vnode_vtype(vp) == VDIR) && np->n_cookiecache) {
1090 FREE_ZONE(np->n_cookiecache, sizeof(struct nfsdmap), M_NFSDIROFF);
1091 }
1092 if (np->n_fhsize > NFS_SMALLFH) {
1093 FREE_ZONE(np->n_fhp, np->n_fhsize, M_NFSBIGFH);
1094 }
1095 if (np->n_vattr.nva_acl) {
1096 kauth_acl_free(np->n_vattr.nva_acl);
1097 }
1098 nfs_node_unlock(np);
1099 vnode_clearfsnode(vp);
1100
1101 if (np->n_parent) {
1102 if (!vnode_get(np->n_parent)) {
1103 vnode_rele(np->n_parent);
1104 vnode_put(np->n_parent);
1105 }
1106 np->n_parent = NULL;
1107 }
1108
1109 lck_mtx_destroy(&np->n_lock, nfs_node_lck_grp);
1110 lck_rw_destroy(&np->n_datalock, nfs_data_lck_grp);
1111 lck_mtx_destroy(&np->n_openlock, nfs_open_grp);
1112
1113 FSDBG_BOT(265, vp, np, np->n_flag, 0xd1ed1e);
1114 FREE_ZONE(np, sizeof(struct nfsnode), M_NFSNODE);
1115 return 0;
1116 }
1117
1118 /*
1119 * Acquire an NFS node lock
1120 */
1121
1122 int
1123 nfs_node_lock_internal(nfsnode_t np, int force)
1124 {
1125 FSDBG_TOP(268, np, force, 0, 0);
1126 lck_mtx_lock(&np->n_lock);
1127 if (!force && !(np->n_hflag && NHHASHED)) {
1128 FSDBG_BOT(268, np, 0xdead, 0, 0);
1129 lck_mtx_unlock(&np->n_lock);
1130 return ENOENT;
1131 }
1132 FSDBG_BOT(268, np, force, 0, 0);
1133 return 0;
1134 }
1135
1136 int
1137 nfs_node_lock(nfsnode_t np)
1138 {
1139 return nfs_node_lock_internal(np, 0);
1140 }
1141
1142 void
1143 nfs_node_lock_force(nfsnode_t np)
1144 {
1145 nfs_node_lock_internal(np, 1);
1146 }
1147
1148 /*
1149 * Release an NFS node lock
1150 */
1151 void
1152 nfs_node_unlock(nfsnode_t np)
1153 {
1154 FSDBG(269, np, current_thread(), 0, 0);
1155 lck_mtx_unlock(&np->n_lock);
1156 }
1157
1158 /*
1159 * Acquire 2 NFS node locks
1160 * - locks taken in reverse address order
1161 * - both or neither of the locks are taken
1162 * - only one lock taken per node (dup nodes are skipped)
1163 */
1164 int
1165 nfs_node_lock2(nfsnode_t np1, nfsnode_t np2)
1166 {
1167 nfsnode_t first, second;
1168 int error;
1169
1170 first = (np1 > np2) ? np1 : np2;
1171 second = (np1 > np2) ? np2 : np1;
1172 if ((error = nfs_node_lock(first))) {
1173 return error;
1174 }
1175 if (np1 == np2) {
1176 return error;
1177 }
1178 if ((error = nfs_node_lock(second))) {
1179 nfs_node_unlock(first);
1180 }
1181 return error;
1182 }
1183
1184 void
1185 nfs_node_unlock2(nfsnode_t np1, nfsnode_t np2)
1186 {
1187 nfs_node_unlock(np1);
1188 if (np1 != np2) {
1189 nfs_node_unlock(np2);
1190 }
1191 }
1192
1193 /*
1194 * Manage NFS node busy state.
1195 * (Similar to NFS node locks above)
1196 */
1197 int
1198 nfs_node_set_busy(nfsnode_t np, thread_t thd)
1199 {
1200 struct timespec ts = { .tv_sec = 2, .tv_nsec = 0 };
1201 int error;
1202
1203 if ((error = nfs_node_lock(np))) {
1204 return error;
1205 }
1206 while (ISSET(np->n_flag, NBUSY)) {
1207 SET(np->n_flag, NBUSYWANT);
1208 msleep(np, &np->n_lock, PZERO - 1, "nfsbusywant", &ts);
1209 if ((error = nfs_sigintr(NFSTONMP(np), NULL, thd, 0))) {
1210 break;
1211 }
1212 }
1213 if (!error) {
1214 SET(np->n_flag, NBUSY);
1215 }
1216 nfs_node_unlock(np);
1217 return error;
1218 }
1219
1220 void
1221 nfs_node_clear_busy(nfsnode_t np)
1222 {
1223 int wanted;
1224
1225 nfs_node_lock_force(np);
1226 wanted = ISSET(np->n_flag, NBUSYWANT);
1227 CLR(np->n_flag, NBUSY | NBUSYWANT);
1228 nfs_node_unlock(np);
1229 if (wanted) {
1230 wakeup(np);
1231 }
1232 }
1233
1234 int
1235 nfs_node_set_busy2(nfsnode_t np1, nfsnode_t np2, thread_t thd)
1236 {
1237 nfsnode_t first, second;
1238 int error;
1239
1240 first = (np1 > np2) ? np1 : np2;
1241 second = (np1 > np2) ? np2 : np1;
1242 if ((error = nfs_node_set_busy(first, thd))) {
1243 return error;
1244 }
1245 if (np1 == np2) {
1246 return error;
1247 }
1248 if ((error = nfs_node_set_busy(second, thd))) {
1249 nfs_node_clear_busy(first);
1250 }
1251 return error;
1252 }
1253
1254 void
1255 nfs_node_clear_busy2(nfsnode_t np1, nfsnode_t np2)
1256 {
1257 nfs_node_clear_busy(np1);
1258 if (np1 != np2) {
1259 nfs_node_clear_busy(np2);
1260 }
1261 }
1262
1263 /* helper function to sort four nodes in reverse address order (no dupes) */
1264 static void
1265 nfs_node_sort4(nfsnode_t np1, nfsnode_t np2, nfsnode_t np3, nfsnode_t np4, nfsnode_t *list, int *lcntp)
1266 {
1267 nfsnode_t na[2], nb[2];
1268 int a, b, i, lcnt;
1269
1270 /* sort pairs then merge */
1271 na[0] = (np1 > np2) ? np1 : np2;
1272 na[1] = (np1 > np2) ? np2 : np1;
1273 nb[0] = (np3 > np4) ? np3 : np4;
1274 nb[1] = (np3 > np4) ? np4 : np3;
1275 for (a = b = i = lcnt = 0; i < 4; i++) {
1276 if (a >= 2) {
1277 list[lcnt] = nb[b++];
1278 } else if ((b >= 2) || (na[a] >= nb[b])) {
1279 list[lcnt] = na[a++];
1280 } else {
1281 list[lcnt] = nb[b++];
1282 }
1283 if ((lcnt <= 0) || (list[lcnt] != list[lcnt - 1])) {
1284 lcnt++; /* omit dups */
1285 }
1286 }
1287 if (list[lcnt - 1] == NULL) {
1288 lcnt--;
1289 }
1290 *lcntp = lcnt;
1291 }
1292
1293 int
1294 nfs_node_set_busy4(nfsnode_t np1, nfsnode_t np2, nfsnode_t np3, nfsnode_t np4, thread_t thd)
1295 {
1296 nfsnode_t list[4];
1297 int i, lcnt, error;
1298
1299 nfs_node_sort4(np1, np2, np3, np4, list, &lcnt);
1300
1301 /* Now we can lock using list[0 - lcnt-1] */
1302 for (i = 0; i < lcnt; ++i) {
1303 if ((error = nfs_node_set_busy(list[i], thd))) {
1304 /* Drop any locks we acquired. */
1305 while (--i >= 0) {
1306 nfs_node_clear_busy(list[i]);
1307 }
1308 return error;
1309 }
1310 }
1311 return 0;
1312 }
1313
1314 void
1315 nfs_node_clear_busy4(nfsnode_t np1, nfsnode_t np2, nfsnode_t np3, nfsnode_t np4)
1316 {
1317 nfsnode_t list[4];
1318 int lcnt;
1319
1320 nfs_node_sort4(np1, np2, np3, np4, list, &lcnt);
1321 while (--lcnt >= 0) {
1322 nfs_node_clear_busy(list[lcnt]);
1323 }
1324 }
1325
1326 /*
1327 * Acquire an NFS node data lock
1328 */
1329 void
1330 nfs_data_lock(nfsnode_t np, int locktype)
1331 {
1332 nfs_data_lock_internal(np, locktype, 1);
1333 }
1334 void
1335 nfs_data_lock_noupdate(nfsnode_t np, int locktype)
1336 {
1337 nfs_data_lock_internal(np, locktype, 0);
1338 }
1339 void
1340 nfs_data_lock_internal(nfsnode_t np, int locktype, int updatesize)
1341 {
1342 FSDBG_TOP(270, np, locktype, np->n_datalockowner, 0);
1343 if (locktype == NFS_DATA_LOCK_SHARED) {
1344 if (updatesize && ISSET(np->n_flag, NUPDATESIZE)) {
1345 nfs_data_update_size(np, 0);
1346 }
1347 lck_rw_lock_shared(&np->n_datalock);
1348 } else {
1349 lck_rw_lock_exclusive(&np->n_datalock);
1350 np->n_datalockowner = current_thread();
1351 if (updatesize && ISSET(np->n_flag, NUPDATESIZE)) {
1352 nfs_data_update_size(np, 1);
1353 }
1354 }
1355 FSDBG_BOT(270, np, locktype, np->n_datalockowner, 0);
1356 }
1357
1358 /*
1359 * Release an NFS node data lock
1360 */
1361 void
1362 nfs_data_unlock(nfsnode_t np)
1363 {
1364 nfs_data_unlock_internal(np, 1);
1365 }
1366 void
1367 nfs_data_unlock_noupdate(nfsnode_t np)
1368 {
1369 nfs_data_unlock_internal(np, 0);
1370 }
1371 void
1372 nfs_data_unlock_internal(nfsnode_t np, int updatesize)
1373 {
1374 int mine = (np->n_datalockowner == current_thread());
1375 FSDBG_TOP(271, np, np->n_datalockowner, current_thread(), 0);
1376 if (updatesize && mine && ISSET(np->n_flag, NUPDATESIZE)) {
1377 nfs_data_update_size(np, 1);
1378 }
1379 np->n_datalockowner = NULL;
1380 lck_rw_done(&np->n_datalock);
1381 if (updatesize && !mine && ISSET(np->n_flag, NUPDATESIZE)) {
1382 nfs_data_update_size(np, 0);
1383 }
1384 FSDBG_BOT(271, np, np->n_datalockowner, current_thread(), 0);
1385 }
1386
1387
1388 /*
1389 * update an NFS node's size
1390 */
1391 void
1392 nfs_data_update_size(nfsnode_t np, int datalocked)
1393 {
1394 int error;
1395
1396 FSDBG_TOP(272, np, np->n_flag, np->n_size, np->n_newsize);
1397 if (!datalocked) {
1398 nfs_data_lock(np, NFS_DATA_LOCK_EXCLUSIVE);
1399 /* grabbing data lock will automatically update size */
1400 nfs_data_unlock(np);
1401 FSDBG_BOT(272, np, np->n_flag, np->n_size, np->n_newsize);
1402 return;
1403 }
1404 error = nfs_node_lock(np);
1405 if (error || !ISSET(np->n_flag, NUPDATESIZE)) {
1406 if (!error) {
1407 nfs_node_unlock(np);
1408 }
1409 FSDBG_BOT(272, np, np->n_flag, np->n_size, np->n_newsize);
1410 return;
1411 }
1412 CLR(np->n_flag, NUPDATESIZE);
1413 np->n_size = np->n_newsize;
1414 /* make sure we invalidate buffers the next chance we get */
1415 SET(np->n_flag, NNEEDINVALIDATE);
1416 nfs_node_unlock(np);
1417 ubc_setsize(NFSTOV(np), (off_t)np->n_size); /* XXX error? */
1418 FSDBG_BOT(272, np, np->n_flag, np->n_size, np->n_newsize);
1419 }
1420
1421 #define DODEBUG 1
1422
1423 int
1424 nfs_mount_is_dirty(mount_t mp)
1425 {
1426 u_long i;
1427 nfsnode_t np;
1428 #ifdef DODEBUG
1429 struct timeval now, then, diff;
1430 u_long ncnt = 0;
1431 microuptime(&now);
1432 #endif
1433 lck_mtx_lock(nfs_node_hash_mutex);
1434 for (i = 0; i <= nfsnodehash; i++) {
1435 LIST_FOREACH(np, &nfsnodehashtbl[i], n_hash) {
1436 #ifdef DODEBUG
1437 ncnt++;
1438 #endif
1439 if (np->n_mount == mp && !LIST_EMPTY(&np->n_dirtyblkhd)) {
1440 goto out;
1441 }
1442 }
1443 }
1444 out:
1445 lck_mtx_unlock(nfs_node_hash_mutex);
1446 #ifdef DODEBUG
1447 microuptime(&then);
1448 timersub(&then, &now, &diff);
1449
1450 NFS_DBG(NFS_FAC_SOCK, 7, "mount_is_dirty for %s took %lld mics for %ld slots and %ld nodes return %d\n",
1451 vfs_statfs(mp)->f_mntfromname, (uint64_t)diff.tv_sec * 1000000LL + diff.tv_usec, i, ncnt, (i <= nfsnodehash));
1452 #endif
1453
1454 return i <= nfsnodehash;
1455 }