X-Git-Url: https://git.saurik.com/apple/xnu.git/blobdiff_plain/de355530ae67247cbd0da700edb3a2a1dae884c2..db6096698656d32db7df630594bd9617ee54f828:/bsd/nfs/nfs_node.c diff --git a/bsd/nfs/nfs_node.c b/bsd/nfs/nfs_node.c index f12696b36..7fc9ddaef 100644 --- a/bsd/nfs/nfs_node.c +++ b/bsd/nfs/nfs_node.c @@ -1,23 +1,29 @@ /* - * Copyright (c) 2000 Apple Computer, Inc. All rights reserved. + * Copyright (c) 2000-2011 Apple Inc. All rights reserved. * - * @APPLE_LICENSE_HEADER_START@ + * @APPLE_OSREFERENCE_LICENSE_HEADER_START@ * - * The contents of this file constitute Original Code as defined in and - * are subject to the Apple Public Source License Version 1.1 (the - * "License"). You may not use this file except in compliance with the - * License. Please obtain a copy of the License at - * http://www.apple.com/publicsource and read it before using this file. + * This file contains Original Code and/or Modifications of Original Code + * as defined in and that are subject to the Apple Public Source License + * Version 2.0 (the 'License'). You may not use this file except in + * compliance with the License. The rights granted to you under the License + * may not be used to create, or enable the creation or redistribution of, + * unlawful or unlicensed copies of an Apple operating system, or to + * circumvent, violate, or enable the circumvention or violation of, any + * terms of an Apple operating system software license agreement. * - * This Original Code and all software distributed under the License are - * distributed on an "AS IS" basis, WITHOUT WARRANTY OF ANY KIND, EITHER + * Please obtain a copy of the License at + * http://www.opensource.apple.com/apsl/ and read it before using this file. + * + * The Original Code and all software distributed under the License are + * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES, * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY, - * FITNESS FOR A PARTICULAR PURPOSE OR NON-INFRINGEMENT. Please see the - * License for the specific language governing rights and limitations - * under the License. + * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT. + * Please see the License for the specific language governing rights and + * limitations under the License. * - * @APPLE_LICENSE_HEADER_END@ + * @APPLE_OSREFERENCE_LICENSE_HEADER_END@ */ /* Copyright (c) 1995 NeXT Computer, Inc. All Rights Reserved */ /* @@ -61,55 +67,68 @@ #include +#include #include #include -#include -#include +#include +#include #include +#include #include +#include +#include #include #include #include #include +#include #include -#ifdef MALLOC_DEFINE -static MALLOC_DEFINE(M_NFSNODE, "NFS node", "NFS vnode private part"); -#endif - -LIST_HEAD(nfsnodehashhead, nfsnode) *nfsnodehashtbl; -u_long nfsnodehash; +#define NFSNOHASH(fhsum) \ + (&nfsnodehashtbl[(fhsum) & nfsnodehash]) +static LIST_HEAD(nfsnodehashhead, nfsnode) *nfsnodehashtbl; +static u_long nfsnodehash; -#define TRUE 1 -#define FALSE 0 +static lck_grp_t *nfs_node_hash_lck_grp; +static lck_grp_t *nfs_node_lck_grp; +static lck_grp_t *nfs_data_lck_grp; +lck_mtx_t *nfs_node_hash_mutex; /* * Initialize hash links for nfsnodes * and build nfsnode free list. */ void -nfs_nhinit() +nfs_nhinit(void) +{ + nfs_node_hash_lck_grp = lck_grp_alloc_init("nfs_node_hash", LCK_GRP_ATTR_NULL); + nfs_node_hash_mutex = lck_mtx_alloc_init(nfs_node_hash_lck_grp, LCK_ATTR_NULL); + nfs_node_lck_grp = lck_grp_alloc_init("nfs_node", LCK_GRP_ATTR_NULL); + nfs_data_lck_grp = lck_grp_alloc_init("nfs_data", LCK_GRP_ATTR_NULL); +} + +void +nfs_nhinit_finish(void) { - nfsnodehashtbl = hashinit(desiredvnodes, M_NFSNODE, &nfsnodehash); + lck_mtx_lock(nfs_node_hash_mutex); + if (!nfsnodehashtbl) + nfsnodehashtbl = hashinit(desiredvnodes, M_NFSNODE, &nfsnodehash); + lck_mtx_unlock(nfs_node_hash_mutex); } /* * Compute an entry in the NFS hash table structure */ u_long -nfs_hash(fhp, fhsize) - register nfsfh_t *fhp; - int fhsize; +nfs_hash(u_char *fhp, int fhsize) { - register u_char *fhpp; - register u_long fhsum; - register int i; + u_long fhsum; + int i; - fhpp = &fhp->fh_bytes[0]; fhsum = 0; for (i = 0; i < fhsize; i++) - fhsum += *fhpp++; + fhsum += *fhp++; return (fhsum); } @@ -119,168 +138,565 @@ nfs_hash(fhp, fhsize) * In all cases, a pointer to a * nfsnode structure is returned. */ -int nfs_node_hash_lock; - int -nfs_nget(mntp, fhp, fhsize, npp) - struct mount *mntp; - register nfsfh_t *fhp; - int fhsize; - struct nfsnode **npp; +nfs_nget( + mount_t mp, + nfsnode_t dnp, + struct componentname *cnp, + u_char *fhp, + int fhsize, + struct nfs_vattr *nvap, + u_int64_t *xidp, + uint32_t auth, + int flags, + nfsnode_t *npp) { - struct proc *p = current_proc(); /* XXX */ - struct nfsnode *np; + nfsnode_t np; struct nfsnodehashhead *nhpp; - register struct vnode *vp; - struct vnode *nvp; - int error; + vnode_t vp; + int error, nfsvers; + mount_t mp2; + struct vnode_fsparam vfsp; + uint32_t vid; + + FSDBG_TOP(263, mp, dnp, flags, npp); /* Check for unmount in progress */ - if (mntp->mnt_kern_flag & MNTK_UNMOUNT) { - *npp = 0; - return (EPERM); + if (!mp || (mp->mnt_kern_flag & MNTK_FRCUNMOUNT)) { + *npp = NULL; + error = ENXIO; + FSDBG_BOT(263, mp, dnp, 0xd1e, error); + return (error); } + nfsvers = VFSTONFS(mp)->nm_vers; nhpp = NFSNOHASH(nfs_hash(fhp, fhsize)); loop: + lck_mtx_lock(nfs_node_hash_mutex); for (np = nhpp->lh_first; np != 0; np = np->n_hash.le_next) { - if (mntp != NFSTOV(np)->v_mount || np->n_fhsize != fhsize || - bcmp((caddr_t)fhp, (caddr_t)np->n_fhp, fhsize)) + mp2 = (np->n_hflag & NHINIT) ? np->n_mount : NFSTOMP(np); + if (mp != mp2 || np->n_fhsize != fhsize || + bcmp(fhp, np->n_fhp, fhsize)) continue; + if (nvap && (nvap->nva_flags & NFS_FFLAG_TRIGGER_REFERRAL) && + cnp && (cnp->cn_namelen > (fhsize - (int)sizeof(dnp)))) { + /* The name was too long to fit in the file handle. Check it against the node's name. */ + int namecmp = 0; + const char *vname = vnode_getname(NFSTOV(np)); + if (vname) { + if (cnp->cn_namelen != (int)strlen(vname)) + namecmp = 1; + else + namecmp = strncmp(vname, cnp->cn_nameptr, cnp->cn_namelen); + vnode_putname(vname); + } + if (namecmp) /* full name didn't match */ + continue; + } + FSDBG(263, dnp, np, np->n_flag, 0xcace0000); + /* if the node is locked, sleep on it */ + if ((np->n_hflag & NHLOCKED) && !(flags & NG_NOCREATE)) { + np->n_hflag |= NHLOCKWANT; + FSDBG(263, dnp, np, np->n_flag, 0xcace2222); + msleep(np, nfs_node_hash_mutex, PDROP | PINOD, "nfs_nget", NULL); + FSDBG(263, dnp, np, np->n_flag, 0xcace3333); + goto loop; + } vp = NFSTOV(np); - if (vget(vp, LK_EXCLUSIVE, p)) + vid = vnode_vid(vp); + lck_mtx_unlock(nfs_node_hash_mutex); + if ((error = vnode_getwithvid(vp, vid))) { + /* + * If vnode is being reclaimed or has already + * changed identity, no need to wait. + */ + FSDBG_BOT(263, dnp, *npp, 0xcace0d1e, error); + return (error); + } + if ((error = nfs_node_lock(np))) { + /* this only fails if the node is now unhashed */ + /* so let's see if we can find/create it again */ + FSDBG(263, dnp, *npp, 0xcaced1e2, error); + vnode_put(vp); + if (flags & NG_NOCREATE) { + *npp = 0; + FSDBG_BOT(263, dnp, *npp, 0xcaced1e0, ENOENT); + return (ENOENT); + } goto loop; - *npp = np; - return(0); + } + /* update attributes */ + if (nvap) + error = nfs_loadattrcache(np, nvap, xidp, 0); + if (error) { + nfs_node_unlock(np); + vnode_put(vp); + } else { + if (dnp && cnp && (flags & NG_MAKEENTRY)) + cache_enter(NFSTOV(dnp), vp, cnp); + *npp = np; + } + FSDBG_BOT(263, dnp, *npp, 0xcace0000, error); + return(error); + } + + FSDBG(263, mp, dnp, npp, 0xaaaaaaaa); + + if (flags & NG_NOCREATE) { + lck_mtx_unlock(nfs_node_hash_mutex); + *npp = 0; + FSDBG_BOT(263, dnp, *npp, 0x80000001, ENOENT); + return (ENOENT); } + /* - * Obtain a lock to prevent a race condition if the getnewvnode() - * or MALLOC() below happens to block. + * allocate and initialize nfsnode and stick it in the hash + * before calling getnewvnode(). Anyone finding it in the + * hash before initialization is complete will wait for it. */ - if (nfs_node_hash_lock) { - while (nfs_node_hash_lock) { - nfs_node_hash_lock = -1; - tsleep(&nfs_node_hash_lock, PVM, "nfsngt", 0); + MALLOC_ZONE(np, nfsnode_t, sizeof *np, M_NFSNODE, M_WAITOK); + if (!np) { + lck_mtx_unlock(nfs_node_hash_mutex); + *npp = 0; + FSDBG_BOT(263, dnp, *npp, 0x80000001, ENOMEM); + return (ENOMEM); + } + bzero(np, sizeof *np); + np->n_hflag |= (NHINIT | NHLOCKED); + np->n_mount = mp; + np->n_auth = auth; + TAILQ_INIT(&np->n_opens); + TAILQ_INIT(&np->n_lock_owners); + TAILQ_INIT(&np->n_locks); + np->n_dlink.tqe_next = NFSNOLIST; + np->n_dreturn.tqe_next = NFSNOLIST; + np->n_monlink.le_next = NFSNOLIST; + + /* ugh... need to keep track of ".zfs" directories to workaround server bugs */ + if ((nvap->nva_type == VDIR) && cnp && (cnp->cn_namelen == 4) && + (cnp->cn_nameptr[0] == '.') && (cnp->cn_nameptr[1] == 'z') && + (cnp->cn_nameptr[2] == 'f') && (cnp->cn_nameptr[3] == 's')) + np->n_flag |= NISDOTZFS; + if (dnp && (dnp->n_flag & NISDOTZFS)) + np->n_flag |= NISDOTZFSCHILD; + + if (dnp && cnp && ((cnp->cn_namelen != 2) || + (cnp->cn_nameptr[0] != '.') || (cnp->cn_nameptr[1] != '.'))) { + vnode_t dvp = NFSTOV(dnp); + if (!vnode_get(dvp)) { + if (!vnode_ref(dvp)) + np->n_parent = dvp; + vnode_put(dvp); } - goto loop; } - nfs_node_hash_lock = 1; - /* - * Do the MALLOC before the getnewvnode since doing so afterward - * might cause a bogus v_data pointer to get dereferenced - * elsewhere if MALLOC should block. - */ - MALLOC_ZONE(np, struct nfsnode *, sizeof *np, M_NFSNODE, M_WAITOK); - - error = getnewvnode(VT_NFS, mntp, nfsv2_vnodeop_p, &nvp); + /* setup node's file handle */ + if (fhsize > NFS_SMALLFH) { + MALLOC_ZONE(np->n_fhp, u_char *, + fhsize, M_NFSBIGFH, M_WAITOK); + if (!np->n_fhp) { + lck_mtx_unlock(nfs_node_hash_mutex); + FREE_ZONE(np, sizeof *np, M_NFSNODE); + *npp = 0; + FSDBG_BOT(263, dnp, *npp, 0x80000002, ENOMEM); + return (ENOMEM); + } + } else { + np->n_fhp = &np->n_fh[0]; + } + bcopy(fhp, np->n_fhp, fhsize); + np->n_fhsize = fhsize; + + /* Insert the nfsnode in the hash queue for its new file handle */ + LIST_INSERT_HEAD(nhpp, np, n_hash); + np->n_hflag |= NHHASHED; + FSDBG(266, 0, np, np->n_flag, np->n_hflag); + + /* lock the new nfsnode */ + lck_mtx_init(&np->n_lock, nfs_node_lck_grp, LCK_ATTR_NULL); + lck_rw_init(&np->n_datalock, nfs_data_lck_grp, LCK_ATTR_NULL); + lck_mtx_init(&np->n_openlock, nfs_open_grp, LCK_ATTR_NULL); + lck_mtx_lock(&np->n_lock); + + /* release lock on hash table */ + lck_mtx_unlock(nfs_node_hash_mutex); + + /* do initial loading of attributes */ + NACLINVALIDATE(np); + NACCESSINVALIDATE(np); + error = nfs_loadattrcache(np, nvap, xidp, 1); if (error) { - if (nfs_node_hash_lock < 0) - wakeup(&nfs_node_hash_lock); - nfs_node_hash_lock = 0; - *npp = 0; + FSDBG(266, 0, np, np->n_flag, 0xb1eb1e); + nfs_node_unlock(np); + lck_mtx_lock(nfs_node_hash_mutex); + LIST_REMOVE(np, n_hash); + np->n_hflag &= ~(NHHASHED|NHINIT|NHLOCKED); + if (np->n_hflag & NHLOCKWANT) { + np->n_hflag &= ~NHLOCKWANT; + wakeup(np); + } + lck_mtx_unlock(nfs_node_hash_mutex); + if (np->n_parent) { + if (!vnode_get(np->n_parent)) { + vnode_rele(np->n_parent); + vnode_put(np->n_parent); + } + np->n_parent = NULL; + } + lck_mtx_destroy(&np->n_lock, nfs_node_lck_grp); + lck_rw_destroy(&np->n_datalock, nfs_data_lck_grp); + lck_mtx_destroy(&np->n_openlock, nfs_open_grp); + if (np->n_fhsize > NFS_SMALLFH) + FREE_ZONE(np->n_fhp, np->n_fhsize, M_NFSBIGFH); FREE_ZONE(np, sizeof *np, M_NFSNODE); + *npp = 0; + FSDBG_BOT(263, dnp, *npp, 0x80000003, error); return (error); } - vp = nvp; - bzero((caddr_t)np, sizeof *np); - vp->v_data = np; - np->n_vnode = vp; - /* - * Insert the nfsnode in the hash queue for its new file handle - */ - LIST_INSERT_HEAD(nhpp, np, n_hash); - if (fhsize > NFS_SMALLFH) { - MALLOC_ZONE(np->n_fhp, nfsfh_t *, - fhsize, M_NFSBIGFH, M_WAITOK); + NFS_CHANGED_UPDATE(nfsvers, np, nvap); + if (nvap->nva_type == VDIR) + NFS_CHANGED_UPDATE_NC(nfsvers, np, nvap); + + /* now, attempt to get a new vnode */ + vfsp.vnfs_mp = mp; + vfsp.vnfs_vtype = nvap->nva_type; + vfsp.vnfs_str = "nfs"; + vfsp.vnfs_dvp = dnp ? NFSTOV(dnp) : NULL; + vfsp.vnfs_fsnode = np; + if (nfsvers == NFS_VER4) { +#if FIFO + if (nvap->nva_type == VFIFO) + vfsp.vnfs_vops = fifo_nfsv4nodeop_p; + else +#endif /* FIFO */ + if (nvap->nva_type == VBLK || nvap->nva_type == VCHR) + vfsp.vnfs_vops = spec_nfsv4nodeop_p; + else + vfsp.vnfs_vops = nfsv4_vnodeop_p; + } else { +#if FIFO + if (nvap->nva_type == VFIFO) + vfsp.vnfs_vops = fifo_nfsv2nodeop_p; + else +#endif /* FIFO */ + if (nvap->nva_type == VBLK || nvap->nva_type == VCHR) + vfsp.vnfs_vops = spec_nfsv2nodeop_p; + else + vfsp.vnfs_vops = nfsv2_vnodeop_p; + } + vfsp.vnfs_markroot = (flags & NG_MARKROOT) ? 1 : 0; + vfsp.vnfs_marksystem = 0; + vfsp.vnfs_rdev = 0; + vfsp.vnfs_filesize = nvap->nva_size; + vfsp.vnfs_cnp = cnp; + vfsp.vnfs_flags = VNFS_ADDFSREF; + if (!dnp || !cnp || !(flags & NG_MAKEENTRY)) + vfsp.vnfs_flags |= VNFS_NOCACHE; + +#if CONFIG_TRIGGERS + if ((nfsvers >= NFS_VER4) && (nvap->nva_type == VDIR) && (np->n_vattr.nva_flags & NFS_FFLAG_TRIGGER)) { + struct vnode_trigger_param vtp; + bzero(&vtp, sizeof(vtp)); + bcopy(&vfsp, &vtp.vnt_params, sizeof(vfsp)); + vtp.vnt_resolve_func = nfs_mirror_mount_trigger_resolve; + vtp.vnt_unresolve_func = nfs_mirror_mount_trigger_unresolve; + vtp.vnt_rearm_func = nfs_mirror_mount_trigger_rearm; + vtp.vnt_flags = VNT_AUTO_REARM; + error = vnode_create(VNCREATE_TRIGGER, VNCREATE_TRIGGER_SIZE, &vtp, &np->n_vnode); } else - np->n_fhp = &np->n_fh; - bcopy((caddr_t)fhp, (caddr_t)np->n_fhp, fhsize); - np->n_fhsize = fhsize; - *npp = np; +#endif + { + error = vnode_create(VNCREATE_FLAVOR, VCREATESIZE, &vfsp, &np->n_vnode); + } + if (error) { + FSDBG(266, 0, np, np->n_flag, 0xb1eb1e); + nfs_node_unlock(np); + lck_mtx_lock(nfs_node_hash_mutex); + LIST_REMOVE(np, n_hash); + np->n_hflag &= ~(NHHASHED|NHINIT|NHLOCKED); + if (np->n_hflag & NHLOCKWANT) { + np->n_hflag &= ~NHLOCKWANT; + wakeup(np); + } + lck_mtx_unlock(nfs_node_hash_mutex); + if (np->n_parent) { + if (!vnode_get(np->n_parent)) { + vnode_rele(np->n_parent); + vnode_put(np->n_parent); + } + np->n_parent = NULL; + } + lck_mtx_destroy(&np->n_lock, nfs_node_lck_grp); + lck_rw_destroy(&np->n_datalock, nfs_data_lck_grp); + lck_mtx_destroy(&np->n_openlock, nfs_open_grp); + if (np->n_fhsize > NFS_SMALLFH) + FREE_ZONE(np->n_fhp, np->n_fhsize, M_NFSBIGFH); + FREE_ZONE(np, sizeof *np, M_NFSNODE); + *npp = 0; + FSDBG_BOT(263, dnp, *npp, 0x80000004, error); + return (error); + } + vp = np->n_vnode; + vnode_settag(vp, VT_NFS); + /* node is now initialized */ - if (nfs_node_hash_lock < 0) - wakeup(&nfs_node_hash_lock); - nfs_node_hash_lock = 0; + /* check if anyone's waiting on this node */ + lck_mtx_lock(nfs_node_hash_mutex); + np->n_hflag &= ~(NHINIT|NHLOCKED); + if (np->n_hflag & NHLOCKWANT) { + np->n_hflag &= ~NHLOCKWANT; + wakeup(np); + } + lck_mtx_unlock(nfs_node_hash_mutex); - /* - * Lock the new nfsnode. - */ - error = vn_lock(vp, LK_EXCLUSIVE | LK_RETRY, p); + *npp = np; + FSDBG_BOT(263, dnp, vp, *npp, error); return (error); } + int -nfs_inactive(ap) - struct vop_inactive_args /* { - struct vnode *a_vp; - struct proc *a_p; +nfs_vnop_inactive(ap) + struct vnop_inactive_args /* { + struct vnodeop_desc *a_desc; + vnode_t a_vp; + vfs_context_t a_context; } */ *ap; { - register struct nfsnode *np; - register struct sillyrename *sp; - struct proc *p = current_proc(); /* XXX */ - extern int prtactive; - struct ucred *cred; - - np = VTONFS(ap->a_vp); - if (prtactive && ap->a_vp->v_usecount != 0) - vprint("nfs_inactive: pushing active", ap->a_vp); - if (ap->a_vp->v_type != VDIR) { - sp = np->n_sillyrename; - np->n_sillyrename = (struct sillyrename *)0; - } else - sp = (struct sillyrename *)0; + vnode_t vp = ap->a_vp; + vfs_context_t ctx = ap->a_context; + nfsnode_t np = VTONFS(ap->a_vp); + struct nfs_sillyrename *nsp; + struct nfs_vattr nvattr; + int unhash, attrerr, busyerror, error, inuse, busied, force; + struct nfs_open_file *nofp; + struct componentname cn; + struct nfsmount *nmp = NFSTONMP(np); + mount_t mp = vnode_mount(vp); - if (sp) { +restart: + force = (!mp || (mp->mnt_kern_flag & MNTK_FRCUNMOUNT)); + error = 0; + inuse = (nfs_mount_state_in_use_start(nmp, NULL) == 0); + + /* There shouldn't be any open or lock state at this point */ + lck_mtx_lock(&np->n_openlock); + if (np->n_openrefcnt && !force) + NP(np, "nfs_vnop_inactive: still open: %d", np->n_openrefcnt); + TAILQ_FOREACH(nofp, &np->n_opens, nof_link) { + lck_mtx_lock(&nofp->nof_lock); + if (nofp->nof_flags & NFS_OPEN_FILE_BUSY) { + if (!force) + NP(np, "nfs_vnop_inactive: open file busy"); + busied = 0; + } else { + nofp->nof_flags |= NFS_OPEN_FILE_BUSY; + busied = 1; + } + lck_mtx_unlock(&nofp->nof_lock); + if ((np->n_flag & NREVOKE) || (nofp->nof_flags & NFS_OPEN_FILE_LOST)) { + if (busied) + nfs_open_file_clear_busy(nofp); + continue; + } /* - * Remove the silly file that was rename'd earlier + * If we just created the file, we already had it open in + * anticipation of getting a subsequent open call. If the + * node has gone inactive without being open, we need to + * clean up (close) the open done in the create. */ -#if DIAGNOSTIC - kprintf("nfs_inactive removing %s, dvp=%x, a_vp=%x, ap=%x, np=%x, sp=%x\n", &sp->s_name[0], (unsigned)sp->s_dvp, (unsigned)ap->a_vp, (unsigned)ap, (unsigned)np, (unsigned)sp); -#endif + if ((nofp->nof_flags & NFS_OPEN_FILE_CREATE) && nofp->nof_creator && !force) { + if (nofp->nof_flags & NFS_OPEN_FILE_REOPEN) { + lck_mtx_unlock(&np->n_openlock); + if (busied) + nfs_open_file_clear_busy(nofp); + if (inuse) + nfs_mount_state_in_use_end(nmp, 0); + if (!nfs4_reopen(nofp, NULL)) + goto restart; + } + nofp->nof_flags &= ~NFS_OPEN_FILE_CREATE; + lck_mtx_unlock(&np->n_openlock); + error = nfs_close(np, nofp, NFS_OPEN_SHARE_ACCESS_BOTH, NFS_OPEN_SHARE_DENY_NONE, ctx); + if (error) { + NP(np, "nfs_vnop_inactive: create close error: %d", error); + nofp->nof_flags |= NFS_OPEN_FILE_CREATE; + } + if (busied) + nfs_open_file_clear_busy(nofp); + if (inuse) + nfs_mount_state_in_use_end(nmp, error); + goto restart; + } + if (nofp->nof_flags & NFS_OPEN_FILE_NEEDCLOSE) { + /* + * If the file is marked as needing reopen, but this was the only + * open on the file, just drop the open. + */ + nofp->nof_flags &= ~NFS_OPEN_FILE_NEEDCLOSE; + if ((nofp->nof_flags & NFS_OPEN_FILE_REOPEN) && (nofp->nof_opencnt == 1)) { + nofp->nof_flags &= ~NFS_OPEN_FILE_REOPEN; + nofp->nof_r--; + nofp->nof_opencnt--; + nofp->nof_access = 0; + } else if (!force) { + lck_mtx_unlock(&np->n_openlock); + if (nofp->nof_flags & NFS_OPEN_FILE_REOPEN) { + if (busied) + nfs_open_file_clear_busy(nofp); + if (inuse) + nfs_mount_state_in_use_end(nmp, 0); + if (!nfs4_reopen(nofp, NULL)) + goto restart; + } + error = nfs_close(np, nofp, NFS_OPEN_SHARE_ACCESS_READ, NFS_OPEN_SHARE_DENY_NONE, ctx); + if (error) { + NP(np, "nfs_vnop_inactive: need close error: %d", error); + nofp->nof_flags |= NFS_OPEN_FILE_NEEDCLOSE; + } + if (busied) + nfs_open_file_clear_busy(nofp); + if (inuse) + nfs_mount_state_in_use_end(nmp, error); + goto restart; + } + } + if (nofp->nof_opencnt && !force) + NP(np, "nfs_vnop_inactive: file still open: %d", nofp->nof_opencnt); + if (!force && (nofp->nof_access || nofp->nof_deny || + nofp->nof_mmap_access || nofp->nof_mmap_deny || + nofp->nof_r || nofp->nof_w || nofp->nof_rw || + nofp->nof_r_dw || nofp->nof_w_dw || nofp->nof_rw_dw || + nofp->nof_r_drw || nofp->nof_w_drw || nofp->nof_rw_drw || + nofp->nof_d_r || nofp->nof_d_w || nofp->nof_d_rw || + nofp->nof_d_r_dw || nofp->nof_d_w_dw || nofp->nof_d_rw_dw || + nofp->nof_d_r_drw || nofp->nof_d_w_drw || nofp->nof_d_rw_drw)) { + NP(np, "nfs_vnop_inactive: non-zero access: %d %d %d %d # %u.%u %u.%u %u.%u dw %u.%u %u.%u %u.%u drw %u.%u %u.%u %u.%u", + nofp->nof_access, nofp->nof_deny, + nofp->nof_mmap_access, nofp->nof_mmap_deny, + nofp->nof_r, nofp->nof_d_r, + nofp->nof_w, nofp->nof_d_w, + nofp->nof_rw, nofp->nof_d_rw, + nofp->nof_r_dw, nofp->nof_d_r_dw, + nofp->nof_w_dw, nofp->nof_d_w_dw, + nofp->nof_rw_dw, nofp->nof_d_rw_dw, + nofp->nof_r_drw, nofp->nof_d_r_drw, + nofp->nof_w_drw, nofp->nof_d_w_drw, + nofp->nof_rw_drw, nofp->nof_d_rw_drw); + } + if (busied) + nfs_open_file_clear_busy(nofp); + } + lck_mtx_unlock(&np->n_openlock); + + if (inuse && nfs_mount_state_in_use_end(nmp, error)) + goto restart; + + nfs_node_lock_force(np); + + if (vnode_vtype(vp) != VDIR) { + nsp = np->n_sillyrename; + np->n_sillyrename = NULL; + } else { + nsp = NULL; + } + + FSDBG_TOP(264, vp, np, np->n_flag, nsp); + + if (!nsp) { + /* no silly file to clean up... */ + /* clear all flags other than these */ + np->n_flag &= (NMODIFIED); + nfs_node_unlock(np); + FSDBG_BOT(264, vp, np, np->n_flag, 0); + return (0); + } + nfs_node_unlock(np); + + /* Remove the silly file that was rename'd earlier */ + + /* flush all the buffers */ + nfs_vinvalbuf2(vp, V_SAVE, vfs_context_thread(ctx), nsp->nsr_cred, 1); + + /* try to get the latest attributes */ + attrerr = nfs_getattr(np, &nvattr, ctx, NGA_UNCACHED); + + /* Check if we should remove it from the node hash. */ + /* Leave it if inuse or it has multiple hard links. */ + if (vnode_isinuse(vp, 0) || (!attrerr && (nvattr.nva_nlink > 1))) { + unhash = 0; + } else { + unhash = 1; + ubc_setsize(vp, 0); + } + + /* mark this node and the directory busy while we do the remove */ + busyerror = nfs_node_set_busy2(nsp->nsr_dnp, np, vfs_context_thread(ctx)); + + /* lock the node while we remove the silly file */ + lck_mtx_lock(nfs_node_hash_mutex); + while (np->n_hflag & NHLOCKED) { + np->n_hflag |= NHLOCKWANT; + msleep(np, nfs_node_hash_mutex, PINOD, "nfs_inactive", NULL); + } + np->n_hflag |= NHLOCKED; + lck_mtx_unlock(nfs_node_hash_mutex); + + /* purge the name cache to deter others from finding it */ + bzero(&cn, sizeof(cn)); + cn.cn_nameptr = nsp->nsr_name; + cn.cn_namelen = nsp->nsr_namlen; + nfs_name_cache_purge(nsp->nsr_dnp, np, &cn, ctx); + + FSDBG(264, np, np->n_size, np->n_vattr.nva_size, 0xf00d00f1); + + /* now remove the silly file */ + nfs_removeit(nsp); + + /* clear all flags other than these */ + nfs_node_lock_force(np); + np->n_flag &= (NMODIFIED); + nfs_node_unlock(np); + + if (!busyerror) + nfs_node_clear_busy2(nsp->nsr_dnp, np); + + if (unhash && vnode_isinuse(vp, 0)) { + /* vnode now inuse after silly remove? */ + unhash = 0; + ubc_setsize(vp, np->n_size); + } + + lck_mtx_lock(nfs_node_hash_mutex); + if (unhash) { /* - * We get a reference (vget) to ensure getnewvnode() - * doesn't recycle vp while we're asleep awaiting I/O. - * Note we don't need the reference unless usecount is - * already zero. In the case of a forcible unmount it - * wont be zero and doing a vget would fail because - * vclean holds VXLOCK. + * remove nfsnode from hash now so we can't accidentally find it + * again if another object gets created with the same filehandle + * before this vnode gets reclaimed */ - if (ap->a_vp->v_usecount > 0) { - VREF(ap->a_vp); - } else if (vget(ap->a_vp, 0, ap->a_p)) - panic("nfs_inactive: vget failed"); - (void) nfs_vinvalbuf(ap->a_vp, 0, sp->s_cred, p, 1); - np->n_size = 0; - ubc_setsize(ap->a_vp, (off_t)0); - - /* We have a problem. The dvp could have gone away on us while - * in the unmount path. Thus it appears as VBAD and we cannot - * use it. If we tried locking the parent (future), for silly - * rename files, it is unclear where we would lock. The unmount - * code just pulls unlocked vnodes as it goes thru its list and - * yanks them. Could unmount be smarter to see if a busy reg vnode has - * a parent, and not yank it yet? Put in more passes at unmount - * time? In the meantime, just check if it went away on us. - * Could have gone away during the nfs_vinvalbuf or ubc_setsize - * which block. Or perhaps even before nfs_inactive got called. - */ - if ((sp->s_dvp)->v_type != VBAD) - nfs_removeit(sp); /* uses the dvp */ - cred = sp->s_cred; - if (cred != NOCRED) { - sp->s_cred = NOCRED; - crfree(cred); + if (np->n_hflag & NHHASHED) { + LIST_REMOVE(np, n_hash); + np->n_hflag &= ~NHHASHED; + FSDBG(266, 0, np, np->n_flag, 0xb1eb1e); } - vrele(sp->s_dvp); - FREE_ZONE((caddr_t)sp, sizeof (struct sillyrename), M_NFSREQ); - vrele(ap->a_vp); + vnode_recycle(vp); } - np->n_flag &= (NMODIFIED | NFLUSHINPROG | NFLUSHWANT | NQNFSEVICTED | - NQNFSNONCACHE | NQNFSWRITE); - VOP_UNLOCK(ap->a_vp, 0, ap->a_p); + /* unlock the node */ + np->n_hflag &= ~NHLOCKED; + if (np->n_hflag & NHLOCKWANT) { + np->n_hflag &= ~NHLOCKWANT; + wakeup(np); + } + lck_mtx_unlock(nfs_node_hash_mutex); + + /* cleanup sillyrename info */ + if (nsp->nsr_cred != NOCRED) + kauth_cred_unref(&nsp->nsr_cred); + vnode_rele(NFSTOV(nsp->nsr_dnp)); + FREE_ZONE(nsp, sizeof(*nsp), M_NFSREQ); + + FSDBG_BOT(264, vp, np, np->n_flag, 0); return (0); } @@ -288,129 +704,510 @@ nfs_inactive(ap) * Reclaim an nfsnode so that it can be used for other purposes. */ int -nfs_reclaim(ap) - struct vop_reclaim_args /* { - struct vnode *a_vp; +nfs_vnop_reclaim(ap) + struct vnop_reclaim_args /* { + struct vnodeop_desc *a_desc; + vnode_t a_vp; + vfs_context_t a_context; } */ *ap; { - register struct vnode *vp = ap->a_vp; - register struct nfsnode *np = VTONFS(vp); - register struct nfsmount *nmp = VFSTONFS(vp->v_mount); - register struct nfsdmap *dp, *dp2; - extern int prtactive; + vnode_t vp = ap->a_vp; + nfsnode_t np = VTONFS(vp); + vfs_context_t ctx = ap->a_context; + struct nfs_open_file *nofp, *nextnofp; + struct nfs_file_lock *nflp, *nextnflp; + struct nfs_lock_owner *nlop, *nextnlop; + struct nfsmount *nmp = np->n_mount ? VFSTONFS(np->n_mount) : NFSTONMP(np); + mount_t mp = vnode_mount(vp); + int force; - if (prtactive && vp->v_usecount != 0) - vprint("nfs_reclaim: pushing active", vp); + FSDBG_TOP(265, vp, np, np->n_flag, 0); + force = (!mp || (mp->mnt_kern_flag & MNTK_FRCUNMOUNT)); - LIST_REMOVE(np, n_hash); + /* There shouldn't be any open or lock state at this point */ + lck_mtx_lock(&np->n_openlock); - /* - * In case we block during FREE_ZONEs below, get the entry out - * of tbe name cache now so subsequent lookups won't find it. - */ - cache_purge(vp); + if (nmp && (nmp->nm_vers >= NFS_VER4)) { + /* need to drop a delegation */ + if (np->n_dreturn.tqe_next != NFSNOLIST) { + /* remove this node from the delegation return list */ + lck_mtx_lock(&nmp->nm_lock); + if (np->n_dreturn.tqe_next != NFSNOLIST) { + TAILQ_REMOVE(&nmp->nm_dreturnq, np, n_dreturn); + np->n_dreturn.tqe_next = NFSNOLIST; + } + lck_mtx_unlock(&nmp->nm_lock); + } + if (np->n_dlink.tqe_next != NFSNOLIST) { + /* remove this node from the delegation list */ + lck_mtx_lock(&nmp->nm_lock); + if (np->n_dlink.tqe_next != NFSNOLIST) { + TAILQ_REMOVE(&nmp->nm_delegations, np, n_dlink); + np->n_dlink.tqe_next = NFSNOLIST; + } + lck_mtx_unlock(&nmp->nm_lock); + } + if ((np->n_openflags & N_DELEG_MASK) && !force) { + /* try to return the delegation */ + np->n_openflags &= ~N_DELEG_MASK; + nfs4_delegreturn_rpc(nmp, np->n_fhp, np->n_fhsize, &np->n_dstateid, + R_RECOVER, vfs_context_thread(ctx), vfs_context_ucred(ctx)); + } + if (np->n_attrdirfh) { + FREE(np->n_attrdirfh, M_TEMP); + np->n_attrdirfh = NULL; + } + } - /* - * For nqnfs, take it off the timer queue as required. - */ - if ((nmp->nm_flag & NFSMNT_NQNFS) && np->n_timer.cqe_next != 0) { - CIRCLEQ_REMOVE(&nmp->nm_timerhead, np, n_timer); + /* clean up file locks */ + TAILQ_FOREACH_SAFE(nflp, &np->n_locks, nfl_link, nextnflp) { + if (!(nflp->nfl_flags & NFS_FILE_LOCK_DEAD) && !force) { + NP(np, "nfs_vnop_reclaim: lock 0x%llx 0x%llx 0x%x (bc %d)", + nflp->nfl_start, nflp->nfl_end, nflp->nfl_flags, nflp->nfl_blockcnt); + } + if (!(nflp->nfl_flags & (NFS_FILE_LOCK_BLOCKED|NFS_FILE_LOCK_DEAD))) { + /* try sending an unlock RPC if it wasn't delegated */ + if (!(nflp->nfl_flags & NFS_FILE_LOCK_DELEGATED) && !force) + nmp->nm_funcs->nf_unlock_rpc(np, nflp->nfl_owner, F_WRLCK, nflp->nfl_start, nflp->nfl_end, R_RECOVER, + NULL, nflp->nfl_owner->nlo_open_owner->noo_cred); + lck_mtx_lock(&nflp->nfl_owner->nlo_lock); + TAILQ_REMOVE(&nflp->nfl_owner->nlo_locks, nflp, nfl_lolink); + lck_mtx_unlock(&nflp->nfl_owner->nlo_lock); + } + TAILQ_REMOVE(&np->n_locks, nflp, nfl_link); + nfs_file_lock_destroy(nflp); + } + /* clean up lock owners */ + TAILQ_FOREACH_SAFE(nlop, &np->n_lock_owners, nlo_link, nextnlop) { + if (!TAILQ_EMPTY(&nlop->nlo_locks) && !force) + NP(np, "nfs_vnop_reclaim: lock owner with locks"); + TAILQ_REMOVE(&np->n_lock_owners, nlop, nlo_link); + nfs_lock_owner_destroy(nlop); + } + /* clean up open state */ + if (np->n_openrefcnt && !force) + NP(np, "nfs_vnop_reclaim: still open: %d", np->n_openrefcnt); + TAILQ_FOREACH_SAFE(nofp, &np->n_opens, nof_link, nextnofp) { + if (nofp->nof_flags & NFS_OPEN_FILE_BUSY) + NP(np, "nfs_vnop_reclaim: open file busy"); + if (!(np->n_flag & NREVOKE) && !(nofp->nof_flags & NFS_OPEN_FILE_LOST)) { + if (nofp->nof_opencnt && !force) + NP(np, "nfs_vnop_reclaim: file still open: %d", nofp->nof_opencnt); + if (!force && (nofp->nof_access || nofp->nof_deny || + nofp->nof_mmap_access || nofp->nof_mmap_deny || + nofp->nof_r || nofp->nof_w || nofp->nof_rw || + nofp->nof_r_dw || nofp->nof_w_dw || nofp->nof_rw_dw || + nofp->nof_r_drw || nofp->nof_w_drw || nofp->nof_rw_drw || + nofp->nof_d_r || nofp->nof_d_w || nofp->nof_d_rw || + nofp->nof_d_r_dw || nofp->nof_d_w_dw || nofp->nof_d_rw_dw || + nofp->nof_d_r_drw || nofp->nof_d_w_drw || nofp->nof_d_rw_drw)) { + NP(np, "nfs_vnop_reclaim: non-zero access: %d %d %d %d # %u.%u %u.%u %u.%u dw %u.%u %u.%u %u.%u drw %u.%u %u.%u %u.%u", + nofp->nof_access, nofp->nof_deny, + nofp->nof_mmap_access, nofp->nof_mmap_deny, + nofp->nof_r, nofp->nof_d_r, + nofp->nof_w, nofp->nof_d_w, + nofp->nof_rw, nofp->nof_d_rw, + nofp->nof_r_dw, nofp->nof_d_r_dw, + nofp->nof_w_dw, nofp->nof_d_w_dw, + nofp->nof_rw_dw, nofp->nof_d_rw_dw, + nofp->nof_r_drw, nofp->nof_d_r_drw, + nofp->nof_w_drw, nofp->nof_d_w_drw, + nofp->nof_rw_drw, nofp->nof_d_rw_drw); + /* try sending a close RPC if it wasn't delegated */ + if (nofp->nof_r || nofp->nof_w || nofp->nof_rw || + nofp->nof_r_dw || nofp->nof_w_dw || nofp->nof_rw_dw || + nofp->nof_r_drw || nofp->nof_w_drw || nofp->nof_rw_drw) + nfs4_close_rpc(np, nofp, NULL, nofp->nof_owner->noo_cred, R_RECOVER); + } + } + TAILQ_REMOVE(&np->n_opens, nofp, nof_link); + nfs_open_file_destroy(nofp); + } + lck_mtx_unlock(&np->n_openlock); + + if (np->n_monlink.le_next != NFSNOLIST) { + /* Wait for any in-progress getattr to complete, */ + /* then remove this node from the monitored node list. */ + lck_mtx_lock(&nmp->nm_lock); + while (np->n_mflag & NMMONSCANINPROG) { + struct timespec ts = { 1, 0 }; + np->n_mflag |= NMMONSCANWANT; + msleep(&np->n_mflag, &nmp->nm_lock, PZERO-1, "nfswaitmonscan", &ts); + } + if (np->n_monlink.le_next != NFSNOLIST) { + LIST_REMOVE(np, n_monlink); + np->n_monlink.le_next = NFSNOLIST; + } + lck_mtx_unlock(&nmp->nm_lock); } + lck_mtx_lock(nfs_buf_mutex); + if (!force && (!LIST_EMPTY(&np->n_dirtyblkhd) || !LIST_EMPTY(&np->n_cleanblkhd))) + NP(np, "nfs_reclaim: dropping %s buffers", (!LIST_EMPTY(&np->n_dirtyblkhd) ? "dirty" : "clean")); + lck_mtx_unlock(nfs_buf_mutex); + nfs_vinvalbuf(vp, V_IGNORE_WRITEERR, ap->a_context, 0); + + lck_mtx_lock(nfs_node_hash_mutex); + + if ((vnode_vtype(vp) != VDIR) && np->n_sillyrename) { + if (!force) + NP(np, "nfs_reclaim: leaving unlinked file %s", np->n_sillyrename->nsr_name); + if (np->n_sillyrename->nsr_cred != NOCRED) + kauth_cred_unref(&np->n_sillyrename->nsr_cred); + vnode_rele(NFSTOV(np->n_sillyrename->nsr_dnp)); + FREE_ZONE(np->n_sillyrename, sizeof(*np->n_sillyrename), M_NFSREQ); + } + + vnode_removefsref(vp); + + if (np->n_hflag & NHHASHED) { + LIST_REMOVE(np, n_hash); + np->n_hflag &= ~NHHASHED; + FSDBG(266, 0, np, np->n_flag, 0xb1eb1e); + } + lck_mtx_unlock(nfs_node_hash_mutex); + /* - * Free up any directory cookie structures and - * large file handle structures that might be associated with - * this nfs node. + * Free up any directory cookie structures and large file handle + * structures that might be associated with this nfs node. */ - if (vp->v_type == VDIR) { - dp = np->n_cookies.lh_first; - while (dp) { - dp2 = dp; - dp = dp->ndm_list.le_next; - FREE_ZONE((caddr_t)dp2, - sizeof (struct nfsdmap), M_NFSDIROFF); + nfs_node_lock_force(np); + if ((vnode_vtype(vp) == VDIR) && np->n_cookiecache) + FREE_ZONE(np->n_cookiecache, sizeof(struct nfsdmap), M_NFSDIROFF); + if (np->n_fhsize > NFS_SMALLFH) + FREE_ZONE(np->n_fhp, np->n_fhsize, M_NFSBIGFH); + if (np->n_vattr.nva_acl) + kauth_acl_free(np->n_vattr.nva_acl); + nfs_node_unlock(np); + vnode_clearfsnode(vp); + + if (np->n_parent) { + if (!vnode_get(np->n_parent)) { + vnode_rele(np->n_parent); + vnode_put(np->n_parent); } + np->n_parent = NULL; } - if (np->n_fhsize > NFS_SMALLFH) { - FREE_ZONE((caddr_t)np->n_fhp, np->n_fhsize, M_NFSBIGFH); - } - FREE_ZONE(vp->v_data, sizeof (struct nfsnode), M_NFSNODE); - vp->v_data = (void *)0; + lck_mtx_destroy(&np->n_lock, nfs_node_lck_grp); + lck_rw_destroy(&np->n_datalock, nfs_data_lck_grp); + lck_mtx_destroy(&np->n_openlock, nfs_open_grp); + + FSDBG_BOT(265, vp, np, np->n_flag, 0xd1ed1e); + FREE_ZONE(np, sizeof(struct nfsnode), M_NFSNODE); return (0); } /* - * Lock an nfsnode + * Acquire an NFS node lock */ + int -nfs_lock(ap) - struct vop_lock_args /* { - struct vnode *a_vp; - int a_flags; - struct proc *a_p; - } */ *ap; +nfs_node_lock_internal(nfsnode_t np, int force) +{ + FSDBG_TOP(268, np, force, 0, 0); + lck_mtx_lock(&np->n_lock); + if (!force && !(np->n_hflag && NHHASHED)) { + FSDBG_BOT(268, np, 0xdead, 0, 0); + lck_mtx_unlock(&np->n_lock); + return (ENOENT); + } + FSDBG_BOT(268, np, force, 0, 0); + return (0); +} + +int +nfs_node_lock(nfsnode_t np) { - register struct vnode *vp = ap->a_vp; + return nfs_node_lock_internal(np, 0); +} - /* - * Ugh, another place where interruptible mounts will get hung. - * If you make this call interruptible, then you have to fix all - * the VOP_LOCK() calls to expect interruptibility. - */ - if (vp->v_tag == VT_NON) - return (ENOENT); /* ??? -- got to check something and error, but what? */ - - return(lockmgr(&VTONFS(vp)->n_lock, ap->a_flags, &vp->v_interlock, - ap->a_p)); - +void +nfs_node_lock_force(nfsnode_t np) +{ + nfs_node_lock_internal(np, 1); +} + +/* + * Release an NFS node lock + */ +void +nfs_node_unlock(nfsnode_t np) +{ + FSDBG(269, np, current_thread(), 0, 0); + lck_mtx_unlock(&np->n_lock); } /* - * Unlock an nfsnode + * Acquire 2 NFS node locks + * - locks taken in reverse address order + * - both or neither of the locks are taken + * - only one lock taken per node (dup nodes are skipped) */ int -nfs_unlock(ap) - struct vop_unlock_args /* { - struct vnode *a_vp; - int a_flags; - struct proc *a_p; - } */ *ap; +nfs_node_lock2(nfsnode_t np1, nfsnode_t np2) { - struct vnode *vp = ap->a_vp; + nfsnode_t first, second; + int error; + + first = (np1 > np2) ? np1 : np2; + second = (np1 > np2) ? np2 : np1; + if ((error = nfs_node_lock(first))) + return (error); + if (np1 == np2) + return (error); + if ((error = nfs_node_lock(second))) + nfs_node_unlock(first); + return (error); +} - return (lockmgr(&VTONFS(vp)->n_lock, ap->a_flags | LK_RELEASE, - &vp->v_interlock, ap->a_p)); +void +nfs_node_unlock2(nfsnode_t np1, nfsnode_t np2) +{ + nfs_node_unlock(np1); + if (np1 != np2) + nfs_node_unlock(np2); } /* - * Check for a locked nfsnode + * Manage NFS node busy state. + * (Similar to NFS node locks above) */ int -nfs_islocked(ap) - struct vop_islocked_args /* { - struct vnode *a_vp; - } */ *ap; +nfs_node_set_busy(nfsnode_t np, thread_t thd) +{ + struct timespec ts = { 2, 0 }; + int error; + + if ((error = nfs_node_lock(np))) + return (error); + while (ISSET(np->n_flag, NBUSY)) { + SET(np->n_flag, NBUSYWANT); + msleep(np, &np->n_lock, PZERO-1, "nfsbusywant", &ts); + if ((error = nfs_sigintr(NFSTONMP(np), NULL, thd, 0))) + break; + } + if (!error) + SET(np->n_flag, NBUSY); + nfs_node_unlock(np); + return (error); +} + +void +nfs_node_clear_busy(nfsnode_t np) { - return (lockstatus(&VTONFS(ap->a_vp)->n_lock)); + int wanted; + nfs_node_lock_force(np); + wanted = ISSET(np->n_flag, NBUSYWANT); + CLR(np->n_flag, NBUSY|NBUSYWANT); + nfs_node_unlock(np); + if (wanted) + wakeup(np); } +int +nfs_node_set_busy2(nfsnode_t np1, nfsnode_t np2, thread_t thd) +{ + nfsnode_t first, second; + int error; + + first = (np1 > np2) ? np1 : np2; + second = (np1 > np2) ? np2 : np1; + if ((error = nfs_node_set_busy(first, thd))) + return (error); + if (np1 == np2) + return (error); + if ((error = nfs_node_set_busy(second, thd))) + nfs_node_clear_busy(first); + return (error); +} + +void +nfs_node_clear_busy2(nfsnode_t np1, nfsnode_t np2) +{ + nfs_node_clear_busy(np1); + if (np1 != np2) + nfs_node_clear_busy(np2); +} + +/* helper function to sort four nodes in reverse address order (no dupes) */ +static void +nfs_node_sort4(nfsnode_t np1, nfsnode_t np2, nfsnode_t np3, nfsnode_t np4, nfsnode_t *list, int *lcntp) +{ + nfsnode_t na[2], nb[2]; + int a, b, i, lcnt; + + /* sort pairs then merge */ + na[0] = (np1 > np2) ? np1 : np2; + na[1] = (np1 > np2) ? np2 : np1; + nb[0] = (np3 > np4) ? np3 : np4; + nb[1] = (np3 > np4) ? np4 : np3; + for (a = b = i = lcnt = 0; i < 4; i++) { + if (a >= 2) + list[lcnt] = nb[b++]; + else if ((b >= 2) || (na[a] >= nb[b])) + list[lcnt] = na[a++]; + else + list[lcnt] = nb[b++]; + if ((lcnt <= 0) || (list[lcnt] != list[lcnt-1])) + lcnt++; /* omit dups */ + } + if (list[lcnt-1] == NULL) + lcnt--; + *lcntp = lcnt; +} + +int +nfs_node_set_busy4(nfsnode_t np1, nfsnode_t np2, nfsnode_t np3, nfsnode_t np4, thread_t thd) +{ + nfsnode_t list[4]; + int i, lcnt, error; + + nfs_node_sort4(np1, np2, np3, np4, list, &lcnt); + + /* Now we can lock using list[0 - lcnt-1] */ + for (i = 0; i < lcnt; ++i) + if ((error = nfs_node_set_busy(list[i], thd))) { + /* Drop any locks we acquired. */ + while (--i >= 0) + nfs_node_clear_busy(list[i]); + return (error); + } + return (0); +} + +void +nfs_node_clear_busy4(nfsnode_t np1, nfsnode_t np2, nfsnode_t np3, nfsnode_t np4) +{ + nfsnode_t list[4]; + int lcnt; + + nfs_node_sort4(np1, np2, np3, np4, list, &lcnt); + while (--lcnt >= 0) + nfs_node_clear_busy(list[lcnt]); +} /* - * Nfs abort op, called after namei() when a CREATE/DELETE isn't actually - * done. Currently nothing to do. + * Acquire an NFS node data lock */ -/* ARGSUSED */ +void +nfs_data_lock(nfsnode_t np, int locktype) +{ + nfs_data_lock_internal(np, locktype, 1); +} +void +nfs_data_lock_noupdate(nfsnode_t np, int locktype) +{ + nfs_data_lock_internal(np, locktype, 0); +} +void +nfs_data_lock_internal(nfsnode_t np, int locktype, int updatesize) +{ + FSDBG_TOP(270, np, locktype, np->n_datalockowner, 0); + if (locktype == NFS_DATA_LOCK_SHARED) { + if (updatesize && ISSET(np->n_flag, NUPDATESIZE)) + nfs_data_update_size(np, 0); + lck_rw_lock_shared(&np->n_datalock); + } else { + lck_rw_lock_exclusive(&np->n_datalock); + np->n_datalockowner = current_thread(); + if (updatesize && ISSET(np->n_flag, NUPDATESIZE)) + nfs_data_update_size(np, 1); + } + FSDBG_BOT(270, np, locktype, np->n_datalockowner, 0); +} + +/* + * Release an NFS node data lock + */ +void +nfs_data_unlock(nfsnode_t np) +{ + nfs_data_unlock_internal(np, 1); +} +void +nfs_data_unlock_noupdate(nfsnode_t np) +{ + nfs_data_unlock_internal(np, 0); +} +void +nfs_data_unlock_internal(nfsnode_t np, int updatesize) +{ + int mine = (np->n_datalockowner == current_thread()); + FSDBG_TOP(271, np, np->n_datalockowner, current_thread(), 0); + if (updatesize && mine && ISSET(np->n_flag, NUPDATESIZE)) + nfs_data_update_size(np, 1); + np->n_datalockowner = NULL; + lck_rw_done(&np->n_datalock); + if (updatesize && !mine && ISSET(np->n_flag, NUPDATESIZE)) + nfs_data_update_size(np, 0); + FSDBG_BOT(271, np, np->n_datalockowner, current_thread(), 0); +} + + +/* + * update an NFS node's size + */ +void +nfs_data_update_size(nfsnode_t np, int datalocked) +{ + int error; + + FSDBG_TOP(272, np, np->n_flag, np->n_size, np->n_newsize); + if (!datalocked) { + nfs_data_lock(np, NFS_DATA_LOCK_EXCLUSIVE); + /* grabbing data lock will automatically update size */ + nfs_data_unlock(np); + FSDBG_BOT(272, np, np->n_flag, np->n_size, np->n_newsize); + return; + } + error = nfs_node_lock(np); + if (error || !ISSET(np->n_flag, NUPDATESIZE)) { + if (!error) + nfs_node_unlock(np); + FSDBG_BOT(272, np, np->n_flag, np->n_size, np->n_newsize); + return; + } + CLR(np->n_flag, NUPDATESIZE); + np->n_size = np->n_newsize; + /* make sure we invalidate buffers the next chance we get */ + SET(np->n_flag, NNEEDINVALIDATE); + nfs_node_unlock(np); + ubc_setsize(NFSTOV(np), (off_t)np->n_size); /* XXX error? */ + FSDBG_BOT(272, np, np->n_flag, np->n_size, np->n_newsize); +} + +#define DODEBUG 1 int -nfs_abortop(ap) - struct vop_abortop_args /* { - struct vnode *a_dvp; - struct componentname *a_cnp; - } */ *ap; +nfs_mount_is_dirty(mount_t mp) { + u_long i; + nfsnode_t np; +#ifdef DODEBUG + struct timeval now, then, diff; + u_long ncnt = 0; + microuptime(&now); +#endif + lck_mtx_lock(nfs_node_hash_mutex); + for (i = 0; i <= nfsnodehash; i++) { + LIST_FOREACH(np, &nfsnodehashtbl[i], n_hash) { +#ifdef DODEBUG + ncnt++; +#endif + if (np->n_mount == mp && !LIST_EMPTY(&np->n_dirtyblkhd)) + goto out; + } + } +out: + lck_mtx_unlock(nfs_node_hash_mutex); +#ifdef DODEBUG + microuptime(&then); + timersub(&then, &now, &diff); + + printf("nfs_mount_is_dirty took %lld mics for %ld slots and %ld nodes return %d\n", + (uint64_t)diff.tv_sec * 1000000LL + diff.tv_usec, i, ncnt, (i <= nfsnodehash)); +#endif - if ((ap->a_cnp->cn_flags & (HASBUF | SAVESTART)) == HASBUF) - FREE_ZONE(ap->a_cnp->cn_pnbuf, ap->a_cnp->cn_pnlen, M_NAMEI); - return (0); + return (i <= nfsnodehash); }