95270f7c84
vnode EXCLUSIVE lock. This prevents threads from adding pages to the vnode while an invalidation is in progress, closing potential races. In the bioread() path, callers acquire the SHARED vnode lock - so while an invalidate was in progress, it was possible to fault in new pages onto the vnode causing the invalidation to take a while or fail. We saw these races at Yahoo! with very large files+heavy concurrent access. Forcing an upgrade to EXCLUSIVE lock before doing the invalidation closes all these races. Submitted by: Mohan Srinivasan mohans at yahoo-inc dot com
361 lines
9.2 KiB
C
361 lines
9.2 KiB
C
/*
|
|
* Copyright (c) 1989, 1993
|
|
* The Regents of the University of California. All rights reserved.
|
|
*
|
|
* This code is derived from software contributed to Berkeley by
|
|
* Rick Macklem at The University of Guelph.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
* 4. Neither the name of the University nor the names of its contributors
|
|
* may be used to endorse or promote products derived from this software
|
|
* without specific prior written permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
|
|
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
|
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
|
|
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
|
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
|
|
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
|
* SUCH DAMAGE.
|
|
*
|
|
* @(#)nfs_node.c 8.6 (Berkeley) 5/22/95
|
|
*/
|
|
|
|
#include <sys/cdefs.h>
|
|
__FBSDID("$FreeBSD$");
|
|
|
|
#include <sys/param.h>
|
|
#include <sys/systm.h>
|
|
#include <sys/fnv_hash.h>
|
|
#include <sys/lock.h>
|
|
#include <sys/malloc.h>
|
|
#include <sys/mount.h>
|
|
#include <sys/namei.h>
|
|
#include <sys/proc.h>
|
|
#include <sys/socket.h>
|
|
#include <sys/sysctl.h>
|
|
#include <sys/vnode.h>
|
|
|
|
#include <vm/uma.h>
|
|
|
|
#include <rpc/rpcclnt.h>
|
|
|
|
#include <nfs/rpcv2.h>
|
|
#include <nfs/nfsproto.h>
|
|
#include <nfsclient/nfs.h>
|
|
#include <nfsclient/nfsnode.h>
|
|
#include <nfsclient/nfsmount.h>
|
|
|
|
static uma_zone_t nfsnode_zone;
|
|
static LIST_HEAD(nfsnodehashhead, nfsnode) *nfsnodehashtbl;
|
|
static u_long nfsnodehash;
|
|
static int nfs_node_hash_lock;
|
|
|
|
#define NFSNOHASH(fhsum) \
|
|
(&nfsnodehashtbl[(fhsum) & nfsnodehash])
|
|
|
|
#define TRUE 1
|
|
#define FALSE 0
|
|
|
|
SYSCTL_DECL(_debug_hashstat);
|
|
|
|
/*
|
|
* Grab an atomic snapshot of the nfsnode hash chain lengths
|
|
*/
|
|
static int
|
|
sysctl_debug_hashstat_rawnfsnode(SYSCTL_HANDLER_ARGS)
|
|
{
|
|
int error;
|
|
struct nfsnodehashhead *nnpp;
|
|
struct nfsnode *nnp;
|
|
int n_nfsnode;
|
|
int count;
|
|
|
|
n_nfsnode = nfsnodehash + 1; /* nfsnodehash = max index, not count */
|
|
if (!req->oldptr)
|
|
return SYSCTL_OUT(req, 0, n_nfsnode * sizeof(int));
|
|
|
|
/* Scan hash tables for applicable entries */
|
|
for (nnpp = nfsnodehashtbl; n_nfsnode > 0; n_nfsnode--, nnpp++) {
|
|
count = 0;
|
|
LIST_FOREACH(nnp, nnpp, n_hash) {
|
|
count++;
|
|
}
|
|
error = SYSCTL_OUT(req, (caddr_t)&count, sizeof(count));
|
|
if (error)
|
|
return (error);
|
|
}
|
|
return (0);
|
|
}
|
|
SYSCTL_PROC(_debug_hashstat, OID_AUTO, rawnfsnode, CTLTYPE_INT|CTLFLAG_RD, 0, 0,
|
|
sysctl_debug_hashstat_rawnfsnode, "S,int", "nfsnode chain lengths");
|
|
|
|
static int
|
|
sysctl_debug_hashstat_nfsnode(SYSCTL_HANDLER_ARGS)
|
|
{
|
|
int error;
|
|
struct nfsnodehashhead *nnpp;
|
|
struct nfsnode *nnp;
|
|
int n_nfsnode;
|
|
int count, maxlength, used, pct;
|
|
|
|
if (!req->oldptr)
|
|
return SYSCTL_OUT(req, 0, 4 * sizeof(int));
|
|
|
|
n_nfsnode = nfsnodehash + 1; /* nfsnodehash = max index, not count */
|
|
used = 0;
|
|
maxlength = 0;
|
|
|
|
/* Scan hash tables for applicable entries */
|
|
for (nnpp = nfsnodehashtbl; n_nfsnode > 0; n_nfsnode--, nnpp++) {
|
|
count = 0;
|
|
LIST_FOREACH(nnp, nnpp, n_hash) {
|
|
count++;
|
|
}
|
|
if (count)
|
|
used++;
|
|
if (maxlength < count)
|
|
maxlength = count;
|
|
}
|
|
n_nfsnode = nfsnodehash + 1;
|
|
pct = (used * 100 * 100) / n_nfsnode;
|
|
error = SYSCTL_OUT(req, (caddr_t)&n_nfsnode, sizeof(n_nfsnode));
|
|
if (error)
|
|
return (error);
|
|
error = SYSCTL_OUT(req, (caddr_t)&used, sizeof(used));
|
|
if (error)
|
|
return (error);
|
|
error = SYSCTL_OUT(req, (caddr_t)&maxlength, sizeof(maxlength));
|
|
if (error)
|
|
return (error);
|
|
error = SYSCTL_OUT(req, (caddr_t)&pct, sizeof(pct));
|
|
if (error)
|
|
return (error);
|
|
return (0);
|
|
}
|
|
SYSCTL_PROC(_debug_hashstat, OID_AUTO, nfsnode, CTLTYPE_INT|CTLFLAG_RD,
|
|
0, 0, sysctl_debug_hashstat_nfsnode, "I", "nfsnode chain lengths");
|
|
|
|
/*
|
|
* Initialize hash links for nfsnodes
|
|
* and build nfsnode free list.
|
|
*/
|
|
void
|
|
nfs_nhinit(void)
|
|
{
|
|
|
|
nfsnode_zone = uma_zcreate("NFSNODE", sizeof(struct nfsnode), NULL,
|
|
NULL, NULL, NULL, UMA_ALIGN_PTR, 0);
|
|
nfsnodehashtbl = hashinit(desiredvnodes, M_NFSHASH, &nfsnodehash);
|
|
}
|
|
|
|
/*
|
|
* Release hash table resources
|
|
*/
|
|
void
|
|
nfs_nhuninit(void)
|
|
{
|
|
hashdestroy(nfsnodehashtbl, M_NFSHASH, nfsnodehash);
|
|
uma_zdestroy(nfsnode_zone);
|
|
}
|
|
|
|
/*
|
|
* Look up a vnode/nfsnode by file handle.
|
|
* Callers must check for mount points!!
|
|
* In all cases, a pointer to a
|
|
* nfsnode structure is returned.
|
|
*/
|
|
int
|
|
nfs_nget(struct mount *mntp, nfsfh_t *fhp, int fhsize, struct nfsnode **npp)
|
|
{
|
|
struct thread *td = curthread; /* XXX */
|
|
struct nfsnode *np, *np2;
|
|
struct nfsnodehashhead *nhpp;
|
|
struct vnode *vp;
|
|
struct vnode *nvp;
|
|
int error;
|
|
int rsflags;
|
|
struct nfsmount *nmp;
|
|
|
|
/*
|
|
* Calculate nfs mount point and figure out whether the rslock should
|
|
* be interruptible or not.
|
|
*/
|
|
nmp = VFSTONFS(mntp);
|
|
if (nmp->nm_flag & NFSMNT_INT)
|
|
rsflags = PCATCH;
|
|
else
|
|
rsflags = 0;
|
|
|
|
retry:
|
|
nhpp = NFSNOHASH(fnv_32_buf(fhp->fh_bytes, fhsize, FNV1_32_INIT));
|
|
loop:
|
|
LIST_FOREACH(np, nhpp, n_hash) {
|
|
if (mntp != NFSTOV(np)->v_mount || np->n_fhsize != fhsize ||
|
|
bcmp((caddr_t)fhp, (caddr_t)np->n_fhp, fhsize))
|
|
continue;
|
|
vp = NFSTOV(np);
|
|
/*
|
|
* np or vp may become invalid if vget() blocks, so loop
|
|
*/
|
|
if (vget(vp, LK_EXCLUSIVE|LK_SLEEPFAIL, td))
|
|
goto loop;
|
|
*npp = np;
|
|
return(0);
|
|
}
|
|
/*
|
|
* Obtain a lock to prevent a race condition if the getnewvnode()
|
|
* or MALLOC() below happens to block.
|
|
*/
|
|
if (nfs_node_hash_lock) {
|
|
while (nfs_node_hash_lock) {
|
|
nfs_node_hash_lock = -1;
|
|
tsleep(&nfs_node_hash_lock, PVM, "nfsngt", 0);
|
|
}
|
|
goto loop;
|
|
}
|
|
nfs_node_hash_lock = 1;
|
|
|
|
/*
|
|
* Allocate before getnewvnode since doing so afterward
|
|
* might cause a bogus v_data pointer to get dereferenced
|
|
* elsewhere if zalloc should block.
|
|
*/
|
|
np = uma_zalloc(nfsnode_zone, M_WAITOK);
|
|
|
|
if (nmp->nm_flag & NFSMNT_NFSV4)
|
|
error = getnewvnode("nfs4", mntp, &nfs4_vnodeops, &nvp);
|
|
else
|
|
error = getnewvnode("nfs", mntp, &nfs_vnodeops, &nvp);
|
|
if (error) {
|
|
if (nfs_node_hash_lock < 0)
|
|
wakeup(&nfs_node_hash_lock);
|
|
nfs_node_hash_lock = 0;
|
|
*npp = 0;
|
|
uma_zfree(nfsnode_zone, np);
|
|
return (error);
|
|
}
|
|
vp = nvp;
|
|
if (nmp->nm_flag & NFSMNT_NFSV4)
|
|
vp->v_bufobj.bo_ops = &buf_ops_nfs4;
|
|
else
|
|
vp->v_bufobj.bo_ops = &buf_ops_nfs;
|
|
bzero((caddr_t)np, sizeof *np);
|
|
vp->v_data = np;
|
|
np->n_vnode = vp;
|
|
/*
|
|
* Insert the nfsnode in the hash queue for its new file handle
|
|
*/
|
|
LIST_FOREACH(np2, nhpp, n_hash) {
|
|
if (mntp != NFSTOV(np2)->v_mount || np2->n_fhsize != fhsize ||
|
|
bcmp((caddr_t)fhp, (caddr_t)np2->n_fhp, fhsize))
|
|
continue;
|
|
vrele(vp);
|
|
if (nfs_node_hash_lock < 0)
|
|
wakeup(&nfs_node_hash_lock);
|
|
nfs_node_hash_lock = 0;
|
|
uma_zfree(nfsnode_zone, np);
|
|
goto retry;
|
|
}
|
|
LIST_INSERT_HEAD(nhpp, np, n_hash);
|
|
if (fhsize > NFS_SMALLFH) {
|
|
MALLOC(np->n_fhp, nfsfh_t *, fhsize, M_NFSBIGFH, M_WAITOK);
|
|
} else
|
|
np->n_fhp = &np->n_fh;
|
|
bcopy((caddr_t)fhp, (caddr_t)np->n_fhp, fhsize);
|
|
np->n_fhsize = fhsize;
|
|
lockinit(&np->n_rslock, PVFS | rsflags, "nfrslk", 0, LK_NOPAUSE);
|
|
*npp = np;
|
|
|
|
if (nfs_node_hash_lock < 0)
|
|
wakeup(&nfs_node_hash_lock);
|
|
nfs_node_hash_lock = 0;
|
|
|
|
/*
|
|
* Lock the new nfsnode.
|
|
*/
|
|
vn_lock(vp, LK_EXCLUSIVE | LK_RETRY, td);
|
|
|
|
return (0);
|
|
}
|
|
|
|
int
|
|
nfs_inactive(struct vop_inactive_args *ap)
|
|
{
|
|
struct nfsnode *np;
|
|
struct sillyrename *sp;
|
|
struct thread *td = curthread; /* XXX */
|
|
|
|
np = VTONFS(ap->a_vp);
|
|
if (prtactive && vrefcnt(ap->a_vp) != 0)
|
|
vprint("nfs_inactive: pushing active", ap->a_vp);
|
|
if (ap->a_vp->v_type != VDIR) {
|
|
sp = np->n_sillyrename;
|
|
np->n_sillyrename = NULL;
|
|
} else
|
|
sp = NULL;
|
|
if (sp) {
|
|
(void)nfs_vinvalbuf(ap->a_vp, 0, sp->s_cred, td, 1);
|
|
/*
|
|
* Remove the silly file that was rename'd earlier
|
|
*/
|
|
(sp->s_removeit)(sp);
|
|
crfree(sp->s_cred);
|
|
vrele(sp->s_dvp);
|
|
FREE((caddr_t)sp, M_NFSREQ);
|
|
}
|
|
np->n_flag &= NMODIFIED;
|
|
VOP_UNLOCK(ap->a_vp, 0, ap->a_td);
|
|
return (0);
|
|
}
|
|
|
|
/*
|
|
* Reclaim an nfsnode so that it can be used for other purposes.
|
|
*/
|
|
int
|
|
nfs_reclaim(struct vop_reclaim_args *ap)
|
|
{
|
|
struct vnode *vp = ap->a_vp;
|
|
struct nfsnode *np = VTONFS(vp);
|
|
struct nfsdmap *dp, *dp2;
|
|
|
|
if (prtactive && vrefcnt(vp) != 0)
|
|
vprint("nfs_reclaim: pushing active", vp);
|
|
|
|
if (np->n_hash.le_prev != NULL) /* XXX beware */
|
|
LIST_REMOVE(np, n_hash);
|
|
|
|
/*
|
|
* Free up any directory cookie structures and
|
|
* large file handle structures that might be associated with
|
|
* this nfs node.
|
|
*/
|
|
if (vp->v_type == VDIR) {
|
|
dp = LIST_FIRST(&np->n_cookies);
|
|
while (dp) {
|
|
dp2 = dp;
|
|
dp = LIST_NEXT(dp, ndm_list);
|
|
FREE((caddr_t)dp2, M_NFSDIROFF);
|
|
}
|
|
}
|
|
if (np->n_fhsize > NFS_SMALLFH) {
|
|
FREE((caddr_t)np->n_fhp, M_NFSBIGFH);
|
|
}
|
|
|
|
lockdestroy(&np->n_rslock);
|
|
uma_zfree(nfsnode_zone, vp->v_data);
|
|
vp->v_data = NULL;
|
|
return (0);
|
|
}
|