1994-05-24 10:09:53 +00:00
|
|
|
/*
|
1997-02-10 02:22:35 +00:00
|
|
|
* Copyright (c) 1989, 1991, 1993, 1995
|
1994-05-24 10:09:53 +00:00
|
|
|
* The Regents of the University of California. All rights reserved.
|
|
|
|
*
|
|
|
|
* This code is derived from software contributed to Berkeley by
|
|
|
|
* Rick Macklem at The University of Guelph.
|
|
|
|
*
|
|
|
|
* Redistribution and use in source and binary forms, with or without
|
|
|
|
* modification, are permitted provided that the following conditions
|
|
|
|
* are met:
|
|
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
|
|
* notice, this list of conditions and the following disclaimer.
|
|
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
|
|
* documentation and/or other materials provided with the distribution.
|
|
|
|
* 3. All advertising materials mentioning features or use of this software
|
|
|
|
* must display the following acknowledgement:
|
|
|
|
* This product includes software developed by the University of
|
|
|
|
* California, Berkeley and its contributors.
|
|
|
|
* 4. Neither the name of the University nor the names of its contributors
|
|
|
|
* may be used to endorse or promote products derived from this software
|
|
|
|
* without specific prior written permission.
|
|
|
|
*
|
|
|
|
* THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
|
|
|
|
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
|
|
|
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
|
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
|
|
|
|
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
|
|
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
|
|
|
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
|
|
|
|
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
|
|
|
* SUCH DAMAGE.
|
|
|
|
*
|
1997-02-10 02:22:35 +00:00
|
|
|
* @(#)nfs_socket.c 8.5 (Berkeley) 3/30/95
|
1994-05-24 10:09:53 +00:00
|
|
|
*/
|
|
|
|
|
2001-09-18 23:32:09 +00:00
|
|
|
#include <sys/cdefs.h>
|
|
|
|
__FBSDID("$FreeBSD$");
|
|
|
|
|
1994-05-24 10:09:53 +00:00
|
|
|
/*
|
|
|
|
* Socket operations for use by nfs
|
|
|
|
*/
|
|
|
|
|
2002-07-15 19:40:23 +00:00
|
|
|
#include "opt_inet6.h"
|
|
|
|
|
1994-05-24 10:09:53 +00:00
|
|
|
#include <sys/param.h>
|
|
|
|
#include <sys/systm.h>
|
|
|
|
#include <sys/kernel.h>
|
2001-05-01 08:13:21 +00:00
|
|
|
#include <sys/lock.h>
|
|
|
|
#include <sys/malloc.h>
|
1994-05-24 10:09:53 +00:00
|
|
|
#include <sys/mbuf.h>
|
2001-05-01 08:13:21 +00:00
|
|
|
#include <sys/mount.h>
|
|
|
|
#include <sys/mutex.h>
|
|
|
|
#include <sys/proc.h>
|
1994-05-24 10:09:53 +00:00
|
|
|
#include <sys/protosw.h>
|
2001-05-01 08:13:21 +00:00
|
|
|
#include <sys/signalvar.h>
|
1994-05-24 10:09:53 +00:00
|
|
|
#include <sys/socket.h>
|
|
|
|
#include <sys/socketvar.h>
|
The VFS/BIO subsystem contained a number of hacks in order to optimize
piecemeal, middle-of-file writes for NFS. These hacks have caused no
end of trouble, especially when combined with mmap(). I've removed
them. Instead, NFS will issue a read-before-write to fully
instantiate the struct buf containing the write. NFS does, however,
optimize piecemeal appends to files. For most common file operations,
you will not notice the difference. The sole remaining fragment in
the VFS/BIO system is b_dirtyoff/end, which NFS uses to avoid cache
coherency issues with read-merge-write style operations. NFS also
optimizes the write-covers-entire-buffer case by avoiding the
read-before-write. There is quite a bit of room for further
optimization in these areas.
The VM system marks pages fully-valid (AKA vm_page_t->valid =
VM_PAGE_BITS_ALL) in several places, most noteably in vm_fault. This
is not correct operation. The vm_pager_get_pages() code is now
responsible for marking VM pages all-valid. A number of VM helper
routines have been added to aid in zeroing-out the invalid portions of
a VM page prior to the page being marked all-valid. This operation is
necessary to properly support mmap(). The zeroing occurs most often
when dealing with file-EOF situations. Several bugs have been fixed
in the NFS subsystem, including bits handling file and directory EOF
situations and buf->b_flags consistancy issues relating to clearing
B_ERROR & B_INVAL, and handling B_DONE.
getblk() and allocbuf() have been rewritten. B_CACHE operation is now
formally defined in comments and more straightforward in
implementation. B_CACHE for VMIO buffers is based on the validity of
the backing store. B_CACHE for non-VMIO buffers is based simply on
whether the buffer is B_INVAL or not (B_CACHE set if B_INVAL clear,
and vise-versa). biodone() is now responsible for setting B_CACHE
when a successful read completes. B_CACHE is also set when a bdwrite()
is initiated and when a bwrite() is initiated. VFS VOP_BWRITE
routines (there are only two - nfs_bwrite() and bwrite()) are now
expected to set B_CACHE. This means that bowrite() and bawrite() also
set B_CACHE indirectly.
There are a number of places in the code which were previously using
buf->b_bufsize (which is DEV_BSIZE aligned) when they should have
been using buf->b_bcount. These have been fixed. getblk() now clears
B_DONE on return because the rest of the system is so bad about
dealing with B_DONE.
Major fixes to NFS/TCP have been made. A server-side bug could cause
requests to be lost by the server due to nfs_realign() overwriting
other rpc's in the same TCP mbuf chain. The server's kernel must be
recompiled to get the benefit of the fixes.
Submitted by: Matthew Dillon <dillon@apollo.backplane.com>
1999-05-02 23:57:16 +00:00
|
|
|
#include <sys/sysctl.h>
|
2001-05-01 08:13:21 +00:00
|
|
|
#include <sys/syslog.h>
|
|
|
|
#include <sys/vnode.h>
|
1994-05-24 10:09:53 +00:00
|
|
|
|
|
|
|
#include <netinet/in.h>
|
|
|
|
#include <netinet/tcp.h>
|
1994-10-17 17:47:45 +00:00
|
|
|
|
2003-11-14 20:54:10 +00:00
|
|
|
#include <rpc/rpcclnt.h>
|
|
|
|
|
1994-05-24 10:09:53 +00:00
|
|
|
#include <nfs/rpcv2.h>
|
1995-06-27 11:07:30 +00:00
|
|
|
#include <nfs/nfsproto.h>
|
2001-09-18 23:32:09 +00:00
|
|
|
#include <nfsclient/nfs.h>
|
1994-05-24 10:09:53 +00:00
|
|
|
#include <nfs/xdr_subs.h>
|
2001-09-18 23:32:09 +00:00
|
|
|
#include <nfsclient/nfsm_subs.h>
|
|
|
|
#include <nfsclient/nfsmount.h>
|
|
|
|
#include <nfsclient/nfsnode.h>
|
1994-05-24 10:09:53 +00:00
|
|
|
|
2003-11-14 20:54:10 +00:00
|
|
|
#include <nfs4client/nfs4.h>
|
|
|
|
|
1994-05-24 10:09:53 +00:00
|
|
|
#define TRUE 1
|
|
|
|
#define FALSE 0
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Estimate rto for an nfs rpc sent via. an unreliable datagram.
|
|
|
|
* Use the mean and mean deviation of rtt for the appropriate type of rpc
|
|
|
|
* for the frequent rpcs and a default for the others.
|
|
|
|
* The justification for doing "other" this way is that these rpcs
|
|
|
|
* happen so infrequently that timer est. would probably be stale.
|
|
|
|
* Also, since many of these rpcs are
|
|
|
|
* non-idempotent, a conservative timeout is desired.
|
|
|
|
* getattr, lookup - A+2D
|
|
|
|
* read, write - A+4D
|
|
|
|
* other - nm_timeo
|
|
|
|
*/
|
|
|
|
#define NFS_RTO(n, t) \
|
|
|
|
((t) == 0 ? (n)->nm_timeo : \
|
|
|
|
((t) < 3 ? \
|
|
|
|
(((((n)->nm_srtt[t-1] + 3) >> 2) + (n)->nm_sdrtt[t-1] + 1) >> 1) : \
|
|
|
|
((((n)->nm_srtt[t-1] + 7) >> 3) + (n)->nm_sdrtt[t-1] + 1)))
|
|
|
|
#define NFS_SRTT(r) (r)->r_nmp->nm_srtt[proct[(r)->r_procnum] - 1]
|
|
|
|
#define NFS_SDRTT(r) (r)->r_nmp->nm_sdrtt[proct[(r)->r_procnum] - 1]
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Defines which timer to use for the procnum.
|
|
|
|
* 0 - default
|
|
|
|
* 1 - getattr
|
|
|
|
* 2 - lookup
|
|
|
|
* 3 - read
|
|
|
|
* 4 - write
|
|
|
|
*/
|
|
|
|
static int proct[NFS_NPROCS] = {
|
1995-06-27 11:07:30 +00:00
|
|
|
0, 1, 0, 2, 1, 3, 3, 4, 0, 0, 0, 0, 0, 0, 0, 0, 3, 3, 0, 0, 0, 0, 0,
|
1994-05-24 10:09:53 +00:00
|
|
|
};
|
|
|
|
|
2001-09-18 23:32:09 +00:00
|
|
|
static int nfs_realign_test;
|
|
|
|
static int nfs_realign_count;
|
|
|
|
static int nfs_bufpackets = 4;
|
The VFS/BIO subsystem contained a number of hacks in order to optimize
piecemeal, middle-of-file writes for NFS. These hacks have caused no
end of trouble, especially when combined with mmap(). I've removed
them. Instead, NFS will issue a read-before-write to fully
instantiate the struct buf containing the write. NFS does, however,
optimize piecemeal appends to files. For most common file operations,
you will not notice the difference. The sole remaining fragment in
the VFS/BIO system is b_dirtyoff/end, which NFS uses to avoid cache
coherency issues with read-merge-write style operations. NFS also
optimizes the write-covers-entire-buffer case by avoiding the
read-before-write. There is quite a bit of room for further
optimization in these areas.
The VM system marks pages fully-valid (AKA vm_page_t->valid =
VM_PAGE_BITS_ALL) in several places, most noteably in vm_fault. This
is not correct operation. The vm_pager_get_pages() code is now
responsible for marking VM pages all-valid. A number of VM helper
routines have been added to aid in zeroing-out the invalid portions of
a VM page prior to the page being marked all-valid. This operation is
necessary to properly support mmap(). The zeroing occurs most often
when dealing with file-EOF situations. Several bugs have been fixed
in the NFS subsystem, including bits handling file and directory EOF
situations and buf->b_flags consistancy issues relating to clearing
B_ERROR & B_INVAL, and handling B_DONE.
getblk() and allocbuf() have been rewritten. B_CACHE operation is now
formally defined in comments and more straightforward in
implementation. B_CACHE for VMIO buffers is based on the validity of
the backing store. B_CACHE for non-VMIO buffers is based simply on
whether the buffer is B_INVAL or not (B_CACHE set if B_INVAL clear,
and vise-versa). biodone() is now responsible for setting B_CACHE
when a successful read completes. B_CACHE is also set when a bdwrite()
is initiated and when a bwrite() is initiated. VFS VOP_BWRITE
routines (there are only two - nfs_bwrite() and bwrite()) are now
expected to set B_CACHE. This means that bowrite() and bawrite() also
set B_CACHE indirectly.
There are a number of places in the code which were previously using
buf->b_bufsize (which is DEV_BSIZE aligned) when they should have
been using buf->b_bcount. These have been fixed. getblk() now clears
B_DONE on return because the rest of the system is so bad about
dealing with B_DONE.
Major fixes to NFS/TCP have been made. A server-side bug could cause
requests to be lost by the server due to nfs_realign() overwriting
other rpc's in the same TCP mbuf chain. The server's kernel must be
recompiled to get the benefit of the fixes.
Submitted by: Matthew Dillon <dillon@apollo.backplane.com>
1999-05-02 23:57:16 +00:00
|
|
|
|
|
|
|
SYSCTL_DECL(_vfs_nfs);
|
|
|
|
|
2000-03-27 21:38:35 +00:00
|
|
|
SYSCTL_INT(_vfs_nfs, OID_AUTO, realign_test, CTLFLAG_RW, &nfs_realign_test, 0, "");
|
|
|
|
SYSCTL_INT(_vfs_nfs, OID_AUTO, realign_count, CTLFLAG_RW, &nfs_realign_count, 0, "");
|
|
|
|
SYSCTL_INT(_vfs_nfs, OID_AUTO, bufpackets, CTLFLAG_RW, &nfs_bufpackets, 0, "");
|
The VFS/BIO subsystem contained a number of hacks in order to optimize
piecemeal, middle-of-file writes for NFS. These hacks have caused no
end of trouble, especially when combined with mmap(). I've removed
them. Instead, NFS will issue a read-before-write to fully
instantiate the struct buf containing the write. NFS does, however,
optimize piecemeal appends to files. For most common file operations,
you will not notice the difference. The sole remaining fragment in
the VFS/BIO system is b_dirtyoff/end, which NFS uses to avoid cache
coherency issues with read-merge-write style operations. NFS also
optimizes the write-covers-entire-buffer case by avoiding the
read-before-write. There is quite a bit of room for further
optimization in these areas.
The VM system marks pages fully-valid (AKA vm_page_t->valid =
VM_PAGE_BITS_ALL) in several places, most noteably in vm_fault. This
is not correct operation. The vm_pager_get_pages() code is now
responsible for marking VM pages all-valid. A number of VM helper
routines have been added to aid in zeroing-out the invalid portions of
a VM page prior to the page being marked all-valid. This operation is
necessary to properly support mmap(). The zeroing occurs most often
when dealing with file-EOF situations. Several bugs have been fixed
in the NFS subsystem, including bits handling file and directory EOF
situations and buf->b_flags consistancy issues relating to clearing
B_ERROR & B_INVAL, and handling B_DONE.
getblk() and allocbuf() have been rewritten. B_CACHE operation is now
formally defined in comments and more straightforward in
implementation. B_CACHE for VMIO buffers is based on the validity of
the backing store. B_CACHE for non-VMIO buffers is based simply on
whether the buffer is B_INVAL or not (B_CACHE set if B_INVAL clear,
and vise-versa). biodone() is now responsible for setting B_CACHE
when a successful read completes. B_CACHE is also set when a bdwrite()
is initiated and when a bwrite() is initiated. VFS VOP_BWRITE
routines (there are only two - nfs_bwrite() and bwrite()) are now
expected to set B_CACHE. This means that bowrite() and bawrite() also
set B_CACHE indirectly.
There are a number of places in the code which were previously using
buf->b_bufsize (which is DEV_BSIZE aligned) when they should have
been using buf->b_bcount. These have been fixed. getblk() now clears
B_DONE on return because the rest of the system is so bad about
dealing with B_DONE.
Major fixes to NFS/TCP have been made. A server-side bug could cause
requests to be lost by the server due to nfs_realign() overwriting
other rpc's in the same TCP mbuf chain. The server's kernel must be
recompiled to get the benefit of the fixes.
Submitted by: Matthew Dillon <dillon@apollo.backplane.com>
1999-05-02 23:57:16 +00:00
|
|
|
|
|
|
|
|
1994-05-24 10:09:53 +00:00
|
|
|
/*
|
|
|
|
* There is a congestion window for outstanding rpcs maintained per mount
|
|
|
|
* point. The cwnd size is adjusted in roughly the way that:
|
|
|
|
* Van Jacobson, Congestion avoidance and Control, In "Proceedings of
|
|
|
|
* SIGCOMM '88". ACM, August 1988.
|
|
|
|
* describes for TCP. The cwnd size is chopped in half on a retransmit timeout
|
|
|
|
* and incremented by 1/cwnd when each rpc reply is received and a full cwnd
|
|
|
|
* of rpcs is in progress.
|
|
|
|
* (The sent count and cwnd are scaled for integer arith.)
|
|
|
|
* Variants of "slow start" were tried and were found to be too much of a
|
|
|
|
* performance hit (ave. rtt 3 times larger),
|
|
|
|
* I suspect due to the large rtt that nfs rpcs have.
|
|
|
|
*/
|
|
|
|
#define NFS_CWNDSCALE 256
|
|
|
|
#define NFS_MAXCWND (NFS_CWNDSCALE * 32)
|
2001-12-30 18:41:52 +00:00
|
|
|
#define NFS_NBACKOFF 8
|
|
|
|
static int nfs_backoff[NFS_NBACKOFF] = { 2, 4, 8, 16, 32, 64, 128, 256, };
|
2004-03-25 21:48:09 +00:00
|
|
|
struct callout nfs_callout;
|
1994-05-24 10:09:53 +00:00
|
|
|
|
2001-09-18 23:32:09 +00:00
|
|
|
static int nfs_msg(struct thread *, char *, char *);
|
|
|
|
static int nfs_rcvlock(struct nfsreq *);
|
|
|
|
static void nfs_rcvunlock(struct nfsreq *);
|
|
|
|
static void nfs_realign(struct mbuf **pm, int hsiz);
|
|
|
|
static int nfs_receive(struct nfsreq *rep, struct sockaddr **aname,
|
|
|
|
struct mbuf **mp);
|
|
|
|
static int nfs_reply(struct nfsreq *);
|
|
|
|
static void nfs_softterm(struct nfsreq *rep);
|
|
|
|
static int nfs_reconnect(struct nfsreq *rep);
|
1994-05-24 10:09:53 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Initialize sockets and congestion for a new NFS connection.
|
|
|
|
* We do not free the sockaddr if error.
|
|
|
|
*/
|
1994-05-25 09:21:21 +00:00
|
|
|
int
|
2001-09-18 23:32:09 +00:00
|
|
|
nfs_connect(struct nfsmount *nmp, struct nfsreq *rep)
|
1994-05-24 10:09:53 +00:00
|
|
|
{
|
2001-09-18 23:32:09 +00:00
|
|
|
struct socket *so;
|
1994-05-24 10:09:53 +00:00
|
|
|
int s, error, rcvreserve, sndreserve;
|
2000-03-27 21:38:35 +00:00
|
|
|
int pktscale;
|
1994-05-24 10:09:53 +00:00
|
|
|
struct sockaddr *saddr;
|
2002-02-07 20:58:47 +00:00
|
|
|
struct thread *td = &thread0; /* only used for socreate and sobind */
|
1994-05-24 10:09:53 +00:00
|
|
|
|
2003-11-07 22:57:09 +00:00
|
|
|
GIANT_REQUIRED; /* XXX until socket locking done */
|
|
|
|
|
2002-07-11 17:54:58 +00:00
|
|
|
nmp->nm_so = NULL;
|
1997-08-16 19:16:27 +00:00
|
|
|
saddr = nmp->nm_nam;
|
1995-05-30 08:16:23 +00:00
|
|
|
error = socreate(saddr->sa_family, &nmp->nm_so, nmp->nm_sotype,
|
2002-09-08 15:11:18 +00:00
|
|
|
nmp->nm_soproto, nmp->nm_mountp->mnt_cred, td);
|
1994-10-02 17:27:07 +00:00
|
|
|
if (error)
|
1994-05-24 10:09:53 +00:00
|
|
|
goto bad;
|
|
|
|
so = nmp->nm_so;
|
|
|
|
nmp->nm_soflags = so->so_proto->pr_flags;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Some servers require that the client port be a reserved port number.
|
|
|
|
*/
|
2002-07-15 19:40:23 +00:00
|
|
|
if (nmp->nm_flag & NFSMNT_RESVPORT) {
|
1999-06-05 05:35:03 +00:00
|
|
|
struct sockopt sopt;
|
2002-07-15 19:40:23 +00:00
|
|
|
int ip, ip2, len;
|
|
|
|
struct sockaddr_in6 ssin;
|
|
|
|
struct sockaddr *sa;
|
1999-06-05 05:35:03 +00:00
|
|
|
|
|
|
|
bzero(&sopt, sizeof sopt);
|
2002-07-15 19:40:23 +00:00
|
|
|
switch(saddr->sa_family) {
|
|
|
|
case AF_INET:
|
|
|
|
sopt.sopt_level = IPPROTO_IP;
|
|
|
|
sopt.sopt_name = IP_PORTRANGE;
|
|
|
|
ip = IP_PORTRANGE_LOW;
|
|
|
|
ip2 = IP_PORTRANGE_DEFAULT;
|
|
|
|
len = sizeof (struct sockaddr_in);
|
|
|
|
break;
|
|
|
|
#ifdef INET6
|
|
|
|
case AF_INET6:
|
|
|
|
sopt.sopt_level = IPPROTO_IPV6;
|
|
|
|
sopt.sopt_name = IPV6_PORTRANGE;
|
|
|
|
ip = IPV6_PORTRANGE_LOW;
|
|
|
|
ip2 = IPV6_PORTRANGE_DEFAULT;
|
|
|
|
len = sizeof (struct sockaddr_in6);
|
|
|
|
break;
|
|
|
|
#endif
|
|
|
|
default:
|
|
|
|
goto noresvport;
|
|
|
|
}
|
|
|
|
sa = (struct sockaddr *)&ssin;
|
|
|
|
bzero(sa, len);
|
|
|
|
sa->sa_len = len;
|
|
|
|
sa->sa_family = saddr->sa_family;
|
1999-06-05 05:35:03 +00:00
|
|
|
sopt.sopt_dir = SOPT_SET;
|
|
|
|
sopt.sopt_val = (void *)&ip;
|
|
|
|
sopt.sopt_valsize = sizeof(ip);
|
|
|
|
error = sosetopt(so, &sopt);
|
|
|
|
if (error)
|
|
|
|
goto bad;
|
2002-07-15 19:40:23 +00:00
|
|
|
error = sobind(so, sa, td);
|
1999-06-05 05:35:03 +00:00
|
|
|
if (error)
|
|
|
|
goto bad;
|
2002-07-15 19:40:23 +00:00
|
|
|
ip = ip2;
|
1999-06-05 05:35:03 +00:00
|
|
|
error = sosetopt(so, &sopt);
|
1994-05-24 10:09:53 +00:00
|
|
|
if (error)
|
|
|
|
goto bad;
|
2002-07-15 19:40:23 +00:00
|
|
|
noresvport: ;
|
1994-05-24 10:09:53 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Protocols that do not require connections may be optionally left
|
|
|
|
* unconnected for servers that reply from a port other than NFS_PORT.
|
|
|
|
*/
|
|
|
|
if (nmp->nm_flag & NFSMNT_NOCONN) {
|
|
|
|
if (nmp->nm_soflags & PR_CONNREQUIRED) {
|
|
|
|
error = ENOTCONN;
|
|
|
|
goto bad;
|
|
|
|
}
|
|
|
|
} else {
|
2001-09-12 08:38:13 +00:00
|
|
|
error = soconnect(so, nmp->nm_nam, td);
|
1994-10-02 17:27:07 +00:00
|
|
|
if (error)
|
1994-05-24 10:09:53 +00:00
|
|
|
goto bad;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Wait for the connection to complete. Cribbed from the
|
|
|
|
* connect system call but with the wait timing out so
|
|
|
|
* that interruptible mounts don't hang here for a long time.
|
|
|
|
*/
|
|
|
|
s = splnet();
|
|
|
|
while ((so->so_state & SS_ISCONNECTING) && so->so_error == 0) {
|
2003-03-02 16:54:40 +00:00
|
|
|
(void) tsleep(&so->so_timeo,
|
2002-05-27 05:20:15 +00:00
|
|
|
PSOCK, "nfscon", 2 * hz);
|
1994-05-24 10:09:53 +00:00
|
|
|
if ((so->so_state & SS_ISCONNECTING) &&
|
|
|
|
so->so_error == 0 && rep &&
|
2002-06-28 21:53:08 +00:00
|
|
|
(error = nfs_sigintr(nmp, rep, rep->r_td)) != 0) {
|
1994-05-24 10:09:53 +00:00
|
|
|
so->so_state &= ~SS_ISCONNECTING;
|
|
|
|
splx(s);
|
|
|
|
goto bad;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if (so->so_error) {
|
|
|
|
error = so->so_error;
|
|
|
|
so->so_error = 0;
|
|
|
|
splx(s);
|
|
|
|
goto bad;
|
|
|
|
}
|
|
|
|
splx(s);
|
|
|
|
}
|
2002-01-02 00:41:26 +00:00
|
|
|
so->so_rcv.sb_timeo = 5 * hz;
|
|
|
|
so->so_snd.sb_timeo = 5 * hz;
|
2000-03-27 21:38:35 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Get buffer reservation size from sysctl, but impose reasonable
|
|
|
|
* limits.
|
|
|
|
*/
|
|
|
|
pktscale = nfs_bufpackets;
|
|
|
|
if (pktscale < 2)
|
|
|
|
pktscale = 2;
|
|
|
|
if (pktscale > 64)
|
|
|
|
pktscale = 64;
|
|
|
|
|
1994-05-24 10:09:53 +00:00
|
|
|
if (nmp->nm_sotype == SOCK_DGRAM) {
|
2000-03-27 21:38:35 +00:00
|
|
|
sndreserve = (nmp->nm_wsize + NFS_MAXPKTHDR) * pktscale;
|
1998-05-31 17:57:43 +00:00
|
|
|
rcvreserve = (max(nmp->nm_rsize, nmp->nm_readdirsize) +
|
2000-03-27 21:38:35 +00:00
|
|
|
NFS_MAXPKTHDR) * pktscale;
|
1994-05-24 10:09:53 +00:00
|
|
|
} else if (nmp->nm_sotype == SOCK_SEQPACKET) {
|
2000-03-27 21:38:35 +00:00
|
|
|
sndreserve = (nmp->nm_wsize + NFS_MAXPKTHDR) * pktscale;
|
1998-05-31 17:57:43 +00:00
|
|
|
rcvreserve = (max(nmp->nm_rsize, nmp->nm_readdirsize) +
|
2000-03-27 21:38:35 +00:00
|
|
|
NFS_MAXPKTHDR) * pktscale;
|
1994-05-24 10:09:53 +00:00
|
|
|
} else {
|
|
|
|
if (nmp->nm_sotype != SOCK_STREAM)
|
|
|
|
panic("nfscon sotype");
|
|
|
|
if (so->so_proto->pr_flags & PR_CONNREQUIRED) {
|
1998-08-23 03:07:17 +00:00
|
|
|
struct sockopt sopt;
|
|
|
|
int val;
|
|
|
|
|
|
|
|
bzero(&sopt, sizeof sopt);
|
2003-10-04 17:41:59 +00:00
|
|
|
sopt.sopt_dir = SOPT_SET;
|
1998-08-23 03:07:17 +00:00
|
|
|
sopt.sopt_level = SOL_SOCKET;
|
|
|
|
sopt.sopt_name = SO_KEEPALIVE;
|
|
|
|
sopt.sopt_val = &val;
|
|
|
|
sopt.sopt_valsize = sizeof val;
|
|
|
|
val = 1;
|
|
|
|
sosetopt(so, &sopt);
|
1994-05-24 10:09:53 +00:00
|
|
|
}
|
|
|
|
if (so->so_proto->pr_protocol == IPPROTO_TCP) {
|
1998-08-23 03:07:17 +00:00
|
|
|
struct sockopt sopt;
|
|
|
|
int val;
|
|
|
|
|
|
|
|
bzero(&sopt, sizeof sopt);
|
2003-10-04 17:41:59 +00:00
|
|
|
sopt.sopt_dir = SOPT_SET;
|
1998-08-23 03:07:17 +00:00
|
|
|
sopt.sopt_level = IPPROTO_TCP;
|
|
|
|
sopt.sopt_name = TCP_NODELAY;
|
|
|
|
sopt.sopt_val = &val;
|
|
|
|
sopt.sopt_valsize = sizeof val;
|
|
|
|
val = 1;
|
|
|
|
sosetopt(so, &sopt);
|
1994-05-24 10:09:53 +00:00
|
|
|
}
|
1998-05-31 20:09:01 +00:00
|
|
|
sndreserve = (nmp->nm_wsize + NFS_MAXPKTHDR +
|
2000-03-27 21:38:35 +00:00
|
|
|
sizeof (u_int32_t)) * pktscale;
|
1998-05-31 20:09:01 +00:00
|
|
|
rcvreserve = (nmp->nm_rsize + NFS_MAXPKTHDR +
|
2000-03-27 21:38:35 +00:00
|
|
|
sizeof (u_int32_t)) * pktscale;
|
1994-05-24 10:09:53 +00:00
|
|
|
}
|
1994-10-02 17:27:07 +00:00
|
|
|
error = soreserve(so, sndreserve, rcvreserve);
|
|
|
|
if (error)
|
1994-05-24 10:09:53 +00:00
|
|
|
goto bad;
|
|
|
|
so->so_rcv.sb_flags |= SB_NOINTR;
|
|
|
|
so->so_snd.sb_flags |= SB_NOINTR;
|
|
|
|
|
|
|
|
/* Initialize other non-zero congestion variables */
|
2001-09-18 23:32:09 +00:00
|
|
|
nmp->nm_srtt[0] = nmp->nm_srtt[1] = nmp->nm_srtt[2] =
|
1999-11-22 04:50:09 +00:00
|
|
|
nmp->nm_srtt[3] = (NFS_TIMEO << 3);
|
1994-05-24 10:09:53 +00:00
|
|
|
nmp->nm_sdrtt[0] = nmp->nm_sdrtt[1] = nmp->nm_sdrtt[2] =
|
1999-11-22 04:50:09 +00:00
|
|
|
nmp->nm_sdrtt[3] = 0;
|
1994-05-24 10:09:53 +00:00
|
|
|
nmp->nm_cwnd = NFS_MAXCWND / 2; /* Initial send window */
|
|
|
|
nmp->nm_sent = 0;
|
|
|
|
nmp->nm_timeouts = 0;
|
|
|
|
return (0);
|
|
|
|
|
|
|
|
bad:
|
|
|
|
nfs_disconnect(nmp);
|
|
|
|
return (error);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Reconnect routine:
|
|
|
|
* Called when a connection is broken on a reliable protocol.
|
|
|
|
* - clean up the old socket
|
|
|
|
* - nfs_connect() again
|
|
|
|
* - set R_MUSTRESEND for all outstanding requests on mount point
|
|
|
|
* If this fails the mount point is DEAD!
|
|
|
|
* nb: Must be called with the nfs_sndlock() set on the mount point.
|
|
|
|
*/
|
1995-12-17 21:14:36 +00:00
|
|
|
static int
|
2001-09-18 23:32:09 +00:00
|
|
|
nfs_reconnect(struct nfsreq *rep)
|
1994-05-24 10:09:53 +00:00
|
|
|
{
|
2001-09-18 23:32:09 +00:00
|
|
|
struct nfsreq *rp;
|
|
|
|
struct nfsmount *nmp = rep->r_nmp;
|
1994-05-24 10:09:53 +00:00
|
|
|
int error;
|
|
|
|
|
|
|
|
nfs_disconnect(nmp);
|
1998-05-31 17:27:58 +00:00
|
|
|
while ((error = nfs_connect(nmp, rep)) != 0) {
|
1994-05-24 10:09:53 +00:00
|
|
|
if (error == EINTR || error == ERESTART)
|
|
|
|
return (EINTR);
|
2003-03-02 16:54:40 +00:00
|
|
|
(void) tsleep(&lbolt, PSOCK, "nfscon", 0);
|
1994-05-24 10:09:53 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Loop through outstanding request list and fix up all requests
|
|
|
|
* on old socket.
|
|
|
|
*/
|
2001-09-18 23:32:09 +00:00
|
|
|
TAILQ_FOREACH(rp, &nfs_reqq, r_chain) {
|
1994-05-24 10:09:53 +00:00
|
|
|
if (rp->r_nmp == nmp)
|
|
|
|
rp->r_flags |= R_MUSTRESEND;
|
|
|
|
}
|
|
|
|
return (0);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* NFS disconnect. Clean up and unlink.
|
|
|
|
*/
|
|
|
|
void
|
2001-09-18 23:32:09 +00:00
|
|
|
nfs_disconnect(struct nfsmount *nmp)
|
1994-05-24 10:09:53 +00:00
|
|
|
{
|
2001-09-18 23:32:09 +00:00
|
|
|
struct socket *so;
|
1994-05-24 10:09:53 +00:00
|
|
|
|
2003-11-07 22:57:09 +00:00
|
|
|
GIANT_REQUIRED; /* XXX until socket locking done */
|
|
|
|
|
1994-05-24 10:09:53 +00:00
|
|
|
if (nmp->nm_so) {
|
|
|
|
so = nmp->nm_so;
|
2002-07-11 17:54:58 +00:00
|
|
|
nmp->nm_so = NULL;
|
1994-05-24 10:09:53 +00:00
|
|
|
soshutdown(so, 2);
|
|
|
|
soclose(so);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
1998-05-31 20:09:01 +00:00
|
|
|
void
|
2001-09-18 23:32:09 +00:00
|
|
|
nfs_safedisconnect(struct nfsmount *nmp)
|
1998-05-31 19:49:31 +00:00
|
|
|
{
|
1998-05-31 20:09:01 +00:00
|
|
|
struct nfsreq dummyreq;
|
1998-05-31 19:49:31 +00:00
|
|
|
|
|
|
|
bzero(&dummyreq, sizeof(dummyreq));
|
|
|
|
dummyreq.r_nmp = nmp;
|
|
|
|
nfs_rcvlock(&dummyreq);
|
|
|
|
nfs_disconnect(nmp);
|
1999-02-25 00:03:51 +00:00
|
|
|
nfs_rcvunlock(&dummyreq);
|
1998-05-31 19:49:31 +00:00
|
|
|
}
|
|
|
|
|
1994-05-24 10:09:53 +00:00
|
|
|
/*
|
|
|
|
* This is the nfs send routine. For connection based socket types, it
|
|
|
|
* must be called with an nfs_sndlock() on the socket.
|
|
|
|
* - return EINTR if the RPC is terminated, 0 otherwise
|
|
|
|
* - set R_MUSTRESEND if the send fails for any reason
|
1998-12-30 00:37:44 +00:00
|
|
|
* - do any cleanup required by recoverable socket errors (?)
|
1994-05-24 10:09:53 +00:00
|
|
|
*/
|
1994-05-25 09:21:21 +00:00
|
|
|
int
|
2001-09-18 23:32:09 +00:00
|
|
|
nfs_send(struct socket *so, struct sockaddr *nam, struct mbuf *top,
|
|
|
|
struct nfsreq *rep)
|
1994-05-24 10:09:53 +00:00
|
|
|
{
|
1997-08-16 19:16:27 +00:00
|
|
|
struct sockaddr *sendnam;
|
1994-05-24 10:09:53 +00:00
|
|
|
int error, soflags, flags;
|
|
|
|
|
2003-11-07 22:57:09 +00:00
|
|
|
GIANT_REQUIRED; /* XXX until socket locking done */
|
|
|
|
|
2001-09-18 23:32:09 +00:00
|
|
|
KASSERT(rep, ("nfs_send: called with rep == NULL"));
|
|
|
|
|
|
|
|
if (rep->r_flags & R_SOFTTERM) {
|
|
|
|
m_freem(top);
|
|
|
|
return (EINTR);
|
|
|
|
}
|
|
|
|
if ((so = rep->r_nmp->nm_so) == NULL) {
|
|
|
|
rep->r_flags |= R_MUSTRESEND;
|
|
|
|
m_freem(top);
|
|
|
|
return (0);
|
|
|
|
}
|
|
|
|
rep->r_flags &= ~R_MUSTRESEND;
|
|
|
|
soflags = rep->r_nmp->nm_soflags;
|
|
|
|
|
2002-05-31 11:52:35 +00:00
|
|
|
if ((soflags & PR_CONNREQUIRED) || (so->so_state & SS_ISCONNECTED))
|
2002-07-11 17:54:58 +00:00
|
|
|
sendnam = NULL;
|
2002-05-31 11:52:35 +00:00
|
|
|
else
|
1994-05-24 10:09:53 +00:00
|
|
|
sendnam = nam;
|
|
|
|
if (so->so_type == SOCK_SEQPACKET)
|
|
|
|
flags = MSG_EOR;
|
|
|
|
else
|
|
|
|
flags = 0;
|
|
|
|
|
1997-04-27 20:01:29 +00:00
|
|
|
error = so->so_proto->pr_usrreqs->pru_sosend(so, sendnam, 0, top, 0,
|
2001-09-12 08:38:13 +00:00
|
|
|
flags, curthread /*XXX*/);
|
1998-08-01 09:04:02 +00:00
|
|
|
if (error == ENOBUFS && so->so_type == SOCK_DGRAM) {
|
|
|
|
error = 0;
|
2001-09-18 23:32:09 +00:00
|
|
|
rep->r_flags |= R_MUSTRESEND;
|
1998-08-01 09:04:02 +00:00
|
|
|
}
|
|
|
|
|
1994-05-24 10:09:53 +00:00
|
|
|
if (error) {
|
2004-03-17 21:10:20 +00:00
|
|
|
/*
|
|
|
|
* Don't report EPIPE errors on nfs sockets.
|
|
|
|
* These can be due to idle tcp mounts which will be closed by
|
|
|
|
* netapp, solaris, etc. if left idle too long.
|
|
|
|
*/
|
2004-03-17 18:10:38 +00:00
|
|
|
if (error != EPIPE) {
|
|
|
|
log(LOG_INFO, "nfs send error %d for server %s\n",
|
|
|
|
error,
|
|
|
|
rep->r_nmp->nm_mountp->mnt_stat.f_mntfromname);
|
|
|
|
}
|
2001-09-18 23:32:09 +00:00
|
|
|
/*
|
|
|
|
* Deal with errors for the client side.
|
|
|
|
*/
|
|
|
|
if (rep->r_flags & R_SOFTTERM)
|
|
|
|
error = EINTR;
|
|
|
|
else
|
|
|
|
rep->r_flags |= R_MUSTRESEND;
|
1994-05-24 10:09:53 +00:00
|
|
|
|
|
|
|
/*
|
1998-12-30 00:37:44 +00:00
|
|
|
* Handle any recoverable (soft) socket errors here. (?)
|
1994-05-24 10:09:53 +00:00
|
|
|
*/
|
|
|
|
if (error != EINTR && error != ERESTART &&
|
|
|
|
error != EWOULDBLOCK && error != EPIPE)
|
|
|
|
error = 0;
|
|
|
|
}
|
|
|
|
return (error);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Receive a Sun RPC Request/Reply. For SOCK_DGRAM, the work is all
|
|
|
|
* done by soreceive(), but for SOCK_STREAM we must deal with the Record
|
|
|
|
* Mark and consolidate the data into a new mbuf list.
|
|
|
|
* nb: Sometimes TCP passes the data up to soreceive() in long lists of
|
|
|
|
* small mbufs.
|
|
|
|
* For SOCK_STREAM we must be very careful to read an entire record once
|
|
|
|
* we have read any of it, even if the system call has been interrupted.
|
|
|
|
*/
|
1995-12-17 21:14:36 +00:00
|
|
|
static int
|
2001-09-18 23:32:09 +00:00
|
|
|
nfs_receive(struct nfsreq *rep, struct sockaddr **aname, struct mbuf **mp)
|
1994-05-24 10:09:53 +00:00
|
|
|
{
|
2001-09-18 23:32:09 +00:00
|
|
|
struct socket *so;
|
1994-05-24 10:09:53 +00:00
|
|
|
struct uio auio;
|
|
|
|
struct iovec aio;
|
2001-09-18 23:32:09 +00:00
|
|
|
struct mbuf *m;
|
1994-05-24 10:09:53 +00:00
|
|
|
struct mbuf *control;
|
1998-05-31 20:09:01 +00:00
|
|
|
u_int32_t len;
|
1997-08-16 19:16:27 +00:00
|
|
|
struct sockaddr **getnam;
|
1994-05-24 10:09:53 +00:00
|
|
|
int error, sotype, rcvflg;
|
2001-09-12 08:38:13 +00:00
|
|
|
struct thread *td = curthread; /* XXX */
|
1994-05-24 10:09:53 +00:00
|
|
|
|
2003-11-07 22:57:09 +00:00
|
|
|
GIANT_REQUIRED; /* XXX until socket locking done */
|
|
|
|
|
1994-05-24 10:09:53 +00:00
|
|
|
/*
|
|
|
|
* Set up arguments for soreceive()
|
|
|
|
*/
|
2002-07-11 17:54:58 +00:00
|
|
|
*mp = NULL;
|
|
|
|
*aname = NULL;
|
1994-05-24 10:09:53 +00:00
|
|
|
sotype = rep->r_nmp->nm_sotype;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* For reliable protocols, lock against other senders/receivers
|
|
|
|
* in case a reconnect is necessary.
|
|
|
|
* For SOCK_STREAM, first get the Record Mark to find out how much
|
|
|
|
* more there is to get.
|
|
|
|
* We must lock the socket against other receivers
|
|
|
|
* until we have an entire rpc request/reply.
|
|
|
|
*/
|
|
|
|
if (sotype != SOCK_DGRAM) {
|
1999-02-25 00:03:51 +00:00
|
|
|
error = nfs_sndlock(rep);
|
1994-10-02 17:27:07 +00:00
|
|
|
if (error)
|
1994-05-24 10:09:53 +00:00
|
|
|
return (error);
|
|
|
|
tryagain:
|
|
|
|
/*
|
|
|
|
* Check for fatal errors and resending request.
|
|
|
|
*/
|
|
|
|
/*
|
|
|
|
* Ugh: If a reconnect attempt just happened, nm_so
|
|
|
|
* would have changed. NULL indicates a failed
|
|
|
|
* attempt that has essentially shut down this
|
|
|
|
* mount point.
|
|
|
|
*/
|
|
|
|
if (rep->r_mrep || (rep->r_flags & R_SOFTTERM)) {
|
1999-02-25 00:03:51 +00:00
|
|
|
nfs_sndunlock(rep);
|
1994-05-24 10:09:53 +00:00
|
|
|
return (EINTR);
|
|
|
|
}
|
1994-10-02 17:27:07 +00:00
|
|
|
so = rep->r_nmp->nm_so;
|
|
|
|
if (!so) {
|
1995-05-30 08:16:23 +00:00
|
|
|
error = nfs_reconnect(rep);
|
1994-10-02 17:27:07 +00:00
|
|
|
if (error) {
|
1999-02-25 00:03:51 +00:00
|
|
|
nfs_sndunlock(rep);
|
1994-05-24 10:09:53 +00:00
|
|
|
return (error);
|
|
|
|
}
|
|
|
|
goto tryagain;
|
|
|
|
}
|
|
|
|
while (rep->r_flags & R_MUSTRESEND) {
|
2003-02-19 05:47:46 +00:00
|
|
|
m = m_copym(rep->r_mreq, 0, M_COPYALL, M_TRYWAIT);
|
1994-05-24 10:09:53 +00:00
|
|
|
nfsstats.rpcretries++;
|
1994-10-02 17:27:07 +00:00
|
|
|
error = nfs_send(so, rep->r_nmp->nm_nam, m, rep);
|
|
|
|
if (error) {
|
1994-05-24 10:09:53 +00:00
|
|
|
if (error == EINTR || error == ERESTART ||
|
1998-05-31 17:27:58 +00:00
|
|
|
(error = nfs_reconnect(rep)) != 0) {
|
1999-02-25 00:03:51 +00:00
|
|
|
nfs_sndunlock(rep);
|
1994-05-24 10:09:53 +00:00
|
|
|
return (error);
|
|
|
|
}
|
|
|
|
goto tryagain;
|
|
|
|
}
|
|
|
|
}
|
1999-02-25 00:03:51 +00:00
|
|
|
nfs_sndunlock(rep);
|
1994-05-24 10:09:53 +00:00
|
|
|
if (sotype == SOCK_STREAM) {
|
|
|
|
aio.iov_base = (caddr_t) &len;
|
1998-05-31 20:09:01 +00:00
|
|
|
aio.iov_len = sizeof(u_int32_t);
|
1994-05-24 10:09:53 +00:00
|
|
|
auio.uio_iov = &aio;
|
|
|
|
auio.uio_iovcnt = 1;
|
|
|
|
auio.uio_segflg = UIO_SYSSPACE;
|
|
|
|
auio.uio_rw = UIO_READ;
|
|
|
|
auio.uio_offset = 0;
|
1998-05-31 20:09:01 +00:00
|
|
|
auio.uio_resid = sizeof(u_int32_t);
|
2001-09-12 08:38:13 +00:00
|
|
|
auio.uio_td = td;
|
1994-05-24 10:09:53 +00:00
|
|
|
do {
|
|
|
|
rcvflg = MSG_WAITALL;
|
1997-04-27 20:01:29 +00:00
|
|
|
error = so->so_proto->pr_usrreqs->pru_soreceive
|
2002-07-11 17:54:58 +00:00
|
|
|
(so, NULL, &auio, NULL, NULL, &rcvflg);
|
1994-05-24 10:09:53 +00:00
|
|
|
if (error == EWOULDBLOCK && rep) {
|
|
|
|
if (rep->r_flags & R_SOFTTERM)
|
|
|
|
return (EINTR);
|
|
|
|
}
|
|
|
|
} while (error == EWOULDBLOCK);
|
|
|
|
if (!error && auio.uio_resid > 0) {
|
1998-05-31 18:00:46 +00:00
|
|
|
/*
|
|
|
|
* Don't log a 0 byte receive; it means
|
|
|
|
* that the socket has been closed, and
|
|
|
|
* can happen during normal operation
|
|
|
|
* (forcible unmount or Solaris server).
|
|
|
|
*/
|
|
|
|
if (auio.uio_resid != sizeof (u_int32_t))
|
1994-05-24 10:09:53 +00:00
|
|
|
log(LOG_INFO,
|
|
|
|
"short receive (%d/%d) from nfs server %s\n",
|
1999-04-24 11:29:48 +00:00
|
|
|
(int)(sizeof(u_int32_t) - auio.uio_resid),
|
|
|
|
(int)sizeof(u_int32_t),
|
1994-05-24 10:09:53 +00:00
|
|
|
rep->r_nmp->nm_mountp->mnt_stat.f_mntfromname);
|
|
|
|
error = EPIPE;
|
|
|
|
}
|
|
|
|
if (error)
|
|
|
|
goto errout;
|
|
|
|
len = ntohl(len) & ~0x80000000;
|
|
|
|
/*
|
|
|
|
* This is SERIOUS! We are out of sync with the sender
|
|
|
|
* and forcing a disconnect/reconnect is all I can do.
|
|
|
|
*/
|
|
|
|
if (len > NFS_MAXPACKET) {
|
|
|
|
log(LOG_ERR, "%s (%d) from nfs server %s\n",
|
|
|
|
"impossible packet length",
|
|
|
|
len,
|
|
|
|
rep->r_nmp->nm_mountp->mnt_stat.f_mntfromname);
|
|
|
|
error = EFBIG;
|
|
|
|
goto errout;
|
|
|
|
}
|
|
|
|
auio.uio_resid = len;
|
|
|
|
do {
|
|
|
|
rcvflg = MSG_WAITALL;
|
1997-04-27 20:01:29 +00:00
|
|
|
error = so->so_proto->pr_usrreqs->pru_soreceive
|
2002-07-11 17:54:58 +00:00
|
|
|
(so, NULL,
|
|
|
|
&auio, mp, NULL, &rcvflg);
|
1994-05-24 10:09:53 +00:00
|
|
|
} while (error == EWOULDBLOCK || error == EINTR ||
|
|
|
|
error == ERESTART);
|
|
|
|
if (!error && auio.uio_resid > 0) {
|
1998-05-31 18:00:46 +00:00
|
|
|
if (len != auio.uio_resid)
|
1994-05-24 10:09:53 +00:00
|
|
|
log(LOG_INFO,
|
|
|
|
"short receive (%d/%d) from nfs server %s\n",
|
|
|
|
len - auio.uio_resid, len,
|
|
|
|
rep->r_nmp->nm_mountp->mnt_stat.f_mntfromname);
|
|
|
|
error = EPIPE;
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
/*
|
|
|
|
* NB: Since uio_resid is big, MSG_WAITALL is ignored
|
|
|
|
* and soreceive() will return when it has either a
|
|
|
|
* control msg or a data msg.
|
|
|
|
* We have no use for control msg., but must grab them
|
|
|
|
* and then throw them away so we know what is going
|
|
|
|
* on.
|
|
|
|
*/
|
|
|
|
auio.uio_resid = len = 100000000; /* Anything Big */
|
2001-09-12 08:38:13 +00:00
|
|
|
auio.uio_td = td;
|
1994-05-24 10:09:53 +00:00
|
|
|
do {
|
|
|
|
rcvflg = 0;
|
1997-04-27 20:01:29 +00:00
|
|
|
error = so->so_proto->pr_usrreqs->pru_soreceive
|
2002-07-11 17:54:58 +00:00
|
|
|
(so, NULL,
|
1994-05-24 10:09:53 +00:00
|
|
|
&auio, mp, &control, &rcvflg);
|
|
|
|
if (control)
|
|
|
|
m_freem(control);
|
|
|
|
if (error == EWOULDBLOCK && rep) {
|
|
|
|
if (rep->r_flags & R_SOFTTERM)
|
|
|
|
return (EINTR);
|
|
|
|
}
|
|
|
|
} while (error == EWOULDBLOCK ||
|
|
|
|
(!error && *mp == NULL && control));
|
|
|
|
if ((rcvflg & MSG_EOR) == 0)
|
|
|
|
printf("Egad!!\n");
|
|
|
|
if (!error && *mp == NULL)
|
|
|
|
error = EPIPE;
|
|
|
|
len -= auio.uio_resid;
|
|
|
|
}
|
|
|
|
errout:
|
|
|
|
if (error && error != EINTR && error != ERESTART) {
|
|
|
|
m_freem(*mp);
|
2002-07-11 17:54:58 +00:00
|
|
|
*mp = NULL;
|
1994-05-24 10:09:53 +00:00
|
|
|
if (error != EPIPE)
|
|
|
|
log(LOG_INFO,
|
|
|
|
"receive error %d from nfs server %s\n",
|
|
|
|
error,
|
|
|
|
rep->r_nmp->nm_mountp->mnt_stat.f_mntfromname);
|
1999-02-25 00:03:51 +00:00
|
|
|
error = nfs_sndlock(rep);
|
2002-07-16 15:12:07 +00:00
|
|
|
if (!error) {
|
1994-05-24 10:09:53 +00:00
|
|
|
error = nfs_reconnect(rep);
|
2002-07-16 15:12:07 +00:00
|
|
|
if (!error)
|
|
|
|
goto tryagain;
|
|
|
|
else
|
|
|
|
nfs_sndunlock(rep);
|
|
|
|
}
|
1994-05-24 10:09:53 +00:00
|
|
|
}
|
|
|
|
} else {
|
|
|
|
if ((so = rep->r_nmp->nm_so) == NULL)
|
|
|
|
return (EACCES);
|
2002-05-31 11:52:35 +00:00
|
|
|
if (so->so_state & SS_ISCONNECTED)
|
2002-07-11 17:54:58 +00:00
|
|
|
getnam = NULL;
|
2002-05-31 11:52:35 +00:00
|
|
|
else
|
1994-05-24 10:09:53 +00:00
|
|
|
getnam = aname;
|
|
|
|
auio.uio_resid = len = 1000000;
|
2001-09-12 08:38:13 +00:00
|
|
|
auio.uio_td = td;
|
1994-05-24 10:09:53 +00:00
|
|
|
do {
|
|
|
|
rcvflg = 0;
|
1997-04-27 20:01:29 +00:00
|
|
|
error = so->so_proto->pr_usrreqs->pru_soreceive
|
|
|
|
(so, getnam, &auio, mp,
|
2002-07-11 17:54:58 +00:00
|
|
|
NULL, &rcvflg);
|
1994-05-24 10:09:53 +00:00
|
|
|
if (error == EWOULDBLOCK &&
|
|
|
|
(rep->r_flags & R_SOFTTERM))
|
|
|
|
return (EINTR);
|
|
|
|
} while (error == EWOULDBLOCK);
|
|
|
|
len -= auio.uio_resid;
|
|
|
|
}
|
|
|
|
if (error) {
|
|
|
|
m_freem(*mp);
|
2002-07-11 17:54:58 +00:00
|
|
|
*mp = NULL;
|
1994-05-24 10:09:53 +00:00
|
|
|
}
|
|
|
|
/*
|
|
|
|
* Search for any mbufs that are not a multiple of 4 bytes long
|
|
|
|
* or with m_data not longword aligned.
|
|
|
|
* These could cause pointer alignment problems, so copy them to
|
|
|
|
* well aligned mbufs.
|
|
|
|
*/
|
The VFS/BIO subsystem contained a number of hacks in order to optimize
piecemeal, middle-of-file writes for NFS. These hacks have caused no
end of trouble, especially when combined with mmap(). I've removed
them. Instead, NFS will issue a read-before-write to fully
instantiate the struct buf containing the write. NFS does, however,
optimize piecemeal appends to files. For most common file operations,
you will not notice the difference. The sole remaining fragment in
the VFS/BIO system is b_dirtyoff/end, which NFS uses to avoid cache
coherency issues with read-merge-write style operations. NFS also
optimizes the write-covers-entire-buffer case by avoiding the
read-before-write. There is quite a bit of room for further
optimization in these areas.
The VM system marks pages fully-valid (AKA vm_page_t->valid =
VM_PAGE_BITS_ALL) in several places, most noteably in vm_fault. This
is not correct operation. The vm_pager_get_pages() code is now
responsible for marking VM pages all-valid. A number of VM helper
routines have been added to aid in zeroing-out the invalid portions of
a VM page prior to the page being marked all-valid. This operation is
necessary to properly support mmap(). The zeroing occurs most often
when dealing with file-EOF situations. Several bugs have been fixed
in the NFS subsystem, including bits handling file and directory EOF
situations and buf->b_flags consistancy issues relating to clearing
B_ERROR & B_INVAL, and handling B_DONE.
getblk() and allocbuf() have been rewritten. B_CACHE operation is now
formally defined in comments and more straightforward in
implementation. B_CACHE for VMIO buffers is based on the validity of
the backing store. B_CACHE for non-VMIO buffers is based simply on
whether the buffer is B_INVAL or not (B_CACHE set if B_INVAL clear,
and vise-versa). biodone() is now responsible for setting B_CACHE
when a successful read completes. B_CACHE is also set when a bdwrite()
is initiated and when a bwrite() is initiated. VFS VOP_BWRITE
routines (there are only two - nfs_bwrite() and bwrite()) are now
expected to set B_CACHE. This means that bowrite() and bawrite() also
set B_CACHE indirectly.
There are a number of places in the code which were previously using
buf->b_bufsize (which is DEV_BSIZE aligned) when they should have
been using buf->b_bcount. These have been fixed. getblk() now clears
B_DONE on return because the rest of the system is so bad about
dealing with B_DONE.
Major fixes to NFS/TCP have been made. A server-side bug could cause
requests to be lost by the server due to nfs_realign() overwriting
other rpc's in the same TCP mbuf chain. The server's kernel must be
recompiled to get the benefit of the fixes.
Submitted by: Matthew Dillon <dillon@apollo.backplane.com>
1999-05-02 23:57:16 +00:00
|
|
|
nfs_realign(mp, 5 * NFSX_UNSIGNED);
|
1994-05-24 10:09:53 +00:00
|
|
|
return (error);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Implement receipt of reply on a socket.
|
|
|
|
* We must search through the list of received datagrams matching them
|
|
|
|
* with outstanding requests using the xid, until ours is found.
|
|
|
|
*/
|
|
|
|
/* ARGSUSED */
|
2001-09-18 23:32:09 +00:00
|
|
|
static int
|
|
|
|
nfs_reply(struct nfsreq *myrep)
|
1994-05-24 10:09:53 +00:00
|
|
|
{
|
2001-09-18 23:32:09 +00:00
|
|
|
struct nfsreq *rep;
|
|
|
|
struct nfsmount *nmp = myrep->r_nmp;
|
|
|
|
int32_t t1;
|
1997-08-16 19:16:27 +00:00
|
|
|
struct mbuf *mrep, *md;
|
|
|
|
struct sockaddr *nam;
|
1998-05-31 20:09:01 +00:00
|
|
|
u_int32_t rxid, *tl;
|
2001-09-18 23:32:09 +00:00
|
|
|
caddr_t dpos;
|
1994-05-24 10:09:53 +00:00
|
|
|
int error;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Loop around until we get our own reply
|
|
|
|
*/
|
|
|
|
for (;;) {
|
|
|
|
/*
|
|
|
|
* Lock against other receivers so that I don't get stuck in
|
|
|
|
* sbwait() after someone else has received my reply for me.
|
|
|
|
* Also necessary for connection based protocols to avoid
|
|
|
|
* race conditions during a reconnect.
|
1996-10-11 10:15:33 +00:00
|
|
|
* If nfs_rcvlock() returns EALREADY, that means that
|
|
|
|
* the reply has already been recieved by another
|
|
|
|
* process and we can return immediately. In this
|
|
|
|
* case, the lock is not taken to avoid races with
|
|
|
|
* other processes.
|
1994-05-24 10:09:53 +00:00
|
|
|
*/
|
1994-10-02 17:27:07 +00:00
|
|
|
error = nfs_rcvlock(myrep);
|
1996-10-11 10:15:33 +00:00
|
|
|
if (error == EALREADY)
|
|
|
|
return (0);
|
1994-10-02 17:27:07 +00:00
|
|
|
if (error)
|
1994-05-24 10:09:53 +00:00
|
|
|
return (error);
|
|
|
|
/*
|
|
|
|
* Get the next Rpc reply off the socket
|
|
|
|
*/
|
|
|
|
error = nfs_receive(myrep, &nam, &mrep);
|
1999-02-25 00:03:51 +00:00
|
|
|
nfs_rcvunlock(myrep);
|
1994-05-24 10:09:53 +00:00
|
|
|
if (error) {
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Ignore routing errors on connectionless protocols??
|
|
|
|
*/
|
|
|
|
if (NFSIGNORE_SOERROR(nmp->nm_soflags, error)) {
|
|
|
|
nmp->nm_so->so_error = 0;
|
|
|
|
if (myrep->r_flags & R_GETONEREP)
|
|
|
|
return (0);
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
return (error);
|
|
|
|
}
|
|
|
|
if (nam)
|
1997-08-16 19:16:27 +00:00
|
|
|
FREE(nam, M_SONAME);
|
1995-05-30 08:16:23 +00:00
|
|
|
|
1994-05-24 10:09:53 +00:00
|
|
|
/*
|
|
|
|
* Get the xid and check that it is an rpc reply
|
|
|
|
*/
|
|
|
|
md = mrep;
|
|
|
|
dpos = mtod(md, caddr_t);
|
2001-09-27 22:40:38 +00:00
|
|
|
tl = nfsm_dissect(u_int32_t *, 2 * NFSX_UNSIGNED);
|
1994-05-24 10:09:53 +00:00
|
|
|
rxid = *tl++;
|
|
|
|
if (*tl != rpc_reply) {
|
1996-01-13 23:27:58 +00:00
|
|
|
nfsstats.rpcinvalid++;
|
|
|
|
m_freem(mrep);
|
1994-05-24 10:09:53 +00:00
|
|
|
nfsmout:
|
|
|
|
if (myrep->r_flags & R_GETONEREP)
|
|
|
|
return (0);
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Loop through the request list to match up the reply
|
|
|
|
* Iff no match, just drop the datagram
|
|
|
|
*/
|
2001-09-18 23:32:09 +00:00
|
|
|
TAILQ_FOREACH(rep, &nfs_reqq, r_chain) {
|
1994-05-24 10:09:53 +00:00
|
|
|
if (rep->r_mrep == NULL && rxid == rep->r_xid) {
|
|
|
|
/* Found it.. */
|
|
|
|
rep->r_mrep = mrep;
|
|
|
|
rep->r_md = md;
|
|
|
|
rep->r_dpos = dpos;
|
|
|
|
/*
|
|
|
|
* Update congestion window.
|
|
|
|
* Do the additive increase of
|
|
|
|
* one rpc/rtt.
|
|
|
|
*/
|
|
|
|
if (nmp->nm_cwnd <= nmp->nm_sent) {
|
|
|
|
nmp->nm_cwnd +=
|
|
|
|
(NFS_CWNDSCALE * NFS_CWNDSCALE +
|
|
|
|
(nmp->nm_cwnd >> 1)) / nmp->nm_cwnd;
|
|
|
|
if (nmp->nm_cwnd > NFS_MAXCWND)
|
|
|
|
nmp->nm_cwnd = NFS_MAXCWND;
|
|
|
|
}
|
1999-12-13 04:24:55 +00:00
|
|
|
if (rep->r_flags & R_SENT) {
|
|
|
|
rep->r_flags &= ~R_SENT;
|
|
|
|
nmp->nm_sent -= NFS_CWNDSCALE;
|
|
|
|
}
|
1994-05-24 10:09:53 +00:00
|
|
|
/*
|
|
|
|
* Update rtt using a gain of 0.125 on the mean
|
|
|
|
* and a gain of 0.25 on the deviation.
|
|
|
|
*/
|
|
|
|
if (rep->r_flags & R_TIMING) {
|
|
|
|
/*
|
|
|
|
* Since the timer resolution of
|
|
|
|
* NFS_HZ is so course, it can often
|
|
|
|
* result in r_rtt == 0. Since
|
|
|
|
* r_rtt == N means that the actual
|
|
|
|
* rtt is between N+dt and N+2-dt ticks,
|
|
|
|
* add 1.
|
|
|
|
*/
|
|
|
|
t1 = rep->r_rtt + 1;
|
|
|
|
t1 -= (NFS_SRTT(rep) >> 3);
|
|
|
|
NFS_SRTT(rep) += t1;
|
|
|
|
if (t1 < 0)
|
|
|
|
t1 = -t1;
|
|
|
|
t1 -= (NFS_SDRTT(rep) >> 2);
|
|
|
|
NFS_SDRTT(rep) += t1;
|
|
|
|
}
|
|
|
|
nmp->nm_timeouts = 0;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* If not matched to a request, drop it.
|
|
|
|
* If it's mine, get out.
|
|
|
|
*/
|
1994-10-17 17:47:45 +00:00
|
|
|
if (rep == 0) {
|
1994-05-24 10:09:53 +00:00
|
|
|
nfsstats.rpcunexpected++;
|
|
|
|
m_freem(mrep);
|
|
|
|
} else if (rep == myrep) {
|
|
|
|
if (rep->r_mrep == NULL)
|
|
|
|
panic("nfsreply nil");
|
|
|
|
return (0);
|
|
|
|
}
|
|
|
|
if (myrep->r_flags & R_GETONEREP)
|
|
|
|
return (0);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* nfs_request - goes something like this
|
|
|
|
* - fill in request struct
|
|
|
|
* - links it into list
|
|
|
|
* - calls nfs_send() for first transmit
|
|
|
|
* - calls nfs_receive() to get reply
|
|
|
|
* - break down rpc header and return with nfs reply pointed to
|
|
|
|
* by mrep or error
|
|
|
|
* nb: always frees up mreq mbuf list
|
|
|
|
*/
|
2001-09-18 23:32:09 +00:00
|
|
|
/* XXX overloaded before */
|
|
|
|
#define NQ_TRYLATERDEL 15 /* Initial try later delay (sec) */
|
|
|
|
|
1994-05-25 09:21:21 +00:00
|
|
|
int
|
2001-09-18 23:32:09 +00:00
|
|
|
nfs_request(struct vnode *vp, struct mbuf *mrest, int procnum,
|
|
|
|
struct thread *td, struct ucred *cred, struct mbuf **mrp,
|
|
|
|
struct mbuf **mdp, caddr_t *dposp)
|
1994-05-24 10:09:53 +00:00
|
|
|
{
|
2001-09-18 23:32:09 +00:00
|
|
|
struct mbuf *mrep, *m2;
|
|
|
|
struct nfsreq *rep;
|
|
|
|
u_int32_t *tl;
|
|
|
|
int i;
|
1994-05-24 10:09:53 +00:00
|
|
|
struct nfsmount *nmp;
|
1999-12-19 01:55:37 +00:00
|
|
|
struct mbuf *m, *md, *mheadend;
|
2001-09-18 23:32:09 +00:00
|
|
|
time_t waituntil;
|
|
|
|
caddr_t dpos;
|
|
|
|
int s, error = 0, mrest_len, auth_len, auth_type;
|
|
|
|
int trylater_delay = NQ_TRYLATERDEL, trylater_cnt = 0;
|
1998-05-31 20:09:01 +00:00
|
|
|
u_int32_t xid;
|
1994-05-24 10:09:53 +00:00
|
|
|
|
2002-04-17 01:07:29 +00:00
|
|
|
/* Reject requests while attempting a forced unmount. */
|
|
|
|
if (vp->v_mount->mnt_kern_flag & MNTK_UNMOUNTF) {
|
2002-01-02 00:41:26 +00:00
|
|
|
m_freem(mrest);
|
|
|
|
return (ESTALE);
|
|
|
|
}
|
1994-05-24 10:09:53 +00:00
|
|
|
nmp = VFSTONFS(vp->v_mount);
|
2003-11-14 20:54:10 +00:00
|
|
|
if ((nmp->nm_flag & NFSMNT_NFSV4) != 0)
|
|
|
|
return nfs4_request(vp, mrest, procnum, td, cred, mrp, mdp, dposp);
|
2003-02-19 05:47:46 +00:00
|
|
|
MALLOC(rep, struct nfsreq *, sizeof(struct nfsreq), M_NFSREQ, M_WAITOK);
|
1994-05-24 10:09:53 +00:00
|
|
|
rep->r_nmp = nmp;
|
|
|
|
rep->r_vp = vp;
|
2001-09-12 08:38:13 +00:00
|
|
|
rep->r_td = td;
|
1994-05-24 10:09:53 +00:00
|
|
|
rep->r_procnum = procnum;
|
2002-09-18 19:44:14 +00:00
|
|
|
mrest_len = m_length(mrest, NULL);
|
1994-05-24 10:09:53 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Get the RPC header with authorization.
|
|
|
|
*/
|
2001-09-18 23:32:09 +00:00
|
|
|
auth_type = RPCAUTH_UNIX;
|
|
|
|
if (cred->cr_ngroups < 1)
|
|
|
|
panic("nfsreq nogrps");
|
|
|
|
auth_len = ((((cred->cr_ngroups - 1) > nmp->nm_numgrps) ?
|
|
|
|
nmp->nm_numgrps : (cred->cr_ngroups - 1)) << 2) +
|
|
|
|
5 * NFSX_UNSIGNED;
|
1995-06-27 11:07:30 +00:00
|
|
|
m = nfsm_rpchead(cred, nmp->nm_flag, procnum, auth_type, auth_len,
|
2001-09-18 23:32:09 +00:00
|
|
|
mrest, mrest_len, &mheadend, &xid);
|
1994-05-24 10:09:53 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* For stream protocols, insert a Sun RPC Record Mark.
|
|
|
|
*/
|
|
|
|
if (nmp->nm_sotype == SOCK_STREAM) {
|
2003-02-19 05:47:46 +00:00
|
|
|
M_PREPEND(m, NFSX_UNSIGNED, M_TRYWAIT);
|
1998-05-31 20:09:01 +00:00
|
|
|
*mtod(m, u_int32_t *) = htonl(0x80000000 |
|
1994-05-24 10:09:53 +00:00
|
|
|
(m->m_pkthdr.len - NFSX_UNSIGNED));
|
|
|
|
}
|
|
|
|
rep->r_mreq = m;
|
|
|
|
rep->r_xid = xid;
|
|
|
|
tryagain:
|
|
|
|
if (nmp->nm_flag & NFSMNT_SOFT)
|
|
|
|
rep->r_retry = nmp->nm_retry;
|
|
|
|
else
|
|
|
|
rep->r_retry = NFS_MAXREXMIT + 1; /* past clip limit */
|
|
|
|
rep->r_rtt = rep->r_rexmit = 0;
|
|
|
|
if (proct[procnum] > 0)
|
|
|
|
rep->r_flags = R_TIMING;
|
|
|
|
else
|
|
|
|
rep->r_flags = 0;
|
|
|
|
rep->r_mrep = NULL;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Do the client side RPC.
|
|
|
|
*/
|
|
|
|
nfsstats.rpcrequests++;
|
|
|
|
/*
|
|
|
|
* Chain request into list of outstanding requests. Be sure
|
|
|
|
* to put it LAST so timer finds oldest requests first.
|
|
|
|
*/
|
|
|
|
s = splsoftclock();
|
2004-03-25 21:48:09 +00:00
|
|
|
if (TAILQ_EMPTY(&nfs_reqq))
|
|
|
|
callout_reset(&nfs_callout, nfs_ticks, nfs_timer, NULL);
|
1994-10-17 17:47:45 +00:00
|
|
|
TAILQ_INSERT_TAIL(&nfs_reqq, rep, r_chain);
|
1994-05-24 10:09:53 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* If backing off another request or avoiding congestion, don't
|
|
|
|
* send this one now but let timer do it. If not timing a request,
|
|
|
|
* do it now.
|
|
|
|
*/
|
|
|
|
if (nmp->nm_so && (nmp->nm_sotype != SOCK_DGRAM ||
|
|
|
|
(nmp->nm_flag & NFSMNT_DUMBTIMR) ||
|
|
|
|
nmp->nm_sent < nmp->nm_cwnd)) {
|
|
|
|
splx(s);
|
|
|
|
if (nmp->nm_soflags & PR_CONNREQUIRED)
|
1999-02-25 00:03:51 +00:00
|
|
|
error = nfs_sndlock(rep);
|
1994-05-24 10:09:53 +00:00
|
|
|
if (!error) {
|
2003-02-19 05:47:46 +00:00
|
|
|
m2 = m_copym(m, 0, M_COPYALL, M_TRYWAIT);
|
1999-06-05 05:35:03 +00:00
|
|
|
error = nfs_send(nmp->nm_so, nmp->nm_nam, m2, rep);
|
1994-05-24 10:09:53 +00:00
|
|
|
if (nmp->nm_soflags & PR_CONNREQUIRED)
|
1999-02-25 00:03:51 +00:00
|
|
|
nfs_sndunlock(rep);
|
1994-05-24 10:09:53 +00:00
|
|
|
}
|
|
|
|
if (!error && (rep->r_flags & R_MUSTRESEND) == 0) {
|
|
|
|
nmp->nm_sent += NFS_CWNDSCALE;
|
|
|
|
rep->r_flags |= R_SENT;
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
splx(s);
|
|
|
|
rep->r_rtt = -1;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Wait for the reply from our send or the timer's.
|
|
|
|
*/
|
|
|
|
if (!error || error == EPIPE)
|
|
|
|
error = nfs_reply(rep);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* RPC done, unlink the request.
|
|
|
|
*/
|
|
|
|
s = splsoftclock();
|
1994-10-17 17:47:45 +00:00
|
|
|
TAILQ_REMOVE(&nfs_reqq, rep, r_chain);
|
2004-03-25 21:48:09 +00:00
|
|
|
if (TAILQ_EMPTY(&nfs_reqq))
|
|
|
|
callout_stop(&nfs_callout);
|
1994-05-24 10:09:53 +00:00
|
|
|
splx(s);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Decrement the outstanding request count.
|
|
|
|
*/
|
|
|
|
if (rep->r_flags & R_SENT) {
|
|
|
|
rep->r_flags &= ~R_SENT; /* paranoia */
|
|
|
|
nmp->nm_sent -= NFS_CWNDSCALE;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If there was a successful reply and a tprintf msg.
|
|
|
|
* tprintf a response.
|
|
|
|
*/
|
|
|
|
if (!error && (rep->r_flags & R_TPRINTFMSG))
|
2001-09-12 08:38:13 +00:00
|
|
|
nfs_msg(rep->r_td, nmp->nm_mountp->mnt_stat.f_mntfromname,
|
1994-05-24 10:09:53 +00:00
|
|
|
"is alive again");
|
|
|
|
mrep = rep->r_mrep;
|
|
|
|
md = rep->r_md;
|
|
|
|
dpos = rep->r_dpos;
|
|
|
|
if (error) {
|
|
|
|
m_freem(rep->r_mreq);
|
|
|
|
free((caddr_t)rep, M_NFSREQ);
|
|
|
|
return (error);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* break down the rpc header and check if ok
|
|
|
|
*/
|
2001-09-27 22:40:38 +00:00
|
|
|
tl = nfsm_dissect(u_int32_t *, 3 * NFSX_UNSIGNED);
|
1994-05-24 10:09:53 +00:00
|
|
|
if (*tl++ == rpc_msgdenied) {
|
|
|
|
if (*tl == rpc_mismatch)
|
|
|
|
error = EOPNOTSUPP;
|
2001-09-18 23:32:09 +00:00
|
|
|
else
|
1994-05-24 10:09:53 +00:00
|
|
|
error = EACCES;
|
|
|
|
m_freem(mrep);
|
|
|
|
m_freem(rep->r_mreq);
|
|
|
|
free((caddr_t)rep, M_NFSREQ);
|
|
|
|
return (error);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
2001-09-18 23:32:09 +00:00
|
|
|
* Just throw away any verifyer (ie: kerberos etc).
|
1994-05-24 10:09:53 +00:00
|
|
|
*/
|
2001-09-18 23:32:09 +00:00
|
|
|
i = fxdr_unsigned(int, *tl++); /* verf type */
|
|
|
|
i = fxdr_unsigned(int32_t, *tl); /* len */
|
|
|
|
if (i > 0)
|
1995-06-27 11:07:30 +00:00
|
|
|
nfsm_adv(nfsm_rndup(i));
|
2001-09-27 22:40:38 +00:00
|
|
|
tl = nfsm_dissect(u_int32_t *, NFSX_UNSIGNED);
|
1994-05-24 10:09:53 +00:00
|
|
|
/* 0 == ok */
|
|
|
|
if (*tl == 0) {
|
2001-09-27 22:40:38 +00:00
|
|
|
tl = nfsm_dissect(u_int32_t *, NFSX_UNSIGNED);
|
1994-05-24 10:09:53 +00:00
|
|
|
if (*tl != 0) {
|
|
|
|
error = fxdr_unsigned(int, *tl);
|
1995-06-27 11:07:30 +00:00
|
|
|
if ((nmp->nm_flag & NFSMNT_NFSV3) &&
|
|
|
|
error == NFSERR_TRYLATER) {
|
|
|
|
m_freem(mrep);
|
1994-05-24 10:09:53 +00:00
|
|
|
error = 0;
|
1998-03-30 09:56:58 +00:00
|
|
|
waituntil = time_second + trylater_delay;
|
|
|
|
while (time_second < waituntil)
|
2003-03-02 16:54:40 +00:00
|
|
|
(void) tsleep(&lbolt,
|
1994-05-24 10:09:53 +00:00
|
|
|
PSOCK, "nqnfstry", 0);
|
|
|
|
trylater_delay *= nfs_backoff[trylater_cnt];
|
2001-12-30 18:41:52 +00:00
|
|
|
if (trylater_cnt < NFS_NBACKOFF - 1)
|
1994-05-24 10:09:53 +00:00
|
|
|
trylater_cnt++;
|
|
|
|
goto tryagain;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If the File Handle was stale, invalidate the
|
|
|
|
* lookup cache, just in case.
|
|
|
|
*/
|
|
|
|
if (error == ESTALE)
|
|
|
|
cache_purge(vp);
|
1995-06-27 11:07:30 +00:00
|
|
|
if (nmp->nm_flag & NFSMNT_NFSV3) {
|
|
|
|
*mrp = mrep;
|
|
|
|
*mdp = md;
|
|
|
|
*dposp = dpos;
|
|
|
|
error |= NFSERR_RETERR;
|
|
|
|
} else
|
|
|
|
m_freem(mrep);
|
1994-05-24 10:09:53 +00:00
|
|
|
m_freem(rep->r_mreq);
|
|
|
|
free((caddr_t)rep, M_NFSREQ);
|
|
|
|
return (error);
|
|
|
|
}
|
|
|
|
|
|
|
|
*mrp = mrep;
|
|
|
|
*mdp = md;
|
|
|
|
*dposp = dpos;
|
|
|
|
m_freem(rep->r_mreq);
|
|
|
|
FREE((caddr_t)rep, M_NFSREQ);
|
|
|
|
return (0);
|
|
|
|
}
|
|
|
|
m_freem(mrep);
|
|
|
|
error = EPROTONOSUPPORT;
|
|
|
|
nfsmout:
|
1995-06-27 11:07:30 +00:00
|
|
|
m_freem(rep->r_mreq);
|
|
|
|
free((caddr_t)rep, M_NFSREQ);
|
1994-05-24 10:09:53 +00:00
|
|
|
return (error);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Nfs timer routine
|
|
|
|
* Scan the nfsreq list and retranmit any requests that have timed out
|
|
|
|
* To avoid retransmission attempts on STREAM sockets (in the future) make
|
|
|
|
* sure to set the r_retry field to 0 (implies nm_retry == 0).
|
|
|
|
*/
|
|
|
|
void
|
2001-09-18 23:32:09 +00:00
|
|
|
nfs_timer(void *arg)
|
1994-05-24 10:09:53 +00:00
|
|
|
{
|
2001-09-18 23:32:09 +00:00
|
|
|
struct nfsreq *rep;
|
|
|
|
struct mbuf *m;
|
|
|
|
struct socket *so;
|
|
|
|
struct nfsmount *nmp;
|
|
|
|
int timeo;
|
1994-05-24 10:09:53 +00:00
|
|
|
int s, error;
|
2002-02-07 23:03:41 +00:00
|
|
|
struct thread *td;
|
1994-05-24 10:09:53 +00:00
|
|
|
|
2002-02-07 23:03:41 +00:00
|
|
|
td = &thread0; /* XXX for credentials, may break if sleep */
|
1994-05-24 10:09:53 +00:00
|
|
|
s = splnet();
|
2001-09-18 23:32:09 +00:00
|
|
|
TAILQ_FOREACH(rep, &nfs_reqq, r_chain) {
|
1994-05-24 10:09:53 +00:00
|
|
|
nmp = rep->r_nmp;
|
|
|
|
if (rep->r_mrep || (rep->r_flags & R_SOFTTERM))
|
|
|
|
continue;
|
2002-06-28 21:53:08 +00:00
|
|
|
if (nfs_sigintr(nmp, rep, rep->r_td)) {
|
1999-12-13 04:24:55 +00:00
|
|
|
nfs_softterm(rep);
|
1994-05-24 10:09:53 +00:00
|
|
|
continue;
|
|
|
|
}
|
|
|
|
if (rep->r_rtt >= 0) {
|
|
|
|
rep->r_rtt++;
|
|
|
|
if (nmp->nm_flag & NFSMNT_DUMBTIMR)
|
|
|
|
timeo = nmp->nm_timeo;
|
|
|
|
else
|
|
|
|
timeo = NFS_RTO(nmp, proct[rep->r_procnum]);
|
|
|
|
if (nmp->nm_timeouts > 0)
|
|
|
|
timeo *= nfs_backoff[nmp->nm_timeouts - 1];
|
|
|
|
if (rep->r_rtt <= timeo)
|
|
|
|
continue;
|
2001-12-30 18:41:52 +00:00
|
|
|
if (nmp->nm_timeouts < NFS_NBACKOFF)
|
1994-05-24 10:09:53 +00:00
|
|
|
nmp->nm_timeouts++;
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* Check for server not responding
|
|
|
|
*/
|
|
|
|
if ((rep->r_flags & R_TPRINTFMSG) == 0 &&
|
|
|
|
rep->r_rexmit > nmp->nm_deadthresh) {
|
2001-09-12 08:38:13 +00:00
|
|
|
nfs_msg(rep->r_td,
|
1994-05-24 10:09:53 +00:00
|
|
|
nmp->nm_mountp->mnt_stat.f_mntfromname,
|
2003-02-18 23:45:01 +00:00
|
|
|
"not responding");
|
1994-05-24 10:09:53 +00:00
|
|
|
rep->r_flags |= R_TPRINTFMSG;
|
|
|
|
}
|
|
|
|
if (rep->r_rexmit >= rep->r_retry) { /* too many */
|
|
|
|
nfsstats.rpctimeouts++;
|
1999-12-13 04:24:55 +00:00
|
|
|
nfs_softterm(rep);
|
1994-05-24 10:09:53 +00:00
|
|
|
continue;
|
|
|
|
}
|
|
|
|
if (nmp->nm_sotype != SOCK_DGRAM) {
|
|
|
|
if (++rep->r_rexmit > NFS_MAXREXMIT)
|
|
|
|
rep->r_rexmit = NFS_MAXREXMIT;
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
if ((so = nmp->nm_so) == NULL)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If there is enough space and the window allows..
|
|
|
|
* Resend it
|
|
|
|
* Set r_rtt to -1 in case we fail to send it now.
|
|
|
|
*/
|
|
|
|
rep->r_rtt = -1;
|
|
|
|
if (sbspace(&so->so_snd) >= rep->r_mreq->m_pkthdr.len &&
|
|
|
|
((nmp->nm_flag & NFSMNT_DUMBTIMR) ||
|
|
|
|
(rep->r_flags & R_SENT) ||
|
|
|
|
nmp->nm_sent < nmp->nm_cwnd) &&
|
2003-02-19 05:47:46 +00:00
|
|
|
(m = m_copym(rep->r_mreq, 0, M_COPYALL, M_DONTWAIT))){
|
1994-05-24 10:09:53 +00:00
|
|
|
if ((nmp->nm_flag & NFSMNT_NOCONN) == 0)
|
1996-07-11 16:32:50 +00:00
|
|
|
error = (*so->so_proto->pr_usrreqs->pru_send)
|
2002-07-11 17:54:58 +00:00
|
|
|
(so, 0, m, NULL, NULL, td);
|
1994-05-24 10:09:53 +00:00
|
|
|
else
|
1996-07-11 16:32:50 +00:00
|
|
|
error = (*so->so_proto->pr_usrreqs->pru_send)
|
2002-07-11 17:54:58 +00:00
|
|
|
(so, 0, m, nmp->nm_nam, NULL, td);
|
1994-05-24 10:09:53 +00:00
|
|
|
if (error) {
|
|
|
|
if (NFSIGNORE_SOERROR(nmp->nm_soflags, error))
|
|
|
|
so->so_error = 0;
|
|
|
|
} else {
|
|
|
|
/*
|
|
|
|
* Iff first send, start timing
|
|
|
|
* else turn timing off, backoff timer
|
|
|
|
* and divide congestion window by 2.
|
|
|
|
*/
|
|
|
|
if (rep->r_flags & R_SENT) {
|
|
|
|
rep->r_flags &= ~R_TIMING;
|
|
|
|
if (++rep->r_rexmit > NFS_MAXREXMIT)
|
|
|
|
rep->r_rexmit = NFS_MAXREXMIT;
|
|
|
|
nmp->nm_cwnd >>= 1;
|
|
|
|
if (nmp->nm_cwnd < NFS_CWNDSCALE)
|
|
|
|
nmp->nm_cwnd = NFS_CWNDSCALE;
|
|
|
|
nfsstats.rpcretries++;
|
|
|
|
} else {
|
|
|
|
rep->r_flags |= R_SENT;
|
|
|
|
nmp->nm_sent += NFS_CWNDSCALE;
|
|
|
|
}
|
|
|
|
rep->r_rtt = 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
splx(s);
|
|
|
|
}
|
|
|
|
|
2002-01-02 00:41:26 +00:00
|
|
|
/*
|
|
|
|
* Mark all of an nfs mount's outstanding requests with R_SOFTTERM and
|
|
|
|
* wait for all requests to complete. This is used by forced unmounts
|
|
|
|
* to terminate any outstanding RPCs.
|
|
|
|
*/
|
|
|
|
int
|
|
|
|
nfs_nmcancelreqs(nmp)
|
|
|
|
struct nfsmount *nmp;
|
|
|
|
{
|
|
|
|
struct nfsreq *req;
|
|
|
|
int i, s;
|
|
|
|
|
|
|
|
s = splnet();
|
|
|
|
TAILQ_FOREACH(req, &nfs_reqq, r_chain) {
|
|
|
|
if (nmp != req->r_nmp || req->r_mrep != NULL ||
|
|
|
|
(req->r_flags & R_SOFTTERM))
|
|
|
|
continue;
|
|
|
|
nfs_softterm(req);
|
|
|
|
}
|
|
|
|
splx(s);
|
|
|
|
|
|
|
|
for (i = 0; i < 30; i++) {
|
|
|
|
s = splnet();
|
|
|
|
TAILQ_FOREACH(req, &nfs_reqq, r_chain) {
|
|
|
|
if (nmp == req->r_nmp)
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
splx(s);
|
|
|
|
if (req == NULL)
|
|
|
|
return (0);
|
2002-01-10 02:15:35 +00:00
|
|
|
tsleep(&lbolt, PSOCK, "nfscancel", 0);
|
2002-01-02 00:41:26 +00:00
|
|
|
}
|
|
|
|
return (EBUSY);
|
|
|
|
}
|
|
|
|
|
1999-12-13 04:24:55 +00:00
|
|
|
/*
|
|
|
|
* Flag a request as being about to terminate (due to NFSMNT_INT/NFSMNT_SOFT).
|
|
|
|
* The nm_send count is decremented now to avoid deadlocks when the process in
|
|
|
|
* soreceive() hasn't yet managed to send its own request.
|
|
|
|
*/
|
|
|
|
|
|
|
|
static void
|
2001-09-18 23:32:09 +00:00
|
|
|
nfs_softterm(struct nfsreq *rep)
|
1999-12-13 04:24:55 +00:00
|
|
|
{
|
|
|
|
|
2001-09-18 23:32:09 +00:00
|
|
|
rep->r_flags |= R_SOFTTERM;
|
1999-12-13 04:24:55 +00:00
|
|
|
if (rep->r_flags & R_SENT) {
|
|
|
|
rep->r_nmp->nm_sent -= NFS_CWNDSCALE;
|
|
|
|
rep->r_flags &= ~R_SENT;
|
|
|
|
}
|
|
|
|
}
|
1996-01-13 23:27:58 +00:00
|
|
|
|
1994-05-24 10:09:53 +00:00
|
|
|
/*
|
|
|
|
* Test for a termination condition pending on the process.
|
|
|
|
* This is used for NFSMNT_INT mounts.
|
|
|
|
*/
|
1994-05-25 09:21:21 +00:00
|
|
|
int
|
2002-06-28 21:53:08 +00:00
|
|
|
nfs_sigintr(struct nfsmount *nmp, struct nfsreq *rep, struct thread *td)
|
1994-05-24 10:09:53 +00:00
|
|
|
{
|
2002-06-28 21:53:08 +00:00
|
|
|
struct proc *p;
|
1999-09-29 15:03:48 +00:00
|
|
|
sigset_t tmpset;
|
1994-05-24 10:09:53 +00:00
|
|
|
|
2003-11-14 20:54:10 +00:00
|
|
|
if ((nmp->nm_flag & NFSMNT_NFSV4) != 0)
|
|
|
|
return nfs4_sigintr(nmp, rep, td);
|
1994-05-24 10:09:53 +00:00
|
|
|
if (rep && (rep->r_flags & R_SOFTTERM))
|
|
|
|
return (EINTR);
|
2002-04-17 01:07:29 +00:00
|
|
|
/* Terminate all requests while attempting a forced unmount. */
|
|
|
|
if (nmp->nm_mountp->mnt_kern_flag & MNTK_UNMOUNTF)
|
2002-01-10 02:15:35 +00:00
|
|
|
return (EINTR);
|
1994-05-24 10:09:53 +00:00
|
|
|
if (!(nmp->nm_flag & NFSMNT_INT))
|
|
|
|
return (0);
|
2002-06-28 21:53:08 +00:00
|
|
|
if (td == NULL)
|
1999-09-29 20:12:39 +00:00
|
|
|
return (0);
|
|
|
|
|
2002-06-28 21:53:08 +00:00
|
|
|
p = td->td_proc;
|
2003-02-15 08:25:57 +00:00
|
|
|
PROC_LOCK(p);
|
2002-10-01 17:15:53 +00:00
|
|
|
tmpset = p->p_siglist;
|
2003-03-31 22:49:17 +00:00
|
|
|
SIGSETNAND(tmpset, td->td_sigmask);
|
- Merge struct procsig with struct sigacts.
- Move struct sigacts out of the u-area and malloc() it using the
M_SUBPROC malloc bucket.
- Add a small sigacts_*() API for managing sigacts structures: sigacts_alloc(),
sigacts_free(), sigacts_copy(), sigacts_share(), and sigacts_shared().
- Remove the p_sigignore, p_sigacts, and p_sigcatch macros.
- Add a mutex to struct sigacts that protects all the members of the struct.
- Add sigacts locking.
- Remove Giant from nosys(), kill(), killpg(), and kern_sigaction() now
that sigacts is locked.
- Several in-kernel functions such as psignal(), tdsignal(), trapsignal(),
and thread_stopped() are now MP safe.
Reviewed by: arch@
Approved by: re (rwatson)
2003-05-13 20:36:02 +00:00
|
|
|
mtx_lock(&p->p_sigacts->ps_mtx);
|
|
|
|
SIGSETNAND(tmpset, p->p_sigacts->ps_sigignore);
|
|
|
|
mtx_unlock(&p->p_sigacts->ps_mtx);
|
2003-02-15 08:25:57 +00:00
|
|
|
if (SIGNOTEMPTY(p->p_siglist) && NFSINT_SIGMASK(tmpset)) {
|
|
|
|
PROC_UNLOCK(p);
|
1994-05-24 10:09:53 +00:00
|
|
|
return (EINTR);
|
2003-02-15 08:25:57 +00:00
|
|
|
}
|
|
|
|
PROC_UNLOCK(p);
|
1999-09-29 20:12:39 +00:00
|
|
|
|
1994-05-24 10:09:53 +00:00
|
|
|
return (0);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Lock a socket against others.
|
|
|
|
* Necessary for STREAM sockets to ensure you get an entire rpc request/reply
|
|
|
|
* and also to avoid race conditions between the processes with nfs requests
|
|
|
|
* in progress when a reconnect is necessary.
|
|
|
|
*/
|
1994-05-25 09:21:21 +00:00
|
|
|
int
|
2001-09-18 23:32:09 +00:00
|
|
|
nfs_sndlock(struct nfsreq *rep)
|
1994-05-24 10:09:53 +00:00
|
|
|
{
|
2001-09-18 23:32:09 +00:00
|
|
|
int *statep = &rep->r_nmp->nm_state;
|
2001-09-12 08:38:13 +00:00
|
|
|
struct thread *td;
|
1994-05-24 10:09:53 +00:00
|
|
|
int slpflag = 0, slptimeo = 0;
|
|
|
|
|
2003-03-26 01:46:11 +00:00
|
|
|
td = rep->r_td;
|
|
|
|
if (rep->r_nmp->nm_flag & NFSMNT_INT)
|
|
|
|
slpflag = PCATCH;
|
1998-05-19 07:11:27 +00:00
|
|
|
while (*statep & NFSSTA_SNDLOCK) {
|
2002-06-28 21:53:08 +00:00
|
|
|
if (nfs_sigintr(rep->r_nmp, rep, td))
|
1994-05-24 10:09:53 +00:00
|
|
|
return (EINTR);
|
1998-05-19 07:11:27 +00:00
|
|
|
*statep |= NFSSTA_WANTSND;
|
2003-03-02 16:54:40 +00:00
|
|
|
(void) tsleep(statep, slpflag | (PZERO - 1),
|
1998-05-19 07:11:27 +00:00
|
|
|
"nfsndlck", slptimeo);
|
1994-05-24 10:09:53 +00:00
|
|
|
if (slpflag == PCATCH) {
|
|
|
|
slpflag = 0;
|
|
|
|
slptimeo = 2 * hz;
|
|
|
|
}
|
|
|
|
}
|
1998-05-19 07:11:27 +00:00
|
|
|
*statep |= NFSSTA_SNDLOCK;
|
1994-05-24 10:09:53 +00:00
|
|
|
return (0);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Unlock the stream socket for others.
|
|
|
|
*/
|
|
|
|
void
|
2001-09-18 23:32:09 +00:00
|
|
|
nfs_sndunlock(struct nfsreq *rep)
|
1994-05-24 10:09:53 +00:00
|
|
|
{
|
2001-09-18 23:32:09 +00:00
|
|
|
int *statep = &rep->r_nmp->nm_state;
|
1994-05-24 10:09:53 +00:00
|
|
|
|
1998-05-19 07:11:27 +00:00
|
|
|
if ((*statep & NFSSTA_SNDLOCK) == 0)
|
1994-05-24 10:09:53 +00:00
|
|
|
panic("nfs sndunlock");
|
1998-05-19 07:11:27 +00:00
|
|
|
*statep &= ~NFSSTA_SNDLOCK;
|
|
|
|
if (*statep & NFSSTA_WANTSND) {
|
|
|
|
*statep &= ~NFSSTA_WANTSND;
|
2003-03-02 16:54:40 +00:00
|
|
|
wakeup(statep);
|
1994-05-24 10:09:53 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
1995-12-17 21:14:36 +00:00
|
|
|
static int
|
2001-09-18 23:32:09 +00:00
|
|
|
nfs_rcvlock(struct nfsreq *rep)
|
1994-05-24 10:09:53 +00:00
|
|
|
{
|
2001-09-18 23:32:09 +00:00
|
|
|
int *statep = &rep->r_nmp->nm_state;
|
1994-05-24 10:09:53 +00:00
|
|
|
int slpflag, slptimeo = 0;
|
|
|
|
|
1999-02-25 00:03:51 +00:00
|
|
|
if (rep->r_nmp->nm_flag & NFSMNT_INT)
|
1994-05-24 10:09:53 +00:00
|
|
|
slpflag = PCATCH;
|
|
|
|
else
|
|
|
|
slpflag = 0;
|
1998-05-19 07:11:27 +00:00
|
|
|
while (*statep & NFSSTA_RCVLOCK) {
|
2002-06-28 21:53:08 +00:00
|
|
|
if (nfs_sigintr(rep->r_nmp, rep, rep->r_td))
|
1994-05-24 10:09:53 +00:00
|
|
|
return (EINTR);
|
1998-05-19 07:11:27 +00:00
|
|
|
*statep |= NFSSTA_WANTRCV;
|
2003-03-02 16:54:40 +00:00
|
|
|
(void) tsleep(statep, slpflag | (PZERO - 1), "nfsrcvlk",
|
1994-05-24 10:09:53 +00:00
|
|
|
slptimeo);
|
1996-10-11 10:15:33 +00:00
|
|
|
/*
|
|
|
|
* If our reply was recieved while we were sleeping,
|
|
|
|
* then just return without taking the lock to avoid a
|
|
|
|
* situation where a single iod could 'capture' the
|
|
|
|
* recieve lock.
|
|
|
|
*/
|
|
|
|
if (rep->r_mrep != NULL)
|
|
|
|
return (EALREADY);
|
1994-05-24 10:09:53 +00:00
|
|
|
if (slpflag == PCATCH) {
|
|
|
|
slpflag = 0;
|
|
|
|
slptimeo = 2 * hz;
|
|
|
|
}
|
|
|
|
}
|
2002-01-02 00:41:26 +00:00
|
|
|
/* Always fail if our request has been cancelled. */
|
|
|
|
if (rep != NULL && (rep->r_flags & R_SOFTTERM))
|
|
|
|
return (EINTR);
|
1998-05-19 07:11:27 +00:00
|
|
|
*statep |= NFSSTA_RCVLOCK;
|
1994-05-24 10:09:53 +00:00
|
|
|
return (0);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Unlock the stream socket for others.
|
|
|
|
*/
|
1995-12-17 21:14:36 +00:00
|
|
|
static void
|
2001-09-18 23:32:09 +00:00
|
|
|
nfs_rcvunlock(struct nfsreq *rep)
|
1994-05-24 10:09:53 +00:00
|
|
|
{
|
2001-09-18 23:32:09 +00:00
|
|
|
int *statep = &rep->r_nmp->nm_state;
|
1994-05-24 10:09:53 +00:00
|
|
|
|
1998-05-19 07:11:27 +00:00
|
|
|
if ((*statep & NFSSTA_RCVLOCK) == 0)
|
1994-05-24 10:09:53 +00:00
|
|
|
panic("nfs rcvunlock");
|
1998-05-19 07:11:27 +00:00
|
|
|
*statep &= ~NFSSTA_RCVLOCK;
|
|
|
|
if (*statep & NFSSTA_WANTRCV) {
|
|
|
|
*statep &= ~NFSSTA_WANTRCV;
|
2003-03-02 16:54:40 +00:00
|
|
|
wakeup(statep);
|
1994-05-24 10:09:53 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
The VFS/BIO subsystem contained a number of hacks in order to optimize
piecemeal, middle-of-file writes for NFS. These hacks have caused no
end of trouble, especially when combined with mmap(). I've removed
them. Instead, NFS will issue a read-before-write to fully
instantiate the struct buf containing the write. NFS does, however,
optimize piecemeal appends to files. For most common file operations,
you will not notice the difference. The sole remaining fragment in
the VFS/BIO system is b_dirtyoff/end, which NFS uses to avoid cache
coherency issues with read-merge-write style operations. NFS also
optimizes the write-covers-entire-buffer case by avoiding the
read-before-write. There is quite a bit of room for further
optimization in these areas.
The VM system marks pages fully-valid (AKA vm_page_t->valid =
VM_PAGE_BITS_ALL) in several places, most noteably in vm_fault. This
is not correct operation. The vm_pager_get_pages() code is now
responsible for marking VM pages all-valid. A number of VM helper
routines have been added to aid in zeroing-out the invalid portions of
a VM page prior to the page being marked all-valid. This operation is
necessary to properly support mmap(). The zeroing occurs most often
when dealing with file-EOF situations. Several bugs have been fixed
in the NFS subsystem, including bits handling file and directory EOF
situations and buf->b_flags consistancy issues relating to clearing
B_ERROR & B_INVAL, and handling B_DONE.
getblk() and allocbuf() have been rewritten. B_CACHE operation is now
formally defined in comments and more straightforward in
implementation. B_CACHE for VMIO buffers is based on the validity of
the backing store. B_CACHE for non-VMIO buffers is based simply on
whether the buffer is B_INVAL or not (B_CACHE set if B_INVAL clear,
and vise-versa). biodone() is now responsible for setting B_CACHE
when a successful read completes. B_CACHE is also set when a bdwrite()
is initiated and when a bwrite() is initiated. VFS VOP_BWRITE
routines (there are only two - nfs_bwrite() and bwrite()) are now
expected to set B_CACHE. This means that bowrite() and bawrite() also
set B_CACHE indirectly.
There are a number of places in the code which were previously using
buf->b_bufsize (which is DEV_BSIZE aligned) when they should have
been using buf->b_bcount. These have been fixed. getblk() now clears
B_DONE on return because the rest of the system is so bad about
dealing with B_DONE.
Major fixes to NFS/TCP have been made. A server-side bug could cause
requests to be lost by the server due to nfs_realign() overwriting
other rpc's in the same TCP mbuf chain. The server's kernel must be
recompiled to get the benefit of the fixes.
Submitted by: Matthew Dillon <dillon@apollo.backplane.com>
1999-05-02 23:57:16 +00:00
|
|
|
* nfs_realign:
|
|
|
|
*
|
|
|
|
* Check for badly aligned mbuf data and realign by copying the unaligned
|
|
|
|
* portion of the data into a new mbuf chain and freeing the portions
|
|
|
|
* of the old chain that were replaced.
|
|
|
|
*
|
|
|
|
* We cannot simply realign the data within the existing mbuf chain
|
|
|
|
* because the underlying buffers may contain other rpc commands and
|
|
|
|
* we cannot afford to overwrite them.
|
|
|
|
*
|
|
|
|
* We would prefer to avoid this situation entirely. The situation does
|
|
|
|
* not occur with NFS/UDP and is supposed to only occassionally occur
|
|
|
|
* with TCP. Use vfs.nfs.realign_count and realign_test to check this.
|
1994-05-24 10:09:53 +00:00
|
|
|
*/
|
1995-12-17 21:14:36 +00:00
|
|
|
static void
|
2001-09-18 23:32:09 +00:00
|
|
|
nfs_realign(struct mbuf **pm, int hsiz)
|
1994-05-24 10:09:53 +00:00
|
|
|
{
|
The VFS/BIO subsystem contained a number of hacks in order to optimize
piecemeal, middle-of-file writes for NFS. These hacks have caused no
end of trouble, especially when combined with mmap(). I've removed
them. Instead, NFS will issue a read-before-write to fully
instantiate the struct buf containing the write. NFS does, however,
optimize piecemeal appends to files. For most common file operations,
you will not notice the difference. The sole remaining fragment in
the VFS/BIO system is b_dirtyoff/end, which NFS uses to avoid cache
coherency issues with read-merge-write style operations. NFS also
optimizes the write-covers-entire-buffer case by avoiding the
read-before-write. There is quite a bit of room for further
optimization in these areas.
The VM system marks pages fully-valid (AKA vm_page_t->valid =
VM_PAGE_BITS_ALL) in several places, most noteably in vm_fault. This
is not correct operation. The vm_pager_get_pages() code is now
responsible for marking VM pages all-valid. A number of VM helper
routines have been added to aid in zeroing-out the invalid portions of
a VM page prior to the page being marked all-valid. This operation is
necessary to properly support mmap(). The zeroing occurs most often
when dealing with file-EOF situations. Several bugs have been fixed
in the NFS subsystem, including bits handling file and directory EOF
situations and buf->b_flags consistancy issues relating to clearing
B_ERROR & B_INVAL, and handling B_DONE.
getblk() and allocbuf() have been rewritten. B_CACHE operation is now
formally defined in comments and more straightforward in
implementation. B_CACHE for VMIO buffers is based on the validity of
the backing store. B_CACHE for non-VMIO buffers is based simply on
whether the buffer is B_INVAL or not (B_CACHE set if B_INVAL clear,
and vise-versa). biodone() is now responsible for setting B_CACHE
when a successful read completes. B_CACHE is also set when a bdwrite()
is initiated and when a bwrite() is initiated. VFS VOP_BWRITE
routines (there are only two - nfs_bwrite() and bwrite()) are now
expected to set B_CACHE. This means that bowrite() and bawrite() also
set B_CACHE indirectly.
There are a number of places in the code which were previously using
buf->b_bufsize (which is DEV_BSIZE aligned) when they should have
been using buf->b_bcount. These have been fixed. getblk() now clears
B_DONE on return because the rest of the system is so bad about
dealing with B_DONE.
Major fixes to NFS/TCP have been made. A server-side bug could cause
requests to be lost by the server due to nfs_realign() overwriting
other rpc's in the same TCP mbuf chain. The server's kernel must be
recompiled to get the benefit of the fixes.
Submitted by: Matthew Dillon <dillon@apollo.backplane.com>
1999-05-02 23:57:16 +00:00
|
|
|
struct mbuf *m;
|
|
|
|
struct mbuf *n = NULL;
|
|
|
|
int off = 0;
|
1994-05-24 10:09:53 +00:00
|
|
|
|
The VFS/BIO subsystem contained a number of hacks in order to optimize
piecemeal, middle-of-file writes for NFS. These hacks have caused no
end of trouble, especially when combined with mmap(). I've removed
them. Instead, NFS will issue a read-before-write to fully
instantiate the struct buf containing the write. NFS does, however,
optimize piecemeal appends to files. For most common file operations,
you will not notice the difference. The sole remaining fragment in
the VFS/BIO system is b_dirtyoff/end, which NFS uses to avoid cache
coherency issues with read-merge-write style operations. NFS also
optimizes the write-covers-entire-buffer case by avoiding the
read-before-write. There is quite a bit of room for further
optimization in these areas.
The VM system marks pages fully-valid (AKA vm_page_t->valid =
VM_PAGE_BITS_ALL) in several places, most noteably in vm_fault. This
is not correct operation. The vm_pager_get_pages() code is now
responsible for marking VM pages all-valid. A number of VM helper
routines have been added to aid in zeroing-out the invalid portions of
a VM page prior to the page being marked all-valid. This operation is
necessary to properly support mmap(). The zeroing occurs most often
when dealing with file-EOF situations. Several bugs have been fixed
in the NFS subsystem, including bits handling file and directory EOF
situations and buf->b_flags consistancy issues relating to clearing
B_ERROR & B_INVAL, and handling B_DONE.
getblk() and allocbuf() have been rewritten. B_CACHE operation is now
formally defined in comments and more straightforward in
implementation. B_CACHE for VMIO buffers is based on the validity of
the backing store. B_CACHE for non-VMIO buffers is based simply on
whether the buffer is B_INVAL or not (B_CACHE set if B_INVAL clear,
and vise-versa). biodone() is now responsible for setting B_CACHE
when a successful read completes. B_CACHE is also set when a bdwrite()
is initiated and when a bwrite() is initiated. VFS VOP_BWRITE
routines (there are only two - nfs_bwrite() and bwrite()) are now
expected to set B_CACHE. This means that bowrite() and bawrite() also
set B_CACHE indirectly.
There are a number of places in the code which were previously using
buf->b_bufsize (which is DEV_BSIZE aligned) when they should have
been using buf->b_bcount. These have been fixed. getblk() now clears
B_DONE on return because the rest of the system is so bad about
dealing with B_DONE.
Major fixes to NFS/TCP have been made. A server-side bug could cause
requests to be lost by the server due to nfs_realign() overwriting
other rpc's in the same TCP mbuf chain. The server's kernel must be
recompiled to get the benefit of the fixes.
Submitted by: Matthew Dillon <dillon@apollo.backplane.com>
1999-05-02 23:57:16 +00:00
|
|
|
++nfs_realign_test;
|
|
|
|
while ((m = *pm) != NULL) {
|
|
|
|
if ((m->m_len & 0x3) || (mtod(m, intptr_t) & 0x3)) {
|
2003-02-19 05:47:46 +00:00
|
|
|
MGET(n, M_TRYWAIT, MT_DATA);
|
The VFS/BIO subsystem contained a number of hacks in order to optimize
piecemeal, middle-of-file writes for NFS. These hacks have caused no
end of trouble, especially when combined with mmap(). I've removed
them. Instead, NFS will issue a read-before-write to fully
instantiate the struct buf containing the write. NFS does, however,
optimize piecemeal appends to files. For most common file operations,
you will not notice the difference. The sole remaining fragment in
the VFS/BIO system is b_dirtyoff/end, which NFS uses to avoid cache
coherency issues with read-merge-write style operations. NFS also
optimizes the write-covers-entire-buffer case by avoiding the
read-before-write. There is quite a bit of room for further
optimization in these areas.
The VM system marks pages fully-valid (AKA vm_page_t->valid =
VM_PAGE_BITS_ALL) in several places, most noteably in vm_fault. This
is not correct operation. The vm_pager_get_pages() code is now
responsible for marking VM pages all-valid. A number of VM helper
routines have been added to aid in zeroing-out the invalid portions of
a VM page prior to the page being marked all-valid. This operation is
necessary to properly support mmap(). The zeroing occurs most often
when dealing with file-EOF situations. Several bugs have been fixed
in the NFS subsystem, including bits handling file and directory EOF
situations and buf->b_flags consistancy issues relating to clearing
B_ERROR & B_INVAL, and handling B_DONE.
getblk() and allocbuf() have been rewritten. B_CACHE operation is now
formally defined in comments and more straightforward in
implementation. B_CACHE for VMIO buffers is based on the validity of
the backing store. B_CACHE for non-VMIO buffers is based simply on
whether the buffer is B_INVAL or not (B_CACHE set if B_INVAL clear,
and vise-versa). biodone() is now responsible for setting B_CACHE
when a successful read completes. B_CACHE is also set when a bdwrite()
is initiated and when a bwrite() is initiated. VFS VOP_BWRITE
routines (there are only two - nfs_bwrite() and bwrite()) are now
expected to set B_CACHE. This means that bowrite() and bawrite() also
set B_CACHE indirectly.
There are a number of places in the code which were previously using
buf->b_bufsize (which is DEV_BSIZE aligned) when they should have
been using buf->b_bcount. These have been fixed. getblk() now clears
B_DONE on return because the rest of the system is so bad about
dealing with B_DONE.
Major fixes to NFS/TCP have been made. A server-side bug could cause
requests to be lost by the server due to nfs_realign() overwriting
other rpc's in the same TCP mbuf chain. The server's kernel must be
recompiled to get the benefit of the fixes.
Submitted by: Matthew Dillon <dillon@apollo.backplane.com>
1999-05-02 23:57:16 +00:00
|
|
|
if (m->m_len >= MINCLSIZE) {
|
2003-02-19 05:47:46 +00:00
|
|
|
MCLGET(n, M_TRYWAIT);
|
1994-05-24 10:09:53 +00:00
|
|
|
}
|
The VFS/BIO subsystem contained a number of hacks in order to optimize
piecemeal, middle-of-file writes for NFS. These hacks have caused no
end of trouble, especially when combined with mmap(). I've removed
them. Instead, NFS will issue a read-before-write to fully
instantiate the struct buf containing the write. NFS does, however,
optimize piecemeal appends to files. For most common file operations,
you will not notice the difference. The sole remaining fragment in
the VFS/BIO system is b_dirtyoff/end, which NFS uses to avoid cache
coherency issues with read-merge-write style operations. NFS also
optimizes the write-covers-entire-buffer case by avoiding the
read-before-write. There is quite a bit of room for further
optimization in these areas.
The VM system marks pages fully-valid (AKA vm_page_t->valid =
VM_PAGE_BITS_ALL) in several places, most noteably in vm_fault. This
is not correct operation. The vm_pager_get_pages() code is now
responsible for marking VM pages all-valid. A number of VM helper
routines have been added to aid in zeroing-out the invalid portions of
a VM page prior to the page being marked all-valid. This operation is
necessary to properly support mmap(). The zeroing occurs most often
when dealing with file-EOF situations. Several bugs have been fixed
in the NFS subsystem, including bits handling file and directory EOF
situations and buf->b_flags consistancy issues relating to clearing
B_ERROR & B_INVAL, and handling B_DONE.
getblk() and allocbuf() have been rewritten. B_CACHE operation is now
formally defined in comments and more straightforward in
implementation. B_CACHE for VMIO buffers is based on the validity of
the backing store. B_CACHE for non-VMIO buffers is based simply on
whether the buffer is B_INVAL or not (B_CACHE set if B_INVAL clear,
and vise-versa). biodone() is now responsible for setting B_CACHE
when a successful read completes. B_CACHE is also set when a bdwrite()
is initiated and when a bwrite() is initiated. VFS VOP_BWRITE
routines (there are only two - nfs_bwrite() and bwrite()) are now
expected to set B_CACHE. This means that bowrite() and bawrite() also
set B_CACHE indirectly.
There are a number of places in the code which were previously using
buf->b_bufsize (which is DEV_BSIZE aligned) when they should have
been using buf->b_bcount. These have been fixed. getblk() now clears
B_DONE on return because the rest of the system is so bad about
dealing with B_DONE.
Major fixes to NFS/TCP have been made. A server-side bug could cause
requests to be lost by the server due to nfs_realign() overwriting
other rpc's in the same TCP mbuf chain. The server's kernel must be
recompiled to get the benefit of the fixes.
Submitted by: Matthew Dillon <dillon@apollo.backplane.com>
1999-05-02 23:57:16 +00:00
|
|
|
n->m_len = 0;
|
|
|
|
break;
|
1994-05-24 10:09:53 +00:00
|
|
|
}
|
The VFS/BIO subsystem contained a number of hacks in order to optimize
piecemeal, middle-of-file writes for NFS. These hacks have caused no
end of trouble, especially when combined with mmap(). I've removed
them. Instead, NFS will issue a read-before-write to fully
instantiate the struct buf containing the write. NFS does, however,
optimize piecemeal appends to files. For most common file operations,
you will not notice the difference. The sole remaining fragment in
the VFS/BIO system is b_dirtyoff/end, which NFS uses to avoid cache
coherency issues with read-merge-write style operations. NFS also
optimizes the write-covers-entire-buffer case by avoiding the
read-before-write. There is quite a bit of room for further
optimization in these areas.
The VM system marks pages fully-valid (AKA vm_page_t->valid =
VM_PAGE_BITS_ALL) in several places, most noteably in vm_fault. This
is not correct operation. The vm_pager_get_pages() code is now
responsible for marking VM pages all-valid. A number of VM helper
routines have been added to aid in zeroing-out the invalid portions of
a VM page prior to the page being marked all-valid. This operation is
necessary to properly support mmap(). The zeroing occurs most often
when dealing with file-EOF situations. Several bugs have been fixed
in the NFS subsystem, including bits handling file and directory EOF
situations and buf->b_flags consistancy issues relating to clearing
B_ERROR & B_INVAL, and handling B_DONE.
getblk() and allocbuf() have been rewritten. B_CACHE operation is now
formally defined in comments and more straightforward in
implementation. B_CACHE for VMIO buffers is based on the validity of
the backing store. B_CACHE for non-VMIO buffers is based simply on
whether the buffer is B_INVAL or not (B_CACHE set if B_INVAL clear,
and vise-versa). biodone() is now responsible for setting B_CACHE
when a successful read completes. B_CACHE is also set when a bdwrite()
is initiated and when a bwrite() is initiated. VFS VOP_BWRITE
routines (there are only two - nfs_bwrite() and bwrite()) are now
expected to set B_CACHE. This means that bowrite() and bawrite() also
set B_CACHE indirectly.
There are a number of places in the code which were previously using
buf->b_bufsize (which is DEV_BSIZE aligned) when they should have
been using buf->b_bcount. These have been fixed. getblk() now clears
B_DONE on return because the rest of the system is so bad about
dealing with B_DONE.
Major fixes to NFS/TCP have been made. A server-side bug could cause
requests to be lost by the server due to nfs_realign() overwriting
other rpc's in the same TCP mbuf chain. The server's kernel must be
recompiled to get the benefit of the fixes.
Submitted by: Matthew Dillon <dillon@apollo.backplane.com>
1999-05-02 23:57:16 +00:00
|
|
|
pm = &m->m_next;
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* If n is non-NULL, loop on m copying data, then replace the
|
|
|
|
* portion of the chain that had to be realigned.
|
|
|
|
*/
|
|
|
|
if (n != NULL) {
|
|
|
|
++nfs_realign_count;
|
|
|
|
while (m) {
|
|
|
|
m_copyback(n, off, m->m_len, mtod(m, caddr_t));
|
|
|
|
off += m->m_len;
|
|
|
|
m = m->m_next;
|
1994-05-24 10:09:53 +00:00
|
|
|
}
|
The VFS/BIO subsystem contained a number of hacks in order to optimize
piecemeal, middle-of-file writes for NFS. These hacks have caused no
end of trouble, especially when combined with mmap(). I've removed
them. Instead, NFS will issue a read-before-write to fully
instantiate the struct buf containing the write. NFS does, however,
optimize piecemeal appends to files. For most common file operations,
you will not notice the difference. The sole remaining fragment in
the VFS/BIO system is b_dirtyoff/end, which NFS uses to avoid cache
coherency issues with read-merge-write style operations. NFS also
optimizes the write-covers-entire-buffer case by avoiding the
read-before-write. There is quite a bit of room for further
optimization in these areas.
The VM system marks pages fully-valid (AKA vm_page_t->valid =
VM_PAGE_BITS_ALL) in several places, most noteably in vm_fault. This
is not correct operation. The vm_pager_get_pages() code is now
responsible for marking VM pages all-valid. A number of VM helper
routines have been added to aid in zeroing-out the invalid portions of
a VM page prior to the page being marked all-valid. This operation is
necessary to properly support mmap(). The zeroing occurs most often
when dealing with file-EOF situations. Several bugs have been fixed
in the NFS subsystem, including bits handling file and directory EOF
situations and buf->b_flags consistancy issues relating to clearing
B_ERROR & B_INVAL, and handling B_DONE.
getblk() and allocbuf() have been rewritten. B_CACHE operation is now
formally defined in comments and more straightforward in
implementation. B_CACHE for VMIO buffers is based on the validity of
the backing store. B_CACHE for non-VMIO buffers is based simply on
whether the buffer is B_INVAL or not (B_CACHE set if B_INVAL clear,
and vise-versa). biodone() is now responsible for setting B_CACHE
when a successful read completes. B_CACHE is also set when a bdwrite()
is initiated and when a bwrite() is initiated. VFS VOP_BWRITE
routines (there are only two - nfs_bwrite() and bwrite()) are now
expected to set B_CACHE. This means that bowrite() and bawrite() also
set B_CACHE indirectly.
There are a number of places in the code which were previously using
buf->b_bufsize (which is DEV_BSIZE aligned) when they should have
been using buf->b_bcount. These have been fixed. getblk() now clears
B_DONE on return because the rest of the system is so bad about
dealing with B_DONE.
Major fixes to NFS/TCP have been made. A server-side bug could cause
requests to be lost by the server due to nfs_realign() overwriting
other rpc's in the same TCP mbuf chain. The server's kernel must be
recompiled to get the benefit of the fixes.
Submitted by: Matthew Dillon <dillon@apollo.backplane.com>
1999-05-02 23:57:16 +00:00
|
|
|
m_freem(*pm);
|
|
|
|
*pm = n;
|
1994-05-24 10:09:53 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
1998-05-31 17:27:58 +00:00
|
|
|
|
|
|
|
static int
|
2001-09-18 23:32:09 +00:00
|
|
|
nfs_msg(struct thread *td, char *server, char *msg)
|
1998-05-31 17:27:58 +00:00
|
|
|
{
|
|
|
|
|
2001-10-09 02:40:45 +00:00
|
|
|
tprintf(td ? td->td_proc : NULL, LOG_INFO,
|
2001-10-08 23:47:44 +00:00
|
|
|
"nfs server %s: %s\n", server, msg);
|
1998-05-31 17:27:58 +00:00
|
|
|
return (0);
|
|
|
|
}
|