a9148abd9d
and server. This replaces the RPC implementation of the NFS client and server with the newer RPC implementation originally developed (actually ported from the userland sunrpc code) to support the NFS Lock Manager. I have tested this code extensively and I believe it is stable and that performance is at least equal to the legacy RPC implementation. The NFS code currently contains support for both the new RPC implementation and the older legacy implementation inherited from the original NFS codebase. The default is to use the new implementation - add the NFS_LEGACYRPC option to fall back to the old code. When I merge this support back to RELENG_7, I will probably change this so that users have to 'opt in' to get the new code. To use RPCSEC_GSS on either client or server, you must build a kernel which includes the KGSSAPI option and the crypto device. On the userland side, you must build at least a new libc, mountd, mount_nfs and gssd. You must install new versions of /etc/rc.d/gssd and /etc/rc.d/nfsd and add 'gssd_enable=YES' to /etc/rc.conf. As long as gssd is running, you should be able to mount an NFS filesystem from a server that requires RPCSEC_GSS authentication. The mount itself can happen without any kerberos credentials but all access to the filesystem will be denied unless the accessing user has a valid ticket file in the standard place (/tmp/krb5cc_<uid>). There is currently no support for situations where the ticket file is in a different place, such as when the user logged in via SSH and has delegated credentials from that login. This restriction is also present in Solaris and Linux. In theory, we could improve this in future, possibly using Brooks Davis' implementation of variant symlinks. Supporting RPCSEC_GSS on a server is nearly as simple. You must create service creds for the server in the form 'nfs/<fqdn>@<REALM>' and install them in /etc/krb5.keytab. The standard heimdal utility ktutil makes this fairly easy. After the service creds have been created, you can add a '-sec=krb5' option to /etc/exports and restart both mountd and nfsd. The only other difference an administrator should notice is that nfsd doesn't fork to create service threads any more. In normal operation, there will be two nfsd processes, one in userland waiting for TCP connections and one in the kernel handling requests. The latter process will create as many kthreads as required - these should be visible via 'top -H'. The code has some support for varying the number of service threads according to load but initially at least, nfsd uses a fixed number of threads according to the value supplied to its '-n' option. Sponsored by: Isilon Systems MFC after: 1 month
823 lines
26 KiB
C
823 lines
26 KiB
C
/* $NetBSD: svc.h,v 1.17 2000/06/02 22:57:56 fvdl Exp $ */
|
|
|
|
/*
|
|
* Sun RPC is a product of Sun Microsystems, Inc. and is provided for
|
|
* unrestricted use provided that this legend is included on all tape
|
|
* media and as a part of the software program in whole or part. Users
|
|
* may copy or modify Sun RPC without charge, but are not authorized
|
|
* to license or distribute it to anyone else except as part of a product or
|
|
* program developed by the user.
|
|
*
|
|
* SUN RPC IS PROVIDED AS IS WITH NO WARRANTIES OF ANY KIND INCLUDING THE
|
|
* WARRANTIES OF DESIGN, MERCHANTABILITY AND FITNESS FOR A PARTICULAR
|
|
* PURPOSE, OR ARISING FROM A COURSE OF DEALING, USAGE OR TRADE PRACTICE.
|
|
*
|
|
* Sun RPC is provided with no support and without any obligation on the
|
|
* part of Sun Microsystems, Inc. to assist in its use, correction,
|
|
* modification or enhancement.
|
|
*
|
|
* SUN MICROSYSTEMS, INC. SHALL HAVE NO LIABILITY WITH RESPECT TO THE
|
|
* INFRINGEMENT OF COPYRIGHTS, TRADE SECRETS OR ANY PATENTS BY SUN RPC
|
|
* OR ANY PART THEREOF.
|
|
*
|
|
* In no event will Sun Microsystems, Inc. be liable for any lost revenue
|
|
* or profits or other special, indirect and consequential damages, even if
|
|
* Sun has been advised of the possibility of such damages.
|
|
*
|
|
* Sun Microsystems, Inc.
|
|
* 2550 Garcia Avenue
|
|
* Mountain View, California 94043
|
|
*
|
|
* from: @(#)svc.h 1.35 88/12/17 SMI
|
|
* from: @(#)svc.h 1.27 94/04/25 SMI
|
|
* $FreeBSD$
|
|
*/
|
|
|
|
/*
|
|
* svc.h, Server-side remote procedure call interface.
|
|
*
|
|
* Copyright (C) 1986-1993 by Sun Microsystems, Inc.
|
|
*/
|
|
|
|
#ifndef _RPC_SVC_H
|
|
#define _RPC_SVC_H
|
|
#include <sys/cdefs.h>
|
|
|
|
#ifdef _KERNEL
|
|
#include <sys/queue.h>
|
|
#include <sys/_lock.h>
|
|
#include <sys/_mutex.h>
|
|
#include <sys/_sx.h>
|
|
#include <sys/condvar.h>
|
|
#include <sys/sysctl.h>
|
|
#endif
|
|
|
|
/*
|
|
* This interface must manage two items concerning remote procedure calling:
|
|
*
|
|
* 1) An arbitrary number of transport connections upon which rpc requests
|
|
* are received. The two most notable transports are TCP and UDP; they are
|
|
* created and registered by routines in svc_tcp.c and svc_udp.c, respectively;
|
|
* they in turn call xprt_register and xprt_unregister.
|
|
*
|
|
* 2) An arbitrary number of locally registered services. Services are
|
|
* described by the following four data: program number, version number,
|
|
* "service dispatch" function, a transport handle, and a boolean that
|
|
* indicates whether or not the exported program should be registered with a
|
|
* local binder service; if true the program's number and version and the
|
|
* port number from the transport handle are registered with the binder.
|
|
* These data are registered with the rpc svc system via svc_register.
|
|
*
|
|
* A service's dispatch function is called whenever an rpc request comes in
|
|
* on a transport. The request's program and version numbers must match
|
|
* those of the registered service. The dispatch function is passed two
|
|
* parameters, struct svc_req * and SVCXPRT *, defined below.
|
|
*/
|
|
|
|
/*
|
|
* Service control requests
|
|
*/
|
|
#define SVCGET_VERSQUIET 1
|
|
#define SVCSET_VERSQUIET 2
|
|
#define SVCGET_CONNMAXREC 3
|
|
#define SVCSET_CONNMAXREC 4
|
|
|
|
/*
|
|
* Operations for rpc_control().
|
|
*/
|
|
#define RPC_SVC_CONNMAXREC_SET 0 /* set max rec size, enable nonblock */
|
|
#define RPC_SVC_CONNMAXREC_GET 1
|
|
|
|
enum xprt_stat {
|
|
XPRT_DIED,
|
|
XPRT_MOREREQS,
|
|
XPRT_IDLE
|
|
};
|
|
|
|
struct __rpc_svcxprt;
|
|
struct mbuf;
|
|
|
|
struct xp_ops {
|
|
#ifdef _KERNEL
|
|
/* receive incoming requests */
|
|
bool_t (*xp_recv)(struct __rpc_svcxprt *, struct rpc_msg *,
|
|
struct sockaddr **, struct mbuf **);
|
|
/* get transport status */
|
|
enum xprt_stat (*xp_stat)(struct __rpc_svcxprt *);
|
|
/* send reply */
|
|
bool_t (*xp_reply)(struct __rpc_svcxprt *, struct rpc_msg *,
|
|
struct sockaddr *, struct mbuf *);
|
|
/* destroy this struct */
|
|
void (*xp_destroy)(struct __rpc_svcxprt *);
|
|
/* catch-all function */
|
|
bool_t (*xp_control)(struct __rpc_svcxprt *, const u_int, void *);
|
|
#else
|
|
/* receive incoming requests */
|
|
bool_t (*xp_recv)(struct __rpc_svcxprt *, struct rpc_msg *);
|
|
/* get transport status */
|
|
enum xprt_stat (*xp_stat)(struct __rpc_svcxprt *);
|
|
/* get arguments */
|
|
bool_t (*xp_getargs)(struct __rpc_svcxprt *, xdrproc_t, void *);
|
|
/* send reply */
|
|
bool_t (*xp_reply)(struct __rpc_svcxprt *, struct rpc_msg *);
|
|
/* free mem allocated for args */
|
|
bool_t (*xp_freeargs)(struct __rpc_svcxprt *, xdrproc_t, void *);
|
|
/* destroy this struct */
|
|
void (*xp_destroy)(struct __rpc_svcxprt *);
|
|
#endif
|
|
};
|
|
|
|
#ifndef _KERNEL
|
|
struct xp_ops2 {
|
|
/* catch-all function */
|
|
bool_t (*xp_control)(struct __rpc_svcxprt *, const u_int, void *);
|
|
};
|
|
#endif
|
|
|
|
#ifdef _KERNEL
|
|
struct __rpc_svcpool;
|
|
struct __rpc_svcthread;
|
|
#endif
|
|
|
|
/*
|
|
* Server side transport handle. In the kernel, transports have a
|
|
* reference count which tracks the number of currently assigned
|
|
* worker threads plus one for the service pool's reference.
|
|
*/
|
|
typedef struct __rpc_svcxprt {
|
|
#ifdef _KERNEL
|
|
volatile u_int xp_refs;
|
|
struct sx xp_lock;
|
|
struct __rpc_svcpool *xp_pool; /* owning pool (see below) */
|
|
TAILQ_ENTRY(__rpc_svcxprt) xp_link;
|
|
TAILQ_ENTRY(__rpc_svcxprt) xp_alink;
|
|
bool_t xp_registered; /* xprt_register has been called */
|
|
bool_t xp_active; /* xprt_active has been called */
|
|
struct __rpc_svcthread *xp_thread; /* assigned service thread */
|
|
struct socket* xp_socket;
|
|
const struct xp_ops *xp_ops;
|
|
char *xp_netid; /* network token */
|
|
struct sockaddr_storage xp_ltaddr; /* local transport address */
|
|
struct sockaddr_storage xp_rtaddr; /* remote transport address */
|
|
void *xp_p1; /* private: for use by svc ops */
|
|
void *xp_p2; /* private: for use by svc ops */
|
|
void *xp_p3; /* private: for use by svc lib */
|
|
int xp_type; /* transport type */
|
|
int xp_idletimeout; /* idle time before closing */
|
|
time_t xp_lastactive; /* time of last RPC */
|
|
#else
|
|
int xp_fd;
|
|
u_short xp_port; /* associated port number */
|
|
const struct xp_ops *xp_ops;
|
|
int xp_addrlen; /* length of remote address */
|
|
struct sockaddr_in xp_raddr; /* remote addr. (backward ABI compat) */
|
|
/* XXX - fvdl stick this here for ABI backward compat reasons */
|
|
const struct xp_ops2 *xp_ops2;
|
|
char *xp_tp; /* transport provider device name */
|
|
char *xp_netid; /* network token */
|
|
struct netbuf xp_ltaddr; /* local transport address */
|
|
struct netbuf xp_rtaddr; /* remote transport address */
|
|
struct opaque_auth xp_verf; /* raw response verifier */
|
|
void *xp_p1; /* private: for use by svc ops */
|
|
void *xp_p2; /* private: for use by svc ops */
|
|
void *xp_p3; /* private: for use by svc lib */
|
|
int xp_type; /* transport type */
|
|
#endif
|
|
} SVCXPRT;
|
|
|
|
/*
|
|
* Interface to server-side authentication flavors.
|
|
*/
|
|
typedef struct __rpc_svcauth {
|
|
struct svc_auth_ops {
|
|
#ifdef _KERNEL
|
|
int (*svc_ah_wrap)(struct __rpc_svcauth *, struct mbuf **);
|
|
int (*svc_ah_unwrap)(struct __rpc_svcauth *, struct mbuf **);
|
|
void (*svc_ah_release)(struct __rpc_svcauth *);
|
|
#else
|
|
int (*svc_ah_wrap)(struct __rpc_svcauth *, XDR *,
|
|
xdrproc_t, caddr_t);
|
|
int (*svc_ah_unwrap)(struct __rpc_svcauth *, XDR *,
|
|
xdrproc_t, caddr_t);
|
|
#endif
|
|
} *svc_ah_ops;
|
|
void *svc_ah_private;
|
|
} SVCAUTH;
|
|
|
|
/*
|
|
* Server transport extensions (accessed via xp_p3).
|
|
*/
|
|
typedef struct __rpc_svcxprt_ext {
|
|
int xp_flags; /* versquiet */
|
|
SVCAUTH xp_auth; /* interface to auth methods */
|
|
} SVCXPRT_EXT;
|
|
|
|
#ifdef _KERNEL
|
|
|
|
/*
|
|
* The services list
|
|
* Each entry represents a set of procedures (an rpc program).
|
|
* The dispatch routine takes request structs and runs the
|
|
* apropriate procedure.
|
|
*/
|
|
struct svc_callout {
|
|
TAILQ_ENTRY(svc_callout) sc_link;
|
|
rpcprog_t sc_prog;
|
|
rpcvers_t sc_vers;
|
|
char *sc_netid;
|
|
void (*sc_dispatch)(struct svc_req *, SVCXPRT *);
|
|
};
|
|
TAILQ_HEAD(svc_callout_list, svc_callout);
|
|
|
|
struct __rpc_svcthread;
|
|
|
|
/*
|
|
* Service request
|
|
*/
|
|
struct svc_req {
|
|
STAILQ_ENTRY(svc_req) rq_link; /* list of requests for a thread */
|
|
struct __rpc_svcthread *rq_thread; /* thread which is to execute this */
|
|
uint32_t rq_xid; /* RPC transaction ID */
|
|
uint32_t rq_prog; /* service program number */
|
|
uint32_t rq_vers; /* service protocol version */
|
|
uint32_t rq_proc; /* the desired procedure */
|
|
size_t rq_size; /* space used by request */
|
|
struct mbuf *rq_args; /* XDR-encoded procedure arguments */
|
|
struct opaque_auth rq_cred; /* raw creds from the wire */
|
|
struct opaque_auth rq_verf; /* verifier for the reply */
|
|
void *rq_clntcred; /* read only cooked cred */
|
|
SVCAUTH rq_auth; /* interface to auth methods */
|
|
SVCXPRT *rq_xprt; /* associated transport */
|
|
struct sockaddr *rq_addr; /* reply address or NULL if connected */
|
|
void *rq_p1; /* application workspace */
|
|
int rq_p2; /* application workspace */
|
|
uint64_t rq_p3; /* application workspace */
|
|
char rq_credarea[3*MAX_AUTH_BYTES];
|
|
};
|
|
STAILQ_HEAD(svc_reqlist, svc_req);
|
|
|
|
#define svc_getrpccaller(rq) \
|
|
((rq)->rq_addr ? (rq)->rq_addr : \
|
|
(struct sockaddr *) &(rq)->rq_xprt->xp_rtaddr)
|
|
|
|
/*
|
|
* This structure is used to manage a thread which is executing
|
|
* requests from a service pool. A service thread is in one of three
|
|
* states:
|
|
*
|
|
* SVCTHREAD_SLEEPING waiting for a request to process
|
|
* SVCTHREAD_ACTIVE processing a request
|
|
* SVCTHREAD_EXITING exiting after finishing current request
|
|
*
|
|
* Threads which have no work to process sleep on the pool's sp_active
|
|
* list. When a transport becomes active, it is assigned a service
|
|
* thread to read and execute pending RPCs.
|
|
*/
|
|
typedef struct __rpc_svcthread {
|
|
SVCXPRT *st_xprt; /* transport we are processing */
|
|
struct svc_reqlist st_reqs; /* RPC requests to execute */
|
|
int st_reqcount; /* number of queued reqs */
|
|
struct cv st_cond; /* sleeping for work */
|
|
LIST_ENTRY(__rpc_svcthread) st_link; /* all threads list */
|
|
LIST_ENTRY(__rpc_svcthread) st_ilink; /* idle threads list */
|
|
LIST_ENTRY(__rpc_svcthread) st_alink; /* application thread list */
|
|
} SVCTHREAD;
|
|
LIST_HEAD(svcthread_list, __rpc_svcthread);
|
|
|
|
/*
|
|
* In the kernel, we can't use global variables to store lists of
|
|
* transports etc. since otherwise we could not have two unrelated RPC
|
|
* services running, each on its own thread. We solve this by
|
|
* importing a tiny part of a Solaris kernel concept, SVCPOOL.
|
|
*
|
|
* A service pool contains a set of transports and service callbacks
|
|
* for a set of related RPC services. The pool handle should be passed
|
|
* when creating new transports etc. Future work may include extending
|
|
* this to support something similar to the Solaris multi-threaded RPC
|
|
* server.
|
|
*/
|
|
TAILQ_HEAD(svcxprt_list, __rpc_svcxprt);
|
|
enum svcpool_state {
|
|
SVCPOOL_INIT, /* svc_run not called yet */
|
|
SVCPOOL_ACTIVE, /* normal running state */
|
|
SVCPOOL_THREADWANTED, /* new service thread requested */
|
|
SVCPOOL_THREADSTARTING, /* new service thread started */
|
|
SVCPOOL_CLOSING /* svc_exit called */
|
|
};
|
|
typedef SVCTHREAD *pool_assign_fn(SVCTHREAD *, struct svc_req *);
|
|
typedef void pool_done_fn(SVCTHREAD *, struct svc_req *);
|
|
typedef struct __rpc_svcpool {
|
|
struct mtx sp_lock; /* protect the transport lists */
|
|
const char *sp_name; /* pool name (e.g. "nfsd", "NLM" */
|
|
enum svcpool_state sp_state; /* current pool state */
|
|
struct proc *sp_proc; /* process which is in svc_run */
|
|
struct svcxprt_list sp_xlist; /* all transports in the pool */
|
|
struct svcxprt_list sp_active; /* transports needing service */
|
|
struct svc_callout_list sp_callouts; /* (prog,vers)->dispatch list */
|
|
struct svcthread_list sp_threads; /* service threads */
|
|
struct svcthread_list sp_idlethreads; /* idle service threads */
|
|
int sp_minthreads; /* minimum service thread count */
|
|
int sp_maxthreads; /* maximum service thread count */
|
|
int sp_threadcount; /* current service thread count */
|
|
time_t sp_lastcreatetime; /* when we last started a thread */
|
|
time_t sp_lastidlecheck; /* when we last checked idle transports */
|
|
|
|
/*
|
|
* Hooks to allow an application to control request to thread
|
|
* placement.
|
|
*/
|
|
pool_assign_fn *sp_assign;
|
|
pool_done_fn *sp_done;
|
|
|
|
/*
|
|
* These variables are used to put an upper bound on the
|
|
* amount of memory used by RPC requests which are queued
|
|
* waiting for execution.
|
|
*/
|
|
unsigned int sp_space_low;
|
|
unsigned int sp_space_high;
|
|
unsigned int sp_space_used;
|
|
unsigned int sp_space_used_highest;
|
|
bool_t sp_space_throttled;
|
|
int sp_space_throttle_count;
|
|
|
|
struct replay_cache *sp_rcache; /* optional replay cache */
|
|
struct sysctl_ctx_list sp_sysctl;
|
|
} SVCPOOL;
|
|
|
|
#else
|
|
|
|
/*
|
|
* Service request
|
|
*/
|
|
struct svc_req {
|
|
uint32_t rq_prog; /* service program number */
|
|
uint32_t rq_vers; /* service protocol version */
|
|
uint32_t rq_proc; /* the desired procedure */
|
|
struct opaque_auth rq_cred; /* raw creds from the wire */
|
|
void *rq_clntcred; /* read only cooked cred */
|
|
SVCXPRT *rq_xprt; /* associated transport */
|
|
};
|
|
|
|
/*
|
|
* Approved way of getting address of caller
|
|
*/
|
|
#define svc_getrpccaller(x) (&(x)->xp_rtaddr)
|
|
|
|
#endif
|
|
|
|
/*
|
|
* Operations defined on an SVCXPRT handle
|
|
*
|
|
* SVCXPRT *xprt;
|
|
* struct rpc_msg *msg;
|
|
* xdrproc_t xargs;
|
|
* void * argsp;
|
|
*/
|
|
#ifdef _KERNEL
|
|
|
|
#define SVC_ACQUIRE(xprt) \
|
|
refcount_acquire(&(xprt)->xp_refs)
|
|
|
|
#define SVC_RELEASE(xprt) \
|
|
if (refcount_release(&(xprt)->xp_refs)) \
|
|
SVC_DESTROY(xprt)
|
|
|
|
#define SVC_RECV(xprt, msg, addr, args) \
|
|
(*(xprt)->xp_ops->xp_recv)((xprt), (msg), (addr), (args))
|
|
|
|
#define SVC_STAT(xprt) \
|
|
(*(xprt)->xp_ops->xp_stat)(xprt)
|
|
|
|
#define SVC_REPLY(xprt, msg, addr, m) \
|
|
(*(xprt)->xp_ops->xp_reply) ((xprt), (msg), (addr), (m))
|
|
|
|
#define SVC_DESTROY(xprt) \
|
|
(*(xprt)->xp_ops->xp_destroy)(xprt)
|
|
|
|
#define SVC_CONTROL(xprt, rq, in) \
|
|
(*(xprt)->xp_ops->xp_control)((xprt), (rq), (in))
|
|
|
|
#else
|
|
|
|
#define SVC_RECV(xprt, msg) \
|
|
(*(xprt)->xp_ops->xp_recv)((xprt), (msg))
|
|
#define svc_recv(xprt, msg) \
|
|
(*(xprt)->xp_ops->xp_recv)((xprt), (msg))
|
|
|
|
#define SVC_STAT(xprt) \
|
|
(*(xprt)->xp_ops->xp_stat)(xprt)
|
|
#define svc_stat(xprt) \
|
|
(*(xprt)->xp_ops->xp_stat)(xprt)
|
|
|
|
#define SVC_GETARGS(xprt, xargs, argsp) \
|
|
(*(xprt)->xp_ops->xp_getargs)((xprt), (xargs), (argsp))
|
|
#define svc_getargs(xprt, xargs, argsp) \
|
|
(*(xprt)->xp_ops->xp_getargs)((xprt), (xargs), (argsp))
|
|
|
|
#define SVC_REPLY(xprt, msg) \
|
|
(*(xprt)->xp_ops->xp_reply) ((xprt), (msg))
|
|
#define svc_reply(xprt, msg) \
|
|
(*(xprt)->xp_ops->xp_reply) ((xprt), (msg))
|
|
|
|
#define SVC_FREEARGS(xprt, xargs, argsp) \
|
|
(*(xprt)->xp_ops->xp_freeargs)((xprt), (xargs), (argsp))
|
|
#define svc_freeargs(xprt, xargs, argsp) \
|
|
(*(xprt)->xp_ops->xp_freeargs)((xprt), (xargs), (argsp))
|
|
|
|
#define SVC_DESTROY(xprt) \
|
|
(*(xprt)->xp_ops->xp_destroy)(xprt)
|
|
#define svc_destroy(xprt) \
|
|
(*(xprt)->xp_ops->xp_destroy)(xprt)
|
|
|
|
#define SVC_CONTROL(xprt, rq, in) \
|
|
(*(xprt)->xp_ops2->xp_control)((xprt), (rq), (in))
|
|
|
|
#endif
|
|
|
|
#define SVC_EXT(xprt) \
|
|
((SVCXPRT_EXT *) xprt->xp_p3)
|
|
|
|
#define SVC_AUTH(xprt) \
|
|
(SVC_EXT(xprt)->xp_auth)
|
|
|
|
/*
|
|
* Operations defined on an SVCAUTH handle
|
|
*/
|
|
#ifdef _KERNEL
|
|
#define SVCAUTH_WRAP(auth, mp) \
|
|
((auth)->svc_ah_ops->svc_ah_wrap(auth, mp))
|
|
#define SVCAUTH_UNWRAP(auth, mp) \
|
|
((auth)->svc_ah_ops->svc_ah_unwrap(auth, mp))
|
|
#define SVCAUTH_RELEASE(auth) \
|
|
((auth)->svc_ah_ops->svc_ah_release(auth))
|
|
#else
|
|
#define SVCAUTH_WRAP(auth, xdrs, xfunc, xwhere) \
|
|
((auth)->svc_ah_ops->svc_ah_wrap(auth, xdrs, xfunc, xwhere))
|
|
#define SVCAUTH_UNWRAP(auth, xdrs, xfunc, xwhere) \
|
|
((auth)->svc_ah_ops->svc_ah_unwrap(auth, xdrs, xfunc, xwhere))
|
|
#endif
|
|
|
|
/*
|
|
* Service registration
|
|
*
|
|
* svc_reg(xprt, prog, vers, dispatch, nconf)
|
|
* const SVCXPRT *xprt;
|
|
* const rpcprog_t prog;
|
|
* const rpcvers_t vers;
|
|
* const void (*dispatch)();
|
|
* const struct netconfig *nconf;
|
|
*/
|
|
|
|
__BEGIN_DECLS
|
|
extern bool_t svc_reg(SVCXPRT *, const rpcprog_t, const rpcvers_t,
|
|
void (*)(struct svc_req *, SVCXPRT *),
|
|
const struct netconfig *);
|
|
__END_DECLS
|
|
|
|
/*
|
|
* Service un-registration
|
|
*
|
|
* svc_unreg(prog, vers)
|
|
* const rpcprog_t prog;
|
|
* const rpcvers_t vers;
|
|
*/
|
|
|
|
__BEGIN_DECLS
|
|
#ifdef _KERNEL
|
|
extern void svc_unreg(SVCPOOL *, const rpcprog_t, const rpcvers_t);
|
|
#else
|
|
extern void svc_unreg(const rpcprog_t, const rpcvers_t);
|
|
#endif
|
|
__END_DECLS
|
|
|
|
/*
|
|
* Transport registration.
|
|
*
|
|
* xprt_register(xprt)
|
|
* SVCXPRT *xprt;
|
|
*/
|
|
__BEGIN_DECLS
|
|
extern void xprt_register(SVCXPRT *);
|
|
__END_DECLS
|
|
|
|
/*
|
|
* Transport un-register
|
|
*
|
|
* xprt_unregister(xprt)
|
|
* SVCXPRT *xprt;
|
|
*/
|
|
__BEGIN_DECLS
|
|
extern void xprt_unregister(SVCXPRT *);
|
|
extern void __xprt_unregister_unlocked(SVCXPRT *);
|
|
__END_DECLS
|
|
|
|
#ifdef _KERNEL
|
|
|
|
/*
|
|
* Called when a transport has pending requests.
|
|
*/
|
|
__BEGIN_DECLS
|
|
extern void xprt_active(SVCXPRT *);
|
|
extern void xprt_inactive(SVCXPRT *);
|
|
extern void xprt_inactive_locked(SVCXPRT *);
|
|
__END_DECLS
|
|
|
|
#endif
|
|
|
|
/*
|
|
* When the service routine is called, it must first check to see if it
|
|
* knows about the procedure; if not, it should call svcerr_noproc
|
|
* and return. If so, it should deserialize its arguments via
|
|
* SVC_GETARGS (defined above). If the deserialization does not work,
|
|
* svcerr_decode should be called followed by a return. Successful
|
|
* decoding of the arguments should be followed the execution of the
|
|
* procedure's code and a call to svc_sendreply.
|
|
*
|
|
* Also, if the service refuses to execute the procedure due to too-
|
|
* weak authentication parameters, svcerr_weakauth should be called.
|
|
* Note: do not confuse access-control failure with weak authentication!
|
|
*
|
|
* NB: In pure implementations of rpc, the caller always waits for a reply
|
|
* msg. This message is sent when svc_sendreply is called.
|
|
* Therefore pure service implementations should always call
|
|
* svc_sendreply even if the function logically returns void; use
|
|
* xdr.h - xdr_void for the xdr routine. HOWEVER, tcp based rpc allows
|
|
* for the abuse of pure rpc via batched calling or pipelining. In the
|
|
* case of a batched call, svc_sendreply should NOT be called since
|
|
* this would send a return message, which is what batching tries to avoid.
|
|
* It is the service/protocol writer's responsibility to know which calls are
|
|
* batched and which are not. Warning: responding to batch calls may
|
|
* deadlock the caller and server processes!
|
|
*/
|
|
|
|
__BEGIN_DECLS
|
|
#ifdef _KERNEL
|
|
extern bool_t svc_sendreply(struct svc_req *, xdrproc_t, void *);
|
|
extern bool_t svc_sendreply_mbuf(struct svc_req *, struct mbuf *);
|
|
extern void svcerr_decode(struct svc_req *);
|
|
extern void svcerr_weakauth(struct svc_req *);
|
|
extern void svcerr_noproc(struct svc_req *);
|
|
extern void svcerr_progvers(struct svc_req *, rpcvers_t, rpcvers_t);
|
|
extern void svcerr_auth(struct svc_req *, enum auth_stat);
|
|
extern void svcerr_noprog(struct svc_req *);
|
|
extern void svcerr_systemerr(struct svc_req *);
|
|
#else
|
|
extern bool_t svc_sendreply(SVCXPRT *, xdrproc_t, void *);
|
|
extern void svcerr_decode(SVCXPRT *);
|
|
extern void svcerr_weakauth(SVCXPRT *);
|
|
extern void svcerr_noproc(SVCXPRT *);
|
|
extern void svcerr_progvers(SVCXPRT *, rpcvers_t, rpcvers_t);
|
|
extern void svcerr_auth(SVCXPRT *, enum auth_stat);
|
|
extern void svcerr_noprog(SVCXPRT *);
|
|
extern void svcerr_systemerr(SVCXPRT *);
|
|
#endif
|
|
extern int rpc_reg(rpcprog_t, rpcvers_t, rpcproc_t,
|
|
char *(*)(char *), xdrproc_t, xdrproc_t,
|
|
char *);
|
|
__END_DECLS
|
|
|
|
/*
|
|
* Lowest level dispatching -OR- who owns this process anyway.
|
|
* Somebody has to wait for incoming requests and then call the correct
|
|
* service routine. The routine svc_run does infinite waiting; i.e.,
|
|
* svc_run never returns.
|
|
* Since another (co-existant) package may wish to selectively wait for
|
|
* incoming calls or other events outside of the rpc architecture, the
|
|
* routine svc_getreq is provided. It must be passed readfds, the
|
|
* "in-place" results of a select system call (see select, section 2).
|
|
*/
|
|
|
|
#ifndef _KERNEL
|
|
/*
|
|
* Global keeper of rpc service descriptors in use
|
|
* dynamic; must be inspected before each call to select
|
|
*/
|
|
extern int svc_maxfd;
|
|
#ifdef FD_SETSIZE
|
|
extern fd_set svc_fdset;
|
|
#define svc_fds svc_fdset.fds_bits[0] /* compatibility */
|
|
#else
|
|
extern int svc_fds;
|
|
#endif /* def FD_SETSIZE */
|
|
#endif
|
|
|
|
/*
|
|
* a small program implemented by the svc_rpc implementation itself;
|
|
* also see clnt.h for protocol numbers.
|
|
*/
|
|
__BEGIN_DECLS
|
|
extern void rpctest_service(void);
|
|
__END_DECLS
|
|
|
|
__BEGIN_DECLS
|
|
extern SVCXPRT *svc_xprt_alloc(void);
|
|
extern void svc_xprt_free(SVCXPRT *);
|
|
#ifndef _KERNEL
|
|
extern void svc_getreq(int);
|
|
extern void svc_getreqset(fd_set *);
|
|
extern void svc_getreq_common(int);
|
|
struct pollfd;
|
|
extern void svc_getreq_poll(struct pollfd *, int);
|
|
extern void svc_run(void);
|
|
extern void svc_exit(void);
|
|
#else
|
|
extern void svc_run(SVCPOOL *);
|
|
extern void svc_exit(SVCPOOL *);
|
|
extern bool_t svc_getargs(struct svc_req *, xdrproc_t, void *);
|
|
extern bool_t svc_freeargs(struct svc_req *, xdrproc_t, void *);
|
|
extern void svc_freereq(struct svc_req *);
|
|
|
|
#endif
|
|
__END_DECLS
|
|
|
|
/*
|
|
* Socket to use on svcxxx_create call to get default socket
|
|
*/
|
|
#define RPC_ANYSOCK -1
|
|
#define RPC_ANYFD RPC_ANYSOCK
|
|
|
|
/*
|
|
* These are the existing service side transport implementations
|
|
*/
|
|
|
|
__BEGIN_DECLS
|
|
|
|
#ifdef _KERNEL
|
|
|
|
/*
|
|
* Create a new service pool.
|
|
*/
|
|
extern SVCPOOL* svcpool_create(const char *name,
|
|
struct sysctl_oid_list *sysctl_base);
|
|
|
|
/*
|
|
* Destroy a service pool, including all registered transports.
|
|
*/
|
|
extern void svcpool_destroy(SVCPOOL *pool);
|
|
|
|
/*
|
|
* Transport independent svc_create routine.
|
|
*/
|
|
extern int svc_create(SVCPOOL *, void (*)(struct svc_req *, SVCXPRT *),
|
|
const rpcprog_t, const rpcvers_t, const char *);
|
|
/*
|
|
* void (*dispatch)(); -- dispatch routine
|
|
* const rpcprog_t prognum; -- program number
|
|
* const rpcvers_t versnum; -- version number
|
|
* const char *nettype; -- network type
|
|
*/
|
|
|
|
|
|
/*
|
|
* Generic server creation routine. It takes a netconfig structure
|
|
* instead of a nettype.
|
|
*/
|
|
|
|
extern SVCXPRT *svc_tp_create(SVCPOOL *, void (*)(struct svc_req *, SVCXPRT *),
|
|
const rpcprog_t, const rpcvers_t, const char *uaddr,
|
|
const struct netconfig *);
|
|
/*
|
|
* void (*dispatch)(); -- dispatch routine
|
|
* const rpcprog_t prognum; -- program number
|
|
* const rpcvers_t versnum; -- version number
|
|
* const char *uaddr; -- universal address of service
|
|
* const struct netconfig *nconf; -- netconfig structure
|
|
*/
|
|
|
|
extern SVCXPRT *svc_dg_create(SVCPOOL *, struct socket *,
|
|
const size_t, const size_t);
|
|
/*
|
|
* struct socket *; -- open connection
|
|
* const size_t sendsize; -- max send size
|
|
* const size_t recvsize; -- max recv size
|
|
*/
|
|
|
|
extern SVCXPRT *svc_vc_create(SVCPOOL *, struct socket *,
|
|
const size_t, const size_t);
|
|
/*
|
|
* struct socket *; -- open connection
|
|
* const size_t sendsize; -- max send size
|
|
* const size_t recvsize; -- max recv size
|
|
*/
|
|
|
|
/*
|
|
* Generic TLI create routine
|
|
*/
|
|
extern SVCXPRT *svc_tli_create(SVCPOOL *, struct socket *,
|
|
const struct netconfig *, const struct t_bind *, const size_t, const size_t);
|
|
/*
|
|
* struct socket * so; -- connection end point
|
|
* const struct netconfig *nconf; -- netconfig structure for network
|
|
* const struct t_bind *bindaddr; -- local bind address
|
|
* const size_t sendsz; -- max sendsize
|
|
* const size_t recvsz; -- max recvsize
|
|
*/
|
|
|
|
#else /* !_KERNEL */
|
|
|
|
/*
|
|
* Transport independent svc_create routine.
|
|
*/
|
|
extern int svc_create(void (*)(struct svc_req *, SVCXPRT *),
|
|
const rpcprog_t, const rpcvers_t, const char *);
|
|
/*
|
|
* void (*dispatch)(); -- dispatch routine
|
|
* const rpcprog_t prognum; -- program number
|
|
* const rpcvers_t versnum; -- version number
|
|
* const char *nettype; -- network type
|
|
*/
|
|
|
|
|
|
/*
|
|
* Generic server creation routine. It takes a netconfig structure
|
|
* instead of a nettype.
|
|
*/
|
|
|
|
extern SVCXPRT *svc_tp_create(void (*)(struct svc_req *, SVCXPRT *),
|
|
const rpcprog_t, const rpcvers_t,
|
|
const struct netconfig *);
|
|
/*
|
|
* void (*dispatch)(); -- dispatch routine
|
|
* const rpcprog_t prognum; -- program number
|
|
* const rpcvers_t versnum; -- version number
|
|
* const struct netconfig *nconf; -- netconfig structure
|
|
*/
|
|
|
|
/*
|
|
* Generic TLI create routine
|
|
*/
|
|
extern SVCXPRT *svc_tli_create(const int, const struct netconfig *,
|
|
const struct t_bind *, const u_int,
|
|
const u_int);
|
|
/*
|
|
* const int fd; -- connection end point
|
|
* const struct netconfig *nconf; -- netconfig structure for network
|
|
* const struct t_bind *bindaddr; -- local bind address
|
|
* const u_int sendsz; -- max sendsize
|
|
* const u_int recvsz; -- max recvsize
|
|
*/
|
|
|
|
/*
|
|
* Connectionless and connectionful create routines
|
|
*/
|
|
|
|
extern SVCXPRT *svc_vc_create(const int, const u_int, const u_int);
|
|
/*
|
|
* const int fd; -- open connection end point
|
|
* const u_int sendsize; -- max send size
|
|
* const u_int recvsize; -- max recv size
|
|
*/
|
|
|
|
/*
|
|
* Added for compatibility to old rpc 4.0. Obsoleted by svc_vc_create().
|
|
*/
|
|
extern SVCXPRT *svcunix_create(int, u_int, u_int, char *);
|
|
|
|
extern SVCXPRT *svc_dg_create(const int, const u_int, const u_int);
|
|
/*
|
|
* const int fd; -- open connection
|
|
* const u_int sendsize; -- max send size
|
|
* const u_int recvsize; -- max recv size
|
|
*/
|
|
|
|
|
|
/*
|
|
* the routine takes any *open* connection
|
|
* descriptor as its first input and is used for open connections.
|
|
*/
|
|
extern SVCXPRT *svc_fd_create(const int, const u_int, const u_int);
|
|
/*
|
|
* const int fd; -- open connection end point
|
|
* const u_int sendsize; -- max send size
|
|
* const u_int recvsize; -- max recv size
|
|
*/
|
|
|
|
/*
|
|
* Added for compatibility to old rpc 4.0. Obsoleted by svc_fd_create().
|
|
*/
|
|
extern SVCXPRT *svcunixfd_create(int, u_int, u_int);
|
|
|
|
/*
|
|
* Memory based rpc (for speed check and testing)
|
|
*/
|
|
extern SVCXPRT *svc_raw_create(void);
|
|
|
|
/*
|
|
* svc_dg_enable_cache() enables the cache on dg transports.
|
|
*/
|
|
int svc_dg_enablecache(SVCXPRT *, const u_int);
|
|
|
|
int __rpc_get_local_uid(SVCXPRT *_transp, uid_t *_uid);
|
|
|
|
#endif /* !_KERNEL */
|
|
|
|
__END_DECLS
|
|
|
|
#ifndef _KERNEL
|
|
/* for backward compatibility */
|
|
#include <rpc/svc_soc.h>
|
|
#endif
|
|
|
|
#endif /* !_RPC_SVC_H */
|