svc.h revision 7c478bd95313f5f23a4c958a745db2134aa03244
/*
* CDDL HEADER START
*
* The contents of this file are subject to the terms of the
* Common Development and Distribution License, Version 1.0 only
* (the "License"). You may not use this file except in compliance
* with the License.
*
* You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
* or http://www.opensolaris.org/os/licensing.
* See the License for the specific language governing permissions
* and limitations under the License.
*
* When distributing Covered Code, include this CDDL HEADER in each
* file and include the License file at usr/src/OPENSOLARIS.LICENSE.
* If applicable, add the following below this CDDL HEADER, with the
* fields enclosed by brackets "[]" replaced with your own identifying
* information: Portions Copyright [yyyy] [name of copyright owner]
*
* CDDL HEADER END
*/
/*
* Copyright 2005 Sun Microsystems, Inc. All rights reserved.
* Use is subject to license terms.
*/
/* Copyright (c) 1983, 1984, 1985, 1986, 1987, 1988, 1989 AT&T */
/* All Rights Reserved */
/*
* Portions of this source code were derived from Berkeley
* 4.3 BSD under license from the Regents of the University of
* California.
*/
/*
* svc.h, Server-side remote procedure call interface.
*/
#ifndef _RPC_SVC_H
#define _RPC_SVC_H
#pragma ident "%Z%%M% %I% %E% SMI"
#include <rpc/rpc_com.h>
#include <rpc/rpc_msg.h>
#include <sys/tihdr.h>
#include <sys/poll.h>
#ifdef _KERNEL
#include <rpc/svc_auth.h>
#include <sys/callb.h>
#endif /* _KERNEL */
/*
* This interface must manage two items concerning remote procedure calling:
*
* 1) An arbitrary number of transport connections upon which rpc requests
* are received. They are created and registered by routines in svc_generic.c,
* svc_vc.c and svc_dg.c; they in turn call xprt_register and
* xprt_unregister.
*
* 2) An arbitrary number of locally registered services. Services are
* described by the following four data: program number, version number,
* "service dispatch" function, a transport handle, and a boolean that
* indicates whether or not the exported program should be registered with a
* local binder service; if true the program's number and version and the
* address from the transport handle are registered with the binder.
* These data are registered with rpcbind via svc_reg().
*
* A service's dispatch function is called whenever an rpc request comes in
* on a transport. The request's program and version numbers must match
* those of the registered service. The dispatch function is passed two
* parameters, struct svc_req * and SVCXPRT *, defined below.
*/
#ifdef __cplusplus
extern "C" {
#endif
/*
* Server-side transport handles.
* The actual type definitions are below.
*/
#ifdef _KERNEL
typedef struct __svcmasterxprt SVCMASTERXPRT; /* Master transport handle */
typedef struct __svcxprt SVCXPRT; /* Per-thread clone handle */
typedef struct __svcpool SVCPOOL; /* Kernel thread pool */
#else /* _KERNEL */
typedef struct __svcxprt SVCXPRT; /* Server transport handle */
#endif /* _KERNEL */
/*
* Prototype of error handler callback
*/
#ifndef _KERNEL
typedef void (*svc_errorhandler_t)(const SVCXPRT* svc, const bool_t isAConn);
#endif
/*
* Service request.
*
* PSARC 2003/523 Contract Private Interface
* svc_req
* Changes must be reviewed by Solaris File Sharing
* Changes must be communicated to contract-2003-523@sun.com
*/
struct svc_req {
rpcprog_t rq_prog; /* service program number */
rpcvers_t rq_vers; /* service protocol version */
rpcproc_t rq_proc; /* the desired procedure */
struct opaque_auth rq_cred; /* raw creds from the wire */
caddr_t rq_clntcred; /* read only cooked cred */
SVCXPRT *rq_xprt; /* associated transport */
};
#ifdef _KERNEL
struct dupreq {
uint32_t dr_xid;
rpcproc_t dr_proc;
rpcvers_t dr_vers;
rpcprog_t dr_prog;
struct netbuf dr_addr;
struct netbuf dr_resp;
void (*dr_resfree)();
int dr_status;
struct dupreq *dr_next;
struct dupreq *dr_chain;
};
/*
* States of requests for duplicate request caching.
*/
#define DUP_NEW 0x00 /* new entry */
#define DUP_INPROGRESS 0x01 /* request already going */
#define DUP_DONE 0x02 /* request done */
#define DUP_DROP 0x03 /* request dropped */
#define DUP_ERROR 0x04 /* error in dup req cache */
/*
* Prototype for a service dispatch routine.
*/
typedef void (SVC_DISPATCH)(struct svc_req *, SVCXPRT *);
/*
* The service provider callout.
* Each entry identifies a dispatch routine to be called
* for a given RPC program number and a version fitting
* into the registered range.
*/
typedef struct {
rpcprog_t sc_prog; /* RPC Program number */
rpcvers_t sc_versmin; /* Min version number */
rpcvers_t sc_versmax; /* Max version number */
SVC_DISPATCH *sc_dispatch; /* Dispatch routine */
} SVC_CALLOUT;
/*
* Table of service provider `callouts' for an RPC
* transport handle. If sct_free is TRUE then transport
* destructor is supposed to deallocate this table.
*/
typedef struct {
size_t sct_size; /* Number of entries */
bool_t sct_free; /* Deallocate if true */
SVC_CALLOUT *sct_sc; /* Callout entries */
} SVC_CALLOUT_TABLE;
struct svc_ops {
bool_t (*xp_recv)(SVCXPRT *, mblk_t *, struct rpc_msg *);
/* receive incoming requests */
bool_t (*xp_getargs)(SVCXPRT *, xdrproc_t, caddr_t);
/* get arguments */
bool_t (*xp_reply)(SVCXPRT *, struct rpc_msg *);
/* send reply */
bool_t (*xp_freeargs)(SVCXPRT *, xdrproc_t, caddr_t);
/* free mem allocated for args */
void (*xp_destroy)(SVCMASTERXPRT *);
/* destroy this struct */
int (*xp_dup)(struct svc_req *, caddr_t, int,
struct dupreq **, bool_t *);
/* check for dup */
void (*xp_dupdone)(struct dupreq *, caddr_t, void (*)(), int, int);
/* mark dup entry as completed */
int32_t *(*xp_getres)(SVCXPRT *, int);
/* get pointer to response buffer */
void (*xp_freeres)(SVCXPRT *);
/* destroy pre-serialized response */
void (*xp_clone_destroy)(SVCXPRT *);
/* destroy a clone xprt */
void (*xp_start)(SVCMASTERXPRT *);
/* `ready-to-receive' */
};
#else /* _KERNEL */
/*
* Service control requests
*/
#define SVCGET_VERSQUIET 1
#define SVCSET_VERSQUIET 2
#define SVCGET_XID 4
#define SVCSET_KEEPALIVE 5
#define SVCSET_CONNMAXREC 6
#define SVCGET_CONNMAXREC 7
#define SVCGET_RECVERRHANDLER 8
#define SVCSET_RECVERRHANDLER 9
enum xprt_stat {
XPRT_DIED,
XPRT_MOREREQS,
XPRT_IDLE
};
struct xp_ops {
#ifdef __STDC__
bool_t (*xp_recv)(SVCXPRT *, struct rpc_msg *);
/* receive incoming requests */
enum xprt_stat (*xp_stat)(SVCXPRT *);
/* get transport status */
bool_t (*xp_getargs)(SVCXPRT *, xdrproc_t, caddr_t);
/* get arguments */
bool_t (*xp_reply)(SVCXPRT *, struct rpc_msg *);
/* send reply */
bool_t (*xp_freeargs)(SVCXPRT *, xdrproc_t, caddr_t);
/* free mem allocated for args */
void (*xp_destroy)(SVCXPRT *);
/* destroy this struct */
bool_t (*xp_control)(SVCXPRT *, const uint_t, void *);
/* catch-all control function */
#else /* __STDC__ */
bool_t (*xp_recv)(); /* receive incoming requests */
enum xprt_stat (*xp_stat)(); /* get transport status */
bool_t (*xp_getargs)(); /* get arguments */
bool_t (*xp_reply)(); /* send reply */
bool_t (*xp_freeargs)(); /* free mem allocated for args */
void (*xp_destroy)(); /* destroy this struct */
bool_t (*xp_control)(); /* catch-all control function */
#endif /* __STDC__ */
};
#endif /* _KERNEL */
#ifdef _KERNEL
/*
* SVCPOOL
* Kernel RPC server-side thread pool structure.
*/
typedef struct __svcxprt_qnode __SVCXPRT_QNODE; /* Defined in svc.c */
struct __svcpool {
/*
* Thread pool variables.
*
* The pool's thread lock p_thread_lock protects:
* - p_threads, p_detached_threads, p_reserved_threads and p_closing
* The pool's request lock protects:
* - p_asleep, p_drowsy, p_reqs, p_walkers, p_req_cv.
* The following fields are `initialized constants':
* - p_id, p_stksize, p_timeout.
* Access to p_next and p_prev is protected by the pool
* list lock.
*/
SVCPOOL *p_next; /* Next pool in the list */
SVCPOOL *p_prev; /* Prev pool in the list */
int p_id; /* Pool id */
int p_threads; /* Non-detached threads */
int p_detached_threads; /* Detached threads */
int p_maxthreads; /* Max threads in the pool */
int p_redline; /* `Redline' for the pool */
int p_reserved_threads; /* Reserved threads */
kmutex_t p_thread_lock; /* Thread lock */
int p_asleep; /* Asleep threads */
int p_drowsy; /* Drowsy flag */
kcondvar_t p_req_cv; /* svc_poll() sleep var. */
clock_t p_timeout; /* svc_poll() timeout */
kmutex_t p_req_lock; /* Request lock */
int p_reqs; /* Pending requests */
int p_walkers; /* Walking threads */
int p_max_same_xprt; /* Max reqs from the xprt */
int p_stksize; /* Stack size for svc_run */
bool_t p_closing : 1; /* Pool is closing */
/*
* Thread creator variables.
* The `creator signaled' flag is turned on when a signal is send
* to the creator thread (to create a new service thread). The
* creator clears when the thread is created. The protocol is not
* to signal the creator thread when the flag is on. However,
* a new thread should signal the creator if there are more
* requests in the queue.
*
* When the pool is closing (ie it has been already unregistered from
* the pool list) the last thread on the last transport should turn
* the p_creator_exit flag on. This tells the creator thread to
* free the pool structure and exit.
*/
bool_t p_creator_signaled : 1; /* Create requested flag */
bool_t p_creator_exit : 1; /* If true creator exits */
kcondvar_t p_creator_cv; /* Creator cond. variable */
kmutex_t p_creator_lock; /* Creator lock */
/*
* Doubly linked list containing `registered' master transport handles.
* There is no special structure for a list node. Instead the
* SVCMASTERXPRT structure has the xp_next and xp_prev fields.
*
* The p_lrwlock protects access to xprt->xp_next and xprt->xp_prev.
* A service thread should also acquire a reader lock before accessing
* any transports it is no longer linked to (to prevent them from
* being destroyed).
*
* The list lock governs also the `pool is closing' flag.
*/
size_t p_lcount; /* Current count */
SVCMASTERXPRT *p_lhead; /* List head */
krwlock_t p_lrwlock; /* R/W lock */
/*
* Circular linked list for the `xprt-ready' queue (FIFO).
* Must be initialized with svc_xprt_qinit() before it is used.
*
* The writer's end is protected by the pool's request lock
* (pool->p_req_lock). The reader's end is protected by q_end_lock.
*
* When the queue is full the p_qoverflow flag is raised. It stays
* on until all the pending request are drained.
*/
size_t p_qsize; /* Number of queue nodes */
int p_qoverflow : 1; /* Overflow flag */
__SVCXPRT_QNODE *p_qbody; /* Queue body (array) */
__SVCXPRT_QNODE *p_qtop; /* Writer's end of FIFO */
__SVCXPRT_QNODE *p_qend; /* Reader's end of FIFO */
kmutex_t p_qend_lock; /* Reader's end lock */
/*
* Userspace thread creator variables.
* Thread creation is actually done in userland, via a thread
* that is parked in the kernel. When that thread is signaled,
* it returns back down to the daemon from whence it came and
* does the lwp create.
*
* A parallel "creator" thread runs in the kernel. That is the
* thread that will signal for the user thread to return to
* userland and do its work.
*
* Since the thread doesn't always exist (there could be a race
* if two threads are created in rapid succession), we set
* p_signal_create_thread to FALSE when we're ready to accept work.
*
* p_user_exit is set to true when the service pool is about
* to close. This is done so that the user creation thread
* can be informed and cleanup any userland state.
*/
bool_t p_signal_create_thread : 1; /* Create requested flag */
bool_t p_user_exit : 1; /* If true creator exits */
bool_t p_user_waiting : 1; /* Thread waiting for work */
kcondvar_t p_user_cv; /* Creator cond. variable */
kmutex_t p_user_lock; /* Creator lock */
void (*p_offline)(); /* callout for unregister */
void (*p_shutdown)(); /* callout for shutdown */
};
/*
* Server side transport handle (SVCMASTERXPRT).
* xprt->xp_req_lock governs the following fields in xprt:
* xp_req_head, xp_req_tail.
* xprt->xp_thread_lock governs the following fields in xprt:
* xp_threads, xp_detached_threads.
*
* xp_req_tail is only valid if xp_req_head is non-NULL
*
* The xp_threads count is the number of attached threads. These threads
* are able to handle new requests, and it is expected that they will not
* block for a very long time handling a given request. The
* xp_detached_threads count is the number of threads that have detached
* themselves from the transport. These threads can block indefinitely
* while handling a request. Once they complete the request, they exit.
*
* A kernel service provider may register a callback function "closeproc"
* for a transport. When the transport is closing the last exiting attached
* thread - xp_threads goes to zero - it calls the callback function, passing
* it a reference to the transport. This call is made with xp_thread_lock
* held, so any cleanup bookkeeping it does should be done quickly.
*
* When the transport is closing the last exiting thread is supposed
* to destroy/free the data structure.
*/
typedef struct __svcxprt_common {
struct file *xpc_fp;
struct svc_ops *xpc_ops;
queue_t *xpc_wq; /* queue to write onto */
cred_t *xpc_cred; /* cached cred for server to use */
int32_t xpc_type; /* transport type */
int xpc_msg_size; /* TSDU or TIDU size */
struct netbuf xpc_rtaddr; /* remote transport address */
SVC_CALLOUT_TABLE *xpc_sct;
} __SVCXPRT_COMMON;
#define xp_fp xp_xpc.xpc_fp
#define xp_ops xp_xpc.xpc_ops
#define xp_wq xp_xpc.xpc_wq
#define xp_cred xp_xpc.xpc_cred
#define xp_type xp_xpc.xpc_type
#define xp_msg_size xp_xpc.xpc_msg_size
#define xp_rtaddr xp_xpc.xpc_rtaddr
#define xp_sct xp_xpc.xpc_sct
struct __svcmasterxprt {
SVCMASTERXPRT *xp_next; /* Next transport in the list */
SVCMASTERXPRT *xp_prev; /* Prev transport in the list */
__SVCXPRT_COMMON xp_xpc; /* Fields common with the clone */
SVCPOOL *xp_pool; /* Pointer to the pool */
mblk_t *xp_req_head; /* Request queue head */
mblk_t *xp_req_tail; /* Request queue tail */
kmutex_t xp_req_lock; /* Request lock */
int xp_threads; /* Current num. of attached threads */
int xp_detached_threads; /* num. of detached threads */
kmutex_t xp_thread_lock; /* Thread count lock */
void (*xp_closeproc)(const SVCMASTERXPRT *);
/* optional; see comments above */
char *xp_netid; /* network token */
struct netbuf xp_addrmask; /* address mask */
caddr_t xp_p2; /* private: for use by svc ops */
};
/*
* Service thread `clone' transport handle (SVCXPRT)
*
* PSARC 2003/523 Contract Private Interface
* SVCXPRT
* Changes must be reviewed by Solaris File Sharing
* Changes must be communicated to contract-2003-523@sun.com
*
* The xp_p2buf buffer is used as the storage for a transport type
* specific structure. It is private for the svc ops for a given
* transport type.
*/
#define SVC_P2LEN 64
struct __svcxprt {
__SVCXPRT_COMMON xp_xpc;
SVCMASTERXPRT *xp_master; /* back ptr to master */
/* The following fileds are on a per-thread basis */
callb_cpr_t *xp_cprp; /* unused padding for Contract */
bool_t xp_reserved : 1; /* is thread reserved? */
bool_t xp_detached : 1; /* is thread detached? */
int xp_same_xprt; /* Reqs from the same xprt */
/* The following fields are used on a per-request basis */
struct opaque_auth xp_verf; /* raw response verifier */
SVCAUTH xp_auth; /* auth flavor of current req */
void *xp_cookie; /* a cookie */
uint32_t xp_xid; /* id */
XDR xp_xdrin; /* input xdr stream */
XDR xp_xdrout; /* output xdr stream */
/* Private for svc ops */
char xp_p2buf[SVC_P2LEN]; /* udp_data or cots_data_t */
/* or clone_rdma_data_t */
};
#else /* _KERNEL */
struct __svcxprt {
int xp_fd;
#define xp_sock xp_fd
ushort_t xp_port;
/*
* associated port number.
* Obsolete, but still used to
* specify whether rendezvouser
* or normal connection
*/
struct xp_ops *xp_ops;
int xp_addrlen; /* length of remote addr. Obsoleted */
char *xp_tp; /* transport provider device name */
char *xp_netid; /* network token */
struct netbuf xp_ltaddr; /* local transport address */
struct netbuf xp_rtaddr; /* remote transport address */
char xp_raddr[16]; /* remote address. Now obsoleted */
struct opaque_auth xp_verf; /* raw response verifier */
caddr_t xp_p1; /* private: for use by svc ops */
caddr_t xp_p2; /* private: for use by svc ops */
caddr_t xp_p3; /* private: for use by svc lib */
int xp_type; /* transport type */
/*
* callback on client death
* First parameter is the current structure,
* Second parameter :
* - FALSE for the service listener
* - TRUE for a real connected socket
*/
svc_errorhandler_t xp_closeclnt;
};
#endif /* _KERNEL */
/*
* Approved way of getting address of caller,
* address mask, and netid of transport.
*/
#define svc_getrpccaller(x) (&(x)->xp_rtaddr)
#ifdef _KERNEL
#define svc_getcaller(x) (&(x)->xp_rtaddr.buf)
#define svc_getaddrmask(x) (&(x)->xp_master->xp_addrmask)
#define svc_getnetid(x) ((x)->xp_master->xp_netid)
#endif /* _KERNEL */
/*
* Operations defined on an SVCXPRT handle
*/
#ifdef _KERNEL
#define SVC_RECV(clone_xprt, mp, msg) \
(*(clone_xprt)->xp_ops->xp_recv)((clone_xprt), (mp), (msg))
/*
* PSARC 2003/523 Contract Private Interface
* SVC_GETARGS
* Changes must be reviewed by Solaris File Sharing
* Changes must be communicated to contract-2003-523@sun.com
*/
#define SVC_GETARGS(clone_xprt, xargs, argsp) \
(*(clone_xprt)->xp_ops->xp_getargs)((clone_xprt), (xargs), (argsp))
#define SVC_REPLY(clone_xprt, msg) \
(*(clone_xprt)->xp_ops->xp_reply) ((clone_xprt), (msg))
#define SVC_FREEARGS(clone_xprt, xargs, argsp) \
(*(clone_xprt)->xp_ops->xp_freeargs)((clone_xprt), (xargs), (argsp))
#define SVC_GETRES(clone_xprt, size) \
(*(clone_xprt)->xp_ops->xp_getres)((clone_xprt), (size))
#define SVC_FREERES(clone_xprt) \
(*(clone_xprt)->xp_ops->xp_freeres)(clone_xprt)
#define SVC_DESTROY(xprt) \
(*(xprt)->xp_ops->xp_destroy)(xprt)
/*
* PSARC 2003/523 Contract Private Interfaces
* SVC_DUP, SVC_DUPDONE, SVC_DUP_EXT, SVC_DUPDONE_EXT
* Changes must be reviewed by Solaris File Sharing
* Changes must be communicated to contract-2003-523@sun.com
*
* SVC_DUP and SVC_DUPDONE are defined here for backward compatibility.
*/
#define SVC_DUP_EXT(clone_xprt, req, res, size, drpp, dupcachedp) \
(*(clone_xprt)->xp_ops->xp_dup)(req, res, size, drpp, dupcachedp)
#define SVC_DUPDONE_EXT(clone_xprt, dr, res, resfree, size, status) \
(*(clone_xprt)->xp_ops->xp_dupdone)(dr, res, resfree, size, status)
#define SVC_DUP(clone_xprt, req, res, size, drpp) \
(*(clone_xprt)->xp_ops->xp_dup)(req, res, size, drpp, NULL)
#define SVC_DUPDONE(clone_xprt, dr, res, size, status) \
(*(clone_xprt)->xp_ops->xp_dupdone)(dr, res, NULL, size, status)
#define SVC_CLONE_DESTROY(clone_xprt) \
(*(clone_xprt)->xp_ops->xp_clone_destroy)(clone_xprt)
#define SVC_START(xprt) \
(*(xprt)->xp_ops->xp_start)(xprt)
#else /* _KERNEL */
#define SVC_RECV(xprt, msg) \
(*(xprt)->xp_ops->xp_recv)((xprt), (msg))
#define svc_recv(xprt, msg) \
(*(xprt)->xp_ops->xp_recv)((xprt), (msg))
#define SVC_STAT(xprt) \
(*(xprt)->xp_ops->xp_stat)(xprt)
#define svc_stat(xprt) \
(*(xprt)->xp_ops->xp_stat)(xprt)
#define SVC_GETARGS(xprt, xargs, argsp) \
(*(xprt)->xp_ops->xp_getargs)((xprt), (xargs), (argsp))
#define svc_getargs(xprt, xargs, argsp) \
(*(xprt)->xp_ops->xp_getargs)((xprt), (xargs), (argsp))
#define SVC_REPLY(xprt, msg) \
(*(xprt)->xp_ops->xp_reply) ((xprt), (msg))
#define svc_reply(xprt, msg) \
(*(xprt)->xp_ops->xp_reply) ((xprt), (msg))
#define SVC_FREEARGS(xprt, xargs, argsp) \
(*(xprt)->xp_ops->xp_freeargs)((xprt), (xargs), (argsp))
#define svc_freeargs(xprt, xargs, argsp) \
(*(xprt)->xp_ops->xp_freeargs)((xprt), (xargs), (argsp))
#define SVC_GETRES(xprt, size) \
(*(xprt)->xp_ops->xp_getres)((xprt), (size))
#define svc_getres(xprt, size) \
(*(xprt)->xp_ops->xp_getres)((xprt), (size))
#define SVC_FREERES(xprt) \
(*(xprt)->xp_ops->xp_freeres)(xprt)
#define svc_freeres(xprt) \
(*(xprt)->xp_ops->xp_freeres)(xprt)
#define SVC_DESTROY(xprt) \
(*(xprt)->xp_ops->xp_destroy)(xprt)
#define svc_destroy(xprt) \
(*(xprt)->xp_ops->xp_destroy)(xprt)
/*
* PSARC 2003/523 Contract Private Interface
* SVC_CONTROL
* Changes must be reviewed by Solaris File Sharing
* Changes must be communicated to contract-2003-523@sun.com
*/
#define SVC_CONTROL(xprt, rq, in) \
(*(xprt)->xp_ops->xp_control)((xprt), (rq), (in))
#endif /* _KERNEL */
/*
* Pool id's reserved for NFS, NLM, and the NFSv4 callback program.
*/
#define NFS_SVCPOOL_ID 0x01
#define NLM_SVCPOOL_ID 0x02
#define NFS_CB_SVCPOOL_ID 0x03
#define RDC_SVCPOOL_ID 0x05 /* SNDR, PSARC 2001/699 */
struct svcpool_args {
uint32_t id; /* Pool id */
uint32_t maxthreads; /* Max threads in the pool */
uint32_t redline; /* `Redline' for the pool */
uint32_t qsize; /* `xprt-ready' queue size */
uint32_t timeout; /* svc_poll() timeout */
uint32_t stksize; /* svc_run() stack size */
uint32_t max_same_xprt; /* Max reqs from the same xprt */
};
#ifdef _KERNEL
/*
* Transport registration and thread pool creation.
*/
extern int svc_xprt_register(SVCMASTERXPRT *, int);
extern void svc_xprt_unregister(SVCMASTERXPRT *);
extern int svc_pool_create(struct svcpool_args *);
extern int svc_wait(int);
extern int svc_do_run(int);
#define SVCPSET_SHUTDOWN_PROC 1
#define SVCPSET_UNREGISTER_PROC 2
extern int svc_pool_control(int, int, void *);
#else /* _KERNEL */
#ifdef __STDC__
extern bool_t rpc_reg(const rpcprog_t, const rpcvers_t, const rpcproc_t,
char *(*)(char *), const xdrproc_t, const xdrproc_t,
const char *);
/*
* Service registration
*
* svc_reg(xprt, prog, vers, dispatch, nconf)
* const SVCXPRT *xprt;
* const rpcprog_t prog;
* const rpcvers_t vers;
* const void (*dispatch)();
* const struct netconfig *nconf;
*/
extern bool_t svc_reg(const SVCXPRT *, const rpcprog_t, const rpcvers_t,
void (*)(struct svc_req *, SVCXPRT *),
const struct netconfig *);
/*
* Service authentication registration
*
* svc_auth_reg(cred_flavor, handler)
* int cred_flavor;
* enum auth_stat (*handler)();
*/
extern int svc_auth_reg(int, enum auth_stat (*)());
/*
* Service un-registration
*
* svc_unreg(prog, vers)
* const rpcprog_t prog;
* const rpcvers_t vers;
*/
extern void svc_unreg(const rpcprog_t, const rpcvers_t);
/*
* Transport registration/unregistration.
*
* xprt_register(xprt)
* const SVCXPRT *xprt;
*
* xprt_unregister(xprt)
* const SVCXPRT *xprt;
*/
extern void xprt_register(const SVCXPRT *);
extern void xprt_unregister(const SVCXPRT *);
#else /* __STDC__ */
extern bool_t rpc_reg();
extern bool_t svc_reg();
extern bool_t svc_auth_reg();
extern void svc_unreg();
extern void xprt_register();
extern void xprt_unregister();
#endif /* __STDC__ */
#endif /* _KERNEL */
/*
* When the service routine is called, it must first check to see if it
* knows about the procedure; if not, it should call svcerr_noproc
* and return. If so, it should deserialize its arguments via
* SVC_GETARGS (defined above). If the deserialization does not work,
* svcerr_decode should be called followed by a return. Successful
* decoding of the arguments should be followed the execution of the
* procedure's code and a call to svc_sendreply.
*
* Also, if the service refuses to execute the procedure due to too-
* weak authentication parameters, svcerr_weakauth should be called.
* Note: do not confuse access-control failure with weak authentication!
*
* NB: In pure implementations of rpc, the caller always waits for a reply
* msg. This message is sent when svc_sendreply is called.
* Therefore pure service implementations should always call
* svc_sendreply even if the function logically returns void; use
* xdr.h - xdr_void for the xdr routine. HOWEVER, connectionful rpc allows
* for the abuse of pure rpc via batched calling or pipelining. In the
* case of a batched call, svc_sendreply should NOT be called since
* this would send a return message, which is what batching tries to avoid.
* It is the service/protocol writer's responsibility to know which calls are
* batched and which are not. Warning: responding to batch calls may
* deadlock the caller and server processes!
*/
#ifdef __STDC__
extern bool_t svc_sendreply(const SVCXPRT *, const xdrproc_t, const caddr_t);
extern void svcerr_decode(const SVCXPRT *);
extern void svcerr_weakauth(const SVCXPRT *);
extern void svcerr_noproc(const SVCXPRT *);
extern void svcerr_progvers(const SVCXPRT *, const rpcvers_t,
const rpcvers_t);
extern void svcerr_auth(const SVCXPRT *, const enum auth_stat);
extern void svcerr_noprog(const SVCXPRT *);
extern void svcerr_systemerr(const SVCXPRT *);
#else /* __STDC__ */
extern bool_t svc_sendreply();
extern void svcerr_decode();
extern void svcerr_weakauth();
extern void svcerr_noproc();
extern void svcerr_progvers();
extern void svcerr_auth();
extern void svcerr_noprog();
extern void svcerr_systemerr();
#endif /* __STDC__ */
#ifdef _KERNEL
/*
* Kernel RPC functions.
*/
extern void svc_init(void);
extern void svc_cots_init(void);
extern void svc_clts_init(void);
extern void mt_kstat_init(void);
extern void mt_kstat_fini(void);
extern int svc_tli_kcreate(struct file *, uint_t, char *,
struct netbuf *, SVCMASTERXPRT **,
SVC_CALLOUT_TABLE *,
void (*closeproc)(const SVCMASTERXPRT *),
int, bool_t);
extern int svc_clts_kcreate(struct file *, uint_t, struct T_info_ack *,
SVCMASTERXPRT **);
extern int svc_cots_kcreate(struct file *, uint_t, struct T_info_ack *,
SVCMASTERXPRT **);
extern void svc_queuereq(queue_t *, mblk_t *);
extern void svc_queueclean(queue_t *);
extern void svc_queueclose(queue_t *);
extern int svc_reserve_thread(SVCXPRT *);
extern void svc_unreserve_thread(SVCXPRT *);
extern callb_cpr_t *svc_detach_thread(SVCXPRT *);
/*
* For RDMA based kRPC.
* "rdma_xprt_record" is a reference to master transport handles
* in kRPC thread pools. This is an easy way of tracking and shuting
* down rdma based kRPC transports on demand.
* "rdma_xprt_group" is a list of RDMA based mster transport handles
* or records in a kRPC thread pool.
*/
typedef struct rdma_xprt_record rdma_xprt_record_t;
struct rdma_xprt_record {
int rtr_type; /* Type of rdma; IB/VI/RDDP */
SVCMASTERXPRT *rtr_xprt_ptr; /* Ptr to master xprt handle */
rdma_xprt_record_t *rtr_next; /* Ptr to next record */
};
typedef struct {
int rtg_count; /* Number transport records */
int rtg_poolid; /* Pool Id for this group */
rdma_xprt_record_t *rtg_listhead; /* Head of the records list */
} rdma_xprt_group_t;
extern int svc_rdma_kcreate(char *, SVC_CALLOUT_TABLE *, int,
rdma_xprt_group_t *);
extern void svc_rdma_kstop(SVCMASTERXPRT *);
extern void svc_rdma_kdestroy(SVCMASTERXPRT *);
extern void rdma_stop(rdma_xprt_group_t);
/*
* GSS cleanup method.
*/
extern void rpc_gss_cleanup(SVCXPRT *);
#else /* _KERNEL */
/*
* Lowest level dispatching -OR- who owns this process anyway.
* Somebody has to wait for incoming requests and then call the correct
* service routine. The routine svc_run does infinite waiting; i.e.,
* svc_run never returns.
* Since another (co-existant) package may wish to selectively wait for
* incoming calls or other events outside of the rpc architecture, the
* routine svc_getreq_poll is provided. It must be passed pollfds, the
* "in-place" results of a poll call (see poll, section 2).
*/
/*
* Global keeper of rpc service descriptors in use
* dynamic; must be inspected before each call to select or poll
*/
extern pollfd_t *svc_pollfd;
extern int svc_max_pollfd;
extern fd_set svc_fdset;
#if !defined(_LP64) && FD_SETSIZE > 1024
extern fd_set _new_svc_fdset;
#ifdef __PRAGMA_REDEFINE_EXTNAME
#pragma redefine_extname svc_fdset _new_svc_fdset
#else /* __PRAGMA_REDEFINE_EXTNAME */
#define svc_fdset _new_svc_fdset
#endif /* __PRAGMA_REDEFINE_EXTNAME */
#endif /* LP64 && FD_SETSIZE > 1024 */
#define svc_fds svc_fdset.fds_bits[0] /* compatibility */
/*
* A small program implemented by the svc_rpc implementation itself.
* Also see clnt.h for protocol numbers.
*/
#ifdef __STDC__
extern void svc_getreq(int);
extern void svc_getreq_common(const int);
extern void svc_getreqset(fd_set *); /* takes fdset instead of int */
extern void svc_getreq_poll(struct pollfd *, const int);
extern void svc_run(void);
extern void svc_exit(void);
#else /* __STDC__ */
extern void rpctest_service();
extern void svc_getreqset();
extern void svc_getreq();
extern void svc_getreq_common();
extern void svc_getreqset(); /* takes fdset instead of int */
extern void svc_getreq_poll();
extern void svc_run();
extern void svc_exit();
#endif /* __STDC__ */
/*
* Functions used to manage user file descriptors
*/
typedef int svc_input_id_t;
typedef void (*svc_callback_t)(svc_input_id_t id, int fd,
unsigned int events, void* cookie);
#ifdef __STDC__
extern svc_input_id_t svc_add_input(int fd, unsigned int events,
svc_callback_t user_callback,
void* cookie);
extern int svc_remove_input(svc_input_id_t id);
#else /* __STDC__ */
extern svc_input_id_t svc_add_input();
extern int svc_remove_input();
#endif
/*
* These are the existing service side transport implementations.
*
* Transport independent svc_create routine.
*/
#ifdef __STDC__
extern int svc_create(void (*)(struct svc_req *, SVCXPRT *),
const rpcprog_t, const rpcvers_t,
const char *);
/*
* void (*dispatch)(); -- dispatch routine
* const rpcprog_t prognum; -- program number
* const rpcvers_t versnum; -- version number
* const char *nettype; -- network type
*/
/*
* Generic server creation routine. It takes a netconfig structure
* instead of a nettype.
*/
extern SVCXPRT *svc_tp_create(void (*)(struct svc_req *, SVCXPRT *),
const rpcprog_t, const rpcvers_t,
const struct netconfig *);
/*
* void (*dispatch)(); -- dispatch routine
* const rpcprog_t prognum; -- program number
* const rpcvers_t versnum; -- version number
* const struct netconfig *nconf; -- netconfig structure
*/
/*
* Generic TLI create routine
*/
extern SVCXPRT *svc_tli_create(const int, const struct netconfig *,
const struct t_bind *, const uint_t,
const uint_t);
/*
* const int fd; -- connection end point
* const struct netconfig *nconf; -- netconfig structure
* const struct t_bind *bindaddr; -- local bind address
* const uint_t sendsz; -- max sendsize
* const uint_t recvsz; -- max recvsize
*/
/*
* Connectionless and connectionful create routines.
*/
extern SVCXPRT *svc_vc_create(const int, const uint_t, const uint_t);
/*
* const int fd; -- open connection end point
* const uint_t sendsize; -- max send size
* const uint_t recvsize; -- max recv size
*/
extern SVCXPRT *svc_dg_create(const int, const uint_t, const uint_t);
/*
* const int fd; -- open connection
* const uint_t sendsize; -- max send size
* const uint_t recvsize; -- max recv size
*/
/*
* the routine takes any *open* TLI file
* descriptor as its first input and is used for open connections.
*/
extern SVCXPRT *svc_fd_create(const int, const uint_t, const uint_t);
/*
* const int fd; -- open connection end point
* const uint_t sendsize; -- max send size
* const uint_t recvsize; -- max recv size
*/
/*
* Memory based rpc (for speed check and testing)
*/
extern SVCXPRT *svc_raw_create(void);
/*
* Creation of service over doors transport.
*/
extern SVCXPRT *svc_door_create(void (*)(struct svc_req *, SVCXPRT *),
const rpcprog_t, const rpcvers_t,
const uint_t);
/*
* void (*dispatch)(); -- dispatch routine
* const rpcprog_t prognum; -- program number
* const rpcvers_t versnum; -- version number
* const uint_t sendsize; -- send buffer size
*/
/*
* Service control interface
*/
extern bool_t svc_control(SVCXPRT *, const uint_t, void *);
/*
* SVCXPRT *svc; -- service to manipulate
* const uint_t req; -- request
* void *info; -- argument to request
*/
/*
* svc_dg_enable_cache() enables the cache on dg transports.
*/
extern int svc_dg_enablecache(SVCXPRT *, const uint_t);
#else /* __STDC__ */
extern int svc_create();
extern SVCXPRT *svc_tp_create();
extern SVCXPRT *svc_tli_create();
extern SVCXPRT *svc_vc_create();
extern SVCXPRT *svc_dg_create();
extern SVCXPRT *svc_fd_create();
extern SVCXPRT *svc_raw_create();
extern SVCXPRT *svc_door_create();
extern int svc_dg_enablecache();
#endif /* __STDC__ */
#ifdef PORTMAP
/* For backward compatibility */
#include <rpc/svc_soc.h>
#endif /* PORTMAP */
/*
* For user level MT hot server functions
*/
/*
* Different MT modes
*/
#define RPC_SVC_MT_NONE 0 /* default, single-threaded */
#define RPC_SVC_MT_AUTO 1 /* automatic MT mode */
#define RPC_SVC_MT_USER 2 /* user MT mode */
#ifdef __STDC__
extern void svc_done(SVCXPRT *);
#else
extern void svc_done();
#endif /* __STDC__ */
/*
* Obtaining local credentials.
*/
typedef struct __svc_local_cred_t {
uid_t euid; /* effective uid */
gid_t egid; /* effective gid */
uid_t ruid; /* real uid */
gid_t rgid; /* real gid */
pid_t pid; /* caller's pid, or -1 if not available */
} svc_local_cred_t;
#ifdef __STDC__
struct ucred_s;
extern void svc_fd_negotiate_ucred(int);
extern int svc_getcallerucred(const SVCXPRT *, struct ucred_s **);
extern bool_t svc_get_local_cred(SVCXPRT *, svc_local_cred_t *);
#else
extern void svc_fd_negotiate_ucred();
extern int svc_getcallerucred();
extern bool_t svc_get_local_cred();
#endif /* __STDC__ */
/*
* Private interfaces and structures for user level duplicate request caching.
* The interfaces and data structures are not committed and subject to
* change in future releases. Currently only intended for use by automountd.
*/
struct dupreq {
uint32_t dr_xid;
rpcproc_t dr_proc;
rpcvers_t dr_vers;
rpcprog_t dr_prog;
struct netbuf dr_addr;
struct netbuf dr_resp;
int dr_status;
time_t dr_time;
uint_t dr_hash;
struct dupreq *dr_next;
struct dupreq *dr_prev;
struct dupreq *dr_chain;
struct dupreq *dr_prevchain;
};
/*
* The fixedtime state is defined if we want to expand the routines to
* handle and encompass fixed size caches.
*/
#define DUPCACHE_FIXEDTIME 0
/*
* States of requests for duplicate request caching.
* These are the same as defined for the kernel.
*/
#define DUP_NEW 0x00 /* new entry */
#define DUP_INPROGRESS 0x01 /* request already going */
#define DUP_DONE 0x02 /* request done */
#define DUP_DROP 0x03 /* request dropped */
#define DUP_ERROR 0x04 /* error in dup req cache */
#ifdef __STDC__
extern bool_t __svc_dupcache_init(void *, int, char **);
extern int __svc_dup(struct svc_req *, caddr_t *, uint_t *, char *);
extern int __svc_dupdone(struct svc_req *, caddr_t, uint_t, int, char *);
extern bool_t __svc_vc_dupcache_init(SVCXPRT *, void *, int);
extern int __svc_vc_dup(struct svc_req *, caddr_t *, uint_t *);
extern int __svc_vc_dupdone(struct svc_req *, caddr_t, uint_t, int);
#else
extern bool_t __svc_dupcache_init();
extern int __svc_dup();
extern int __svc_dupdone();
extern bool_t __svc_vc_dupcache_init();
extern int __svc_vc_dup();
extern int __svc_vc_dupdone();
#endif /* __STDC__ */
#endif /* _KERNEL */
#ifdef __cplusplus
}
#endif
#endif /* !_RPC_SVC_H */