summaryrefslogtreecommitdiffstats
path: root/sys/nlm/nlm_prot_impl.c
diff options
context:
space:
mode:
authordfr <dfr@FreeBSD.org>2008-03-26 15:23:12 +0000
committerdfr <dfr@FreeBSD.org>2008-03-26 15:23:12 +0000
commit79d2dfdaa69db38c43daed9744a6dbd0568189b5 (patch)
treeeaf6a0fa52bc76253126814ddab4cbf78722a8a5 /sys/nlm/nlm_prot_impl.c
parent632e5d39f769e15274575347233e9d7aa364c0d6 (diff)
downloadFreeBSD-src-79d2dfdaa69db38c43daed9744a6dbd0568189b5.zip
FreeBSD-src-79d2dfdaa69db38c43daed9744a6dbd0568189b5.tar.gz
Add the new kernel-mode NFS Lock Manager. To use it instead of the
user-mode lock manager, build a kernel with the NFSLOCKD option and add '-k' to 'rpc_lockd_flags' in rc.conf. Highlights include: * Thread-safe kernel RPC client - many threads can use the same RPC client handle safely with replies being de-multiplexed at the socket upcall (typically driven directly by the NIC interrupt) and handed off to whichever thread matches the reply. For UDP sockets, many RPC clients can share the same socket. This allows the use of a single privileged UDP port number to talk to an arbitrary number of remote hosts. * Single-threaded kernel RPC server. Adding support for multi-threaded server would be relatively straightforward and would follow approximately the Solaris KPI. A single thread should be sufficient for the NLM since it should rarely block in normal operation. * Kernel mode NLM server supporting cancel requests and granted callbacks. I've tested the NLM server reasonably extensively - it passes both my own tests and the NFS Connectathon locking tests running on Solaris, Mac OS X and Ubuntu Linux. * Userland NLM client supported. While the NLM server doesn't have support for the local NFS client's locking needs, it does have to field async replies and granted callbacks from remote NLMs that the local client has contacted. We relay these replies to the userland rpc.lockd over a local domain RPC socket. * Robust deadlock detection for the local lock manager. In particular it will detect deadlocks caused by a lock request that covers more than one blocking request. As required by the NLM protocol, all deadlock detection happens synchronously - a user is guaranteed that if a lock request isn't rejected immediately, the lock will eventually be granted. The old system allowed for a 'deferred deadlock' condition where a blocked lock request could wake up and find that some other deadlock-causing lock owner had beaten them to the lock. * Since both local and remote locks are managed by the same kernel locking code, local and remote processes can safely use file locks for mutual exclusion. Local processes have no fairness advantage compared to remote processes when contending to lock a region that has just been unlocked - the local lock manager enforces a strict first-come first-served model for both local and remote lockers. Sponsored by: Isilon Systems PR: 95247 107555 115524 116679 MFC after: 2 weeks
Diffstat (limited to 'sys/nlm/nlm_prot_impl.c')
-rw-r--r--sys/nlm/nlm_prot_impl.c1783
1 files changed, 1783 insertions, 0 deletions
diff --git a/sys/nlm/nlm_prot_impl.c b/sys/nlm/nlm_prot_impl.c
new file mode 100644
index 0000000..106f4b5
--- /dev/null
+++ b/sys/nlm/nlm_prot_impl.c
@@ -0,0 +1,1783 @@
+/*-
+ * Copyright (c) 2008 Isilon Inc http://www.isilon.com/
+ * Authors: Doug Rabson <dfr@rabson.org>
+ * Developed with Red Inc: Alfred Perlstein <alfred@freebsd.org>
+ *
+ * Redistribution and use in source and binary forms, with or without
+ * modification, are permitted provided that the following conditions
+ * are met:
+ * 1. Redistributions of source code must retain the above copyright
+ * notice, this list of conditions and the following disclaimer.
+ * 2. Redistributions in binary form must reproduce the above copyright
+ * notice, this list of conditions and the following disclaimer in the
+ * documentation and/or other materials provided with the distribution.
+ *
+ * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
+ * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
+ * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
+ * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
+ * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
+ * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
+ * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
+ * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
+ * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
+ * SUCH DAMAGE.
+ */
+
+#include "opt_inet6.h"
+
+#include <sys/cdefs.h>
+__FBSDID("$FreeBSD$");
+
+#include <sys/param.h>
+#include <sys/fcntl.h>
+#include <sys/kernel.h>
+#include <sys/lockf.h>
+#include <sys/malloc.h>
+#include <sys/mount.h>
+#include <sys/priv.h>
+#include <sys/proc.h>
+#include <sys/socket.h>
+#include <sys/socketvar.h>
+#include <sys/syscall.h>
+#include <sys/sysctl.h>
+#include <sys/sysent.h>
+#include <sys/sysproto.h>
+#include <sys/systm.h>
+#include <sys/taskqueue.h>
+#include <sys/unistd.h>
+#include <sys/vnode.h>
+
+#include "nlm_prot.h"
+#include "sm_inter.h"
+#include "nlm.h"
+#include <rpc/rpc_com.h>
+#include <rpc/rpcb_prot.h>
+
+MALLOC_DEFINE(M_NLM, "NLM", "Network Lock Manager");
+
+/*
+ * If a host is inactive (and holds no locks) for this amount of
+ * seconds, we consider it idle and stop tracking it.
+ */
+#define NLM_IDLE_TIMEOUT 30
+
+/*
+ * We check the host list for idle every few seconds.
+ */
+#define NLM_IDLE_PERIOD 5
+
+/*
+ * Support for sysctl vfs.nlm.sysid
+ */
+SYSCTL_NODE(_vfs, OID_AUTO, nlm, CTLFLAG_RW, NULL, "Network Lock Manager");
+SYSCTL_NODE(_vfs_nlm, OID_AUTO, sysid, CTLFLAG_RW, NULL, "");
+
+/*
+ * Syscall hooks
+ */
+static int nlm_syscall_offset = SYS_nlm_syscall;
+static struct sysent nlm_syscall_prev_sysent;
+MAKE_SYSENT(nlm_syscall);
+static bool_t nlm_syscall_registered = FALSE;
+
+/*
+ * Debug level passed in from userland. We also support a sysctl hook
+ * so that it can be changed on a live system.
+ */
+static int nlm_debug_level;
+SYSCTL_INT(_debug, OID_AUTO, nlm_debug, CTLFLAG_RW, &nlm_debug_level, 0, "");
+
+/*
+ * Grace period handling. The value of nlm_grace_threshold is the
+ * value of time_uptime after which we are serving requests normally.
+ */
+static time_t nlm_grace_threshold;
+
+/*
+ * We check for idle hosts if time_uptime is greater than
+ * nlm_next_idle_check,
+ */
+static time_t nlm_next_idle_check;
+
+/*
+ * A socket to use for RPC - shared by all IPv4 RPC clients.
+ */
+static struct socket *nlm_socket;
+
+#ifdef INET6
+
+/*
+ * A socket to use for RPC - shared by all IPv6 RPC clients.
+ */
+static struct socket *nlm_socket6;
+
+#endif
+
+/*
+ * An RPC client handle that can be used to communicate with the local
+ * NSM.
+ */
+static CLIENT *nlm_nsm;
+
+/*
+ * An RPC client handle that can be used to communicate with the
+ * userland part of lockd.
+ */
+static CLIENT *nlm_lockd;
+
+/*
+ * Locks:
+ * (l) locked by nh_lock
+ * (s) only accessed via server RPC which is single threaded
+ * (c) const until freeing
+ */
+
+/*
+ * A pending asynchronous lock request, stored on the nc_pending list
+ * of the NLM host.
+ */
+struct nlm_async_lock {
+ TAILQ_ENTRY(nlm_async_lock) af_link; /* (l) host's list of locks */
+ struct task af_task; /* (c) async callback details */
+ void *af_cookie; /* (l) lock manager cancel token */
+ struct vnode *af_vp; /* (l) vnode to lock */
+ struct flock af_fl; /* (c) lock details */
+ struct nlm_host *af_host; /* (c) host which is locking */
+ nlm4_testargs af_granted; /* (c) notification details */
+};
+TAILQ_HEAD(nlm_async_lock_list, nlm_async_lock);
+
+/*
+ * NLM host.
+ */
+struct nlm_host {
+ struct mtx nh_lock;
+ TAILQ_ENTRY(nlm_host) nh_link; /* (s) global list of hosts */
+ char *nh_caller_name; /* (c) printable name of host */
+ uint32_t nh_sysid; /* (c) our allocaed system ID */
+ char nh_sysid_string[10]; /* (c) string rep. of sysid */
+ struct sockaddr_storage nh_addr; /* (s) remote address of host */
+ CLIENT *nh_rpc; /* (s) RPC handle to send to host */
+ rpcvers_t nh_vers; /* (s) NLM version of host */
+ int nh_state; /* (s) last seen NSM state of host */
+ bool_t nh_monitored; /* (s) TRUE if local NSM is monitoring */
+ time_t nh_idle_timeout; /* (s) Time at which host is idle */
+ struct sysctl_ctx_list nh_sysctl; /* (c) vfs.nlm.sysid nodes */
+ struct nlm_async_lock_list nh_pending; /* (l) pending async locks */
+ struct nlm_async_lock_list nh_finished; /* (l) finished async locks */
+};
+TAILQ_HEAD(nlm_host_list, nlm_host);
+
+static struct nlm_host_list nlm_hosts;
+static uint32_t nlm_next_sysid = 1;
+
+static void nlm_host_unmonitor(struct nlm_host *);
+
+/**********************************************************************/
+
+/*
+ * Initialise NLM globals.
+ */
+static void
+nlm_init(void *dummy)
+{
+ int error;
+
+ TAILQ_INIT(&nlm_hosts);
+
+ error = syscall_register(&nlm_syscall_offset, &nlm_syscall_sysent,
+ &nlm_syscall_prev_sysent);
+ if (error)
+ printf("Can't register NLM syscall\n");
+ else
+ nlm_syscall_registered = TRUE;
+}
+SYSINIT(nlm_init, SI_SUB_LOCK, SI_ORDER_FIRST, nlm_init, NULL);
+
+static void
+nlm_uninit(void *dummy)
+{
+
+ if (nlm_syscall_registered)
+ syscall_deregister(&nlm_syscall_offset,
+ &nlm_syscall_prev_sysent);
+}
+SYSUNINIT(nlm_uninit, SI_SUB_LOCK, SI_ORDER_FIRST, nlm_uninit, NULL);
+
+/*
+ * Copy a struct netobj.
+ */
+void
+nlm_copy_netobj(struct netobj *dst, struct netobj *src,
+ struct malloc_type *type)
+{
+
+ dst->n_len = src->n_len;
+ dst->n_bytes = malloc(src->n_len, type, M_WAITOK);
+ memcpy(dst->n_bytes, src->n_bytes, src->n_len);
+}
+
+/*
+ * Create an RPC client handle for the given (address,prog,vers)
+ * triple using UDP.
+ */
+static CLIENT *
+nlm_get_rpc(struct sockaddr *sa, rpcprog_t prog, rpcvers_t vers)
+{
+ const char *wchan = "nlmrcv";
+ const char* protofmly;
+ struct sockaddr_storage ss;
+ struct socket *so;
+ CLIENT *rpcb;
+ struct timeval timo;
+ RPCB parms;
+ char *uaddr;
+ enum clnt_stat stat;
+ int rpcvers;
+
+ /*
+ * First we need to contact the remote RPCBIND service to find
+ * the right port.
+ */
+ memcpy(&ss, sa, sa->sa_len);
+ switch (ss.ss_family) {
+ case AF_INET:
+ ((struct sockaddr_in *)&ss)->sin_port = htons(111);
+ protofmly = "inet";
+ so = nlm_socket;
+ break;
+
+#ifdef INET6
+ case AF_INET6:
+ ((struct sockaddr_in6 *)&ss)->sin6_port = htons(111);
+ protofmly = "inet6";
+ so = nlm_socket6;
+ break;
+#endif
+
+ default:
+ /*
+ * Unsupported address family - fail.
+ */
+ return (NULL);
+ }
+
+ rpcb = clnt_dg_create(so, (struct sockaddr *)&ss,
+ RPCBPROG, RPCBVERS4, 0, 0);
+ if (!rpcb)
+ return (NULL);
+
+ parms.r_prog = prog;
+ parms.r_vers = vers;
+ parms.r_netid = "udp";
+ parms.r_addr = "";
+ parms.r_owner = "";
+
+ /*
+ * Use the default timeout.
+ */
+ timo.tv_sec = 25;
+ timo.tv_usec = 0;
+again:
+ uaddr = NULL;
+ stat = CLNT_CALL(rpcb, (rpcprog_t) RPCBPROC_GETADDR,
+ (xdrproc_t) xdr_rpcb, &parms,
+ (xdrproc_t) xdr_wrapstring, &uaddr, timo);
+ if (stat == RPC_PROGVERSMISMATCH) {
+ /*
+ * Try RPCBIND version 3 if we haven't already.
+ *
+ * XXX fall back to portmap?
+ */
+ CLNT_CONTROL(rpcb, CLGET_VERS, &rpcvers);
+ if (rpcvers == RPCBVERS4) {
+ rpcvers = RPCBVERS;
+ CLNT_CONTROL(rpcb, CLSET_VERS, &rpcvers);
+ goto again;
+ }
+ }
+
+ if (stat == RPC_SUCCESS) {
+ /*
+ * We have a reply from the remote RPCBIND - turn it into an
+ * appropriate address and make a new client that can talk to
+ * the remote NLM.
+ *
+ * XXX fixup IPv6 scope ID.
+ */
+ struct netbuf *a;
+ a = __rpc_uaddr2taddr_af(ss.ss_family, uaddr);
+ memcpy(&ss, a->buf, a->len);
+ free(a->buf, M_RPC);
+ free(a, M_RPC);
+ xdr_free((xdrproc_t) xdr_wrapstring, &uaddr);
+ } else if (stat == RPC_PROGVERSMISMATCH) {
+ /*
+ * Try portmap.
+ */
+ struct pmap mapping;
+ u_short port;
+
+ rpcvers = PMAPVERS;
+ CLNT_CONTROL(rpcb, CLSET_VERS, &rpcvers);
+
+
+ mapping.pm_prog = parms.r_prog;
+ mapping.pm_vers = parms.r_vers;
+ mapping.pm_prot = IPPROTO_UDP;
+ mapping.pm_port = 0;
+
+ stat = CLNT_CALL(rpcb, (rpcprog_t) PMAPPROC_GETPORT,
+ (xdrproc_t) xdr_pmap, &mapping,
+ (xdrproc_t) xdr_u_short, &port, timo);
+
+ if (stat == RPC_SUCCESS) {
+ switch (ss.ss_family) {
+ case AF_INET:
+ ((struct sockaddr_in *)&ss)->sin_port =
+ htons(port);
+ break;
+
+#ifdef INET6
+ case AF_INET6:
+ ((struct sockaddr_in6 *)&ss)->sin6_port =
+ htons(port);
+ break;
+#endif
+ }
+ }
+ }
+ if (stat != RPC_SUCCESS) {
+ printf("NLM: failed to contact remote rpcbind, stat = %d\n",
+ (int) stat);
+ return (NULL);
+ }
+
+ /*
+ * Re-use the client we used to speak to rpcbind.
+ */
+ CLNT_CONTROL(rpcb, CLSET_SVC_ADDR, &ss);
+ CLNT_CONTROL(rpcb, CLSET_PROG, &prog);
+ CLNT_CONTROL(rpcb, CLSET_VERS, &vers);
+ CLNT_CONTROL(rpcb, CLSET_WAITCHAN, &wchan);
+ rpcb->cl_auth = authunix_create(curthread->td_ucred);
+
+ return (rpcb);
+}
+
+/*
+ * This async callback after when an async lock request has been
+ * granted. We notify the host which initiated the request.
+ */
+static void
+nlm_lock_callback(void *arg, int pending)
+{
+ struct nlm_async_lock *af = (struct nlm_async_lock *) arg;
+
+ if (nlm_debug_level >= 2)
+ printf("NLM: async lock %p for %s (sysid %d) granted\n",
+ af, af->af_host->nh_caller_name,
+ af->af_host->nh_sysid);
+
+ /*
+ * Send the results back to the host.
+ *
+ * Note: there is a possible race here with nlm_host_notify
+ * destroying teh RPC client. To avoid problems, the first
+ * thing nlm_host_notify does is to cancel pending async lock
+ * requests.
+ */
+ if (af->af_host->nh_vers == NLM_VERS4) {
+ nlm4_granted_msg_4(&af->af_granted,
+ NULL, af->af_host->nh_rpc);
+ } else {
+ /*
+ * Back-convert to legacy protocol
+ */
+ nlm_testargs granted;
+ granted.cookie = af->af_granted.cookie;
+ granted.exclusive = af->af_granted.exclusive;
+ granted.alock.caller_name =
+ af->af_granted.alock.caller_name;
+ granted.alock.fh = af->af_granted.alock.fh;
+ granted.alock.oh = af->af_granted.alock.oh;
+ granted.alock.svid = af->af_granted.alock.svid;
+ granted.alock.l_offset =
+ af->af_granted.alock.l_offset;
+ granted.alock.l_len =
+ af->af_granted.alock.l_len;
+
+ nlm_granted_msg_1(&granted,
+ NULL, af->af_host->nh_rpc);
+ }
+
+ /*
+ * Move this entry to the nh_finished list. Someone else will
+ * free it later - its too hard to do it here safely without
+ * racing with cancel.
+ *
+ * XXX possibly we should have a third "granted sent but not
+ * ack'ed" list so that we can re-send the granted message.
+ */
+ mtx_lock(&af->af_host->nh_lock);
+ TAILQ_REMOVE(&af->af_host->nh_pending, af, af_link);
+ TAILQ_INSERT_TAIL(&af->af_host->nh_finished, af, af_link);
+ mtx_unlock(&af->af_host->nh_lock);
+}
+
+/*
+ * Free an async lock request. The request must have been removed from
+ * any list.
+ */
+static void
+nlm_free_async_lock(struct nlm_async_lock *af)
+{
+ /*
+ * Free an async lock.
+ */
+ xdr_free((xdrproc_t) xdr_nlm4_testargs, &af->af_granted);
+ if (af->af_vp)
+ vrele(af->af_vp);
+ free(af, M_NLM);
+}
+
+/*
+ * Cancel our async request - this must be called with
+ * af->nh_host->nh_lock held. This is slightly complicated by a
+ * potential race with our own callback. If we fail to cancel the
+ * lock, it must already have been granted - we make sure our async
+ * task has completed by calling taskqueue_drain in this case.
+ */
+static int
+nlm_cancel_async_lock(struct nlm_async_lock *af)
+{
+ struct nlm_host *host = af->af_host;
+ int error;
+
+ mtx_assert(&host->nh_lock, MA_OWNED);
+
+ mtx_unlock(&host->nh_lock);
+
+ error = VOP_ADVLOCKASYNC(af->af_vp, NULL, F_CANCEL, &af->af_fl,
+ F_REMOTE, NULL, &af->af_cookie);
+
+ if (error) {
+ /*
+ * We failed to cancel - make sure our callback has
+ * completed before we continue.
+ */
+ taskqueue_drain(taskqueue_thread, &af->af_task);
+ }
+
+ mtx_lock(&host->nh_lock);
+
+ if (!error) {
+ if (nlm_debug_level >= 2)
+ printf("NLM: async lock %p for %s (sysid %d) "
+ "cancelled\n",
+ af, host->nh_caller_name, host->nh_sysid);
+
+ /*
+ * Remove from the nh_pending list and free now that
+ * we are safe from the callback.
+ */
+ TAILQ_REMOVE(&host->nh_pending, af, af_link);
+ mtx_unlock(&host->nh_lock);
+ nlm_free_async_lock(af);
+ mtx_lock(&host->nh_lock);
+ }
+
+ return (error);
+}
+
+static void
+nlm_free_finished_locks(struct nlm_host *host)
+{
+ struct nlm_async_lock *af;
+
+ mtx_lock(&host->nh_lock);
+ while ((af = TAILQ_FIRST(&host->nh_finished)) != NULL) {
+ TAILQ_REMOVE(&host->nh_finished, af, af_link);
+ mtx_unlock(&host->nh_lock);
+ nlm_free_async_lock(af);
+ mtx_lock(&host->nh_lock);
+ }
+ mtx_unlock(&host->nh_lock);
+}
+
+/*
+ * This is called when we receive a host state change
+ * notification. We unlock any active locks owned by the host.
+ */
+static void
+nlm_host_notify(struct nlm_host *host, int newstate, bool_t destroy)
+{
+ struct nlm_async_lock *af;
+
+ if (newstate) {
+ if (nlm_debug_level >= 1)
+ printf("NLM: host %s (sysid %d) rebooted, new "
+ "state is %d\n",
+ host->nh_caller_name, host->nh_sysid, newstate);
+ }
+
+ /*
+ * Cancel any pending async locks for this host.
+ */
+ mtx_lock(&host->nh_lock);
+ while ((af = TAILQ_FIRST(&host->nh_pending)) != NULL) {
+ /*
+ * nlm_cancel_async_lock will remove the entry from
+ * nh_pending and free it.
+ */
+ nlm_cancel_async_lock(af);
+ }
+ mtx_unlock(&host->nh_lock);
+ nlm_free_finished_locks(host);
+
+ /*
+ * The host just rebooted - trash its locks and forget any
+ * RPC client handle that we may have for it.
+ */
+ lf_clearremotesys(host->nh_sysid);
+ if (host->nh_rpc) {
+ AUTH_DESTROY(host->nh_rpc->cl_auth);
+ CLNT_DESTROY(host->nh_rpc);
+ host->nh_rpc = NULL;
+ }
+ host->nh_state = newstate;
+
+ /*
+ * Destroy the host if the caller believes that it won't be
+ * used again. This is safe enough - if we see the same name
+ * again, we will just create a new host.
+ */
+ if (destroy) {
+ TAILQ_REMOVE(&nlm_hosts, host, nh_link);
+ mtx_destroy(&host->nh_lock);
+ sysctl_ctx_free(&host->nh_sysctl);
+ free(host->nh_caller_name, M_NLM);
+ free(host, M_NLM);
+ }
+}
+
+/*
+ * Sysctl handler to count the number of locks for a sysid.
+ */
+static int
+nlm_host_lock_count_sysctl(SYSCTL_HANDLER_ARGS)
+{
+ struct nlm_host *host;
+ int count;
+
+ host = oidp->oid_arg1;
+ count = lf_countlocks(host->nh_sysid);
+ return sysctl_handle_int(oidp, &count, 0, req);
+}
+
+/*
+ * Create a new NLM host.
+ */
+static struct nlm_host *
+nlm_create_host(const char* caller_name)
+{
+ struct nlm_host *host;
+ struct sysctl_oid *oid;
+
+ if (nlm_debug_level >= 1)
+ printf("NLM: new host %s (sysid %d)\n",
+ caller_name, nlm_next_sysid);
+ host = malloc(sizeof(struct nlm_host), M_NLM, M_WAITOK|M_ZERO);
+ mtx_init(&host->nh_lock, "nh_lock", NULL, MTX_DEF);
+ host->nh_caller_name = strdup(caller_name, M_NLM);
+ host->nh_sysid = nlm_next_sysid++;
+ snprintf(host->nh_sysid_string, sizeof(host->nh_sysid_string),
+ "%d", host->nh_sysid);
+ host->nh_rpc = NULL;
+ host->nh_vers = 0;
+ host->nh_state = 0;
+ host->nh_monitored = FALSE;
+ TAILQ_INIT(&host->nh_pending);
+ TAILQ_INIT(&host->nh_finished);
+ TAILQ_INSERT_TAIL(&nlm_hosts, host, nh_link);
+
+ sysctl_ctx_init(&host->nh_sysctl);
+ oid = SYSCTL_ADD_NODE(&host->nh_sysctl,
+ SYSCTL_STATIC_CHILDREN(_vfs_nlm_sysid),
+ OID_AUTO, host->nh_sysid_string, CTLFLAG_RD, NULL, "");
+ SYSCTL_ADD_STRING(&host->nh_sysctl, SYSCTL_CHILDREN(oid), OID_AUTO,
+ "hostname", CTLFLAG_RD, host->nh_caller_name, 0, "");
+ SYSCTL_ADD_INT(&host->nh_sysctl, SYSCTL_CHILDREN(oid), OID_AUTO,
+ "version", CTLFLAG_RD, &host->nh_vers, 0, "");
+ SYSCTL_ADD_INT(&host->nh_sysctl, SYSCTL_CHILDREN(oid), OID_AUTO,
+ "monitored", CTLFLAG_RD, &host->nh_monitored, 0, "");
+ SYSCTL_ADD_PROC(&host->nh_sysctl, SYSCTL_CHILDREN(oid), OID_AUTO,
+ "lock_count", CTLTYPE_INT | CTLFLAG_RD, host, 0,
+ nlm_host_lock_count_sysctl, "I", "");
+
+ return (host);
+}
+
+/*
+ * Return non-zero if the address parts of the two sockaddrs are the
+ * same.
+ */
+static int
+nlm_compare_addr(const struct sockaddr *a, const struct sockaddr *b)
+{
+ const struct sockaddr_in *a4, *b4;
+#ifdef INET6
+ const struct sockaddr_in6 *a6, *b6;
+#endif
+
+ if (a->sa_family != b->sa_family)
+ return (FALSE);
+
+ switch (a->sa_family) {
+ case AF_INET:
+ a4 = (const struct sockaddr_in *) a;
+ b4 = (const struct sockaddr_in *) b;
+ return !memcmp(&a4->sin_addr, &b4->sin_addr,
+ sizeof(a4->sin_addr));
+#ifdef INET6
+ case AF_INET6:
+ a6 = (const struct sockaddr_in6 *) a;
+ b6 = (const struct sockaddr_in6 *) b;
+ return !memcmp(&a6->sin6_addr, &b6->sin6_addr,
+ sizeof(a6->sin6_addr));
+#endif
+ }
+
+ return (0);
+}
+
+/*
+ * Check for idle hosts and stop monitoring them. We could also free
+ * the host structure here, possibly after a larger timeout but that
+ * would require some care to avoid races with
+ * e.g. nlm_host_lock_count_sysctl.
+ */
+static void
+nlm_check_idle(void)
+{
+ struct nlm_host *host;
+
+ if (time_uptime <= nlm_next_idle_check)
+ return;
+
+ nlm_next_idle_check = time_uptime + NLM_IDLE_PERIOD;
+
+ TAILQ_FOREACH(host, &nlm_hosts, nh_link) {
+ if (host->nh_monitored
+ && time_uptime > host->nh_idle_timeout) {
+ if (lf_countlocks(host->nh_sysid) > 0) {
+ host->nh_idle_timeout =
+ time_uptime + NLM_IDLE_TIMEOUT;
+ continue;
+ }
+ nlm_host_unmonitor(host);
+ }
+ }
+}
+
+/*
+ * Search for an existing NLM host that matches the given name
+ * (typically the caller_name element of an nlm4_lock). If none is
+ * found, create a new host. If 'rqstp' is non-NULL, record the remote
+ * address of the host so that we can call it back for async
+ * responses.
+ */
+struct nlm_host *
+nlm_find_host_by_name(const char *name, struct svc_req *rqstp)
+{
+ struct nlm_host *host;
+
+ nlm_check_idle();
+
+ /*
+ * The remote host is determined by caller_name.
+ */
+ TAILQ_FOREACH(host, &nlm_hosts, nh_link) {
+ if (!strcmp(host->nh_caller_name, name))
+ break;
+ }
+
+ if (!host)
+ host = nlm_create_host(name);
+ host->nh_idle_timeout = time_uptime + NLM_IDLE_TIMEOUT;
+
+ /*
+ * If we have an RPC request, record the remote address so
+ * that can send async replies etc.
+ */
+ if (rqstp) {
+ struct netbuf *addr = &rqstp->rq_xprt->xp_rtaddr;
+
+ KASSERT(addr->len < sizeof(struct sockaddr_storage),
+ ("Strange remote transport address length"));
+
+ /*
+ * If we have seen an address before and we currently
+ * have an RPC client handle, make sure the address is
+ * the same, otherwise discard the client handle.
+ */
+ if (host->nh_addr.ss_len && host->nh_rpc) {
+ if (!nlm_compare_addr(
+ (struct sockaddr *) &host->nh_addr,
+ (struct sockaddr *) addr->buf)
+ || host->nh_vers != rqstp->rq_vers) {
+ AUTH_DESTROY(host->nh_rpc->cl_auth);
+ CLNT_DESTROY(host->nh_rpc);
+ host->nh_rpc = NULL;
+ }
+ }
+ memcpy(&host->nh_addr, addr->buf, addr->len);
+ host->nh_vers = rqstp->rq_vers;
+ }
+
+ return (host);
+}
+
+/*
+ * Search for an existing NLM host that matches the given remote
+ * address. If none is found, create a new host with the requested
+ * address and remember 'vers' as the NLM protocol version to use for
+ * that host.
+ */
+struct nlm_host *
+nlm_find_host_by_addr(const struct sockaddr *addr, int vers)
+{
+ struct nlm_host *host;
+
+ nlm_check_idle();
+
+ /*
+ * The remote host is determined by caller_name.
+ */
+ TAILQ_FOREACH(host, &nlm_hosts, nh_link) {
+ if (nlm_compare_addr(addr,
+ (const struct sockaddr *) &host->nh_addr))
+ break;
+ }
+
+ if (!host) {
+ /*
+ * Fake up a name using inet_ntop. This buffer is
+ * large enough for an IPv6 address.
+ */
+ char tmp[sizeof "ffff:ffff:ffff:ffff:ffff:ffff:255.255.255.255"];
+ switch (addr->sa_family) {
+ case AF_INET:
+ __rpc_inet_ntop(AF_INET,
+ &((const struct sockaddr_in *) addr)->sin_addr,
+ tmp, sizeof tmp);
+ break;
+#ifdef INET6
+ case AF_INET6:
+ __rpc_inet_ntop(AF_INET6,
+ &((const struct sockaddr_in6 *) addr)->sin6_addr,
+ tmp, sizeof tmp);
+ break;
+#endif
+ default:
+ strcmp(tmp, "<unknown>");
+ }
+ host = nlm_create_host(tmp);
+ memcpy(&host->nh_addr, addr, addr->sa_len);
+ host->nh_vers = vers;
+ }
+ host->nh_idle_timeout = time_uptime + NLM_IDLE_TIMEOUT;
+
+ return (host);
+}
+
+/*
+ * Find the NLM host that matches the value of 'sysid'. If none
+ * exists, return NULL.
+ */
+static struct nlm_host *
+nlm_find_host_by_sysid(int sysid)
+{
+ struct nlm_host *host;
+
+ TAILQ_FOREACH(host, &nlm_hosts, nh_link) {
+ if (host->nh_sysid == sysid)
+ return (host);
+ }
+
+ return (NULL);
+}
+
+/*
+ * Unregister this NLM host with the local NSM due to idleness.
+ */
+static void
+nlm_host_unmonitor(struct nlm_host *host)
+{
+ mon_id smmonid;
+ sm_stat_res smstat;
+ struct timeval timo;
+ enum clnt_stat stat;
+
+ if (nlm_debug_level >= 1)
+ printf("NLM: unmonitoring %s (sysid %d)\n",
+ host->nh_caller_name, host->nh_sysid);
+
+ /*
+ * We put our assigned system ID value in the priv field to
+ * make it simpler to find the host if we are notified of a
+ * host restart.
+ */
+ smmonid.mon_name = host->nh_caller_name;
+ smmonid.my_id.my_name = "localhost";
+ smmonid.my_id.my_prog = NLM_PROG;
+ smmonid.my_id.my_vers = NLM_SM;
+ smmonid.my_id.my_proc = NLM_SM_NOTIFY;
+
+ timo.tv_sec = 25;
+ timo.tv_usec = 0;
+ stat = CLNT_CALL(nlm_nsm, SM_UNMON,
+ (xdrproc_t) xdr_mon, &smmonid,
+ (xdrproc_t) xdr_sm_stat, &smstat, timo);
+
+ if (stat != RPC_SUCCESS) {
+ printf("Failed to contact local NSM - rpc error %d\n", stat);
+ return;
+ }
+ if (smstat.res_stat == stat_fail) {
+ printf("Local NSM refuses to unmonitor %s\n",
+ host->nh_caller_name);
+ return;
+ }
+
+ host->nh_monitored = FALSE;
+}
+
+/*
+ * Register this NLM host with the local NSM so that we can be
+ * notified if it reboots.
+ */
+static void
+nlm_host_monitor(struct nlm_host *host, int state)
+{
+ mon smmon;
+ sm_stat_res smstat;
+ struct timeval timo;
+ enum clnt_stat stat;
+
+ if (host->nh_state && state && host->nh_state != state) {
+ /*
+ * The host rebooted without telling us. Trash its
+ * locks.
+ */
+ nlm_host_notify(host, state, FALSE);
+ }
+
+ if (state && !host->nh_state) {
+ /*
+ * This is the first time we have seen an NSM state
+ * value for this host. We record it here to help
+ * detect host reboots.
+ */
+ host->nh_state = state;
+ if (nlm_debug_level >= 1)
+ printf("NLM: host %s (sysid %d) has NSM state %d\n",
+ host->nh_caller_name, host->nh_sysid, state);
+ }
+
+ if (host->nh_monitored)
+ return;
+
+ if (nlm_debug_level >= 1)
+ printf("NLM: monitoring %s (sysid %d)\n",
+ host->nh_caller_name, host->nh_sysid);
+
+ /*
+ * We put our assigned system ID value in the priv field to
+ * make it simpler to find the host if we are notified of a
+ * host restart.
+ */
+ smmon.mon_id.mon_name = host->nh_caller_name;
+ smmon.mon_id.my_id.my_name = "localhost";
+ smmon.mon_id.my_id.my_prog = NLM_PROG;
+ smmon.mon_id.my_id.my_vers = NLM_SM;
+ smmon.mon_id.my_id.my_proc = NLM_SM_NOTIFY;
+ memcpy(smmon.priv, &host->nh_sysid, sizeof(host->nh_sysid));
+
+ timo.tv_sec = 25;
+ timo.tv_usec = 0;
+ stat = CLNT_CALL(nlm_nsm, SM_MON,
+ (xdrproc_t) xdr_mon, &smmon,
+ (xdrproc_t) xdr_sm_stat, &smstat, timo);
+
+ if (stat != RPC_SUCCESS) {
+ printf("Failed to contact local NSM - rpc error %d\n", stat);
+ return;
+ }
+ if (smstat.res_stat == stat_fail) {
+ printf("Local NSM refuses to monitor %s\n",
+ host->nh_caller_name);
+ return;
+ }
+
+ host->nh_monitored = TRUE;
+}
+
+/*
+ * Return an RPC client handle that can be used to talk to the NLM
+ * running on the given host.
+ */
+CLIENT *
+nlm_host_get_rpc(struct nlm_host *host)
+{
+ struct timeval zero;
+
+ if (host->nh_rpc)
+ return (host->nh_rpc);
+
+ /*
+ * Set the send timeout to zero - we only use this rpc handle
+ * for sending async replies which have no return value.
+ */
+ host->nh_rpc = nlm_get_rpc((struct sockaddr *)&host->nh_addr,
+ NLM_PROG, host->nh_vers);
+
+ if (host->nh_rpc) {
+ zero.tv_sec = 0;
+ zero.tv_usec = 0;
+ CLNT_CONTROL(host->nh_rpc, CLSET_TIMEOUT, &zero);
+
+ /*
+ * Monitor the host - if it reboots, the address of
+ * its NSM might change so we must discard our RPC
+ * handle.
+ */
+ nlm_host_monitor(host, 0);
+ }
+
+ return (host->nh_rpc);
+}
+
+/**********************************************************************/
+
+/*
+ * Syscall interface with userland.
+ */
+
+extern void nlm_prog_0(struct svc_req *rqstp, SVCXPRT *transp);
+extern void nlm_prog_1(struct svc_req *rqstp, SVCXPRT *transp);
+extern void nlm_prog_3(struct svc_req *rqstp, SVCXPRT *transp);
+extern void nlm_prog_4(struct svc_req *rqstp, SVCXPRT *transp);
+
+static int
+nlm_register_services(SVCPOOL *pool, int addr_count, char **addrs)
+{
+ static rpcvers_t versions[] = {
+ NLM_SM, NLM_VERS, NLM_VERSX, NLM_VERS4
+ };
+ static void (*dispatchers[])(struct svc_req *, SVCXPRT *) = {
+ nlm_prog_0, nlm_prog_1, nlm_prog_3, nlm_prog_4
+ };
+ static const int version_count = sizeof(versions) / sizeof(versions[0]);
+
+ SVCXPRT **xprts;
+ char netid[16];
+ char uaddr[128];
+ struct netconfig *nconf;
+ int i, j, error;
+
+ if (!addr_count) {
+ printf("NLM: no service addresses given - can't start server");
+ return (EINVAL);
+ }
+
+ xprts = malloc(addr_count * sizeof(SVCXPRT *), M_NLM, M_WAITOK);
+ for (i = 0; i < version_count; i++) {
+ for (j = 0; j < addr_count; j++) {
+ /*
+ * Create transports for the first version and
+ * then just register everything else to the
+ * same transports.
+ */
+ if (i == 0) {
+ char *up;
+
+ error = copyin(&addrs[2*j], &up,
+ sizeof(char*));
+ if (error)
+ goto out;
+ error = copyinstr(up, netid, sizeof(netid),
+ NULL);
+ if (error)
+ goto out;
+ error = copyin(&addrs[2*j+1], &up,
+ sizeof(char*));
+ if (error)
+ goto out;
+ error = copyinstr(up, uaddr, sizeof(uaddr),
+ NULL);
+ if (error)
+ goto out;
+ nconf = getnetconfigent(netid);
+ if (!nconf) {
+ printf("Can't lookup netid %s\n",
+ netid);
+ error = EINVAL;
+ goto out;
+ }
+ xprts[j] = svc_tp_create(pool, dispatchers[i],
+ NLM_PROG, versions[i], uaddr, nconf);
+ if (!xprts[j]) {
+ printf("NLM: unable to create "
+ "(NLM_PROG, %d).\n", versions[i]);
+ error = EINVAL;
+ goto out;
+ }
+ freenetconfigent(nconf);
+ } else {
+ nconf = getnetconfigent(xprts[j]->xp_netid);
+ rpcb_unset(NLM_PROG, versions[i], nconf);
+ if (!svc_reg(xprts[j], NLM_PROG, versions[i],
+ dispatchers[i], nconf)) {
+ printf("NLM: can't register "
+ "(NLM_PROG, %d)\n", versions[i]);
+ error = EINVAL;
+ goto out;
+ }
+ }
+ }
+ }
+ error = 0;
+out:
+ free(xprts, M_NLM);
+ return (error);
+}
+
+/*
+ * Main server entry point. Contacts the local NSM to get its current
+ * state and send SM_UNMON_ALL. Registers the NLM services and then
+ * services requests. Does not return until the server is interrupted
+ * by a signal.
+ */
+static int
+nlm_server_main(int addr_count, char **addrs)
+{
+ struct thread *td = curthread;
+ int error;
+ SVCPOOL *pool;
+ struct sockopt opt;
+ int portlow;
+#ifdef INET6
+ struct sockaddr_in6 sin6;
+#endif
+ struct sockaddr_in sin;
+ my_id id;
+ sm_stat smstat;
+ struct timeval timo;
+ enum clnt_stat stat;
+ struct nlm_host *host;
+
+ if (nlm_socket) {
+ printf("NLM: can't start server - it appears to be running already\n");
+ return (EPERM);
+ }
+
+ memset(&opt, 0, sizeof(opt));
+
+ nlm_socket = NULL;
+ error = socreate(AF_INET, &nlm_socket, SOCK_DGRAM, 0,
+ td->td_ucred, td);
+ if (error) {
+ printf("NLM: can't create IPv4 socket - error %d\n", error);
+ return (error);
+ }
+ opt.sopt_dir = SOPT_SET;
+ opt.sopt_level = IPPROTO_IP;
+ opt.sopt_name = IP_PORTRANGE;
+ portlow = IP_PORTRANGE_LOW;
+ opt.sopt_val = &portlow;
+ opt.sopt_valsize = sizeof(portlow);
+ sosetopt(nlm_socket, &opt);
+
+#ifdef INET6
+ nlm_socket6 = NULL;
+ error = socreate(AF_INET6, &nlm_socket6, SOCK_DGRAM, 0,
+ td->td_ucred, td);
+ if (error) {
+ printf("NLM: can't create IPv6 socket - error %d\n", error);
+ return (error);
+ }
+ opt.sopt_dir = SOPT_SET;
+ opt.sopt_level = IPPROTO_IPV6;
+ opt.sopt_name = IPV6_PORTRANGE;
+ portlow = IPV6_PORTRANGE_LOW;
+ opt.sopt_val = &portlow;
+ opt.sopt_valsize = sizeof(portlow);
+ sosetopt(nlm_socket6, &opt);
+#endif
+
+#ifdef INET6
+ memset(&sin6, 0, sizeof(sin6));
+ sin6.sin6_len = sizeof(sin6);
+ sin6.sin6_family = AF_INET6;
+ sin6.sin6_addr = in6addr_loopback;
+ nlm_nsm = nlm_get_rpc((struct sockaddr *) &sin6, SM_PROG, SM_VERS);
+ if (!nlm_nsm) {
+#endif
+ memset(&sin, 0, sizeof(sin));
+ sin.sin_len = sizeof(sin);
+ sin.sin_family = AF_INET6;
+ sin.sin_addr.s_addr = htonl(INADDR_LOOPBACK);
+ nlm_nsm = nlm_get_rpc((struct sockaddr *) &sin, SM_PROG,
+ SM_VERS);
+#ifdef INET6
+ }
+#endif
+
+ if (!nlm_nsm) {
+ printf("Can't start NLM - unable to contact NSM\n");
+ return (EINVAL);
+ }
+
+ pool = svcpool_create();
+
+ error = nlm_register_services(pool, addr_count, addrs);
+ if (error)
+ goto out;
+
+ memset(&id, 0, sizeof(id));
+ id.my_name = "NFS NLM";
+
+ timo.tv_sec = 25;
+ timo.tv_usec = 0;
+ stat = CLNT_CALL(nlm_nsm, SM_UNMON_ALL,
+ (xdrproc_t) xdr_my_id, &id,
+ (xdrproc_t) xdr_sm_stat, &smstat, timo);
+
+ if (stat != RPC_SUCCESS) {
+ struct rpc_err err;
+
+ CLNT_GETERR(nlm_nsm, &err);
+ printf("NLM: unexpected error contacting NSM, stat=%d, errno=%d\n",
+ stat, err.re_errno);
+ error = EINVAL;
+ goto out;
+ }
+
+ if (nlm_debug_level >= 1)
+ printf("NLM: local NSM state is %d\n", smstat.state);
+
+ svc_run(pool);
+ error = 0;
+
+out:
+ if (pool)
+ svcpool_destroy(pool);
+
+ /*
+ * Trash all the existing state so that if the server
+ * restarts, it gets a clean slate.
+ */
+ while ((host = TAILQ_FIRST(&nlm_hosts)) != NULL) {
+ nlm_host_notify(host, 0, TRUE);
+ }
+ if (nlm_nsm) {
+ AUTH_DESTROY(nlm_nsm->cl_auth);
+ CLNT_DESTROY(nlm_nsm);
+ nlm_nsm = NULL;
+ }
+ if (nlm_lockd) {
+ AUTH_DESTROY(nlm_lockd->cl_auth);
+ CLNT_DESTROY(nlm_lockd);
+ nlm_lockd = NULL;
+ }
+
+ soclose(nlm_socket);
+ nlm_socket = NULL;
+#ifdef INET6
+ soclose(nlm_socket6);
+ nlm_socket6 = NULL;
+#endif
+
+ return (error);
+}
+
+int
+nlm_syscall(struct thread *td, struct nlm_syscall_args *uap)
+{
+ int error;
+
+ error = priv_check(td, PRIV_NFS_LOCKD);
+ if (error)
+ return (error);
+
+ nlm_debug_level = uap->debug_level;
+ nlm_grace_threshold = time_uptime + uap->grace_period;
+ nlm_next_idle_check = time_uptime + NLM_IDLE_PERIOD;
+
+ return nlm_server_main(uap->addr_count, uap->addrs);
+}
+
+/**********************************************************************/
+
+/*
+ * NLM implementation details, called from the RPC stubs.
+ */
+
+
+void
+nlm_sm_notify(struct nlm_sm_status *argp)
+{
+ uint32_t sysid;
+ struct nlm_host *host;
+
+ if (nlm_debug_level >= 3)
+ printf("nlm_sm_notify(): mon_name = %s\n", argp->mon_name);
+ memcpy(&sysid, &argp->priv, sizeof(sysid));
+ host = nlm_find_host_by_sysid(sysid);
+ if (host)
+ nlm_host_notify(host, argp->state, FALSE);
+}
+
+static void
+nlm_convert_to_fhandle_t(fhandle_t *fhp, struct netobj *p)
+{
+ memcpy(fhp, p->n_bytes, sizeof(fhandle_t));
+}
+
+struct vfs_state {
+ struct mount *vs_mp;
+ struct vnode *vs_vp;
+ int vs_vfslocked;
+};
+
+static int
+nlm_get_vfs_state(struct nlm_host *host, struct svc_req *rqstp,
+ fhandle_t *fhp, struct vfs_state *vs)
+{
+ int error, exflags, freecred;
+ struct ucred *cred = NULL, *credanon;
+
+ memset(vs, 0, sizeof(*vs));
+ freecred = FALSE;
+
+ vs->vs_mp = vfs_getvfs(&fhp->fh_fsid);
+ if (!vs->vs_mp) {
+ return (ESTALE);
+ }
+ vs->vs_vfslocked = VFS_LOCK_GIANT(vs->vs_mp);
+
+ error = VFS_CHECKEXP(vs->vs_mp, (struct sockaddr *)&host->nh_addr,
+ &exflags, &credanon);
+ if (error)
+ goto out;
+
+ if (exflags & MNT_EXRDONLY || (vs->vs_mp->mnt_flag & MNT_RDONLY)) {
+ error = EROFS;
+ goto out;
+ }
+
+ error = VFS_FHTOVP(vs->vs_mp, &fhp->fh_fid, &vs->vs_vp);
+ if (error)
+ goto out;
+
+ cred = crget();
+ freecred = TRUE;
+ if (!svc_getcred(rqstp, cred, NULL)) {
+ error = EINVAL;
+ goto out;
+ }
+ if (cred->cr_uid == 0 || (exflags & MNT_EXPORTANON)) {
+ crfree(cred);
+ cred = credanon;
+ freecred = FALSE;
+ }
+#if __FreeBSD_version < 800011
+ VOP_UNLOCK(vs->vs_vp, 0, curthread);
+#else
+ VOP_UNLOCK(vs->vs_vp, 0);
+#endif
+
+ /*
+ * Check cred.
+ */
+ error = VOP_ACCESS(vs->vs_vp, VWRITE, cred, curthread);
+ if (error)
+ goto out;
+
+out:
+ if (freecred)
+ crfree(cred);
+
+ return (error);
+}
+
+static void
+nlm_release_vfs_state(struct vfs_state *vs)
+{
+
+ if (vs->vs_vp)
+ vrele(vs->vs_vp);
+ if (vs->vs_mp)
+ vfs_rel(vs->vs_mp);
+ VFS_UNLOCK_GIANT(vs->vs_vfslocked);
+}
+
+static nlm4_stats
+nlm_convert_error(int error)
+{
+
+ if (error == ESTALE)
+ return nlm4_stale_fh;
+ else if (error == EROFS)
+ return nlm4_rofs;
+ else
+ return nlm4_failed;
+}
+
+struct nlm_host *
+nlm_do_test(nlm4_testargs *argp, nlm4_testres *result, struct svc_req *rqstp)
+{
+ fhandle_t fh;
+ struct vfs_state vs;
+ struct nlm_host *host, *bhost;
+ int error, sysid;
+ struct flock fl;
+
+ memset(result, 0, sizeof(*result));
+
+ host = nlm_find_host_by_name(argp->alock.caller_name, rqstp);
+ if (!host) {
+ result->stat.stat = nlm4_denied_nolocks;
+ return (NULL);
+ }
+
+ if (nlm_debug_level >= 3)
+ printf("nlm_do_test(): caller_name = %s (sysid = %d)\n",
+ host->nh_caller_name, host->nh_sysid);
+
+ nlm_free_finished_locks(host);
+ sysid = host->nh_sysid;
+
+ nlm_convert_to_fhandle_t(&fh, &argp->alock.fh);
+ nlm_copy_netobj(&result->cookie, &argp->cookie, M_RPC);
+
+ if (time_uptime < nlm_grace_threshold) {
+ result->stat.stat = nlm4_denied_grace_period;
+ return (host);
+ }
+
+ error = nlm_get_vfs_state(host, rqstp, &fh, &vs);
+ if (error) {
+ result->stat.stat = nlm_convert_error(error);
+ goto out;
+ }
+
+ fl.l_start = argp->alock.l_offset;
+ fl.l_len = argp->alock.l_len;
+ fl.l_pid = argp->alock.svid;
+ fl.l_sysid = sysid;
+ fl.l_whence = SEEK_SET;
+ if (argp->exclusive)
+ fl.l_type = F_WRLCK;
+ else
+ fl.l_type = F_RDLCK;
+ error = VOP_ADVLOCK(vs.vs_vp, NULL, F_GETLK, &fl, F_REMOTE);
+ if (error) {
+ result->stat.stat = nlm4_failed;
+ goto out;
+ }
+
+ if (fl.l_type == F_UNLCK) {
+ result->stat.stat = nlm4_granted;
+ } else {
+ result->stat.stat = nlm4_denied;
+ result->stat.nlm4_testrply_u.holder.exclusive =
+ (fl.l_type == F_WRLCK);
+ result->stat.nlm4_testrply_u.holder.svid = fl.l_pid;
+ bhost = nlm_find_host_by_sysid(fl.l_sysid);
+ if (bhost) {
+ /*
+ * We don't have any useful way of recording
+ * the value of oh used in the original lock
+ * request. Ideally, the test reply would have
+ * a space for the owning host's name allowing
+ * our caller's NLM to keep track.
+ *
+ * As far as I can see, Solaris uses an eight
+ * byte structure for oh which contains a four
+ * byte pid encoded in local byte order and
+ * the first four bytes of the host
+ * name. Linux uses a variable length string
+ * 'pid@hostname' in ascii but doesn't even
+ * return that in test replies.
+ *
+ * For the moment, return nothing in oh
+ * (already zero'ed above).
+ */
+ }
+ result->stat.nlm4_testrply_u.holder.l_offset = fl.l_start;
+ result->stat.nlm4_testrply_u.holder.l_len = fl.l_len;
+ }
+
+out:
+ nlm_release_vfs_state(&vs);
+ return (host);
+}
+
+struct nlm_host *
+nlm_do_lock(nlm4_lockargs *argp, nlm4_res *result, struct svc_req *rqstp,
+ bool_t monitor)
+{
+ fhandle_t fh;
+ struct vfs_state vs;
+ struct nlm_host *host;
+ int error, sysid;
+ struct flock fl;
+
+ memset(result, 0, sizeof(*result));
+
+ host = nlm_find_host_by_name(argp->alock.caller_name, rqstp);
+ if (!host) {
+ result->stat.stat = nlm4_denied_nolocks;
+ return (NULL);
+ }
+
+ if (nlm_debug_level >= 3)
+ printf("nlm_do_lock(): caller_name = %s (sysid = %d)\n",
+ host->nh_caller_name, host->nh_sysid);
+
+ nlm_free_finished_locks(host);
+ sysid = host->nh_sysid;
+
+ nlm_convert_to_fhandle_t(&fh, &argp->alock.fh);
+ nlm_copy_netobj(&result->cookie, &argp->cookie, M_RPC);
+
+ if (time_uptime < nlm_grace_threshold && !argp->reclaim) {
+ result->stat.stat = nlm4_denied_grace_period;
+ return (host);
+ }
+
+ error = nlm_get_vfs_state(host, rqstp, &fh, &vs);
+ if (error) {
+ result->stat.stat = nlm_convert_error(error);
+ goto out;
+ }
+
+ fl.l_start = argp->alock.l_offset;
+ fl.l_len = argp->alock.l_len;
+ fl.l_pid = argp->alock.svid;
+ fl.l_sysid = sysid;
+ fl.l_whence = SEEK_SET;
+ if (argp->exclusive)
+ fl.l_type = F_WRLCK;
+ else
+ fl.l_type = F_RDLCK;
+ if (argp->block) {
+ struct nlm_async_lock *af;
+
+ /*
+ * First, make sure we can contact the host's NLM.
+ */
+ if (!nlm_host_get_rpc(host)) {
+ result->stat.stat = nlm4_failed;
+ goto out;
+ }
+
+ /*
+ * First we need to check and see if there is an
+ * existing blocked lock that matches. This could be a
+ * badly behaved client or an RPC re-send. If we find
+ * one, just return nlm4_blocked.
+ */
+ mtx_lock(&host->nh_lock);
+ TAILQ_FOREACH(af, &host->nh_pending, af_link) {
+ if (af->af_fl.l_start == fl.l_start
+ && af->af_fl.l_len == fl.l_len
+ && af->af_fl.l_pid == fl.l_pid
+ && af->af_fl.l_type == fl.l_type) {
+ break;
+ }
+ }
+ mtx_unlock(&host->nh_lock);
+ if (af) {
+ result->stat.stat = nlm4_blocked;
+ goto out;
+ }
+
+ af = malloc(sizeof(struct nlm_async_lock), M_NLM,
+ M_WAITOK|M_ZERO);
+ TASK_INIT(&af->af_task, 0, nlm_lock_callback, af);
+ af->af_vp = vs.vs_vp;
+ af->af_fl = fl;
+ af->af_host = host;
+ /*
+ * We use M_RPC here so that we can xdr_free the thing
+ * later.
+ */
+ af->af_granted.exclusive = argp->exclusive;
+ af->af_granted.alock.caller_name =
+ strdup(argp->alock.caller_name, M_RPC);
+ nlm_copy_netobj(&af->af_granted.alock.fh,
+ &argp->alock.fh, M_RPC);
+ nlm_copy_netobj(&af->af_granted.alock.oh,
+ &argp->alock.oh, M_RPC);
+ af->af_granted.alock.svid = argp->alock.svid;
+ af->af_granted.alock.l_offset = argp->alock.l_offset;
+ af->af_granted.alock.l_len = argp->alock.l_len;
+
+ /*
+ * Put the entry on the pending list before calling
+ * VOP_ADVLOCKASYNC. We do this in case the lock
+ * request was blocked (returning EINPROGRESS) but
+ * then granted before we manage to run again. The
+ * client may receive the granted message before we
+ * send our blocked reply but thats their problem.
+ */
+ mtx_lock(&host->nh_lock);
+ TAILQ_INSERT_TAIL(&host->nh_pending, af, af_link);
+ mtx_unlock(&host->nh_lock);
+
+ error = VOP_ADVLOCKASYNC(vs.vs_vp, NULL, F_SETLK, &fl, F_REMOTE,
+ &af->af_task, &af->af_cookie);
+
+ /*
+ * If the lock completed synchronously, just free the
+ * tracking structure now.
+ */
+ if (error != EINPROGRESS) {
+ mtx_lock(&host->nh_lock);
+ TAILQ_REMOVE(&host->nh_pending, af, af_link);
+ mtx_unlock(&host->nh_lock);
+ xdr_free((xdrproc_t) xdr_nlm4_testargs,
+ &af->af_granted);
+ free(af, M_NLM);
+ } else {
+ if (nlm_debug_level >= 2)
+ printf("NLM: pending async lock %p for %s "
+ "(sysid %d)\n",
+ af, host->nh_caller_name, sysid);
+ /*
+ * Don't vrele the vnode just yet - this must
+ * wait until either the async callback
+ * happens or the lock is cancelled.
+ */
+ vs.vs_vp = NULL;
+ }
+ } else {
+ error = VOP_ADVLOCK(vs.vs_vp, NULL, F_SETLK, &fl, F_REMOTE);
+ }
+
+ if (error) {
+ if (error == EINPROGRESS) {
+ result->stat.stat = nlm4_blocked;
+ } else if (error == EDEADLK) {
+ result->stat.stat = nlm4_deadlck;
+ } else if (error == EAGAIN) {
+ result->stat.stat = nlm4_denied;
+ } else {
+ result->stat.stat = nlm4_failed;
+ }
+ } else {
+ if (monitor)
+ nlm_host_monitor(host, argp->state);
+ result->stat.stat = nlm4_granted;
+ }
+
+out:
+ nlm_release_vfs_state(&vs);
+
+ return (host);
+}
+
+struct nlm_host *
+nlm_do_cancel(nlm4_cancargs *argp, nlm4_res *result, struct svc_req *rqstp)
+{
+ fhandle_t fh;
+ struct vfs_state vs;
+ struct nlm_host *host;
+ int error, sysid;
+ struct flock fl;
+ struct nlm_async_lock *af;
+
+ memset(result, 0, sizeof(*result));
+
+ host = nlm_find_host_by_name(argp->alock.caller_name, rqstp);
+ if (!host) {
+ result->stat.stat = nlm4_denied_nolocks;
+ return (NULL);
+ }
+
+ if (nlm_debug_level >= 3)
+ printf("nlm_do_cancel(): caller_name = %s (sysid = %d)\n",
+ host->nh_caller_name, host->nh_sysid);
+
+ nlm_free_finished_locks(host);
+ sysid = host->nh_sysid;
+
+ nlm_convert_to_fhandle_t(&fh, &argp->alock.fh);
+ nlm_copy_netobj(&result->cookie, &argp->cookie, M_RPC);
+
+ if (time_uptime < nlm_grace_threshold) {
+ result->stat.stat = nlm4_denied_grace_period;
+ return (host);
+ }
+
+ error = nlm_get_vfs_state(host, rqstp, &fh, &vs);
+ if (error) {
+ result->stat.stat = nlm_convert_error(error);
+ goto out;
+ }
+
+ fl.l_start = argp->alock.l_offset;
+ fl.l_len = argp->alock.l_len;
+ fl.l_pid = argp->alock.svid;
+ fl.l_sysid = sysid;
+ fl.l_whence = SEEK_SET;
+ if (argp->exclusive)
+ fl.l_type = F_WRLCK;
+ else
+ fl.l_type = F_RDLCK;
+
+ /*
+ * First we need to try and find the async lock request - if
+ * there isn't one, we give up and return nlm4_denied.
+ */
+ mtx_lock(&host->nh_lock);
+
+ TAILQ_FOREACH(af, &host->nh_pending, af_link) {
+ if (af->af_fl.l_start == fl.l_start
+ && af->af_fl.l_len == fl.l_len
+ && af->af_fl.l_pid == fl.l_pid
+ && af->af_fl.l_type == fl.l_type) {
+ break;
+ }
+ }
+
+ if (!af) {
+ mtx_unlock(&host->nh_lock);
+ result->stat.stat = nlm4_denied;
+ goto out;
+ }
+
+ error = nlm_cancel_async_lock(af);
+
+ if (error) {
+ result->stat.stat = nlm4_denied;
+ } else {
+ result->stat.stat = nlm4_granted;
+ }
+
+ mtx_unlock(&host->nh_lock);
+
+out:
+ nlm_release_vfs_state(&vs);
+
+ return (host);
+}
+
+struct nlm_host *
+nlm_do_unlock(nlm4_unlockargs *argp, nlm4_res *result, struct svc_req *rqstp)
+{
+ fhandle_t fh;
+ struct vfs_state vs;
+ struct nlm_host *host;
+ int error, sysid;
+ struct flock fl;
+
+ memset(result, 0, sizeof(*result));
+
+ host = nlm_find_host_by_name(argp->alock.caller_name, rqstp);
+ if (!host) {
+ result->stat.stat = nlm4_denied_nolocks;
+ return (NULL);
+ }
+
+ if (nlm_debug_level >= 3)
+ printf("nlm_do_unlock(): caller_name = %s (sysid = %d)\n",
+ host->nh_caller_name, host->nh_sysid);
+
+ nlm_free_finished_locks(host);
+ sysid = host->nh_sysid;
+
+ nlm_convert_to_fhandle_t(&fh, &argp->alock.fh);
+ nlm_copy_netobj(&result->cookie, &argp->cookie, M_RPC);
+
+ if (time_uptime < nlm_grace_threshold) {
+ result->stat.stat = nlm4_denied_grace_period;
+ return (host);
+ }
+
+ error = nlm_get_vfs_state(host, rqstp, &fh, &vs);
+ if (error) {
+ result->stat.stat = nlm_convert_error(error);
+ goto out;
+ }
+
+ fl.l_start = argp->alock.l_offset;
+ fl.l_len = argp->alock.l_len;
+ fl.l_pid = argp->alock.svid;
+ fl.l_sysid = sysid;
+ fl.l_whence = SEEK_SET;
+ fl.l_type = F_UNLCK;
+ error = VOP_ADVLOCK(vs.vs_vp, NULL, F_UNLCK, &fl, F_REMOTE);
+
+ /*
+ * Ignore the error - there is no result code for failure,
+ * only for grace period.
+ */
+ result->stat.stat = nlm4_granted;
+
+out:
+ nlm_release_vfs_state(&vs);
+
+ return (host);
+}
+
+void
+nlm_do_free_all(nlm4_notify *argp)
+{
+ struct nlm_host *host, *thost;
+
+ TAILQ_FOREACH_SAFE(host, &nlm_hosts, nh_link, thost) {
+ if (!strcmp(host->nh_caller_name, argp->name))
+ nlm_host_notify(host, argp->state, FALSE);
+ }
+}
+
+#define _PATH_RPCLOCKDSOCK "/var/run/rpclockd.sock"
+
+/*
+ * Make a connection to the userland lockd - we push anything we can't
+ * handle out to userland.
+ */
+CLIENT *
+nlm_user_lockd(void)
+{
+ struct sockaddr_un sun;
+ struct netconfig *nconf;
+ struct timeval zero;
+
+ if (nlm_lockd)
+ return (nlm_lockd);
+
+ sun.sun_family = AF_LOCAL;
+ strcpy(sun.sun_path, _PATH_RPCLOCKDSOCK);
+ sun.sun_len = SUN_LEN(&sun);
+
+ nconf = getnetconfigent("local");
+ nlm_lockd = clnt_reconnect_create(nconf, (struct sockaddr *) &sun,
+ NLM_PROG, NLM_VERS4, RPC_MAXDATASIZE, RPC_MAXDATASIZE);
+
+ /*
+ * Set the send timeout to zero - we only use this rpc handle
+ * for sending async replies which have no return value.
+ */
+ zero.tv_sec = 0;
+ zero.tv_usec = 0;
+ CLNT_CONTROL(nlm_lockd, CLSET_TIMEOUT, &zero);
+
+ return (nlm_lockd);
+}
OpenPOWER on IntegriCloud