]> git.saurik.com Git - apple/xnu.git/blame - bsd/nfs/nfs_syscalls.c
xnu-3789.1.32.tar.gz
[apple/xnu.git] / bsd / nfs / nfs_syscalls.c
CommitLineData
1c79356b 1/*
fe8ab488 2 * Copyright (c) 2000-2014 Apple Inc. All rights reserved.
5d5c5d0d 3 *
2d21ac55 4 * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
1c79356b 5 *
2d21ac55
A
6 * This file contains Original Code and/or Modifications of Original Code
7 * as defined in and that are subject to the Apple Public Source License
8 * Version 2.0 (the 'License'). You may not use this file except in
9 * compliance with the License. The rights granted to you under the License
10 * may not be used to create, or enable the creation or redistribution of,
11 * unlawful or unlicensed copies of an Apple operating system, or to
12 * circumvent, violate, or enable the circumvention or violation of, any
13 * terms of an Apple operating system software license agreement.
8f6c56a5 14 *
2d21ac55
A
15 * Please obtain a copy of the License at
16 * http://www.opensource.apple.com/apsl/ and read it before using this file.
17 *
18 * The Original Code and all software distributed under the License are
19 * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
8f6c56a5
A
20 * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
21 * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
2d21ac55
A
22 * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
23 * Please see the License for the specific language governing rights and
24 * limitations under the License.
8f6c56a5 25 *
2d21ac55 26 * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
1c79356b
A
27 */
28/* Copyright (c) 1995 NeXT Computer, Inc. All Rights Reserved */
29/*
30 * Copyright (c) 1989, 1993
31 * The Regents of the University of California. All rights reserved.
32 *
33 * This code is derived from software contributed to Berkeley by
34 * Rick Macklem at The University of Guelph.
35 *
36 * Redistribution and use in source and binary forms, with or without
37 * modification, are permitted provided that the following conditions
38 * are met:
39 * 1. Redistributions of source code must retain the above copyright
40 * notice, this list of conditions and the following disclaimer.
41 * 2. Redistributions in binary form must reproduce the above copyright
42 * notice, this list of conditions and the following disclaimer in the
43 * documentation and/or other materials provided with the distribution.
44 * 3. All advertising materials mentioning features or use of this software
45 * must display the following acknowledgement:
46 * This product includes software developed by the University of
47 * California, Berkeley and its contributors.
48 * 4. Neither the name of the University nor the names of its contributors
49 * may be used to endorse or promote products derived from this software
50 * without specific prior written permission.
51 *
52 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
53 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
54 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
55 * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
56 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
57 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
58 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
59 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
60 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
61 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
62 * SUCH DAMAGE.
63 *
64 * @(#)nfs_syscalls.c 8.5 (Berkeley) 3/30/95
65 * FreeBSD-Id: nfs_syscalls.c,v 1.32 1997/11/07 08:53:25 phk Exp $
66 */
2d21ac55
A
67/*
68 * NOTICE: This file was modified by SPARTA, Inc. in 2005 to introduce
69 * support for mandatory and extensible security protections. This notice
70 * is included in support of clause 2.2 (b) of the Apple Public License,
71 * Version 2.0.
72 */
1c79356b
A
73
74#include <sys/param.h>
75#include <sys/systm.h>
1c79356b 76#include <sys/kernel.h>
91447636 77#include <sys/file_internal.h>
1c79356b
A
78#include <sys/filedesc.h>
79#include <sys/stat.h>
91447636
A
80#include <sys/vnode_internal.h>
81#include <sys/mount_internal.h>
82#include <sys/proc_internal.h> /* for fdflags */
83#include <sys/kauth.h>
1c79356b 84#include <sys/sysctl.h>
55e303ae 85#include <sys/ubc.h>
1c79356b
A
86#include <sys/uio.h>
87#include <sys/malloc.h>
91447636 88#include <sys/kpi_mbuf.h>
1c79356b
A
89#include <sys/socket.h>
90#include <sys/socketvar.h>
91#include <sys/domain.h>
92#include <sys/protosw.h>
55e303ae
A
93#include <sys/fcntl.h>
94#include <sys/lockf.h>
1c79356b
A
95#include <sys/syslog.h>
96#include <sys/user.h>
91447636
A
97#include <sys/sysproto.h>
98#include <sys/kpi_socket.h>
2d21ac55 99#include <sys/fsevents.h>
91447636 100#include <libkern/OSAtomic.h>
2d21ac55
A
101#include <kern/thread_call.h>
102#include <kern/task.h>
1c79356b 103
b0d623f7 104#include <security/audit/audit.h>
ccc36f2f 105
1c79356b
A
106#include <netinet/in.h>
107#include <netinet/tcp.h>
1c79356b
A
108#include <nfs/xdr_subs.h>
109#include <nfs/rpcv2.h>
110#include <nfs/nfsproto.h>
111#include <nfs/nfs.h>
112#include <nfs/nfsm_subs.h>
113#include <nfs/nfsrvcache.h>
2d21ac55 114#include <nfs/nfs_gss.h>
1c79356b
A
115#include <nfs/nfsmount.h>
116#include <nfs/nfsnode.h>
55e303ae 117#include <nfs/nfs_lock.h>
2d21ac55
A
118#if CONFIG_MACF
119#include <security/mac_framework.h>
1c79356b
A
120#endif
121
2d21ac55
A
122kern_return_t thread_terminate(thread_t); /* XXX */
123
124#if NFSSERVER
125
126extern int (*nfsrv_procs[NFS_NPROCS])(struct nfsrv_descript *nd,
127 struct nfsrv_sock *slp,
128 vfs_context_t ctx,
129 mbuf_t *mrepp);
130extern int nfsrv_wg_delay;
131extern int nfsrv_wg_delay_v3;
132
133static int nfsrv_require_resv_port = 0;
3e170ce0
A
134static time_t nfsrv_idlesock_timer_on = 0;
135static int nfsrv_sock_tcp_cnt = 0;
136#define NFSD_MIN_IDLE_TIMEOUT 30
137static int nfsrv_sock_idle_timeout = 3600; /* One hour */
2d21ac55 138
b0d623f7
A
139int nfssvc_export(user_addr_t argp);
140int nfssvc_nfsd(void);
141int nfssvc_addsock(socket_t, mbuf_t);
142void nfsrv_zapsock(struct nfsrv_sock *);
143void nfsrv_slpderef(struct nfsrv_sock *);
144void nfsrv_slpfree(struct nfsrv_sock *);
2d21ac55
A
145
146#endif /* NFSSERVER */
147
148/*
149 * sysctl stuff
150 */
151SYSCTL_DECL(_vfs_generic);
152SYSCTL_NODE(_vfs_generic, OID_AUTO, nfs, CTLFLAG_RW|CTLFLAG_LOCKED, 0, "nfs hinge");
153
154#if NFSCLIENT
155SYSCTL_NODE(_vfs_generic_nfs, OID_AUTO, client, CTLFLAG_RW|CTLFLAG_LOCKED, 0, "nfs client hinge");
6d2010ae
A
156SYSCTL_INT(_vfs_generic_nfs_client, OID_AUTO, initialdowndelay, CTLFLAG_RW | CTLFLAG_LOCKED, &nfs_tprintf_initial_delay, 0, "");
157SYSCTL_INT(_vfs_generic_nfs_client, OID_AUTO, nextdowndelay, CTLFLAG_RW | CTLFLAG_LOCKED, &nfs_tprintf_delay, 0, "");
158SYSCTL_INT(_vfs_generic_nfs_client, OID_AUTO, iosize, CTLFLAG_RW | CTLFLAG_LOCKED, &nfs_iosize, 0, "");
159SYSCTL_INT(_vfs_generic_nfs_client, OID_AUTO, access_cache_timeout, CTLFLAG_RW | CTLFLAG_LOCKED, &nfs_access_cache_timeout, 0, "");
160SYSCTL_INT(_vfs_generic_nfs_client, OID_AUTO, allow_async, CTLFLAG_RW | CTLFLAG_LOCKED, &nfs_allow_async, 0, "");
161SYSCTL_INT(_vfs_generic_nfs_client, OID_AUTO, statfs_rate_limit, CTLFLAG_RW | CTLFLAG_LOCKED, &nfs_statfs_rate_limit, 0, "");
162SYSCTL_INT(_vfs_generic_nfs_client, OID_AUTO, nfsiod_thread_max, CTLFLAG_RW | CTLFLAG_LOCKED, &nfsiod_thread_max, 0, "");
163SYSCTL_INT(_vfs_generic_nfs_client, OID_AUTO, nfsiod_thread_count, CTLFLAG_RD | CTLFLAG_LOCKED, &nfsiod_thread_count, 0, "");
164SYSCTL_INT(_vfs_generic_nfs_client, OID_AUTO, lockd_mounts, CTLFLAG_RD | CTLFLAG_LOCKED, &nfs_lockd_mounts, 0, "");
165SYSCTL_INT(_vfs_generic_nfs_client, OID_AUTO, max_async_writes, CTLFLAG_RW | CTLFLAG_LOCKED, &nfs_max_async_writes, 0, "");
6d2010ae
A
166SYSCTL_INT(_vfs_generic_nfs_client, OID_AUTO, access_delete, CTLFLAG_RW | CTLFLAG_LOCKED, &nfs_access_delete, 0, "");
167SYSCTL_INT(_vfs_generic_nfs_client, OID_AUTO, access_dotzfs, CTLFLAG_RW | CTLFLAG_LOCKED, &nfs_access_dotzfs, 0, "");
168SYSCTL_INT(_vfs_generic_nfs_client, OID_AUTO, access_for_getattr, CTLFLAG_RW | CTLFLAG_LOCKED, &nfs_access_for_getattr, 0, "");
169SYSCTL_INT(_vfs_generic_nfs_client, OID_AUTO, idmap_ctrl, CTLFLAG_RW | CTLFLAG_LOCKED, &nfs_idmap_ctrl, 0, "");
170SYSCTL_INT(_vfs_generic_nfs_client, OID_AUTO, callback_port, CTLFLAG_RW | CTLFLAG_LOCKED, &nfs_callback_port, 0, "");
316670eb
A
171SYSCTL_INT(_vfs_generic_nfs_client, OID_AUTO, is_mobile, CTLFLAG_RW | CTLFLAG_LOCKED, &nfs_is_mobile, 0, "");
172SYSCTL_INT(_vfs_generic_nfs_client, OID_AUTO, squishy_flags, CTLFLAG_RW | CTLFLAG_LOCKED, &nfs_squishy_flags, 0, "");
39236c6e 173SYSCTL_UINT(_vfs_generic_nfs_client, OID_AUTO, debug_ctl, CTLFLAG_RW | CTLFLAG_LOCKED, &nfs_debug_ctl, 0, "");
fe8ab488 174SYSCTL_INT(_vfs_generic_nfs_client, OID_AUTO, readlink_nocache, CTLFLAG_RW | CTLFLAG_LOCKED, &nfs_readlink_nocache, 0, "");
3e170ce0 175SYSCTL_INT(_vfs_generic_nfs_client, OID_AUTO, root_steals_gss_context, CTLFLAG_RW | CTLFLAG_LOCKED, &nfs_root_steals_ctx, 0, "");
39037602 176SYSCTL_STRING(_vfs_generic_nfs_client, OID_AUTO, default_nfs4domain, CTLFLAG_RW | CTLFLAG_LOCKED, nfs4_domain, sizeof(nfs4_domain), "");
2d21ac55
A
177#endif /* NFSCLIENT */
178
179#if NFSSERVER
180SYSCTL_NODE(_vfs_generic_nfs, OID_AUTO, server, CTLFLAG_RW|CTLFLAG_LOCKED, 0, "nfs server hinge");
6d2010ae
A
181SYSCTL_INT(_vfs_generic_nfs_server, OID_AUTO, wg_delay, CTLFLAG_RW | CTLFLAG_LOCKED, &nfsrv_wg_delay, 0, "");
182SYSCTL_INT(_vfs_generic_nfs_server, OID_AUTO, wg_delay_v3, CTLFLAG_RW | CTLFLAG_LOCKED, &nfsrv_wg_delay_v3, 0, "");
183SYSCTL_INT(_vfs_generic_nfs_server, OID_AUTO, require_resv_port, CTLFLAG_RW | CTLFLAG_LOCKED, &nfsrv_require_resv_port, 0, "");
184SYSCTL_INT(_vfs_generic_nfs_server, OID_AUTO, async, CTLFLAG_RW | CTLFLAG_LOCKED, &nfsrv_async, 0, "");
185SYSCTL_INT(_vfs_generic_nfs_server, OID_AUTO, export_hash_size, CTLFLAG_RW | CTLFLAG_LOCKED, &nfsrv_export_hash_size, 0, "");
186SYSCTL_INT(_vfs_generic_nfs_server, OID_AUTO, reqcache_size, CTLFLAG_RW | CTLFLAG_LOCKED, &nfsrv_reqcache_size, 0, "");
187SYSCTL_INT(_vfs_generic_nfs_server, OID_AUTO, request_queue_length, CTLFLAG_RW | CTLFLAG_LOCKED, &nfsrv_sock_max_rec_queue_length, 0, "");
188SYSCTL_INT(_vfs_generic_nfs_server, OID_AUTO, user_stats, CTLFLAG_RW | CTLFLAG_LOCKED, &nfsrv_user_stat_enabled, 0, "");
189SYSCTL_UINT(_vfs_generic_nfs_server, OID_AUTO, gss_context_ttl, CTLFLAG_RW | CTLFLAG_LOCKED, &nfsrv_gss_context_ttl, 0, "");
b0d623f7 190#if CONFIG_FSE
6d2010ae 191SYSCTL_INT(_vfs_generic_nfs_server, OID_AUTO, fsevents, CTLFLAG_RW | CTLFLAG_LOCKED, &nfsrv_fsevents_enabled, 0, "");
b0d623f7 192#endif
6d2010ae
A
193SYSCTL_INT(_vfs_generic_nfs_server, OID_AUTO, nfsd_thread_max, CTLFLAG_RW | CTLFLAG_LOCKED, &nfsd_thread_max, 0, "");
194SYSCTL_INT(_vfs_generic_nfs_server, OID_AUTO, nfsd_thread_count, CTLFLAG_RD | CTLFLAG_LOCKED, &nfsd_thread_count, 0, "");
3e170ce0
A
195SYSCTL_INT(_vfs_generic_nfs_server, OID_AUTO, nfsd_sock_idle_timeout, CTLFLAG_RW | CTLFLAG_LOCKED, &nfsrv_sock_idle_timeout, 0, "");
196SYSCTL_INT(_vfs_generic_nfs_server, OID_AUTO, nfsd_tcp_connections, CTLFLAG_RD | CTLFLAG_LOCKED, &nfsrv_sock_tcp_cnt, 0, "");
316670eb
A
197#ifdef NFS_UC_Q_DEBUG
198SYSCTL_INT(_vfs_generic_nfs_server, OID_AUTO, use_upcall_svc, CTLFLAG_RW | CTLFLAG_LOCKED, &nfsrv_uc_use_proxy, 0, "");
199SYSCTL_INT(_vfs_generic_nfs_server, OID_AUTO, upcall_queue_limit, CTLFLAG_RW | CTLFLAG_LOCKED, &nfsrv_uc_queue_limit, 0, "");
200SYSCTL_INT(_vfs_generic_nfs_server, OID_AUTO, upcall_queue_max_seen, CTLFLAG_RW | CTLFLAG_LOCKED, &nfsrv_uc_queue_max_seen, 0, "");
3e170ce0 201SYSCTL_INT(_vfs_generic_nfs_server, OID_AUTO, upcall_queue_count, CTLFLAG_RD | CTLFLAG_LOCKED, __DECONST(int *, &nfsrv_uc_queue_count), 0, "");
316670eb 202#endif
2d21ac55
A
203#endif /* NFSSERVER */
204
205
206#if NFSCLIENT
207
fe8ab488
A
208static int
209mapname2id(struct nfs_testmapid *map)
210{
211 int error;
212
213 error = nfs4_id2guid(map->ntm_name, &map->ntm_guid, map->ntm_grpflag);
214 if (error)
215 return (error);
216
217 if (map->ntm_grpflag)
218 error = kauth_cred_guid2gid(&map->ntm_guid, (gid_t *)&map->ntm_id);
219 else
220 error = kauth_cred_guid2uid(&map->ntm_guid, (uid_t *)&map->ntm_id);
221
222 return (error);
223}
224
225static int
226mapid2name(struct nfs_testmapid *map)
227{
228 int error;
229 int len = sizeof(map->ntm_name);
230
231 if (map->ntm_grpflag)
232 error = kauth_cred_gid2guid((gid_t)map->ntm_id, &map->ntm_guid);
233 else
234 error = kauth_cred_uid2guid((uid_t)map->ntm_id, &map->ntm_guid);
235
236 if (error)
237 return (error);
238
239 error = nfs4_guid2id(&map->ntm_guid, map->ntm_name, &len, map->ntm_grpflag);
240
241 return (error);
242
243}
244
245
246static int
247nfsclnt_testidmap(proc_t p, user_addr_t argp)
248{
249 struct nfs_testmapid mapid;
250 int error, coerror;
251
252 /* Let root make this call. */
253 error = proc_suser(p);
254 if (error)
255 return (error);
256
257 error = copyin(argp, &mapid, sizeof(mapid));
258 if (error)
259 return (error);
260 if (mapid.ntm_name2id)
261 error = mapname2id(&mapid);
262 else
263 error = mapid2name(&mapid);
264
265 coerror = copyout(&mapid, argp, sizeof(mapid));
266
267 return (error ? error : coerror);
268}
269
2d21ac55
A
270int
271nfsclnt(proc_t p, struct nfsclnt_args *uap, __unused int *retval)
272{
273 struct lockd_ans la;
274 int error;
275
6d2010ae
A
276 switch (uap->flag) {
277 case NFSCLNT_LOCKDANS:
2d21ac55 278 error = copyin(uap->argp, &la, sizeof(la));
6d2010ae
A
279 if (!error)
280 error = nfslockdans(p, &la);
281 break;
282 case NFSCLNT_LOCKDNOTIFY:
283 error = nfslockdnotify(p, uap->argp);
284 break;
fe8ab488
A
285 case NFSCLNT_TESTIDMAP:
286 error = nfsclnt_testidmap(p, uap->argp);
287 break;
6d2010ae
A
288 default:
289 error = EINVAL;
2d21ac55 290 }
6d2010ae 291 return (error);
2d21ac55
A
292}
293
fe8ab488 294
2d21ac55
A
295/*
296 * Asynchronous I/O threads for client NFS.
297 * They do read-ahead and write-behind operations on the block I/O cache.
298 *
299 * The pool of up to nfsiod_thread_max threads is launched on demand and exit
300 * when unused for a while. There are as many nfsiod structs as there are
301 * nfsiod threads; however there's no strict tie between a thread and a struct.
302 * Each thread puts an nfsiod on the free list and sleeps on it. When it wakes
303 * up, it removes the next struct nfsiod from the queue and services it. Then
304 * it will put the struct at the head of free list and sleep on it.
305 * Async requests will pull the next struct nfsiod from the head of the free list,
306 * put it on the work queue, and wake whatever thread is waiting on that struct.
307 */
2d21ac55
A
308
309/*
310 * nfsiod thread exit routine
311 *
312 * Must be called with nfsiod_mutex held so that the
313 * decision to terminate is atomic with the termination.
314 */
b0d623f7 315void
2d21ac55
A
316nfsiod_terminate(struct nfsiod *niod)
317{
318 nfsiod_thread_count--;
319 lck_mtx_unlock(nfsiod_mutex);
320 if (niod)
321 FREE(niod, M_TEMP);
322 else
323 printf("nfsiod: terminating without niod\n");
324 thread_terminate(current_thread());
325 /*NOTREACHED*/
326}
327
328/* nfsiod thread startup routine */
b0d623f7 329void
2d21ac55
A
330nfsiod_thread(void)
331{
332 struct nfsiod *niod;
333 int error;
334
335 MALLOC(niod, struct nfsiod *, sizeof(struct nfsiod), M_TEMP, M_WAITOK);
336 if (!niod) {
337 lck_mtx_lock(nfsiod_mutex);
338 nfsiod_thread_count--;
b0d623f7 339 wakeup(current_thread());
2d21ac55
A
340 lck_mtx_unlock(nfsiod_mutex);
341 thread_terminate(current_thread());
342 /*NOTREACHED*/
343 }
344 bzero(niod, sizeof(*niod));
345 lck_mtx_lock(nfsiod_mutex);
346 TAILQ_INSERT_HEAD(&nfsiodfree, niod, niod_link);
347 wakeup(current_thread());
348 error = msleep0(niod, nfsiod_mutex, PWAIT | PDROP, "nfsiod", NFS_ASYNCTHREADMAXIDLE*hz, nfsiod_continue);
349 /* shouldn't return... so we have an error */
350 /* remove an old nfsiod struct and terminate */
351 lck_mtx_lock(nfsiod_mutex);
352 if ((niod = TAILQ_LAST(&nfsiodfree, nfsiodlist)))
353 TAILQ_REMOVE(&nfsiodfree, niod, niod_link);
354 nfsiod_terminate(niod);
355 /*NOTREACHED*/
356}
357
358/*
359 * Start up another nfsiod thread.
360 * (unless we're already maxed out and there are nfsiods running)
361 */
362int
363nfsiod_start(void)
364{
b0d623f7 365 thread_t thd = THREAD_NULL;
2d21ac55
A
366
367 lck_mtx_lock(nfsiod_mutex);
368 if ((nfsiod_thread_count >= NFSIOD_MAX) && (nfsiod_thread_count > 0)) {
369 lck_mtx_unlock(nfsiod_mutex);
370 return (EBUSY);
371 }
372 nfsiod_thread_count++;
b0d623f7
A
373 if (kernel_thread_start((thread_continue_t)nfsiod_thread, NULL, &thd) != KERN_SUCCESS) {
374 lck_mtx_unlock(nfsiod_mutex);
375 return (EBUSY);
376 }
2d21ac55
A
377 /* wait for the thread to complete startup */
378 msleep(thd, nfsiod_mutex, PWAIT | PDROP, "nfsiodw", NULL);
b0d623f7 379 thread_deallocate(thd);
2d21ac55
A
380 return (0);
381}
382
383/*
384 * Continuation for Asynchronous I/O threads for NFS client.
385 *
386 * Grab an nfsiod struct to work on, do some work, then drop it
387 */
b0d623f7 388int
2d21ac55
A
389nfsiod_continue(int error)
390{
391 struct nfsiod *niod;
392 struct nfsmount *nmp;
393 struct nfsreq *req, *treq;
394 struct nfs_reqqhead iodq;
395 int morework;
396
397 lck_mtx_lock(nfsiod_mutex);
398 niod = TAILQ_FIRST(&nfsiodwork);
399 if (!niod) {
400 /* there's no work queued up */
2d21ac55
A
401 /* remove an old nfsiod struct and terminate */
402 if ((niod = TAILQ_LAST(&nfsiodfree, nfsiodlist)))
403 TAILQ_REMOVE(&nfsiodfree, niod, niod_link);
404 nfsiod_terminate(niod);
405 /*NOTREACHED*/
406 }
407 TAILQ_REMOVE(&nfsiodwork, niod, niod_link);
408
409worktodo:
410 while ((nmp = niod->niod_nmp)) {
fe8ab488
A
411 if (nmp == NULL){
412 niod->niod_nmp = NULL;
413 break;
414 }
415
2d21ac55
A
416 /*
417 * Service this mount's async I/O queue.
418 *
419 * In order to ensure some level of fairness between mounts,
420 * we grab all the work up front before processing it so any
421 * new work that arrives will be serviced on a subsequent
422 * iteration - and we have a chance to see if other work needs
423 * to be done (e.g. the delayed write queue needs to be pushed
424 * or other mounts are waiting for an nfsiod).
425 */
426 /* grab the current contents of the queue */
427 TAILQ_INIT(&iodq);
428 TAILQ_CONCAT(&iodq, &nmp->nm_iodq, r_achain);
3e170ce0
A
429 /* Mark each iod request as being managed by an iod */
430 TAILQ_FOREACH(req, &iodq, r_achain) {
431 lck_mtx_lock(&req->r_mtx);
432 assert(!(req->r_flags & R_IOD));
433 req->r_flags |= R_IOD;
434 lck_mtx_unlock(&req->r_mtx);
435 }
2d21ac55
A
436 lck_mtx_unlock(nfsiod_mutex);
437
438 /* process the queue */
439 TAILQ_FOREACH_SAFE(req, &iodq, r_achain, treq) {
440 TAILQ_REMOVE(&iodq, req, r_achain);
3e170ce0 441 req->r_achain.tqe_next = NFSREQNOLIST;
2d21ac55
A
442 req->r_callback.rcb_func(req);
443 }
444
445 /* now check if there's more/other work to be done */
446 lck_mtx_lock(nfsiod_mutex);
447 morework = !TAILQ_EMPTY(&nmp->nm_iodq);
448 if (!morework || !TAILQ_EMPTY(&nfsiodmounts)) {
fe8ab488
A
449 /*
450 * we're going to stop working on this mount but if the
451 * mount still needs more work so queue it up
452 */
453 if (morework && nmp->nm_iodlink.tqe_next == NFSNOLIST)
2d21ac55
A
454 TAILQ_INSERT_TAIL(&nfsiodmounts, nmp, nm_iodlink);
455 nmp->nm_niod = NULL;
456 niod->niod_nmp = NULL;
457 }
458 }
459
460 /* loop if there's still a mount to work on */
461 if (!niod->niod_nmp && !TAILQ_EMPTY(&nfsiodmounts)) {
462 niod->niod_nmp = TAILQ_FIRST(&nfsiodmounts);
463 TAILQ_REMOVE(&nfsiodmounts, niod->niod_nmp, nm_iodlink);
fe8ab488 464 niod->niod_nmp->nm_iodlink.tqe_next = NFSNOLIST;
2d21ac55
A
465 }
466 if (niod->niod_nmp)
467 goto worktodo;
468
469 /* queue ourselves back up - if there aren't too many threads running */
470 if (nfsiod_thread_count <= NFSIOD_MAX) {
471 TAILQ_INSERT_HEAD(&nfsiodfree, niod, niod_link);
472 error = msleep0(niod, nfsiod_mutex, PWAIT | PDROP, "nfsiod", NFS_ASYNCTHREADMAXIDLE*hz, nfsiod_continue);
473 /* shouldn't return... so we have an error */
474 /* remove an old nfsiod struct and terminate */
475 lck_mtx_lock(nfsiod_mutex);
476 if ((niod = TAILQ_LAST(&nfsiodfree, nfsiodlist)))
477 TAILQ_REMOVE(&nfsiodfree, niod, niod_link);
478 }
479 nfsiod_terminate(niod);
480 /*NOTREACHED*/
481 return (0);
482}
483
484#endif /* NFSCLIENT */
485
486
487#if NFSSERVER
488
1c79356b
A
489/*
490 * NFS server system calls
491 * getfh() lives here too, but maybe should move to kern/vfs_syscalls.c
492 */
493
494/*
495 * Get file handle system call
496 */
1c79356b 497int
91447636 498getfh(proc_t p, struct getfh_args *uap, __unused int *retval)
1c79356b 499{
91447636
A
500 vnode_t vp;
501 struct nfs_filehandle nfh;
6d2010ae 502 int error, fhlen, fidlen;
1c79356b 503 struct nameidata nd;
91447636 504 char path[MAXPATHLEN], *ptr;
6d2010ae 505 size_t pathlen;
91447636
A
506 struct nfs_exportfs *nxfs;
507 struct nfs_export *nx;
508
1c79356b
A
509 /*
510 * Must be super user
511 */
91447636
A
512 error = proc_suser(p);
513 if (error)
514 return (error);
515
6d2010ae
A
516 error = copyinstr(uap->fname, path, MAXPATHLEN, &pathlen);
517 if (!error)
518 error = copyin(uap->fhp, &fhlen, sizeof(fhlen));
91447636 519 if (error)
1c79356b 520 return (error);
6d2010ae
A
521 /* limit fh size to length specified (or v3 size by default) */
522 if ((fhlen != NFSV2_MAX_FH_SIZE) && (fhlen != NFSV3_MAX_FH_SIZE))
523 fhlen = NFSV3_MAX_FH_SIZE;
524 fidlen = fhlen - sizeof(struct nfs_exphandle);
91447636 525
2d21ac55
A
526 if (!nfsrv_is_initialized())
527 return (EINVAL);
528
6d2010ae 529 NDINIT(&nd, LOOKUP, OP_LOOKUP, FOLLOW | LOCKLEAF | AUDITVNPATH1,
2d21ac55 530 UIO_SYSSPACE, CAST_USER_ADDR_T(path), vfs_context_current());
1c79356b
A
531 error = namei(&nd);
532 if (error)
533 return (error);
91447636
A
534 nameidone(&nd);
535
1c79356b 536 vp = nd.ni_vp;
91447636
A
537
538 // find exportfs that matches f_mntonname
2d21ac55 539 lck_rw_lock_shared(&nfsrv_export_rwlock);
91447636 540 ptr = vnode_mount(vp)->mnt_vfsstat.f_mntonname;
2d21ac55
A
541 LIST_FOREACH(nxfs, &nfsrv_exports, nxfs_next) {
542 if (!strncmp(nxfs->nxfs_path, ptr, MAXPATHLEN))
91447636
A
543 break;
544 }
545 if (!nxfs || strncmp(nxfs->nxfs_path, path, strlen(nxfs->nxfs_path))) {
546 error = EINVAL;
547 goto out;
548 }
549 // find export that best matches remainder of path
550 ptr = path + strlen(nxfs->nxfs_path);
551 while (*ptr && (*ptr == '/'))
552 ptr++;
553 LIST_FOREACH(nx, &nxfs->nxfs_exports, nx_next) {
554 int len = strlen(nx->nx_path);
555 if (len == 0) // we've hit the export entry for the root directory
556 break;
557 if (!strncmp(nx->nx_path, ptr, len))
558 break;
559 }
560 if (!nx) {
561 error = EINVAL;
562 goto out;
563 }
564
565 bzero(&nfh, sizeof(nfh));
0c530ab8
A
566 nfh.nfh_xh.nxh_version = htonl(NFS_FH_VERSION);
567 nfh.nfh_xh.nxh_fsid = htonl(nxfs->nxfs_id);
568 nfh.nfh_xh.nxh_expid = htonl(nx->nx_id);
91447636
A
569 nfh.nfh_xh.nxh_flags = 0;
570 nfh.nfh_xh.nxh_reserved = 0;
6d2010ae 571 nfh.nfh_len = fidlen;
2d21ac55 572 error = VFS_VPTOFH(vp, (int*)&nfh.nfh_len, &nfh.nfh_fid[0], NULL);
6d2010ae 573 if (nfh.nfh_len > (uint32_t)fidlen)
91447636
A
574 error = EOVERFLOW;
575 nfh.nfh_xh.nxh_fidlen = nfh.nfh_len;
576 nfh.nfh_len += sizeof(nfh.nfh_xh);
2d21ac55 577 nfh.nfh_fhp = (u_char*)&nfh.nfh_xh;
91447636
A
578
579out:
2d21ac55 580 lck_rw_done(&nfsrv_export_rwlock);
91447636 581 vnode_put(vp);
1c79356b
A
582 if (error)
583 return (error);
6d2010ae 584 error = copyout((caddr_t)&nfh, uap->fhp, sizeof(fhandle_t));
1c79356b
A
585 return (error);
586}
587
39236c6e 588extern const struct fileops vnops;
91447636 589
55e303ae
A
590/*
591 * syscall for the rpc.lockd to use to translate a NFS file handle into
592 * an open descriptor.
593 *
594 * warning: do not remove the suser() call or this becomes one giant
595 * security hole.
596 */
55e303ae 597int
91447636
A
598fhopen( proc_t p,
599 struct fhopen_args *uap,
b0d623f7 600 int32_t *retval)
55e303ae 601{
91447636
A
602 vnode_t vp;
603 struct nfs_filehandle nfh;
604 struct nfs_export *nx;
605 struct nfs_export_options *nxo;
55e303ae 606 struct flock lf;
91447636
A
607 struct fileproc *fp, *nfp;
608 int fmode, error, type;
55e303ae 609 int indx;
2d21ac55 610 vfs_context_t ctx = vfs_context_current();
91447636
A
611 kauth_action_t action;
612
55e303ae
A
613 /*
614 * Must be super user
615 */
2d21ac55 616 error = suser(vfs_context_ucred(ctx), 0);
0c530ab8 617 if (error) {
55e303ae 618 return (error);
0c530ab8 619 }
55e303ae 620
2d21ac55
A
621 if (!nfsrv_is_initialized()) {
622 return (EINVAL);
623 }
624
55e303ae
A
625 fmode = FFLAGS(uap->flags);
626 /* why not allow a non-read/write open for our lockd? */
2d21ac55 627 if (((fmode & (FREAD | FWRITE)) == 0) || (fmode & O_CREAT))
55e303ae 628 return (EINVAL);
91447636
A
629
630 error = copyin(uap->u_fhp, &nfh.nfh_len, sizeof(nfh.nfh_len));
2d21ac55 631 if (error)
91447636
A
632 return (error);
633 if ((nfh.nfh_len < (int)sizeof(struct nfs_exphandle)) ||
2d21ac55 634 (nfh.nfh_len > (int)NFSV3_MAX_FH_SIZE))
91447636
A
635 return (EINVAL);
636 error = copyin(uap->u_fhp, &nfh, sizeof(nfh.nfh_len) + nfh.nfh_len);
2d21ac55 637 if (error)
55e303ae 638 return (error);
2d21ac55 639 nfh.nfh_fhp = (u_char*)&nfh.nfh_xh;
91447636 640
2d21ac55 641 lck_rw_lock_shared(&nfsrv_export_rwlock);
91447636 642 /* now give me my vnode, it gets returned to me with a reference */
2d21ac55
A
643 error = nfsrv_fhtovp(&nfh, NULL, &vp, &nx, &nxo);
644 lck_rw_done(&nfsrv_export_rwlock);
0c530ab8 645 if (error) {
2d21ac55
A
646 if (error == NFSERR_TRYLATER)
647 error = EAGAIN; // XXX EBUSY? Or just leave as TRYLATER?
55e303ae 648 return (error);
0c530ab8 649 }
91447636 650
55e303ae 651 /*
91447636
A
652 * From now on we have to make sure not
653 * to forget about the vnode.
654 * Any error that causes an abort must vnode_put(vp).
655 * Just set error = err and 'goto bad;'.
55e303ae
A
656 */
657
658 /*
659 * from vn_open
660 */
91447636 661 if (vnode_vtype(vp) == VSOCK) {
55e303ae
A
662 error = EOPNOTSUPP;
663 goto bad;
664 }
665
91447636
A
666 /* disallow write operations on directories */
667 if (vnode_isdir(vp) && (fmode & (FWRITE | O_TRUNC))) {
668 error = EISDIR;
55e303ae
A
669 goto bad;
670 }
671
4bd07ac2
A
672#if CONFIG_MACF
673 if ((error = mac_vnode_check_open(ctx, vp, fmode)))
674 goto bad;
675#endif
676
91447636
A
677 /* compute action to be authorized */
678 action = 0;
679 if (fmode & FREAD)
680 action |= KAUTH_VNODE_READ_DATA;
681 if (fmode & (FWRITE | O_TRUNC))
682 action |= KAUTH_VNODE_WRITE_DATA;
2d21ac55 683 if ((error = vnode_authorize(vp, NULL, action, ctx)) != 0)
91447636 684 goto bad;
55e303ae 685
2d21ac55 686 if ((error = VNOP_OPEN(vp, fmode, ctx)))
91447636 687 goto bad;
6d2010ae 688 if ((error = vnode_ref_ext(vp, fmode, 0)))
55e303ae
A
689 goto bad;
690
55e303ae
A
691 /*
692 * end of vn_open code
693 */
694
91447636 695 // starting here... error paths should call vn_close/vnode_put
2d21ac55
A
696 if ((error = falloc(p, &nfp, &indx, ctx)) != 0) {
697 vn_close(vp, fmode & FMASK, ctx);
55e303ae
A
698 goto bad;
699 }
700 fp = nfp;
701
91447636 702 fp->f_fglob->fg_flag = fmode & FMASK;
91447636
A
703 fp->f_fglob->fg_ops = &vnops;
704 fp->f_fglob->fg_data = (caddr_t)vp;
705
706 // XXX do we really need to support this with fhopen()?
55e303ae
A
707 if (fmode & (O_EXLOCK | O_SHLOCK)) {
708 lf.l_whence = SEEK_SET;
709 lf.l_start = 0;
710 lf.l_len = 0;
711 if (fmode & O_EXLOCK)
712 lf.l_type = F_WRLCK;
713 else
714 lf.l_type = F_RDLCK;
715 type = F_FLOCK;
716 if ((fmode & FNONBLOCK) == 0)
717 type |= F_WAIT;
39236c6e 718 if ((error = VNOP_ADVLOCK(vp, (caddr_t)fp->f_fglob, F_SETLK, &lf, type, ctx, NULL))) {
2d21ac55
A
719 struct vfs_context context = *vfs_context_current();
720 /* Modify local copy (to not damage thread copy) */
721 context.vc_ucred = fp->f_fglob->fg_cred;
722
723 vn_close(vp, fp->f_fglob->fg_flag, &context);
91447636 724 fp_free(p, indx, fp);
55e303ae
A
725 return (error);
726 }
91447636 727 fp->f_fglob->fg_flag |= FHASLOCK;
55e303ae
A
728 }
729
91447636
A
730 vnode_put(vp);
731
732 proc_fdlock(p);
6601e61a 733 procfdtbl_releasefd(p, indx, NULL);
91447636
A
734 fp_drop(p, indx, fp, 1);
735 proc_fdunlock(p);
736
55e303ae
A
737 *retval = indx;
738 return (0);
739
740bad:
91447636 741 vnode_put(vp);
55e303ae
A
742 return (error);
743}
744
1c79356b 745/*
2d21ac55 746 * NFS server pseudo system call
1c79356b 747 */
1c79356b 748int
91447636 749nfssvc(proc_t p, struct nfssvc_args *uap, __unused int *retval)
1c79356b 750{
91447636
A
751 mbuf_t nam;
752 struct user_nfsd_args user_nfsdarg;
91447636 753 socket_t so;
1c79356b
A
754 int error;
755
ccc36f2f
A
756 AUDIT_ARG(cmd, uap->flag);
757
1c79356b 758 /*
b0d623f7 759 * Must be super user for most operations (export ops checked later).
1c79356b 760 */
b0d623f7 761 if ((uap->flag != NFSSVC_EXPORT) && ((error = proc_suser(p))))
1c79356b 762 return (error);
2d21ac55
A
763#if CONFIG_MACF
764 error = mac_system_check_nfsd(kauth_cred_get());
765 if (error)
766 return (error);
767#endif
91447636 768
2d21ac55
A
769 /* make sure NFS server data structures have been initialized */
770 nfsrv_init();
1c79356b 771
2d21ac55 772 if (uap->flag & NFSSVC_ADDSOCK) {
91447636
A
773 if (IS_64BIT_PROCESS(p)) {
774 error = copyin(uap->argp, (caddr_t)&user_nfsdarg, sizeof(user_nfsdarg));
775 } else {
776 struct nfsd_args tmp_args;
777 error = copyin(uap->argp, (caddr_t)&tmp_args, sizeof(tmp_args));
778 if (error == 0) {
779 user_nfsdarg.sock = tmp_args.sock;
780 user_nfsdarg.name = CAST_USER_ADDR_T(tmp_args.name);
781 user_nfsdarg.namelen = tmp_args.namelen;
782 }
783 }
1c79356b
A
784 if (error)
785 return (error);
91447636
A
786 /* get the socket */
787 error = file_socket(user_nfsdarg.sock, &so);
1c79356b
A
788 if (error)
789 return (error);
91447636
A
790 /* Get the client address for connected sockets. */
791 if (user_nfsdarg.name == USER_ADDR_NULL || user_nfsdarg.namelen == 0) {
792 nam = NULL;
793 } else {
794 error = sockargs(&nam, user_nfsdarg.name, user_nfsdarg.namelen, MBUF_TYPE_SONAME);
795 if (error) {
796 /* drop the iocount file_socket() grabbed on the file descriptor */
797 file_drop(user_nfsdarg.sock);
1c79356b 798 return (error);
91447636 799 }
1c79356b 800 }
91447636
A
801 /*
802 * nfssvc_addsock() will grab a retain count on the socket
803 * to keep the socket from being closed when nfsd closes its
804 * file descriptor for it.
805 */
2d21ac55 806 error = nfssvc_addsock(so, nam);
91447636
A
807 /* drop the iocount file_socket() grabbed on the file descriptor */
808 file_drop(user_nfsdarg.sock);
809 } else if (uap->flag & NFSSVC_NFSD) {
2d21ac55 810 error = nfssvc_nfsd();
91447636 811 } else if (uap->flag & NFSSVC_EXPORT) {
2d21ac55 812 error = nfssvc_export(uap->argp);
91447636
A
813 } else {
814 error = EINVAL;
1c79356b 815 }
1c79356b
A
816 if (error == EINTR || error == ERESTART)
817 error = 0;
818 return (error);
819}
820
1c79356b
A
821/*
822 * Adds a socket to the list for servicing by nfsds.
823 */
b0d623f7 824int
2d21ac55 825nfssvc_addsock(socket_t so, mbuf_t mynam)
1c79356b 826{
2d21ac55
A
827 struct nfsrv_sock *slp;
828 int error = 0, sodomain, sotype, soprotocol, on = 1;
316670eb 829 int first;
91447636
A
830 struct timeval timeo;
831
832 /* make sure mbuf constants are set up */
2d21ac55 833 if (!nfs_mbuf_mhlen)
91447636
A
834 nfs_mbuf_init();
835
836 sock_gettype(so, &sodomain, &sotype, &soprotocol);
837
6d2010ae
A
838 /* There should be only one UDP socket for each of IPv4 and IPv6 */
839 if ((sodomain == AF_INET) && (soprotocol == IPPROTO_UDP) && nfsrv_udpsock) {
840 mbuf_freem(mynam);
841 return (EEXIST);
842 }
843 if ((sodomain == AF_INET6) && (soprotocol == IPPROTO_UDP) && nfsrv_udp6sock) {
91447636 844 mbuf_freem(mynam);
2d21ac55 845 return (EEXIST);
1c79356b
A
846 }
847
2d21ac55 848 /* Set protocol options and reserve some space (for UDP). */
3e170ce0
A
849 if (sotype == SOCK_STREAM) {
850 error = nfsrv_check_exports_allow_address(mynam);
851 if (error)
852 return (error);
91447636 853 sock_setsockopt(so, SOL_SOCKET, SO_KEEPALIVE, &on, sizeof(on));
3e170ce0 854 }
2d21ac55 855 if ((sodomain == AF_INET) && (soprotocol == IPPROTO_TCP))
91447636 856 sock_setsockopt(so, IPPROTO_TCP, TCP_NODELAY, &on, sizeof(on));
2d21ac55
A
857 if (sotype == SOCK_DGRAM) { /* set socket buffer sizes for UDP */
858 int reserve = NFS_UDPSOCKBUF;
859 error |= sock_setsockopt(so, SOL_SOCKET, SO_SNDBUF, &reserve, sizeof(reserve));
860 error |= sock_setsockopt(so, SOL_SOCKET, SO_RCVBUF, &reserve, sizeof(reserve));
861 if (error) {
862 log(LOG_INFO, "nfssvc_addsock: UDP socket buffer setting error(s) %d\n", error);
863 error = 0;
864 }
1c79356b 865 }
91447636
A
866 sock_nointerrupt(so, 0);
867
2d21ac55
A
868 /*
869 * Set socket send/receive timeouts.
870 * Receive timeout shouldn't matter, but setting the send timeout
871 * will make sure that an unresponsive client can't hang the server.
872 */
91447636 873 timeo.tv_usec = 0;
2d21ac55
A
874 timeo.tv_sec = 1;
875 error |= sock_setsockopt(so, SOL_SOCKET, SO_RCVTIMEO, &timeo, sizeof(timeo));
876 timeo.tv_sec = 30;
877 error |= sock_setsockopt(so, SOL_SOCKET, SO_SNDTIMEO, &timeo, sizeof(timeo));
878 if (error) {
879 log(LOG_INFO, "nfssvc_addsock: socket timeout setting error(s) %d\n", error);
880 error = 0;
881 }
91447636 882
2d21ac55
A
883 MALLOC(slp, struct nfsrv_sock *, sizeof(struct nfsrv_sock), M_NFSSVC, M_WAITOK);
884 if (!slp) {
885 mbuf_freem(mynam);
886 return (ENOMEM);
887 }
888 bzero((caddr_t)slp, sizeof (struct nfsrv_sock));
889 lck_rw_init(&slp->ns_rwlock, nfsrv_slp_rwlock_group, LCK_ATTR_NULL);
890 lck_mtx_init(&slp->ns_wgmutex, nfsrv_slp_mutex_group, LCK_ATTR_NULL);
891
892 lck_mtx_lock(nfsd_mutex);
893
894 if (soprotocol == IPPROTO_UDP) {
6d2010ae
A
895 if (sodomain == AF_INET) {
896 /* There should be only one UDP/IPv4 socket */
897 if (nfsrv_udpsock) {
898 lck_mtx_unlock(nfsd_mutex);
899 nfsrv_slpfree(slp);
900 mbuf_freem(mynam);
901 return (EEXIST);
902 }
903 nfsrv_udpsock = slp;
904 }
905 if (sodomain == AF_INET6) {
906 /* There should be only one UDP/IPv6 socket */
907 if (nfsrv_udp6sock) {
908 lck_mtx_unlock(nfsd_mutex);
909 nfsrv_slpfree(slp);
910 mbuf_freem(mynam);
911 return (EEXIST);
912 }
913 nfsrv_udp6sock = slp;
91447636 914 }
1c79356b 915 }
91447636 916
2d21ac55 917 /* add the socket to the list */
316670eb 918 first = TAILQ_EMPTY(&nfsrv_socklist);
2d21ac55 919 TAILQ_INSERT_TAIL(&nfsrv_socklist, slp, ns_chain);
3e170ce0
A
920 if (soprotocol == IPPROTO_TCP) {
921 nfsrv_sock_tcp_cnt++;
922 if (nfsrv_sock_idle_timeout < 0)
923 nfsrv_sock_idle_timeout = 0;
924 if (nfsrv_sock_idle_timeout && (nfsrv_sock_idle_timeout < NFSD_MIN_IDLE_TIMEOUT))
925 nfsrv_sock_idle_timeout = NFSD_MIN_IDLE_TIMEOUT;
926 /*
927 * Possibly start or stop the idle timer. We only start the idle timer when
928 * we have more than 2 * nfsd_thread_max connections. If the idle timer is
929 * on then we may need to turn it off based on the nvsrv_sock_idle_timeout or
930 * the number of connections.
931 */
932 if ((nfsrv_sock_tcp_cnt > 2 * nfsd_thread_max) || nfsrv_idlesock_timer_on) {
933 if (nfsrv_sock_idle_timeout == 0 || nfsrv_sock_tcp_cnt <= 2 * nfsd_thread_max) {
934 if (nfsrv_idlesock_timer_on) {
935 thread_call_cancel(nfsrv_idlesock_timer_call);
936 nfsrv_idlesock_timer_on = 0;
937 }
938 } else {
939 struct nfsrv_sock *old_slp;
940 struct timeval now;
941 time_t time_to_wait = nfsrv_sock_idle_timeout;
942 /*
943 * Get the oldest tcp socket and calculate the
944 * earliest time for the next idle timer to fire
945 * based on the possibly updated nfsrv_sock_idle_timeout
946 */
947 TAILQ_FOREACH(old_slp, &nfsrv_socklist, ns_chain) {
948 if (old_slp->ns_sotype == SOCK_STREAM) {
949 microuptime(&now);
950 time_to_wait -= now.tv_sec - old_slp->ns_timestamp;
951 if (time_to_wait < 1)
952 time_to_wait = 1;
953 break;
954 }
955 }
956 /*
957 * If we have a timer scheduled, but if its going to fire too late,
958 * turn it off.
959 */
960 if (nfsrv_idlesock_timer_on > now.tv_sec + time_to_wait) {
961 thread_call_cancel(nfsrv_idlesock_timer_call);
962 nfsrv_idlesock_timer_on = 0;
963 }
964 /* Schedule the idle thread if it isn't already */
965 if (!nfsrv_idlesock_timer_on) {
966 nfs_interval_timer_start(nfsrv_idlesock_timer_call, time_to_wait * 1000);
967 nfsrv_idlesock_timer_on = now.tv_sec + time_to_wait;
968 }
969 }
970 }
971 }
2d21ac55 972
91447636 973 sock_retain(so); /* grab a retain count on the socket */
1c79356b 974 slp->ns_so = so;
91447636 975 slp->ns_sotype = sotype;
1c79356b 976 slp->ns_nam = mynam;
91447636 977
316670eb
A
978 /* set up the socket up-call */
979 nfsrv_uc_addsock(slp, first);
91447636 980
2d21ac55
A
981 /* mark that the socket is not in the nfsrv_sockwg list */
982 slp->ns_wgq.tqe_next = SLPNOLIST;
3e170ce0 983
91447636
A
984 slp->ns_flag = SLP_VALID | SLP_NEEDQ;
985
1c79356b 986 nfsrv_wakenfsd(slp);
91447636
A
987 lck_mtx_unlock(nfsd_mutex);
988
1c79356b
A
989 return (0);
990}
991
992/*
2d21ac55
A
993 * nfssvc_nfsd()
994 *
995 * nfsd theory of operation:
996 *
997 * The first nfsd thread stays in user mode accepting new TCP connections
998 * which are then added via the "addsock" call. The rest of the nfsd threads
999 * simply call into the kernel and remain there in a loop handling NFS
1000 * requests until killed by a signal.
1001 *
1002 * There's a list of nfsd threads (nfsd_head).
1003 * There's an nfsd queue that contains only those nfsds that are
1004 * waiting for work to do (nfsd_queue).
1005 *
1006 * There's a list of all NFS sockets (nfsrv_socklist) and two queues for
1007 * managing the work on the sockets:
1008 * nfsrv_sockwait - sockets w/new data waiting to be worked on
1009 * nfsrv_sockwork - sockets being worked on which may have more work to do
1010 * nfsrv_sockwg -- sockets which have pending write gather data
1011 * When a socket receives data, if it is not currently queued, it
1012 * will be placed at the end of the "wait" queue.
1013 * Whenever a socket needs servicing we make sure it is queued and
1014 * wake up a waiting nfsd (if there is one).
1015 *
1016 * nfsds will service at most 8 requests from the same socket before
1017 * defecting to work on another socket.
1018 * nfsds will defect immediately if there are any sockets in the "wait" queue
1019 * nfsds looking for a socket to work on check the "wait" queue first and
1020 * then check the "work" queue.
1021 * When an nfsd starts working on a socket, it removes it from the head of
1022 * the queue it's currently on and moves it to the end of the "work" queue.
1023 * When nfsds are checking the queues for work, any sockets found not to
1024 * have any work are simply dropped from the queue.
1025 *
1c79356b 1026 */
b0d623f7 1027int
2d21ac55 1028nfssvc_nfsd(void)
1c79356b 1029{
2d21ac55
A
1030 mbuf_t m, mrep;
1031 struct nfsrv_sock *slp;
1032 struct nfsd *nfsd;
1c79356b 1033 struct nfsrv_descript *nd = NULL;
91447636 1034 int error = 0, cacherep, writes_todo;
2d21ac55 1035 int siz, procrastinate, opcnt = 0;
1c79356b 1036 u_quad_t cur_usec;
55e303ae 1037 struct timeval now;
2d21ac55 1038 struct vfs_context context;
316670eb 1039 struct timespec to;
1c79356b
A
1040
1041#ifndef nolint
1042 cacherep = RC_DOIT;
1043 writes_todo = 0;
1044#endif
91447636 1045
2d21ac55
A
1046 MALLOC(nfsd, struct nfsd *, sizeof(struct nfsd), M_NFSD, M_WAITOK);
1047 if (!nfsd)
1048 return (ENOMEM);
1049 bzero(nfsd, sizeof(struct nfsd));
1050 lck_mtx_lock(nfsd_mutex);
1051 if (nfsd_thread_count++ == 0)
1052 nfsrv_initcache(); /* Init the server request cache */
316670eb 1053
2d21ac55
A
1054 TAILQ_INSERT_TAIL(&nfsd_head, nfsd, nfsd_chain);
1055 lck_mtx_unlock(nfsd_mutex);
1056
1057 context.vc_thread = current_thread();
91447636 1058
316670eb
A
1059 /* Set time out so that nfsd threads can wake up a see if they are still needed. */
1060 to.tv_sec = 5;
1061 to.tv_nsec = 0;
1062
1c79356b
A
1063 /*
1064 * Loop getting rpc requests until SIGKILL.
1065 */
1066 for (;;) {
2d21ac55
A
1067 if (nfsd_thread_max <= 0) {
1068 /* NFS server shutting down, get out ASAP */
1069 error = EINTR;
1070 slp = nfsd->nfsd_slp;
1071 } else if (nfsd->nfsd_flag & NFSD_REQINPROG) {
1072 /* already have some work to do */
1073 error = 0;
1074 slp = nfsd->nfsd_slp;
1075 } else {
1076 /* need to find work to do */
1077 error = 0;
91447636 1078 lck_mtx_lock(nfsd_mutex);
2d21ac55
A
1079 while (!nfsd->nfsd_slp && TAILQ_EMPTY(&nfsrv_sockwait) && TAILQ_EMPTY(&nfsrv_sockwork)) {
1080 if (nfsd_thread_count > nfsd_thread_max) {
1081 /*
1082 * If we have no socket and there are more
1083 * nfsd threads than configured, let's exit.
1084 */
1085 error = 0;
1086 goto done;
1087 }
1c79356b 1088 nfsd->nfsd_flag |= NFSD_WAITING;
2d21ac55 1089 TAILQ_INSERT_HEAD(&nfsd_queue, nfsd, nfsd_queue);
316670eb 1090 error = msleep(nfsd, nfsd_mutex, PSOCK | PCATCH, "nfsd", &to);
91447636 1091 if (error) {
2d21ac55
A
1092 if (nfsd->nfsd_flag & NFSD_WAITING) {
1093 TAILQ_REMOVE(&nfsd_queue, nfsd, nfsd_queue);
1094 nfsd->nfsd_flag &= ~NFSD_WAITING;
1095 }
316670eb
A
1096 if (error == EWOULDBLOCK)
1097 continue;
1c79356b 1098 goto done;
91447636 1099 }
1c79356b 1100 }
2d21ac55
A
1101 slp = nfsd->nfsd_slp;
1102 if (!slp && !TAILQ_EMPTY(&nfsrv_sockwait)) {
1103 /* look for a socket to work on in the wait queue */
1104 while ((slp = TAILQ_FIRST(&nfsrv_sockwait))) {
1105 lck_rw_lock_exclusive(&slp->ns_rwlock);
1106 /* remove from the head of the queue */
1107 TAILQ_REMOVE(&nfsrv_sockwait, slp, ns_svcq);
1108 slp->ns_flag &= ~SLP_WAITQ;
1109 if ((slp->ns_flag & SLP_VALID) && (slp->ns_flag & SLP_WORKTODO))
1110 break;
1111 /* nothing to do, so skip this socket */
1112 lck_rw_done(&slp->ns_rwlock);
1c79356b 1113 }
2d21ac55
A
1114 }
1115 if (!slp && !TAILQ_EMPTY(&nfsrv_sockwork)) {
1116 /* look for a socket to work on in the work queue */
1117 while ((slp = TAILQ_FIRST(&nfsrv_sockwork))) {
1118 lck_rw_lock_exclusive(&slp->ns_rwlock);
1119 /* remove from the head of the queue */
1120 TAILQ_REMOVE(&nfsrv_sockwork, slp, ns_svcq);
1121 slp->ns_flag &= ~SLP_WORKQ;
1122 if ((slp->ns_flag & SLP_VALID) && (slp->ns_flag & SLP_WORKTODO))
1123 break;
1124 /* nothing to do, so skip this socket */
1125 lck_rw_done(&slp->ns_rwlock);
1126 }
1127 }
1128 if (!nfsd->nfsd_slp && slp) {
1129 /* we found a socket to work on, grab a reference */
1130 slp->ns_sref++;
3e170ce0
A
1131 microuptime(&now);
1132 slp->ns_timestamp = now.tv_sec;
1133 /* We keep the socket list in least recently used order for reaping idle sockets */
1134 TAILQ_REMOVE(&nfsrv_socklist, slp, ns_chain);
1135 TAILQ_INSERT_TAIL(&nfsrv_socklist, slp, ns_chain);
2d21ac55
A
1136 nfsd->nfsd_slp = slp;
1137 opcnt = 0;
1138 /* and put it at the back of the work queue */
1139 TAILQ_INSERT_TAIL(&nfsrv_sockwork, slp, ns_svcq);
1140 slp->ns_flag |= SLP_WORKQ;
1141 lck_rw_done(&slp->ns_rwlock);
1c79356b 1142 }
91447636 1143 lck_mtx_unlock(nfsd_mutex);
2d21ac55 1144 if (!slp)
1c79356b 1145 continue;
91447636 1146 lck_rw_lock_exclusive(&slp->ns_rwlock);
1c79356b 1147 if (slp->ns_flag & SLP_VALID) {
743b1565 1148 if ((slp->ns_flag & (SLP_NEEDQ|SLP_DISCONN)) == SLP_NEEDQ) {
91447636
A
1149 slp->ns_flag &= ~SLP_NEEDQ;
1150 nfsrv_rcv_locked(slp->ns_so, slp, MBUF_WAITOK);
1151 }
743b1565
A
1152 if (slp->ns_flag & SLP_DISCONN)
1153 nfsrv_zapsock(slp);
1c79356b 1154 error = nfsrv_dorec(slp, nfsd, &nd);
2d21ac55
A
1155 if (error == EINVAL) { // RPCSEC_GSS drop
1156 if (slp->ns_sotype == SOCK_STREAM)
1157 nfsrv_zapsock(slp); // drop connection
1158 }
1159 writes_todo = 0;
1160 if (error && (slp->ns_wgtime || (slp->ns_flag & SLP_DOWRITES))) {
1161 microuptime(&now);
1162 cur_usec = (u_quad_t)now.tv_sec * 1000000 +
1163 (u_quad_t)now.tv_usec;
1164 if (slp->ns_wgtime <= cur_usec) {
1165 error = 0;
1166 cacherep = RC_DOIT;
1167 writes_todo = 1;
1168 }
1169 slp->ns_flag &= ~SLP_DOWRITES;
1170 }
1c79356b
A
1171 nfsd->nfsd_flag |= NFSD_REQINPROG;
1172 }
91447636 1173 lck_rw_done(&slp->ns_rwlock);
1c79356b 1174 }
2d21ac55 1175 if (error || (slp && !(slp->ns_flag & SLP_VALID))) {
1c79356b 1176 if (nd) {
2d21ac55 1177 nfsm_chain_cleanup(&nd->nd_nmreq);
55e303ae 1178 if (nd->nd_nam2)
91447636 1179 mbuf_freem(nd->nd_nam2);
0c530ab8
A
1180 if (IS_VALID_CRED(nd->nd_cr))
1181 kauth_cred_unref(&nd->nd_cr);
6d2010ae
A
1182 if (nd->nd_gss_context)
1183 nfs_gss_svc_ctx_deref(nd->nd_gss_context);
2d21ac55 1184 FREE_ZONE(nd, sizeof(*nd), M_NFSRVDESC);
1c79356b
A
1185 nd = NULL;
1186 }
91447636 1187 nfsd->nfsd_slp = NULL;
1c79356b 1188 nfsd->nfsd_flag &= ~NFSD_REQINPROG;
2d21ac55
A
1189 if (slp)
1190 nfsrv_slpderef(slp);
1191 if (nfsd_thread_max <= 0)
1192 break;
1c79356b
A
1193 continue;
1194 }
1c79356b 1195 if (nd) {
55e303ae 1196 microuptime(&nd->nd_starttime);
1c79356b
A
1197 if (nd->nd_nam2)
1198 nd->nd_nam = nd->nd_nam2;
1199 else
1200 nd->nd_nam = slp->ns_nam;
1201
2d21ac55
A
1202 cacherep = nfsrv_getcache(nd, slp, &mrep);
1203
1204 if (nfsrv_require_resv_port) {
1205 /* Check if source port is a reserved port */
6d2010ae
A
1206 in_port_t port = 0;
1207 struct sockaddr *saddr = mbuf_data(nd->nd_nam);
1208
1209 if (saddr->sa_family == AF_INET)
1210 port = ntohs(((struct sockaddr_in*)saddr)->sin_port);
1211 else if (saddr->sa_family == AF_INET6)
1212 port = ntohs(((struct sockaddr_in6*)saddr)->sin6_port);
1213 if ((port >= IPPORT_RESERVED) && (nd->nd_procnum != NFSPROC_NULL)) {
1c79356b
A
1214 nd->nd_procnum = NFSPROC_NOOP;
1215 nd->nd_repstat = (NFSERR_AUTHERR | AUTH_TOOWEAK);
1216 cacherep = RC_DOIT;
1c79356b
A
1217 }
1218 }
1219
1220 }
1221
1222 /*
2d21ac55 1223 * Loop to get all the write RPC replies that have been
1c79356b
A
1224 * gathered together.
1225 */
1226 do {
1227 switch (cacherep) {
1228 case RC_DOIT:
2d21ac55
A
1229 if (nd && (nd->nd_vers == NFS_VER3))
1230 procrastinate = nfsrv_wg_delay_v3;
1c79356b 1231 else
2d21ac55
A
1232 procrastinate = nfsrv_wg_delay;
1233 lck_rw_lock_shared(&nfsrv_export_rwlock);
1234 context.vc_ucred = NULL;
91447636 1235 if (writes_todo || ((nd->nd_procnum == NFSPROC_WRITE) && (procrastinate > 0)))
2d21ac55
A
1236 error = nfsrv_writegather(&nd, slp, &context, &mrep);
1237 else
1238 error = (*(nfsrv_procs[nd->nd_procnum]))(nd, slp, &context, &mrep);
1239 lck_rw_done(&nfsrv_export_rwlock);
1240 if (mrep == NULL) {
1241 /*
1242 * If this is a stream socket and we are not going
1243 * to send a reply we better close the connection
1244 * so the client doesn't hang.
1245 */
1246 if (error && slp->ns_sotype == SOCK_STREAM) {
1247 lck_rw_lock_exclusive(&slp->ns_rwlock);
1248 nfsrv_zapsock(slp);
1249 lck_rw_done(&slp->ns_rwlock);
1250 printf("NFS server: NULL reply from proc = %d error = %d\n",
1251 nd->nd_procnum, error);
1252 }
1c79356b 1253 break;
2d21ac55
A
1254
1255 }
1c79356b 1256 if (error) {
316670eb 1257 OSAddAtomic64(1, &nfsstats.srv_errs);
2d21ac55 1258 nfsrv_updatecache(nd, FALSE, mrep);
55e303ae 1259 if (nd->nd_nam2) {
91447636 1260 mbuf_freem(nd->nd_nam2);
55e303ae
A
1261 nd->nd_nam2 = NULL;
1262 }
1c79356b
A
1263 break;
1264 }
316670eb 1265 OSAddAtomic64(1, &nfsstats.srvrpccnt[nd->nd_procnum]);
2d21ac55
A
1266 nfsrv_updatecache(nd, TRUE, mrep);
1267 /* FALLTHRU */
1268
1c79356b 1269 case RC_REPLY:
2d21ac55
A
1270 if (nd->nd_gss_mb != NULL) { // It's RPCSEC_GSS
1271 /*
1272 * Need to checksum or encrypt the reply
1273 */
1274 error = nfs_gss_svc_protect_reply(nd, mrep);
1275 if (error) {
1276 mbuf_freem(mrep);
1277 break;
1278 }
1279 }
1280
1281 /*
1282 * Get the total size of the reply
1283 */
1284 m = mrep;
1c79356b
A
1285 siz = 0;
1286 while (m) {
91447636
A
1287 siz += mbuf_len(m);
1288 m = mbuf_next(m);
1c79356b
A
1289 }
1290 if (siz <= 0 || siz > NFS_MAXPACKET) {
1291 printf("mbuf siz=%d\n",siz);
1292 panic("Bad nfs svc reply");
1293 }
2d21ac55 1294 m = mrep;
91447636
A
1295 mbuf_pkthdr_setlen(m, siz);
1296 error = mbuf_pkthdr_setrcvif(m, NULL);
1297 if (error)
1298 panic("nfsd setrcvif failed: %d", error);
1c79356b
A
1299 /*
1300 * For stream protocols, prepend a Sun RPC
1301 * Record Mark.
1302 */
91447636
A
1303 if (slp->ns_sotype == SOCK_STREAM) {
1304 error = mbuf_prepend(&m, NFSX_UNSIGNED, MBUF_WAITOK);
1305 if (!error)
b0d623f7 1306 *(u_int32_t*)mbuf_data(m) = htonl(0x80000000 | siz);
1c79356b 1307 }
91447636
A
1308 if (!error) {
1309 if (slp->ns_flag & SLP_VALID) {
2d21ac55 1310 error = nfsrv_send(slp, nd->nd_nam2, m);
91447636
A
1311 } else {
1312 error = EPIPE;
1313 mbuf_freem(m);
1314 }
1315 } else {
1316 mbuf_freem(m);
1c79356b 1317 }
2d21ac55 1318 mrep = NULL;
55e303ae 1319 if (nd->nd_nam2) {
91447636 1320 mbuf_freem(nd->nd_nam2);
55e303ae
A
1321 nd->nd_nam2 = NULL;
1322 }
91447636
A
1323 if (error == EPIPE) {
1324 lck_rw_lock_exclusive(&slp->ns_rwlock);
1c79356b 1325 nfsrv_zapsock(slp);
91447636
A
1326 lck_rw_done(&slp->ns_rwlock);
1327 }
1c79356b 1328 if (error == EINTR || error == ERESTART) {
2d21ac55 1329 nfsm_chain_cleanup(&nd->nd_nmreq);
0c530ab8
A
1330 if (IS_VALID_CRED(nd->nd_cr))
1331 kauth_cred_unref(&nd->nd_cr);
6d2010ae
A
1332 if (nd->nd_gss_context)
1333 nfs_gss_svc_ctx_deref(nd->nd_gss_context);
2d21ac55 1334 FREE_ZONE(nd, sizeof(*nd), M_NFSRVDESC);
1c79356b 1335 nfsrv_slpderef(slp);
2d21ac55 1336 lck_mtx_lock(nfsd_mutex);
1c79356b
A
1337 goto done;
1338 }
1339 break;
1340 case RC_DROPIT:
91447636 1341 mbuf_freem(nd->nd_nam2);
2d21ac55 1342 nd->nd_nam2 = NULL;
1c79356b
A
1343 break;
1344 };
2d21ac55 1345 opcnt++;
1c79356b 1346 if (nd) {
2d21ac55 1347 nfsm_chain_cleanup(&nd->nd_nmreq);
55e303ae 1348 if (nd->nd_nam2)
91447636 1349 mbuf_freem(nd->nd_nam2);
0c530ab8
A
1350 if (IS_VALID_CRED(nd->nd_cr))
1351 kauth_cred_unref(&nd->nd_cr);
6d2010ae
A
1352 if (nd->nd_gss_context)
1353 nfs_gss_svc_ctx_deref(nd->nd_gss_context);
2d21ac55 1354 FREE_ZONE(nd, sizeof(*nd), M_NFSRVDESC);
1c79356b
A
1355 nd = NULL;
1356 }
1357
1358 /*
1359 * Check to see if there are outstanding writes that
1360 * need to be serviced.
1361 */
2d21ac55
A
1362 writes_todo = 0;
1363 if (slp->ns_wgtime) {
1364 microuptime(&now);
1365 cur_usec = (u_quad_t)now.tv_sec * 1000000 +
1366 (u_quad_t)now.tv_usec;
1367 if (slp->ns_wgtime <= cur_usec) {
1368 cacherep = RC_DOIT;
1369 writes_todo = 1;
1370 }
91447636 1371 }
1c79356b 1372 } while (writes_todo);
2d21ac55
A
1373
1374 nd = NULL;
1375 if (TAILQ_EMPTY(&nfsrv_sockwait) && (opcnt < 8)) {
1376 lck_rw_lock_exclusive(&slp->ns_rwlock);
1377 error = nfsrv_dorec(slp, nfsd, &nd);
1378 if (error == EINVAL) { // RPCSEC_GSS drop
1379 if (slp->ns_sotype == SOCK_STREAM)
1380 nfsrv_zapsock(slp); // drop connection
1381 }
91447636 1382 lck_rw_done(&slp->ns_rwlock);
2d21ac55
A
1383 }
1384 if (!nd) {
1385 /* drop our reference on the socket */
1c79356b
A
1386 nfsd->nfsd_flag &= ~NFSD_REQINPROG;
1387 nfsd->nfsd_slp = NULL;
1388 nfsrv_slpderef(slp);
91447636 1389 }
1c79356b 1390 }
91447636 1391 lck_mtx_lock(nfsd_mutex);
2d21ac55 1392done:
1c79356b 1393 TAILQ_REMOVE(&nfsd_head, nfsd, nfsd_chain);
91447636 1394 FREE(nfsd, M_NFSD);
2d21ac55
A
1395 if (--nfsd_thread_count == 0)
1396 nfsrv_cleanup();
91447636 1397 lck_mtx_unlock(nfsd_mutex);
1c79356b
A
1398 return (error);
1399}
91447636 1400
b0d623f7 1401int
2d21ac55 1402nfssvc_export(user_addr_t argp)
91447636
A
1403{
1404 int error = 0, is_64bit;
1405 struct user_nfs_export_args unxa;
2d21ac55 1406 vfs_context_t ctx = vfs_context_current();
91447636 1407
2d21ac55 1408 is_64bit = IS_64BIT_PROCESS(vfs_context_proc(ctx));
91447636
A
1409
1410 /* copy in pointers to path and export args */
1411 if (is_64bit) {
1412 error = copyin(argp, (caddr_t)&unxa, sizeof(unxa));
1413 } else {
1414 struct nfs_export_args tnxa;
1415 error = copyin(argp, (caddr_t)&tnxa, sizeof(tnxa));
1416 if (error == 0) {
1417 /* munge into LP64 version of nfs_export_args structure */
1418 unxa.nxa_fsid = tnxa.nxa_fsid;
1419 unxa.nxa_expid = tnxa.nxa_expid;
1420 unxa.nxa_fspath = CAST_USER_ADDR_T(tnxa.nxa_fspath);
1421 unxa.nxa_exppath = CAST_USER_ADDR_T(tnxa.nxa_exppath);
1422 unxa.nxa_flags = tnxa.nxa_flags;
1423 unxa.nxa_netcount = tnxa.nxa_netcount;
1424 unxa.nxa_nets = CAST_USER_ADDR_T(tnxa.nxa_nets);
1425 }
1426 }
1427 if (error)
1428 return (error);
1429
2d21ac55 1430 error = nfsrv_export(&unxa, ctx);
91447636
A
1431
1432 return (error);
1433}
1434
1c79356b 1435/*
2d21ac55 1436 * Shut down a socket associated with an nfsrv_sock structure.
1c79356b
A
1437 * Should be called with the send lock set, if required.
1438 * The trick here is to increment the sref at the start, so that the nfsds
1439 * will stop using it and clear ns_flag at the end so that it will not be
1440 * reassigned during cleanup.
1441 */
b0d623f7 1442void
2d21ac55 1443nfsrv_zapsock(struct nfsrv_sock *slp)
1c79356b 1444{
91447636 1445 socket_t so;
1c79356b 1446
91447636
A
1447 if ((slp->ns_flag & SLP_VALID) == 0)
1448 return;
1c79356b 1449 slp->ns_flag &= ~SLP_ALLFLAGS;
91447636
A
1450
1451 so = slp->ns_so;
1452 if (so == NULL)
1453 return;
1454
3e170ce0 1455 sock_setupcall(so, NULL, NULL);
91447636 1456 sock_shutdown(so, SHUT_RDWR);
316670eb
A
1457
1458 /*
1459 * Remove from the up-call queue
1460 */
1461 nfsrv_uc_dequeue(slp);
1c79356b
A
1462}
1463
1c79356b 1464/*
91447636 1465 * cleanup and release a server socket structure.
1c79356b 1466 */
b0d623f7 1467void
2d21ac55 1468nfsrv_slpfree(struct nfsrv_sock *slp)
1c79356b 1469{
91447636 1470 struct nfsrv_descript *nwp, *nnwp;
1c79356b 1471
91447636
A
1472 if (slp->ns_so) {
1473 sock_release(slp->ns_so);
1474 slp->ns_so = NULL;
1475 }
1476 if (slp->ns_nam)
1477 mbuf_free(slp->ns_nam);
1478 if (slp->ns_raw)
1479 mbuf_freem(slp->ns_raw);
1480 if (slp->ns_rec)
1481 mbuf_freem(slp->ns_rec);
2d21ac55
A
1482 if (slp->ns_frag)
1483 mbuf_freem(slp->ns_frag);
1484 slp->ns_nam = slp->ns_raw = slp->ns_rec = slp->ns_frag = NULL;
1485 slp->ns_reccnt = 0;
55e303ae 1486
91447636
A
1487 for (nwp = slp->ns_tq.lh_first; nwp; nwp = nnwp) {
1488 nnwp = nwp->nd_tq.le_next;
1489 LIST_REMOVE(nwp, nd_tq);
2d21ac55
A
1490 nfsm_chain_cleanup(&nwp->nd_nmreq);
1491 if (nwp->nd_mrep)
1492 mbuf_freem(nwp->nd_mrep);
1493 if (nwp->nd_nam2)
1494 mbuf_freem(nwp->nd_nam2);
0c530ab8
A
1495 if (IS_VALID_CRED(nwp->nd_cr))
1496 kauth_cred_unref(&nwp->nd_cr);
6d2010ae
A
1497 if (nwp->nd_gss_context)
1498 nfs_gss_svc_ctx_deref(nwp->nd_gss_context);
2d21ac55 1499 FREE_ZONE(nwp, sizeof(*nwp), M_NFSRVDESC);
55e303ae 1500 }
91447636
A
1501 LIST_INIT(&slp->ns_tq);
1502
2d21ac55
A
1503 lck_rw_destroy(&slp->ns_rwlock, nfsrv_slp_rwlock_group);
1504 lck_mtx_destroy(&slp->ns_wgmutex, nfsrv_slp_mutex_group);
91447636 1505 FREE(slp, M_NFSSVC);
55e303ae
A
1506}
1507
1508/*
91447636
A
1509 * Derefence a server socket structure. If it has no more references and
1510 * is no longer valid, you can throw it away.
55e303ae 1511 */
3e170ce0
A
1512static void
1513nfsrv_slpderef_locked(struct nfsrv_sock *slp)
55e303ae 1514{
91447636
A
1515 lck_rw_lock_exclusive(&slp->ns_rwlock);
1516 slp->ns_sref--;
2d21ac55 1517
91447636 1518 if (slp->ns_sref || (slp->ns_flag & SLP_VALID)) {
2d21ac55
A
1519 if ((slp->ns_flag & SLP_QUEUED) && !(slp->ns_flag & SLP_WORKTODO)) {
1520 /* remove socket from queue since there's no work */
1521 if (slp->ns_flag & SLP_WAITQ)
1522 TAILQ_REMOVE(&nfsrv_sockwait, slp, ns_svcq);
1523 else
1524 TAILQ_REMOVE(&nfsrv_sockwork, slp, ns_svcq);
1525 slp->ns_flag &= ~SLP_QUEUED;
1526 }
91447636 1527 lck_rw_done(&slp->ns_rwlock);
91447636 1528 return;
55e303ae 1529 }
91447636 1530
2d21ac55
A
1531 /* This socket is no longer valid, so we'll get rid of it */
1532
1533 if (slp->ns_flag & SLP_QUEUED) {
1534 if (slp->ns_flag & SLP_WAITQ)
1535 TAILQ_REMOVE(&nfsrv_sockwait, slp, ns_svcq);
1536 else
1537 TAILQ_REMOVE(&nfsrv_sockwork, slp, ns_svcq);
1538 slp->ns_flag &= ~SLP_QUEUED;
1539 }
3e170ce0 1540 lck_rw_done(&slp->ns_rwlock);
2d21ac55 1541
2d21ac55 1542 TAILQ_REMOVE(&nfsrv_socklist, slp, ns_chain);
3e170ce0
A
1543 if (slp->ns_sotype == SOCK_STREAM)
1544 nfsrv_sock_tcp_cnt--;
2d21ac55 1545
2d21ac55
A
1546 /* now remove from the write gather socket list */
1547 if (slp->ns_wgq.tqe_next != SLPNOLIST) {
1548 TAILQ_REMOVE(&nfsrv_sockwg, slp, ns_wgq);
1549 slp->ns_wgq.tqe_next = SLPNOLIST;
1550 }
3e170ce0
A
1551 nfsrv_slpfree(slp);
1552}
1553
1554void
1555nfsrv_slpderef(struct nfsrv_sock *slp)
1556{
1557 lck_mtx_lock(nfsd_mutex);
1558 nfsrv_slpderef_locked(slp);
91447636 1559 lck_mtx_unlock(nfsd_mutex);
55e303ae
A
1560}
1561
1c79356b 1562/*
3e170ce0
A
1563 * Check periodically for idle sockest if needed and
1564 * zap them.
1c79356b
A
1565 */
1566void
3e170ce0 1567nfsrv_idlesock_timer(__unused void *param0, __unused void *param1)
1c79356b 1568{
3e170ce0 1569 struct nfsrv_sock *slp, *tslp;
743b1565 1570 struct timeval now;
3e170ce0 1571 time_t time_to_wait = nfsrv_sock_idle_timeout;
1c79356b 1572
2d21ac55
A
1573 microuptime(&now);
1574 lck_mtx_lock(nfsd_mutex);
1c79356b 1575
3e170ce0
A
1576 /* Turn off the timer if we're suppose to and get out */
1577 if (nfsrv_sock_idle_timeout < NFSD_MIN_IDLE_TIMEOUT)
1578 nfsrv_sock_idle_timeout = 0;
1579 if ((nfsrv_sock_tcp_cnt <= 2 * nfsd_thread_max) || (nfsrv_sock_idle_timeout == 0)) {
1580 nfsrv_idlesock_timer_on = 0;
2d21ac55
A
1581 lck_mtx_unlock(nfsd_mutex);
1582 return;
91447636 1583 }
1c79356b 1584
3e170ce0
A
1585 TAILQ_FOREACH_SAFE(slp, &nfsrv_socklist, ns_chain, tslp) {
1586 lck_rw_lock_exclusive(&slp->ns_rwlock);
1587 /* Skip udp and referenced sockets */
1588 if (slp->ns_sotype == SOCK_DGRAM || slp->ns_sref) {
1589 lck_rw_done(&slp->ns_rwlock);
1590 continue;
1591 }
1592 /*
1593 * If this is the first non-referenced socket that hasn't idle out,
1594 * use its time stamp to calculate the earlist time in the future
1595 * to start the next invocation of the timer. Since the nfsrv_socklist
1596 * is sorted oldest access to newest. Once we find the first one,
1597 * we're done and break out of the loop.
1598 */
1599 if (((slp->ns_timestamp + nfsrv_sock_idle_timeout) > now.tv_sec) ||
1600 nfsrv_sock_tcp_cnt <= 2 * nfsd_thread_max) {
1601 time_to_wait -= now.tv_sec - slp->ns_timestamp;
1602 if (time_to_wait < 1)
1603 time_to_wait = 1;
1604 lck_rw_done(&slp->ns_rwlock);
1605 break;
1606 }
1607 /*
1608 * Bump the ref count. nfsrv_slpderef below will destroy
1609 * the socket, since nfsrv_zapsock has closed it.
1610 */
1611 slp->ns_sref++;
1612 nfsrv_zapsock(slp);
1613 lck_rw_done(&slp->ns_rwlock);
1614 nfsrv_slpderef_locked(slp);
1615 }
2d21ac55 1616
3e170ce0
A
1617 /* Start ourself back up */
1618 nfs_interval_timer_start(nfsrv_idlesock_timer_call, time_to_wait * 1000);
1619 /* Remember when the next timer will fire for nfssvc_addsock. */
1620 nfsrv_idlesock_timer_on = now.tv_sec + time_to_wait;
1621 lck_mtx_unlock(nfsd_mutex);
1c79356b
A
1622}
1623
1624/*
2d21ac55 1625 * Clean up the data structures for the server.
1c79356b 1626 */
2d21ac55
A
1627void
1628nfsrv_cleanup(void)
1c79356b 1629{
2d21ac55 1630 struct nfsrv_sock *slp, *nslp;
55e303ae 1631 struct timeval now;
b0d623f7 1632#if CONFIG_FSE
2d21ac55
A
1633 struct nfsrv_fmod *fp, *nfp;
1634 int i;
b0d623f7 1635#endif
1c79356b 1636
55e303ae 1637 microuptime(&now);
2d21ac55
A
1638 for (slp = TAILQ_FIRST(&nfsrv_socklist); slp != 0; slp = nslp) {
1639 nslp = TAILQ_NEXT(slp, ns_chain);
3e170ce0
A
1640 lck_rw_lock_exclusive(&slp->ns_rwlock);
1641 slp->ns_sref++;
1642 if (slp->ns_flag & SLP_VALID)
2d21ac55 1643 nfsrv_zapsock(slp);
3e170ce0
A
1644 lck_rw_done(&slp->ns_rwlock);
1645 nfsrv_slpderef_locked(slp);
2d21ac55 1646 }
3e170ce0 1647#
b0d623f7 1648#if CONFIG_FSE
2d21ac55
A
1649 /*
1650 * Flush pending file write fsevents
1651 */
1652 lck_mtx_lock(nfsrv_fmod_mutex);
1653 for (i = 0; i < NFSRVFMODHASHSZ; i++) {
1654 for (fp = LIST_FIRST(&nfsrv_fmod_hashtbl[i]); fp; fp = nfp) {
1655 /*
1656 * Fire off the content modified fsevent for each
1657 * entry, remove it from the list, and free it.
1658 */
6d2010ae
A
1659 if (nfsrv_fsevents_enabled) {
1660 fp->fm_context.vc_thread = current_thread();
2d21ac55
A
1661 add_fsevent(FSE_CONTENT_MODIFIED, &fp->fm_context,
1662 FSE_ARG_VNODE, fp->fm_vp,
1663 FSE_ARG_DONE);
6d2010ae 1664 }
2d21ac55
A
1665 vnode_put(fp->fm_vp);
1666 kauth_cred_unref(&fp->fm_context.vc_ucred);
1667 nfp = LIST_NEXT(fp, fm_link);
1668 LIST_REMOVE(fp, fm_link);
1669 FREE(fp, M_TEMP);
1670 }
1671 }
1672 nfsrv_fmod_pending = 0;
1673 lck_mtx_unlock(nfsrv_fmod_mutex);
b0d623f7 1674#endif
2d21ac55 1675
316670eb
A
1676 nfsrv_uc_cleanup(); /* Stop nfs socket up-call threads */
1677
2d21ac55
A
1678 nfs_gss_svc_cleanup(); /* Remove any RPCSEC_GSS contexts */
1679
1680 nfsrv_cleancache(); /* And clear out server cache */
1681
1682 nfsrv_udpsock = NULL;
6d2010ae 1683 nfsrv_udp6sock = NULL;
1c79356b 1684}
2d21ac55 1685
1c79356b 1686#endif /* NFS_NOSERVER */