nfs_clnode.c revision 302421
1/*-
2 * Copyright (c) 1989, 1993
3 *	The Regents of the University of California.  All rights reserved.
4 *
5 * This code is derived from software contributed to Berkeley by
6 * Rick Macklem at The University of Guelph.
7 *
8 * Redistribution and use in source and binary forms, with or without
9 * modification, are permitted provided that the following conditions
10 * are met:
11 * 1. Redistributions of source code must retain the above copyright
12 *    notice, this list of conditions and the following disclaimer.
13 * 2. Redistributions in binary form must reproduce the above copyright
14 *    notice, this list of conditions and the following disclaimer in the
15 *    documentation and/or other materials provided with the distribution.
16 * 4. Neither the name of the University nor the names of its contributors
17 *    may be used to endorse or promote products derived from this software
18 *    without specific prior written permission.
19 *
20 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
21 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
22 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
23 * ARE DISCLAIMED.  IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
24 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
25 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
26 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
27 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
28 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
29 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
30 * SUCH DAMAGE.
31 *
32 *	from nfs_node.c	8.6 (Berkeley) 5/22/95
33 */
34
35#include <sys/cdefs.h>
36__FBSDID("$FreeBSD: stable/10/sys/fs/nfsclient/nfs_clnode.c 302421 2016-07-08 02:34:04Z kib $");
37
38#include "opt_kdtrace.h"
39
40#include <sys/param.h>
41#include <sys/systm.h>
42#include <sys/fcntl.h>
43#include <sys/lock.h>
44#include <sys/malloc.h>
45#include <sys/mount.h>
46#include <sys/namei.h>
47#include <sys/proc.h>
48#include <sys/socket.h>
49#include <sys/sysctl.h>
50#include <sys/taskqueue.h>
51#include <sys/vnode.h>
52
53#include <vm/uma.h>
54
55#include <fs/nfs/nfsport.h>
56#include <fs/nfsclient/nfsnode.h>
57#include <fs/nfsclient/nfsmount.h>
58#include <fs/nfsclient/nfs.h>
59#include <fs/nfsclient/nfs_kdtrace.h>
60
61#include <nfs/nfs_lock.h>
62
63extern struct vop_vector newnfs_vnodeops;
64extern struct buf_ops buf_ops_newnfs;
65MALLOC_DECLARE(M_NEWNFSREQ);
66
67uma_zone_t newnfsnode_zone;
68
69const char nfs_vnode_tag[] = "newnfs";
70
71static void	nfs_freesillyrename(void *arg, __unused int pending);
72
73void
74ncl_nhinit(void)
75{
76
77	newnfsnode_zone = uma_zcreate("NCLNODE", sizeof(struct nfsnode), NULL,
78	    NULL, NULL, NULL, UMA_ALIGN_PTR, 0);
79}
80
81void
82ncl_nhuninit(void)
83{
84	uma_zdestroy(newnfsnode_zone);
85}
86
87/*
88 * ONLY USED FOR THE ROOT DIRECTORY. nfscl_nget() does the rest. If this
89 * function is going to be used to get Regular Files, code must be added
90 * to fill in the "struct nfsv4node".
91 * Look up a vnode/nfsnode by file handle.
92 * Callers must check for mount points!!
93 * In all cases, a pointer to a
94 * nfsnode structure is returned.
95 */
96int
97ncl_nget(struct mount *mntp, u_int8_t *fhp, int fhsize, struct nfsnode **npp,
98    int lkflags)
99{
100	struct thread *td = curthread;	/* XXX */
101	struct nfsnode *np;
102	struct vnode *vp;
103	struct vnode *nvp;
104	int error;
105	u_int hash;
106	struct nfsmount *nmp;
107	struct nfsfh *nfhp;
108
109	nmp = VFSTONFS(mntp);
110	*npp = NULL;
111
112	hash = fnv_32_buf(fhp, fhsize, FNV1_32_INIT);
113
114	MALLOC(nfhp, struct nfsfh *, sizeof (struct nfsfh) + fhsize,
115	    M_NFSFH, M_WAITOK);
116	bcopy(fhp, &nfhp->nfh_fh[0], fhsize);
117	nfhp->nfh_len = fhsize;
118	error = vfs_hash_get(mntp, hash, lkflags,
119	    td, &nvp, newnfs_vncmpf, nfhp);
120	FREE(nfhp, M_NFSFH);
121	if (error)
122		return (error);
123	if (nvp != NULL) {
124		*npp = VTONFS(nvp);
125		return (0);
126	}
127	np = uma_zalloc(newnfsnode_zone, M_WAITOK | M_ZERO);
128
129	error = getnewvnode(nfs_vnode_tag, mntp, &newnfs_vnodeops, &nvp);
130	if (error) {
131		uma_zfree(newnfsnode_zone, np);
132		return (error);
133	}
134	vp = nvp;
135	KASSERT(vp->v_bufobj.bo_bsize != 0, ("ncl_nget: bo_bsize == 0"));
136	vp->v_bufobj.bo_ops = &buf_ops_newnfs;
137	vp->v_data = np;
138	np->n_vnode = vp;
139	/*
140	 * Initialize the mutex even if the vnode is going to be a loser.
141	 * This simplifies the logic in reclaim, which can then unconditionally
142	 * destroy the mutex (in the case of the loser, or if hash_insert
143	 * happened to return an error no special casing is needed).
144	 */
145	mtx_init(&np->n_mtx, "NEWNFSnode lock", NULL, MTX_DEF | MTX_DUPOK);
146	/*
147	 * NFS supports recursive and shared locking.
148	 */
149	lockmgr(vp->v_vnlock, LK_EXCLUSIVE | LK_NOWITNESS, NULL);
150	VN_LOCK_AREC(vp);
151	VN_LOCK_ASHARE(vp);
152	/*
153	 * Are we getting the root? If so, make sure the vnode flags
154	 * are correct
155	 */
156	if ((fhsize == nmp->nm_fhsize) &&
157	    !bcmp(fhp, nmp->nm_fh, fhsize)) {
158		if (vp->v_type == VNON)
159			vp->v_type = VDIR;
160		vp->v_vflag |= VV_ROOT;
161	}
162
163	MALLOC(np->n_fhp, struct nfsfh *, sizeof (struct nfsfh) + fhsize,
164	    M_NFSFH, M_WAITOK);
165	bcopy(fhp, np->n_fhp->nfh_fh, fhsize);
166	np->n_fhp->nfh_len = fhsize;
167	error = insmntque(vp, mntp);
168	if (error != 0) {
169		*npp = NULL;
170		FREE((caddr_t)np->n_fhp, M_NFSFH);
171		mtx_destroy(&np->n_mtx);
172		uma_zfree(newnfsnode_zone, np);
173		return (error);
174	}
175	error = vfs_hash_insert(vp, hash, lkflags,
176	    td, &nvp, newnfs_vncmpf, np->n_fhp);
177	if (error)
178		return (error);
179	if (nvp != NULL) {
180		*npp = VTONFS(nvp);
181		/* vfs_hash_insert() vput()'s the losing vnode */
182		return (0);
183	}
184	*npp = np;
185
186	return (0);
187}
188
189/*
190 * Do the vrele(sp->s_dvp) as a separate task in order to avoid a
191 * deadlock because of a LOR when vrele() locks the directory vnode.
192 */
193static void
194nfs_freesillyrename(void *arg, __unused int pending)
195{
196	struct sillyrename *sp;
197
198	sp = arg;
199	vrele(sp->s_dvp);
200	free(sp, M_NEWNFSREQ);
201}
202
203static void
204ncl_releasesillyrename(struct vnode *vp, struct thread *td)
205{
206	struct nfsnode *np;
207	struct sillyrename *sp;
208
209	ASSERT_VOP_ELOCKED(vp, "releasesillyrename");
210	np = VTONFS(vp);
211	mtx_assert(&np->n_mtx, MA_OWNED);
212	if (vp->v_type != VDIR) {
213		sp = np->n_sillyrename;
214		np->n_sillyrename = NULL;
215	} else
216		sp = NULL;
217	if (sp != NULL) {
218		mtx_unlock(&np->n_mtx);
219		(void) ncl_vinvalbuf(vp, 0, td, 1);
220		/*
221		 * Remove the silly file that was rename'd earlier
222		 */
223		ncl_removeit(sp, vp);
224		crfree(sp->s_cred);
225		TASK_INIT(&sp->s_task, 0, nfs_freesillyrename, sp);
226		taskqueue_enqueue(taskqueue_thread, &sp->s_task);
227		mtx_lock(&np->n_mtx);
228	}
229}
230
231int
232ncl_inactive(struct vop_inactive_args *ap)
233{
234	struct vnode *vp = ap->a_vp;
235	struct nfsnode *np;
236	boolean_t retv;
237
238	if (NFS_ISV4(vp) && vp->v_type == VREG) {
239		/*
240		 * Since mmap()'d files do I/O after VOP_CLOSE(), the NFSv4
241		 * Close operations are delayed until now. Any dirty
242		 * buffers/pages must be flushed before the close, so that the
243		 * stateid is available for the writes.
244		 */
245		if (vp->v_object != NULL) {
246			VM_OBJECT_WLOCK(vp->v_object);
247			retv = vm_object_page_clean(vp->v_object, 0, 0,
248			    OBJPC_SYNC);
249			VM_OBJECT_WUNLOCK(vp->v_object);
250		} else
251			retv = TRUE;
252		if (retv == TRUE) {
253			(void)ncl_flush(vp, MNT_WAIT, NULL, ap->a_td, 1, 0);
254			(void)nfsrpc_close(vp, 1, ap->a_td);
255		}
256	}
257
258	np = VTONFS(vp);
259	mtx_lock(&np->n_mtx);
260	ncl_releasesillyrename(vp, ap->a_td);
261
262	/*
263	 * NMODIFIED means that there might be dirty/stale buffers
264	 * associated with the NFS vnode.  None of the other flags are
265	 * meaningful after the vnode is unused.
266	 */
267	np->n_flag &= NMODIFIED;
268	mtx_unlock(&np->n_mtx);
269	return (0);
270}
271
272/*
273 * Reclaim an nfsnode so that it can be used for other purposes.
274 */
275int
276ncl_reclaim(struct vop_reclaim_args *ap)
277{
278	struct vnode *vp = ap->a_vp;
279	struct nfsnode *np = VTONFS(vp);
280	struct nfsdmap *dp, *dp2;
281
282	/*
283	 * If the NLM is running, give it a chance to abort pending
284	 * locks.
285	 */
286	if (nfs_reclaim_p != NULL)
287		nfs_reclaim_p(ap);
288
289	mtx_lock(&np->n_mtx);
290	ncl_releasesillyrename(vp, ap->a_td);
291	mtx_unlock(&np->n_mtx);
292
293	/*
294	 * Destroy the vm object and flush associated pages.
295	 */
296	vnode_destroy_vobject(vp);
297
298	if (NFS_ISV4(vp) && vp->v_type == VREG)
299		/*
300		 * We can now safely close any remaining NFSv4 Opens for
301		 * this file. Most opens will have already been closed by
302		 * ncl_inactive(), but there are cases where it is not
303		 * called, so we need to do it again here.
304		 */
305		(void) nfsrpc_close(vp, 1, ap->a_td);
306
307	vfs_hash_remove(vp);
308
309	/*
310	 * Call nfscl_reclaimnode() to save attributes in the delegation,
311	 * as required.
312	 */
313	if (vp->v_type == VREG)
314		nfscl_reclaimnode(vp);
315
316	/*
317	 * Free up any directory cookie structures and
318	 * large file handle structures that might be associated with
319	 * this nfs node.
320	 */
321	if (vp->v_type == VDIR) {
322		dp = LIST_FIRST(&np->n_cookies);
323		while (dp) {
324			dp2 = dp;
325			dp = LIST_NEXT(dp, ndm_list);
326			FREE((caddr_t)dp2, M_NFSDIROFF);
327		}
328	}
329	if (np->n_writecred != NULL)
330		crfree(np->n_writecred);
331	FREE((caddr_t)np->n_fhp, M_NFSFH);
332	if (np->n_v4 != NULL)
333		FREE((caddr_t)np->n_v4, M_NFSV4NODE);
334	mtx_destroy(&np->n_mtx);
335	uma_zfree(newnfsnode_zone, vp->v_data);
336	vp->v_data = NULL;
337	return (0);
338}
339
340/*
341 * Invalidate both the access and attribute caches for this vnode.
342 */
343void
344ncl_invalcaches(struct vnode *vp)
345{
346	struct nfsnode *np = VTONFS(vp);
347	int i;
348
349	mtx_lock(&np->n_mtx);
350	for (i = 0; i < NFS_ACCESSCACHESIZE; i++)
351		np->n_accesscache[i].stamp = 0;
352	KDTRACE_NFS_ACCESSCACHE_FLUSH_DONE(vp);
353	np->n_attrstamp = 0;
354	KDTRACE_NFS_ATTRCACHE_FLUSH_DONE(vp);
355	mtx_unlock(&np->n_mtx);
356}
357