nfs_node.c revision 1.101.10.4 1 1.101.10.4 yamt /* $NetBSD: nfs_node.c,v 1.101.10.4 2010/08/11 22:54:59 yamt Exp $ */
2 1.12 cgd
3 1.1 cgd /*
4 1.9 mycroft * Copyright (c) 1989, 1993
5 1.9 mycroft * The Regents of the University of California. All rights reserved.
6 1.1 cgd *
7 1.1 cgd * This code is derived from software contributed to Berkeley by
8 1.1 cgd * Rick Macklem at The University of Guelph.
9 1.1 cgd *
10 1.1 cgd * Redistribution and use in source and binary forms, with or without
11 1.1 cgd * modification, are permitted provided that the following conditions
12 1.1 cgd * are met:
13 1.1 cgd * 1. Redistributions of source code must retain the above copyright
14 1.1 cgd * notice, this list of conditions and the following disclaimer.
15 1.1 cgd * 2. Redistributions in binary form must reproduce the above copyright
16 1.1 cgd * notice, this list of conditions and the following disclaimer in the
17 1.1 cgd * documentation and/or other materials provided with the distribution.
18 1.70 agc * 3. Neither the name of the University nor the names of its contributors
19 1.1 cgd * may be used to endorse or promote products derived from this software
20 1.1 cgd * without specific prior written permission.
21 1.1 cgd *
22 1.1 cgd * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
23 1.1 cgd * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
24 1.1 cgd * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
25 1.1 cgd * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
26 1.1 cgd * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
27 1.1 cgd * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
28 1.1 cgd * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
29 1.1 cgd * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
30 1.1 cgd * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
31 1.1 cgd * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
32 1.1 cgd * SUCH DAMAGE.
33 1.1 cgd *
34 1.16 fvdl * @(#)nfs_node.c 8.6 (Berkeley) 5/22/95
35 1.1 cgd */
36 1.47 lukem
37 1.47 lukem #include <sys/cdefs.h>
38 1.101.10.4 yamt __KERNEL_RCSID(0, "$NetBSD: nfs_node.c,v 1.101.10.4 2010/08/11 22:54:59 yamt Exp $");
39 1.1 cgd
40 1.101.10.3 yamt #ifdef _KERNEL_OPT
41 1.35 bjh21 #include "opt_nfs.h"
42 1.101.10.3 yamt #endif
43 1.16 fvdl
44 1.4 mycroft #include <sys/param.h>
45 1.4 mycroft #include <sys/systm.h>
46 1.4 mycroft #include <sys/proc.h>
47 1.4 mycroft #include <sys/mount.h>
48 1.4 mycroft #include <sys/namei.h>
49 1.4 mycroft #include <sys/vnode.h>
50 1.4 mycroft #include <sys/kernel.h>
51 1.28 thorpej #include <sys/pool.h>
52 1.22 fvdl #include <sys/lock.h>
53 1.48 lukem #include <sys/hash.h>
54 1.84 elad #include <sys/kauth.h>
55 1.1 cgd
56 1.9 mycroft #include <nfs/rpcv2.h>
57 1.16 fvdl #include <nfs/nfsproto.h>
58 1.4 mycroft #include <nfs/nfs.h>
59 1.4 mycroft #include <nfs/nfsnode.h>
60 1.4 mycroft #include <nfs/nfsmount.h>
61 1.15 christos #include <nfs/nfs_var.h>
62 1.1 cgd
63 1.101.10.3 yamt struct pool nfs_node_pool;
64 1.101.10.3 yamt struct pool nfs_vattr_pool;
65 1.101.10.3 yamt static struct workqueue *nfs_sillyworkq;
66 1.28 thorpej
67 1.41 tsutsui extern int prtactive;
68 1.1 cgd
69 1.101.10.3 yamt static void nfs_gop_size(struct vnode *, off_t, off_t *, int);
70 1.101.10.3 yamt static int nfs_gop_alloc(struct vnode *, off_t, off_t, int, kauth_cred_t);
71 1.101.10.3 yamt static int nfs_gop_write(struct vnode *, struct vm_page **, int, int);
72 1.101.10.3 yamt static void nfs_sillyworker(struct work *, void *);
73 1.46 chs
74 1.80 yamt static const struct genfs_ops nfs_genfsops = {
75 1.80 yamt .gop_size = nfs_gop_size,
76 1.80 yamt .gop_alloc = nfs_gop_alloc,
77 1.80 yamt .gop_write = nfs_gop_write,
78 1.46 chs };
79 1.46 chs
80 1.1 cgd /*
81 1.101.10.3 yamt * Reinitialize inode hash table.
82 1.1 cgd */
83 1.15 christos void
84 1.101.10.3 yamt nfs_node_init(void)
85 1.1 cgd {
86 1.1 cgd
87 1.101.10.3 yamt pool_init(&nfs_node_pool, sizeof(struct nfsnode), 0, 0, 0, "nfsnodepl",
88 1.101.10.3 yamt &pool_allocator_nointr, IPL_NONE);
89 1.101.10.3 yamt pool_init(&nfs_vattr_pool, sizeof(struct vattr), 0, 0, 0, "nfsvapl",
90 1.101.10.3 yamt &pool_allocator_nointr, IPL_NONE);
91 1.101.10.3 yamt if (workqueue_create(&nfs_sillyworkq, "nfssilly", nfs_sillyworker,
92 1.101.10.3 yamt NULL, PRI_NONE, IPL_NONE, 0) != 0) {
93 1.101.10.3 yamt panic("nfs_node_init");
94 1.101.10.3 yamt }
95 1.31 jdolecek }
96 1.31 jdolecek
97 1.31 jdolecek /*
98 1.101.10.3 yamt * Free resources previously allocated in nfs_node_reinit().
99 1.45 chs */
100 1.45 chs void
101 1.101.10.3 yamt nfs_node_done(void)
102 1.45 chs {
103 1.45 chs
104 1.101.10.3 yamt pool_destroy(&nfs_node_pool);
105 1.101.10.3 yamt pool_destroy(&nfs_vattr_pool);
106 1.101.10.3 yamt workqueue_destroy(nfs_sillyworkq);
107 1.101.10.3 yamt }
108 1.79 perry
109 1.101.10.3 yamt #define RBTONFSNODE(node) \
110 1.101.10.3 yamt (void *)((uintptr_t)(node) - offsetof(struct nfsnode, n_rbnode))
111 1.101.10.3 yamt
112 1.101.10.3 yamt struct fh_match {
113 1.101.10.3 yamt nfsfh_t *fhm_fhp;
114 1.101.10.3 yamt size_t fhm_fhsize;
115 1.101.10.3 yamt size_t fhm_fhoffset;
116 1.101.10.3 yamt };
117 1.101.10.3 yamt
118 1.101.10.3 yamt static int
119 1.101.10.3 yamt nfs_compare_nodes(const struct rb_node *parent, const struct rb_node *node)
120 1.101.10.3 yamt {
121 1.101.10.3 yamt const struct nfsnode * const pnp = RBTONFSNODE(parent);
122 1.101.10.3 yamt const struct nfsnode * const np = RBTONFSNODE(node);
123 1.101.10.3 yamt
124 1.101.10.3 yamt if (pnp->n_fhsize != np->n_fhsize)
125 1.101.10.3 yamt return np->n_fhsize - pnp->n_fhsize;
126 1.101.10.3 yamt
127 1.101.10.3 yamt return memcmp(np->n_fhp, pnp->n_fhp, np->n_fhsize);
128 1.45 chs }
129 1.45 chs
130 1.101.10.3 yamt static int
131 1.101.10.3 yamt nfs_compare_node_fh(const struct rb_node *b, const void *key)
132 1.101.10.3 yamt {
133 1.101.10.3 yamt const struct nfsnode * const pnp = RBTONFSNODE(b);
134 1.101.10.3 yamt const struct fh_match * const fhm = key;
135 1.101.10.3 yamt
136 1.101.10.3 yamt if (pnp->n_fhsize != fhm->fhm_fhsize)
137 1.101.10.3 yamt return fhm->fhm_fhsize - pnp->n_fhsize;
138 1.101.10.3 yamt
139 1.101.10.3 yamt return memcmp(fhm->fhm_fhp, pnp->n_fhp, pnp->n_fhsize);
140 1.101.10.3 yamt }
141 1.101.10.3 yamt
142 1.101.10.3 yamt static const struct rb_tree_ops nfs_node_rbtree_ops = {
143 1.101.10.3 yamt .rbto_compare_nodes = nfs_compare_nodes,
144 1.101.10.3 yamt .rbto_compare_key = nfs_compare_node_fh,
145 1.101.10.3 yamt };
146 1.101.10.3 yamt
147 1.31 jdolecek void
148 1.101.10.3 yamt nfs_rbtinit(struct nfsmount *nmp)
149 1.31 jdolecek {
150 1.101.10.3 yamt rb_tree_init(&nmp->nm_rbtree, &nfs_node_rbtree_ops);
151 1.1 cgd }
152 1.1 cgd
153 1.101.10.3 yamt
154 1.1 cgd /*
155 1.1 cgd * Look up a vnode/nfsnode by file handle.
156 1.1 cgd * Callers must check for mount points!!
157 1.1 cgd * In all cases, a pointer to a
158 1.1 cgd * nfsnode structure is returned.
159 1.1 cgd */
160 1.15 christos int
161 1.101.10.3 yamt nfs_nget1(struct mount *mntp, nfsfh_t *fhp, int fhsize, struct nfsnode **npp, int lkflags)
162 1.1 cgd {
163 1.101.10.3 yamt struct nfsnode *np;
164 1.33 augustss struct vnode *vp;
165 1.101.10.3 yamt struct nfsmount *nmp = VFSTONFS(mntp);
166 1.1 cgd int error;
167 1.101.10.3 yamt struct fh_match fhm;
168 1.101.10.3 yamt struct rb_node *node;
169 1.101.10.3 yamt
170 1.101.10.3 yamt fhm.fhm_fhp = fhp;
171 1.101.10.3 yamt fhm.fhm_fhsize = fhsize;
172 1.1 cgd
173 1.1 cgd loop:
174 1.101.10.3 yamt rw_enter(&nmp->nm_rbtlock, RW_READER);
175 1.101.10.3 yamt node = rb_tree_find_node(&nmp->nm_rbtree, &fhm);
176 1.101.10.3 yamt if (node != NULL) {
177 1.101.10.3 yamt np = RBTONFSNODE(node);
178 1.1 cgd vp = NFSTOV(np);
179 1.100 ad mutex_enter(&vp->v_interlock);
180 1.101.10.3 yamt rw_exit(&nmp->nm_rbtlock);
181 1.101.10.4 yamt error = vget(vp, LK_EXCLUSIVE | lkflags);
182 1.75 yamt if (error == EBUSY)
183 1.75 yamt return error;
184 1.75 yamt if (error)
185 1.1 cgd goto loop;
186 1.1 cgd *npp = np;
187 1.1 cgd return(0);
188 1.1 cgd }
189 1.101.10.3 yamt rw_exit(&nmp->nm_rbtlock);
190 1.100 ad
191 1.73 yamt error = getnewvnode(VT_NFS, mntp, nfsv2_vnodeop_p, &vp);
192 1.15 christos if (error) {
193 1.1 cgd *npp = 0;
194 1.1 cgd return (error);
195 1.1 cgd }
196 1.28 thorpej np = pool_get(&nfs_node_pool, PR_WAITOK);
197 1.38 chs memset(np, 0, sizeof *np);
198 1.1 cgd np->n_vnode = vp;
199 1.38 chs
200 1.1 cgd /*
201 1.1 cgd * Insert the nfsnode in the hash queue for its new file handle
202 1.1 cgd */
203 1.46 chs
204 1.16 fvdl if (fhsize > NFS_SMALLFH) {
205 1.98 yamt np->n_fhp = kmem_alloc(fhsize, KM_SLEEP);
206 1.16 fvdl } else
207 1.16 fvdl np->n_fhp = &np->n_fh;
208 1.38 chs memcpy(np->n_fhp, fhp, fhsize);
209 1.16 fvdl np->n_fhsize = fhsize;
210 1.30 fvdl np->n_accstamp = -1;
211 1.28 thorpej np->n_vattr = pool_get(&nfs_vattr_pool, PR_WAITOK);
212 1.71 fvdl
213 1.101.10.3 yamt rw_enter(&nmp->nm_rbtlock, RW_WRITER);
214 1.101.10.3 yamt if (NULL != rb_tree_find_node(&nmp->nm_rbtree, &fhm)) {
215 1.101.10.3 yamt rw_exit(&nmp->nm_rbtlock);
216 1.100 ad if (fhsize > NFS_SMALLFH) {
217 1.100 ad kmem_free(np->n_fhp, fhsize);
218 1.100 ad }
219 1.100 ad pool_put(&nfs_vattr_pool, np->n_vattr);
220 1.100 ad pool_put(&nfs_node_pool, np);
221 1.100 ad ungetnewvnode(vp);
222 1.100 ad goto loop;
223 1.100 ad }
224 1.101.10.1 yamt mutex_init(&np->n_attrlock, MUTEX_DEFAULT, IPL_NONE);
225 1.100 ad vp->v_data = np;
226 1.100 ad genfs_node_init(vp, &nfs_genfsops);
227 1.71 fvdl /*
228 1.71 fvdl * Initalize read/write creds to useful values. VOP_OPEN will
229 1.71 fvdl * overwrite these.
230 1.71 fvdl */
231 1.85 ad np->n_rcred = curlwp->l_cred;
232 1.84 elad kauth_cred_hold(np->n_rcred);
233 1.85 ad np->n_wcred = curlwp->l_cred;
234 1.84 elad kauth_cred_hold(np->n_wcred);
235 1.101.10.4 yamt VOP_LOCK(vp, LK_EXCLUSIVE);
236 1.74 yamt NFS_INVALIDATE_ATTRCACHE(np);
237 1.74 yamt uvm_vnp_setsize(vp, 0);
238 1.101.10.3 yamt rb_tree_insert_node(&nmp->nm_rbtree, &np->n_rbnode);
239 1.101.10.3 yamt rw_exit(&nmp->nm_rbtlock);
240 1.100 ad
241 1.1 cgd *npp = np;
242 1.1 cgd return (0);
243 1.1 cgd }
244 1.1 cgd
245 1.15 christos int
246 1.101.10.3 yamt nfs_inactive(void *v)
247 1.15 christos {
248 1.9 mycroft struct vop_inactive_args /* {
249 1.9 mycroft struct vnode *a_vp;
250 1.97 ad bool *a_recycle;
251 1.15 christos } */ *ap = v;
252 1.33 augustss struct nfsnode *np;
253 1.33 augustss struct sillyrename *sp;
254 1.40 fvdl struct vnode *vp = ap->a_vp;
255 1.1 cgd
256 1.40 fvdl np = VTONFS(vp);
257 1.40 fvdl if (vp->v_type != VDIR) {
258 1.16 fvdl sp = np->n_sillyrename;
259 1.18 fvdl np->n_sillyrename = (struct sillyrename *)0;
260 1.18 fvdl } else
261 1.44 fvdl sp = NULL;
262 1.44 fvdl if (sp != NULL)
263 1.97 ad nfs_vinvalbuf(vp, 0, sp->s_cred, curlwp, 1);
264 1.97 ad *ap->a_recycle = (np->n_flag & NREMOVED) != 0;
265 1.94 yamt np->n_flag &=
266 1.94 yamt (NMODIFIED | NFLUSHINPROG | NFLUSHWANT | NEOFVALID | NTRUNCDELAYED);
267 1.76 yamt
268 1.76 yamt if (vp->v_type == VDIR && np->n_dircache)
269 1.78 yamt nfs_invaldircache(vp,
270 1.78 yamt NFS_INVALDIRCACHE_FORCE | NFS_INVALDIRCACHE_KEEPEOF);
271 1.76 yamt
272 1.101.10.4 yamt VOP_UNLOCK(vp);
273 1.76 yamt
274 1.44 fvdl if (sp != NULL) {
275 1.101.10.3 yamt workqueue_enqueue(nfs_sillyworkq, &sp->s_work, NULL);
276 1.1 cgd }
277 1.59 fvdl
278 1.1 cgd return (0);
279 1.1 cgd }
280 1.1 cgd
281 1.1 cgd /*
282 1.1 cgd * Reclaim an nfsnode so that it can be used for other purposes.
283 1.1 cgd */
284 1.15 christos int
285 1.101.10.3 yamt nfs_reclaim(void *v)
286 1.15 christos {
287 1.9 mycroft struct vop_reclaim_args /* {
288 1.9 mycroft struct vnode *a_vp;
289 1.15 christos } */ *ap = v;
290 1.33 augustss struct vnode *vp = ap->a_vp;
291 1.33 augustss struct nfsnode *np = VTONFS(vp);
292 1.101.10.3 yamt struct nfsmount *nmp = VFSTONFS(vp->v_mount);
293 1.1 cgd
294 1.99 ad if (prtactive && vp->v_usecount > 1)
295 1.1 cgd vprint("nfs_reclaim: pushing active", vp);
296 1.60 drochner
297 1.101.10.3 yamt rw_enter(&nmp->nm_rbtlock, RW_WRITER);
298 1.101.10.3 yamt rb_tree_remove_node(&nmp->nm_rbtree, &np->n_rbnode);
299 1.101.10.3 yamt rw_exit(&nmp->nm_rbtlock);
300 1.16 fvdl
301 1.16 fvdl /*
302 1.16 fvdl * Free up any directory cookie structures and
303 1.16 fvdl * large file handle structures that might be associated with
304 1.16 fvdl * this nfs node.
305 1.16 fvdl */
306 1.101.10.3 yamt if (vp->v_type == VDIR && np->n_dircache != NULL) {
307 1.101.10.3 yamt nfs_invaldircache(vp, NFS_INVALDIRCACHE_FORCE);
308 1.101.10.2 yamt hashdone(np->n_dircache, HASH_LIST, nfsdirhashmask);
309 1.101.10.3 yamt }
310 1.65 yamt KASSERT(np->n_dirgens == NULL);
311 1.59 fvdl
312 1.59 fvdl if (np->n_fhsize > NFS_SMALLFH)
313 1.98 yamt kmem_free(np->n_fhp, np->n_fhsize);
314 1.16 fvdl
315 1.28 thorpej pool_put(&nfs_vattr_pool, np->n_vattr);
316 1.59 fvdl if (np->n_rcred)
317 1.84 elad kauth_cred_free(np->n_rcred);
318 1.59 fvdl
319 1.59 fvdl if (np->n_wcred)
320 1.84 elad kauth_cred_free(np->n_wcred);
321 1.59 fvdl
322 1.1 cgd cache_purge(vp);
323 1.90 yamt if (vp->v_type == VREG) {
324 1.90 yamt mutex_destroy(&np->n_commitlock);
325 1.90 yamt }
326 1.101.10.1 yamt mutex_destroy(&np->n_attrlock);
327 1.91 ad genfs_node_destroy(vp);
328 1.90 yamt pool_put(&nfs_node_pool, np);
329 1.38 chs vp->v_data = NULL;
330 1.1 cgd return (0);
331 1.46 chs }
332 1.46 chs
333 1.46 chs void
334 1.87 yamt nfs_gop_size(struct vnode *vp, off_t size, off_t *eobp, int flags)
335 1.46 chs {
336 1.83 yamt
337 1.46 chs *eobp = MAX(size, vp->v_size);
338 1.46 chs }
339 1.46 chs
340 1.46 chs int
341 1.87 yamt nfs_gop_alloc(struct vnode *vp, off_t off, off_t len, int flags,
342 1.87 yamt kauth_cred_t cred)
343 1.46 chs {
344 1.87 yamt
345 1.46 chs return 0;
346 1.53 chs }
347 1.53 chs
348 1.53 chs int
349 1.53 chs nfs_gop_write(struct vnode *vp, struct vm_page **pgs, int npages, int flags)
350 1.53 chs {
351 1.53 chs int i;
352 1.53 chs
353 1.53 chs for (i = 0; i < npages; i++) {
354 1.53 chs pmap_page_protect(pgs[i], VM_PROT_READ);
355 1.53 chs }
356 1.53 chs return genfs_gop_write(vp, pgs, npages, flags);
357 1.1 cgd }
358 1.101.10.3 yamt
359 1.101.10.3 yamt /*
360 1.101.10.3 yamt * Remove a silly file that was rename'd earlier
361 1.101.10.3 yamt */
362 1.101.10.3 yamt static void
363 1.101.10.3 yamt nfs_sillyworker(struct work *work, void *arg)
364 1.101.10.3 yamt {
365 1.101.10.3 yamt struct sillyrename *sp;
366 1.101.10.3 yamt int error;
367 1.101.10.3 yamt
368 1.101.10.3 yamt sp = (struct sillyrename *)work;
369 1.101.10.3 yamt error = vn_lock(sp->s_dvp, LK_EXCLUSIVE);
370 1.101.10.3 yamt if (error || sp->s_dvp->v_data == NULL) {
371 1.101.10.3 yamt /* XXX should recover */
372 1.101.10.3 yamt printf("%s: vp=%p error=%d\n", __func__, sp->s_dvp, error);
373 1.101.10.3 yamt if (error == 0) {
374 1.101.10.3 yamt vput(sp->s_dvp);
375 1.101.10.3 yamt } else {
376 1.101.10.3 yamt vrele(sp->s_dvp);
377 1.101.10.3 yamt }
378 1.101.10.3 yamt } else {
379 1.101.10.3 yamt nfs_removeit(sp);
380 1.101.10.3 yamt vput(sp->s_dvp);
381 1.101.10.3 yamt }
382 1.101.10.3 yamt kauth_cred_free(sp->s_cred);
383 1.101.10.3 yamt kmem_free(sp, sizeof(*sp));
384 1.101.10.3 yamt }
385