lfs_vnops.c revision 1.25.2.1.2.5 1 /* $NetBSD: lfs_vnops.c,v 1.25.2.1.2.5 1999/08/31 21:03:46 perseant Exp $ */
2
3 /*-
4 * Copyright (c) 1999 The NetBSD Foundation, Inc.
5 * All rights reserved.
6 *
7 * This code is derived from software contributed to The NetBSD Foundation
8 * by Konrad E. Schroder <perseant (at) hhhh.org>.
9 *
10 * Redistribution and use in source and binary forms, with or without
11 * modification, are permitted provided that the following conditions
12 * are met:
13 * 1. Redistributions of source code must retain the above copyright
14 * notice, this list of conditions and the following disclaimer.
15 * 2. Redistributions in binary form must reproduce the above copyright
16 * notice, this list of conditions and the following disclaimer in the
17 * documentation and/or other materials provided with the distribution.
18 * 3. All advertising materials mentioning features or use of this software
19 * must display the following acknowledgement:
20 * This product includes software developed by the NetBSD
21 * Foundation, Inc. and its contributors.
22 * 4. Neither the name of The NetBSD Foundation nor the names of its
23 * contributors may be used to endorse or promote products derived
24 * from this software without specific prior written permission.
25 *
26 * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
27 * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
28 * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
29 * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
30 * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
31 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
32 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
33 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
34 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
35 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
36 * POSSIBILITY OF SUCH DAMAGE.
37 */
38 /*
39 * Copyright (c) 1986, 1989, 1991, 1993, 1995
40 * The Regents of the University of California. All rights reserved.
41 *
42 * Redistribution and use in source and binary forms, with or without
43 * modification, are permitted provided that the following conditions
44 * are met:
45 * 1. Redistributions of source code must retain the above copyright
46 * notice, this list of conditions and the following disclaimer.
47 * 2. Redistributions in binary form must reproduce the above copyright
48 * notice, this list of conditions and the following disclaimer in the
49 * documentation and/or other materials provided with the distribution.
50 * 3. All advertising materials mentioning features or use of this software
51 * must display the following acknowledgement:
52 * This product includes software developed by the University of
53 * California, Berkeley and its contributors.
54 * 4. Neither the name of the University nor the names of its contributors
55 * may be used to endorse or promote products derived from this software
56 * without specific prior written permission.
57 *
58 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
59 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
60 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
61 * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
62 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
63 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
64 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
65 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
66 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
67 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
68 * SUCH DAMAGE.
69 *
70 * @(#)lfs_vnops.c 8.13 (Berkeley) 6/10/95
71 */
72
73 #include <sys/param.h>
74 #include <sys/systm.h>
75 #include <sys/namei.h>
76 #include <sys/resourcevar.h>
77 #include <sys/kernel.h>
78 #include <sys/file.h>
79 #include <sys/stat.h>
80 #include <sys/buf.h>
81 #include <sys/proc.h>
82 #include <sys/conf.h>
83 #include <sys/mount.h>
84 #include <sys/vnode.h>
85 #include <sys/malloc.h>
86 #include <sys/pool.h>
87 #include <sys/signalvar.h>
88
89 #include <vm/vm.h>
90 #include <uvm/uvm.h>
91 #include <uvm/uvm_pager.h> /* For PGO_* flags */
92
93 #include <miscfs/fifofs/fifo.h>
94 #include <miscfs/genfs/genfs.h>
95 #include <miscfs/specfs/specdev.h>
96
97 #include <ufs/ufs/quota.h>
98 #include <ufs/ufs/inode.h>
99 #include <ufs/ufs/dir.h>
100 #include <ufs/ufs/ufsmount.h>
101 #include <ufs/ufs/ufs_extern.h>
102
103 #include <ufs/lfs/lfs.h>
104 #include <ufs/lfs/lfs_extern.h>
105
106 /* Global vfs data structures for lfs. */
107 int (**lfs_vnodeop_p) __P((void *));
108 struct vnodeopv_entry_desc lfs_vnodeop_entries[] = {
109 { &vop_default_desc, vn_default_error },
110 { &vop_lookup_desc, ufs_lookup }, /* lookup */
111 { &vop_create_desc, lfs_create }, /* create */
112 { &vop_whiteout_desc, lfs_whiteout }, /* whiteout */
113 { &vop_mknod_desc, lfs_mknod }, /* mknod */
114 { &vop_open_desc, ufs_open }, /* open */
115 { &vop_close_desc, lfs_close }, /* close */
116 { &vop_access_desc, ufs_access }, /* access */
117 { &vop_getattr_desc, lfs_getattr }, /* getattr */
118 { &vop_setattr_desc, ufs_setattr }, /* setattr */
119 { &vop_read_desc, lfs_read }, /* read */
120 { &vop_write_desc, lfs_write }, /* write */
121 { &vop_lease_desc, ufs_lease_check }, /* lease */
122 { &vop_ioctl_desc, ufs_ioctl }, /* ioctl */
123 { &vop_poll_desc, ufs_poll }, /* poll */
124 { &vop_revoke_desc, ufs_revoke }, /* revoke */
125 { &vop_mmap_desc, ufs_mmap }, /* mmap */
126 { &vop_fsync_desc, lfs_fsync }, /* fsync */
127 { &vop_seek_desc, ufs_seek }, /* seek */
128 { &vop_remove_desc, lfs_remove }, /* remove */
129 { &vop_link_desc, lfs_link }, /* link */
130 { &vop_rename_desc, lfs_rename }, /* rename */
131 { &vop_mkdir_desc, lfs_mkdir }, /* mkdir */
132 { &vop_rmdir_desc, lfs_rmdir }, /* rmdir */
133 { &vop_symlink_desc, lfs_symlink }, /* symlink */
134 { &vop_readdir_desc, ufs_readdir }, /* readdir */
135 { &vop_readlink_desc, ufs_readlink }, /* readlink */
136 { &vop_abortop_desc, ufs_abortop }, /* abortop */
137 { &vop_inactive_desc, ufs_inactive }, /* inactive */
138 { &vop_reclaim_desc, lfs_reclaim }, /* reclaim */
139 { &vop_lock_desc, ufs_lock }, /* lock */
140 { &vop_unlock_desc, ufs_unlock }, /* unlock */
141 { &vop_bmap_desc, ufs_bmap }, /* bmap */
142 { &vop_strategy_desc, ufs_strategy }, /* strategy */
143 { &vop_print_desc, ufs_print }, /* print */
144 { &vop_islocked_desc, ufs_islocked }, /* islocked */
145 { &vop_pathconf_desc, ufs_pathconf }, /* pathconf */
146 { &vop_advlock_desc, ufs_advlock }, /* advlock */
147 { &vop_blkatoff_desc, lfs_blkatoff }, /* blkatoff */
148 { &vop_valloc_desc, lfs_valloc }, /* valloc */
149 { &vop_balloc_desc, lfs_balloc }, /* balloc */
150 { &vop_vfree_desc, lfs_vfree }, /* vfree */
151 { &vop_truncate_desc, lfs_truncate }, /* truncate */
152 { &vop_update_desc, lfs_update }, /* update */
153 { &vop_bwrite_desc, lfs_bwrite }, /* bwrite */
154 { &vop_getpages_desc, lfs_getpages }, /* getpages */
155 { &vop_putpages_desc, lfs_putpages }, /* putpages */
156 { (struct vnodeop_desc*)NULL, (int(*) __P((void *)))NULL }
157 };
158 struct vnodeopv_desc lfs_vnodeop_opv_desc =
159 { &lfs_vnodeop_p, lfs_vnodeop_entries };
160
161 int (**lfs_specop_p) __P((void *));
162 struct vnodeopv_entry_desc lfs_specop_entries[] = {
163 { &vop_default_desc, vn_default_error },
164 { &vop_lookup_desc, spec_lookup }, /* lookup */
165 { &vop_create_desc, spec_create }, /* create */
166 { &vop_mknod_desc, spec_mknod }, /* mknod */
167 { &vop_open_desc, spec_open }, /* open */
168 { &vop_close_desc, ufsspec_close }, /* close */
169 { &vop_access_desc, ufs_access }, /* access */
170 { &vop_getattr_desc, lfs_getattr }, /* getattr */
171 { &vop_setattr_desc, ufs_setattr }, /* setattr */
172 { &vop_read_desc, ufsspec_read }, /* read */
173 { &vop_write_desc, ufsspec_write }, /* write */
174 { &vop_lease_desc, spec_lease_check }, /* lease */
175 { &vop_ioctl_desc, spec_ioctl }, /* ioctl */
176 { &vop_poll_desc, spec_poll }, /* poll */
177 { &vop_revoke_desc, spec_revoke }, /* revoke */
178 { &vop_mmap_desc, spec_mmap }, /* mmap */
179 { &vop_fsync_desc, spec_fsync }, /* fsync */
180 { &vop_seek_desc, spec_seek }, /* seek */
181 { &vop_remove_desc, spec_remove }, /* remove */
182 { &vop_link_desc, spec_link }, /* link */
183 { &vop_rename_desc, spec_rename }, /* rename */
184 { &vop_mkdir_desc, spec_mkdir }, /* mkdir */
185 { &vop_rmdir_desc, spec_rmdir }, /* rmdir */
186 { &vop_symlink_desc, spec_symlink }, /* symlink */
187 { &vop_readdir_desc, spec_readdir }, /* readdir */
188 { &vop_readlink_desc, spec_readlink }, /* readlink */
189 { &vop_abortop_desc, spec_abortop }, /* abortop */
190 { &vop_inactive_desc, ufs_inactive }, /* inactive */
191 { &vop_reclaim_desc, lfs_reclaim }, /* reclaim */
192 { &vop_lock_desc, ufs_lock }, /* lock */
193 { &vop_unlock_desc, ufs_unlock }, /* unlock */
194 { &vop_bmap_desc, spec_bmap }, /* bmap */
195 { &vop_strategy_desc, spec_strategy }, /* strategy */
196 { &vop_print_desc, ufs_print }, /* print */
197 { &vop_islocked_desc, ufs_islocked }, /* islocked */
198 { &vop_pathconf_desc, spec_pathconf }, /* pathconf */
199 { &vop_advlock_desc, spec_advlock }, /* advlock */
200 { &vop_blkatoff_desc, spec_blkatoff }, /* blkatoff */
201 { &vop_valloc_desc, spec_valloc }, /* valloc */
202 { &vop_vfree_desc, lfs_vfree }, /* vfree */
203 { &vop_truncate_desc, spec_truncate }, /* truncate */
204 { &vop_update_desc, lfs_update }, /* update */
205 { &vop_bwrite_desc, lfs_bwrite }, /* bwrite */
206 { (struct vnodeop_desc*)NULL, (int(*) __P((void *)))NULL }
207 };
208 struct vnodeopv_desc lfs_specop_opv_desc =
209 { &lfs_specop_p, lfs_specop_entries };
210
211 int (**lfs_fifoop_p) __P((void *));
212 struct vnodeopv_entry_desc lfs_fifoop_entries[] = {
213 { &vop_default_desc, vn_default_error },
214 { &vop_lookup_desc, fifo_lookup }, /* lookup */
215 { &vop_create_desc, fifo_create }, /* create */
216 { &vop_mknod_desc, fifo_mknod }, /* mknod */
217 { &vop_open_desc, fifo_open }, /* open */
218 { &vop_close_desc, ufsfifo_close }, /* close */
219 { &vop_access_desc, ufs_access }, /* access */
220 { &vop_getattr_desc, lfs_getattr }, /* getattr */
221 { &vop_setattr_desc, ufs_setattr }, /* setattr */
222 { &vop_read_desc, ufsfifo_read }, /* read */
223 { &vop_write_desc, ufsfifo_write }, /* write */
224 { &vop_lease_desc, fifo_lease_check }, /* lease */
225 { &vop_ioctl_desc, fifo_ioctl }, /* ioctl */
226 { &vop_poll_desc, fifo_poll }, /* poll */
227 { &vop_revoke_desc, fifo_revoke }, /* revoke */
228 { &vop_mmap_desc, fifo_mmap }, /* mmap */
229 { &vop_fsync_desc, fifo_fsync }, /* fsync */
230 { &vop_seek_desc, fifo_seek }, /* seek */
231 { &vop_remove_desc, fifo_remove }, /* remove */
232 { &vop_link_desc, fifo_link }, /* link */
233 { &vop_rename_desc, fifo_rename }, /* rename */
234 { &vop_mkdir_desc, fifo_mkdir }, /* mkdir */
235 { &vop_rmdir_desc, fifo_rmdir }, /* rmdir */
236 { &vop_symlink_desc, fifo_symlink }, /* symlink */
237 { &vop_readdir_desc, fifo_readdir }, /* readdir */
238 { &vop_readlink_desc, fifo_readlink }, /* readlink */
239 { &vop_abortop_desc, fifo_abortop }, /* abortop */
240 { &vop_inactive_desc, ufs_inactive }, /* inactive */
241 { &vop_reclaim_desc, lfs_reclaim }, /* reclaim */
242 { &vop_lock_desc, ufs_lock }, /* lock */
243 { &vop_unlock_desc, ufs_unlock }, /* unlock */
244 { &vop_bmap_desc, fifo_bmap }, /* bmap */
245 { &vop_strategy_desc, fifo_strategy }, /* strategy */
246 { &vop_print_desc, ufs_print }, /* print */
247 { &vop_islocked_desc, ufs_islocked }, /* islocked */
248 { &vop_pathconf_desc, fifo_pathconf }, /* pathconf */
249 { &vop_advlock_desc, fifo_advlock }, /* advlock */
250 { &vop_blkatoff_desc, fifo_blkatoff }, /* blkatoff */
251 { &vop_valloc_desc, fifo_valloc }, /* valloc */
252 { &vop_vfree_desc, lfs_vfree }, /* vfree */
253 { &vop_truncate_desc, fifo_truncate }, /* truncate */
254 { &vop_update_desc, lfs_update }, /* update */
255 { &vop_bwrite_desc, lfs_bwrite }, /* bwrite */
256 { (struct vnodeop_desc*)NULL, (int(*) __P((void *)))NULL }
257 };
258 struct vnodeopv_desc lfs_fifoop_opv_desc =
259 { &lfs_fifoop_p, lfs_fifoop_entries };
260
261 #define LFS_READWRITE
262 #include <ufs/ufs/ufs_readwrite.c>
263 #undef LFS_READWRITE
264
265 /*
266 * Synch an open file.
267 */
268 /* ARGSUSED */
269 int
270 lfs_fsync(v)
271 void *v;
272 {
273 struct vop_fsync_args /* {
274 struct vnode *a_vp;
275 struct ucred *a_cred;
276 int a_flags;
277 struct proc *a_p;
278 } */ *ap = v;
279
280 return (VOP_UPDATE(ap->a_vp, NULL, NULL,
281 (ap->a_flags & FSYNC_WAIT) != 0 ? LFS_SYNC : 0)); /* XXX */
282 }
283
284 /*
285 * These macros are used to bracket UFS directory ops, so that we can
286 * identify all the pages touched during directory ops which need to
287 * be ordered and flushed atomically, so that they may be recovered.
288 */
289 /*
290 * XXX KS - Because we have to mark nodes VDIROP in order to prevent
291 * the cache from reclaiming them while a dirop is in progress, we must
292 * also manage the number of nodes so marked (otherwise we can run out).
293 * We do this by setting lfs_dirvcount to the number of marked vnodes; it
294 * is decremented during segment write, when VDIROP is taken off.
295 */
296 #define SET_DIROP(fs) lfs_set_dirop(fs)
297 static int lfs_set_dirop __P((struct lfs *));
298
299 static int lfs_set_dirop(fs)
300 struct lfs *fs;
301 {
302 int error;
303
304 while (fs->lfs_writer || fs->lfs_dirvcount>LFS_MAXDIROP) {
305 if(fs->lfs_writer)
306 tsleep(&fs->lfs_dirops, PRIBIO + 1, "lfs_dirop", 0);
307 if(fs->lfs_dirvcount > LFS_MAXDIROP) {
308 #ifdef DEBUG_LFS
309 printf("(dirvcount=%d)\n",fs->lfs_dirvcount);
310 #endif
311 if((error=tsleep(&fs->lfs_dirvcount, PCATCH|PUSER, "lfs_maxdirop", 0))!=0)
312 return error;
313 }
314 }
315 ++fs->lfs_dirops;
316 fs->lfs_doifile = 1;
317
318 return 0;
319 }
320
321 #define SET_ENDOP(fs,vp,str) { \
322 --(fs)->lfs_dirops; \
323 if (!(fs)->lfs_dirops) { \
324 wakeup(&(fs)->lfs_writer); \
325 lfs_check((vp),LFS_UNUSED_LBN,0); \
326 } \
327 }
328
329 #define MARK_VNODE(dvp) do { \
330 if(!((dvp)->v_flag & VDIROP)) { \
331 lfs_vref(dvp); \
332 ++VTOI((dvp))->i_lfs->lfs_dirvcount; \
333 } \
334 (dvp)->v_flag |= VDIROP; \
335 } while(0)
336
337 #define MAYBE_INACTIVE(fs,vp) do { \
338 if((vp) && ((vp)->v_flag & VDIROP) && (vp)->v_usecount == 1 \
339 && VTOI(vp) && VTOI(vp)->i_ffs_nlink == 0) \
340 { \
341 if (VOP_LOCK((vp), LK_EXCLUSIVE) == 0) { \
342 VOP_INACTIVE((vp),curproc); \
343 } \
344 } \
345 } while(0)
346
347 int
348 lfs_symlink(v)
349 void *v;
350 {
351 struct vop_symlink_args /* {
352 struct vnode *a_dvp;
353 struct vnode **a_vpp;
354 struct componentname *a_cnp;
355 struct vattr *a_vap;
356 char *a_target;
357 } */ *ap = v;
358 int ret;
359
360 if((ret=SET_DIROP(VTOI(ap->a_dvp)->i_lfs))!=0)
361 return ret;
362 MARK_VNODE(ap->a_dvp);
363 ret = ufs_symlink(ap);
364 MAYBE_INACTIVE(VTOI(ap->a_dvp)->i_lfs,*(ap->a_vpp)); /* XXX KS */
365 SET_ENDOP(VTOI(ap->a_dvp)->i_lfs,ap->a_dvp,"symilnk");
366 return (ret);
367 }
368
369 int
370 lfs_mknod(v)
371 void *v;
372 {
373 struct vop_mknod_args /* {
374 struct vnode *a_dvp;
375 struct vnode **a_vpp;
376 struct componentname *a_cnp;
377 struct vattr *a_vap;
378 } */ *ap = v;
379 int ret;
380
381 if((ret=SET_DIROP(VTOI(ap->a_dvp)->i_lfs))!=0)
382 return ret;
383 MARK_VNODE(ap->a_dvp);
384 ret = ufs_mknod(ap);
385 MAYBE_INACTIVE(VTOI(ap->a_dvp)->i_lfs,*(ap->a_vpp)); /* XXX KS */
386 SET_ENDOP(VTOI(ap->a_dvp)->i_lfs,ap->a_dvp,"mknod");
387 return (ret);
388 }
389
390 int
391 lfs_create(v)
392 void *v;
393 {
394 struct vop_create_args /* {
395 struct vnode *a_dvp;
396 struct vnode **a_vpp;
397 struct componentname *a_cnp;
398 struct vattr *a_vap;
399 } */ *ap = v;
400 int ret;
401
402 if((ret=SET_DIROP(VTOI(ap->a_dvp)->i_lfs))!=0)
403 return ret;
404 MARK_VNODE(ap->a_dvp);
405 ret = ufs_create(ap);
406 MAYBE_INACTIVE(VTOI(ap->a_dvp)->i_lfs,*(ap->a_vpp)); /* XXX KS */
407 SET_ENDOP(VTOI(ap->a_dvp)->i_lfs,ap->a_dvp,"create");
408 return (ret);
409 }
410
411 int
412 lfs_whiteout(v)
413 void *v;
414 {
415 struct vop_whiteout_args /* {
416 struct vnode *a_dvp;
417 struct componentname *a_cnp;
418 int a_flags;
419 } */ *ap = v;
420 int ret;
421
422 if((ret=SET_DIROP(VTOI(ap->a_dvp)->i_lfs))!=0)
423 return ret;
424 MARK_VNODE(ap->a_dvp);
425 ret = ufs_whiteout(ap);
426 SET_ENDOP(VTOI(ap->a_dvp)->i_lfs,ap->a_dvp,"whiteout");
427 return (ret);
428 }
429
430 int
431 lfs_mkdir(v)
432 void *v;
433 {
434 struct vop_mkdir_args /* {
435 struct vnode *a_dvp;
436 struct vnode **a_vpp;
437 struct componentname *a_cnp;
438 struct vattr *a_vap;
439 } */ *ap = v;
440 int ret;
441
442 if((ret=SET_DIROP(VTOI(ap->a_dvp)->i_lfs))!=0)
443 return ret;
444 MARK_VNODE(ap->a_dvp);
445 ret = ufs_mkdir(ap);
446 MAYBE_INACTIVE(VTOI(ap->a_dvp)->i_lfs,*(ap->a_vpp)); /* XXX KS */
447 SET_ENDOP(VTOI(ap->a_dvp)->i_lfs,ap->a_dvp,"mkdir");
448 return (ret);
449 }
450
451 int
452 lfs_remove(v)
453 void *v;
454 {
455 struct vop_remove_args /* {
456 struct vnode *a_dvp;
457 struct vnode *a_vp;
458 struct componentname *a_cnp;
459 } */ *ap = v;
460 int ret;
461 if((ret=SET_DIROP(VTOI(ap->a_dvp)->i_lfs))!=0)
462 return ret;
463 MARK_VNODE(ap->a_dvp);
464 MARK_VNODE(ap->a_vp);
465 ret = ufs_remove(ap);
466 MAYBE_INACTIVE(VTOI(ap->a_dvp)->i_lfs,ap->a_vp);
467 SET_ENDOP(VTOI(ap->a_dvp)->i_lfs,ap->a_dvp,"remove");
468 return (ret);
469 }
470
471 int
472 lfs_rmdir(v)
473 void *v;
474 {
475 struct vop_rmdir_args /* {
476 struct vnodeop_desc *a_desc;
477 struct vnode *a_dvp;
478 struct vnode *a_vp;
479 struct componentname *a_cnp;
480 } */ *ap = v;
481 int ret;
482
483 if((ret=SET_DIROP(VTOI(ap->a_dvp)->i_lfs))!=0)
484 return ret;
485 MARK_VNODE(ap->a_dvp);
486 MARK_VNODE(ap->a_vp);
487 ret = ufs_rmdir(ap);
488 MAYBE_INACTIVE(VTOI(ap->a_dvp)->i_lfs,ap->a_vp);
489 SET_ENDOP(VTOI(ap->a_dvp)->i_lfs,ap->a_dvp,"rmdir");
490 return (ret);
491 }
492
493 int
494 lfs_link(v)
495 void *v;
496 {
497 struct vop_link_args /* {
498 struct vnode *a_dvp;
499 struct vnode *a_vp;
500 struct componentname *a_cnp;
501 } */ *ap = v;
502 int ret;
503
504 if((ret=SET_DIROP(VTOI(ap->a_dvp)->i_lfs))!=0)
505 return ret;
506 MARK_VNODE(ap->a_dvp);
507 ret = ufs_link(ap);
508 SET_ENDOP(VTOI(ap->a_dvp)->i_lfs,ap->a_dvp,"link");
509 return (ret);
510 }
511
512 int
513 lfs_rename(v)
514 void *v;
515 {
516 struct vop_rename_args /* {
517 struct vnode *a_fdvp;
518 struct vnode *a_fvp;
519 struct componentname *a_fcnp;
520 struct vnode *a_tdvp;
521 struct vnode *a_tvp;
522 struct componentname *a_tcnp;
523 } */ *ap = v;
524 int ret;
525
526 if((ret=SET_DIROP(VTOI(ap->a_fdvp)->i_lfs))!=0)
527 return ret;
528 MARK_VNODE(ap->a_fdvp);
529 MARK_VNODE(ap->a_tdvp);
530 ret = ufs_rename(ap);
531 MAYBE_INACTIVE(VTOI(ap->a_dvp)->i_lfs,ap->a_fvp);
532 MAYBE_INACTIVE(VTOI(ap->a_dvp)->i_lfs,ap->a_tvp);
533 SET_ENDOP(VTOI(ap->a_fdvp)->i_lfs,ap->a_fdvp,"rename");
534 return (ret);
535 }
536
537 /* XXX hack to avoid calling ITIMES in getattr */
538 int
539 lfs_getattr(v)
540 void *v;
541 {
542 struct vop_getattr_args /* {
543 struct vnode *a_vp;
544 struct vattr *a_vap;
545 struct ucred *a_cred;
546 struct proc *a_p;
547 } */ *ap = v;
548 register struct vnode *vp = ap->a_vp;
549 register struct inode *ip = VTOI(vp);
550 register struct vattr *vap = ap->a_vap;
551 /*
552 * Copy from inode table
553 */
554 vap->va_fsid = ip->i_dev;
555 vap->va_fileid = ip->i_number;
556 vap->va_mode = ip->i_ffs_mode & ~IFMT;
557 vap->va_nlink = ip->i_ffs_nlink;
558 vap->va_uid = ip->i_ffs_uid;
559 vap->va_gid = ip->i_ffs_gid;
560 vap->va_rdev = (dev_t)ip->i_ffs_rdev;
561 vap->va_size = ip->i_ffs_size;
562 vap->va_atime.tv_sec = ip->i_ffs_atime;
563 vap->va_atime.tv_nsec = ip->i_ffs_atimensec;
564 vap->va_mtime.tv_sec = ip->i_ffs_mtime;
565 vap->va_mtime.tv_nsec = ip->i_ffs_mtimensec;
566 vap->va_ctime.tv_sec = ip->i_ffs_ctime;
567 vap->va_ctime.tv_nsec = ip->i_ffs_ctimensec;
568 vap->va_flags = ip->i_ffs_flags;
569 vap->va_gen = ip->i_ffs_gen;
570 /* this doesn't belong here */
571 if (vp->v_type == VBLK)
572 vap->va_blocksize = BLKDEV_IOSIZE;
573 else if (vp->v_type == VCHR)
574 vap->va_blocksize = MAXBSIZE;
575 else
576 vap->va_blocksize = vp->v_mount->mnt_stat.f_iosize;
577 vap->va_bytes = dbtob((u_quad_t)ip->i_ffs_blocks);
578 vap->va_type = vp->v_type;
579 vap->va_filerev = ip->i_modrev;
580 return (0);
581 }
582
583 /*
584 * Close called
585 *
586 * XXX -- we were using ufs_close, but since it updates the
587 * times on the inode, we might need to bump the uinodes
588 * count.
589 */
590 /* ARGSUSED */
591 int
592 lfs_close(v)
593 void *v;
594 {
595 struct vop_close_args /* {
596 struct vnode *a_vp;
597 int a_fflag;
598 struct ucred *a_cred;
599 struct proc *a_p;
600 } */ *ap = v;
601 register struct vnode *vp = ap->a_vp;
602 register struct inode *ip = VTOI(vp);
603 int mod;
604 struct timespec ts;
605
606 simple_lock(&vp->v_interlock);
607 if (vp->v_usecount > 1) {
608 mod = ip->i_flag & IN_MODIFIED;
609 TIMEVAL_TO_TIMESPEC(&time, &ts);
610 LFS_ITIMES(ip, &ts, &ts, &ts);
611 if (!mod && ip->i_flag & IN_MODIFIED)
612 ip->i_lfs->lfs_uinodes++;
613 }
614 simple_unlock(&vp->v_interlock);
615 return (0);
616 }
617
618 /*
619 * Reclaim an inode so that it can be used for other purposes.
620 */
621 int lfs_no_inactive = 0;
622
623 int
624 lfs_reclaim(v)
625 void *v;
626 {
627 struct vop_reclaim_args /* {
628 struct vnode *a_vp;
629 struct proc *a_p;
630 } */ *ap = v;
631 struct vnode *vp = ap->a_vp;
632 int error;
633
634 if ((error = ufs_reclaim(vp, ap->a_p)))
635 return (error);
636 pool_put(&lfs_inode_pool, vp->v_data);
637 vp->v_data = NULL;
638 return (0);
639 }
640
641 #define BUFHEAD_THROTTLE 128
642 static unsigned long bufhead_count, bufhead_rem;
643
644 static void
645 lfs_bufcallback(struct buf *bp)
646 {
647 int s;
648
649 s = splbio();
650 if(bp->b_vp)
651 brelvp(bp);
652 --bufhead_count;
653 splx(s);
654
655 uvm_aio_biodone1(bp); /* maybe set mbp->b_iodone = aio_aiodone */
656 return;
657 }
658
659 /*
660 * LFS-specific VM putpages routine.
661 * Encapsulate the pages in buffer headers, and call lfs_segwrite to write
662 * them out.
663 *
664 * Mostly taken from genfs_putpages.
665 *
666 * XXX Assumes that fs block size is <= page size, in order to avoid
667 * XXX having to read partial blocks in from disk.
668 *
669 * Ideally we would use pagemove() and not have this requirement
670 * (or just call lfs_segwrite with some notation of which pages
671 * were busy).
672 */
673
674 #define SET_FLUSHING(fs,vp) (fs)->lfs_flushvp = (vp)
675 #define IS_FLUSHING(fs,vp) ((fs)->lfs_flushvp == (vp))
676 #define CLR_FLUSHING(fs,vp) (fs)->lfs_flushvp = NULL
677
678 /* Statistics Counters */
679 extern int lfs_dostats;
680 extern struct lfs_stats lfs_stats;
681
682 /*
683 * Utility function for flushing a single vnode, in a case where
684 * lfs_segwrite might deadlock. Assumes that we hold the seglock.
685 * Excerpted from lfs_vflush. (XXX should be combined)
686 */
687 static inline void
688 lfs_writevnode(struct lfs *fs, struct vnode *vp)
689 {
690 struct segment *sp;
691 struct inode *ip;
692
693 ip = VTOI(vp);
694 sp = fs->lfs_sp;
695
696 do {
697 do {
698 if (vp->v_dirtyblkhd.lh_first != NULL)
699 lfs_writefile(fs, sp, vp);
700 } while (lfs_writeinode(fs, sp, ip));
701 } while (lfs_writeseg(fs, sp) && ip->i_number == LFS_IFILE_INUM);
702
703 if(lfs_dostats) {
704 ++lfs_stats.nwrites;
705 if (sp->seg_flags & SEGM_SYNC)
706 ++lfs_stats.nsync_writes;
707 if (sp->seg_flags & SEGM_CKP)
708 ++lfs_stats.ncheckpoints;
709 }
710 }
711
712 int
713 lfs_putpages(v)
714 void *v;
715 {
716 struct vop_putpages_args /* {
717 struct vnode *a_vp;
718 struct vm_page **a_m;
719 int a_count;
720 int a_flags;
721 int *a_rtvals;
722 } */ *ap = v;
723
724 int s, error, npages, bshift, dev_bshift, dev_bsize;
725 char * kva;
726 off_t offset, startoffset;
727 size_t bytes, iobytes, skipbytes;
728 daddr_t lbn;
729 struct vm_page *pg;
730 #if 0
731 struct vm_page *vmpp, *vmpp2;
732 #endif
733 struct buf *mbp, *bp;
734 struct vnode *vp = ap->a_vp;
735 struct lfs *fs;
736 struct inode *ip;
737
738 UVMHIST_FUNC("lfs_putpages"); UVMHIST_CALLED(ubchist);
739
740 ip = VTOI(vp);
741 fs = ip->i_lfs;
742
743 lfs_seglock(fs,((ap->a_flags & PGO_SYNCIO) ? SEGM_SYNC : 0));
744
745 /* If we're flushing, segwrite needs to know not to wait */
746 if(ap->a_flags & PGO_ALLPAGES) {
747 SET_FLUSHING(fs,vp);
748 }
749
750 error = 0;
751 npages = ap->a_count;
752 bshift = vp->v_mount->mnt_fs_bshift;
753 dev_bshift = vp->v_mount->mnt_dev_bshift;
754 dev_bsize = 1 << dev_bshift;
755
756 pg = ap->a_m[0];
757 startoffset = pg->offset;
758 bytes = npages << PAGE_SHIFT;
759 skipbytes = 0;
760
761 if (bytes == 0) {
762 panic("lfs_putpages: bytes == 0??? vp %p", vp);
763 }
764
765 #if 0
766 /*
767 * Fix start/end addresses to write entire blocks/fragments
768 * XXX This can deadlock us; can the pager do this klustering
769 * for us?
770 */
771 lbn = startoffset >> bshift;
772 if(startoffset != (lbn << bshift)) {
773 printf("Reading %d bytes at beginning of putpages\n",
774 startoffset - (lbn<<bshift));
775 /* Make sure the request is valid - XXX deadlock */
776 VOP_GETPAGES(vp, (lbn<<bshift), &vmpp, (startoffset - (lbn<<bshift)) >> pgshift, &newpages,
777 0, VM_PROT_DEFAULT, 0, 0);
778 bytes += (startoffset - (lbn << bshift));
779 startoffset = lbn << bshift;
780 pg = vmpp[0];
781 printf("Done reading beginning.\n");
782 }
783 if(bytes & fs->lfs_bmask) {
784 new_lastsize = blksize(fs, ip, startoffset + bytes);
785 diff = new_lastsize - (bytes & fs->lfs_bmask);
786 if(diff) {
787 printf("Reading %d bytes at end of putpages\n", diff);
788 /* Make sure request is valid - XXX deadlock */
789 VOP_GETPAGES(vp, startoffset+bytes, &vmpp2, diff, &newpages, 0,
790 VM_PROT_DEFAULT, 0, 0);
791 bytes += diff;
792 printf("Done reading end.\n");
793 }
794 }
795 #endif
796
797 kva = (void *)uvm_pagermapin(ap->a_m, npages, M_WAITOK);
798
799 s = splbio();
800 vp->v_numoutput++;
801 mbp = pool_get(&bufpool, PR_WAITOK);
802 UVMHIST_LOG(ubchist, "master vp %p bp %p num now %d",
803 vp, mbp, vp->v_numoutput, 0);
804 splx(s);
805 mbp->b_bufsize = npages << PAGE_SHIFT;
806 mbp->b_data = kva;
807 mbp->b_resid = mbp->b_bcount = bytes;
808 mbp->b_flags = B_BUSY|B_WRITE|(ap->a_flags & PGO_SYNCIO ? 0 : B_CALL)|
809 (curproc == uvm.pagedaemon_proc ? B_PDAEMON : 0);
810 mbp->b_iodone = uvm_aio_biodone;
811 mbp->b_vp = vp;
812
813 bp = NULL;
814
815 offset = startoffset;
816
817 for (; bytes > 0; offset += iobytes, bytes -= iobytes) {
818 lbn = offset >> bshift;
819 iobytes = min((1<<bshift), bytes);
820
821 if(bufhead_rem > BUFHEAD_THROTTLE) {
822 if(ap->a_flags & PGO_ALLPAGES)
823 lfs_writevnode(fs, vp);
824 else
825 lfs_segwrite(vp->v_mount, 0);
826 bufhead_rem = 0;
827 /* Don't sleep here, there's no point XXX ? */
828 }
829
830 ++bufhead_rem;
831 s = splbio();
832 ++bufhead_count;
833 /* vp->v_numoutput++; */
834 bp = pool_get(&bufpool, PR_WAITOK);
835 UVMHIST_LOG(ubchist, "vp %p bp %p num now %d",
836 vp, bp, vp->v_numoutput, 0);
837 memset(bp,0,sizeof(*bp));
838 bgetvp(vp, bp);
839 splx(s);
840
841 bp->b_data = kva + offset - pg->offset;
842 bp->b_resid = bp->b_bcount = bp->b_bufsize = iobytes;
843 bp->b_flags = B_BUSY|B_WRITE|B_CALL;
844 bp->b_iodone = lfs_bufcallback /* uvm_aio_biodone1 */ ;
845 bp->b_lblkno = bp->b_blkno = lbn;
846 bp->b_private = mbp;
847
848 #ifdef DIAGNOSTIC
849 if(offset != (lbn<<bshift)) {
850 printf("off=0x%lx, lbn<<bshift=0x%lx\n",
851 (unsigned long)offset,
852 (unsigned long)(lbn<<bshift));
853 panic("Oops, fs bsize is %u but page size is apparently %u\n", (unsigned)1<<bshift, (unsigned)bp->b_bcount);
854 }
855 if(iobytes != (1<<bshift)) {
856 printf("iobytes=0x%lx, 1<<bshift=0x%lx\n",
857 (unsigned long)iobytes,
858 (unsigned long)(1<<bshift));
859 panic("Oops, fs bsize is %u but page size is apparently %u\n", (unsigned)1<<bshift, (unsigned)bp->b_bcount);
860 }
861 #endif
862 VOP_BWRITE(bp);
863 }
864
865 /*
866 * The pager wants these buffers to be written right now (async may
867 * be okay, but definitely start right away).
868 *
869 * Call segwrite to start the write.
870 */
871 if (skipbytes) {
872 s = splbio();
873 mbp->b_resid -= skipbytes;
874 if (mbp->b_resid == 0)
875 biodone(mbp);
876 splx(s);
877 }
878 if(bufhead_rem) {
879 if(ap->a_flags & PGO_ALLPAGES)
880 lfs_writevnode(fs, vp);
881 else
882 lfs_segwrite(vp->v_mount,
883 ((ap->a_flags & PGO_SYNCIO) ?
884 SEGM_SYNC : 0));
885 bufhead_rem = 0;
886 if (!(ap->a_flags & PGO_SYNCIO)) {
887 CLR_FLUSHING(fs,vp);
888 lfs_segunlock(fs);
889 return EINPROGRESS;
890 }
891 }
892
893 /*
894 * By the time we reach this point, all buffers should be recovered,
895 * and the mbp should be waiting for aiodoned to reclaim it. We're writing
896 * synchronously, so we have to wait for this to happen.
897 */
898 /* errout: */
899 CLR_FLUSHING(fs,vp);
900 lfs_segunlock(fs);
901
902 if (bufhead_count) {
903 error = biowait(mbp);
904 }
905 s = splbio();
906 pool_put(&bufpool, mbp);
907 splx(s);
908
909 uvm_pagermapout((vaddr_t)kva, npages);
910 UVMHIST_LOG(ubchist, "returning, error %d", error,0,0,0);
911 return error;
912 }
913