rumpfs.c revision 1.132 1 /* $NetBSD: rumpfs.c,v 1.132 2015/01/03 16:30:32 hannken Exp $ */
2
3 /*
4 * Copyright (c) 2009, 2010, 2011 Antti Kantee. All Rights Reserved.
5 *
6 * Redistribution and use in source and binary forms, with or without
7 * modification, are permitted provided that the following conditions
8 * are met:
9 * 1. Redistributions of source code must retain the above copyright
10 * notice, this list of conditions and the following disclaimer.
11 * 2. Redistributions in binary form must reproduce the above copyright
12 * notice, this list of conditions and the following disclaimer in the
13 * documentation and/or other materials provided with the distribution.
14 *
15 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS
16 * OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
17 * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
18 * DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
19 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
20 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
21 * SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
22 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
23 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
24 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
25 * SUCH DAMAGE.
26 */
27
28 #include <sys/cdefs.h>
29 __KERNEL_RCSID(0, "$NetBSD: rumpfs.c,v 1.132 2015/01/03 16:30:32 hannken Exp $");
30
31 #include <sys/param.h>
32 #include <sys/atomic.h>
33 #include <sys/buf.h>
34 #include <sys/dirent.h>
35 #include <sys/errno.h>
36 #include <sys/filedesc.h>
37 #include <sys/fcntl.h>
38 #include <sys/kauth.h>
39 #include <sys/malloc.h>
40 #include <sys/module.h>
41 #include <sys/mount.h>
42 #include <sys/namei.h>
43 #include <sys/lock.h>
44 #include <sys/lockf.h>
45 #include <sys/queue.h>
46 #include <sys/stat.h>
47 #include <sys/syscallargs.h>
48 #include <sys/vnode.h>
49 #include <sys/unistd.h>
50
51 #include <miscfs/fifofs/fifo.h>
52 #include <miscfs/specfs/specdev.h>
53 #include <miscfs/genfs/genfs.h>
54 #include <miscfs/genfs/genfs_node.h>
55
56 #include <uvm/uvm_extern.h>
57
58 #include <rump/rumpuser.h>
59
60 #include "rump_private.h"
61 #include "rump_vfs_private.h"
62
63 static int rump_vop_lookup(void *);
64 static int rump_vop_getattr(void *);
65 static int rump_vop_setattr(void *);
66 static int rump_vop_mkdir(void *);
67 static int rump_vop_rmdir(void *);
68 static int rump_vop_remove(void *);
69 static int rump_vop_mknod(void *);
70 static int rump_vop_create(void *);
71 static int rump_vop_inactive(void *);
72 static int rump_vop_reclaim(void *);
73 static int rump_vop_success(void *);
74 static int rump_vop_readdir(void *);
75 static int rump_vop_spec(void *);
76 static int rump_vop_read(void *);
77 static int rump_vop_write(void *);
78 static int rump_vop_open(void *);
79 static int rump_vop_symlink(void *);
80 static int rump_vop_readlink(void *);
81 static int rump_vop_whiteout(void *);
82 static int rump_vop_pathconf(void *);
83 static int rump_vop_bmap(void *);
84 static int rump_vop_strategy(void *);
85 static int rump_vop_advlock(void *);
86 static int rump_vop_access(void *);
87
88 int (**fifo_vnodeop_p)(void *);
89 const struct vnodeopv_entry_desc fifo_vnodeop_entries[] = {
90 { &vop_default_desc, vn_default_error },
91 { &vop_putpages_desc, genfs_null_putpages },
92 { NULL, NULL }
93 };
94 const struct vnodeopv_desc fifo_vnodeop_opv_desc =
95 { &fifo_vnodeop_p, fifo_vnodeop_entries };
96
97 int (**rump_vnodeop_p)(void *);
98 const struct vnodeopv_entry_desc rump_vnodeop_entries[] = {
99 { &vop_default_desc, vn_default_error },
100 { &vop_lookup_desc, rump_vop_lookup },
101 { &vop_getattr_desc, rump_vop_getattr },
102 { &vop_setattr_desc, rump_vop_setattr },
103 { &vop_mkdir_desc, rump_vop_mkdir },
104 { &vop_rmdir_desc, rump_vop_rmdir },
105 { &vop_remove_desc, rump_vop_remove },
106 { &vop_mknod_desc, rump_vop_mknod },
107 { &vop_create_desc, rump_vop_create },
108 { &vop_symlink_desc, rump_vop_symlink },
109 { &vop_readlink_desc, rump_vop_readlink },
110 { &vop_access_desc, rump_vop_access },
111 { &vop_readdir_desc, rump_vop_readdir },
112 { &vop_read_desc, rump_vop_read },
113 { &vop_write_desc, rump_vop_write },
114 { &vop_open_desc, rump_vop_open },
115 { &vop_close_desc, genfs_nullop },
116 { &vop_seek_desc, genfs_seek },
117 { &vop_getpages_desc, genfs_getpages },
118 { &vop_putpages_desc, genfs_putpages },
119 { &vop_whiteout_desc, rump_vop_whiteout },
120 { &vop_fsync_desc, rump_vop_success },
121 { &vop_lock_desc, genfs_lock },
122 { &vop_unlock_desc, genfs_unlock },
123 { &vop_islocked_desc, genfs_islocked },
124 { &vop_inactive_desc, rump_vop_inactive },
125 { &vop_reclaim_desc, rump_vop_reclaim },
126 { &vop_link_desc, genfs_eopnotsupp },
127 { &vop_pathconf_desc, rump_vop_pathconf },
128 { &vop_bmap_desc, rump_vop_bmap },
129 { &vop_strategy_desc, rump_vop_strategy },
130 { &vop_advlock_desc, rump_vop_advlock },
131 { NULL, NULL }
132 };
133 const struct vnodeopv_desc rump_vnodeop_opv_desc =
134 { &rump_vnodeop_p, rump_vnodeop_entries };
135
136 int (**rump_specop_p)(void *);
137 const struct vnodeopv_entry_desc rump_specop_entries[] = {
138 { &vop_default_desc, rump_vop_spec },
139 { NULL, NULL }
140 };
141 const struct vnodeopv_desc rump_specop_opv_desc =
142 { &rump_specop_p, rump_specop_entries };
143
144 const struct vnodeopv_desc * const rump_opv_descs[] = {
145 &rump_vnodeop_opv_desc,
146 &rump_specop_opv_desc,
147 NULL
148 };
149
150 #define RUMPFS_WHITEOUT ((void *)-1)
151 #define RDENT_ISWHITEOUT(rdp) (rdp->rd_node == RUMPFS_WHITEOUT)
152 struct rumpfs_dent {
153 char *rd_name;
154 int rd_namelen;
155 struct rumpfs_node *rd_node;
156
157 LIST_ENTRY(rumpfs_dent) rd_entries;
158 };
159
160 struct genfs_ops rumpfs_genfsops = {
161 .gop_size = genfs_size,
162 .gop_write = genfs_gop_write,
163
164 /* optional */
165 .gop_alloc = NULL,
166 .gop_markupdate = NULL,
167 };
168
169 struct rumpfs_node {
170 struct genfs_node rn_gn;
171 struct vattr rn_va;
172 struct vnode *rn_vp;
173 char *rn_hostpath;
174 int rn_flags;
175 struct lockf *rn_lockf;
176
177 union {
178 struct { /* VREG */
179 int readfd;
180 int writefd;
181 uint64_t offset;
182 } reg;
183 struct {
184 void *data;
185 size_t dlen;
186 } reg_noet;
187 struct { /* VDIR */
188 LIST_HEAD(, rumpfs_dent) dents;
189 struct rumpfs_node *parent;
190 int flags;
191 } dir;
192 struct {
193 char *target;
194 size_t len;
195 } link;
196 } rn_u;
197 };
198 #define rn_readfd rn_u.reg.readfd
199 #define rn_writefd rn_u.reg.writefd
200 #define rn_offset rn_u.reg.offset
201 #define rn_data rn_u.reg_noet.data
202 #define rn_dlen rn_u.reg_noet.dlen
203 #define rn_dir rn_u.dir.dents
204 #define rn_parent rn_u.dir.parent
205 #define rn_linktarg rn_u.link.target
206 #define rn_linklen rn_u.link.len
207
208 #define RUMPNODE_CANRECLAIM 0x01
209 #define RUMPNODE_DIR_ET 0x02
210 #define RUMPNODE_DIR_ETSUBS 0x04
211 #define RUMPNODE_ET_PHONE_HOST 0x10
212
213 struct rumpfs_mount {
214 struct vnode *rfsmp_rvp;
215 };
216
217 #define INO_WHITEOUT 1
218 static int lastino = 2;
219 static kmutex_t reclock;
220
221 #define RUMPFS_DEFAULTMODE 0755
222 static void freedir(struct rumpfs_node *, struct componentname *);
223 static struct rumpfs_node *makeprivate(enum vtype, mode_t, dev_t, off_t, bool);
224
225 /*
226 * Extra Terrestrial stuff. We map a given key (pathname) to a file on
227 * the host FS. ET phones home only from the root node of rumpfs.
228 *
229 * When an etfs node is removed, a vnode potentially behind it is not
230 * immediately recycled.
231 */
232
233 struct etfs {
234 char et_key[MAXPATHLEN];
235 size_t et_keylen;
236 bool et_prefixkey;
237 bool et_removing;
238 devminor_t et_blkmin;
239
240 LIST_ENTRY(etfs) et_entries;
241
242 struct rumpfs_node *et_rn;
243 };
244 static kmutex_t etfs_lock;
245 static LIST_HEAD(, etfs) etfs_list = LIST_HEAD_INITIALIZER(etfs_list);
246
247 static enum vtype
248 ettype_to_vtype(enum rump_etfs_type et)
249 {
250 enum vtype vt;
251
252 switch (et) {
253 case RUMP_ETFS_REG:
254 vt = VREG;
255 break;
256 case RUMP_ETFS_BLK:
257 vt = VBLK;
258 break;
259 case RUMP_ETFS_CHR:
260 vt = VCHR;
261 break;
262 case RUMP_ETFS_DIR:
263 vt = VDIR;
264 break;
265 case RUMP_ETFS_DIR_SUBDIRS:
266 vt = VDIR;
267 break;
268 default:
269 panic("invalid et type: %d", et);
270 }
271
272 return vt;
273 }
274
275 static enum vtype
276 hft_to_vtype(int hft)
277 {
278 enum vtype vt;
279
280 switch (hft) {
281 case RUMPUSER_FT_OTHER:
282 vt = VNON;
283 break;
284 case RUMPUSER_FT_DIR:
285 vt = VDIR;
286 break;
287 case RUMPUSER_FT_REG:
288 vt = VREG;
289 break;
290 case RUMPUSER_FT_BLK:
291 vt = VBLK;
292 break;
293 case RUMPUSER_FT_CHR:
294 vt = VCHR;
295 break;
296 default:
297 vt = VNON;
298 break;
299 }
300
301 return vt;
302 }
303
304 static bool
305 etfs_find(const char *key, struct etfs **etp, bool forceprefix)
306 {
307 struct etfs *et;
308 size_t keylen = strlen(key);
309
310 KASSERT(mutex_owned(&etfs_lock));
311
312 LIST_FOREACH(et, &etfs_list, et_entries) {
313 if ((keylen == et->et_keylen || et->et_prefixkey || forceprefix)
314 && strncmp(key, et->et_key, et->et_keylen) == 0) {
315 if (etp)
316 *etp = et;
317 return true;
318 }
319 }
320
321 return false;
322 }
323
324 #define REGDIR(ftype) \
325 ((ftype) == RUMP_ETFS_DIR || (ftype) == RUMP_ETFS_DIR_SUBDIRS)
326 static int
327 etfsregister(const char *key, const char *hostpath,
328 enum rump_etfs_type ftype, uint64_t begin, uint64_t size)
329 {
330 char buf[9];
331 struct etfs *et;
332 struct rumpfs_node *rn;
333 uint64_t fsize;
334 dev_t rdev = NODEV;
335 devminor_t dmin = -1;
336 int hft, error;
337
338 if (key[0] != '/') {
339 return EINVAL;
340 }
341 while (key[0] == '/') {
342 key++;
343 }
344
345 if ((error = rumpuser_getfileinfo(hostpath, &fsize, &hft)) != 0)
346 return error;
347
348 /* etfs directory requires a directory on the host */
349 if (REGDIR(ftype)) {
350 if (hft != RUMPUSER_FT_DIR)
351 return ENOTDIR;
352 if (begin != 0)
353 return EISDIR;
354 if (size != RUMP_ETFS_SIZE_ENDOFF)
355 return EISDIR;
356 size = fsize;
357 } else {
358 if (begin > fsize)
359 return EINVAL;
360 if (size == RUMP_ETFS_SIZE_ENDOFF)
361 size = fsize - begin;
362 if (begin + size > fsize)
363 return EINVAL;
364 }
365
366 if (ftype == RUMP_ETFS_BLK || ftype == RUMP_ETFS_CHR) {
367 error = rumpblk_register(hostpath, &dmin, begin, size);
368 if (error != 0) {
369 return error;
370 }
371 rdev = makedev(RUMPBLK_DEVMAJOR, dmin);
372 }
373
374 et = kmem_alloc(sizeof(*et), KM_SLEEP);
375 strcpy(et->et_key, key);
376 et->et_keylen = strlen(et->et_key);
377 et->et_rn = rn = makeprivate(ettype_to_vtype(ftype), RUMPFS_DEFAULTMODE,
378 rdev, size, true);
379 et->et_removing = false;
380 et->et_blkmin = dmin;
381
382 rn->rn_flags |= RUMPNODE_ET_PHONE_HOST;
383
384 if (ftype == RUMP_ETFS_REG || REGDIR(ftype) || et->et_blkmin != -1) {
385 size_t len = strlen(hostpath)+1;
386
387 rn->rn_hostpath = malloc(len, M_TEMP, M_WAITOK | M_ZERO);
388 memcpy(rn->rn_hostpath, hostpath, len);
389 rn->rn_offset = begin;
390 }
391
392 if (REGDIR(ftype)) {
393 rn->rn_flags |= RUMPNODE_DIR_ET;
394 et->et_prefixkey = true;
395 } else {
396 et->et_prefixkey = false;
397 }
398
399 if (ftype == RUMP_ETFS_DIR_SUBDIRS)
400 rn->rn_flags |= RUMPNODE_DIR_ETSUBS;
401
402 mutex_enter(&etfs_lock);
403 if (etfs_find(key, NULL, REGDIR(ftype))) {
404 mutex_exit(&etfs_lock);
405 if (et->et_blkmin != -1)
406 rumpblk_deregister(hostpath);
407 if (et->et_rn->rn_hostpath != NULL)
408 free(et->et_rn->rn_hostpath, M_TEMP);
409 kmem_free(et->et_rn, sizeof(*et->et_rn));
410 kmem_free(et, sizeof(*et));
411 return EEXIST;
412 }
413 LIST_INSERT_HEAD(&etfs_list, et, et_entries);
414 mutex_exit(&etfs_lock);
415
416 if (ftype == RUMP_ETFS_BLK) {
417 format_bytes(buf, sizeof(buf), size);
418 aprint_verbose("/%s: hostpath %s (%s)\n", key, hostpath, buf);
419 }
420
421 return 0;
422 }
423 #undef REGDIR
424
425 /* remove etfs mapping. caller's responsibility to make sure it's not in use */
426 static int
427 etfsremove(const char *key)
428 {
429 struct etfs *et;
430 size_t keylen;
431 int rv __diagused;
432
433 if (key[0] != '/') {
434 return EINVAL;
435 }
436 while (key[0] == '/') {
437 key++;
438 }
439
440 keylen = strlen(key);
441
442 mutex_enter(&etfs_lock);
443 LIST_FOREACH(et, &etfs_list, et_entries) {
444 if (keylen == et->et_keylen && strcmp(et->et_key, key) == 0) {
445 if (et->et_removing)
446 et = NULL;
447 else
448 et->et_removing = true;
449 break;
450 }
451 }
452 mutex_exit(&etfs_lock);
453 if (!et)
454 return ENOENT;
455
456 /*
457 * ok, we know what we want to remove and have signalled there
458 * actually are men at work. first, unregister from rumpblk
459 */
460 if (et->et_blkmin != -1) {
461 rv = rumpblk_deregister(et->et_rn->rn_hostpath);
462 } else {
463 rv = 0;
464 }
465 KASSERT(rv == 0);
466
467 /* then do the actual removal */
468 mutex_enter(&etfs_lock);
469 LIST_REMOVE(et, et_entries);
470 mutex_exit(&etfs_lock);
471
472 /* node is unreachable, safe to nuke all device copies */
473 if (et->et_blkmin != -1) {
474 vdevgone(RUMPBLK_DEVMAJOR, et->et_blkmin, et->et_blkmin, VBLK);
475 } else {
476 struct vnode *vp;
477 struct mount *mp;
478 struct rumpfs_node *rn;
479
480 mutex_enter(&reclock);
481 if ((vp = et->et_rn->rn_vp) != NULL) {
482 mp = vp->v_mount;
483 rn = vp->v_data;
484 KASSERT(rn == et->et_rn);
485 } else {
486 mp = NULL;
487 }
488 mutex_exit(&reclock);
489 if (mp && vcache_get(mp, &rn, sizeof(rn), &vp) == 0)
490 vgone(vp);
491 }
492
493 if (et->et_rn->rn_hostpath != NULL)
494 free(et->et_rn->rn_hostpath, M_TEMP);
495 kmem_free(et->et_rn, sizeof(*et->et_rn));
496 kmem_free(et, sizeof(*et));
497
498 return 0;
499 }
500
501 /*
502 * rumpfs
503 */
504
505 static struct rumpfs_node *
506 makeprivate(enum vtype vt, mode_t mode, dev_t rdev, off_t size, bool et)
507 {
508 struct rumpfs_node *rn;
509 struct vattr *va;
510 struct timespec ts;
511
512 KASSERT((mode & ~ALLPERMS) == 0);
513 rn = kmem_zalloc(sizeof(*rn), KM_SLEEP);
514
515 switch (vt) {
516 case VDIR:
517 LIST_INIT(&rn->rn_dir);
518 break;
519 case VREG:
520 if (et) {
521 rn->rn_readfd = -1;
522 rn->rn_writefd = -1;
523 }
524 break;
525 default:
526 break;
527 }
528
529 nanotime(&ts);
530
531 va = &rn->rn_va;
532 va->va_type = vt;
533 va->va_mode = mode;
534 if (vt == VDIR)
535 va->va_nlink = 2;
536 else
537 va->va_nlink = 1;
538 va->va_uid = 0;
539 va->va_gid = 0;
540 va->va_fsid =
541 va->va_fileid = atomic_inc_uint_nv(&lastino);
542 va->va_size = size;
543 va->va_blocksize = 512;
544 va->va_atime = ts;
545 va->va_mtime = ts;
546 va->va_ctime = ts;
547 va->va_birthtime = ts;
548 va->va_gen = 0;
549 va->va_flags = 0;
550 va->va_rdev = rdev;
551 va->va_bytes = 512;
552 va->va_filerev = 0;
553 va->va_vaflags = 0;
554
555 return rn;
556 }
557
558 static void
559 makedir(struct rumpfs_node *rnd,
560 struct componentname *cnp, struct rumpfs_node *rn)
561 {
562 struct rumpfs_dent *rdent;
563
564 rdent = kmem_alloc(sizeof(*rdent), KM_SLEEP);
565 rdent->rd_name = kmem_alloc(cnp->cn_namelen+1, KM_SLEEP);
566 rdent->rd_node = rn;
567 strlcpy(rdent->rd_name, cnp->cn_nameptr, cnp->cn_namelen+1);
568 rdent->rd_namelen = strlen(rdent->rd_name);
569
570 if ((cnp->cn_flags & ISWHITEOUT) != 0) {
571 KASSERT((cnp->cn_flags & DOWHITEOUT) == 0);
572 freedir(rnd, cnp);
573 }
574 LIST_INSERT_HEAD(&rnd->rn_dir, rdent, rd_entries);
575 }
576
577 static void
578 freedir(struct rumpfs_node *rnd, struct componentname *cnp)
579 {
580 struct rumpfs_dent *rd = NULL;
581
582 LIST_FOREACH(rd, &rnd->rn_dir, rd_entries) {
583 if (rd->rd_namelen == cnp->cn_namelen &&
584 strncmp(rd->rd_name, cnp->cn_nameptr,
585 cnp->cn_namelen) == 0)
586 break;
587 }
588 if (rd == NULL)
589 panic("could not find directory entry: %s", cnp->cn_nameptr);
590
591 if (cnp->cn_flags & DOWHITEOUT) {
592 rd->rd_node = RUMPFS_WHITEOUT;
593 } else {
594 LIST_REMOVE(rd, rd_entries);
595 kmem_free(rd->rd_name, rd->rd_namelen+1);
596 kmem_free(rd, sizeof(*rd));
597 }
598 }
599
600 /*
601 * Simple lookup for rump file systems.
602 *
603 * uhm, this is twisted. C F C C, hope of C C F C looming
604 */
605 static int
606 rump_vop_lookup(void *v)
607 {
608 struct vop_lookup_v2_args /* {
609 struct vnode *a_dvp;
610 struct vnode **a_vpp;
611 struct componentname *a_cnp;
612 }; */ *ap = v;
613 struct componentname *cnp = ap->a_cnp;
614 struct vnode *dvp = ap->a_dvp;
615 struct vnode **vpp = ap->a_vpp;
616 struct rumpfs_node *rnd = dvp->v_data, *rn;
617 struct rumpfs_dent *rd = NULL;
618 struct etfs *et;
619 bool dotdot = (cnp->cn_flags & ISDOTDOT) != 0;
620 int rv = 0;
621 const char *cp;
622
623 *vpp = NULL;
624
625 rv = VOP_ACCESS(dvp, VEXEC, cnp->cn_cred);
626 if (rv)
627 return rv;
628
629 if ((cnp->cn_flags & ISLASTCN)
630 && (dvp->v_mount->mnt_flag & MNT_RDONLY)
631 && (cnp->cn_nameiop == DELETE || cnp->cn_nameiop == RENAME))
632 return EROFS;
633
634 /* check for dot, return directly if the case */
635 if (cnp->cn_namelen == 1 && cnp->cn_nameptr[0] == '.') {
636 vref(dvp);
637 *vpp = dvp;
638 return 0;
639 }
640
641 /* we don't do rename */
642 if (!(((cnp->cn_flags & ISLASTCN) == 0) || (cnp->cn_nameiop != RENAME)))
643 return EOPNOTSUPP;
644
645 /* check for etfs */
646 if (dvp == rootvnode &&
647 (cnp->cn_nameiop == LOOKUP || cnp->cn_nameiop == CREATE)) {
648 bool found;
649 mutex_enter(&etfs_lock);
650 found = etfs_find(cnp->cn_nameptr, &et, false);
651 mutex_exit(&etfs_lock);
652
653 if (found) {
654 rn = et->et_rn;
655 cnp->cn_consume += et->et_keylen - cnp->cn_namelen;
656 /*
657 * consume trailing slashes if any and clear
658 * REQUIREDIR if we consumed the full path.
659 */
660 cp = &cnp->cn_nameptr[cnp->cn_namelen];
661 cp += cnp->cn_consume;
662 KASSERT(*cp == '\0' || *cp == '/');
663 if (*cp == '\0' && rn->rn_va.va_type != VDIR)
664 cnp->cn_flags &= ~REQUIREDIR;
665 while (*cp++ == '/')
666 cnp->cn_consume++;
667 goto getvnode;
668 }
669 }
670
671 if (rnd->rn_flags & RUMPNODE_DIR_ET) {
672 uint64_t fsize;
673 char *newpath;
674 size_t newpathlen;
675 int hft, error;
676
677 if (dotdot)
678 return EOPNOTSUPP;
679
680 newpathlen = strlen(rnd->rn_hostpath) + 1 + cnp->cn_namelen + 1;
681 newpath = malloc(newpathlen, M_TEMP, M_WAITOK);
682
683 strlcpy(newpath, rnd->rn_hostpath, newpathlen);
684 strlcat(newpath, "/", newpathlen);
685 strlcat(newpath, cnp->cn_nameptr, newpathlen);
686
687 if ((error = rumpuser_getfileinfo(newpath, &fsize, &hft)) != 0){
688 free(newpath, M_TEMP);
689 return error;
690 }
691
692 /* allow only dirs and regular files */
693 if (hft != RUMPUSER_FT_REG && hft != RUMPUSER_FT_DIR) {
694 free(newpath, M_TEMP);
695 return ENOENT;
696 }
697
698 rn = makeprivate(hft_to_vtype(hft), RUMPFS_DEFAULTMODE,
699 NODEV, fsize, true);
700 rn->rn_flags |= RUMPNODE_CANRECLAIM;
701 if (rnd->rn_flags & RUMPNODE_DIR_ETSUBS) {
702 rn->rn_flags |= RUMPNODE_DIR_ET | RUMPNODE_DIR_ETSUBS;
703 rn->rn_flags |= RUMPNODE_ET_PHONE_HOST;
704 }
705 rn->rn_hostpath = newpath;
706
707 goto getvnode;
708 } else {
709 if (dotdot) {
710 if ((rn = rnd->rn_parent) != NULL)
711 goto getvnode;
712 } else {
713 LIST_FOREACH(rd, &rnd->rn_dir, rd_entries) {
714 if (rd->rd_namelen == cnp->cn_namelen &&
715 strncmp(rd->rd_name, cnp->cn_nameptr,
716 cnp->cn_namelen) == 0)
717 break;
718 }
719 }
720 }
721
722 if (!rd && ((cnp->cn_flags & ISLASTCN) == 0||cnp->cn_nameiop != CREATE))
723 return ENOENT;
724
725 if (!rd && (cnp->cn_flags & ISLASTCN) && cnp->cn_nameiop == CREATE) {
726 if (dvp->v_mount->mnt_flag & MNT_RDONLY)
727 return EROFS;
728 rv = VOP_ACCESS(dvp, VWRITE, cnp->cn_cred);
729 if (rv)
730 return rv;
731 return EJUSTRETURN;
732 }
733
734 if ((cnp->cn_flags & ISLASTCN) && cnp->cn_nameiop == DELETE) {
735 rv = VOP_ACCESS(dvp, VWRITE, cnp->cn_cred);
736 if (rv)
737 return rv;
738 }
739
740 if (RDENT_ISWHITEOUT(rd)) {
741 cnp->cn_flags |= ISWHITEOUT;
742 if ((cnp->cn_flags & ISLASTCN) && cnp->cn_nameiop == CREATE)
743 return EJUSTRETURN;
744 return ENOENT;
745 }
746
747 rn = rd->rd_node;
748
749 getvnode:
750 KASSERT(rn);
751 rv = vcache_get(dvp->v_mount, &rn, sizeof(rn), vpp);
752
753 return rv;
754 }
755
756 static int
757 rump_check_possible(struct vnode *vp, struct rumpfs_node *rnode,
758 mode_t mode)
759 {
760
761 if ((mode & VWRITE) == 0)
762 return 0;
763
764 switch (vp->v_type) {
765 case VDIR:
766 case VLNK:
767 case VREG:
768 break;
769 default:
770 /* special file is always writable. */
771 return 0;
772 }
773
774 return vp->v_mount->mnt_flag & MNT_RDONLY ? EROFS : 0;
775 }
776
777 static int
778 rump_check_permitted(struct vnode *vp, struct rumpfs_node *rnode,
779 mode_t mode, kauth_cred_t cred)
780 {
781 struct vattr *attr = &rnode->rn_va;
782
783 return kauth_authorize_vnode(cred, KAUTH_ACCESS_ACTION(mode,
784 vp->v_type, attr->va_mode), vp, NULL, genfs_can_access(vp->v_type,
785 attr->va_mode, attr->va_uid, attr->va_gid, mode, cred));
786 }
787
788 int
789 rump_vop_access(void *v)
790 {
791 struct vop_access_args /* {
792 const struct vnodeop_desc *a_desc;
793 struct vnode *a_vp;
794 int a_mode;
795 kauth_cred_t a_cred;
796 } */ *ap = v;
797 struct vnode *vp = ap->a_vp;
798 struct rumpfs_node *rn = vp->v_data;
799 int error;
800
801 error = rump_check_possible(vp, rn, ap->a_mode);
802 if (error)
803 return error;
804
805 error = rump_check_permitted(vp, rn, ap->a_mode, ap->a_cred);
806
807 return error;
808 }
809
810 static int
811 rump_vop_getattr(void *v)
812 {
813 struct vop_getattr_args /* {
814 struct vnode *a_vp;
815 struct vattr *a_vap;
816 kauth_cred_t a_cred;
817 } */ *ap = v;
818 struct vnode *vp = ap->a_vp;
819 struct rumpfs_node *rn = vp->v_data;
820 struct vattr *vap = ap->a_vap;
821
822 memcpy(vap, &rn->rn_va, sizeof(struct vattr));
823 vap->va_size = vp->v_size;
824 return 0;
825 }
826
827 static int
828 rump_vop_setattr(void *v)
829 {
830 struct vop_setattr_args /* {
831 struct vnode *a_vp;
832 struct vattr *a_vap;
833 kauth_cred_t a_cred;
834 } */ *ap = v;
835 struct vnode *vp = ap->a_vp;
836 struct vattr *vap = ap->a_vap;
837 struct rumpfs_node *rn = vp->v_data;
838 struct vattr *attr = &rn->rn_va;
839 kauth_cred_t cred = ap->a_cred;
840 int error;
841
842 #define CHANGED(a, t) (vap->a != (t)VNOVAL)
843 #define SETIFVAL(a,t) if (CHANGED(a, t)) rn->rn_va.a = vap->a
844 if (CHANGED(va_atime.tv_sec, time_t) ||
845 CHANGED(va_ctime.tv_sec, time_t) ||
846 CHANGED(va_mtime.tv_sec, time_t) ||
847 CHANGED(va_birthtime.tv_sec, time_t) ||
848 CHANGED(va_atime.tv_nsec, long) ||
849 CHANGED(va_ctime.tv_nsec, long) ||
850 CHANGED(va_mtime.tv_nsec, long) ||
851 CHANGED(va_birthtime.tv_nsec, long)) {
852 error = kauth_authorize_vnode(cred, KAUTH_VNODE_WRITE_TIMES, vp,
853 NULL, genfs_can_chtimes(vp, vap->va_vaflags, attr->va_uid,
854 cred));
855 if (error)
856 return error;
857 }
858
859 SETIFVAL(va_atime.tv_sec, time_t);
860 SETIFVAL(va_ctime.tv_sec, time_t);
861 SETIFVAL(va_mtime.tv_sec, time_t);
862 SETIFVAL(va_birthtime.tv_sec, time_t);
863 SETIFVAL(va_atime.tv_nsec, long);
864 SETIFVAL(va_ctime.tv_nsec, long);
865 SETIFVAL(va_mtime.tv_nsec, long);
866 SETIFVAL(va_birthtime.tv_nsec, long);
867
868 if (CHANGED(va_flags, u_long)) {
869 /* XXX Can we handle system flags here...? */
870 error = kauth_authorize_vnode(cred, KAUTH_VNODE_WRITE_FLAGS, vp,
871 NULL, genfs_can_chflags(cred, vp->v_type, attr->va_uid,
872 false));
873 if (error)
874 return error;
875 }
876
877 SETIFVAL(va_flags, u_long);
878 #undef SETIFVAL
879 #undef CHANGED
880
881 if (vap->va_uid != (uid_t)VNOVAL || vap->va_gid != (uid_t)VNOVAL) {
882 uid_t uid =
883 (vap->va_uid != (uid_t)VNOVAL) ? vap->va_uid : attr->va_uid;
884 gid_t gid =
885 (vap->va_gid != (gid_t)VNOVAL) ? vap->va_gid : attr->va_gid;
886 error = kauth_authorize_vnode(cred,
887 KAUTH_VNODE_CHANGE_OWNERSHIP, vp, NULL,
888 genfs_can_chown(cred, attr->va_uid, attr->va_gid, uid,
889 gid));
890 if (error)
891 return error;
892 attr->va_uid = uid;
893 attr->va_gid = gid;
894 }
895
896 if (vap->va_mode != (mode_t)VNOVAL) {
897 mode_t mode = vap->va_mode;
898 error = kauth_authorize_vnode(cred, KAUTH_VNODE_WRITE_SECURITY,
899 vp, NULL, genfs_can_chmod(vp->v_type, cred, attr->va_uid,
900 attr->va_gid, mode));
901 if (error)
902 return error;
903 attr->va_mode = mode;
904 }
905
906 if (vp->v_type == VREG &&
907 vap->va_size != VSIZENOTSET &&
908 vap->va_size != rn->rn_dlen &&
909 (rn->rn_flags & RUMPNODE_ET_PHONE_HOST) == 0) {
910 void *newdata;
911 size_t copylen, newlen;
912
913 newlen = vap->va_size;
914 newdata = rump_hypermalloc(newlen, 0, false, "rumpfs");
915 if (newdata == NULL)
916 return ENOSPC;
917
918 copylen = MIN(rn->rn_dlen, newlen);
919 memset(newdata, 0, newlen);
920 memcpy(newdata, rn->rn_data, copylen);
921 rump_hyperfree(rn->rn_data, rn->rn_dlen);
922
923 rn->rn_data = newdata;
924 rn->rn_dlen = newlen;
925 uvm_vnp_setsize(vp, newlen);
926 }
927 return 0;
928 }
929
930 static int
931 rump_vop_mkdir(void *v)
932 {
933 struct vop_mkdir_v3_args /* {
934 struct vnode *a_dvp;
935 struct vnode **a_vpp;
936 struct componentname *a_cnp;
937 struct vattr *a_vap;
938 }; */ *ap = v;
939 struct vnode *dvp = ap->a_dvp;
940 struct vnode **vpp = ap->a_vpp;
941 struct componentname *cnp = ap->a_cnp;
942 struct vattr *va = ap->a_vap;
943 struct rumpfs_node *rnd = dvp->v_data, *rn;
944 int rv = 0;
945
946 rn = makeprivate(VDIR, va->va_mode & ALLPERMS, NODEV, DEV_BSIZE, false);
947 if ((cnp->cn_flags & ISWHITEOUT) != 0)
948 rn->rn_va.va_flags |= UF_OPAQUE;
949 rn->rn_parent = rnd;
950 rv = vcache_get(dvp->v_mount, &rn, sizeof(rn), vpp);
951 if (rv)
952 return rv;
953
954 makedir(rnd, cnp, rn);
955
956 return rv;
957 }
958
959 static int
960 rump_vop_rmdir(void *v)
961 {
962 struct vop_rmdir_args /* {
963 struct vnode *a_dvp;
964 struct vnode *a_vp;
965 struct componentname *a_cnp;
966 }; */ *ap = v;
967 struct vnode *dvp = ap->a_dvp;
968 struct vnode *vp = ap->a_vp;
969 struct componentname *cnp = ap->a_cnp;
970 struct rumpfs_node *rnd = dvp->v_data;
971 struct rumpfs_node *rn = vp->v_data;
972 struct rumpfs_dent *rd;
973 int rv = 0;
974
975 LIST_FOREACH(rd, &rn->rn_dir, rd_entries) {
976 if (rd->rd_node != RUMPFS_WHITEOUT) {
977 rv = ENOTEMPTY;
978 goto out;
979 }
980 }
981 while ((rd = LIST_FIRST(&rn->rn_dir)) != NULL) {
982 KASSERT(rd->rd_node == RUMPFS_WHITEOUT);
983 LIST_REMOVE(rd, rd_entries);
984 kmem_free(rd->rd_name, rd->rd_namelen+1);
985 kmem_free(rd, sizeof(*rd));
986 }
987
988 freedir(rnd, cnp);
989 rn->rn_flags |= RUMPNODE_CANRECLAIM;
990 rn->rn_parent = NULL;
991
992 out:
993 vput(dvp);
994 vput(vp);
995
996 return rv;
997 }
998
999 static int
1000 rump_vop_remove(void *v)
1001 {
1002 struct vop_remove_args /* {
1003 struct vnode *a_dvp;
1004 struct vnode *a_vp;
1005 struct componentname *a_cnp;
1006 }; */ *ap = v;
1007 struct vnode *dvp = ap->a_dvp;
1008 struct vnode *vp = ap->a_vp;
1009 struct componentname *cnp = ap->a_cnp;
1010 struct rumpfs_node *rnd = dvp->v_data;
1011 struct rumpfs_node *rn = vp->v_data;
1012 int rv = 0;
1013
1014 if (rn->rn_flags & RUMPNODE_ET_PHONE_HOST)
1015 return EOPNOTSUPP;
1016
1017 freedir(rnd, cnp);
1018 rn->rn_flags |= RUMPNODE_CANRECLAIM;
1019
1020 vput(dvp);
1021 vput(vp);
1022
1023 return rv;
1024 }
1025
1026 static int
1027 rump_vop_mknod(void *v)
1028 {
1029 struct vop_mknod_v3_args /* {
1030 struct vnode *a_dvp;
1031 struct vnode **a_vpp;
1032 struct componentname *a_cnp;
1033 struct vattr *a_vap;
1034 }; */ *ap = v;
1035 struct vnode *dvp = ap->a_dvp;
1036 struct vnode **vpp = ap->a_vpp;
1037 struct componentname *cnp = ap->a_cnp;
1038 struct vattr *va = ap->a_vap;
1039 struct rumpfs_node *rnd = dvp->v_data, *rn;
1040 int rv;
1041
1042 rn = makeprivate(va->va_type, va->va_mode & ALLPERMS, va->va_rdev,
1043 DEV_BSIZE, false);
1044 if ((cnp->cn_flags & ISWHITEOUT) != 0)
1045 rn->rn_va.va_flags |= UF_OPAQUE;
1046 rv = vcache_get(dvp->v_mount, &rn, sizeof(rn), vpp);
1047 if (rv)
1048 return rv;
1049
1050 makedir(rnd, cnp, rn);
1051
1052 return rv;
1053 }
1054
1055 static int
1056 rump_vop_create(void *v)
1057 {
1058 struct vop_create_v3_args /* {
1059 struct vnode *a_dvp;
1060 struct vnode **a_vpp;
1061 struct componentname *a_cnp;
1062 struct vattr *a_vap;
1063 }; */ *ap = v;
1064 struct vnode *dvp = ap->a_dvp;
1065 struct vnode **vpp = ap->a_vpp;
1066 struct componentname *cnp = ap->a_cnp;
1067 struct vattr *va = ap->a_vap;
1068 struct rumpfs_node *rnd = dvp->v_data, *rn;
1069 off_t newsize;
1070 int rv;
1071
1072 newsize = va->va_type == VSOCK ? DEV_BSIZE : 0;
1073 rn = makeprivate(va->va_type, va->va_mode & ALLPERMS, NODEV,
1074 newsize, false);
1075 if ((cnp->cn_flags & ISWHITEOUT) != 0)
1076 rn->rn_va.va_flags |= UF_OPAQUE;
1077 rv = vcache_get(dvp->v_mount, &rn, sizeof(rn), vpp);
1078 if (rv)
1079 return rv;
1080
1081 makedir(rnd, cnp, rn);
1082
1083 return rv;
1084 }
1085
1086 static int
1087 rump_vop_symlink(void *v)
1088 {
1089 struct vop_symlink_v3_args /* {
1090 struct vnode *a_dvp;
1091 struct vnode **a_vpp;
1092 struct componentname *a_cnp;
1093 struct vattr *a_vap;
1094 char *a_target;
1095 }; */ *ap = v;
1096 struct vnode *dvp = ap->a_dvp;
1097 struct vnode **vpp = ap->a_vpp;
1098 struct componentname *cnp = ap->a_cnp;
1099 struct vattr *va = ap->a_vap;
1100 struct rumpfs_node *rnd = dvp->v_data, *rn;
1101 const char *target = ap->a_target;
1102 size_t linklen;
1103 int rv;
1104
1105 linklen = strlen(target);
1106 KASSERT(linklen < MAXPATHLEN);
1107 rn = makeprivate(VLNK, va->va_mode & ALLPERMS, NODEV, linklen, false);
1108 if ((cnp->cn_flags & ISWHITEOUT) != 0)
1109 rn->rn_va.va_flags |= UF_OPAQUE;
1110 rv = vcache_get(dvp->v_mount, &rn, sizeof(rn), vpp);
1111 if (rv)
1112 return rv;
1113
1114 makedir(rnd, cnp, rn);
1115
1116 KASSERT(linklen < MAXPATHLEN);
1117 rn->rn_linktarg = PNBUF_GET();
1118 rn->rn_linklen = linklen;
1119 strcpy(rn->rn_linktarg, target);
1120
1121 return rv;
1122 }
1123
1124 static int
1125 rump_vop_readlink(void *v)
1126 {
1127 struct vop_readlink_args /* {
1128 struct vnode *a_vp;
1129 struct uio *a_uio;
1130 kauth_cred_t a_cred;
1131 }; */ *ap = v;
1132 struct vnode *vp = ap->a_vp;
1133 struct rumpfs_node *rn = vp->v_data;
1134 struct uio *uio = ap->a_uio;
1135
1136 return uiomove(rn->rn_linktarg, rn->rn_linklen, uio);
1137 }
1138
1139 static int
1140 rump_vop_whiteout(void *v)
1141 {
1142 struct vop_whiteout_args /* {
1143 struct vnode *a_dvp;
1144 struct componentname *a_cnp;
1145 int a_flags;
1146 } */ *ap = v;
1147 struct vnode *dvp = ap->a_dvp;
1148 struct rumpfs_node *rnd = dvp->v_data;
1149 struct componentname *cnp = ap->a_cnp;
1150 int flags = ap->a_flags;
1151
1152 switch (flags) {
1153 case LOOKUP:
1154 break;
1155 case CREATE:
1156 makedir(rnd, cnp, RUMPFS_WHITEOUT);
1157 break;
1158 case DELETE:
1159 cnp->cn_flags &= ~DOWHITEOUT; /* cargo culting never fails ? */
1160 freedir(rnd, cnp);
1161 break;
1162 default:
1163 panic("unknown whiteout op %d", flags);
1164 }
1165
1166 return 0;
1167 }
1168
1169 static int
1170 rump_vop_open(void *v)
1171 {
1172 struct vop_open_args /* {
1173 struct vnode *a_vp;
1174 int a_mode;
1175 kauth_cred_t a_cred;
1176 } */ *ap = v;
1177 struct vnode *vp = ap->a_vp;
1178 struct rumpfs_node *rn = vp->v_data;
1179 int mode = ap->a_mode;
1180 int error = EINVAL;
1181
1182 if (vp->v_type != VREG || (rn->rn_flags & RUMPNODE_ET_PHONE_HOST) == 0)
1183 return 0;
1184
1185 if (mode & FREAD) {
1186 if (rn->rn_readfd != -1)
1187 return 0;
1188 error = rumpuser_open(rn->rn_hostpath,
1189 RUMPUSER_OPEN_RDONLY, &rn->rn_readfd);
1190 }
1191
1192 if (mode & FWRITE) {
1193 if (rn->rn_writefd != -1)
1194 return 0;
1195 error = rumpuser_open(rn->rn_hostpath,
1196 RUMPUSER_OPEN_WRONLY, &rn->rn_writefd);
1197 }
1198
1199 return error;
1200 }
1201
1202 /* simple readdir. even omits dotstuff and periods */
1203 static int
1204 rump_vop_readdir(void *v)
1205 {
1206 struct vop_readdir_args /* {
1207 struct vnode *a_vp;
1208 struct uio *a_uio;
1209 kauth_cred_t a_cred;
1210 int *a_eofflag;
1211 off_t **a_cookies;
1212 int *a_ncookies;
1213 } */ *ap = v;
1214 struct vnode *vp = ap->a_vp;
1215 struct uio *uio = ap->a_uio;
1216 struct rumpfs_node *rnd = vp->v_data;
1217 struct rumpfs_dent *rdent;
1218 struct dirent *dentp = NULL;
1219 unsigned i;
1220 int rv = 0;
1221
1222 /* seek to current entry */
1223 for (i = 0, rdent = LIST_FIRST(&rnd->rn_dir);
1224 (i < uio->uio_offset) && rdent;
1225 i++, rdent = LIST_NEXT(rdent, rd_entries))
1226 continue;
1227 if (!rdent)
1228 goto out;
1229
1230 /* copy entries */
1231 dentp = kmem_alloc(sizeof(*dentp), KM_SLEEP);
1232 for (; rdent && uio->uio_resid > 0;
1233 rdent = LIST_NEXT(rdent, rd_entries), i++) {
1234 strlcpy(dentp->d_name, rdent->rd_name, sizeof(dentp->d_name));
1235 dentp->d_namlen = strlen(dentp->d_name);
1236 dentp->d_reclen = _DIRENT_RECLEN(dentp, dentp->d_namlen);
1237
1238 if (__predict_false(RDENT_ISWHITEOUT(rdent))) {
1239 dentp->d_fileno = INO_WHITEOUT;
1240 dentp->d_type = DT_WHT;
1241 } else {
1242 dentp->d_fileno = rdent->rd_node->rn_va.va_fileid;
1243 dentp->d_type = vtype2dt(rdent->rd_node->rn_va.va_type);
1244 }
1245
1246 if (uio->uio_resid < dentp->d_reclen) {
1247 i--;
1248 break;
1249 }
1250
1251 rv = uiomove(dentp, dentp->d_reclen, uio);
1252 if (rv) {
1253 i--;
1254 break;
1255 }
1256 }
1257 kmem_free(dentp, sizeof(*dentp));
1258 dentp = NULL;
1259
1260 out:
1261 KASSERT(dentp == NULL);
1262 if (ap->a_cookies) {
1263 *ap->a_ncookies = 0;
1264 *ap->a_cookies = NULL;
1265 }
1266 if (rdent)
1267 *ap->a_eofflag = 0;
1268 else
1269 *ap->a_eofflag = 1;
1270 uio->uio_offset = i;
1271
1272 return rv;
1273 }
1274
1275 static int
1276 etread(struct rumpfs_node *rn, struct uio *uio)
1277 {
1278 struct rumpuser_iovec iov;
1279 uint8_t *buf;
1280 size_t bufsize, n;
1281 int error = 0;
1282
1283 bufsize = uio->uio_resid;
1284 if (bufsize == 0)
1285 return 0;
1286 buf = kmem_alloc(bufsize, KM_SLEEP);
1287
1288 iov.iov_base = buf;
1289 iov.iov_len = bufsize;
1290 if ((error = rumpuser_iovread(rn->rn_readfd, &iov, 1,
1291 uio->uio_offset + rn->rn_offset, &n)) == 0) {
1292 KASSERT(n <= bufsize);
1293 error = uiomove(buf, n, uio);
1294 }
1295
1296 kmem_free(buf, bufsize);
1297 return error;
1298 }
1299
1300 static int
1301 rump_vop_read(void *v)
1302 {
1303 struct vop_read_args /* {
1304 struct vnode *a_vp;
1305 struct uio *a_uio;
1306 int ioflags a_ioflag;
1307 kauth_cred_t a_cred;
1308 }; */ *ap = v;
1309 struct vnode *vp = ap->a_vp;
1310 struct rumpfs_node *rn = vp->v_data;
1311 struct uio *uio = ap->a_uio;
1312 const int advice = IO_ADV_DECODE(ap->a_ioflag);
1313 off_t chunk;
1314 int error = 0;
1315
1316 if (vp->v_type == VDIR)
1317 return EISDIR;
1318
1319 /* et op? */
1320 if (rn->rn_flags & RUMPNODE_ET_PHONE_HOST)
1321 return etread(rn, uio);
1322
1323 /* otherwise, it's off to ubc with us */
1324 while (uio->uio_resid > 0) {
1325 chunk = MIN(uio->uio_resid, (off_t)rn->rn_dlen-uio->uio_offset);
1326 if (chunk == 0)
1327 break;
1328 error = ubc_uiomove(&vp->v_uobj, uio, chunk, advice,
1329 UBC_READ | UBC_PARTIALOK | UBC_UNMAP_FLAG(vp));
1330 if (error)
1331 break;
1332 }
1333
1334 return error;
1335 }
1336
1337 static int
1338 etwrite(struct rumpfs_node *rn, struct uio *uio)
1339 {
1340 struct rumpuser_iovec iov;
1341 uint8_t *buf;
1342 size_t bufsize, n;
1343 int error = 0;
1344
1345 bufsize = uio->uio_resid;
1346 if (bufsize == 0)
1347 return 0;
1348 buf = kmem_alloc(bufsize, KM_SLEEP);
1349 error = uiomove(buf, bufsize, uio);
1350 if (error)
1351 goto out;
1352
1353 KASSERT(uio->uio_resid == 0);
1354 iov.iov_base = buf;
1355 iov.iov_len = bufsize;
1356 if ((error = rumpuser_iovwrite(rn->rn_writefd, &iov, 1,
1357 (uio->uio_offset-bufsize) + rn->rn_offset, &n)) == 0) {
1358 KASSERT(n <= bufsize);
1359 uio->uio_resid = bufsize - n;
1360 }
1361
1362 out:
1363 kmem_free(buf, bufsize);
1364 return error;
1365 }
1366
1367 static int
1368 rump_vop_write(void *v)
1369 {
1370 struct vop_write_args /* {
1371 struct vnode *a_vp;
1372 struct uio *a_uio;
1373 int ioflags a_ioflag;
1374 kauth_cred_t a_cred;
1375 }; */ *ap = v;
1376 struct vnode *vp = ap->a_vp;
1377 struct rumpfs_node *rn = vp->v_data;
1378 struct uio *uio = ap->a_uio;
1379 const int advice = IO_ADV_DECODE(ap->a_ioflag);
1380 void *olddata;
1381 size_t oldlen, newlen;
1382 off_t chunk;
1383 int error = 0;
1384 bool allocd = false;
1385
1386 if (ap->a_ioflag & IO_APPEND)
1387 uio->uio_offset = vp->v_size;
1388
1389 /* consult et? */
1390 if (rn->rn_flags & RUMPNODE_ET_PHONE_HOST)
1391 return etwrite(rn, uio);
1392
1393 /*
1394 * Otherwise, it's a case of ubcmove.
1395 */
1396
1397 /*
1398 * First, make sure we have enough storage.
1399 *
1400 * No, you don't need to tell me it's not very efficient.
1401 * No, it doesn't really support sparse files, just fakes it.
1402 */
1403 newlen = uio->uio_offset + uio->uio_resid;
1404 oldlen = 0; /* XXXgcc */
1405 olddata = NULL;
1406 if (rn->rn_dlen < newlen) {
1407 oldlen = rn->rn_dlen;
1408 olddata = rn->rn_data;
1409
1410 rn->rn_data = rump_hypermalloc(newlen, 0, false, "rumpfs");
1411 if (rn->rn_data == NULL)
1412 return ENOSPC;
1413 rn->rn_dlen = newlen;
1414 memset(rn->rn_data, 0, newlen);
1415 memcpy(rn->rn_data, olddata, oldlen);
1416 allocd = true;
1417 uvm_vnp_setsize(vp, newlen);
1418 }
1419
1420 /* ok, we have enough stooorage. write */
1421 while (uio->uio_resid > 0) {
1422 chunk = MIN(uio->uio_resid, (off_t)rn->rn_dlen-uio->uio_offset);
1423 if (chunk == 0)
1424 break;
1425 error = ubc_uiomove(&vp->v_uobj, uio, chunk, advice,
1426 UBC_WRITE | UBC_PARTIALOK | UBC_UNMAP_FLAG(vp));
1427 if (error)
1428 break;
1429 }
1430
1431 if (allocd) {
1432 if (error) {
1433 rump_hyperfree(rn->rn_data, newlen);
1434 rn->rn_data = olddata;
1435 rn->rn_dlen = oldlen;
1436 uvm_vnp_setsize(vp, oldlen);
1437 } else {
1438 rump_hyperfree(olddata, oldlen);
1439 }
1440 }
1441
1442 return error;
1443 }
1444
1445 static int
1446 rump_vop_bmap(void *v)
1447 {
1448 struct vop_bmap_args /* {
1449 struct vnode *a_vp;
1450 daddr_t a_bn;
1451 struct vnode **a_vpp;
1452 daddr_t *a_bnp;
1453 int *a_runp;
1454 } */ *ap = v;
1455
1456 /* 1:1 mapping */
1457 if (ap->a_vpp)
1458 *ap->a_vpp = ap->a_vp;
1459 if (ap->a_bnp)
1460 *ap->a_bnp = ap->a_bn;
1461 if (ap->a_runp)
1462 *ap->a_runp = 16;
1463
1464 return 0;
1465 }
1466
1467 static int
1468 rump_vop_strategy(void *v)
1469 {
1470 struct vop_strategy_args /* {
1471 struct vnode *a_vp;
1472 struct buf *a_bp;
1473 } */ *ap = v;
1474 struct vnode *vp = ap->a_vp;
1475 struct rumpfs_node *rn = vp->v_data;
1476 struct buf *bp = ap->a_bp;
1477 off_t copylen, copyoff;
1478 int error;
1479
1480 if (vp->v_type != VREG || rn->rn_flags & RUMPNODE_ET_PHONE_HOST) {
1481 error = EINVAL;
1482 goto out;
1483 }
1484
1485 copyoff = bp->b_blkno << DEV_BSHIFT;
1486 copylen = MIN(rn->rn_dlen - copyoff, bp->b_bcount);
1487 if (BUF_ISWRITE(bp)) {
1488 memcpy((uint8_t *)rn->rn_data + copyoff, bp->b_data, copylen);
1489 } else {
1490 memset((uint8_t*)bp->b_data + copylen, 0, bp->b_bcount-copylen);
1491 memcpy(bp->b_data, (uint8_t *)rn->rn_data + copyoff, copylen);
1492 }
1493 bp->b_resid = 0;
1494 error = 0;
1495
1496 out:
1497 bp->b_error = error;
1498 biodone(bp);
1499 return 0;
1500 }
1501
1502 static int
1503 rump_vop_pathconf(void *v)
1504 {
1505 struct vop_pathconf_args /* {
1506 struct vnode *a_vp;
1507 int a_name;
1508 register_t *a_retval;
1509 }; */ *ap = v;
1510 int name = ap->a_name;
1511 register_t *retval = ap->a_retval;
1512
1513 switch (name) {
1514 case _PC_LINK_MAX:
1515 *retval = LINK_MAX;
1516 return 0;
1517 case _PC_NAME_MAX:
1518 *retval = RUMPFS_MAXNAMLEN;
1519 return 0;
1520 case _PC_PATH_MAX:
1521 *retval = PATH_MAX;
1522 return 0;
1523 case _PC_PIPE_BUF:
1524 *retval = PIPE_BUF;
1525 return 0;
1526 case _PC_CHOWN_RESTRICTED:
1527 *retval = 1;
1528 return 0;
1529 case _PC_NO_TRUNC:
1530 *retval = 1;
1531 return 0;
1532 case _PC_SYNC_IO:
1533 *retval = 1;
1534 return 0;
1535 case _PC_FILESIZEBITS:
1536 *retval = 43; /* this one goes to 11 */
1537 return 0;
1538 case _PC_SYMLINK_MAX:
1539 *retval = MAXPATHLEN;
1540 return 0;
1541 case _PC_2_SYMLINKS:
1542 *retval = 1;
1543 return 0;
1544 default:
1545 return EINVAL;
1546 }
1547 }
1548
1549 static int
1550 rump_vop_success(void *v)
1551 {
1552
1553 return 0;
1554 }
1555
1556 static int
1557 rump_vop_inactive(void *v)
1558 {
1559 struct vop_inactive_args /* {
1560 struct vnode *a_vp;
1561 bool *a_recycle;
1562 } */ *ap = v;
1563 struct vnode *vp = ap->a_vp;
1564 struct rumpfs_node *rn = vp->v_data;
1565
1566 if (rn->rn_flags & RUMPNODE_ET_PHONE_HOST && vp->v_type == VREG) {
1567 if (rn->rn_readfd != -1) {
1568 rumpuser_close(rn->rn_readfd);
1569 rn->rn_readfd = -1;
1570 }
1571 if (rn->rn_writefd != -1) {
1572 rumpuser_close(rn->rn_writefd);
1573 rn->rn_writefd = -1;
1574 }
1575 }
1576 *ap->a_recycle = (rn->rn_flags & RUMPNODE_CANRECLAIM) ? true : false;
1577
1578 VOP_UNLOCK(vp);
1579 return 0;
1580 }
1581
1582 static int
1583 rump_vop_reclaim(void *v)
1584 {
1585 struct vop_reclaim_args /* {
1586 struct vnode *a_vp;
1587 } */ *ap = v;
1588 struct vnode *vp = ap->a_vp;
1589 struct rumpfs_node *rn = vp->v_data;
1590
1591 vcache_remove(vp->v_mount, &rn, sizeof(rn));
1592 mutex_enter(&reclock);
1593 rn->rn_vp = NULL;
1594 mutex_exit(&reclock);
1595 genfs_node_destroy(vp);
1596 vp->v_data = NULL;
1597
1598 if (rn->rn_flags & RUMPNODE_CANRECLAIM) {
1599 if (vp->v_type == VREG
1600 && (rn->rn_flags & RUMPNODE_ET_PHONE_HOST) == 0
1601 && rn->rn_data) {
1602 rump_hyperfree(rn->rn_data, rn->rn_dlen);
1603 rn->rn_data = NULL;
1604 }
1605
1606 if (vp->v_type == VLNK)
1607 PNBUF_PUT(rn->rn_linktarg);
1608 if (rn->rn_hostpath)
1609 free(rn->rn_hostpath, M_TEMP);
1610 kmem_free(rn, sizeof(*rn));
1611 }
1612
1613 return 0;
1614 }
1615
1616 static int
1617 rump_vop_spec(void *v)
1618 {
1619 struct vop_generic_args *ap = v;
1620 int (**opvec)(void *);
1621
1622 switch (ap->a_desc->vdesc_offset) {
1623 case VOP_ACCESS_DESCOFFSET:
1624 case VOP_GETATTR_DESCOFFSET:
1625 case VOP_SETATTR_DESCOFFSET:
1626 case VOP_LOCK_DESCOFFSET:
1627 case VOP_UNLOCK_DESCOFFSET:
1628 case VOP_ISLOCKED_DESCOFFSET:
1629 case VOP_RECLAIM_DESCOFFSET:
1630 opvec = rump_vnodeop_p;
1631 break;
1632 default:
1633 opvec = spec_vnodeop_p;
1634 break;
1635 }
1636
1637 return VOCALL(opvec, ap->a_desc->vdesc_offset, v);
1638 }
1639
1640 static int
1641 rump_vop_advlock(void *v)
1642 {
1643 struct vop_advlock_args /* {
1644 const struct vnodeop_desc *a_desc;
1645 struct vnode *a_vp;
1646 void *a_id;
1647 int a_op;
1648 struct flock *a_fl;
1649 int a_flags;
1650 } */ *ap = v;
1651 struct vnode *vp = ap->a_vp;
1652 struct rumpfs_node *rn = vp->v_data;
1653
1654 return lf_advlock(ap, &rn->rn_lockf, vp->v_size);
1655 }
1656
1657 /*
1658 * Begin vfs-level stuff
1659 */
1660
1661 VFS_PROTOS(rumpfs);
1662 struct vfsops rumpfs_vfsops = {
1663 .vfs_name = MOUNT_RUMPFS,
1664 .vfs_min_mount_data = 0,
1665 .vfs_mount = rumpfs_mount,
1666 .vfs_start = (void *)nullop,
1667 .vfs_unmount = rumpfs_unmount,
1668 .vfs_root = rumpfs_root,
1669 .vfs_quotactl = (void *)eopnotsupp,
1670 .vfs_statvfs = genfs_statvfs,
1671 .vfs_sync = (void *)nullop,
1672 .vfs_vget = rumpfs_vget,
1673 .vfs_loadvnode = rumpfs_loadvnode,
1674 .vfs_fhtovp = (void *)eopnotsupp,
1675 .vfs_vptofh = (void *)eopnotsupp,
1676 .vfs_init = rumpfs_init,
1677 .vfs_reinit = NULL,
1678 .vfs_done = rumpfs_done,
1679 .vfs_mountroot = rumpfs_mountroot,
1680 .vfs_snapshot = (void *)eopnotsupp,
1681 .vfs_extattrctl = (void *)eopnotsupp,
1682 .vfs_suspendctl = (void *)eopnotsupp,
1683 .vfs_renamelock_enter = genfs_renamelock_enter,
1684 .vfs_renamelock_exit = genfs_renamelock_exit,
1685 .vfs_opv_descs = rump_opv_descs,
1686 /* vfs_refcount */
1687 /* vfs_list */
1688 };
1689
1690 static int
1691 rumpfs_mountfs(struct mount *mp)
1692 {
1693 struct rumpfs_mount *rfsmp;
1694 struct rumpfs_node *rn;
1695 int error;
1696
1697 rfsmp = kmem_alloc(sizeof(*rfsmp), KM_SLEEP);
1698
1699 rn = makeprivate(VDIR, RUMPFS_DEFAULTMODE, NODEV, DEV_BSIZE, false);
1700 rn->rn_parent = rn;
1701 if ((error = vcache_get(mp, &rn, sizeof(rn), &rfsmp->rfsmp_rvp)) != 0) {
1702 kmem_free(rfsmp, sizeof(*rfsmp));
1703 return error;
1704 }
1705
1706 rfsmp->rfsmp_rvp->v_vflag |= VV_ROOT;
1707
1708 mp->mnt_data = rfsmp;
1709 mp->mnt_stat.f_namemax = RUMPFS_MAXNAMLEN;
1710 mp->mnt_stat.f_iosize = 512;
1711 mp->mnt_flag |= MNT_LOCAL;
1712 mp->mnt_iflag |= IMNT_MPSAFE | IMNT_CAN_RWTORO;
1713 mp->mnt_fs_bshift = DEV_BSHIFT;
1714 vfs_getnewfsid(mp);
1715
1716 return 0;
1717 }
1718
1719 int
1720 rumpfs_mount(struct mount *mp, const char *mntpath, void *arg, size_t *alen)
1721 {
1722 int error;
1723
1724 if (mp->mnt_flag & MNT_UPDATE) {
1725 return 0;
1726 }
1727
1728 error = set_statvfs_info(mntpath, UIO_USERSPACE, "rumpfs", UIO_SYSSPACE,
1729 mp->mnt_op->vfs_name, mp, curlwp);
1730 if (error)
1731 return error;
1732
1733 return rumpfs_mountfs(mp);
1734 }
1735
1736 int
1737 rumpfs_unmount(struct mount *mp, int mntflags)
1738 {
1739 struct rumpfs_mount *rfsmp = mp->mnt_data;
1740 int flags = 0, error;
1741
1742 if (panicstr || mntflags & MNT_FORCE)
1743 flags |= FORCECLOSE;
1744
1745 if (rfsmp->rfsmp_rvp->v_usecount > 1 && (flags & FORCECLOSE) == 0)
1746 return EBUSY;
1747
1748 if ((error = vflush(mp, rfsmp->rfsmp_rvp, flags)) != 0)
1749 return error;
1750 vgone(rfsmp->rfsmp_rvp);
1751
1752 kmem_free(rfsmp, sizeof(*rfsmp));
1753
1754 return 0;
1755 }
1756
1757 int
1758 rumpfs_root(struct mount *mp, struct vnode **vpp)
1759 {
1760 struct rumpfs_mount *rfsmp = mp->mnt_data;
1761
1762 vref(rfsmp->rfsmp_rvp);
1763 vn_lock(rfsmp->rfsmp_rvp, LK_EXCLUSIVE | LK_RETRY);
1764 *vpp = rfsmp->rfsmp_rvp;
1765 return 0;
1766 }
1767
1768 int
1769 rumpfs_vget(struct mount *mp, ino_t ino, struct vnode **vpp)
1770 {
1771
1772 return EOPNOTSUPP;
1773 }
1774
1775 int
1776 rumpfs_loadvnode(struct mount *mp, struct vnode *vp,
1777 const void *key, size_t key_len, const void **new_key)
1778 {
1779 struct rumpfs_node *rn;
1780 struct vattr *va;
1781
1782 KASSERT(!mutex_owned(&reclock));
1783
1784 KASSERT(key_len == sizeof(rn));
1785 memcpy(&rn, key, key_len);
1786
1787 va = &rn->rn_va;
1788
1789 vp->v_tag = VT_RUMP;
1790 vp->v_type = va->va_type;
1791 switch (vp->v_type) {
1792 case VCHR:
1793 case VBLK:
1794 vp->v_op = rump_specop_p;
1795 spec_node_init(vp, va->va_rdev);
1796 break;
1797 default:
1798 vp->v_op = rump_vnodeop_p;
1799 break;
1800 }
1801 vp->v_size = vp->v_writesize = va->va_size;
1802 vp->v_data = rn;
1803
1804 genfs_node_init(vp, &rumpfs_genfsops);
1805 mutex_enter(&reclock);
1806 rn->rn_vp = vp;
1807 mutex_exit(&reclock);
1808
1809 *new_key = &vp->v_data;
1810
1811 return 0;
1812 }
1813
1814 void
1815 rumpfs_init()
1816 {
1817 extern rump_etfs_register_withsize_fn rump__etfs_register;
1818 extern rump_etfs_remove_fn rump__etfs_remove;
1819 extern struct rump_boot_etfs *ebstart;
1820 struct rump_boot_etfs *eb;
1821
1822 CTASSERT(RUMP_ETFS_SIZE_ENDOFF == RUMPBLK_SIZENOTSET);
1823
1824 mutex_init(&reclock, MUTEX_DEFAULT, IPL_NONE);
1825 mutex_init(&etfs_lock, MUTEX_DEFAULT, IPL_NONE);
1826
1827 rump__etfs_register = etfsregister;
1828 rump__etfs_remove = etfsremove;
1829
1830 for (eb = ebstart; eb; eb = eb->_eb_next) {
1831 eb->eb_status = etfsregister(eb->eb_key, eb->eb_hostpath,
1832 eb->eb_type, eb->eb_begin, eb->eb_size);
1833 }
1834 }
1835
1836 void
1837 rumpfs_done()
1838 {
1839
1840 mutex_destroy(&reclock);
1841 mutex_destroy(&etfs_lock);
1842 }
1843
1844 int
1845 rumpfs_mountroot()
1846 {
1847 struct mount *mp;
1848 int error;
1849
1850 if ((error = vfs_rootmountalloc(MOUNT_RUMPFS, "rootdev", &mp)) != 0) {
1851 vrele(rootvp);
1852 return error;
1853 }
1854
1855 if ((error = rumpfs_mountfs(mp)) != 0)
1856 panic("mounting rootfs failed: %d", error);
1857
1858 mountlist_append(mp);
1859
1860 error = set_statvfs_info("/", UIO_SYSSPACE, "rumpfs", UIO_SYSSPACE,
1861 mp->mnt_op->vfs_name, mp, curlwp);
1862 if (error)
1863 panic("set_statvfs_info failed for rootfs: %d", error);
1864
1865 mp->mnt_flag &= ~MNT_RDONLY;
1866 vfs_unbusy(mp, false, NULL);
1867
1868 return 0;
1869 }
1870