cgd.c revision 1.108.2.11 1 /* $NetBSD: cgd.c,v 1.108.2.11 2016/07/23 03:20:37 pgoyette Exp $ */
2
3 /*-
4 * Copyright (c) 2002 The NetBSD Foundation, Inc.
5 * All rights reserved.
6 *
7 * This code is derived from software contributed to The NetBSD Foundation
8 * by Roland C. Dowdeswell.
9 *
10 * Redistribution and use in source and binary forms, with or without
11 * modification, are permitted provided that the following conditions
12 * are met:
13 * 1. Redistributions of source code must retain the above copyright
14 * notice, this list of conditions and the following disclaimer.
15 * 2. Redistributions in binary form must reproduce the above copyright
16 * notice, this list of conditions and the following disclaimer in the
17 * documentation and/or other materials provided with the distribution.
18 *
19 * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
20 * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
21 * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
22 * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
23 * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
24 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
25 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
26 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
27 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
28 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
29 * POSSIBILITY OF SUCH DAMAGE.
30 */
31
32 #include <sys/cdefs.h>
33 __KERNEL_RCSID(0, "$NetBSD: cgd.c,v 1.108.2.11 2016/07/23 03:20:37 pgoyette Exp $");
34
35 #include <sys/types.h>
36 #include <sys/param.h>
37 #include <sys/systm.h>
38 #include <sys/proc.h>
39 #include <sys/errno.h>
40 #include <sys/buf.h>
41 #include <sys/bufq.h>
42 #include <sys/malloc.h>
43 #include <sys/module.h>
44 #include <sys/pool.h>
45 #include <sys/ioctl.h>
46 #include <sys/device.h>
47 #include <sys/disk.h>
48 #include <sys/disklabel.h>
49 #include <sys/fcntl.h>
50 #include <sys/namei.h> /* for pathbuf */
51 #include <sys/vnode.h>
52 #include <sys/conf.h>
53 #include <sys/syslog.h>
54 #include <sys/localcount.h>
55
56 #include <dev/dkvar.h>
57 #include <dev/cgdvar.h>
58
59 #include <miscfs/specfs/specdev.h> /* for v_rdev */
60
61 #include "ioconf.h"
62
63 /* Entry Point Functions */
64
65 static dev_type_open(cgdopen);
66 static dev_type_close(cgdclose);
67 static dev_type_read(cgdread);
68 static dev_type_write(cgdwrite);
69 static dev_type_ioctl(cgdioctl);
70 static dev_type_strategy(cgdstrategy);
71 static dev_type_dump(cgddump);
72 static dev_type_size(cgdsize);
73
74 const struct bdevsw cgd_bdevsw = {
75 LOCALCOUNT_INITIALIZER
76 .d_open = cgdopen,
77 .d_close = cgdclose,
78 .d_strategy = cgdstrategy,
79 .d_ioctl = cgdioctl,
80 .d_dump = cgddump,
81 .d_psize = cgdsize,
82 .d_discard = nodiscard,
83 .d_flag = D_DISK
84 };
85
86 const struct cdevsw cgd_cdevsw = {
87 LOCALCOUNT_INITIALIZER
88 .d_open = cgdopen,
89 .d_close = cgdclose,
90 .d_read = cgdread,
91 .d_write = cgdwrite,
92 .d_ioctl = cgdioctl,
93 .d_stop = nostop,
94 .d_tty = notty,
95 .d_poll = nopoll,
96 .d_mmap = nommap,
97 .d_kqfilter = nokqfilter,
98 .d_discard = nodiscard,
99 .d_flag = D_DISK
100 };
101
102 static int cgd_match(device_t, cfdata_t, void *);
103 static void cgd_attach(device_t, device_t, void *);
104 static int cgd_detach(device_t, int);
105 static struct cgd_softc *cgd_spawn(int, device_t *);
106 static int cgd_destroy(device_t);
107
108 /* Internal Functions */
109
110 static int cgd_diskstart(device_t, struct buf *);
111 static void cgdiodone(struct buf *);
112 static int cgd_dumpblocks(device_t, void *, daddr_t, int);
113
114 static int cgd_ioctl_set(struct cgd_softc *, void *, struct lwp *);
115 static int cgd_ioctl_clr(struct cgd_softc *, struct lwp *);
116 static int cgd_ioctl_get(dev_t, void *, struct lwp *);
117 static int cgdinit(struct cgd_softc *, const char *, struct vnode *,
118 struct lwp *);
119 static void cgd_cipher(struct cgd_softc *, void *, void *,
120 size_t, daddr_t, size_t, int);
121
122 static struct dkdriver cgddkdriver = {
123 .d_minphys = minphys,
124 .d_open = cgdopen,
125 .d_close = cgdclose,
126 .d_strategy = cgdstrategy,
127 .d_iosize = NULL,
128 .d_diskstart = cgd_diskstart,
129 .d_dumpblocks = cgd_dumpblocks,
130 .d_lastclose = NULL
131 };
132
133 CFATTACH_DECL3_NEW(cgd, sizeof(struct cgd_softc),
134 cgd_match, cgd_attach, cgd_detach, NULL, NULL, NULL, DVF_DETACH_SHUTDOWN);
135 extern struct cfdriver cgd_cd;
136
137 /* DIAGNOSTIC and DEBUG definitions */
138
139 #if defined(CGDDEBUG) && !defined(DEBUG)
140 #define DEBUG
141 #endif
142
143 #ifdef DEBUG
144 int cgddebug = 0;
145
146 #define CGDB_FOLLOW 0x1
147 #define CGDB_IO 0x2
148 #define CGDB_CRYPTO 0x4
149
150 #define IFDEBUG(x,y) if (cgddebug & (x)) y
151 #define DPRINTF(x,y) IFDEBUG(x, printf y)
152 #define DPRINTF_FOLLOW(y) DPRINTF(CGDB_FOLLOW, y)
153
154 static void hexprint(const char *, void *, int);
155
156 #else
157 #define IFDEBUG(x,y)
158 #define DPRINTF(x,y)
159 #define DPRINTF_FOLLOW(y)
160 #endif
161
162 #ifdef DIAGNOSTIC
163 #define DIAGPANIC(x) panic x
164 #define DIAGCONDPANIC(x,y) if (x) panic y
165 #else
166 #define DIAGPANIC(x)
167 #define DIAGCONDPANIC(x,y)
168 #endif
169
170 /* Global variables */
171
172 /* Utility Functions */
173
174 #define CGDUNIT(x) DISKUNIT(x)
175 #define GETCGD_SOFTC(_cs, x, _dv) \
176 printf("%s: GETCGD_SOFTC\n", __func__); \
177 if (!((_cs) = getcgd_softc(x, &_dv))) { \
178 printf("%s: cs NULL\n", __func__); \
179 return ENXIO; \
180 }
181
182 /* The code */
183
184 static void
185 cgd_release(dev_t dev)
186 {
187 int unit = CGDUNIT(dev);
188 device_t self;
189
190 self = device_lookup_acquire(&cgd_cd, unit);
191 if (self != NULL)
192 device_release(self);
193 }
194
195 static struct cgd_softc *
196 getcgd_softc(dev_t dev, device_t *self)
197 {
198 int unit = CGDUNIT(dev);
199 struct cgd_softc *sc;
200
201 DPRINTF_FOLLOW(("getcgd_softc(0x%"PRIx64"): unit = %d\n", dev, unit));
202
203 *self = device_lookup_acquire(&cgd_cd, unit);
204
205 if (*self == NULL) {
206 sc = cgd_spawn(unit, self);
207 } else {
208 sc = device_private(*self);
209 }
210
211 return sc;
212 }
213
214 static int
215 cgd_match(device_t self, cfdata_t cfdata, void *aux)
216 {
217
218 return 1;
219 }
220
221 static void
222 cgd_attach(device_t parent, device_t self, void *aux)
223 {
224 struct cgd_softc *sc = device_private(self);
225
226 mutex_init(&sc->sc_lock, MUTEX_DEFAULT, IPL_BIO);
227 dk_init(&sc->sc_dksc, self, DKTYPE_CGD);
228 disk_init(&sc->sc_dksc.sc_dkdev, sc->sc_dksc.sc_xname, &cgddkdriver);
229
230 if (!pmf_device_register(self, NULL, NULL))
231 aprint_error_dev(self,
232 "unable to register power management hooks\n");
233 }
234
235
236 static int
237 cgd_detach(device_t self, int flags)
238 {
239 int ret;
240 const int pmask = 1 << RAW_PART;
241 struct cgd_softc *sc = device_private(self);
242 struct dk_softc *dksc = &sc->sc_dksc;
243
244 if (DK_BUSY(dksc, pmask))
245 return EBUSY;
246
247 if (DK_ATTACHED(dksc) &&
248 (ret = cgd_ioctl_clr(sc, curlwp)) != 0)
249 return ret;
250
251 disk_destroy(&dksc->sc_dkdev);
252 mutex_destroy(&sc->sc_lock);
253
254 return 0;
255 }
256
257 void
258 cgdattach(int num)
259 {
260 /*
261 * We don't need to do anything here - the config database is updated
262 * in module initialization code.
263
264 int error;
265
266 error = config_cfattach_attach(cgd_cd.cd_name, &cgd_ca);
267 if (error != 0)
268 aprint_error("%s: unable to register cfattach\n",
269 cgd_cd.cd_name);
270 *
271 */
272 }
273
274 static struct cgd_softc *
275 cgd_spawn(int unit, device_t *self)
276 {
277 cfdata_t cf;
278
279 cf = malloc(sizeof(*cf), M_DEVBUF, M_WAITOK);
280 cf->cf_name = cgd_cd.cd_name;
281 cf->cf_atname = cgd_cd.cd_name;
282 cf->cf_unit = unit;
283 cf->cf_fstate = FSTATE_STAR;
284
285 if (config_attach_pseudo(cf) == NULL)
286 return NULL;
287
288 *self = device_lookup_acquire(&cgd_cd, unit);
289 if (self == NULL)
290 return NULL;
291 else
292 /*
293 * Note that we return while still holding a reference
294 * to the device!
295 */
296 return device_private(*self);
297 }
298
299 static int
300 cgd_destroy(device_t dev)
301 {
302 int error;
303 cfdata_t cf;
304
305 cf = device_cfdata(dev);
306 error = config_detach(dev, DETACH_QUIET);
307 if (error)
308 return error;
309 free(cf, M_DEVBUF);
310 return 0;
311 }
312
313 static int
314 cgdopen(dev_t dev, int flags, int fmt, struct lwp *l)
315 {
316 device_t self;
317 int error;
318 struct cgd_softc *cs;
319
320 DPRINTF_FOLLOW(("cgdopen(0x%"PRIx64", %d)\n", dev, flags));
321 GETCGD_SOFTC(cs, dev, self);
322 error = dk_open(&cs->sc_dksc, dev, flags, fmt, l);
323 device_release(self);
324 return error;
325 }
326
327 static int
328 cgdclose(dev_t dev, int flags, int fmt, struct lwp *l)
329 {
330 int error;
331 device_t self;
332 struct cgd_softc *cs;
333 struct dk_softc *dksc;
334
335 DPRINTF_FOLLOW(("cgdclose(0x%"PRIx64", %d)\n", dev, flags));
336 GETCGD_SOFTC(cs, dev, self);
337 dksc = &cs->sc_dksc;
338 if ((error = dk_close(dksc, dev, flags, fmt, l)) != 0) {
339 device_release(self);
340 return error;
341 }
342
343 if (!DK_ATTACHED(dksc)) {
344 if ((error = cgd_destroy(cs->sc_dksc.sc_dev)) != 0) {
345 aprint_error_dev(dksc->sc_dev,
346 "unable to detach instance\n");
347 device_release(self);
348 return error;
349 }
350 }
351 device_release(self);
352 return error;
353 }
354
355 static void
356 cgdstrategy(struct buf *bp)
357 {
358 device_t self;
359 struct cgd_softc *cs = getcgd_softc(bp->b_dev, &self);
360 struct dk_softc *dksc = &cs->sc_dksc;
361 struct disk_geom *dg = &dksc->sc_dkdev.dk_geom;
362
363 DPRINTF_FOLLOW(("cgdstrategy(%p): b_bcount = %ld\n", bp,
364 (long)bp->b_bcount));
365
366 /*
367 * Reject unaligned writes. We can encrypt and decrypt only
368 * complete disk sectors, and we let the ciphers require their
369 * buffers to be aligned to 32-bit boundaries.
370 */
371 if (bp->b_blkno < 0 ||
372 (bp->b_bcount % dg->dg_secsize) != 0 ||
373 ((uintptr_t)bp->b_data & 3) != 0) {
374 bp->b_error = EINVAL;
375 bp->b_resid = bp->b_bcount;
376 biodone(bp);
377 cgd_release(bp->b_dev);
378 device_release(self);
379 return;
380 }
381
382 /* XXXrcd: Should we test for (cs != NULL)? */
383 dk_strategy(&cs->sc_dksc, bp);
384 cgd_release(bp->b_dev);
385 device_release(self);
386 return;
387 }
388
389 static int
390 cgdsize(dev_t dev)
391 {
392 int retval;
393 device_t self;
394 struct cgd_softc *cs = getcgd_softc(dev, &self);
395
396 DPRINTF_FOLLOW(("cgdsize(0x%"PRIx64")\n", dev));
397 if (!cs)
398 retval = -1;
399 else
400 retval = dk_size(&cs->sc_dksc, dev);
401
402 cgd_release(dev);
403 device_release(self);
404 return retval;
405 }
406
407 /*
408 * cgd_{get,put}data are functions that deal with getting a buffer
409 * for the new encrypted data. We have a buffer per device so that
410 * we can ensure that we can always have a transaction in flight.
411 * We use this buffer first so that we have one less piece of
412 * malloc'ed data at any given point.
413 */
414
415 static void *
416 cgd_getdata(struct dk_softc *dksc, unsigned long size)
417 {
418 struct cgd_softc *cs = (struct cgd_softc *)dksc;
419 void * data = NULL;
420
421 mutex_enter(&cs->sc_lock);
422 if (cs->sc_data_used == 0) {
423 cs->sc_data_used = 1;
424 data = cs->sc_data;
425 }
426 mutex_exit(&cs->sc_lock);
427
428 if (data)
429 return data;
430
431 return malloc(size, M_DEVBUF, M_NOWAIT);
432 }
433
434 static void
435 cgd_putdata(struct dk_softc *dksc, void *data)
436 {
437 struct cgd_softc *cs = (struct cgd_softc *)dksc;
438
439 if (data == cs->sc_data) {
440 mutex_enter(&cs->sc_lock);
441 cs->sc_data_used = 0;
442 mutex_exit(&cs->sc_lock);
443 } else {
444 free(data, M_DEVBUF);
445 }
446 }
447
448 static int
449 cgd_diskstart(device_t dev, struct buf *bp)
450 {
451 struct cgd_softc *cs = device_private(dev);
452 struct dk_softc *dksc = &cs->sc_dksc;
453 struct disk_geom *dg = &dksc->sc_dkdev.dk_geom;
454 struct buf *nbp;
455 void * addr;
456 void * newaddr;
457 daddr_t bn;
458 struct vnode *vp;
459
460 DPRINTF_FOLLOW(("cgd_diskstart(%p, %p)\n", dksc, bp));
461
462 bn = bp->b_rawblkno;
463
464 /*
465 * We attempt to allocate all of our resources up front, so that
466 * we can fail quickly if they are unavailable.
467 */
468 nbp = getiobuf(cs->sc_tvn, false);
469 if (nbp == NULL)
470 return EAGAIN;
471
472 /*
473 * If we are writing, then we need to encrypt the outgoing
474 * block into a new block of memory.
475 */
476 newaddr = addr = bp->b_data;
477 if ((bp->b_flags & B_READ) == 0) {
478 newaddr = cgd_getdata(dksc, bp->b_bcount);
479 if (!newaddr) {
480 putiobuf(nbp);
481 return EAGAIN;
482 }
483 cgd_cipher(cs, newaddr, addr, bp->b_bcount, bn,
484 dg->dg_secsize, CGD_CIPHER_ENCRYPT);
485 }
486
487 nbp->b_data = newaddr;
488 nbp->b_flags = bp->b_flags;
489 nbp->b_oflags = bp->b_oflags;
490 nbp->b_cflags = bp->b_cflags;
491 nbp->b_iodone = cgdiodone;
492 nbp->b_proc = bp->b_proc;
493 nbp->b_blkno = btodb(bn * dg->dg_secsize);
494 nbp->b_bcount = bp->b_bcount;
495 nbp->b_private = bp;
496
497 BIO_COPYPRIO(nbp, bp);
498
499 if ((nbp->b_flags & B_READ) == 0) {
500 vp = nbp->b_vp;
501 mutex_enter(vp->v_interlock);
502 vp->v_numoutput++;
503 mutex_exit(vp->v_interlock);
504 }
505 VOP_STRATEGY(cs->sc_tvn, nbp);
506
507 return 0;
508 }
509
510 static void
511 cgdiodone(struct buf *nbp)
512 {
513 dev_t dev;
514 device_t self;
515 struct buf *obp = nbp->b_private;
516 struct cgd_softc *cs = getcgd_softc(obp->b_dev, &self);
517 struct dk_softc *dksc = &cs->sc_dksc;
518 struct disk_geom *dg = &dksc->sc_dkdev.dk_geom;
519 daddr_t bn;
520
521 KDASSERT(cs);
522
523 DPRINTF_FOLLOW(("cgdiodone(%p)\n", nbp));
524 DPRINTF(CGDB_IO, ("cgdiodone: bp %p bcount %d resid %d\n",
525 obp, obp->b_bcount, obp->b_resid));
526 DPRINTF(CGDB_IO, (" dev 0x%"PRIx64", nbp %p bn %" PRId64
527 " addr %p bcnt %d\n", nbp->b_dev, nbp, nbp->b_blkno, nbp->b_data,
528 nbp->b_bcount));
529 if (nbp->b_error != 0) {
530 obp->b_error = nbp->b_error;
531 DPRINTF(CGDB_IO, ("%s: error %d\n", dksc->sc_xname,
532 obp->b_error));
533 }
534
535 /* Perform the decryption if we are reading.
536 *
537 * Note: use the blocknumber from nbp, since it is what
538 * we used to encrypt the blocks.
539 */
540
541 if (nbp->b_flags & B_READ) {
542 bn = dbtob(nbp->b_blkno) / dg->dg_secsize;
543 cgd_cipher(cs, obp->b_data, obp->b_data, obp->b_bcount,
544 bn, dg->dg_secsize, CGD_CIPHER_DECRYPT);
545 }
546
547 /* If we allocated memory, free it now... */
548 if (nbp->b_data != obp->b_data)
549 cgd_putdata(dksc, nbp->b_data);
550
551 putiobuf(nbp);
552
553 /* Request is complete for whatever reason */
554 obp->b_resid = 0;
555 if (obp->b_error != 0)
556 obp->b_resid = obp->b_bcount;
557
558 /*
559 * copy the dev_t, finish the disk operation, and release the
560 * reference we're holding on to (from getcgd_softc() earlier)
561 */
562 dev = obp->b_dev;
563 dk_done(dksc, obp);
564 cgd_release(dev);
565 device_release(self);
566
567 dk_start(dksc, NULL);
568 }
569
570 static int
571 cgd_dumpblocks(device_t dev, void *va, daddr_t blkno, int nblk)
572 {
573 struct cgd_softc *sc = device_private(dev);
574 struct dk_softc *dksc = &sc->sc_dksc;
575 struct disk_geom *dg = &dksc->sc_dkdev.dk_geom;
576 size_t nbytes, blksize;
577 void *buf;
578 int error;
579
580 /*
581 * dk_dump gives us units of disklabel sectors. Everything
582 * else in cgd uses units of diskgeom sectors. These had
583 * better agree; otherwise we need to figure out how to convert
584 * between them.
585 */
586 KASSERTMSG((dg->dg_secsize == dksc->sc_dkdev.dk_label->d_secsize),
587 "diskgeom secsize %"PRIu32" != disklabel secsize %"PRIu32,
588 dg->dg_secsize, dksc->sc_dkdev.dk_label->d_secsize);
589 blksize = dg->dg_secsize;
590
591 /*
592 * Compute the number of bytes in this request, which dk_dump
593 * has `helpfully' converted to a number of blocks for us.
594 */
595 nbytes = nblk*blksize;
596
597 /* Try to acquire a buffer to store the ciphertext. */
598 buf = cgd_getdata(dksc, nbytes);
599 if (buf == NULL)
600 /* Out of memory: give up. */
601 return ENOMEM;
602
603 /* Encrypt the caller's data into the temporary buffer. */
604 cgd_cipher(sc, buf, va, nbytes, blkno, blksize, CGD_CIPHER_ENCRYPT);
605
606 /* Pass it on to the underlying disk device. */
607 error = bdev_dump(sc->sc_tdev, blkno, buf, nbytes);
608
609 /* Release the buffer. */
610 cgd_putdata(dksc, buf);
611
612 /* Return any error from the underlying disk device. */
613 return error;
614 }
615
616 /* XXX: we should probably put these into dksubr.c, mostly */
617 static int
618 cgdread(dev_t dev, struct uio *uio, int flags)
619 {
620 device_t self;
621 int error;
622 struct cgd_softc *cs;
623 struct dk_softc *dksc;
624
625 DPRINTF_FOLLOW(("cgdread(0x%llx, %p, %d)\n",
626 (unsigned long long)dev, uio, flags));
627 GETCGD_SOFTC(cs, dev, self);
628 dksc = &cs->sc_dksc;
629 if (!DK_ATTACHED(dksc))
630 return ENXIO;
631 error = physio(cgdstrategy, NULL, dev, B_READ, minphys, uio);
632 device_release(self);
633 return error;
634 }
635
636 /* XXX: we should probably put these into dksubr.c, mostly */
637 static int
638 cgdwrite(dev_t dev, struct uio *uio, int flags)
639 {
640 device_t self;
641 int error;
642 struct cgd_softc *cs;
643 struct dk_softc *dksc;
644
645 DPRINTF_FOLLOW(("cgdwrite(0x%"PRIx64", %p, %d)\n", dev, uio, flags));
646 GETCGD_SOFTC(cs, dev, self);
647 dksc = &cs->sc_dksc;
648 if (!DK_ATTACHED(dksc)) {
649 device_release(self);
650 return ENXIO;
651 }
652 error = physio(cgdstrategy, NULL, dev, B_WRITE, minphys, uio);
653 device_release(self);
654 return error;
655 }
656
657 static int
658 cgdioctl(dev_t dev, u_long cmd, void *data, int flag, struct lwp *l)
659 {
660 device_t self;
661 struct cgd_softc *cs;
662 struct dk_softc *dksc;
663 int part = DISKPART(dev);
664 int pmask = 1 << part;
665 int error = 0;
666
667 DPRINTF_FOLLOW(("cgdioctl(0x%"PRIx64", %ld, %p, %d, %p)\n",
668 dev, cmd, data, flag, l));
669
670 switch (cmd) {
671 case CGDIOCGET:
672 return cgd_ioctl_get(dev, data, l);
673 case CGDIOCSET:
674 case CGDIOCCLR:
675 if ((flag & FWRITE) == 0)
676 return EBADF;
677 /* FALLTHROUGH */
678 default:
679 GETCGD_SOFTC(cs, dev, self);
680 dksc = &cs->sc_dksc;
681 break;
682 }
683
684 switch (cmd) {
685 case CGDIOCSET:
686 if (DK_ATTACHED(dksc))
687 error = EBUSY;
688 else
689 error = cgd_ioctl_set(cs, data, l);
690 break;
691 case CGDIOCCLR:
692 if (DK_BUSY(&cs->sc_dksc, pmask))
693 error = EBUSY;
694 else
695 error = cgd_ioctl_clr(cs, l);
696 break;
697 case DIOCCACHESYNC:
698 /*
699 * XXX Do we really need to care about having a writable
700 * file descriptor here?
701 */
702 if ((flag & FWRITE) == 0)
703 error = (EBADF);
704
705 /*
706 * We pass this call down to the underlying disk.
707 */
708 else
709 error = VOP_IOCTL(cs->sc_tvn, cmd, data, flag,
710 l->l_cred);
711 break;
712 case DIOCGSTRATEGY:
713 case DIOCSSTRATEGY:
714 if (!DK_ATTACHED(dksc)) {
715 error = ENOENT;
716 break;
717 }
718 /*FALLTHROUGH*/
719 default:
720 error = dk_ioctl(dksc, dev, cmd, data, flag, l);
721 break;
722 case CGDIOCGET:
723 KASSERT(0);
724 error = EINVAL;
725 break;
726 }
727 device_release(self);
728 return error;
729 }
730
731 static int
732 cgddump(dev_t dev, daddr_t blkno, void *va, size_t size)
733 {
734 device_t self;
735 int error;
736 struct cgd_softc *cs;
737
738 DPRINTF_FOLLOW(("cgddump(0x%"PRIx64", %" PRId64 ", %p, %lu)\n",
739 dev, blkno, va, (unsigned long)size));
740 GETCGD_SOFTC(cs, dev, self);
741 error = dk_dump(&cs->sc_dksc, dev, blkno, va, size);
742 device_release(self);
743 return error;
744 }
745
746 /*
747 * XXXrcd:
748 * for now we hardcode the maximum key length.
749 */
750 #define MAX_KEYSIZE 1024
751
752 static const struct {
753 const char *n;
754 int v;
755 int d;
756 } encblkno[] = {
757 { "encblkno", CGD_CIPHER_CBC_ENCBLKNO8, 1 },
758 { "encblkno8", CGD_CIPHER_CBC_ENCBLKNO8, 1 },
759 { "encblkno1", CGD_CIPHER_CBC_ENCBLKNO1, 8 },
760 };
761
762 /* ARGSUSED */
763 static int
764 cgd_ioctl_set(struct cgd_softc *cs, void *data, struct lwp *l)
765 {
766 struct cgd_ioctl *ci = data;
767 struct vnode *vp;
768 int ret;
769 size_t i;
770 size_t keybytes; /* key length in bytes */
771 const char *cp;
772 struct pathbuf *pb;
773 char *inbuf;
774 struct dk_softc *dksc = &cs->sc_dksc;
775
776 cp = ci->ci_disk;
777
778 ret = pathbuf_copyin(ci->ci_disk, &pb);
779 if (ret != 0) {
780 return ret;
781 }
782 ret = dk_lookup(pb, l, &vp);
783 pathbuf_destroy(pb);
784 if (ret != 0) {
785 return ret;
786 }
787
788 inbuf = malloc(MAX_KEYSIZE, M_TEMP, M_WAITOK);
789
790 if ((ret = cgdinit(cs, cp, vp, l)) != 0)
791 goto bail;
792
793 (void)memset(inbuf, 0, MAX_KEYSIZE);
794 ret = copyinstr(ci->ci_alg, inbuf, 256, NULL);
795 if (ret)
796 goto bail;
797 cs->sc_cfuncs = cryptfuncs_find(inbuf);
798 if (!cs->sc_cfuncs) {
799 ret = EINVAL;
800 goto bail;
801 }
802
803 (void)memset(inbuf, 0, MAX_KEYSIZE);
804 ret = copyinstr(ci->ci_ivmethod, inbuf, MAX_KEYSIZE, NULL);
805 if (ret)
806 goto bail;
807
808 for (i = 0; i < __arraycount(encblkno); i++)
809 if (strcmp(encblkno[i].n, inbuf) == 0)
810 break;
811
812 if (i == __arraycount(encblkno)) {
813 ret = EINVAL;
814 goto bail;
815 }
816
817 keybytes = ci->ci_keylen / 8 + 1;
818 if (keybytes > MAX_KEYSIZE) {
819 ret = EINVAL;
820 goto bail;
821 }
822
823 (void)memset(inbuf, 0, MAX_KEYSIZE);
824 ret = copyin(ci->ci_key, inbuf, keybytes);
825 if (ret)
826 goto bail;
827
828 cs->sc_cdata.cf_blocksize = ci->ci_blocksize;
829 cs->sc_cdata.cf_mode = encblkno[i].v;
830 cs->sc_cdata.cf_keylen = ci->ci_keylen;
831 cs->sc_cdata.cf_priv = cs->sc_cfuncs->cf_init(ci->ci_keylen, inbuf,
832 &cs->sc_cdata.cf_blocksize);
833 if (cs->sc_cdata.cf_blocksize > CGD_MAXBLOCKSIZE) {
834 log(LOG_WARNING, "cgd: Disallowed cipher with blocksize %zu > %u\n",
835 cs->sc_cdata.cf_blocksize, CGD_MAXBLOCKSIZE);
836 cs->sc_cdata.cf_priv = NULL;
837 }
838
839 /*
840 * The blocksize is supposed to be in bytes. Unfortunately originally
841 * it was expressed in bits. For compatibility we maintain encblkno
842 * and encblkno8.
843 */
844 cs->sc_cdata.cf_blocksize /= encblkno[i].d;
845 (void)explicit_memset(inbuf, 0, MAX_KEYSIZE);
846 if (!cs->sc_cdata.cf_priv) {
847 ret = EINVAL; /* XXX is this the right error? */
848 goto bail;
849 }
850 free(inbuf, M_TEMP);
851
852 bufq_alloc(&dksc->sc_bufq, "fcfs", 0);
853
854 cs->sc_data = malloc(MAXPHYS, M_DEVBUF, M_WAITOK);
855 cs->sc_data_used = 0;
856
857 /* Attach the disk. */
858 dk_attach(dksc);
859 disk_attach(&dksc->sc_dkdev);
860
861 disk_set_info(dksc->sc_dev, &dksc->sc_dkdev, NULL);
862
863 /* Discover wedges on this disk. */
864 dkwedge_discover(&dksc->sc_dkdev);
865
866 return 0;
867
868 bail:
869 free(inbuf, M_TEMP);
870 (void)vn_close(vp, FREAD|FWRITE, l->l_cred);
871 return ret;
872 }
873
874 /* ARGSUSED */
875 static int
876 cgd_ioctl_clr(struct cgd_softc *cs, struct lwp *l)
877 {
878 struct dk_softc *dksc = &cs->sc_dksc;
879
880 if (!DK_ATTACHED(dksc))
881 return ENXIO;
882
883 /* Delete all of our wedges. */
884 dkwedge_delall(&dksc->sc_dkdev);
885
886 /* Kill off any queued buffers. */
887 dk_drain(dksc);
888 bufq_free(dksc->sc_bufq);
889
890 (void)vn_close(cs->sc_tvn, FREAD|FWRITE, l->l_cred);
891 cs->sc_cfuncs->cf_destroy(cs->sc_cdata.cf_priv);
892 free(cs->sc_tpath, M_DEVBUF);
893 free(cs->sc_data, M_DEVBUF);
894 cs->sc_data_used = 0;
895 dk_detach(dksc);
896 disk_detach(&dksc->sc_dkdev);
897
898 return 0;
899 }
900
901 static int
902 cgd_ioctl_get(dev_t dev, void *data, struct lwp *l)
903 {
904 device_t self;
905 struct cgd_softc *cs = getcgd_softc(dev, &self);
906 struct cgd_user *cgu;
907 int unit;
908 struct dk_softc *dksc = &cs->sc_dksc;
909
910 unit = CGDUNIT(dev);
911 cgu = (struct cgd_user *)data;
912
913 DPRINTF_FOLLOW(("cgd_ioctl_get(0x%"PRIx64", %d, %p, %p)\n",
914 dev, unit, data, l));
915
916 if (cgu->cgu_unit == -1)
917 cgu->cgu_unit = unit;
918
919 if (cgu->cgu_unit < 0) {
920 cgd_release(dev);
921 device_release(self);
922 return EINVAL; /* XXX: should this be ENXIO? */
923 }
924
925 cs = device_lookup_private(&cgd_cd, unit);
926 if (cs == NULL || !DK_ATTACHED(dksc)) {
927 cgu->cgu_dev = 0;
928 cgu->cgu_alg[0] = '\0';
929 cgu->cgu_blocksize = 0;
930 cgu->cgu_mode = 0;
931 cgu->cgu_keylen = 0;
932 }
933 else {
934 cgu->cgu_dev = cs->sc_tdev;
935 strlcpy(cgu->cgu_alg, cs->sc_cfuncs->cf_name,
936 sizeof(cgu->cgu_alg));
937 cgu->cgu_blocksize = cs->sc_cdata.cf_blocksize;
938 cgu->cgu_mode = cs->sc_cdata.cf_mode;
939 cgu->cgu_keylen = cs->sc_cdata.cf_keylen;
940 }
941 cgd_release(dev);
942 device_release(self);
943 return 0;
944 }
945
946 static int
947 cgdinit(struct cgd_softc *cs, const char *cpath, struct vnode *vp,
948 struct lwp *l)
949 {
950 struct disk_geom *dg;
951 int ret;
952 char *tmppath;
953 uint64_t psize;
954 unsigned secsize;
955 struct dk_softc *dksc = &cs->sc_dksc;
956
957 cs->sc_tvn = vp;
958 cs->sc_tpath = NULL;
959
960 tmppath = malloc(MAXPATHLEN, M_TEMP, M_WAITOK);
961 ret = copyinstr(cpath, tmppath, MAXPATHLEN, &cs->sc_tpathlen);
962 if (ret)
963 goto bail;
964 cs->sc_tpath = malloc(cs->sc_tpathlen, M_DEVBUF, M_WAITOK);
965 memcpy(cs->sc_tpath, tmppath, cs->sc_tpathlen);
966
967 cs->sc_tdev = vp->v_rdev;
968
969 if ((ret = getdisksize(vp, &psize, &secsize)) != 0)
970 goto bail;
971
972 if (psize == 0) {
973 ret = ENODEV;
974 goto bail;
975 }
976
977 /*
978 * XXX here we should probe the underlying device. If we
979 * are accessing a partition of type RAW_PART, then
980 * we should populate our initial geometry with the
981 * geometry that we discover from the device.
982 */
983 dg = &dksc->sc_dkdev.dk_geom;
984 memset(dg, 0, sizeof(*dg));
985 dg->dg_secperunit = psize;
986 dg->dg_secsize = secsize;
987 dg->dg_ntracks = 1;
988 dg->dg_nsectors = 1024 * 1024 / dg->dg_secsize;
989 dg->dg_ncylinders = dg->dg_secperunit / dg->dg_nsectors;
990
991 bail:
992 free(tmppath, M_TEMP);
993 if (ret && cs->sc_tpath)
994 free(cs->sc_tpath, M_DEVBUF);
995 return ret;
996 }
997
998 /*
999 * Our generic cipher entry point. This takes care of the
1000 * IV mode and passes off the work to the specific cipher.
1001 * We implement here the IV method ``encrypted block
1002 * number''.
1003 *
1004 * For the encryption case, we accomplish this by setting
1005 * up a struct uio where the first iovec of the source is
1006 * the blocknumber and the first iovec of the dest is a
1007 * sink. We then call the cipher with an IV of zero, and
1008 * the right thing happens.
1009 *
1010 * For the decryption case, we use the same basic mechanism
1011 * for symmetry, but we encrypt the block number in the
1012 * first iovec.
1013 *
1014 * We mainly do this to avoid requiring the definition of
1015 * an ECB mode.
1016 *
1017 * XXXrcd: for now we rely on our own crypto framework defined
1018 * in dev/cgd_crypto.c. This will change when we
1019 * get a generic kernel crypto framework.
1020 */
1021
1022 static void
1023 blkno2blkno_buf(char *sbuf, daddr_t blkno)
1024 {
1025 int i;
1026
1027 /* Set up the blkno in blkno_buf, here we do not care much
1028 * about the final layout of the information as long as we
1029 * can guarantee that each sector will have a different IV
1030 * and that the endianness of the machine will not affect
1031 * the representation that we have chosen.
1032 *
1033 * We choose this representation, because it does not rely
1034 * on the size of buf (which is the blocksize of the cipher),
1035 * but allows daddr_t to grow without breaking existing
1036 * disks.
1037 *
1038 * Note that blkno2blkno_buf does not take a size as input,
1039 * and hence must be called on a pre-zeroed buffer of length
1040 * greater than or equal to sizeof(daddr_t).
1041 */
1042 for (i=0; i < sizeof(daddr_t); i++) {
1043 *sbuf++ = blkno & 0xff;
1044 blkno >>= 8;
1045 }
1046 }
1047
1048 static void
1049 cgd_cipher(struct cgd_softc *cs, void *dstv, void *srcv,
1050 size_t len, daddr_t blkno, size_t secsize, int dir)
1051 {
1052 char *dst = dstv;
1053 char *src = srcv;
1054 cfunc_cipher *cipher = cs->sc_cfuncs->cf_cipher;
1055 struct uio dstuio;
1056 struct uio srcuio;
1057 struct iovec dstiov[2];
1058 struct iovec srciov[2];
1059 size_t blocksize = cs->sc_cdata.cf_blocksize;
1060 size_t todo;
1061 char sink[CGD_MAXBLOCKSIZE];
1062 char zero_iv[CGD_MAXBLOCKSIZE];
1063 char blkno_buf[CGD_MAXBLOCKSIZE];
1064
1065 DPRINTF_FOLLOW(("cgd_cipher() dir=%d\n", dir));
1066
1067 DIAGCONDPANIC(len % blocksize != 0,
1068 ("cgd_cipher: len %% blocksize != 0"));
1069
1070 /* ensure that sizeof(daddr_t) <= blocksize (for encblkno IVing) */
1071 DIAGCONDPANIC(sizeof(daddr_t) > blocksize,
1072 ("cgd_cipher: sizeof(daddr_t) > blocksize"));
1073
1074 memset(zero_iv, 0x0, blocksize);
1075
1076 dstuio.uio_iov = dstiov;
1077 dstuio.uio_iovcnt = 2;
1078
1079 srcuio.uio_iov = srciov;
1080 srcuio.uio_iovcnt = 2;
1081
1082 dstiov[0].iov_base = sink;
1083 dstiov[0].iov_len = blocksize;
1084 srciov[0].iov_base = blkno_buf;
1085 srciov[0].iov_len = blocksize;
1086
1087 for (; len > 0; len -= todo) {
1088 todo = MIN(len, secsize);
1089
1090 dstiov[1].iov_base = dst;
1091 srciov[1].iov_base = src;
1092 dstiov[1].iov_len = todo;
1093 srciov[1].iov_len = todo;
1094
1095 memset(blkno_buf, 0x0, blocksize);
1096 blkno2blkno_buf(blkno_buf, blkno);
1097 if (dir == CGD_CIPHER_DECRYPT) {
1098 dstuio.uio_iovcnt = 1;
1099 srcuio.uio_iovcnt = 1;
1100 IFDEBUG(CGDB_CRYPTO, hexprint("step 0: blkno_buf",
1101 blkno_buf, blocksize));
1102 cipher(cs->sc_cdata.cf_priv, &dstuio, &srcuio,
1103 zero_iv, CGD_CIPHER_ENCRYPT);
1104 memcpy(blkno_buf, sink, blocksize);
1105 dstuio.uio_iovcnt = 2;
1106 srcuio.uio_iovcnt = 2;
1107 }
1108
1109 IFDEBUG(CGDB_CRYPTO, hexprint("step 1: blkno_buf",
1110 blkno_buf, blocksize));
1111 cipher(cs->sc_cdata.cf_priv, &dstuio, &srcuio, zero_iv, dir);
1112 IFDEBUG(CGDB_CRYPTO, hexprint("step 2: sink",
1113 sink, blocksize));
1114
1115 dst += todo;
1116 src += todo;
1117 blkno++;
1118 }
1119 }
1120
1121 #ifdef DEBUG
1122 static void
1123 hexprint(const char *start, void *buf, int len)
1124 {
1125 char *c = buf;
1126
1127 DIAGCONDPANIC(len < 0, ("hexprint: called with len < 0"));
1128 printf("%s: len=%06d 0x", start, len);
1129 while (len--)
1130 printf("%02x", (unsigned char) *c++);
1131 }
1132 #endif
1133
1134 MODULE(MODULE_CLASS_DRIVER, cgd, "dk_subr");
1135
1136 #ifdef _MODULE
1137 CFDRIVER_DECL(cgd, DV_DISK, NULL);
1138 #endif
1139
1140 static int
1141 cgd_modcmd(modcmd_t cmd, void *arg)
1142 {
1143 int error = 0;
1144
1145 #ifdef _MODULE
1146 devmajor_t bmajor = -1, cmajor = -1;
1147 #endif
1148
1149 switch (cmd) {
1150 case MODULE_CMD_INIT:
1151 #ifdef _MODULE
1152 error = config_cfdriver_attach(&cgd_cd);
1153 if (error)
1154 break;
1155
1156 error = config_cfattach_attach(cgd_cd.cd_name, &cgd_ca);
1157 if (error) {
1158 config_cfdriver_detach(&cgd_cd);
1159 aprint_error("%s: unable to register cfattach for "
1160 "%s, error %d", __func__, cgd_cd.cd_name, error);
1161 break;
1162 }
1163
1164 /*
1165 * Attach the {b,c}devsw's
1166 */
1167 error = devsw_attach("cgd", &cgd_bdevsw, &bmajor,
1168 &cgd_cdevsw, &cmajor);
1169
1170 /*
1171 * If devsw_attach fails, remove from autoconf database
1172 */
1173 if (error) {
1174 config_cfattach_detach(cgd_cd.cd_name, &cgd_ca);
1175 config_cfdriver_detach(&cgd_cd);
1176 aprint_error("%s: unable to attach %s devsw, "
1177 "error %d", __func__, cgd_cd.cd_name, error);
1178 }
1179 #endif
1180 break;
1181
1182 case MODULE_CMD_FINI:
1183 #ifdef _MODULE
1184 /*
1185 * Remove {b,c}devsw's
1186 */
1187 devsw_detach(&cgd_bdevsw, &cgd_cdevsw);
1188
1189 /*
1190 * Now remove device from autoconf database
1191 */
1192 error = config_cfattach_detach(cgd_cd.cd_name, &cgd_ca);
1193 if (error) {
1194 error = devsw_attach("cgd", &cgd_bdevsw, &bmajor,
1195 &cgd_cdevsw, &cmajor);
1196 aprint_error("%s: failed to detach %s cfattach, "
1197 "error %d\n", __func__, cgd_cd.cd_name, error);
1198 break;
1199 }
1200 error = config_cfdriver_detach(&cgd_cd);
1201 if (error) {
1202 config_cfattach_attach(cgd_cd.cd_name, &cgd_ca);
1203 devsw_attach("cgd", &cgd_bdevsw, &bmajor,
1204 &cgd_cdevsw, &cmajor);
1205 aprint_error("%s: failed to detach %s cfdriver, "
1206 "error %d\n", __func__, cgd_cd.cd_name, error);
1207 break;
1208 }
1209 #endif
1210 break;
1211
1212 case MODULE_CMD_STAT:
1213 error = ENOTTY;
1214 break;
1215 default:
1216 error = ENOTTY;
1217 break;
1218 }
1219
1220 return error;
1221 }
1222