cgd.c revision 1.91.2.4 1 /* $NetBSD: cgd.c,v 1.91.2.4 2015/12/27 12:09:48 skrll Exp $ */
2
3 /*-
4 * Copyright (c) 2002 The NetBSD Foundation, Inc.
5 * All rights reserved.
6 *
7 * This code is derived from software contributed to The NetBSD Foundation
8 * by Roland C. Dowdeswell.
9 *
10 * Redistribution and use in source and binary forms, with or without
11 * modification, are permitted provided that the following conditions
12 * are met:
13 * 1. Redistributions of source code must retain the above copyright
14 * notice, this list of conditions and the following disclaimer.
15 * 2. Redistributions in binary form must reproduce the above copyright
16 * notice, this list of conditions and the following disclaimer in the
17 * documentation and/or other materials provided with the distribution.
18 *
19 * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
20 * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
21 * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
22 * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
23 * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
24 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
25 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
26 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
27 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
28 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
29 * POSSIBILITY OF SUCH DAMAGE.
30 */
31
32 #include <sys/cdefs.h>
33 __KERNEL_RCSID(0, "$NetBSD: cgd.c,v 1.91.2.4 2015/12/27 12:09:48 skrll Exp $");
34
35 #include <sys/types.h>
36 #include <sys/param.h>
37 #include <sys/systm.h>
38 #include <sys/proc.h>
39 #include <sys/errno.h>
40 #include <sys/buf.h>
41 #include <sys/bufq.h>
42 #include <sys/malloc.h>
43 #include <sys/module.h>
44 #include <sys/pool.h>
45 #include <sys/ioctl.h>
46 #include <sys/device.h>
47 #include <sys/disk.h>
48 #include <sys/disklabel.h>
49 #include <sys/fcntl.h>
50 #include <sys/namei.h> /* for pathbuf */
51 #include <sys/vnode.h>
52 #include <sys/conf.h>
53 #include <sys/syslog.h>
54
55 #include <dev/dkvar.h>
56 #include <dev/cgdvar.h>
57
58 #include <miscfs/specfs/specdev.h> /* for v_rdev */
59
60 #include "ioconf.h"
61
62 /* Entry Point Functions */
63
64 static dev_type_open(cgdopen);
65 static dev_type_close(cgdclose);
66 static dev_type_read(cgdread);
67 static dev_type_write(cgdwrite);
68 static dev_type_ioctl(cgdioctl);
69 static dev_type_strategy(cgdstrategy);
70 static dev_type_dump(cgddump);
71 static dev_type_size(cgdsize);
72
73 const struct bdevsw cgd_bdevsw = {
74 .d_open = cgdopen,
75 .d_close = cgdclose,
76 .d_strategy = cgdstrategy,
77 .d_ioctl = cgdioctl,
78 .d_dump = cgddump,
79 .d_psize = cgdsize,
80 .d_discard = nodiscard,
81 .d_flag = D_DISK
82 };
83
84 const struct cdevsw cgd_cdevsw = {
85 .d_open = cgdopen,
86 .d_close = cgdclose,
87 .d_read = cgdread,
88 .d_write = cgdwrite,
89 .d_ioctl = cgdioctl,
90 .d_stop = nostop,
91 .d_tty = notty,
92 .d_poll = nopoll,
93 .d_mmap = nommap,
94 .d_kqfilter = nokqfilter,
95 .d_discard = nodiscard,
96 .d_flag = D_DISK
97 };
98
99 static int cgd_match(device_t, cfdata_t, void *);
100 static void cgd_attach(device_t, device_t, void *);
101 static int cgd_detach(device_t, int);
102 static struct cgd_softc *cgd_spawn(int);
103 static int cgd_destroy(device_t);
104
105 /* Internal Functions */
106
107 static int cgd_diskstart(device_t, struct buf *);
108 static void cgdiodone(struct buf *);
109
110 static int cgd_ioctl_set(struct cgd_softc *, void *, struct lwp *);
111 static int cgd_ioctl_clr(struct cgd_softc *, struct lwp *);
112 static int cgd_ioctl_get(dev_t, void *, struct lwp *);
113 static int cgdinit(struct cgd_softc *, const char *, struct vnode *,
114 struct lwp *);
115 static void cgd_cipher(struct cgd_softc *, void *, void *,
116 size_t, daddr_t, size_t, int);
117
118 static struct dkdriver cgddkdriver = {
119 .d_minphys = minphys,
120 .d_open = cgdopen,
121 .d_close = cgdclose,
122 .d_strategy = cgdstrategy,
123 .d_iosize = NULL,
124 .d_diskstart = cgd_diskstart,
125 .d_dumpblocks = NULL,
126 .d_lastclose = NULL
127 };
128
129 CFATTACH_DECL3_NEW(cgd, sizeof(struct cgd_softc),
130 cgd_match, cgd_attach, cgd_detach, NULL, NULL, NULL, DVF_DETACH_SHUTDOWN);
131 extern struct cfdriver cgd_cd;
132
133 /* DIAGNOSTIC and DEBUG definitions */
134
135 #if defined(CGDDEBUG) && !defined(DEBUG)
136 #define DEBUG
137 #endif
138
139 #ifdef DEBUG
140 int cgddebug = 0;
141
142 #define CGDB_FOLLOW 0x1
143 #define CGDB_IO 0x2
144 #define CGDB_CRYPTO 0x4
145
146 #define IFDEBUG(x,y) if (cgddebug & (x)) y
147 #define DPRINTF(x,y) IFDEBUG(x, printf y)
148 #define DPRINTF_FOLLOW(y) DPRINTF(CGDB_FOLLOW, y)
149
150 static void hexprint(const char *, void *, int);
151
152 #else
153 #define IFDEBUG(x,y)
154 #define DPRINTF(x,y)
155 #define DPRINTF_FOLLOW(y)
156 #endif
157
158 #ifdef DIAGNOSTIC
159 #define DIAGPANIC(x) panic x
160 #define DIAGCONDPANIC(x,y) if (x) panic y
161 #else
162 #define DIAGPANIC(x)
163 #define DIAGCONDPANIC(x,y)
164 #endif
165
166 /* Global variables */
167
168 /* Utility Functions */
169
170 #define CGDUNIT(x) DISKUNIT(x)
171 #define GETCGD_SOFTC(_cs, x) if (!((_cs) = getcgd_softc(x))) return ENXIO
172
173 /* The code */
174
175 static struct cgd_softc *
176 getcgd_softc(dev_t dev)
177 {
178 int unit = CGDUNIT(dev);
179 struct cgd_softc *sc;
180
181 DPRINTF_FOLLOW(("getcgd_softc(0x%"PRIx64"): unit = %d\n", dev, unit));
182
183 sc = device_lookup_private(&cgd_cd, unit);
184 if (sc == NULL)
185 sc = cgd_spawn(unit);
186 return sc;
187 }
188
189 static int
190 cgd_match(device_t self, cfdata_t cfdata, void *aux)
191 {
192
193 return 1;
194 }
195
196 static void
197 cgd_attach(device_t parent, device_t self, void *aux)
198 {
199 struct cgd_softc *sc = device_private(self);
200
201 mutex_init(&sc->sc_lock, MUTEX_DEFAULT, IPL_BIO);
202 dk_init(&sc->sc_dksc, self, DKTYPE_CGD);
203 disk_init(&sc->sc_dksc.sc_dkdev, sc->sc_dksc.sc_xname, &cgddkdriver);
204
205 if (!pmf_device_register(self, NULL, NULL))
206 aprint_error_dev(self, "unable to register power management hooks\n");
207 }
208
209
210 static int
211 cgd_detach(device_t self, int flags)
212 {
213 int ret;
214 const int pmask = 1 << RAW_PART;
215 struct cgd_softc *sc = device_private(self);
216 struct dk_softc *dksc = &sc->sc_dksc;
217
218 if (DK_BUSY(dksc, pmask))
219 return EBUSY;
220
221 if (DK_ATTACHED(dksc) &&
222 (ret = cgd_ioctl_clr(sc, curlwp)) != 0)
223 return ret;
224
225 disk_destroy(&dksc->sc_dkdev);
226 mutex_destroy(&sc->sc_lock);
227
228 return 0;
229 }
230
231 void
232 cgdattach(int num)
233 {
234 int error;
235
236 error = config_cfattach_attach(cgd_cd.cd_name, &cgd_ca);
237 if (error != 0)
238 aprint_error("%s: unable to register cfattach\n",
239 cgd_cd.cd_name);
240 }
241
242 static struct cgd_softc *
243 cgd_spawn(int unit)
244 {
245 cfdata_t cf;
246
247 cf = malloc(sizeof(*cf), M_DEVBUF, M_WAITOK);
248 cf->cf_name = cgd_cd.cd_name;
249 cf->cf_atname = cgd_cd.cd_name;
250 cf->cf_unit = unit;
251 cf->cf_fstate = FSTATE_STAR;
252
253 return device_private(config_attach_pseudo(cf));
254 }
255
256 static int
257 cgd_destroy(device_t dev)
258 {
259 int error;
260 cfdata_t cf;
261
262 cf = device_cfdata(dev);
263 error = config_detach(dev, DETACH_QUIET);
264 if (error)
265 return error;
266 free(cf, M_DEVBUF);
267 return 0;
268 }
269
270 static int
271 cgdopen(dev_t dev, int flags, int fmt, struct lwp *l)
272 {
273 struct cgd_softc *cs;
274
275 DPRINTF_FOLLOW(("cgdopen(0x%"PRIx64", %d)\n", dev, flags));
276 GETCGD_SOFTC(cs, dev);
277 return dk_open(&cs->sc_dksc, dev, flags, fmt, l);
278 }
279
280 static int
281 cgdclose(dev_t dev, int flags, int fmt, struct lwp *l)
282 {
283 int error;
284 struct cgd_softc *cs;
285 struct dk_softc *dksc;
286
287 DPRINTF_FOLLOW(("cgdclose(0x%"PRIx64", %d)\n", dev, flags));
288 GETCGD_SOFTC(cs, dev);
289 dksc = &cs->sc_dksc;
290 if ((error = dk_close(dksc, dev, flags, fmt, l)) != 0)
291 return error;
292
293 if (!DK_ATTACHED(dksc)) {
294 if ((error = cgd_destroy(cs->sc_dksc.sc_dev)) != 0) {
295 aprint_error_dev(dksc->sc_dev,
296 "unable to detach instance\n");
297 return error;
298 }
299 }
300 return 0;
301 }
302
303 static void
304 cgdstrategy(struct buf *bp)
305 {
306 struct cgd_softc *cs = getcgd_softc(bp->b_dev);
307 struct dk_softc *dksc = &cs->sc_dksc;
308 struct disk_geom *dg = &dksc->sc_dkdev.dk_geom;
309
310 DPRINTF_FOLLOW(("cgdstrategy(%p): b_bcount = %ld\n", bp,
311 (long)bp->b_bcount));
312
313 /*
314 * Reject unaligned writes. We can encrypt and decrypt only
315 * complete disk sectors, and we let the ciphers require their
316 * buffers to be aligned to 32-bit boundaries.
317 */
318 if (bp->b_blkno < 0 ||
319 (bp->b_bcount % dg->dg_secsize) != 0 ||
320 ((uintptr_t)bp->b_data & 3) != 0) {
321 bp->b_error = EINVAL;
322 bp->b_resid = bp->b_bcount;
323 biodone(bp);
324 return;
325 }
326
327 /* XXXrcd: Should we test for (cs != NULL)? */
328 dk_strategy(&cs->sc_dksc, bp);
329 return;
330 }
331
332 static int
333 cgdsize(dev_t dev)
334 {
335 struct cgd_softc *cs = getcgd_softc(dev);
336
337 DPRINTF_FOLLOW(("cgdsize(0x%"PRIx64")\n", dev));
338 if (!cs)
339 return -1;
340 return dk_size(&cs->sc_dksc, dev);
341 }
342
343 /*
344 * cgd_{get,put}data are functions that deal with getting a buffer
345 * for the new encrypted data. We have a buffer per device so that
346 * we can ensure that we can always have a transaction in flight.
347 * We use this buffer first so that we have one less piece of
348 * malloc'ed data at any given point.
349 */
350
351 static void *
352 cgd_getdata(struct dk_softc *dksc, unsigned long size)
353 {
354 struct cgd_softc *cs = (struct cgd_softc *)dksc;
355 void * data = NULL;
356
357 mutex_enter(&cs->sc_lock);
358 if (cs->sc_data_used == 0) {
359 cs->sc_data_used = 1;
360 data = cs->sc_data;
361 }
362 mutex_exit(&cs->sc_lock);
363
364 if (data)
365 return data;
366
367 return malloc(size, M_DEVBUF, M_NOWAIT);
368 }
369
370 static void
371 cgd_putdata(struct dk_softc *dksc, void *data)
372 {
373 struct cgd_softc *cs = (struct cgd_softc *)dksc;
374
375 if (data == cs->sc_data) {
376 mutex_enter(&cs->sc_lock);
377 cs->sc_data_used = 0;
378 mutex_exit(&cs->sc_lock);
379 } else {
380 free(data, M_DEVBUF);
381 }
382 }
383
384 static int
385 cgd_diskstart(device_t dev, struct buf *bp)
386 {
387 struct cgd_softc *cs = device_private(dev);
388 struct dk_softc *dksc = &cs->sc_dksc;
389 struct disk_geom *dg = &dksc->sc_dkdev.dk_geom;
390 struct buf *nbp;
391 void * addr;
392 void * newaddr;
393 daddr_t bn;
394 struct vnode *vp;
395
396 DPRINTF_FOLLOW(("cgd_diskstart(%p, %p)\n", dksc, bp));
397
398 bn = bp->b_rawblkno;
399
400 /*
401 * We attempt to allocate all of our resources up front, so that
402 * we can fail quickly if they are unavailable.
403 */
404 nbp = getiobuf(cs->sc_tvn, false);
405 if (nbp == NULL)
406 return EAGAIN;
407
408 /*
409 * If we are writing, then we need to encrypt the outgoing
410 * block into a new block of memory.
411 */
412 newaddr = addr = bp->b_data;
413 if ((bp->b_flags & B_READ) == 0) {
414 newaddr = cgd_getdata(dksc, bp->b_bcount);
415 if (!newaddr) {
416 putiobuf(nbp);
417 return EAGAIN;
418 }
419 cgd_cipher(cs, newaddr, addr, bp->b_bcount, bn,
420 dg->dg_secsize, CGD_CIPHER_ENCRYPT);
421 }
422
423 nbp->b_data = newaddr;
424 nbp->b_flags = bp->b_flags;
425 nbp->b_oflags = bp->b_oflags;
426 nbp->b_cflags = bp->b_cflags;
427 nbp->b_iodone = cgdiodone;
428 nbp->b_proc = bp->b_proc;
429 nbp->b_blkno = btodb(bn * dg->dg_secsize);
430 nbp->b_bcount = bp->b_bcount;
431 nbp->b_private = bp;
432
433 BIO_COPYPRIO(nbp, bp);
434
435 if ((nbp->b_flags & B_READ) == 0) {
436 vp = nbp->b_vp;
437 mutex_enter(vp->v_interlock);
438 vp->v_numoutput++;
439 mutex_exit(vp->v_interlock);
440 }
441 VOP_STRATEGY(cs->sc_tvn, nbp);
442
443 return 0;
444 }
445
446 static void
447 cgdiodone(struct buf *nbp)
448 {
449 struct buf *obp = nbp->b_private;
450 struct cgd_softc *cs = getcgd_softc(obp->b_dev);
451 struct dk_softc *dksc = &cs->sc_dksc;
452 struct disk_geom *dg = &dksc->sc_dkdev.dk_geom;
453 daddr_t bn;
454
455 KDASSERT(cs);
456
457 DPRINTF_FOLLOW(("cgdiodone(%p)\n", nbp));
458 DPRINTF(CGDB_IO, ("cgdiodone: bp %p bcount %d resid %d\n",
459 obp, obp->b_bcount, obp->b_resid));
460 DPRINTF(CGDB_IO, (" dev 0x%"PRIx64", nbp %p bn %" PRId64 " addr %p bcnt %d\n",
461 nbp->b_dev, nbp, nbp->b_blkno, nbp->b_data,
462 nbp->b_bcount));
463 if (nbp->b_error != 0) {
464 obp->b_error = nbp->b_error;
465 DPRINTF(CGDB_IO, ("%s: error %d\n", dksc->sc_xname,
466 obp->b_error));
467 }
468
469 /* Perform the decryption if we are reading.
470 *
471 * Note: use the blocknumber from nbp, since it is what
472 * we used to encrypt the blocks.
473 */
474
475 if (nbp->b_flags & B_READ) {
476 bn = dbtob(nbp->b_blkno) / dg->dg_secsize;
477 cgd_cipher(cs, obp->b_data, obp->b_data, obp->b_bcount,
478 bn, dg->dg_secsize, CGD_CIPHER_DECRYPT);
479 }
480
481 /* If we allocated memory, free it now... */
482 if (nbp->b_data != obp->b_data)
483 cgd_putdata(dksc, nbp->b_data);
484
485 putiobuf(nbp);
486
487 /* Request is complete for whatever reason */
488 obp->b_resid = 0;
489 if (obp->b_error != 0)
490 obp->b_resid = obp->b_bcount;
491
492 dk_done(dksc, obp);
493 dk_start(dksc, NULL);
494 }
495
496 /* XXX: we should probably put these into dksubr.c, mostly */
497 static int
498 cgdread(dev_t dev, struct uio *uio, int flags)
499 {
500 struct cgd_softc *cs;
501 struct dk_softc *dksc;
502
503 DPRINTF_FOLLOW(("cgdread(0x%llx, %p, %d)\n",
504 (unsigned long long)dev, uio, flags));
505 GETCGD_SOFTC(cs, dev);
506 dksc = &cs->sc_dksc;
507 if (!DK_ATTACHED(dksc))
508 return ENXIO;
509 return physio(cgdstrategy, NULL, dev, B_READ, minphys, uio);
510 }
511
512 /* XXX: we should probably put these into dksubr.c, mostly */
513 static int
514 cgdwrite(dev_t dev, struct uio *uio, int flags)
515 {
516 struct cgd_softc *cs;
517 struct dk_softc *dksc;
518
519 DPRINTF_FOLLOW(("cgdwrite(0x%"PRIx64", %p, %d)\n", dev, uio, flags));
520 GETCGD_SOFTC(cs, dev);
521 dksc = &cs->sc_dksc;
522 if (!DK_ATTACHED(dksc))
523 return ENXIO;
524 return physio(cgdstrategy, NULL, dev, B_WRITE, minphys, uio);
525 }
526
527 static int
528 cgdioctl(dev_t dev, u_long cmd, void *data, int flag, struct lwp *l)
529 {
530 struct cgd_softc *cs;
531 struct dk_softc *dksc;
532 int part = DISKPART(dev);
533 int pmask = 1 << part;
534
535 DPRINTF_FOLLOW(("cgdioctl(0x%"PRIx64", %ld, %p, %d, %p)\n",
536 dev, cmd, data, flag, l));
537
538 switch (cmd) {
539 case CGDIOCGET:
540 return cgd_ioctl_get(dev, data, l);
541 case CGDIOCSET:
542 case CGDIOCCLR:
543 if ((flag & FWRITE) == 0)
544 return EBADF;
545 /* FALLTHROUGH */
546 default:
547 GETCGD_SOFTC(cs, dev);
548 dksc = &cs->sc_dksc;
549 break;
550 }
551
552 switch (cmd) {
553 case CGDIOCSET:
554 if (DK_ATTACHED(dksc))
555 return EBUSY;
556 return cgd_ioctl_set(cs, data, l);
557 case CGDIOCCLR:
558 if (DK_BUSY(&cs->sc_dksc, pmask))
559 return EBUSY;
560 return cgd_ioctl_clr(cs, l);
561 case DIOCCACHESYNC:
562 /*
563 * XXX Do we really need to care about having a writable
564 * file descriptor here?
565 */
566 if ((flag & FWRITE) == 0)
567 return (EBADF);
568
569 /*
570 * We pass this call down to the underlying disk.
571 */
572 return VOP_IOCTL(cs->sc_tvn, cmd, data, flag, l->l_cred);
573 case DIOCGSTRATEGY:
574 case DIOCSSTRATEGY:
575 if (!DK_ATTACHED(dksc))
576 return ENOENT;
577 /*FALLTHROUGH*/
578 default:
579 return dk_ioctl(dksc, dev, cmd, data, flag, l);
580 case CGDIOCGET:
581 KASSERT(0);
582 return EINVAL;
583 }
584 }
585
586 static int
587 cgddump(dev_t dev, daddr_t blkno, void *va, size_t size)
588 {
589 struct cgd_softc *cs;
590
591 DPRINTF_FOLLOW(("cgddump(0x%"PRIx64", %" PRId64 ", %p, %lu)\n",
592 dev, blkno, va, (unsigned long)size));
593 GETCGD_SOFTC(cs, dev);
594 return dk_dump(&cs->sc_dksc, dev, blkno, va, size);
595 }
596
597 /*
598 * XXXrcd:
599 * for now we hardcode the maximum key length.
600 */
601 #define MAX_KEYSIZE 1024
602
603 static const struct {
604 const char *n;
605 int v;
606 int d;
607 } encblkno[] = {
608 { "encblkno", CGD_CIPHER_CBC_ENCBLKNO8, 1 },
609 { "encblkno8", CGD_CIPHER_CBC_ENCBLKNO8, 1 },
610 { "encblkno1", CGD_CIPHER_CBC_ENCBLKNO1, 8 },
611 };
612
613 /* ARGSUSED */
614 static int
615 cgd_ioctl_set(struct cgd_softc *cs, void *data, struct lwp *l)
616 {
617 struct cgd_ioctl *ci = data;
618 struct vnode *vp;
619 int ret;
620 size_t i;
621 size_t keybytes; /* key length in bytes */
622 const char *cp;
623 struct pathbuf *pb;
624 char *inbuf;
625 struct dk_softc *dksc = &cs->sc_dksc;
626
627 cp = ci->ci_disk;
628
629 ret = pathbuf_copyin(ci->ci_disk, &pb);
630 if (ret != 0) {
631 return ret;
632 }
633 ret = dk_lookup(pb, l, &vp);
634 pathbuf_destroy(pb);
635 if (ret != 0) {
636 return ret;
637 }
638
639 inbuf = malloc(MAX_KEYSIZE, M_TEMP, M_WAITOK);
640
641 if ((ret = cgdinit(cs, cp, vp, l)) != 0)
642 goto bail;
643
644 (void)memset(inbuf, 0, MAX_KEYSIZE);
645 ret = copyinstr(ci->ci_alg, inbuf, 256, NULL);
646 if (ret)
647 goto bail;
648 cs->sc_cfuncs = cryptfuncs_find(inbuf);
649 if (!cs->sc_cfuncs) {
650 ret = EINVAL;
651 goto bail;
652 }
653
654 (void)memset(inbuf, 0, MAX_KEYSIZE);
655 ret = copyinstr(ci->ci_ivmethod, inbuf, MAX_KEYSIZE, NULL);
656 if (ret)
657 goto bail;
658
659 for (i = 0; i < __arraycount(encblkno); i++)
660 if (strcmp(encblkno[i].n, inbuf) == 0)
661 break;
662
663 if (i == __arraycount(encblkno)) {
664 ret = EINVAL;
665 goto bail;
666 }
667
668 keybytes = ci->ci_keylen / 8 + 1;
669 if (keybytes > MAX_KEYSIZE) {
670 ret = EINVAL;
671 goto bail;
672 }
673
674 (void)memset(inbuf, 0, MAX_KEYSIZE);
675 ret = copyin(ci->ci_key, inbuf, keybytes);
676 if (ret)
677 goto bail;
678
679 cs->sc_cdata.cf_blocksize = ci->ci_blocksize;
680 cs->sc_cdata.cf_mode = encblkno[i].v;
681 cs->sc_cdata.cf_keylen = ci->ci_keylen;
682 cs->sc_cdata.cf_priv = cs->sc_cfuncs->cf_init(ci->ci_keylen, inbuf,
683 &cs->sc_cdata.cf_blocksize);
684 if (cs->sc_cdata.cf_blocksize > CGD_MAXBLOCKSIZE) {
685 log(LOG_WARNING, "cgd: Disallowed cipher with blocksize %zu > %u\n",
686 cs->sc_cdata.cf_blocksize, CGD_MAXBLOCKSIZE);
687 cs->sc_cdata.cf_priv = NULL;
688 }
689
690 /*
691 * The blocksize is supposed to be in bytes. Unfortunately originally
692 * it was expressed in bits. For compatibility we maintain encblkno
693 * and encblkno8.
694 */
695 cs->sc_cdata.cf_blocksize /= encblkno[i].d;
696 (void)explicit_memset(inbuf, 0, MAX_KEYSIZE);
697 if (!cs->sc_cdata.cf_priv) {
698 ret = EINVAL; /* XXX is this the right error? */
699 goto bail;
700 }
701 free(inbuf, M_TEMP);
702
703 bufq_alloc(&dksc->sc_bufq, "fcfs", 0);
704
705 cs->sc_data = malloc(MAXPHYS, M_DEVBUF, M_WAITOK);
706 cs->sc_data_used = 0;
707
708 /* Attach the disk. */
709 dk_attach(dksc);
710 disk_attach(&dksc->sc_dkdev);
711
712 disk_set_info(dksc->sc_dev, &dksc->sc_dkdev, NULL);
713
714 /* Discover wedges on this disk. */
715 dkwedge_discover(&dksc->sc_dkdev);
716
717 return 0;
718
719 bail:
720 free(inbuf, M_TEMP);
721 (void)vn_close(vp, FREAD|FWRITE, l->l_cred);
722 return ret;
723 }
724
725 /* ARGSUSED */
726 static int
727 cgd_ioctl_clr(struct cgd_softc *cs, struct lwp *l)
728 {
729 struct dk_softc *dksc = &cs->sc_dksc;
730
731 if (!DK_ATTACHED(dksc))
732 return ENXIO;
733
734 /* Delete all of our wedges. */
735 dkwedge_delall(&dksc->sc_dkdev);
736
737 /* Kill off any queued buffers. */
738 dk_drain(dksc);
739 bufq_free(dksc->sc_bufq);
740
741 (void)vn_close(cs->sc_tvn, FREAD|FWRITE, l->l_cred);
742 cs->sc_cfuncs->cf_destroy(cs->sc_cdata.cf_priv);
743 free(cs->sc_tpath, M_DEVBUF);
744 free(cs->sc_data, M_DEVBUF);
745 cs->sc_data_used = 0;
746 dk_detach(dksc);
747 disk_detach(&dksc->sc_dkdev);
748
749 return 0;
750 }
751
752 static int
753 cgd_ioctl_get(dev_t dev, void *data, struct lwp *l)
754 {
755 struct cgd_softc *cs = getcgd_softc(dev);
756 struct cgd_user *cgu;
757 int unit;
758 struct dk_softc *dksc = &cs->sc_dksc;
759
760 unit = CGDUNIT(dev);
761 cgu = (struct cgd_user *)data;
762
763 DPRINTF_FOLLOW(("cgd_ioctl_get(0x%"PRIx64", %d, %p, %p)\n",
764 dev, unit, data, l));
765
766 if (cgu->cgu_unit == -1)
767 cgu->cgu_unit = unit;
768
769 if (cgu->cgu_unit < 0)
770 return EINVAL; /* XXX: should this be ENXIO? */
771
772 cs = device_lookup_private(&cgd_cd, unit);
773 if (cs == NULL || !DK_ATTACHED(dksc)) {
774 cgu->cgu_dev = 0;
775 cgu->cgu_alg[0] = '\0';
776 cgu->cgu_blocksize = 0;
777 cgu->cgu_mode = 0;
778 cgu->cgu_keylen = 0;
779 }
780 else {
781 cgu->cgu_dev = cs->sc_tdev;
782 strlcpy(cgu->cgu_alg, cs->sc_cfuncs->cf_name,
783 sizeof(cgu->cgu_alg));
784 cgu->cgu_blocksize = cs->sc_cdata.cf_blocksize;
785 cgu->cgu_mode = cs->sc_cdata.cf_mode;
786 cgu->cgu_keylen = cs->sc_cdata.cf_keylen;
787 }
788 return 0;
789 }
790
791 static int
792 cgdinit(struct cgd_softc *cs, const char *cpath, struct vnode *vp,
793 struct lwp *l)
794 {
795 struct disk_geom *dg;
796 int ret;
797 char *tmppath;
798 uint64_t psize;
799 unsigned secsize;
800 struct dk_softc *dksc = &cs->sc_dksc;
801
802 cs->sc_tvn = vp;
803 cs->sc_tpath = NULL;
804
805 tmppath = malloc(MAXPATHLEN, M_TEMP, M_WAITOK);
806 ret = copyinstr(cpath, tmppath, MAXPATHLEN, &cs->sc_tpathlen);
807 if (ret)
808 goto bail;
809 cs->sc_tpath = malloc(cs->sc_tpathlen, M_DEVBUF, M_WAITOK);
810 memcpy(cs->sc_tpath, tmppath, cs->sc_tpathlen);
811
812 cs->sc_tdev = vp->v_rdev;
813
814 if ((ret = getdisksize(vp, &psize, &secsize)) != 0)
815 goto bail;
816
817 if (psize == 0) {
818 ret = ENODEV;
819 goto bail;
820 }
821
822 /*
823 * XXX here we should probe the underlying device. If we
824 * are accessing a partition of type RAW_PART, then
825 * we should populate our initial geometry with the
826 * geometry that we discover from the device.
827 */
828 dg = &dksc->sc_dkdev.dk_geom;
829 memset(dg, 0, sizeof(*dg));
830 dg->dg_secperunit = psize;
831 dg->dg_secsize = secsize;
832 dg->dg_ntracks = 1;
833 dg->dg_nsectors = 1024 * 1024 / dg->dg_secsize;
834 dg->dg_ncylinders = dg->dg_secperunit / dg->dg_nsectors;
835
836 bail:
837 free(tmppath, M_TEMP);
838 if (ret && cs->sc_tpath)
839 free(cs->sc_tpath, M_DEVBUF);
840 return ret;
841 }
842
843 /*
844 * Our generic cipher entry point. This takes care of the
845 * IV mode and passes off the work to the specific cipher.
846 * We implement here the IV method ``encrypted block
847 * number''.
848 *
849 * For the encryption case, we accomplish this by setting
850 * up a struct uio where the first iovec of the source is
851 * the blocknumber and the first iovec of the dest is a
852 * sink. We then call the cipher with an IV of zero, and
853 * the right thing happens.
854 *
855 * For the decryption case, we use the same basic mechanism
856 * for symmetry, but we encrypt the block number in the
857 * first iovec.
858 *
859 * We mainly do this to avoid requiring the definition of
860 * an ECB mode.
861 *
862 * XXXrcd: for now we rely on our own crypto framework defined
863 * in dev/cgd_crypto.c. This will change when we
864 * get a generic kernel crypto framework.
865 */
866
867 static void
868 blkno2blkno_buf(char *sbuf, daddr_t blkno)
869 {
870 int i;
871
872 /* Set up the blkno in blkno_buf, here we do not care much
873 * about the final layout of the information as long as we
874 * can guarantee that each sector will have a different IV
875 * and that the endianness of the machine will not affect
876 * the representation that we have chosen.
877 *
878 * We choose this representation, because it does not rely
879 * on the size of buf (which is the blocksize of the cipher),
880 * but allows daddr_t to grow without breaking existing
881 * disks.
882 *
883 * Note that blkno2blkno_buf does not take a size as input,
884 * and hence must be called on a pre-zeroed buffer of length
885 * greater than or equal to sizeof(daddr_t).
886 */
887 for (i=0; i < sizeof(daddr_t); i++) {
888 *sbuf++ = blkno & 0xff;
889 blkno >>= 8;
890 }
891 }
892
893 static void
894 cgd_cipher(struct cgd_softc *cs, void *dstv, void *srcv,
895 size_t len, daddr_t blkno, size_t secsize, int dir)
896 {
897 char *dst = dstv;
898 char *src = srcv;
899 cfunc_cipher *cipher = cs->sc_cfuncs->cf_cipher;
900 struct uio dstuio;
901 struct uio srcuio;
902 struct iovec dstiov[2];
903 struct iovec srciov[2];
904 size_t blocksize = cs->sc_cdata.cf_blocksize;
905 size_t todo;
906 char sink[CGD_MAXBLOCKSIZE];
907 char zero_iv[CGD_MAXBLOCKSIZE];
908 char blkno_buf[CGD_MAXBLOCKSIZE];
909
910 DPRINTF_FOLLOW(("cgd_cipher() dir=%d\n", dir));
911
912 DIAGCONDPANIC(len % blocksize != 0,
913 ("cgd_cipher: len %% blocksize != 0"));
914
915 /* ensure that sizeof(daddr_t) <= blocksize (for encblkno IVing) */
916 DIAGCONDPANIC(sizeof(daddr_t) > blocksize,
917 ("cgd_cipher: sizeof(daddr_t) > blocksize"));
918
919 memset(zero_iv, 0x0, blocksize);
920
921 dstuio.uio_iov = dstiov;
922 dstuio.uio_iovcnt = 2;
923
924 srcuio.uio_iov = srciov;
925 srcuio.uio_iovcnt = 2;
926
927 dstiov[0].iov_base = sink;
928 dstiov[0].iov_len = blocksize;
929 srciov[0].iov_base = blkno_buf;
930 srciov[0].iov_len = blocksize;
931
932 for (; len > 0; len -= todo) {
933 todo = MIN(len, secsize);
934
935 dstiov[1].iov_base = dst;
936 srciov[1].iov_base = src;
937 dstiov[1].iov_len = todo;
938 srciov[1].iov_len = todo;
939
940 memset(blkno_buf, 0x0, blocksize);
941 blkno2blkno_buf(blkno_buf, blkno);
942 if (dir == CGD_CIPHER_DECRYPT) {
943 dstuio.uio_iovcnt = 1;
944 srcuio.uio_iovcnt = 1;
945 IFDEBUG(CGDB_CRYPTO, hexprint("step 0: blkno_buf",
946 blkno_buf, blocksize));
947 cipher(cs->sc_cdata.cf_priv, &dstuio, &srcuio,
948 zero_iv, CGD_CIPHER_ENCRYPT);
949 memcpy(blkno_buf, sink, blocksize);
950 dstuio.uio_iovcnt = 2;
951 srcuio.uio_iovcnt = 2;
952 }
953
954 IFDEBUG(CGDB_CRYPTO, hexprint("step 1: blkno_buf",
955 blkno_buf, blocksize));
956 cipher(cs->sc_cdata.cf_priv, &dstuio, &srcuio, zero_iv, dir);
957 IFDEBUG(CGDB_CRYPTO, hexprint("step 2: sink",
958 sink, blocksize));
959
960 dst += todo;
961 src += todo;
962 blkno++;
963 }
964 }
965
966 #ifdef DEBUG
967 static void
968 hexprint(const char *start, void *buf, int len)
969 {
970 char *c = buf;
971
972 DIAGCONDPANIC(len < 0, ("hexprint: called with len < 0"));
973 printf("%s: len=%06d 0x", start, len);
974 while (len--)
975 printf("%02x", (unsigned char) *c++);
976 }
977 #endif
978
979 MODULE(MODULE_CLASS_DRIVER, cgd, "dk_subr");
980
981 #ifdef _MODULE
982 CFDRIVER_DECL(cgd, DV_DISK, NULL);
983 #endif
984
985 static int
986 cgd_modcmd(modcmd_t cmd, void *arg)
987 {
988 int error = 0;
989
990 #ifdef _MODULE
991 devmajor_t bmajor = -1, cmajor = -1;
992 #endif
993
994 switch (cmd) {
995 case MODULE_CMD_INIT:
996 #ifdef _MODULE
997 error = config_cfdriver_attach(&cgd_cd);
998 if (error)
999 break;
1000
1001 error = config_cfattach_attach(cgd_cd.cd_name, &cgd_ca);
1002 if (error) {
1003 config_cfdriver_detach(&cgd_cd);
1004 aprint_error("%s: unable to register cfattach\n",
1005 cgd_cd.cd_name);
1006 break;
1007 }
1008
1009 error = devsw_attach("cgd", &cgd_bdevsw, &bmajor,
1010 &cgd_cdevsw, &cmajor);
1011 if (error) {
1012 config_cfattach_detach(cgd_cd.cd_name, &cgd_ca);
1013 config_cfdriver_detach(&cgd_cd);
1014 break;
1015 }
1016 #endif
1017 break;
1018
1019 case MODULE_CMD_FINI:
1020 #ifdef _MODULE
1021 error = config_cfattach_detach(cgd_cd.cd_name, &cgd_ca);
1022 if (error)
1023 break;
1024 config_cfdriver_detach(&cgd_cd);
1025 devsw_detach(&cgd_bdevsw, &cgd_cdevsw);
1026 #endif
1027 break;
1028
1029 case MODULE_CMD_STAT:
1030 return ENOTTY;
1031
1032 default:
1033 return ENOTTY;
1034 }
1035
1036 return error;
1037 }
1038