Home | History | Annotate | Line # | Download | only in tc
if_le_ioasic.c revision 1.26
      1 /*	$NetBSD: if_le_ioasic.c,v 1.26 2006/03/30 18:18:39 thorpej Exp $	*/
      2 
      3 /*
      4  * Copyright (c) 1996 Carnegie-Mellon University.
      5  * All rights reserved.
      6  *
      7  * Author: Chris G. Demetriou
      8  *
      9  * Permission to use, copy, modify and distribute this software and
     10  * its documentation is hereby granted, provided that both the copyright
     11  * notice and this permission notice appear in all copies of the
     12  * software, derivative works or modified versions, and any portions
     13  * thereof, and that both notices appear in supporting documentation.
     14  *
     15  * CARNEGIE MELLON ALLOWS FREE USE OF THIS SOFTWARE IN ITS "AS IS"
     16  * CONDITION.  CARNEGIE MELLON DISCLAIMS ANY LIABILITY OF ANY KIND
     17  * FOR ANY DAMAGES WHATSOEVER RESULTING FROM THE USE OF THIS SOFTWARE.
     18  *
     19  * Carnegie Mellon requests users of this software to return to
     20  *
     21  *  Software Distribution Coordinator  or  Software.Distribution (at) CS.CMU.EDU
     22  *  School of Computer Science
     23  *  Carnegie Mellon University
     24  *  Pittsburgh PA 15213-3890
     25  *
     26  * any improvements or extensions that they make and grant Carnegie the
     27  * rights to redistribute these changes.
     28  */
     29 
     30 /*
     31  * LANCE on DEC IOCTL ASIC.
     32  */
     33 
     34 #include <sys/cdefs.h>
     35 __KERNEL_RCSID(0, "$NetBSD: if_le_ioasic.c,v 1.26 2006/03/30 18:18:39 thorpej Exp $");
     36 
     37 #include "opt_inet.h"
     38 
     39 #include <sys/param.h>
     40 #include <sys/systm.h>
     41 #include <sys/mbuf.h>
     42 #include <sys/syslog.h>
     43 #include <sys/socket.h>
     44 #include <sys/device.h>
     45 
     46 #include <net/if.h>
     47 #include <net/if_ether.h>
     48 #include <net/if_media.h>
     49 
     50 #ifdef INET
     51 #include <netinet/in.h>
     52 #include <netinet/if_inarp.h>
     53 #endif
     54 
     55 #include <dev/ic/lancereg.h>
     56 #include <dev/ic/lancevar.h>
     57 #include <dev/ic/am7990reg.h>
     58 #include <dev/ic/am7990var.h>
     59 
     60 #include <dev/tc/if_levar.h>
     61 #include <dev/tc/tcvar.h>
     62 #include <dev/tc/ioasicreg.h>
     63 #include <dev/tc/ioasicvar.h>
     64 
     65 struct le_ioasic_softc {
     66 	struct	am7990_softc sc_am7990;	/* glue to MI code */
     67 	struct	lereg1 *sc_r1;		/* LANCE registers */
     68 	/* XXX must match with le_softc of if_levar.h XXX */
     69 
     70 	bus_dma_tag_t sc_dmat;		/* bus dma tag */
     71 	bus_dmamap_t sc_dmamap;		/* bus dmamap */
     72 };
     73 
     74 static int  le_ioasic_match(struct device *, struct cfdata *, void *);
     75 static void le_ioasic_attach(struct device *, struct device *, void *);
     76 
     77 CFATTACH_DECL(le_ioasic, sizeof(struct le_softc),
     78     le_ioasic_match, le_ioasic_attach, NULL, NULL);
     79 
     80 static void le_ioasic_copytobuf_gap2(struct lance_softc *, void *, int, int);
     81 static void le_ioasic_copyfrombuf_gap2(struct lance_softc *, void *, int, int);
     82 static void le_ioasic_copytobuf_gap16(struct lance_softc *, void *, int, int);
     83 static void le_ioasic_copyfrombuf_gap16(struct lance_softc *, void *,
     84 	    int, int);
     85 static void le_ioasic_zerobuf_gap16(struct lance_softc *, int, int);
     86 
     87 static int
     88 le_ioasic_match(parent, match, aux)
     89 	struct device *parent;
     90 	struct cfdata *match;
     91 	void *aux;
     92 {
     93 	struct ioasicdev_attach_args *d = aux;
     94 
     95 	if (strncmp("PMAD-BA ", d->iada_modname, TC_ROM_LLEN) != 0)
     96 		return 0;
     97 
     98 	return 1;
     99 }
    100 
    101 /* IOASIC LANCE DMA needs 128KB boundary aligned 128KB chunk */
    102 #define	LE_IOASIC_MEMSIZE	(128*1024)
    103 #define	LE_IOASIC_MEMALIGN	(128*1024)
    104 
    105 static void
    106 le_ioasic_attach(parent, self, aux)
    107 	struct device *parent, *self;
    108 	void *aux;
    109 {
    110 	struct le_ioasic_softc *sc = device_private(self);
    111 	struct ioasicdev_attach_args *d = aux;
    112 	struct lance_softc *le = &sc->sc_am7990.lsc;
    113 	struct ioasic_softc *iosc = device_private(parent);
    114 	bus_space_tag_t ioasic_bst;
    115 	bus_space_handle_t ioasic_bsh;
    116 	bus_dma_tag_t dmat;
    117 	bus_dma_segment_t seg;
    118 	tc_addr_t tca;
    119 	u_int32_t ssr;
    120 	int rseg;
    121 	caddr_t le_iomem;
    122 
    123 	ioasic_bst = iosc->sc_bst;
    124 	ioasic_bsh = iosc->sc_bsh;
    125 	dmat = sc->sc_dmat = iosc->sc_dmat;
    126 	/*
    127 	 * Allocate a DMA area for the chip.
    128 	 */
    129 	if (bus_dmamem_alloc(dmat, LE_IOASIC_MEMSIZE, LE_IOASIC_MEMALIGN,
    130 	    0, &seg, 1, &rseg, BUS_DMA_NOWAIT)) {
    131 		printf("can't allocate DMA area for LANCE\n");
    132 		return;
    133 	}
    134 	if (bus_dmamem_map(dmat, &seg, rseg, LE_IOASIC_MEMSIZE,
    135 	    &le_iomem, BUS_DMA_NOWAIT|BUS_DMA_COHERENT)) {
    136 		printf("can't map DMA area for LANCE\n");
    137 		bus_dmamem_free(dmat, &seg, rseg);
    138 		return;
    139 	}
    140 	/*
    141 	 * Create and load the DMA map for the DMA area.
    142 	 */
    143 	if (bus_dmamap_create(dmat, LE_IOASIC_MEMSIZE, 1,
    144 	    LE_IOASIC_MEMSIZE, 0, BUS_DMA_NOWAIT, &sc->sc_dmamap)) {
    145 		printf("can't create DMA map\n");
    146 		goto bad;
    147 	}
    148 	if (bus_dmamap_load(dmat, sc->sc_dmamap,
    149 	    le_iomem, LE_IOASIC_MEMSIZE, NULL, BUS_DMA_NOWAIT)) {
    150 		printf("can't load DMA map\n");
    151 		goto bad;
    152 	}
    153 	/*
    154 	 * Bind 128KB buffer with IOASIC DMA.
    155 	 */
    156 	tca = IOASIC_DMA_ADDR(sc->sc_dmamap->dm_segs[0].ds_addr);
    157 	bus_space_write_4(ioasic_bst, ioasic_bsh, IOASIC_LANCE_DMAPTR, tca);
    158 	ssr = bus_space_read_4(ioasic_bst, ioasic_bsh, IOASIC_CSR);
    159 	ssr |= IOASIC_CSR_DMAEN_LANCE;
    160 	bus_space_write_4(ioasic_bst, ioasic_bsh, IOASIC_CSR, ssr);
    161 
    162 	sc->sc_r1 = (struct lereg1 *)
    163 		TC_DENSE_TO_SPARSE(TC_PHYS_TO_UNCACHED(d->iada_addr));
    164 	le->sc_mem = (void *)TC_PHYS_TO_UNCACHED(le_iomem);
    165 	le->sc_copytodesc = le_ioasic_copytobuf_gap2;
    166 	le->sc_copyfromdesc = le_ioasic_copyfrombuf_gap2;
    167 	le->sc_copytobuf = le_ioasic_copytobuf_gap16;
    168 	le->sc_copyfrombuf = le_ioasic_copyfrombuf_gap16;
    169 	le->sc_zerobuf = le_ioasic_zerobuf_gap16;
    170 
    171 	dec_le_common_attach(&sc->sc_am7990,
    172 	    (u_char *)iosc->sc_base + IOASIC_SLOT_2_START);
    173 
    174 	ioasic_intr_establish(parent, d->iada_cookie, TC_IPL_NET,
    175 	    am7990_intr, sc);
    176 	return;
    177 
    178  bad:
    179 	bus_dmamem_unmap(dmat, le_iomem, LE_IOASIC_MEMSIZE);
    180 	bus_dmamem_free(dmat, &seg, rseg);
    181 }
    182 
    183 /*
    184  * Special memory access functions needed by ioasic-attached LANCE
    185  * chips.
    186  */
    187 
    188 /*
    189  * gap2: two bytes of data followed by two bytes of pad.
    190  *
    191  * Buffers must be 4-byte aligned.  The code doesn't worry about
    192  * doing an extra byte.
    193  */
    194 
    195 void
    196 le_ioasic_copytobuf_gap2(sc, fromv, boff, len)
    197 	struct lance_softc *sc;
    198 	void *fromv;
    199 	int boff;
    200 	int len;
    201 {
    202 	volatile caddr_t buf = sc->sc_mem;
    203 	caddr_t from = fromv;
    204 	volatile u_int16_t *bptr;
    205 
    206 	if (boff & 0x1) {
    207 		/* handle unaligned first byte */
    208 		bptr = ((volatile u_int16_t *)buf) + (boff - 1);
    209 		*bptr = (*from++ << 8) | (*bptr & 0xff);
    210 		bptr += 2;
    211 		len--;
    212 	} else
    213 		bptr = ((volatile u_int16_t *)buf) + boff;
    214 	while (len > 1) {
    215 		*bptr = (from[1] << 8) | (from[0] & 0xff);
    216 		bptr += 2;
    217 		from += 2;
    218 		len -= 2;
    219 	}
    220 	if (len == 1)
    221 		*bptr = (u_int16_t)*from;
    222 }
    223 
    224 void
    225 le_ioasic_copyfrombuf_gap2(sc, tov, boff, len)
    226 	struct lance_softc *sc;
    227 	void *tov;
    228 	int boff, len;
    229 {
    230 	volatile caddr_t buf = sc->sc_mem;
    231 	caddr_t to = tov;
    232 	volatile u_int16_t *bptr;
    233 	u_int16_t tmp;
    234 
    235 	if (boff & 0x1) {
    236 		/* handle unaligned first byte */
    237 		bptr = ((volatile u_int16_t *)buf) + (boff - 1);
    238 		*to++ = (*bptr >> 8) & 0xff;
    239 		bptr += 2;
    240 		len--;
    241 	} else
    242 		bptr = ((volatile u_int16_t *)buf) + boff;
    243 	while (len > 1) {
    244 		tmp = *bptr;
    245 		*to++ = tmp & 0xff;
    246 		*to++ = (tmp >> 8) & 0xff;
    247 		bptr += 2;
    248 		len -= 2;
    249 	}
    250 	if (len == 1)
    251 		*to = *bptr & 0xff;
    252 }
    253 
    254 /*
    255  * gap16: 16 bytes of data followed by 16 bytes of pad.
    256  *
    257  * Buffers must be 32-byte aligned.
    258  */
    259 
    260 void
    261 le_ioasic_copytobuf_gap16(sc, fromv, boff, len)
    262 	struct lance_softc *sc;
    263 	void *fromv;
    264 	int boff;
    265 	int len;
    266 {
    267 	volatile caddr_t buf = sc->sc_mem;
    268 	caddr_t from = fromv;
    269 	caddr_t bptr;
    270 
    271 	bptr = buf + ((boff << 1) & ~0x1f);
    272 	boff &= 0xf;
    273 
    274 	/*
    275 	 * Dispose of boff so destination of subsequent copies is
    276 	 * 16-byte aligned.
    277 	 */
    278 	if (boff) {
    279 		int xfer;
    280 		xfer = min(len, 16 - boff);
    281 		bcopy(from, bptr + boff, xfer);
    282 		from += xfer;
    283 		bptr += 32;
    284 		len -= xfer;
    285 	}
    286 
    287 	/* Destination of  copies is now 16-byte aligned. */
    288 	if (len >= 16)
    289 		switch ((u_long)from & (sizeof(u_int32_t) -1)) {
    290 		case 2:
    291 			/*  Ethernet headers make this the dominant case. */
    292 		do {
    293 			u_int32_t *dst = (u_int32_t*)bptr;
    294 			u_int16_t t0;
    295 			u_int32_t t1,  t2, t3, t4;
    296 
    297 			/* read from odd-16-bit-aligned, cached src */
    298 			t0 = *(u_int16_t*)from;
    299 			t1 = *(u_int32_t*)(from+2);
    300 			t2 = *(u_int32_t*)(from+6);
    301 			t3 = *(u_int32_t*)(from+10);
    302 			t4 = *(u_int16_t*)(from+14);
    303 
    304 			/* DMA buffer is uncached on mips */
    305 			dst[0] =         t0 |  (t1 << 16);
    306 			dst[1] = (t1 >> 16) |  (t2 << 16);
    307 			dst[2] = (t2 >> 16) |  (t3 << 16);
    308 			dst[3] = (t3 >> 16) |  (t4 << 16);
    309 
    310 			from += 16;
    311 			bptr += 32;
    312 			len -= 16;
    313 		} while (len >= 16);
    314 		break;
    315 
    316 		case 0:
    317 		do {
    318 			u_int32_t *src = (u_int32_t*)from;
    319 			u_int32_t *dst = (u_int32_t*)bptr;
    320 			u_int32_t t0, t1, t2, t3;
    321 
    322 			t0 = src[0]; t1 = src[1]; t2 = src[2]; t3 = src[3];
    323 			dst[0] = t0; dst[1] = t1; dst[2] = t2; dst[3] = t3;
    324 
    325 			from += 16;
    326 			bptr += 32;
    327 			len -= 16;
    328 		} while (len >= 16);
    329 		break;
    330 
    331 		default:
    332 		/* Does odd-aligned case ever happen? */
    333 		do {
    334 			bcopy(from, bptr, 16);
    335 			from += 16;
    336 			bptr += 32;
    337 			len -= 16;
    338 		} while (len >= 16);
    339 		break;
    340 	}
    341 	if (len)
    342 		bcopy(from, bptr, len);
    343 }
    344 
    345 void
    346 le_ioasic_copyfrombuf_gap16(sc, tov, boff, len)
    347 	struct lance_softc *sc;
    348 	void *tov;
    349 	int boff, len;
    350 {
    351 	volatile caddr_t buf = sc->sc_mem;
    352 	caddr_t to = tov;
    353 	caddr_t bptr;
    354 
    355 	bptr = buf + ((boff << 1) & ~0x1f);
    356 	boff &= 0xf;
    357 
    358 	/* Dispose of boff. source of copy is subsequently 16-byte aligned. */
    359 	if (boff) {
    360 		int xfer;
    361 		xfer = min(len, 16 - boff);
    362 		bcopy(bptr+boff, to, xfer);
    363 		to += xfer;
    364 		bptr += 32;
    365 		len -= xfer;
    366 	}
    367 	if (len >= 16)
    368 	switch ((u_long)to & (sizeof(u_int32_t) -1)) {
    369 	case 2:
    370 		/*
    371 		 * to is aligned to an odd 16-bit boundary.  Ethernet headers
    372 		 * make this the dominant case (98% or more).
    373 		 */
    374 		do {
    375 			u_int32_t *src = (u_int32_t*)bptr;
    376 			u_int32_t t0, t1, t2, t3;
    377 
    378 			/* read from uncached aligned DMA buf */
    379 			t0 = src[0]; t1 = src[1]; t2 = src[2]; t3 = src[3];
    380 
    381 			/* write to odd-16-bit-word aligned dst */
    382 			*(u_int16_t *) (to+0)  = (u_short)  t0;
    383 			*(u_int32_t *) (to+2)  = (t0 >> 16) |  (t1 << 16);
    384 			*(u_int32_t *) (to+6)  = (t1 >> 16) |  (t2 << 16);
    385 			*(u_int32_t *) (to+10) = (t2 >> 16) |  (t3 << 16);
    386 			*(u_int16_t *) (to+14) = (t3 >> 16);
    387 			bptr += 32;
    388 			to += 16;
    389 			len -= 16;
    390 		} while (len > 16);
    391 		break;
    392 	case 0:
    393 		/* 32-bit aligned aligned copy. Rare. */
    394 		do {
    395 			u_int32_t *src = (u_int32_t*)bptr;
    396 			u_int32_t *dst = (u_int32_t*)to;
    397 			u_int32_t t0, t1, t2, t3;
    398 
    399 			t0 = src[0]; t1 = src[1]; t2 = src[2]; t3 = src[3];
    400 			dst[0] = t0; dst[1] = t1; dst[2] = t2; dst[3] = t3;
    401 			to += 16;
    402 			bptr += 32;
    403 			len -= 16;
    404 		} while (len  > 16);
    405 		break;
    406 
    407 	/* XXX Does odd-byte-aligned case ever happen? */
    408 	default:
    409 		do {
    410 			bcopy(bptr, to, 16);
    411 			to += 16;
    412 			bptr += 32;
    413 			len -= 16;
    414 		} while (len  > 16);
    415 		break;
    416 	}
    417 	if (len)
    418 		bcopy(bptr, to, len);
    419 }
    420 
    421 void
    422 le_ioasic_zerobuf_gap16(sc, boff, len)
    423 	struct lance_softc *sc;
    424 	int boff, len;
    425 {
    426 	volatile caddr_t buf = sc->sc_mem;
    427 	caddr_t bptr;
    428 	int xfer;
    429 
    430 	bptr = buf + ((boff << 1) & ~0x1f);
    431 	boff &= 0xf;
    432 	xfer = min(len, 16 - boff);
    433 	while (len > 0) {
    434 		bzero(bptr + boff, xfer);
    435 		bptr += 32;
    436 		boff = 0;
    437 		len -= xfer;
    438 		xfer = min(len, 16);
    439 	}
    440 }
    441