if_kse.c revision 1.2.4.2 1 1.2.4.2 yamt /* $NetBSD: if_kse.c,v 1.2.4.2 2006/12/30 20:48:44 yamt Exp $ */
2 1.2.4.2 yamt
3 1.2.4.2 yamt /*
4 1.2.4.2 yamt * Copyright (c) 2006 Tohru Nishimura
5 1.2.4.2 yamt *
6 1.2.4.2 yamt * Redistribution and use in source and binary forms, with or without
7 1.2.4.2 yamt * modification, are permitted provided that the following conditions
8 1.2.4.2 yamt * are met:
9 1.2.4.2 yamt * 1. Redistributions of source code must retain the above copyright
10 1.2.4.2 yamt * notice, this list of conditions and the following disclaimer.
11 1.2.4.2 yamt * 2. Redistributions in binary form must reproduce the above copyright
12 1.2.4.2 yamt * notice, this list of conditions and the following disclaimer in the
13 1.2.4.2 yamt * documentation and/or other materials provided with the distribution.
14 1.2.4.2 yamt * 3. All advertising materials mentioning features or use of this software
15 1.2.4.2 yamt * must display the following acknowledgement:
16 1.2.4.2 yamt * This product includes software developed by Tohru Nishimura.
17 1.2.4.2 yamt * 4. The name of the author may not be used to endorse or promote products
18 1.2.4.2 yamt * derived from this software without specific prior written permission.
19 1.2.4.2 yamt *
20 1.2.4.2 yamt * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
21 1.2.4.2 yamt * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
22 1.2.4.2 yamt * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
23 1.2.4.2 yamt * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
24 1.2.4.2 yamt * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
25 1.2.4.2 yamt * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
26 1.2.4.2 yamt * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
27 1.2.4.2 yamt * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
28 1.2.4.2 yamt * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
29 1.2.4.2 yamt * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
30 1.2.4.2 yamt */
31 1.2.4.2 yamt
32 1.2.4.2 yamt #include <sys/cdefs.h>
33 1.2.4.2 yamt __KERNEL_RCSID(0, "$NetBSD: if_kse.c,v 1.2.4.2 2006/12/30 20:48:44 yamt Exp $");
34 1.2.4.2 yamt
35 1.2.4.2 yamt #include "bpfilter.h"
36 1.2.4.2 yamt
37 1.2.4.2 yamt #include <sys/param.h>
38 1.2.4.2 yamt #include <sys/systm.h>
39 1.2.4.2 yamt #include <sys/callout.h>
40 1.2.4.2 yamt #include <sys/mbuf.h>
41 1.2.4.2 yamt #include <sys/malloc.h>
42 1.2.4.2 yamt #include <sys/kernel.h>
43 1.2.4.2 yamt #include <sys/ioctl.h>
44 1.2.4.2 yamt #include <sys/errno.h>
45 1.2.4.2 yamt #include <sys/device.h>
46 1.2.4.2 yamt #include <sys/queue.h>
47 1.2.4.2 yamt
48 1.2.4.2 yamt #include <machine/endian.h>
49 1.2.4.2 yamt #include <machine/bus.h>
50 1.2.4.2 yamt #include <machine/intr.h>
51 1.2.4.2 yamt
52 1.2.4.2 yamt #include <net/if.h>
53 1.2.4.2 yamt #include <net/if_media.h>
54 1.2.4.2 yamt #include <net/if_dl.h>
55 1.2.4.2 yamt #include <net/if_ether.h>
56 1.2.4.2 yamt
57 1.2.4.2 yamt #if NBPFILTER > 0
58 1.2.4.2 yamt #include <net/bpf.h>
59 1.2.4.2 yamt #endif
60 1.2.4.2 yamt
61 1.2.4.2 yamt #include <dev/pci/pcivar.h>
62 1.2.4.2 yamt #include <dev/pci/pcireg.h>
63 1.2.4.2 yamt #include <dev/pci/pcidevs.h>
64 1.2.4.2 yamt
65 1.2.4.2 yamt #define CSR_READ_4(sc, off) \
66 1.2.4.2 yamt bus_space_read_4(sc->sc_st, sc->sc_sh, off)
67 1.2.4.2 yamt #define CSR_WRITE_4(sc, off, val) \
68 1.2.4.2 yamt bus_space_write_4(sc->sc_st, sc->sc_sh, off, val)
69 1.2.4.2 yamt #define CSR_READ_2(sc, off) \
70 1.2.4.2 yamt bus_space_read_2(sc->sc_st, sc->sc_sh, off)
71 1.2.4.2 yamt #define CSR_WRITE_2(sc, off, val) \
72 1.2.4.2 yamt bus_space_write_2(sc->sc_st, sc->sc_sh, off, val)
73 1.2.4.2 yamt
74 1.2.4.2 yamt #define MDTXC 0x000 /* DMA transmit control */
75 1.2.4.2 yamt #define MDRXC 0x004 /* DMA receive control */
76 1.2.4.2 yamt #define MDTSC 0x008 /* DMA transmit start */
77 1.2.4.2 yamt #define MDRSC 0x00c /* DMA receive start */
78 1.2.4.2 yamt #define TDLB 0x010 /* transmit descriptor list base */
79 1.2.4.2 yamt #define RDLB 0x014 /* receive descriptor list base */
80 1.2.4.2 yamt #define INTEN 0x028 /* interrupt enable */
81 1.2.4.2 yamt #define INTST 0x02c /* interrupt status */
82 1.2.4.2 yamt #define MARL 0x200 /* MAC address low */
83 1.2.4.2 yamt #define MARM 0x202 /* MAC address middle */
84 1.2.4.2 yamt #define MARH 0x204 /* MAC address high */
85 1.2.4.2 yamt #define GRR 0x216 /* global reset */
86 1.2.4.2 yamt #define CIDR 0x400 /* chip ID and enable */
87 1.2.4.2 yamt #define CGCR 0x40a /* chip global control */
88 1.2.4.2 yamt #define P1CR4 0x512 /* port 1 control 4 */
89 1.2.4.2 yamt #define P1SR 0x514 /* port 1 status */
90 1.2.4.2 yamt
91 1.2.4.2 yamt #define TXC_BS_MSK 0x3f000000 /* burst size */
92 1.2.4.2 yamt #define TXC_BS_SFT (24) /* 1,2,4,8,16,32 or 0 for unlimited */
93 1.2.4.2 yamt #define TXC_UCG (1U<<18) /* generate UDP checksum */
94 1.2.4.2 yamt #define TXC_TCG (1U<<17) /* generate TCP checksum */
95 1.2.4.2 yamt #define TXC_ICG (1U<<16) /* generate IP checksum */
96 1.2.4.2 yamt #define TXC_FCE (1U<<9) /* enable flowcontrol */
97 1.2.4.2 yamt #define TXC_EP (1U<<2) /* enable automatic padding */
98 1.2.4.2 yamt #define TXC_AC (1U<<1) /* add CRC to frame */
99 1.2.4.2 yamt #define TXC_TEN (1) /* enable DMA to run */
100 1.2.4.2 yamt
101 1.2.4.2 yamt #define RXC_BS_MSK 0x3f000000 /* burst size */
102 1.2.4.2 yamt #define RXC_BS_SFT (24) /* 1,2,4,8,16,32 or 0 for unlimited */
103 1.2.4.2 yamt #define RXC_UCG (1U<<18) /* run UDP checksum */
104 1.2.4.2 yamt #define RXC_TCG (1U<<17) /* run TDP checksum */
105 1.2.4.2 yamt #define RXC_ICG (1U<<16) /* run IP checksum */
106 1.2.4.2 yamt #define RXC_FCE (1U<<9) /* enable flowcontrol */
107 1.2.4.2 yamt #define RXC_RB (1U<<6) /* receive broadcast frame */
108 1.2.4.2 yamt #define RXC_RM (1U<<5) /* receive multicast frame */
109 1.2.4.2 yamt #define RXC_RU (1U<<4) /* receive unicast frame */
110 1.2.4.2 yamt #define RXC_RE (1U<<3) /* accept error frame */
111 1.2.4.2 yamt #define RXC_RA (1U<<2) /* receive all frame */
112 1.2.4.2 yamt #define RXC_MA (1U<<1) /* receive through hash filter */
113 1.2.4.2 yamt #define RXC_REN (1) /* enable DMA to run */
114 1.2.4.2 yamt
115 1.2.4.2 yamt #define INT_DMLCS (1U<<31) /* link status change */
116 1.2.4.2 yamt #define INT_DMTS (1U<<30) /* sending desc. has posted Tx done */
117 1.2.4.2 yamt #define INT_DMRS (1U<<29) /* frame was received */
118 1.2.4.2 yamt #define INT_DMRBUS (1U<<27) /* Rx descriptor pool is full */
119 1.2.4.2 yamt
120 1.2.4.2 yamt #define T0_OWN (1U<<31) /* desc is ready to Tx */
121 1.2.4.2 yamt
122 1.2.4.2 yamt #define R0_OWN (1U<<31) /* desc is empty */
123 1.2.4.2 yamt #define R0_FS (1U<<30) /* first segment of frame */
124 1.2.4.2 yamt #define R0_LS (1U<<29) /* last segment of frame */
125 1.2.4.2 yamt #define R0_IPE (1U<<28) /* IP checksum error */
126 1.2.4.2 yamt #define R0_TCPE (1U<<27) /* TCP checksum error */
127 1.2.4.2 yamt #define R0_UDPE (1U<<26) /* UDP checksum error */
128 1.2.4.2 yamt #define R0_ES (1U<<25) /* error summary */
129 1.2.4.2 yamt #define R0_MF (1U<<24) /* multicast frame */
130 1.2.4.2 yamt #define R0_RE (1U<<19) /* framing error */
131 1.2.4.2 yamt #define R0_TL (1U<<18) /* too long frame */
132 1.2.4.2 yamt #define R0_RF (1U<<17) /* damaged runt frame */
133 1.2.4.2 yamt #define R0_CE (1U<<16) /* CRC error */
134 1.2.4.2 yamt #define R0_FT (1U<<15) /* frame type */
135 1.2.4.2 yamt #define R0_FL_MASK 0x7ff /* frame length 10:0 */
136 1.2.4.2 yamt
137 1.2.4.2 yamt #define T1_IC (1U<<31) /* post interrupt on complete */
138 1.2.4.2 yamt #define T1_FS (1U<<30) /* first segment of frame */
139 1.2.4.2 yamt #define T1_LS (1U<<29) /* last segment of frame */
140 1.2.4.2 yamt #define T1_IPCKG (1U<<28) /* generate IP checksum */
141 1.2.4.2 yamt #define T1_TCPCKG (1U<<27) /* generate TCP checksum */
142 1.2.4.2 yamt #define T1_UDPCKG (1U<<26) /* generate UDP checksum */
143 1.2.4.2 yamt #define T1_TER (1U<<25) /* end of ring */
144 1.2.4.2 yamt #define T1_TBS_MASK 0x7ff /* segment size 10:0 */
145 1.2.4.2 yamt
146 1.2.4.2 yamt #define R1_RER (1U<<25) /* end of ring */
147 1.2.4.2 yamt #define R1_RBS_MASK 0x7ff /* segment size 10:0 */
148 1.2.4.2 yamt
149 1.2.4.2 yamt #define KSE_NTXSEGS 16
150 1.2.4.2 yamt #define KSE_TXQUEUELEN 64
151 1.2.4.2 yamt #define KSE_TXQUEUELEN_MASK (KSE_TXQUEUELEN - 1)
152 1.2.4.2 yamt #define KSE_TXQUEUE_GC (KSE_TXQUEUELEN / 4)
153 1.2.4.2 yamt #define KSE_NTXDESC 256
154 1.2.4.2 yamt #define KSE_NTXDESC_MASK (KSE_NTXDESC - 1)
155 1.2.4.2 yamt #define KSE_NEXTTX(x) (((x) + 1) & KSE_NTXDESC_MASK)
156 1.2.4.2 yamt #define KSE_NEXTTXS(x) (((x) + 1) & KSE_TXQUEUELEN_MASK)
157 1.2.4.2 yamt
158 1.2.4.2 yamt #define KSE_NRXDESC 64
159 1.2.4.2 yamt #define KSE_NRXDESC_MASK (KSE_NRXDESC - 1)
160 1.2.4.2 yamt #define KSE_NEXTRX(x) (((x) + 1) & KSE_NRXDESC_MASK)
161 1.2.4.2 yamt
162 1.2.4.2 yamt struct tdes {
163 1.2.4.2 yamt uint32_t t0, t1, t2, t3;
164 1.2.4.2 yamt };
165 1.2.4.2 yamt
166 1.2.4.2 yamt struct rdes {
167 1.2.4.2 yamt uint32_t r0, r1, r2, r3;
168 1.2.4.2 yamt };
169 1.2.4.2 yamt
170 1.2.4.2 yamt struct kse_control_data {
171 1.2.4.2 yamt struct tdes kcd_txdescs[KSE_NTXDESC];
172 1.2.4.2 yamt struct rdes kcd_rxdescs[KSE_NRXDESC];
173 1.2.4.2 yamt };
174 1.2.4.2 yamt #define KSE_CDOFF(x) offsetof(struct kse_control_data, x)
175 1.2.4.2 yamt #define KSE_CDTXOFF(x) KSE_CDOFF(kcd_txdescs[(x)])
176 1.2.4.2 yamt #define KSE_CDRXOFF(x) KSE_CDOFF(kcd_rxdescs[(x)])
177 1.2.4.2 yamt
178 1.2.4.2 yamt struct kse_txsoft {
179 1.2.4.2 yamt struct mbuf *txs_mbuf; /* head of our mbuf chain */
180 1.2.4.2 yamt bus_dmamap_t txs_dmamap; /* our DMA map */
181 1.2.4.2 yamt int txs_firstdesc; /* first descriptor in packet */
182 1.2.4.2 yamt int txs_lastdesc; /* last descriptor in packet */
183 1.2.4.2 yamt int txs_ndesc; /* # of descriptors used */
184 1.2.4.2 yamt };
185 1.2.4.2 yamt
186 1.2.4.2 yamt struct kse_rxsoft {
187 1.2.4.2 yamt struct mbuf *rxs_mbuf; /* head of our mbuf chain */
188 1.2.4.2 yamt bus_dmamap_t rxs_dmamap; /* our DMA map */
189 1.2.4.2 yamt };
190 1.2.4.2 yamt
191 1.2.4.2 yamt struct kse_softc {
192 1.2.4.2 yamt struct device sc_dev; /* generic device information */
193 1.2.4.2 yamt bus_space_tag_t sc_st; /* bus space tag */
194 1.2.4.2 yamt bus_space_handle_t sc_sh; /* bus space handle */
195 1.2.4.2 yamt bus_dma_tag_t sc_dmat; /* bus DMA tag */
196 1.2.4.2 yamt struct ethercom sc_ethercom; /* Ethernet common data */
197 1.2.4.2 yamt void *sc_ih; /* interrupt cookie */
198 1.2.4.2 yamt
199 1.2.4.2 yamt struct ifmedia sc_media; /* ifmedia information */
200 1.2.4.2 yamt int sc_media_status; /* PHY */
201 1.2.4.2 yamt int sc_media_active; /* PHY */
202 1.2.4.2 yamt struct callout sc_callout; /* tick callout */
203 1.2.4.2 yamt
204 1.2.4.2 yamt bus_dmamap_t sc_cddmamap; /* control data DMA map */
205 1.2.4.2 yamt #define sc_cddma sc_cddmamap->dm_segs[0].ds_addr
206 1.2.4.2 yamt
207 1.2.4.2 yamt struct kse_control_data *sc_control_data;
208 1.2.4.2 yamt #define sc_txdescs sc_control_data->kcd_txdescs
209 1.2.4.2 yamt #define sc_rxdescs sc_control_data->kcd_rxdescs
210 1.2.4.2 yamt
211 1.2.4.2 yamt struct kse_txsoft sc_txsoft[KSE_TXQUEUELEN];
212 1.2.4.2 yamt struct kse_rxsoft sc_rxsoft[KSE_NRXDESC];
213 1.2.4.2 yamt int sc_txfree; /* number of free Tx descriptors */
214 1.2.4.2 yamt int sc_txnext; /* next ready Tx descriptor */
215 1.2.4.2 yamt int sc_txsfree; /* number of free Tx jobs */
216 1.2.4.2 yamt int sc_txsnext; /* next ready Tx job */
217 1.2.4.2 yamt int sc_txsdirty; /* dirty Tx jobs */
218 1.2.4.2 yamt int sc_rxptr; /* next ready Rx descriptor/descsoft */
219 1.2.4.2 yamt
220 1.2.4.2 yamt uint32_t sc_txc, sc_rxc;
221 1.2.4.2 yamt uint32_t sc_t1csum;
222 1.2.4.2 yamt int sc_mcsum;
223 1.2.4.2 yamt uint32_t sc_chip;
224 1.2.4.2 yamt };
225 1.2.4.2 yamt
226 1.2.4.2 yamt #define KSE_CDTXADDR(sc, x) ((sc)->sc_cddma + KSE_CDTXOFF((x)))
227 1.2.4.2 yamt #define KSE_CDRXADDR(sc, x) ((sc)->sc_cddma + KSE_CDRXOFF((x)))
228 1.2.4.2 yamt
229 1.2.4.2 yamt #define KSE_CDTXSYNC(sc, x, n, ops) \
230 1.2.4.2 yamt do { \
231 1.2.4.2 yamt int __x, __n; \
232 1.2.4.2 yamt \
233 1.2.4.2 yamt __x = (x); \
234 1.2.4.2 yamt __n = (n); \
235 1.2.4.2 yamt \
236 1.2.4.2 yamt /* If it will wrap around, sync to the end of the ring. */ \
237 1.2.4.2 yamt if ((__x + __n) > KSE_NTXDESC) { \
238 1.2.4.2 yamt bus_dmamap_sync((sc)->sc_dmat, (sc)->sc_cddmamap, \
239 1.2.4.2 yamt KSE_CDTXOFF(__x), sizeof(struct tdes) * \
240 1.2.4.2 yamt (KSE_NTXDESC - __x), (ops)); \
241 1.2.4.2 yamt __n -= (KSE_NTXDESC - __x); \
242 1.2.4.2 yamt __x = 0; \
243 1.2.4.2 yamt } \
244 1.2.4.2 yamt \
245 1.2.4.2 yamt /* Now sync whatever is left. */ \
246 1.2.4.2 yamt bus_dmamap_sync((sc)->sc_dmat, (sc)->sc_cddmamap, \
247 1.2.4.2 yamt KSE_CDTXOFF(__x), sizeof(struct tdes) * __n, (ops)); \
248 1.2.4.2 yamt } while (/*CONSTCOND*/0)
249 1.2.4.2 yamt
250 1.2.4.2 yamt #define KSE_CDRXSYNC(sc, x, ops) \
251 1.2.4.2 yamt do { \
252 1.2.4.2 yamt bus_dmamap_sync((sc)->sc_dmat, (sc)->sc_cddmamap, \
253 1.2.4.2 yamt KSE_CDRXOFF((x)), sizeof(struct rdes), (ops)); \
254 1.2.4.2 yamt } while (/*CONSTCOND*/0)
255 1.2.4.2 yamt
256 1.2.4.2 yamt #define KSE_INIT_RXDESC(sc, x) \
257 1.2.4.2 yamt do { \
258 1.2.4.2 yamt struct kse_rxsoft *__rxs = &(sc)->sc_rxsoft[(x)]; \
259 1.2.4.2 yamt struct rdes *__rxd = &(sc)->sc_rxdescs[(x)]; \
260 1.2.4.2 yamt struct mbuf *__m = __rxs->rxs_mbuf; \
261 1.2.4.2 yamt \
262 1.2.4.2 yamt /* \
263 1.2.4.2 yamt * Note: may be able to scoot the packet forward 2 bytes for \
264 1.2.4.2 yamt * the alignment. Unclear KS8842 Rx DMA really mandates to have \
265 1.2.4.2 yamt * 32-bit buffer boundary. Tx DMA has no alignment limitation. \
266 1.2.4.2 yamt */ \
267 1.2.4.2 yamt __m->m_data = __m->m_ext.ext_buf; \
268 1.2.4.2 yamt __rxd->r2 = __rxs->rxs_dmamap->dm_segs[0].ds_addr; \
269 1.2.4.2 yamt __rxd->r1 = R1_RBS_MASK /* __m->m_ext.ext_size */; \
270 1.2.4.2 yamt __rxd->r0 = R0_OWN; \
271 1.2.4.2 yamt KSE_CDRXSYNC((sc), (x), BUS_DMASYNC_PREREAD|BUS_DMASYNC_PREWRITE); \
272 1.2.4.2 yamt } while (/*CONSTCOND*/0)
273 1.2.4.2 yamt
274 1.2.4.2 yamt u_int kse_burstsize = 16; /* DMA burst length tuning knob */
275 1.2.4.2 yamt
276 1.2.4.2 yamt #ifdef KSEDIAGNOSTIC
277 1.2.4.2 yamt u_int kse_monitor_rxintr; /* fragmented UDP csum HW bug hook */
278 1.2.4.2 yamt #endif
279 1.2.4.2 yamt
280 1.2.4.2 yamt static int kse_match(struct device *, struct cfdata *, void *);
281 1.2.4.2 yamt static void kse_attach(struct device *, struct device *, void *);
282 1.2.4.2 yamt
283 1.2.4.2 yamt CFATTACH_DECL(kse, sizeof(struct kse_softc),
284 1.2.4.2 yamt kse_match, kse_attach, NULL, NULL);
285 1.2.4.2 yamt
286 1.2.4.2 yamt static int kse_ioctl(struct ifnet *, u_long, caddr_t);
287 1.2.4.2 yamt static void kse_start(struct ifnet *);
288 1.2.4.2 yamt static void kse_watchdog(struct ifnet *);
289 1.2.4.2 yamt static int kse_init(struct ifnet *);
290 1.2.4.2 yamt static void kse_stop(struct ifnet *, int);
291 1.2.4.2 yamt static void kse_reset(struct kse_softc *);
292 1.2.4.2 yamt static void kse_set_filter(struct kse_softc *);
293 1.2.4.2 yamt static int add_rxbuf(struct kse_softc *, int);
294 1.2.4.2 yamt static void rxdrain(struct kse_softc *);
295 1.2.4.2 yamt static int kse_intr(void *);
296 1.2.4.2 yamt static void rxintr(struct kse_softc *);
297 1.2.4.2 yamt static void txreap(struct kse_softc *);
298 1.2.4.2 yamt static void lnkchg(struct kse_softc *);
299 1.2.4.2 yamt static int ifmedia_upd(struct ifnet *);
300 1.2.4.2 yamt static void ifmedia_sts(struct ifnet *, struct ifmediareq *);
301 1.2.4.2 yamt static void phy_tick(void *);
302 1.2.4.2 yamt
303 1.2.4.2 yamt static int
304 1.2.4.2 yamt kse_match(struct device *parent, struct cfdata *match, void *aux)
305 1.2.4.2 yamt {
306 1.2.4.2 yamt struct pci_attach_args *pa = (struct pci_attach_args *)aux;
307 1.2.4.2 yamt
308 1.2.4.2 yamt if (PCI_VENDOR(pa->pa_id) == PCI_VENDOR_MICREL &&
309 1.2.4.2 yamt (PCI_PRODUCT(pa->pa_id) == PCI_PRODUCT_MICREL_KSZ8842 ||
310 1.2.4.2 yamt PCI_PRODUCT(pa->pa_id) == PCI_PRODUCT_MICREL_KSZ8841) &&
311 1.2.4.2 yamt PCI_CLASS(pa->pa_class) == PCI_CLASS_NETWORK)
312 1.2.4.2 yamt return 1;
313 1.2.4.2 yamt
314 1.2.4.2 yamt return 0;
315 1.2.4.2 yamt }
316 1.2.4.2 yamt
317 1.2.4.2 yamt static void
318 1.2.4.2 yamt kse_attach(struct device *parent, struct device *self, void *aux)
319 1.2.4.2 yamt {
320 1.2.4.2 yamt struct kse_softc *sc = (struct kse_softc *)self;
321 1.2.4.2 yamt struct pci_attach_args *pa = aux;
322 1.2.4.2 yamt pci_chipset_tag_t pc = pa->pa_pc;
323 1.2.4.2 yamt pci_intr_handle_t ih;
324 1.2.4.2 yamt const char *intrstr;
325 1.2.4.2 yamt struct ifnet *ifp;
326 1.2.4.2 yamt uint8_t enaddr[ETHER_ADDR_LEN];
327 1.2.4.2 yamt bus_dma_segment_t seg;
328 1.2.4.2 yamt int error, i, nseg;
329 1.2.4.2 yamt pcireg_t pmode;
330 1.2.4.2 yamt int pmreg;
331 1.2.4.2 yamt
332 1.2.4.2 yamt if (pci_mapreg_map(pa, 0x10,
333 1.2.4.2 yamt PCI_MAPREG_TYPE_MEM | PCI_MAPREG_MEM_TYPE_32BIT,
334 1.2.4.2 yamt 0, &sc->sc_st, &sc->sc_sh, NULL, NULL) != 0) {
335 1.2.4.2 yamt printf(": unable to map device registers\n");
336 1.2.4.2 yamt return;
337 1.2.4.2 yamt }
338 1.2.4.2 yamt
339 1.2.4.2 yamt sc->sc_dmat = pa->pa_dmat;
340 1.2.4.2 yamt
341 1.2.4.2 yamt /* Make sure bus mastering is enabled. */
342 1.2.4.2 yamt pci_conf_write(pc, pa->pa_tag, PCI_COMMAND_STATUS_REG,
343 1.2.4.2 yamt pci_conf_read(pc, pa->pa_tag, PCI_COMMAND_STATUS_REG) |
344 1.2.4.2 yamt PCI_COMMAND_MASTER_ENABLE);
345 1.2.4.2 yamt
346 1.2.4.2 yamt /* Get it out of power save mode, if needed. */
347 1.2.4.2 yamt if (pci_get_capability(pc, pa->pa_tag, PCI_CAP_PWRMGMT, &pmreg, 0)) {
348 1.2.4.2 yamt pmode = pci_conf_read(pc, pa->pa_tag, pmreg + PCI_PMCSR) &
349 1.2.4.2 yamt PCI_PMCSR_STATE_MASK;
350 1.2.4.2 yamt if (pmode == PCI_PMCSR_STATE_D3) {
351 1.2.4.2 yamt /*
352 1.2.4.2 yamt * The card has lost all configuration data in
353 1.2.4.2 yamt * this state, so punt.
354 1.2.4.2 yamt */
355 1.2.4.2 yamt printf("%s: unable to wake from power state D3\n",
356 1.2.4.2 yamt sc->sc_dev.dv_xname);
357 1.2.4.2 yamt return;
358 1.2.4.2 yamt }
359 1.2.4.2 yamt if (pmode != PCI_PMCSR_STATE_D0) {
360 1.2.4.2 yamt printf("%s: waking up from power date D%d\n",
361 1.2.4.2 yamt sc->sc_dev.dv_xname, pmode);
362 1.2.4.2 yamt pci_conf_write(pc, pa->pa_tag, pmreg + PCI_PMCSR,
363 1.2.4.2 yamt PCI_PMCSR_STATE_D0);
364 1.2.4.2 yamt }
365 1.2.4.2 yamt }
366 1.2.4.2 yamt
367 1.2.4.2 yamt sc->sc_chip = PCI_PRODUCT(pa->pa_id);
368 1.2.4.2 yamt printf(": Micrel KSZ%04x Ethernet (rev. 0x%02x)\n",
369 1.2.4.2 yamt sc->sc_chip, PCI_REVISION(pa->pa_class));
370 1.2.4.2 yamt
371 1.2.4.2 yamt /*
372 1.2.4.2 yamt * Read the Ethernet address from the EEPROM.
373 1.2.4.2 yamt */
374 1.2.4.2 yamt i = CSR_READ_2(sc, MARL);
375 1.2.4.2 yamt enaddr[5] = i; enaddr[4] = i >> 8;
376 1.2.4.2 yamt i = CSR_READ_2(sc, MARM);
377 1.2.4.2 yamt enaddr[3] = i; enaddr[2] = i >> 8;
378 1.2.4.2 yamt i = CSR_READ_2(sc, MARH);
379 1.2.4.2 yamt enaddr[1] = i; enaddr[0] = i >> 8;
380 1.2.4.2 yamt printf("%s: Ethernet address: %s\n",
381 1.2.4.2 yamt sc->sc_dev.dv_xname, ether_sprintf(enaddr));
382 1.2.4.2 yamt
383 1.2.4.2 yamt /*
384 1.2.4.2 yamt * Enable chip function.
385 1.2.4.2 yamt */
386 1.2.4.2 yamt CSR_WRITE_2(sc, CIDR, 1);
387 1.2.4.2 yamt
388 1.2.4.2 yamt /*
389 1.2.4.2 yamt * Map and establish our interrupt.
390 1.2.4.2 yamt */
391 1.2.4.2 yamt if (pci_intr_map(pa, &ih)) {
392 1.2.4.2 yamt printf("%s: unable to map interrupt\n", sc->sc_dev.dv_xname);
393 1.2.4.2 yamt return;
394 1.2.4.2 yamt }
395 1.2.4.2 yamt intrstr = pci_intr_string(pc, ih);
396 1.2.4.2 yamt sc->sc_ih = pci_intr_establish(pc, ih, IPL_NET, kse_intr, sc);
397 1.2.4.2 yamt if (sc->sc_ih == NULL) {
398 1.2.4.2 yamt printf("%s: unable to establish interrupt",
399 1.2.4.2 yamt sc->sc_dev.dv_xname);
400 1.2.4.2 yamt if (intrstr != NULL)
401 1.2.4.2 yamt printf(" at %s", intrstr);
402 1.2.4.2 yamt printf("\n");
403 1.2.4.2 yamt return;
404 1.2.4.2 yamt }
405 1.2.4.2 yamt printf("%s: interrupting at %s\n", sc->sc_dev.dv_xname, intrstr);
406 1.2.4.2 yamt
407 1.2.4.2 yamt /*
408 1.2.4.2 yamt * Allocate the control data structures, and create and load the
409 1.2.4.2 yamt * DMA map for it.
410 1.2.4.2 yamt */
411 1.2.4.2 yamt error = bus_dmamem_alloc(sc->sc_dmat,
412 1.2.4.2 yamt sizeof(struct kse_control_data), PAGE_SIZE, 0, &seg, 1, &nseg, 0);
413 1.2.4.2 yamt if (error != 0) {
414 1.2.4.2 yamt printf("%s: unable to allocate control data, error = %d\n",
415 1.2.4.2 yamt sc->sc_dev.dv_xname, error);
416 1.2.4.2 yamt goto fail_0;
417 1.2.4.2 yamt }
418 1.2.4.2 yamt error = bus_dmamem_map(sc->sc_dmat, &seg, nseg,
419 1.2.4.2 yamt sizeof(struct kse_control_data), (caddr_t *)&sc->sc_control_data,
420 1.2.4.2 yamt BUS_DMA_COHERENT);
421 1.2.4.2 yamt if (error != 0) {
422 1.2.4.2 yamt printf("%s: unable to map control data, error = %d\n",
423 1.2.4.2 yamt sc->sc_dev.dv_xname, error);
424 1.2.4.2 yamt goto fail_1;
425 1.2.4.2 yamt }
426 1.2.4.2 yamt error = bus_dmamap_create(sc->sc_dmat,
427 1.2.4.2 yamt sizeof(struct kse_control_data), 1,
428 1.2.4.2 yamt sizeof(struct kse_control_data), 0, 0, &sc->sc_cddmamap);
429 1.2.4.2 yamt if (error != 0) {
430 1.2.4.2 yamt printf("%s: unable to create control data DMA map, "
431 1.2.4.2 yamt "error = %d\n", sc->sc_dev.dv_xname, error);
432 1.2.4.2 yamt goto fail_2;
433 1.2.4.2 yamt }
434 1.2.4.2 yamt error = bus_dmamap_load(sc->sc_dmat, sc->sc_cddmamap,
435 1.2.4.2 yamt sc->sc_control_data, sizeof(struct kse_control_data), NULL, 0);
436 1.2.4.2 yamt if (error != 0) {
437 1.2.4.2 yamt printf("%s: unable to load control data DMA map, error = %d\n",
438 1.2.4.2 yamt sc->sc_dev.dv_xname, error);
439 1.2.4.2 yamt goto fail_3;
440 1.2.4.2 yamt }
441 1.2.4.2 yamt for (i = 0; i < KSE_TXQUEUELEN; i++) {
442 1.2.4.2 yamt if ((error = bus_dmamap_create(sc->sc_dmat, MCLBYTES,
443 1.2.4.2 yamt KSE_NTXSEGS, MCLBYTES, 0, 0,
444 1.2.4.2 yamt &sc->sc_txsoft[i].txs_dmamap)) != 0) {
445 1.2.4.2 yamt printf("%s: unable to create tx DMA map %d, "
446 1.2.4.2 yamt "error = %d\n", sc->sc_dev.dv_xname, i, error);
447 1.2.4.2 yamt goto fail_4;
448 1.2.4.2 yamt }
449 1.2.4.2 yamt }
450 1.2.4.2 yamt for (i = 0; i < KSE_NRXDESC; i++) {
451 1.2.4.2 yamt if ((error = bus_dmamap_create(sc->sc_dmat, MCLBYTES,
452 1.2.4.2 yamt 1, MCLBYTES, 0, 0, &sc->sc_rxsoft[i].rxs_dmamap)) != 0) {
453 1.2.4.2 yamt printf("%s: unable to create rx DMA map %d, "
454 1.2.4.2 yamt "error = %d\n", sc->sc_dev.dv_xname, i, error);
455 1.2.4.2 yamt goto fail_5;
456 1.2.4.2 yamt }
457 1.2.4.2 yamt sc->sc_rxsoft[i].rxs_mbuf = NULL;
458 1.2.4.2 yamt }
459 1.2.4.2 yamt
460 1.2.4.2 yamt callout_init(&sc->sc_callout);
461 1.2.4.2 yamt
462 1.2.4.2 yamt ifmedia_init(&sc->sc_media, 0, ifmedia_upd, ifmedia_sts);
463 1.2.4.2 yamt ifmedia_add(&sc->sc_media, IFM_ETHER|IFM_10_T, 0, NULL);
464 1.2.4.2 yamt ifmedia_add(&sc->sc_media, IFM_ETHER|IFM_10_T|IFM_FDX, 0, NULL);
465 1.2.4.2 yamt ifmedia_add(&sc->sc_media, IFM_ETHER|IFM_100_TX, 0, NULL);
466 1.2.4.2 yamt ifmedia_add(&sc->sc_media, IFM_ETHER|IFM_100_TX|IFM_FDX, 0, NULL);
467 1.2.4.2 yamt ifmedia_add(&sc->sc_media, IFM_ETHER|IFM_AUTO, 0, NULL);
468 1.2.4.2 yamt ifmedia_set(&sc->sc_media, IFM_ETHER|IFM_AUTO);
469 1.2.4.2 yamt
470 1.2.4.2 yamt printf("%s: 10baseT, 10baseT-FDX, 100baseTX, 100baseTX-FDX, auto\n",
471 1.2.4.2 yamt sc->sc_dev.dv_xname);
472 1.2.4.2 yamt
473 1.2.4.2 yamt ifp = &sc->sc_ethercom.ec_if;
474 1.2.4.2 yamt strcpy(ifp->if_xname, sc->sc_dev.dv_xname);
475 1.2.4.2 yamt ifp->if_softc = sc;
476 1.2.4.2 yamt ifp->if_flags = IFF_BROADCAST | IFF_SIMPLEX | IFF_MULTICAST;
477 1.2.4.2 yamt ifp->if_ioctl = kse_ioctl;
478 1.2.4.2 yamt ifp->if_start = kse_start;
479 1.2.4.2 yamt ifp->if_watchdog = kse_watchdog;
480 1.2.4.2 yamt ifp->if_init = kse_init;
481 1.2.4.2 yamt ifp->if_stop = kse_stop;
482 1.2.4.2 yamt IFQ_SET_READY(&ifp->if_snd);
483 1.2.4.2 yamt
484 1.2.4.2 yamt /*
485 1.2.4.2 yamt * KSZ8842 can handle 802.1Q VLAN-sized frames,
486 1.2.4.2 yamt * can do IPv4, TCPv4, and UDPv4 checksums in hardware.
487 1.2.4.2 yamt */
488 1.2.4.2 yamt sc->sc_ethercom.ec_capabilities |= ETHERCAP_VLAN_MTU;
489 1.2.4.2 yamt ifp->if_capabilities |=
490 1.2.4.2 yamt IFCAP_CSUM_IPv4_Tx | IFCAP_CSUM_IPv4_Rx |
491 1.2.4.2 yamt IFCAP_CSUM_TCPv4_Tx | IFCAP_CSUM_TCPv4_Rx |
492 1.2.4.2 yamt IFCAP_CSUM_UDPv4_Tx | IFCAP_CSUM_UDPv4_Rx;
493 1.2.4.2 yamt
494 1.2.4.2 yamt if_attach(ifp);
495 1.2.4.2 yamt ether_ifattach(ifp, enaddr);
496 1.2.4.2 yamt return;
497 1.2.4.2 yamt
498 1.2.4.2 yamt fail_5:
499 1.2.4.2 yamt for (i = 0; i < KSE_NRXDESC; i++) {
500 1.2.4.2 yamt if (sc->sc_rxsoft[i].rxs_dmamap != NULL)
501 1.2.4.2 yamt bus_dmamap_destroy(sc->sc_dmat,
502 1.2.4.2 yamt sc->sc_rxsoft[i].rxs_dmamap);
503 1.2.4.2 yamt }
504 1.2.4.2 yamt fail_4:
505 1.2.4.2 yamt for (i = 0; i < KSE_TXQUEUELEN; i++) {
506 1.2.4.2 yamt if (sc->sc_txsoft[i].txs_dmamap != NULL)
507 1.2.4.2 yamt bus_dmamap_destroy(sc->sc_dmat,
508 1.2.4.2 yamt sc->sc_txsoft[i].txs_dmamap);
509 1.2.4.2 yamt }
510 1.2.4.2 yamt bus_dmamap_unload(sc->sc_dmat, sc->sc_cddmamap);
511 1.2.4.2 yamt fail_3:
512 1.2.4.2 yamt bus_dmamap_destroy(sc->sc_dmat, sc->sc_cddmamap);
513 1.2.4.2 yamt fail_2:
514 1.2.4.2 yamt bus_dmamem_unmap(sc->sc_dmat, (caddr_t)sc->sc_control_data,
515 1.2.4.2 yamt sizeof(struct kse_control_data));
516 1.2.4.2 yamt fail_1:
517 1.2.4.2 yamt bus_dmamem_free(sc->sc_dmat, &seg, nseg);
518 1.2.4.2 yamt fail_0:
519 1.2.4.2 yamt return;
520 1.2.4.2 yamt }
521 1.2.4.2 yamt
522 1.2.4.2 yamt static int
523 1.2.4.2 yamt kse_ioctl(struct ifnet *ifp, u_long cmd, caddr_t data)
524 1.2.4.2 yamt {
525 1.2.4.2 yamt struct kse_softc *sc = ifp->if_softc;
526 1.2.4.2 yamt struct ifreq *ifr = (struct ifreq *)data;
527 1.2.4.2 yamt int s, error;
528 1.2.4.2 yamt
529 1.2.4.2 yamt s = splnet();
530 1.2.4.2 yamt
531 1.2.4.2 yamt switch (cmd) {
532 1.2.4.2 yamt case SIOCSIFMEDIA:
533 1.2.4.2 yamt case SIOCGIFMEDIA:
534 1.2.4.2 yamt error = ifmedia_ioctl(ifp, ifr, &sc->sc_media, cmd);
535 1.2.4.2 yamt break;
536 1.2.4.2 yamt
537 1.2.4.2 yamt default:
538 1.2.4.2 yamt error = ether_ioctl(ifp, cmd, data);
539 1.2.4.2 yamt if (cmd == ENETRESET) {
540 1.2.4.2 yamt /*
541 1.2.4.2 yamt * Multicast list has changed; set the hardware filter
542 1.2.4.2 yamt * accordingly.
543 1.2.4.2 yamt */
544 1.2.4.2 yamt kse_set_filter(sc);
545 1.2.4.2 yamt error = 0;
546 1.2.4.2 yamt }
547 1.2.4.2 yamt break;
548 1.2.4.2 yamt }
549 1.2.4.2 yamt
550 1.2.4.2 yamt kse_start(ifp);
551 1.2.4.2 yamt
552 1.2.4.2 yamt splx(s);
553 1.2.4.2 yamt return error;
554 1.2.4.2 yamt }
555 1.2.4.2 yamt
556 1.2.4.2 yamt #define KSE_INTRS (INT_DMLCS|INT_DMTS|INT_DMRS|INT_DMRBUS)
557 1.2.4.2 yamt
558 1.2.4.2 yamt static int
559 1.2.4.2 yamt kse_init(struct ifnet *ifp)
560 1.2.4.2 yamt {
561 1.2.4.2 yamt struct kse_softc *sc = ifp->if_softc;
562 1.2.4.2 yamt uint32_t paddr;
563 1.2.4.2 yamt int i, error = 0;
564 1.2.4.2 yamt
565 1.2.4.2 yamt /* cancel pending I/O */
566 1.2.4.2 yamt kse_stop(ifp, 0);
567 1.2.4.2 yamt
568 1.2.4.2 yamt /* reset all registers but PCI configuration */
569 1.2.4.2 yamt kse_reset(sc);
570 1.2.4.2 yamt
571 1.2.4.2 yamt /* craft Tx descriptor ring */
572 1.2.4.2 yamt memset(sc->sc_txdescs, 0, sizeof(sc->sc_txdescs));
573 1.2.4.2 yamt for (i = 0, paddr = KSE_CDTXADDR(sc, 1); i < KSE_NTXDESC - 1; i++) {
574 1.2.4.2 yamt sc->sc_txdescs[i].t3 = paddr;
575 1.2.4.2 yamt paddr += sizeof(struct tdes);
576 1.2.4.2 yamt }
577 1.2.4.2 yamt sc->sc_txdescs[KSE_NTXDESC - 1].t3 = KSE_CDTXADDR(sc, 0);
578 1.2.4.2 yamt KSE_CDTXSYNC(sc, 0, KSE_NTXDESC,
579 1.2.4.2 yamt BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
580 1.2.4.2 yamt sc->sc_txfree = KSE_NTXDESC;
581 1.2.4.2 yamt sc->sc_txnext = 0;
582 1.2.4.2 yamt
583 1.2.4.2 yamt for (i = 0; i < KSE_TXQUEUELEN; i++)
584 1.2.4.2 yamt sc->sc_txsoft[i].txs_mbuf = NULL;
585 1.2.4.2 yamt sc->sc_txsfree = KSE_TXQUEUELEN;
586 1.2.4.2 yamt sc->sc_txsnext = 0;
587 1.2.4.2 yamt sc->sc_txsdirty = 0;
588 1.2.4.2 yamt
589 1.2.4.2 yamt /* craft Rx descriptor ring */
590 1.2.4.2 yamt memset(sc->sc_rxdescs, 0, sizeof(sc->sc_rxdescs));
591 1.2.4.2 yamt for (i = 0, paddr = KSE_CDRXADDR(sc, 1); i < KSE_NRXDESC - 1; i++) {
592 1.2.4.2 yamt sc->sc_rxdescs[i].r3 = paddr;
593 1.2.4.2 yamt paddr += sizeof(struct rdes);
594 1.2.4.2 yamt }
595 1.2.4.2 yamt sc->sc_rxdescs[KSE_NRXDESC - 1].r3 = KSE_CDRXADDR(sc, 0);
596 1.2.4.2 yamt for (i = 0; i < KSE_NRXDESC; i++) {
597 1.2.4.2 yamt if (sc->sc_rxsoft[i].rxs_mbuf == NULL) {
598 1.2.4.2 yamt if ((error = add_rxbuf(sc, i)) != 0) {
599 1.2.4.2 yamt printf("%s: unable to allocate or map rx "
600 1.2.4.2 yamt "buffer %d, error = %d\n",
601 1.2.4.2 yamt sc->sc_dev.dv_xname, i, error);
602 1.2.4.2 yamt rxdrain(sc);
603 1.2.4.2 yamt goto out;
604 1.2.4.2 yamt }
605 1.2.4.2 yamt }
606 1.2.4.2 yamt else
607 1.2.4.2 yamt KSE_INIT_RXDESC(sc, i);
608 1.2.4.2 yamt }
609 1.2.4.2 yamt sc->sc_rxptr = 0;
610 1.2.4.2 yamt
611 1.2.4.2 yamt /* hand Tx/Rx rings to HW */
612 1.2.4.2 yamt CSR_WRITE_4(sc, TDLB, KSE_CDTXADDR(sc, 0));
613 1.2.4.2 yamt CSR_WRITE_4(sc, RDLB, KSE_CDRXADDR(sc, 0));
614 1.2.4.2 yamt
615 1.2.4.2 yamt sc->sc_txc = TXC_TEN | TXC_EP | TXC_AC | TXC_FCE;
616 1.2.4.2 yamt sc->sc_rxc = RXC_REN | RXC_RU | RXC_FCE;
617 1.2.4.2 yamt if (ifp->if_flags & IFF_PROMISC)
618 1.2.4.2 yamt sc->sc_rxc |= RXC_RA;
619 1.2.4.2 yamt if (ifp->if_flags & IFF_BROADCAST)
620 1.2.4.2 yamt sc->sc_rxc |= RXC_RB;
621 1.2.4.2 yamt
622 1.2.4.2 yamt sc->sc_t1csum = sc->sc_mcsum = 0;
623 1.2.4.2 yamt if (ifp->if_capenable & IFCAP_CSUM_IPv4_Rx) {
624 1.2.4.2 yamt sc->sc_rxc |= RXC_ICG;
625 1.2.4.2 yamt sc->sc_mcsum |= M_CSUM_IPv4;
626 1.2.4.2 yamt }
627 1.2.4.2 yamt if (ifp->if_capenable & IFCAP_CSUM_IPv4_Tx) {
628 1.2.4.2 yamt sc->sc_txc |= TXC_ICG;
629 1.2.4.2 yamt sc->sc_t1csum |= T1_IPCKG;
630 1.2.4.2 yamt }
631 1.2.4.2 yamt if (ifp->if_capenable & IFCAP_CSUM_TCPv4_Rx) {
632 1.2.4.2 yamt sc->sc_rxc |= RXC_TCG;
633 1.2.4.2 yamt sc->sc_mcsum |= M_CSUM_TCPv4;
634 1.2.4.2 yamt }
635 1.2.4.2 yamt if (ifp->if_capenable & IFCAP_CSUM_TCPv4_Tx) {
636 1.2.4.2 yamt sc->sc_txc |= TXC_TCG;
637 1.2.4.2 yamt sc->sc_t1csum |= T1_TCPCKG;
638 1.2.4.2 yamt }
639 1.2.4.2 yamt if (ifp->if_capenable & IFCAP_CSUM_UDPv4_Rx) {
640 1.2.4.2 yamt sc->sc_rxc |= RXC_UCG;
641 1.2.4.2 yamt sc->sc_mcsum |= M_CSUM_UDPv4;
642 1.2.4.2 yamt }
643 1.2.4.2 yamt if (ifp->if_capenable & IFCAP_CSUM_UDPv4_Tx) {
644 1.2.4.2 yamt sc->sc_txc |= TXC_UCG;
645 1.2.4.2 yamt sc->sc_t1csum |= T1_UDPCKG;
646 1.2.4.2 yamt }
647 1.2.4.2 yamt sc->sc_txc |= (kse_burstsize << TXC_BS_SFT);
648 1.2.4.2 yamt sc->sc_rxc |= (kse_burstsize << RXC_BS_SFT);
649 1.2.4.2 yamt
650 1.2.4.2 yamt /* set current media */
651 1.2.4.2 yamt (void)ifmedia_upd(ifp);
652 1.2.4.2 yamt
653 1.2.4.2 yamt /* enable transmitter and receiver */
654 1.2.4.2 yamt CSR_WRITE_4(sc, MDTXC, sc->sc_txc);
655 1.2.4.2 yamt CSR_WRITE_4(sc, MDRXC, sc->sc_rxc);
656 1.2.4.2 yamt CSR_WRITE_4(sc, MDRSC, 1);
657 1.2.4.2 yamt
658 1.2.4.2 yamt /* enable interrupts */
659 1.2.4.2 yamt CSR_WRITE_4(sc, INTST, ~0);
660 1.2.4.2 yamt CSR_WRITE_4(sc, INTEN, KSE_INTRS);
661 1.2.4.2 yamt
662 1.2.4.2 yamt ifp->if_flags |= IFF_RUNNING;
663 1.2.4.2 yamt ifp->if_flags &= ~IFF_OACTIVE;
664 1.2.4.2 yamt
665 1.2.4.2 yamt /* start one second timer */
666 1.2.4.2 yamt callout_reset(&sc->sc_callout, hz, phy_tick, sc);
667 1.2.4.2 yamt
668 1.2.4.2 yamt out:
669 1.2.4.2 yamt if (error) {
670 1.2.4.2 yamt ifp->if_flags &= ~(IFF_RUNNING | IFF_OACTIVE);
671 1.2.4.2 yamt ifp->if_timer = 0;
672 1.2.4.2 yamt printf("%s: interface not running\n", sc->sc_dev.dv_xname);
673 1.2.4.2 yamt }
674 1.2.4.2 yamt return error;
675 1.2.4.2 yamt }
676 1.2.4.2 yamt
677 1.2.4.2 yamt static void
678 1.2.4.2 yamt kse_stop(struct ifnet *ifp, int disable)
679 1.2.4.2 yamt {
680 1.2.4.2 yamt struct kse_softc *sc = ifp->if_softc;
681 1.2.4.2 yamt struct kse_txsoft *txs;
682 1.2.4.2 yamt int i;
683 1.2.4.2 yamt
684 1.2.4.2 yamt callout_stop(&sc->sc_callout);
685 1.2.4.2 yamt
686 1.2.4.2 yamt sc->sc_txc &= ~TXC_TEN;
687 1.2.4.2 yamt sc->sc_rxc &= ~RXC_REN;
688 1.2.4.2 yamt CSR_WRITE_4(sc, MDTXC, sc->sc_txc);
689 1.2.4.2 yamt CSR_WRITE_4(sc, MDRXC, sc->sc_rxc);
690 1.2.4.2 yamt
691 1.2.4.2 yamt for (i = 0; i < KSE_TXQUEUELEN; i++) {
692 1.2.4.2 yamt txs = &sc->sc_txsoft[i];
693 1.2.4.2 yamt if (txs->txs_mbuf != NULL) {
694 1.2.4.2 yamt bus_dmamap_unload(sc->sc_dmat, txs->txs_dmamap);
695 1.2.4.2 yamt m_freem(txs->txs_mbuf);
696 1.2.4.2 yamt txs->txs_mbuf = NULL;
697 1.2.4.2 yamt }
698 1.2.4.2 yamt }
699 1.2.4.2 yamt
700 1.2.4.2 yamt if (disable)
701 1.2.4.2 yamt rxdrain(sc);
702 1.2.4.2 yamt
703 1.2.4.2 yamt ifp->if_flags &= ~(IFF_RUNNING | IFF_OACTIVE);
704 1.2.4.2 yamt ifp->if_timer = 0;
705 1.2.4.2 yamt }
706 1.2.4.2 yamt
707 1.2.4.2 yamt static void
708 1.2.4.2 yamt kse_reset(struct kse_softc *sc)
709 1.2.4.2 yamt {
710 1.2.4.2 yamt
711 1.2.4.2 yamt CSR_WRITE_2(sc, GRR, 1);
712 1.2.4.2 yamt delay(1000); /* PDF does not mention the delay amount */
713 1.2.4.2 yamt CSR_WRITE_2(sc, GRR, 0);
714 1.2.4.2 yamt
715 1.2.4.2 yamt CSR_WRITE_2(sc, CIDR, 1);
716 1.2.4.2 yamt }
717 1.2.4.2 yamt
718 1.2.4.2 yamt static void
719 1.2.4.2 yamt kse_watchdog(struct ifnet *ifp)
720 1.2.4.2 yamt {
721 1.2.4.2 yamt struct kse_softc *sc = ifp->if_softc;
722 1.2.4.2 yamt
723 1.2.4.2 yamt /*
724 1.2.4.2 yamt * Since we're not interrupting every packet, sweep
725 1.2.4.2 yamt * up before we report an error.
726 1.2.4.2 yamt */
727 1.2.4.2 yamt txreap(sc);
728 1.2.4.2 yamt
729 1.2.4.2 yamt if (sc->sc_txfree != KSE_NTXDESC) {
730 1.2.4.2 yamt printf("%s: device timeout (txfree %d txsfree %d txnext %d)\n",
731 1.2.4.2 yamt sc->sc_dev.dv_xname, sc->sc_txfree, sc->sc_txsfree,
732 1.2.4.2 yamt sc->sc_txnext);
733 1.2.4.2 yamt ifp->if_oerrors++;
734 1.2.4.2 yamt
735 1.2.4.2 yamt /* Reset the interface. */
736 1.2.4.2 yamt kse_init(ifp);
737 1.2.4.2 yamt }
738 1.2.4.2 yamt else if (ifp->if_flags & IFF_DEBUG)
739 1.2.4.2 yamt printf("%s: recovered from device timeout\n",
740 1.2.4.2 yamt sc->sc_dev.dv_xname);
741 1.2.4.2 yamt
742 1.2.4.2 yamt /* Try to get more packets going. */
743 1.2.4.2 yamt kse_start(ifp);
744 1.2.4.2 yamt }
745 1.2.4.2 yamt
746 1.2.4.2 yamt static void
747 1.2.4.2 yamt kse_start(struct ifnet *ifp)
748 1.2.4.2 yamt {
749 1.2.4.2 yamt struct kse_softc *sc = ifp->if_softc;
750 1.2.4.2 yamt struct mbuf *m0;
751 1.2.4.2 yamt struct kse_txsoft *txs;
752 1.2.4.2 yamt bus_dmamap_t dmamap;
753 1.2.4.2 yamt int error, nexttx, lasttx, ofree, seg;
754 1.2.4.2 yamt
755 1.2.4.2 yamt lasttx = -1;
756 1.2.4.2 yamt
757 1.2.4.2 yamt if ((ifp->if_flags & (IFF_RUNNING|IFF_OACTIVE)) != IFF_RUNNING)
758 1.2.4.2 yamt return;
759 1.2.4.2 yamt
760 1.2.4.2 yamt /*
761 1.2.4.2 yamt * Remember the previous number of free descriptors.
762 1.2.4.2 yamt */
763 1.2.4.2 yamt ofree = sc->sc_txfree;
764 1.2.4.2 yamt
765 1.2.4.2 yamt /*
766 1.2.4.2 yamt * Loop through the send queue, setting up transmit descriptors
767 1.2.4.2 yamt * until we drain the queue, or use up all available transmit
768 1.2.4.2 yamt * descriptors.
769 1.2.4.2 yamt */
770 1.2.4.2 yamt for (;;) {
771 1.2.4.2 yamt IFQ_POLL(&ifp->if_snd, m0);
772 1.2.4.2 yamt if (m0 == NULL)
773 1.2.4.2 yamt break;
774 1.2.4.2 yamt
775 1.2.4.2 yamt if (sc->sc_txsfree < KSE_TXQUEUE_GC) {
776 1.2.4.2 yamt txreap(sc);
777 1.2.4.2 yamt if (sc->sc_txsfree == 0)
778 1.2.4.2 yamt break;
779 1.2.4.2 yamt }
780 1.2.4.2 yamt txs = &sc->sc_txsoft[sc->sc_txsnext];
781 1.2.4.2 yamt dmamap = txs->txs_dmamap;
782 1.2.4.2 yamt
783 1.2.4.2 yamt error = bus_dmamap_load_mbuf(sc->sc_dmat, dmamap, m0,
784 1.2.4.2 yamt BUS_DMA_WRITE|BUS_DMA_NOWAIT);
785 1.2.4.2 yamt if (error) {
786 1.2.4.2 yamt if (error == EFBIG) {
787 1.2.4.2 yamt printf("%s: Tx packet consumes too many "
788 1.2.4.2 yamt "DMA segments, dropping...\n",
789 1.2.4.2 yamt sc->sc_dev.dv_xname);
790 1.2.4.2 yamt IFQ_DEQUEUE(&ifp->if_snd, m0);
791 1.2.4.2 yamt m_freem(m0);
792 1.2.4.2 yamt continue;
793 1.2.4.2 yamt }
794 1.2.4.2 yamt /* Short on resources, just stop for now. */
795 1.2.4.2 yamt break;
796 1.2.4.2 yamt }
797 1.2.4.2 yamt
798 1.2.4.2 yamt if (dmamap->dm_nsegs > sc->sc_txfree) {
799 1.2.4.2 yamt /*
800 1.2.4.2 yamt * Not enough free descriptors to transmit this
801 1.2.4.2 yamt * packet. We haven't committed anything yet,
802 1.2.4.2 yamt * so just unload the DMA map, put the packet
803 1.2.4.2 yamt * back on the queue, and punt. Notify the upper
804 1.2.4.2 yamt * layer that there are not more slots left.
805 1.2.4.2 yamt */
806 1.2.4.2 yamt ifp->if_flags |= IFF_OACTIVE;
807 1.2.4.2 yamt bus_dmamap_unload(sc->sc_dmat, dmamap);
808 1.2.4.2 yamt break;
809 1.2.4.2 yamt }
810 1.2.4.2 yamt
811 1.2.4.2 yamt IFQ_DEQUEUE(&ifp->if_snd, m0);
812 1.2.4.2 yamt
813 1.2.4.2 yamt /*
814 1.2.4.2 yamt * WE ARE NOW COMMITTED TO TRANSMITTING THE PACKET.
815 1.2.4.2 yamt */
816 1.2.4.2 yamt
817 1.2.4.2 yamt bus_dmamap_sync(sc->sc_dmat, dmamap, 0, dmamap->dm_mapsize,
818 1.2.4.2 yamt BUS_DMASYNC_PREWRITE);
819 1.2.4.2 yamt
820 1.2.4.2 yamt for (nexttx = sc->sc_txnext, seg = 0;
821 1.2.4.2 yamt seg < dmamap->dm_nsegs;
822 1.2.4.2 yamt seg++, nexttx = KSE_NEXTTX(nexttx)) {
823 1.2.4.2 yamt struct tdes *tdes = &sc->sc_txdescs[nexttx];
824 1.2.4.2 yamt /*
825 1.2.4.2 yamt * If this is the first descriptor we're
826 1.2.4.2 yamt * enqueueing, don't set the OWN bit just
827 1.2.4.2 yamt * yet. That could cause a race condition.
828 1.2.4.2 yamt * We'll do it below.
829 1.2.4.2 yamt */
830 1.2.4.2 yamt tdes->t2 = dmamap->dm_segs[seg].ds_addr;
831 1.2.4.2 yamt tdes->t1 = sc->sc_t1csum
832 1.2.4.2 yamt | (dmamap->dm_segs[seg].ds_len & T1_TBS_MASK);
833 1.2.4.2 yamt if (nexttx != sc->sc_txnext)
834 1.2.4.2 yamt tdes->t0 = T0_OWN;
835 1.2.4.2 yamt lasttx = nexttx;
836 1.2.4.2 yamt }
837 1.2.4.2 yamt #if 0
838 1.2.4.2 yamt /*
839 1.2.4.2 yamt * T1_IC bit could schedule Tx frame done interrupt here,
840 1.2.4.2 yamt * but this driver takes a "shoot away" Tx strategy.
841 1.2.4.2 yamt */
842 1.2.4.2 yamt #else
843 1.2.4.2 yamt {
844 1.2.4.2 yamt /*
845 1.2.4.2 yamt * Outgoing NFS mbuf must be unloaded when Tx completed.
846 1.2.4.2 yamt * Without T1_IC NFS mbuf is left unack'ed for excessive
847 1.2.4.2 yamt * time and NFS stops to proceed until kse_watchdog()
848 1.2.4.2 yamt * calls txreap() to reclaim the unack'ed mbuf.
849 1.2.4.2 yamt * It's painful to tranverse every mbuf chain to determine
850 1.2.4.2 yamt * whether someone is waiting for Tx completion.
851 1.2.4.2 yamt */
852 1.2.4.2 yamt struct mbuf *m = m0;
853 1.2.4.2 yamt do {
854 1.2.4.2 yamt if ((m->m_flags & M_EXT) && m->m_ext.ext_free) {
855 1.2.4.2 yamt sc->sc_txdescs[lasttx].t1 |= T1_IC;
856 1.2.4.2 yamt break;
857 1.2.4.2 yamt }
858 1.2.4.2 yamt } while ((m = m->m_next) != NULL);
859 1.2.4.2 yamt }
860 1.2.4.2 yamt #endif
861 1.2.4.2 yamt
862 1.2.4.2 yamt /* write last T0_OWN bit of the 1st segment */
863 1.2.4.2 yamt sc->sc_txdescs[lasttx].t1 |= T1_LS;
864 1.2.4.2 yamt sc->sc_txdescs[sc->sc_txnext].t1 |= T1_FS;
865 1.2.4.2 yamt sc->sc_txdescs[sc->sc_txnext].t0 = T0_OWN;
866 1.2.4.2 yamt KSE_CDTXSYNC(sc, sc->sc_txnext, dmamap->dm_nsegs,
867 1.2.4.2 yamt BUS_DMASYNC_PREREAD|BUS_DMASYNC_PREWRITE);
868 1.2.4.2 yamt
869 1.2.4.2 yamt /* tell DMA start transmit */
870 1.2.4.2 yamt CSR_WRITE_4(sc, MDTSC, 1);
871 1.2.4.2 yamt
872 1.2.4.2 yamt txs->txs_mbuf = m0;
873 1.2.4.2 yamt txs->txs_firstdesc = sc->sc_txnext;
874 1.2.4.2 yamt txs->txs_lastdesc = lasttx;
875 1.2.4.2 yamt txs->txs_ndesc = dmamap->dm_nsegs;
876 1.2.4.2 yamt
877 1.2.4.2 yamt sc->sc_txfree -= txs->txs_ndesc;
878 1.2.4.2 yamt sc->sc_txnext = nexttx;
879 1.2.4.2 yamt sc->sc_txsfree--;
880 1.2.4.2 yamt sc->sc_txsnext = KSE_NEXTTXS(sc->sc_txsnext);
881 1.2.4.2 yamt #if NBPFILTER > 0
882 1.2.4.2 yamt /*
883 1.2.4.2 yamt * Pass the packet to any BPF listeners.
884 1.2.4.2 yamt */
885 1.2.4.2 yamt if (ifp->if_bpf)
886 1.2.4.2 yamt bpf_mtap(ifp->if_bpf, m0);
887 1.2.4.2 yamt #endif /* NBPFILTER > 0 */
888 1.2.4.2 yamt }
889 1.2.4.2 yamt
890 1.2.4.2 yamt if (sc->sc_txsfree == 0 || sc->sc_txfree == 0) {
891 1.2.4.2 yamt /* No more slots left; notify upper layer. */
892 1.2.4.2 yamt ifp->if_flags |= IFF_OACTIVE;
893 1.2.4.2 yamt }
894 1.2.4.2 yamt if (sc->sc_txfree != ofree) {
895 1.2.4.2 yamt /* Set a watchdog timer in case the chip flakes out. */
896 1.2.4.2 yamt ifp->if_timer = 5;
897 1.2.4.2 yamt }
898 1.2.4.2 yamt }
899 1.2.4.2 yamt
900 1.2.4.2 yamt static void
901 1.2.4.2 yamt kse_set_filter(struct kse_softc *sc)
902 1.2.4.2 yamt {
903 1.2.4.2 yamt #if 0 /* later */
904 1.2.4.2 yamt struct ether_multistep step;
905 1.2.4.2 yamt struct ether_multi *enm;
906 1.2.4.2 yamt struct ifnet *ifp = &sc->sc_ethercom.ec_if;
907 1.2.4.2 yamt int cnt = 0;
908 1.2.4.2 yamt
909 1.2.4.2 yamt ETHER_FIRST_MULTI(step, &sc->sc_ethercom, enm);
910 1.2.4.2 yamt while (enm != NULL) {
911 1.2.4.2 yamt if (memcmp(enm->enm_addrlo,
912 1.2.4.2 yamt enm->enm_addrhi, ETHER_ADDR_LEN) != 0) {
913 1.2.4.2 yamt ;
914 1.2.4.2 yamt }
915 1.2.4.2 yamt ETHER_NEXT_MULTI(step, enm);
916 1.2.4.2 yamt cnt++;
917 1.2.4.2 yamt }
918 1.2.4.2 yamt return;
919 1.2.4.2 yamt #endif
920 1.2.4.2 yamt }
921 1.2.4.2 yamt
922 1.2.4.2 yamt static int
923 1.2.4.2 yamt add_rxbuf(struct kse_softc *sc, int idx)
924 1.2.4.2 yamt {
925 1.2.4.2 yamt struct kse_rxsoft *rxs = &sc->sc_rxsoft[idx];
926 1.2.4.2 yamt struct mbuf *m;
927 1.2.4.2 yamt int error;
928 1.2.4.2 yamt
929 1.2.4.2 yamt MGETHDR(m, M_DONTWAIT, MT_DATA);
930 1.2.4.2 yamt if (m == NULL)
931 1.2.4.2 yamt return ENOBUFS;
932 1.2.4.2 yamt
933 1.2.4.2 yamt MCLGET(m, M_DONTWAIT);
934 1.2.4.2 yamt if ((m->m_flags & M_EXT) == 0) {
935 1.2.4.2 yamt m_freem(m);
936 1.2.4.2 yamt return ENOBUFS;
937 1.2.4.2 yamt }
938 1.2.4.2 yamt
939 1.2.4.2 yamt if (rxs->rxs_mbuf != NULL)
940 1.2.4.2 yamt bus_dmamap_unload(sc->sc_dmat, rxs->rxs_dmamap);
941 1.2.4.2 yamt
942 1.2.4.2 yamt rxs->rxs_mbuf = m;
943 1.2.4.2 yamt
944 1.2.4.2 yamt error = bus_dmamap_load(sc->sc_dmat, rxs->rxs_dmamap,
945 1.2.4.2 yamt m->m_ext.ext_buf, m->m_ext.ext_size, NULL, BUS_DMA_NOWAIT);
946 1.2.4.2 yamt if (error) {
947 1.2.4.2 yamt printf("%s: can't load rx DMA map %d, error = %d\n",
948 1.2.4.2 yamt sc->sc_dev.dv_xname, idx, error);
949 1.2.4.2 yamt panic("kse_add_rxbuf");
950 1.2.4.2 yamt }
951 1.2.4.2 yamt
952 1.2.4.2 yamt bus_dmamap_sync(sc->sc_dmat, rxs->rxs_dmamap, 0,
953 1.2.4.2 yamt rxs->rxs_dmamap->dm_mapsize, BUS_DMASYNC_PREREAD);
954 1.2.4.2 yamt
955 1.2.4.2 yamt KSE_INIT_RXDESC(sc, idx);
956 1.2.4.2 yamt
957 1.2.4.2 yamt return 0;
958 1.2.4.2 yamt }
959 1.2.4.2 yamt
960 1.2.4.2 yamt static void
961 1.2.4.2 yamt rxdrain(struct kse_softc *sc)
962 1.2.4.2 yamt {
963 1.2.4.2 yamt struct kse_rxsoft *rxs;
964 1.2.4.2 yamt int i;
965 1.2.4.2 yamt
966 1.2.4.2 yamt for (i = 0; i < KSE_NRXDESC; i++) {
967 1.2.4.2 yamt rxs = &sc->sc_rxsoft[i];
968 1.2.4.2 yamt if (rxs->rxs_mbuf != NULL) {
969 1.2.4.2 yamt bus_dmamap_unload(sc->sc_dmat, rxs->rxs_dmamap);
970 1.2.4.2 yamt m_freem(rxs->rxs_mbuf);
971 1.2.4.2 yamt rxs->rxs_mbuf = NULL;
972 1.2.4.2 yamt }
973 1.2.4.2 yamt }
974 1.2.4.2 yamt }
975 1.2.4.2 yamt
976 1.2.4.2 yamt static int
977 1.2.4.2 yamt kse_intr(void *arg)
978 1.2.4.2 yamt {
979 1.2.4.2 yamt struct kse_softc *sc = arg;
980 1.2.4.2 yamt uint32_t isr;
981 1.2.4.2 yamt
982 1.2.4.2 yamt if ((isr = CSR_READ_4(sc, INTST)) == 0)
983 1.2.4.2 yamt return 0;
984 1.2.4.2 yamt
985 1.2.4.2 yamt if (isr & INT_DMRS)
986 1.2.4.2 yamt rxintr(sc);
987 1.2.4.2 yamt if (isr & INT_DMTS)
988 1.2.4.2 yamt txreap(sc);
989 1.2.4.2 yamt if (isr & INT_DMLCS)
990 1.2.4.2 yamt lnkchg(sc);
991 1.2.4.2 yamt if (isr & INT_DMRBUS)
992 1.2.4.2 yamt printf("%s: Rx descriptor full\n", sc->sc_dev.dv_xname);
993 1.2.4.2 yamt
994 1.2.4.2 yamt CSR_WRITE_4(sc, INTST, isr);
995 1.2.4.2 yamt return 1;
996 1.2.4.2 yamt }
997 1.2.4.2 yamt
998 1.2.4.2 yamt static void
999 1.2.4.2 yamt rxintr(struct kse_softc *sc)
1000 1.2.4.2 yamt {
1001 1.2.4.2 yamt struct ifnet *ifp = &sc->sc_ethercom.ec_if;
1002 1.2.4.2 yamt struct kse_rxsoft *rxs;
1003 1.2.4.2 yamt struct mbuf *m;
1004 1.2.4.2 yamt uint32_t rxstat;
1005 1.2.4.2 yamt int i, len;
1006 1.2.4.2 yamt
1007 1.2.4.2 yamt for (i = sc->sc_rxptr; /*CONSTCOND*/ 1; i = KSE_NEXTRX(i)) {
1008 1.2.4.2 yamt rxs = &sc->sc_rxsoft[i];
1009 1.2.4.2 yamt
1010 1.2.4.2 yamt KSE_CDRXSYNC(sc, i,
1011 1.2.4.2 yamt BUS_DMASYNC_POSTREAD|BUS_DMASYNC_POSTWRITE);
1012 1.2.4.2 yamt
1013 1.2.4.2 yamt rxstat = sc->sc_rxdescs[i].r0;
1014 1.2.4.2 yamt
1015 1.2.4.2 yamt if (rxstat & R0_OWN) /* desc is left empty */
1016 1.2.4.2 yamt break;
1017 1.2.4.2 yamt
1018 1.2.4.2 yamt /* R0_FS|R0_LS must have been marked for this desc */
1019 1.2.4.2 yamt
1020 1.2.4.2 yamt if (rxstat & R0_ES) {
1021 1.2.4.2 yamt ifp->if_ierrors++;
1022 1.2.4.2 yamt #define PRINTERR(bit, str) \
1023 1.2.4.2 yamt if (rxstat & (bit)) \
1024 1.2.4.2 yamt printf("%s: receive error: %s\n", \
1025 1.2.4.2 yamt sc->sc_dev.dv_xname, str)
1026 1.2.4.2 yamt PRINTERR(R0_TL, "frame too long");
1027 1.2.4.2 yamt PRINTERR(R0_RF, "runt frame");
1028 1.2.4.2 yamt PRINTERR(R0_CE, "bad FCS");
1029 1.2.4.2 yamt #undef PRINTERR
1030 1.2.4.2 yamt KSE_INIT_RXDESC(sc, i);
1031 1.2.4.2 yamt continue;
1032 1.2.4.2 yamt }
1033 1.2.4.2 yamt
1034 1.2.4.2 yamt /* HW errata; frame might be too small or too large */
1035 1.2.4.2 yamt
1036 1.2.4.2 yamt bus_dmamap_sync(sc->sc_dmat, rxs->rxs_dmamap, 0,
1037 1.2.4.2 yamt rxs->rxs_dmamap->dm_mapsize, BUS_DMASYNC_POSTREAD);
1038 1.2.4.2 yamt
1039 1.2.4.2 yamt len = rxstat & R0_FL_MASK;
1040 1.2.4.2 yamt len -= ETHER_CRC_LEN; /* trim CRC off */
1041 1.2.4.2 yamt m = rxs->rxs_mbuf;
1042 1.2.4.2 yamt
1043 1.2.4.2 yamt if (add_rxbuf(sc, i) != 0) {
1044 1.2.4.2 yamt ifp->if_ierrors++;
1045 1.2.4.2 yamt KSE_INIT_RXDESC(sc, i);
1046 1.2.4.2 yamt bus_dmamap_sync(sc->sc_dmat,
1047 1.2.4.2 yamt rxs->rxs_dmamap, 0,
1048 1.2.4.2 yamt rxs->rxs_dmamap->dm_mapsize,
1049 1.2.4.2 yamt BUS_DMASYNC_PREREAD);
1050 1.2.4.2 yamt continue;
1051 1.2.4.2 yamt }
1052 1.2.4.2 yamt
1053 1.2.4.2 yamt ifp->if_ipackets++;
1054 1.2.4.2 yamt m->m_pkthdr.rcvif = ifp;
1055 1.2.4.2 yamt m->m_pkthdr.len = m->m_len = len;
1056 1.2.4.2 yamt
1057 1.2.4.2 yamt if (sc->sc_mcsum) {
1058 1.2.4.2 yamt m->m_pkthdr.csum_flags |= sc->sc_mcsum;
1059 1.2.4.2 yamt if (rxstat & R0_IPE)
1060 1.2.4.2 yamt m->m_pkthdr.csum_flags |= M_CSUM_IPv4_BAD;
1061 1.2.4.2 yamt if (rxstat & (R0_TCPE | R0_UDPE))
1062 1.2.4.2 yamt m->m_pkthdr.csum_flags |= M_CSUM_TCP_UDP_BAD;
1063 1.2.4.2 yamt }
1064 1.2.4.2 yamt #if NBPFILTER > 0
1065 1.2.4.2 yamt if (ifp->if_bpf)
1066 1.2.4.2 yamt bpf_mtap(ifp->if_bpf, m);
1067 1.2.4.2 yamt #endif /* NBPFILTER > 0 */
1068 1.2.4.2 yamt (*ifp->if_input)(ifp, m);
1069 1.2.4.2 yamt #ifdef KSEDIAGNOSTIC
1070 1.2.4.2 yamt if (kse_monitor_rxintr > 0) {
1071 1.2.4.2 yamt printf("m stat %x data %p len %d\n",
1072 1.2.4.2 yamt rxstat, m->m_data, m->m_len);
1073 1.2.4.2 yamt }
1074 1.2.4.2 yamt #endif
1075 1.2.4.2 yamt }
1076 1.2.4.2 yamt sc->sc_rxptr = i;
1077 1.2.4.2 yamt }
1078 1.2.4.2 yamt
1079 1.2.4.2 yamt static void
1080 1.2.4.2 yamt txreap(struct kse_softc *sc)
1081 1.2.4.2 yamt {
1082 1.2.4.2 yamt struct ifnet *ifp = &sc->sc_ethercom.ec_if;
1083 1.2.4.2 yamt struct kse_txsoft *txs;
1084 1.2.4.2 yamt uint32_t txstat;
1085 1.2.4.2 yamt int i;
1086 1.2.4.2 yamt
1087 1.2.4.2 yamt ifp->if_flags &= ~IFF_OACTIVE;
1088 1.2.4.2 yamt
1089 1.2.4.2 yamt for (i = sc->sc_txsdirty; sc->sc_txsfree != KSE_TXQUEUELEN;
1090 1.2.4.2 yamt i = KSE_NEXTTXS(i), sc->sc_txsfree++) {
1091 1.2.4.2 yamt txs = &sc->sc_txsoft[i];
1092 1.2.4.2 yamt
1093 1.2.4.2 yamt KSE_CDTXSYNC(sc, txs->txs_firstdesc, txs->txs_ndesc,
1094 1.2.4.2 yamt BUS_DMASYNC_POSTREAD|BUS_DMASYNC_POSTWRITE);
1095 1.2.4.2 yamt
1096 1.2.4.2 yamt txstat = sc->sc_txdescs[txs->txs_lastdesc].t0;
1097 1.2.4.2 yamt
1098 1.2.4.2 yamt if (txstat & T0_OWN) /* desc is still in use */
1099 1.2.4.2 yamt break;
1100 1.2.4.2 yamt
1101 1.2.4.2 yamt /* there is no way to tell transmission status per frame */
1102 1.2.4.2 yamt
1103 1.2.4.2 yamt ifp->if_opackets++;
1104 1.2.4.2 yamt
1105 1.2.4.2 yamt sc->sc_txfree += txs->txs_ndesc;
1106 1.2.4.2 yamt bus_dmamap_sync(sc->sc_dmat, txs->txs_dmamap,
1107 1.2.4.2 yamt 0, txs->txs_dmamap->dm_mapsize, BUS_DMASYNC_POSTWRITE);
1108 1.2.4.2 yamt bus_dmamap_unload(sc->sc_dmat, txs->txs_dmamap);
1109 1.2.4.2 yamt m_freem(txs->txs_mbuf);
1110 1.2.4.2 yamt txs->txs_mbuf = NULL;
1111 1.2.4.2 yamt }
1112 1.2.4.2 yamt sc->sc_txsdirty = i;
1113 1.2.4.2 yamt if (sc->sc_txsfree == KSE_TXQUEUELEN)
1114 1.2.4.2 yamt ifp->if_timer = 0;
1115 1.2.4.2 yamt }
1116 1.2.4.2 yamt
1117 1.2.4.2 yamt static void
1118 1.2.4.2 yamt lnkchg(struct kse_softc *sc)
1119 1.2.4.2 yamt {
1120 1.2.4.2 yamt struct ifmediareq ifmr;
1121 1.2.4.2 yamt
1122 1.2.4.2 yamt #if 0 /* rambling link status */
1123 1.2.4.2 yamt printf("%s: link %s\n", sc->sc_dev.dv_xname,
1124 1.2.4.2 yamt (CSR_READ_2(sc, P1SR) & (1U << 5)) ? "up" : "down");
1125 1.2.4.2 yamt #endif
1126 1.2.4.2 yamt ifmedia_sts(&sc->sc_ethercom.ec_if, &ifmr);
1127 1.2.4.2 yamt }
1128 1.2.4.2 yamt
1129 1.2.4.2 yamt static int
1130 1.2.4.2 yamt ifmedia_upd(struct ifnet *ifp)
1131 1.2.4.2 yamt {
1132 1.2.4.2 yamt struct kse_softc *sc = ifp->if_softc;
1133 1.2.4.2 yamt struct ifmedia *ifm = &sc->sc_media;
1134 1.2.4.2 yamt uint16_t ctl;
1135 1.2.4.2 yamt
1136 1.2.4.2 yamt ctl = 0;
1137 1.2.4.2 yamt if (IFM_SUBTYPE(ifm->ifm_media) == IFM_AUTO) {
1138 1.2.4.2 yamt ctl |= (1U << 13); /* restart AN */
1139 1.2.4.2 yamt ctl |= (1U << 7); /* enable AN */
1140 1.2.4.2 yamt ctl |= (1U << 4); /* advertise flow control pause */
1141 1.2.4.2 yamt ctl |= (1U << 3) | (1U << 2) | (1U << 1) | (1U << 0);
1142 1.2.4.2 yamt }
1143 1.2.4.2 yamt else {
1144 1.2.4.2 yamt if (IFM_SUBTYPE(ifm->ifm_media) == IFM_100_TX)
1145 1.2.4.2 yamt ctl |= (1U << 6);
1146 1.2.4.2 yamt if (ifm->ifm_media & IFM_FDX)
1147 1.2.4.2 yamt ctl |= (1U << 5);
1148 1.2.4.2 yamt }
1149 1.2.4.2 yamt CSR_WRITE_2(sc, P1CR4, ctl);
1150 1.2.4.2 yamt
1151 1.2.4.2 yamt sc->sc_media_active = IFM_NONE;
1152 1.2.4.2 yamt sc->sc_media_status = IFM_AVALID;
1153 1.2.4.2 yamt
1154 1.2.4.2 yamt return 0;
1155 1.2.4.2 yamt }
1156 1.2.4.2 yamt
1157 1.2.4.2 yamt static void
1158 1.2.4.2 yamt ifmedia_sts(struct ifnet *ifp, struct ifmediareq *ifmr)
1159 1.2.4.2 yamt {
1160 1.2.4.2 yamt struct kse_softc *sc = ifp->if_softc;
1161 1.2.4.2 yamt struct ifmedia *ifm = &sc->sc_media;
1162 1.2.4.2 yamt uint16_t ctl, sts, result;
1163 1.2.4.2 yamt
1164 1.2.4.2 yamt ifmr->ifm_status = IFM_AVALID;
1165 1.2.4.2 yamt ifmr->ifm_active = IFM_ETHER;
1166 1.2.4.2 yamt
1167 1.2.4.2 yamt ctl = CSR_READ_2(sc, P1CR4);
1168 1.2.4.2 yamt sts = CSR_READ_2(sc, P1SR);
1169 1.2.4.2 yamt if ((sts & (1U << 5)) == 0) {
1170 1.2.4.2 yamt ifmr->ifm_active |= IFM_NONE;
1171 1.2.4.2 yamt goto out; /* link is down */
1172 1.2.4.2 yamt }
1173 1.2.4.2 yamt ifmr->ifm_status |= IFM_ACTIVE;
1174 1.2.4.2 yamt if (IFM_SUBTYPE(ifm->ifm_media) == IFM_AUTO) {
1175 1.2.4.2 yamt if ((sts & (1U << 6)) == 0) {
1176 1.2.4.2 yamt ifmr->ifm_active |= IFM_NONE;
1177 1.2.4.2 yamt goto out; /* negotiation in progress */
1178 1.2.4.2 yamt }
1179 1.2.4.2 yamt result = ctl & sts & 017;
1180 1.2.4.2 yamt if (result & (1U << 3))
1181 1.2.4.2 yamt ifmr->ifm_active |= IFM_100_TX|IFM_FDX;
1182 1.2.4.2 yamt else if (result & (1U << 2))
1183 1.2.4.2 yamt ifmr->ifm_active |= IFM_100_TX;
1184 1.2.4.2 yamt else if (result & (1U << 1))
1185 1.2.4.2 yamt ifmr->ifm_active |= IFM_10_T|IFM_FDX;
1186 1.2.4.2 yamt else if (result & (1U << 0))
1187 1.2.4.2 yamt ifmr->ifm_active |= IFM_10_T;
1188 1.2.4.2 yamt else
1189 1.2.4.2 yamt ifmr->ifm_active |= IFM_NONE;
1190 1.2.4.2 yamt if (ctl & (1U << 4))
1191 1.2.4.2 yamt ifmr->ifm_active |= IFM_FLOW | IFM_ETH_RXPAUSE;
1192 1.2.4.2 yamt if (sts & (1U << 4))
1193 1.2.4.2 yamt ifmr->ifm_active |= IFM_FLOW | IFM_ETH_TXPAUSE;
1194 1.2.4.2 yamt }
1195 1.2.4.2 yamt else {
1196 1.2.4.2 yamt ifmr->ifm_active |= (sts & (1U << 10)) ? IFM_100_TX : IFM_10_T;
1197 1.2.4.2 yamt if (sts & (1U << 9))
1198 1.2.4.2 yamt ifmr->ifm_active |= IFM_FDX;
1199 1.2.4.2 yamt if (sts & (1U << 12))
1200 1.2.4.2 yamt ifmr->ifm_active |= IFM_FLOW | IFM_ETH_RXPAUSE;
1201 1.2.4.2 yamt if (sts & (1U << 11))
1202 1.2.4.2 yamt ifmr->ifm_active |= IFM_FLOW | IFM_ETH_TXPAUSE;
1203 1.2.4.2 yamt }
1204 1.2.4.2 yamt
1205 1.2.4.2 yamt out:
1206 1.2.4.2 yamt sc->sc_media_status = ifmr->ifm_status;
1207 1.2.4.2 yamt sc->sc_media_active = ifmr->ifm_active;
1208 1.2.4.2 yamt }
1209 1.2.4.2 yamt
1210 1.2.4.2 yamt static void
1211 1.2.4.2 yamt phy_tick(void *arg)
1212 1.2.4.2 yamt {
1213 1.2.4.2 yamt struct kse_softc *sc = arg;
1214 1.2.4.2 yamt struct ifmediareq ifmr;
1215 1.2.4.2 yamt int s;
1216 1.2.4.2 yamt
1217 1.2.4.2 yamt s = splnet();
1218 1.2.4.2 yamt ifmedia_sts(&sc->sc_ethercom.ec_if, &ifmr);
1219 1.2.4.2 yamt splx(s);
1220 1.2.4.2 yamt
1221 1.2.4.2 yamt callout_reset(&sc->sc_callout, hz, phy_tick, sc);
1222 1.2.4.2 yamt }
1223