if_kse.c revision 1.2.4.6 1 1.2.4.6 yamt /* $NetBSD: if_kse.c,v 1.2.4.6 2008/02/11 14:59:38 yamt Exp $ */
2 1.2.4.2 yamt
3 1.2.4.2 yamt /*
4 1.2.4.2 yamt * Copyright (c) 2006 Tohru Nishimura
5 1.2.4.2 yamt *
6 1.2.4.2 yamt * Redistribution and use in source and binary forms, with or without
7 1.2.4.2 yamt * modification, are permitted provided that the following conditions
8 1.2.4.2 yamt * are met:
9 1.2.4.2 yamt * 1. Redistributions of source code must retain the above copyright
10 1.2.4.2 yamt * notice, this list of conditions and the following disclaimer.
11 1.2.4.2 yamt * 2. Redistributions in binary form must reproduce the above copyright
12 1.2.4.2 yamt * notice, this list of conditions and the following disclaimer in the
13 1.2.4.2 yamt * documentation and/or other materials provided with the distribution.
14 1.2.4.2 yamt * 3. All advertising materials mentioning features or use of this software
15 1.2.4.2 yamt * must display the following acknowledgement:
16 1.2.4.2 yamt * This product includes software developed by Tohru Nishimura.
17 1.2.4.2 yamt * 4. The name of the author may not be used to endorse or promote products
18 1.2.4.2 yamt * derived from this software without specific prior written permission.
19 1.2.4.2 yamt *
20 1.2.4.2 yamt * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
21 1.2.4.2 yamt * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
22 1.2.4.2 yamt * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
23 1.2.4.2 yamt * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
24 1.2.4.2 yamt * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
25 1.2.4.2 yamt * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
26 1.2.4.2 yamt * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
27 1.2.4.2 yamt * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
28 1.2.4.2 yamt * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
29 1.2.4.2 yamt * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
30 1.2.4.2 yamt */
31 1.2.4.2 yamt
32 1.2.4.2 yamt #include <sys/cdefs.h>
33 1.2.4.6 yamt __KERNEL_RCSID(0, "$NetBSD: if_kse.c,v 1.2.4.6 2008/02/11 14:59:38 yamt Exp $");
34 1.2.4.2 yamt
35 1.2.4.2 yamt #include "bpfilter.h"
36 1.2.4.2 yamt
37 1.2.4.2 yamt #include <sys/param.h>
38 1.2.4.2 yamt #include <sys/systm.h>
39 1.2.4.2 yamt #include <sys/callout.h>
40 1.2.4.2 yamt #include <sys/mbuf.h>
41 1.2.4.2 yamt #include <sys/malloc.h>
42 1.2.4.2 yamt #include <sys/kernel.h>
43 1.2.4.2 yamt #include <sys/ioctl.h>
44 1.2.4.2 yamt #include <sys/errno.h>
45 1.2.4.2 yamt #include <sys/device.h>
46 1.2.4.2 yamt #include <sys/queue.h>
47 1.2.4.2 yamt
48 1.2.4.2 yamt #include <machine/endian.h>
49 1.2.4.4 yamt #include <sys/bus.h>
50 1.2.4.4 yamt #include <sys/intr.h>
51 1.2.4.2 yamt
52 1.2.4.2 yamt #include <net/if.h>
53 1.2.4.2 yamt #include <net/if_media.h>
54 1.2.4.2 yamt #include <net/if_dl.h>
55 1.2.4.2 yamt #include <net/if_ether.h>
56 1.2.4.2 yamt
57 1.2.4.2 yamt #if NBPFILTER > 0
58 1.2.4.2 yamt #include <net/bpf.h>
59 1.2.4.2 yamt #endif
60 1.2.4.2 yamt
61 1.2.4.2 yamt #include <dev/pci/pcivar.h>
62 1.2.4.2 yamt #include <dev/pci/pcireg.h>
63 1.2.4.2 yamt #include <dev/pci/pcidevs.h>
64 1.2.4.2 yamt
65 1.2.4.2 yamt #define CSR_READ_4(sc, off) \
66 1.2.4.2 yamt bus_space_read_4(sc->sc_st, sc->sc_sh, off)
67 1.2.4.2 yamt #define CSR_WRITE_4(sc, off, val) \
68 1.2.4.2 yamt bus_space_write_4(sc->sc_st, sc->sc_sh, off, val)
69 1.2.4.2 yamt #define CSR_READ_2(sc, off) \
70 1.2.4.2 yamt bus_space_read_2(sc->sc_st, sc->sc_sh, off)
71 1.2.4.2 yamt #define CSR_WRITE_2(sc, off, val) \
72 1.2.4.2 yamt bus_space_write_2(sc->sc_st, sc->sc_sh, off, val)
73 1.2.4.2 yamt
74 1.2.4.2 yamt #define MDTXC 0x000 /* DMA transmit control */
75 1.2.4.2 yamt #define MDRXC 0x004 /* DMA receive control */
76 1.2.4.2 yamt #define MDTSC 0x008 /* DMA transmit start */
77 1.2.4.2 yamt #define MDRSC 0x00c /* DMA receive start */
78 1.2.4.2 yamt #define TDLB 0x010 /* transmit descriptor list base */
79 1.2.4.2 yamt #define RDLB 0x014 /* receive descriptor list base */
80 1.2.4.4 yamt #define MTR0 0x020 /* multicast table 31:0 */
81 1.2.4.4 yamt #define MTR1 0x024 /* multicast table 63:32 */
82 1.2.4.2 yamt #define INTEN 0x028 /* interrupt enable */
83 1.2.4.2 yamt #define INTST 0x02c /* interrupt status */
84 1.2.4.2 yamt #define MARL 0x200 /* MAC address low */
85 1.2.4.2 yamt #define MARM 0x202 /* MAC address middle */
86 1.2.4.2 yamt #define MARH 0x204 /* MAC address high */
87 1.2.4.2 yamt #define GRR 0x216 /* global reset */
88 1.2.4.2 yamt #define CIDR 0x400 /* chip ID and enable */
89 1.2.4.2 yamt #define CGCR 0x40a /* chip global control */
90 1.2.4.4 yamt #define IACR 0x4a0 /* indirect access control */
91 1.2.4.4 yamt #define IADR1 0x4a2 /* indirect access data 66:63 */
92 1.2.4.4 yamt #define IADR2 0x4a4 /* indirect access data 47:32 */
93 1.2.4.4 yamt #define IADR3 0x4a6 /* indirect access data 63:48 */
94 1.2.4.4 yamt #define IADR4 0x4a8 /* indirect access data 15:0 */
95 1.2.4.4 yamt #define IADR5 0x4aa /* indirect access data 31:16 */
96 1.2.4.2 yamt #define P1CR4 0x512 /* port 1 control 4 */
97 1.2.4.2 yamt #define P1SR 0x514 /* port 1 status */
98 1.2.4.4 yamt #define P2CR4 0x532 /* port 2 control 4 */
99 1.2.4.4 yamt #define P2SR 0x534 /* port 2 status */
100 1.2.4.2 yamt
101 1.2.4.2 yamt #define TXC_BS_MSK 0x3f000000 /* burst size */
102 1.2.4.2 yamt #define TXC_BS_SFT (24) /* 1,2,4,8,16,32 or 0 for unlimited */
103 1.2.4.2 yamt #define TXC_UCG (1U<<18) /* generate UDP checksum */
104 1.2.4.2 yamt #define TXC_TCG (1U<<17) /* generate TCP checksum */
105 1.2.4.2 yamt #define TXC_ICG (1U<<16) /* generate IP checksum */
106 1.2.4.2 yamt #define TXC_FCE (1U<<9) /* enable flowcontrol */
107 1.2.4.2 yamt #define TXC_EP (1U<<2) /* enable automatic padding */
108 1.2.4.2 yamt #define TXC_AC (1U<<1) /* add CRC to frame */
109 1.2.4.2 yamt #define TXC_TEN (1) /* enable DMA to run */
110 1.2.4.2 yamt
111 1.2.4.2 yamt #define RXC_BS_MSK 0x3f000000 /* burst size */
112 1.2.4.2 yamt #define RXC_BS_SFT (24) /* 1,2,4,8,16,32 or 0 for unlimited */
113 1.2.4.4 yamt #define RXC_IHAE (1U<<19) /* IP header alignment enable */
114 1.2.4.4 yamt #define RXC_UCC (1U<<18) /* run UDP checksum */
115 1.2.4.4 yamt #define RXC_TCC (1U<<17) /* run TDP checksum */
116 1.2.4.4 yamt #define RXC_ICC (1U<<16) /* run IP checksum */
117 1.2.4.2 yamt #define RXC_FCE (1U<<9) /* enable flowcontrol */
118 1.2.4.2 yamt #define RXC_RB (1U<<6) /* receive broadcast frame */
119 1.2.4.2 yamt #define RXC_RM (1U<<5) /* receive multicast frame */
120 1.2.4.2 yamt #define RXC_RU (1U<<4) /* receive unicast frame */
121 1.2.4.2 yamt #define RXC_RE (1U<<3) /* accept error frame */
122 1.2.4.2 yamt #define RXC_RA (1U<<2) /* receive all frame */
123 1.2.4.4 yamt #define RXC_MHTE (1U<<1) /* use multicast hash table */
124 1.2.4.2 yamt #define RXC_REN (1) /* enable DMA to run */
125 1.2.4.2 yamt
126 1.2.4.2 yamt #define INT_DMLCS (1U<<31) /* link status change */
127 1.2.4.2 yamt #define INT_DMTS (1U<<30) /* sending desc. has posted Tx done */
128 1.2.4.2 yamt #define INT_DMRS (1U<<29) /* frame was received */
129 1.2.4.2 yamt #define INT_DMRBUS (1U<<27) /* Rx descriptor pool is full */
130 1.2.4.2 yamt
131 1.2.4.2 yamt #define T0_OWN (1U<<31) /* desc is ready to Tx */
132 1.2.4.2 yamt
133 1.2.4.2 yamt #define R0_OWN (1U<<31) /* desc is empty */
134 1.2.4.2 yamt #define R0_FS (1U<<30) /* first segment of frame */
135 1.2.4.2 yamt #define R0_LS (1U<<29) /* last segment of frame */
136 1.2.4.2 yamt #define R0_IPE (1U<<28) /* IP checksum error */
137 1.2.4.2 yamt #define R0_TCPE (1U<<27) /* TCP checksum error */
138 1.2.4.2 yamt #define R0_UDPE (1U<<26) /* UDP checksum error */
139 1.2.4.2 yamt #define R0_ES (1U<<25) /* error summary */
140 1.2.4.2 yamt #define R0_MF (1U<<24) /* multicast frame */
141 1.2.4.4 yamt #define R0_SPN 0x00300000 /* 21:20 switch port 1/2 */
142 1.2.4.4 yamt #define R0_ALIGN 0x00300000 /* 21:20 (KSZ8692P) Rx align amount */
143 1.2.4.4 yamt #define R0_RE (1U<<19) /* MII reported error */
144 1.2.4.4 yamt #define R0_TL (1U<<18) /* frame too long, beyond 1518 */
145 1.2.4.2 yamt #define R0_RF (1U<<17) /* damaged runt frame */
146 1.2.4.2 yamt #define R0_CE (1U<<16) /* CRC error */
147 1.2.4.2 yamt #define R0_FT (1U<<15) /* frame type */
148 1.2.4.2 yamt #define R0_FL_MASK 0x7ff /* frame length 10:0 */
149 1.2.4.2 yamt
150 1.2.4.2 yamt #define T1_IC (1U<<31) /* post interrupt on complete */
151 1.2.4.2 yamt #define T1_FS (1U<<30) /* first segment of frame */
152 1.2.4.2 yamt #define T1_LS (1U<<29) /* last segment of frame */
153 1.2.4.2 yamt #define T1_IPCKG (1U<<28) /* generate IP checksum */
154 1.2.4.2 yamt #define T1_TCPCKG (1U<<27) /* generate TCP checksum */
155 1.2.4.2 yamt #define T1_UDPCKG (1U<<26) /* generate UDP checksum */
156 1.2.4.2 yamt #define T1_TER (1U<<25) /* end of ring */
157 1.2.4.4 yamt #define T1_SPN 0x00300000 /* 21:20 switch port 1/2 */
158 1.2.4.2 yamt #define T1_TBS_MASK 0x7ff /* segment size 10:0 */
159 1.2.4.2 yamt
160 1.2.4.2 yamt #define R1_RER (1U<<25) /* end of ring */
161 1.2.4.4 yamt #define R1_RBS_MASK 0x7fc /* segment size 10:0 */
162 1.2.4.2 yamt
163 1.2.4.2 yamt #define KSE_NTXSEGS 16
164 1.2.4.2 yamt #define KSE_TXQUEUELEN 64
165 1.2.4.2 yamt #define KSE_TXQUEUELEN_MASK (KSE_TXQUEUELEN - 1)
166 1.2.4.2 yamt #define KSE_TXQUEUE_GC (KSE_TXQUEUELEN / 4)
167 1.2.4.2 yamt #define KSE_NTXDESC 256
168 1.2.4.2 yamt #define KSE_NTXDESC_MASK (KSE_NTXDESC - 1)
169 1.2.4.2 yamt #define KSE_NEXTTX(x) (((x) + 1) & KSE_NTXDESC_MASK)
170 1.2.4.2 yamt #define KSE_NEXTTXS(x) (((x) + 1) & KSE_TXQUEUELEN_MASK)
171 1.2.4.2 yamt
172 1.2.4.2 yamt #define KSE_NRXDESC 64
173 1.2.4.2 yamt #define KSE_NRXDESC_MASK (KSE_NRXDESC - 1)
174 1.2.4.2 yamt #define KSE_NEXTRX(x) (((x) + 1) & KSE_NRXDESC_MASK)
175 1.2.4.2 yamt
176 1.2.4.2 yamt struct tdes {
177 1.2.4.2 yamt uint32_t t0, t1, t2, t3;
178 1.2.4.2 yamt };
179 1.2.4.2 yamt
180 1.2.4.2 yamt struct rdes {
181 1.2.4.2 yamt uint32_t r0, r1, r2, r3;
182 1.2.4.2 yamt };
183 1.2.4.2 yamt
184 1.2.4.2 yamt struct kse_control_data {
185 1.2.4.2 yamt struct tdes kcd_txdescs[KSE_NTXDESC];
186 1.2.4.2 yamt struct rdes kcd_rxdescs[KSE_NRXDESC];
187 1.2.4.2 yamt };
188 1.2.4.2 yamt #define KSE_CDOFF(x) offsetof(struct kse_control_data, x)
189 1.2.4.2 yamt #define KSE_CDTXOFF(x) KSE_CDOFF(kcd_txdescs[(x)])
190 1.2.4.2 yamt #define KSE_CDRXOFF(x) KSE_CDOFF(kcd_rxdescs[(x)])
191 1.2.4.2 yamt
192 1.2.4.2 yamt struct kse_txsoft {
193 1.2.4.2 yamt struct mbuf *txs_mbuf; /* head of our mbuf chain */
194 1.2.4.2 yamt bus_dmamap_t txs_dmamap; /* our DMA map */
195 1.2.4.2 yamt int txs_firstdesc; /* first descriptor in packet */
196 1.2.4.2 yamt int txs_lastdesc; /* last descriptor in packet */
197 1.2.4.2 yamt int txs_ndesc; /* # of descriptors used */
198 1.2.4.2 yamt };
199 1.2.4.2 yamt
200 1.2.4.2 yamt struct kse_rxsoft {
201 1.2.4.2 yamt struct mbuf *rxs_mbuf; /* head of our mbuf chain */
202 1.2.4.2 yamt bus_dmamap_t rxs_dmamap; /* our DMA map */
203 1.2.4.2 yamt };
204 1.2.4.2 yamt
205 1.2.4.2 yamt struct kse_softc {
206 1.2.4.2 yamt struct device sc_dev; /* generic device information */
207 1.2.4.2 yamt bus_space_tag_t sc_st; /* bus space tag */
208 1.2.4.2 yamt bus_space_handle_t sc_sh; /* bus space handle */
209 1.2.4.2 yamt bus_dma_tag_t sc_dmat; /* bus DMA tag */
210 1.2.4.2 yamt struct ethercom sc_ethercom; /* Ethernet common data */
211 1.2.4.2 yamt void *sc_ih; /* interrupt cookie */
212 1.2.4.2 yamt
213 1.2.4.2 yamt struct ifmedia sc_media; /* ifmedia information */
214 1.2.4.2 yamt int sc_media_status; /* PHY */
215 1.2.4.2 yamt int sc_media_active; /* PHY */
216 1.2.4.4 yamt callout_t sc_callout; /* MII tick callout */
217 1.2.4.4 yamt callout_t sc_stat_ch; /* statistics counter callout */
218 1.2.4.2 yamt
219 1.2.4.2 yamt bus_dmamap_t sc_cddmamap; /* control data DMA map */
220 1.2.4.2 yamt #define sc_cddma sc_cddmamap->dm_segs[0].ds_addr
221 1.2.4.2 yamt
222 1.2.4.2 yamt struct kse_control_data *sc_control_data;
223 1.2.4.4 yamt #define sc_txdescs sc_control_data->kcd_txdescs
224 1.2.4.4 yamt #define sc_rxdescs sc_control_data->kcd_rxdescs
225 1.2.4.2 yamt
226 1.2.4.2 yamt struct kse_txsoft sc_txsoft[KSE_TXQUEUELEN];
227 1.2.4.2 yamt struct kse_rxsoft sc_rxsoft[KSE_NRXDESC];
228 1.2.4.2 yamt int sc_txfree; /* number of free Tx descriptors */
229 1.2.4.2 yamt int sc_txnext; /* next ready Tx descriptor */
230 1.2.4.2 yamt int sc_txsfree; /* number of free Tx jobs */
231 1.2.4.2 yamt int sc_txsnext; /* next ready Tx job */
232 1.2.4.2 yamt int sc_txsdirty; /* dirty Tx jobs */
233 1.2.4.2 yamt int sc_rxptr; /* next ready Rx descriptor/descsoft */
234 1.2.4.2 yamt
235 1.2.4.2 yamt uint32_t sc_txc, sc_rxc;
236 1.2.4.2 yamt uint32_t sc_t1csum;
237 1.2.4.2 yamt int sc_mcsum;
238 1.2.4.4 yamt uint32_t sc_inten;
239 1.2.4.4 yamt
240 1.2.4.2 yamt uint32_t sc_chip;
241 1.2.4.4 yamt uint8_t sc_altmac[16][ETHER_ADDR_LEN];
242 1.2.4.4 yamt uint16_t sc_vlan[16];
243 1.2.4.4 yamt
244 1.2.4.4 yamt #ifdef KSE_EVENT_COUNTERS
245 1.2.4.4 yamt struct ksext {
246 1.2.4.4 yamt char evcntname[3][8];
247 1.2.4.4 yamt struct evcnt pev[3][34];
248 1.2.4.4 yamt } sc_ext; /* switch statistics */
249 1.2.4.4 yamt #endif
250 1.2.4.2 yamt };
251 1.2.4.2 yamt
252 1.2.4.2 yamt #define KSE_CDTXADDR(sc, x) ((sc)->sc_cddma + KSE_CDTXOFF((x)))
253 1.2.4.2 yamt #define KSE_CDRXADDR(sc, x) ((sc)->sc_cddma + KSE_CDRXOFF((x)))
254 1.2.4.2 yamt
255 1.2.4.2 yamt #define KSE_CDTXSYNC(sc, x, n, ops) \
256 1.2.4.2 yamt do { \
257 1.2.4.2 yamt int __x, __n; \
258 1.2.4.2 yamt \
259 1.2.4.2 yamt __x = (x); \
260 1.2.4.2 yamt __n = (n); \
261 1.2.4.2 yamt \
262 1.2.4.2 yamt /* If it will wrap around, sync to the end of the ring. */ \
263 1.2.4.2 yamt if ((__x + __n) > KSE_NTXDESC) { \
264 1.2.4.2 yamt bus_dmamap_sync((sc)->sc_dmat, (sc)->sc_cddmamap, \
265 1.2.4.2 yamt KSE_CDTXOFF(__x), sizeof(struct tdes) * \
266 1.2.4.2 yamt (KSE_NTXDESC - __x), (ops)); \
267 1.2.4.2 yamt __n -= (KSE_NTXDESC - __x); \
268 1.2.4.2 yamt __x = 0; \
269 1.2.4.2 yamt } \
270 1.2.4.2 yamt \
271 1.2.4.2 yamt /* Now sync whatever is left. */ \
272 1.2.4.2 yamt bus_dmamap_sync((sc)->sc_dmat, (sc)->sc_cddmamap, \
273 1.2.4.2 yamt KSE_CDTXOFF(__x), sizeof(struct tdes) * __n, (ops)); \
274 1.2.4.2 yamt } while (/*CONSTCOND*/0)
275 1.2.4.2 yamt
276 1.2.4.2 yamt #define KSE_CDRXSYNC(sc, x, ops) \
277 1.2.4.2 yamt do { \
278 1.2.4.2 yamt bus_dmamap_sync((sc)->sc_dmat, (sc)->sc_cddmamap, \
279 1.2.4.2 yamt KSE_CDRXOFF((x)), sizeof(struct rdes), (ops)); \
280 1.2.4.2 yamt } while (/*CONSTCOND*/0)
281 1.2.4.2 yamt
282 1.2.4.2 yamt #define KSE_INIT_RXDESC(sc, x) \
283 1.2.4.2 yamt do { \
284 1.2.4.2 yamt struct kse_rxsoft *__rxs = &(sc)->sc_rxsoft[(x)]; \
285 1.2.4.2 yamt struct rdes *__rxd = &(sc)->sc_rxdescs[(x)]; \
286 1.2.4.2 yamt struct mbuf *__m = __rxs->rxs_mbuf; \
287 1.2.4.2 yamt \
288 1.2.4.2 yamt __m->m_data = __m->m_ext.ext_buf; \
289 1.2.4.2 yamt __rxd->r2 = __rxs->rxs_dmamap->dm_segs[0].ds_addr; \
290 1.2.4.2 yamt __rxd->r1 = R1_RBS_MASK /* __m->m_ext.ext_size */; \
291 1.2.4.2 yamt __rxd->r0 = R0_OWN; \
292 1.2.4.2 yamt KSE_CDRXSYNC((sc), (x), BUS_DMASYNC_PREREAD|BUS_DMASYNC_PREWRITE); \
293 1.2.4.2 yamt } while (/*CONSTCOND*/0)
294 1.2.4.2 yamt
295 1.2.4.5 yamt u_int kse_burstsize = 8; /* DMA burst length tuning knob */
296 1.2.4.2 yamt
297 1.2.4.2 yamt #ifdef KSEDIAGNOSTIC
298 1.2.4.2 yamt u_int kse_monitor_rxintr; /* fragmented UDP csum HW bug hook */
299 1.2.4.2 yamt #endif
300 1.2.4.2 yamt
301 1.2.4.2 yamt static int kse_match(struct device *, struct cfdata *, void *);
302 1.2.4.2 yamt static void kse_attach(struct device *, struct device *, void *);
303 1.2.4.2 yamt
304 1.2.4.2 yamt CFATTACH_DECL(kse, sizeof(struct kse_softc),
305 1.2.4.2 yamt kse_match, kse_attach, NULL, NULL);
306 1.2.4.2 yamt
307 1.2.4.3 yamt static int kse_ioctl(struct ifnet *, u_long, void *);
308 1.2.4.2 yamt static void kse_start(struct ifnet *);
309 1.2.4.2 yamt static void kse_watchdog(struct ifnet *);
310 1.2.4.2 yamt static int kse_init(struct ifnet *);
311 1.2.4.2 yamt static void kse_stop(struct ifnet *, int);
312 1.2.4.2 yamt static void kse_reset(struct kse_softc *);
313 1.2.4.2 yamt static void kse_set_filter(struct kse_softc *);
314 1.2.4.2 yamt static int add_rxbuf(struct kse_softc *, int);
315 1.2.4.2 yamt static void rxdrain(struct kse_softc *);
316 1.2.4.2 yamt static int kse_intr(void *);
317 1.2.4.2 yamt static void rxintr(struct kse_softc *);
318 1.2.4.2 yamt static void txreap(struct kse_softc *);
319 1.2.4.2 yamt static void lnkchg(struct kse_softc *);
320 1.2.4.2 yamt static int ifmedia_upd(struct ifnet *);
321 1.2.4.2 yamt static void ifmedia_sts(struct ifnet *, struct ifmediareq *);
322 1.2.4.2 yamt static void phy_tick(void *);
323 1.2.4.4 yamt static int ifmedia2_upd(struct ifnet *);
324 1.2.4.4 yamt static void ifmedia2_sts(struct ifnet *, struct ifmediareq *);
325 1.2.4.4 yamt #ifdef KSE_EVENT_COUNTERS
326 1.2.4.4 yamt static void stat_tick(void *);
327 1.2.4.4 yamt static void zerostats(struct kse_softc *);
328 1.2.4.4 yamt #endif
329 1.2.4.2 yamt
330 1.2.4.2 yamt static int
331 1.2.4.2 yamt kse_match(struct device *parent, struct cfdata *match, void *aux)
332 1.2.4.2 yamt {
333 1.2.4.2 yamt struct pci_attach_args *pa = (struct pci_attach_args *)aux;
334 1.2.4.2 yamt
335 1.2.4.2 yamt if (PCI_VENDOR(pa->pa_id) == PCI_VENDOR_MICREL &&
336 1.2.4.2 yamt (PCI_PRODUCT(pa->pa_id) == PCI_PRODUCT_MICREL_KSZ8842 ||
337 1.2.4.2 yamt PCI_PRODUCT(pa->pa_id) == PCI_PRODUCT_MICREL_KSZ8841) &&
338 1.2.4.2 yamt PCI_CLASS(pa->pa_class) == PCI_CLASS_NETWORK)
339 1.2.4.2 yamt return 1;
340 1.2.4.2 yamt
341 1.2.4.2 yamt return 0;
342 1.2.4.2 yamt }
343 1.2.4.2 yamt
344 1.2.4.2 yamt static void
345 1.2.4.2 yamt kse_attach(struct device *parent, struct device *self, void *aux)
346 1.2.4.2 yamt {
347 1.2.4.2 yamt struct kse_softc *sc = (struct kse_softc *)self;
348 1.2.4.2 yamt struct pci_attach_args *pa = aux;
349 1.2.4.2 yamt pci_chipset_tag_t pc = pa->pa_pc;
350 1.2.4.2 yamt pci_intr_handle_t ih;
351 1.2.4.2 yamt const char *intrstr;
352 1.2.4.2 yamt struct ifnet *ifp;
353 1.2.4.4 yamt struct ifmedia *ifm;
354 1.2.4.2 yamt uint8_t enaddr[ETHER_ADDR_LEN];
355 1.2.4.2 yamt bus_dma_segment_t seg;
356 1.2.4.4 yamt int i, p, error, nseg;
357 1.2.4.2 yamt pcireg_t pmode;
358 1.2.4.2 yamt int pmreg;
359 1.2.4.2 yamt
360 1.2.4.2 yamt if (pci_mapreg_map(pa, 0x10,
361 1.2.4.2 yamt PCI_MAPREG_TYPE_MEM | PCI_MAPREG_MEM_TYPE_32BIT,
362 1.2.4.2 yamt 0, &sc->sc_st, &sc->sc_sh, NULL, NULL) != 0) {
363 1.2.4.2 yamt printf(": unable to map device registers\n");
364 1.2.4.2 yamt return;
365 1.2.4.2 yamt }
366 1.2.4.2 yamt
367 1.2.4.2 yamt sc->sc_dmat = pa->pa_dmat;
368 1.2.4.2 yamt
369 1.2.4.2 yamt /* Make sure bus mastering is enabled. */
370 1.2.4.2 yamt pci_conf_write(pc, pa->pa_tag, PCI_COMMAND_STATUS_REG,
371 1.2.4.2 yamt pci_conf_read(pc, pa->pa_tag, PCI_COMMAND_STATUS_REG) |
372 1.2.4.2 yamt PCI_COMMAND_MASTER_ENABLE);
373 1.2.4.2 yamt
374 1.2.4.2 yamt /* Get it out of power save mode, if needed. */
375 1.2.4.2 yamt if (pci_get_capability(pc, pa->pa_tag, PCI_CAP_PWRMGMT, &pmreg, 0)) {
376 1.2.4.2 yamt pmode = pci_conf_read(pc, pa->pa_tag, pmreg + PCI_PMCSR) &
377 1.2.4.2 yamt PCI_PMCSR_STATE_MASK;
378 1.2.4.2 yamt if (pmode == PCI_PMCSR_STATE_D3) {
379 1.2.4.2 yamt /*
380 1.2.4.2 yamt * The card has lost all configuration data in
381 1.2.4.2 yamt * this state, so punt.
382 1.2.4.2 yamt */
383 1.2.4.2 yamt printf("%s: unable to wake from power state D3\n",
384 1.2.4.2 yamt sc->sc_dev.dv_xname);
385 1.2.4.2 yamt return;
386 1.2.4.2 yamt }
387 1.2.4.2 yamt if (pmode != PCI_PMCSR_STATE_D0) {
388 1.2.4.2 yamt printf("%s: waking up from power date D%d\n",
389 1.2.4.2 yamt sc->sc_dev.dv_xname, pmode);
390 1.2.4.2 yamt pci_conf_write(pc, pa->pa_tag, pmreg + PCI_PMCSR,
391 1.2.4.2 yamt PCI_PMCSR_STATE_D0);
392 1.2.4.2 yamt }
393 1.2.4.2 yamt }
394 1.2.4.2 yamt
395 1.2.4.2 yamt sc->sc_chip = PCI_PRODUCT(pa->pa_id);
396 1.2.4.2 yamt printf(": Micrel KSZ%04x Ethernet (rev. 0x%02x)\n",
397 1.2.4.2 yamt sc->sc_chip, PCI_REVISION(pa->pa_class));
398 1.2.4.2 yamt
399 1.2.4.2 yamt /*
400 1.2.4.2 yamt * Read the Ethernet address from the EEPROM.
401 1.2.4.2 yamt */
402 1.2.4.2 yamt i = CSR_READ_2(sc, MARL);
403 1.2.4.2 yamt enaddr[5] = i; enaddr[4] = i >> 8;
404 1.2.4.2 yamt i = CSR_READ_2(sc, MARM);
405 1.2.4.2 yamt enaddr[3] = i; enaddr[2] = i >> 8;
406 1.2.4.2 yamt i = CSR_READ_2(sc, MARH);
407 1.2.4.2 yamt enaddr[1] = i; enaddr[0] = i >> 8;
408 1.2.4.2 yamt printf("%s: Ethernet address: %s\n",
409 1.2.4.2 yamt sc->sc_dev.dv_xname, ether_sprintf(enaddr));
410 1.2.4.2 yamt
411 1.2.4.2 yamt /*
412 1.2.4.2 yamt * Enable chip function.
413 1.2.4.2 yamt */
414 1.2.4.2 yamt CSR_WRITE_2(sc, CIDR, 1);
415 1.2.4.2 yamt
416 1.2.4.2 yamt /*
417 1.2.4.2 yamt * Map and establish our interrupt.
418 1.2.4.2 yamt */
419 1.2.4.2 yamt if (pci_intr_map(pa, &ih)) {
420 1.2.4.2 yamt printf("%s: unable to map interrupt\n", sc->sc_dev.dv_xname);
421 1.2.4.2 yamt return;
422 1.2.4.2 yamt }
423 1.2.4.2 yamt intrstr = pci_intr_string(pc, ih);
424 1.2.4.2 yamt sc->sc_ih = pci_intr_establish(pc, ih, IPL_NET, kse_intr, sc);
425 1.2.4.2 yamt if (sc->sc_ih == NULL) {
426 1.2.4.2 yamt printf("%s: unable to establish interrupt",
427 1.2.4.2 yamt sc->sc_dev.dv_xname);
428 1.2.4.2 yamt if (intrstr != NULL)
429 1.2.4.2 yamt printf(" at %s", intrstr);
430 1.2.4.2 yamt printf("\n");
431 1.2.4.2 yamt return;
432 1.2.4.2 yamt }
433 1.2.4.2 yamt printf("%s: interrupting at %s\n", sc->sc_dev.dv_xname, intrstr);
434 1.2.4.2 yamt
435 1.2.4.2 yamt /*
436 1.2.4.2 yamt * Allocate the control data structures, and create and load the
437 1.2.4.2 yamt * DMA map for it.
438 1.2.4.2 yamt */
439 1.2.4.2 yamt error = bus_dmamem_alloc(sc->sc_dmat,
440 1.2.4.2 yamt sizeof(struct kse_control_data), PAGE_SIZE, 0, &seg, 1, &nseg, 0);
441 1.2.4.2 yamt if (error != 0) {
442 1.2.4.2 yamt printf("%s: unable to allocate control data, error = %d\n",
443 1.2.4.2 yamt sc->sc_dev.dv_xname, error);
444 1.2.4.2 yamt goto fail_0;
445 1.2.4.2 yamt }
446 1.2.4.2 yamt error = bus_dmamem_map(sc->sc_dmat, &seg, nseg,
447 1.2.4.3 yamt sizeof(struct kse_control_data), (void **)&sc->sc_control_data,
448 1.2.4.2 yamt BUS_DMA_COHERENT);
449 1.2.4.2 yamt if (error != 0) {
450 1.2.4.2 yamt printf("%s: unable to map control data, error = %d\n",
451 1.2.4.2 yamt sc->sc_dev.dv_xname, error);
452 1.2.4.2 yamt goto fail_1;
453 1.2.4.2 yamt }
454 1.2.4.2 yamt error = bus_dmamap_create(sc->sc_dmat,
455 1.2.4.2 yamt sizeof(struct kse_control_data), 1,
456 1.2.4.2 yamt sizeof(struct kse_control_data), 0, 0, &sc->sc_cddmamap);
457 1.2.4.2 yamt if (error != 0) {
458 1.2.4.2 yamt printf("%s: unable to create control data DMA map, "
459 1.2.4.2 yamt "error = %d\n", sc->sc_dev.dv_xname, error);
460 1.2.4.2 yamt goto fail_2;
461 1.2.4.2 yamt }
462 1.2.4.2 yamt error = bus_dmamap_load(sc->sc_dmat, sc->sc_cddmamap,
463 1.2.4.2 yamt sc->sc_control_data, sizeof(struct kse_control_data), NULL, 0);
464 1.2.4.2 yamt if (error != 0) {
465 1.2.4.2 yamt printf("%s: unable to load control data DMA map, error = %d\n",
466 1.2.4.2 yamt sc->sc_dev.dv_xname, error);
467 1.2.4.2 yamt goto fail_3;
468 1.2.4.2 yamt }
469 1.2.4.2 yamt for (i = 0; i < KSE_TXQUEUELEN; i++) {
470 1.2.4.2 yamt if ((error = bus_dmamap_create(sc->sc_dmat, MCLBYTES,
471 1.2.4.2 yamt KSE_NTXSEGS, MCLBYTES, 0, 0,
472 1.2.4.2 yamt &sc->sc_txsoft[i].txs_dmamap)) != 0) {
473 1.2.4.2 yamt printf("%s: unable to create tx DMA map %d, "
474 1.2.4.2 yamt "error = %d\n", sc->sc_dev.dv_xname, i, error);
475 1.2.4.2 yamt goto fail_4;
476 1.2.4.2 yamt }
477 1.2.4.2 yamt }
478 1.2.4.2 yamt for (i = 0; i < KSE_NRXDESC; i++) {
479 1.2.4.2 yamt if ((error = bus_dmamap_create(sc->sc_dmat, MCLBYTES,
480 1.2.4.2 yamt 1, MCLBYTES, 0, 0, &sc->sc_rxsoft[i].rxs_dmamap)) != 0) {
481 1.2.4.2 yamt printf("%s: unable to create rx DMA map %d, "
482 1.2.4.2 yamt "error = %d\n", sc->sc_dev.dv_xname, i, error);
483 1.2.4.2 yamt goto fail_5;
484 1.2.4.2 yamt }
485 1.2.4.2 yamt sc->sc_rxsoft[i].rxs_mbuf = NULL;
486 1.2.4.2 yamt }
487 1.2.4.2 yamt
488 1.2.4.3 yamt callout_init(&sc->sc_callout, 0);
489 1.2.4.4 yamt callout_init(&sc->sc_stat_ch, 0);
490 1.2.4.2 yamt
491 1.2.4.4 yamt ifm = &sc->sc_media;
492 1.2.4.4 yamt if (sc->sc_chip == 0x8841) {
493 1.2.4.4 yamt ifmedia_init(ifm, 0, ifmedia_upd, ifmedia_sts);
494 1.2.4.4 yamt ifmedia_add(ifm, IFM_ETHER|IFM_10_T, 0, NULL);
495 1.2.4.4 yamt ifmedia_add(ifm, IFM_ETHER|IFM_10_T|IFM_FDX, 0, NULL);
496 1.2.4.4 yamt ifmedia_add(ifm, IFM_ETHER|IFM_100_TX, 0, NULL);
497 1.2.4.4 yamt ifmedia_add(ifm, IFM_ETHER|IFM_100_TX|IFM_FDX, 0, NULL);
498 1.2.4.4 yamt ifmedia_add(ifm, IFM_ETHER|IFM_AUTO, 0, NULL);
499 1.2.4.4 yamt ifmedia_set(ifm, IFM_ETHER|IFM_AUTO);
500 1.2.4.4 yamt }
501 1.2.4.4 yamt else {
502 1.2.4.4 yamt ifmedia_init(ifm, 0, ifmedia2_upd, ifmedia2_sts);
503 1.2.4.4 yamt ifmedia_add(ifm, IFM_ETHER|IFM_AUTO, 0, NULL);
504 1.2.4.4 yamt ifmedia_set(ifm, IFM_ETHER|IFM_AUTO);
505 1.2.4.4 yamt }
506 1.2.4.2 yamt
507 1.2.4.2 yamt printf("%s: 10baseT, 10baseT-FDX, 100baseTX, 100baseTX-FDX, auto\n",
508 1.2.4.2 yamt sc->sc_dev.dv_xname);
509 1.2.4.2 yamt
510 1.2.4.2 yamt ifp = &sc->sc_ethercom.ec_if;
511 1.2.4.2 yamt strcpy(ifp->if_xname, sc->sc_dev.dv_xname);
512 1.2.4.2 yamt ifp->if_softc = sc;
513 1.2.4.2 yamt ifp->if_flags = IFF_BROADCAST | IFF_SIMPLEX | IFF_MULTICAST;
514 1.2.4.2 yamt ifp->if_ioctl = kse_ioctl;
515 1.2.4.2 yamt ifp->if_start = kse_start;
516 1.2.4.2 yamt ifp->if_watchdog = kse_watchdog;
517 1.2.4.2 yamt ifp->if_init = kse_init;
518 1.2.4.2 yamt ifp->if_stop = kse_stop;
519 1.2.4.2 yamt IFQ_SET_READY(&ifp->if_snd);
520 1.2.4.2 yamt
521 1.2.4.2 yamt /*
522 1.2.4.2 yamt * KSZ8842 can handle 802.1Q VLAN-sized frames,
523 1.2.4.2 yamt * can do IPv4, TCPv4, and UDPv4 checksums in hardware.
524 1.2.4.2 yamt */
525 1.2.4.2 yamt sc->sc_ethercom.ec_capabilities |= ETHERCAP_VLAN_MTU;
526 1.2.4.2 yamt ifp->if_capabilities |=
527 1.2.4.2 yamt IFCAP_CSUM_IPv4_Tx | IFCAP_CSUM_IPv4_Rx |
528 1.2.4.2 yamt IFCAP_CSUM_TCPv4_Tx | IFCAP_CSUM_TCPv4_Rx |
529 1.2.4.2 yamt IFCAP_CSUM_UDPv4_Tx | IFCAP_CSUM_UDPv4_Rx;
530 1.2.4.2 yamt
531 1.2.4.2 yamt if_attach(ifp);
532 1.2.4.2 yamt ether_ifattach(ifp, enaddr);
533 1.2.4.4 yamt
534 1.2.4.4 yamt p = (sc->sc_chip == 0x8842) ? 3 : 1;
535 1.2.4.4 yamt #ifdef KSE_EVENT_COUNTERS
536 1.2.4.4 yamt for (i = 0; i < p; i++) {
537 1.2.4.4 yamt struct ksext *ee = &sc->sc_ext;
538 1.2.4.4 yamt sprintf(ee->evcntname[i], "%s.%d", sc->sc_dev.dv_xname, i+1);
539 1.2.4.4 yamt evcnt_attach_dynamic(&ee->pev[i][0], EVCNT_TYPE_MISC,
540 1.2.4.4 yamt NULL, ee->evcntname[i], "RxLoPriotyByte");
541 1.2.4.4 yamt evcnt_attach_dynamic(&ee->pev[i][1], EVCNT_TYPE_MISC,
542 1.2.4.4 yamt NULL, ee->evcntname[i], "RxHiPriotyByte");
543 1.2.4.4 yamt evcnt_attach_dynamic(&ee->pev[i][2], EVCNT_TYPE_MISC,
544 1.2.4.4 yamt NULL, ee->evcntname[i], "RxUndersizePkt");
545 1.2.4.4 yamt evcnt_attach_dynamic(&ee->pev[i][3], EVCNT_TYPE_MISC,
546 1.2.4.4 yamt NULL, ee->evcntname[i], "RxFragments");
547 1.2.4.4 yamt evcnt_attach_dynamic(&ee->pev[i][4], EVCNT_TYPE_MISC,
548 1.2.4.4 yamt NULL, ee->evcntname[i], "RxOversize");
549 1.2.4.4 yamt evcnt_attach_dynamic(&ee->pev[i][5], EVCNT_TYPE_MISC,
550 1.2.4.4 yamt NULL, ee->evcntname[i], "RxJabbers");
551 1.2.4.4 yamt evcnt_attach_dynamic(&ee->pev[i][6], EVCNT_TYPE_MISC,
552 1.2.4.4 yamt NULL, ee->evcntname[i], "RxSymbolError");
553 1.2.4.4 yamt evcnt_attach_dynamic(&ee->pev[i][7], EVCNT_TYPE_MISC,
554 1.2.4.4 yamt NULL, ee->evcntname[i], "RxCRCError");
555 1.2.4.4 yamt evcnt_attach_dynamic(&ee->pev[i][8], EVCNT_TYPE_MISC,
556 1.2.4.4 yamt NULL, ee->evcntname[i], "RxAlignmentError");
557 1.2.4.4 yamt evcnt_attach_dynamic(&ee->pev[i][9], EVCNT_TYPE_MISC,
558 1.2.4.4 yamt NULL, ee->evcntname[i], "RxControl8808Pkts");
559 1.2.4.4 yamt evcnt_attach_dynamic(&ee->pev[i][10], EVCNT_TYPE_MISC,
560 1.2.4.4 yamt NULL, ee->evcntname[i], "RxPausePkts");
561 1.2.4.4 yamt evcnt_attach_dynamic(&ee->pev[i][11], EVCNT_TYPE_MISC,
562 1.2.4.4 yamt NULL, ee->evcntname[i], "RxBroadcast");
563 1.2.4.4 yamt evcnt_attach_dynamic(&ee->pev[i][12], EVCNT_TYPE_MISC,
564 1.2.4.4 yamt NULL, ee->evcntname[i], "RxMulticast");
565 1.2.4.4 yamt evcnt_attach_dynamic(&ee->pev[i][13], EVCNT_TYPE_MISC,
566 1.2.4.4 yamt NULL, ee->evcntname[i], "RxUnicast");
567 1.2.4.4 yamt evcnt_attach_dynamic(&ee->pev[i][14], EVCNT_TYPE_MISC,
568 1.2.4.4 yamt NULL, ee->evcntname[i], "Rx64Octets");
569 1.2.4.4 yamt evcnt_attach_dynamic(&ee->pev[i][15], EVCNT_TYPE_MISC,
570 1.2.4.4 yamt NULL, ee->evcntname[i], "Rx65To127Octets");
571 1.2.4.4 yamt evcnt_attach_dynamic(&ee->pev[i][16], EVCNT_TYPE_MISC,
572 1.2.4.4 yamt NULL, ee->evcntname[i], "Rx128To255Octets");
573 1.2.4.4 yamt evcnt_attach_dynamic(&ee->pev[i][17], EVCNT_TYPE_MISC,
574 1.2.4.4 yamt NULL, ee->evcntname[i], "Rx255To511Octets");
575 1.2.4.4 yamt evcnt_attach_dynamic(&ee->pev[i][18], EVCNT_TYPE_MISC,
576 1.2.4.4 yamt NULL, ee->evcntname[i], "Rx512To1023Octets");
577 1.2.4.4 yamt evcnt_attach_dynamic(&ee->pev[i][19], EVCNT_TYPE_MISC,
578 1.2.4.4 yamt NULL, ee->evcntname[i], "Rx1024To1522Octets");
579 1.2.4.4 yamt evcnt_attach_dynamic(&ee->pev[i][20], EVCNT_TYPE_MISC,
580 1.2.4.4 yamt NULL, ee->evcntname[i], "TxLoPriotyByte");
581 1.2.4.4 yamt evcnt_attach_dynamic(&ee->pev[i][21], EVCNT_TYPE_MISC,
582 1.2.4.4 yamt NULL, ee->evcntname[i], "TxHiPriotyByte");
583 1.2.4.4 yamt evcnt_attach_dynamic(&ee->pev[i][22], EVCNT_TYPE_MISC,
584 1.2.4.4 yamt NULL, ee->evcntname[i], "TxLateCollision");
585 1.2.4.4 yamt evcnt_attach_dynamic(&ee->pev[i][23], EVCNT_TYPE_MISC,
586 1.2.4.4 yamt NULL, ee->evcntname[i], "TxPausePkts");
587 1.2.4.4 yamt evcnt_attach_dynamic(&ee->pev[i][24], EVCNT_TYPE_MISC,
588 1.2.4.4 yamt NULL, ee->evcntname[i], "TxBroadcastPkts");
589 1.2.4.4 yamt evcnt_attach_dynamic(&ee->pev[i][25], EVCNT_TYPE_MISC,
590 1.2.4.4 yamt NULL, ee->evcntname[i], "TxMulticastPkts");
591 1.2.4.4 yamt evcnt_attach_dynamic(&ee->pev[i][26], EVCNT_TYPE_MISC,
592 1.2.4.4 yamt NULL, ee->evcntname[i], "TxUnicastPkts");
593 1.2.4.4 yamt evcnt_attach_dynamic(&ee->pev[i][27], EVCNT_TYPE_MISC,
594 1.2.4.4 yamt NULL, ee->evcntname[i], "TxDeferred");
595 1.2.4.4 yamt evcnt_attach_dynamic(&ee->pev[i][28], EVCNT_TYPE_MISC,
596 1.2.4.4 yamt NULL, ee->evcntname[i], "TxTotalCollision");
597 1.2.4.4 yamt evcnt_attach_dynamic(&ee->pev[i][29], EVCNT_TYPE_MISC,
598 1.2.4.4 yamt NULL, ee->evcntname[i], "TxExcessiveCollision");
599 1.2.4.4 yamt evcnt_attach_dynamic(&ee->pev[i][30], EVCNT_TYPE_MISC,
600 1.2.4.4 yamt NULL, ee->evcntname[i], "TxSingleCollision");
601 1.2.4.4 yamt evcnt_attach_dynamic(&ee->pev[i][31], EVCNT_TYPE_MISC,
602 1.2.4.4 yamt NULL, ee->evcntname[i], "TxMultipleCollision");
603 1.2.4.4 yamt evcnt_attach_dynamic(&ee->pev[i][32], EVCNT_TYPE_MISC,
604 1.2.4.4 yamt NULL, ee->evcntname[i], "TxDropPkts");
605 1.2.4.4 yamt evcnt_attach_dynamic(&ee->pev[i][33], EVCNT_TYPE_MISC,
606 1.2.4.4 yamt NULL, ee->evcntname[i], "RxDropPkts");
607 1.2.4.4 yamt }
608 1.2.4.4 yamt #endif
609 1.2.4.2 yamt return;
610 1.2.4.2 yamt
611 1.2.4.2 yamt fail_5:
612 1.2.4.2 yamt for (i = 0; i < KSE_NRXDESC; i++) {
613 1.2.4.2 yamt if (sc->sc_rxsoft[i].rxs_dmamap != NULL)
614 1.2.4.2 yamt bus_dmamap_destroy(sc->sc_dmat,
615 1.2.4.2 yamt sc->sc_rxsoft[i].rxs_dmamap);
616 1.2.4.2 yamt }
617 1.2.4.2 yamt fail_4:
618 1.2.4.2 yamt for (i = 0; i < KSE_TXQUEUELEN; i++) {
619 1.2.4.2 yamt if (sc->sc_txsoft[i].txs_dmamap != NULL)
620 1.2.4.2 yamt bus_dmamap_destroy(sc->sc_dmat,
621 1.2.4.2 yamt sc->sc_txsoft[i].txs_dmamap);
622 1.2.4.2 yamt }
623 1.2.4.2 yamt bus_dmamap_unload(sc->sc_dmat, sc->sc_cddmamap);
624 1.2.4.2 yamt fail_3:
625 1.2.4.2 yamt bus_dmamap_destroy(sc->sc_dmat, sc->sc_cddmamap);
626 1.2.4.2 yamt fail_2:
627 1.2.4.3 yamt bus_dmamem_unmap(sc->sc_dmat, (void *)sc->sc_control_data,
628 1.2.4.2 yamt sizeof(struct kse_control_data));
629 1.2.4.2 yamt fail_1:
630 1.2.4.2 yamt bus_dmamem_free(sc->sc_dmat, &seg, nseg);
631 1.2.4.2 yamt fail_0:
632 1.2.4.2 yamt return;
633 1.2.4.2 yamt }
634 1.2.4.2 yamt
635 1.2.4.2 yamt static int
636 1.2.4.3 yamt kse_ioctl(struct ifnet *ifp, u_long cmd, void *data)
637 1.2.4.2 yamt {
638 1.2.4.2 yamt struct kse_softc *sc = ifp->if_softc;
639 1.2.4.2 yamt struct ifreq *ifr = (struct ifreq *)data;
640 1.2.4.2 yamt int s, error;
641 1.2.4.2 yamt
642 1.2.4.2 yamt s = splnet();
643 1.2.4.2 yamt
644 1.2.4.2 yamt switch (cmd) {
645 1.2.4.2 yamt case SIOCSIFMEDIA:
646 1.2.4.2 yamt case SIOCGIFMEDIA:
647 1.2.4.2 yamt error = ifmedia_ioctl(ifp, ifr, &sc->sc_media, cmd);
648 1.2.4.2 yamt break;
649 1.2.4.2 yamt
650 1.2.4.2 yamt default:
651 1.2.4.6 yamt if ((error = ether_ioctl(ifp, cmd, data)) != ENETRESET)
652 1.2.4.6 yamt break;
653 1.2.4.6 yamt
654 1.2.4.6 yamt error = 0;
655 1.2.4.6 yamt
656 1.2.4.6 yamt if (cmd == SIOCSIFCAP)
657 1.2.4.6 yamt error = (*ifp->if_init)(ifp);
658 1.2.4.6 yamt if (cmd != SIOCADDMULTI && cmd != SIOCDELMULTI)
659 1.2.4.6 yamt ;
660 1.2.4.6 yamt else if (ifp->if_flags & IFF_RUNNING) {
661 1.2.4.2 yamt /*
662 1.2.4.2 yamt * Multicast list has changed; set the hardware filter
663 1.2.4.2 yamt * accordingly.
664 1.2.4.2 yamt */
665 1.2.4.6 yamt kse_set_filter(sc);
666 1.2.4.2 yamt }
667 1.2.4.2 yamt break;
668 1.2.4.2 yamt }
669 1.2.4.2 yamt
670 1.2.4.2 yamt kse_start(ifp);
671 1.2.4.2 yamt
672 1.2.4.2 yamt splx(s);
673 1.2.4.2 yamt return error;
674 1.2.4.2 yamt }
675 1.2.4.2 yamt
676 1.2.4.2 yamt static int
677 1.2.4.2 yamt kse_init(struct ifnet *ifp)
678 1.2.4.2 yamt {
679 1.2.4.2 yamt struct kse_softc *sc = ifp->if_softc;
680 1.2.4.2 yamt uint32_t paddr;
681 1.2.4.2 yamt int i, error = 0;
682 1.2.4.2 yamt
683 1.2.4.2 yamt /* cancel pending I/O */
684 1.2.4.2 yamt kse_stop(ifp, 0);
685 1.2.4.2 yamt
686 1.2.4.2 yamt /* reset all registers but PCI configuration */
687 1.2.4.2 yamt kse_reset(sc);
688 1.2.4.2 yamt
689 1.2.4.2 yamt /* craft Tx descriptor ring */
690 1.2.4.2 yamt memset(sc->sc_txdescs, 0, sizeof(sc->sc_txdescs));
691 1.2.4.2 yamt for (i = 0, paddr = KSE_CDTXADDR(sc, 1); i < KSE_NTXDESC - 1; i++) {
692 1.2.4.2 yamt sc->sc_txdescs[i].t3 = paddr;
693 1.2.4.2 yamt paddr += sizeof(struct tdes);
694 1.2.4.2 yamt }
695 1.2.4.2 yamt sc->sc_txdescs[KSE_NTXDESC - 1].t3 = KSE_CDTXADDR(sc, 0);
696 1.2.4.2 yamt KSE_CDTXSYNC(sc, 0, KSE_NTXDESC,
697 1.2.4.2 yamt BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
698 1.2.4.2 yamt sc->sc_txfree = KSE_NTXDESC;
699 1.2.4.2 yamt sc->sc_txnext = 0;
700 1.2.4.2 yamt
701 1.2.4.2 yamt for (i = 0; i < KSE_TXQUEUELEN; i++)
702 1.2.4.2 yamt sc->sc_txsoft[i].txs_mbuf = NULL;
703 1.2.4.2 yamt sc->sc_txsfree = KSE_TXQUEUELEN;
704 1.2.4.2 yamt sc->sc_txsnext = 0;
705 1.2.4.2 yamt sc->sc_txsdirty = 0;
706 1.2.4.2 yamt
707 1.2.4.2 yamt /* craft Rx descriptor ring */
708 1.2.4.2 yamt memset(sc->sc_rxdescs, 0, sizeof(sc->sc_rxdescs));
709 1.2.4.2 yamt for (i = 0, paddr = KSE_CDRXADDR(sc, 1); i < KSE_NRXDESC - 1; i++) {
710 1.2.4.2 yamt sc->sc_rxdescs[i].r3 = paddr;
711 1.2.4.2 yamt paddr += sizeof(struct rdes);
712 1.2.4.2 yamt }
713 1.2.4.2 yamt sc->sc_rxdescs[KSE_NRXDESC - 1].r3 = KSE_CDRXADDR(sc, 0);
714 1.2.4.2 yamt for (i = 0; i < KSE_NRXDESC; i++) {
715 1.2.4.2 yamt if (sc->sc_rxsoft[i].rxs_mbuf == NULL) {
716 1.2.4.2 yamt if ((error = add_rxbuf(sc, i)) != 0) {
717 1.2.4.2 yamt printf("%s: unable to allocate or map rx "
718 1.2.4.2 yamt "buffer %d, error = %d\n",
719 1.2.4.2 yamt sc->sc_dev.dv_xname, i, error);
720 1.2.4.2 yamt rxdrain(sc);
721 1.2.4.2 yamt goto out;
722 1.2.4.2 yamt }
723 1.2.4.2 yamt }
724 1.2.4.2 yamt else
725 1.2.4.2 yamt KSE_INIT_RXDESC(sc, i);
726 1.2.4.2 yamt }
727 1.2.4.2 yamt sc->sc_rxptr = 0;
728 1.2.4.2 yamt
729 1.2.4.2 yamt /* hand Tx/Rx rings to HW */
730 1.2.4.2 yamt CSR_WRITE_4(sc, TDLB, KSE_CDTXADDR(sc, 0));
731 1.2.4.2 yamt CSR_WRITE_4(sc, RDLB, KSE_CDRXADDR(sc, 0));
732 1.2.4.2 yamt
733 1.2.4.2 yamt sc->sc_txc = TXC_TEN | TXC_EP | TXC_AC | TXC_FCE;
734 1.2.4.2 yamt sc->sc_rxc = RXC_REN | RXC_RU | RXC_FCE;
735 1.2.4.2 yamt if (ifp->if_flags & IFF_PROMISC)
736 1.2.4.2 yamt sc->sc_rxc |= RXC_RA;
737 1.2.4.2 yamt if (ifp->if_flags & IFF_BROADCAST)
738 1.2.4.2 yamt sc->sc_rxc |= RXC_RB;
739 1.2.4.2 yamt sc->sc_t1csum = sc->sc_mcsum = 0;
740 1.2.4.2 yamt if (ifp->if_capenable & IFCAP_CSUM_IPv4_Rx) {
741 1.2.4.4 yamt sc->sc_rxc |= RXC_ICC;
742 1.2.4.2 yamt sc->sc_mcsum |= M_CSUM_IPv4;
743 1.2.4.2 yamt }
744 1.2.4.2 yamt if (ifp->if_capenable & IFCAP_CSUM_IPv4_Tx) {
745 1.2.4.2 yamt sc->sc_txc |= TXC_ICG;
746 1.2.4.2 yamt sc->sc_t1csum |= T1_IPCKG;
747 1.2.4.2 yamt }
748 1.2.4.2 yamt if (ifp->if_capenable & IFCAP_CSUM_TCPv4_Rx) {
749 1.2.4.4 yamt sc->sc_rxc |= RXC_TCC;
750 1.2.4.2 yamt sc->sc_mcsum |= M_CSUM_TCPv4;
751 1.2.4.2 yamt }
752 1.2.4.2 yamt if (ifp->if_capenable & IFCAP_CSUM_TCPv4_Tx) {
753 1.2.4.2 yamt sc->sc_txc |= TXC_TCG;
754 1.2.4.2 yamt sc->sc_t1csum |= T1_TCPCKG;
755 1.2.4.2 yamt }
756 1.2.4.2 yamt if (ifp->if_capenable & IFCAP_CSUM_UDPv4_Rx) {
757 1.2.4.4 yamt sc->sc_rxc |= RXC_UCC;
758 1.2.4.2 yamt sc->sc_mcsum |= M_CSUM_UDPv4;
759 1.2.4.2 yamt }
760 1.2.4.2 yamt if (ifp->if_capenable & IFCAP_CSUM_UDPv4_Tx) {
761 1.2.4.2 yamt sc->sc_txc |= TXC_UCG;
762 1.2.4.2 yamt sc->sc_t1csum |= T1_UDPCKG;
763 1.2.4.2 yamt }
764 1.2.4.2 yamt sc->sc_txc |= (kse_burstsize << TXC_BS_SFT);
765 1.2.4.2 yamt sc->sc_rxc |= (kse_burstsize << RXC_BS_SFT);
766 1.2.4.2 yamt
767 1.2.4.4 yamt /* build multicast hash filter if necessary */
768 1.2.4.4 yamt kse_set_filter(sc);
769 1.2.4.4 yamt
770 1.2.4.2 yamt /* set current media */
771 1.2.4.2 yamt (void)ifmedia_upd(ifp);
772 1.2.4.2 yamt
773 1.2.4.2 yamt /* enable transmitter and receiver */
774 1.2.4.2 yamt CSR_WRITE_4(sc, MDTXC, sc->sc_txc);
775 1.2.4.2 yamt CSR_WRITE_4(sc, MDRXC, sc->sc_rxc);
776 1.2.4.2 yamt CSR_WRITE_4(sc, MDRSC, 1);
777 1.2.4.2 yamt
778 1.2.4.2 yamt /* enable interrupts */
779 1.2.4.4 yamt sc->sc_inten = INT_DMTS|INT_DMRS|INT_DMRBUS;
780 1.2.4.4 yamt if (sc->sc_chip == 0x8841)
781 1.2.4.4 yamt sc->sc_inten |= INT_DMLCS;
782 1.2.4.2 yamt CSR_WRITE_4(sc, INTST, ~0);
783 1.2.4.4 yamt CSR_WRITE_4(sc, INTEN, sc->sc_inten);
784 1.2.4.2 yamt
785 1.2.4.2 yamt ifp->if_flags |= IFF_RUNNING;
786 1.2.4.2 yamt ifp->if_flags &= ~IFF_OACTIVE;
787 1.2.4.2 yamt
788 1.2.4.4 yamt if (sc->sc_chip == 0x8841) {
789 1.2.4.4 yamt /* start one second timer */
790 1.2.4.4 yamt callout_reset(&sc->sc_callout, hz, phy_tick, sc);
791 1.2.4.4 yamt }
792 1.2.4.4 yamt #ifdef KSE_EVENT_COUNTERS
793 1.2.4.4 yamt /* start statistics gather 1 minute timer */
794 1.2.4.4 yamt zerostats(sc);
795 1.2.4.4 yamt callout_reset(&sc->sc_stat_ch, hz * 60, stat_tick, sc);
796 1.2.4.4 yamt #endif
797 1.2.4.2 yamt
798 1.2.4.2 yamt out:
799 1.2.4.2 yamt if (error) {
800 1.2.4.2 yamt ifp->if_flags &= ~(IFF_RUNNING | IFF_OACTIVE);
801 1.2.4.2 yamt ifp->if_timer = 0;
802 1.2.4.2 yamt printf("%s: interface not running\n", sc->sc_dev.dv_xname);
803 1.2.4.2 yamt }
804 1.2.4.2 yamt return error;
805 1.2.4.2 yamt }
806 1.2.4.2 yamt
807 1.2.4.2 yamt static void
808 1.2.4.2 yamt kse_stop(struct ifnet *ifp, int disable)
809 1.2.4.2 yamt {
810 1.2.4.2 yamt struct kse_softc *sc = ifp->if_softc;
811 1.2.4.2 yamt struct kse_txsoft *txs;
812 1.2.4.2 yamt int i;
813 1.2.4.2 yamt
814 1.2.4.4 yamt if (sc->sc_chip == 0x8841)
815 1.2.4.4 yamt callout_stop(&sc->sc_callout);
816 1.2.4.4 yamt callout_stop(&sc->sc_stat_ch);
817 1.2.4.2 yamt
818 1.2.4.2 yamt sc->sc_txc &= ~TXC_TEN;
819 1.2.4.2 yamt sc->sc_rxc &= ~RXC_REN;
820 1.2.4.2 yamt CSR_WRITE_4(sc, MDTXC, sc->sc_txc);
821 1.2.4.2 yamt CSR_WRITE_4(sc, MDRXC, sc->sc_rxc);
822 1.2.4.2 yamt
823 1.2.4.2 yamt for (i = 0; i < KSE_TXQUEUELEN; i++) {
824 1.2.4.2 yamt txs = &sc->sc_txsoft[i];
825 1.2.4.2 yamt if (txs->txs_mbuf != NULL) {
826 1.2.4.2 yamt bus_dmamap_unload(sc->sc_dmat, txs->txs_dmamap);
827 1.2.4.2 yamt m_freem(txs->txs_mbuf);
828 1.2.4.2 yamt txs->txs_mbuf = NULL;
829 1.2.4.2 yamt }
830 1.2.4.2 yamt }
831 1.2.4.2 yamt
832 1.2.4.2 yamt if (disable)
833 1.2.4.2 yamt rxdrain(sc);
834 1.2.4.2 yamt
835 1.2.4.2 yamt ifp->if_flags &= ~(IFF_RUNNING | IFF_OACTIVE);
836 1.2.4.2 yamt ifp->if_timer = 0;
837 1.2.4.2 yamt }
838 1.2.4.2 yamt
839 1.2.4.2 yamt static void
840 1.2.4.2 yamt kse_reset(struct kse_softc *sc)
841 1.2.4.2 yamt {
842 1.2.4.2 yamt
843 1.2.4.2 yamt CSR_WRITE_2(sc, GRR, 1);
844 1.2.4.2 yamt delay(1000); /* PDF does not mention the delay amount */
845 1.2.4.2 yamt CSR_WRITE_2(sc, GRR, 0);
846 1.2.4.2 yamt
847 1.2.4.2 yamt CSR_WRITE_2(sc, CIDR, 1);
848 1.2.4.2 yamt }
849 1.2.4.2 yamt
850 1.2.4.2 yamt static void
851 1.2.4.2 yamt kse_watchdog(struct ifnet *ifp)
852 1.2.4.2 yamt {
853 1.2.4.2 yamt struct kse_softc *sc = ifp->if_softc;
854 1.2.4.2 yamt
855 1.2.4.2 yamt /*
856 1.2.4.2 yamt * Since we're not interrupting every packet, sweep
857 1.2.4.2 yamt * up before we report an error.
858 1.2.4.2 yamt */
859 1.2.4.2 yamt txreap(sc);
860 1.2.4.2 yamt
861 1.2.4.2 yamt if (sc->sc_txfree != KSE_NTXDESC) {
862 1.2.4.2 yamt printf("%s: device timeout (txfree %d txsfree %d txnext %d)\n",
863 1.2.4.2 yamt sc->sc_dev.dv_xname, sc->sc_txfree, sc->sc_txsfree,
864 1.2.4.2 yamt sc->sc_txnext);
865 1.2.4.2 yamt ifp->if_oerrors++;
866 1.2.4.2 yamt
867 1.2.4.2 yamt /* Reset the interface. */
868 1.2.4.2 yamt kse_init(ifp);
869 1.2.4.2 yamt }
870 1.2.4.2 yamt else if (ifp->if_flags & IFF_DEBUG)
871 1.2.4.2 yamt printf("%s: recovered from device timeout\n",
872 1.2.4.2 yamt sc->sc_dev.dv_xname);
873 1.2.4.2 yamt
874 1.2.4.2 yamt /* Try to get more packets going. */
875 1.2.4.2 yamt kse_start(ifp);
876 1.2.4.2 yamt }
877 1.2.4.2 yamt
878 1.2.4.2 yamt static void
879 1.2.4.2 yamt kse_start(struct ifnet *ifp)
880 1.2.4.2 yamt {
881 1.2.4.2 yamt struct kse_softc *sc = ifp->if_softc;
882 1.2.4.4 yamt struct mbuf *m0, *m;
883 1.2.4.2 yamt struct kse_txsoft *txs;
884 1.2.4.2 yamt bus_dmamap_t dmamap;
885 1.2.4.2 yamt int error, nexttx, lasttx, ofree, seg;
886 1.2.4.4 yamt uint32_t tdes0;
887 1.2.4.2 yamt
888 1.2.4.2 yamt if ((ifp->if_flags & (IFF_RUNNING|IFF_OACTIVE)) != IFF_RUNNING)
889 1.2.4.2 yamt return;
890 1.2.4.2 yamt
891 1.2.4.2 yamt /*
892 1.2.4.2 yamt * Remember the previous number of free descriptors.
893 1.2.4.2 yamt */
894 1.2.4.2 yamt ofree = sc->sc_txfree;
895 1.2.4.2 yamt
896 1.2.4.2 yamt /*
897 1.2.4.2 yamt * Loop through the send queue, setting up transmit descriptors
898 1.2.4.2 yamt * until we drain the queue, or use up all available transmit
899 1.2.4.2 yamt * descriptors.
900 1.2.4.2 yamt */
901 1.2.4.2 yamt for (;;) {
902 1.2.4.2 yamt IFQ_POLL(&ifp->if_snd, m0);
903 1.2.4.2 yamt if (m0 == NULL)
904 1.2.4.2 yamt break;
905 1.2.4.2 yamt
906 1.2.4.2 yamt if (sc->sc_txsfree < KSE_TXQUEUE_GC) {
907 1.2.4.2 yamt txreap(sc);
908 1.2.4.2 yamt if (sc->sc_txsfree == 0)
909 1.2.4.2 yamt break;
910 1.2.4.2 yamt }
911 1.2.4.2 yamt txs = &sc->sc_txsoft[sc->sc_txsnext];
912 1.2.4.2 yamt dmamap = txs->txs_dmamap;
913 1.2.4.2 yamt
914 1.2.4.2 yamt error = bus_dmamap_load_mbuf(sc->sc_dmat, dmamap, m0,
915 1.2.4.2 yamt BUS_DMA_WRITE|BUS_DMA_NOWAIT);
916 1.2.4.2 yamt if (error) {
917 1.2.4.2 yamt if (error == EFBIG) {
918 1.2.4.2 yamt printf("%s: Tx packet consumes too many "
919 1.2.4.2 yamt "DMA segments, dropping...\n",
920 1.2.4.2 yamt sc->sc_dev.dv_xname);
921 1.2.4.2 yamt IFQ_DEQUEUE(&ifp->if_snd, m0);
922 1.2.4.2 yamt m_freem(m0);
923 1.2.4.2 yamt continue;
924 1.2.4.2 yamt }
925 1.2.4.2 yamt /* Short on resources, just stop for now. */
926 1.2.4.2 yamt break;
927 1.2.4.2 yamt }
928 1.2.4.2 yamt
929 1.2.4.2 yamt if (dmamap->dm_nsegs > sc->sc_txfree) {
930 1.2.4.2 yamt /*
931 1.2.4.2 yamt * Not enough free descriptors to transmit this
932 1.2.4.2 yamt * packet. We haven't committed anything yet,
933 1.2.4.2 yamt * so just unload the DMA map, put the packet
934 1.2.4.2 yamt * back on the queue, and punt. Notify the upper
935 1.2.4.2 yamt * layer that there are not more slots left.
936 1.2.4.2 yamt */
937 1.2.4.2 yamt ifp->if_flags |= IFF_OACTIVE;
938 1.2.4.2 yamt bus_dmamap_unload(sc->sc_dmat, dmamap);
939 1.2.4.2 yamt break;
940 1.2.4.2 yamt }
941 1.2.4.2 yamt
942 1.2.4.2 yamt IFQ_DEQUEUE(&ifp->if_snd, m0);
943 1.2.4.2 yamt
944 1.2.4.2 yamt /*
945 1.2.4.2 yamt * WE ARE NOW COMMITTED TO TRANSMITTING THE PACKET.
946 1.2.4.2 yamt */
947 1.2.4.2 yamt
948 1.2.4.2 yamt bus_dmamap_sync(sc->sc_dmat, dmamap, 0, dmamap->dm_mapsize,
949 1.2.4.2 yamt BUS_DMASYNC_PREWRITE);
950 1.2.4.2 yamt
951 1.2.4.4 yamt lasttx = -1; tdes0 = 0;
952 1.2.4.2 yamt for (nexttx = sc->sc_txnext, seg = 0;
953 1.2.4.2 yamt seg < dmamap->dm_nsegs;
954 1.2.4.2 yamt seg++, nexttx = KSE_NEXTTX(nexttx)) {
955 1.2.4.2 yamt struct tdes *tdes = &sc->sc_txdescs[nexttx];
956 1.2.4.2 yamt /*
957 1.2.4.2 yamt * If this is the first descriptor we're
958 1.2.4.2 yamt * enqueueing, don't set the OWN bit just
959 1.2.4.2 yamt * yet. That could cause a race condition.
960 1.2.4.2 yamt * We'll do it below.
961 1.2.4.2 yamt */
962 1.2.4.2 yamt tdes->t2 = dmamap->dm_segs[seg].ds_addr;
963 1.2.4.2 yamt tdes->t1 = sc->sc_t1csum
964 1.2.4.2 yamt | (dmamap->dm_segs[seg].ds_len & T1_TBS_MASK);
965 1.2.4.4 yamt tdes->t0 = tdes0;
966 1.2.4.4 yamt tdes0 |= T0_OWN;
967 1.2.4.2 yamt lasttx = nexttx;
968 1.2.4.2 yamt }
969 1.2.4.4 yamt
970 1.2.4.2 yamt /*
971 1.2.4.2 yamt * Outgoing NFS mbuf must be unloaded when Tx completed.
972 1.2.4.2 yamt * Without T1_IC NFS mbuf is left unack'ed for excessive
973 1.2.4.2 yamt * time and NFS stops to proceed until kse_watchdog()
974 1.2.4.2 yamt * calls txreap() to reclaim the unack'ed mbuf.
975 1.2.4.4 yamt * It's painful to traverse every mbuf chain to determine
976 1.2.4.2 yamt * whether someone is waiting for Tx completion.
977 1.2.4.2 yamt */
978 1.2.4.4 yamt m = m0;
979 1.2.4.2 yamt do {
980 1.2.4.2 yamt if ((m->m_flags & M_EXT) && m->m_ext.ext_free) {
981 1.2.4.2 yamt sc->sc_txdescs[lasttx].t1 |= T1_IC;
982 1.2.4.2 yamt break;
983 1.2.4.2 yamt }
984 1.2.4.2 yamt } while ((m = m->m_next) != NULL);
985 1.2.4.2 yamt
986 1.2.4.2 yamt /* write last T0_OWN bit of the 1st segment */
987 1.2.4.2 yamt sc->sc_txdescs[lasttx].t1 |= T1_LS;
988 1.2.4.2 yamt sc->sc_txdescs[sc->sc_txnext].t1 |= T1_FS;
989 1.2.4.2 yamt sc->sc_txdescs[sc->sc_txnext].t0 = T0_OWN;
990 1.2.4.2 yamt KSE_CDTXSYNC(sc, sc->sc_txnext, dmamap->dm_nsegs,
991 1.2.4.2 yamt BUS_DMASYNC_PREREAD|BUS_DMASYNC_PREWRITE);
992 1.2.4.2 yamt
993 1.2.4.2 yamt /* tell DMA start transmit */
994 1.2.4.2 yamt CSR_WRITE_4(sc, MDTSC, 1);
995 1.2.4.2 yamt
996 1.2.4.2 yamt txs->txs_mbuf = m0;
997 1.2.4.2 yamt txs->txs_firstdesc = sc->sc_txnext;
998 1.2.4.2 yamt txs->txs_lastdesc = lasttx;
999 1.2.4.2 yamt txs->txs_ndesc = dmamap->dm_nsegs;
1000 1.2.4.2 yamt
1001 1.2.4.2 yamt sc->sc_txfree -= txs->txs_ndesc;
1002 1.2.4.2 yamt sc->sc_txnext = nexttx;
1003 1.2.4.2 yamt sc->sc_txsfree--;
1004 1.2.4.2 yamt sc->sc_txsnext = KSE_NEXTTXS(sc->sc_txsnext);
1005 1.2.4.2 yamt #if NBPFILTER > 0
1006 1.2.4.2 yamt /*
1007 1.2.4.2 yamt * Pass the packet to any BPF listeners.
1008 1.2.4.2 yamt */
1009 1.2.4.2 yamt if (ifp->if_bpf)
1010 1.2.4.2 yamt bpf_mtap(ifp->if_bpf, m0);
1011 1.2.4.2 yamt #endif /* NBPFILTER > 0 */
1012 1.2.4.2 yamt }
1013 1.2.4.2 yamt
1014 1.2.4.2 yamt if (sc->sc_txsfree == 0 || sc->sc_txfree == 0) {
1015 1.2.4.2 yamt /* No more slots left; notify upper layer. */
1016 1.2.4.2 yamt ifp->if_flags |= IFF_OACTIVE;
1017 1.2.4.2 yamt }
1018 1.2.4.2 yamt if (sc->sc_txfree != ofree) {
1019 1.2.4.2 yamt /* Set a watchdog timer in case the chip flakes out. */
1020 1.2.4.2 yamt ifp->if_timer = 5;
1021 1.2.4.2 yamt }
1022 1.2.4.2 yamt }
1023 1.2.4.2 yamt
1024 1.2.4.2 yamt static void
1025 1.2.4.2 yamt kse_set_filter(struct kse_softc *sc)
1026 1.2.4.2 yamt {
1027 1.2.4.2 yamt struct ether_multistep step;
1028 1.2.4.2 yamt struct ether_multi *enm;
1029 1.2.4.2 yamt struct ifnet *ifp = &sc->sc_ethercom.ec_if;
1030 1.2.4.4 yamt uint32_t h, hashes[2];
1031 1.2.4.4 yamt
1032 1.2.4.4 yamt sc->sc_rxc &= ~(RXC_MHTE | RXC_RM);
1033 1.2.4.4 yamt ifp->if_flags &= ~IFF_ALLMULTI;
1034 1.2.4.4 yamt if (ifp->if_flags & IFF_PROMISC)
1035 1.2.4.4 yamt return;
1036 1.2.4.2 yamt
1037 1.2.4.2 yamt ETHER_FIRST_MULTI(step, &sc->sc_ethercom, enm);
1038 1.2.4.4 yamt if (enm == NULL)
1039 1.2.4.4 yamt return;
1040 1.2.4.4 yamt hashes[0] = hashes[1] = 0;
1041 1.2.4.4 yamt do {
1042 1.2.4.4 yamt if (memcmp(enm->enm_addrlo, enm->enm_addrhi, ETHER_ADDR_LEN)) {
1043 1.2.4.4 yamt /*
1044 1.2.4.4 yamt * We must listen to a range of multicast addresses.
1045 1.2.4.4 yamt * For now, just accept all multicasts, rather than
1046 1.2.4.4 yamt * trying to set only those filter bits needed to match
1047 1.2.4.4 yamt * the range. (At this time, the only use of address
1048 1.2.4.4 yamt * ranges is for IP multicast routing, for which the
1049 1.2.4.4 yamt * range is big enough to require all bits set.)
1050 1.2.4.4 yamt */
1051 1.2.4.4 yamt goto allmulti;
1052 1.2.4.2 yamt }
1053 1.2.4.4 yamt h = ether_crc32_le(enm->enm_addrlo, ETHER_ADDR_LEN) >> 26;
1054 1.2.4.4 yamt hashes[h >> 5] |= 1 << (h & 0x1f);
1055 1.2.4.2 yamt ETHER_NEXT_MULTI(step, enm);
1056 1.2.4.4 yamt } while (enm != NULL);
1057 1.2.4.4 yamt sc->sc_rxc |= RXC_MHTE;
1058 1.2.4.4 yamt CSR_WRITE_4(sc, MTR0, hashes[0]);
1059 1.2.4.4 yamt CSR_WRITE_4(sc, MTR1, hashes[1]);
1060 1.2.4.2 yamt return;
1061 1.2.4.4 yamt allmulti:
1062 1.2.4.4 yamt sc->sc_rxc |= RXC_RM;
1063 1.2.4.4 yamt ifp->if_flags |= IFF_ALLMULTI;
1064 1.2.4.2 yamt }
1065 1.2.4.2 yamt
1066 1.2.4.2 yamt static int
1067 1.2.4.2 yamt add_rxbuf(struct kse_softc *sc, int idx)
1068 1.2.4.2 yamt {
1069 1.2.4.2 yamt struct kse_rxsoft *rxs = &sc->sc_rxsoft[idx];
1070 1.2.4.2 yamt struct mbuf *m;
1071 1.2.4.2 yamt int error;
1072 1.2.4.2 yamt
1073 1.2.4.2 yamt MGETHDR(m, M_DONTWAIT, MT_DATA);
1074 1.2.4.2 yamt if (m == NULL)
1075 1.2.4.2 yamt return ENOBUFS;
1076 1.2.4.2 yamt
1077 1.2.4.2 yamt MCLGET(m, M_DONTWAIT);
1078 1.2.4.2 yamt if ((m->m_flags & M_EXT) == 0) {
1079 1.2.4.2 yamt m_freem(m);
1080 1.2.4.2 yamt return ENOBUFS;
1081 1.2.4.2 yamt }
1082 1.2.4.2 yamt
1083 1.2.4.2 yamt if (rxs->rxs_mbuf != NULL)
1084 1.2.4.2 yamt bus_dmamap_unload(sc->sc_dmat, rxs->rxs_dmamap);
1085 1.2.4.2 yamt
1086 1.2.4.2 yamt rxs->rxs_mbuf = m;
1087 1.2.4.2 yamt
1088 1.2.4.2 yamt error = bus_dmamap_load(sc->sc_dmat, rxs->rxs_dmamap,
1089 1.2.4.2 yamt m->m_ext.ext_buf, m->m_ext.ext_size, NULL, BUS_DMA_NOWAIT);
1090 1.2.4.2 yamt if (error) {
1091 1.2.4.2 yamt printf("%s: can't load rx DMA map %d, error = %d\n",
1092 1.2.4.2 yamt sc->sc_dev.dv_xname, idx, error);
1093 1.2.4.2 yamt panic("kse_add_rxbuf");
1094 1.2.4.2 yamt }
1095 1.2.4.2 yamt
1096 1.2.4.2 yamt bus_dmamap_sync(sc->sc_dmat, rxs->rxs_dmamap, 0,
1097 1.2.4.2 yamt rxs->rxs_dmamap->dm_mapsize, BUS_DMASYNC_PREREAD);
1098 1.2.4.2 yamt
1099 1.2.4.2 yamt KSE_INIT_RXDESC(sc, idx);
1100 1.2.4.2 yamt
1101 1.2.4.2 yamt return 0;
1102 1.2.4.2 yamt }
1103 1.2.4.2 yamt
1104 1.2.4.2 yamt static void
1105 1.2.4.2 yamt rxdrain(struct kse_softc *sc)
1106 1.2.4.2 yamt {
1107 1.2.4.2 yamt struct kse_rxsoft *rxs;
1108 1.2.4.2 yamt int i;
1109 1.2.4.2 yamt
1110 1.2.4.2 yamt for (i = 0; i < KSE_NRXDESC; i++) {
1111 1.2.4.2 yamt rxs = &sc->sc_rxsoft[i];
1112 1.2.4.2 yamt if (rxs->rxs_mbuf != NULL) {
1113 1.2.4.2 yamt bus_dmamap_unload(sc->sc_dmat, rxs->rxs_dmamap);
1114 1.2.4.2 yamt m_freem(rxs->rxs_mbuf);
1115 1.2.4.2 yamt rxs->rxs_mbuf = NULL;
1116 1.2.4.2 yamt }
1117 1.2.4.2 yamt }
1118 1.2.4.2 yamt }
1119 1.2.4.2 yamt
1120 1.2.4.2 yamt static int
1121 1.2.4.2 yamt kse_intr(void *arg)
1122 1.2.4.2 yamt {
1123 1.2.4.2 yamt struct kse_softc *sc = arg;
1124 1.2.4.2 yamt uint32_t isr;
1125 1.2.4.2 yamt
1126 1.2.4.2 yamt if ((isr = CSR_READ_4(sc, INTST)) == 0)
1127 1.2.4.2 yamt return 0;
1128 1.2.4.2 yamt
1129 1.2.4.2 yamt if (isr & INT_DMRS)
1130 1.2.4.2 yamt rxintr(sc);
1131 1.2.4.2 yamt if (isr & INT_DMTS)
1132 1.2.4.2 yamt txreap(sc);
1133 1.2.4.2 yamt if (isr & INT_DMLCS)
1134 1.2.4.2 yamt lnkchg(sc);
1135 1.2.4.2 yamt if (isr & INT_DMRBUS)
1136 1.2.4.2 yamt printf("%s: Rx descriptor full\n", sc->sc_dev.dv_xname);
1137 1.2.4.2 yamt
1138 1.2.4.2 yamt CSR_WRITE_4(sc, INTST, isr);
1139 1.2.4.2 yamt return 1;
1140 1.2.4.2 yamt }
1141 1.2.4.2 yamt
1142 1.2.4.2 yamt static void
1143 1.2.4.2 yamt rxintr(struct kse_softc *sc)
1144 1.2.4.2 yamt {
1145 1.2.4.2 yamt struct ifnet *ifp = &sc->sc_ethercom.ec_if;
1146 1.2.4.2 yamt struct kse_rxsoft *rxs;
1147 1.2.4.2 yamt struct mbuf *m;
1148 1.2.4.2 yamt uint32_t rxstat;
1149 1.2.4.2 yamt int i, len;
1150 1.2.4.2 yamt
1151 1.2.4.2 yamt for (i = sc->sc_rxptr; /*CONSTCOND*/ 1; i = KSE_NEXTRX(i)) {
1152 1.2.4.2 yamt rxs = &sc->sc_rxsoft[i];
1153 1.2.4.2 yamt
1154 1.2.4.2 yamt KSE_CDRXSYNC(sc, i,
1155 1.2.4.2 yamt BUS_DMASYNC_POSTREAD|BUS_DMASYNC_POSTWRITE);
1156 1.2.4.2 yamt
1157 1.2.4.2 yamt rxstat = sc->sc_rxdescs[i].r0;
1158 1.2.4.2 yamt
1159 1.2.4.2 yamt if (rxstat & R0_OWN) /* desc is left empty */
1160 1.2.4.2 yamt break;
1161 1.2.4.2 yamt
1162 1.2.4.2 yamt /* R0_FS|R0_LS must have been marked for this desc */
1163 1.2.4.2 yamt
1164 1.2.4.2 yamt if (rxstat & R0_ES) {
1165 1.2.4.2 yamt ifp->if_ierrors++;
1166 1.2.4.2 yamt #define PRINTERR(bit, str) \
1167 1.2.4.2 yamt if (rxstat & (bit)) \
1168 1.2.4.2 yamt printf("%s: receive error: %s\n", \
1169 1.2.4.2 yamt sc->sc_dev.dv_xname, str)
1170 1.2.4.2 yamt PRINTERR(R0_TL, "frame too long");
1171 1.2.4.2 yamt PRINTERR(R0_RF, "runt frame");
1172 1.2.4.2 yamt PRINTERR(R0_CE, "bad FCS");
1173 1.2.4.2 yamt #undef PRINTERR
1174 1.2.4.2 yamt KSE_INIT_RXDESC(sc, i);
1175 1.2.4.2 yamt continue;
1176 1.2.4.2 yamt }
1177 1.2.4.2 yamt
1178 1.2.4.2 yamt /* HW errata; frame might be too small or too large */
1179 1.2.4.2 yamt
1180 1.2.4.2 yamt bus_dmamap_sync(sc->sc_dmat, rxs->rxs_dmamap, 0,
1181 1.2.4.2 yamt rxs->rxs_dmamap->dm_mapsize, BUS_DMASYNC_POSTREAD);
1182 1.2.4.2 yamt
1183 1.2.4.2 yamt len = rxstat & R0_FL_MASK;
1184 1.2.4.2 yamt len -= ETHER_CRC_LEN; /* trim CRC off */
1185 1.2.4.2 yamt m = rxs->rxs_mbuf;
1186 1.2.4.2 yamt
1187 1.2.4.2 yamt if (add_rxbuf(sc, i) != 0) {
1188 1.2.4.2 yamt ifp->if_ierrors++;
1189 1.2.4.2 yamt KSE_INIT_RXDESC(sc, i);
1190 1.2.4.2 yamt bus_dmamap_sync(sc->sc_dmat,
1191 1.2.4.2 yamt rxs->rxs_dmamap, 0,
1192 1.2.4.2 yamt rxs->rxs_dmamap->dm_mapsize,
1193 1.2.4.2 yamt BUS_DMASYNC_PREREAD);
1194 1.2.4.2 yamt continue;
1195 1.2.4.2 yamt }
1196 1.2.4.2 yamt
1197 1.2.4.2 yamt ifp->if_ipackets++;
1198 1.2.4.2 yamt m->m_pkthdr.rcvif = ifp;
1199 1.2.4.2 yamt m->m_pkthdr.len = m->m_len = len;
1200 1.2.4.2 yamt
1201 1.2.4.2 yamt if (sc->sc_mcsum) {
1202 1.2.4.2 yamt m->m_pkthdr.csum_flags |= sc->sc_mcsum;
1203 1.2.4.2 yamt if (rxstat & R0_IPE)
1204 1.2.4.2 yamt m->m_pkthdr.csum_flags |= M_CSUM_IPv4_BAD;
1205 1.2.4.2 yamt if (rxstat & (R0_TCPE | R0_UDPE))
1206 1.2.4.2 yamt m->m_pkthdr.csum_flags |= M_CSUM_TCP_UDP_BAD;
1207 1.2.4.2 yamt }
1208 1.2.4.2 yamt #if NBPFILTER > 0
1209 1.2.4.2 yamt if (ifp->if_bpf)
1210 1.2.4.2 yamt bpf_mtap(ifp->if_bpf, m);
1211 1.2.4.2 yamt #endif /* NBPFILTER > 0 */
1212 1.2.4.2 yamt (*ifp->if_input)(ifp, m);
1213 1.2.4.2 yamt #ifdef KSEDIAGNOSTIC
1214 1.2.4.2 yamt if (kse_monitor_rxintr > 0) {
1215 1.2.4.2 yamt printf("m stat %x data %p len %d\n",
1216 1.2.4.2 yamt rxstat, m->m_data, m->m_len);
1217 1.2.4.2 yamt }
1218 1.2.4.2 yamt #endif
1219 1.2.4.2 yamt }
1220 1.2.4.2 yamt sc->sc_rxptr = i;
1221 1.2.4.2 yamt }
1222 1.2.4.2 yamt
1223 1.2.4.2 yamt static void
1224 1.2.4.2 yamt txreap(struct kse_softc *sc)
1225 1.2.4.2 yamt {
1226 1.2.4.2 yamt struct ifnet *ifp = &sc->sc_ethercom.ec_if;
1227 1.2.4.2 yamt struct kse_txsoft *txs;
1228 1.2.4.2 yamt uint32_t txstat;
1229 1.2.4.2 yamt int i;
1230 1.2.4.2 yamt
1231 1.2.4.2 yamt ifp->if_flags &= ~IFF_OACTIVE;
1232 1.2.4.2 yamt
1233 1.2.4.2 yamt for (i = sc->sc_txsdirty; sc->sc_txsfree != KSE_TXQUEUELEN;
1234 1.2.4.2 yamt i = KSE_NEXTTXS(i), sc->sc_txsfree++) {
1235 1.2.4.2 yamt txs = &sc->sc_txsoft[i];
1236 1.2.4.2 yamt
1237 1.2.4.2 yamt KSE_CDTXSYNC(sc, txs->txs_firstdesc, txs->txs_ndesc,
1238 1.2.4.2 yamt BUS_DMASYNC_POSTREAD|BUS_DMASYNC_POSTWRITE);
1239 1.2.4.2 yamt
1240 1.2.4.2 yamt txstat = sc->sc_txdescs[txs->txs_lastdesc].t0;
1241 1.2.4.2 yamt
1242 1.2.4.2 yamt if (txstat & T0_OWN) /* desc is still in use */
1243 1.2.4.2 yamt break;
1244 1.2.4.2 yamt
1245 1.2.4.2 yamt /* there is no way to tell transmission status per frame */
1246 1.2.4.2 yamt
1247 1.2.4.2 yamt ifp->if_opackets++;
1248 1.2.4.2 yamt
1249 1.2.4.2 yamt sc->sc_txfree += txs->txs_ndesc;
1250 1.2.4.2 yamt bus_dmamap_sync(sc->sc_dmat, txs->txs_dmamap,
1251 1.2.4.2 yamt 0, txs->txs_dmamap->dm_mapsize, BUS_DMASYNC_POSTWRITE);
1252 1.2.4.2 yamt bus_dmamap_unload(sc->sc_dmat, txs->txs_dmamap);
1253 1.2.4.2 yamt m_freem(txs->txs_mbuf);
1254 1.2.4.2 yamt txs->txs_mbuf = NULL;
1255 1.2.4.2 yamt }
1256 1.2.4.2 yamt sc->sc_txsdirty = i;
1257 1.2.4.2 yamt if (sc->sc_txsfree == KSE_TXQUEUELEN)
1258 1.2.4.2 yamt ifp->if_timer = 0;
1259 1.2.4.2 yamt }
1260 1.2.4.2 yamt
1261 1.2.4.2 yamt static void
1262 1.2.4.2 yamt lnkchg(struct kse_softc *sc)
1263 1.2.4.2 yamt {
1264 1.2.4.2 yamt struct ifmediareq ifmr;
1265 1.2.4.2 yamt
1266 1.2.4.2 yamt #if 0 /* rambling link status */
1267 1.2.4.2 yamt printf("%s: link %s\n", sc->sc_dev.dv_xname,
1268 1.2.4.2 yamt (CSR_READ_2(sc, P1SR) & (1U << 5)) ? "up" : "down");
1269 1.2.4.2 yamt #endif
1270 1.2.4.2 yamt ifmedia_sts(&sc->sc_ethercom.ec_if, &ifmr);
1271 1.2.4.2 yamt }
1272 1.2.4.2 yamt
1273 1.2.4.2 yamt static int
1274 1.2.4.2 yamt ifmedia_upd(struct ifnet *ifp)
1275 1.2.4.2 yamt {
1276 1.2.4.2 yamt struct kse_softc *sc = ifp->if_softc;
1277 1.2.4.2 yamt struct ifmedia *ifm = &sc->sc_media;
1278 1.2.4.2 yamt uint16_t ctl;
1279 1.2.4.2 yamt
1280 1.2.4.2 yamt ctl = 0;
1281 1.2.4.2 yamt if (IFM_SUBTYPE(ifm->ifm_media) == IFM_AUTO) {
1282 1.2.4.2 yamt ctl |= (1U << 13); /* restart AN */
1283 1.2.4.2 yamt ctl |= (1U << 7); /* enable AN */
1284 1.2.4.2 yamt ctl |= (1U << 4); /* advertise flow control pause */
1285 1.2.4.2 yamt ctl |= (1U << 3) | (1U << 2) | (1U << 1) | (1U << 0);
1286 1.2.4.2 yamt }
1287 1.2.4.2 yamt else {
1288 1.2.4.2 yamt if (IFM_SUBTYPE(ifm->ifm_media) == IFM_100_TX)
1289 1.2.4.2 yamt ctl |= (1U << 6);
1290 1.2.4.2 yamt if (ifm->ifm_media & IFM_FDX)
1291 1.2.4.2 yamt ctl |= (1U << 5);
1292 1.2.4.2 yamt }
1293 1.2.4.2 yamt CSR_WRITE_2(sc, P1CR4, ctl);
1294 1.2.4.2 yamt
1295 1.2.4.2 yamt sc->sc_media_active = IFM_NONE;
1296 1.2.4.2 yamt sc->sc_media_status = IFM_AVALID;
1297 1.2.4.2 yamt
1298 1.2.4.2 yamt return 0;
1299 1.2.4.2 yamt }
1300 1.2.4.2 yamt
1301 1.2.4.2 yamt static void
1302 1.2.4.2 yamt ifmedia_sts(struct ifnet *ifp, struct ifmediareq *ifmr)
1303 1.2.4.2 yamt {
1304 1.2.4.2 yamt struct kse_softc *sc = ifp->if_softc;
1305 1.2.4.2 yamt struct ifmedia *ifm = &sc->sc_media;
1306 1.2.4.2 yamt uint16_t ctl, sts, result;
1307 1.2.4.2 yamt
1308 1.2.4.2 yamt ifmr->ifm_status = IFM_AVALID;
1309 1.2.4.2 yamt ifmr->ifm_active = IFM_ETHER;
1310 1.2.4.2 yamt
1311 1.2.4.2 yamt ctl = CSR_READ_2(sc, P1CR4);
1312 1.2.4.2 yamt sts = CSR_READ_2(sc, P1SR);
1313 1.2.4.2 yamt if ((sts & (1U << 5)) == 0) {
1314 1.2.4.2 yamt ifmr->ifm_active |= IFM_NONE;
1315 1.2.4.2 yamt goto out; /* link is down */
1316 1.2.4.2 yamt }
1317 1.2.4.2 yamt ifmr->ifm_status |= IFM_ACTIVE;
1318 1.2.4.2 yamt if (IFM_SUBTYPE(ifm->ifm_media) == IFM_AUTO) {
1319 1.2.4.2 yamt if ((sts & (1U << 6)) == 0) {
1320 1.2.4.2 yamt ifmr->ifm_active |= IFM_NONE;
1321 1.2.4.2 yamt goto out; /* negotiation in progress */
1322 1.2.4.2 yamt }
1323 1.2.4.2 yamt result = ctl & sts & 017;
1324 1.2.4.2 yamt if (result & (1U << 3))
1325 1.2.4.2 yamt ifmr->ifm_active |= IFM_100_TX|IFM_FDX;
1326 1.2.4.2 yamt else if (result & (1U << 2))
1327 1.2.4.2 yamt ifmr->ifm_active |= IFM_100_TX;
1328 1.2.4.2 yamt else if (result & (1U << 1))
1329 1.2.4.2 yamt ifmr->ifm_active |= IFM_10_T|IFM_FDX;
1330 1.2.4.2 yamt else if (result & (1U << 0))
1331 1.2.4.2 yamt ifmr->ifm_active |= IFM_10_T;
1332 1.2.4.2 yamt else
1333 1.2.4.2 yamt ifmr->ifm_active |= IFM_NONE;
1334 1.2.4.2 yamt if (ctl & (1U << 4))
1335 1.2.4.2 yamt ifmr->ifm_active |= IFM_FLOW | IFM_ETH_RXPAUSE;
1336 1.2.4.2 yamt if (sts & (1U << 4))
1337 1.2.4.2 yamt ifmr->ifm_active |= IFM_FLOW | IFM_ETH_TXPAUSE;
1338 1.2.4.2 yamt }
1339 1.2.4.2 yamt else {
1340 1.2.4.2 yamt ifmr->ifm_active |= (sts & (1U << 10)) ? IFM_100_TX : IFM_10_T;
1341 1.2.4.2 yamt if (sts & (1U << 9))
1342 1.2.4.2 yamt ifmr->ifm_active |= IFM_FDX;
1343 1.2.4.2 yamt if (sts & (1U << 12))
1344 1.2.4.2 yamt ifmr->ifm_active |= IFM_FLOW | IFM_ETH_RXPAUSE;
1345 1.2.4.2 yamt if (sts & (1U << 11))
1346 1.2.4.2 yamt ifmr->ifm_active |= IFM_FLOW | IFM_ETH_TXPAUSE;
1347 1.2.4.2 yamt }
1348 1.2.4.2 yamt
1349 1.2.4.2 yamt out:
1350 1.2.4.2 yamt sc->sc_media_status = ifmr->ifm_status;
1351 1.2.4.2 yamt sc->sc_media_active = ifmr->ifm_active;
1352 1.2.4.2 yamt }
1353 1.2.4.2 yamt
1354 1.2.4.2 yamt static void
1355 1.2.4.2 yamt phy_tick(void *arg)
1356 1.2.4.2 yamt {
1357 1.2.4.2 yamt struct kse_softc *sc = arg;
1358 1.2.4.2 yamt struct ifmediareq ifmr;
1359 1.2.4.2 yamt int s;
1360 1.2.4.2 yamt
1361 1.2.4.2 yamt s = splnet();
1362 1.2.4.2 yamt ifmedia_sts(&sc->sc_ethercom.ec_if, &ifmr);
1363 1.2.4.2 yamt splx(s);
1364 1.2.4.2 yamt
1365 1.2.4.2 yamt callout_reset(&sc->sc_callout, hz, phy_tick, sc);
1366 1.2.4.2 yamt }
1367 1.2.4.4 yamt
1368 1.2.4.4 yamt static int
1369 1.2.4.4 yamt ifmedia2_upd(struct ifnet *ifp)
1370 1.2.4.4 yamt {
1371 1.2.4.4 yamt struct kse_softc *sc = ifp->if_softc;
1372 1.2.4.4 yamt
1373 1.2.4.4 yamt sc->sc_media_status = IFM_AVALID;
1374 1.2.4.4 yamt sc->sc_media_active = IFM_NONE;
1375 1.2.4.4 yamt return 0;
1376 1.2.4.4 yamt }
1377 1.2.4.4 yamt
1378 1.2.4.4 yamt static void
1379 1.2.4.4 yamt ifmedia2_sts(struct ifnet *ifp, struct ifmediareq *ifmr)
1380 1.2.4.4 yamt {
1381 1.2.4.4 yamt struct kse_softc *sc = ifp->if_softc;
1382 1.2.4.4 yamt int p1sts, p2sts;
1383 1.2.4.4 yamt
1384 1.2.4.4 yamt ifmr->ifm_status = IFM_AVALID;
1385 1.2.4.4 yamt ifmr->ifm_active = IFM_ETHER;
1386 1.2.4.4 yamt p1sts = CSR_READ_2(sc, P1SR);
1387 1.2.4.4 yamt p2sts = CSR_READ_2(sc, P2SR);
1388 1.2.4.4 yamt if (((p1sts | p2sts) & (1U << 5)) == 0)
1389 1.2.4.4 yamt ifmr->ifm_active |= IFM_NONE;
1390 1.2.4.4 yamt else {
1391 1.2.4.4 yamt ifmr->ifm_status |= IFM_ACTIVE;
1392 1.2.4.4 yamt ifmr->ifm_active |= IFM_100_TX|IFM_FDX;
1393 1.2.4.4 yamt ifmr->ifm_active |= IFM_FLOW|IFM_ETH_RXPAUSE|IFM_ETH_TXPAUSE;
1394 1.2.4.4 yamt }
1395 1.2.4.4 yamt sc->sc_media_status = ifmr->ifm_status;
1396 1.2.4.4 yamt sc->sc_media_active = ifmr->ifm_active;
1397 1.2.4.4 yamt }
1398 1.2.4.4 yamt
1399 1.2.4.4 yamt #ifdef KSE_EVENT_COUNTERS
1400 1.2.4.4 yamt static void
1401 1.2.4.4 yamt stat_tick(arg)
1402 1.2.4.4 yamt void *arg;
1403 1.2.4.4 yamt {
1404 1.2.4.4 yamt struct kse_softc *sc = arg;
1405 1.2.4.4 yamt struct ksext *ee = &sc->sc_ext;
1406 1.2.4.4 yamt int nport, p, i, val;
1407 1.2.4.4 yamt
1408 1.2.4.4 yamt nport = (sc->sc_chip == 0x8842) ? 3 : 1;
1409 1.2.4.4 yamt for (p = 0; p < nport; p++) {
1410 1.2.4.4 yamt for (i = 0; i < 32; i++) {
1411 1.2.4.4 yamt val = 0x1c00 | (p * 0x20 + i);
1412 1.2.4.4 yamt CSR_WRITE_2(sc, IACR, val);
1413 1.2.4.4 yamt do {
1414 1.2.4.4 yamt val = CSR_READ_2(sc, IADR5) << 16;
1415 1.2.4.4 yamt } while ((val & (1U << 30)) == 0);
1416 1.2.4.4 yamt if (val & (1U << 31)) {
1417 1.2.4.4 yamt (void)CSR_READ_2(sc, IADR4);
1418 1.2.4.4 yamt val = 0x3fffffff; /* has made overflow */
1419 1.2.4.4 yamt }
1420 1.2.4.4 yamt else {
1421 1.2.4.4 yamt val &= 0x3fff0000; /* 29:16 */
1422 1.2.4.4 yamt val |= CSR_READ_2(sc, IADR4); /* 15:0 */
1423 1.2.4.4 yamt }
1424 1.2.4.4 yamt ee->pev[p][i].ev_count += val; /* i (0-31) */
1425 1.2.4.4 yamt }
1426 1.2.4.4 yamt CSR_WRITE_2(sc, IACR, 0x1c00 + 0x100 + p);
1427 1.2.4.4 yamt ee->pev[p][32].ev_count = CSR_READ_2(sc, IADR4); /* 32 */
1428 1.2.4.4 yamt CSR_WRITE_2(sc, IACR, 0x1c00 + 0x100 + p * 3 + 1);
1429 1.2.4.4 yamt ee->pev[p][33].ev_count = CSR_READ_2(sc, IADR4); /* 33 */
1430 1.2.4.4 yamt }
1431 1.2.4.4 yamt callout_reset(&sc->sc_stat_ch, hz * 60, stat_tick, arg);
1432 1.2.4.4 yamt }
1433 1.2.4.4 yamt
1434 1.2.4.4 yamt static void
1435 1.2.4.4 yamt zerostats(struct kse_softc *sc)
1436 1.2.4.4 yamt {
1437 1.2.4.4 yamt struct ksext *ee = &sc->sc_ext;
1438 1.2.4.4 yamt int nport, p, i, val;
1439 1.2.4.4 yamt
1440 1.2.4.4 yamt /* make sure all the HW counters get zero */
1441 1.2.4.4 yamt nport = (sc->sc_chip == 0x8842) ? 3 : 1;
1442 1.2.4.4 yamt for (p = 0; p < nport; p++) {
1443 1.2.4.4 yamt for (i = 0; i < 31; i++) {
1444 1.2.4.4 yamt val = 0x1c00 | (p * 0x20 + i);
1445 1.2.4.4 yamt CSR_WRITE_2(sc, IACR, val);
1446 1.2.4.4 yamt do {
1447 1.2.4.4 yamt val = CSR_READ_2(sc, IADR5) << 16;
1448 1.2.4.4 yamt } while ((val & (1U << 30)) == 0);
1449 1.2.4.4 yamt (void)CSR_READ_2(sc, IADR4);
1450 1.2.4.4 yamt ee->pev[p][i].ev_count = 0;
1451 1.2.4.4 yamt }
1452 1.2.4.4 yamt }
1453 1.2.4.4 yamt }
1454 1.2.4.4 yamt #endif
1455