dwc_gmac.c revision 1.75 1 1.75 andvar /* $NetBSD: dwc_gmac.c,v 1.75 2021/09/11 20:28:06 andvar Exp $ */
2 1.18 jmcneill
3 1.1 martin /*-
4 1.1 martin * Copyright (c) 2013, 2014 The NetBSD Foundation, Inc.
5 1.1 martin * All rights reserved.
6 1.1 martin *
7 1.1 martin * This code is derived from software contributed to The NetBSD Foundation
8 1.1 martin * by Matt Thomas of 3am Software Foundry and Martin Husemann.
9 1.1 martin *
10 1.1 martin * Redistribution and use in source and binary forms, with or without
11 1.1 martin * modification, are permitted provided that the following conditions
12 1.1 martin * are met:
13 1.1 martin * 1. Redistributions of source code must retain the above copyright
14 1.1 martin * notice, this list of conditions and the following disclaimer.
15 1.1 martin * 2. Redistributions in binary form must reproduce the above copyright
16 1.1 martin * notice, this list of conditions and the following disclaimer in the
17 1.1 martin * documentation and/or other materials provided with the distribution.
18 1.1 martin *
19 1.1 martin * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
20 1.1 martin * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
21 1.1 martin * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
22 1.1 martin * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
23 1.1 martin * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
24 1.1 martin * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
25 1.1 martin * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
26 1.1 martin * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
27 1.1 martin * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
28 1.1 martin * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
29 1.1 martin * POSSIBILITY OF SUCH DAMAGE.
30 1.1 martin */
31 1.1 martin
32 1.1 martin /*
33 1.1 martin * This driver supports the Synopsis Designware GMAC core, as found
34 1.1 martin * on Allwinner A20 cores and others.
35 1.1 martin *
36 1.1 martin * Real documentation seems to not be available, the marketing product
37 1.1 martin * documents could be found here:
38 1.1 martin *
39 1.1 martin * http://www.synopsys.com/dw/ipdir.php?ds=dwc_ether_mac10_100_1000_unive
40 1.1 martin */
41 1.1 martin
42 1.1 martin #include <sys/cdefs.h>
43 1.1 martin
44 1.75 andvar __KERNEL_RCSID(1, "$NetBSD: dwc_gmac.c,v 1.75 2021/09/11 20:28:06 andvar Exp $");
45 1.7 martin
46 1.7 martin /* #define DWC_GMAC_DEBUG 1 */
47 1.1 martin
48 1.38 skrll #ifdef _KERNEL_OPT
49 1.1 martin #include "opt_inet.h"
50 1.38 skrll #include "opt_net_mpsafe.h"
51 1.38 skrll #endif
52 1.1 martin
53 1.1 martin #include <sys/param.h>
54 1.1 martin #include <sys/bus.h>
55 1.1 martin #include <sys/device.h>
56 1.1 martin #include <sys/intr.h>
57 1.1 martin #include <sys/systm.h>
58 1.1 martin #include <sys/sockio.h>
59 1.29 jmcneill #include <sys/cprng.h>
60 1.63 msaitoh #include <sys/rndsource.h>
61 1.1 martin
62 1.1 martin #include <net/if.h>
63 1.1 martin #include <net/if_ether.h>
64 1.1 martin #include <net/if_media.h>
65 1.1 martin #include <net/bpf.h>
66 1.1 martin #ifdef INET
67 1.1 martin #include <netinet/if_inarp.h>
68 1.1 martin #endif
69 1.1 martin
70 1.1 martin #include <dev/mii/miivar.h>
71 1.1 martin
72 1.1 martin #include <dev/ic/dwc_gmac_reg.h>
73 1.1 martin #include <dev/ic/dwc_gmac_var.h>
74 1.1 martin
75 1.56 msaitoh static int dwc_gmac_miibus_read_reg(device_t, int, int, uint16_t *);
76 1.56 msaitoh static int dwc_gmac_miibus_write_reg(device_t, int, int, uint16_t);
77 1.1 martin static void dwc_gmac_miibus_statchg(struct ifnet *);
78 1.1 martin
79 1.61 msaitoh static int dwc_gmac_reset(struct dwc_gmac_softc *);
80 1.61 msaitoh static void dwc_gmac_write_hwaddr(struct dwc_gmac_softc *, uint8_t *);
81 1.61 msaitoh static int dwc_gmac_alloc_dma_rings(struct dwc_gmac_softc *);
82 1.61 msaitoh static void dwc_gmac_free_dma_rings(struct dwc_gmac_softc *);
83 1.61 msaitoh static int dwc_gmac_alloc_rx_ring(struct dwc_gmac_softc *, struct dwc_gmac_rx_ring *);
84 1.61 msaitoh static void dwc_gmac_reset_rx_ring(struct dwc_gmac_softc *, struct dwc_gmac_rx_ring *);
85 1.61 msaitoh static void dwc_gmac_free_rx_ring(struct dwc_gmac_softc *, struct dwc_gmac_rx_ring *);
86 1.61 msaitoh static int dwc_gmac_alloc_tx_ring(struct dwc_gmac_softc *, struct dwc_gmac_tx_ring *);
87 1.61 msaitoh static void dwc_gmac_reset_tx_ring(struct dwc_gmac_softc *, struct dwc_gmac_tx_ring *);
88 1.61 msaitoh static void dwc_gmac_free_tx_ring(struct dwc_gmac_softc *, struct dwc_gmac_tx_ring *);
89 1.61 msaitoh static void dwc_gmac_txdesc_sync(struct dwc_gmac_softc *, int, int, int);
90 1.61 msaitoh static int dwc_gmac_init(struct ifnet *);
91 1.61 msaitoh static int dwc_gmac_init_locked(struct ifnet *);
92 1.61 msaitoh static void dwc_gmac_stop(struct ifnet *, int);
93 1.61 msaitoh static void dwc_gmac_stop_locked(struct ifnet *, int);
94 1.61 msaitoh static void dwc_gmac_start(struct ifnet *);
95 1.61 msaitoh static void dwc_gmac_start_locked(struct ifnet *);
96 1.61 msaitoh static int dwc_gmac_queue(struct dwc_gmac_softc *, struct mbuf *);
97 1.1 martin static int dwc_gmac_ioctl(struct ifnet *, u_long, void *);
98 1.61 msaitoh static void dwc_gmac_tx_intr(struct dwc_gmac_softc *);
99 1.61 msaitoh static void dwc_gmac_rx_intr(struct dwc_gmac_softc *);
100 1.61 msaitoh static void dwc_gmac_setmulti(struct dwc_gmac_softc *);
101 1.22 martin static int dwc_gmac_ifflags_cb(struct ethercom *);
102 1.61 msaitoh static uint32_t bitrev32(uint32_t);
103 1.55 martin static void dwc_gmac_desc_set_owned_by_dev(struct dwc_gmac_dev_dmadesc *);
104 1.55 martin static int dwc_gmac_desc_is_owned_by_dev(struct dwc_gmac_dev_dmadesc *);
105 1.55 martin static void dwc_gmac_desc_std_set_len(struct dwc_gmac_dev_dmadesc *, int);
106 1.55 martin static uint32_t dwc_gmac_desc_std_get_len(struct dwc_gmac_dev_dmadesc *);
107 1.55 martin static void dwc_gmac_desc_std_tx_init_flags(struct dwc_gmac_dev_dmadesc *);
108 1.55 martin static void dwc_gmac_desc_std_tx_set_first_frag(struct dwc_gmac_dev_dmadesc *);
109 1.55 martin static void dwc_gmac_desc_std_tx_set_last_frag(struct dwc_gmac_dev_dmadesc *);
110 1.55 martin static void dwc_gmac_desc_std_rx_init_flags(struct dwc_gmac_dev_dmadesc *);
111 1.55 martin static int dwc_gmac_desc_std_rx_has_error(struct dwc_gmac_dev_dmadesc *);
112 1.55 martin static void dwc_gmac_desc_enh_set_len(struct dwc_gmac_dev_dmadesc *, int);
113 1.55 martin static uint32_t dwc_gmac_desc_enh_get_len(struct dwc_gmac_dev_dmadesc *);
114 1.55 martin static void dwc_gmac_desc_enh_tx_init_flags(struct dwc_gmac_dev_dmadesc *);
115 1.55 martin static void dwc_gmac_desc_enh_tx_set_first_frag(struct dwc_gmac_dev_dmadesc *);
116 1.55 martin static void dwc_gmac_desc_enh_tx_set_last_frag(struct dwc_gmac_dev_dmadesc *);
117 1.55 martin static void dwc_gmac_desc_enh_rx_init_flags(struct dwc_gmac_dev_dmadesc *);
118 1.55 martin static int dwc_gmac_desc_enh_rx_has_error(struct dwc_gmac_dev_dmadesc *);
119 1.55 martin
120 1.55 martin static const struct dwc_gmac_desc_methods desc_methods_standard = {
121 1.55 martin .tx_init_flags = dwc_gmac_desc_std_tx_init_flags,
122 1.55 martin .tx_set_owned_by_dev = dwc_gmac_desc_set_owned_by_dev,
123 1.55 martin .tx_is_owned_by_dev = dwc_gmac_desc_is_owned_by_dev,
124 1.55 martin .tx_set_len = dwc_gmac_desc_std_set_len,
125 1.55 martin .tx_set_first_frag = dwc_gmac_desc_std_tx_set_first_frag,
126 1.55 martin .tx_set_last_frag = dwc_gmac_desc_std_tx_set_last_frag,
127 1.55 martin .rx_init_flags = dwc_gmac_desc_std_rx_init_flags,
128 1.55 martin .rx_set_owned_by_dev = dwc_gmac_desc_set_owned_by_dev,
129 1.55 martin .rx_is_owned_by_dev = dwc_gmac_desc_is_owned_by_dev,
130 1.55 martin .rx_set_len = dwc_gmac_desc_std_set_len,
131 1.55 martin .rx_get_len = dwc_gmac_desc_std_get_len,
132 1.55 martin .rx_has_error = dwc_gmac_desc_std_rx_has_error
133 1.55 martin };
134 1.55 martin
135 1.55 martin static const struct dwc_gmac_desc_methods desc_methods_enhanced = {
136 1.55 martin .tx_init_flags = dwc_gmac_desc_enh_tx_init_flags,
137 1.55 martin .tx_set_owned_by_dev = dwc_gmac_desc_set_owned_by_dev,
138 1.55 martin .tx_is_owned_by_dev = dwc_gmac_desc_is_owned_by_dev,
139 1.55 martin .tx_set_len = dwc_gmac_desc_enh_set_len,
140 1.55 martin .tx_set_first_frag = dwc_gmac_desc_enh_tx_set_first_frag,
141 1.55 martin .tx_set_last_frag = dwc_gmac_desc_enh_tx_set_last_frag,
142 1.55 martin .rx_init_flags = dwc_gmac_desc_enh_rx_init_flags,
143 1.55 martin .rx_set_owned_by_dev = dwc_gmac_desc_set_owned_by_dev,
144 1.55 martin .rx_is_owned_by_dev = dwc_gmac_desc_is_owned_by_dev,
145 1.55 martin .rx_set_len = dwc_gmac_desc_enh_set_len,
146 1.55 martin .rx_get_len = dwc_gmac_desc_enh_get_len,
147 1.55 martin .rx_has_error = dwc_gmac_desc_enh_rx_has_error
148 1.55 martin };
149 1.55 martin
150 1.1 martin
151 1.1 martin #define TX_DESC_OFFSET(N) ((AWGE_RX_RING_COUNT+(N)) \
152 1.1 martin *sizeof(struct dwc_gmac_dev_dmadesc))
153 1.8 martin #define TX_NEXT(N) (((N)+1) & (AWGE_TX_RING_COUNT-1))
154 1.1 martin
155 1.1 martin #define RX_DESC_OFFSET(N) ((N)*sizeof(struct dwc_gmac_dev_dmadesc))
156 1.8 martin #define RX_NEXT(N) (((N)+1) & (AWGE_RX_RING_COUNT-1))
157 1.8 martin
158 1.8 martin
159 1.8 martin
160 1.61 msaitoh #define GMAC_DEF_DMA_INT_MASK (GMAC_DMA_INT_TIE | GMAC_DMA_INT_RIE | \
161 1.61 msaitoh GMAC_DMA_INT_NIE | GMAC_DMA_INT_AIE | \
162 1.61 msaitoh GMAC_DMA_INT_FBE | GMAC_DMA_INT_UNE)
163 1.61 msaitoh
164 1.61 msaitoh #define GMAC_DMA_INT_ERRORS (GMAC_DMA_INT_AIE | GMAC_DMA_INT_ERE | \
165 1.61 msaitoh GMAC_DMA_INT_FBE | \
166 1.61 msaitoh GMAC_DMA_INT_RWE | GMAC_DMA_INT_RUE | \
167 1.61 msaitoh GMAC_DMA_INT_UNE | GMAC_DMA_INT_OVE | \
168 1.10 martin GMAC_DMA_INT_TJE)
169 1.8 martin
170 1.8 martin #define AWIN_DEF_MAC_INTRMASK \
171 1.8 martin (AWIN_GMAC_MAC_INT_TSI | AWIN_GMAC_MAC_INT_ANEG | \
172 1.55 martin AWIN_GMAC_MAC_INT_LINKCHG)
173 1.1 martin
174 1.7 martin #ifdef DWC_GMAC_DEBUG
175 1.61 msaitoh static void dwc_gmac_dump_dma(struct dwc_gmac_softc *);
176 1.61 msaitoh static void dwc_gmac_dump_tx_desc(struct dwc_gmac_softc *);
177 1.61 msaitoh static void dwc_gmac_dump_rx_desc(struct dwc_gmac_softc *);
178 1.61 msaitoh static void dwc_dump_and_abort(struct dwc_gmac_softc *, const char *);
179 1.61 msaitoh static void dwc_dump_status(struct dwc_gmac_softc *);
180 1.61 msaitoh static void dwc_gmac_dump_ffilt(struct dwc_gmac_softc *, uint32_t);
181 1.7 martin #endif
182 1.7 martin
183 1.51 jmcneill int
184 1.57 martin dwc_gmac_attach(struct dwc_gmac_softc *sc, int phy_id, uint32_t mii_clk)
185 1.1 martin {
186 1.1 martin uint8_t enaddr[ETHER_ADDR_LEN];
187 1.55 martin uint32_t maclo, machi, ver, hwft;
188 1.1 martin struct mii_data * const mii = &sc->sc_mii;
189 1.1 martin struct ifnet * const ifp = &sc->sc_ec.ec_if;
190 1.5 martin prop_dictionary_t dict;
191 1.1 martin
192 1.1 martin mutex_init(&sc->sc_mdio_lock, MUTEX_DEFAULT, IPL_NET);
193 1.3 martin sc->sc_mii_clk = mii_clk & 7;
194 1.1 martin
195 1.5 martin dict = device_properties(sc->sc_dev);
196 1.5 martin prop_data_t ea = dict ? prop_dictionary_get(dict, "mac-address") : NULL;
197 1.5 martin if (ea != NULL) {
198 1.5 martin /*
199 1.75 andvar * If the MAC address is overridden by a device property,
200 1.5 martin * use that.
201 1.5 martin */
202 1.5 martin KASSERT(prop_object_type(ea) == PROP_TYPE_DATA);
203 1.5 martin KASSERT(prop_data_size(ea) == ETHER_ADDR_LEN);
204 1.71 jmcneill memcpy(enaddr, prop_data_value(ea), ETHER_ADDR_LEN);
205 1.5 martin } else {
206 1.5 martin /*
207 1.5 martin * If we did not get an externaly configure address,
208 1.5 martin * try to read one from the current filter setup,
209 1.5 martin * before resetting the chip.
210 1.5 martin */
211 1.8 martin maclo = bus_space_read_4(sc->sc_bst, sc->sc_bsh,
212 1.8 martin AWIN_GMAC_MAC_ADDR0LO);
213 1.8 martin machi = bus_space_read_4(sc->sc_bst, sc->sc_bsh,
214 1.8 martin AWIN_GMAC_MAC_ADDR0HI);
215 1.14 jmcneill
216 1.14 jmcneill if (maclo == 0xffffffff && (machi & 0xffff) == 0xffff) {
217 1.29 jmcneill /* fake MAC address */
218 1.29 jmcneill maclo = 0x00f2 | (cprng_strong32() << 16);
219 1.29 jmcneill machi = cprng_strong32();
220 1.14 jmcneill }
221 1.14 jmcneill
222 1.1 martin enaddr[0] = maclo & 0x0ff;
223 1.1 martin enaddr[1] = (maclo >> 8) & 0x0ff;
224 1.1 martin enaddr[2] = (maclo >> 16) & 0x0ff;
225 1.1 martin enaddr[3] = (maclo >> 24) & 0x0ff;
226 1.1 martin enaddr[4] = machi & 0x0ff;
227 1.1 martin enaddr[5] = (machi >> 8) & 0x0ff;
228 1.1 martin }
229 1.1 martin
230 1.55 martin ver = bus_space_read_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_MAC_VERSION);
231 1.55 martin aprint_normal_dev(sc->sc_dev, "Core version: %08x\n", ver);
232 1.55 martin
233 1.1 martin /*
234 1.21 joerg * Init chip and do initial setup
235 1.1 martin */
236 1.1 martin if (dwc_gmac_reset(sc) != 0)
237 1.51 jmcneill return ENXIO; /* not much to cleanup, haven't attached yet */
238 1.5 martin dwc_gmac_write_hwaddr(sc, enaddr);
239 1.52 sevan aprint_normal_dev(sc->sc_dev, "Ethernet address %s\n",
240 1.1 martin ether_sprintf(enaddr));
241 1.1 martin
242 1.55 martin hwft = 0;
243 1.55 martin if (ver >= 0x35) {
244 1.55 martin hwft = bus_space_read_4(sc->sc_bst, sc->sc_bsh,
245 1.55 martin AWIN_GMAC_DMA_HWFEATURES);
246 1.55 martin aprint_normal_dev(sc->sc_dev,
247 1.55 martin "HW feature mask: %x\n", hwft);
248 1.55 martin }
249 1.55 martin if (hwft & GMAC_DMA_FEAT_ENHANCED_DESC) {
250 1.55 martin aprint_normal_dev(sc->sc_dev,
251 1.55 martin "Using enhanced descriptor format\n");
252 1.55 martin sc->sc_descm = &desc_methods_enhanced;
253 1.55 martin } else {
254 1.55 martin sc->sc_descm = &desc_methods_standard;
255 1.55 martin }
256 1.70 chs if (hwft & GMAC_DMA_FEAT_RMON) {
257 1.70 chs uint32_t val;
258 1.70 chs
259 1.70 chs /* Mask all MMC interrupts */
260 1.70 chs val = 0xffffffff;
261 1.70 chs bus_space_write_4(sc->sc_bst, sc->sc_bsh,
262 1.70 chs GMAC_MMC_RX_INT_MSK, val);
263 1.70 chs bus_space_write_4(sc->sc_bst, sc->sc_bsh,
264 1.70 chs GMAC_MMC_TX_INT_MSK, val);
265 1.70 chs }
266 1.55 martin
267 1.1 martin /*
268 1.1 martin * Allocate Tx and Rx rings
269 1.1 martin */
270 1.1 martin if (dwc_gmac_alloc_dma_rings(sc) != 0) {
271 1.1 martin aprint_error_dev(sc->sc_dev, "could not allocate DMA rings\n");
272 1.1 martin goto fail;
273 1.1 martin }
274 1.38 skrll
275 1.1 martin if (dwc_gmac_alloc_tx_ring(sc, &sc->sc_txq) != 0) {
276 1.1 martin aprint_error_dev(sc->sc_dev, "could not allocate Tx ring\n");
277 1.1 martin goto fail;
278 1.1 martin }
279 1.1 martin
280 1.1 martin if (dwc_gmac_alloc_rx_ring(sc, &sc->sc_rxq) != 0) {
281 1.1 martin aprint_error_dev(sc->sc_dev, "could not allocate Rx ring\n");
282 1.1 martin goto fail;
283 1.1 martin }
284 1.1 martin
285 1.38 skrll sc->sc_lock = mutex_obj_alloc(MUTEX_DEFAULT, IPL_NET);
286 1.38 skrll mutex_init(&sc->sc_txq.t_mtx, MUTEX_DEFAULT, IPL_NET);
287 1.38 skrll mutex_init(&sc->sc_rxq.r_mtx, MUTEX_DEFAULT, IPL_NET);
288 1.38 skrll
289 1.1 martin /*
290 1.1 martin * Prepare interface data
291 1.1 martin */
292 1.1 martin ifp->if_softc = sc;
293 1.1 martin strlcpy(ifp->if_xname, device_xname(sc->sc_dev), IFNAMSIZ);
294 1.1 martin ifp->if_flags = IFF_BROADCAST | IFF_SIMPLEX | IFF_MULTICAST;
295 1.44 ozaki #ifdef DWCGMAC_MPSAFE
296 1.43 ozaki ifp->if_extflags = IFEF_MPSAFE;
297 1.44 ozaki #endif
298 1.1 martin ifp->if_ioctl = dwc_gmac_ioctl;
299 1.1 martin ifp->if_start = dwc_gmac_start;
300 1.1 martin ifp->if_init = dwc_gmac_init;
301 1.1 martin ifp->if_stop = dwc_gmac_stop;
302 1.1 martin IFQ_SET_MAXLEN(&ifp->if_snd, IFQ_MAXLEN);
303 1.1 martin IFQ_SET_READY(&ifp->if_snd);
304 1.1 martin
305 1.1 martin /*
306 1.1 martin * Attach MII subdevices
307 1.1 martin */
308 1.2 martin sc->sc_ec.ec_mii = &sc->sc_mii;
309 1.1 martin ifmedia_init(&mii->mii_media, 0, ether_mediachange, ether_mediastatus);
310 1.62 msaitoh mii->mii_ifp = ifp;
311 1.62 msaitoh mii->mii_readreg = dwc_gmac_miibus_read_reg;
312 1.62 msaitoh mii->mii_writereg = dwc_gmac_miibus_write_reg;
313 1.62 msaitoh mii->mii_statchg = dwc_gmac_miibus_statchg;
314 1.62 msaitoh mii_attach(sc->sc_dev, mii, 0xffffffff, phy_id, MII_OFFSET_ANY,
315 1.25 jmcneill MIIF_DOPAUSE);
316 1.1 martin
317 1.62 msaitoh if (LIST_EMPTY(&mii->mii_phys)) {
318 1.62 msaitoh aprint_error_dev(sc->sc_dev, "no PHY found!\n");
319 1.62 msaitoh ifmedia_add(&mii->mii_media, IFM_ETHER | IFM_MANUAL, 0, NULL);
320 1.62 msaitoh ifmedia_set(&mii->mii_media, IFM_ETHER | IFM_MANUAL);
321 1.62 msaitoh } else {
322 1.62 msaitoh ifmedia_set(&mii->mii_media, IFM_ETHER | IFM_AUTO);
323 1.62 msaitoh }
324 1.1 martin
325 1.1 martin /*
326 1.33 tnn * We can support 802.1Q VLAN-sized frames.
327 1.33 tnn */
328 1.33 tnn sc->sc_ec.ec_capabilities |= ETHERCAP_VLAN_MTU;
329 1.33 tnn
330 1.33 tnn /*
331 1.1 martin * Ready, attach interface
332 1.1 martin */
333 1.38 skrll /* Attach the interface. */
334 1.74 riastrad if_initialize(ifp);
335 1.38 skrll sc->sc_ipq = if_percpuq_create(&sc->sc_ec.ec_if);
336 1.40 ozaki if_deferred_start_init(ifp, NULL);
337 1.1 martin ether_ifattach(ifp, enaddr);
338 1.22 martin ether_set_ifflags_cb(&sc->sc_ec, dwc_gmac_ifflags_cb);
339 1.38 skrll if_register(ifp);
340 1.63 msaitoh rnd_attach_source(&sc->rnd_source, device_xname(sc->sc_dev),
341 1.63 msaitoh RND_TYPE_NET, RND_FLAG_DEFAULT);
342 1.1 martin
343 1.1 martin /*
344 1.1 martin * Enable interrupts
345 1.1 martin */
346 1.38 skrll mutex_enter(sc->sc_lock);
347 1.25 jmcneill bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_MAC_INTMASK,
348 1.8 martin AWIN_DEF_MAC_INTRMASK);
349 1.8 martin bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_DMA_INTENABLE,
350 1.8 martin GMAC_DEF_DMA_INT_MASK);
351 1.38 skrll mutex_exit(sc->sc_lock);
352 1.1 martin
353 1.51 jmcneill return 0;
354 1.51 jmcneill
355 1.1 martin fail:
356 1.1 martin dwc_gmac_free_rx_ring(sc, &sc->sc_rxq);
357 1.1 martin dwc_gmac_free_tx_ring(sc, &sc->sc_txq);
358 1.41 msaitoh dwc_gmac_free_dma_rings(sc);
359 1.41 msaitoh mutex_destroy(&sc->sc_mdio_lock);
360 1.51 jmcneill
361 1.51 jmcneill return ENXIO;
362 1.1 martin }
363 1.1 martin
364 1.1 martin
365 1.1 martin
366 1.1 martin static int
367 1.1 martin dwc_gmac_reset(struct dwc_gmac_softc *sc)
368 1.1 martin {
369 1.1 martin size_t cnt;
370 1.1 martin bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_DMA_BUSMODE,
371 1.61 msaitoh bus_space_read_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_DMA_BUSMODE)
372 1.61 msaitoh | GMAC_BUSMODE_RESET);
373 1.72 ryo for (cnt = 0; cnt < 30000; cnt++) {
374 1.1 martin if ((bus_space_read_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_DMA_BUSMODE)
375 1.1 martin & GMAC_BUSMODE_RESET) == 0)
376 1.1 martin return 0;
377 1.1 martin delay(10);
378 1.1 martin }
379 1.1 martin
380 1.1 martin aprint_error_dev(sc->sc_dev, "reset timed out\n");
381 1.1 martin return EIO;
382 1.1 martin }
383 1.1 martin
384 1.1 martin static void
385 1.1 martin dwc_gmac_write_hwaddr(struct dwc_gmac_softc *sc,
386 1.1 martin uint8_t enaddr[ETHER_ADDR_LEN])
387 1.1 martin {
388 1.49 jmcneill uint32_t hi, lo;
389 1.1 martin
390 1.49 jmcneill hi = enaddr[4] | (enaddr[5] << 8);
391 1.1 martin lo = enaddr[0] | (enaddr[1] << 8) | (enaddr[2] << 16)
392 1.73 msaitoh | ((uint32_t)enaddr[3] << 24);
393 1.49 jmcneill bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_MAC_ADDR0HI, hi);
394 1.1 martin bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_MAC_ADDR0LO, lo);
395 1.1 martin }
396 1.1 martin
397 1.1 martin static int
398 1.56 msaitoh dwc_gmac_miibus_read_reg(device_t self, int phy, int reg, uint16_t *val)
399 1.1 martin {
400 1.1 martin struct dwc_gmac_softc * const sc = device_private(self);
401 1.6 martin uint16_t mii;
402 1.1 martin size_t cnt;
403 1.1 martin
404 1.61 msaitoh mii = __SHIFTIN(phy, GMAC_MII_PHY_MASK)
405 1.61 msaitoh | __SHIFTIN(reg, GMAC_MII_REG_MASK)
406 1.61 msaitoh | __SHIFTIN(sc->sc_mii_clk, GMAC_MII_CLKMASK)
407 1.6 martin | GMAC_MII_BUSY;
408 1.1 martin
409 1.1 martin mutex_enter(&sc->sc_mdio_lock);
410 1.6 martin bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_MAC_MIIADDR, mii);
411 1.1 martin
412 1.1 martin for (cnt = 0; cnt < 1000; cnt++) {
413 1.3 martin if (!(bus_space_read_4(sc->sc_bst, sc->sc_bsh,
414 1.3 martin AWIN_GMAC_MAC_MIIADDR) & GMAC_MII_BUSY)) {
415 1.56 msaitoh *val = bus_space_read_4(sc->sc_bst, sc->sc_bsh,
416 1.3 martin AWIN_GMAC_MAC_MIIDATA);
417 1.1 martin break;
418 1.1 martin }
419 1.1 martin delay(10);
420 1.1 martin }
421 1.1 martin
422 1.1 martin mutex_exit(&sc->sc_mdio_lock);
423 1.1 martin
424 1.56 msaitoh if (cnt >= 1000)
425 1.56 msaitoh return ETIMEDOUT;
426 1.61 msaitoh
427 1.56 msaitoh return 0;
428 1.1 martin }
429 1.1 martin
430 1.56 msaitoh static int
431 1.56 msaitoh dwc_gmac_miibus_write_reg(device_t self, int phy, int reg, uint16_t val)
432 1.1 martin {
433 1.1 martin struct dwc_gmac_softc * const sc = device_private(self);
434 1.6 martin uint16_t mii;
435 1.1 martin size_t cnt;
436 1.1 martin
437 1.61 msaitoh mii = __SHIFTIN(phy, GMAC_MII_PHY_MASK)
438 1.61 msaitoh | __SHIFTIN(reg, GMAC_MII_REG_MASK)
439 1.61 msaitoh | __SHIFTIN(sc->sc_mii_clk, GMAC_MII_CLKMASK)
440 1.6 martin | GMAC_MII_BUSY | GMAC_MII_WRITE;
441 1.1 martin
442 1.1 martin mutex_enter(&sc->sc_mdio_lock);
443 1.1 martin bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_MAC_MIIDATA, val);
444 1.6 martin bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_MAC_MIIADDR, mii);
445 1.1 martin
446 1.1 martin for (cnt = 0; cnt < 1000; cnt++) {
447 1.3 martin if (!(bus_space_read_4(sc->sc_bst, sc->sc_bsh,
448 1.3 martin AWIN_GMAC_MAC_MIIADDR) & GMAC_MII_BUSY))
449 1.1 martin break;
450 1.1 martin delay(10);
451 1.1 martin }
452 1.38 skrll
453 1.1 martin mutex_exit(&sc->sc_mdio_lock);
454 1.56 msaitoh
455 1.56 msaitoh if (cnt >= 1000)
456 1.56 msaitoh return ETIMEDOUT;
457 1.56 msaitoh
458 1.56 msaitoh return 0;
459 1.1 martin }
460 1.1 martin
461 1.1 martin static int
462 1.1 martin dwc_gmac_alloc_rx_ring(struct dwc_gmac_softc *sc,
463 1.1 martin struct dwc_gmac_rx_ring *ring)
464 1.1 martin {
465 1.1 martin struct dwc_gmac_rx_data *data;
466 1.1 martin bus_addr_t physaddr;
467 1.6 martin const size_t descsize = AWGE_RX_RING_COUNT * sizeof(*ring->r_desc);
468 1.1 martin int error, i, next;
469 1.1 martin
470 1.1 martin ring->r_cur = ring->r_next = 0;
471 1.1 martin memset(ring->r_desc, 0, descsize);
472 1.1 martin
473 1.1 martin /*
474 1.1 martin * Pre-allocate Rx buffers and populate Rx ring.
475 1.1 martin */
476 1.1 martin for (i = 0; i < AWGE_RX_RING_COUNT; i++) {
477 1.1 martin struct dwc_gmac_dev_dmadesc *desc;
478 1.1 martin
479 1.1 martin data = &sc->sc_rxq.r_data[i];
480 1.1 martin
481 1.1 martin MGETHDR(data->rd_m, M_DONTWAIT, MT_DATA);
482 1.1 martin if (data->rd_m == NULL) {
483 1.1 martin aprint_error_dev(sc->sc_dev,
484 1.1 martin "could not allocate rx mbuf #%d\n", i);
485 1.1 martin error = ENOMEM;
486 1.1 martin goto fail;
487 1.1 martin }
488 1.1 martin error = bus_dmamap_create(sc->sc_dmat, MCLBYTES, 1,
489 1.1 martin MCLBYTES, 0, BUS_DMA_NOWAIT, &data->rd_map);
490 1.1 martin if (error != 0) {
491 1.1 martin aprint_error_dev(sc->sc_dev,
492 1.1 martin "could not create DMA map\n");
493 1.1 martin data->rd_map = NULL;
494 1.1 martin goto fail;
495 1.1 martin }
496 1.1 martin MCLGET(data->rd_m, M_DONTWAIT);
497 1.1 martin if (!(data->rd_m->m_flags & M_EXT)) {
498 1.1 martin aprint_error_dev(sc->sc_dev,
499 1.1 martin "could not allocate mbuf cluster #%d\n", i);
500 1.1 martin error = ENOMEM;
501 1.1 martin goto fail;
502 1.1 martin }
503 1.66 tnn data->rd_m->m_len = data->rd_m->m_pkthdr.len
504 1.66 tnn = data->rd_m->m_ext.ext_size;
505 1.66 tnn if (data->rd_m->m_len > AWGE_MAX_PACKET) {
506 1.66 tnn data->rd_m->m_len = data->rd_m->m_pkthdr.len
507 1.66 tnn = AWGE_MAX_PACKET;
508 1.66 tnn }
509 1.1 martin
510 1.66 tnn error = bus_dmamap_load_mbuf(sc->sc_dmat, data->rd_map,
511 1.66 tnn data->rd_m, BUS_DMA_READ | BUS_DMA_NOWAIT);
512 1.1 martin if (error != 0) {
513 1.1 martin aprint_error_dev(sc->sc_dev,
514 1.1 martin "could not load rx buf DMA map #%d", i);
515 1.1 martin goto fail;
516 1.1 martin }
517 1.66 tnn bus_dmamap_sync(sc->sc_dmat, data->rd_map, 0,
518 1.66 tnn data->rd_map->dm_mapsize, BUS_DMASYNC_PREREAD);
519 1.1 martin physaddr = data->rd_map->dm_segs[0].ds_addr;
520 1.1 martin
521 1.1 martin desc = &sc->sc_rxq.r_desc[i];
522 1.1 martin desc->ddesc_data = htole32(physaddr);
523 1.8 martin next = RX_NEXT(i);
524 1.38 skrll desc->ddesc_next = htole32(ring->r_physaddr
525 1.1 martin + next * sizeof(*desc));
526 1.55 martin sc->sc_descm->rx_init_flags(desc);
527 1.66 tnn sc->sc_descm->rx_set_len(desc, data->rd_m->m_len);
528 1.55 martin sc->sc_descm->rx_set_owned_by_dev(desc);
529 1.1 martin }
530 1.1 martin
531 1.1 martin bus_dmamap_sync(sc->sc_dmat, sc->sc_dma_ring_map, 0,
532 1.1 martin AWGE_RX_RING_COUNT*sizeof(struct dwc_gmac_dev_dmadesc),
533 1.61 msaitoh BUS_DMASYNC_PREWRITE | BUS_DMASYNC_PREREAD);
534 1.1 martin bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_DMA_RX_ADDR,
535 1.6 martin ring->r_physaddr);
536 1.1 martin
537 1.1 martin return 0;
538 1.1 martin
539 1.1 martin fail:
540 1.1 martin dwc_gmac_free_rx_ring(sc, ring);
541 1.1 martin return error;
542 1.1 martin }
543 1.1 martin
544 1.1 martin static void
545 1.1 martin dwc_gmac_reset_rx_ring(struct dwc_gmac_softc *sc,
546 1.1 martin struct dwc_gmac_rx_ring *ring)
547 1.1 martin {
548 1.1 martin struct dwc_gmac_dev_dmadesc *desc;
549 1.66 tnn struct dwc_gmac_rx_data *data;
550 1.1 martin int i;
551 1.1 martin
552 1.38 skrll mutex_enter(&ring->r_mtx);
553 1.1 martin for (i = 0; i < AWGE_RX_RING_COUNT; i++) {
554 1.1 martin desc = &sc->sc_rxq.r_desc[i];
555 1.66 tnn data = &sc->sc_rxq.r_data[i];
556 1.55 martin sc->sc_descm->rx_init_flags(desc);
557 1.66 tnn sc->sc_descm->rx_set_len(desc, data->rd_m->m_len);
558 1.55 martin sc->sc_descm->rx_set_owned_by_dev(desc);
559 1.1 martin }
560 1.1 martin
561 1.1 martin bus_dmamap_sync(sc->sc_dmat, sc->sc_dma_ring_map, 0,
562 1.1 martin AWGE_RX_RING_COUNT*sizeof(struct dwc_gmac_dev_dmadesc),
563 1.61 msaitoh BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
564 1.1 martin
565 1.1 martin ring->r_cur = ring->r_next = 0;
566 1.11 martin /* reset DMA address to start of ring */
567 1.11 martin bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_DMA_RX_ADDR,
568 1.11 martin sc->sc_rxq.r_physaddr);
569 1.38 skrll mutex_exit(&ring->r_mtx);
570 1.1 martin }
571 1.1 martin
572 1.1 martin static int
573 1.1 martin dwc_gmac_alloc_dma_rings(struct dwc_gmac_softc *sc)
574 1.1 martin {
575 1.1 martin const size_t descsize = AWGE_TOTAL_RING_COUNT *
576 1.1 martin sizeof(struct dwc_gmac_dev_dmadesc);
577 1.1 martin int error, nsegs;
578 1.1 martin void *rings;
579 1.1 martin
580 1.1 martin error = bus_dmamap_create(sc->sc_dmat, descsize, 1, descsize, 0,
581 1.1 martin BUS_DMA_NOWAIT, &sc->sc_dma_ring_map);
582 1.1 martin if (error != 0) {
583 1.1 martin aprint_error_dev(sc->sc_dev,
584 1.1 martin "could not create desc DMA map\n");
585 1.1 martin sc->sc_dma_ring_map = NULL;
586 1.1 martin goto fail;
587 1.1 martin }
588 1.1 martin
589 1.1 martin error = bus_dmamem_alloc(sc->sc_dmat, descsize, PAGE_SIZE, 0,
590 1.61 msaitoh &sc->sc_dma_ring_seg, 1, &nsegs, BUS_DMA_NOWAIT |BUS_DMA_COHERENT);
591 1.1 martin if (error != 0) {
592 1.1 martin aprint_error_dev(sc->sc_dev,
593 1.1 martin "could not map DMA memory\n");
594 1.1 martin goto fail;
595 1.1 martin }
596 1.1 martin
597 1.1 martin error = bus_dmamem_map(sc->sc_dmat, &sc->sc_dma_ring_seg, nsegs,
598 1.61 msaitoh descsize, &rings, BUS_DMA_NOWAIT | BUS_DMA_COHERENT);
599 1.1 martin if (error != 0) {
600 1.1 martin aprint_error_dev(sc->sc_dev,
601 1.1 martin "could not allocate DMA memory\n");
602 1.1 martin goto fail;
603 1.1 martin }
604 1.1 martin
605 1.1 martin error = bus_dmamap_load(sc->sc_dmat, sc->sc_dma_ring_map, rings,
606 1.61 msaitoh descsize, NULL, BUS_DMA_NOWAIT | BUS_DMA_COHERENT);
607 1.1 martin if (error != 0) {
608 1.1 martin aprint_error_dev(sc->sc_dev,
609 1.1 martin "could not load desc DMA map\n");
610 1.1 martin goto fail;
611 1.1 martin }
612 1.1 martin
613 1.1 martin /* give first AWGE_RX_RING_COUNT to the RX side */
614 1.1 martin sc->sc_rxq.r_desc = rings;
615 1.1 martin sc->sc_rxq.r_physaddr = sc->sc_dma_ring_map->dm_segs[0].ds_addr;
616 1.1 martin
617 1.1 martin /* and next rings to the TX side */
618 1.1 martin sc->sc_txq.t_desc = sc->sc_rxq.r_desc + AWGE_RX_RING_COUNT;
619 1.38 skrll sc->sc_txq.t_physaddr = sc->sc_rxq.r_physaddr +
620 1.1 martin AWGE_RX_RING_COUNT*sizeof(struct dwc_gmac_dev_dmadesc);
621 1.1 martin
622 1.1 martin return 0;
623 1.1 martin
624 1.1 martin fail:
625 1.1 martin dwc_gmac_free_dma_rings(sc);
626 1.1 martin return error;
627 1.1 martin }
628 1.1 martin
629 1.1 martin static void
630 1.1 martin dwc_gmac_free_dma_rings(struct dwc_gmac_softc *sc)
631 1.1 martin {
632 1.1 martin bus_dmamap_sync(sc->sc_dmat, sc->sc_dma_ring_map, 0,
633 1.1 martin sc->sc_dma_ring_map->dm_mapsize, BUS_DMASYNC_POSTWRITE);
634 1.1 martin bus_dmamap_unload(sc->sc_dmat, sc->sc_dma_ring_map);
635 1.1 martin bus_dmamem_unmap(sc->sc_dmat, sc->sc_rxq.r_desc,
636 1.1 martin AWGE_TOTAL_RING_COUNT * sizeof(struct dwc_gmac_dev_dmadesc));
637 1.1 martin bus_dmamem_free(sc->sc_dmat, &sc->sc_dma_ring_seg, 1);
638 1.1 martin }
639 1.1 martin
640 1.1 martin static void
641 1.1 martin dwc_gmac_free_rx_ring(struct dwc_gmac_softc *sc, struct dwc_gmac_rx_ring *ring)
642 1.1 martin {
643 1.1 martin struct dwc_gmac_rx_data *data;
644 1.1 martin int i;
645 1.1 martin
646 1.1 martin if (ring->r_desc == NULL)
647 1.1 martin return;
648 1.1 martin
649 1.1 martin
650 1.1 martin for (i = 0; i < AWGE_RX_RING_COUNT; i++) {
651 1.1 martin data = &ring->r_data[i];
652 1.1 martin
653 1.1 martin if (data->rd_map != NULL) {
654 1.1 martin bus_dmamap_sync(sc->sc_dmat, data->rd_map, 0,
655 1.1 martin AWGE_RX_RING_COUNT
656 1.1 martin *sizeof(struct dwc_gmac_dev_dmadesc),
657 1.1 martin BUS_DMASYNC_POSTREAD);
658 1.1 martin bus_dmamap_unload(sc->sc_dmat, data->rd_map);
659 1.1 martin bus_dmamap_destroy(sc->sc_dmat, data->rd_map);
660 1.1 martin }
661 1.1 martin if (data->rd_m != NULL)
662 1.1 martin m_freem(data->rd_m);
663 1.1 martin }
664 1.1 martin }
665 1.1 martin
666 1.1 martin static int
667 1.1 martin dwc_gmac_alloc_tx_ring(struct dwc_gmac_softc *sc,
668 1.1 martin struct dwc_gmac_tx_ring *ring)
669 1.1 martin {
670 1.1 martin int i, error = 0;
671 1.1 martin
672 1.1 martin ring->t_queued = 0;
673 1.1 martin ring->t_cur = ring->t_next = 0;
674 1.1 martin
675 1.1 martin memset(ring->t_desc, 0, AWGE_TX_RING_COUNT*sizeof(*ring->t_desc));
676 1.1 martin bus_dmamap_sync(sc->sc_dmat, sc->sc_dma_ring_map,
677 1.1 martin TX_DESC_OFFSET(0),
678 1.1 martin AWGE_TX_RING_COUNT*sizeof(struct dwc_gmac_dev_dmadesc),
679 1.1 martin BUS_DMASYNC_POSTWRITE);
680 1.1 martin
681 1.1 martin for (i = 0; i < AWGE_TX_RING_COUNT; i++) {
682 1.1 martin error = bus_dmamap_create(sc->sc_dmat, MCLBYTES,
683 1.1 martin AWGE_TX_RING_COUNT, MCLBYTES, 0,
684 1.61 msaitoh BUS_DMA_NOWAIT | BUS_DMA_COHERENT,
685 1.1 martin &ring->t_data[i].td_map);
686 1.1 martin if (error != 0) {
687 1.1 martin aprint_error_dev(sc->sc_dev,
688 1.1 martin "could not create TX DMA map #%d\n", i);
689 1.1 martin ring->t_data[i].td_map = NULL;
690 1.1 martin goto fail;
691 1.1 martin }
692 1.1 martin ring->t_desc[i].ddesc_next = htole32(
693 1.1 martin ring->t_physaddr + sizeof(struct dwc_gmac_dev_dmadesc)
694 1.8 martin *TX_NEXT(i));
695 1.1 martin }
696 1.1 martin
697 1.1 martin return 0;
698 1.1 martin
699 1.1 martin fail:
700 1.1 martin dwc_gmac_free_tx_ring(sc, ring);
701 1.1 martin return error;
702 1.1 martin }
703 1.1 martin
704 1.1 martin static void
705 1.1 martin dwc_gmac_txdesc_sync(struct dwc_gmac_softc *sc, int start, int end, int ops)
706 1.1 martin {
707 1.64 mrg /* 'end' is pointing one descriptor beyond the last we want to sync */
708 1.1 martin if (end > start) {
709 1.1 martin bus_dmamap_sync(sc->sc_dmat, sc->sc_dma_ring_map,
710 1.1 martin TX_DESC_OFFSET(start),
711 1.1 martin TX_DESC_OFFSET(end)-TX_DESC_OFFSET(start),
712 1.1 martin ops);
713 1.1 martin return;
714 1.1 martin }
715 1.1 martin /* sync from 'start' to end of ring */
716 1.1 martin bus_dmamap_sync(sc->sc_dmat, sc->sc_dma_ring_map,
717 1.1 martin TX_DESC_OFFSET(start),
718 1.31 jmcneill TX_DESC_OFFSET(AWGE_TX_RING_COUNT)-TX_DESC_OFFSET(start),
719 1.1 martin ops);
720 1.47 jmcneill if (TX_DESC_OFFSET(end) - TX_DESC_OFFSET(0) > 0) {
721 1.47 jmcneill /* sync from start of ring to 'end' */
722 1.47 jmcneill bus_dmamap_sync(sc->sc_dmat, sc->sc_dma_ring_map,
723 1.47 jmcneill TX_DESC_OFFSET(0),
724 1.47 jmcneill TX_DESC_OFFSET(end)-TX_DESC_OFFSET(0),
725 1.47 jmcneill ops);
726 1.47 jmcneill }
727 1.1 martin }
728 1.1 martin
729 1.1 martin static void
730 1.1 martin dwc_gmac_reset_tx_ring(struct dwc_gmac_softc *sc,
731 1.1 martin struct dwc_gmac_tx_ring *ring)
732 1.1 martin {
733 1.1 martin int i;
734 1.1 martin
735 1.38 skrll mutex_enter(&ring->t_mtx);
736 1.1 martin for (i = 0; i < AWGE_TX_RING_COUNT; i++) {
737 1.1 martin struct dwc_gmac_tx_data *data = &ring->t_data[i];
738 1.1 martin
739 1.1 martin if (data->td_m != NULL) {
740 1.1 martin bus_dmamap_sync(sc->sc_dmat, data->td_active,
741 1.1 martin 0, data->td_active->dm_mapsize,
742 1.1 martin BUS_DMASYNC_POSTWRITE);
743 1.1 martin bus_dmamap_unload(sc->sc_dmat, data->td_active);
744 1.1 martin m_freem(data->td_m);
745 1.1 martin data->td_m = NULL;
746 1.1 martin }
747 1.1 martin }
748 1.1 martin
749 1.1 martin bus_dmamap_sync(sc->sc_dmat, sc->sc_dma_ring_map,
750 1.1 martin TX_DESC_OFFSET(0),
751 1.1 martin AWGE_TX_RING_COUNT*sizeof(struct dwc_gmac_dev_dmadesc),
752 1.61 msaitoh BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
753 1.6 martin bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_DMA_TX_ADDR,
754 1.6 martin sc->sc_txq.t_physaddr);
755 1.1 martin
756 1.1 martin ring->t_queued = 0;
757 1.1 martin ring->t_cur = ring->t_next = 0;
758 1.38 skrll mutex_exit(&ring->t_mtx);
759 1.1 martin }
760 1.1 martin
761 1.1 martin static void
762 1.1 martin dwc_gmac_free_tx_ring(struct dwc_gmac_softc *sc,
763 1.1 martin struct dwc_gmac_tx_ring *ring)
764 1.1 martin {
765 1.1 martin int i;
766 1.1 martin
767 1.1 martin /* unload the maps */
768 1.1 martin for (i = 0; i < AWGE_TX_RING_COUNT; i++) {
769 1.1 martin struct dwc_gmac_tx_data *data = &ring->t_data[i];
770 1.1 martin
771 1.1 martin if (data->td_m != NULL) {
772 1.1 martin bus_dmamap_sync(sc->sc_dmat, data->td_active,
773 1.1 martin 0, data->td_map->dm_mapsize,
774 1.1 martin BUS_DMASYNC_POSTWRITE);
775 1.1 martin bus_dmamap_unload(sc->sc_dmat, data->td_active);
776 1.1 martin m_freem(data->td_m);
777 1.1 martin data->td_m = NULL;
778 1.1 martin }
779 1.1 martin }
780 1.1 martin
781 1.1 martin /* and actually free them */
782 1.1 martin for (i = 0; i < AWGE_TX_RING_COUNT; i++) {
783 1.1 martin struct dwc_gmac_tx_data *data = &ring->t_data[i];
784 1.1 martin
785 1.1 martin bus_dmamap_destroy(sc->sc_dmat, data->td_map);
786 1.1 martin }
787 1.1 martin }
788 1.1 martin
789 1.1 martin static void
790 1.1 martin dwc_gmac_miibus_statchg(struct ifnet *ifp)
791 1.1 martin {
792 1.1 martin struct dwc_gmac_softc * const sc = ifp->if_softc;
793 1.1 martin struct mii_data * const mii = &sc->sc_mii;
794 1.25 jmcneill uint32_t conf, flow;
795 1.1 martin
796 1.1 martin /*
797 1.1 martin * Set MII or GMII interface based on the speed
798 1.38 skrll * negotiated by the PHY.
799 1.9 martin */
800 1.9 martin conf = bus_space_read_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_MAC_CONF);
801 1.61 msaitoh conf &= ~(AWIN_GMAC_MAC_CONF_FES100 | AWIN_GMAC_MAC_CONF_MIISEL
802 1.61 msaitoh | AWIN_GMAC_MAC_CONF_FULLDPLX);
803 1.11 martin conf |= AWIN_GMAC_MAC_CONF_FRAMEBURST
804 1.11 martin | AWIN_GMAC_MAC_CONF_DISABLERXOWN
805 1.25 jmcneill | AWIN_GMAC_MAC_CONF_DISABLEJABBER
806 1.25 jmcneill | AWIN_GMAC_MAC_CONF_ACS
807 1.11 martin | AWIN_GMAC_MAC_CONF_RXENABLE
808 1.11 martin | AWIN_GMAC_MAC_CONF_TXENABLE;
809 1.1 martin switch (IFM_SUBTYPE(mii->mii_media_active)) {
810 1.1 martin case IFM_10_T:
811 1.12 jmcneill conf |= AWIN_GMAC_MAC_CONF_MIISEL;
812 1.9 martin break;
813 1.1 martin case IFM_100_TX:
814 1.12 jmcneill conf |= AWIN_GMAC_MAC_CONF_FES100 |
815 1.12 jmcneill AWIN_GMAC_MAC_CONF_MIISEL;
816 1.1 martin break;
817 1.1 martin case IFM_1000_T:
818 1.1 martin break;
819 1.1 martin }
820 1.46 jmcneill if (sc->sc_set_speed)
821 1.46 jmcneill sc->sc_set_speed(sc, IFM_SUBTYPE(mii->mii_media_active));
822 1.25 jmcneill
823 1.25 jmcneill flow = 0;
824 1.25 jmcneill if (IFM_OPTIONS(mii->mii_media_active) & IFM_FDX) {
825 1.9 martin conf |= AWIN_GMAC_MAC_CONF_FULLDPLX;
826 1.25 jmcneill flow |= __SHIFTIN(0x200, AWIN_GMAC_MAC_FLOWCTRL_PAUSE);
827 1.25 jmcneill }
828 1.25 jmcneill if (mii->mii_media_active & IFM_ETH_TXPAUSE) {
829 1.25 jmcneill flow |= AWIN_GMAC_MAC_FLOWCTRL_TFE;
830 1.25 jmcneill }
831 1.25 jmcneill if (mii->mii_media_active & IFM_ETH_RXPAUSE) {
832 1.25 jmcneill flow |= AWIN_GMAC_MAC_FLOWCTRL_RFE;
833 1.25 jmcneill }
834 1.25 jmcneill bus_space_write_4(sc->sc_bst, sc->sc_bsh,
835 1.25 jmcneill AWIN_GMAC_MAC_FLOWCTRL, flow);
836 1.9 martin
837 1.9 martin #ifdef DWC_GMAC_DEBUG
838 1.9 martin aprint_normal_dev(sc->sc_dev,
839 1.9 martin "setting MAC conf register: %08x\n", conf);
840 1.9 martin #endif
841 1.9 martin
842 1.9 martin bus_space_write_4(sc->sc_bst, sc->sc_bsh,
843 1.9 martin AWIN_GMAC_MAC_CONF, conf);
844 1.1 martin }
845 1.1 martin
846 1.1 martin static int
847 1.1 martin dwc_gmac_init(struct ifnet *ifp)
848 1.1 martin {
849 1.1 martin struct dwc_gmac_softc *sc = ifp->if_softc;
850 1.38 skrll
851 1.38 skrll mutex_enter(sc->sc_lock);
852 1.38 skrll int ret = dwc_gmac_init_locked(ifp);
853 1.38 skrll mutex_exit(sc->sc_lock);
854 1.38 skrll
855 1.38 skrll return ret;
856 1.38 skrll }
857 1.38 skrll
858 1.38 skrll static int
859 1.38 skrll dwc_gmac_init_locked(struct ifnet *ifp)
860 1.38 skrll {
861 1.38 skrll struct dwc_gmac_softc *sc = ifp->if_softc;
862 1.13 jmcneill uint32_t ffilt;
863 1.1 martin
864 1.1 martin if (ifp->if_flags & IFF_RUNNING)
865 1.1 martin return 0;
866 1.1 martin
867 1.38 skrll dwc_gmac_stop_locked(ifp, 0);
868 1.1 martin
869 1.1 martin /*
870 1.11 martin * Configure DMA burst/transfer mode and RX/TX priorities.
871 1.11 martin * XXX - the GMAC_BUSMODE_PRIORXTX bits are undocumented.
872 1.11 martin */
873 1.11 martin bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_DMA_BUSMODE,
874 1.25 jmcneill GMAC_BUSMODE_FIXEDBURST | GMAC_BUSMODE_4PBL |
875 1.25 jmcneill __SHIFTIN(2, GMAC_BUSMODE_RPBL) |
876 1.25 jmcneill __SHIFTIN(2, GMAC_BUSMODE_PBL));
877 1.11 martin
878 1.11 martin /*
879 1.13 jmcneill * Set up address filter
880 1.11 martin */
881 1.20 jmcneill ffilt = bus_space_read_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_MAC_FFILT);
882 1.20 jmcneill if (ifp->if_flags & IFF_PROMISC) {
883 1.13 jmcneill ffilt |= AWIN_GMAC_MAC_FFILT_PR;
884 1.20 jmcneill } else {
885 1.20 jmcneill ffilt &= ~AWIN_GMAC_MAC_FFILT_PR;
886 1.20 jmcneill }
887 1.20 jmcneill if (ifp->if_flags & IFF_BROADCAST) {
888 1.20 jmcneill ffilt &= ~AWIN_GMAC_MAC_FFILT_DBF;
889 1.20 jmcneill } else {
890 1.20 jmcneill ffilt |= AWIN_GMAC_MAC_FFILT_DBF;
891 1.20 jmcneill }
892 1.13 jmcneill bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_MAC_FFILT, ffilt);
893 1.11 martin
894 1.11 martin /*
895 1.20 jmcneill * Set up multicast filter
896 1.20 jmcneill */
897 1.20 jmcneill dwc_gmac_setmulti(sc);
898 1.20 jmcneill
899 1.20 jmcneill /*
900 1.6 martin * Set up dma pointer for RX and TX ring
901 1.1 martin */
902 1.6 martin bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_DMA_RX_ADDR,
903 1.6 martin sc->sc_rxq.r_physaddr);
904 1.6 martin bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_DMA_TX_ADDR,
905 1.6 martin sc->sc_txq.t_physaddr);
906 1.6 martin
907 1.6 martin /*
908 1.10 martin * Start RX/TX part
909 1.6 martin */
910 1.46 jmcneill uint32_t opmode = GMAC_DMA_OP_RXSTART | GMAC_DMA_OP_TXSTART;
911 1.46 jmcneill if ((sc->sc_flags & DWC_GMAC_FORCE_THRESH_DMA_MODE) == 0) {
912 1.46 jmcneill opmode |= GMAC_DMA_OP_RXSTOREFORWARD | GMAC_DMA_OP_TXSTOREFORWARD;
913 1.46 jmcneill }
914 1.46 jmcneill bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_DMA_OPMODE, opmode);
915 1.1 martin
916 1.38 skrll sc->sc_stopping = false;
917 1.38 skrll
918 1.1 martin ifp->if_flags |= IFF_RUNNING;
919 1.1 martin ifp->if_flags &= ~IFF_OACTIVE;
920 1.1 martin
921 1.1 martin return 0;
922 1.1 martin }
923 1.1 martin
924 1.1 martin static void
925 1.1 martin dwc_gmac_start(struct ifnet *ifp)
926 1.1 martin {
927 1.1 martin struct dwc_gmac_softc *sc = ifp->if_softc;
928 1.45 martin #ifdef DWCGMAC_MPSAFE
929 1.43 ozaki KASSERT(if_is_mpsafe(ifp));
930 1.45 martin #endif
931 1.38 skrll
932 1.38 skrll mutex_enter(sc->sc_lock);
933 1.38 skrll if (!sc->sc_stopping) {
934 1.38 skrll mutex_enter(&sc->sc_txq.t_mtx);
935 1.38 skrll dwc_gmac_start_locked(ifp);
936 1.38 skrll mutex_exit(&sc->sc_txq.t_mtx);
937 1.38 skrll }
938 1.38 skrll mutex_exit(sc->sc_lock);
939 1.38 skrll }
940 1.38 skrll
941 1.38 skrll static void
942 1.38 skrll dwc_gmac_start_locked(struct ifnet *ifp)
943 1.38 skrll {
944 1.38 skrll struct dwc_gmac_softc *sc = ifp->if_softc;
945 1.1 martin int old = sc->sc_txq.t_queued;
946 1.30 martin int start = sc->sc_txq.t_cur;
947 1.1 martin struct mbuf *m0;
948 1.1 martin
949 1.1 martin if ((ifp->if_flags & (IFF_RUNNING | IFF_OACTIVE)) != IFF_RUNNING)
950 1.1 martin return;
951 1.1 martin
952 1.1 martin for (;;) {
953 1.1 martin IFQ_POLL(&ifp->if_snd, m0);
954 1.1 martin if (m0 == NULL)
955 1.1 martin break;
956 1.1 martin if (dwc_gmac_queue(sc, m0) != 0) {
957 1.1 martin ifp->if_flags |= IFF_OACTIVE;
958 1.1 martin break;
959 1.1 martin }
960 1.1 martin IFQ_DEQUEUE(&ifp->if_snd, m0);
961 1.50 msaitoh bpf_mtap(ifp, m0, BPF_D_OUT);
962 1.32 martin if (sc->sc_txq.t_queued == AWGE_TX_RING_COUNT) {
963 1.32 martin ifp->if_flags |= IFF_OACTIVE;
964 1.32 martin break;
965 1.32 martin }
966 1.1 martin }
967 1.1 martin
968 1.1 martin if (sc->sc_txq.t_queued != old) {
969 1.1 martin /* packets have been queued, kick it off */
970 1.30 martin dwc_gmac_txdesc_sync(sc, start, sc->sc_txq.t_cur,
971 1.61 msaitoh BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
972 1.10 martin
973 1.10 martin #ifdef DWC_GMAC_DEBUG
974 1.10 martin dwc_dump_status(sc);
975 1.10 martin #endif
976 1.55 martin bus_space_write_4(sc->sc_bst, sc->sc_bsh,
977 1.55 martin AWIN_GMAC_DMA_TXPOLL, ~0U);
978 1.1 martin }
979 1.1 martin }
980 1.1 martin
981 1.1 martin static void
982 1.1 martin dwc_gmac_stop(struct ifnet *ifp, int disable)
983 1.1 martin {
984 1.1 martin struct dwc_gmac_softc *sc = ifp->if_softc;
985 1.1 martin
986 1.38 skrll mutex_enter(sc->sc_lock);
987 1.38 skrll dwc_gmac_stop_locked(ifp, disable);
988 1.38 skrll mutex_exit(sc->sc_lock);
989 1.38 skrll }
990 1.38 skrll
991 1.38 skrll static void
992 1.38 skrll dwc_gmac_stop_locked(struct ifnet *ifp, int disable)
993 1.38 skrll {
994 1.38 skrll struct dwc_gmac_softc *sc = ifp->if_softc;
995 1.38 skrll
996 1.38 skrll sc->sc_stopping = true;
997 1.38 skrll
998 1.6 martin bus_space_write_4(sc->sc_bst, sc->sc_bsh,
999 1.6 martin AWIN_GMAC_DMA_OPMODE,
1000 1.6 martin bus_space_read_4(sc->sc_bst, sc->sc_bsh,
1001 1.62 msaitoh AWIN_GMAC_DMA_OPMODE)
1002 1.61 msaitoh & ~(GMAC_DMA_OP_TXSTART | GMAC_DMA_OP_RXSTART));
1003 1.6 martin bus_space_write_4(sc->sc_bst, sc->sc_bsh,
1004 1.6 martin AWIN_GMAC_DMA_OPMODE,
1005 1.6 martin bus_space_read_4(sc->sc_bst, sc->sc_bsh,
1006 1.62 msaitoh AWIN_GMAC_DMA_OPMODE) | GMAC_DMA_OP_FLUSHTX);
1007 1.6 martin
1008 1.1 martin mii_down(&sc->sc_mii);
1009 1.1 martin dwc_gmac_reset_tx_ring(sc, &sc->sc_txq);
1010 1.1 martin dwc_gmac_reset_rx_ring(sc, &sc->sc_rxq);
1011 1.48 jmcneill
1012 1.48 jmcneill ifp->if_flags &= ~(IFF_RUNNING | IFF_OACTIVE);
1013 1.1 martin }
1014 1.1 martin
1015 1.1 martin /*
1016 1.1 martin * Add m0 to the TX ring
1017 1.1 martin */
1018 1.1 martin static int
1019 1.1 martin dwc_gmac_queue(struct dwc_gmac_softc *sc, struct mbuf *m0)
1020 1.1 martin {
1021 1.1 martin struct dwc_gmac_dev_dmadesc *desc = NULL;
1022 1.1 martin struct dwc_gmac_tx_data *data = NULL;
1023 1.1 martin bus_dmamap_t map;
1024 1.1 martin int error, i, first;
1025 1.1 martin
1026 1.8 martin #ifdef DWC_GMAC_DEBUG
1027 1.8 martin aprint_normal_dev(sc->sc_dev,
1028 1.8 martin "dwc_gmac_queue: adding mbuf chain %p\n", m0);
1029 1.8 martin #endif
1030 1.8 martin
1031 1.1 martin first = sc->sc_txq.t_cur;
1032 1.1 martin map = sc->sc_txq.t_data[first].td_map;
1033 1.1 martin
1034 1.1 martin error = bus_dmamap_load_mbuf(sc->sc_dmat, map, m0,
1035 1.61 msaitoh BUS_DMA_WRITE | BUS_DMA_NOWAIT);
1036 1.1 martin if (error != 0) {
1037 1.1 martin aprint_error_dev(sc->sc_dev, "could not map mbuf "
1038 1.1 martin "(len: %d, error %d)\n", m0->m_pkthdr.len, error);
1039 1.1 martin return error;
1040 1.1 martin }
1041 1.1 martin
1042 1.32 martin if (sc->sc_txq.t_queued + map->dm_nsegs > AWGE_TX_RING_COUNT) {
1043 1.1 martin bus_dmamap_unload(sc->sc_dmat, map);
1044 1.1 martin return ENOBUFS;
1045 1.1 martin }
1046 1.1 martin
1047 1.1 martin for (i = 0; i < map->dm_nsegs; i++) {
1048 1.1 martin data = &sc->sc_txq.t_data[sc->sc_txq.t_cur];
1049 1.8 martin desc = &sc->sc_txq.t_desc[sc->sc_txq.t_cur];
1050 1.8 martin
1051 1.8 martin desc->ddesc_data = htole32(map->dm_segs[i].ds_addr);
1052 1.7 martin
1053 1.7 martin #ifdef DWC_GMAC_DEBUG
1054 1.7 martin aprint_normal_dev(sc->sc_dev, "enqueing desc #%d data %08lx "
1055 1.55 martin "len %lu\n", sc->sc_txq.t_cur,
1056 1.7 martin (unsigned long)map->dm_segs[i].ds_addr,
1057 1.55 martin (unsigned long)map->dm_segs[i].ds_len);
1058 1.7 martin #endif
1059 1.7 martin
1060 1.55 martin sc->sc_descm->tx_init_flags(desc);
1061 1.55 martin sc->sc_descm->tx_set_len(desc, map->dm_segs[i].ds_len);
1062 1.55 martin
1063 1.55 martin if (i == 0)
1064 1.55 martin sc->sc_descm->tx_set_first_frag(desc);
1065 1.1 martin
1066 1.1 martin /*
1067 1.1 martin * Defer passing ownership of the first descriptor
1068 1.23 joerg * until we are done.
1069 1.1 martin */
1070 1.55 martin if (i != 0)
1071 1.55 martin sc->sc_descm->tx_set_owned_by_dev(desc);
1072 1.8 martin
1073 1.6 martin sc->sc_txq.t_queued++;
1074 1.8 martin sc->sc_txq.t_cur = TX_NEXT(sc->sc_txq.t_cur);
1075 1.1 martin }
1076 1.1 martin
1077 1.55 martin sc->sc_descm->tx_set_last_frag(desc);
1078 1.1 martin
1079 1.1 martin data->td_m = m0;
1080 1.1 martin data->td_active = map;
1081 1.1 martin
1082 1.1 martin bus_dmamap_sync(sc->sc_dmat, map, 0, map->dm_mapsize,
1083 1.34 jmcneill BUS_DMASYNC_PREWRITE);
1084 1.1 martin
1085 1.32 martin /* Pass first to device */
1086 1.55 martin sc->sc_descm->tx_set_owned_by_dev(&sc->sc_txq.t_desc[first]);
1087 1.55 martin
1088 1.55 martin bus_dmamap_sync(sc->sc_dmat, map, 0, map->dm_mapsize,
1089 1.55 martin BUS_DMASYNC_PREWRITE);
1090 1.32 martin
1091 1.1 martin return 0;
1092 1.1 martin }
1093 1.1 martin
1094 1.22 martin /*
1095 1.22 martin * If the interface is up and running, only modify the receive
1096 1.22 martin * filter when setting promiscuous or debug mode. Otherwise fall
1097 1.22 martin * through to ether_ioctl, which will reset the chip.
1098 1.22 martin */
1099 1.22 martin static int
1100 1.22 martin dwc_gmac_ifflags_cb(struct ethercom *ec)
1101 1.22 martin {
1102 1.22 martin struct ifnet *ifp = &ec->ec_if;
1103 1.22 martin struct dwc_gmac_softc *sc = ifp->if_softc;
1104 1.38 skrll int ret = 0;
1105 1.38 skrll
1106 1.38 skrll mutex_enter(sc->sc_lock);
1107 1.65 msaitoh u_short change = ifp->if_flags ^ sc->sc_if_flags;
1108 1.38 skrll sc->sc_if_flags = ifp->if_flags;
1109 1.22 martin
1110 1.61 msaitoh if ((change & ~(IFF_CANTCHANGE | IFF_DEBUG)) != 0) {
1111 1.38 skrll ret = ENETRESET;
1112 1.38 skrll goto out;
1113 1.38 skrll }
1114 1.38 skrll if ((change & IFF_PROMISC) != 0) {
1115 1.22 martin dwc_gmac_setmulti(sc);
1116 1.38 skrll }
1117 1.38 skrll out:
1118 1.38 skrll mutex_exit(sc->sc_lock);
1119 1.38 skrll
1120 1.38 skrll return ret;
1121 1.22 martin }
1122 1.22 martin
1123 1.1 martin static int
1124 1.1 martin dwc_gmac_ioctl(struct ifnet *ifp, u_long cmd, void *data)
1125 1.1 martin {
1126 1.20 jmcneill struct dwc_gmac_softc *sc = ifp->if_softc;
1127 1.38 skrll int error = 0;
1128 1.38 skrll
1129 1.38 skrll int s = splnet();
1130 1.38 skrll error = ether_ioctl(ifp, cmd, data);
1131 1.1 martin
1132 1.38 skrll #ifdef DWCGMAC_MPSAFE
1133 1.38 skrll splx(s);
1134 1.38 skrll #endif
1135 1.1 martin
1136 1.38 skrll if (error == ENETRESET) {
1137 1.1 martin error = 0;
1138 1.1 martin if (cmd != SIOCADDMULTI && cmd != SIOCDELMULTI)
1139 1.1 martin ;
1140 1.22 martin else if (ifp->if_flags & IFF_RUNNING) {
1141 1.22 martin /*
1142 1.22 martin * Multicast list has changed; set the hardware filter
1143 1.22 martin * accordingly.
1144 1.22 martin */
1145 1.38 skrll mutex_enter(sc->sc_lock);
1146 1.20 jmcneill dwc_gmac_setmulti(sc);
1147 1.38 skrll mutex_exit(sc->sc_lock);
1148 1.22 martin }
1149 1.1 martin }
1150 1.1 martin
1151 1.22 martin /* Try to get things going again */
1152 1.22 martin if (ifp->if_flags & IFF_UP)
1153 1.22 martin dwc_gmac_start(ifp);
1154 1.22 martin sc->sc_if_flags = sc->sc_ec.ec_if.if_flags;
1155 1.38 skrll
1156 1.38 skrll #ifndef DWCGMAC_MPSAFE
1157 1.1 martin splx(s);
1158 1.38 skrll #endif
1159 1.38 skrll
1160 1.1 martin return error;
1161 1.1 martin }
1162 1.1 martin
1163 1.8 martin static void
1164 1.8 martin dwc_gmac_tx_intr(struct dwc_gmac_softc *sc)
1165 1.8 martin {
1166 1.32 martin struct ifnet *ifp = &sc->sc_ec.ec_if;
1167 1.8 martin struct dwc_gmac_tx_data *data;
1168 1.8 martin struct dwc_gmac_dev_dmadesc *desc;
1169 1.32 martin int i, nsegs;
1170 1.8 martin
1171 1.38 skrll mutex_enter(&sc->sc_txq.t_mtx);
1172 1.38 skrll
1173 1.32 martin for (i = sc->sc_txq.t_next; sc->sc_txq.t_queued > 0; i = TX_NEXT(i)) {
1174 1.8 martin #ifdef DWC_GMAC_DEBUG
1175 1.8 martin aprint_normal_dev(sc->sc_dev,
1176 1.8 martin "dwc_gmac_tx_intr: checking desc #%d (t_queued: %d)\n",
1177 1.8 martin i, sc->sc_txq.t_queued);
1178 1.8 martin #endif
1179 1.8 martin
1180 1.26 martin /*
1181 1.26 martin * i+1 does not need to be a valid descriptor,
1182 1.26 martin * this is just a special notion to just sync
1183 1.26 martin * a single tx descriptor (i)
1184 1.26 martin */
1185 1.26 martin dwc_gmac_txdesc_sync(sc, i, i+1,
1186 1.61 msaitoh BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE);
1187 1.11 martin
1188 1.32 martin desc = &sc->sc_txq.t_desc[i];
1189 1.55 martin if (sc->sc_descm->tx_is_owned_by_dev(desc))
1190 1.8 martin break;
1191 1.11 martin
1192 1.8 martin data = &sc->sc_txq.t_data[i];
1193 1.8 martin if (data->td_m == NULL)
1194 1.8 martin continue;
1195 1.32 martin
1196 1.69 thorpej if_statinc(ifp, if_opackets);
1197 1.32 martin nsegs = data->td_active->dm_nsegs;
1198 1.8 martin bus_dmamap_sync(sc->sc_dmat, data->td_active, 0,
1199 1.8 martin data->td_active->dm_mapsize, BUS_DMASYNC_POSTWRITE);
1200 1.8 martin bus_dmamap_unload(sc->sc_dmat, data->td_active);
1201 1.8 martin
1202 1.8 martin #ifdef DWC_GMAC_DEBUG
1203 1.8 martin aprint_normal_dev(sc->sc_dev,
1204 1.8 martin "dwc_gmac_tx_intr: done with packet at desc #%d, "
1205 1.8 martin "freeing mbuf %p\n", i, data->td_m);
1206 1.8 martin #endif
1207 1.8 martin
1208 1.8 martin m_freem(data->td_m);
1209 1.8 martin data->td_m = NULL;
1210 1.32 martin
1211 1.32 martin sc->sc_txq.t_queued -= nsegs;
1212 1.8 martin }
1213 1.8 martin
1214 1.8 martin sc->sc_txq.t_next = i;
1215 1.8 martin
1216 1.8 martin if (sc->sc_txq.t_queued < AWGE_TX_RING_COUNT) {
1217 1.32 martin ifp->if_flags &= ~IFF_OACTIVE;
1218 1.8 martin }
1219 1.38 skrll mutex_exit(&sc->sc_txq.t_mtx);
1220 1.8 martin }
1221 1.8 martin
1222 1.8 martin static void
1223 1.8 martin dwc_gmac_rx_intr(struct dwc_gmac_softc *sc)
1224 1.8 martin {
1225 1.11 martin struct ifnet *ifp = &sc->sc_ec.ec_if;
1226 1.11 martin struct dwc_gmac_dev_dmadesc *desc;
1227 1.11 martin struct dwc_gmac_rx_data *data;
1228 1.11 martin bus_addr_t physaddr;
1229 1.11 martin struct mbuf *m, *mnew;
1230 1.11 martin int i, len, error;
1231 1.11 martin
1232 1.38 skrll mutex_enter(&sc->sc_rxq.r_mtx);
1233 1.11 martin for (i = sc->sc_rxq.r_cur; ; i = RX_NEXT(i)) {
1234 1.11 martin bus_dmamap_sync(sc->sc_dmat, sc->sc_dma_ring_map,
1235 1.11 martin RX_DESC_OFFSET(i), sizeof(*desc),
1236 1.61 msaitoh BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE);
1237 1.11 martin desc = &sc->sc_rxq.r_desc[i];
1238 1.11 martin data = &sc->sc_rxq.r_data[i];
1239 1.11 martin
1240 1.55 martin if (sc->sc_descm->rx_is_owned_by_dev(desc))
1241 1.11 martin break;
1242 1.11 martin
1243 1.55 martin if (sc->sc_descm->rx_has_error(desc)) {
1244 1.11 martin #ifdef DWC_GMAC_DEBUG
1245 1.15 martin aprint_normal_dev(sc->sc_dev,
1246 1.15 martin "RX error: descriptor status %08x, skipping\n",
1247 1.55 martin le32toh(desc->ddesc_status0));
1248 1.11 martin #endif
1249 1.69 thorpej if_statinc(ifp, if_ierrors);
1250 1.11 martin goto skip;
1251 1.11 martin }
1252 1.11 martin
1253 1.55 martin len = sc->sc_descm->rx_get_len(desc);
1254 1.11 martin
1255 1.11 martin #ifdef DWC_GMAC_DEBUG
1256 1.15 martin aprint_normal_dev(sc->sc_dev,
1257 1.15 martin "rx int: device is done with descriptor #%d, len: %d\n",
1258 1.15 martin i, len);
1259 1.11 martin #endif
1260 1.11 martin
1261 1.11 martin /*
1262 1.11 martin * Try to get a new mbuf before passing this one
1263 1.11 martin * up, if that fails, drop the packet and reuse
1264 1.11 martin * the existing one.
1265 1.11 martin */
1266 1.11 martin MGETHDR(mnew, M_DONTWAIT, MT_DATA);
1267 1.11 martin if (mnew == NULL) {
1268 1.69 thorpej if_statinc(ifp, if_ierrors);
1269 1.11 martin goto skip;
1270 1.11 martin }
1271 1.11 martin MCLGET(mnew, M_DONTWAIT);
1272 1.11 martin if ((mnew->m_flags & M_EXT) == 0) {
1273 1.11 martin m_freem(mnew);
1274 1.69 thorpej if_statinc(ifp, if_ierrors);
1275 1.11 martin goto skip;
1276 1.11 martin }
1277 1.66 tnn mnew->m_len = mnew->m_pkthdr.len = mnew->m_ext.ext_size;
1278 1.66 tnn if (mnew->m_len > AWGE_MAX_PACKET) {
1279 1.66 tnn mnew->m_len = mnew->m_pkthdr.len = AWGE_MAX_PACKET;
1280 1.66 tnn }
1281 1.11 martin
1282 1.11 martin /* unload old DMA map */
1283 1.11 martin bus_dmamap_sync(sc->sc_dmat, data->rd_map, 0,
1284 1.11 martin data->rd_map->dm_mapsize, BUS_DMASYNC_POSTREAD);
1285 1.11 martin bus_dmamap_unload(sc->sc_dmat, data->rd_map);
1286 1.11 martin
1287 1.11 martin /* and reload with new mbuf */
1288 1.66 tnn error = bus_dmamap_load_mbuf(sc->sc_dmat, data->rd_map,
1289 1.66 tnn mnew, BUS_DMA_READ | BUS_DMA_NOWAIT);
1290 1.11 martin if (error != 0) {
1291 1.11 martin m_freem(mnew);
1292 1.11 martin /* try to reload old mbuf */
1293 1.66 tnn error = bus_dmamap_load_mbuf(sc->sc_dmat, data->rd_map,
1294 1.66 tnn data->rd_m, BUS_DMA_READ | BUS_DMA_NOWAIT);
1295 1.11 martin if (error != 0) {
1296 1.11 martin panic("%s: could not load old rx mbuf",
1297 1.11 martin device_xname(sc->sc_dev));
1298 1.11 martin }
1299 1.69 thorpej if_statinc(ifp, if_ierrors);
1300 1.11 martin goto skip;
1301 1.11 martin }
1302 1.11 martin physaddr = data->rd_map->dm_segs[0].ds_addr;
1303 1.11 martin
1304 1.11 martin /*
1305 1.11 martin * New mbuf loaded, update RX ring and continue
1306 1.11 martin */
1307 1.11 martin m = data->rd_m;
1308 1.11 martin data->rd_m = mnew;
1309 1.11 martin desc->ddesc_data = htole32(physaddr);
1310 1.11 martin
1311 1.11 martin /* finalize mbuf */
1312 1.11 martin m->m_pkthdr.len = m->m_len = len;
1313 1.36 ozaki m_set_rcvif(m, ifp);
1314 1.19 matt m->m_flags |= M_HASFCS;
1315 1.11 martin
1316 1.39 skrll if_percpuq_enqueue(sc->sc_ipq, m);
1317 1.11 martin
1318 1.11 martin skip:
1319 1.27 matt bus_dmamap_sync(sc->sc_dmat, data->rd_map, 0,
1320 1.27 matt data->rd_map->dm_mapsize, BUS_DMASYNC_PREREAD);
1321 1.55 martin
1322 1.55 martin sc->sc_descm->rx_init_flags(desc);
1323 1.66 tnn sc->sc_descm->rx_set_len(desc, data->rd_m->m_len);
1324 1.55 martin sc->sc_descm->rx_set_owned_by_dev(desc);
1325 1.55 martin
1326 1.11 martin bus_dmamap_sync(sc->sc_dmat, sc->sc_dma_ring_map,
1327 1.11 martin RX_DESC_OFFSET(i), sizeof(*desc),
1328 1.61 msaitoh BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
1329 1.11 martin }
1330 1.11 martin
1331 1.11 martin /* update RX pointer */
1332 1.11 martin sc->sc_rxq.r_cur = i;
1333 1.11 martin
1334 1.38 skrll mutex_exit(&sc->sc_rxq.r_mtx);
1335 1.8 martin }
1336 1.8 martin
1337 1.22 martin /*
1338 1.24 skrll * Reverse order of bits - http://aggregate.org/MAGIC/#Bit%20Reversal
1339 1.22 martin */
1340 1.22 martin static uint32_t
1341 1.22 martin bitrev32(uint32_t x)
1342 1.22 martin {
1343 1.22 martin x = (((x & 0xaaaaaaaa) >> 1) | ((x & 0x55555555) << 1));
1344 1.22 martin x = (((x & 0xcccccccc) >> 2) | ((x & 0x33333333) << 2));
1345 1.22 martin x = (((x & 0xf0f0f0f0) >> 4) | ((x & 0x0f0f0f0f) << 4));
1346 1.22 martin x = (((x & 0xff00ff00) >> 8) | ((x & 0x00ff00ff) << 8));
1347 1.22 martin
1348 1.22 martin return (x >> 16) | (x << 16);
1349 1.22 martin }
1350 1.22 martin
1351 1.20 jmcneill static void
1352 1.20 jmcneill dwc_gmac_setmulti(struct dwc_gmac_softc *sc)
1353 1.20 jmcneill {
1354 1.20 jmcneill struct ifnet * const ifp = &sc->sc_ec.ec_if;
1355 1.20 jmcneill struct ether_multi *enm;
1356 1.20 jmcneill struct ether_multistep step;
1357 1.59 ozaki struct ethercom *ec = &sc->sc_ec;
1358 1.20 jmcneill uint32_t hashes[2] = { 0, 0 };
1359 1.22 martin uint32_t ffilt, h;
1360 1.38 skrll int mcnt;
1361 1.22 martin
1362 1.38 skrll KASSERT(mutex_owned(sc->sc_lock));
1363 1.20 jmcneill
1364 1.20 jmcneill ffilt = bus_space_read_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_MAC_FFILT);
1365 1.38 skrll
1366 1.20 jmcneill if (ifp->if_flags & IFF_PROMISC) {
1367 1.22 martin ffilt |= AWIN_GMAC_MAC_FFILT_PR;
1368 1.22 martin goto special_filter;
1369 1.20 jmcneill }
1370 1.20 jmcneill
1371 1.61 msaitoh ffilt &= ~(AWIN_GMAC_MAC_FFILT_PM | AWIN_GMAC_MAC_FFILT_PR);
1372 1.20 jmcneill
1373 1.20 jmcneill bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_MAC_HTLOW, 0);
1374 1.20 jmcneill bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_MAC_HTHIGH, 0);
1375 1.20 jmcneill
1376 1.59 ozaki ETHER_LOCK(ec);
1377 1.60 ozaki ec->ec_flags &= ~ETHER_F_ALLMULTI;
1378 1.59 ozaki ETHER_FIRST_MULTI(step, ec, enm);
1379 1.20 jmcneill mcnt = 0;
1380 1.20 jmcneill while (enm != NULL) {
1381 1.20 jmcneill if (memcmp(enm->enm_addrlo, enm->enm_addrhi,
1382 1.22 martin ETHER_ADDR_LEN) != 0) {
1383 1.60 ozaki ffilt |= AWIN_GMAC_MAC_FFILT_PM;
1384 1.60 ozaki ec->ec_flags |= ETHER_F_ALLMULTI;
1385 1.59 ozaki ETHER_UNLOCK(ec);
1386 1.22 martin goto special_filter;
1387 1.22 martin }
1388 1.20 jmcneill
1389 1.22 martin h = bitrev32(
1390 1.22 martin ~ether_crc32_le(enm->enm_addrlo, ETHER_ADDR_LEN)
1391 1.22 martin ) >> 26;
1392 1.20 jmcneill hashes[h >> 5] |= (1 << (h & 0x1f));
1393 1.20 jmcneill
1394 1.20 jmcneill mcnt++;
1395 1.20 jmcneill ETHER_NEXT_MULTI(step, enm);
1396 1.20 jmcneill }
1397 1.59 ozaki ETHER_UNLOCK(ec);
1398 1.20 jmcneill
1399 1.20 jmcneill if (mcnt)
1400 1.20 jmcneill ffilt |= AWIN_GMAC_MAC_FFILT_HMC;
1401 1.20 jmcneill else
1402 1.20 jmcneill ffilt &= ~AWIN_GMAC_MAC_FFILT_HMC;
1403 1.20 jmcneill
1404 1.20 jmcneill bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_MAC_FFILT, ffilt);
1405 1.20 jmcneill bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_MAC_HTLOW,
1406 1.20 jmcneill hashes[0]);
1407 1.20 jmcneill bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_MAC_HTHIGH,
1408 1.20 jmcneill hashes[1]);
1409 1.60 ozaki sc->sc_if_flags = ifp->if_flags;
1410 1.22 martin
1411 1.22 martin #ifdef DWC_GMAC_DEBUG
1412 1.22 martin dwc_gmac_dump_ffilt(sc, ffilt);
1413 1.22 martin #endif
1414 1.22 martin return;
1415 1.22 martin
1416 1.22 martin special_filter:
1417 1.22 martin #ifdef DWC_GMAC_DEBUG
1418 1.22 martin dwc_gmac_dump_ffilt(sc, ffilt);
1419 1.22 martin #endif
1420 1.22 martin /* no MAC hashes, ALLMULTI or PROMISC */
1421 1.22 martin bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_MAC_FFILT,
1422 1.22 martin ffilt);
1423 1.22 martin bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_MAC_HTLOW,
1424 1.22 martin 0xffffffff);
1425 1.22 martin bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_MAC_HTHIGH,
1426 1.22 martin 0xffffffff);
1427 1.22 martin sc->sc_if_flags = sc->sc_ec.ec_if.if_flags;
1428 1.20 jmcneill }
1429 1.20 jmcneill
1430 1.1 martin int
1431 1.1 martin dwc_gmac_intr(struct dwc_gmac_softc *sc)
1432 1.1 martin {
1433 1.1 martin uint32_t status, dma_status;
1434 1.8 martin int rv = 0;
1435 1.1 martin
1436 1.38 skrll if (sc->sc_stopping)
1437 1.38 skrll return 0;
1438 1.38 skrll
1439 1.1 martin status = bus_space_read_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_MAC_INTR);
1440 1.2 martin if (status & AWIN_GMAC_MII_IRQ) {
1441 1.1 martin (void)bus_space_read_4(sc->sc_bst, sc->sc_bsh,
1442 1.1 martin AWIN_GMAC_MII_STATUS);
1443 1.8 martin rv = 1;
1444 1.2 martin mii_pollstat(&sc->sc_mii);
1445 1.2 martin }
1446 1.1 martin
1447 1.1 martin dma_status = bus_space_read_4(sc->sc_bst, sc->sc_bsh,
1448 1.1 martin AWIN_GMAC_DMA_STATUS);
1449 1.1 martin
1450 1.61 msaitoh if (dma_status & (GMAC_DMA_INT_NIE | GMAC_DMA_INT_AIE))
1451 1.8 martin rv = 1;
1452 1.1 martin
1453 1.8 martin if (dma_status & GMAC_DMA_INT_TIE)
1454 1.8 martin dwc_gmac_tx_intr(sc);
1455 1.1 martin
1456 1.8 martin if (dma_status & GMAC_DMA_INT_RIE)
1457 1.8 martin dwc_gmac_rx_intr(sc);
1458 1.8 martin
1459 1.8 martin /*
1460 1.8 martin * Check error conditions
1461 1.8 martin */
1462 1.8 martin if (dma_status & GMAC_DMA_INT_ERRORS) {
1463 1.69 thorpej if_statinc(&sc->sc_ec.ec_if, if_oerrors);
1464 1.8 martin #ifdef DWC_GMAC_DEBUG
1465 1.8 martin dwc_dump_and_abort(sc, "interrupt error condition");
1466 1.8 martin #endif
1467 1.8 martin }
1468 1.8 martin
1469 1.63 msaitoh rnd_add_uint32(&sc->rnd_source, dma_status);
1470 1.63 msaitoh
1471 1.8 martin /* ack interrupt */
1472 1.8 martin if (dma_status)
1473 1.8 martin bus_space_write_4(sc->sc_bst, sc->sc_bsh,
1474 1.8 martin AWIN_GMAC_DMA_STATUS, dma_status & GMAC_DMA_INT_MASK);
1475 1.8 martin
1476 1.28 martin /*
1477 1.28 martin * Get more packets
1478 1.28 martin */
1479 1.28 martin if (rv)
1480 1.40 ozaki if_schedule_deferred_start(&sc->sc_ec.ec_if);
1481 1.28 martin
1482 1.8 martin return rv;
1483 1.1 martin }
1484 1.7 martin
1485 1.55 martin static void
1486 1.55 martin dwc_gmac_desc_set_owned_by_dev(struct dwc_gmac_dev_dmadesc *desc)
1487 1.55 martin {
1488 1.55 martin
1489 1.55 martin desc->ddesc_status0 |= htole32(DDESC_STATUS_OWNEDBYDEV);
1490 1.55 martin }
1491 1.55 martin
1492 1.55 martin static int
1493 1.55 martin dwc_gmac_desc_is_owned_by_dev(struct dwc_gmac_dev_dmadesc *desc)
1494 1.55 martin {
1495 1.55 martin
1496 1.55 martin return !!(le32toh(desc->ddesc_status0) & DDESC_STATUS_OWNEDBYDEV);
1497 1.55 martin }
1498 1.55 martin
1499 1.55 martin static void
1500 1.55 martin dwc_gmac_desc_std_set_len(struct dwc_gmac_dev_dmadesc *desc, int len)
1501 1.55 martin {
1502 1.55 martin uint32_t cntl = le32toh(desc->ddesc_cntl1);
1503 1.55 martin
1504 1.55 martin desc->ddesc_cntl1 = htole32((cntl & ~DDESC_CNTL_SIZE1MASK) |
1505 1.55 martin __SHIFTIN(len, DDESC_CNTL_SIZE1MASK));
1506 1.55 martin }
1507 1.55 martin
1508 1.55 martin static uint32_t
1509 1.55 martin dwc_gmac_desc_std_get_len(struct dwc_gmac_dev_dmadesc *desc)
1510 1.55 martin {
1511 1.55 martin
1512 1.55 martin return __SHIFTOUT(le32toh(desc->ddesc_status0), DDESC_STATUS_FRMLENMSK);
1513 1.55 martin }
1514 1.55 martin
1515 1.55 martin static void
1516 1.55 martin dwc_gmac_desc_std_tx_init_flags(struct dwc_gmac_dev_dmadesc *desc)
1517 1.55 martin {
1518 1.55 martin
1519 1.55 martin desc->ddesc_status0 = 0;
1520 1.55 martin desc->ddesc_cntl1 = htole32(DDESC_CNTL_TXCHAIN);
1521 1.55 martin }
1522 1.55 martin
1523 1.55 martin static void
1524 1.55 martin dwc_gmac_desc_std_tx_set_first_frag(struct dwc_gmac_dev_dmadesc *desc)
1525 1.55 martin {
1526 1.55 martin uint32_t cntl = le32toh(desc->ddesc_cntl1);
1527 1.55 martin
1528 1.55 martin desc->ddesc_cntl1 = htole32(cntl | DDESC_CNTL_TXFIRST);
1529 1.55 martin }
1530 1.55 martin
1531 1.55 martin static void
1532 1.55 martin dwc_gmac_desc_std_tx_set_last_frag(struct dwc_gmac_dev_dmadesc *desc)
1533 1.55 martin {
1534 1.55 martin uint32_t cntl = le32toh(desc->ddesc_cntl1);
1535 1.55 martin
1536 1.55 martin desc->ddesc_cntl1 = htole32(cntl |
1537 1.55 martin DDESC_CNTL_TXLAST | DDESC_CNTL_TXINT);
1538 1.55 martin }
1539 1.55 martin
1540 1.55 martin static void
1541 1.55 martin dwc_gmac_desc_std_rx_init_flags(struct dwc_gmac_dev_dmadesc *desc)
1542 1.55 martin {
1543 1.55 martin
1544 1.55 martin desc->ddesc_status0 = 0;
1545 1.55 martin desc->ddesc_cntl1 = htole32(DDESC_CNTL_TXCHAIN);
1546 1.55 martin }
1547 1.55 martin
1548 1.55 martin static int
1549 1.55 martin dwc_gmac_desc_std_rx_has_error(struct dwc_gmac_dev_dmadesc *desc) {
1550 1.55 martin return !!(le32toh(desc->ddesc_status0) &
1551 1.55 martin (DDESC_STATUS_RXERROR | DDESC_STATUS_RXTRUNCATED));
1552 1.55 martin }
1553 1.55 martin
1554 1.55 martin static void
1555 1.55 martin dwc_gmac_desc_enh_set_len(struct dwc_gmac_dev_dmadesc *desc, int len)
1556 1.55 martin {
1557 1.55 martin uint32_t tdes1 = le32toh(desc->ddesc_cntl1);
1558 1.55 martin
1559 1.55 martin desc->ddesc_cntl1 = htole32((tdes1 & ~DDESC_DES1_SIZE1MASK) |
1560 1.55 martin __SHIFTIN(len, DDESC_DES1_SIZE1MASK));
1561 1.55 martin }
1562 1.55 martin
1563 1.55 martin static uint32_t
1564 1.55 martin dwc_gmac_desc_enh_get_len(struct dwc_gmac_dev_dmadesc *desc)
1565 1.55 martin {
1566 1.55 martin
1567 1.55 martin return __SHIFTOUT(le32toh(desc->ddesc_status0), DDESC_RDES0_FL);
1568 1.55 martin }
1569 1.55 martin
1570 1.55 martin static void
1571 1.55 martin dwc_gmac_desc_enh_tx_init_flags(struct dwc_gmac_dev_dmadesc *desc)
1572 1.55 martin {
1573 1.55 martin
1574 1.55 martin desc->ddesc_status0 = htole32(DDESC_TDES0_TCH);
1575 1.55 martin desc->ddesc_cntl1 = 0;
1576 1.55 martin }
1577 1.55 martin
1578 1.55 martin static void
1579 1.55 martin dwc_gmac_desc_enh_tx_set_first_frag(struct dwc_gmac_dev_dmadesc *desc)
1580 1.55 martin {
1581 1.55 martin uint32_t tdes0 = le32toh(desc->ddesc_status0);
1582 1.55 martin
1583 1.55 martin desc->ddesc_status0 = htole32(tdes0 | DDESC_TDES0_FS);
1584 1.55 martin }
1585 1.55 martin
1586 1.55 martin static void
1587 1.55 martin dwc_gmac_desc_enh_tx_set_last_frag(struct dwc_gmac_dev_dmadesc *desc)
1588 1.55 martin {
1589 1.55 martin uint32_t tdes0 = le32toh(desc->ddesc_status0);
1590 1.55 martin
1591 1.55 martin desc->ddesc_status0 = htole32(tdes0 | DDESC_TDES0_LS | DDESC_TDES0_IC);
1592 1.55 martin }
1593 1.55 martin
1594 1.55 martin static void
1595 1.55 martin dwc_gmac_desc_enh_rx_init_flags(struct dwc_gmac_dev_dmadesc *desc)
1596 1.55 martin {
1597 1.55 martin
1598 1.55 martin desc->ddesc_status0 = 0;
1599 1.55 martin desc->ddesc_cntl1 = htole32(DDESC_RDES1_RCH);
1600 1.55 martin }
1601 1.55 martin
1602 1.55 martin static int
1603 1.55 martin dwc_gmac_desc_enh_rx_has_error(struct dwc_gmac_dev_dmadesc *desc)
1604 1.55 martin {
1605 1.55 martin
1606 1.55 martin return !!(le32toh(desc->ddesc_status0) &
1607 1.55 martin (DDESC_RDES0_ES | DDESC_RDES0_LE));
1608 1.55 martin }
1609 1.55 martin
1610 1.7 martin #ifdef DWC_GMAC_DEBUG
1611 1.7 martin static void
1612 1.7 martin dwc_gmac_dump_dma(struct dwc_gmac_softc *sc)
1613 1.7 martin {
1614 1.7 martin aprint_normal_dev(sc->sc_dev, "busmode: %08x\n",
1615 1.7 martin bus_space_read_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_DMA_BUSMODE));
1616 1.7 martin aprint_normal_dev(sc->sc_dev, "tx poll: %08x\n",
1617 1.7 martin bus_space_read_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_DMA_TXPOLL));
1618 1.7 martin aprint_normal_dev(sc->sc_dev, "rx poll: %08x\n",
1619 1.7 martin bus_space_read_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_DMA_RXPOLL));
1620 1.7 martin aprint_normal_dev(sc->sc_dev, "rx descriptors: %08x\n",
1621 1.7 martin bus_space_read_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_DMA_RX_ADDR));
1622 1.7 martin aprint_normal_dev(sc->sc_dev, "tx descriptors: %08x\n",
1623 1.7 martin bus_space_read_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_DMA_TX_ADDR));
1624 1.7 martin aprint_normal_dev(sc->sc_dev, "status: %08x\n",
1625 1.7 martin bus_space_read_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_DMA_STATUS));
1626 1.7 martin aprint_normal_dev(sc->sc_dev, "op mode: %08x\n",
1627 1.7 martin bus_space_read_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_DMA_OPMODE));
1628 1.7 martin aprint_normal_dev(sc->sc_dev, "int enable: %08x\n",
1629 1.7 martin bus_space_read_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_DMA_INTENABLE));
1630 1.7 martin aprint_normal_dev(sc->sc_dev, "cur tx: %08x\n",
1631 1.7 martin bus_space_read_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_DMA_CUR_TX_DESC));
1632 1.7 martin aprint_normal_dev(sc->sc_dev, "cur rx: %08x\n",
1633 1.7 martin bus_space_read_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_DMA_CUR_RX_DESC));
1634 1.7 martin aprint_normal_dev(sc->sc_dev, "cur tx buffer: %08x\n",
1635 1.7 martin bus_space_read_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_DMA_CUR_TX_BUFADDR));
1636 1.7 martin aprint_normal_dev(sc->sc_dev, "cur rx buffer: %08x\n",
1637 1.7 martin bus_space_read_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_DMA_CUR_RX_BUFADDR));
1638 1.7 martin }
1639 1.7 martin
1640 1.7 martin static void
1641 1.7 martin dwc_gmac_dump_tx_desc(struct dwc_gmac_softc *sc)
1642 1.7 martin {
1643 1.7 martin int i;
1644 1.7 martin
1645 1.8 martin aprint_normal_dev(sc->sc_dev, "TX queue: cur=%d, next=%d, queued=%d\n",
1646 1.8 martin sc->sc_txq.t_cur, sc->sc_txq.t_next, sc->sc_txq.t_queued);
1647 1.8 martin aprint_normal_dev(sc->sc_dev, "TX DMA descriptors:\n");
1648 1.7 martin for (i = 0; i < AWGE_TX_RING_COUNT; i++) {
1649 1.7 martin struct dwc_gmac_dev_dmadesc *desc = &sc->sc_txq.t_desc[i];
1650 1.15 martin aprint_normal("#%d (%08lx): status: %08x cntl: %08x "
1651 1.15 martin "data: %08x next: %08x\n",
1652 1.15 martin i, sc->sc_txq.t_physaddr +
1653 1.15 martin i*sizeof(struct dwc_gmac_dev_dmadesc),
1654 1.55 martin le32toh(desc->ddesc_status0), le32toh(desc->ddesc_cntl1),
1655 1.7 martin le32toh(desc->ddesc_data), le32toh(desc->ddesc_next));
1656 1.7 martin }
1657 1.7 martin }
1658 1.8 martin
1659 1.8 martin static void
1660 1.11 martin dwc_gmac_dump_rx_desc(struct dwc_gmac_softc *sc)
1661 1.11 martin {
1662 1.11 martin int i;
1663 1.11 martin
1664 1.11 martin aprint_normal_dev(sc->sc_dev, "RX queue: cur=%d, next=%d\n",
1665 1.11 martin sc->sc_rxq.r_cur, sc->sc_rxq.r_next);
1666 1.11 martin aprint_normal_dev(sc->sc_dev, "RX DMA descriptors:\n");
1667 1.11 martin for (i = 0; i < AWGE_RX_RING_COUNT; i++) {
1668 1.11 martin struct dwc_gmac_dev_dmadesc *desc = &sc->sc_rxq.r_desc[i];
1669 1.15 martin aprint_normal("#%d (%08lx): status: %08x cntl: %08x "
1670 1.15 martin "data: %08x next: %08x\n",
1671 1.15 martin i, sc->sc_rxq.r_physaddr +
1672 1.15 martin i*sizeof(struct dwc_gmac_dev_dmadesc),
1673 1.55 martin le32toh(desc->ddesc_status0), le32toh(desc->ddesc_cntl1),
1674 1.11 martin le32toh(desc->ddesc_data), le32toh(desc->ddesc_next));
1675 1.11 martin }
1676 1.11 martin }
1677 1.11 martin
1678 1.11 martin static void
1679 1.10 martin dwc_dump_status(struct dwc_gmac_softc *sc)
1680 1.8 martin {
1681 1.8 martin uint32_t status = bus_space_read_4(sc->sc_bst, sc->sc_bsh,
1682 1.8 martin AWIN_GMAC_MAC_INTR);
1683 1.8 martin uint32_t dma_status = bus_space_read_4(sc->sc_bst, sc->sc_bsh,
1684 1.8 martin AWIN_GMAC_DMA_STATUS);
1685 1.8 martin char buf[200];
1686 1.8 martin
1687 1.8 martin /* print interrupt state */
1688 1.8 martin snprintb(buf, sizeof(buf), "\177\20"
1689 1.10 martin "b\x10""NI\0"
1690 1.10 martin "b\x0f""AI\0"
1691 1.10 martin "b\x0e""ER\0"
1692 1.10 martin "b\x0d""FB\0"
1693 1.10 martin "b\x0a""ET\0"
1694 1.10 martin "b\x09""RW\0"
1695 1.10 martin "b\x08""RS\0"
1696 1.10 martin "b\x07""RU\0"
1697 1.10 martin "b\x06""RI\0"
1698 1.10 martin "b\x05""UN\0"
1699 1.10 martin "b\x04""OV\0"
1700 1.10 martin "b\x03""TJ\0"
1701 1.10 martin "b\x02""TU\0"
1702 1.10 martin "b\x01""TS\0"
1703 1.10 martin "b\x00""TI\0"
1704 1.8 martin "\0", dma_status);
1705 1.10 martin aprint_normal_dev(sc->sc_dev, "INTR status: %08x, DMA status: %s\n",
1706 1.8 martin status, buf);
1707 1.10 martin }
1708 1.8 martin
1709 1.10 martin static void
1710 1.10 martin dwc_dump_and_abort(struct dwc_gmac_softc *sc, const char *msg)
1711 1.10 martin {
1712 1.10 martin dwc_dump_status(sc);
1713 1.22 martin dwc_gmac_dump_ffilt(sc,
1714 1.22 martin bus_space_read_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_MAC_FFILT));
1715 1.8 martin dwc_gmac_dump_dma(sc);
1716 1.8 martin dwc_gmac_dump_tx_desc(sc);
1717 1.11 martin dwc_gmac_dump_rx_desc(sc);
1718 1.8 martin
1719 1.21 joerg panic("%s", msg);
1720 1.8 martin }
1721 1.22 martin
1722 1.22 martin static void dwc_gmac_dump_ffilt(struct dwc_gmac_softc *sc, uint32_t ffilt)
1723 1.22 martin {
1724 1.22 martin char buf[200];
1725 1.22 martin
1726 1.22 martin /* print filter setup */
1727 1.22 martin snprintb(buf, sizeof(buf), "\177\20"
1728 1.22 martin "b\x1f""RA\0"
1729 1.22 martin "b\x0a""HPF\0"
1730 1.22 martin "b\x09""SAF\0"
1731 1.22 martin "b\x08""SAIF\0"
1732 1.22 martin "b\x05""DBF\0"
1733 1.22 martin "b\x04""PM\0"
1734 1.22 martin "b\x03""DAIF\0"
1735 1.22 martin "b\x02""HMC\0"
1736 1.22 martin "b\x01""HUC\0"
1737 1.22 martin "b\x00""PR\0"
1738 1.22 martin "\0", ffilt);
1739 1.22 martin aprint_normal_dev(sc->sc_dev, "FFILT: %s\n", buf);
1740 1.22 martin }
1741 1.7 martin #endif
1742