ixv.c revision 1.191 1 1.191 msaitoh /* $NetBSD: ixv.c,v 1.191 2023/10/12 08:06:13 msaitoh Exp $ */
2 1.58 msaitoh
3 1.1 dyoung /******************************************************************************
4 1.1 dyoung
5 1.58 msaitoh Copyright (c) 2001-2017, Intel Corporation
6 1.1 dyoung All rights reserved.
7 1.58 msaitoh
8 1.58 msaitoh Redistribution and use in source and binary forms, with or without
9 1.1 dyoung modification, are permitted provided that the following conditions are met:
10 1.58 msaitoh
11 1.58 msaitoh 1. Redistributions of source code must retain the above copyright notice,
12 1.1 dyoung this list of conditions and the following disclaimer.
13 1.58 msaitoh
14 1.58 msaitoh 2. Redistributions in binary form must reproduce the above copyright
15 1.58 msaitoh notice, this list of conditions and the following disclaimer in the
16 1.1 dyoung documentation and/or other materials provided with the distribution.
17 1.58 msaitoh
18 1.58 msaitoh 3. Neither the name of the Intel Corporation nor the names of its
19 1.58 msaitoh contributors may be used to endorse or promote products derived from
20 1.1 dyoung this software without specific prior written permission.
21 1.58 msaitoh
22 1.1 dyoung THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
23 1.58 msaitoh AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
24 1.58 msaitoh IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
25 1.58 msaitoh ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
26 1.58 msaitoh LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
27 1.58 msaitoh CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
28 1.58 msaitoh SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
29 1.58 msaitoh INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
30 1.58 msaitoh CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
31 1.1 dyoung ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
32 1.1 dyoung POSSIBILITY OF SUCH DAMAGE.
33 1.1 dyoung
34 1.1 dyoung ******************************************************************************/
35 1.92 msaitoh /*$FreeBSD: head/sys/dev/ixgbe/if_ixv.c 331224 2018-03-19 20:55:05Z erj $*/
36 1.1 dyoung
37 1.159 msaitoh #include <sys/cdefs.h>
38 1.191 msaitoh __KERNEL_RCSID(0, "$NetBSD: ixv.c,v 1.191 2023/10/12 08:06:13 msaitoh Exp $");
39 1.159 msaitoh
40 1.55 msaitoh #ifdef _KERNEL_OPT
41 1.1 dyoung #include "opt_inet.h"
42 1.4 msaitoh #include "opt_inet6.h"
43 1.55 msaitoh #include "opt_net_mpsafe.h"
44 1.55 msaitoh #endif
45 1.1 dyoung
46 1.21 msaitoh #include "ixgbe.h"
47 1.1 dyoung
48 1.58 msaitoh /************************************************************************
49 1.58 msaitoh * Driver version
50 1.58 msaitoh ************************************************************************/
51 1.103 maxv static const char ixv_driver_version[] = "2.0.1-k";
52 1.117 msaitoh /* XXX NetBSD: + 1.5.17 */
53 1.58 msaitoh
54 1.58 msaitoh /************************************************************************
55 1.58 msaitoh * PCI Device ID Table
56 1.58 msaitoh *
57 1.58 msaitoh * Used by probe to select devices to load on
58 1.58 msaitoh * Last field stores an index into ixv_strings
59 1.58 msaitoh * Last entry must be all 0s
60 1.1 dyoung *
61 1.58 msaitoh * { Vendor ID, Device ID, SubVendor ID, SubDevice ID, String Index }
62 1.58 msaitoh ************************************************************************/
63 1.103 maxv static const ixgbe_vendor_info_t ixv_vendor_info_array[] =
64 1.1 dyoung {
65 1.1 dyoung {IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_82599_VF, 0, 0, 0},
66 1.5 msaitoh {IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_X540_VF, 0, 0, 0},
67 1.21 msaitoh {IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_X550_VF, 0, 0, 0},
68 1.21 msaitoh {IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_X550EM_X_VF, 0, 0, 0},
69 1.58 msaitoh {IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_X550EM_A_VF, 0, 0, 0},
70 1.1 dyoung /* required last entry */
71 1.1 dyoung {0, 0, 0, 0, 0}
72 1.1 dyoung };
73 1.1 dyoung
74 1.58 msaitoh /************************************************************************
75 1.58 msaitoh * Table of branding strings
76 1.58 msaitoh ************************************************************************/
77 1.58 msaitoh static const char *ixv_strings[] = {
78 1.1 dyoung "Intel(R) PRO/10GbE Virtual Function Network Driver"
79 1.1 dyoung };
80 1.1 dyoung
81 1.1 dyoung /*********************************************************************
82 1.1 dyoung * Function prototypes
83 1.1 dyoung *********************************************************************/
84 1.114 msaitoh static int ixv_probe(device_t, cfdata_t, void *);
85 1.22 msaitoh static void ixv_attach(device_t, device_t, void *);
86 1.114 msaitoh static int ixv_detach(device_t, int);
87 1.3 msaitoh #if 0
88 1.114 msaitoh static int ixv_shutdown(device_t);
89 1.3 msaitoh #endif
90 1.57 msaitoh static int ixv_ifflags_cb(struct ethercom *);
91 1.114 msaitoh static int ixv_ioctl(struct ifnet *, u_long, void *);
92 1.3 msaitoh static int ixv_init(struct ifnet *);
93 1.186 msaitoh static void ixv_init_locked(struct ixgbe_softc *);
94 1.56 msaitoh static void ixv_ifstop(struct ifnet *, int);
95 1.153 msaitoh static void ixv_stop_locked(void *);
96 1.186 msaitoh static void ixv_init_device_features(struct ixgbe_softc *);
97 1.114 msaitoh static void ixv_media_status(struct ifnet *, struct ifmediareq *);
98 1.114 msaitoh static int ixv_media_change(struct ifnet *);
99 1.186 msaitoh static int ixv_allocate_pci_resources(struct ixgbe_softc *,
100 1.3 msaitoh const struct pci_attach_args *);
101 1.186 msaitoh static void ixv_free_deferred_handlers(struct ixgbe_softc *);
102 1.186 msaitoh static int ixv_allocate_msix(struct ixgbe_softc *,
103 1.11 msaitoh const struct pci_attach_args *);
104 1.186 msaitoh static int ixv_configure_interrupts(struct ixgbe_softc *);
105 1.186 msaitoh static void ixv_free_pci_resources(struct ixgbe_softc *);
106 1.114 msaitoh static void ixv_local_timer(void *);
107 1.151 msaitoh static void ixv_handle_timer(struct work *, void *);
108 1.186 msaitoh static int ixv_setup_interface(device_t, struct ixgbe_softc *);
109 1.186 msaitoh static void ixv_schedule_admin_tasklet(struct ixgbe_softc *);
110 1.186 msaitoh static int ixv_negotiate_api(struct ixgbe_softc *);
111 1.186 msaitoh
112 1.186 msaitoh static void ixv_initialize_transmit_units(struct ixgbe_softc *);
113 1.186 msaitoh static void ixv_initialize_receive_units(struct ixgbe_softc *);
114 1.186 msaitoh static void ixv_initialize_rss_mapping(struct ixgbe_softc *);
115 1.186 msaitoh static s32 ixv_check_link(struct ixgbe_softc *);
116 1.186 msaitoh
117 1.186 msaitoh static void ixv_enable_intr(struct ixgbe_softc *);
118 1.186 msaitoh static void ixv_disable_intr(struct ixgbe_softc *);
119 1.186 msaitoh static int ixv_set_rxfilter(struct ixgbe_softc *);
120 1.186 msaitoh static void ixv_update_link_status(struct ixgbe_softc *);
121 1.3 msaitoh static int ixv_sysctl_debug(SYSCTLFN_PROTO);
122 1.186 msaitoh static void ixv_set_ivar(struct ixgbe_softc *, u8, u8, s8);
123 1.186 msaitoh static void ixv_configure_ivars(struct ixgbe_softc *);
124 1.1 dyoung static u8 * ixv_mc_array_itr(struct ixgbe_hw *, u8 **, u32 *);
125 1.186 msaitoh static void ixv_eitr_write(struct ixgbe_softc *, uint32_t, uint32_t);
126 1.1 dyoung
127 1.186 msaitoh static void ixv_setup_vlan_tagging(struct ixgbe_softc *);
128 1.186 msaitoh static int ixv_setup_vlan_support(struct ixgbe_softc *);
129 1.120 msaitoh static int ixv_vlan_cb(struct ethercom *, uint16_t, bool);
130 1.186 msaitoh static int ixv_register_vlan(struct ixgbe_softc *, u16);
131 1.186 msaitoh static int ixv_unregister_vlan(struct ixgbe_softc *, u16);
132 1.1 dyoung
133 1.186 msaitoh static void ixv_add_device_sysctls(struct ixgbe_softc *);
134 1.186 msaitoh static void ixv_init_stats(struct ixgbe_softc *);
135 1.186 msaitoh static void ixv_update_stats(struct ixgbe_softc *);
136 1.186 msaitoh static void ixv_add_stats_sysctls(struct ixgbe_softc *);
137 1.186 msaitoh static void ixv_clear_evcnt(struct ixgbe_softc *);
138 1.83 msaitoh
139 1.83 msaitoh /* Sysctl handlers */
140 1.114 msaitoh static int ixv_sysctl_interrupt_rate_handler(SYSCTLFN_PROTO);
141 1.114 msaitoh static int ixv_sysctl_next_to_check_handler(SYSCTLFN_PROTO);
142 1.164 msaitoh static int ixv_sysctl_next_to_refresh_handler(SYSCTLFN_PROTO);
143 1.114 msaitoh static int ixv_sysctl_rdh_handler(SYSCTLFN_PROTO);
144 1.114 msaitoh static int ixv_sysctl_rdt_handler(SYSCTLFN_PROTO);
145 1.114 msaitoh static int ixv_sysctl_tdt_handler(SYSCTLFN_PROTO);
146 1.114 msaitoh static int ixv_sysctl_tdh_handler(SYSCTLFN_PROTO);
147 1.179 msaitoh static int ixv_sysctl_tx_process_limit(SYSCTLFN_PROTO);
148 1.179 msaitoh static int ixv_sysctl_rx_process_limit(SYSCTLFN_PROTO);
149 1.163 msaitoh static int ixv_sysctl_rx_copy_len(SYSCTLFN_PROTO);
150 1.1 dyoung
151 1.58 msaitoh /* The MSI-X Interrupt handlers */
152 1.11 msaitoh static int ixv_msix_que(void *);
153 1.11 msaitoh static int ixv_msix_mbx(void *);
154 1.1 dyoung
155 1.151 msaitoh /* Event handlers running on workqueue */
156 1.3 msaitoh static void ixv_handle_que(void *);
157 1.3 msaitoh
158 1.151 msaitoh /* Deferred workqueue handlers */
159 1.151 msaitoh static void ixv_handle_admin(struct work *, void *);
160 1.84 knakahar static void ixv_handle_que_work(struct work *, void *);
161 1.84 knakahar
162 1.186 msaitoh const struct sysctlnode *ixv_sysctl_instance(struct ixgbe_softc *);
163 1.103 maxv static const ixgbe_vendor_info_t *ixv_lookup(const struct pci_attach_args *);
164 1.1 dyoung
165 1.58 msaitoh /************************************************************************
166 1.150 msaitoh * NetBSD Device Interface Entry Points
167 1.58 msaitoh ************************************************************************/
168 1.186 msaitoh CFATTACH_DECL3_NEW(ixv, sizeof(struct ixgbe_softc),
169 1.3 msaitoh ixv_probe, ixv_attach, ixv_detach, NULL, NULL, NULL,
170 1.3 msaitoh DVF_DETACH_SHUTDOWN);
171 1.3 msaitoh
172 1.1 dyoung #if 0
173 1.1 dyoung static driver_t ixv_driver = {
174 1.186 msaitoh "ixv", ixv_methods, sizeof(struct ixgbe_softc),
175 1.1 dyoung };
176 1.1 dyoung
177 1.22 msaitoh devclass_t ixv_devclass;
178 1.22 msaitoh DRIVER_MODULE(ixv, pci, ixv_driver, ixv_devclass, 0, 0);
179 1.1 dyoung MODULE_DEPEND(ixv, pci, 1, 1, 1);
180 1.1 dyoung MODULE_DEPEND(ixv, ether, 1, 1, 1);
181 1.1 dyoung #endif
182 1.1 dyoung
183 1.1 dyoung /*
184 1.58 msaitoh * TUNEABLE PARAMETERS:
185 1.58 msaitoh */
186 1.1 dyoung
187 1.58 msaitoh /* Number of Queues - do not exceed MSI-X vectors - 1 */
188 1.44 msaitoh static int ixv_num_queues = 0;
189 1.23 msaitoh #define TUNABLE_INT(__x, __y)
190 1.23 msaitoh TUNABLE_INT("hw.ixv.num_queues", &ixv_num_queues);
191 1.23 msaitoh
192 1.1 dyoung /*
193 1.58 msaitoh * AIM: Adaptive Interrupt Moderation
194 1.58 msaitoh * which means that the interrupt rate
195 1.58 msaitoh * is varied over time based on the
196 1.58 msaitoh * traffic for that interrupt vector
197 1.58 msaitoh */
198 1.50 msaitoh static bool ixv_enable_aim = false;
199 1.1 dyoung TUNABLE_INT("hw.ixv.enable_aim", &ixv_enable_aim);
200 1.1 dyoung
201 1.83 msaitoh static int ixv_max_interrupt_rate = (4000000 / IXGBE_LOW_LATENCY);
202 1.83 msaitoh TUNABLE_INT("hw.ixv.max_interrupt_rate", &ixv_max_interrupt_rate);
203 1.83 msaitoh
204 1.1 dyoung /* How many packets rxeof tries to clean at a time */
205 1.21 msaitoh static int ixv_rx_process_limit = 256;
206 1.1 dyoung TUNABLE_INT("hw.ixv.rx_process_limit", &ixv_rx_process_limit);
207 1.1 dyoung
208 1.21 msaitoh /* How many packets txeof tries to clean at a time */
209 1.21 msaitoh static int ixv_tx_process_limit = 256;
210 1.21 msaitoh TUNABLE_INT("hw.ixv.tx_process_limit", &ixv_tx_process_limit);
211 1.1 dyoung
212 1.112 msaitoh /* Which packet processing uses workqueue or softint */
213 1.84 knakahar static bool ixv_txrx_workqueue = false;
214 1.84 knakahar
215 1.1 dyoung /*
216 1.58 msaitoh * Number of TX descriptors per ring,
217 1.58 msaitoh * setting higher than RX as this seems
218 1.58 msaitoh * the better performing choice.
219 1.58 msaitoh */
220 1.187 msaitoh static int ixv_txd = DEFAULT_TXD;
221 1.1 dyoung TUNABLE_INT("hw.ixv.txd", &ixv_txd);
222 1.1 dyoung
223 1.1 dyoung /* Number of RX descriptors per ring */
224 1.187 msaitoh static int ixv_rxd = DEFAULT_RXD;
225 1.1 dyoung TUNABLE_INT("hw.ixv.rxd", &ixv_rxd);
226 1.1 dyoung
227 1.58 msaitoh /* Legacy Transmit (single queue) */
228 1.58 msaitoh static int ixv_enable_legacy_tx = 0;
229 1.58 msaitoh TUNABLE_INT("hw.ixv.enable_legacy_tx", &ixv_enable_legacy_tx);
230 1.58 msaitoh
231 1.55 msaitoh #ifdef NET_MPSAFE
232 1.55 msaitoh #define IXGBE_MPSAFE 1
233 1.55 msaitoh #define IXGBE_CALLOUT_FLAGS CALLOUT_MPSAFE
234 1.149 msaitoh #define IXGBE_SOFTINT_FLAGS SOFTINT_MPSAFE
235 1.84 knakahar #define IXGBE_WORKQUEUE_FLAGS WQ_PERCPU | WQ_MPSAFE
236 1.151 msaitoh #define IXGBE_TASKLET_WQ_FLAGS WQ_MPSAFE
237 1.55 msaitoh #else
238 1.55 msaitoh #define IXGBE_CALLOUT_FLAGS 0
239 1.149 msaitoh #define IXGBE_SOFTINT_FLAGS 0
240 1.84 knakahar #define IXGBE_WORKQUEUE_FLAGS WQ_PERCPU
241 1.151 msaitoh #define IXGBE_TASKLET_WQ_FLAGS 0
242 1.55 msaitoh #endif
243 1.84 knakahar #define IXGBE_WORKQUEUE_PRI PRI_SOFTNET
244 1.55 msaitoh
245 1.58 msaitoh #if 0
246 1.58 msaitoh static int (*ixv_start_locked)(struct ifnet *, struct tx_ring *);
247 1.58 msaitoh static int (*ixv_ring_empty)(struct ifnet *, struct buf_ring *);
248 1.58 msaitoh #endif
249 1.58 msaitoh
250 1.58 msaitoh /************************************************************************
251 1.58 msaitoh * ixv_probe - Device identification routine
252 1.1 dyoung *
253 1.58 msaitoh * Determines if the driver should be loaded on
254 1.58 msaitoh * adapter based on its PCI vendor/device ID.
255 1.1 dyoung *
256 1.58 msaitoh * return BUS_PROBE_DEFAULT on success, positive on failure
257 1.58 msaitoh ************************************************************************/
258 1.1 dyoung static int
259 1.3 msaitoh ixv_probe(device_t dev, cfdata_t cf, void *aux)
260 1.3 msaitoh {
261 1.19 knakahar #ifdef __HAVE_PCI_MSI_MSIX
262 1.3 msaitoh const struct pci_attach_args *pa = aux;
263 1.3 msaitoh
264 1.3 msaitoh return (ixv_lookup(pa) != NULL) ? 1 : 0;
265 1.18 msaitoh #else
266 1.18 msaitoh return 0;
267 1.18 msaitoh #endif
268 1.58 msaitoh } /* ixv_probe */
269 1.3 msaitoh
270 1.103 maxv static const ixgbe_vendor_info_t *
271 1.3 msaitoh ixv_lookup(const struct pci_attach_args *pa)
272 1.1 dyoung {
273 1.103 maxv const ixgbe_vendor_info_t *ent;
274 1.3 msaitoh pcireg_t subid;
275 1.1 dyoung
276 1.31 msaitoh INIT_DEBUGOUT("ixv_lookup: begin");
277 1.1 dyoung
278 1.3 msaitoh if (PCI_VENDOR(pa->pa_id) != IXGBE_INTEL_VENDOR_ID)
279 1.3 msaitoh return NULL;
280 1.1 dyoung
281 1.3 msaitoh subid = pci_conf_read(pa->pa_pc, pa->pa_tag, PCI_SUBSYS_ID_REG);
282 1.1 dyoung
283 1.3 msaitoh for (ent = ixv_vendor_info_array; ent->vendor_id != 0; ent++) {
284 1.3 msaitoh if ((PCI_VENDOR(pa->pa_id) == ent->vendor_id) &&
285 1.3 msaitoh (PCI_PRODUCT(pa->pa_id) == ent->device_id) &&
286 1.3 msaitoh ((PCI_SUBSYS_VENDOR(subid) == ent->subvendor_id) ||
287 1.1 dyoung (ent->subvendor_id == 0)) &&
288 1.3 msaitoh ((PCI_SUBSYS_ID(subid) == ent->subdevice_id) ||
289 1.1 dyoung (ent->subdevice_id == 0))) {
290 1.3 msaitoh return ent;
291 1.1 dyoung }
292 1.1 dyoung }
293 1.58 msaitoh
294 1.3 msaitoh return NULL;
295 1.3 msaitoh }
296 1.3 msaitoh
297 1.58 msaitoh /************************************************************************
298 1.58 msaitoh * ixv_attach - Device initialization routine
299 1.57 msaitoh *
300 1.58 msaitoh * Called when the driver is being loaded.
301 1.58 msaitoh * Identifies the type of hardware, allocates all resources
302 1.58 msaitoh * and initializes the hardware.
303 1.57 msaitoh *
304 1.58 msaitoh * return 0 on success, positive on failure
305 1.58 msaitoh ************************************************************************/
306 1.3 msaitoh static void
307 1.3 msaitoh ixv_attach(device_t parent, device_t dev, void *aux)
308 1.1 dyoung {
309 1.186 msaitoh struct ixgbe_softc *sc;
310 1.1 dyoung struct ixgbe_hw *hw;
311 1.114 msaitoh int error = 0;
312 1.58 msaitoh pcireg_t id, subid;
313 1.103 maxv const ixgbe_vendor_info_t *ent;
314 1.3 msaitoh const struct pci_attach_args *pa = aux;
315 1.60 msaitoh const char *apivstr;
316 1.66 msaitoh const char *str;
317 1.151 msaitoh char wqname[MAXCOMLEN];
318 1.63 msaitoh char buf[256];
319 1.63 msaitoh
320 1.1 dyoung INIT_DEBUGOUT("ixv_attach: begin");
321 1.1 dyoung
322 1.58 msaitoh /*
323 1.58 msaitoh * Make sure BUSMASTER is set, on a VM under
324 1.58 msaitoh * KVM it may not be and will break things.
325 1.58 msaitoh */
326 1.58 msaitoh ixgbe_pci_enable_busmaster(pa->pa_pc, pa->pa_tag);
327 1.58 msaitoh
328 1.1 dyoung /* Allocate, clear, and link in our adapter structure */
329 1.186 msaitoh sc = device_private(dev);
330 1.186 msaitoh sc->hw.back = sc;
331 1.186 msaitoh sc->dev = dev;
332 1.186 msaitoh hw = &sc->hw;
333 1.26 msaitoh
334 1.186 msaitoh sc->init_locked = ixv_init_locked;
335 1.186 msaitoh sc->stop_locked = ixv_stop_locked;
336 1.26 msaitoh
337 1.186 msaitoh sc->osdep.pc = pa->pa_pc;
338 1.186 msaitoh sc->osdep.tag = pa->pa_tag;
339 1.43 msaitoh if (pci_dma64_available(pa))
340 1.186 msaitoh sc->osdep.dmat = pa->pa_dmat64;
341 1.43 msaitoh else
342 1.186 msaitoh sc->osdep.dmat = pa->pa_dmat;
343 1.186 msaitoh sc->osdep.attached = false;
344 1.1 dyoung
345 1.3 msaitoh ent = ixv_lookup(pa);
346 1.3 msaitoh
347 1.3 msaitoh KASSERT(ent != NULL);
348 1.3 msaitoh
349 1.3 msaitoh aprint_normal(": %s, Version - %s\n",
350 1.3 msaitoh ixv_strings[ent->index], ixv_driver_version);
351 1.3 msaitoh
352 1.150 msaitoh /* Core Lock Init */
353 1.186 msaitoh IXGBE_CORE_LOCK_INIT(sc, device_xname(dev));
354 1.1 dyoung
355 1.1 dyoung /* Do base PCI setup - map BAR0 */
356 1.186 msaitoh if (ixv_allocate_pci_resources(sc, pa)) {
357 1.26 msaitoh aprint_error_dev(dev, "ixv_allocate_pci_resources() failed!\n");
358 1.1 dyoung error = ENXIO;
359 1.1 dyoung goto err_out;
360 1.1 dyoung }
361 1.1 dyoung
362 1.58 msaitoh /* SYSCTL APIs */
363 1.186 msaitoh ixv_add_device_sysctls(sc);
364 1.25 msaitoh
365 1.151 msaitoh /* Set up the timer callout and workqueue */
366 1.186 msaitoh callout_init(&sc->timer, IXGBE_CALLOUT_FLAGS);
367 1.151 msaitoh snprintf(wqname, sizeof(wqname), "%s-timer", device_xname(dev));
368 1.186 msaitoh error = workqueue_create(&sc->timer_wq, wqname,
369 1.186 msaitoh ixv_handle_timer, sc, IXGBE_WORKQUEUE_PRI, IPL_NET,
370 1.151 msaitoh IXGBE_TASKLET_WQ_FLAGS);
371 1.151 msaitoh if (error) {
372 1.151 msaitoh aprint_error_dev(dev,
373 1.151 msaitoh "could not create timer workqueue (%d)\n", error);
374 1.151 msaitoh goto err_out;
375 1.151 msaitoh }
376 1.25 msaitoh
377 1.58 msaitoh /* Save off the information about this board */
378 1.58 msaitoh id = pci_conf_read(pa->pa_pc, pa->pa_tag, PCI_ID_REG);
379 1.58 msaitoh subid = pci_conf_read(pa->pa_pc, pa->pa_tag, PCI_SUBSYS_ID_REG);
380 1.58 msaitoh hw->vendor_id = PCI_VENDOR(id);
381 1.58 msaitoh hw->device_id = PCI_PRODUCT(id);
382 1.58 msaitoh hw->revision_id =
383 1.58 msaitoh PCI_REVISION(pci_conf_read(pa->pa_pc, pa->pa_tag, PCI_CLASS_REG));
384 1.58 msaitoh hw->subsystem_vendor_id = PCI_SUBSYS_VENDOR(subid);
385 1.58 msaitoh hw->subsystem_device_id = PCI_SUBSYS_ID(subid);
386 1.1 dyoung
387 1.58 msaitoh /* A subset of set_mac_type */
388 1.58 msaitoh switch (hw->device_id) {
389 1.58 msaitoh case IXGBE_DEV_ID_82599_VF:
390 1.58 msaitoh hw->mac.type = ixgbe_mac_82599_vf;
391 1.66 msaitoh str = "82599 VF";
392 1.58 msaitoh break;
393 1.58 msaitoh case IXGBE_DEV_ID_X540_VF:
394 1.58 msaitoh hw->mac.type = ixgbe_mac_X540_vf;
395 1.66 msaitoh str = "X540 VF";
396 1.58 msaitoh break;
397 1.58 msaitoh case IXGBE_DEV_ID_X550_VF:
398 1.58 msaitoh hw->mac.type = ixgbe_mac_X550_vf;
399 1.66 msaitoh str = "X550 VF";
400 1.58 msaitoh break;
401 1.58 msaitoh case IXGBE_DEV_ID_X550EM_X_VF:
402 1.58 msaitoh hw->mac.type = ixgbe_mac_X550EM_x_vf;
403 1.66 msaitoh str = "X550EM X VF";
404 1.58 msaitoh break;
405 1.58 msaitoh case IXGBE_DEV_ID_X550EM_A_VF:
406 1.58 msaitoh hw->mac.type = ixgbe_mac_X550EM_a_vf;
407 1.66 msaitoh str = "X550EM A VF";
408 1.58 msaitoh break;
409 1.58 msaitoh default:
410 1.58 msaitoh /* Shouldn't get here since probe succeeded */
411 1.58 msaitoh aprint_error_dev(dev, "Unknown device ID!\n");
412 1.58 msaitoh error = ENXIO;
413 1.1 dyoung goto err_out;
414 1.58 msaitoh break;
415 1.1 dyoung }
416 1.66 msaitoh aprint_normal_dev(dev, "device %s\n", str);
417 1.1 dyoung
418 1.186 msaitoh ixv_init_device_features(sc);
419 1.58 msaitoh
420 1.58 msaitoh /* Initialize the shared code */
421 1.58 msaitoh error = ixgbe_init_ops_vf(hw);
422 1.1 dyoung if (error) {
423 1.58 msaitoh aprint_error_dev(dev, "ixgbe_init_ops_vf() failed!\n");
424 1.1 dyoung error = EIO;
425 1.58 msaitoh goto err_out;
426 1.1 dyoung }
427 1.1 dyoung
428 1.1 dyoung /* Setup the mailbox */
429 1.1 dyoung ixgbe_init_mbx_params_vf(hw);
430 1.1 dyoung
431 1.58 msaitoh /* Set the right number of segments */
432 1.168 msaitoh KASSERT(IXGBE_82599_SCATTER_MAX >= IXGBE_SCATTER_DEFAULT);
433 1.186 msaitoh sc->num_segs = IXGBE_SCATTER_DEFAULT;
434 1.58 msaitoh
435 1.26 msaitoh /* Reset mbox api to 1.0 */
436 1.58 msaitoh error = hw->mac.ops.reset_hw(hw);
437 1.26 msaitoh if (error == IXGBE_ERR_RESET_FAILED)
438 1.58 msaitoh aprint_error_dev(dev, "...reset_hw() failure: Reset Failed!\n");
439 1.26 msaitoh else if (error)
440 1.58 msaitoh aprint_error_dev(dev, "...reset_hw() failed with error %d\n",
441 1.58 msaitoh error);
442 1.26 msaitoh if (error) {
443 1.26 msaitoh error = EIO;
444 1.58 msaitoh goto err_out;
445 1.26 msaitoh }
446 1.1 dyoung
447 1.58 msaitoh error = hw->mac.ops.init_hw(hw);
448 1.1 dyoung if (error) {
449 1.58 msaitoh aprint_error_dev(dev, "...init_hw() failed!\n");
450 1.1 dyoung error = EIO;
451 1.58 msaitoh goto err_out;
452 1.1 dyoung }
453 1.63 msaitoh
454 1.58 msaitoh /* Negotiate mailbox API version */
455 1.186 msaitoh error = ixv_negotiate_api(sc);
456 1.58 msaitoh if (error)
457 1.58 msaitoh aprint_normal_dev(dev,
458 1.58 msaitoh "MBX API negotiation failed during attach!\n");
459 1.60 msaitoh switch (hw->api_version) {
460 1.60 msaitoh case ixgbe_mbox_api_10:
461 1.60 msaitoh apivstr = "1.0";
462 1.60 msaitoh break;
463 1.60 msaitoh case ixgbe_mbox_api_20:
464 1.60 msaitoh apivstr = "2.0";
465 1.60 msaitoh break;
466 1.60 msaitoh case ixgbe_mbox_api_11:
467 1.60 msaitoh apivstr = "1.1";
468 1.60 msaitoh break;
469 1.60 msaitoh case ixgbe_mbox_api_12:
470 1.60 msaitoh apivstr = "1.2";
471 1.60 msaitoh break;
472 1.60 msaitoh case ixgbe_mbox_api_13:
473 1.60 msaitoh apivstr = "1.3";
474 1.60 msaitoh break;
475 1.172 msaitoh case ixgbe_mbox_api_14:
476 1.172 msaitoh apivstr = "1.4";
477 1.172 msaitoh break;
478 1.172 msaitoh case ixgbe_mbox_api_15:
479 1.172 msaitoh apivstr = "1.5";
480 1.172 msaitoh break;
481 1.60 msaitoh default:
482 1.60 msaitoh apivstr = "unknown";
483 1.60 msaitoh break;
484 1.60 msaitoh }
485 1.60 msaitoh aprint_normal_dev(dev, "Mailbox API %s\n", apivstr);
486 1.1 dyoung
487 1.21 msaitoh /* If no mac address was assigned, make a random one */
488 1.21 msaitoh if (!ixv_check_ether_addr(hw->mac.addr)) {
489 1.21 msaitoh u8 addr[ETHER_ADDR_LEN];
490 1.59 msaitoh uint64_t rndval = cprng_strong64();
491 1.21 msaitoh
492 1.21 msaitoh memcpy(addr, &rndval, sizeof(addr));
493 1.21 msaitoh addr[0] &= 0xFE;
494 1.21 msaitoh addr[0] |= 0x02;
495 1.21 msaitoh bcopy(addr, hw->mac.addr, sizeof(addr));
496 1.21 msaitoh }
497 1.21 msaitoh
498 1.58 msaitoh /* Register for VLAN events */
499 1.186 msaitoh ether_set_vlan_cb(&sc->osdep.ec, ixv_vlan_cb);
500 1.58 msaitoh
501 1.58 msaitoh /* Do descriptor calc and sanity checks */
502 1.58 msaitoh if (((ixv_txd * sizeof(union ixgbe_adv_tx_desc)) % DBA_ALIGN) != 0 ||
503 1.58 msaitoh ixv_txd < MIN_TXD || ixv_txd > MAX_TXD) {
504 1.188 msaitoh aprint_error_dev(dev, "Invalid TX ring size (%d). "
505 1.188 msaitoh "It must be between %d and %d, "
506 1.188 msaitoh "inclusive, and must be a multiple of %zu. "
507 1.188 msaitoh "Using default value of %d instead.\n",
508 1.188 msaitoh ixv_txd, MIN_TXD, MAX_TXD,
509 1.188 msaitoh DBA_ALIGN / sizeof(union ixgbe_adv_tx_desc),
510 1.188 msaitoh DEFAULT_TXD);
511 1.186 msaitoh sc->num_tx_desc = DEFAULT_TXD;
512 1.58 msaitoh } else
513 1.186 msaitoh sc->num_tx_desc = ixv_txd;
514 1.58 msaitoh
515 1.58 msaitoh if (((ixv_rxd * sizeof(union ixgbe_adv_rx_desc)) % DBA_ALIGN) != 0 ||
516 1.58 msaitoh ixv_rxd < MIN_RXD || ixv_rxd > MAX_RXD) {
517 1.188 msaitoh aprint_error_dev(dev, "Invalid RX ring size (%d). "
518 1.188 msaitoh "It must be between %d and %d, "
519 1.188 msaitoh "inclusive, and must be a multiple of %zu. "
520 1.188 msaitoh "Using default value of %d instead.\n",
521 1.188 msaitoh ixv_rxd, MIN_RXD, MAX_RXD,
522 1.188 msaitoh DBA_ALIGN / sizeof(union ixgbe_adv_rx_desc),
523 1.188 msaitoh DEFAULT_RXD);
524 1.186 msaitoh sc->num_rx_desc = DEFAULT_RXD;
525 1.58 msaitoh } else
526 1.186 msaitoh sc->num_rx_desc = ixv_rxd;
527 1.58 msaitoh
528 1.179 msaitoh /* Sysctls for limiting the amount of work done in the taskqueues */
529 1.186 msaitoh sc->rx_process_limit
530 1.186 msaitoh = (ixv_rx_process_limit <= sc->num_rx_desc)
531 1.186 msaitoh ? ixv_rx_process_limit : sc->num_rx_desc;
532 1.186 msaitoh sc->tx_process_limit
533 1.186 msaitoh = (ixv_tx_process_limit <= sc->num_tx_desc)
534 1.186 msaitoh ? ixv_tx_process_limit : sc->num_tx_desc;
535 1.179 msaitoh
536 1.163 msaitoh /* Set default high limit of copying mbuf in rxeof */
537 1.186 msaitoh sc->rx_copy_len = IXGBE_RX_COPY_LEN_MAX;
538 1.163 msaitoh
539 1.58 msaitoh /* Setup MSI-X */
540 1.186 msaitoh error = ixv_configure_interrupts(sc);
541 1.58 msaitoh if (error)
542 1.58 msaitoh goto err_out;
543 1.58 msaitoh
544 1.58 msaitoh /* Allocate our TX/RX Queues */
545 1.186 msaitoh if (ixgbe_allocate_queues(sc)) {
546 1.58 msaitoh aprint_error_dev(dev, "ixgbe_allocate_queues() failed!\n");
547 1.58 msaitoh error = ENOMEM;
548 1.58 msaitoh goto err_out;
549 1.58 msaitoh }
550 1.58 msaitoh
551 1.50 msaitoh /* hw.ix defaults init */
552 1.186 msaitoh sc->enable_aim = ixv_enable_aim;
553 1.191 msaitoh sc->max_interrupt_rate = ixv_max_interrupt_rate;
554 1.50 msaitoh
555 1.186 msaitoh sc->txrx_use_workqueue = ixv_txrx_workqueue;
556 1.84 knakahar
557 1.186 msaitoh error = ixv_allocate_msix(sc, pa);
558 1.76 msaitoh if (error) {
559 1.130 msaitoh aprint_error_dev(dev, "ixv_allocate_msix() failed!\n");
560 1.76 msaitoh goto err_late;
561 1.76 msaitoh }
562 1.76 msaitoh
563 1.1 dyoung /* Setup OS specific network interface */
564 1.186 msaitoh error = ixv_setup_interface(dev, sc);
565 1.73 msaitoh if (error != 0) {
566 1.73 msaitoh aprint_error_dev(dev, "ixv_setup_interface() failed!\n");
567 1.73 msaitoh goto err_late;
568 1.73 msaitoh }
569 1.1 dyoung
570 1.170 msaitoh /* Allocate multicast array memory */
571 1.186 msaitoh sc->mta = malloc(sizeof(*sc->mta) *
572 1.170 msaitoh IXGBE_MAX_VF_MC, M_DEVBUF, M_WAITOK);
573 1.170 msaitoh
574 1.185 msaitoh /* Check if VF was disabled by PF */
575 1.186 msaitoh error = hw->mac.ops.get_link_state(hw, &sc->link_enabled);
576 1.185 msaitoh if (error) {
577 1.185 msaitoh /* PF is not capable of controlling VF state. Enable the link. */
578 1.186 msaitoh sc->link_enabled = TRUE;
579 1.185 msaitoh }
580 1.185 msaitoh
581 1.1 dyoung /* Do the stats setup */
582 1.186 msaitoh ixv_init_stats(sc);
583 1.186 msaitoh ixv_add_stats_sysctls(sc);
584 1.1 dyoung
585 1.186 msaitoh if (sc->feat_en & IXGBE_FEATURE_NETMAP)
586 1.186 msaitoh ixgbe_netmap_attach(sc);
587 1.48 msaitoh
588 1.186 msaitoh snprintb(buf, sizeof(buf), IXGBE_FEATURE_FLAGS, sc->feat_cap);
589 1.63 msaitoh aprint_verbose_dev(dev, "feature cap %s\n", buf);
590 1.186 msaitoh snprintb(buf, sizeof(buf), IXGBE_FEATURE_FLAGS, sc->feat_en);
591 1.63 msaitoh aprint_verbose_dev(dev, "feature ena %s\n", buf);
592 1.63 msaitoh
593 1.1 dyoung INIT_DEBUGOUT("ixv_attach: end");
594 1.186 msaitoh sc->osdep.attached = true;
595 1.57 msaitoh
596 1.3 msaitoh return;
597 1.1 dyoung
598 1.1 dyoung err_late:
599 1.186 msaitoh ixgbe_free_queues(sc);
600 1.1 dyoung err_out:
601 1.186 msaitoh ixv_free_pci_resources(sc);
602 1.186 msaitoh IXGBE_CORE_LOCK_DESTROY(sc);
603 1.58 msaitoh
604 1.3 msaitoh return;
605 1.58 msaitoh } /* ixv_attach */
606 1.1 dyoung
607 1.58 msaitoh /************************************************************************
608 1.58 msaitoh * ixv_detach - Device removal routine
609 1.1 dyoung *
610 1.58 msaitoh * Called when the driver is being removed.
611 1.58 msaitoh * Stops the adapter and deallocates all the resources
612 1.58 msaitoh * that were allocated for driver operation.
613 1.1 dyoung *
614 1.58 msaitoh * return 0 on success, positive on failure
615 1.58 msaitoh ************************************************************************/
616 1.1 dyoung static int
617 1.3 msaitoh ixv_detach(device_t dev, int flags)
618 1.1 dyoung {
619 1.186 msaitoh struct ixgbe_softc *sc = device_private(dev);
620 1.186 msaitoh struct ixgbe_hw *hw = &sc->hw;
621 1.186 msaitoh struct tx_ring *txr = sc->tx_rings;
622 1.186 msaitoh struct rx_ring *rxr = sc->rx_rings;
623 1.186 msaitoh struct ixgbevf_hw_stats *stats = &sc->stats.vf;
624 1.1 dyoung
625 1.1 dyoung INIT_DEBUGOUT("ixv_detach: begin");
626 1.186 msaitoh if (sc->osdep.attached == false)
627 1.13 msaitoh return 0;
628 1.1 dyoung
629 1.56 msaitoh /* Stop the interface. Callouts are stopped in it. */
630 1.186 msaitoh ixv_ifstop(sc->ifp, 1);
631 1.56 msaitoh
632 1.186 msaitoh if (VLAN_ATTACHED(&sc->osdep.ec) &&
633 1.169 yamaguch (flags & (DETACH_SHUTDOWN | DETACH_FORCE)) == 0) {
634 1.26 msaitoh aprint_error_dev(dev, "VLANs in use, detach first\n");
635 1.3 msaitoh return EBUSY;
636 1.1 dyoung }
637 1.1 dyoung
638 1.186 msaitoh ether_ifdetach(sc->ifp);
639 1.186 msaitoh callout_halt(&sc->timer, NULL);
640 1.186 msaitoh ixv_free_deferred_handlers(sc);
641 1.58 msaitoh
642 1.186 msaitoh if (sc->feat_en & IXGBE_FEATURE_NETMAP)
643 1.186 msaitoh netmap_detach(sc->ifp);
644 1.58 msaitoh
645 1.186 msaitoh ixv_free_pci_resources(sc);
646 1.3 msaitoh #if 0 /* XXX the NetBSD port is probably missing something here */
647 1.1 dyoung bus_generic_detach(dev);
648 1.3 msaitoh #endif
649 1.186 msaitoh if_detach(sc->ifp);
650 1.186 msaitoh ifmedia_fini(&sc->media);
651 1.186 msaitoh if_percpuq_destroy(sc->ipq);
652 1.186 msaitoh
653 1.186 msaitoh sysctl_teardown(&sc->sysctllog);
654 1.186 msaitoh evcnt_detach(&sc->efbig_tx_dma_setup);
655 1.186 msaitoh evcnt_detach(&sc->mbuf_defrag_failed);
656 1.186 msaitoh evcnt_detach(&sc->efbig2_tx_dma_setup);
657 1.186 msaitoh evcnt_detach(&sc->einval_tx_dma_setup);
658 1.186 msaitoh evcnt_detach(&sc->other_tx_dma_setup);
659 1.186 msaitoh evcnt_detach(&sc->eagain_tx_dma_setup);
660 1.186 msaitoh evcnt_detach(&sc->enomem_tx_dma_setup);
661 1.186 msaitoh evcnt_detach(&sc->watchdog_events);
662 1.186 msaitoh evcnt_detach(&sc->tso_err);
663 1.186 msaitoh evcnt_detach(&sc->admin_irqev);
664 1.186 msaitoh evcnt_detach(&sc->link_workev);
665 1.186 msaitoh
666 1.186 msaitoh txr = sc->tx_rings;
667 1.186 msaitoh for (int i = 0; i < sc->num_queues; i++, rxr++, txr++) {
668 1.186 msaitoh evcnt_detach(&sc->queues[i].irqs);
669 1.186 msaitoh evcnt_detach(&sc->queues[i].handleq);
670 1.186 msaitoh evcnt_detach(&sc->queues[i].req);
671 1.49 msaitoh evcnt_detach(&txr->total_packets);
672 1.49 msaitoh #ifndef IXGBE_LEGACY_TX
673 1.49 msaitoh evcnt_detach(&txr->pcq_drops);
674 1.49 msaitoh #endif
675 1.184 msaitoh evcnt_detach(&txr->no_desc_avail);
676 1.184 msaitoh evcnt_detach(&txr->tso_tx);
677 1.49 msaitoh
678 1.49 msaitoh evcnt_detach(&rxr->rx_packets);
679 1.49 msaitoh evcnt_detach(&rxr->rx_bytes);
680 1.49 msaitoh evcnt_detach(&rxr->rx_copies);
681 1.166 msaitoh evcnt_detach(&rxr->no_mbuf);
682 1.49 msaitoh evcnt_detach(&rxr->rx_discarded);
683 1.49 msaitoh }
684 1.49 msaitoh evcnt_detach(&stats->ipcs);
685 1.49 msaitoh evcnt_detach(&stats->l4cs);
686 1.49 msaitoh evcnt_detach(&stats->ipcs_bad);
687 1.49 msaitoh evcnt_detach(&stats->l4cs_bad);
688 1.49 msaitoh
689 1.49 msaitoh /* Packet Reception Stats */
690 1.49 msaitoh evcnt_detach(&stats->vfgorc);
691 1.49 msaitoh evcnt_detach(&stats->vfgprc);
692 1.49 msaitoh evcnt_detach(&stats->vfmprc);
693 1.49 msaitoh
694 1.49 msaitoh /* Packet Transmission Stats */
695 1.49 msaitoh evcnt_detach(&stats->vfgotc);
696 1.49 msaitoh evcnt_detach(&stats->vfgptc);
697 1.41 msaitoh
698 1.67 msaitoh /* Mailbox Stats */
699 1.67 msaitoh evcnt_detach(&hw->mbx.stats.msgs_tx);
700 1.67 msaitoh evcnt_detach(&hw->mbx.stats.msgs_rx);
701 1.67 msaitoh evcnt_detach(&hw->mbx.stats.acks);
702 1.67 msaitoh evcnt_detach(&hw->mbx.stats.reqs);
703 1.67 msaitoh evcnt_detach(&hw->mbx.stats.rsts);
704 1.67 msaitoh
705 1.186 msaitoh ixgbe_free_queues(sc);
706 1.1 dyoung
707 1.186 msaitoh IXGBE_CORE_LOCK_DESTROY(sc);
708 1.58 msaitoh
709 1.1 dyoung return (0);
710 1.58 msaitoh } /* ixv_detach */
711 1.1 dyoung
712 1.58 msaitoh /************************************************************************
713 1.58 msaitoh * ixv_init_locked - Init entry point
714 1.58 msaitoh *
715 1.58 msaitoh * Used in two ways: It is used by the stack as an init entry
716 1.58 msaitoh * point in network interface structure. It is also used
717 1.58 msaitoh * by the driver as a hw/sw initialization routine to get
718 1.58 msaitoh * to a consistent state.
719 1.1 dyoung *
720 1.58 msaitoh * return 0 on success, positive on failure
721 1.58 msaitoh ************************************************************************/
722 1.1 dyoung static void
723 1.186 msaitoh ixv_init_locked(struct ixgbe_softc *sc)
724 1.1 dyoung {
725 1.186 msaitoh struct ifnet *ifp = sc->ifp;
726 1.186 msaitoh device_t dev = sc->dev;
727 1.186 msaitoh struct ixgbe_hw *hw = &sc->hw;
728 1.102 msaitoh struct ix_queue *que;
729 1.114 msaitoh int error = 0;
730 1.68 msaitoh uint32_t mask;
731 1.68 msaitoh int i;
732 1.1 dyoung
733 1.26 msaitoh INIT_DEBUGOUT("ixv_init_locked: begin");
734 1.186 msaitoh KASSERT(mutex_owned(&sc->core_mtx));
735 1.1 dyoung hw->adapter_stopped = FALSE;
736 1.58 msaitoh hw->mac.ops.stop_adapter(hw);
737 1.186 msaitoh callout_stop(&sc->timer);
738 1.186 msaitoh for (i = 0, que = sc->queues; i < sc->num_queues; i++, que++)
739 1.102 msaitoh que->disabled_count = 0;
740 1.1 dyoung
741 1.186 msaitoh sc->max_frame_size =
742 1.139 msaitoh ifp->if_mtu + ETHER_HDR_LEN + ETHER_CRC_LEN;
743 1.139 msaitoh
744 1.57 msaitoh /* reprogram the RAR[0] in case user changed it. */
745 1.58 msaitoh hw->mac.ops.set_rar(hw, 0, hw->mac.addr, 0, IXGBE_RAH_AV);
746 1.1 dyoung
747 1.1 dyoung /* Get the latest mac address, User can use a LAA */
748 1.91 msaitoh memcpy(hw->mac.addr, CLLADDR(ifp->if_sadl),
749 1.1 dyoung IXGBE_ETH_LENGTH_OF_ADDRESS);
750 1.58 msaitoh hw->mac.ops.set_rar(hw, 0, hw->mac.addr, 0, 1);
751 1.1 dyoung
752 1.1 dyoung /* Prepare transmit descriptors and buffers */
753 1.186 msaitoh if (ixgbe_setup_transmit_structures(sc)) {
754 1.26 msaitoh aprint_error_dev(dev, "Could not setup transmit structures\n");
755 1.186 msaitoh ixv_stop_locked(sc);
756 1.1 dyoung return;
757 1.1 dyoung }
758 1.1 dyoung
759 1.26 msaitoh /* Reset VF and renegotiate mailbox API version */
760 1.58 msaitoh hw->mac.ops.reset_hw(hw);
761 1.92 msaitoh hw->mac.ops.start_hw(hw);
762 1.186 msaitoh error = ixv_negotiate_api(sc);
763 1.26 msaitoh if (error)
764 1.58 msaitoh device_printf(dev,
765 1.58 msaitoh "Mailbox API negotiation failed in init_locked!\n");
766 1.26 msaitoh
767 1.186 msaitoh ixv_initialize_transmit_units(sc);
768 1.1 dyoung
769 1.1 dyoung /* Setup Multicast table */
770 1.186 msaitoh ixv_set_rxfilter(sc);
771 1.1 dyoung
772 1.165 msaitoh /* Use fixed buffer size, even for jumbo frames */
773 1.186 msaitoh sc->rx_mbuf_sz = MCLBYTES;
774 1.1 dyoung
775 1.1 dyoung /* Prepare receive descriptors and buffers */
776 1.186 msaitoh error = ixgbe_setup_receive_structures(sc);
777 1.160 msaitoh if (error) {
778 1.160 msaitoh device_printf(dev,
779 1.160 msaitoh "Could not setup receive structures (err = %d)\n", error);
780 1.186 msaitoh ixv_stop_locked(sc);
781 1.1 dyoung return;
782 1.1 dyoung }
783 1.1 dyoung
784 1.1 dyoung /* Configure RX settings */
785 1.186 msaitoh ixv_initialize_receive_units(sc);
786 1.1 dyoung
787 1.151 msaitoh /* Initialize variable holding task enqueue requests interrupts */
788 1.186 msaitoh sc->task_requests = 0;
789 1.151 msaitoh
790 1.1 dyoung /* Set up VLAN offload and filter */
791 1.186 msaitoh ixv_setup_vlan_support(sc);
792 1.1 dyoung
793 1.58 msaitoh /* Set up MSI-X routing */
794 1.186 msaitoh ixv_configure_ivars(sc);
795 1.1 dyoung
796 1.1 dyoung /* Set up auto-mask */
797 1.186 msaitoh mask = (1 << sc->vector);
798 1.186 msaitoh for (i = 0, que = sc->queues; i < sc->num_queues; i++, que++)
799 1.68 msaitoh mask |= (1 << que->msix);
800 1.68 msaitoh IXGBE_WRITE_REG(hw, IXGBE_VTEIAM, mask);
801 1.1 dyoung
802 1.57 msaitoh /* Set moderation on the Link interrupt */
803 1.186 msaitoh ixv_eitr_write(sc, sc->vector, IXGBE_LINK_ITR);
804 1.1 dyoung
805 1.1 dyoung /* Stats init */
806 1.186 msaitoh ixv_init_stats(sc);
807 1.1 dyoung
808 1.1 dyoung /* Config/Enable Link */
809 1.186 msaitoh error = hw->mac.ops.get_link_state(hw, &sc->link_enabled);
810 1.185 msaitoh if (error) {
811 1.185 msaitoh /* PF is not capable of controlling VF state. Enable the link. */
812 1.186 msaitoh sc->link_enabled = TRUE;
813 1.186 msaitoh } else if (sc->link_enabled == FALSE)
814 1.185 msaitoh device_printf(dev, "VF is disabled by PF\n");
815 1.185 msaitoh
816 1.62 msaitoh hw->mac.get_link_status = TRUE;
817 1.186 msaitoh hw->mac.ops.check_link(hw, &sc->link_speed, &sc->link_up,
818 1.58 msaitoh FALSE);
819 1.1 dyoung
820 1.26 msaitoh /* Start watchdog */
821 1.186 msaitoh callout_reset(&sc->timer, hz, ixv_local_timer, sc);
822 1.186 msaitoh atomic_store_relaxed(&sc->timer_pending, 0);
823 1.151 msaitoh
824 1.151 msaitoh /* OK to schedule workqueues. */
825 1.186 msaitoh sc->schedule_wqs_ok = true;
826 1.26 msaitoh
827 1.79 msaitoh /* Update saved flags. See ixgbe_ifflags_cb() */
828 1.186 msaitoh sc->if_flags = ifp->if_flags;
829 1.186 msaitoh sc->ec_capenable = sc->osdep.ec.ec_capenable;
830 1.79 msaitoh
831 1.189 msaitoh /* Inform the stack we're ready */
832 1.3 msaitoh ifp->if_flags |= IFF_RUNNING;
833 1.3 msaitoh ifp->if_flags &= ~IFF_OACTIVE;
834 1.1 dyoung
835 1.189 msaitoh /* And now turn on interrupts */
836 1.189 msaitoh ixv_enable_intr(sc);
837 1.189 msaitoh
838 1.1 dyoung return;
839 1.58 msaitoh } /* ixv_init_locked */
840 1.1 dyoung
841 1.88 msaitoh /************************************************************************
842 1.88 msaitoh * ixv_enable_queue
843 1.88 msaitoh ************************************************************************/
844 1.1 dyoung static inline void
845 1.186 msaitoh ixv_enable_queue(struct ixgbe_softc *sc, u32 vector)
846 1.1 dyoung {
847 1.186 msaitoh struct ixgbe_hw *hw = &sc->hw;
848 1.186 msaitoh struct ix_queue *que = &sc->queues[vector];
849 1.124 msaitoh u32 queue = 1UL << vector;
850 1.114 msaitoh u32 mask;
851 1.1 dyoung
852 1.90 knakahar mutex_enter(&que->dc_mtx);
853 1.90 knakahar if (que->disabled_count > 0 && --que->disabled_count > 0)
854 1.82 knakahar goto out;
855 1.82 knakahar
856 1.1 dyoung mask = (IXGBE_EIMS_RTX_QUEUE & queue);
857 1.1 dyoung IXGBE_WRITE_REG(hw, IXGBE_VTEIMS, mask);
858 1.82 knakahar out:
859 1.90 knakahar mutex_exit(&que->dc_mtx);
860 1.58 msaitoh } /* ixv_enable_queue */
861 1.1 dyoung
862 1.88 msaitoh /************************************************************************
863 1.88 msaitoh * ixv_disable_queue
864 1.88 msaitoh ************************************************************************/
865 1.1 dyoung static inline void
866 1.186 msaitoh ixv_disable_queue(struct ixgbe_softc *sc, u32 vector)
867 1.1 dyoung {
868 1.186 msaitoh struct ixgbe_hw *hw = &sc->hw;
869 1.186 msaitoh struct ix_queue *que = &sc->queues[vector];
870 1.124 msaitoh u32 queue = 1UL << vector;
871 1.114 msaitoh u32 mask;
872 1.1 dyoung
873 1.90 knakahar mutex_enter(&que->dc_mtx);
874 1.90 knakahar if (que->disabled_count++ > 0)
875 1.82 knakahar goto out;
876 1.82 knakahar
877 1.1 dyoung mask = (IXGBE_EIMS_RTX_QUEUE & queue);
878 1.1 dyoung IXGBE_WRITE_REG(hw, IXGBE_VTEIMC, mask);
879 1.82 knakahar out:
880 1.90 knakahar mutex_exit(&que->dc_mtx);
881 1.58 msaitoh } /* ixv_disable_queue */
882 1.1 dyoung
883 1.105 kamil #if 0
884 1.1 dyoung static inline void
885 1.186 msaitoh ixv_rearm_queues(struct ixgbe_softc *sc, u64 queues)
886 1.1 dyoung {
887 1.1 dyoung u32 mask = (IXGBE_EIMS_RTX_QUEUE & queues);
888 1.186 msaitoh IXGBE_WRITE_REG(&sc->hw, IXGBE_VTEICS, mask);
889 1.58 msaitoh } /* ixv_rearm_queues */
890 1.105 kamil #endif
891 1.1 dyoung
892 1.1 dyoung
893 1.58 msaitoh /************************************************************************
894 1.91 msaitoh * ixv_msix_que - MSI-X Queue Interrupt Service routine
895 1.58 msaitoh ************************************************************************/
896 1.58 msaitoh static int
897 1.1 dyoung ixv_msix_que(void *arg)
898 1.1 dyoung {
899 1.1 dyoung struct ix_queue *que = arg;
900 1.186 msaitoh struct ixgbe_softc *sc = que->sc;
901 1.1 dyoung struct tx_ring *txr = que->txr;
902 1.1 dyoung struct rx_ring *rxr = que->rxr;
903 1.21 msaitoh bool more;
904 1.1 dyoung u32 newitr = 0;
905 1.1 dyoung
906 1.186 msaitoh ixv_disable_queue(sc, que->msix);
907 1.178 msaitoh IXGBE_EVC_ADD(&que->irqs, 1);
908 1.1 dyoung
909 1.34 msaitoh #ifdef __NetBSD__
910 1.34 msaitoh /* Don't run ixgbe_rxeof in interrupt context */
911 1.34 msaitoh more = true;
912 1.34 msaitoh #else
913 1.21 msaitoh more = ixgbe_rxeof(que);
914 1.34 msaitoh #endif
915 1.1 dyoung
916 1.21 msaitoh IXGBE_TX_LOCK(txr);
917 1.21 msaitoh ixgbe_txeof(txr);
918 1.21 msaitoh IXGBE_TX_UNLOCK(txr);
919 1.1 dyoung
920 1.1 dyoung /* Do AIM now? */
921 1.1 dyoung
922 1.186 msaitoh if (sc->enable_aim == false)
923 1.1 dyoung goto no_calc;
924 1.1 dyoung /*
925 1.58 msaitoh * Do Adaptive Interrupt Moderation:
926 1.58 msaitoh * - Write out last calculated setting
927 1.58 msaitoh * - Calculate based on average size over
928 1.58 msaitoh * the last interval.
929 1.58 msaitoh */
930 1.63 msaitoh if (que->eitr_setting)
931 1.186 msaitoh ixv_eitr_write(sc, que->msix, que->eitr_setting);
932 1.58 msaitoh
933 1.57 msaitoh que->eitr_setting = 0;
934 1.57 msaitoh
935 1.57 msaitoh /* Idle, do nothing */
936 1.57 msaitoh if ((txr->bytes == 0) && (rxr->bytes == 0))
937 1.57 msaitoh goto no_calc;
938 1.1 dyoung
939 1.1 dyoung if ((txr->bytes) && (txr->packets))
940 1.57 msaitoh newitr = txr->bytes/txr->packets;
941 1.1 dyoung if ((rxr->bytes) && (rxr->packets))
942 1.106 riastrad newitr = uimax(newitr, (rxr->bytes / rxr->packets));
943 1.1 dyoung newitr += 24; /* account for hardware frame, crc */
944 1.1 dyoung
945 1.1 dyoung /* set an upper boundary */
946 1.106 riastrad newitr = uimin(newitr, 3000);
947 1.1 dyoung
948 1.1 dyoung /* Be nice to the mid range */
949 1.1 dyoung if ((newitr > 300) && (newitr < 1200))
950 1.1 dyoung newitr = (newitr / 3);
951 1.1 dyoung else
952 1.1 dyoung newitr = (newitr / 2);
953 1.1 dyoung
954 1.80 msaitoh /*
955 1.80 msaitoh * When RSC is used, ITR interval must be larger than RSC_DELAY.
956 1.80 msaitoh * Currently, we use 2us for RSC_DELAY. The minimum value is always
957 1.80 msaitoh * greater than 2us on 100M (and 10M?(not documented)), but it's not
958 1.80 msaitoh * on 1G and higher.
959 1.80 msaitoh */
960 1.186 msaitoh if ((sc->link_speed != IXGBE_LINK_SPEED_100_FULL)
961 1.186 msaitoh && (sc->link_speed != IXGBE_LINK_SPEED_10_FULL)) {
962 1.80 msaitoh if (newitr < IXGBE_MIN_RSC_EITR_10G1G)
963 1.80 msaitoh newitr = IXGBE_MIN_RSC_EITR_10G1G;
964 1.80 msaitoh }
965 1.58 msaitoh
966 1.58 msaitoh /* save for next interrupt */
967 1.58 msaitoh que->eitr_setting = newitr;
968 1.1 dyoung
969 1.57 msaitoh /* Reset state */
970 1.57 msaitoh txr->bytes = 0;
971 1.57 msaitoh txr->packets = 0;
972 1.57 msaitoh rxr->bytes = 0;
973 1.57 msaitoh rxr->packets = 0;
974 1.1 dyoung
975 1.1 dyoung no_calc:
976 1.86 msaitoh if (more)
977 1.3 msaitoh softint_schedule(que->que_si);
978 1.86 msaitoh else /* Re-enable this interrupt */
979 1.186 msaitoh ixv_enable_queue(sc, que->msix);
980 1.58 msaitoh
981 1.11 msaitoh return 1;
982 1.58 msaitoh } /* ixv_msix_que */
983 1.1 dyoung
984 1.58 msaitoh /************************************************************************
985 1.58 msaitoh * ixv_msix_mbx
986 1.58 msaitoh ************************************************************************/
987 1.11 msaitoh static int
988 1.1 dyoung ixv_msix_mbx(void *arg)
989 1.1 dyoung {
990 1.186 msaitoh struct ixgbe_softc *sc = arg;
991 1.186 msaitoh struct ixgbe_hw *hw = &sc->hw;
992 1.1 dyoung
993 1.186 msaitoh IXGBE_EVC_ADD(&sc->admin_irqev, 1);
994 1.69 msaitoh /* NetBSD: We use auto-clear, so it's not required to write VTEICR */
995 1.1 dyoung
996 1.1 dyoung /* Link status change */
997 1.69 msaitoh hw->mac.get_link_status = TRUE;
998 1.186 msaitoh atomic_or_32(&sc->task_requests, IXGBE_REQUEST_TASK_MBX);
999 1.186 msaitoh ixv_schedule_admin_tasklet(sc);
1000 1.57 msaitoh
1001 1.11 msaitoh return 1;
1002 1.58 msaitoh } /* ixv_msix_mbx */
1003 1.1 dyoung
1004 1.80 msaitoh static void
1005 1.186 msaitoh ixv_eitr_write(struct ixgbe_softc *sc, uint32_t index, uint32_t itr)
1006 1.80 msaitoh {
1007 1.80 msaitoh
1008 1.80 msaitoh /*
1009 1.80 msaitoh * Newer devices than 82598 have VF function, so this function is
1010 1.80 msaitoh * simple.
1011 1.80 msaitoh */
1012 1.80 msaitoh itr |= IXGBE_EITR_CNT_WDIS;
1013 1.80 msaitoh
1014 1.186 msaitoh IXGBE_WRITE_REG(&sc->hw, IXGBE_VTEITR(index), itr);
1015 1.80 msaitoh }
1016 1.80 msaitoh
1017 1.80 msaitoh
1018 1.58 msaitoh /************************************************************************
1019 1.58 msaitoh * ixv_media_status - Media Ioctl callback
1020 1.1 dyoung *
1021 1.58 msaitoh * Called whenever the user queries the status of
1022 1.58 msaitoh * the interface using ifconfig.
1023 1.58 msaitoh ************************************************************************/
1024 1.1 dyoung static void
1025 1.63 msaitoh ixv_media_status(struct ifnet *ifp, struct ifmediareq *ifmr)
1026 1.1 dyoung {
1027 1.186 msaitoh struct ixgbe_softc *sc = ifp->if_softc;
1028 1.1 dyoung
1029 1.1 dyoung INIT_DEBUGOUT("ixv_media_status: begin");
1030 1.186 msaitoh ixv_update_link_status(sc);
1031 1.1 dyoung
1032 1.1 dyoung ifmr->ifm_status = IFM_AVALID;
1033 1.1 dyoung ifmr->ifm_active = IFM_ETHER;
1034 1.1 dyoung
1035 1.186 msaitoh if (sc->link_active != LINK_STATE_UP) {
1036 1.39 msaitoh ifmr->ifm_active |= IFM_NONE;
1037 1.1 dyoung return;
1038 1.1 dyoung }
1039 1.1 dyoung
1040 1.1 dyoung ifmr->ifm_status |= IFM_ACTIVE;
1041 1.1 dyoung
1042 1.186 msaitoh switch (sc->link_speed) {
1043 1.42 msaitoh case IXGBE_LINK_SPEED_10GB_FULL:
1044 1.42 msaitoh ifmr->ifm_active |= IFM_10G_T | IFM_FDX;
1045 1.42 msaitoh break;
1046 1.71 msaitoh case IXGBE_LINK_SPEED_5GB_FULL:
1047 1.71 msaitoh ifmr->ifm_active |= IFM_5000_T | IFM_FDX;
1048 1.71 msaitoh break;
1049 1.71 msaitoh case IXGBE_LINK_SPEED_2_5GB_FULL:
1050 1.71 msaitoh ifmr->ifm_active |= IFM_2500_T | IFM_FDX;
1051 1.71 msaitoh break;
1052 1.1 dyoung case IXGBE_LINK_SPEED_1GB_FULL:
1053 1.1 dyoung ifmr->ifm_active |= IFM_1000_T | IFM_FDX;
1054 1.1 dyoung break;
1055 1.42 msaitoh case IXGBE_LINK_SPEED_100_FULL:
1056 1.42 msaitoh ifmr->ifm_active |= IFM_100_TX | IFM_FDX;
1057 1.1 dyoung break;
1058 1.58 msaitoh case IXGBE_LINK_SPEED_10_FULL:
1059 1.58 msaitoh ifmr->ifm_active |= IFM_10_T | IFM_FDX;
1060 1.58 msaitoh break;
1061 1.1 dyoung }
1062 1.1 dyoung
1063 1.70 msaitoh ifp->if_baudrate = ifmedia_baudrate(ifmr->ifm_active);
1064 1.58 msaitoh } /* ixv_media_status */
1065 1.1 dyoung
1066 1.58 msaitoh /************************************************************************
1067 1.58 msaitoh * ixv_media_change - Media Ioctl callback
1068 1.1 dyoung *
1069 1.58 msaitoh * Called when the user changes speed/duplex using
1070 1.58 msaitoh * media/mediopt option with ifconfig.
1071 1.58 msaitoh ************************************************************************/
1072 1.1 dyoung static int
1073 1.57 msaitoh ixv_media_change(struct ifnet *ifp)
1074 1.1 dyoung {
1075 1.186 msaitoh struct ixgbe_softc *sc = ifp->if_softc;
1076 1.186 msaitoh struct ifmedia *ifm = &sc->media;
1077 1.1 dyoung
1078 1.1 dyoung INIT_DEBUGOUT("ixv_media_change: begin");
1079 1.1 dyoung
1080 1.1 dyoung if (IFM_TYPE(ifm->ifm_media) != IFM_ETHER)
1081 1.1 dyoung return (EINVAL);
1082 1.1 dyoung
1083 1.57 msaitoh switch (IFM_SUBTYPE(ifm->ifm_media)) {
1084 1.57 msaitoh case IFM_AUTO:
1085 1.57 msaitoh break;
1086 1.57 msaitoh default:
1087 1.186 msaitoh device_printf(sc->dev, "Only auto media type\n");
1088 1.1 dyoung return (EINVAL);
1089 1.57 msaitoh }
1090 1.1 dyoung
1091 1.1 dyoung return (0);
1092 1.58 msaitoh } /* ixv_media_change */
1093 1.1 dyoung
1094 1.151 msaitoh static void
1095 1.186 msaitoh ixv_schedule_admin_tasklet(struct ixgbe_softc *sc)
1096 1.151 msaitoh {
1097 1.186 msaitoh if (sc->schedule_wqs_ok) {
1098 1.186 msaitoh if (atomic_cas_uint(&sc->admin_pending, 0, 1) == 0)
1099 1.186 msaitoh workqueue_enqueue(sc->admin_wq,
1100 1.186 msaitoh &sc->admin_wc, NULL);
1101 1.151 msaitoh }
1102 1.151 msaitoh }
1103 1.151 msaitoh
1104 1.137 msaitoh /************************************************************************
1105 1.58 msaitoh * ixv_negotiate_api
1106 1.1 dyoung *
1107 1.58 msaitoh * Negotiate the Mailbox API with the PF;
1108 1.58 msaitoh * start with the most featured API first.
1109 1.58 msaitoh ************************************************************************/
1110 1.58 msaitoh static int
1111 1.186 msaitoh ixv_negotiate_api(struct ixgbe_softc *sc)
1112 1.58 msaitoh {
1113 1.186 msaitoh struct ixgbe_hw *hw = &sc->hw;
1114 1.173 msaitoh int mbx_api[] = { ixgbe_mbox_api_15,
1115 1.173 msaitoh ixgbe_mbox_api_13,
1116 1.134 msaitoh ixgbe_mbox_api_12,
1117 1.134 msaitoh ixgbe_mbox_api_11,
1118 1.114 msaitoh ixgbe_mbox_api_10,
1119 1.114 msaitoh ixgbe_mbox_api_unknown };
1120 1.114 msaitoh int i = 0;
1121 1.58 msaitoh
1122 1.58 msaitoh while (mbx_api[i] != ixgbe_mbox_api_unknown) {
1123 1.173 msaitoh if (ixgbevf_negotiate_api_version(hw, mbx_api[i]) == 0) {
1124 1.173 msaitoh if (hw->api_version >= ixgbe_mbox_api_15)
1125 1.173 msaitoh ixgbe_upgrade_mbx_params_vf(hw);
1126 1.58 msaitoh return (0);
1127 1.173 msaitoh }
1128 1.58 msaitoh i++;
1129 1.58 msaitoh }
1130 1.58 msaitoh
1131 1.58 msaitoh return (EINVAL);
1132 1.58 msaitoh } /* ixv_negotiate_api */
1133 1.58 msaitoh
1134 1.58 msaitoh
1135 1.58 msaitoh /************************************************************************
1136 1.150 msaitoh * ixv_set_rxfilter - Multicast Update
1137 1.1 dyoung *
1138 1.58 msaitoh * Called whenever multicast address list is updated.
1139 1.58 msaitoh ************************************************************************/
1140 1.138 msaitoh static int
1141 1.186 msaitoh ixv_set_rxfilter(struct ixgbe_softc *sc)
1142 1.1 dyoung {
1143 1.170 msaitoh struct ixgbe_mc_addr *mta;
1144 1.186 msaitoh struct ifnet *ifp = sc->ifp;
1145 1.186 msaitoh struct ixgbe_hw *hw = &sc->hw;
1146 1.138 msaitoh u8 *update_ptr;
1147 1.138 msaitoh int mcnt = 0;
1148 1.186 msaitoh struct ethercom *ec = &sc->osdep.ec;
1149 1.138 msaitoh struct ether_multi *enm;
1150 1.138 msaitoh struct ether_multistep step;
1151 1.138 msaitoh bool overflow = false;
1152 1.138 msaitoh int error, rc = 0;
1153 1.1 dyoung
1154 1.186 msaitoh KASSERT(mutex_owned(&sc->core_mtx));
1155 1.138 msaitoh IOCTL_DEBUGOUT("ixv_set_rxfilter: begin");
1156 1.1 dyoung
1157 1.186 msaitoh mta = sc->mta;
1158 1.170 msaitoh bzero(mta, sizeof(*mta) * IXGBE_MAX_VF_MC);
1159 1.170 msaitoh
1160 1.138 msaitoh /* 1: For PROMISC */
1161 1.138 msaitoh if (ifp->if_flags & IFF_PROMISC) {
1162 1.138 msaitoh error = hw->mac.ops.update_xcast_mode(hw,
1163 1.138 msaitoh IXGBEVF_XCAST_MODE_PROMISC);
1164 1.138 msaitoh if (error == IXGBE_ERR_NOT_TRUSTED) {
1165 1.186 msaitoh device_printf(sc->dev,
1166 1.138 msaitoh "this interface is not trusted\n");
1167 1.138 msaitoh error = EPERM;
1168 1.138 msaitoh } else if (error == IXGBE_ERR_FEATURE_NOT_SUPPORTED) {
1169 1.186 msaitoh device_printf(sc->dev,
1170 1.138 msaitoh "the PF doesn't support promisc mode\n");
1171 1.138 msaitoh error = EOPNOTSUPP;
1172 1.138 msaitoh } else if (error == IXGBE_ERR_NOT_IN_PROMISC) {
1173 1.186 msaitoh device_printf(sc->dev,
1174 1.138 msaitoh "the PF may not in promisc mode\n");
1175 1.138 msaitoh error = EINVAL;
1176 1.138 msaitoh } else if (error) {
1177 1.186 msaitoh device_printf(sc->dev,
1178 1.138 msaitoh "failed to set promisc mode. error = %d\n",
1179 1.138 msaitoh error);
1180 1.138 msaitoh error = EIO;
1181 1.138 msaitoh } else
1182 1.138 msaitoh return 0;
1183 1.138 msaitoh rc = error;
1184 1.138 msaitoh }
1185 1.138 msaitoh
1186 1.138 msaitoh /* 2: For ALLMULTI or normal */
1187 1.72 msaitoh ETHER_LOCK(ec);
1188 1.3 msaitoh ETHER_FIRST_MULTI(step, ec, enm);
1189 1.3 msaitoh while (enm != NULL) {
1190 1.138 msaitoh if ((mcnt >= IXGBE_MAX_VF_MC) ||
1191 1.138 msaitoh (memcmp(enm->enm_addrlo, enm->enm_addrhi,
1192 1.138 msaitoh ETHER_ADDR_LEN) != 0)) {
1193 1.134 msaitoh overflow = true;
1194 1.134 msaitoh break;
1195 1.134 msaitoh }
1196 1.3 msaitoh bcopy(enm->enm_addrlo,
1197 1.170 msaitoh mta[mcnt].addr, IXGBE_ETH_LENGTH_OF_ADDRESS);
1198 1.1 dyoung mcnt++;
1199 1.3 msaitoh ETHER_NEXT_MULTI(step, enm);
1200 1.1 dyoung }
1201 1.72 msaitoh ETHER_UNLOCK(ec);
1202 1.1 dyoung
1203 1.138 msaitoh /* 3: For ALLMULTI */
1204 1.134 msaitoh if (overflow) {
1205 1.134 msaitoh error = hw->mac.ops.update_xcast_mode(hw,
1206 1.134 msaitoh IXGBEVF_XCAST_MODE_ALLMULTI);
1207 1.134 msaitoh if (error == IXGBE_ERR_NOT_TRUSTED) {
1208 1.186 msaitoh device_printf(sc->dev,
1209 1.134 msaitoh "this interface is not trusted\n");
1210 1.135 msaitoh error = EPERM;
1211 1.135 msaitoh } else if (error == IXGBE_ERR_FEATURE_NOT_SUPPORTED) {
1212 1.186 msaitoh device_printf(sc->dev,
1213 1.135 msaitoh "the PF doesn't support allmulti mode\n");
1214 1.135 msaitoh error = EOPNOTSUPP;
1215 1.134 msaitoh } else if (error) {
1216 1.186 msaitoh device_printf(sc->dev,
1217 1.134 msaitoh "number of Ethernet multicast addresses "
1218 1.134 msaitoh "exceeds the limit (%d). error = %d\n",
1219 1.134 msaitoh IXGBE_MAX_VF_MC, error);
1220 1.134 msaitoh error = ENOSPC;
1221 1.134 msaitoh } else {
1222 1.138 msaitoh ETHER_LOCK(ec);
1223 1.134 msaitoh ec->ec_flags |= ETHER_F_ALLMULTI;
1224 1.138 msaitoh ETHER_UNLOCK(ec);
1225 1.148 msaitoh return rc; /* Promisc might have failed */
1226 1.134 msaitoh }
1227 1.138 msaitoh
1228 1.138 msaitoh if (rc == 0)
1229 1.138 msaitoh rc = error;
1230 1.138 msaitoh
1231 1.138 msaitoh /* Continue to update the multicast table as many as we can */
1232 1.134 msaitoh }
1233 1.134 msaitoh
1234 1.138 msaitoh /* 4: For normal operation */
1235 1.138 msaitoh error = hw->mac.ops.update_xcast_mode(hw, IXGBEVF_XCAST_MODE_MULTI);
1236 1.138 msaitoh if ((error == IXGBE_ERR_FEATURE_NOT_SUPPORTED) || (error == 0)) {
1237 1.138 msaitoh /* Normal operation */
1238 1.138 msaitoh ETHER_LOCK(ec);
1239 1.134 msaitoh ec->ec_flags &= ~ETHER_F_ALLMULTI;
1240 1.138 msaitoh ETHER_UNLOCK(ec);
1241 1.138 msaitoh error = 0;
1242 1.138 msaitoh } else if (error) {
1243 1.186 msaitoh device_printf(sc->dev,
1244 1.138 msaitoh "failed to set Ethernet multicast address "
1245 1.138 msaitoh "operation to normal. error = %d\n", error);
1246 1.134 msaitoh }
1247 1.134 msaitoh
1248 1.170 msaitoh update_ptr = (u8 *)mta;
1249 1.186 msaitoh error = sc->hw.mac.ops.update_mc_addr_list(&sc->hw,
1250 1.138 msaitoh update_ptr, mcnt, ixv_mc_array_itr, TRUE);
1251 1.138 msaitoh if (rc == 0)
1252 1.138 msaitoh rc = error;
1253 1.138 msaitoh
1254 1.138 msaitoh return rc;
1255 1.138 msaitoh } /* ixv_set_rxfilter */
1256 1.1 dyoung
1257 1.58 msaitoh /************************************************************************
1258 1.58 msaitoh * ixv_mc_array_itr
1259 1.58 msaitoh *
1260 1.58 msaitoh * An iterator function needed by the multicast shared code.
1261 1.58 msaitoh * It feeds the shared code routine the addresses in the
1262 1.138 msaitoh * array of ixv_set_rxfilter() one by one.
1263 1.58 msaitoh ************************************************************************/
1264 1.1 dyoung static u8 *
1265 1.1 dyoung ixv_mc_array_itr(struct ixgbe_hw *hw, u8 **update_ptr, u32 *vmdq)
1266 1.1 dyoung {
1267 1.170 msaitoh struct ixgbe_mc_addr *mta;
1268 1.170 msaitoh
1269 1.170 msaitoh mta = (struct ixgbe_mc_addr *)*update_ptr;
1270 1.88 msaitoh
1271 1.1 dyoung *vmdq = 0;
1272 1.170 msaitoh *update_ptr = (u8*)(mta + 1);
1273 1.1 dyoung
1274 1.170 msaitoh return (mta->addr);
1275 1.58 msaitoh } /* ixv_mc_array_itr */
1276 1.1 dyoung
1277 1.58 msaitoh /************************************************************************
1278 1.58 msaitoh * ixv_local_timer - Timer routine
1279 1.1 dyoung *
1280 1.58 msaitoh * Checks for link status, updates statistics,
1281 1.58 msaitoh * and runs the watchdog check.
1282 1.58 msaitoh ************************************************************************/
1283 1.1 dyoung static void
1284 1.22 msaitoh ixv_local_timer(void *arg)
1285 1.22 msaitoh {
1286 1.186 msaitoh struct ixgbe_softc *sc = arg;
1287 1.22 msaitoh
1288 1.186 msaitoh if (sc->schedule_wqs_ok) {
1289 1.186 msaitoh if (atomic_cas_uint(&sc->timer_pending, 0, 1) == 0)
1290 1.186 msaitoh workqueue_enqueue(sc->timer_wq,
1291 1.186 msaitoh &sc->timer_wc, NULL);
1292 1.151 msaitoh }
1293 1.22 msaitoh }
1294 1.22 msaitoh
1295 1.22 msaitoh static void
1296 1.151 msaitoh ixv_handle_timer(struct work *wk, void *context)
1297 1.1 dyoung {
1298 1.186 msaitoh struct ixgbe_softc *sc = context;
1299 1.186 msaitoh device_t dev = sc->dev;
1300 1.186 msaitoh struct ix_queue *que = sc->queues;
1301 1.98 msaitoh u64 queues = 0;
1302 1.87 msaitoh u64 v0, v1, v2, v3, v4, v5, v6, v7;
1303 1.98 msaitoh int hung = 0;
1304 1.87 msaitoh int i;
1305 1.1 dyoung
1306 1.186 msaitoh IXGBE_CORE_LOCK(sc);
1307 1.1 dyoung
1308 1.186 msaitoh if (ixv_check_link(sc)) {
1309 1.186 msaitoh ixv_init_locked(sc);
1310 1.186 msaitoh IXGBE_CORE_UNLOCK(sc);
1311 1.117 msaitoh return;
1312 1.117 msaitoh }
1313 1.1 dyoung
1314 1.1 dyoung /* Stats Update */
1315 1.186 msaitoh ixv_update_stats(sc);
1316 1.1 dyoung
1317 1.87 msaitoh /* Update some event counters */
1318 1.87 msaitoh v0 = v1 = v2 = v3 = v4 = v5 = v6 = v7 = 0;
1319 1.186 msaitoh que = sc->queues;
1320 1.186 msaitoh for (i = 0; i < sc->num_queues; i++, que++) {
1321 1.114 msaitoh struct tx_ring *txr = que->txr;
1322 1.87 msaitoh
1323 1.87 msaitoh v0 += txr->q_efbig_tx_dma_setup;
1324 1.87 msaitoh v1 += txr->q_mbuf_defrag_failed;
1325 1.87 msaitoh v2 += txr->q_efbig2_tx_dma_setup;
1326 1.87 msaitoh v3 += txr->q_einval_tx_dma_setup;
1327 1.87 msaitoh v4 += txr->q_other_tx_dma_setup;
1328 1.87 msaitoh v5 += txr->q_eagain_tx_dma_setup;
1329 1.87 msaitoh v6 += txr->q_enomem_tx_dma_setup;
1330 1.87 msaitoh v7 += txr->q_tso_err;
1331 1.87 msaitoh }
1332 1.186 msaitoh IXGBE_EVC_STORE(&sc->efbig_tx_dma_setup, v0);
1333 1.186 msaitoh IXGBE_EVC_STORE(&sc->mbuf_defrag_failed, v1);
1334 1.186 msaitoh IXGBE_EVC_STORE(&sc->efbig2_tx_dma_setup, v2);
1335 1.186 msaitoh IXGBE_EVC_STORE(&sc->einval_tx_dma_setup, v3);
1336 1.186 msaitoh IXGBE_EVC_STORE(&sc->other_tx_dma_setup, v4);
1337 1.186 msaitoh IXGBE_EVC_STORE(&sc->eagain_tx_dma_setup, v5);
1338 1.186 msaitoh IXGBE_EVC_STORE(&sc->enomem_tx_dma_setup, v6);
1339 1.186 msaitoh IXGBE_EVC_STORE(&sc->tso_err, v7);
1340 1.87 msaitoh
1341 1.98 msaitoh /*
1342 1.98 msaitoh * Check the TX queues status
1343 1.114 msaitoh * - mark hung queues so we don't schedule on them
1344 1.114 msaitoh * - watchdog only if all queues show hung
1345 1.98 msaitoh */
1346 1.186 msaitoh que = sc->queues;
1347 1.186 msaitoh for (i = 0; i < sc->num_queues; i++, que++) {
1348 1.98 msaitoh /* Keep track of queues with work for soft irq */
1349 1.98 msaitoh if (que->txr->busy)
1350 1.98 msaitoh queues |= ((u64)1 << que->me);
1351 1.21 msaitoh /*
1352 1.98 msaitoh * Each time txeof runs without cleaning, but there
1353 1.98 msaitoh * are uncleaned descriptors it increments busy. If
1354 1.98 msaitoh * we get to the MAX we declare it hung.
1355 1.58 msaitoh */
1356 1.98 msaitoh if (que->busy == IXGBE_QUEUE_HUNG) {
1357 1.98 msaitoh ++hung;
1358 1.98 msaitoh /* Mark the queue as inactive */
1359 1.186 msaitoh sc->active_queues &= ~((u64)1 << que->me);
1360 1.98 msaitoh continue;
1361 1.98 msaitoh } else {
1362 1.98 msaitoh /* Check if we've come back from hung */
1363 1.186 msaitoh if ((sc->active_queues & ((u64)1 << que->me)) == 0)
1364 1.186 msaitoh sc->active_queues |= ((u64)1 << que->me);
1365 1.1 dyoung }
1366 1.98 msaitoh if (que->busy >= IXGBE_MAX_TX_BUSY) {
1367 1.98 msaitoh device_printf(dev,
1368 1.98 msaitoh "Warning queue %d appears to be hung!\n", i);
1369 1.98 msaitoh que->txr->busy = IXGBE_QUEUE_HUNG;
1370 1.98 msaitoh ++hung;
1371 1.98 msaitoh }
1372 1.98 msaitoh }
1373 1.98 msaitoh
1374 1.98 msaitoh /* Only truly watchdog if all queues show hung */
1375 1.186 msaitoh if (hung == sc->num_queues)
1376 1.98 msaitoh goto watchdog;
1377 1.104 msaitoh #if 0
1378 1.98 msaitoh else if (queues != 0) { /* Force an IRQ on queues with work */
1379 1.186 msaitoh ixv_rearm_queues(sc, queues);
1380 1.21 msaitoh }
1381 1.104 msaitoh #endif
1382 1.21 msaitoh
1383 1.186 msaitoh atomic_store_relaxed(&sc->timer_pending, 0);
1384 1.186 msaitoh IXGBE_CORE_UNLOCK(sc);
1385 1.186 msaitoh callout_reset(&sc->timer, hz, ixv_local_timer, sc);
1386 1.98 msaitoh
1387 1.98 msaitoh return;
1388 1.96 msaitoh
1389 1.98 msaitoh watchdog:
1390 1.186 msaitoh device_printf(sc->dev, "Watchdog timeout -- resetting\n");
1391 1.186 msaitoh sc->ifp->if_flags &= ~IFF_RUNNING;
1392 1.186 msaitoh IXGBE_EVC_ADD(&sc->watchdog_events, 1);
1393 1.186 msaitoh ixv_init_locked(sc);
1394 1.186 msaitoh IXGBE_CORE_UNLOCK(sc);
1395 1.151 msaitoh } /* ixv_handle_timer */
1396 1.1 dyoung
1397 1.58 msaitoh /************************************************************************
1398 1.58 msaitoh * ixv_update_link_status - Update OS on link state
1399 1.58 msaitoh *
1400 1.58 msaitoh * Note: Only updates the OS on the cached link state.
1401 1.114 msaitoh * The real check of the hardware only happens with
1402 1.114 msaitoh * a link interrupt.
1403 1.58 msaitoh ************************************************************************/
1404 1.1 dyoung static void
1405 1.186 msaitoh ixv_update_link_status(struct ixgbe_softc *sc)
1406 1.1 dyoung {
1407 1.186 msaitoh struct ifnet *ifp = sc->ifp;
1408 1.186 msaitoh device_t dev = sc->dev;
1409 1.1 dyoung
1410 1.186 msaitoh KASSERT(mutex_owned(&sc->core_mtx));
1411 1.89 knakahar
1412 1.186 msaitoh if (sc->link_up && sc->link_enabled) {
1413 1.186 msaitoh if (sc->link_active != LINK_STATE_UP) {
1414 1.42 msaitoh if (bootverbose) {
1415 1.42 msaitoh const char *bpsmsg;
1416 1.42 msaitoh
1417 1.186 msaitoh switch (sc->link_speed) {
1418 1.42 msaitoh case IXGBE_LINK_SPEED_10GB_FULL:
1419 1.42 msaitoh bpsmsg = "10 Gbps";
1420 1.42 msaitoh break;
1421 1.58 msaitoh case IXGBE_LINK_SPEED_5GB_FULL:
1422 1.58 msaitoh bpsmsg = "5 Gbps";
1423 1.58 msaitoh break;
1424 1.58 msaitoh case IXGBE_LINK_SPEED_2_5GB_FULL:
1425 1.58 msaitoh bpsmsg = "2.5 Gbps";
1426 1.58 msaitoh break;
1427 1.42 msaitoh case IXGBE_LINK_SPEED_1GB_FULL:
1428 1.42 msaitoh bpsmsg = "1 Gbps";
1429 1.42 msaitoh break;
1430 1.42 msaitoh case IXGBE_LINK_SPEED_100_FULL:
1431 1.42 msaitoh bpsmsg = "100 Mbps";
1432 1.42 msaitoh break;
1433 1.58 msaitoh case IXGBE_LINK_SPEED_10_FULL:
1434 1.58 msaitoh bpsmsg = "10 Mbps";
1435 1.58 msaitoh break;
1436 1.42 msaitoh default:
1437 1.42 msaitoh bpsmsg = "unknown speed";
1438 1.42 msaitoh break;
1439 1.42 msaitoh }
1440 1.63 msaitoh device_printf(dev, "Link is up %s %s \n",
1441 1.42 msaitoh bpsmsg, "Full Duplex");
1442 1.42 msaitoh }
1443 1.186 msaitoh sc->link_active = LINK_STATE_UP;
1444 1.1 dyoung if_link_state_change(ifp, LINK_STATE_UP);
1445 1.1 dyoung }
1446 1.109 msaitoh } else {
1447 1.109 msaitoh /*
1448 1.109 msaitoh * Do it when link active changes to DOWN. i.e.
1449 1.109 msaitoh * a) LINK_STATE_UNKNOWN -> LINK_STATE_DOWN
1450 1.114 msaitoh * b) LINK_STATE_UP -> LINK_STATE_DOWN
1451 1.109 msaitoh */
1452 1.186 msaitoh if (sc->link_active != LINK_STATE_DOWN) {
1453 1.1 dyoung if (bootverbose)
1454 1.63 msaitoh device_printf(dev, "Link is Down\n");
1455 1.1 dyoung if_link_state_change(ifp, LINK_STATE_DOWN);
1456 1.186 msaitoh sc->link_active = LINK_STATE_DOWN;
1457 1.1 dyoung }
1458 1.1 dyoung }
1459 1.58 msaitoh } /* ixv_update_link_status */
1460 1.1 dyoung
1461 1.1 dyoung
1462 1.58 msaitoh /************************************************************************
1463 1.58 msaitoh * ixv_stop - Stop the hardware
1464 1.58 msaitoh *
1465 1.58 msaitoh * Disables all traffic on the adapter by issuing a
1466 1.58 msaitoh * global reset on the MAC and deallocates TX/RX buffers.
1467 1.58 msaitoh ************************************************************************/
1468 1.3 msaitoh static void
1469 1.3 msaitoh ixv_ifstop(struct ifnet *ifp, int disable)
1470 1.3 msaitoh {
1471 1.186 msaitoh struct ixgbe_softc *sc = ifp->if_softc;
1472 1.3 msaitoh
1473 1.186 msaitoh IXGBE_CORE_LOCK(sc);
1474 1.186 msaitoh ixv_stop_locked(sc);
1475 1.186 msaitoh IXGBE_CORE_UNLOCK(sc);
1476 1.186 msaitoh
1477 1.186 msaitoh workqueue_wait(sc->admin_wq, &sc->admin_wc);
1478 1.186 msaitoh atomic_store_relaxed(&sc->admin_pending, 0);
1479 1.186 msaitoh workqueue_wait(sc->timer_wq, &sc->timer_wc);
1480 1.186 msaitoh atomic_store_relaxed(&sc->timer_pending, 0);
1481 1.3 msaitoh }
1482 1.3 msaitoh
1483 1.1 dyoung static void
1484 1.153 msaitoh ixv_stop_locked(void *arg)
1485 1.1 dyoung {
1486 1.114 msaitoh struct ifnet *ifp;
1487 1.186 msaitoh struct ixgbe_softc *sc = arg;
1488 1.186 msaitoh struct ixgbe_hw *hw = &sc->hw;
1489 1.58 msaitoh
1490 1.186 msaitoh ifp = sc->ifp;
1491 1.1 dyoung
1492 1.186 msaitoh KASSERT(mutex_owned(&sc->core_mtx));
1493 1.1 dyoung
1494 1.153 msaitoh INIT_DEBUGOUT("ixv_stop_locked: begin\n");
1495 1.186 msaitoh ixv_disable_intr(sc);
1496 1.1 dyoung
1497 1.1 dyoung /* Tell the stack that the interface is no longer active */
1498 1.3 msaitoh ifp->if_flags &= ~(IFF_RUNNING | IFF_OACTIVE);
1499 1.1 dyoung
1500 1.58 msaitoh hw->mac.ops.reset_hw(hw);
1501 1.186 msaitoh sc->hw.adapter_stopped = FALSE;
1502 1.58 msaitoh hw->mac.ops.stop_adapter(hw);
1503 1.186 msaitoh callout_stop(&sc->timer);
1504 1.1 dyoung
1505 1.151 msaitoh /* Don't schedule workqueues. */
1506 1.186 msaitoh sc->schedule_wqs_ok = false;
1507 1.151 msaitoh
1508 1.1 dyoung /* reprogram the RAR[0] in case user changed it. */
1509 1.58 msaitoh hw->mac.ops.set_rar(hw, 0, hw->mac.addr, 0, IXGBE_RAH_AV);
1510 1.1 dyoung
1511 1.1 dyoung return;
1512 1.153 msaitoh } /* ixv_stop_locked */
1513 1.1 dyoung
1514 1.1 dyoung
1515 1.58 msaitoh /************************************************************************
1516 1.58 msaitoh * ixv_allocate_pci_resources
1517 1.58 msaitoh ************************************************************************/
1518 1.57 msaitoh static int
1519 1.186 msaitoh ixv_allocate_pci_resources(struct ixgbe_softc *sc,
1520 1.57 msaitoh const struct pci_attach_args *pa)
1521 1.1 dyoung {
1522 1.108 msaitoh pcireg_t memtype, csr;
1523 1.186 msaitoh device_t dev = sc->dev;
1524 1.57 msaitoh bus_addr_t addr;
1525 1.57 msaitoh int flags;
1526 1.3 msaitoh
1527 1.57 msaitoh memtype = pci_mapreg_type(pa->pa_pc, pa->pa_tag, PCI_BAR(0));
1528 1.57 msaitoh switch (memtype) {
1529 1.57 msaitoh case PCI_MAPREG_TYPE_MEM | PCI_MAPREG_MEM_TYPE_32BIT:
1530 1.57 msaitoh case PCI_MAPREG_TYPE_MEM | PCI_MAPREG_MEM_TYPE_64BIT:
1531 1.186 msaitoh sc->osdep.mem_bus_space_tag = pa->pa_memt;
1532 1.57 msaitoh if (pci_mapreg_info(pa->pa_pc, pa->pa_tag, PCI_BAR(0),
1533 1.186 msaitoh memtype, &addr, &sc->osdep.mem_size, &flags) != 0)
1534 1.57 msaitoh goto map_err;
1535 1.57 msaitoh if ((flags & BUS_SPACE_MAP_PREFETCHABLE) != 0) {
1536 1.57 msaitoh aprint_normal_dev(dev, "clearing prefetchable bit\n");
1537 1.57 msaitoh flags &= ~BUS_SPACE_MAP_PREFETCHABLE;
1538 1.57 msaitoh }
1539 1.186 msaitoh if (bus_space_map(sc->osdep.mem_bus_space_tag, addr,
1540 1.186 msaitoh sc->osdep.mem_size, flags,
1541 1.186 msaitoh &sc->osdep.mem_bus_space_handle) != 0) {
1542 1.3 msaitoh map_err:
1543 1.186 msaitoh sc->osdep.mem_size = 0;
1544 1.3 msaitoh aprint_error_dev(dev, "unable to map BAR0\n");
1545 1.3 msaitoh return ENXIO;
1546 1.3 msaitoh }
1547 1.108 msaitoh /*
1548 1.108 msaitoh * Enable address decoding for memory range in case it's not
1549 1.108 msaitoh * set.
1550 1.108 msaitoh */
1551 1.108 msaitoh csr = pci_conf_read(pa->pa_pc, pa->pa_tag,
1552 1.108 msaitoh PCI_COMMAND_STATUS_REG);
1553 1.108 msaitoh csr |= PCI_COMMAND_MEM_ENABLE;
1554 1.108 msaitoh pci_conf_write(pa->pa_pc, pa->pa_tag, PCI_COMMAND_STATUS_REG,
1555 1.108 msaitoh csr);
1556 1.3 msaitoh break;
1557 1.3 msaitoh default:
1558 1.3 msaitoh aprint_error_dev(dev, "unexpected type on BAR0\n");
1559 1.3 msaitoh return ENXIO;
1560 1.1 dyoung }
1561 1.1 dyoung
1562 1.23 msaitoh /* Pick up the tuneable queues */
1563 1.186 msaitoh sc->num_queues = ixv_num_queues;
1564 1.1 dyoung
1565 1.58 msaitoh return (0);
1566 1.58 msaitoh } /* ixv_allocate_pci_resources */
1567 1.1 dyoung
1568 1.151 msaitoh static void
1569 1.186 msaitoh ixv_free_deferred_handlers(struct ixgbe_softc *sc)
1570 1.151 msaitoh {
1571 1.186 msaitoh struct ix_queue *que = sc->queues;
1572 1.186 msaitoh struct tx_ring *txr = sc->tx_rings;
1573 1.151 msaitoh int i;
1574 1.151 msaitoh
1575 1.186 msaitoh for (i = 0; i < sc->num_queues; i++, que++, txr++) {
1576 1.186 msaitoh if (!(sc->feat_en & IXGBE_FEATURE_LEGACY_TX)) {
1577 1.151 msaitoh if (txr->txr_si != NULL)
1578 1.151 msaitoh softint_disestablish(txr->txr_si);
1579 1.151 msaitoh }
1580 1.151 msaitoh if (que->que_si != NULL)
1581 1.151 msaitoh softint_disestablish(que->que_si);
1582 1.151 msaitoh }
1583 1.186 msaitoh if (sc->txr_wq != NULL)
1584 1.186 msaitoh workqueue_destroy(sc->txr_wq);
1585 1.186 msaitoh if (sc->txr_wq_enqueued != NULL)
1586 1.186 msaitoh percpu_free(sc->txr_wq_enqueued, sizeof(u_int));
1587 1.186 msaitoh if (sc->que_wq != NULL)
1588 1.186 msaitoh workqueue_destroy(sc->que_wq);
1589 1.151 msaitoh
1590 1.151 msaitoh /* Drain the Mailbox(link) queue */
1591 1.186 msaitoh if (sc->admin_wq != NULL) {
1592 1.186 msaitoh workqueue_destroy(sc->admin_wq);
1593 1.186 msaitoh sc->admin_wq = NULL;
1594 1.186 msaitoh }
1595 1.186 msaitoh if (sc->timer_wq != NULL) {
1596 1.186 msaitoh workqueue_destroy(sc->timer_wq);
1597 1.186 msaitoh sc->timer_wq = NULL;
1598 1.151 msaitoh }
1599 1.154 msaitoh } /* ixv_free_deferred_handlers */
1600 1.151 msaitoh
1601 1.58 msaitoh /************************************************************************
1602 1.58 msaitoh * ixv_free_pci_resources
1603 1.58 msaitoh ************************************************************************/
1604 1.1 dyoung static void
1605 1.186 msaitoh ixv_free_pci_resources(struct ixgbe_softc *sc)
1606 1.1 dyoung {
1607 1.190 msaitoh struct ix_queue *que = sc->queues;
1608 1.11 msaitoh int rid;
1609 1.1 dyoung
1610 1.1 dyoung /*
1611 1.58 msaitoh * Release all msix queue resources:
1612 1.58 msaitoh */
1613 1.186 msaitoh for (int i = 0; i < sc->num_queues; i++, que++) {
1614 1.1 dyoung if (que->res != NULL)
1615 1.186 msaitoh pci_intr_disestablish(sc->osdep.pc,
1616 1.186 msaitoh sc->osdep.ihs[i]);
1617 1.1 dyoung }
1618 1.1 dyoung
1619 1.12 msaitoh
1620 1.58 msaitoh /* Clean the Mailbox interrupt last */
1621 1.186 msaitoh rid = sc->vector;
1622 1.1 dyoung
1623 1.186 msaitoh if (sc->osdep.ihs[rid] != NULL) {
1624 1.186 msaitoh pci_intr_disestablish(sc->osdep.pc,
1625 1.186 msaitoh sc->osdep.ihs[rid]);
1626 1.186 msaitoh sc->osdep.ihs[rid] = NULL;
1627 1.41 msaitoh }
1628 1.11 msaitoh
1629 1.186 msaitoh pci_intr_release(sc->osdep.pc, sc->osdep.intrs,
1630 1.186 msaitoh sc->osdep.nintrs);
1631 1.11 msaitoh
1632 1.186 msaitoh if (sc->osdep.mem_size != 0) {
1633 1.186 msaitoh bus_space_unmap(sc->osdep.mem_bus_space_tag,
1634 1.186 msaitoh sc->osdep.mem_bus_space_handle,
1635 1.186 msaitoh sc->osdep.mem_size);
1636 1.11 msaitoh }
1637 1.1 dyoung
1638 1.1 dyoung return;
1639 1.58 msaitoh } /* ixv_free_pci_resources */
1640 1.1 dyoung
1641 1.58 msaitoh /************************************************************************
1642 1.58 msaitoh * ixv_setup_interface
1643 1.1 dyoung *
1644 1.58 msaitoh * Setup networking device structure and register an interface.
1645 1.58 msaitoh ************************************************************************/
1646 1.73 msaitoh static int
1647 1.186 msaitoh ixv_setup_interface(device_t dev, struct ixgbe_softc *sc)
1648 1.1 dyoung {
1649 1.186 msaitoh struct ethercom *ec = &sc->osdep.ec;
1650 1.1 dyoung struct ifnet *ifp;
1651 1.1 dyoung
1652 1.1 dyoung INIT_DEBUGOUT("ixv_setup_interface: begin");
1653 1.1 dyoung
1654 1.186 msaitoh ifp = sc->ifp = &ec->ec_if;
1655 1.3 msaitoh strlcpy(ifp->if_xname, device_xname(dev), IFNAMSIZ);
1656 1.46 msaitoh ifp->if_baudrate = IF_Gbps(10);
1657 1.1 dyoung ifp->if_init = ixv_init;
1658 1.3 msaitoh ifp->if_stop = ixv_ifstop;
1659 1.186 msaitoh ifp->if_softc = sc;
1660 1.1 dyoung ifp->if_flags = IFF_BROADCAST | IFF_SIMPLEX | IFF_MULTICAST;
1661 1.55 msaitoh #ifdef IXGBE_MPSAFE
1662 1.74 ozaki ifp->if_extflags = IFEF_MPSAFE;
1663 1.55 msaitoh #endif
1664 1.1 dyoung ifp->if_ioctl = ixv_ioctl;
1665 1.186 msaitoh if (sc->feat_en & IXGBE_FEATURE_LEGACY_TX) {
1666 1.58 msaitoh #if 0
1667 1.58 msaitoh ixv_start_locked = ixgbe_legacy_start_locked;
1668 1.58 msaitoh #endif
1669 1.58 msaitoh } else {
1670 1.58 msaitoh ifp->if_transmit = ixgbe_mq_start;
1671 1.58 msaitoh #if 0
1672 1.58 msaitoh ixv_start_locked = ixgbe_mq_start_locked;
1673 1.35 msaitoh #endif
1674 1.58 msaitoh }
1675 1.58 msaitoh ifp->if_start = ixgbe_legacy_start;
1676 1.186 msaitoh IFQ_SET_MAXLEN(&ifp->if_snd, sc->num_tx_desc - 2);
1677 1.45 msaitoh IFQ_SET_READY(&ifp->if_snd);
1678 1.1 dyoung
1679 1.162 riastrad if_initialize(ifp);
1680 1.186 msaitoh sc->ipq = if_percpuq_create(&sc->osdep.ec.ec_if);
1681 1.186 msaitoh ether_ifattach(ifp, sc->hw.mac.addr);
1682 1.141 msaitoh aprint_normal_dev(dev, "Ethernet address %s\n",
1683 1.186 msaitoh ether_sprintf(sc->hw.mac.addr));
1684 1.51 msaitoh /*
1685 1.51 msaitoh * We use per TX queue softint, so if_deferred_start_init() isn't
1686 1.51 msaitoh * used.
1687 1.51 msaitoh */
1688 1.3 msaitoh ether_set_ifflags_cb(ec, ixv_ifflags_cb);
1689 1.1 dyoung
1690 1.186 msaitoh sc->max_frame_size = ifp->if_mtu + IXGBE_MTU_HDR;
1691 1.1 dyoung
1692 1.1 dyoung /*
1693 1.1 dyoung * Tell the upper layer(s) we support long frames.
1694 1.1 dyoung */
1695 1.3 msaitoh ifp->if_hdrlen = sizeof(struct ether_vlan_header);
1696 1.3 msaitoh
1697 1.58 msaitoh /* Set capability flags */
1698 1.58 msaitoh ifp->if_capabilities |= IFCAP_HWCSUM
1699 1.114 msaitoh | IFCAP_TSOv4
1700 1.114 msaitoh | IFCAP_TSOv6;
1701 1.3 msaitoh ifp->if_capenable = 0;
1702 1.1 dyoung
1703 1.120 msaitoh ec->ec_capabilities |= ETHERCAP_VLAN_HWFILTER
1704 1.120 msaitoh | ETHERCAP_VLAN_HWTAGGING
1705 1.58 msaitoh | ETHERCAP_VLAN_HWCSUM
1706 1.58 msaitoh | ETHERCAP_JUMBO_MTU
1707 1.58 msaitoh | ETHERCAP_VLAN_MTU;
1708 1.58 msaitoh
1709 1.58 msaitoh /* Enable the above capabilities by default */
1710 1.3 msaitoh ec->ec_capenable = ec->ec_capabilities;
1711 1.1 dyoung
1712 1.3 msaitoh /* Don't enable LRO by default */
1713 1.107 msaitoh #if 0
1714 1.107 msaitoh /* NetBSD doesn't support LRO yet */
1715 1.3 msaitoh ifp->if_capabilities |= IFCAP_LRO;
1716 1.21 msaitoh #endif
1717 1.3 msaitoh
1718 1.3 msaitoh /*
1719 1.1 dyoung * Specify the media types supported by this adapter and register
1720 1.1 dyoung * callbacks to update media and link information
1721 1.1 dyoung */
1722 1.186 msaitoh ec->ec_ifmedia = &sc->media;
1723 1.186 msaitoh ifmedia_init_with_lock(&sc->media, IFM_IMASK, ixv_media_change,
1724 1.186 msaitoh ixv_media_status, &sc->core_mtx);
1725 1.186 msaitoh ifmedia_add(&sc->media, IFM_ETHER | IFM_AUTO, 0, NULL);
1726 1.186 msaitoh ifmedia_set(&sc->media, IFM_ETHER | IFM_AUTO);
1727 1.1 dyoung
1728 1.101 ozaki if_register(ifp);
1729 1.101 ozaki
1730 1.73 msaitoh return 0;
1731 1.58 msaitoh } /* ixv_setup_interface */
1732 1.58 msaitoh
1733 1.58 msaitoh
1734 1.58 msaitoh /************************************************************************
1735 1.58 msaitoh * ixv_initialize_transmit_units - Enable transmit unit.
1736 1.58 msaitoh ************************************************************************/
1737 1.21 msaitoh static void
1738 1.186 msaitoh ixv_initialize_transmit_units(struct ixgbe_softc *sc)
1739 1.1 dyoung {
1740 1.186 msaitoh struct tx_ring *txr = sc->tx_rings;
1741 1.186 msaitoh struct ixgbe_hw *hw = &sc->hw;
1742 1.91 msaitoh int i;
1743 1.1 dyoung
1744 1.186 msaitoh for (i = 0; i < sc->num_queues; i++, txr++) {
1745 1.58 msaitoh u64 tdba = txr->txdma.dma_paddr;
1746 1.58 msaitoh u32 txctrl, txdctl;
1747 1.91 msaitoh int j = txr->me;
1748 1.1 dyoung
1749 1.21 msaitoh /* Set WTHRESH to 8, burst writeback */
1750 1.91 msaitoh txdctl = IXGBE_READ_REG(hw, IXGBE_VFTXDCTL(j));
1751 1.168 msaitoh txdctl |= IXGBE_TX_WTHRESH << IXGBE_TXDCTL_WTHRESH_SHIFT;
1752 1.91 msaitoh IXGBE_WRITE_REG(hw, IXGBE_VFTXDCTL(j), txdctl);
1753 1.1 dyoung
1754 1.21 msaitoh /* Set the HW Tx Head and Tail indices */
1755 1.186 msaitoh IXGBE_WRITE_REG(&sc->hw, IXGBE_VFTDH(j), 0);
1756 1.186 msaitoh IXGBE_WRITE_REG(&sc->hw, IXGBE_VFTDT(j), 0);
1757 1.1 dyoung
1758 1.21 msaitoh /* Set Tx Tail register */
1759 1.91 msaitoh txr->tail = IXGBE_VFTDT(j);
1760 1.1 dyoung
1761 1.100 msaitoh txr->txr_no_space = false;
1762 1.100 msaitoh
1763 1.21 msaitoh /* Set Ring parameters */
1764 1.91 msaitoh IXGBE_WRITE_REG(hw, IXGBE_VFTDBAL(j),
1765 1.57 msaitoh (tdba & 0x00000000ffffffffULL));
1766 1.91 msaitoh IXGBE_WRITE_REG(hw, IXGBE_VFTDBAH(j), (tdba >> 32));
1767 1.91 msaitoh IXGBE_WRITE_REG(hw, IXGBE_VFTDLEN(j),
1768 1.186 msaitoh sc->num_tx_desc * sizeof(struct ixgbe_legacy_tx_desc));
1769 1.91 msaitoh txctrl = IXGBE_READ_REG(hw, IXGBE_VFDCA_TXCTRL(j));
1770 1.21 msaitoh txctrl &= ~IXGBE_DCA_TXCTRL_DESC_WRO_EN;
1771 1.91 msaitoh IXGBE_WRITE_REG(hw, IXGBE_VFDCA_TXCTRL(j), txctrl);
1772 1.1 dyoung
1773 1.21 msaitoh /* Now enable */
1774 1.91 msaitoh txdctl = IXGBE_READ_REG(hw, IXGBE_VFTXDCTL(j));
1775 1.21 msaitoh txdctl |= IXGBE_TXDCTL_ENABLE;
1776 1.91 msaitoh IXGBE_WRITE_REG(hw, IXGBE_VFTXDCTL(j), txdctl);
1777 1.1 dyoung }
1778 1.1 dyoung
1779 1.21 msaitoh return;
1780 1.58 msaitoh } /* ixv_initialize_transmit_units */
1781 1.58 msaitoh
1782 1.58 msaitoh
1783 1.58 msaitoh /************************************************************************
1784 1.58 msaitoh * ixv_initialize_rss_mapping
1785 1.58 msaitoh ************************************************************************/
1786 1.58 msaitoh static void
1787 1.186 msaitoh ixv_initialize_rss_mapping(struct ixgbe_softc *sc)
1788 1.58 msaitoh {
1789 1.186 msaitoh struct ixgbe_hw *hw = &sc->hw;
1790 1.114 msaitoh u32 reta = 0, mrqc, rss_key[10];
1791 1.114 msaitoh int queue_id;
1792 1.114 msaitoh int i, j;
1793 1.114 msaitoh u32 rss_hash_config;
1794 1.58 msaitoh
1795 1.78 knakahar /* force use default RSS key. */
1796 1.78 knakahar #ifdef __NetBSD__
1797 1.78 knakahar rss_getkey((uint8_t *) &rss_key);
1798 1.78 knakahar #else
1799 1.186 msaitoh if (sc->feat_en & IXGBE_FEATURE_RSS) {
1800 1.58 msaitoh /* Fetch the configured RSS key */
1801 1.58 msaitoh rss_getkey((uint8_t *)&rss_key);
1802 1.58 msaitoh } else {
1803 1.58 msaitoh /* set up random bits */
1804 1.58 msaitoh cprng_fast(&rss_key, sizeof(rss_key));
1805 1.58 msaitoh }
1806 1.78 knakahar #endif
1807 1.58 msaitoh
1808 1.58 msaitoh /* Now fill out hash function seeds */
1809 1.58 msaitoh for (i = 0; i < 10; i++)
1810 1.58 msaitoh IXGBE_WRITE_REG(hw, IXGBE_VFRSSRK(i), rss_key[i]);
1811 1.58 msaitoh
1812 1.58 msaitoh /* Set up the redirection table */
1813 1.58 msaitoh for (i = 0, j = 0; i < 64; i++, j++) {
1814 1.186 msaitoh if (j == sc->num_queues)
1815 1.58 msaitoh j = 0;
1816 1.1 dyoung
1817 1.186 msaitoh if (sc->feat_en & IXGBE_FEATURE_RSS) {
1818 1.58 msaitoh /*
1819 1.58 msaitoh * Fetch the RSS bucket id for the given indirection
1820 1.58 msaitoh * entry. Cap it at the number of configured buckets
1821 1.58 msaitoh * (which is num_queues.)
1822 1.58 msaitoh */
1823 1.58 msaitoh queue_id = rss_get_indirection_to_bucket(i);
1824 1.186 msaitoh queue_id = queue_id % sc->num_queues;
1825 1.58 msaitoh } else
1826 1.58 msaitoh queue_id = j;
1827 1.1 dyoung
1828 1.58 msaitoh /*
1829 1.58 msaitoh * The low 8 bits are for hash value (n+0);
1830 1.58 msaitoh * The next 8 bits are for hash value (n+1), etc.
1831 1.58 msaitoh */
1832 1.58 msaitoh reta >>= 8;
1833 1.58 msaitoh reta |= ((uint32_t)queue_id) << 24;
1834 1.58 msaitoh if ((i & 3) == 3) {
1835 1.58 msaitoh IXGBE_WRITE_REG(hw, IXGBE_VFRETA(i >> 2), reta);
1836 1.58 msaitoh reta = 0;
1837 1.58 msaitoh }
1838 1.58 msaitoh }
1839 1.21 msaitoh
1840 1.58 msaitoh /* Perform hash on these packet types */
1841 1.186 msaitoh if (sc->feat_en & IXGBE_FEATURE_RSS)
1842 1.58 msaitoh rss_hash_config = rss_gethashconfig();
1843 1.58 msaitoh else {
1844 1.58 msaitoh /*
1845 1.58 msaitoh * Disable UDP - IP fragments aren't currently being handled
1846 1.58 msaitoh * and so we end up with a mix of 2-tuple and 4-tuple
1847 1.58 msaitoh * traffic.
1848 1.58 msaitoh */
1849 1.58 msaitoh rss_hash_config = RSS_HASHTYPE_RSS_IPV4
1850 1.114 msaitoh | RSS_HASHTYPE_RSS_TCP_IPV4
1851 1.114 msaitoh | RSS_HASHTYPE_RSS_IPV6
1852 1.114 msaitoh | RSS_HASHTYPE_RSS_TCP_IPV6;
1853 1.58 msaitoh }
1854 1.58 msaitoh
1855 1.58 msaitoh mrqc = IXGBE_MRQC_RSSEN;
1856 1.58 msaitoh if (rss_hash_config & RSS_HASHTYPE_RSS_IPV4)
1857 1.58 msaitoh mrqc |= IXGBE_MRQC_RSS_FIELD_IPV4;
1858 1.58 msaitoh if (rss_hash_config & RSS_HASHTYPE_RSS_TCP_IPV4)
1859 1.58 msaitoh mrqc |= IXGBE_MRQC_RSS_FIELD_IPV4_TCP;
1860 1.58 msaitoh if (rss_hash_config & RSS_HASHTYPE_RSS_IPV6)
1861 1.58 msaitoh mrqc |= IXGBE_MRQC_RSS_FIELD_IPV6;
1862 1.58 msaitoh if (rss_hash_config & RSS_HASHTYPE_RSS_TCP_IPV6)
1863 1.58 msaitoh mrqc |= IXGBE_MRQC_RSS_FIELD_IPV6_TCP;
1864 1.58 msaitoh if (rss_hash_config & RSS_HASHTYPE_RSS_IPV6_EX)
1865 1.186 msaitoh device_printf(sc->dev, "%s: RSS_HASHTYPE_RSS_IPV6_EX "
1866 1.182 msaitoh "defined, but not supported\n", __func__);
1867 1.58 msaitoh if (rss_hash_config & RSS_HASHTYPE_RSS_TCP_IPV6_EX)
1868 1.186 msaitoh device_printf(sc->dev, "%s: RSS_HASHTYPE_RSS_TCP_IPV6_EX "
1869 1.182 msaitoh "defined, but not supported\n", __func__);
1870 1.58 msaitoh if (rss_hash_config & RSS_HASHTYPE_RSS_UDP_IPV4)
1871 1.58 msaitoh mrqc |= IXGBE_MRQC_RSS_FIELD_IPV4_UDP;
1872 1.58 msaitoh if (rss_hash_config & RSS_HASHTYPE_RSS_UDP_IPV6)
1873 1.58 msaitoh mrqc |= IXGBE_MRQC_RSS_FIELD_IPV6_UDP;
1874 1.58 msaitoh if (rss_hash_config & RSS_HASHTYPE_RSS_UDP_IPV6_EX)
1875 1.186 msaitoh device_printf(sc->dev, "%s: RSS_HASHTYPE_RSS_UDP_IPV6_EX "
1876 1.182 msaitoh "defined, but not supported\n", __func__);
1877 1.58 msaitoh IXGBE_WRITE_REG(hw, IXGBE_VFMRQC, mrqc);
1878 1.58 msaitoh } /* ixv_initialize_rss_mapping */
1879 1.58 msaitoh
1880 1.58 msaitoh
1881 1.58 msaitoh /************************************************************************
1882 1.58 msaitoh * ixv_initialize_receive_units - Setup receive registers and features.
1883 1.58 msaitoh ************************************************************************/
1884 1.21 msaitoh static void
1885 1.186 msaitoh ixv_initialize_receive_units(struct ixgbe_softc *sc)
1886 1.1 dyoung {
1887 1.186 msaitoh struct rx_ring *rxr = sc->rx_rings;
1888 1.186 msaitoh struct ixgbe_hw *hw = &sc->hw;
1889 1.186 msaitoh struct ifnet *ifp = sc->ifp;
1890 1.117 msaitoh u32 bufsz, psrtype;
1891 1.1 dyoung
1892 1.23 msaitoh if (ifp->if_mtu > ETHERMTU)
1893 1.23 msaitoh bufsz = 4096 >> IXGBE_SRRCTL_BSIZEPKT_SHIFT;
1894 1.23 msaitoh else
1895 1.23 msaitoh bufsz = 2048 >> IXGBE_SRRCTL_BSIZEPKT_SHIFT;
1896 1.1 dyoung
1897 1.58 msaitoh psrtype = IXGBE_PSRTYPE_TCPHDR
1898 1.114 msaitoh | IXGBE_PSRTYPE_UDPHDR
1899 1.114 msaitoh | IXGBE_PSRTYPE_IPV4HDR
1900 1.114 msaitoh | IXGBE_PSRTYPE_IPV6HDR
1901 1.114 msaitoh | IXGBE_PSRTYPE_L2HDR;
1902 1.58 msaitoh
1903 1.186 msaitoh if (sc->num_queues > 1)
1904 1.58 msaitoh psrtype |= 1 << 29;
1905 1.1 dyoung
1906 1.23 msaitoh IXGBE_WRITE_REG(hw, IXGBE_VFPSRTYPE, psrtype);
1907 1.23 msaitoh
1908 1.26 msaitoh /* Tell PF our max_frame size */
1909 1.186 msaitoh if (ixgbevf_rlpml_set_vf(hw, sc->max_frame_size) != 0) {
1910 1.186 msaitoh device_printf(sc->dev, "There is a problem with the PF "
1911 1.182 msaitoh "setup. It is likely the receive unit for this VF will "
1912 1.182 msaitoh "not function correctly.\n");
1913 1.58 msaitoh }
1914 1.1 dyoung
1915 1.186 msaitoh for (int i = 0; i < sc->num_queues; i++, rxr++) {
1916 1.1 dyoung u64 rdba = rxr->rxdma.dma_paddr;
1917 1.1 dyoung u32 reg, rxdctl;
1918 1.91 msaitoh int j = rxr->me;
1919 1.1 dyoung
1920 1.23 msaitoh /* Disable the queue */
1921 1.91 msaitoh rxdctl = IXGBE_READ_REG(hw, IXGBE_VFRXDCTL(j));
1922 1.28 msaitoh rxdctl &= ~IXGBE_RXDCTL_ENABLE;
1923 1.91 msaitoh IXGBE_WRITE_REG(hw, IXGBE_VFRXDCTL(j), rxdctl);
1924 1.91 msaitoh for (int k = 0; k < 10; k++) {
1925 1.91 msaitoh if (IXGBE_READ_REG(hw, IXGBE_VFRXDCTL(j)) &
1926 1.23 msaitoh IXGBE_RXDCTL_ENABLE)
1927 1.23 msaitoh msec_delay(1);
1928 1.23 msaitoh else
1929 1.23 msaitoh break;
1930 1.23 msaitoh }
1931 1.143 msaitoh IXGBE_WRITE_BARRIER(hw);
1932 1.1 dyoung /* Setup the Base and Length of the Rx Descriptor Ring */
1933 1.91 msaitoh IXGBE_WRITE_REG(hw, IXGBE_VFRDBAL(j),
1934 1.1 dyoung (rdba & 0x00000000ffffffffULL));
1935 1.91 msaitoh IXGBE_WRITE_REG(hw, IXGBE_VFRDBAH(j), (rdba >> 32));
1936 1.91 msaitoh IXGBE_WRITE_REG(hw, IXGBE_VFRDLEN(j),
1937 1.186 msaitoh sc->num_rx_desc * sizeof(union ixgbe_adv_rx_desc));
1938 1.1 dyoung
1939 1.23 msaitoh /* Reset the ring indices */
1940 1.23 msaitoh IXGBE_WRITE_REG(hw, IXGBE_VFRDH(rxr->me), 0);
1941 1.23 msaitoh IXGBE_WRITE_REG(hw, IXGBE_VFRDT(rxr->me), 0);
1942 1.23 msaitoh
1943 1.1 dyoung /* Set up the SRRCTL register */
1944 1.91 msaitoh reg = IXGBE_READ_REG(hw, IXGBE_VFSRRCTL(j));
1945 1.1 dyoung reg &= ~IXGBE_SRRCTL_BSIZEHDR_MASK;
1946 1.1 dyoung reg &= ~IXGBE_SRRCTL_BSIZEPKT_MASK;
1947 1.1 dyoung reg |= bufsz;
1948 1.21 msaitoh reg |= IXGBE_SRRCTL_DESCTYPE_ADV_ONEBUF;
1949 1.91 msaitoh IXGBE_WRITE_REG(hw, IXGBE_VFSRRCTL(j), reg);
1950 1.1 dyoung
1951 1.23 msaitoh /* Capture Rx Tail index */
1952 1.21 msaitoh rxr->tail = IXGBE_VFRDT(rxr->me);
1953 1.21 msaitoh
1954 1.21 msaitoh /* Do the queue enabling last */
1955 1.28 msaitoh rxdctl |= IXGBE_RXDCTL_ENABLE | IXGBE_RXDCTL_VME;
1956 1.91 msaitoh IXGBE_WRITE_REG(hw, IXGBE_VFRXDCTL(j), rxdctl);
1957 1.21 msaitoh for (int k = 0; k < 10; k++) {
1958 1.91 msaitoh if (IXGBE_READ_REG(hw, IXGBE_VFRXDCTL(j)) &
1959 1.21 msaitoh IXGBE_RXDCTL_ENABLE)
1960 1.21 msaitoh break;
1961 1.58 msaitoh msec_delay(1);
1962 1.21 msaitoh }
1963 1.143 msaitoh IXGBE_WRITE_BARRIER(hw);
1964 1.24 msaitoh
1965 1.24 msaitoh /* Set the Tail Pointer */
1966 1.88 msaitoh #ifdef DEV_NETMAP
1967 1.25 msaitoh /*
1968 1.25 msaitoh * In netmap mode, we must preserve the buffers made
1969 1.25 msaitoh * available to userspace before the if_init()
1970 1.25 msaitoh * (this is true by default on the TX side, because
1971 1.25 msaitoh * init makes all buffers available to userspace).
1972 1.25 msaitoh *
1973 1.25 msaitoh * netmap_reset() and the device specific routines
1974 1.25 msaitoh * (e.g. ixgbe_setup_receive_rings()) map these
1975 1.25 msaitoh * buffers at the end of the NIC ring, so here we
1976 1.25 msaitoh * must set the RDT (tail) register to make sure
1977 1.25 msaitoh * they are not overwritten.
1978 1.25 msaitoh *
1979 1.25 msaitoh * In this driver the NIC ring starts at RDH = 0,
1980 1.25 msaitoh * RDT points to the last slot available for reception (?),
1981 1.25 msaitoh * so RDT = num_rx_desc - 1 means the whole ring is available.
1982 1.25 msaitoh */
1983 1.186 msaitoh if ((sc->feat_en & IXGBE_FEATURE_NETMAP) &&
1984 1.58 msaitoh (ifp->if_capenable & IFCAP_NETMAP)) {
1985 1.186 msaitoh struct netmap_adapter *na = NA(sc->ifp);
1986 1.117 msaitoh struct netmap_kring *kring = na->rx_rings[i];
1987 1.25 msaitoh int t = na->num_rx_desc - 1 - nm_kr_rxspace(kring);
1988 1.25 msaitoh
1989 1.25 msaitoh IXGBE_WRITE_REG(hw, IXGBE_VFRDT(rxr->me), t);
1990 1.25 msaitoh } else
1991 1.25 msaitoh #endif /* DEV_NETMAP */
1992 1.25 msaitoh IXGBE_WRITE_REG(hw, IXGBE_VFRDT(rxr->me),
1993 1.186 msaitoh sc->num_rx_desc - 1);
1994 1.1 dyoung }
1995 1.1 dyoung
1996 1.186 msaitoh if (sc->hw.mac.type >= ixgbe_mac_X550_vf)
1997 1.186 msaitoh ixv_initialize_rss_mapping(sc);
1998 1.58 msaitoh } /* ixv_initialize_receive_units */
1999 1.1 dyoung
2000 1.58 msaitoh /************************************************************************
2001 1.83 msaitoh * ixv_sysctl_tdh_handler - Transmit Descriptor Head handler function
2002 1.83 msaitoh *
2003 1.83 msaitoh * Retrieves the TDH value from the hardware
2004 1.83 msaitoh ************************************************************************/
2005 1.113 msaitoh static int
2006 1.83 msaitoh ixv_sysctl_tdh_handler(SYSCTLFN_ARGS)
2007 1.83 msaitoh {
2008 1.83 msaitoh struct sysctlnode node = *rnode;
2009 1.83 msaitoh struct tx_ring *txr = (struct tx_ring *)node.sysctl_data;
2010 1.83 msaitoh uint32_t val;
2011 1.83 msaitoh
2012 1.83 msaitoh if (!txr)
2013 1.83 msaitoh return (0);
2014 1.83 msaitoh
2015 1.186 msaitoh val = IXGBE_READ_REG(&txr->sc->hw, IXGBE_VFTDH(txr->me));
2016 1.83 msaitoh node.sysctl_data = &val;
2017 1.83 msaitoh return sysctl_lookup(SYSCTLFN_CALL(&node));
2018 1.83 msaitoh } /* ixv_sysctl_tdh_handler */
2019 1.83 msaitoh
2020 1.83 msaitoh /************************************************************************
2021 1.83 msaitoh * ixgbe_sysctl_tdt_handler - Transmit Descriptor Tail handler function
2022 1.83 msaitoh *
2023 1.83 msaitoh * Retrieves the TDT value from the hardware
2024 1.83 msaitoh ************************************************************************/
2025 1.113 msaitoh static int
2026 1.83 msaitoh ixv_sysctl_tdt_handler(SYSCTLFN_ARGS)
2027 1.83 msaitoh {
2028 1.83 msaitoh struct sysctlnode node = *rnode;
2029 1.83 msaitoh struct tx_ring *txr = (struct tx_ring *)node.sysctl_data;
2030 1.83 msaitoh uint32_t val;
2031 1.83 msaitoh
2032 1.83 msaitoh if (!txr)
2033 1.83 msaitoh return (0);
2034 1.83 msaitoh
2035 1.186 msaitoh val = IXGBE_READ_REG(&txr->sc->hw, IXGBE_VFTDT(txr->me));
2036 1.83 msaitoh node.sysctl_data = &val;
2037 1.83 msaitoh return sysctl_lookup(SYSCTLFN_CALL(&node));
2038 1.83 msaitoh } /* ixv_sysctl_tdt_handler */
2039 1.83 msaitoh
2040 1.83 msaitoh /************************************************************************
2041 1.99 msaitoh * ixv_sysctl_next_to_check_handler - Receive Descriptor next to check
2042 1.99 msaitoh * handler function
2043 1.99 msaitoh *
2044 1.99 msaitoh * Retrieves the next_to_check value
2045 1.99 msaitoh ************************************************************************/
2046 1.113 msaitoh static int
2047 1.99 msaitoh ixv_sysctl_next_to_check_handler(SYSCTLFN_ARGS)
2048 1.99 msaitoh {
2049 1.99 msaitoh struct sysctlnode node = *rnode;
2050 1.99 msaitoh struct rx_ring *rxr = (struct rx_ring *)node.sysctl_data;
2051 1.99 msaitoh uint32_t val;
2052 1.99 msaitoh
2053 1.99 msaitoh if (!rxr)
2054 1.99 msaitoh return (0);
2055 1.99 msaitoh
2056 1.99 msaitoh val = rxr->next_to_check;
2057 1.99 msaitoh node.sysctl_data = &val;
2058 1.99 msaitoh return sysctl_lookup(SYSCTLFN_CALL(&node));
2059 1.99 msaitoh } /* ixv_sysctl_next_to_check_handler */
2060 1.99 msaitoh
2061 1.99 msaitoh /************************************************************************
2062 1.164 msaitoh * ixv_sysctl_next_to_refresh_handler - Receive Descriptor next to refresh
2063 1.164 msaitoh * handler function
2064 1.164 msaitoh *
2065 1.164 msaitoh * Retrieves the next_to_refresh value
2066 1.164 msaitoh ************************************************************************/
2067 1.164 msaitoh static int
2068 1.164 msaitoh ixv_sysctl_next_to_refresh_handler(SYSCTLFN_ARGS)
2069 1.164 msaitoh {
2070 1.164 msaitoh struct sysctlnode node = *rnode;
2071 1.164 msaitoh struct rx_ring *rxr = (struct rx_ring *)node.sysctl_data;
2072 1.186 msaitoh struct ixgbe_softc *sc;
2073 1.164 msaitoh uint32_t val;
2074 1.164 msaitoh
2075 1.164 msaitoh if (!rxr)
2076 1.164 msaitoh return (0);
2077 1.164 msaitoh
2078 1.186 msaitoh sc = rxr->sc;
2079 1.186 msaitoh if (ixgbe_fw_recovery_mode_swflag(sc))
2080 1.164 msaitoh return (EPERM);
2081 1.164 msaitoh
2082 1.164 msaitoh val = rxr->next_to_refresh;
2083 1.164 msaitoh node.sysctl_data = &val;
2084 1.164 msaitoh return sysctl_lookup(SYSCTLFN_CALL(&node));
2085 1.164 msaitoh } /* ixv_sysctl_next_to_refresh_handler */
2086 1.164 msaitoh
2087 1.164 msaitoh /************************************************************************
2088 1.83 msaitoh * ixv_sysctl_rdh_handler - Receive Descriptor Head handler function
2089 1.83 msaitoh *
2090 1.83 msaitoh * Retrieves the RDH value from the hardware
2091 1.83 msaitoh ************************************************************************/
2092 1.113 msaitoh static int
2093 1.83 msaitoh ixv_sysctl_rdh_handler(SYSCTLFN_ARGS)
2094 1.83 msaitoh {
2095 1.83 msaitoh struct sysctlnode node = *rnode;
2096 1.83 msaitoh struct rx_ring *rxr = (struct rx_ring *)node.sysctl_data;
2097 1.83 msaitoh uint32_t val;
2098 1.83 msaitoh
2099 1.83 msaitoh if (!rxr)
2100 1.83 msaitoh return (0);
2101 1.83 msaitoh
2102 1.186 msaitoh val = IXGBE_READ_REG(&rxr->sc->hw, IXGBE_VFRDH(rxr->me));
2103 1.83 msaitoh node.sysctl_data = &val;
2104 1.83 msaitoh return sysctl_lookup(SYSCTLFN_CALL(&node));
2105 1.83 msaitoh } /* ixv_sysctl_rdh_handler */
2106 1.83 msaitoh
2107 1.83 msaitoh /************************************************************************
2108 1.83 msaitoh * ixv_sysctl_rdt_handler - Receive Descriptor Tail handler function
2109 1.83 msaitoh *
2110 1.83 msaitoh * Retrieves the RDT value from the hardware
2111 1.83 msaitoh ************************************************************************/
2112 1.113 msaitoh static int
2113 1.83 msaitoh ixv_sysctl_rdt_handler(SYSCTLFN_ARGS)
2114 1.83 msaitoh {
2115 1.83 msaitoh struct sysctlnode node = *rnode;
2116 1.83 msaitoh struct rx_ring *rxr = (struct rx_ring *)node.sysctl_data;
2117 1.83 msaitoh uint32_t val;
2118 1.83 msaitoh
2119 1.83 msaitoh if (!rxr)
2120 1.83 msaitoh return (0);
2121 1.83 msaitoh
2122 1.186 msaitoh val = IXGBE_READ_REG(&rxr->sc->hw, IXGBE_VFRDT(rxr->me));
2123 1.83 msaitoh node.sysctl_data = &val;
2124 1.83 msaitoh return sysctl_lookup(SYSCTLFN_CALL(&node));
2125 1.83 msaitoh } /* ixv_sysctl_rdt_handler */
2126 1.83 msaitoh
2127 1.126 msaitoh static void
2128 1.186 msaitoh ixv_setup_vlan_tagging(struct ixgbe_softc *sc)
2129 1.1 dyoung {
2130 1.186 msaitoh struct ethercom *ec = &sc->osdep.ec;
2131 1.186 msaitoh struct ixgbe_hw *hw = &sc->hw;
2132 1.114 msaitoh struct rx_ring *rxr;
2133 1.126 msaitoh u32 ctrl;
2134 1.126 msaitoh int i;
2135 1.110 msaitoh bool hwtagging;
2136 1.1 dyoung
2137 1.111 msaitoh /* Enable HW tagging only if any vlan is attached */
2138 1.110 msaitoh hwtagging = (ec->ec_capenable & ETHERCAP_VLAN_HWTAGGING)
2139 1.111 msaitoh && VLAN_ATTACHED(ec);
2140 1.1 dyoung
2141 1.1 dyoung /* Enable the queues */
2142 1.186 msaitoh for (i = 0; i < sc->num_queues; i++) {
2143 1.186 msaitoh rxr = &sc->rx_rings[i];
2144 1.65 msaitoh ctrl = IXGBE_READ_REG(hw, IXGBE_VFRXDCTL(rxr->me));
2145 1.110 msaitoh if (hwtagging)
2146 1.110 msaitoh ctrl |= IXGBE_RXDCTL_VME;
2147 1.110 msaitoh else
2148 1.110 msaitoh ctrl &= ~IXGBE_RXDCTL_VME;
2149 1.65 msaitoh IXGBE_WRITE_REG(hw, IXGBE_VFRXDCTL(rxr->me), ctrl);
2150 1.26 msaitoh /*
2151 1.26 msaitoh * Let Rx path know that it needs to store VLAN tag
2152 1.26 msaitoh * as part of extra mbuf info.
2153 1.26 msaitoh */
2154 1.110 msaitoh rxr->vtag_strip = hwtagging ? TRUE : FALSE;
2155 1.1 dyoung }
2156 1.126 msaitoh } /* ixv_setup_vlan_tagging */
2157 1.126 msaitoh
2158 1.126 msaitoh /************************************************************************
2159 1.126 msaitoh * ixv_setup_vlan_support
2160 1.126 msaitoh ************************************************************************/
2161 1.126 msaitoh static int
2162 1.186 msaitoh ixv_setup_vlan_support(struct ixgbe_softc *sc)
2163 1.126 msaitoh {
2164 1.186 msaitoh struct ethercom *ec = &sc->osdep.ec;
2165 1.186 msaitoh struct ixgbe_hw *hw = &sc->hw;
2166 1.126 msaitoh u32 vid, vfta, retry;
2167 1.126 msaitoh struct vlanid_list *vlanidp;
2168 1.126 msaitoh int rv, error = 0;
2169 1.1 dyoung
2170 1.126 msaitoh /*
2171 1.126 msaitoh * This function is called from both if_init and ifflags_cb()
2172 1.126 msaitoh * on NetBSD.
2173 1.126 msaitoh */
2174 1.126 msaitoh
2175 1.126 msaitoh /*
2176 1.126 msaitoh * Part 1:
2177 1.126 msaitoh * Setup VLAN HW tagging
2178 1.126 msaitoh */
2179 1.186 msaitoh ixv_setup_vlan_tagging(sc);
2180 1.126 msaitoh
2181 1.126 msaitoh if (!VLAN_ATTACHED(ec))
2182 1.120 msaitoh return 0;
2183 1.120 msaitoh
2184 1.126 msaitoh /*
2185 1.126 msaitoh * Part 2:
2186 1.126 msaitoh * Setup VLAN HW filter
2187 1.126 msaitoh */
2188 1.120 msaitoh /* Cleanup shadow_vfta */
2189 1.65 msaitoh for (int i = 0; i < IXGBE_VFTA_SIZE; i++)
2190 1.186 msaitoh sc->shadow_vfta[i] = 0;
2191 1.120 msaitoh /* Generate shadow_vfta from ec_vids */
2192 1.127 msaitoh ETHER_LOCK(ec);
2193 1.120 msaitoh SIMPLEQ_FOREACH(vlanidp, &ec->ec_vids, vid_list) {
2194 1.120 msaitoh uint32_t idx;
2195 1.120 msaitoh
2196 1.120 msaitoh idx = vlanidp->vid / 32;
2197 1.120 msaitoh KASSERT(idx < IXGBE_VFTA_SIZE);
2198 1.186 msaitoh sc->shadow_vfta[idx] |= (u32)1 << (vlanidp->vid % 32);
2199 1.120 msaitoh }
2200 1.127 msaitoh ETHER_UNLOCK(ec);
2201 1.142 msaitoh
2202 1.1 dyoung /*
2203 1.58 msaitoh * A soft reset zero's out the VFTA, so
2204 1.58 msaitoh * we need to repopulate it now.
2205 1.58 msaitoh */
2206 1.21 msaitoh for (int i = 0; i < IXGBE_VFTA_SIZE; i++) {
2207 1.186 msaitoh if (sc->shadow_vfta[i] == 0)
2208 1.1 dyoung continue;
2209 1.186 msaitoh vfta = sc->shadow_vfta[i];
2210 1.1 dyoung /*
2211 1.58 msaitoh * Reconstruct the vlan id's
2212 1.58 msaitoh * based on the bits set in each
2213 1.58 msaitoh * of the array ints.
2214 1.58 msaitoh */
2215 1.26 msaitoh for (int j = 0; j < 32; j++) {
2216 1.1 dyoung retry = 0;
2217 1.123 msaitoh if ((vfta & ((u32)1 << j)) == 0)
2218 1.1 dyoung continue;
2219 1.1 dyoung vid = (i * 32) + j;
2220 1.142 msaitoh
2221 1.1 dyoung /* Call the shared code mailbox routine */
2222 1.120 msaitoh while ((rv = hw->mac.ops.set_vfta(hw, vid, 0, TRUE,
2223 1.120 msaitoh FALSE)) != 0) {
2224 1.120 msaitoh if (++retry > 5) {
2225 1.186 msaitoh device_printf(sc->dev,
2226 1.120 msaitoh "%s: max retry exceeded\n",
2227 1.120 msaitoh __func__);
2228 1.1 dyoung break;
2229 1.120 msaitoh }
2230 1.120 msaitoh }
2231 1.120 msaitoh if (rv != 0) {
2232 1.186 msaitoh device_printf(sc->dev,
2233 1.120 msaitoh "failed to set vlan %d\n", vid);
2234 1.120 msaitoh error = EACCES;
2235 1.1 dyoung }
2236 1.1 dyoung }
2237 1.1 dyoung }
2238 1.120 msaitoh return error;
2239 1.58 msaitoh } /* ixv_setup_vlan_support */
2240 1.1 dyoung
2241 1.120 msaitoh static int
2242 1.120 msaitoh ixv_vlan_cb(struct ethercom *ec, uint16_t vid, bool set)
2243 1.120 msaitoh {
2244 1.120 msaitoh struct ifnet *ifp = &ec->ec_if;
2245 1.186 msaitoh struct ixgbe_softc *sc = ifp->if_softc;
2246 1.120 msaitoh int rv;
2247 1.120 msaitoh
2248 1.120 msaitoh if (set)
2249 1.186 msaitoh rv = ixv_register_vlan(sc, vid);
2250 1.120 msaitoh else
2251 1.186 msaitoh rv = ixv_unregister_vlan(sc, vid);
2252 1.120 msaitoh
2253 1.126 msaitoh if (rv != 0)
2254 1.126 msaitoh return rv;
2255 1.126 msaitoh
2256 1.126 msaitoh /*
2257 1.126 msaitoh * Control VLAN HW tagging when ec_nvlan is changed from 1 to 0
2258 1.126 msaitoh * or 0 to 1.
2259 1.126 msaitoh */
2260 1.126 msaitoh if ((set && (ec->ec_nvlans == 1)) || (!set && (ec->ec_nvlans == 0)))
2261 1.186 msaitoh ixv_setup_vlan_tagging(sc);
2262 1.126 msaitoh
2263 1.120 msaitoh return rv;
2264 1.120 msaitoh }
2265 1.120 msaitoh
2266 1.58 msaitoh /************************************************************************
2267 1.58 msaitoh * ixv_register_vlan
2268 1.58 msaitoh *
2269 1.58 msaitoh * Run via a vlan config EVENT, it enables us to use the
2270 1.58 msaitoh * HW Filter table since we can get the vlan id. This just
2271 1.58 msaitoh * creates the entry in the soft version of the VFTA, init
2272 1.58 msaitoh * will repopulate the real table.
2273 1.58 msaitoh ************************************************************************/
2274 1.120 msaitoh static int
2275 1.186 msaitoh ixv_register_vlan(struct ixgbe_softc *sc, u16 vtag)
2276 1.1 dyoung {
2277 1.186 msaitoh struct ixgbe_hw *hw = &sc->hw;
2278 1.1 dyoung u16 index, bit;
2279 1.120 msaitoh int error;
2280 1.1 dyoung
2281 1.26 msaitoh if ((vtag == 0) || (vtag > 4095)) /* Invalid */
2282 1.120 msaitoh return EINVAL;
2283 1.186 msaitoh IXGBE_CORE_LOCK(sc);
2284 1.1 dyoung index = (vtag >> 5) & 0x7F;
2285 1.1 dyoung bit = vtag & 0x1F;
2286 1.186 msaitoh sc->shadow_vfta[index] |= ((u32)1 << bit);
2287 1.120 msaitoh error = hw->mac.ops.set_vfta(hw, vtag, 0, true, false);
2288 1.186 msaitoh IXGBE_CORE_UNLOCK(sc);
2289 1.120 msaitoh
2290 1.120 msaitoh if (error != 0) {
2291 1.186 msaitoh device_printf(sc->dev, "failed to register vlan %hu\n", vtag);
2292 1.120 msaitoh error = EACCES;
2293 1.120 msaitoh }
2294 1.120 msaitoh return error;
2295 1.58 msaitoh } /* ixv_register_vlan */
2296 1.1 dyoung
2297 1.58 msaitoh /************************************************************************
2298 1.58 msaitoh * ixv_unregister_vlan
2299 1.58 msaitoh *
2300 1.58 msaitoh * Run via a vlan unconfig EVENT, remove our entry
2301 1.58 msaitoh * in the soft vfta.
2302 1.58 msaitoh ************************************************************************/
2303 1.120 msaitoh static int
2304 1.186 msaitoh ixv_unregister_vlan(struct ixgbe_softc *sc, u16 vtag)
2305 1.1 dyoung {
2306 1.186 msaitoh struct ixgbe_hw *hw = &sc->hw;
2307 1.1 dyoung u16 index, bit;
2308 1.142 msaitoh int error;
2309 1.1 dyoung
2310 1.58 msaitoh if ((vtag == 0) || (vtag > 4095)) /* Invalid */
2311 1.120 msaitoh return EINVAL;
2312 1.1 dyoung
2313 1.186 msaitoh IXGBE_CORE_LOCK(sc);
2314 1.1 dyoung index = (vtag >> 5) & 0x7F;
2315 1.1 dyoung bit = vtag & 0x1F;
2316 1.186 msaitoh sc->shadow_vfta[index] &= ~((u32)1 << bit);
2317 1.120 msaitoh error = hw->mac.ops.set_vfta(hw, vtag, 0, false, false);
2318 1.186 msaitoh IXGBE_CORE_UNLOCK(sc);
2319 1.120 msaitoh
2320 1.120 msaitoh if (error != 0) {
2321 1.186 msaitoh device_printf(sc->dev, "failed to unregister vlan %hu\n",
2322 1.120 msaitoh vtag);
2323 1.120 msaitoh error = EIO;
2324 1.120 msaitoh }
2325 1.120 msaitoh return error;
2326 1.58 msaitoh } /* ixv_unregister_vlan */
2327 1.1 dyoung
2328 1.58 msaitoh /************************************************************************
2329 1.58 msaitoh * ixv_enable_intr
2330 1.58 msaitoh ************************************************************************/
2331 1.1 dyoung static void
2332 1.186 msaitoh ixv_enable_intr(struct ixgbe_softc *sc)
2333 1.1 dyoung {
2334 1.186 msaitoh struct ixgbe_hw *hw = &sc->hw;
2335 1.186 msaitoh struct ix_queue *que = sc->queues;
2336 1.114 msaitoh u32 mask;
2337 1.68 msaitoh int i;
2338 1.1 dyoung
2339 1.68 msaitoh /* For VTEIAC */
2340 1.186 msaitoh mask = (1 << sc->vector);
2341 1.186 msaitoh for (i = 0; i < sc->num_queues; i++, que++)
2342 1.68 msaitoh mask |= (1 << que->msix);
2343 1.1 dyoung IXGBE_WRITE_REG(hw, IXGBE_VTEIAC, mask);
2344 1.1 dyoung
2345 1.68 msaitoh /* For VTEIMS */
2346 1.186 msaitoh IXGBE_WRITE_REG(hw, IXGBE_VTEIMS, (1 << sc->vector));
2347 1.186 msaitoh que = sc->queues;
2348 1.186 msaitoh for (i = 0; i < sc->num_queues; i++, que++)
2349 1.186 msaitoh ixv_enable_queue(sc, que->msix);
2350 1.1 dyoung
2351 1.1 dyoung IXGBE_WRITE_FLUSH(hw);
2352 1.58 msaitoh } /* ixv_enable_intr */
2353 1.1 dyoung
2354 1.58 msaitoh /************************************************************************
2355 1.58 msaitoh * ixv_disable_intr
2356 1.58 msaitoh ************************************************************************/
2357 1.1 dyoung static void
2358 1.186 msaitoh ixv_disable_intr(struct ixgbe_softc *sc)
2359 1.1 dyoung {
2360 1.186 msaitoh struct ix_queue *que = sc->queues;
2361 1.82 knakahar
2362 1.186 msaitoh IXGBE_WRITE_REG(&sc->hw, IXGBE_VTEIAC, 0);
2363 1.82 knakahar
2364 1.82 knakahar /* disable interrupts other than queues */
2365 1.186 msaitoh IXGBE_WRITE_REG(&sc->hw, IXGBE_VTEIMC, sc->vector);
2366 1.82 knakahar
2367 1.186 msaitoh for (int i = 0; i < sc->num_queues; i++, que++)
2368 1.186 msaitoh ixv_disable_queue(sc, que->msix);
2369 1.82 knakahar
2370 1.186 msaitoh IXGBE_WRITE_FLUSH(&sc->hw);
2371 1.58 msaitoh } /* ixv_disable_intr */
2372 1.1 dyoung
2373 1.58 msaitoh /************************************************************************
2374 1.58 msaitoh * ixv_set_ivar
2375 1.58 msaitoh *
2376 1.58 msaitoh * Setup the correct IVAR register for a particular MSI-X interrupt
2377 1.58 msaitoh * - entry is the register array entry
2378 1.58 msaitoh * - vector is the MSI-X vector for this queue
2379 1.58 msaitoh * - type is RX/TX/MISC
2380 1.58 msaitoh ************************************************************************/
2381 1.1 dyoung static void
2382 1.186 msaitoh ixv_set_ivar(struct ixgbe_softc *sc, u8 entry, u8 vector, s8 type)
2383 1.1 dyoung {
2384 1.186 msaitoh struct ixgbe_hw *hw = &sc->hw;
2385 1.114 msaitoh u32 ivar, index;
2386 1.1 dyoung
2387 1.1 dyoung vector |= IXGBE_IVAR_ALLOC_VAL;
2388 1.1 dyoung
2389 1.1 dyoung if (type == -1) { /* MISC IVAR */
2390 1.1 dyoung ivar = IXGBE_READ_REG(hw, IXGBE_VTIVAR_MISC);
2391 1.1 dyoung ivar &= ~0xFF;
2392 1.1 dyoung ivar |= vector;
2393 1.1 dyoung IXGBE_WRITE_REG(hw, IXGBE_VTIVAR_MISC, ivar);
2394 1.114 msaitoh } else { /* RX/TX IVARS */
2395 1.1 dyoung index = (16 * (entry & 1)) + (8 * type);
2396 1.1 dyoung ivar = IXGBE_READ_REG(hw, IXGBE_VTIVAR(entry >> 1));
2397 1.121 msaitoh ivar &= ~(0xffUL << index);
2398 1.121 msaitoh ivar |= ((u32)vector << index);
2399 1.1 dyoung IXGBE_WRITE_REG(hw, IXGBE_VTIVAR(entry >> 1), ivar);
2400 1.1 dyoung }
2401 1.58 msaitoh } /* ixv_set_ivar */
2402 1.1 dyoung
2403 1.58 msaitoh /************************************************************************
2404 1.58 msaitoh * ixv_configure_ivars
2405 1.58 msaitoh ************************************************************************/
2406 1.1 dyoung static void
2407 1.186 msaitoh ixv_configure_ivars(struct ixgbe_softc *sc)
2408 1.1 dyoung {
2409 1.186 msaitoh struct ix_queue *que = sc->queues;
2410 1.1 dyoung
2411 1.80 msaitoh /* XXX We should sync EITR value calculation with ixgbe.c? */
2412 1.80 msaitoh
2413 1.186 msaitoh for (int i = 0; i < sc->num_queues; i++, que++) {
2414 1.1 dyoung /* First the RX queue entry */
2415 1.186 msaitoh ixv_set_ivar(sc, i, que->msix, 0);
2416 1.1 dyoung /* ... and the TX */
2417 1.186 msaitoh ixv_set_ivar(sc, i, que->msix, 1);
2418 1.1 dyoung /* Set an initial value in EITR */
2419 1.186 msaitoh ixv_eitr_write(sc, que->msix, IXGBE_EITR_DEFAULT);
2420 1.1 dyoung }
2421 1.1 dyoung
2422 1.21 msaitoh /* For the mailbox interrupt */
2423 1.186 msaitoh ixv_set_ivar(sc, 1, sc->vector, -1);
2424 1.58 msaitoh } /* ixv_configure_ivars */
2425 1.1 dyoung
2426 1.1 dyoung
2427 1.58 msaitoh /************************************************************************
2428 1.176 msaitoh * ixv_init_stats
2429 1.58 msaitoh *
2430 1.58 msaitoh * The VF stats registers never have a truly virgin
2431 1.176 msaitoh * starting point, so this routine save initial vaules to
2432 1.176 msaitoh * last_<REGNAME>.
2433 1.58 msaitoh ************************************************************************/
2434 1.1 dyoung static void
2435 1.186 msaitoh ixv_init_stats(struct ixgbe_softc *sc)
2436 1.1 dyoung {
2437 1.186 msaitoh struct ixgbe_hw *hw = &sc->hw;
2438 1.57 msaitoh
2439 1.186 msaitoh sc->stats.vf.last_vfgprc = IXGBE_READ_REG(hw, IXGBE_VFGPRC);
2440 1.186 msaitoh sc->stats.vf.last_vfgorc = IXGBE_READ_REG(hw, IXGBE_VFGORC_LSB);
2441 1.186 msaitoh sc->stats.vf.last_vfgorc |=
2442 1.1 dyoung (((u64)(IXGBE_READ_REG(hw, IXGBE_VFGORC_MSB))) << 32);
2443 1.1 dyoung
2444 1.186 msaitoh sc->stats.vf.last_vfgptc = IXGBE_READ_REG(hw, IXGBE_VFGPTC);
2445 1.186 msaitoh sc->stats.vf.last_vfgotc = IXGBE_READ_REG(hw, IXGBE_VFGOTC_LSB);
2446 1.186 msaitoh sc->stats.vf.last_vfgotc |=
2447 1.1 dyoung (((u64)(IXGBE_READ_REG(hw, IXGBE_VFGOTC_MSB))) << 32);
2448 1.1 dyoung
2449 1.186 msaitoh sc->stats.vf.last_vfmprc = IXGBE_READ_REG(hw, IXGBE_VFMPRC);
2450 1.58 msaitoh } /* ixv_init_stats */
2451 1.1 dyoung
2452 1.1 dyoung #define UPDATE_STAT_32(reg, last, count) \
2453 1.114 msaitoh { \
2454 1.58 msaitoh u32 current = IXGBE_READ_REG(hw, (reg)); \
2455 1.178 msaitoh IXGBE_EVC_ADD(&count, current - (last)); \
2456 1.58 msaitoh (last) = current; \
2457 1.1 dyoung }
2458 1.1 dyoung
2459 1.178 msaitoh #define UPDATE_STAT_36(lsb, msb, last, count) \
2460 1.178 msaitoh { \
2461 1.178 msaitoh u64 cur_lsb = IXGBE_READ_REG(hw, (lsb)); \
2462 1.178 msaitoh u64 cur_msb = IXGBE_READ_REG(hw, (msb)); \
2463 1.178 msaitoh u64 current = ((cur_msb << 32) | cur_lsb); \
2464 1.178 msaitoh if (current < (last)) \
2465 1.178 msaitoh IXGBE_EVC_ADD(&count, current + __BIT(36) - (last)); \
2466 1.178 msaitoh else \
2467 1.178 msaitoh IXGBE_EVC_ADD(&count, current - (last)); \
2468 1.178 msaitoh (last) = current; \
2469 1.1 dyoung }
2470 1.1 dyoung
2471 1.58 msaitoh /************************************************************************
2472 1.58 msaitoh * ixv_update_stats - Update the board statistics counters.
2473 1.58 msaitoh ************************************************************************/
2474 1.1 dyoung void
2475 1.186 msaitoh ixv_update_stats(struct ixgbe_softc *sc)
2476 1.1 dyoung {
2477 1.186 msaitoh struct ixgbe_hw *hw = &sc->hw;
2478 1.186 msaitoh struct ixgbevf_hw_stats *stats = &sc->stats.vf;
2479 1.1 dyoung
2480 1.88 msaitoh UPDATE_STAT_32(IXGBE_VFGPRC, stats->last_vfgprc, stats->vfgprc);
2481 1.88 msaitoh UPDATE_STAT_32(IXGBE_VFGPTC, stats->last_vfgptc, stats->vfgptc);
2482 1.88 msaitoh UPDATE_STAT_36(IXGBE_VFGORC_LSB, IXGBE_VFGORC_MSB, stats->last_vfgorc,
2483 1.58 msaitoh stats->vfgorc);
2484 1.88 msaitoh UPDATE_STAT_36(IXGBE_VFGOTC_LSB, IXGBE_VFGOTC_MSB, stats->last_vfgotc,
2485 1.58 msaitoh stats->vfgotc);
2486 1.88 msaitoh UPDATE_STAT_32(IXGBE_VFMPRC, stats->last_vfmprc, stats->vfmprc);
2487 1.58 msaitoh
2488 1.146 msaitoh /* VF doesn't count errors by hardware */
2489 1.146 msaitoh
2490 1.58 msaitoh } /* ixv_update_stats */
2491 1.1 dyoung
2492 1.83 msaitoh /************************************************************************
2493 1.83 msaitoh * ixv_sysctl_interrupt_rate_handler
2494 1.83 msaitoh ************************************************************************/
2495 1.83 msaitoh static int
2496 1.83 msaitoh ixv_sysctl_interrupt_rate_handler(SYSCTLFN_ARGS)
2497 1.83 msaitoh {
2498 1.83 msaitoh struct sysctlnode node = *rnode;
2499 1.83 msaitoh struct ix_queue *que = (struct ix_queue *)node.sysctl_data;
2500 1.186 msaitoh struct ixgbe_softc *sc = que->sc;
2501 1.83 msaitoh uint32_t reg, usec, rate;
2502 1.83 msaitoh int error;
2503 1.83 msaitoh
2504 1.83 msaitoh if (que == NULL)
2505 1.83 msaitoh return 0;
2506 1.186 msaitoh reg = IXGBE_READ_REG(&que->sc->hw, IXGBE_VTEITR(que->msix));
2507 1.83 msaitoh usec = ((reg & 0x0FF8) >> 3);
2508 1.83 msaitoh if (usec > 0)
2509 1.83 msaitoh rate = 500000 / usec;
2510 1.83 msaitoh else
2511 1.83 msaitoh rate = 0;
2512 1.83 msaitoh node.sysctl_data = &rate;
2513 1.83 msaitoh error = sysctl_lookup(SYSCTLFN_CALL(&node));
2514 1.83 msaitoh if (error || newp == NULL)
2515 1.83 msaitoh return error;
2516 1.83 msaitoh reg &= ~0xfff; /* default, no limitation */
2517 1.83 msaitoh if (rate > 0 && rate < 500000) {
2518 1.83 msaitoh if (rate < 1000)
2519 1.83 msaitoh rate = 1000;
2520 1.148 msaitoh reg |= ((4000000 / rate) & 0xff8);
2521 1.83 msaitoh /*
2522 1.83 msaitoh * When RSC is used, ITR interval must be larger than
2523 1.83 msaitoh * RSC_DELAY. Currently, we use 2us for RSC_DELAY.
2524 1.83 msaitoh * The minimum value is always greater than 2us on 100M
2525 1.83 msaitoh * (and 10M?(not documented)), but it's not on 1G and higher.
2526 1.83 msaitoh */
2527 1.186 msaitoh if ((sc->link_speed != IXGBE_LINK_SPEED_100_FULL)
2528 1.186 msaitoh && (sc->link_speed != IXGBE_LINK_SPEED_10_FULL)) {
2529 1.186 msaitoh if ((sc->num_queues > 1)
2530 1.83 msaitoh && (reg < IXGBE_MIN_RSC_EITR_10G1G))
2531 1.83 msaitoh return EINVAL;
2532 1.83 msaitoh }
2533 1.191 msaitoh sc->max_interrupt_rate = rate;
2534 1.83 msaitoh } else
2535 1.191 msaitoh sc->max_interrupt_rate = 0;
2536 1.186 msaitoh ixv_eitr_write(sc, que->msix, reg);
2537 1.83 msaitoh
2538 1.83 msaitoh return (0);
2539 1.83 msaitoh } /* ixv_sysctl_interrupt_rate_handler */
2540 1.83 msaitoh
2541 1.3 msaitoh const struct sysctlnode *
2542 1.186 msaitoh ixv_sysctl_instance(struct ixgbe_softc *sc)
2543 1.3 msaitoh {
2544 1.3 msaitoh const char *dvname;
2545 1.3 msaitoh struct sysctllog **log;
2546 1.3 msaitoh int rc;
2547 1.3 msaitoh const struct sysctlnode *rnode;
2548 1.3 msaitoh
2549 1.186 msaitoh log = &sc->sysctllog;
2550 1.186 msaitoh dvname = device_xname(sc->dev);
2551 1.3 msaitoh
2552 1.3 msaitoh if ((rc = sysctl_createv(log, 0, NULL, &rnode,
2553 1.3 msaitoh 0, CTLTYPE_NODE, dvname,
2554 1.3 msaitoh SYSCTL_DESCR("ixv information and settings"),
2555 1.3 msaitoh NULL, 0, NULL, 0, CTL_HW, CTL_CREATE, CTL_EOL)) != 0)
2556 1.3 msaitoh goto err;
2557 1.3 msaitoh
2558 1.3 msaitoh return rnode;
2559 1.3 msaitoh err:
2560 1.186 msaitoh device_printf(sc->dev,
2561 1.129 msaitoh "%s: sysctl_createv failed, rc = %d\n", __func__, rc);
2562 1.3 msaitoh return NULL;
2563 1.3 msaitoh }
2564 1.48 msaitoh
2565 1.48 msaitoh static void
2566 1.186 msaitoh ixv_add_device_sysctls(struct ixgbe_softc *sc)
2567 1.48 msaitoh {
2568 1.48 msaitoh struct sysctllog **log;
2569 1.48 msaitoh const struct sysctlnode *rnode, *cnode;
2570 1.48 msaitoh device_t dev;
2571 1.48 msaitoh
2572 1.186 msaitoh dev = sc->dev;
2573 1.186 msaitoh log = &sc->sysctllog;
2574 1.48 msaitoh
2575 1.186 msaitoh if ((rnode = ixv_sysctl_instance(sc)) == NULL) {
2576 1.48 msaitoh aprint_error_dev(dev, "could not create sysctl root\n");
2577 1.48 msaitoh return;
2578 1.48 msaitoh }
2579 1.48 msaitoh
2580 1.48 msaitoh if (sysctl_createv(log, 0, &rnode, &cnode,
2581 1.158 msaitoh CTLFLAG_READWRITE, CTLTYPE_INT, "debug",
2582 1.158 msaitoh SYSCTL_DESCR("Debug Info"),
2583 1.186 msaitoh ixv_sysctl_debug, 0, (void *)sc, 0, CTL_CREATE, CTL_EOL) != 0)
2584 1.48 msaitoh aprint_error_dev(dev, "could not create sysctl\n");
2585 1.48 msaitoh
2586 1.48 msaitoh if (sysctl_createv(log, 0, &rnode, &cnode,
2587 1.163 msaitoh CTLFLAG_READWRITE, CTLTYPE_INT,
2588 1.163 msaitoh "rx_copy_len", SYSCTL_DESCR("RX Copy Length"),
2589 1.163 msaitoh ixv_sysctl_rx_copy_len, 0,
2590 1.186 msaitoh (void *)sc, 0, CTL_CREATE, CTL_EOL) != 0)
2591 1.163 msaitoh aprint_error_dev(dev, "could not create sysctl\n");
2592 1.163 msaitoh
2593 1.163 msaitoh if (sysctl_createv(log, 0, &rnode, &cnode,
2594 1.180 msaitoh CTLFLAG_READONLY, CTLTYPE_INT,
2595 1.180 msaitoh "num_tx_desc", SYSCTL_DESCR("Number of TX descriptors"),
2596 1.186 msaitoh NULL, 0, &sc->num_tx_desc, 0, CTL_CREATE, CTL_EOL) != 0)
2597 1.180 msaitoh aprint_error_dev(dev, "could not create sysctl\n");
2598 1.180 msaitoh
2599 1.180 msaitoh if (sysctl_createv(log, 0, &rnode, &cnode,
2600 1.180 msaitoh CTLFLAG_READONLY, CTLTYPE_INT,
2601 1.180 msaitoh "num_rx_desc", SYSCTL_DESCR("Number of RX descriptors"),
2602 1.186 msaitoh NULL, 0, &sc->num_rx_desc, 0, CTL_CREATE, CTL_EOL) != 0)
2603 1.180 msaitoh aprint_error_dev(dev, "could not create sysctl\n");
2604 1.180 msaitoh
2605 1.180 msaitoh if (sysctl_createv(log, 0, &rnode, &cnode,
2606 1.179 msaitoh CTLFLAG_READWRITE, CTLTYPE_INT, "rx_process_limit",
2607 1.179 msaitoh SYSCTL_DESCR("max number of RX packets to process"),
2608 1.186 msaitoh ixv_sysctl_rx_process_limit, 0, (void *)sc, 0, CTL_CREATE,
2609 1.179 msaitoh CTL_EOL) != 0)
2610 1.179 msaitoh aprint_error_dev(dev, "could not create sysctl\n");
2611 1.179 msaitoh
2612 1.179 msaitoh if (sysctl_createv(log, 0, &rnode, &cnode,
2613 1.179 msaitoh CTLFLAG_READWRITE, CTLTYPE_INT, "tx_process_limit",
2614 1.179 msaitoh SYSCTL_DESCR("max number of TX packets to process"),
2615 1.186 msaitoh ixv_sysctl_tx_process_limit, 0, (void *)sc, 0, CTL_CREATE,
2616 1.179 msaitoh CTL_EOL) != 0)
2617 1.179 msaitoh aprint_error_dev(dev, "could not create sysctl\n");
2618 1.179 msaitoh
2619 1.179 msaitoh if (sysctl_createv(log, 0, &rnode, &cnode,
2620 1.158 msaitoh CTLFLAG_READWRITE, CTLTYPE_BOOL, "enable_aim",
2621 1.158 msaitoh SYSCTL_DESCR("Interrupt Moderation"),
2622 1.186 msaitoh NULL, 0, &sc->enable_aim, 0, CTL_CREATE, CTL_EOL) != 0)
2623 1.48 msaitoh aprint_error_dev(dev, "could not create sysctl\n");
2624 1.84 knakahar
2625 1.84 knakahar if (sysctl_createv(log, 0, &rnode, &cnode,
2626 1.158 msaitoh CTLFLAG_READWRITE, CTLTYPE_BOOL, "txrx_workqueue",
2627 1.158 msaitoh SYSCTL_DESCR("Use workqueue for packet processing"),
2628 1.186 msaitoh NULL, 0, &sc->txrx_use_workqueue, 0, CTL_CREATE, CTL_EOL)
2629 1.158 msaitoh != 0)
2630 1.84 knakahar aprint_error_dev(dev, "could not create sysctl\n");
2631 1.48 msaitoh }
2632 1.48 msaitoh
2633 1.58 msaitoh /************************************************************************
2634 1.58 msaitoh * ixv_add_stats_sysctls - Add statistic sysctls for the VF.
2635 1.58 msaitoh ************************************************************************/
2636 1.48 msaitoh static void
2637 1.186 msaitoh ixv_add_stats_sysctls(struct ixgbe_softc *sc)
2638 1.48 msaitoh {
2639 1.186 msaitoh device_t dev = sc->dev;
2640 1.186 msaitoh struct tx_ring *txr = sc->tx_rings;
2641 1.186 msaitoh struct rx_ring *rxr = sc->rx_rings;
2642 1.186 msaitoh struct ixgbevf_hw_stats *stats = &sc->stats.vf;
2643 1.186 msaitoh struct ixgbe_hw *hw = &sc->hw;
2644 1.83 msaitoh const struct sysctlnode *rnode, *cnode;
2645 1.186 msaitoh struct sysctllog **log = &sc->sysctllog;
2646 1.48 msaitoh const char *xname = device_xname(dev);
2647 1.48 msaitoh
2648 1.48 msaitoh /* Driver Statistics */
2649 1.186 msaitoh evcnt_attach_dynamic(&sc->efbig_tx_dma_setup, EVCNT_TYPE_MISC,
2650 1.49 msaitoh NULL, xname, "Driver tx dma soft fail EFBIG");
2651 1.186 msaitoh evcnt_attach_dynamic(&sc->mbuf_defrag_failed, EVCNT_TYPE_MISC,
2652 1.48 msaitoh NULL, xname, "m_defrag() failed");
2653 1.186 msaitoh evcnt_attach_dynamic(&sc->efbig2_tx_dma_setup, EVCNT_TYPE_MISC,
2654 1.49 msaitoh NULL, xname, "Driver tx dma hard fail EFBIG");
2655 1.186 msaitoh evcnt_attach_dynamic(&sc->einval_tx_dma_setup, EVCNT_TYPE_MISC,
2656 1.49 msaitoh NULL, xname, "Driver tx dma hard fail EINVAL");
2657 1.186 msaitoh evcnt_attach_dynamic(&sc->other_tx_dma_setup, EVCNT_TYPE_MISC,
2658 1.49 msaitoh NULL, xname, "Driver tx dma hard fail other");
2659 1.186 msaitoh evcnt_attach_dynamic(&sc->eagain_tx_dma_setup, EVCNT_TYPE_MISC,
2660 1.49 msaitoh NULL, xname, "Driver tx dma soft fail EAGAIN");
2661 1.186 msaitoh evcnt_attach_dynamic(&sc->enomem_tx_dma_setup, EVCNT_TYPE_MISC,
2662 1.49 msaitoh NULL, xname, "Driver tx dma soft fail ENOMEM");
2663 1.186 msaitoh evcnt_attach_dynamic(&sc->watchdog_events, EVCNT_TYPE_MISC,
2664 1.48 msaitoh NULL, xname, "Watchdog timeouts");
2665 1.186 msaitoh evcnt_attach_dynamic(&sc->tso_err, EVCNT_TYPE_MISC,
2666 1.49 msaitoh NULL, xname, "TSO errors");
2667 1.186 msaitoh evcnt_attach_dynamic(&sc->admin_irqev, EVCNT_TYPE_INTR,
2668 1.151 msaitoh NULL, xname, "Admin MSI-X IRQ Handled");
2669 1.186 msaitoh evcnt_attach_dynamic(&sc->link_workev, EVCNT_TYPE_INTR,
2670 1.151 msaitoh NULL, xname, "Admin event");
2671 1.49 msaitoh
2672 1.186 msaitoh for (int i = 0; i < sc->num_queues; i++, rxr++, txr++) {
2673 1.184 msaitoh #ifdef LRO
2674 1.184 msaitoh struct lro_ctrl *lro = &rxr->lro;
2675 1.184 msaitoh #endif
2676 1.184 msaitoh
2677 1.186 msaitoh snprintf(sc->queues[i].evnamebuf,
2678 1.186 msaitoh sizeof(sc->queues[i].evnamebuf), "%s q%d", xname, i);
2679 1.186 msaitoh snprintf(sc->queues[i].namebuf,
2680 1.186 msaitoh sizeof(sc->queues[i].namebuf), "q%d", i);
2681 1.49 msaitoh
2682 1.186 msaitoh if ((rnode = ixv_sysctl_instance(sc)) == NULL) {
2683 1.182 msaitoh aprint_error_dev(dev,
2684 1.182 msaitoh "could not create sysctl root\n");
2685 1.49 msaitoh break;
2686 1.49 msaitoh }
2687 1.49 msaitoh
2688 1.49 msaitoh if (sysctl_createv(log, 0, &rnode, &rnode,
2689 1.49 msaitoh 0, CTLTYPE_NODE,
2690 1.186 msaitoh sc->queues[i].namebuf, SYSCTL_DESCR("Queue Name"),
2691 1.49 msaitoh NULL, 0, NULL, 0, CTL_CREATE, CTL_EOL) != 0)
2692 1.49 msaitoh break;
2693 1.49 msaitoh
2694 1.49 msaitoh if (sysctl_createv(log, 0, &rnode, &cnode,
2695 1.49 msaitoh CTLFLAG_READWRITE, CTLTYPE_INT,
2696 1.49 msaitoh "interrupt_rate", SYSCTL_DESCR("Interrupt Rate"),
2697 1.83 msaitoh ixv_sysctl_interrupt_rate_handler, 0,
2698 1.186 msaitoh (void *)&sc->queues[i], 0, CTL_CREATE, CTL_EOL) != 0)
2699 1.49 msaitoh break;
2700 1.49 msaitoh
2701 1.49 msaitoh if (sysctl_createv(log, 0, &rnode, &cnode,
2702 1.49 msaitoh CTLFLAG_READONLY, CTLTYPE_INT,
2703 1.49 msaitoh "txd_head", SYSCTL_DESCR("Transmit Descriptor Head"),
2704 1.83 msaitoh ixv_sysctl_tdh_handler, 0, (void *)txr,
2705 1.49 msaitoh 0, CTL_CREATE, CTL_EOL) != 0)
2706 1.49 msaitoh break;
2707 1.49 msaitoh
2708 1.49 msaitoh if (sysctl_createv(log, 0, &rnode, &cnode,
2709 1.49 msaitoh CTLFLAG_READONLY, CTLTYPE_INT,
2710 1.49 msaitoh "txd_tail", SYSCTL_DESCR("Transmit Descriptor Tail"),
2711 1.83 msaitoh ixv_sysctl_tdt_handler, 0, (void *)txr,
2712 1.49 msaitoh 0, CTL_CREATE, CTL_EOL) != 0)
2713 1.49 msaitoh break;
2714 1.83 msaitoh
2715 1.49 msaitoh if (sysctl_createv(log, 0, &rnode, &cnode,
2716 1.158 msaitoh CTLFLAG_READONLY, CTLTYPE_INT, "rxd_nxck",
2717 1.158 msaitoh SYSCTL_DESCR("Receive Descriptor next to check"),
2718 1.158 msaitoh ixv_sysctl_next_to_check_handler, 0, (void *)rxr, 0,
2719 1.99 msaitoh CTL_CREATE, CTL_EOL) != 0)
2720 1.99 msaitoh break;
2721 1.99 msaitoh
2722 1.99 msaitoh if (sysctl_createv(log, 0, &rnode, &cnode,
2723 1.164 msaitoh CTLFLAG_READONLY, CTLTYPE_INT, "rxd_nxrf",
2724 1.164 msaitoh SYSCTL_DESCR("Receive Descriptor next to refresh"),
2725 1.164 msaitoh ixv_sysctl_next_to_refresh_handler, 0, (void *)rxr, 0,
2726 1.164 msaitoh CTL_CREATE, CTL_EOL) != 0)
2727 1.164 msaitoh break;
2728 1.164 msaitoh
2729 1.164 msaitoh if (sysctl_createv(log, 0, &rnode, &cnode,
2730 1.158 msaitoh CTLFLAG_READONLY, CTLTYPE_INT, "rxd_head",
2731 1.158 msaitoh SYSCTL_DESCR("Receive Descriptor Head"),
2732 1.83 msaitoh ixv_sysctl_rdh_handler, 0, (void *)rxr, 0,
2733 1.49 msaitoh CTL_CREATE, CTL_EOL) != 0)
2734 1.49 msaitoh break;
2735 1.49 msaitoh
2736 1.49 msaitoh if (sysctl_createv(log, 0, &rnode, &cnode,
2737 1.158 msaitoh CTLFLAG_READONLY, CTLTYPE_INT, "rxd_tail",
2738 1.158 msaitoh SYSCTL_DESCR("Receive Descriptor Tail"),
2739 1.83 msaitoh ixv_sysctl_rdt_handler, 0, (void *)rxr, 0,
2740 1.49 msaitoh CTL_CREATE, CTL_EOL) != 0)
2741 1.49 msaitoh break;
2742 1.49 msaitoh
2743 1.186 msaitoh evcnt_attach_dynamic(&sc->queues[i].irqs, EVCNT_TYPE_INTR,
2744 1.186 msaitoh NULL, sc->queues[i].evnamebuf, "IRQs on queue");
2745 1.186 msaitoh evcnt_attach_dynamic(&sc->queues[i].handleq,
2746 1.186 msaitoh EVCNT_TYPE_MISC, NULL, sc->queues[i].evnamebuf,
2747 1.184 msaitoh "Handled queue in softint");
2748 1.186 msaitoh evcnt_attach_dynamic(&sc->queues[i].req, EVCNT_TYPE_MISC,
2749 1.186 msaitoh NULL, sc->queues[i].evnamebuf, "Requeued in softint");
2750 1.184 msaitoh evcnt_attach_dynamic(&txr->total_packets, EVCNT_TYPE_MISC,
2751 1.186 msaitoh NULL, sc->queues[i].evnamebuf,
2752 1.184 msaitoh "Queue Packets Transmitted");
2753 1.184 msaitoh #ifndef IXGBE_LEGACY_TX
2754 1.184 msaitoh evcnt_attach_dynamic(&txr->pcq_drops, EVCNT_TYPE_MISC,
2755 1.186 msaitoh NULL, sc->queues[i].evnamebuf,
2756 1.184 msaitoh "Packets dropped in pcq");
2757 1.184 msaitoh #endif
2758 1.184 msaitoh evcnt_attach_dynamic(&txr->no_desc_avail, EVCNT_TYPE_MISC,
2759 1.186 msaitoh NULL, sc->queues[i].evnamebuf,
2760 1.184 msaitoh "TX Queue No Descriptor Available");
2761 1.184 msaitoh evcnt_attach_dynamic(&txr->tso_tx, EVCNT_TYPE_MISC,
2762 1.186 msaitoh NULL, sc->queues[i].evnamebuf, "TSO");
2763 1.184 msaitoh
2764 1.49 msaitoh evcnt_attach_dynamic(&rxr->rx_bytes, EVCNT_TYPE_MISC,
2765 1.186 msaitoh NULL, sc->queues[i].evnamebuf,
2766 1.158 msaitoh "Queue Bytes Received");
2767 1.184 msaitoh evcnt_attach_dynamic(&rxr->rx_packets, EVCNT_TYPE_MISC,
2768 1.186 msaitoh NULL, sc->queues[i].evnamebuf,
2769 1.184 msaitoh "Queue Packets Received");
2770 1.166 msaitoh evcnt_attach_dynamic(&rxr->no_mbuf, EVCNT_TYPE_MISC,
2771 1.186 msaitoh NULL, sc->queues[i].evnamebuf, "Rx no mbuf");
2772 1.49 msaitoh evcnt_attach_dynamic(&rxr->rx_discarded, EVCNT_TYPE_MISC,
2773 1.186 msaitoh NULL, sc->queues[i].evnamebuf, "Rx discarded");
2774 1.184 msaitoh evcnt_attach_dynamic(&rxr->rx_copies, EVCNT_TYPE_MISC,
2775 1.186 msaitoh NULL, sc->queues[i].evnamebuf, "Copied RX Frames");
2776 1.49 msaitoh #ifdef LRO
2777 1.49 msaitoh SYSCTL_ADD_INT(ctx, queue_list, OID_AUTO, "lro_queued",
2778 1.49 msaitoh CTLFLAG_RD, &lro->lro_queued, 0,
2779 1.49 msaitoh "LRO Queued");
2780 1.49 msaitoh SYSCTL_ADD_INT(ctx, queue_list, OID_AUTO, "lro_flushed",
2781 1.49 msaitoh CTLFLAG_RD, &lro->lro_flushed, 0,
2782 1.49 msaitoh "LRO Flushed");
2783 1.49 msaitoh #endif /* LRO */
2784 1.49 msaitoh }
2785 1.49 msaitoh
2786 1.58 msaitoh /* MAC stats get their own sub node */
2787 1.49 msaitoh
2788 1.49 msaitoh snprintf(stats->namebuf,
2789 1.49 msaitoh sizeof(stats->namebuf), "%s MAC Statistics", xname);
2790 1.49 msaitoh
2791 1.49 msaitoh evcnt_attach_dynamic(&stats->ipcs, EVCNT_TYPE_MISC, NULL,
2792 1.49 msaitoh stats->namebuf, "rx csum offload - IP");
2793 1.49 msaitoh evcnt_attach_dynamic(&stats->l4cs, EVCNT_TYPE_MISC, NULL,
2794 1.49 msaitoh stats->namebuf, "rx csum offload - L4");
2795 1.49 msaitoh evcnt_attach_dynamic(&stats->ipcs_bad, EVCNT_TYPE_MISC, NULL,
2796 1.49 msaitoh stats->namebuf, "rx csum offload - IP bad");
2797 1.49 msaitoh evcnt_attach_dynamic(&stats->l4cs_bad, EVCNT_TYPE_MISC, NULL,
2798 1.49 msaitoh stats->namebuf, "rx csum offload - L4 bad");
2799 1.48 msaitoh
2800 1.49 msaitoh /* Packet Reception Stats */
2801 1.48 msaitoh evcnt_attach_dynamic(&stats->vfgprc, EVCNT_TYPE_MISC, NULL,
2802 1.48 msaitoh xname, "Good Packets Received");
2803 1.48 msaitoh evcnt_attach_dynamic(&stats->vfgorc, EVCNT_TYPE_MISC, NULL,
2804 1.48 msaitoh xname, "Good Octets Received");
2805 1.48 msaitoh evcnt_attach_dynamic(&stats->vfmprc, EVCNT_TYPE_MISC, NULL,
2806 1.48 msaitoh xname, "Multicast Packets Received");
2807 1.48 msaitoh evcnt_attach_dynamic(&stats->vfgptc, EVCNT_TYPE_MISC, NULL,
2808 1.48 msaitoh xname, "Good Packets Transmitted");
2809 1.48 msaitoh evcnt_attach_dynamic(&stats->vfgotc, EVCNT_TYPE_MISC, NULL,
2810 1.48 msaitoh xname, "Good Octets Transmitted");
2811 1.67 msaitoh
2812 1.67 msaitoh /* Mailbox Stats */
2813 1.67 msaitoh evcnt_attach_dynamic(&hw->mbx.stats.msgs_tx, EVCNT_TYPE_MISC, NULL,
2814 1.67 msaitoh xname, "message TXs");
2815 1.67 msaitoh evcnt_attach_dynamic(&hw->mbx.stats.msgs_rx, EVCNT_TYPE_MISC, NULL,
2816 1.67 msaitoh xname, "message RXs");
2817 1.67 msaitoh evcnt_attach_dynamic(&hw->mbx.stats.acks, EVCNT_TYPE_MISC, NULL,
2818 1.67 msaitoh xname, "ACKs");
2819 1.67 msaitoh evcnt_attach_dynamic(&hw->mbx.stats.reqs, EVCNT_TYPE_MISC, NULL,
2820 1.67 msaitoh xname, "REQs");
2821 1.67 msaitoh evcnt_attach_dynamic(&hw->mbx.stats.rsts, EVCNT_TYPE_MISC, NULL,
2822 1.67 msaitoh xname, "RSTs");
2823 1.67 msaitoh
2824 1.58 msaitoh } /* ixv_add_stats_sysctls */
2825 1.48 msaitoh
2826 1.131 msaitoh static void
2827 1.186 msaitoh ixv_clear_evcnt(struct ixgbe_softc *sc)
2828 1.131 msaitoh {
2829 1.186 msaitoh struct tx_ring *txr = sc->tx_rings;
2830 1.186 msaitoh struct rx_ring *rxr = sc->rx_rings;
2831 1.186 msaitoh struct ixgbevf_hw_stats *stats = &sc->stats.vf;
2832 1.186 msaitoh struct ixgbe_hw *hw = &sc->hw;
2833 1.131 msaitoh int i;
2834 1.131 msaitoh
2835 1.131 msaitoh /* Driver Statistics */
2836 1.186 msaitoh IXGBE_EVC_STORE(&sc->efbig_tx_dma_setup, 0);
2837 1.186 msaitoh IXGBE_EVC_STORE(&sc->mbuf_defrag_failed, 0);
2838 1.186 msaitoh IXGBE_EVC_STORE(&sc->efbig2_tx_dma_setup, 0);
2839 1.186 msaitoh IXGBE_EVC_STORE(&sc->einval_tx_dma_setup, 0);
2840 1.186 msaitoh IXGBE_EVC_STORE(&sc->other_tx_dma_setup, 0);
2841 1.186 msaitoh IXGBE_EVC_STORE(&sc->eagain_tx_dma_setup, 0);
2842 1.186 msaitoh IXGBE_EVC_STORE(&sc->enomem_tx_dma_setup, 0);
2843 1.186 msaitoh IXGBE_EVC_STORE(&sc->watchdog_events, 0);
2844 1.186 msaitoh IXGBE_EVC_STORE(&sc->tso_err, 0);
2845 1.186 msaitoh IXGBE_EVC_STORE(&sc->admin_irqev, 0);
2846 1.186 msaitoh IXGBE_EVC_STORE(&sc->link_workev, 0);
2847 1.186 msaitoh
2848 1.186 msaitoh for (i = 0; i < sc->num_queues; i++, rxr++, txr++) {
2849 1.186 msaitoh IXGBE_EVC_STORE(&sc->queues[i].irqs, 0);
2850 1.186 msaitoh IXGBE_EVC_STORE(&sc->queues[i].handleq, 0);
2851 1.186 msaitoh IXGBE_EVC_STORE(&sc->queues[i].req, 0);
2852 1.178 msaitoh IXGBE_EVC_STORE(&txr->total_packets, 0);
2853 1.131 msaitoh #ifndef IXGBE_LEGACY_TX
2854 1.178 msaitoh IXGBE_EVC_STORE(&txr->pcq_drops, 0);
2855 1.131 msaitoh #endif
2856 1.184 msaitoh IXGBE_EVC_STORE(&txr->no_desc_avail, 0);
2857 1.184 msaitoh IXGBE_EVC_STORE(&txr->tso_tx, 0);
2858 1.131 msaitoh txr->q_efbig_tx_dma_setup = 0;
2859 1.131 msaitoh txr->q_mbuf_defrag_failed = 0;
2860 1.131 msaitoh txr->q_efbig2_tx_dma_setup = 0;
2861 1.131 msaitoh txr->q_einval_tx_dma_setup = 0;
2862 1.131 msaitoh txr->q_other_tx_dma_setup = 0;
2863 1.131 msaitoh txr->q_eagain_tx_dma_setup = 0;
2864 1.131 msaitoh txr->q_enomem_tx_dma_setup = 0;
2865 1.131 msaitoh txr->q_tso_err = 0;
2866 1.131 msaitoh
2867 1.178 msaitoh IXGBE_EVC_STORE(&rxr->rx_packets, 0);
2868 1.178 msaitoh IXGBE_EVC_STORE(&rxr->rx_bytes, 0);
2869 1.178 msaitoh IXGBE_EVC_STORE(&rxr->rx_copies, 0);
2870 1.178 msaitoh IXGBE_EVC_STORE(&rxr->no_mbuf, 0);
2871 1.178 msaitoh IXGBE_EVC_STORE(&rxr->rx_discarded, 0);
2872 1.131 msaitoh }
2873 1.131 msaitoh
2874 1.131 msaitoh /* MAC stats get their own sub node */
2875 1.131 msaitoh
2876 1.178 msaitoh IXGBE_EVC_STORE(&stats->ipcs, 0);
2877 1.178 msaitoh IXGBE_EVC_STORE(&stats->l4cs, 0);
2878 1.178 msaitoh IXGBE_EVC_STORE(&stats->ipcs_bad, 0);
2879 1.178 msaitoh IXGBE_EVC_STORE(&stats->l4cs_bad, 0);
2880 1.131 msaitoh
2881 1.176 msaitoh /*
2882 1.176 msaitoh * Packet Reception Stats.
2883 1.176 msaitoh * Call ixv_init_stats() to save last VF counters' values.
2884 1.176 msaitoh */
2885 1.186 msaitoh ixv_init_stats(sc);
2886 1.178 msaitoh IXGBE_EVC_STORE(&stats->vfgprc, 0);
2887 1.178 msaitoh IXGBE_EVC_STORE(&stats->vfgorc, 0);
2888 1.178 msaitoh IXGBE_EVC_STORE(&stats->vfmprc, 0);
2889 1.178 msaitoh IXGBE_EVC_STORE(&stats->vfgptc, 0);
2890 1.178 msaitoh IXGBE_EVC_STORE(&stats->vfgotc, 0);
2891 1.131 msaitoh
2892 1.131 msaitoh /* Mailbox Stats */
2893 1.178 msaitoh IXGBE_EVC_STORE(&hw->mbx.stats.msgs_tx, 0);
2894 1.178 msaitoh IXGBE_EVC_STORE(&hw->mbx.stats.msgs_rx, 0);
2895 1.178 msaitoh IXGBE_EVC_STORE(&hw->mbx.stats.acks, 0);
2896 1.178 msaitoh IXGBE_EVC_STORE(&hw->mbx.stats.reqs, 0);
2897 1.178 msaitoh IXGBE_EVC_STORE(&hw->mbx.stats.rsts, 0);
2898 1.131 msaitoh
2899 1.131 msaitoh } /* ixv_clear_evcnt */
2900 1.131 msaitoh
2901 1.186 msaitoh #define PRINTQS(sc, regname) \
2902 1.175 msaitoh do { \
2903 1.186 msaitoh struct ixgbe_hw *_hw = &(sc)->hw; \
2904 1.175 msaitoh int _i; \
2905 1.175 msaitoh \
2906 1.186 msaitoh printf("%s: %s", device_xname((sc)->dev), #regname); \
2907 1.186 msaitoh for (_i = 0; _i < (sc)->num_queues; _i++) { \
2908 1.175 msaitoh printf((_i == 0) ? "\t" : " "); \
2909 1.175 msaitoh printf("%08x", IXGBE_READ_REG(_hw, \
2910 1.175 msaitoh IXGBE_##regname(_i))); \
2911 1.175 msaitoh } \
2912 1.175 msaitoh printf("\n"); \
2913 1.175 msaitoh } while (0)
2914 1.175 msaitoh
2915 1.58 msaitoh /************************************************************************
2916 1.58 msaitoh * ixv_print_debug_info
2917 1.57 msaitoh *
2918 1.58 msaitoh * Provides a way to take a look at important statistics
2919 1.58 msaitoh * maintained by the driver and hardware.
2920 1.58 msaitoh ************************************************************************/
2921 1.57 msaitoh static void
2922 1.186 msaitoh ixv_print_debug_info(struct ixgbe_softc *sc)
2923 1.57 msaitoh {
2924 1.186 msaitoh device_t dev = sc->dev;
2925 1.186 msaitoh struct ixgbe_hw *hw = &sc->hw;
2926 1.175 msaitoh int i;
2927 1.57 msaitoh
2928 1.175 msaitoh device_printf(dev, "queue:");
2929 1.186 msaitoh for (i = 0; i < sc->num_queues; i++) {
2930 1.175 msaitoh printf((i == 0) ? "\t" : " ");
2931 1.175 msaitoh printf("%8d", i);
2932 1.58 msaitoh }
2933 1.175 msaitoh printf("\n");
2934 1.186 msaitoh PRINTQS(sc, VFRDBAL);
2935 1.186 msaitoh PRINTQS(sc, VFRDBAH);
2936 1.186 msaitoh PRINTQS(sc, VFRDLEN);
2937 1.186 msaitoh PRINTQS(sc, VFSRRCTL);
2938 1.186 msaitoh PRINTQS(sc, VFRDH);
2939 1.186 msaitoh PRINTQS(sc, VFRDT);
2940 1.186 msaitoh PRINTQS(sc, VFRXDCTL);
2941 1.175 msaitoh
2942 1.175 msaitoh device_printf(dev, "EIMS:\t%08x\n", IXGBE_READ_REG(hw, IXGBE_VTEIMS));
2943 1.175 msaitoh device_printf(dev, "EIAM:\t%08x\n", IXGBE_READ_REG(hw, IXGBE_VTEIAM));
2944 1.175 msaitoh device_printf(dev, "EIAC:\t%08x\n", IXGBE_READ_REG(hw, IXGBE_VTEIAC));
2945 1.58 msaitoh } /* ixv_print_debug_info */
2946 1.58 msaitoh
2947 1.58 msaitoh /************************************************************************
2948 1.58 msaitoh * ixv_sysctl_debug
2949 1.58 msaitoh ************************************************************************/
2950 1.57 msaitoh static int
2951 1.57 msaitoh ixv_sysctl_debug(SYSCTLFN_ARGS)
2952 1.57 msaitoh {
2953 1.97 msaitoh struct sysctlnode node = *rnode;
2954 1.186 msaitoh struct ixgbe_softc *sc = (struct ixgbe_softc *)node.sysctl_data;
2955 1.174 msaitoh int error, result = 0;
2956 1.57 msaitoh
2957 1.57 msaitoh node.sysctl_data = &result;
2958 1.57 msaitoh error = sysctl_lookup(SYSCTLFN_CALL(&node));
2959 1.57 msaitoh
2960 1.58 msaitoh if (error || newp == NULL)
2961 1.57 msaitoh return error;
2962 1.57 msaitoh
2963 1.97 msaitoh if (result == 1)
2964 1.186 msaitoh ixv_print_debug_info(sc);
2965 1.57 msaitoh
2966 1.57 msaitoh return 0;
2967 1.58 msaitoh } /* ixv_sysctl_debug */
2968 1.58 msaitoh
2969 1.58 msaitoh /************************************************************************
2970 1.163 msaitoh * ixv_sysctl_rx_copy_len
2971 1.163 msaitoh ************************************************************************/
2972 1.163 msaitoh static int
2973 1.163 msaitoh ixv_sysctl_rx_copy_len(SYSCTLFN_ARGS)
2974 1.163 msaitoh {
2975 1.163 msaitoh struct sysctlnode node = *rnode;
2976 1.186 msaitoh struct ixgbe_softc *sc = (struct ixgbe_softc *)node.sysctl_data;
2977 1.163 msaitoh int error;
2978 1.186 msaitoh int result = sc->rx_copy_len;
2979 1.163 msaitoh
2980 1.163 msaitoh node.sysctl_data = &result;
2981 1.163 msaitoh error = sysctl_lookup(SYSCTLFN_CALL(&node));
2982 1.163 msaitoh
2983 1.163 msaitoh if (error || newp == NULL)
2984 1.163 msaitoh return error;
2985 1.163 msaitoh
2986 1.163 msaitoh if ((result < 0) || (result > IXGBE_RX_COPY_LEN_MAX))
2987 1.163 msaitoh return EINVAL;
2988 1.163 msaitoh
2989 1.186 msaitoh sc->rx_copy_len = result;
2990 1.163 msaitoh
2991 1.163 msaitoh return 0;
2992 1.179 msaitoh } /* ixv_sysctl_rx_copy_len */
2993 1.179 msaitoh
2994 1.179 msaitoh /************************************************************************
2995 1.179 msaitoh * ixv_sysctl_tx_process_limit
2996 1.179 msaitoh ************************************************************************/
2997 1.179 msaitoh static int
2998 1.179 msaitoh ixv_sysctl_tx_process_limit(SYSCTLFN_ARGS)
2999 1.179 msaitoh {
3000 1.179 msaitoh struct sysctlnode node = *rnode;
3001 1.186 msaitoh struct ixgbe_softc *sc = (struct ixgbe_softc *)node.sysctl_data;
3002 1.179 msaitoh int error;
3003 1.186 msaitoh int result = sc->tx_process_limit;
3004 1.179 msaitoh
3005 1.179 msaitoh node.sysctl_data = &result;
3006 1.179 msaitoh error = sysctl_lookup(SYSCTLFN_CALL(&node));
3007 1.179 msaitoh
3008 1.179 msaitoh if (error || newp == NULL)
3009 1.179 msaitoh return error;
3010 1.179 msaitoh
3011 1.186 msaitoh if ((result <= 0) || (result > sc->num_tx_desc))
3012 1.179 msaitoh return EINVAL;
3013 1.179 msaitoh
3014 1.186 msaitoh sc->tx_process_limit = result;
3015 1.179 msaitoh
3016 1.179 msaitoh return 0;
3017 1.179 msaitoh } /* ixv_sysctl_tx_process_limit */
3018 1.179 msaitoh
3019 1.179 msaitoh /************************************************************************
3020 1.179 msaitoh * ixv_sysctl_rx_process_limit
3021 1.179 msaitoh ************************************************************************/
3022 1.179 msaitoh static int
3023 1.179 msaitoh ixv_sysctl_rx_process_limit(SYSCTLFN_ARGS)
3024 1.179 msaitoh {
3025 1.179 msaitoh struct sysctlnode node = *rnode;
3026 1.186 msaitoh struct ixgbe_softc *sc = (struct ixgbe_softc *)node.sysctl_data;
3027 1.179 msaitoh int error;
3028 1.186 msaitoh int result = sc->rx_process_limit;
3029 1.179 msaitoh
3030 1.179 msaitoh node.sysctl_data = &result;
3031 1.179 msaitoh error = sysctl_lookup(SYSCTLFN_CALL(&node));
3032 1.179 msaitoh
3033 1.179 msaitoh if (error || newp == NULL)
3034 1.179 msaitoh return error;
3035 1.179 msaitoh
3036 1.186 msaitoh if ((result <= 0) || (result > sc->num_rx_desc))
3037 1.179 msaitoh return EINVAL;
3038 1.179 msaitoh
3039 1.186 msaitoh sc->rx_process_limit = result;
3040 1.179 msaitoh
3041 1.179 msaitoh return 0;
3042 1.179 msaitoh } /* ixv_sysctl_rx_process_limit */
3043 1.163 msaitoh
3044 1.163 msaitoh /************************************************************************
3045 1.58 msaitoh * ixv_init_device_features
3046 1.58 msaitoh ************************************************************************/
3047 1.58 msaitoh static void
3048 1.186 msaitoh ixv_init_device_features(struct ixgbe_softc *sc)
3049 1.58 msaitoh {
3050 1.186 msaitoh sc->feat_cap = IXGBE_FEATURE_NETMAP
3051 1.114 msaitoh | IXGBE_FEATURE_VF
3052 1.114 msaitoh | IXGBE_FEATURE_RSS
3053 1.114 msaitoh | IXGBE_FEATURE_LEGACY_TX;
3054 1.58 msaitoh
3055 1.58 msaitoh /* A tad short on feature flags for VFs, atm. */
3056 1.186 msaitoh switch (sc->hw.mac.type) {
3057 1.58 msaitoh case ixgbe_mac_82599_vf:
3058 1.58 msaitoh break;
3059 1.58 msaitoh case ixgbe_mac_X540_vf:
3060 1.58 msaitoh break;
3061 1.58 msaitoh case ixgbe_mac_X550_vf:
3062 1.58 msaitoh case ixgbe_mac_X550EM_x_vf:
3063 1.58 msaitoh case ixgbe_mac_X550EM_a_vf:
3064 1.186 msaitoh sc->feat_cap |= IXGBE_FEATURE_NEEDS_CTXD;
3065 1.58 msaitoh break;
3066 1.58 msaitoh default:
3067 1.58 msaitoh break;
3068 1.58 msaitoh }
3069 1.57 msaitoh
3070 1.58 msaitoh /* Enabled by default... */
3071 1.58 msaitoh /* Is a virtual function (VF) */
3072 1.186 msaitoh if (sc->feat_cap & IXGBE_FEATURE_VF)
3073 1.186 msaitoh sc->feat_en |= IXGBE_FEATURE_VF;
3074 1.58 msaitoh /* Netmap */
3075 1.186 msaitoh if (sc->feat_cap & IXGBE_FEATURE_NETMAP)
3076 1.186 msaitoh sc->feat_en |= IXGBE_FEATURE_NETMAP;
3077 1.58 msaitoh /* Receive-Side Scaling (RSS) */
3078 1.186 msaitoh if (sc->feat_cap & IXGBE_FEATURE_RSS)
3079 1.186 msaitoh sc->feat_en |= IXGBE_FEATURE_RSS;
3080 1.58 msaitoh /* Needs advanced context descriptor regardless of offloads req'd */
3081 1.186 msaitoh if (sc->feat_cap & IXGBE_FEATURE_NEEDS_CTXD)
3082 1.186 msaitoh sc->feat_en |= IXGBE_FEATURE_NEEDS_CTXD;
3083 1.58 msaitoh
3084 1.58 msaitoh /* Enabled via sysctl... */
3085 1.58 msaitoh /* Legacy (single queue) transmit */
3086 1.186 msaitoh if ((sc->feat_cap & IXGBE_FEATURE_LEGACY_TX) &&
3087 1.58 msaitoh ixv_enable_legacy_tx)
3088 1.186 msaitoh sc->feat_en |= IXGBE_FEATURE_LEGACY_TX;
3089 1.58 msaitoh } /* ixv_init_device_features */
3090 1.58 msaitoh
3091 1.58 msaitoh /************************************************************************
3092 1.58 msaitoh * ixv_shutdown - Shutdown entry point
3093 1.58 msaitoh ************************************************************************/
3094 1.57 msaitoh #if 0 /* XXX NetBSD ought to register something like this through pmf(9) */
3095 1.57 msaitoh static int
3096 1.57 msaitoh ixv_shutdown(device_t dev)
3097 1.57 msaitoh {
3098 1.186 msaitoh struct ixgbe_softc *sc = device_private(dev);
3099 1.186 msaitoh IXGBE_CORE_LOCK(sc);
3100 1.186 msaitoh ixv_stop_locked(sc);
3101 1.186 msaitoh IXGBE_CORE_UNLOCK(sc);
3102 1.57 msaitoh
3103 1.57 msaitoh return (0);
3104 1.58 msaitoh } /* ixv_shutdown */
3105 1.57 msaitoh #endif
3106 1.57 msaitoh
3107 1.57 msaitoh static int
3108 1.57 msaitoh ixv_ifflags_cb(struct ethercom *ec)
3109 1.57 msaitoh {
3110 1.57 msaitoh struct ifnet *ifp = &ec->ec_if;
3111 1.186 msaitoh struct ixgbe_softc *sc = ifp->if_softc;
3112 1.137 msaitoh u_short saved_flags;
3113 1.136 msaitoh u_short change;
3114 1.136 msaitoh int rv = 0;
3115 1.57 msaitoh
3116 1.186 msaitoh IXGBE_CORE_LOCK(sc);
3117 1.57 msaitoh
3118 1.186 msaitoh saved_flags = sc->if_flags;
3119 1.186 msaitoh change = ifp->if_flags ^ sc->if_flags;
3120 1.57 msaitoh if (change != 0)
3121 1.186 msaitoh sc->if_flags = ifp->if_flags;
3122 1.57 msaitoh
3123 1.118 msaitoh if ((change & ~(IFF_CANTCHANGE | IFF_DEBUG)) != 0) {
3124 1.118 msaitoh rv = ENETRESET;
3125 1.118 msaitoh goto out;
3126 1.137 msaitoh } else if ((change & IFF_PROMISC) != 0) {
3127 1.186 msaitoh rv = ixv_set_rxfilter(sc);
3128 1.137 msaitoh if (rv != 0) {
3129 1.137 msaitoh /* Restore previous */
3130 1.186 msaitoh sc->if_flags = saved_flags;
3131 1.137 msaitoh goto out;
3132 1.137 msaitoh }
3133 1.118 msaitoh }
3134 1.57 msaitoh
3135 1.120 msaitoh /* Check for ec_capenable. */
3136 1.186 msaitoh change = ec->ec_capenable ^ sc->ec_capenable;
3137 1.186 msaitoh sc->ec_capenable = ec->ec_capenable;
3138 1.120 msaitoh if ((change & ~(ETHERCAP_VLAN_MTU | ETHERCAP_VLAN_HWTAGGING
3139 1.120 msaitoh | ETHERCAP_VLAN_HWFILTER)) != 0) {
3140 1.120 msaitoh rv = ENETRESET;
3141 1.120 msaitoh goto out;
3142 1.120 msaitoh }
3143 1.120 msaitoh
3144 1.120 msaitoh /*
3145 1.120 msaitoh * Special handling is not required for ETHERCAP_VLAN_MTU.
3146 1.120 msaitoh * PF's MAXFRS(MHADD) does not include the 4bytes of the VLAN header.
3147 1.120 msaitoh */
3148 1.120 msaitoh
3149 1.65 msaitoh /* Set up VLAN support and filter */
3150 1.120 msaitoh if ((change & (ETHERCAP_VLAN_HWTAGGING | ETHERCAP_VLAN_HWFILTER)) != 0)
3151 1.186 msaitoh rv = ixv_setup_vlan_support(sc);
3152 1.65 msaitoh
3153 1.118 msaitoh out:
3154 1.186 msaitoh IXGBE_CORE_UNLOCK(sc);
3155 1.57 msaitoh
3156 1.118 msaitoh return rv;
3157 1.57 msaitoh }
3158 1.57 msaitoh
3159 1.58 msaitoh
3160 1.58 msaitoh /************************************************************************
3161 1.58 msaitoh * ixv_ioctl - Ioctl entry point
3162 1.57 msaitoh *
3163 1.58 msaitoh * Called when the user wants to configure the interface.
3164 1.57 msaitoh *
3165 1.58 msaitoh * return 0 on success, positive on failure
3166 1.58 msaitoh ************************************************************************/
3167 1.57 msaitoh static int
3168 1.58 msaitoh ixv_ioctl(struct ifnet *ifp, u_long command, void *data)
3169 1.57 msaitoh {
3170 1.186 msaitoh struct ixgbe_softc *sc = ifp->if_softc;
3171 1.186 msaitoh struct ixgbe_hw *hw = &sc->hw;
3172 1.57 msaitoh struct ifcapreq *ifcr = data;
3173 1.134 msaitoh int error;
3174 1.57 msaitoh int l4csum_en;
3175 1.113 msaitoh const int l4csum = IFCAP_CSUM_TCPv4_Rx | IFCAP_CSUM_UDPv4_Rx |
3176 1.113 msaitoh IFCAP_CSUM_TCPv6_Rx | IFCAP_CSUM_UDPv6_Rx;
3177 1.57 msaitoh
3178 1.57 msaitoh switch (command) {
3179 1.57 msaitoh case SIOCSIFFLAGS:
3180 1.57 msaitoh IOCTL_DEBUGOUT("ioctl: SIOCSIFFLAGS (Set Interface Flags)");
3181 1.57 msaitoh break;
3182 1.132 msaitoh case SIOCADDMULTI: {
3183 1.132 msaitoh struct ether_multi *enm;
3184 1.132 msaitoh struct ether_multistep step;
3185 1.186 msaitoh struct ethercom *ec = &sc->osdep.ec;
3186 1.134 msaitoh bool overflow = false;
3187 1.132 msaitoh int mcnt = 0;
3188 1.132 msaitoh
3189 1.132 msaitoh /*
3190 1.132 msaitoh * Check the number of multicast address. If it exceeds,
3191 1.132 msaitoh * return ENOSPC.
3192 1.132 msaitoh * Update this code when we support API 1.3.
3193 1.132 msaitoh */
3194 1.132 msaitoh ETHER_LOCK(ec);
3195 1.132 msaitoh ETHER_FIRST_MULTI(step, ec, enm);
3196 1.132 msaitoh while (enm != NULL) {
3197 1.132 msaitoh mcnt++;
3198 1.132 msaitoh
3199 1.132 msaitoh /*
3200 1.132 msaitoh * This code is before adding, so one room is required
3201 1.132 msaitoh * at least.
3202 1.132 msaitoh */
3203 1.132 msaitoh if (mcnt > (IXGBE_MAX_VF_MC - 1)) {
3204 1.134 msaitoh overflow = true;
3205 1.132 msaitoh break;
3206 1.132 msaitoh }
3207 1.132 msaitoh ETHER_NEXT_MULTI(step, enm);
3208 1.132 msaitoh }
3209 1.132 msaitoh ETHER_UNLOCK(ec);
3210 1.134 msaitoh error = 0;
3211 1.134 msaitoh if (overflow && ((ec->ec_flags & ETHER_F_ALLMULTI) == 0)) {
3212 1.134 msaitoh error = hw->mac.ops.update_xcast_mode(hw,
3213 1.134 msaitoh IXGBEVF_XCAST_MODE_ALLMULTI);
3214 1.134 msaitoh if (error == IXGBE_ERR_NOT_TRUSTED) {
3215 1.186 msaitoh device_printf(sc->dev,
3216 1.134 msaitoh "this interface is not trusted\n");
3217 1.135 msaitoh error = EPERM;
3218 1.135 msaitoh } else if (error == IXGBE_ERR_FEATURE_NOT_SUPPORTED) {
3219 1.186 msaitoh device_printf(sc->dev,
3220 1.135 msaitoh "the PF doesn't support allmulti mode\n");
3221 1.135 msaitoh error = EOPNOTSUPP;
3222 1.134 msaitoh } else if (error) {
3223 1.186 msaitoh device_printf(sc->dev,
3224 1.134 msaitoh "number of Ethernet multicast addresses "
3225 1.134 msaitoh "exceeds the limit (%d). error = %d\n",
3226 1.134 msaitoh IXGBE_MAX_VF_MC, error);
3227 1.134 msaitoh error = ENOSPC;
3228 1.134 msaitoh } else
3229 1.134 msaitoh ec->ec_flags |= ETHER_F_ALLMULTI;
3230 1.134 msaitoh }
3231 1.132 msaitoh if (error)
3232 1.132 msaitoh return error;
3233 1.132 msaitoh }
3234 1.132 msaitoh /*FALLTHROUGH*/
3235 1.57 msaitoh case SIOCDELMULTI:
3236 1.57 msaitoh IOCTL_DEBUGOUT("ioctl: SIOC(ADD|DEL)MULTI");
3237 1.57 msaitoh break;
3238 1.57 msaitoh case SIOCSIFMEDIA:
3239 1.57 msaitoh case SIOCGIFMEDIA:
3240 1.57 msaitoh IOCTL_DEBUGOUT("ioctl: SIOCxIFMEDIA (Get/Set Interface Media)");
3241 1.57 msaitoh break;
3242 1.57 msaitoh case SIOCSIFCAP:
3243 1.57 msaitoh IOCTL_DEBUGOUT("ioctl: SIOCSIFCAP (Set Capabilities)");
3244 1.57 msaitoh break;
3245 1.57 msaitoh case SIOCSIFMTU:
3246 1.57 msaitoh IOCTL_DEBUGOUT("ioctl: SIOCSIFMTU (Set Interface MTU)");
3247 1.57 msaitoh break;
3248 1.131 msaitoh case SIOCZIFDATA:
3249 1.131 msaitoh IOCTL_DEBUGOUT("ioctl: SIOCZIFDATA (Zero counter)");
3250 1.186 msaitoh ixv_update_stats(sc);
3251 1.186 msaitoh ixv_clear_evcnt(sc);
3252 1.131 msaitoh break;
3253 1.57 msaitoh default:
3254 1.57 msaitoh IOCTL_DEBUGOUT1("ioctl: UNKNOWN (0x%X)", (int)command);
3255 1.57 msaitoh break;
3256 1.57 msaitoh }
3257 1.57 msaitoh
3258 1.57 msaitoh switch (command) {
3259 1.57 msaitoh case SIOCSIFCAP:
3260 1.57 msaitoh /* Layer-4 Rx checksum offload has to be turned on and
3261 1.57 msaitoh * off as a unit.
3262 1.57 msaitoh */
3263 1.57 msaitoh l4csum_en = ifcr->ifcr_capenable & l4csum;
3264 1.57 msaitoh if (l4csum_en != l4csum && l4csum_en != 0)
3265 1.57 msaitoh return EINVAL;
3266 1.57 msaitoh /*FALLTHROUGH*/
3267 1.57 msaitoh case SIOCADDMULTI:
3268 1.57 msaitoh case SIOCDELMULTI:
3269 1.57 msaitoh case SIOCSIFFLAGS:
3270 1.57 msaitoh case SIOCSIFMTU:
3271 1.57 msaitoh default:
3272 1.57 msaitoh if ((error = ether_ioctl(ifp, command, data)) != ENETRESET)
3273 1.57 msaitoh return error;
3274 1.57 msaitoh if ((ifp->if_flags & IFF_RUNNING) == 0)
3275 1.57 msaitoh ;
3276 1.57 msaitoh else if (command == SIOCSIFCAP || command == SIOCSIFMTU) {
3277 1.186 msaitoh IXGBE_CORE_LOCK(sc);
3278 1.186 msaitoh ixv_init_locked(sc);
3279 1.186 msaitoh IXGBE_CORE_UNLOCK(sc);
3280 1.57 msaitoh } else if (command == SIOCADDMULTI || command == SIOCDELMULTI) {
3281 1.57 msaitoh /*
3282 1.57 msaitoh * Multicast list has changed; set the hardware filter
3283 1.57 msaitoh * accordingly.
3284 1.57 msaitoh */
3285 1.186 msaitoh IXGBE_CORE_LOCK(sc);
3286 1.186 msaitoh ixv_disable_intr(sc);
3287 1.186 msaitoh ixv_set_rxfilter(sc);
3288 1.186 msaitoh ixv_enable_intr(sc);
3289 1.186 msaitoh IXGBE_CORE_UNLOCK(sc);
3290 1.57 msaitoh }
3291 1.57 msaitoh return 0;
3292 1.57 msaitoh }
3293 1.58 msaitoh } /* ixv_ioctl */
3294 1.57 msaitoh
3295 1.58 msaitoh /************************************************************************
3296 1.58 msaitoh * ixv_init
3297 1.58 msaitoh ************************************************************************/
3298 1.57 msaitoh static int
3299 1.57 msaitoh ixv_init(struct ifnet *ifp)
3300 1.57 msaitoh {
3301 1.186 msaitoh struct ixgbe_softc *sc = ifp->if_softc;
3302 1.57 msaitoh
3303 1.186 msaitoh IXGBE_CORE_LOCK(sc);
3304 1.186 msaitoh ixv_init_locked(sc);
3305 1.186 msaitoh IXGBE_CORE_UNLOCK(sc);
3306 1.57 msaitoh
3307 1.57 msaitoh return 0;
3308 1.58 msaitoh } /* ixv_init */
3309 1.57 msaitoh
3310 1.58 msaitoh /************************************************************************
3311 1.58 msaitoh * ixv_handle_que
3312 1.58 msaitoh ************************************************************************/
3313 1.57 msaitoh static void
3314 1.57 msaitoh ixv_handle_que(void *context)
3315 1.57 msaitoh {
3316 1.57 msaitoh struct ix_queue *que = context;
3317 1.186 msaitoh struct ixgbe_softc *sc = que->sc;
3318 1.57 msaitoh struct tx_ring *txr = que->txr;
3319 1.186 msaitoh struct ifnet *ifp = sc->ifp;
3320 1.57 msaitoh bool more;
3321 1.57 msaitoh
3322 1.178 msaitoh IXGBE_EVC_ADD(&que->handleq, 1);
3323 1.57 msaitoh
3324 1.57 msaitoh if (ifp->if_flags & IFF_RUNNING) {
3325 1.57 msaitoh IXGBE_TX_LOCK(txr);
3326 1.183 msaitoh more = ixgbe_txeof(txr);
3327 1.186 msaitoh if (!(sc->feat_en & IXGBE_FEATURE_LEGACY_TX))
3328 1.58 msaitoh if (!ixgbe_mq_ring_empty(ifp, txr->txr_interq))
3329 1.58 msaitoh ixgbe_mq_start_locked(ifp, txr);
3330 1.57 msaitoh /* Only for queue 0 */
3331 1.61 msaitoh /* NetBSD still needs this for CBQ */
3332 1.186 msaitoh if ((&sc->queues[0] == que)
3333 1.58 msaitoh && (!ixgbe_legacy_ring_empty(ifp, NULL)))
3334 1.58 msaitoh ixgbe_legacy_start_locked(ifp, txr);
3335 1.57 msaitoh IXGBE_TX_UNLOCK(txr);
3336 1.183 msaitoh more |= ixgbe_rxeof(que);
3337 1.57 msaitoh if (more) {
3338 1.178 msaitoh IXGBE_EVC_ADD(&que->req, 1);
3339 1.186 msaitoh if (sc->txrx_use_workqueue) {
3340 1.84 knakahar /*
3341 1.84 knakahar * "enqueued flag" is not required here
3342 1.84 knakahar * the same as ixg(4). See ixgbe_msix_que().
3343 1.84 knakahar */
3344 1.186 msaitoh workqueue_enqueue(sc->que_wq,
3345 1.84 knakahar &que->wq_cookie, curcpu());
3346 1.84 knakahar } else
3347 1.84 knakahar softint_schedule(que->que_si);
3348 1.57 msaitoh return;
3349 1.57 msaitoh }
3350 1.57 msaitoh }
3351 1.57 msaitoh
3352 1.58 msaitoh /* Re-enable this interrupt */
3353 1.186 msaitoh ixv_enable_queue(sc, que->msix);
3354 1.57 msaitoh
3355 1.57 msaitoh return;
3356 1.58 msaitoh } /* ixv_handle_que */
3357 1.57 msaitoh
3358 1.58 msaitoh /************************************************************************
3359 1.84 knakahar * ixv_handle_que_work
3360 1.84 knakahar ************************************************************************/
3361 1.84 knakahar static void
3362 1.84 knakahar ixv_handle_que_work(struct work *wk, void *context)
3363 1.84 knakahar {
3364 1.84 knakahar struct ix_queue *que = container_of(wk, struct ix_queue, wq_cookie);
3365 1.84 knakahar
3366 1.84 knakahar /*
3367 1.84 knakahar * "enqueued flag" is not required here the same as ixg(4).
3368 1.84 knakahar * See ixgbe_msix_que().
3369 1.84 knakahar */
3370 1.84 knakahar ixv_handle_que(que);
3371 1.84 knakahar }
3372 1.84 knakahar
3373 1.84 knakahar /************************************************************************
3374 1.58 msaitoh * ixv_allocate_msix - Setup MSI-X Interrupt resources and handlers
3375 1.58 msaitoh ************************************************************************/
3376 1.57 msaitoh static int
3377 1.186 msaitoh ixv_allocate_msix(struct ixgbe_softc *sc, const struct pci_attach_args *pa)
3378 1.57 msaitoh {
3379 1.186 msaitoh device_t dev = sc->dev;
3380 1.186 msaitoh struct ix_queue *que = sc->queues;
3381 1.190 msaitoh struct tx_ring *txr = sc->tx_rings;
3382 1.114 msaitoh int error, msix_ctrl, rid, vector = 0;
3383 1.57 msaitoh pci_chipset_tag_t pc;
3384 1.57 msaitoh pcitag_t tag;
3385 1.57 msaitoh char intrbuf[PCI_INTRSTR_LEN];
3386 1.84 knakahar char wqname[MAXCOMLEN];
3387 1.57 msaitoh char intr_xname[32];
3388 1.57 msaitoh const char *intrstr = NULL;
3389 1.57 msaitoh kcpuset_t *affinity;
3390 1.57 msaitoh int cpu_id = 0;
3391 1.57 msaitoh
3392 1.186 msaitoh pc = sc->osdep.pc;
3393 1.186 msaitoh tag = sc->osdep.tag;
3394 1.57 msaitoh
3395 1.186 msaitoh sc->osdep.nintrs = sc->num_queues + 1;
3396 1.186 msaitoh if (pci_msix_alloc_exact(pa, &sc->osdep.intrs,
3397 1.186 msaitoh sc->osdep.nintrs) != 0) {
3398 1.57 msaitoh aprint_error_dev(dev,
3399 1.57 msaitoh "failed to allocate MSI-X interrupt\n");
3400 1.57 msaitoh return (ENXIO);
3401 1.57 msaitoh }
3402 1.57 msaitoh
3403 1.57 msaitoh kcpuset_create(&affinity, false);
3404 1.186 msaitoh for (int i = 0; i < sc->num_queues; i++, vector++, que++, txr++) {
3405 1.57 msaitoh snprintf(intr_xname, sizeof(intr_xname), "%s TXRX%d",
3406 1.57 msaitoh device_xname(dev), i);
3407 1.186 msaitoh intrstr = pci_intr_string(pc, sc->osdep.intrs[i], intrbuf,
3408 1.57 msaitoh sizeof(intrbuf));
3409 1.57 msaitoh #ifdef IXGBE_MPSAFE
3410 1.186 msaitoh pci_intr_setattr(pc, &sc->osdep.intrs[i], PCI_INTR_MPSAFE,
3411 1.57 msaitoh true);
3412 1.57 msaitoh #endif
3413 1.57 msaitoh /* Set the handler function */
3414 1.186 msaitoh que->res = sc->osdep.ihs[i] = pci_intr_establish_xname(pc,
3415 1.186 msaitoh sc->osdep.intrs[i], IPL_NET, ixv_msix_que, que,
3416 1.57 msaitoh intr_xname);
3417 1.57 msaitoh if (que->res == NULL) {
3418 1.186 msaitoh pci_intr_release(pc, sc->osdep.intrs,
3419 1.186 msaitoh sc->osdep.nintrs);
3420 1.57 msaitoh aprint_error_dev(dev,
3421 1.57 msaitoh "Failed to register QUE handler\n");
3422 1.57 msaitoh kcpuset_destroy(affinity);
3423 1.57 msaitoh return (ENXIO);
3424 1.57 msaitoh }
3425 1.57 msaitoh que->msix = vector;
3426 1.186 msaitoh sc->active_queues |= (u64)(1 << que->msix);
3427 1.57 msaitoh
3428 1.57 msaitoh cpu_id = i;
3429 1.57 msaitoh /* Round-robin affinity */
3430 1.57 msaitoh kcpuset_zero(affinity);
3431 1.57 msaitoh kcpuset_set(affinity, cpu_id % ncpu);
3432 1.186 msaitoh error = interrupt_distribute(sc->osdep.ihs[i], affinity, NULL);
3433 1.57 msaitoh aprint_normal_dev(dev, "for TX/RX, interrupting at %s",
3434 1.57 msaitoh intrstr);
3435 1.57 msaitoh if (error == 0)
3436 1.57 msaitoh aprint_normal(", bound queue %d to cpu %d\n",
3437 1.57 msaitoh i, cpu_id % ncpu);
3438 1.57 msaitoh else
3439 1.57 msaitoh aprint_normal("\n");
3440 1.57 msaitoh
3441 1.57 msaitoh #ifndef IXGBE_LEGACY_TX
3442 1.57 msaitoh txr->txr_si
3443 1.149 msaitoh = softint_establish(SOFTINT_NET | IXGBE_SOFTINT_FLAGS,
3444 1.57 msaitoh ixgbe_deferred_mq_start, txr);
3445 1.57 msaitoh #endif
3446 1.57 msaitoh que->que_si
3447 1.149 msaitoh = softint_establish(SOFTINT_NET | IXGBE_SOFTINT_FLAGS,
3448 1.57 msaitoh ixv_handle_que, que);
3449 1.57 msaitoh if (que->que_si == NULL) {
3450 1.57 msaitoh aprint_error_dev(dev,
3451 1.113 msaitoh "could not establish software interrupt\n");
3452 1.57 msaitoh }
3453 1.57 msaitoh }
3454 1.84 knakahar snprintf(wqname, sizeof(wqname), "%sdeferTx", device_xname(dev));
3455 1.186 msaitoh error = workqueue_create(&sc->txr_wq, wqname,
3456 1.186 msaitoh ixgbe_deferred_mq_start_work, sc, IXGBE_WORKQUEUE_PRI, IPL_NET,
3457 1.84 knakahar IXGBE_WORKQUEUE_FLAGS);
3458 1.84 knakahar if (error) {
3459 1.158 msaitoh aprint_error_dev(dev,
3460 1.158 msaitoh "couldn't create workqueue for deferred Tx\n");
3461 1.84 knakahar }
3462 1.186 msaitoh sc->txr_wq_enqueued = percpu_alloc(sizeof(u_int));
3463 1.84 knakahar
3464 1.84 knakahar snprintf(wqname, sizeof(wqname), "%sTxRx", device_xname(dev));
3465 1.186 msaitoh error = workqueue_create(&sc->que_wq, wqname,
3466 1.186 msaitoh ixv_handle_que_work, sc, IXGBE_WORKQUEUE_PRI, IPL_NET,
3467 1.84 knakahar IXGBE_WORKQUEUE_FLAGS);
3468 1.84 knakahar if (error) {
3469 1.158 msaitoh aprint_error_dev(dev, "couldn't create workqueue for Tx/Rx\n");
3470 1.84 knakahar }
3471 1.57 msaitoh
3472 1.57 msaitoh /* and Mailbox */
3473 1.57 msaitoh cpu_id++;
3474 1.57 msaitoh snprintf(intr_xname, sizeof(intr_xname), "%s link", device_xname(dev));
3475 1.186 msaitoh sc->vector = vector;
3476 1.186 msaitoh intrstr = pci_intr_string(pc, sc->osdep.intrs[vector], intrbuf,
3477 1.57 msaitoh sizeof(intrbuf));
3478 1.57 msaitoh #ifdef IXGBE_MPSAFE
3479 1.186 msaitoh pci_intr_setattr(pc, &sc->osdep.intrs[vector], PCI_INTR_MPSAFE,
3480 1.57 msaitoh true);
3481 1.57 msaitoh #endif
3482 1.57 msaitoh /* Set the mbx handler function */
3483 1.186 msaitoh sc->osdep.ihs[vector] = pci_intr_establish_xname(pc,
3484 1.186 msaitoh sc->osdep.intrs[vector], IPL_NET, ixv_msix_mbx, sc, intr_xname);
3485 1.186 msaitoh if (sc->osdep.ihs[vector] == NULL) {
3486 1.57 msaitoh aprint_error_dev(dev, "Failed to register LINK handler\n");
3487 1.57 msaitoh kcpuset_destroy(affinity);
3488 1.57 msaitoh return (ENXIO);
3489 1.57 msaitoh }
3490 1.57 msaitoh /* Round-robin affinity */
3491 1.57 msaitoh kcpuset_zero(affinity);
3492 1.57 msaitoh kcpuset_set(affinity, cpu_id % ncpu);
3493 1.186 msaitoh error = interrupt_distribute(sc->osdep.ihs[vector], affinity,
3494 1.150 msaitoh NULL);
3495 1.57 msaitoh
3496 1.57 msaitoh aprint_normal_dev(dev,
3497 1.57 msaitoh "for link, interrupting at %s", intrstr);
3498 1.57 msaitoh if (error == 0)
3499 1.57 msaitoh aprint_normal(", affinity to cpu %d\n", cpu_id % ncpu);
3500 1.57 msaitoh else
3501 1.57 msaitoh aprint_normal("\n");
3502 1.57 msaitoh
3503 1.57 msaitoh /* Tasklets for Mailbox */
3504 1.151 msaitoh snprintf(wqname, sizeof(wqname), "%s-admin", device_xname(dev));
3505 1.186 msaitoh error = workqueue_create(&sc->admin_wq, wqname,
3506 1.186 msaitoh ixv_handle_admin, sc, IXGBE_WORKQUEUE_PRI, IPL_NET,
3507 1.151 msaitoh IXGBE_TASKLET_WQ_FLAGS);
3508 1.151 msaitoh if (error) {
3509 1.151 msaitoh aprint_error_dev(dev,
3510 1.151 msaitoh "could not create admin workqueue (%d)\n", error);
3511 1.151 msaitoh goto err_out;
3512 1.151 msaitoh }
3513 1.151 msaitoh
3514 1.57 msaitoh /*
3515 1.58 msaitoh * Due to a broken design QEMU will fail to properly
3516 1.58 msaitoh * enable the guest for MSI-X unless the vectors in
3517 1.58 msaitoh * the table are all set up, so we must rewrite the
3518 1.58 msaitoh * ENABLE in the MSI-X control register again at this
3519 1.58 msaitoh * point to cause it to successfully initialize us.
3520 1.58 msaitoh */
3521 1.186 msaitoh if (sc->hw.mac.type == ixgbe_mac_82599_vf) {
3522 1.57 msaitoh pci_get_capability(pc, tag, PCI_CAP_MSIX, &rid, NULL);
3523 1.57 msaitoh rid += PCI_MSIX_CTL;
3524 1.57 msaitoh msix_ctrl = pci_conf_read(pc, tag, rid);
3525 1.57 msaitoh msix_ctrl |= PCI_MSIX_CTL_ENABLE;
3526 1.57 msaitoh pci_conf_write(pc, tag, rid, msix_ctrl);
3527 1.57 msaitoh }
3528 1.57 msaitoh
3529 1.57 msaitoh kcpuset_destroy(affinity);
3530 1.57 msaitoh return (0);
3531 1.151 msaitoh err_out:
3532 1.151 msaitoh kcpuset_destroy(affinity);
3533 1.186 msaitoh ixv_free_deferred_handlers(sc);
3534 1.186 msaitoh ixv_free_pci_resources(sc);
3535 1.151 msaitoh return (error);
3536 1.58 msaitoh } /* ixv_allocate_msix */
3537 1.57 msaitoh
3538 1.58 msaitoh /************************************************************************
3539 1.58 msaitoh * ixv_configure_interrupts - Setup MSI-X resources
3540 1.58 msaitoh *
3541 1.58 msaitoh * Note: The VF device MUST use MSI-X, there is no fallback.
3542 1.58 msaitoh ************************************************************************/
3543 1.57 msaitoh static int
3544 1.186 msaitoh ixv_configure_interrupts(struct ixgbe_softc *sc)
3545 1.57 msaitoh {
3546 1.186 msaitoh device_t dev = sc->dev;
3547 1.57 msaitoh int want, queues, msgs;
3548 1.57 msaitoh
3549 1.58 msaitoh /* Must have at least 2 MSI-X vectors */
3550 1.186 msaitoh msgs = pci_msix_count(sc->osdep.pc, sc->osdep.tag);
3551 1.57 msaitoh if (msgs < 2) {
3552 1.63 msaitoh aprint_error_dev(dev, "MSIX config error\n");
3553 1.57 msaitoh return (ENXIO);
3554 1.57 msaitoh }
3555 1.57 msaitoh msgs = MIN(msgs, IXG_MAX_NINTR);
3556 1.57 msaitoh
3557 1.57 msaitoh /* Figure out a reasonable auto config value */
3558 1.57 msaitoh queues = (ncpu > (msgs - 1)) ? (msgs - 1) : ncpu;
3559 1.57 msaitoh
3560 1.57 msaitoh if (ixv_num_queues != 0)
3561 1.57 msaitoh queues = ixv_num_queues;
3562 1.57 msaitoh else if ((ixv_num_queues == 0) && (queues > IXGBE_VF_MAX_TX_QUEUES))
3563 1.57 msaitoh queues = IXGBE_VF_MAX_TX_QUEUES;
3564 1.57 msaitoh
3565 1.57 msaitoh /*
3566 1.58 msaitoh * Want vectors for the queues,
3567 1.58 msaitoh * plus an additional for mailbox.
3568 1.58 msaitoh */
3569 1.57 msaitoh want = queues + 1;
3570 1.57 msaitoh if (msgs >= want)
3571 1.57 msaitoh msgs = want;
3572 1.57 msaitoh else {
3573 1.114 msaitoh aprint_error_dev(dev,
3574 1.58 msaitoh "MSI-X Configuration Problem, "
3575 1.182 msaitoh "%d vectors but %d queues wanted!\n", msgs, want);
3576 1.57 msaitoh return -1;
3577 1.57 msaitoh }
3578 1.57 msaitoh
3579 1.57 msaitoh aprint_normal_dev(dev,
3580 1.58 msaitoh "Using MSI-X interrupts with %d vectors\n", msgs);
3581 1.186 msaitoh sc->num_queues = queues;
3582 1.57 msaitoh
3583 1.58 msaitoh return (0);
3584 1.58 msaitoh } /* ixv_configure_interrupts */
3585 1.58 msaitoh
3586 1.58 msaitoh
3587 1.58 msaitoh /************************************************************************
3588 1.151 msaitoh * ixv_handle_admin - Tasklet handler for MSI-X MBX interrupts
3589 1.58 msaitoh *
3590 1.58 msaitoh * Done outside of interrupt context since the driver might sleep
3591 1.58 msaitoh ************************************************************************/
3592 1.57 msaitoh static void
3593 1.151 msaitoh ixv_handle_admin(struct work *wk, void *context)
3594 1.57 msaitoh {
3595 1.186 msaitoh struct ixgbe_softc *sc = context;
3596 1.186 msaitoh struct ixgbe_hw *hw = &sc->hw;
3597 1.57 msaitoh
3598 1.186 msaitoh IXGBE_CORE_LOCK(sc);
3599 1.89 knakahar
3600 1.186 msaitoh IXGBE_EVC_ADD(&sc->link_workev, 1);
3601 1.186 msaitoh sc->hw.mac.ops.check_link(&sc->hw, &sc->link_speed,
3602 1.186 msaitoh &sc->link_up, FALSE);
3603 1.186 msaitoh ixv_update_link_status(sc);
3604 1.89 knakahar
3605 1.186 msaitoh sc->task_requests = 0;
3606 1.186 msaitoh atomic_store_relaxed(&sc->admin_pending, 0);
3607 1.151 msaitoh
3608 1.151 msaitoh /* Re-enable interrupts */
3609 1.186 msaitoh IXGBE_WRITE_REG(hw, IXGBE_VTEIMS, (1 << sc->vector));
3610 1.151 msaitoh
3611 1.186 msaitoh IXGBE_CORE_UNLOCK(sc);
3612 1.151 msaitoh } /* ixv_handle_admin */
3613 1.57 msaitoh
3614 1.58 msaitoh /************************************************************************
3615 1.58 msaitoh * ixv_check_link - Used in the local timer to poll for link changes
3616 1.58 msaitoh ************************************************************************/
3617 1.117 msaitoh static s32
3618 1.186 msaitoh ixv_check_link(struct ixgbe_softc *sc)
3619 1.57 msaitoh {
3620 1.117 msaitoh s32 error;
3621 1.89 knakahar
3622 1.186 msaitoh KASSERT(mutex_owned(&sc->core_mtx));
3623 1.89 knakahar
3624 1.186 msaitoh sc->hw.mac.get_link_status = TRUE;
3625 1.57 msaitoh
3626 1.186 msaitoh error = sc->hw.mac.ops.check_link(&sc->hw,
3627 1.186 msaitoh &sc->link_speed, &sc->link_up, FALSE);
3628 1.186 msaitoh ixv_update_link_status(sc);
3629 1.117 msaitoh
3630 1.117 msaitoh return error;
3631 1.58 msaitoh } /* ixv_check_link */
3632