Home | History | Annotate | Line # | Download | only in ixgbe
ixv.c revision 1.141
      1 /*$NetBSD: ixv.c,v 1.141 2019/11/18 03:17:51 msaitoh Exp $*/
      2 
      3 /******************************************************************************
      4 
      5   Copyright (c) 2001-2017, Intel Corporation
      6   All rights reserved.
      7 
      8   Redistribution and use in source and binary forms, with or without
      9   modification, are permitted provided that the following conditions are met:
     10 
     11    1. Redistributions of source code must retain the above copyright notice,
     12       this list of conditions and the following disclaimer.
     13 
     14    2. Redistributions in binary form must reproduce the above copyright
     15       notice, this list of conditions and the following disclaimer in the
     16       documentation and/or other materials provided with the distribution.
     17 
     18    3. Neither the name of the Intel Corporation nor the names of its
     19       contributors may be used to endorse or promote products derived from
     20       this software without specific prior written permission.
     21 
     22   THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
     23   AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
     24   IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
     25   ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
     26   LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
     27   CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
     28   SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
     29   INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
     30   CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
     31   ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
     32   POSSIBILITY OF SUCH DAMAGE.
     33 
     34 ******************************************************************************/
     35 /*$FreeBSD: head/sys/dev/ixgbe/if_ixv.c 331224 2018-03-19 20:55:05Z erj $*/
     36 
     37 #ifdef _KERNEL_OPT
     38 #include "opt_inet.h"
     39 #include "opt_inet6.h"
     40 #include "opt_net_mpsafe.h"
     41 #endif
     42 
     43 #include "ixgbe.h"
     44 #include "vlan.h"
     45 
     46 /************************************************************************
     47  * Driver version
     48  ************************************************************************/
     49 static const char ixv_driver_version[] = "2.0.1-k";
     50 /* XXX NetBSD: + 1.5.17 */
     51 
     52 /************************************************************************
     53  * PCI Device ID Table
     54  *
     55  *   Used by probe to select devices to load on
     56  *   Last field stores an index into ixv_strings
     57  *   Last entry must be all 0s
     58  *
     59  *   { Vendor ID, Device ID, SubVendor ID, SubDevice ID, String Index }
     60  ************************************************************************/
     61 static const ixgbe_vendor_info_t ixv_vendor_info_array[] =
     62 {
     63 	{IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_82599_VF, 0, 0, 0},
     64 	{IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_X540_VF, 0, 0, 0},
     65 	{IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_X550_VF, 0, 0, 0},
     66 	{IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_X550EM_X_VF, 0, 0, 0},
     67 	{IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_X550EM_A_VF, 0, 0, 0},
     68 	/* required last entry */
     69 	{0, 0, 0, 0, 0}
     70 };
     71 
     72 /************************************************************************
     73  * Table of branding strings
     74  ************************************************************************/
     75 static const char *ixv_strings[] = {
     76 	"Intel(R) PRO/10GbE Virtual Function Network Driver"
     77 };
     78 
     79 /*********************************************************************
     80  *  Function prototypes
     81  *********************************************************************/
     82 static int	ixv_probe(device_t, cfdata_t, void *);
     83 static void	ixv_attach(device_t, device_t, void *);
     84 static int	ixv_detach(device_t, int);
     85 #if 0
     86 static int	ixv_shutdown(device_t);
     87 #endif
     88 static int	ixv_ifflags_cb(struct ethercom *);
     89 static int	ixv_ioctl(struct ifnet *, u_long, void *);
     90 static int	ixv_init(struct ifnet *);
     91 static void	ixv_init_locked(struct adapter *);
     92 static void	ixv_ifstop(struct ifnet *, int);
     93 static void	ixv_stop(void *);
     94 static void	ixv_init_device_features(struct adapter *);
     95 static void	ixv_media_status(struct ifnet *, struct ifmediareq *);
     96 static int	ixv_media_change(struct ifnet *);
     97 static int	ixv_allocate_pci_resources(struct adapter *,
     98 		    const struct pci_attach_args *);
     99 static int	ixv_allocate_msix(struct adapter *,
    100 		    const struct pci_attach_args *);
    101 static int	ixv_configure_interrupts(struct adapter *);
    102 static void	ixv_free_pci_resources(struct adapter *);
    103 static void	ixv_local_timer(void *);
    104 static void	ixv_local_timer_locked(void *);
    105 static int	ixv_setup_interface(device_t, struct adapter *);
    106 static int	ixv_negotiate_api(struct adapter *);
    107 
    108 static void	ixv_initialize_transmit_units(struct adapter *);
    109 static void	ixv_initialize_receive_units(struct adapter *);
    110 static void	ixv_initialize_rss_mapping(struct adapter *);
    111 static s32	ixv_check_link(struct adapter *);
    112 
    113 static void	ixv_enable_intr(struct adapter *);
    114 static void	ixv_disable_intr(struct adapter *);
    115 static int	ixv_set_rxfilter(struct adapter *);
    116 static void	ixv_update_link_status(struct adapter *);
    117 static int	ixv_sysctl_debug(SYSCTLFN_PROTO);
    118 static void	ixv_set_ivar(struct adapter *, u8, u8, s8);
    119 static void	ixv_configure_ivars(struct adapter *);
    120 static u8 *	ixv_mc_array_itr(struct ixgbe_hw *, u8 **, u32 *);
    121 static void	ixv_eitr_write(struct adapter *, uint32_t, uint32_t);
    122 
    123 static void	ixv_setup_vlan_tagging(struct adapter *);
    124 static int	ixv_setup_vlan_support(struct adapter *);
    125 static int	ixv_vlan_cb(struct ethercom *, uint16_t, bool);
    126 static int	ixv_register_vlan(struct adapter *, u16);
    127 static int	ixv_unregister_vlan(struct adapter *, u16);
    128 
    129 static void	ixv_add_device_sysctls(struct adapter *);
    130 static void	ixv_save_stats(struct adapter *);
    131 static void	ixv_init_stats(struct adapter *);
    132 static void	ixv_update_stats(struct adapter *);
    133 static void	ixv_add_stats_sysctls(struct adapter *);
    134 static void	ixv_clear_evcnt(struct adapter *);
    135 
    136 /* Sysctl handlers */
    137 static void	ixv_set_sysctl_value(struct adapter *, const char *,
    138 		    const char *, int *, int);
    139 static int	ixv_sysctl_interrupt_rate_handler(SYSCTLFN_PROTO);
    140 static int	ixv_sysctl_next_to_check_handler(SYSCTLFN_PROTO);
    141 static int	ixv_sysctl_rdh_handler(SYSCTLFN_PROTO);
    142 static int	ixv_sysctl_rdt_handler(SYSCTLFN_PROTO);
    143 static int	ixv_sysctl_tdt_handler(SYSCTLFN_PROTO);
    144 static int	ixv_sysctl_tdh_handler(SYSCTLFN_PROTO);
    145 
    146 /* The MSI-X Interrupt handlers */
    147 static int	ixv_msix_que(void *);
    148 static int	ixv_msix_mbx(void *);
    149 
    150 /* Deferred interrupt tasklets */
    151 static void	ixv_handle_que(void *);
    152 static void	ixv_handle_link(void *);
    153 
    154 /* Workqueue handler for deferred work */
    155 static void	ixv_handle_que_work(struct work *, void *);
    156 
    157 const struct sysctlnode *ixv_sysctl_instance(struct adapter *);
    158 static const ixgbe_vendor_info_t *ixv_lookup(const struct pci_attach_args *);
    159 
    160 /************************************************************************
    161  * FreeBSD Device Interface Entry Points
    162  ************************************************************************/
    163 CFATTACH_DECL3_NEW(ixv, sizeof(struct adapter),
    164     ixv_probe, ixv_attach, ixv_detach, NULL, NULL, NULL,
    165     DVF_DETACH_SHUTDOWN);
    166 
    167 #if 0
    168 static driver_t ixv_driver = {
    169 	"ixv", ixv_methods, sizeof(struct adapter),
    170 };
    171 
    172 devclass_t ixv_devclass;
    173 DRIVER_MODULE(ixv, pci, ixv_driver, ixv_devclass, 0, 0);
    174 MODULE_DEPEND(ixv, pci, 1, 1, 1);
    175 MODULE_DEPEND(ixv, ether, 1, 1, 1);
    176 #endif
    177 
    178 /*
    179  * TUNEABLE PARAMETERS:
    180  */
    181 
    182 /* Number of Queues - do not exceed MSI-X vectors - 1 */
    183 static int ixv_num_queues = 0;
    184 #define	TUNABLE_INT(__x, __y)
    185 TUNABLE_INT("hw.ixv.num_queues", &ixv_num_queues);
    186 
    187 /*
    188  * AIM: Adaptive Interrupt Moderation
    189  * which means that the interrupt rate
    190  * is varied over time based on the
    191  * traffic for that interrupt vector
    192  */
    193 static bool ixv_enable_aim = false;
    194 TUNABLE_INT("hw.ixv.enable_aim", &ixv_enable_aim);
    195 
    196 static int ixv_max_interrupt_rate = (4000000 / IXGBE_LOW_LATENCY);
    197 TUNABLE_INT("hw.ixv.max_interrupt_rate", &ixv_max_interrupt_rate);
    198 
    199 /* How many packets rxeof tries to clean at a time */
    200 static int ixv_rx_process_limit = 256;
    201 TUNABLE_INT("hw.ixv.rx_process_limit", &ixv_rx_process_limit);
    202 
    203 /* How many packets txeof tries to clean at a time */
    204 static int ixv_tx_process_limit = 256;
    205 TUNABLE_INT("hw.ixv.tx_process_limit", &ixv_tx_process_limit);
    206 
    207 /* Which packet processing uses workqueue or softint */
    208 static bool ixv_txrx_workqueue = false;
    209 
    210 /*
    211  * Number of TX descriptors per ring,
    212  * setting higher than RX as this seems
    213  * the better performing choice.
    214  */
    215 static int ixv_txd = PERFORM_TXD;
    216 TUNABLE_INT("hw.ixv.txd", &ixv_txd);
    217 
    218 /* Number of RX descriptors per ring */
    219 static int ixv_rxd = PERFORM_RXD;
    220 TUNABLE_INT("hw.ixv.rxd", &ixv_rxd);
    221 
    222 /* Legacy Transmit (single queue) */
    223 static int ixv_enable_legacy_tx = 0;
    224 TUNABLE_INT("hw.ixv.enable_legacy_tx", &ixv_enable_legacy_tx);
    225 
    226 #ifdef NET_MPSAFE
    227 #define IXGBE_MPSAFE		1
    228 #define IXGBE_CALLOUT_FLAGS	CALLOUT_MPSAFE
    229 #define IXGBE_SOFTINFT_FLAGS	SOFTINT_MPSAFE
    230 #define IXGBE_WORKQUEUE_FLAGS	WQ_PERCPU | WQ_MPSAFE
    231 #else
    232 #define IXGBE_CALLOUT_FLAGS	0
    233 #define IXGBE_SOFTINFT_FLAGS	0
    234 #define IXGBE_WORKQUEUE_FLAGS	WQ_PERCPU
    235 #endif
    236 #define IXGBE_WORKQUEUE_PRI PRI_SOFTNET
    237 
    238 #if 0
    239 static int (*ixv_start_locked)(struct ifnet *, struct tx_ring *);
    240 static int (*ixv_ring_empty)(struct ifnet *, struct buf_ring *);
    241 #endif
    242 
    243 /************************************************************************
    244  * ixv_probe - Device identification routine
    245  *
    246  *   Determines if the driver should be loaded on
    247  *   adapter based on its PCI vendor/device ID.
    248  *
    249  *   return BUS_PROBE_DEFAULT on success, positive on failure
    250  ************************************************************************/
    251 static int
    252 ixv_probe(device_t dev, cfdata_t cf, void *aux)
    253 {
    254 #ifdef __HAVE_PCI_MSI_MSIX
    255 	const struct pci_attach_args *pa = aux;
    256 
    257 	return (ixv_lookup(pa) != NULL) ? 1 : 0;
    258 #else
    259 	return 0;
    260 #endif
    261 } /* ixv_probe */
    262 
    263 static const ixgbe_vendor_info_t *
    264 ixv_lookup(const struct pci_attach_args *pa)
    265 {
    266 	const ixgbe_vendor_info_t *ent;
    267 	pcireg_t subid;
    268 
    269 	INIT_DEBUGOUT("ixv_lookup: begin");
    270 
    271 	if (PCI_VENDOR(pa->pa_id) != IXGBE_INTEL_VENDOR_ID)
    272 		return NULL;
    273 
    274 	subid = pci_conf_read(pa->pa_pc, pa->pa_tag, PCI_SUBSYS_ID_REG);
    275 
    276 	for (ent = ixv_vendor_info_array; ent->vendor_id != 0; ent++) {
    277 		if ((PCI_VENDOR(pa->pa_id) == ent->vendor_id) &&
    278 		    (PCI_PRODUCT(pa->pa_id) == ent->device_id) &&
    279 		    ((PCI_SUBSYS_VENDOR(subid) == ent->subvendor_id) ||
    280 		     (ent->subvendor_id == 0)) &&
    281 		    ((PCI_SUBSYS_ID(subid) == ent->subdevice_id) ||
    282 		     (ent->subdevice_id == 0))) {
    283 			return ent;
    284 		}
    285 	}
    286 
    287 	return NULL;
    288 }
    289 
    290 /************************************************************************
    291  * ixv_attach - Device initialization routine
    292  *
    293  *   Called when the driver is being loaded.
    294  *   Identifies the type of hardware, allocates all resources
    295  *   and initializes the hardware.
    296  *
    297  *   return 0 on success, positive on failure
    298  ************************************************************************/
    299 static void
    300 ixv_attach(device_t parent, device_t dev, void *aux)
    301 {
    302 	struct adapter *adapter;
    303 	struct ixgbe_hw *hw;
    304 	int		error = 0;
    305 	pcireg_t	id, subid;
    306 	const ixgbe_vendor_info_t *ent;
    307 	const struct pci_attach_args *pa = aux;
    308 	const char *apivstr;
    309 	const char *str;
    310 	char buf[256];
    311 
    312 	INIT_DEBUGOUT("ixv_attach: begin");
    313 
    314 	/*
    315 	 * Make sure BUSMASTER is set, on a VM under
    316 	 * KVM it may not be and will break things.
    317 	 */
    318 	ixgbe_pci_enable_busmaster(pa->pa_pc, pa->pa_tag);
    319 
    320 	/* Allocate, clear, and link in our adapter structure */
    321 	adapter = device_private(dev);
    322 	adapter->dev = dev;
    323 	adapter->hw.back = adapter;
    324 	hw = &adapter->hw;
    325 
    326 	adapter->init_locked = ixv_init_locked;
    327 	adapter->stop_locked = ixv_stop;
    328 
    329 	adapter->osdep.pc = pa->pa_pc;
    330 	adapter->osdep.tag = pa->pa_tag;
    331 	if (pci_dma64_available(pa))
    332 		adapter->osdep.dmat = pa->pa_dmat64;
    333 	else
    334 		adapter->osdep.dmat = pa->pa_dmat;
    335 	adapter->osdep.attached = false;
    336 
    337 	ent = ixv_lookup(pa);
    338 
    339 	KASSERT(ent != NULL);
    340 
    341 	aprint_normal(": %s, Version - %s\n",
    342 	    ixv_strings[ent->index], ixv_driver_version);
    343 
    344 	/* Core Lock Init*/
    345 	IXGBE_CORE_LOCK_INIT(adapter, device_xname(dev));
    346 
    347 	/* Do base PCI setup - map BAR0 */
    348 	if (ixv_allocate_pci_resources(adapter, pa)) {
    349 		aprint_error_dev(dev, "ixv_allocate_pci_resources() failed!\n");
    350 		error = ENXIO;
    351 		goto err_out;
    352 	}
    353 
    354 	/* SYSCTL APIs */
    355 	ixv_add_device_sysctls(adapter);
    356 
    357 	/* Set up the timer callout */
    358 	callout_init(&adapter->timer, IXGBE_CALLOUT_FLAGS);
    359 
    360 	/* Save off the information about this board */
    361 	id = pci_conf_read(pa->pa_pc, pa->pa_tag, PCI_ID_REG);
    362 	subid = pci_conf_read(pa->pa_pc, pa->pa_tag, PCI_SUBSYS_ID_REG);
    363 	hw->vendor_id = PCI_VENDOR(id);
    364 	hw->device_id = PCI_PRODUCT(id);
    365 	hw->revision_id =
    366 	    PCI_REVISION(pci_conf_read(pa->pa_pc, pa->pa_tag, PCI_CLASS_REG));
    367 	hw->subsystem_vendor_id = PCI_SUBSYS_VENDOR(subid);
    368 	hw->subsystem_device_id = PCI_SUBSYS_ID(subid);
    369 
    370 	/* A subset of set_mac_type */
    371 	switch (hw->device_id) {
    372 	case IXGBE_DEV_ID_82599_VF:
    373 		hw->mac.type = ixgbe_mac_82599_vf;
    374 		str = "82599 VF";
    375 		break;
    376 	case IXGBE_DEV_ID_X540_VF:
    377 		hw->mac.type = ixgbe_mac_X540_vf;
    378 		str = "X540 VF";
    379 		break;
    380 	case IXGBE_DEV_ID_X550_VF:
    381 		hw->mac.type = ixgbe_mac_X550_vf;
    382 		str = "X550 VF";
    383 		break;
    384 	case IXGBE_DEV_ID_X550EM_X_VF:
    385 		hw->mac.type = ixgbe_mac_X550EM_x_vf;
    386 		str = "X550EM X VF";
    387 		break;
    388 	case IXGBE_DEV_ID_X550EM_A_VF:
    389 		hw->mac.type = ixgbe_mac_X550EM_a_vf;
    390 		str = "X550EM A VF";
    391 		break;
    392 	default:
    393 		/* Shouldn't get here since probe succeeded */
    394 		aprint_error_dev(dev, "Unknown device ID!\n");
    395 		error = ENXIO;
    396 		goto err_out;
    397 		break;
    398 	}
    399 	aprint_normal_dev(dev, "device %s\n", str);
    400 
    401 	ixv_init_device_features(adapter);
    402 
    403 	/* Initialize the shared code */
    404 	error = ixgbe_init_ops_vf(hw);
    405 	if (error) {
    406 		aprint_error_dev(dev, "ixgbe_init_ops_vf() failed!\n");
    407 		error = EIO;
    408 		goto err_out;
    409 	}
    410 
    411 	/* Setup the mailbox */
    412 	ixgbe_init_mbx_params_vf(hw);
    413 
    414 	/* Set the right number of segments */
    415 	adapter->num_segs = IXGBE_82599_SCATTER;
    416 
    417 	/* Reset mbox api to 1.0 */
    418 	error = hw->mac.ops.reset_hw(hw);
    419 	if (error == IXGBE_ERR_RESET_FAILED)
    420 		aprint_error_dev(dev, "...reset_hw() failure: Reset Failed!\n");
    421 	else if (error)
    422 		aprint_error_dev(dev, "...reset_hw() failed with error %d\n",
    423 		    error);
    424 	if (error) {
    425 		error = EIO;
    426 		goto err_out;
    427 	}
    428 
    429 	error = hw->mac.ops.init_hw(hw);
    430 	if (error) {
    431 		aprint_error_dev(dev, "...init_hw() failed!\n");
    432 		error = EIO;
    433 		goto err_out;
    434 	}
    435 
    436 	/* Negotiate mailbox API version */
    437 	error = ixv_negotiate_api(adapter);
    438 	if (error)
    439 		aprint_normal_dev(dev,
    440 		    "MBX API negotiation failed during attach!\n");
    441 	switch (hw->api_version) {
    442 	case ixgbe_mbox_api_10:
    443 		apivstr = "1.0";
    444 		break;
    445 	case ixgbe_mbox_api_20:
    446 		apivstr = "2.0";
    447 		break;
    448 	case ixgbe_mbox_api_11:
    449 		apivstr = "1.1";
    450 		break;
    451 	case ixgbe_mbox_api_12:
    452 		apivstr = "1.2";
    453 		break;
    454 	case ixgbe_mbox_api_13:
    455 		apivstr = "1.3";
    456 		break;
    457 	default:
    458 		apivstr = "unknown";
    459 		break;
    460 	}
    461 	aprint_normal_dev(dev, "Mailbox API %s\n", apivstr);
    462 
    463 	/* If no mac address was assigned, make a random one */
    464 	if (!ixv_check_ether_addr(hw->mac.addr)) {
    465 		u8 addr[ETHER_ADDR_LEN];
    466 		uint64_t rndval = cprng_strong64();
    467 
    468 		memcpy(addr, &rndval, sizeof(addr));
    469 		addr[0] &= 0xFE;
    470 		addr[0] |= 0x02;
    471 		bcopy(addr, hw->mac.addr, sizeof(addr));
    472 	}
    473 
    474 	/* Register for VLAN events */
    475 	ether_set_vlan_cb(&adapter->osdep.ec, ixv_vlan_cb);
    476 
    477 	/* Sysctls for limiting the amount of work done in the taskqueues */
    478 	ixv_set_sysctl_value(adapter, "rx_processing_limit",
    479 	    "max number of rx packets to process",
    480 	    &adapter->rx_process_limit, ixv_rx_process_limit);
    481 
    482 	ixv_set_sysctl_value(adapter, "tx_processing_limit",
    483 	    "max number of tx packets to process",
    484 	    &adapter->tx_process_limit, ixv_tx_process_limit);
    485 
    486 	/* Do descriptor calc and sanity checks */
    487 	if (((ixv_txd * sizeof(union ixgbe_adv_tx_desc)) % DBA_ALIGN) != 0 ||
    488 	    ixv_txd < MIN_TXD || ixv_txd > MAX_TXD) {
    489 		aprint_error_dev(dev, "TXD config issue, using default!\n");
    490 		adapter->num_tx_desc = DEFAULT_TXD;
    491 	} else
    492 		adapter->num_tx_desc = ixv_txd;
    493 
    494 	if (((ixv_rxd * sizeof(union ixgbe_adv_rx_desc)) % DBA_ALIGN) != 0 ||
    495 	    ixv_rxd < MIN_RXD || ixv_rxd > MAX_RXD) {
    496 		aprint_error_dev(dev, "RXD config issue, using default!\n");
    497 		adapter->num_rx_desc = DEFAULT_RXD;
    498 	} else
    499 		adapter->num_rx_desc = ixv_rxd;
    500 
    501 	/* Setup MSI-X */
    502 	error = ixv_configure_interrupts(adapter);
    503 	if (error)
    504 		goto err_out;
    505 
    506 	/* Allocate our TX/RX Queues */
    507 	if (ixgbe_allocate_queues(adapter)) {
    508 		aprint_error_dev(dev, "ixgbe_allocate_queues() failed!\n");
    509 		error = ENOMEM;
    510 		goto err_out;
    511 	}
    512 
    513 	/* hw.ix defaults init */
    514 	adapter->enable_aim = ixv_enable_aim;
    515 
    516 	adapter->txrx_use_workqueue = ixv_txrx_workqueue;
    517 
    518 	error = ixv_allocate_msix(adapter, pa);
    519 	if (error) {
    520 		aprint_error_dev(dev, "ixv_allocate_msix() failed!\n");
    521 		goto err_late;
    522 	}
    523 
    524 	/* Setup OS specific network interface */
    525 	error = ixv_setup_interface(dev, adapter);
    526 	if (error != 0) {
    527 		aprint_error_dev(dev, "ixv_setup_interface() failed!\n");
    528 		goto err_late;
    529 	}
    530 
    531 	/* Do the stats setup */
    532 	ixv_save_stats(adapter);
    533 	ixv_init_stats(adapter);
    534 	ixv_add_stats_sysctls(adapter);
    535 
    536 	if (adapter->feat_en & IXGBE_FEATURE_NETMAP)
    537 		ixgbe_netmap_attach(adapter);
    538 
    539 	snprintb(buf, sizeof(buf), IXGBE_FEATURE_FLAGS, adapter->feat_cap);
    540 	aprint_verbose_dev(dev, "feature cap %s\n", buf);
    541 	snprintb(buf, sizeof(buf), IXGBE_FEATURE_FLAGS, adapter->feat_en);
    542 	aprint_verbose_dev(dev, "feature ena %s\n", buf);
    543 
    544 	INIT_DEBUGOUT("ixv_attach: end");
    545 	adapter->osdep.attached = true;
    546 
    547 	return;
    548 
    549 err_late:
    550 	ixgbe_free_transmit_structures(adapter);
    551 	ixgbe_free_receive_structures(adapter);
    552 	free(adapter->queues, M_DEVBUF);
    553 err_out:
    554 	ixv_free_pci_resources(adapter);
    555 	IXGBE_CORE_LOCK_DESTROY(adapter);
    556 
    557 	return;
    558 } /* ixv_attach */
    559 
    560 /************************************************************************
    561  * ixv_detach - Device removal routine
    562  *
    563  *   Called when the driver is being removed.
    564  *   Stops the adapter and deallocates all the resources
    565  *   that were allocated for driver operation.
    566  *
    567  *   return 0 on success, positive on failure
    568  ************************************************************************/
    569 static int
    570 ixv_detach(device_t dev, int flags)
    571 {
    572 	struct adapter	*adapter = device_private(dev);
    573 	struct ixgbe_hw *hw = &adapter->hw;
    574 	struct ix_queue *que = adapter->queues;
    575 	struct tx_ring *txr = adapter->tx_rings;
    576 	struct rx_ring *rxr = adapter->rx_rings;
    577 	struct ixgbevf_hw_stats *stats = &adapter->stats.vf;
    578 
    579 	INIT_DEBUGOUT("ixv_detach: begin");
    580 	if (adapter->osdep.attached == false)
    581 		return 0;
    582 
    583 	/* Stop the interface. Callouts are stopped in it. */
    584 	ixv_ifstop(adapter->ifp, 1);
    585 
    586 #if NVLAN > 0
    587 	/* Make sure VLANs are not using driver */
    588 	if (!VLAN_ATTACHED(&adapter->osdep.ec))
    589 		;	/* nothing to do: no VLANs */
    590 	else if ((flags & (DETACH_SHUTDOWN | DETACH_FORCE)) != 0)
    591 		vlan_ifdetach(adapter->ifp);
    592 	else {
    593 		aprint_error_dev(dev, "VLANs in use, detach first\n");
    594 		return EBUSY;
    595 	}
    596 #endif
    597 
    598 	for (int i = 0; i < adapter->num_queues; i++, que++, txr++) {
    599 		if (!(adapter->feat_en & IXGBE_FEATURE_LEGACY_TX))
    600 			softint_disestablish(txr->txr_si);
    601 		softint_disestablish(que->que_si);
    602 	}
    603 	if (adapter->txr_wq != NULL)
    604 		workqueue_destroy(adapter->txr_wq);
    605 	if (adapter->txr_wq_enqueued != NULL)
    606 		percpu_free(adapter->txr_wq_enqueued, sizeof(u_int));
    607 	if (adapter->que_wq != NULL)
    608 		workqueue_destroy(adapter->que_wq);
    609 
    610 	/* Drain the Mailbox(link) queue */
    611 	softint_disestablish(adapter->link_si);
    612 
    613 	ether_ifdetach(adapter->ifp);
    614 	callout_halt(&adapter->timer, NULL);
    615 
    616 	if (adapter->feat_en & IXGBE_FEATURE_NETMAP)
    617 		netmap_detach(adapter->ifp);
    618 
    619 	ixv_free_pci_resources(adapter);
    620 #if 0 /* XXX the NetBSD port is probably missing something here */
    621 	bus_generic_detach(dev);
    622 #endif
    623 	if_detach(adapter->ifp);
    624 	if_percpuq_destroy(adapter->ipq);
    625 
    626 	sysctl_teardown(&adapter->sysctllog);
    627 	evcnt_detach(&adapter->efbig_tx_dma_setup);
    628 	evcnt_detach(&adapter->mbuf_defrag_failed);
    629 	evcnt_detach(&adapter->efbig2_tx_dma_setup);
    630 	evcnt_detach(&adapter->einval_tx_dma_setup);
    631 	evcnt_detach(&adapter->other_tx_dma_setup);
    632 	evcnt_detach(&adapter->eagain_tx_dma_setup);
    633 	evcnt_detach(&adapter->enomem_tx_dma_setup);
    634 	evcnt_detach(&adapter->watchdog_events);
    635 	evcnt_detach(&adapter->tso_err);
    636 	evcnt_detach(&adapter->link_irq);
    637 
    638 	txr = adapter->tx_rings;
    639 	for (int i = 0; i < adapter->num_queues; i++, rxr++, txr++) {
    640 		evcnt_detach(&adapter->queues[i].irqs);
    641 		evcnt_detach(&adapter->queues[i].handleq);
    642 		evcnt_detach(&adapter->queues[i].req);
    643 		evcnt_detach(&txr->no_desc_avail);
    644 		evcnt_detach(&txr->total_packets);
    645 		evcnt_detach(&txr->tso_tx);
    646 #ifndef IXGBE_LEGACY_TX
    647 		evcnt_detach(&txr->pcq_drops);
    648 #endif
    649 
    650 		evcnt_detach(&rxr->rx_packets);
    651 		evcnt_detach(&rxr->rx_bytes);
    652 		evcnt_detach(&rxr->rx_copies);
    653 		evcnt_detach(&rxr->no_jmbuf);
    654 		evcnt_detach(&rxr->rx_discarded);
    655 	}
    656 	evcnt_detach(&stats->ipcs);
    657 	evcnt_detach(&stats->l4cs);
    658 	evcnt_detach(&stats->ipcs_bad);
    659 	evcnt_detach(&stats->l4cs_bad);
    660 
    661 	/* Packet Reception Stats */
    662 	evcnt_detach(&stats->vfgorc);
    663 	evcnt_detach(&stats->vfgprc);
    664 	evcnt_detach(&stats->vfmprc);
    665 
    666 	/* Packet Transmission Stats */
    667 	evcnt_detach(&stats->vfgotc);
    668 	evcnt_detach(&stats->vfgptc);
    669 
    670 	/* Mailbox Stats */
    671 	evcnt_detach(&hw->mbx.stats.msgs_tx);
    672 	evcnt_detach(&hw->mbx.stats.msgs_rx);
    673 	evcnt_detach(&hw->mbx.stats.acks);
    674 	evcnt_detach(&hw->mbx.stats.reqs);
    675 	evcnt_detach(&hw->mbx.stats.rsts);
    676 
    677 	ixgbe_free_transmit_structures(adapter);
    678 	ixgbe_free_receive_structures(adapter);
    679 	for (int i = 0; i < adapter->num_queues; i++) {
    680 		struct ix_queue *lque = &adapter->queues[i];
    681 		mutex_destroy(&lque->dc_mtx);
    682 	}
    683 	free(adapter->queues, M_DEVBUF);
    684 
    685 	IXGBE_CORE_LOCK_DESTROY(adapter);
    686 
    687 	return (0);
    688 } /* ixv_detach */
    689 
    690 /************************************************************************
    691  * ixv_init_locked - Init entry point
    692  *
    693  *   Used in two ways: It is used by the stack as an init entry
    694  *   point in network interface structure. It is also used
    695  *   by the driver as a hw/sw initialization routine to get
    696  *   to a consistent state.
    697  *
    698  *   return 0 on success, positive on failure
    699  ************************************************************************/
    700 static void
    701 ixv_init_locked(struct adapter *adapter)
    702 {
    703 	struct ifnet	*ifp = adapter->ifp;
    704 	device_t	dev = adapter->dev;
    705 	struct ixgbe_hw *hw = &adapter->hw;
    706 	struct ix_queue	*que;
    707 	int		error = 0;
    708 	uint32_t mask;
    709 	int i;
    710 
    711 	INIT_DEBUGOUT("ixv_init_locked: begin");
    712 	KASSERT(mutex_owned(&adapter->core_mtx));
    713 	hw->adapter_stopped = FALSE;
    714 	hw->mac.ops.stop_adapter(hw);
    715 	callout_stop(&adapter->timer);
    716 	for (i = 0, que = adapter->queues; i < adapter->num_queues; i++, que++)
    717 		que->disabled_count = 0;
    718 
    719 	adapter->max_frame_size =
    720 	    ifp->if_mtu + ETHER_HDR_LEN + ETHER_CRC_LEN;
    721 
    722 	/* reprogram the RAR[0] in case user changed it. */
    723 	hw->mac.ops.set_rar(hw, 0, hw->mac.addr, 0, IXGBE_RAH_AV);
    724 
    725 	/* Get the latest mac address, User can use a LAA */
    726 	memcpy(hw->mac.addr, CLLADDR(ifp->if_sadl),
    727 	     IXGBE_ETH_LENGTH_OF_ADDRESS);
    728 	hw->mac.ops.set_rar(hw, 0, hw->mac.addr, 0, 1);
    729 
    730 	/* Prepare transmit descriptors and buffers */
    731 	if (ixgbe_setup_transmit_structures(adapter)) {
    732 		aprint_error_dev(dev, "Could not setup transmit structures\n");
    733 		ixv_stop(adapter);
    734 		return;
    735 	}
    736 
    737 	/* Reset VF and renegotiate mailbox API version */
    738 	hw->mac.ops.reset_hw(hw);
    739 	hw->mac.ops.start_hw(hw);
    740 	error = ixv_negotiate_api(adapter);
    741 	if (error)
    742 		device_printf(dev,
    743 		    "Mailbox API negotiation failed in init_locked!\n");
    744 
    745 	ixv_initialize_transmit_units(adapter);
    746 
    747 	/* Setup Multicast table */
    748 	ixv_set_rxfilter(adapter);
    749 
    750 	/*
    751 	 * Determine the correct mbuf pool
    752 	 * for doing jumbo/headersplit
    753 	 */
    754 	if (adapter->max_frame_size <= MCLBYTES)
    755 		adapter->rx_mbuf_sz = MCLBYTES;
    756 	else
    757 		adapter->rx_mbuf_sz = MJUMPAGESIZE;
    758 
    759 	/* Prepare receive descriptors and buffers */
    760 	if (ixgbe_setup_receive_structures(adapter)) {
    761 		device_printf(dev, "Could not setup receive structures\n");
    762 		ixv_stop(adapter);
    763 		return;
    764 	}
    765 
    766 	/* Configure RX settings */
    767 	ixv_initialize_receive_units(adapter);
    768 
    769 #if 0 /* XXX isn't it required? -- msaitoh  */
    770 	/* Set the various hardware offload abilities */
    771 	ifp->if_hwassist = 0;
    772 	if (ifp->if_capenable & IFCAP_TSO4)
    773 		ifp->if_hwassist |= CSUM_TSO;
    774 	if (ifp->if_capenable & IFCAP_TXCSUM) {
    775 		ifp->if_hwassist |= (CSUM_TCP | CSUM_UDP);
    776 #if __FreeBSD_version >= 800000
    777 		ifp->if_hwassist |= CSUM_SCTP;
    778 #endif
    779 	}
    780 #endif
    781 
    782 	/* Set up VLAN offload and filter */
    783 	ixv_setup_vlan_support(adapter);
    784 
    785 	/* Set up MSI-X routing */
    786 	ixv_configure_ivars(adapter);
    787 
    788 	/* Set up auto-mask */
    789 	mask = (1 << adapter->vector);
    790 	for (i = 0, que = adapter->queues; i < adapter->num_queues; i++, que++)
    791 		mask |= (1 << que->msix);
    792 	IXGBE_WRITE_REG(hw, IXGBE_VTEIAM, mask);
    793 
    794 	/* Set moderation on the Link interrupt */
    795 	ixv_eitr_write(adapter, adapter->vector, IXGBE_LINK_ITR);
    796 
    797 	/* Stats init */
    798 	ixv_init_stats(adapter);
    799 
    800 	/* Config/Enable Link */
    801 	hw->mac.get_link_status = TRUE;
    802 	hw->mac.ops.check_link(hw, &adapter->link_speed, &adapter->link_up,
    803 	    FALSE);
    804 
    805 	/* Start watchdog */
    806 	callout_reset(&adapter->timer, hz, ixv_local_timer, adapter);
    807 
    808 	/* And now turn on interrupts */
    809 	ixv_enable_intr(adapter);
    810 
    811 	/* Update saved flags. See ixgbe_ifflags_cb() */
    812 	adapter->if_flags = ifp->if_flags;
    813 	adapter->ec_capenable = adapter->osdep.ec.ec_capenable;
    814 
    815 	/* Now inform the stack we're ready */
    816 	ifp->if_flags |= IFF_RUNNING;
    817 	ifp->if_flags &= ~IFF_OACTIVE;
    818 
    819 	return;
    820 } /* ixv_init_locked */
    821 
    822 /************************************************************************
    823  * ixv_enable_queue
    824  ************************************************************************/
    825 static inline void
    826 ixv_enable_queue(struct adapter *adapter, u32 vector)
    827 {
    828 	struct ixgbe_hw *hw = &adapter->hw;
    829 	struct ix_queue *que = &adapter->queues[vector];
    830 	u32		queue = 1UL << vector;
    831 	u32		mask;
    832 
    833 	mutex_enter(&que->dc_mtx);
    834 	if (que->disabled_count > 0 && --que->disabled_count > 0)
    835 		goto out;
    836 
    837 	mask = (IXGBE_EIMS_RTX_QUEUE & queue);
    838 	IXGBE_WRITE_REG(hw, IXGBE_VTEIMS, mask);
    839 out:
    840 	mutex_exit(&que->dc_mtx);
    841 } /* ixv_enable_queue */
    842 
    843 /************************************************************************
    844  * ixv_disable_queue
    845  ************************************************************************/
    846 static inline void
    847 ixv_disable_queue(struct adapter *adapter, u32 vector)
    848 {
    849 	struct ixgbe_hw *hw = &adapter->hw;
    850 	struct ix_queue *que = &adapter->queues[vector];
    851 	u32		queue = 1UL << vector;
    852 	u32		mask;
    853 
    854 	mutex_enter(&que->dc_mtx);
    855 	if (que->disabled_count++ > 0)
    856 		goto  out;
    857 
    858 	mask = (IXGBE_EIMS_RTX_QUEUE & queue);
    859 	IXGBE_WRITE_REG(hw, IXGBE_VTEIMC, mask);
    860 out:
    861 	mutex_exit(&que->dc_mtx);
    862 } /* ixv_disable_queue */
    863 
    864 #if 0
    865 static inline void
    866 ixv_rearm_queues(struct adapter *adapter, u64 queues)
    867 {
    868 	u32 mask = (IXGBE_EIMS_RTX_QUEUE & queues);
    869 	IXGBE_WRITE_REG(&adapter->hw, IXGBE_VTEICS, mask);
    870 } /* ixv_rearm_queues */
    871 #endif
    872 
    873 
    874 /************************************************************************
    875  * ixv_msix_que - MSI-X Queue Interrupt Service routine
    876  ************************************************************************/
    877 static int
    878 ixv_msix_que(void *arg)
    879 {
    880 	struct ix_queue	*que = arg;
    881 	struct adapter	*adapter = que->adapter;
    882 	struct tx_ring	*txr = que->txr;
    883 	struct rx_ring	*rxr = que->rxr;
    884 	bool		more;
    885 	u32		newitr = 0;
    886 
    887 	ixv_disable_queue(adapter, que->msix);
    888 	++que->irqs.ev_count;
    889 
    890 #ifdef __NetBSD__
    891 	/* Don't run ixgbe_rxeof in interrupt context */
    892 	more = true;
    893 #else
    894 	more = ixgbe_rxeof(que);
    895 #endif
    896 
    897 	IXGBE_TX_LOCK(txr);
    898 	ixgbe_txeof(txr);
    899 	IXGBE_TX_UNLOCK(txr);
    900 
    901 	/* Do AIM now? */
    902 
    903 	if (adapter->enable_aim == false)
    904 		goto no_calc;
    905 	/*
    906 	 * Do Adaptive Interrupt Moderation:
    907 	 *  - Write out last calculated setting
    908 	 *  - Calculate based on average size over
    909 	 *    the last interval.
    910 	 */
    911 	if (que->eitr_setting)
    912 		ixv_eitr_write(adapter, que->msix, que->eitr_setting);
    913 
    914 	que->eitr_setting = 0;
    915 
    916 	/* Idle, do nothing */
    917 	if ((txr->bytes == 0) && (rxr->bytes == 0))
    918 		goto no_calc;
    919 
    920 	if ((txr->bytes) && (txr->packets))
    921 		newitr = txr->bytes/txr->packets;
    922 	if ((rxr->bytes) && (rxr->packets))
    923 		newitr = uimax(newitr, (rxr->bytes / rxr->packets));
    924 	newitr += 24; /* account for hardware frame, crc */
    925 
    926 	/* set an upper boundary */
    927 	newitr = uimin(newitr, 3000);
    928 
    929 	/* Be nice to the mid range */
    930 	if ((newitr > 300) && (newitr < 1200))
    931 		newitr = (newitr / 3);
    932 	else
    933 		newitr = (newitr / 2);
    934 
    935 	/*
    936 	 * When RSC is used, ITR interval must be larger than RSC_DELAY.
    937 	 * Currently, we use 2us for RSC_DELAY. The minimum value is always
    938 	 * greater than 2us on 100M (and 10M?(not documented)), but it's not
    939 	 * on 1G and higher.
    940 	 */
    941 	if ((adapter->link_speed != IXGBE_LINK_SPEED_100_FULL)
    942 	    && (adapter->link_speed != IXGBE_LINK_SPEED_10_FULL)) {
    943 		if (newitr < IXGBE_MIN_RSC_EITR_10G1G)
    944 			newitr = IXGBE_MIN_RSC_EITR_10G1G;
    945 	}
    946 
    947 	/* save for next interrupt */
    948 	que->eitr_setting = newitr;
    949 
    950 	/* Reset state */
    951 	txr->bytes = 0;
    952 	txr->packets = 0;
    953 	rxr->bytes = 0;
    954 	rxr->packets = 0;
    955 
    956 no_calc:
    957 	if (more)
    958 		softint_schedule(que->que_si);
    959 	else /* Re-enable this interrupt */
    960 		ixv_enable_queue(adapter, que->msix);
    961 
    962 	return 1;
    963 } /* ixv_msix_que */
    964 
    965 /************************************************************************
    966  * ixv_msix_mbx
    967  ************************************************************************/
    968 static int
    969 ixv_msix_mbx(void *arg)
    970 {
    971 	struct adapter	*adapter = arg;
    972 	struct ixgbe_hw *hw = &adapter->hw;
    973 
    974 	++adapter->link_irq.ev_count;
    975 	/* NetBSD: We use auto-clear, so it's not required to write VTEICR */
    976 
    977 	/* Link status change */
    978 	hw->mac.get_link_status = TRUE;
    979 	softint_schedule(adapter->link_si);
    980 
    981 	IXGBE_WRITE_REG(hw, IXGBE_VTEIMS, (1 << adapter->vector));
    982 
    983 	return 1;
    984 } /* ixv_msix_mbx */
    985 
    986 static void
    987 ixv_eitr_write(struct adapter *adapter, uint32_t index, uint32_t itr)
    988 {
    989 
    990 	/*
    991 	 * Newer devices than 82598 have VF function, so this function is
    992 	 * simple.
    993 	 */
    994 	itr |= IXGBE_EITR_CNT_WDIS;
    995 
    996 	IXGBE_WRITE_REG(&adapter->hw, IXGBE_VTEITR(index), itr);
    997 }
    998 
    999 
   1000 /************************************************************************
   1001  * ixv_media_status - Media Ioctl callback
   1002  *
   1003  *   Called whenever the user queries the status of
   1004  *   the interface using ifconfig.
   1005  ************************************************************************/
   1006 static void
   1007 ixv_media_status(struct ifnet *ifp, struct ifmediareq *ifmr)
   1008 {
   1009 	struct adapter *adapter = ifp->if_softc;
   1010 
   1011 	INIT_DEBUGOUT("ixv_media_status: begin");
   1012 	IXGBE_CORE_LOCK(adapter);
   1013 	ixv_update_link_status(adapter);
   1014 
   1015 	ifmr->ifm_status = IFM_AVALID;
   1016 	ifmr->ifm_active = IFM_ETHER;
   1017 
   1018 	if (adapter->link_active != LINK_STATE_UP) {
   1019 		ifmr->ifm_active |= IFM_NONE;
   1020 		IXGBE_CORE_UNLOCK(adapter);
   1021 		return;
   1022 	}
   1023 
   1024 	ifmr->ifm_status |= IFM_ACTIVE;
   1025 
   1026 	switch (adapter->link_speed) {
   1027 		case IXGBE_LINK_SPEED_10GB_FULL:
   1028 			ifmr->ifm_active |= IFM_10G_T | IFM_FDX;
   1029 			break;
   1030 		case IXGBE_LINK_SPEED_5GB_FULL:
   1031 			ifmr->ifm_active |= IFM_5000_T | IFM_FDX;
   1032 			break;
   1033 		case IXGBE_LINK_SPEED_2_5GB_FULL:
   1034 			ifmr->ifm_active |= IFM_2500_T | IFM_FDX;
   1035 			break;
   1036 		case IXGBE_LINK_SPEED_1GB_FULL:
   1037 			ifmr->ifm_active |= IFM_1000_T | IFM_FDX;
   1038 			break;
   1039 		case IXGBE_LINK_SPEED_100_FULL:
   1040 			ifmr->ifm_active |= IFM_100_TX | IFM_FDX;
   1041 			break;
   1042 		case IXGBE_LINK_SPEED_10_FULL:
   1043 			ifmr->ifm_active |= IFM_10_T | IFM_FDX;
   1044 			break;
   1045 	}
   1046 
   1047 	ifp->if_baudrate = ifmedia_baudrate(ifmr->ifm_active);
   1048 
   1049 	IXGBE_CORE_UNLOCK(adapter);
   1050 } /* ixv_media_status */
   1051 
   1052 /************************************************************************
   1053  * ixv_media_change - Media Ioctl callback
   1054  *
   1055  *   Called when the user changes speed/duplex using
   1056  *   media/mediopt option with ifconfig.
   1057  ************************************************************************/
   1058 static int
   1059 ixv_media_change(struct ifnet *ifp)
   1060 {
   1061 	struct adapter *adapter = ifp->if_softc;
   1062 	struct ifmedia *ifm = &adapter->media;
   1063 
   1064 	INIT_DEBUGOUT("ixv_media_change: begin");
   1065 
   1066 	if (IFM_TYPE(ifm->ifm_media) != IFM_ETHER)
   1067 		return (EINVAL);
   1068 
   1069 	switch (IFM_SUBTYPE(ifm->ifm_media)) {
   1070 	case IFM_AUTO:
   1071 		break;
   1072 	default:
   1073 		device_printf(adapter->dev, "Only auto media type\n");
   1074 		return (EINVAL);
   1075 	}
   1076 
   1077 	return (0);
   1078 } /* ixv_media_change */
   1079 
   1080 /************************************************************************
   1081  * ixv_negotiate_api
   1082  *
   1083  *   Negotiate the Mailbox API with the PF;
   1084  *   start with the most featured API first.
   1085  ************************************************************************/
   1086 static int
   1087 ixv_negotiate_api(struct adapter *adapter)
   1088 {
   1089 	struct ixgbe_hw *hw = &adapter->hw;
   1090 	int		mbx_api[] = { ixgbe_mbox_api_13,
   1091 				      ixgbe_mbox_api_12,
   1092 				      ixgbe_mbox_api_11,
   1093 				      ixgbe_mbox_api_10,
   1094 				      ixgbe_mbox_api_unknown };
   1095 	int		i = 0;
   1096 
   1097 	while (mbx_api[i] != ixgbe_mbox_api_unknown) {
   1098 		if (ixgbevf_negotiate_api_version(hw, mbx_api[i]) == 0)
   1099 			return (0);
   1100 		i++;
   1101 	}
   1102 
   1103 	return (EINVAL);
   1104 } /* ixv_negotiate_api */
   1105 
   1106 
   1107 /************************************************************************
   1108  * ixv_set_multi - Multicast Update
   1109  *
   1110  *   Called whenever multicast address list is updated.
   1111  ************************************************************************/
   1112 static int
   1113 ixv_set_rxfilter(struct adapter *adapter)
   1114 {
   1115 	u8	mta[IXGBE_MAX_VF_MC * IXGBE_ETH_LENGTH_OF_ADDRESS];
   1116 	struct ifnet		*ifp = adapter->ifp;
   1117 	struct ixgbe_hw		*hw = &adapter->hw;
   1118 	u8			*update_ptr;
   1119 	int			mcnt = 0;
   1120 	struct ethercom		*ec = &adapter->osdep.ec;
   1121 	struct ether_multi	*enm;
   1122 	struct ether_multistep	step;
   1123 	bool			overflow = false;
   1124 	int			error, rc = 0;
   1125 
   1126 	KASSERT(mutex_owned(&adapter->core_mtx));
   1127 	IOCTL_DEBUGOUT("ixv_set_rxfilter: begin");
   1128 
   1129 	/* 1: For PROMISC */
   1130 	if (ifp->if_flags & IFF_PROMISC) {
   1131 		error = hw->mac.ops.update_xcast_mode(hw,
   1132 		    IXGBEVF_XCAST_MODE_PROMISC);
   1133 		if (error == IXGBE_ERR_NOT_TRUSTED) {
   1134 			device_printf(adapter->dev,
   1135 			    "this interface is not trusted\n");
   1136 			error = EPERM;
   1137 		} else if (error == IXGBE_ERR_FEATURE_NOT_SUPPORTED) {
   1138 			device_printf(adapter->dev,
   1139 			    "the PF doesn't support promisc mode\n");
   1140 			error = EOPNOTSUPP;
   1141 		} else if (error == IXGBE_ERR_NOT_IN_PROMISC) {
   1142 			device_printf(adapter->dev,
   1143 			    "the PF may not in promisc mode\n");
   1144 			error = EINVAL;
   1145 		} else if (error) {
   1146 			device_printf(adapter->dev,
   1147 			    "failed to set promisc mode. error = %d\n",
   1148 			    error);
   1149 			error = EIO;
   1150 		} else
   1151 			return 0;
   1152 		rc = error;
   1153 	}
   1154 
   1155 	/* 2: For ALLMULTI or normal */
   1156 	ETHER_LOCK(ec);
   1157 	ETHER_FIRST_MULTI(step, ec, enm);
   1158 	while (enm != NULL) {
   1159 		if ((mcnt >= IXGBE_MAX_VF_MC) ||
   1160 		    (memcmp(enm->enm_addrlo, enm->enm_addrhi,
   1161 			ETHER_ADDR_LEN) != 0)) {
   1162 			overflow = true;
   1163 			break;
   1164 		}
   1165 		bcopy(enm->enm_addrlo,
   1166 		    &mta[mcnt * IXGBE_ETH_LENGTH_OF_ADDRESS],
   1167 		    IXGBE_ETH_LENGTH_OF_ADDRESS);
   1168 		mcnt++;
   1169 		ETHER_NEXT_MULTI(step, enm);
   1170 	}
   1171 	ETHER_UNLOCK(ec);
   1172 
   1173 	/* 3: For ALLMULTI */
   1174 	if (overflow) {
   1175 		error = hw->mac.ops.update_xcast_mode(hw,
   1176 		    IXGBEVF_XCAST_MODE_ALLMULTI);
   1177 		if (error == IXGBE_ERR_NOT_TRUSTED) {
   1178 			device_printf(adapter->dev,
   1179 			    "this interface is not trusted\n");
   1180 			error = EPERM;
   1181 		} else if (error == IXGBE_ERR_FEATURE_NOT_SUPPORTED) {
   1182 			device_printf(adapter->dev,
   1183 			    "the PF doesn't support allmulti mode\n");
   1184 			error = EOPNOTSUPP;
   1185 		} else if (error) {
   1186 			device_printf(adapter->dev,
   1187 			    "number of Ethernet multicast addresses "
   1188 			    "exceeds the limit (%d). error = %d\n",
   1189 			    IXGBE_MAX_VF_MC, error);
   1190 			error = ENOSPC;
   1191 		} else {
   1192 			ETHER_LOCK(ec);
   1193 			ec->ec_flags |= ETHER_F_ALLMULTI;
   1194 			ETHER_UNLOCK(ec);
   1195 			return rc; /* Promisc might failed */
   1196 		}
   1197 
   1198 		if (rc == 0)
   1199 			rc = error;
   1200 
   1201 		/* Continue to update the multicast table as many as we can */
   1202 	}
   1203 
   1204 	/* 4: For normal operation */
   1205 	error = hw->mac.ops.update_xcast_mode(hw, IXGBEVF_XCAST_MODE_MULTI);
   1206 	if ((error == IXGBE_ERR_FEATURE_NOT_SUPPORTED) || (error == 0)) {
   1207 		/* Normal operation */
   1208 		ETHER_LOCK(ec);
   1209 		ec->ec_flags &= ~ETHER_F_ALLMULTI;
   1210 		ETHER_UNLOCK(ec);
   1211 		error = 0;
   1212 	} else if (error) {
   1213 		device_printf(adapter->dev,
   1214 		    "failed to set Ethernet multicast address "
   1215 		    "operation to normal. error = %d\n", error);
   1216 	}
   1217 
   1218 	update_ptr = mta;
   1219 
   1220 	error = adapter->hw.mac.ops.update_mc_addr_list(&adapter->hw,
   1221 	    update_ptr, mcnt, ixv_mc_array_itr, TRUE);
   1222 	if (rc == 0)
   1223 		rc = error;
   1224 
   1225 	return rc;
   1226 } /* ixv_set_rxfilter */
   1227 
   1228 /************************************************************************
   1229  * ixv_mc_array_itr
   1230  *
   1231  *   An iterator function needed by the multicast shared code.
   1232  *   It feeds the shared code routine the addresses in the
   1233  *   array of ixv_set_rxfilter() one by one.
   1234  ************************************************************************/
   1235 static u8 *
   1236 ixv_mc_array_itr(struct ixgbe_hw *hw, u8 **update_ptr, u32 *vmdq)
   1237 {
   1238 	u8 *addr = *update_ptr;
   1239 	u8 *newptr;
   1240 
   1241 	*vmdq = 0;
   1242 
   1243 	newptr = addr + IXGBE_ETH_LENGTH_OF_ADDRESS;
   1244 	*update_ptr = newptr;
   1245 
   1246 	return addr;
   1247 } /* ixv_mc_array_itr */
   1248 
   1249 /************************************************************************
   1250  * ixv_local_timer - Timer routine
   1251  *
   1252  *   Checks for link status, updates statistics,
   1253  *   and runs the watchdog check.
   1254  ************************************************************************/
   1255 static void
   1256 ixv_local_timer(void *arg)
   1257 {
   1258 	struct adapter *adapter = arg;
   1259 
   1260 	IXGBE_CORE_LOCK(adapter);
   1261 	ixv_local_timer_locked(adapter);
   1262 	IXGBE_CORE_UNLOCK(adapter);
   1263 }
   1264 
   1265 static void
   1266 ixv_local_timer_locked(void *arg)
   1267 {
   1268 	struct adapter	*adapter = arg;
   1269 	device_t	dev = adapter->dev;
   1270 	struct ix_queue	*que = adapter->queues;
   1271 	u64		queues = 0;
   1272 	u64		v0, v1, v2, v3, v4, v5, v6, v7;
   1273 	int		hung = 0;
   1274 	int		i;
   1275 
   1276 	KASSERT(mutex_owned(&adapter->core_mtx));
   1277 
   1278 	if (ixv_check_link(adapter)) {
   1279 		ixv_init_locked(adapter);
   1280 		return;
   1281 	}
   1282 
   1283 	/* Stats Update */
   1284 	ixv_update_stats(adapter);
   1285 
   1286 	/* Update some event counters */
   1287 	v0 = v1 = v2 = v3 = v4 = v5 = v6 = v7 = 0;
   1288 	que = adapter->queues;
   1289 	for (i = 0; i < adapter->num_queues; i++, que++) {
   1290 		struct tx_ring	*txr = que->txr;
   1291 
   1292 		v0 += txr->q_efbig_tx_dma_setup;
   1293 		v1 += txr->q_mbuf_defrag_failed;
   1294 		v2 += txr->q_efbig2_tx_dma_setup;
   1295 		v3 += txr->q_einval_tx_dma_setup;
   1296 		v4 += txr->q_other_tx_dma_setup;
   1297 		v5 += txr->q_eagain_tx_dma_setup;
   1298 		v6 += txr->q_enomem_tx_dma_setup;
   1299 		v7 += txr->q_tso_err;
   1300 	}
   1301 	adapter->efbig_tx_dma_setup.ev_count = v0;
   1302 	adapter->mbuf_defrag_failed.ev_count = v1;
   1303 	adapter->efbig2_tx_dma_setup.ev_count = v2;
   1304 	adapter->einval_tx_dma_setup.ev_count = v3;
   1305 	adapter->other_tx_dma_setup.ev_count = v4;
   1306 	adapter->eagain_tx_dma_setup.ev_count = v5;
   1307 	adapter->enomem_tx_dma_setup.ev_count = v6;
   1308 	adapter->tso_err.ev_count = v7;
   1309 
   1310 	/*
   1311 	 * Check the TX queues status
   1312 	 *	- mark hung queues so we don't schedule on them
   1313 	 *	- watchdog only if all queues show hung
   1314 	 */
   1315 	que = adapter->queues;
   1316 	for (i = 0; i < adapter->num_queues; i++, que++) {
   1317 		/* Keep track of queues with work for soft irq */
   1318 		if (que->txr->busy)
   1319 			queues |= ((u64)1 << que->me);
   1320 		/*
   1321 		 * Each time txeof runs without cleaning, but there
   1322 		 * are uncleaned descriptors it increments busy. If
   1323 		 * we get to the MAX we declare it hung.
   1324 		 */
   1325 		if (que->busy == IXGBE_QUEUE_HUNG) {
   1326 			++hung;
   1327 			/* Mark the queue as inactive */
   1328 			adapter->active_queues &= ~((u64)1 << que->me);
   1329 			continue;
   1330 		} else {
   1331 			/* Check if we've come back from hung */
   1332 			if ((adapter->active_queues & ((u64)1 << que->me)) == 0)
   1333 				adapter->active_queues |= ((u64)1 << que->me);
   1334 		}
   1335 		if (que->busy >= IXGBE_MAX_TX_BUSY) {
   1336 			device_printf(dev,
   1337 			    "Warning queue %d appears to be hung!\n", i);
   1338 			que->txr->busy = IXGBE_QUEUE_HUNG;
   1339 			++hung;
   1340 		}
   1341 	}
   1342 
   1343 	/* Only truly watchdog if all queues show hung */
   1344 	if (hung == adapter->num_queues)
   1345 		goto watchdog;
   1346 #if 0
   1347 	else if (queues != 0) { /* Force an IRQ on queues with work */
   1348 		ixv_rearm_queues(adapter, queues);
   1349 	}
   1350 #endif
   1351 
   1352 	callout_reset(&adapter->timer, hz, ixv_local_timer, adapter);
   1353 
   1354 	return;
   1355 
   1356 watchdog:
   1357 
   1358 	device_printf(adapter->dev, "Watchdog timeout -- resetting\n");
   1359 	adapter->ifp->if_flags &= ~IFF_RUNNING;
   1360 	adapter->watchdog_events.ev_count++;
   1361 	ixv_init_locked(adapter);
   1362 } /* ixv_local_timer */
   1363 
   1364 /************************************************************************
   1365  * ixv_update_link_status - Update OS on link state
   1366  *
   1367  * Note: Only updates the OS on the cached link state.
   1368  *	 The real check of the hardware only happens with
   1369  *	 a link interrupt.
   1370  ************************************************************************/
   1371 static void
   1372 ixv_update_link_status(struct adapter *adapter)
   1373 {
   1374 	struct ifnet *ifp = adapter->ifp;
   1375 	device_t     dev = adapter->dev;
   1376 
   1377 	KASSERT(mutex_owned(&adapter->core_mtx));
   1378 
   1379 	if (adapter->link_up) {
   1380 		if (adapter->link_active != LINK_STATE_UP) {
   1381 			if (bootverbose) {
   1382 				const char *bpsmsg;
   1383 
   1384 				switch (adapter->link_speed) {
   1385 				case IXGBE_LINK_SPEED_10GB_FULL:
   1386 					bpsmsg = "10 Gbps";
   1387 					break;
   1388 				case IXGBE_LINK_SPEED_5GB_FULL:
   1389 					bpsmsg = "5 Gbps";
   1390 					break;
   1391 				case IXGBE_LINK_SPEED_2_5GB_FULL:
   1392 					bpsmsg = "2.5 Gbps";
   1393 					break;
   1394 				case IXGBE_LINK_SPEED_1GB_FULL:
   1395 					bpsmsg = "1 Gbps";
   1396 					break;
   1397 				case IXGBE_LINK_SPEED_100_FULL:
   1398 					bpsmsg = "100 Mbps";
   1399 					break;
   1400 				case IXGBE_LINK_SPEED_10_FULL:
   1401 					bpsmsg = "10 Mbps";
   1402 					break;
   1403 				default:
   1404 					bpsmsg = "unknown speed";
   1405 					break;
   1406 				}
   1407 				device_printf(dev, "Link is up %s %s \n",
   1408 				    bpsmsg, "Full Duplex");
   1409 			}
   1410 			adapter->link_active = LINK_STATE_UP;
   1411 			if_link_state_change(ifp, LINK_STATE_UP);
   1412 		}
   1413 	} else {
   1414 		/*
   1415 		 * Do it when link active changes to DOWN. i.e.
   1416 		 * a) LINK_STATE_UNKNOWN -> LINK_STATE_DOWN
   1417 		 * b) LINK_STATE_UP	 -> LINK_STATE_DOWN
   1418 		 */
   1419 		if (adapter->link_active != LINK_STATE_DOWN) {
   1420 			if (bootverbose)
   1421 				device_printf(dev, "Link is Down\n");
   1422 			if_link_state_change(ifp, LINK_STATE_DOWN);
   1423 			adapter->link_active = LINK_STATE_DOWN;
   1424 		}
   1425 	}
   1426 } /* ixv_update_link_status */
   1427 
   1428 
   1429 /************************************************************************
   1430  * ixv_stop - Stop the hardware
   1431  *
   1432  *   Disables all traffic on the adapter by issuing a
   1433  *   global reset on the MAC and deallocates TX/RX buffers.
   1434  ************************************************************************/
   1435 static void
   1436 ixv_ifstop(struct ifnet *ifp, int disable)
   1437 {
   1438 	struct adapter *adapter = ifp->if_softc;
   1439 
   1440 	IXGBE_CORE_LOCK(adapter);
   1441 	ixv_stop(adapter);
   1442 	IXGBE_CORE_UNLOCK(adapter);
   1443 }
   1444 
   1445 static void
   1446 ixv_stop(void *arg)
   1447 {
   1448 	struct ifnet	*ifp;
   1449 	struct adapter	*adapter = arg;
   1450 	struct ixgbe_hw *hw = &adapter->hw;
   1451 
   1452 	ifp = adapter->ifp;
   1453 
   1454 	KASSERT(mutex_owned(&adapter->core_mtx));
   1455 
   1456 	INIT_DEBUGOUT("ixv_stop: begin\n");
   1457 	ixv_disable_intr(adapter);
   1458 
   1459 	/* Tell the stack that the interface is no longer active */
   1460 	ifp->if_flags &= ~(IFF_RUNNING | IFF_OACTIVE);
   1461 
   1462 	hw->mac.ops.reset_hw(hw);
   1463 	adapter->hw.adapter_stopped = FALSE;
   1464 	hw->mac.ops.stop_adapter(hw);
   1465 	callout_stop(&adapter->timer);
   1466 
   1467 	/* reprogram the RAR[0] in case user changed it. */
   1468 	hw->mac.ops.set_rar(hw, 0, hw->mac.addr, 0, IXGBE_RAH_AV);
   1469 
   1470 	return;
   1471 } /* ixv_stop */
   1472 
   1473 
   1474 /************************************************************************
   1475  * ixv_allocate_pci_resources
   1476  ************************************************************************/
   1477 static int
   1478 ixv_allocate_pci_resources(struct adapter *adapter,
   1479     const struct pci_attach_args *pa)
   1480 {
   1481 	pcireg_t	memtype, csr;
   1482 	device_t	dev = adapter->dev;
   1483 	bus_addr_t addr;
   1484 	int flags;
   1485 
   1486 	memtype = pci_mapreg_type(pa->pa_pc, pa->pa_tag, PCI_BAR(0));
   1487 	switch (memtype) {
   1488 	case PCI_MAPREG_TYPE_MEM | PCI_MAPREG_MEM_TYPE_32BIT:
   1489 	case PCI_MAPREG_TYPE_MEM | PCI_MAPREG_MEM_TYPE_64BIT:
   1490 		adapter->osdep.mem_bus_space_tag = pa->pa_memt;
   1491 		if (pci_mapreg_info(pa->pa_pc, pa->pa_tag, PCI_BAR(0),
   1492 		      memtype, &addr, &adapter->osdep.mem_size, &flags) != 0)
   1493 			goto map_err;
   1494 		if ((flags & BUS_SPACE_MAP_PREFETCHABLE) != 0) {
   1495 			aprint_normal_dev(dev, "clearing prefetchable bit\n");
   1496 			flags &= ~BUS_SPACE_MAP_PREFETCHABLE;
   1497 		}
   1498 		if (bus_space_map(adapter->osdep.mem_bus_space_tag, addr,
   1499 		     adapter->osdep.mem_size, flags,
   1500 		     &adapter->osdep.mem_bus_space_handle) != 0) {
   1501 map_err:
   1502 			adapter->osdep.mem_size = 0;
   1503 			aprint_error_dev(dev, "unable to map BAR0\n");
   1504 			return ENXIO;
   1505 		}
   1506 		/*
   1507 		 * Enable address decoding for memory range in case it's not
   1508 		 * set.
   1509 		 */
   1510 		csr = pci_conf_read(pa->pa_pc, pa->pa_tag,
   1511 		    PCI_COMMAND_STATUS_REG);
   1512 		csr |= PCI_COMMAND_MEM_ENABLE;
   1513 		pci_conf_write(pa->pa_pc, pa->pa_tag, PCI_COMMAND_STATUS_REG,
   1514 		    csr);
   1515 		break;
   1516 	default:
   1517 		aprint_error_dev(dev, "unexpected type on BAR0\n");
   1518 		return ENXIO;
   1519 	}
   1520 
   1521 	/* Pick up the tuneable queues */
   1522 	adapter->num_queues = ixv_num_queues;
   1523 
   1524 	return (0);
   1525 } /* ixv_allocate_pci_resources */
   1526 
   1527 /************************************************************************
   1528  * ixv_free_pci_resources
   1529  ************************************************************************/
   1530 static void
   1531 ixv_free_pci_resources(struct adapter * adapter)
   1532 {
   1533 	struct		ix_queue *que = adapter->queues;
   1534 	int		rid;
   1535 
   1536 	/*
   1537 	 *  Release all msix queue resources:
   1538 	 */
   1539 	for (int i = 0; i < adapter->num_queues; i++, que++) {
   1540 		if (que->res != NULL)
   1541 			pci_intr_disestablish(adapter->osdep.pc,
   1542 			    adapter->osdep.ihs[i]);
   1543 	}
   1544 
   1545 
   1546 	/* Clean the Mailbox interrupt last */
   1547 	rid = adapter->vector;
   1548 
   1549 	if (adapter->osdep.ihs[rid] != NULL) {
   1550 		pci_intr_disestablish(adapter->osdep.pc,
   1551 		    adapter->osdep.ihs[rid]);
   1552 		adapter->osdep.ihs[rid] = NULL;
   1553 	}
   1554 
   1555 	pci_intr_release(adapter->osdep.pc, adapter->osdep.intrs,
   1556 	    adapter->osdep.nintrs);
   1557 
   1558 	if (adapter->osdep.mem_size != 0) {
   1559 		bus_space_unmap(adapter->osdep.mem_bus_space_tag,
   1560 		    adapter->osdep.mem_bus_space_handle,
   1561 		    adapter->osdep.mem_size);
   1562 	}
   1563 
   1564 	return;
   1565 } /* ixv_free_pci_resources */
   1566 
   1567 /************************************************************************
   1568  * ixv_setup_interface
   1569  *
   1570  *   Setup networking device structure and register an interface.
   1571  ************************************************************************/
   1572 static int
   1573 ixv_setup_interface(device_t dev, struct adapter *adapter)
   1574 {
   1575 	struct ethercom *ec = &adapter->osdep.ec;
   1576 	struct ifnet   *ifp;
   1577 	int rv;
   1578 
   1579 	INIT_DEBUGOUT("ixv_setup_interface: begin");
   1580 
   1581 	ifp = adapter->ifp = &ec->ec_if;
   1582 	strlcpy(ifp->if_xname, device_xname(dev), IFNAMSIZ);
   1583 	ifp->if_baudrate = IF_Gbps(10);
   1584 	ifp->if_init = ixv_init;
   1585 	ifp->if_stop = ixv_ifstop;
   1586 	ifp->if_softc = adapter;
   1587 	ifp->if_flags = IFF_BROADCAST | IFF_SIMPLEX | IFF_MULTICAST;
   1588 #ifdef IXGBE_MPSAFE
   1589 	ifp->if_extflags = IFEF_MPSAFE;
   1590 #endif
   1591 	ifp->if_ioctl = ixv_ioctl;
   1592 	if (adapter->feat_en & IXGBE_FEATURE_LEGACY_TX) {
   1593 #if 0
   1594 		ixv_start_locked = ixgbe_legacy_start_locked;
   1595 #endif
   1596 	} else {
   1597 		ifp->if_transmit = ixgbe_mq_start;
   1598 #if 0
   1599 		ixv_start_locked = ixgbe_mq_start_locked;
   1600 #endif
   1601 	}
   1602 	ifp->if_start = ixgbe_legacy_start;
   1603 	IFQ_SET_MAXLEN(&ifp->if_snd, adapter->num_tx_desc - 2);
   1604 	IFQ_SET_READY(&ifp->if_snd);
   1605 
   1606 	rv = if_initialize(ifp);
   1607 	if (rv != 0) {
   1608 		aprint_error_dev(dev, "if_initialize failed(%d)\n", rv);
   1609 		return rv;
   1610 	}
   1611 	adapter->ipq = if_percpuq_create(&adapter->osdep.ec.ec_if);
   1612 	ether_ifattach(ifp, adapter->hw.mac.addr);
   1613 	aprint_normal_dev(dev, "Ethernet address %s\n",
   1614 	    ether_sprintf(adapter->hw.mac.addr));
   1615 	/*
   1616 	 * We use per TX queue softint, so if_deferred_start_init() isn't
   1617 	 * used.
   1618 	 */
   1619 	ether_set_ifflags_cb(ec, ixv_ifflags_cb);
   1620 
   1621 	adapter->max_frame_size = ifp->if_mtu + IXGBE_MTU_HDR;
   1622 
   1623 	/*
   1624 	 * Tell the upper layer(s) we support long frames.
   1625 	 */
   1626 	ifp->if_hdrlen = sizeof(struct ether_vlan_header);
   1627 
   1628 	/* Set capability flags */
   1629 	ifp->if_capabilities |= IFCAP_HWCSUM
   1630 			     |	IFCAP_TSOv4
   1631 			     |	IFCAP_TSOv6;
   1632 	ifp->if_capenable = 0;
   1633 
   1634 	ec->ec_capabilities |= ETHERCAP_VLAN_HWFILTER
   1635 			    |  ETHERCAP_VLAN_HWTAGGING
   1636 			    |  ETHERCAP_VLAN_HWCSUM
   1637 			    |  ETHERCAP_JUMBO_MTU
   1638 			    |  ETHERCAP_VLAN_MTU;
   1639 
   1640 	/* Enable the above capabilities by default */
   1641 	ec->ec_capenable = ec->ec_capabilities;
   1642 
   1643 	/* Don't enable LRO by default */
   1644 #if 0
   1645 	/* NetBSD doesn't support LRO yet */
   1646 	ifp->if_capabilities |= IFCAP_LRO;
   1647 #endif
   1648 
   1649 	/*
   1650 	 * Specify the media types supported by this adapter and register
   1651 	 * callbacks to update media and link information
   1652 	 */
   1653 	ec->ec_ifmedia = &adapter->media;
   1654 	ifmedia_init(&adapter->media, IFM_IMASK, ixv_media_change,
   1655 	    ixv_media_status);
   1656 	ifmedia_add(&adapter->media, IFM_ETHER | IFM_AUTO, 0, NULL);
   1657 	ifmedia_set(&adapter->media, IFM_ETHER | IFM_AUTO);
   1658 
   1659 	if_register(ifp);
   1660 
   1661 	return 0;
   1662 } /* ixv_setup_interface */
   1663 
   1664 
   1665 /************************************************************************
   1666  * ixv_initialize_transmit_units - Enable transmit unit.
   1667  ************************************************************************/
   1668 static void
   1669 ixv_initialize_transmit_units(struct adapter *adapter)
   1670 {
   1671 	struct tx_ring	*txr = adapter->tx_rings;
   1672 	struct ixgbe_hw	*hw = &adapter->hw;
   1673 	int i;
   1674 
   1675 	for (i = 0; i < adapter->num_queues; i++, txr++) {
   1676 		u64 tdba = txr->txdma.dma_paddr;
   1677 		u32 txctrl, txdctl;
   1678 		int j = txr->me;
   1679 
   1680 		/* Set WTHRESH to 8, burst writeback */
   1681 		txdctl = IXGBE_READ_REG(hw, IXGBE_VFTXDCTL(j));
   1682 		txdctl |= (8 << 16);
   1683 		IXGBE_WRITE_REG(hw, IXGBE_VFTXDCTL(j), txdctl);
   1684 
   1685 		/* Set the HW Tx Head and Tail indices */
   1686 		IXGBE_WRITE_REG(&adapter->hw, IXGBE_VFTDH(j), 0);
   1687 		IXGBE_WRITE_REG(&adapter->hw, IXGBE_VFTDT(j), 0);
   1688 
   1689 		/* Set Tx Tail register */
   1690 		txr->tail = IXGBE_VFTDT(j);
   1691 
   1692 		txr->txr_no_space = false;
   1693 
   1694 		/* Set Ring parameters */
   1695 		IXGBE_WRITE_REG(hw, IXGBE_VFTDBAL(j),
   1696 		    (tdba & 0x00000000ffffffffULL));
   1697 		IXGBE_WRITE_REG(hw, IXGBE_VFTDBAH(j), (tdba >> 32));
   1698 		IXGBE_WRITE_REG(hw, IXGBE_VFTDLEN(j),
   1699 		    adapter->num_tx_desc * sizeof(struct ixgbe_legacy_tx_desc));
   1700 		txctrl = IXGBE_READ_REG(hw, IXGBE_VFDCA_TXCTRL(j));
   1701 		txctrl &= ~IXGBE_DCA_TXCTRL_DESC_WRO_EN;
   1702 		IXGBE_WRITE_REG(hw, IXGBE_VFDCA_TXCTRL(j), txctrl);
   1703 
   1704 		/* Now enable */
   1705 		txdctl = IXGBE_READ_REG(hw, IXGBE_VFTXDCTL(j));
   1706 		txdctl |= IXGBE_TXDCTL_ENABLE;
   1707 		IXGBE_WRITE_REG(hw, IXGBE_VFTXDCTL(j), txdctl);
   1708 	}
   1709 
   1710 	return;
   1711 } /* ixv_initialize_transmit_units */
   1712 
   1713 
   1714 /************************************************************************
   1715  * ixv_initialize_rss_mapping
   1716  ************************************************************************/
   1717 static void
   1718 ixv_initialize_rss_mapping(struct adapter *adapter)
   1719 {
   1720 	struct ixgbe_hw *hw = &adapter->hw;
   1721 	u32		reta = 0, mrqc, rss_key[10];
   1722 	int		queue_id;
   1723 	int		i, j;
   1724 	u32		rss_hash_config;
   1725 
   1726 	/* force use default RSS key. */
   1727 #ifdef __NetBSD__
   1728 	rss_getkey((uint8_t *) &rss_key);
   1729 #else
   1730 	if (adapter->feat_en & IXGBE_FEATURE_RSS) {
   1731 		/* Fetch the configured RSS key */
   1732 		rss_getkey((uint8_t *)&rss_key);
   1733 	} else {
   1734 		/* set up random bits */
   1735 		cprng_fast(&rss_key, sizeof(rss_key));
   1736 	}
   1737 #endif
   1738 
   1739 	/* Now fill out hash function seeds */
   1740 	for (i = 0; i < 10; i++)
   1741 		IXGBE_WRITE_REG(hw, IXGBE_VFRSSRK(i), rss_key[i]);
   1742 
   1743 	/* Set up the redirection table */
   1744 	for (i = 0, j = 0; i < 64; i++, j++) {
   1745 		if (j == adapter->num_queues)
   1746 			j = 0;
   1747 
   1748 		if (adapter->feat_en & IXGBE_FEATURE_RSS) {
   1749 			/*
   1750 			 * Fetch the RSS bucket id for the given indirection
   1751 			 * entry. Cap it at the number of configured buckets
   1752 			 * (which is num_queues.)
   1753 			 */
   1754 			queue_id = rss_get_indirection_to_bucket(i);
   1755 			queue_id = queue_id % adapter->num_queues;
   1756 		} else
   1757 			queue_id = j;
   1758 
   1759 		/*
   1760 		 * The low 8 bits are for hash value (n+0);
   1761 		 * The next 8 bits are for hash value (n+1), etc.
   1762 		 */
   1763 		reta >>= 8;
   1764 		reta |= ((uint32_t)queue_id) << 24;
   1765 		if ((i & 3) == 3) {
   1766 			IXGBE_WRITE_REG(hw, IXGBE_VFRETA(i >> 2), reta);
   1767 			reta = 0;
   1768 		}
   1769 	}
   1770 
   1771 	/* Perform hash on these packet types */
   1772 	if (adapter->feat_en & IXGBE_FEATURE_RSS)
   1773 		rss_hash_config = rss_gethashconfig();
   1774 	else {
   1775 		/*
   1776 		 * Disable UDP - IP fragments aren't currently being handled
   1777 		 * and so we end up with a mix of 2-tuple and 4-tuple
   1778 		 * traffic.
   1779 		 */
   1780 		rss_hash_config = RSS_HASHTYPE_RSS_IPV4
   1781 				| RSS_HASHTYPE_RSS_TCP_IPV4
   1782 				| RSS_HASHTYPE_RSS_IPV6
   1783 				| RSS_HASHTYPE_RSS_TCP_IPV6;
   1784 	}
   1785 
   1786 	mrqc = IXGBE_MRQC_RSSEN;
   1787 	if (rss_hash_config & RSS_HASHTYPE_RSS_IPV4)
   1788 		mrqc |= IXGBE_MRQC_RSS_FIELD_IPV4;
   1789 	if (rss_hash_config & RSS_HASHTYPE_RSS_TCP_IPV4)
   1790 		mrqc |= IXGBE_MRQC_RSS_FIELD_IPV4_TCP;
   1791 	if (rss_hash_config & RSS_HASHTYPE_RSS_IPV6)
   1792 		mrqc |= IXGBE_MRQC_RSS_FIELD_IPV6;
   1793 	if (rss_hash_config & RSS_HASHTYPE_RSS_TCP_IPV6)
   1794 		mrqc |= IXGBE_MRQC_RSS_FIELD_IPV6_TCP;
   1795 	if (rss_hash_config & RSS_HASHTYPE_RSS_IPV6_EX)
   1796 		device_printf(adapter->dev, "%s: RSS_HASHTYPE_RSS_IPV6_EX defined, but not supported\n",
   1797 		    __func__);
   1798 	if (rss_hash_config & RSS_HASHTYPE_RSS_TCP_IPV6_EX)
   1799 		device_printf(adapter->dev, "%s: RSS_HASHTYPE_RSS_TCP_IPV6_EX defined, but not supported\n",
   1800 		    __func__);
   1801 	if (rss_hash_config & RSS_HASHTYPE_RSS_UDP_IPV4)
   1802 		mrqc |= IXGBE_MRQC_RSS_FIELD_IPV4_UDP;
   1803 	if (rss_hash_config & RSS_HASHTYPE_RSS_UDP_IPV6)
   1804 		mrqc |= IXGBE_MRQC_RSS_FIELD_IPV6_UDP;
   1805 	if (rss_hash_config & RSS_HASHTYPE_RSS_UDP_IPV6_EX)
   1806 		device_printf(adapter->dev, "%s: RSS_HASHTYPE_RSS_UDP_IPV6_EX defined, but not supported\n",
   1807 		    __func__);
   1808 	IXGBE_WRITE_REG(hw, IXGBE_VFMRQC, mrqc);
   1809 } /* ixv_initialize_rss_mapping */
   1810 
   1811 
   1812 /************************************************************************
   1813  * ixv_initialize_receive_units - Setup receive registers and features.
   1814  ************************************************************************/
   1815 static void
   1816 ixv_initialize_receive_units(struct adapter *adapter)
   1817 {
   1818 	struct	rx_ring	*rxr = adapter->rx_rings;
   1819 	struct ixgbe_hw	*hw = &adapter->hw;
   1820 	struct ifnet	*ifp = adapter->ifp;
   1821 	u32		bufsz, psrtype;
   1822 
   1823 	if (ifp->if_mtu > ETHERMTU)
   1824 		bufsz = 4096 >> IXGBE_SRRCTL_BSIZEPKT_SHIFT;
   1825 	else
   1826 		bufsz = 2048 >> IXGBE_SRRCTL_BSIZEPKT_SHIFT;
   1827 
   1828 	psrtype = IXGBE_PSRTYPE_TCPHDR
   1829 		| IXGBE_PSRTYPE_UDPHDR
   1830 		| IXGBE_PSRTYPE_IPV4HDR
   1831 		| IXGBE_PSRTYPE_IPV6HDR
   1832 		| IXGBE_PSRTYPE_L2HDR;
   1833 
   1834 	if (adapter->num_queues > 1)
   1835 		psrtype |= 1 << 29;
   1836 
   1837 	IXGBE_WRITE_REG(hw, IXGBE_VFPSRTYPE, psrtype);
   1838 
   1839 	/* Tell PF our max_frame size */
   1840 	if (ixgbevf_rlpml_set_vf(hw, adapter->max_frame_size) != 0) {
   1841 		device_printf(adapter->dev, "There is a problem with the PF setup.  It is likely the receive unit for this VF will not function correctly.\n");
   1842 	}
   1843 
   1844 	for (int i = 0; i < adapter->num_queues; i++, rxr++) {
   1845 		u64 rdba = rxr->rxdma.dma_paddr;
   1846 		u32 reg, rxdctl;
   1847 		int j = rxr->me;
   1848 
   1849 		/* Disable the queue */
   1850 		rxdctl = IXGBE_READ_REG(hw, IXGBE_VFRXDCTL(j));
   1851 		rxdctl &= ~IXGBE_RXDCTL_ENABLE;
   1852 		IXGBE_WRITE_REG(hw, IXGBE_VFRXDCTL(j), rxdctl);
   1853 		for (int k = 0; k < 10; k++) {
   1854 			if (IXGBE_READ_REG(hw, IXGBE_VFRXDCTL(j)) &
   1855 			    IXGBE_RXDCTL_ENABLE)
   1856 				msec_delay(1);
   1857 			else
   1858 				break;
   1859 		}
   1860 		wmb();
   1861 		/* Setup the Base and Length of the Rx Descriptor Ring */
   1862 		IXGBE_WRITE_REG(hw, IXGBE_VFRDBAL(j),
   1863 		    (rdba & 0x00000000ffffffffULL));
   1864 		IXGBE_WRITE_REG(hw, IXGBE_VFRDBAH(j), (rdba >> 32));
   1865 		IXGBE_WRITE_REG(hw, IXGBE_VFRDLEN(j),
   1866 		    adapter->num_rx_desc * sizeof(union ixgbe_adv_rx_desc));
   1867 
   1868 		/* Reset the ring indices */
   1869 		IXGBE_WRITE_REG(hw, IXGBE_VFRDH(rxr->me), 0);
   1870 		IXGBE_WRITE_REG(hw, IXGBE_VFRDT(rxr->me), 0);
   1871 
   1872 		/* Set up the SRRCTL register */
   1873 		reg = IXGBE_READ_REG(hw, IXGBE_VFSRRCTL(j));
   1874 		reg &= ~IXGBE_SRRCTL_BSIZEHDR_MASK;
   1875 		reg &= ~IXGBE_SRRCTL_BSIZEPKT_MASK;
   1876 		reg |= bufsz;
   1877 		reg |= IXGBE_SRRCTL_DESCTYPE_ADV_ONEBUF;
   1878 		IXGBE_WRITE_REG(hw, IXGBE_VFSRRCTL(j), reg);
   1879 
   1880 		/* Capture Rx Tail index */
   1881 		rxr->tail = IXGBE_VFRDT(rxr->me);
   1882 
   1883 		/* Do the queue enabling last */
   1884 		rxdctl |= IXGBE_RXDCTL_ENABLE | IXGBE_RXDCTL_VME;
   1885 		IXGBE_WRITE_REG(hw, IXGBE_VFRXDCTL(j), rxdctl);
   1886 		for (int k = 0; k < 10; k++) {
   1887 			if (IXGBE_READ_REG(hw, IXGBE_VFRXDCTL(j)) &
   1888 			    IXGBE_RXDCTL_ENABLE)
   1889 				break;
   1890 			msec_delay(1);
   1891 		}
   1892 		wmb();
   1893 
   1894 		/* Set the Tail Pointer */
   1895 #ifdef DEV_NETMAP
   1896 		/*
   1897 		 * In netmap mode, we must preserve the buffers made
   1898 		 * available to userspace before the if_init()
   1899 		 * (this is true by default on the TX side, because
   1900 		 * init makes all buffers available to userspace).
   1901 		 *
   1902 		 * netmap_reset() and the device specific routines
   1903 		 * (e.g. ixgbe_setup_receive_rings()) map these
   1904 		 * buffers at the end of the NIC ring, so here we
   1905 		 * must set the RDT (tail) register to make sure
   1906 		 * they are not overwritten.
   1907 		 *
   1908 		 * In this driver the NIC ring starts at RDH = 0,
   1909 		 * RDT points to the last slot available for reception (?),
   1910 		 * so RDT = num_rx_desc - 1 means the whole ring is available.
   1911 		 */
   1912 		if ((adapter->feat_en & IXGBE_FEATURE_NETMAP) &&
   1913 		    (ifp->if_capenable & IFCAP_NETMAP)) {
   1914 			struct netmap_adapter *na = NA(adapter->ifp);
   1915 			struct netmap_kring *kring = na->rx_rings[i];
   1916 			int t = na->num_rx_desc - 1 - nm_kr_rxspace(kring);
   1917 
   1918 			IXGBE_WRITE_REG(hw, IXGBE_VFRDT(rxr->me), t);
   1919 		} else
   1920 #endif /* DEV_NETMAP */
   1921 			IXGBE_WRITE_REG(hw, IXGBE_VFRDT(rxr->me),
   1922 			    adapter->num_rx_desc - 1);
   1923 	}
   1924 
   1925 	if (adapter->hw.mac.type >= ixgbe_mac_X550_vf)
   1926 		ixv_initialize_rss_mapping(adapter);
   1927 } /* ixv_initialize_receive_units */
   1928 
   1929 /************************************************************************
   1930  * ixv_sysctl_tdh_handler - Transmit Descriptor Head handler function
   1931  *
   1932  *   Retrieves the TDH value from the hardware
   1933  ************************************************************************/
   1934 static int
   1935 ixv_sysctl_tdh_handler(SYSCTLFN_ARGS)
   1936 {
   1937 	struct sysctlnode node = *rnode;
   1938 	struct tx_ring *txr = (struct tx_ring *)node.sysctl_data;
   1939 	uint32_t val;
   1940 
   1941 	if (!txr)
   1942 		return (0);
   1943 
   1944 	val = IXGBE_READ_REG(&txr->adapter->hw, IXGBE_VFTDH(txr->me));
   1945 	node.sysctl_data = &val;
   1946 	return sysctl_lookup(SYSCTLFN_CALL(&node));
   1947 } /* ixv_sysctl_tdh_handler */
   1948 
   1949 /************************************************************************
   1950  * ixgbe_sysctl_tdt_handler - Transmit Descriptor Tail handler function
   1951  *
   1952  *   Retrieves the TDT value from the hardware
   1953  ************************************************************************/
   1954 static int
   1955 ixv_sysctl_tdt_handler(SYSCTLFN_ARGS)
   1956 {
   1957 	struct sysctlnode node = *rnode;
   1958 	struct tx_ring *txr = (struct tx_ring *)node.sysctl_data;
   1959 	uint32_t val;
   1960 
   1961 	if (!txr)
   1962 		return (0);
   1963 
   1964 	val = IXGBE_READ_REG(&txr->adapter->hw, IXGBE_VFTDT(txr->me));
   1965 	node.sysctl_data = &val;
   1966 	return sysctl_lookup(SYSCTLFN_CALL(&node));
   1967 } /* ixv_sysctl_tdt_handler */
   1968 
   1969 /************************************************************************
   1970  * ixv_sysctl_next_to_check_handler - Receive Descriptor next to check
   1971  * handler function
   1972  *
   1973  *   Retrieves the next_to_check value
   1974  ************************************************************************/
   1975 static int
   1976 ixv_sysctl_next_to_check_handler(SYSCTLFN_ARGS)
   1977 {
   1978 	struct sysctlnode node = *rnode;
   1979 	struct rx_ring *rxr = (struct rx_ring *)node.sysctl_data;
   1980 	uint32_t val;
   1981 
   1982 	if (!rxr)
   1983 		return (0);
   1984 
   1985 	val = rxr->next_to_check;
   1986 	node.sysctl_data = &val;
   1987 	return sysctl_lookup(SYSCTLFN_CALL(&node));
   1988 } /* ixv_sysctl_next_to_check_handler */
   1989 
   1990 /************************************************************************
   1991  * ixv_sysctl_rdh_handler - Receive Descriptor Head handler function
   1992  *
   1993  *   Retrieves the RDH value from the hardware
   1994  ************************************************************************/
   1995 static int
   1996 ixv_sysctl_rdh_handler(SYSCTLFN_ARGS)
   1997 {
   1998 	struct sysctlnode node = *rnode;
   1999 	struct rx_ring *rxr = (struct rx_ring *)node.sysctl_data;
   2000 	uint32_t val;
   2001 
   2002 	if (!rxr)
   2003 		return (0);
   2004 
   2005 	val = IXGBE_READ_REG(&rxr->adapter->hw, IXGBE_VFRDH(rxr->me));
   2006 	node.sysctl_data = &val;
   2007 	return sysctl_lookup(SYSCTLFN_CALL(&node));
   2008 } /* ixv_sysctl_rdh_handler */
   2009 
   2010 /************************************************************************
   2011  * ixv_sysctl_rdt_handler - Receive Descriptor Tail handler function
   2012  *
   2013  *   Retrieves the RDT value from the hardware
   2014  ************************************************************************/
   2015 static int
   2016 ixv_sysctl_rdt_handler(SYSCTLFN_ARGS)
   2017 {
   2018 	struct sysctlnode node = *rnode;
   2019 	struct rx_ring *rxr = (struct rx_ring *)node.sysctl_data;
   2020 	uint32_t val;
   2021 
   2022 	if (!rxr)
   2023 		return (0);
   2024 
   2025 	val = IXGBE_READ_REG(&rxr->adapter->hw, IXGBE_VFRDT(rxr->me));
   2026 	node.sysctl_data = &val;
   2027 	return sysctl_lookup(SYSCTLFN_CALL(&node));
   2028 } /* ixv_sysctl_rdt_handler */
   2029 
   2030 static void
   2031 ixv_setup_vlan_tagging(struct adapter *adapter)
   2032 {
   2033 	struct ethercom *ec = &adapter->osdep.ec;
   2034 	struct ixgbe_hw *hw = &adapter->hw;
   2035 	struct rx_ring	*rxr;
   2036 	u32		ctrl;
   2037 	int		i;
   2038 	bool		hwtagging;
   2039 
   2040 	/* Enable HW tagging only if any vlan is attached */
   2041 	hwtagging = (ec->ec_capenable & ETHERCAP_VLAN_HWTAGGING)
   2042 	    && VLAN_ATTACHED(ec);
   2043 
   2044 	/* Enable the queues */
   2045 	for (i = 0; i < adapter->num_queues; i++) {
   2046 		rxr = &adapter->rx_rings[i];
   2047 		ctrl = IXGBE_READ_REG(hw, IXGBE_VFRXDCTL(rxr->me));
   2048 		if (hwtagging)
   2049 			ctrl |= IXGBE_RXDCTL_VME;
   2050 		else
   2051 			ctrl &= ~IXGBE_RXDCTL_VME;
   2052 		IXGBE_WRITE_REG(hw, IXGBE_VFRXDCTL(rxr->me), ctrl);
   2053 		/*
   2054 		 * Let Rx path know that it needs to store VLAN tag
   2055 		 * as part of extra mbuf info.
   2056 		 */
   2057 		rxr->vtag_strip = hwtagging ? TRUE : FALSE;
   2058 	}
   2059 } /* ixv_setup_vlan_tagging */
   2060 
   2061 /************************************************************************
   2062  * ixv_setup_vlan_support
   2063  ************************************************************************/
   2064 static int
   2065 ixv_setup_vlan_support(struct adapter *adapter)
   2066 {
   2067 	struct ethercom *ec = &adapter->osdep.ec;
   2068 	struct ixgbe_hw *hw = &adapter->hw;
   2069 	u32		vid, vfta, retry;
   2070 	struct vlanid_list *vlanidp;
   2071 	int rv, error = 0;
   2072 
   2073 	/*
   2074 	 *  This function is called from both if_init and ifflags_cb()
   2075 	 * on NetBSD.
   2076 	 */
   2077 
   2078 	/*
   2079 	 * Part 1:
   2080 	 * Setup VLAN HW tagging
   2081 	 */
   2082 	ixv_setup_vlan_tagging(adapter);
   2083 
   2084 	if (!VLAN_ATTACHED(ec))
   2085 		return 0;
   2086 
   2087 	/*
   2088 	 * Part 2:
   2089 	 * Setup VLAN HW filter
   2090 	 */
   2091 	/* Cleanup shadow_vfta */
   2092 	for (int i = 0; i < IXGBE_VFTA_SIZE; i++)
   2093 		adapter->shadow_vfta[i] = 0;
   2094 	/* Generate shadow_vfta from ec_vids */
   2095 	ETHER_LOCK(ec);
   2096 	SIMPLEQ_FOREACH(vlanidp, &ec->ec_vids, vid_list) {
   2097 		uint32_t idx;
   2098 
   2099 		idx = vlanidp->vid / 32;
   2100 		KASSERT(idx < IXGBE_VFTA_SIZE);
   2101 		adapter->shadow_vfta[idx] |= (u32)1 << (vlanidp->vid % 32);
   2102 	}
   2103 	ETHER_UNLOCK(ec);
   2104 
   2105 	/*
   2106 	 * A soft reset zero's out the VFTA, so
   2107 	 * we need to repopulate it now.
   2108 	 */
   2109 	for (int i = 0; i < IXGBE_VFTA_SIZE; i++) {
   2110 		if (adapter->shadow_vfta[i] == 0)
   2111 			continue;
   2112 		vfta = adapter->shadow_vfta[i];
   2113 		/*
   2114 		 * Reconstruct the vlan id's
   2115 		 * based on the bits set in each
   2116 		 * of the array ints.
   2117 		 */
   2118 		for (int j = 0; j < 32; j++) {
   2119 			retry = 0;
   2120 			if ((vfta & ((u32)1 << j)) == 0)
   2121 				continue;
   2122 			vid = (i * 32) + j;
   2123 
   2124 			/* Call the shared code mailbox routine */
   2125 			while ((rv = hw->mac.ops.set_vfta(hw, vid, 0, TRUE,
   2126 			    FALSE)) != 0) {
   2127 				if (++retry > 5) {
   2128 					device_printf(adapter->dev,
   2129 					    "%s: max retry exceeded\n",
   2130 						__func__);
   2131 					break;
   2132 				}
   2133 			}
   2134 			if (rv != 0) {
   2135 				device_printf(adapter->dev,
   2136 				    "failed to set vlan %d\n", vid);
   2137 				error = EACCES;
   2138 			}
   2139 		}
   2140 	}
   2141 	return error;
   2142 } /* ixv_setup_vlan_support */
   2143 
   2144 static int
   2145 ixv_vlan_cb(struct ethercom *ec, uint16_t vid, bool set)
   2146 {
   2147 	struct ifnet *ifp = &ec->ec_if;
   2148 	struct adapter *adapter = ifp->if_softc;
   2149 	int rv;
   2150 
   2151 	if (set)
   2152 		rv = ixv_register_vlan(adapter, vid);
   2153 	else
   2154 		rv = ixv_unregister_vlan(adapter, vid);
   2155 
   2156 	if (rv != 0)
   2157 		return rv;
   2158 
   2159 	/*
   2160 	 * Control VLAN HW tagging when ec_nvlan is changed from 1 to 0
   2161 	 * or 0 to 1.
   2162 	 */
   2163 	if ((set && (ec->ec_nvlans == 1)) || (!set && (ec->ec_nvlans == 0)))
   2164 		ixv_setup_vlan_tagging(adapter);
   2165 
   2166 	return rv;
   2167 }
   2168 
   2169 /************************************************************************
   2170  * ixv_register_vlan
   2171  *
   2172  *   Run via a vlan config EVENT, it enables us to use the
   2173  *   HW Filter table since we can get the vlan id. This just
   2174  *   creates the entry in the soft version of the VFTA, init
   2175  *   will repopulate the real table.
   2176  ************************************************************************/
   2177 static int
   2178 ixv_register_vlan(struct adapter *adapter, u16 vtag)
   2179 {
   2180 	struct ixgbe_hw *hw = &adapter->hw;
   2181 	u16		index, bit;
   2182 	int error;
   2183 
   2184 	if ((vtag == 0) || (vtag > 4095)) /* Invalid */
   2185 		return EINVAL;
   2186 	IXGBE_CORE_LOCK(adapter);
   2187 	index = (vtag >> 5) & 0x7F;
   2188 	bit = vtag & 0x1F;
   2189 	adapter->shadow_vfta[index] |= ((u32)1 << bit);
   2190 	error = hw->mac.ops.set_vfta(hw, vtag, 0, true, false);
   2191 	IXGBE_CORE_UNLOCK(adapter);
   2192 
   2193 	if (error != 0) {
   2194 		device_printf(adapter->dev, "failed to register vlan %hu\n",
   2195 		    vtag);
   2196 		error = EACCES;
   2197 	}
   2198 	return error;
   2199 } /* ixv_register_vlan */
   2200 
   2201 /************************************************************************
   2202  * ixv_unregister_vlan
   2203  *
   2204  *   Run via a vlan unconfig EVENT, remove our entry
   2205  *   in the soft vfta.
   2206  ************************************************************************/
   2207 static int
   2208 ixv_unregister_vlan(struct adapter *adapter, u16 vtag)
   2209 {
   2210 	struct ixgbe_hw *hw = &adapter->hw;
   2211 	u16		index, bit;
   2212 	int 		error;
   2213 
   2214 	if ((vtag == 0) || (vtag > 4095))  /* Invalid */
   2215 		return EINVAL;
   2216 
   2217 	IXGBE_CORE_LOCK(adapter);
   2218 	index = (vtag >> 5) & 0x7F;
   2219 	bit = vtag & 0x1F;
   2220 	adapter->shadow_vfta[index] &= ~((u32)1 << bit);
   2221 	error = hw->mac.ops.set_vfta(hw, vtag, 0, false, false);
   2222 	IXGBE_CORE_UNLOCK(adapter);
   2223 
   2224 	if (error != 0) {
   2225 		device_printf(adapter->dev, "failed to unregister vlan %hu\n",
   2226 		    vtag);
   2227 		error = EIO;
   2228 	}
   2229 	return error;
   2230 } /* ixv_unregister_vlan */
   2231 
   2232 /************************************************************************
   2233  * ixv_enable_intr
   2234  ************************************************************************/
   2235 static void
   2236 ixv_enable_intr(struct adapter *adapter)
   2237 {
   2238 	struct ixgbe_hw *hw = &adapter->hw;
   2239 	struct ix_queue *que = adapter->queues;
   2240 	u32		mask;
   2241 	int i;
   2242 
   2243 	/* For VTEIAC */
   2244 	mask = (1 << adapter->vector);
   2245 	for (i = 0; i < adapter->num_queues; i++, que++)
   2246 		mask |= (1 << que->msix);
   2247 	IXGBE_WRITE_REG(hw, IXGBE_VTEIAC, mask);
   2248 
   2249 	/* For VTEIMS */
   2250 	IXGBE_WRITE_REG(hw, IXGBE_VTEIMS, (1 << adapter->vector));
   2251 	que = adapter->queues;
   2252 	for (i = 0; i < adapter->num_queues; i++, que++)
   2253 		ixv_enable_queue(adapter, que->msix);
   2254 
   2255 	IXGBE_WRITE_FLUSH(hw);
   2256 } /* ixv_enable_intr */
   2257 
   2258 /************************************************************************
   2259  * ixv_disable_intr
   2260  ************************************************************************/
   2261 static void
   2262 ixv_disable_intr(struct adapter *adapter)
   2263 {
   2264 	struct ix_queue	*que = adapter->queues;
   2265 
   2266 	IXGBE_WRITE_REG(&adapter->hw, IXGBE_VTEIAC, 0);
   2267 
   2268 	/* disable interrupts other than queues */
   2269 	IXGBE_WRITE_REG(&adapter->hw, IXGBE_VTEIMC, adapter->vector);
   2270 
   2271 	for (int i = 0; i < adapter->num_queues; i++, que++)
   2272 		ixv_disable_queue(adapter, que->msix);
   2273 
   2274 	IXGBE_WRITE_FLUSH(&adapter->hw);
   2275 } /* ixv_disable_intr */
   2276 
   2277 /************************************************************************
   2278  * ixv_set_ivar
   2279  *
   2280  *   Setup the correct IVAR register for a particular MSI-X interrupt
   2281  *    - entry is the register array entry
   2282  *    - vector is the MSI-X vector for this queue
   2283  *    - type is RX/TX/MISC
   2284  ************************************************************************/
   2285 static void
   2286 ixv_set_ivar(struct adapter *adapter, u8 entry, u8 vector, s8 type)
   2287 {
   2288 	struct ixgbe_hw *hw = &adapter->hw;
   2289 	u32		ivar, index;
   2290 
   2291 	vector |= IXGBE_IVAR_ALLOC_VAL;
   2292 
   2293 	if (type == -1) { /* MISC IVAR */
   2294 		ivar = IXGBE_READ_REG(hw, IXGBE_VTIVAR_MISC);
   2295 		ivar &= ~0xFF;
   2296 		ivar |= vector;
   2297 		IXGBE_WRITE_REG(hw, IXGBE_VTIVAR_MISC, ivar);
   2298 	} else {	  /* RX/TX IVARS */
   2299 		index = (16 * (entry & 1)) + (8 * type);
   2300 		ivar = IXGBE_READ_REG(hw, IXGBE_VTIVAR(entry >> 1));
   2301 		ivar &= ~(0xffUL << index);
   2302 		ivar |= ((u32)vector << index);
   2303 		IXGBE_WRITE_REG(hw, IXGBE_VTIVAR(entry >> 1), ivar);
   2304 	}
   2305 } /* ixv_set_ivar */
   2306 
   2307 /************************************************************************
   2308  * ixv_configure_ivars
   2309  ************************************************************************/
   2310 static void
   2311 ixv_configure_ivars(struct adapter *adapter)
   2312 {
   2313 	struct ix_queue *que = adapter->queues;
   2314 
   2315 	/* XXX We should sync EITR value calculation with ixgbe.c? */
   2316 
   2317 	for (int i = 0; i < adapter->num_queues; i++, que++) {
   2318 		/* First the RX queue entry */
   2319 		ixv_set_ivar(adapter, i, que->msix, 0);
   2320 		/* ... and the TX */
   2321 		ixv_set_ivar(adapter, i, que->msix, 1);
   2322 		/* Set an initial value in EITR */
   2323 		ixv_eitr_write(adapter, que->msix, IXGBE_EITR_DEFAULT);
   2324 	}
   2325 
   2326 	/* For the mailbox interrupt */
   2327 	ixv_set_ivar(adapter, 1, adapter->vector, -1);
   2328 } /* ixv_configure_ivars */
   2329 
   2330 
   2331 /************************************************************************
   2332  * ixv_save_stats
   2333  *
   2334  *   The VF stats registers never have a truly virgin
   2335  *   starting point, so this routine tries to make an
   2336  *   artificial one, marking ground zero on attach as
   2337  *   it were.
   2338  ************************************************************************/
   2339 static void
   2340 ixv_save_stats(struct adapter *adapter)
   2341 {
   2342 	struct ixgbevf_hw_stats *stats = &adapter->stats.vf;
   2343 
   2344 	if (stats->vfgprc.ev_count || stats->vfgptc.ev_count) {
   2345 		stats->saved_reset_vfgprc +=
   2346 		    stats->vfgprc.ev_count - stats->base_vfgprc;
   2347 		stats->saved_reset_vfgptc +=
   2348 		    stats->vfgptc.ev_count - stats->base_vfgptc;
   2349 		stats->saved_reset_vfgorc +=
   2350 		    stats->vfgorc.ev_count - stats->base_vfgorc;
   2351 		stats->saved_reset_vfgotc +=
   2352 		    stats->vfgotc.ev_count - stats->base_vfgotc;
   2353 		stats->saved_reset_vfmprc +=
   2354 		    stats->vfmprc.ev_count - stats->base_vfmprc;
   2355 	}
   2356 } /* ixv_save_stats */
   2357 
   2358 /************************************************************************
   2359  * ixv_init_stats
   2360  ************************************************************************/
   2361 static void
   2362 ixv_init_stats(struct adapter *adapter)
   2363 {
   2364 	struct ixgbe_hw *hw = &adapter->hw;
   2365 
   2366 	adapter->stats.vf.last_vfgprc = IXGBE_READ_REG(hw, IXGBE_VFGPRC);
   2367 	adapter->stats.vf.last_vfgorc = IXGBE_READ_REG(hw, IXGBE_VFGORC_LSB);
   2368 	adapter->stats.vf.last_vfgorc |=
   2369 	    (((u64)(IXGBE_READ_REG(hw, IXGBE_VFGORC_MSB))) << 32);
   2370 
   2371 	adapter->stats.vf.last_vfgptc = IXGBE_READ_REG(hw, IXGBE_VFGPTC);
   2372 	adapter->stats.vf.last_vfgotc = IXGBE_READ_REG(hw, IXGBE_VFGOTC_LSB);
   2373 	adapter->stats.vf.last_vfgotc |=
   2374 	    (((u64)(IXGBE_READ_REG(hw, IXGBE_VFGOTC_MSB))) << 32);
   2375 
   2376 	adapter->stats.vf.last_vfmprc = IXGBE_READ_REG(hw, IXGBE_VFMPRC);
   2377 
   2378 	adapter->stats.vf.base_vfgprc = adapter->stats.vf.last_vfgprc;
   2379 	adapter->stats.vf.base_vfgorc = adapter->stats.vf.last_vfgorc;
   2380 	adapter->stats.vf.base_vfgptc = adapter->stats.vf.last_vfgptc;
   2381 	adapter->stats.vf.base_vfgotc = adapter->stats.vf.last_vfgotc;
   2382 	adapter->stats.vf.base_vfmprc = adapter->stats.vf.last_vfmprc;
   2383 } /* ixv_init_stats */
   2384 
   2385 #define UPDATE_STAT_32(reg, last, count)		\
   2386 {							\
   2387 	u32 current = IXGBE_READ_REG(hw, (reg));	\
   2388 	if (current < (last))				\
   2389 		count.ev_count += 0x100000000LL;	\
   2390 	(last) = current;				\
   2391 	count.ev_count &= 0xFFFFFFFF00000000LL;		\
   2392 	count.ev_count |= current;			\
   2393 }
   2394 
   2395 #define UPDATE_STAT_36(lsb, msb, last, count)		\
   2396 {							\
   2397 	u64 cur_lsb = IXGBE_READ_REG(hw, (lsb));	\
   2398 	u64 cur_msb = IXGBE_READ_REG(hw, (msb));	\
   2399 	u64 current = ((cur_msb << 32) | cur_lsb);	\
   2400 	if (current < (last))				\
   2401 		count.ev_count += 0x1000000000LL;	\
   2402 	(last) = current;				\
   2403 	count.ev_count &= 0xFFFFFFF000000000LL;		\
   2404 	count.ev_count |= current;			\
   2405 }
   2406 
   2407 /************************************************************************
   2408  * ixv_update_stats - Update the board statistics counters.
   2409  ************************************************************************/
   2410 void
   2411 ixv_update_stats(struct adapter *adapter)
   2412 {
   2413 	struct ixgbe_hw *hw = &adapter->hw;
   2414 	struct ixgbevf_hw_stats *stats = &adapter->stats.vf;
   2415 
   2416 	UPDATE_STAT_32(IXGBE_VFGPRC, stats->last_vfgprc, stats->vfgprc);
   2417 	UPDATE_STAT_32(IXGBE_VFGPTC, stats->last_vfgptc, stats->vfgptc);
   2418 	UPDATE_STAT_36(IXGBE_VFGORC_LSB, IXGBE_VFGORC_MSB, stats->last_vfgorc,
   2419 	    stats->vfgorc);
   2420 	UPDATE_STAT_36(IXGBE_VFGOTC_LSB, IXGBE_VFGOTC_MSB, stats->last_vfgotc,
   2421 	    stats->vfgotc);
   2422 	UPDATE_STAT_32(IXGBE_VFMPRC, stats->last_vfmprc, stats->vfmprc);
   2423 
   2424 	/* Fill out the OS statistics structure */
   2425 	/*
   2426 	 * NetBSD: Don't override if_{i|o}{packets|bytes|mcasts} with
   2427 	 * adapter->stats counters. It's required to make ifconfig -z
   2428 	 * (SOICZIFDATA) work.
   2429 	 */
   2430 } /* ixv_update_stats */
   2431 
   2432 /************************************************************************
   2433  * ixv_sysctl_interrupt_rate_handler
   2434  ************************************************************************/
   2435 static int
   2436 ixv_sysctl_interrupt_rate_handler(SYSCTLFN_ARGS)
   2437 {
   2438 	struct sysctlnode node = *rnode;
   2439 	struct ix_queue *que = (struct ix_queue *)node.sysctl_data;
   2440 	struct adapter	*adapter = que->adapter;
   2441 	uint32_t reg, usec, rate;
   2442 	int error;
   2443 
   2444 	if (que == NULL)
   2445 		return 0;
   2446 	reg = IXGBE_READ_REG(&que->adapter->hw, IXGBE_VTEITR(que->msix));
   2447 	usec = ((reg & 0x0FF8) >> 3);
   2448 	if (usec > 0)
   2449 		rate = 500000 / usec;
   2450 	else
   2451 		rate = 0;
   2452 	node.sysctl_data = &rate;
   2453 	error = sysctl_lookup(SYSCTLFN_CALL(&node));
   2454 	if (error || newp == NULL)
   2455 		return error;
   2456 	reg &= ~0xfff; /* default, no limitation */
   2457 	if (rate > 0 && rate < 500000) {
   2458 		if (rate < 1000)
   2459 			rate = 1000;
   2460 		reg |= ((4000000/rate) & 0xff8);
   2461 		/*
   2462 		 * When RSC is used, ITR interval must be larger than
   2463 		 * RSC_DELAY. Currently, we use 2us for RSC_DELAY.
   2464 		 * The minimum value is always greater than 2us on 100M
   2465 		 * (and 10M?(not documented)), but it's not on 1G and higher.
   2466 		 */
   2467 		if ((adapter->link_speed != IXGBE_LINK_SPEED_100_FULL)
   2468 		    && (adapter->link_speed != IXGBE_LINK_SPEED_10_FULL)) {
   2469 			if ((adapter->num_queues > 1)
   2470 			    && (reg < IXGBE_MIN_RSC_EITR_10G1G))
   2471 				return EINVAL;
   2472 		}
   2473 		ixv_max_interrupt_rate = rate;
   2474 	} else
   2475 		ixv_max_interrupt_rate = 0;
   2476 	ixv_eitr_write(adapter, que->msix, reg);
   2477 
   2478 	return (0);
   2479 } /* ixv_sysctl_interrupt_rate_handler */
   2480 
   2481 const struct sysctlnode *
   2482 ixv_sysctl_instance(struct adapter *adapter)
   2483 {
   2484 	const char *dvname;
   2485 	struct sysctllog **log;
   2486 	int rc;
   2487 	const struct sysctlnode *rnode;
   2488 
   2489 	log = &adapter->sysctllog;
   2490 	dvname = device_xname(adapter->dev);
   2491 
   2492 	if ((rc = sysctl_createv(log, 0, NULL, &rnode,
   2493 	    0, CTLTYPE_NODE, dvname,
   2494 	    SYSCTL_DESCR("ixv information and settings"),
   2495 	    NULL, 0, NULL, 0, CTL_HW, CTL_CREATE, CTL_EOL)) != 0)
   2496 		goto err;
   2497 
   2498 	return rnode;
   2499 err:
   2500 	device_printf(adapter->dev,
   2501 	    "%s: sysctl_createv failed, rc = %d\n", __func__, rc);
   2502 	return NULL;
   2503 }
   2504 
   2505 static void
   2506 ixv_add_device_sysctls(struct adapter *adapter)
   2507 {
   2508 	struct sysctllog **log;
   2509 	const struct sysctlnode *rnode, *cnode;
   2510 	device_t dev;
   2511 
   2512 	dev = adapter->dev;
   2513 	log = &adapter->sysctllog;
   2514 
   2515 	if ((rnode = ixv_sysctl_instance(adapter)) == NULL) {
   2516 		aprint_error_dev(dev, "could not create sysctl root\n");
   2517 		return;
   2518 	}
   2519 
   2520 	if (sysctl_createv(log, 0, &rnode, &cnode,
   2521 	    CTLFLAG_READWRITE, CTLTYPE_INT,
   2522 	    "debug", SYSCTL_DESCR("Debug Info"),
   2523 	    ixv_sysctl_debug, 0, (void *)adapter, 0, CTL_CREATE, CTL_EOL) != 0)
   2524 		aprint_error_dev(dev, "could not create sysctl\n");
   2525 
   2526 	if (sysctl_createv(log, 0, &rnode, &cnode,
   2527 	    CTLFLAG_READWRITE, CTLTYPE_BOOL,
   2528 	    "enable_aim", SYSCTL_DESCR("Interrupt Moderation"),
   2529 	    NULL, 0, &adapter->enable_aim, 0, CTL_CREATE, CTL_EOL) != 0)
   2530 		aprint_error_dev(dev, "could not create sysctl\n");
   2531 
   2532 	if (sysctl_createv(log, 0, &rnode, &cnode,
   2533 	    CTLFLAG_READWRITE, CTLTYPE_BOOL,
   2534 	    "txrx_workqueue", SYSCTL_DESCR("Use workqueue for packet processing"),
   2535 		NULL, 0, &adapter->txrx_use_workqueue, 0, CTL_CREATE, CTL_EOL) != 0)
   2536 		aprint_error_dev(dev, "could not create sysctl\n");
   2537 }
   2538 
   2539 /************************************************************************
   2540  * ixv_add_stats_sysctls - Add statistic sysctls for the VF.
   2541  ************************************************************************/
   2542 static void
   2543 ixv_add_stats_sysctls(struct adapter *adapter)
   2544 {
   2545 	device_t		dev = adapter->dev;
   2546 	struct tx_ring		*txr = adapter->tx_rings;
   2547 	struct rx_ring		*rxr = adapter->rx_rings;
   2548 	struct ixgbevf_hw_stats *stats = &adapter->stats.vf;
   2549 	struct ixgbe_hw *hw = &adapter->hw;
   2550 	const struct sysctlnode *rnode, *cnode;
   2551 	struct sysctllog **log = &adapter->sysctllog;
   2552 	const char *xname = device_xname(dev);
   2553 
   2554 	/* Driver Statistics */
   2555 	evcnt_attach_dynamic(&adapter->efbig_tx_dma_setup, EVCNT_TYPE_MISC,
   2556 	    NULL, xname, "Driver tx dma soft fail EFBIG");
   2557 	evcnt_attach_dynamic(&adapter->mbuf_defrag_failed, EVCNT_TYPE_MISC,
   2558 	    NULL, xname, "m_defrag() failed");
   2559 	evcnt_attach_dynamic(&adapter->efbig2_tx_dma_setup, EVCNT_TYPE_MISC,
   2560 	    NULL, xname, "Driver tx dma hard fail EFBIG");
   2561 	evcnt_attach_dynamic(&adapter->einval_tx_dma_setup, EVCNT_TYPE_MISC,
   2562 	    NULL, xname, "Driver tx dma hard fail EINVAL");
   2563 	evcnt_attach_dynamic(&adapter->other_tx_dma_setup, EVCNT_TYPE_MISC,
   2564 	    NULL, xname, "Driver tx dma hard fail other");
   2565 	evcnt_attach_dynamic(&adapter->eagain_tx_dma_setup, EVCNT_TYPE_MISC,
   2566 	    NULL, xname, "Driver tx dma soft fail EAGAIN");
   2567 	evcnt_attach_dynamic(&adapter->enomem_tx_dma_setup, EVCNT_TYPE_MISC,
   2568 	    NULL, xname, "Driver tx dma soft fail ENOMEM");
   2569 	evcnt_attach_dynamic(&adapter->watchdog_events, EVCNT_TYPE_MISC,
   2570 	    NULL, xname, "Watchdog timeouts");
   2571 	evcnt_attach_dynamic(&adapter->tso_err, EVCNT_TYPE_MISC,
   2572 	    NULL, xname, "TSO errors");
   2573 	evcnt_attach_dynamic(&adapter->link_irq, EVCNT_TYPE_INTR,
   2574 	    NULL, xname, "Link MSI-X IRQ Handled");
   2575 
   2576 	for (int i = 0; i < adapter->num_queues; i++, rxr++, txr++) {
   2577 		snprintf(adapter->queues[i].evnamebuf,
   2578 		    sizeof(adapter->queues[i].evnamebuf), "%s q%d",
   2579 		    xname, i);
   2580 		snprintf(adapter->queues[i].namebuf,
   2581 		    sizeof(adapter->queues[i].namebuf), "q%d", i);
   2582 
   2583 		if ((rnode = ixv_sysctl_instance(adapter)) == NULL) {
   2584 			aprint_error_dev(dev, "could not create sysctl root\n");
   2585 			break;
   2586 		}
   2587 
   2588 		if (sysctl_createv(log, 0, &rnode, &rnode,
   2589 		    0, CTLTYPE_NODE,
   2590 		    adapter->queues[i].namebuf, SYSCTL_DESCR("Queue Name"),
   2591 		    NULL, 0, NULL, 0, CTL_CREATE, CTL_EOL) != 0)
   2592 			break;
   2593 
   2594 		if (sysctl_createv(log, 0, &rnode, &cnode,
   2595 		    CTLFLAG_READWRITE, CTLTYPE_INT,
   2596 		    "interrupt_rate", SYSCTL_DESCR("Interrupt Rate"),
   2597 		    ixv_sysctl_interrupt_rate_handler, 0,
   2598 		    (void *)&adapter->queues[i], 0, CTL_CREATE, CTL_EOL) != 0)
   2599 			break;
   2600 
   2601 		if (sysctl_createv(log, 0, &rnode, &cnode,
   2602 		    CTLFLAG_READONLY, CTLTYPE_INT,
   2603 		    "txd_head", SYSCTL_DESCR("Transmit Descriptor Head"),
   2604 		    ixv_sysctl_tdh_handler, 0, (void *)txr,
   2605 		    0, CTL_CREATE, CTL_EOL) != 0)
   2606 			break;
   2607 
   2608 		if (sysctl_createv(log, 0, &rnode, &cnode,
   2609 		    CTLFLAG_READONLY, CTLTYPE_INT,
   2610 		    "txd_tail", SYSCTL_DESCR("Transmit Descriptor Tail"),
   2611 		    ixv_sysctl_tdt_handler, 0, (void *)txr,
   2612 		    0, CTL_CREATE, CTL_EOL) != 0)
   2613 			break;
   2614 
   2615 		evcnt_attach_dynamic(&adapter->queues[i].irqs, EVCNT_TYPE_INTR,
   2616 		    NULL, adapter->queues[i].evnamebuf, "IRQs on queue");
   2617 		evcnt_attach_dynamic(&adapter->queues[i].handleq,
   2618 		    EVCNT_TYPE_MISC, NULL, adapter->queues[i].evnamebuf,
   2619 		    "Handled queue in softint");
   2620 		evcnt_attach_dynamic(&adapter->queues[i].req, EVCNT_TYPE_MISC,
   2621 		    NULL, adapter->queues[i].evnamebuf, "Requeued in softint");
   2622 		evcnt_attach_dynamic(&txr->tso_tx, EVCNT_TYPE_MISC,
   2623 		    NULL, adapter->queues[i].evnamebuf, "TSO");
   2624 		evcnt_attach_dynamic(&txr->no_desc_avail, EVCNT_TYPE_MISC,
   2625 		    NULL, adapter->queues[i].evnamebuf,
   2626 		    "Queue No Descriptor Available");
   2627 		evcnt_attach_dynamic(&txr->total_packets, EVCNT_TYPE_MISC,
   2628 		    NULL, adapter->queues[i].evnamebuf,
   2629 		    "Queue Packets Transmitted");
   2630 #ifndef IXGBE_LEGACY_TX
   2631 		evcnt_attach_dynamic(&txr->pcq_drops, EVCNT_TYPE_MISC,
   2632 		    NULL, adapter->queues[i].evnamebuf,
   2633 		    "Packets dropped in pcq");
   2634 #endif
   2635 
   2636 #ifdef LRO
   2637 		struct lro_ctrl *lro = &rxr->lro;
   2638 #endif /* LRO */
   2639 
   2640 		if (sysctl_createv(log, 0, &rnode, &cnode,
   2641 		    CTLFLAG_READONLY,
   2642 		    CTLTYPE_INT,
   2643 		    "rxd_nxck", SYSCTL_DESCR("Receive Descriptor next to check"),
   2644 			ixv_sysctl_next_to_check_handler, 0, (void *)rxr, 0,
   2645 		    CTL_CREATE, CTL_EOL) != 0)
   2646 			break;
   2647 
   2648 		if (sysctl_createv(log, 0, &rnode, &cnode,
   2649 		    CTLFLAG_READONLY,
   2650 		    CTLTYPE_INT,
   2651 		    "rxd_head", SYSCTL_DESCR("Receive Descriptor Head"),
   2652 		    ixv_sysctl_rdh_handler, 0, (void *)rxr, 0,
   2653 		    CTL_CREATE, CTL_EOL) != 0)
   2654 			break;
   2655 
   2656 		if (sysctl_createv(log, 0, &rnode, &cnode,
   2657 		    CTLFLAG_READONLY,
   2658 		    CTLTYPE_INT,
   2659 		    "rxd_tail", SYSCTL_DESCR("Receive Descriptor Tail"),
   2660 		    ixv_sysctl_rdt_handler, 0, (void *)rxr, 0,
   2661 		    CTL_CREATE, CTL_EOL) != 0)
   2662 			break;
   2663 
   2664 		evcnt_attach_dynamic(&rxr->rx_packets, EVCNT_TYPE_MISC,
   2665 		    NULL, adapter->queues[i].evnamebuf, "Queue Packets Received");
   2666 		evcnt_attach_dynamic(&rxr->rx_bytes, EVCNT_TYPE_MISC,
   2667 		    NULL, adapter->queues[i].evnamebuf, "Queue Bytes Received");
   2668 		evcnt_attach_dynamic(&rxr->rx_copies, EVCNT_TYPE_MISC,
   2669 		    NULL, adapter->queues[i].evnamebuf, "Copied RX Frames");
   2670 		evcnt_attach_dynamic(&rxr->no_jmbuf, EVCNT_TYPE_MISC,
   2671 		    NULL, adapter->queues[i].evnamebuf, "Rx no jumbo mbuf");
   2672 		evcnt_attach_dynamic(&rxr->rx_discarded, EVCNT_TYPE_MISC,
   2673 		    NULL, adapter->queues[i].evnamebuf, "Rx discarded");
   2674 #ifdef LRO
   2675 		SYSCTL_ADD_INT(ctx, queue_list, OID_AUTO, "lro_queued",
   2676 				CTLFLAG_RD, &lro->lro_queued, 0,
   2677 				"LRO Queued");
   2678 		SYSCTL_ADD_INT(ctx, queue_list, OID_AUTO, "lro_flushed",
   2679 				CTLFLAG_RD, &lro->lro_flushed, 0,
   2680 				"LRO Flushed");
   2681 #endif /* LRO */
   2682 	}
   2683 
   2684 	/* MAC stats get their own sub node */
   2685 
   2686 	snprintf(stats->namebuf,
   2687 	    sizeof(stats->namebuf), "%s MAC Statistics", xname);
   2688 
   2689 	evcnt_attach_dynamic(&stats->ipcs, EVCNT_TYPE_MISC, NULL,
   2690 	    stats->namebuf, "rx csum offload - IP");
   2691 	evcnt_attach_dynamic(&stats->l4cs, EVCNT_TYPE_MISC, NULL,
   2692 	    stats->namebuf, "rx csum offload - L4");
   2693 	evcnt_attach_dynamic(&stats->ipcs_bad, EVCNT_TYPE_MISC, NULL,
   2694 	    stats->namebuf, "rx csum offload - IP bad");
   2695 	evcnt_attach_dynamic(&stats->l4cs_bad, EVCNT_TYPE_MISC, NULL,
   2696 	    stats->namebuf, "rx csum offload - L4 bad");
   2697 
   2698 	/* Packet Reception Stats */
   2699 	evcnt_attach_dynamic(&stats->vfgprc, EVCNT_TYPE_MISC, NULL,
   2700 	    xname, "Good Packets Received");
   2701 	evcnt_attach_dynamic(&stats->vfgorc, EVCNT_TYPE_MISC, NULL,
   2702 	    xname, "Good Octets Received");
   2703 	evcnt_attach_dynamic(&stats->vfmprc, EVCNT_TYPE_MISC, NULL,
   2704 	    xname, "Multicast Packets Received");
   2705 	evcnt_attach_dynamic(&stats->vfgptc, EVCNT_TYPE_MISC, NULL,
   2706 	    xname, "Good Packets Transmitted");
   2707 	evcnt_attach_dynamic(&stats->vfgotc, EVCNT_TYPE_MISC, NULL,
   2708 	    xname, "Good Octets Transmitted");
   2709 
   2710 	/* Mailbox Stats */
   2711 	evcnt_attach_dynamic(&hw->mbx.stats.msgs_tx, EVCNT_TYPE_MISC, NULL,
   2712 	    xname, "message TXs");
   2713 	evcnt_attach_dynamic(&hw->mbx.stats.msgs_rx, EVCNT_TYPE_MISC, NULL,
   2714 	    xname, "message RXs");
   2715 	evcnt_attach_dynamic(&hw->mbx.stats.acks, EVCNT_TYPE_MISC, NULL,
   2716 	    xname, "ACKs");
   2717 	evcnt_attach_dynamic(&hw->mbx.stats.reqs, EVCNT_TYPE_MISC, NULL,
   2718 	    xname, "REQs");
   2719 	evcnt_attach_dynamic(&hw->mbx.stats.rsts, EVCNT_TYPE_MISC, NULL,
   2720 	    xname, "RSTs");
   2721 
   2722 } /* ixv_add_stats_sysctls */
   2723 
   2724 static void
   2725 ixv_clear_evcnt(struct adapter *adapter)
   2726 {
   2727 	struct tx_ring		*txr = adapter->tx_rings;
   2728 	struct rx_ring		*rxr = adapter->rx_rings;
   2729 	struct ixgbevf_hw_stats *stats = &adapter->stats.vf;
   2730 	struct ixgbe_hw *hw = &adapter->hw;
   2731 	int i;
   2732 
   2733 	/* Driver Statistics */
   2734 	adapter->efbig_tx_dma_setup.ev_count = 0;
   2735 	adapter->mbuf_defrag_failed.ev_count = 0;
   2736 	adapter->efbig2_tx_dma_setup.ev_count = 0;
   2737 	adapter->einval_tx_dma_setup.ev_count = 0;
   2738 	adapter->other_tx_dma_setup.ev_count = 0;
   2739 	adapter->eagain_tx_dma_setup.ev_count = 0;
   2740 	adapter->enomem_tx_dma_setup.ev_count = 0;
   2741 	adapter->watchdog_events.ev_count = 0;
   2742 	adapter->tso_err.ev_count = 0;
   2743 	adapter->link_irq.ev_count = 0;
   2744 
   2745 	for (i = 0; i < adapter->num_queues; i++, rxr++, txr++) {
   2746 		adapter->queues[i].irqs.ev_count = 0;
   2747 		adapter->queues[i].handleq.ev_count = 0;
   2748 		adapter->queues[i].req.ev_count = 0;
   2749 		txr->tso_tx.ev_count = 0;
   2750 		txr->no_desc_avail.ev_count = 0;
   2751 		txr->total_packets.ev_count = 0;
   2752 #ifndef IXGBE_LEGACY_TX
   2753 		txr->pcq_drops.ev_count = 0;
   2754 #endif
   2755 		txr->q_efbig_tx_dma_setup = 0;
   2756 		txr->q_mbuf_defrag_failed = 0;
   2757 		txr->q_efbig2_tx_dma_setup = 0;
   2758 		txr->q_einval_tx_dma_setup = 0;
   2759 		txr->q_other_tx_dma_setup = 0;
   2760 		txr->q_eagain_tx_dma_setup = 0;
   2761 		txr->q_enomem_tx_dma_setup = 0;
   2762 		txr->q_tso_err = 0;
   2763 
   2764 		rxr->rx_packets.ev_count = 0;
   2765 		rxr->rx_bytes.ev_count = 0;
   2766 		rxr->rx_copies.ev_count = 0;
   2767 		rxr->no_jmbuf.ev_count = 0;
   2768 		rxr->rx_discarded.ev_count = 0;
   2769 	}
   2770 
   2771 	/* MAC stats get their own sub node */
   2772 
   2773 	stats->ipcs.ev_count = 0;
   2774 	stats->l4cs.ev_count = 0;
   2775 	stats->ipcs_bad.ev_count = 0;
   2776 	stats->l4cs_bad.ev_count = 0;
   2777 
   2778 	/* Packet Reception Stats */
   2779 	stats->vfgprc.ev_count = 0;
   2780 	stats->vfgorc.ev_count = 0;
   2781 	stats->vfmprc.ev_count = 0;
   2782 	stats->vfgptc.ev_count = 0;
   2783 	stats->vfgotc.ev_count = 0;
   2784 
   2785 	/* Mailbox Stats */
   2786 	hw->mbx.stats.msgs_tx.ev_count = 0;
   2787 	hw->mbx.stats.msgs_rx.ev_count = 0;
   2788 	hw->mbx.stats.acks.ev_count = 0;
   2789 	hw->mbx.stats.reqs.ev_count = 0;
   2790 	hw->mbx.stats.rsts.ev_count = 0;
   2791 
   2792 } /* ixv_clear_evcnt */
   2793 
   2794 /************************************************************************
   2795  * ixv_set_sysctl_value
   2796  ************************************************************************/
   2797 static void
   2798 ixv_set_sysctl_value(struct adapter *adapter, const char *name,
   2799 	const char *description, int *limit, int value)
   2800 {
   2801 	device_t dev =	adapter->dev;
   2802 	struct sysctllog **log;
   2803 	const struct sysctlnode *rnode, *cnode;
   2804 
   2805 	log = &adapter->sysctllog;
   2806 	if ((rnode = ixv_sysctl_instance(adapter)) == NULL) {
   2807 		aprint_error_dev(dev, "could not create sysctl root\n");
   2808 		return;
   2809 	}
   2810 	if (sysctl_createv(log, 0, &rnode, &cnode,
   2811 	    CTLFLAG_READWRITE, CTLTYPE_INT,
   2812 	    name, SYSCTL_DESCR(description),
   2813 	    NULL, 0, limit, 0, CTL_CREATE, CTL_EOL) != 0)
   2814 		aprint_error_dev(dev, "could not create sysctl\n");
   2815 	*limit = value;
   2816 } /* ixv_set_sysctl_value */
   2817 
   2818 /************************************************************************
   2819  * ixv_print_debug_info
   2820  *
   2821  *   Called only when em_display_debug_stats is enabled.
   2822  *   Provides a way to take a look at important statistics
   2823  *   maintained by the driver and hardware.
   2824  ************************************************************************/
   2825 static void
   2826 ixv_print_debug_info(struct adapter *adapter)
   2827 {
   2828 	device_t	dev = adapter->dev;
   2829 	struct ix_queue *que = adapter->queues;
   2830 	struct rx_ring	*rxr;
   2831 	struct tx_ring	*txr;
   2832 #ifdef LRO
   2833 	struct lro_ctrl *lro;
   2834 #endif /* LRO */
   2835 
   2836 	for (int i = 0; i < adapter->num_queues; i++, que++) {
   2837 		txr = que->txr;
   2838 		rxr = que->rxr;
   2839 #ifdef LRO
   2840 		lro = &rxr->lro;
   2841 #endif /* LRO */
   2842 		device_printf(dev, "QUE(%d) IRQs Handled: %lu\n",
   2843 		    que->msix, (long)que->irqs.ev_count);
   2844 		device_printf(dev, "RX(%d) Packets Received: %lld\n",
   2845 		    rxr->me, (long long)rxr->rx_packets.ev_count);
   2846 		device_printf(dev, "RX(%d) Bytes Received: %lu\n",
   2847 		    rxr->me, (long)rxr->rx_bytes.ev_count);
   2848 #ifdef LRO
   2849 		device_printf(dev, "RX(%d) LRO Queued= %ju\n",
   2850 		    rxr->me, (uintmax_t)lro->lro_queued);
   2851 		device_printf(dev, "RX(%d) LRO Flushed= %ju\n",
   2852 		    rxr->me, (uintmax_t)lro->lro_flushed);
   2853 #endif /* LRO */
   2854 		device_printf(dev, "TX(%d) Packets Sent: %lu\n",
   2855 		    txr->me, (long)txr->total_packets.ev_count);
   2856 		device_printf(dev, "TX(%d) NO Desc Avail: %lu\n",
   2857 		    txr->me, (long)txr->no_desc_avail.ev_count);
   2858 	}
   2859 
   2860 	device_printf(dev, "MBX IRQ Handled: %lu\n",
   2861 	    (long)adapter->link_irq.ev_count);
   2862 } /* ixv_print_debug_info */
   2863 
   2864 /************************************************************************
   2865  * ixv_sysctl_debug
   2866  ************************************************************************/
   2867 static int
   2868 ixv_sysctl_debug(SYSCTLFN_ARGS)
   2869 {
   2870 	struct sysctlnode node = *rnode;
   2871 	struct adapter *adapter = (struct adapter *)node.sysctl_data;
   2872 	int	       error, result;
   2873 
   2874 	node.sysctl_data = &result;
   2875 	error = sysctl_lookup(SYSCTLFN_CALL(&node));
   2876 
   2877 	if (error || newp == NULL)
   2878 		return error;
   2879 
   2880 	if (result == 1)
   2881 		ixv_print_debug_info(adapter);
   2882 
   2883 	return 0;
   2884 } /* ixv_sysctl_debug */
   2885 
   2886 /************************************************************************
   2887  * ixv_init_device_features
   2888  ************************************************************************/
   2889 static void
   2890 ixv_init_device_features(struct adapter *adapter)
   2891 {
   2892 	adapter->feat_cap = IXGBE_FEATURE_NETMAP
   2893 			  | IXGBE_FEATURE_VF
   2894 			  | IXGBE_FEATURE_RSS
   2895 			  | IXGBE_FEATURE_LEGACY_TX;
   2896 
   2897 	/* A tad short on feature flags for VFs, atm. */
   2898 	switch (adapter->hw.mac.type) {
   2899 	case ixgbe_mac_82599_vf:
   2900 		break;
   2901 	case ixgbe_mac_X540_vf:
   2902 		break;
   2903 	case ixgbe_mac_X550_vf:
   2904 	case ixgbe_mac_X550EM_x_vf:
   2905 	case ixgbe_mac_X550EM_a_vf:
   2906 		adapter->feat_cap |= IXGBE_FEATURE_NEEDS_CTXD;
   2907 		break;
   2908 	default:
   2909 		break;
   2910 	}
   2911 
   2912 	/* Enabled by default... */
   2913 	/* Is a virtual function (VF) */
   2914 	if (adapter->feat_cap & IXGBE_FEATURE_VF)
   2915 		adapter->feat_en |= IXGBE_FEATURE_VF;
   2916 	/* Netmap */
   2917 	if (adapter->feat_cap & IXGBE_FEATURE_NETMAP)
   2918 		adapter->feat_en |= IXGBE_FEATURE_NETMAP;
   2919 	/* Receive-Side Scaling (RSS) */
   2920 	if (adapter->feat_cap & IXGBE_FEATURE_RSS)
   2921 		adapter->feat_en |= IXGBE_FEATURE_RSS;
   2922 	/* Needs advanced context descriptor regardless of offloads req'd */
   2923 	if (adapter->feat_cap & IXGBE_FEATURE_NEEDS_CTXD)
   2924 		adapter->feat_en |= IXGBE_FEATURE_NEEDS_CTXD;
   2925 
   2926 	/* Enabled via sysctl... */
   2927 	/* Legacy (single queue) transmit */
   2928 	if ((adapter->feat_cap & IXGBE_FEATURE_LEGACY_TX) &&
   2929 	    ixv_enable_legacy_tx)
   2930 		adapter->feat_en |= IXGBE_FEATURE_LEGACY_TX;
   2931 } /* ixv_init_device_features */
   2932 
   2933 /************************************************************************
   2934  * ixv_shutdown - Shutdown entry point
   2935  ************************************************************************/
   2936 #if 0 /* XXX NetBSD ought to register something like this through pmf(9) */
   2937 static int
   2938 ixv_shutdown(device_t dev)
   2939 {
   2940 	struct adapter *adapter = device_private(dev);
   2941 	IXGBE_CORE_LOCK(adapter);
   2942 	ixv_stop(adapter);
   2943 	IXGBE_CORE_UNLOCK(adapter);
   2944 
   2945 	return (0);
   2946 } /* ixv_shutdown */
   2947 #endif
   2948 
   2949 static int
   2950 ixv_ifflags_cb(struct ethercom *ec)
   2951 {
   2952 	struct ifnet *ifp = &ec->ec_if;
   2953 	struct adapter *adapter = ifp->if_softc;
   2954 	u_short saved_flags;
   2955 	u_short change;
   2956 	int rv = 0;
   2957 
   2958 	IXGBE_CORE_LOCK(adapter);
   2959 
   2960 	saved_flags = adapter->if_flags;
   2961 	change = ifp->if_flags ^ adapter->if_flags;
   2962 	if (change != 0)
   2963 		adapter->if_flags = ifp->if_flags;
   2964 
   2965 	if ((change & ~(IFF_CANTCHANGE | IFF_DEBUG)) != 0) {
   2966 		rv = ENETRESET;
   2967 		goto out;
   2968 	} else if ((change & IFF_PROMISC) != 0) {
   2969 		rv = ixv_set_rxfilter(adapter);
   2970 		if (rv != 0) {
   2971 			/* Restore previous */
   2972 			adapter->if_flags = saved_flags;
   2973 			goto out;
   2974 		}
   2975 	}
   2976 
   2977 	/* Check for ec_capenable. */
   2978 	change = ec->ec_capenable ^ adapter->ec_capenable;
   2979 	adapter->ec_capenable = ec->ec_capenable;
   2980 	if ((change & ~(ETHERCAP_VLAN_MTU | ETHERCAP_VLAN_HWTAGGING
   2981 	    | ETHERCAP_VLAN_HWFILTER)) != 0) {
   2982 		rv = ENETRESET;
   2983 		goto out;
   2984 	}
   2985 
   2986 	/*
   2987 	 * Special handling is not required for ETHERCAP_VLAN_MTU.
   2988 	 * PF's MAXFRS(MHADD) does not include the 4bytes of the VLAN header.
   2989 	 */
   2990 
   2991 	/* Set up VLAN support and filter */
   2992 	if ((change & (ETHERCAP_VLAN_HWTAGGING | ETHERCAP_VLAN_HWFILTER)) != 0)
   2993 		rv = ixv_setup_vlan_support(adapter);
   2994 
   2995 out:
   2996 	IXGBE_CORE_UNLOCK(adapter);
   2997 
   2998 	return rv;
   2999 }
   3000 
   3001 
   3002 /************************************************************************
   3003  * ixv_ioctl - Ioctl entry point
   3004  *
   3005  *   Called when the user wants to configure the interface.
   3006  *
   3007  *   return 0 on success, positive on failure
   3008  ************************************************************************/
   3009 static int
   3010 ixv_ioctl(struct ifnet *ifp, u_long command, void *data)
   3011 {
   3012 	struct adapter	*adapter = ifp->if_softc;
   3013 	struct ixgbe_hw *hw = &adapter->hw;
   3014 	struct ifcapreq *ifcr = data;
   3015 	int		error;
   3016 	int l4csum_en;
   3017 	const int l4csum = IFCAP_CSUM_TCPv4_Rx | IFCAP_CSUM_UDPv4_Rx |
   3018 	     IFCAP_CSUM_TCPv6_Rx | IFCAP_CSUM_UDPv6_Rx;
   3019 
   3020 	switch (command) {
   3021 	case SIOCSIFFLAGS:
   3022 		IOCTL_DEBUGOUT("ioctl: SIOCSIFFLAGS (Set Interface Flags)");
   3023 		break;
   3024 	case SIOCADDMULTI: {
   3025 		struct ether_multi *enm;
   3026 		struct ether_multistep step;
   3027 		struct ethercom *ec = &adapter->osdep.ec;
   3028 		bool overflow = false;
   3029 		int mcnt = 0;
   3030 
   3031 		/*
   3032 		 * Check the number of multicast address. If it exceeds,
   3033 		 * return ENOSPC.
   3034 		 * Update this code when we support API 1.3.
   3035 		 */
   3036 		ETHER_LOCK(ec);
   3037 		ETHER_FIRST_MULTI(step, ec, enm);
   3038 		while (enm != NULL) {
   3039 			mcnt++;
   3040 
   3041 			/*
   3042 			 * This code is before adding, so one room is required
   3043 			 * at least.
   3044 			 */
   3045 			if (mcnt > (IXGBE_MAX_VF_MC - 1)) {
   3046 				overflow = true;
   3047 				break;
   3048 			}
   3049 			ETHER_NEXT_MULTI(step, enm);
   3050 		}
   3051 		ETHER_UNLOCK(ec);
   3052 		error = 0;
   3053 		if (overflow && ((ec->ec_flags & ETHER_F_ALLMULTI) == 0)) {
   3054 			error = hw->mac.ops.update_xcast_mode(hw,
   3055 			    IXGBEVF_XCAST_MODE_ALLMULTI);
   3056 			if (error == IXGBE_ERR_NOT_TRUSTED) {
   3057 				device_printf(adapter->dev,
   3058 				    "this interface is not trusted\n");
   3059 				error = EPERM;
   3060 			} else if (error == IXGBE_ERR_FEATURE_NOT_SUPPORTED) {
   3061 				device_printf(adapter->dev,
   3062 				    "the PF doesn't support allmulti mode\n");
   3063 				error = EOPNOTSUPP;
   3064 			} else if (error) {
   3065 				device_printf(adapter->dev,
   3066 				    "number of Ethernet multicast addresses "
   3067 				    "exceeds the limit (%d). error = %d\n",
   3068 				    IXGBE_MAX_VF_MC, error);
   3069 				error = ENOSPC;
   3070 			} else
   3071 				ec->ec_flags |= ETHER_F_ALLMULTI;
   3072 		}
   3073 		if (error)
   3074 			return error;
   3075 	}
   3076 		/*FALLTHROUGH*/
   3077 	case SIOCDELMULTI:
   3078 		IOCTL_DEBUGOUT("ioctl: SIOC(ADD|DEL)MULTI");
   3079 		break;
   3080 	case SIOCSIFMEDIA:
   3081 	case SIOCGIFMEDIA:
   3082 		IOCTL_DEBUGOUT("ioctl: SIOCxIFMEDIA (Get/Set Interface Media)");
   3083 		break;
   3084 	case SIOCSIFCAP:
   3085 		IOCTL_DEBUGOUT("ioctl: SIOCSIFCAP (Set Capabilities)");
   3086 		break;
   3087 	case SIOCSIFMTU:
   3088 		IOCTL_DEBUGOUT("ioctl: SIOCSIFMTU (Set Interface MTU)");
   3089 		break;
   3090 	case SIOCZIFDATA:
   3091 		IOCTL_DEBUGOUT("ioctl: SIOCZIFDATA (Zero counter)");
   3092 		ixv_update_stats(adapter);
   3093 		ixv_clear_evcnt(adapter);
   3094 		break;
   3095 	default:
   3096 		IOCTL_DEBUGOUT1("ioctl: UNKNOWN (0x%X)", (int)command);
   3097 		break;
   3098 	}
   3099 
   3100 	switch (command) {
   3101 	case SIOCSIFCAP:
   3102 		/* Layer-4 Rx checksum offload has to be turned on and
   3103 		 * off as a unit.
   3104 		 */
   3105 		l4csum_en = ifcr->ifcr_capenable & l4csum;
   3106 		if (l4csum_en != l4csum && l4csum_en != 0)
   3107 			return EINVAL;
   3108 		/*FALLTHROUGH*/
   3109 	case SIOCADDMULTI:
   3110 	case SIOCDELMULTI:
   3111 	case SIOCSIFFLAGS:
   3112 	case SIOCSIFMTU:
   3113 	default:
   3114 		if ((error = ether_ioctl(ifp, command, data)) != ENETRESET)
   3115 			return error;
   3116 		if ((ifp->if_flags & IFF_RUNNING) == 0)
   3117 			;
   3118 		else if (command == SIOCSIFCAP || command == SIOCSIFMTU) {
   3119 			IXGBE_CORE_LOCK(adapter);
   3120 			ixv_init_locked(adapter);
   3121 			IXGBE_CORE_UNLOCK(adapter);
   3122 		} else if (command == SIOCADDMULTI || command == SIOCDELMULTI) {
   3123 			/*
   3124 			 * Multicast list has changed; set the hardware filter
   3125 			 * accordingly.
   3126 			 */
   3127 			IXGBE_CORE_LOCK(adapter);
   3128 			ixv_disable_intr(adapter);
   3129 			ixv_set_rxfilter(adapter);
   3130 			ixv_enable_intr(adapter);
   3131 			IXGBE_CORE_UNLOCK(adapter);
   3132 		}
   3133 		return 0;
   3134 	}
   3135 } /* ixv_ioctl */
   3136 
   3137 /************************************************************************
   3138  * ixv_init
   3139  ************************************************************************/
   3140 static int
   3141 ixv_init(struct ifnet *ifp)
   3142 {
   3143 	struct adapter *adapter = ifp->if_softc;
   3144 
   3145 	IXGBE_CORE_LOCK(adapter);
   3146 	ixv_init_locked(adapter);
   3147 	IXGBE_CORE_UNLOCK(adapter);
   3148 
   3149 	return 0;
   3150 } /* ixv_init */
   3151 
   3152 /************************************************************************
   3153  * ixv_handle_que
   3154  ************************************************************************/
   3155 static void
   3156 ixv_handle_que(void *context)
   3157 {
   3158 	struct ix_queue *que = context;
   3159 	struct adapter	*adapter = que->adapter;
   3160 	struct tx_ring	*txr = que->txr;
   3161 	struct ifnet	*ifp = adapter->ifp;
   3162 	bool		more;
   3163 
   3164 	que->handleq.ev_count++;
   3165 
   3166 	if (ifp->if_flags & IFF_RUNNING) {
   3167 		more = ixgbe_rxeof(que);
   3168 		IXGBE_TX_LOCK(txr);
   3169 		more |= ixgbe_txeof(txr);
   3170 		if (!(adapter->feat_en & IXGBE_FEATURE_LEGACY_TX))
   3171 			if (!ixgbe_mq_ring_empty(ifp, txr->txr_interq))
   3172 				ixgbe_mq_start_locked(ifp, txr);
   3173 		/* Only for queue 0 */
   3174 		/* NetBSD still needs this for CBQ */
   3175 		if ((&adapter->queues[0] == que)
   3176 		    && (!ixgbe_legacy_ring_empty(ifp, NULL)))
   3177 			ixgbe_legacy_start_locked(ifp, txr);
   3178 		IXGBE_TX_UNLOCK(txr);
   3179 		if (more) {
   3180 			que->req.ev_count++;
   3181 			if (adapter->txrx_use_workqueue) {
   3182 				/*
   3183 				 * "enqueued flag" is not required here
   3184 				 * the same as ixg(4). See ixgbe_msix_que().
   3185 				 */
   3186 				workqueue_enqueue(adapter->que_wq,
   3187 				    &que->wq_cookie, curcpu());
   3188 			} else
   3189 				  softint_schedule(que->que_si);
   3190 			return;
   3191 		}
   3192 	}
   3193 
   3194 	/* Re-enable this interrupt */
   3195 	ixv_enable_queue(adapter, que->msix);
   3196 
   3197 	return;
   3198 } /* ixv_handle_que */
   3199 
   3200 /************************************************************************
   3201  * ixv_handle_que_work
   3202  ************************************************************************/
   3203 static void
   3204 ixv_handle_que_work(struct work *wk, void *context)
   3205 {
   3206 	struct ix_queue *que = container_of(wk, struct ix_queue, wq_cookie);
   3207 
   3208 	/*
   3209 	 * "enqueued flag" is not required here the same as ixg(4).
   3210 	 * See ixgbe_msix_que().
   3211 	 */
   3212 	ixv_handle_que(que);
   3213 }
   3214 
   3215 /************************************************************************
   3216  * ixv_allocate_msix - Setup MSI-X Interrupt resources and handlers
   3217  ************************************************************************/
   3218 static int
   3219 ixv_allocate_msix(struct adapter *adapter, const struct pci_attach_args *pa)
   3220 {
   3221 	device_t	dev = adapter->dev;
   3222 	struct ix_queue *que = adapter->queues;
   3223 	struct		tx_ring *txr = adapter->tx_rings;
   3224 	int		error, msix_ctrl, rid, vector = 0;
   3225 	pci_chipset_tag_t pc;
   3226 	pcitag_t	tag;
   3227 	char		intrbuf[PCI_INTRSTR_LEN];
   3228 	char		wqname[MAXCOMLEN];
   3229 	char		intr_xname[32];
   3230 	const char	*intrstr = NULL;
   3231 	kcpuset_t	*affinity;
   3232 	int		cpu_id = 0;
   3233 
   3234 	pc = adapter->osdep.pc;
   3235 	tag = adapter->osdep.tag;
   3236 
   3237 	adapter->osdep.nintrs = adapter->num_queues + 1;
   3238 	if (pci_msix_alloc_exact(pa, &adapter->osdep.intrs,
   3239 	    adapter->osdep.nintrs) != 0) {
   3240 		aprint_error_dev(dev,
   3241 		    "failed to allocate MSI-X interrupt\n");
   3242 		return (ENXIO);
   3243 	}
   3244 
   3245 	kcpuset_create(&affinity, false);
   3246 	for (int i = 0; i < adapter->num_queues; i++, vector++, que++, txr++) {
   3247 		snprintf(intr_xname, sizeof(intr_xname), "%s TXRX%d",
   3248 		    device_xname(dev), i);
   3249 		intrstr = pci_intr_string(pc, adapter->osdep.intrs[i], intrbuf,
   3250 		    sizeof(intrbuf));
   3251 #ifdef IXGBE_MPSAFE
   3252 		pci_intr_setattr(pc, &adapter->osdep.intrs[i], PCI_INTR_MPSAFE,
   3253 		    true);
   3254 #endif
   3255 		/* Set the handler function */
   3256 		que->res = adapter->osdep.ihs[i] = pci_intr_establish_xname(pc,
   3257 		    adapter->osdep.intrs[i], IPL_NET, ixv_msix_que, que,
   3258 		    intr_xname);
   3259 		if (que->res == NULL) {
   3260 			pci_intr_release(pc, adapter->osdep.intrs,
   3261 			    adapter->osdep.nintrs);
   3262 			aprint_error_dev(dev,
   3263 			    "Failed to register QUE handler\n");
   3264 			kcpuset_destroy(affinity);
   3265 			return (ENXIO);
   3266 		}
   3267 		que->msix = vector;
   3268 		adapter->active_queues |= (u64)(1 << que->msix);
   3269 
   3270 		cpu_id = i;
   3271 		/* Round-robin affinity */
   3272 		kcpuset_zero(affinity);
   3273 		kcpuset_set(affinity, cpu_id % ncpu);
   3274 		error = interrupt_distribute(adapter->osdep.ihs[i], affinity,
   3275 		    NULL);
   3276 		aprint_normal_dev(dev, "for TX/RX, interrupting at %s",
   3277 		    intrstr);
   3278 		if (error == 0)
   3279 			aprint_normal(", bound queue %d to cpu %d\n",
   3280 			    i, cpu_id % ncpu);
   3281 		else
   3282 			aprint_normal("\n");
   3283 
   3284 #ifndef IXGBE_LEGACY_TX
   3285 		txr->txr_si
   3286 		    = softint_establish(SOFTINT_NET | IXGBE_SOFTINFT_FLAGS,
   3287 			ixgbe_deferred_mq_start, txr);
   3288 #endif
   3289 		que->que_si
   3290 		    = softint_establish(SOFTINT_NET | IXGBE_SOFTINFT_FLAGS,
   3291 			ixv_handle_que, que);
   3292 		if (que->que_si == NULL) {
   3293 			aprint_error_dev(dev,
   3294 			    "could not establish software interrupt\n");
   3295 		}
   3296 	}
   3297 	snprintf(wqname, sizeof(wqname), "%sdeferTx", device_xname(dev));
   3298 	error = workqueue_create(&adapter->txr_wq, wqname,
   3299 	    ixgbe_deferred_mq_start_work, adapter, IXGBE_WORKQUEUE_PRI, IPL_NET,
   3300 	    IXGBE_WORKQUEUE_FLAGS);
   3301 	if (error) {
   3302 		aprint_error_dev(dev, "couldn't create workqueue for deferred Tx\n");
   3303 	}
   3304 	adapter->txr_wq_enqueued = percpu_alloc(sizeof(u_int));
   3305 
   3306 	snprintf(wqname, sizeof(wqname), "%sTxRx", device_xname(dev));
   3307 	error = workqueue_create(&adapter->que_wq, wqname,
   3308 	    ixv_handle_que_work, adapter, IXGBE_WORKQUEUE_PRI, IPL_NET,
   3309 	    IXGBE_WORKQUEUE_FLAGS);
   3310 	if (error) {
   3311 		aprint_error_dev(dev,
   3312 		    "couldn't create workqueue\n");
   3313 	}
   3314 
   3315 	/* and Mailbox */
   3316 	cpu_id++;
   3317 	snprintf(intr_xname, sizeof(intr_xname), "%s link", device_xname(dev));
   3318 	adapter->vector = vector;
   3319 	intrstr = pci_intr_string(pc, adapter->osdep.intrs[vector], intrbuf,
   3320 	    sizeof(intrbuf));
   3321 #ifdef IXGBE_MPSAFE
   3322 	pci_intr_setattr(pc, &adapter->osdep.intrs[vector], PCI_INTR_MPSAFE,
   3323 	    true);
   3324 #endif
   3325 	/* Set the mbx handler function */
   3326 	adapter->osdep.ihs[vector] = pci_intr_establish_xname(pc,
   3327 	    adapter->osdep.intrs[vector], IPL_NET, ixv_msix_mbx, adapter,
   3328 	    intr_xname);
   3329 	if (adapter->osdep.ihs[vector] == NULL) {
   3330 		aprint_error_dev(dev, "Failed to register LINK handler\n");
   3331 		kcpuset_destroy(affinity);
   3332 		return (ENXIO);
   3333 	}
   3334 	/* Round-robin affinity */
   3335 	kcpuset_zero(affinity);
   3336 	kcpuset_set(affinity, cpu_id % ncpu);
   3337 	error = interrupt_distribute(adapter->osdep.ihs[vector],
   3338 	    affinity, NULL);
   3339 
   3340 	aprint_normal_dev(dev,
   3341 	    "for link, interrupting at %s", intrstr);
   3342 	if (error == 0)
   3343 		aprint_normal(", affinity to cpu %d\n", cpu_id % ncpu);
   3344 	else
   3345 		aprint_normal("\n");
   3346 
   3347 	/* Tasklets for Mailbox */
   3348 	adapter->link_si = softint_establish(SOFTINT_NET |IXGBE_SOFTINFT_FLAGS,
   3349 	    ixv_handle_link, adapter);
   3350 	/*
   3351 	 * Due to a broken design QEMU will fail to properly
   3352 	 * enable the guest for MSI-X unless the vectors in
   3353 	 * the table are all set up, so we must rewrite the
   3354 	 * ENABLE in the MSI-X control register again at this
   3355 	 * point to cause it to successfully initialize us.
   3356 	 */
   3357 	if (adapter->hw.mac.type == ixgbe_mac_82599_vf) {
   3358 		pci_get_capability(pc, tag, PCI_CAP_MSIX, &rid, NULL);
   3359 		rid += PCI_MSIX_CTL;
   3360 		msix_ctrl = pci_conf_read(pc, tag, rid);
   3361 		msix_ctrl |= PCI_MSIX_CTL_ENABLE;
   3362 		pci_conf_write(pc, tag, rid, msix_ctrl);
   3363 	}
   3364 
   3365 	kcpuset_destroy(affinity);
   3366 	return (0);
   3367 } /* ixv_allocate_msix */
   3368 
   3369 /************************************************************************
   3370  * ixv_configure_interrupts - Setup MSI-X resources
   3371  *
   3372  *   Note: The VF device MUST use MSI-X, there is no fallback.
   3373  ************************************************************************/
   3374 static int
   3375 ixv_configure_interrupts(struct adapter *adapter)
   3376 {
   3377 	device_t dev = adapter->dev;
   3378 	int want, queues, msgs;
   3379 
   3380 	/* Must have at least 2 MSI-X vectors */
   3381 	msgs = pci_msix_count(adapter->osdep.pc, adapter->osdep.tag);
   3382 	if (msgs < 2) {
   3383 		aprint_error_dev(dev, "MSIX config error\n");
   3384 		return (ENXIO);
   3385 	}
   3386 	msgs = MIN(msgs, IXG_MAX_NINTR);
   3387 
   3388 	/* Figure out a reasonable auto config value */
   3389 	queues = (ncpu > (msgs - 1)) ? (msgs - 1) : ncpu;
   3390 
   3391 	if (ixv_num_queues != 0)
   3392 		queues = ixv_num_queues;
   3393 	else if ((ixv_num_queues == 0) && (queues > IXGBE_VF_MAX_TX_QUEUES))
   3394 		queues = IXGBE_VF_MAX_TX_QUEUES;
   3395 
   3396 	/*
   3397 	 * Want vectors for the queues,
   3398 	 * plus an additional for mailbox.
   3399 	 */
   3400 	want = queues + 1;
   3401 	if (msgs >= want)
   3402 		msgs = want;
   3403 	else {
   3404 		aprint_error_dev(dev,
   3405 		    "MSI-X Configuration Problem, "
   3406 		    "%d vectors but %d queues wanted!\n",
   3407 		    msgs, want);
   3408 		return -1;
   3409 	}
   3410 
   3411 	adapter->msix_mem = (void *)1; /* XXX */
   3412 	aprint_normal_dev(dev,
   3413 	    "Using MSI-X interrupts with %d vectors\n", msgs);
   3414 	adapter->num_queues = queues;
   3415 
   3416 	return (0);
   3417 } /* ixv_configure_interrupts */
   3418 
   3419 
   3420 /************************************************************************
   3421  * ixv_handle_link - Tasklet handler for MSI-X MBX interrupts
   3422  *
   3423  *   Done outside of interrupt context since the driver might sleep
   3424  ************************************************************************/
   3425 static void
   3426 ixv_handle_link(void *context)
   3427 {
   3428 	struct adapter *adapter = context;
   3429 
   3430 	IXGBE_CORE_LOCK(adapter);
   3431 
   3432 	adapter->hw.mac.ops.check_link(&adapter->hw, &adapter->link_speed,
   3433 	    &adapter->link_up, FALSE);
   3434 	ixv_update_link_status(adapter);
   3435 
   3436 	IXGBE_CORE_UNLOCK(adapter);
   3437 } /* ixv_handle_link */
   3438 
   3439 /************************************************************************
   3440  * ixv_check_link - Used in the local timer to poll for link changes
   3441  ************************************************************************/
   3442 static s32
   3443 ixv_check_link(struct adapter *adapter)
   3444 {
   3445 	s32 error;
   3446 
   3447 	KASSERT(mutex_owned(&adapter->core_mtx));
   3448 
   3449 	adapter->hw.mac.get_link_status = TRUE;
   3450 
   3451 	error = adapter->hw.mac.ops.check_link(&adapter->hw,
   3452 	    &adapter->link_speed, &adapter->link_up, FALSE);
   3453 	ixv_update_link_status(adapter);
   3454 
   3455 	return error;
   3456 } /* ixv_check_link */
   3457