Lines Matching defs:txr
607 struct tx_ring *txr = sc->tx_rings;
652 txr = sc->tx_rings;
653 for (int i = 0; i < sc->num_queues; i++, rxr++, txr++) {
657 evcnt_detach(&txr->total_packets);
659 evcnt_detach(&txr->pcq_drops);
661 evcnt_detach(&txr->no_desc_avail);
662 evcnt_detach(&txr->tso_tx);
886 struct tx_ring *txr = que->txr;
901 IXGBE_TX_LOCK(txr);
902 ixgbe_txeof(txr);
903 IXGBE_TX_UNLOCK(txr);
921 if ((txr->bytes == 0) && (rxr->bytes == 0))
924 if ((txr->bytes) && (txr->packets))
925 newitr = txr->bytes/txr->packets;
955 txr->bytes = 0;
956 txr->packets = 0;
1306 struct tx_ring *txr = que->txr;
1308 v0 += txr->q_efbig_tx_dma_setup;
1309 v1 += txr->q_mbuf_defrag_failed;
1310 v2 += txr->q_efbig2_tx_dma_setup;
1311 v3 += txr->q_einval_tx_dma_setup;
1312 v4 += txr->q_other_tx_dma_setup;
1313 v5 += txr->q_eagain_tx_dma_setup;
1314 v6 += txr->q_enomem_tx_dma_setup;
1315 v7 += txr->q_tso_err;
1334 if (que->txr->busy)
1354 que->txr->busy = IXGBE_QUEUE_HUNG;
1557 struct tx_ring *txr = sc->tx_rings;
1560 for (i = 0; i < sc->num_queues; i++, que++, txr++) {
1562 if (txr->txr_si != NULL)
1563 softint_disestablish(txr->txr_si);
1724 struct tx_ring *txr = sc->tx_rings;
1728 for (i = 0; i < sc->num_queues; i++, txr++) {
1729 u64 tdba = txr->txdma.dma_paddr;
1731 int j = txr->me;
1744 txr->tail = IXGBE_VFTDT(j);
1746 txr->txr_no_space = false;
1994 struct tx_ring *txr = (struct tx_ring *)node.sysctl_data;
1997 if (!txr)
2000 val = IXGBE_READ_REG(&txr->sc->hw, IXGBE_VFTDH(txr->me));
2014 struct tx_ring *txr = (struct tx_ring *)node.sysctl_data;
2017 if (!txr)
2020 val = IXGBE_READ_REG(&txr->sc->hw, IXGBE_VFTDT(txr->me));
2625 struct tx_ring *txr = sc->tx_rings;
2657 for (int i = 0; i < sc->num_queues; i++, rxr++, txr++) {
2689 ixv_sysctl_tdh_handler, 0, (void *)txr,
2696 ixv_sysctl_tdt_handler, 0, (void *)txr,
2735 evcnt_attach_dynamic(&txr->total_packets, EVCNT_TYPE_MISC,
2739 evcnt_attach_dynamic(&txr->pcq_drops, EVCNT_TYPE_MISC,
2743 evcnt_attach_dynamic(&txr->no_desc_avail, EVCNT_TYPE_MISC,
2746 evcnt_attach_dynamic(&txr->tso_tx, EVCNT_TYPE_MISC,
2814 struct tx_ring *txr = sc->tx_rings;
2833 for (i = 0; i < sc->num_queues; i++, rxr++, txr++) {
2837 IXGBE_EVC_STORE(&txr->total_packets, 0);
2839 IXGBE_EVC_STORE(&txr->pcq_drops, 0);
2841 IXGBE_EVC_STORE(&txr->no_desc_avail, 0);
2842 IXGBE_EVC_STORE(&txr->tso_tx, 0);
2843 txr->q_efbig_tx_dma_setup = 0;
2844 txr->q_mbuf_defrag_failed = 0;
2845 txr->q_efbig2_tx_dma_setup = 0;
2846 txr->q_einval_tx_dma_setup = 0;
2847 txr->q_other_tx_dma_setup = 0;
2848 txr->q_eagain_tx_dma_setup = 0;
2849 txr->q_enomem_tx_dma_setup = 0;
2850 txr->q_tso_err = 0;
3303 struct tx_ring *txr = que->txr;
3310 IXGBE_TX_LOCK(txr);
3311 more = ixgbe_txeof(txr);
3313 if (!ixgbe_mq_ring_empty(ifp, txr->txr_interq))
3314 ixgbe_mq_start_locked(ifp, txr);
3319 ixgbe_legacy_start_locked(ifp, txr);
3320 IXGBE_TX_UNLOCK(txr);
3366 struct tx_ring *txr = sc->tx_rings;
3389 for (int i = 0; i < sc->num_queues; i++, vector++, que++, txr++) {
3426 txr->txr_si
3428 ixgbe_deferred_mq_start, txr);