net/bnx2x: restructure Tx routine

- Process Tx completions based on configured Tx free threshold and
  determine how much TX BDs are required before invoking bnx2x_tx_encap()
- Change bnx2x_tx_encap() to void function as it can now never fail

Signed-off-by: Rasesh Mody <rasesh.mody@qlogic.com>
Signed-off-by: Harish Patil <harish.patil@qlogic.com>
This commit is contained in:
Rasesh Mody 2016-05-11 17:06:23 -07:00 committed by Bruce Richardson
parent ec9c605fd3
commit 827ed2a118
3 changed files with 26 additions and 32 deletions

View File

@ -2113,10 +2113,12 @@ bnx2x_nic_unload(struct bnx2x_softc *sc, uint32_t unload_mode, uint8_t keep_link
* the mbuf and return to the caller.
*
* Returns:
* 0 = Success, !0 = Failure
* void.
*
* Note the side effect that an mbuf may be freed if it causes a problem.
*/
int bnx2x_tx_encap(struct bnx2x_tx_queue *txq, struct rte_mbuf **m_head, int m_pkts)
void bnx2x_tx_encap(struct bnx2x_tx_queue *txq, struct rte_mbuf **m_head,
int m_pkts)
{
struct rte_mbuf *m0;
struct eth_tx_start_bd *tx_start_bd;
@ -2136,12 +2138,6 @@ int bnx2x_tx_encap(struct bnx2x_tx_queue *txq, struct rte_mbuf **m_head, int m_p
m0 = *m_head++;
if (unlikely(txq->nb_tx_avail < 3)) {
PMD_TX_LOG(ERR, "no enough bds %d/%d",
bd_prod, txq->nb_tx_avail);
return -ENOMEM;
}
txq->sw_ring[TX_BD(pkt_prod, txq)] = m0;
tx_start_bd = &txq->tx_ring[TX_BD(bd_prod, txq)].start_bd;
@ -2252,8 +2248,6 @@ int bnx2x_tx_encap(struct bnx2x_tx_queue *txq, struct rte_mbuf **m_head, int m_p
fp->tx_db.data.prod += (m_pkts << 1) + nbds;
DOORBELL(sc, txq->queue_id, fp->tx_db.raw);
mb();
return 0;
}
static uint16_t bnx2x_cid_ilt_lines(struct bnx2x_softc *sc)

View File

@ -167,6 +167,8 @@ struct bnx2x_device_type {
#define TX_PAGE(x) (((x) & ~USABLE_TX_BD_PER_PAGE) >> 8)
#define TX_IDX(x) ((x) & USABLE_TX_BD_PER_PAGE)
#define BDS_PER_TX_PKT (3)
/*
* Trigger pending transmits when the number of available BDs is greater
* than 1/8 of the total number of usable BDs.
@ -1864,7 +1866,8 @@ int bnx2x_alloc_hsi_mem(struct bnx2x_softc *sc);
int bnx2x_alloc_ilt_mem(struct bnx2x_softc *sc);
void bnx2x_free_ilt_mem(struct bnx2x_softc *sc);
void bnx2x_dump_tx_chain(struct bnx2x_fastpath * fp, int bd_prod, int count);
int bnx2x_tx_encap(struct bnx2x_tx_queue *txq, struct rte_mbuf **m_head, int m_pkts);
void bnx2x_tx_encap(struct bnx2x_tx_queue *txq, struct rte_mbuf **m_head,
int m_pkts);
uint8_t bnx2x_txeof(struct bnx2x_softc *sc, struct bnx2x_fastpath *fp);
void bnx2x_print_adapter_info(struct bnx2x_softc *sc);
int bnx2x_intr_legacy(struct bnx2x_softc *sc, int scan_fp);

View File

@ -211,40 +211,37 @@ bnx2x_xmit_pkts(void *p_txq, struct rte_mbuf **tx_pkts, uint16_t nb_pkts)
struct bnx2x_tx_queue *txq;
struct bnx2x_softc *sc;
struct bnx2x_fastpath *fp;
uint32_t burst, nb_tx;
uint32_t burst;
struct rte_mbuf **m = tx_pkts;
int ret;
uint16_t nb_tx_pkts;
uint16_t nb_pkt_sent = 0;
txq = p_txq;
sc = txq->sc;
fp = &sc->fp[txq->queue_id];
nb_tx = nb_pkts;
if ((unlikely((txq->nb_tx_desc - txq->nb_tx_avail) >
txq->tx_free_thresh)))
bnx2x_txeof(sc, fp);
do {
burst = RTE_MIN(nb_pkts, RTE_PMD_BNX2X_TX_MAX_BURST);
nb_tx_pkts = RTE_MIN(nb_pkts, txq->nb_tx_avail / BDS_PER_TX_PKT);
if (unlikely(nb_tx_pkts == 0))
return 0;
ret = bnx2x_tx_encap(txq, m, burst);
if (unlikely(ret)) {
PMD_TX_LOG(ERR, "tx_encap failed!");
}
burst = RTE_MIN(nb_tx_pkts, RTE_PMD_BNX2X_TX_MAX_BURST);
while (nb_tx_pkts--) {
assert(m != NULL);
bnx2x_tx_encap(txq, m, burst);
bnx2x_update_fp_sb_idx(fp);
if ((txq->nb_tx_desc - txq->nb_tx_avail) > txq->tx_free_thresh) {
if ((txq->nb_tx_desc - txq->nb_tx_avail) >
txq->tx_free_thresh)
bnx2x_txeof(sc, fp);
}
if (unlikely(ret == -ENOMEM)) {
break;
}
m += burst;
nb_pkts -= burst;
nb_pkt_sent++;
}
} while (nb_pkts);
return nb_tx - nb_pkts;
return nb_pkt_sent;
}
int