net/sfc: implement representor Tx
Forward traffic that is transmitted from a port representor to the corresponding virtual function using the dedicated TxQ. Signed-off-by: Igor Romanov <igor.romanov@oktetlabs.ru> Signed-off-by: Andrew Rybchenko <andrew.rybchenko@oktetlabs.ru> Reviewed-by: Andy Moreton <amoreton@xilinx.com> Reviewed-by: Ivan Malov <ivan.malov@oktetlabs.ru>
This commit is contained in:
parent
472fa1b17c
commit
788dedcfcb
@ -160,6 +160,49 @@ sfc_repr_tx_queue_stop(void *queue)
|
||||
rte_ring_reset(txq->ring);
|
||||
}
|
||||
|
||||
static uint16_t
|
||||
sfc_repr_tx_burst(void *tx_queue, struct rte_mbuf **tx_pkts, uint16_t nb_pkts)
|
||||
{
|
||||
struct sfc_repr_txq *txq = tx_queue;
|
||||
unsigned int n_tx;
|
||||
void **objs;
|
||||
uint16_t i;
|
||||
|
||||
/*
|
||||
* mbuf is likely cache-hot. Set flag and egress m-port here instead of
|
||||
* doing that in representors proxy. Also, it should help to avoid
|
||||
* cache bounce. Moreover, potentially, it allows to use one
|
||||
* multi-producer single-consumer ring for all representors.
|
||||
*
|
||||
* The only potential problem is doing so many times if enqueue
|
||||
* fails and sender retries.
|
||||
*/
|
||||
for (i = 0; i < nb_pkts; ++i) {
|
||||
struct rte_mbuf *m = tx_pkts[i];
|
||||
|
||||
m->ol_flags |= sfc_dp_mport_override;
|
||||
*RTE_MBUF_DYNFIELD(m, sfc_dp_mport_offset,
|
||||
efx_mport_id_t *) = txq->egress_mport;
|
||||
}
|
||||
|
||||
objs = (void *)&tx_pkts[0];
|
||||
n_tx = rte_ring_sp_enqueue_burst(txq->ring, objs, nb_pkts, NULL);
|
||||
|
||||
/*
|
||||
* Remove m-port override flag from packets that were not enqueued
|
||||
* Setting the flag only for enqueued packets after the burst is
|
||||
* not possible since the ownership of enqueued packets is
|
||||
* transferred to representor proxy.
|
||||
*/
|
||||
for (i = n_tx; i < nb_pkts; ++i) {
|
||||
struct rte_mbuf *m = tx_pkts[i];
|
||||
|
||||
m->ol_flags &= ~sfc_dp_mport_override;
|
||||
}
|
||||
|
||||
return n_tx;
|
||||
}
|
||||
|
||||
static int
|
||||
sfc_repr_start(struct rte_eth_dev *dev)
|
||||
{
|
||||
@ -760,6 +803,7 @@ sfc_repr_dev_close(struct rte_eth_dev *dev)
|
||||
|
||||
(void)sfc_repr_proxy_del_port(srs->pf_port_id, srs->repr_id);
|
||||
|
||||
dev->tx_pkt_burst = NULL;
|
||||
dev->dev_ops = NULL;
|
||||
|
||||
sfc_repr_unlock(sr);
|
||||
@ -880,6 +924,7 @@ sfc_repr_eth_dev_init(struct rte_eth_dev *dev, void *init_params)
|
||||
goto fail_mac_addrs;
|
||||
}
|
||||
|
||||
dev->tx_pkt_burst = sfc_repr_tx_burst;
|
||||
dev->dev_ops = &sfc_repr_dev_ops;
|
||||
|
||||
sr->state = SFC_ETHDEV_INITIALIZED;
|
||||
|
@ -25,6 +25,12 @@
|
||||
*/
|
||||
#define SFC_REPR_PROXY_MBOX_POLL_TIMEOUT_MS 1000
|
||||
|
||||
/**
|
||||
* Amount of time to wait for the representor proxy routine (which is
|
||||
* running on a service core) to terminate after service core is stopped.
|
||||
*/
|
||||
#define SFC_REPR_PROXY_ROUTINE_TERMINATE_TIMEOUT_MS 10000
|
||||
|
||||
static struct sfc_repr_proxy *
|
||||
sfc_repr_proxy_by_adapter(struct sfc_adapter *sa)
|
||||
{
|
||||
@ -148,16 +154,71 @@ sfc_repr_proxy_mbox_handle(struct sfc_repr_proxy *rp)
|
||||
__atomic_store_n(&mbox->ack, true, __ATOMIC_RELEASE);
|
||||
}
|
||||
|
||||
static void
|
||||
sfc_repr_proxy_handle_tx(struct sfc_repr_proxy_dp_txq *rp_txq,
|
||||
struct sfc_repr_proxy_txq *repr_txq)
|
||||
{
|
||||
/*
|
||||
* With multiple representor proxy queues configured it is
|
||||
* possible that not all of the corresponding representor
|
||||
* queues were created. Skip the queues that do not exist.
|
||||
*/
|
||||
if (repr_txq->ring == NULL)
|
||||
return;
|
||||
|
||||
if (rp_txq->available < RTE_DIM(rp_txq->tx_pkts)) {
|
||||
rp_txq->available +=
|
||||
rte_ring_sc_dequeue_burst(repr_txq->ring,
|
||||
(void **)(&rp_txq->tx_pkts[rp_txq->available]),
|
||||
RTE_DIM(rp_txq->tx_pkts) - rp_txq->available,
|
||||
NULL);
|
||||
|
||||
if (rp_txq->available == rp_txq->transmitted)
|
||||
return;
|
||||
}
|
||||
|
||||
rp_txq->transmitted += rp_txq->pkt_burst(rp_txq->dp,
|
||||
&rp_txq->tx_pkts[rp_txq->transmitted],
|
||||
rp_txq->available - rp_txq->transmitted);
|
||||
|
||||
if (rp_txq->available == rp_txq->transmitted) {
|
||||
rp_txq->available = 0;
|
||||
rp_txq->transmitted = 0;
|
||||
}
|
||||
}
|
||||
|
||||
static int32_t
|
||||
sfc_repr_proxy_routine(void *arg)
|
||||
{
|
||||
struct sfc_repr_proxy_port *port;
|
||||
struct sfc_repr_proxy *rp = arg;
|
||||
unsigned int i;
|
||||
|
||||
sfc_repr_proxy_mbox_handle(rp);
|
||||
|
||||
TAILQ_FOREACH(port, &rp->ports, entries) {
|
||||
if (!port->started)
|
||||
continue;
|
||||
|
||||
for (i = 0; i < rp->nb_txq; i++)
|
||||
sfc_repr_proxy_handle_tx(&rp->dp_txq[i], &port->txq[i]);
|
||||
}
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static struct sfc_txq_info *
|
||||
sfc_repr_proxy_txq_info_get(struct sfc_adapter *sa, unsigned int repr_queue_id)
|
||||
{
|
||||
struct sfc_adapter_shared *sas = sfc_sa2shared(sa);
|
||||
struct sfc_repr_proxy_dp_txq *dp_txq;
|
||||
|
||||
SFC_ASSERT(repr_queue_id < sfc_repr_nb_txq(sas));
|
||||
dp_txq = &sa->repr_proxy.dp_txq[repr_queue_id];
|
||||
|
||||
return &sas->txq_info[dp_txq->sw_index];
|
||||
}
|
||||
|
||||
static int
|
||||
sfc_repr_proxy_txq_attach(struct sfc_adapter *sa)
|
||||
{
|
||||
@ -289,11 +350,20 @@ sfc_repr_proxy_txq_fini(struct sfc_adapter *sa)
|
||||
static int
|
||||
sfc_repr_proxy_txq_start(struct sfc_adapter *sa)
|
||||
{
|
||||
struct sfc_adapter_shared * const sas = sfc_sa2shared(sa);
|
||||
struct sfc_repr_proxy *rp = &sa->repr_proxy;
|
||||
unsigned int i;
|
||||
|
||||
sfc_log_init(sa, "entry");
|
||||
|
||||
RTE_SET_USED(rp);
|
||||
for (i = 0; i < sfc_repr_nb_txq(sas); i++) {
|
||||
struct sfc_repr_proxy_dp_txq *txq = &rp->dp_txq[i];
|
||||
|
||||
txq->dp = sfc_repr_proxy_txq_info_get(sa, i)->dp;
|
||||
txq->pkt_burst = sa->eth_dev->tx_pkt_burst;
|
||||
txq->available = 0;
|
||||
txq->transmitted = 0;
|
||||
}
|
||||
|
||||
sfc_log_init(sa, "done");
|
||||
|
||||
@ -922,6 +992,8 @@ sfc_repr_proxy_start(struct sfc_adapter *sa)
|
||||
if (rc != 0)
|
||||
goto fail_txq_start;
|
||||
|
||||
rp->nb_txq = sfc_repr_nb_txq(sas);
|
||||
|
||||
/* Service core may be in "stopped" state, start it */
|
||||
rc = rte_service_lcore_start(rp->service_core_id);
|
||||
if (rc != 0 && rc != -EALREADY) {
|
||||
@ -1007,6 +1079,9 @@ sfc_repr_proxy_stop(struct sfc_adapter *sa)
|
||||
struct sfc_adapter_shared * const sas = sfc_sa2shared(sa);
|
||||
struct sfc_repr_proxy *rp = &sa->repr_proxy;
|
||||
struct sfc_repr_proxy_port *port;
|
||||
const unsigned int wait_ms_total =
|
||||
SFC_REPR_PROXY_ROUTINE_TERMINATE_TIMEOUT_MS;
|
||||
unsigned int i;
|
||||
int rc;
|
||||
|
||||
sfc_log_init(sa, "entry");
|
||||
@ -1050,6 +1125,17 @@ sfc_repr_proxy_stop(struct sfc_adapter *sa)
|
||||
|
||||
/* Service lcore may be shared and we never stop it */
|
||||
|
||||
/*
|
||||
* Wait for the representor proxy routine to finish the last iteration.
|
||||
* Give up on timeout.
|
||||
*/
|
||||
for (i = 0; i < wait_ms_total; i++) {
|
||||
if (rte_service_may_be_active(rp->service_id) == 0)
|
||||
break;
|
||||
|
||||
rte_delay_ms(1);
|
||||
}
|
||||
|
||||
sfc_repr_proxy_rxq_stop(sa);
|
||||
sfc_repr_proxy_txq_stop(sa);
|
||||
|
||||
|
@ -79,6 +79,13 @@ struct sfc_repr_proxy_dp_rxq {
|
||||
};
|
||||
|
||||
struct sfc_repr_proxy_dp_txq {
|
||||
eth_tx_burst_t pkt_burst;
|
||||
struct sfc_dp_txq *dp;
|
||||
|
||||
unsigned int available;
|
||||
unsigned int transmitted;
|
||||
struct rte_mbuf *tx_pkts[SFC_REPR_PROXY_TX_BURST];
|
||||
|
||||
sfc_sw_index_t sw_index;
|
||||
};
|
||||
|
||||
@ -110,6 +117,7 @@ struct sfc_repr_proxy {
|
||||
struct sfc_repr_proxy_filter mport_filter;
|
||||
|
||||
struct sfc_repr_proxy_mbox mbox;
|
||||
unsigned int nb_txq;
|
||||
};
|
||||
|
||||
struct sfc_adapter;
|
||||
|
Loading…
Reference in New Issue
Block a user