a6bd4911ad
This patch removes the existing Tx datapath code as preparation step before introducing the new implementation. The following entities are being removed: - deprecated devargs support - tx_burst() routines - related PRM definitions - SQ configuration code - Tx routine selection code - incompatible Tx completion code The following devargs are deprecated and ignored: - "txq_inline" is going to be converted to "txq_inline_max" for compatibility issue - "tx_vec_en" - "txqs_max_vec" - "txq_mpw_hdr_dseg_en" - "txq_max_inline_len" is going to be converted to "txq_inline_mpw" for compatibility issue The deprecated devarg keys are recognized by PMD and ignored/converted to the new ones in order not to block device probing. Signed-off-by: Viacheslav Ovsiienko <viacheslavo@mellanox.com> Acked-by: Yongseok Koh <yskoh@mellanox.com>
855 lines
21 KiB
C
855 lines
21 KiB
C
/* SPDX-License-Identifier: BSD-3-Clause
|
|
* Copyright 2015 6WIND S.A.
|
|
* Copyright 2015 Mellanox Technologies, Ltd
|
|
*/
|
|
|
|
#include <stddef.h>
|
|
#include <assert.h>
|
|
#include <errno.h>
|
|
#include <string.h>
|
|
#include <stdint.h>
|
|
#include <unistd.h>
|
|
#include <sys/mman.h>
|
|
#include <inttypes.h>
|
|
|
|
/* Verbs header. */
|
|
/* ISO C doesn't support unnamed structs/unions, disabling -pedantic. */
|
|
#ifdef PEDANTIC
|
|
#pragma GCC diagnostic ignored "-Wpedantic"
|
|
#endif
|
|
#include <infiniband/verbs.h>
|
|
#ifdef PEDANTIC
|
|
#pragma GCC diagnostic error "-Wpedantic"
|
|
#endif
|
|
|
|
#include <rte_mbuf.h>
|
|
#include <rte_malloc.h>
|
|
#include <rte_ethdev_driver.h>
|
|
#include <rte_common.h>
|
|
|
|
#include "mlx5_utils.h"
|
|
#include "mlx5_defs.h"
|
|
#include "mlx5.h"
|
|
#include "mlx5_rxtx.h"
|
|
#include "mlx5_autoconf.h"
|
|
#include "mlx5_glue.h"
|
|
|
|
/**
|
|
* Allocate TX queue elements.
|
|
*
|
|
* @param txq_ctrl
|
|
* Pointer to TX queue structure.
|
|
*/
|
|
void
|
|
txq_alloc_elts(struct mlx5_txq_ctrl *txq_ctrl)
|
|
{
|
|
const unsigned int elts_n = 1 << txq_ctrl->txq.elts_n;
|
|
unsigned int i;
|
|
|
|
for (i = 0; (i != elts_n); ++i)
|
|
(*txq_ctrl->txq.elts)[i] = NULL;
|
|
DRV_LOG(DEBUG, "port %u Tx queue %u allocated and configured %u WRs",
|
|
PORT_ID(txq_ctrl->priv), txq_ctrl->txq.idx, elts_n);
|
|
txq_ctrl->txq.elts_head = 0;
|
|
txq_ctrl->txq.elts_tail = 0;
|
|
txq_ctrl->txq.elts_comp = 0;
|
|
}
|
|
|
|
/**
|
|
* Free TX queue elements.
|
|
*
|
|
* @param txq_ctrl
|
|
* Pointer to TX queue structure.
|
|
*/
|
|
static void
|
|
txq_free_elts(struct mlx5_txq_ctrl *txq_ctrl)
|
|
{
|
|
const uint16_t elts_n = 1 << txq_ctrl->txq.elts_n;
|
|
const uint16_t elts_m = elts_n - 1;
|
|
uint16_t elts_head = txq_ctrl->txq.elts_head;
|
|
uint16_t elts_tail = txq_ctrl->txq.elts_tail;
|
|
struct rte_mbuf *(*elts)[elts_n] = txq_ctrl->txq.elts;
|
|
|
|
DRV_LOG(DEBUG, "port %u Tx queue %u freeing WRs",
|
|
PORT_ID(txq_ctrl->priv), txq_ctrl->txq.idx);
|
|
txq_ctrl->txq.elts_head = 0;
|
|
txq_ctrl->txq.elts_tail = 0;
|
|
txq_ctrl->txq.elts_comp = 0;
|
|
|
|
while (elts_tail != elts_head) {
|
|
struct rte_mbuf *elt = (*elts)[elts_tail & elts_m];
|
|
|
|
assert(elt != NULL);
|
|
rte_pktmbuf_free_seg(elt);
|
|
#ifndef NDEBUG
|
|
/* Poisoning. */
|
|
memset(&(*elts)[elts_tail & elts_m],
|
|
0x77,
|
|
sizeof((*elts)[elts_tail & elts_m]));
|
|
#endif
|
|
++elts_tail;
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Returns the per-port supported offloads.
|
|
*
|
|
* @param dev
|
|
* Pointer to Ethernet device.
|
|
*
|
|
* @return
|
|
* Supported Tx offloads.
|
|
*/
|
|
uint64_t
|
|
mlx5_get_tx_port_offloads(struct rte_eth_dev *dev)
|
|
{
|
|
struct mlx5_priv *priv = dev->data->dev_private;
|
|
uint64_t offloads = (DEV_TX_OFFLOAD_MULTI_SEGS |
|
|
DEV_TX_OFFLOAD_VLAN_INSERT);
|
|
struct mlx5_dev_config *config = &priv->config;
|
|
|
|
if (config->hw_csum)
|
|
offloads |= (DEV_TX_OFFLOAD_IPV4_CKSUM |
|
|
DEV_TX_OFFLOAD_UDP_CKSUM |
|
|
DEV_TX_OFFLOAD_TCP_CKSUM);
|
|
if (config->tso)
|
|
offloads |= DEV_TX_OFFLOAD_TCP_TSO;
|
|
if (config->swp) {
|
|
if (config->hw_csum)
|
|
offloads |= DEV_TX_OFFLOAD_OUTER_IPV4_CKSUM;
|
|
if (config->tso)
|
|
offloads |= (DEV_TX_OFFLOAD_IP_TNL_TSO |
|
|
DEV_TX_OFFLOAD_UDP_TNL_TSO);
|
|
}
|
|
if (config->tunnel_en) {
|
|
if (config->hw_csum)
|
|
offloads |= DEV_TX_OFFLOAD_OUTER_IPV4_CKSUM;
|
|
if (config->tso)
|
|
offloads |= (DEV_TX_OFFLOAD_VXLAN_TNL_TSO |
|
|
DEV_TX_OFFLOAD_GRE_TNL_TSO);
|
|
}
|
|
#ifdef HAVE_IBV_FLOW_DV_SUPPORT
|
|
if (config->dv_flow_en)
|
|
offloads |= DEV_TX_OFFLOAD_MATCH_METADATA;
|
|
#endif
|
|
return offloads;
|
|
}
|
|
|
|
/**
|
|
* DPDK callback to configure a TX queue.
|
|
*
|
|
* @param dev
|
|
* Pointer to Ethernet device structure.
|
|
* @param idx
|
|
* TX queue index.
|
|
* @param desc
|
|
* Number of descriptors to configure in queue.
|
|
* @param socket
|
|
* NUMA socket on which memory must be allocated.
|
|
* @param[in] conf
|
|
* Thresholds parameters.
|
|
*
|
|
* @return
|
|
* 0 on success, a negative errno value otherwise and rte_errno is set.
|
|
*/
|
|
int
|
|
mlx5_tx_queue_setup(struct rte_eth_dev *dev, uint16_t idx, uint16_t desc,
|
|
unsigned int socket, const struct rte_eth_txconf *conf)
|
|
{
|
|
struct mlx5_priv *priv = dev->data->dev_private;
|
|
struct mlx5_txq_data *txq = (*priv->txqs)[idx];
|
|
struct mlx5_txq_ctrl *txq_ctrl =
|
|
container_of(txq, struct mlx5_txq_ctrl, txq);
|
|
|
|
if (desc <= MLX5_TX_COMP_THRESH) {
|
|
DRV_LOG(WARNING,
|
|
"port %u number of descriptors requested for Tx queue"
|
|
" %u must be higher than MLX5_TX_COMP_THRESH, using %u"
|
|
" instead of %u",
|
|
dev->data->port_id, idx, MLX5_TX_COMP_THRESH + 1, desc);
|
|
desc = MLX5_TX_COMP_THRESH + 1;
|
|
}
|
|
if (!rte_is_power_of_2(desc)) {
|
|
desc = 1 << log2above(desc);
|
|
DRV_LOG(WARNING,
|
|
"port %u increased number of descriptors in Tx queue"
|
|
" %u to the next power of two (%d)",
|
|
dev->data->port_id, idx, desc);
|
|
}
|
|
DRV_LOG(DEBUG, "port %u configuring queue %u for %u descriptors",
|
|
dev->data->port_id, idx, desc);
|
|
if (idx >= priv->txqs_n) {
|
|
DRV_LOG(ERR, "port %u Tx queue index out of range (%u >= %u)",
|
|
dev->data->port_id, idx, priv->txqs_n);
|
|
rte_errno = EOVERFLOW;
|
|
return -rte_errno;
|
|
}
|
|
if (!mlx5_txq_releasable(dev, idx)) {
|
|
rte_errno = EBUSY;
|
|
DRV_LOG(ERR, "port %u unable to release queue index %u",
|
|
dev->data->port_id, idx);
|
|
return -rte_errno;
|
|
}
|
|
mlx5_txq_release(dev, idx);
|
|
txq_ctrl = mlx5_txq_new(dev, idx, desc, socket, conf);
|
|
if (!txq_ctrl) {
|
|
DRV_LOG(ERR, "port %u unable to allocate queue index %u",
|
|
dev->data->port_id, idx);
|
|
return -rte_errno;
|
|
}
|
|
DRV_LOG(DEBUG, "port %u adding Tx queue %u to list",
|
|
dev->data->port_id, idx);
|
|
(*priv->txqs)[idx] = &txq_ctrl->txq;
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* DPDK callback to release a TX queue.
|
|
*
|
|
* @param dpdk_txq
|
|
* Generic TX queue pointer.
|
|
*/
|
|
void
|
|
mlx5_tx_queue_release(void *dpdk_txq)
|
|
{
|
|
struct mlx5_txq_data *txq = (struct mlx5_txq_data *)dpdk_txq;
|
|
struct mlx5_txq_ctrl *txq_ctrl;
|
|
struct mlx5_priv *priv;
|
|
unsigned int i;
|
|
|
|
if (txq == NULL)
|
|
return;
|
|
txq_ctrl = container_of(txq, struct mlx5_txq_ctrl, txq);
|
|
priv = txq_ctrl->priv;
|
|
for (i = 0; (i != priv->txqs_n); ++i)
|
|
if ((*priv->txqs)[i] == txq) {
|
|
mlx5_txq_release(ETH_DEV(priv), i);
|
|
DRV_LOG(DEBUG, "port %u removing Tx queue %u from list",
|
|
PORT_ID(priv), txq->idx);
|
|
break;
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Initialize Tx UAR registers for primary process.
|
|
*
|
|
* @param txq_ctrl
|
|
* Pointer to Tx queue control structure.
|
|
*/
|
|
static void
|
|
txq_uar_init(struct mlx5_txq_ctrl *txq_ctrl)
|
|
{
|
|
struct mlx5_priv *priv = txq_ctrl->priv;
|
|
struct mlx5_proc_priv *ppriv = MLX5_PROC_PRIV(PORT_ID(priv));
|
|
#ifndef RTE_ARCH_64
|
|
unsigned int lock_idx;
|
|
const size_t page_size = sysconf(_SC_PAGESIZE);
|
|
#endif
|
|
|
|
assert(rte_eal_process_type() == RTE_PROC_PRIMARY);
|
|
assert(ppriv);
|
|
ppriv->uar_table[txq_ctrl->txq.idx] = txq_ctrl->bf_reg;
|
|
#ifndef RTE_ARCH_64
|
|
/* Assign an UAR lock according to UAR page number */
|
|
lock_idx = (txq_ctrl->uar_mmap_offset / page_size) &
|
|
MLX5_UAR_PAGE_NUM_MASK;
|
|
txq_ctrl->txq.uar_lock = &priv->uar_lock[lock_idx];
|
|
#endif
|
|
}
|
|
|
|
/**
|
|
* Remap UAR register of a Tx queue for secondary process.
|
|
*
|
|
* Remapped address is stored at the table in the process private structure of
|
|
* the device, indexed by queue index.
|
|
*
|
|
* @param txq_ctrl
|
|
* Pointer to Tx queue control structure.
|
|
* @param fd
|
|
* Verbs file descriptor to map UAR pages.
|
|
*
|
|
* @return
|
|
* 0 on success, a negative errno value otherwise and rte_errno is set.
|
|
*/
|
|
static int
|
|
txq_uar_init_secondary(struct mlx5_txq_ctrl *txq_ctrl, int fd)
|
|
{
|
|
struct mlx5_priv *priv = txq_ctrl->priv;
|
|
struct mlx5_proc_priv *ppriv = MLX5_PROC_PRIV(PORT_ID(priv));
|
|
struct mlx5_txq_data *txq = &txq_ctrl->txq;
|
|
void *addr;
|
|
uintptr_t uar_va;
|
|
uintptr_t offset;
|
|
const size_t page_size = sysconf(_SC_PAGESIZE);
|
|
|
|
assert(ppriv);
|
|
/*
|
|
* As rdma-core, UARs are mapped in size of OS page
|
|
* size. Ref to libmlx5 function: mlx5_init_context()
|
|
*/
|
|
uar_va = (uintptr_t)txq_ctrl->bf_reg;
|
|
offset = uar_va & (page_size - 1); /* Offset in page. */
|
|
addr = mmap(NULL, page_size, PROT_WRITE, MAP_SHARED, fd,
|
|
txq_ctrl->uar_mmap_offset);
|
|
if (addr == MAP_FAILED) {
|
|
DRV_LOG(ERR,
|
|
"port %u mmap failed for BF reg of txq %u",
|
|
txq->port_id, txq->idx);
|
|
rte_errno = ENXIO;
|
|
return -rte_errno;
|
|
}
|
|
addr = RTE_PTR_ADD(addr, offset);
|
|
ppriv->uar_table[txq->idx] = addr;
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* Unmap UAR register of a Tx queue for secondary process.
|
|
*
|
|
* @param txq_ctrl
|
|
* Pointer to Tx queue control structure.
|
|
*/
|
|
static void
|
|
txq_uar_uninit_secondary(struct mlx5_txq_ctrl *txq_ctrl)
|
|
{
|
|
struct mlx5_proc_priv *ppriv = MLX5_PROC_PRIV(PORT_ID(txq_ctrl->priv));
|
|
const size_t page_size = sysconf(_SC_PAGESIZE);
|
|
void *addr;
|
|
|
|
addr = ppriv->uar_table[txq_ctrl->txq.idx];
|
|
munmap(RTE_PTR_ALIGN_FLOOR(addr, page_size), page_size);
|
|
}
|
|
|
|
/**
|
|
* Initialize Tx UAR registers for secondary process.
|
|
*
|
|
* @param dev
|
|
* Pointer to Ethernet device.
|
|
* @param fd
|
|
* Verbs file descriptor to map UAR pages.
|
|
*
|
|
* @return
|
|
* 0 on success, a negative errno value otherwise and rte_errno is set.
|
|
*/
|
|
int
|
|
mlx5_tx_uar_init_secondary(struct rte_eth_dev *dev, int fd)
|
|
{
|
|
struct mlx5_priv *priv = dev->data->dev_private;
|
|
struct mlx5_txq_data *txq;
|
|
struct mlx5_txq_ctrl *txq_ctrl;
|
|
unsigned int i;
|
|
int ret;
|
|
|
|
assert(rte_eal_process_type() == RTE_PROC_SECONDARY);
|
|
for (i = 0; i != priv->txqs_n; ++i) {
|
|
if (!(*priv->txqs)[i])
|
|
continue;
|
|
txq = (*priv->txqs)[i];
|
|
txq_ctrl = container_of(txq, struct mlx5_txq_ctrl, txq);
|
|
assert(txq->idx == (uint16_t)i);
|
|
ret = txq_uar_init_secondary(txq_ctrl, fd);
|
|
if (ret)
|
|
goto error;
|
|
}
|
|
return 0;
|
|
error:
|
|
/* Rollback. */
|
|
do {
|
|
if (!(*priv->txqs)[i])
|
|
continue;
|
|
txq = (*priv->txqs)[i];
|
|
txq_ctrl = container_of(txq, struct mlx5_txq_ctrl, txq);
|
|
txq_uar_uninit_secondary(txq_ctrl);
|
|
} while (i--);
|
|
return -rte_errno;
|
|
}
|
|
|
|
/**
|
|
* Create the Tx queue Verbs object.
|
|
*
|
|
* @param dev
|
|
* Pointer to Ethernet device.
|
|
* @param idx
|
|
* Queue index in DPDK Tx queue array.
|
|
*
|
|
* @return
|
|
* The Verbs object initialised, NULL otherwise and rte_errno is set.
|
|
*/
|
|
struct mlx5_txq_ibv *
|
|
mlx5_txq_ibv_new(struct rte_eth_dev *dev, uint16_t idx)
|
|
{
|
|
struct mlx5_priv *priv = dev->data->dev_private;
|
|
struct mlx5_txq_data *txq_data = (*priv->txqs)[idx];
|
|
struct mlx5_txq_ctrl *txq_ctrl =
|
|
container_of(txq_data, struct mlx5_txq_ctrl, txq);
|
|
struct mlx5_txq_ibv tmpl;
|
|
struct mlx5_txq_ibv *txq_ibv = NULL;
|
|
union {
|
|
struct ibv_qp_init_attr_ex init;
|
|
struct ibv_cq_init_attr_ex cq;
|
|
struct ibv_qp_attr mod;
|
|
struct ibv_cq_ex cq_attr;
|
|
} attr;
|
|
unsigned int cqe_n;
|
|
struct mlx5dv_qp qp = { .comp_mask = MLX5DV_QP_MASK_UAR_MMAP_OFFSET };
|
|
struct mlx5dv_cq cq_info;
|
|
struct mlx5dv_obj obj;
|
|
const int desc = 1 << txq_data->elts_n;
|
|
int ret = 0;
|
|
|
|
assert(txq_data);
|
|
priv->verbs_alloc_ctx.type = MLX5_VERBS_ALLOC_TYPE_TX_QUEUE;
|
|
priv->verbs_alloc_ctx.obj = txq_ctrl;
|
|
if (mlx5_getenv_int("MLX5_ENABLE_CQE_COMPRESSION")) {
|
|
DRV_LOG(ERR,
|
|
"port %u MLX5_ENABLE_CQE_COMPRESSION must never be set",
|
|
dev->data->port_id);
|
|
rte_errno = EINVAL;
|
|
return NULL;
|
|
}
|
|
memset(&tmpl, 0, sizeof(struct mlx5_txq_ibv));
|
|
attr.cq = (struct ibv_cq_init_attr_ex){
|
|
.comp_mask = 0,
|
|
};
|
|
cqe_n = desc / MLX5_TX_COMP_THRESH + 1;
|
|
tmpl.cq = mlx5_glue->create_cq(priv->sh->ctx, cqe_n, NULL, NULL, 0);
|
|
if (tmpl.cq == NULL) {
|
|
DRV_LOG(ERR, "port %u Tx queue %u CQ creation failure",
|
|
dev->data->port_id, idx);
|
|
rte_errno = errno;
|
|
goto error;
|
|
}
|
|
attr.init = (struct ibv_qp_init_attr_ex){
|
|
/* CQ to be associated with the send queue. */
|
|
.send_cq = tmpl.cq,
|
|
/* CQ to be associated with the receive queue. */
|
|
.recv_cq = tmpl.cq,
|
|
.cap = {
|
|
/* Max number of outstanding WRs. */
|
|
.max_send_wr =
|
|
((priv->sh->device_attr.orig_attr.max_qp_wr <
|
|
desc) ?
|
|
priv->sh->device_attr.orig_attr.max_qp_wr :
|
|
desc),
|
|
/*
|
|
* Max number of scatter/gather elements in a WR,
|
|
* must be 1 to prevent libmlx5 from trying to affect
|
|
* too much memory. TX gather is not impacted by the
|
|
* device_attr.max_sge limit and will still work
|
|
* properly.
|
|
*/
|
|
.max_send_sge = 1,
|
|
},
|
|
.qp_type = IBV_QPT_RAW_PACKET,
|
|
/*
|
|
* Do *NOT* enable this, completions events are managed per
|
|
* Tx burst.
|
|
*/
|
|
.sq_sig_all = 0,
|
|
.pd = priv->sh->pd,
|
|
.comp_mask = IBV_QP_INIT_ATTR_PD,
|
|
};
|
|
if (txq_data->max_inline)
|
|
attr.init.cap.max_inline_data = txq_ctrl->max_inline_data;
|
|
if (txq_data->tso_en) {
|
|
attr.init.max_tso_header = txq_ctrl->max_tso_header;
|
|
attr.init.comp_mask |= IBV_QP_INIT_ATTR_MAX_TSO_HEADER;
|
|
}
|
|
tmpl.qp = mlx5_glue->create_qp_ex(priv->sh->ctx, &attr.init);
|
|
if (tmpl.qp == NULL) {
|
|
DRV_LOG(ERR, "port %u Tx queue %u QP creation failure",
|
|
dev->data->port_id, idx);
|
|
rte_errno = errno;
|
|
goto error;
|
|
}
|
|
attr.mod = (struct ibv_qp_attr){
|
|
/* Move the QP to this state. */
|
|
.qp_state = IBV_QPS_INIT,
|
|
/* IB device port number. */
|
|
.port_num = (uint8_t)priv->ibv_port,
|
|
};
|
|
ret = mlx5_glue->modify_qp(tmpl.qp, &attr.mod,
|
|
(IBV_QP_STATE | IBV_QP_PORT));
|
|
if (ret) {
|
|
DRV_LOG(ERR,
|
|
"port %u Tx queue %u QP state to IBV_QPS_INIT failed",
|
|
dev->data->port_id, idx);
|
|
rte_errno = errno;
|
|
goto error;
|
|
}
|
|
attr.mod = (struct ibv_qp_attr){
|
|
.qp_state = IBV_QPS_RTR
|
|
};
|
|
ret = mlx5_glue->modify_qp(tmpl.qp, &attr.mod, IBV_QP_STATE);
|
|
if (ret) {
|
|
DRV_LOG(ERR,
|
|
"port %u Tx queue %u QP state to IBV_QPS_RTR failed",
|
|
dev->data->port_id, idx);
|
|
rte_errno = errno;
|
|
goto error;
|
|
}
|
|
attr.mod.qp_state = IBV_QPS_RTS;
|
|
ret = mlx5_glue->modify_qp(tmpl.qp, &attr.mod, IBV_QP_STATE);
|
|
if (ret) {
|
|
DRV_LOG(ERR,
|
|
"port %u Tx queue %u QP state to IBV_QPS_RTS failed",
|
|
dev->data->port_id, idx);
|
|
rte_errno = errno;
|
|
goto error;
|
|
}
|
|
txq_ibv = rte_calloc_socket(__func__, 1, sizeof(struct mlx5_txq_ibv), 0,
|
|
txq_ctrl->socket);
|
|
if (!txq_ibv) {
|
|
DRV_LOG(ERR, "port %u Tx queue %u cannot allocate memory",
|
|
dev->data->port_id, idx);
|
|
rte_errno = ENOMEM;
|
|
goto error;
|
|
}
|
|
obj.cq.in = tmpl.cq;
|
|
obj.cq.out = &cq_info;
|
|
obj.qp.in = tmpl.qp;
|
|
obj.qp.out = &qp;
|
|
ret = mlx5_glue->dv_init_obj(&obj, MLX5DV_OBJ_CQ | MLX5DV_OBJ_QP);
|
|
if (ret != 0) {
|
|
rte_errno = errno;
|
|
goto error;
|
|
}
|
|
if (cq_info.cqe_size != RTE_CACHE_LINE_SIZE) {
|
|
DRV_LOG(ERR,
|
|
"port %u wrong MLX5_CQE_SIZE environment variable"
|
|
" value: it should be set to %u",
|
|
dev->data->port_id, RTE_CACHE_LINE_SIZE);
|
|
rte_errno = EINVAL;
|
|
goto error;
|
|
}
|
|
txq_data->cqe_n = log2above(cq_info.cqe_cnt);
|
|
txq_data->qp_num_8s = tmpl.qp->qp_num << 8;
|
|
txq_data->wqes = qp.sq.buf;
|
|
txq_data->wqe_n = log2above(qp.sq.wqe_cnt);
|
|
txq_data->qp_db = &qp.dbrec[MLX5_SND_DBR];
|
|
txq_data->cq_db = cq_info.dbrec;
|
|
txq_data->cqes =
|
|
(volatile struct mlx5_cqe (*)[])
|
|
(uintptr_t)cq_info.buf;
|
|
txq_data->cq_ci = 0;
|
|
#ifndef NDEBUG
|
|
txq_data->cq_pi = 0;
|
|
#endif
|
|
txq_data->wqe_ci = 0;
|
|
txq_data->wqe_pi = 0;
|
|
txq_ibv->qp = tmpl.qp;
|
|
txq_ibv->cq = tmpl.cq;
|
|
rte_atomic32_inc(&txq_ibv->refcnt);
|
|
txq_ctrl->bf_reg = qp.bf.reg;
|
|
txq_ctrl->cqn = cq_info.cqn;
|
|
txq_uar_init(txq_ctrl);
|
|
if (qp.comp_mask & MLX5DV_QP_MASK_UAR_MMAP_OFFSET) {
|
|
txq_ctrl->uar_mmap_offset = qp.uar_mmap_offset;
|
|
DRV_LOG(DEBUG, "port %u: uar_mmap_offset 0x%"PRIx64,
|
|
dev->data->port_id, txq_ctrl->uar_mmap_offset);
|
|
} else {
|
|
DRV_LOG(ERR,
|
|
"port %u failed to retrieve UAR info, invalid"
|
|
" libmlx5.so",
|
|
dev->data->port_id);
|
|
rte_errno = EINVAL;
|
|
goto error;
|
|
}
|
|
LIST_INSERT_HEAD(&priv->txqsibv, txq_ibv, next);
|
|
txq_ibv->txq_ctrl = txq_ctrl;
|
|
priv->verbs_alloc_ctx.type = MLX5_VERBS_ALLOC_TYPE_NONE;
|
|
return txq_ibv;
|
|
error:
|
|
ret = rte_errno; /* Save rte_errno before cleanup. */
|
|
if (tmpl.cq)
|
|
claim_zero(mlx5_glue->destroy_cq(tmpl.cq));
|
|
if (tmpl.qp)
|
|
claim_zero(mlx5_glue->destroy_qp(tmpl.qp));
|
|
if (txq_ibv)
|
|
rte_free(txq_ibv);
|
|
priv->verbs_alloc_ctx.type = MLX5_VERBS_ALLOC_TYPE_NONE;
|
|
rte_errno = ret; /* Restore rte_errno. */
|
|
return NULL;
|
|
}
|
|
|
|
/**
|
|
* Get an Tx queue Verbs object.
|
|
*
|
|
* @param dev
|
|
* Pointer to Ethernet device.
|
|
* @param idx
|
|
* Queue index in DPDK Tx queue array.
|
|
*
|
|
* @return
|
|
* The Verbs object if it exists.
|
|
*/
|
|
struct mlx5_txq_ibv *
|
|
mlx5_txq_ibv_get(struct rte_eth_dev *dev, uint16_t idx)
|
|
{
|
|
struct mlx5_priv *priv = dev->data->dev_private;
|
|
struct mlx5_txq_ctrl *txq_ctrl;
|
|
|
|
if (idx >= priv->txqs_n)
|
|
return NULL;
|
|
if (!(*priv->txqs)[idx])
|
|
return NULL;
|
|
txq_ctrl = container_of((*priv->txqs)[idx], struct mlx5_txq_ctrl, txq);
|
|
if (txq_ctrl->ibv)
|
|
rte_atomic32_inc(&txq_ctrl->ibv->refcnt);
|
|
return txq_ctrl->ibv;
|
|
}
|
|
|
|
/**
|
|
* Release an Tx verbs queue object.
|
|
*
|
|
* @param txq_ibv
|
|
* Verbs Tx queue object.
|
|
*
|
|
* @return
|
|
* 1 while a reference on it exists, 0 when freed.
|
|
*/
|
|
int
|
|
mlx5_txq_ibv_release(struct mlx5_txq_ibv *txq_ibv)
|
|
{
|
|
assert(txq_ibv);
|
|
if (rte_atomic32_dec_and_test(&txq_ibv->refcnt)) {
|
|
claim_zero(mlx5_glue->destroy_qp(txq_ibv->qp));
|
|
claim_zero(mlx5_glue->destroy_cq(txq_ibv->cq));
|
|
LIST_REMOVE(txq_ibv, next);
|
|
rte_free(txq_ibv);
|
|
return 0;
|
|
}
|
|
return 1;
|
|
}
|
|
|
|
/**
|
|
* Verify the Verbs Tx queue list is empty
|
|
*
|
|
* @param dev
|
|
* Pointer to Ethernet device.
|
|
*
|
|
* @return
|
|
* The number of object not released.
|
|
*/
|
|
int
|
|
mlx5_txq_ibv_verify(struct rte_eth_dev *dev)
|
|
{
|
|
struct mlx5_priv *priv = dev->data->dev_private;
|
|
int ret = 0;
|
|
struct mlx5_txq_ibv *txq_ibv;
|
|
|
|
LIST_FOREACH(txq_ibv, &priv->txqsibv, next) {
|
|
DRV_LOG(DEBUG, "port %u Verbs Tx queue %u still referenced",
|
|
dev->data->port_id, txq_ibv->txq_ctrl->txq.idx);
|
|
++ret;
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
/**
|
|
* Calculate the total number of WQEBB for Tx queue.
|
|
*
|
|
* Simplified version of calc_sq_size() in rdma-core.
|
|
*
|
|
* @param txq_ctrl
|
|
* Pointer to Tx queue control structure.
|
|
*
|
|
* @return
|
|
* The number of WQEBB.
|
|
*/
|
|
static int
|
|
txq_calc_wqebb_cnt(struct mlx5_txq_ctrl *txq_ctrl)
|
|
{
|
|
unsigned int wqe_size;
|
|
const unsigned int desc = 1 << txq_ctrl->txq.elts_n;
|
|
|
|
wqe_size = MLX5_WQE_SIZE + txq_ctrl->max_inline_data;
|
|
return rte_align32pow2(wqe_size * desc) / MLX5_WQE_SIZE;
|
|
}
|
|
|
|
/**
|
|
* Set Tx queue parameters from device configuration.
|
|
*
|
|
* @param txq_ctrl
|
|
* Pointer to Tx queue control structure.
|
|
*/
|
|
static void
|
|
txq_set_params(struct mlx5_txq_ctrl *txq_ctrl)
|
|
{
|
|
(void)txq_ctrl;
|
|
}
|
|
|
|
/**
|
|
* Create a DPDK Tx queue.
|
|
*
|
|
* @param dev
|
|
* Pointer to Ethernet device.
|
|
* @param idx
|
|
* TX queue index.
|
|
* @param desc
|
|
* Number of descriptors to configure in queue.
|
|
* @param socket
|
|
* NUMA socket on which memory must be allocated.
|
|
* @param[in] conf
|
|
* Thresholds parameters.
|
|
*
|
|
* @return
|
|
* A DPDK queue object on success, NULL otherwise and rte_errno is set.
|
|
*/
|
|
struct mlx5_txq_ctrl *
|
|
mlx5_txq_new(struct rte_eth_dev *dev, uint16_t idx, uint16_t desc,
|
|
unsigned int socket, const struct rte_eth_txconf *conf)
|
|
{
|
|
struct mlx5_priv *priv = dev->data->dev_private;
|
|
struct mlx5_txq_ctrl *tmpl;
|
|
|
|
tmpl = rte_calloc_socket("TXQ", 1,
|
|
sizeof(*tmpl) +
|
|
desc * sizeof(struct rte_mbuf *),
|
|
0, socket);
|
|
if (!tmpl) {
|
|
rte_errno = ENOMEM;
|
|
return NULL;
|
|
}
|
|
if (mlx5_mr_btree_init(&tmpl->txq.mr_ctrl.cache_bh,
|
|
MLX5_MR_BTREE_CACHE_N, socket)) {
|
|
/* rte_errno is already set. */
|
|
goto error;
|
|
}
|
|
/* Save pointer of global generation number to check memory event. */
|
|
tmpl->txq.mr_ctrl.dev_gen_ptr = &priv->sh->mr.dev_gen;
|
|
assert(desc > MLX5_TX_COMP_THRESH);
|
|
tmpl->txq.offloads = conf->offloads |
|
|
dev->data->dev_conf.txmode.offloads;
|
|
tmpl->priv = priv;
|
|
tmpl->socket = socket;
|
|
tmpl->txq.elts_n = log2above(desc);
|
|
tmpl->txq.port_id = dev->data->port_id;
|
|
tmpl->txq.idx = idx;
|
|
txq_set_params(tmpl);
|
|
if (txq_calc_wqebb_cnt(tmpl) >
|
|
priv->sh->device_attr.orig_attr.max_qp_wr) {
|
|
DRV_LOG(ERR,
|
|
"port %u Tx WQEBB count (%d) exceeds the limit (%d),"
|
|
" try smaller queue size",
|
|
dev->data->port_id, txq_calc_wqebb_cnt(tmpl),
|
|
priv->sh->device_attr.orig_attr.max_qp_wr);
|
|
rte_errno = ENOMEM;
|
|
goto error;
|
|
}
|
|
tmpl->txq.elts =
|
|
(struct rte_mbuf *(*)[1 << tmpl->txq.elts_n])(tmpl + 1);
|
|
rte_atomic32_inc(&tmpl->refcnt);
|
|
LIST_INSERT_HEAD(&priv->txqsctrl, tmpl, next);
|
|
return tmpl;
|
|
error:
|
|
rte_free(tmpl);
|
|
return NULL;
|
|
}
|
|
|
|
/**
|
|
* Get a Tx queue.
|
|
*
|
|
* @param dev
|
|
* Pointer to Ethernet device.
|
|
* @param idx
|
|
* TX queue index.
|
|
*
|
|
* @return
|
|
* A pointer to the queue if it exists.
|
|
*/
|
|
struct mlx5_txq_ctrl *
|
|
mlx5_txq_get(struct rte_eth_dev *dev, uint16_t idx)
|
|
{
|
|
struct mlx5_priv *priv = dev->data->dev_private;
|
|
struct mlx5_txq_ctrl *ctrl = NULL;
|
|
|
|
if ((*priv->txqs)[idx]) {
|
|
ctrl = container_of((*priv->txqs)[idx], struct mlx5_txq_ctrl,
|
|
txq);
|
|
mlx5_txq_ibv_get(dev, idx);
|
|
rte_atomic32_inc(&ctrl->refcnt);
|
|
}
|
|
return ctrl;
|
|
}
|
|
|
|
/**
|
|
* Release a Tx queue.
|
|
*
|
|
* @param dev
|
|
* Pointer to Ethernet device.
|
|
* @param idx
|
|
* TX queue index.
|
|
*
|
|
* @return
|
|
* 1 while a reference on it exists, 0 when freed.
|
|
*/
|
|
int
|
|
mlx5_txq_release(struct rte_eth_dev *dev, uint16_t idx)
|
|
{
|
|
struct mlx5_priv *priv = dev->data->dev_private;
|
|
struct mlx5_txq_ctrl *txq;
|
|
|
|
if (!(*priv->txqs)[idx])
|
|
return 0;
|
|
txq = container_of((*priv->txqs)[idx], struct mlx5_txq_ctrl, txq);
|
|
if (txq->ibv && !mlx5_txq_ibv_release(txq->ibv))
|
|
txq->ibv = NULL;
|
|
if (rte_atomic32_dec_and_test(&txq->refcnt)) {
|
|
txq_free_elts(txq);
|
|
mlx5_mr_btree_free(&txq->txq.mr_ctrl.cache_bh);
|
|
LIST_REMOVE(txq, next);
|
|
rte_free(txq);
|
|
(*priv->txqs)[idx] = NULL;
|
|
return 0;
|
|
}
|
|
return 1;
|
|
}
|
|
|
|
/**
|
|
* Verify if the queue can be released.
|
|
*
|
|
* @param dev
|
|
* Pointer to Ethernet device.
|
|
* @param idx
|
|
* TX queue index.
|
|
*
|
|
* @return
|
|
* 1 if the queue can be released.
|
|
*/
|
|
int
|
|
mlx5_txq_releasable(struct rte_eth_dev *dev, uint16_t idx)
|
|
{
|
|
struct mlx5_priv *priv = dev->data->dev_private;
|
|
struct mlx5_txq_ctrl *txq;
|
|
|
|
if (!(*priv->txqs)[idx])
|
|
return -1;
|
|
txq = container_of((*priv->txqs)[idx], struct mlx5_txq_ctrl, txq);
|
|
return (rte_atomic32_read(&txq->refcnt) == 1);
|
|
}
|
|
|
|
/**
|
|
* Verify the Tx Queue list is empty
|
|
*
|
|
* @param dev
|
|
* Pointer to Ethernet device.
|
|
*
|
|
* @return
|
|
* The number of object not released.
|
|
*/
|
|
int
|
|
mlx5_txq_verify(struct rte_eth_dev *dev)
|
|
{
|
|
struct mlx5_priv *priv = dev->data->dev_private;
|
|
struct mlx5_txq_ctrl *txq_ctrl;
|
|
int ret = 0;
|
|
|
|
LIST_FOREACH(txq_ctrl, &priv->txqsctrl, next) {
|
|
DRV_LOG(DEBUG, "port %u Tx queue %u still referenced",
|
|
dev->data->port_id, txq_ctrl->txq.idx);
|
|
++ret;
|
|
}
|
|
return ret;
|
|
}
|