2018-01-29 13:11:30 +00:00
|
|
|
/* SPDX-License-Identifier: BSD-3-Clause
|
|
|
|
* Copyright 2015 6WIND S.A.
|
2018-03-20 19:20:35 +00:00
|
|
|
* Copyright 2015 Mellanox Technologies, Ltd
|
2015-10-30 18:52:40 +00:00
|
|
|
*/
|
|
|
|
|
|
|
|
#include <stddef.h>
|
|
|
|
#include <errno.h>
|
|
|
|
#include <stdint.h>
|
2020-01-29 12:38:47 +00:00
|
|
|
#include <unistd.h>
|
|
|
|
|
2015-10-30 18:52:40 +00:00
|
|
|
|
2018-03-14 17:51:48 +00:00
|
|
|
/*
|
|
|
|
* Not needed by this file; included to work around the lack of off_t
|
|
|
|
* definition for mlx5dv.h with unpatched rdma-core versions.
|
|
|
|
*/
|
|
|
|
#include <sys/types.h>
|
|
|
|
|
2018-01-30 15:34:56 +00:00
|
|
|
/* Verbs headers do not support -pedantic. */
|
|
|
|
#ifdef PEDANTIC
|
|
|
|
#pragma GCC diagnostic ignored "-Wpedantic"
|
|
|
|
#endif
|
|
|
|
#include <infiniband/mlx5dv.h>
|
|
|
|
#include <infiniband/verbs.h>
|
|
|
|
#ifdef PEDANTIC
|
|
|
|
#pragma GCC diagnostic error "-Wpedantic"
|
|
|
|
#endif
|
|
|
|
|
2018-01-22 00:16:22 +00:00
|
|
|
#include <rte_ethdev_driver.h>
|
2015-10-30 18:52:40 +00:00
|
|
|
#include <rte_common.h>
|
2020-01-29 12:38:47 +00:00
|
|
|
#include <rte_malloc.h>
|
|
|
|
#include <rte_hypervisor.h>
|
2015-10-30 18:52:40 +00:00
|
|
|
|
2020-01-29 12:38:27 +00:00
|
|
|
#include <mlx5_glue.h>
|
|
|
|
#include <mlx5_devx_cmds.h>
|
2020-01-29 12:38:49 +00:00
|
|
|
#include <mlx5_nl.h>
|
2020-01-29 12:38:27 +00:00
|
|
|
|
2015-10-30 18:52:40 +00:00
|
|
|
#include "mlx5.h"
|
2016-03-03 14:26:44 +00:00
|
|
|
#include "mlx5_autoconf.h"
|
2019-04-10 18:41:15 +00:00
|
|
|
#include "mlx5_rxtx.h"
|
|
|
|
#include "mlx5_utils.h"
|
2015-10-30 18:52:40 +00:00
|
|
|
|
|
|
|
/**
|
2017-10-09 14:44:55 +00:00
|
|
|
* DPDK callback to configure a VLAN filter.
|
2015-10-30 18:52:40 +00:00
|
|
|
*
|
|
|
|
* @param dev
|
|
|
|
* Pointer to Ethernet device structure.
|
|
|
|
* @param vlan_id
|
|
|
|
* VLAN ID to filter.
|
|
|
|
* @param on
|
|
|
|
* Toggle filter.
|
|
|
|
*
|
|
|
|
* @return
|
2018-03-05 12:21:06 +00:00
|
|
|
* 0 on success, a negative errno value otherwise and rte_errno is set.
|
2015-10-30 18:52:40 +00:00
|
|
|
*/
|
2017-10-09 14:44:55 +00:00
|
|
|
int
|
|
|
|
mlx5_vlan_filter_set(struct rte_eth_dev *dev, uint16_t vlan_id, int on)
|
2015-10-30 18:52:40 +00:00
|
|
|
{
|
2019-02-21 09:29:14 +00:00
|
|
|
struct mlx5_priv *priv = dev->data->dev_private;
|
2015-10-30 18:52:40 +00:00
|
|
|
unsigned int i;
|
|
|
|
|
2018-03-13 09:23:56 +00:00
|
|
|
DRV_LOG(DEBUG, "port %u %s VLAN filter ID %" PRIu16,
|
|
|
|
dev->data->port_id, (on ? "enable" : "disable"), vlan_id);
|
2020-01-30 16:14:40 +00:00
|
|
|
MLX5_ASSERT(priv->vlan_filter_n <= RTE_DIM(priv->vlan_filter));
|
2015-10-30 18:52:40 +00:00
|
|
|
for (i = 0; (i != priv->vlan_filter_n); ++i)
|
|
|
|
if (priv->vlan_filter[i] == vlan_id)
|
|
|
|
break;
|
2017-11-02 13:30:03 +00:00
|
|
|
/* Check if there's room for another VLAN filter. */
|
|
|
|
if (i == RTE_DIM(priv->vlan_filter)) {
|
2018-03-05 12:21:06 +00:00
|
|
|
rte_errno = ENOMEM;
|
|
|
|
return -rte_errno;
|
2017-11-02 13:30:03 +00:00
|
|
|
}
|
2015-10-30 18:52:40 +00:00
|
|
|
if (i < priv->vlan_filter_n) {
|
2020-01-30 16:14:40 +00:00
|
|
|
MLX5_ASSERT(priv->vlan_filter_n != 0);
|
2015-10-30 18:52:40 +00:00
|
|
|
/* Enabling an existing VLAN filter has no effect. */
|
|
|
|
if (on)
|
2017-10-09 14:44:55 +00:00
|
|
|
goto out;
|
2015-10-30 18:52:40 +00:00
|
|
|
/* Remove VLAN filter from list. */
|
|
|
|
--priv->vlan_filter_n;
|
|
|
|
memmove(&priv->vlan_filter[i],
|
|
|
|
&priv->vlan_filter[i + 1],
|
2016-09-14 11:53:50 +00:00
|
|
|
sizeof(priv->vlan_filter[i]) *
|
|
|
|
(priv->vlan_filter_n - i));
|
2015-10-30 18:52:40 +00:00
|
|
|
priv->vlan_filter[priv->vlan_filter_n] = 0;
|
|
|
|
} else {
|
2020-01-30 16:14:40 +00:00
|
|
|
MLX5_ASSERT(i == priv->vlan_filter_n);
|
2015-10-30 18:52:40 +00:00
|
|
|
/* Disabling an unknown VLAN filter has no effect. */
|
|
|
|
if (!on)
|
2017-10-09 14:44:55 +00:00
|
|
|
goto out;
|
2015-10-30 18:52:40 +00:00
|
|
|
/* Add new VLAN filter. */
|
|
|
|
priv->vlan_filter[priv->vlan_filter_n] = vlan_id;
|
|
|
|
++priv->vlan_filter_n;
|
|
|
|
}
|
2017-10-09 14:44:55 +00:00
|
|
|
out:
|
2018-03-05 12:21:06 +00:00
|
|
|
if (dev->data->dev_started)
|
|
|
|
return mlx5_traffic_restart(dev);
|
|
|
|
return 0;
|
2015-10-30 18:52:40 +00:00
|
|
|
}
|
2016-03-03 14:26:44 +00:00
|
|
|
|
|
|
|
/**
|
2018-03-05 12:21:04 +00:00
|
|
|
* Callback to set/reset VLAN stripping for a specific queue.
|
2016-03-03 14:26:44 +00:00
|
|
|
*
|
2018-03-05 12:21:04 +00:00
|
|
|
* @param dev
|
|
|
|
* Pointer to Ethernet device structure.
|
|
|
|
* @param queue
|
2016-03-03 14:26:44 +00:00
|
|
|
* RX queue index.
|
|
|
|
* @param on
|
|
|
|
* Enable/disable VLAN stripping.
|
|
|
|
*/
|
2018-03-05 12:21:04 +00:00
|
|
|
void
|
|
|
|
mlx5_vlan_strip_queue_set(struct rte_eth_dev *dev, uint16_t queue, int on)
|
2016-03-03 14:26:44 +00:00
|
|
|
{
|
2019-02-21 09:29:14 +00:00
|
|
|
struct mlx5_priv *priv = dev->data->dev_private;
|
2018-03-05 12:21:04 +00:00
|
|
|
struct mlx5_rxq_data *rxq = (*priv->rxqs)[queue];
|
2017-10-09 14:44:39 +00:00
|
|
|
struct mlx5_rxq_ctrl *rxq_ctrl =
|
|
|
|
container_of(rxq, struct mlx5_rxq_ctrl, rxq);
|
2017-09-26 15:38:24 +00:00
|
|
|
struct ibv_wq_attr mod;
|
2016-03-03 14:26:44 +00:00
|
|
|
uint16_t vlan_offloads =
|
2017-09-26 15:38:24 +00:00
|
|
|
(on ? IBV_WQ_FLAGS_CVLAN_STRIPPING : 0) |
|
2016-03-03 14:26:44 +00:00
|
|
|
0;
|
2019-07-22 14:52:18 +00:00
|
|
|
int ret = 0;
|
2016-03-03 14:26:44 +00:00
|
|
|
|
2018-03-05 12:21:04 +00:00
|
|
|
/* Validate hw support */
|
|
|
|
if (!priv->config.hw_vlan_strip) {
|
2018-03-13 09:23:56 +00:00
|
|
|
DRV_LOG(ERR, "port %u VLAN stripping is not supported",
|
|
|
|
dev->data->port_id);
|
2018-03-05 12:21:04 +00:00
|
|
|
return;
|
|
|
|
}
|
|
|
|
/* Validate queue number */
|
|
|
|
if (queue >= priv->rxqs_n) {
|
2018-03-13 09:23:56 +00:00
|
|
|
DRV_LOG(ERR, "port %u VLAN stripping, invalid queue number %d",
|
|
|
|
dev->data->port_id, queue);
|
2018-03-05 12:21:04 +00:00
|
|
|
return;
|
|
|
|
}
|
2018-03-13 09:23:56 +00:00
|
|
|
DRV_LOG(DEBUG, "port %u set VLAN offloads 0x%x for port %uqueue %d",
|
|
|
|
dev->data->port_id, vlan_offloads, rxq->port_id, queue);
|
2019-07-22 14:52:11 +00:00
|
|
|
if (!rxq_ctrl->obj) {
|
2017-12-13 14:03:10 +00:00
|
|
|
/* Update related bits in RX queue. */
|
|
|
|
rxq->vlan_strip = !!on;
|
|
|
|
return;
|
|
|
|
}
|
2019-07-22 14:52:18 +00:00
|
|
|
if (rxq_ctrl->obj->type == MLX5_RXQ_OBJ_TYPE_IBV) {
|
|
|
|
mod = (struct ibv_wq_attr){
|
|
|
|
.attr_mask = IBV_WQ_ATTR_FLAGS,
|
|
|
|
.flags_mask = IBV_WQ_FLAGS_CVLAN_STRIPPING,
|
|
|
|
.flags = vlan_offloads,
|
|
|
|
};
|
|
|
|
ret = mlx5_glue->modify_wq(rxq_ctrl->obj->wq, &mod);
|
|
|
|
} else if (rxq_ctrl->obj->type == MLX5_RXQ_OBJ_TYPE_DEVX_RQ) {
|
|
|
|
struct mlx5_devx_modify_rq_attr rq_attr;
|
|
|
|
|
|
|
|
memset(&rq_attr, 0, sizeof(rq_attr));
|
|
|
|
rq_attr.rq_state = MLX5_RQC_STATE_RDY;
|
|
|
|
rq_attr.state = MLX5_RQC_STATE_RDY;
|
|
|
|
rq_attr.vsd = (on ? 0 : 1);
|
|
|
|
rq_attr.modify_bitmask = MLX5_MODIFY_RQ_IN_MODIFY_BITMASK_VSD;
|
|
|
|
ret = mlx5_devx_cmd_modify_rq(rxq_ctrl->obj->rq, &rq_attr);
|
|
|
|
}
|
2018-03-05 12:21:06 +00:00
|
|
|
if (ret) {
|
2019-07-22 14:52:18 +00:00
|
|
|
DRV_LOG(ERR, "port %u failed to modify object %d stripping "
|
|
|
|
"mode: %s", dev->data->port_id,
|
|
|
|
rxq_ctrl->obj->type, strerror(rte_errno));
|
2016-03-03 14:26:44 +00:00
|
|
|
return;
|
|
|
|
}
|
|
|
|
/* Update related bits in RX queue. */
|
|
|
|
rxq->vlan_strip = !!on;
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* Callback to set/reset VLAN offloads for a port.
|
|
|
|
*
|
|
|
|
* @param dev
|
|
|
|
* Pointer to Ethernet device structure.
|
|
|
|
* @param mask
|
|
|
|
* VLAN offload bit mask.
|
2018-03-05 12:21:06 +00:00
|
|
|
*
|
|
|
|
* @return
|
|
|
|
* 0 on success, a negative errno value otherwise and rte_errno is set.
|
2016-03-03 14:26:44 +00:00
|
|
|
*/
|
2017-09-01 02:36:28 +00:00
|
|
|
int
|
2016-03-03 14:26:44 +00:00
|
|
|
mlx5_vlan_offload_set(struct rte_eth_dev *dev, int mask)
|
|
|
|
{
|
2019-02-21 09:29:14 +00:00
|
|
|
struct mlx5_priv *priv = dev->data->dev_private;
|
2016-03-03 14:26:44 +00:00
|
|
|
unsigned int i;
|
|
|
|
|
|
|
|
if (mask & ETH_VLAN_STRIP_MASK) {
|
2018-01-10 09:17:01 +00:00
|
|
|
int hw_vlan_strip = !!(dev->data->dev_conf.rxmode.offloads &
|
|
|
|
DEV_RX_OFFLOAD_VLAN_STRIP);
|
2016-03-03 14:26:44 +00:00
|
|
|
|
2018-01-10 09:16:58 +00:00
|
|
|
if (!priv->config.hw_vlan_strip) {
|
2018-03-13 09:23:56 +00:00
|
|
|
DRV_LOG(ERR, "port %u VLAN stripping is not supported",
|
|
|
|
dev->data->port_id);
|
2017-09-01 02:36:28 +00:00
|
|
|
return 0;
|
2016-03-03 14:26:44 +00:00
|
|
|
}
|
|
|
|
/* Run on every RX queue and set/reset VLAN stripping. */
|
|
|
|
for (i = 0; (i != priv->rxqs_n); i++)
|
2018-03-05 12:21:04 +00:00
|
|
|
mlx5_vlan_strip_queue_set(dev, i, hw_vlan_strip);
|
2016-03-03 14:26:44 +00:00
|
|
|
}
|
2017-09-01 02:36:28 +00:00
|
|
|
return 0;
|
2016-03-03 14:26:44 +00:00
|
|
|
}
|
2020-01-29 12:38:47 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Release VLAN network device, created for VM workaround.
|
|
|
|
*
|
|
|
|
* @param[in] dev
|
|
|
|
* Ethernet device object, Netlink context provider.
|
|
|
|
* @param[in] vlan
|
|
|
|
* Object representing the network device to release.
|
|
|
|
*/
|
|
|
|
void mlx5_vlan_vmwa_release(struct rte_eth_dev *dev,
|
|
|
|
struct mlx5_vf_vlan *vlan)
|
|
|
|
{
|
|
|
|
struct mlx5_priv *priv = dev->data->dev_private;
|
|
|
|
struct mlx5_nl_vlan_vmwa_context *vmwa = priv->vmwa_context;
|
|
|
|
struct mlx5_nl_vlan_dev *vlan_dev = &vmwa->vlan_dev[0];
|
|
|
|
|
2020-01-30 16:14:40 +00:00
|
|
|
MLX5_ASSERT(vlan->created);
|
|
|
|
MLX5_ASSERT(priv->vmwa_context);
|
2020-01-29 12:38:47 +00:00
|
|
|
if (!vlan->created || !vmwa)
|
|
|
|
return;
|
|
|
|
vlan->created = 0;
|
2020-01-30 16:14:40 +00:00
|
|
|
MLX5_ASSERT(vlan_dev[vlan->tag].refcnt);
|
2020-01-29 12:38:47 +00:00
|
|
|
if (--vlan_dev[vlan->tag].refcnt == 0 &&
|
|
|
|
vlan_dev[vlan->tag].ifindex) {
|
|
|
|
mlx5_nl_vlan_vmwa_delete(vmwa, vlan_dev[vlan->tag].ifindex);
|
|
|
|
vlan_dev[vlan->tag].ifindex = 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* Acquire VLAN interface with specified tag for VM workaround.
|
|
|
|
*
|
|
|
|
* @param[in] dev
|
|
|
|
* Ethernet device object, Netlink context provider.
|
|
|
|
* @param[in] vlan
|
|
|
|
* Object representing the network device to acquire.
|
|
|
|
*/
|
|
|
|
void mlx5_vlan_vmwa_acquire(struct rte_eth_dev *dev,
|
|
|
|
struct mlx5_vf_vlan *vlan)
|
|
|
|
{
|
|
|
|
struct mlx5_priv *priv = dev->data->dev_private;
|
|
|
|
struct mlx5_nl_vlan_vmwa_context *vmwa = priv->vmwa_context;
|
|
|
|
struct mlx5_nl_vlan_dev *vlan_dev = &vmwa->vlan_dev[0];
|
|
|
|
|
2020-01-30 16:14:40 +00:00
|
|
|
MLX5_ASSERT(!vlan->created);
|
|
|
|
MLX5_ASSERT(priv->vmwa_context);
|
2020-01-29 12:38:47 +00:00
|
|
|
if (vlan->created || !vmwa)
|
|
|
|
return;
|
|
|
|
if (vlan_dev[vlan->tag].refcnt == 0) {
|
2020-01-30 16:14:40 +00:00
|
|
|
MLX5_ASSERT(!vlan_dev[vlan->tag].ifindex);
|
2020-01-29 12:38:47 +00:00
|
|
|
vlan_dev[vlan->tag].ifindex =
|
|
|
|
mlx5_nl_vlan_vmwa_create(vmwa, vmwa->vf_ifindex,
|
|
|
|
vlan->tag);
|
|
|
|
}
|
|
|
|
if (vlan_dev[vlan->tag].ifindex) {
|
|
|
|
vlan_dev[vlan->tag].refcnt++;
|
|
|
|
vlan->created = 1;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Create per ethernet device VLAN VM workaround context
|
|
|
|
*/
|
|
|
|
struct mlx5_nl_vlan_vmwa_context *
|
|
|
|
mlx5_vlan_vmwa_init(struct rte_eth_dev *dev, uint32_t ifindex)
|
|
|
|
{
|
|
|
|
struct mlx5_priv *priv = dev->data->dev_private;
|
|
|
|
struct mlx5_dev_config *config = &priv->config;
|
|
|
|
struct mlx5_nl_vlan_vmwa_context *vmwa;
|
|
|
|
enum rte_hypervisor hv_type;
|
|
|
|
|
|
|
|
/* Do not engage workaround over PF. */
|
|
|
|
if (!config->vf)
|
|
|
|
return NULL;
|
|
|
|
/* Check whether there is desired virtual environment */
|
|
|
|
hv_type = rte_hypervisor_get();
|
|
|
|
switch (hv_type) {
|
|
|
|
case RTE_HYPERVISOR_UNKNOWN:
|
|
|
|
case RTE_HYPERVISOR_VMWARE:
|
|
|
|
/*
|
|
|
|
* The "white list" of configurations
|
|
|
|
* to engage the workaround.
|
|
|
|
*/
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
/*
|
|
|
|
* The configuration is not found in the "white list".
|
|
|
|
* We should not engage the VLAN workaround.
|
|
|
|
*/
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
vmwa = rte_zmalloc(__func__, sizeof(*vmwa), sizeof(uint32_t));
|
|
|
|
if (!vmwa) {
|
|
|
|
DRV_LOG(WARNING,
|
|
|
|
"Can not allocate memory"
|
|
|
|
" for VLAN workaround context");
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
vmwa->nl_socket = mlx5_nl_init(NETLINK_ROUTE);
|
|
|
|
if (vmwa->nl_socket < 0) {
|
|
|
|
DRV_LOG(WARNING,
|
|
|
|
"Can not create Netlink socket"
|
|
|
|
" for VLAN workaround context");
|
|
|
|
rte_free(vmwa);
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
vmwa->vf_ifindex = ifindex;
|
|
|
|
/* Cleanup for existing VLAN devices. */
|
|
|
|
return vmwa;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Destroy per ethernet device VLAN VM workaround context
|
|
|
|
*/
|
|
|
|
void mlx5_vlan_vmwa_exit(struct mlx5_nl_vlan_vmwa_context *vmwa)
|
|
|
|
{
|
|
|
|
unsigned int i;
|
|
|
|
|
|
|
|
/* Delete all remaining VLAN devices. */
|
|
|
|
for (i = 0; i < RTE_DIM(vmwa->vlan_dev); i++) {
|
|
|
|
if (vmwa->vlan_dev[i].ifindex)
|
|
|
|
mlx5_nl_vlan_vmwa_delete(vmwa,
|
|
|
|
vmwa->vlan_dev[i].ifindex);
|
|
|
|
}
|
|
|
|
if (vmwa->nl_socket >= 0)
|
|
|
|
close(vmwa->nl_socket);
|
|
|
|
rte_free(vmwa);
|
|
|
|
}
|