a7db3afce7
Added macros to simplify print of MAC address. The six bytes of a MAC address are extracted in a macro here, to improve code readablity. Signed-off-by: Aman Deep Singh <aman.deep.singh@intel.com> Reviewed-by: Ferruh Yigit <ferruh.yigit@intel.com>
1737 lines
47 KiB
C
1737 lines
47 KiB
C
/* SPDX-License-Identifier: BSD-3-Clause
|
|
* Copyright(c) 2017 Intel Corporation
|
|
*/
|
|
|
|
#include <stdio.h>
|
|
#include <errno.h>
|
|
#include <stdint.h>
|
|
#include <string.h>
|
|
#include <unistd.h>
|
|
#include <stdarg.h>
|
|
#include <inttypes.h>
|
|
#include <rte_byteorder.h>
|
|
#include <rte_common.h>
|
|
|
|
#include <rte_debug.h>
|
|
#include <rte_atomic.h>
|
|
#include <rte_eal.h>
|
|
#include <rte_ether.h>
|
|
#include <ethdev_driver.h>
|
|
#include <ethdev_pci.h>
|
|
#include <rte_dev.h>
|
|
|
|
#include "iavf.h"
|
|
#include "iavf_rxtx.h"
|
|
|
|
#define MAX_TRY_TIMES 200
|
|
#define ASQ_DELAY_MS 10
|
|
|
|
static uint32_t
|
|
iavf_convert_link_speed(enum virtchnl_link_speed virt_link_speed)
|
|
{
|
|
uint32_t speed;
|
|
|
|
switch (virt_link_speed) {
|
|
case VIRTCHNL_LINK_SPEED_100MB:
|
|
speed = 100;
|
|
break;
|
|
case VIRTCHNL_LINK_SPEED_1GB:
|
|
speed = 1000;
|
|
break;
|
|
case VIRTCHNL_LINK_SPEED_10GB:
|
|
speed = 10000;
|
|
break;
|
|
case VIRTCHNL_LINK_SPEED_40GB:
|
|
speed = 40000;
|
|
break;
|
|
case VIRTCHNL_LINK_SPEED_20GB:
|
|
speed = 20000;
|
|
break;
|
|
case VIRTCHNL_LINK_SPEED_25GB:
|
|
speed = 25000;
|
|
break;
|
|
case VIRTCHNL_LINK_SPEED_2_5GB:
|
|
speed = 2500;
|
|
break;
|
|
case VIRTCHNL_LINK_SPEED_5GB:
|
|
speed = 5000;
|
|
break;
|
|
default:
|
|
speed = 0;
|
|
break;
|
|
}
|
|
|
|
return speed;
|
|
}
|
|
|
|
/* Read data in admin queue to get msg from pf driver */
|
|
static enum iavf_aq_result
|
|
iavf_read_msg_from_pf(struct iavf_adapter *adapter, uint16_t buf_len,
|
|
uint8_t *buf)
|
|
{
|
|
struct iavf_hw *hw = IAVF_DEV_PRIVATE_TO_HW(adapter);
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct rte_eth_dev *dev = adapter->eth_dev;
|
|
struct iavf_arq_event_info event;
|
|
enum iavf_aq_result result = IAVF_MSG_NON;
|
|
enum virtchnl_ops opcode;
|
|
int ret;
|
|
|
|
event.buf_len = buf_len;
|
|
event.msg_buf = buf;
|
|
ret = iavf_clean_arq_element(hw, &event, NULL);
|
|
/* Can't read any msg from adminQ */
|
|
if (ret) {
|
|
PMD_DRV_LOG(DEBUG, "Can't read msg from AQ");
|
|
if (ret != IAVF_ERR_ADMIN_QUEUE_NO_WORK)
|
|
result = IAVF_MSG_ERR;
|
|
return result;
|
|
}
|
|
|
|
opcode = (enum virtchnl_ops)rte_le_to_cpu_32(event.desc.cookie_high);
|
|
vf->cmd_retval = (enum virtchnl_status_code)rte_le_to_cpu_32(
|
|
event.desc.cookie_low);
|
|
|
|
PMD_DRV_LOG(DEBUG, "AQ from pf carries opcode %u, retval %d",
|
|
opcode, vf->cmd_retval);
|
|
|
|
if (opcode == VIRTCHNL_OP_EVENT) {
|
|
struct virtchnl_pf_event *vpe =
|
|
(struct virtchnl_pf_event *)event.msg_buf;
|
|
|
|
result = IAVF_MSG_SYS;
|
|
switch (vpe->event) {
|
|
case VIRTCHNL_EVENT_LINK_CHANGE:
|
|
vf->link_up =
|
|
vpe->event_data.link_event.link_status;
|
|
if (vf->vf_res->vf_cap_flags &
|
|
VIRTCHNL_VF_CAP_ADV_LINK_SPEED) {
|
|
vf->link_speed =
|
|
vpe->event_data.link_event_adv.link_speed;
|
|
} else {
|
|
enum virtchnl_link_speed speed;
|
|
speed = vpe->event_data.link_event.link_speed;
|
|
vf->link_speed = iavf_convert_link_speed(speed);
|
|
}
|
|
iavf_dev_link_update(dev, 0);
|
|
PMD_DRV_LOG(INFO, "Link status update:%s",
|
|
vf->link_up ? "up" : "down");
|
|
break;
|
|
case VIRTCHNL_EVENT_RESET_IMPENDING:
|
|
vf->vf_reset = true;
|
|
PMD_DRV_LOG(INFO, "VF is resetting");
|
|
break;
|
|
case VIRTCHNL_EVENT_PF_DRIVER_CLOSE:
|
|
vf->dev_closed = true;
|
|
PMD_DRV_LOG(INFO, "PF driver closed");
|
|
break;
|
|
default:
|
|
PMD_DRV_LOG(ERR, "%s: Unknown event %d from pf",
|
|
__func__, vpe->event);
|
|
}
|
|
} else {
|
|
/* async reply msg on command issued by vf previously */
|
|
result = IAVF_MSG_CMD;
|
|
if (opcode != vf->pend_cmd) {
|
|
PMD_DRV_LOG(WARNING, "command mismatch, expect %u, get %u",
|
|
vf->pend_cmd, opcode);
|
|
result = IAVF_MSG_ERR;
|
|
}
|
|
}
|
|
|
|
return result;
|
|
}
|
|
|
|
static int
|
|
iavf_execute_vf_cmd(struct iavf_adapter *adapter, struct iavf_cmd_info *args)
|
|
{
|
|
struct iavf_hw *hw = IAVF_DEV_PRIVATE_TO_HW(adapter);
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
enum iavf_aq_result result;
|
|
enum iavf_status ret;
|
|
int err = 0;
|
|
int i = 0;
|
|
|
|
if (vf->vf_reset)
|
|
return -EIO;
|
|
|
|
if (_atomic_set_cmd(vf, args->ops))
|
|
return -1;
|
|
|
|
ret = iavf_aq_send_msg_to_pf(hw, args->ops, IAVF_SUCCESS,
|
|
args->in_args, args->in_args_size, NULL);
|
|
if (ret) {
|
|
PMD_DRV_LOG(ERR, "fail to send cmd %d", args->ops);
|
|
_clear_cmd(vf);
|
|
return err;
|
|
}
|
|
|
|
switch (args->ops) {
|
|
case VIRTCHNL_OP_RESET_VF:
|
|
/*no need to wait for response */
|
|
_clear_cmd(vf);
|
|
break;
|
|
case VIRTCHNL_OP_VERSION:
|
|
case VIRTCHNL_OP_GET_VF_RESOURCES:
|
|
case VIRTCHNL_OP_GET_SUPPORTED_RXDIDS:
|
|
case VIRTCHNL_OP_GET_OFFLOAD_VLAN_V2_CAPS:
|
|
/* for init virtchnl ops, need to poll the response */
|
|
do {
|
|
result = iavf_read_msg_from_pf(adapter, args->out_size,
|
|
args->out_buffer);
|
|
if (result == IAVF_MSG_CMD)
|
|
break;
|
|
rte_delay_ms(ASQ_DELAY_MS);
|
|
} while (i++ < MAX_TRY_TIMES);
|
|
if (i >= MAX_TRY_TIMES ||
|
|
vf->cmd_retval != VIRTCHNL_STATUS_SUCCESS) {
|
|
err = -1;
|
|
PMD_DRV_LOG(ERR, "No response or return failure (%d)"
|
|
" for cmd %d", vf->cmd_retval, args->ops);
|
|
}
|
|
_clear_cmd(vf);
|
|
break;
|
|
case VIRTCHNL_OP_REQUEST_QUEUES:
|
|
/*
|
|
* ignore async reply, only wait for system message,
|
|
* vf_reset = true if get VIRTCHNL_EVENT_RESET_IMPENDING,
|
|
* if not, means request queues failed.
|
|
*/
|
|
do {
|
|
result = iavf_read_msg_from_pf(adapter, args->out_size,
|
|
args->out_buffer);
|
|
if (result == IAVF_MSG_SYS && vf->vf_reset) {
|
|
break;
|
|
} else if (result == IAVF_MSG_CMD ||
|
|
result == IAVF_MSG_ERR) {
|
|
err = -1;
|
|
break;
|
|
}
|
|
rte_delay_ms(ASQ_DELAY_MS);
|
|
/* If don't read msg or read sys event, continue */
|
|
} while (i++ < MAX_TRY_TIMES);
|
|
if (i >= MAX_TRY_TIMES ||
|
|
vf->cmd_retval != VIRTCHNL_STATUS_SUCCESS) {
|
|
err = -1;
|
|
PMD_DRV_LOG(ERR, "No response or return failure (%d)"
|
|
" for cmd %d", vf->cmd_retval, args->ops);
|
|
}
|
|
_clear_cmd(vf);
|
|
break;
|
|
default:
|
|
/* For other virtchnl ops in running time,
|
|
* wait for the cmd done flag.
|
|
*/
|
|
do {
|
|
if (vf->pend_cmd == VIRTCHNL_OP_UNKNOWN)
|
|
break;
|
|
rte_delay_ms(ASQ_DELAY_MS);
|
|
/* If don't read msg or read sys event, continue */
|
|
} while (i++ < MAX_TRY_TIMES);
|
|
|
|
if (i >= MAX_TRY_TIMES) {
|
|
PMD_DRV_LOG(ERR, "No response for cmd %d", args->ops);
|
|
_clear_cmd(vf);
|
|
err = -EIO;
|
|
} else if (vf->cmd_retval ==
|
|
VIRTCHNL_STATUS_ERR_NOT_SUPPORTED) {
|
|
PMD_DRV_LOG(ERR, "Cmd %d not supported", args->ops);
|
|
err = -ENOTSUP;
|
|
} else if (vf->cmd_retval != VIRTCHNL_STATUS_SUCCESS) {
|
|
PMD_DRV_LOG(ERR, "Return failure %d for cmd %d",
|
|
vf->cmd_retval, args->ops);
|
|
err = -EINVAL;
|
|
}
|
|
break;
|
|
}
|
|
|
|
return err;
|
|
}
|
|
|
|
static void
|
|
iavf_handle_pf_event_msg(struct rte_eth_dev *dev, uint8_t *msg,
|
|
uint16_t msglen)
|
|
{
|
|
struct virtchnl_pf_event *pf_msg =
|
|
(struct virtchnl_pf_event *)msg;
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(dev->data->dev_private);
|
|
|
|
if (msglen < sizeof(struct virtchnl_pf_event)) {
|
|
PMD_DRV_LOG(DEBUG, "Error event");
|
|
return;
|
|
}
|
|
switch (pf_msg->event) {
|
|
case VIRTCHNL_EVENT_RESET_IMPENDING:
|
|
PMD_DRV_LOG(DEBUG, "VIRTCHNL_EVENT_RESET_IMPENDING event");
|
|
vf->vf_reset = true;
|
|
rte_eth_dev_callback_process(dev, RTE_ETH_EVENT_INTR_RESET,
|
|
NULL);
|
|
break;
|
|
case VIRTCHNL_EVENT_LINK_CHANGE:
|
|
PMD_DRV_LOG(DEBUG, "VIRTCHNL_EVENT_LINK_CHANGE event");
|
|
vf->link_up = pf_msg->event_data.link_event.link_status;
|
|
if (vf->vf_res->vf_cap_flags & VIRTCHNL_VF_CAP_ADV_LINK_SPEED) {
|
|
vf->link_speed =
|
|
pf_msg->event_data.link_event_adv.link_speed;
|
|
} else {
|
|
enum virtchnl_link_speed speed;
|
|
speed = pf_msg->event_data.link_event.link_speed;
|
|
vf->link_speed = iavf_convert_link_speed(speed);
|
|
}
|
|
iavf_dev_link_update(dev, 0);
|
|
rte_eth_dev_callback_process(dev, RTE_ETH_EVENT_INTR_LSC, NULL);
|
|
break;
|
|
case VIRTCHNL_EVENT_PF_DRIVER_CLOSE:
|
|
PMD_DRV_LOG(DEBUG, "VIRTCHNL_EVENT_PF_DRIVER_CLOSE event");
|
|
break;
|
|
default:
|
|
PMD_DRV_LOG(ERR, " unknown event received %u", pf_msg->event);
|
|
break;
|
|
}
|
|
}
|
|
|
|
void
|
|
iavf_handle_virtchnl_msg(struct rte_eth_dev *dev)
|
|
{
|
|
struct iavf_hw *hw = IAVF_DEV_PRIVATE_TO_HW(dev->data->dev_private);
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(dev->data->dev_private);
|
|
struct iavf_arq_event_info info;
|
|
uint16_t pending, aq_opc;
|
|
enum virtchnl_ops msg_opc;
|
|
enum iavf_status msg_ret;
|
|
int ret;
|
|
|
|
info.buf_len = IAVF_AQ_BUF_SZ;
|
|
if (!vf->aq_resp) {
|
|
PMD_DRV_LOG(ERR, "Buffer for adminq resp should not be NULL");
|
|
return;
|
|
}
|
|
info.msg_buf = vf->aq_resp;
|
|
|
|
pending = 1;
|
|
while (pending) {
|
|
ret = iavf_clean_arq_element(hw, &info, &pending);
|
|
|
|
if (ret != IAVF_SUCCESS) {
|
|
PMD_DRV_LOG(INFO, "Failed to read msg from AdminQ,"
|
|
"ret: %d", ret);
|
|
break;
|
|
}
|
|
aq_opc = rte_le_to_cpu_16(info.desc.opcode);
|
|
/* For the message sent from pf to vf, opcode is stored in
|
|
* cookie_high of struct iavf_aq_desc, while return error code
|
|
* are stored in cookie_low, Which is done by PF driver.
|
|
*/
|
|
msg_opc = (enum virtchnl_ops)rte_le_to_cpu_32(
|
|
info.desc.cookie_high);
|
|
msg_ret = (enum iavf_status)rte_le_to_cpu_32(
|
|
info.desc.cookie_low);
|
|
switch (aq_opc) {
|
|
case iavf_aqc_opc_send_msg_to_vf:
|
|
if (msg_opc == VIRTCHNL_OP_EVENT) {
|
|
iavf_handle_pf_event_msg(dev, info.msg_buf,
|
|
info.msg_len);
|
|
} else {
|
|
/* read message and it's expected one */
|
|
if (msg_opc == vf->pend_cmd)
|
|
_notify_cmd(vf, msg_ret);
|
|
else
|
|
PMD_DRV_LOG(ERR, "command mismatch,"
|
|
"expect %u, get %u",
|
|
vf->pend_cmd, msg_opc);
|
|
PMD_DRV_LOG(DEBUG,
|
|
"adminq response is received,"
|
|
" opcode = %d", msg_opc);
|
|
}
|
|
break;
|
|
default:
|
|
PMD_DRV_LOG(DEBUG, "Request %u is not supported yet",
|
|
aq_opc);
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
|
|
int
|
|
iavf_enable_vlan_strip(struct iavf_adapter *adapter)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct iavf_cmd_info args;
|
|
int ret;
|
|
|
|
memset(&args, 0, sizeof(args));
|
|
args.ops = VIRTCHNL_OP_ENABLE_VLAN_STRIPPING;
|
|
args.in_args = NULL;
|
|
args.in_args_size = 0;
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
ret = iavf_execute_vf_cmd(adapter, &args);
|
|
if (ret)
|
|
PMD_DRV_LOG(ERR, "Failed to execute command of"
|
|
" OP_ENABLE_VLAN_STRIPPING");
|
|
|
|
return ret;
|
|
}
|
|
|
|
int
|
|
iavf_disable_vlan_strip(struct iavf_adapter *adapter)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct iavf_cmd_info args;
|
|
int ret;
|
|
|
|
memset(&args, 0, sizeof(args));
|
|
args.ops = VIRTCHNL_OP_DISABLE_VLAN_STRIPPING;
|
|
args.in_args = NULL;
|
|
args.in_args_size = 0;
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
ret = iavf_execute_vf_cmd(adapter, &args);
|
|
if (ret)
|
|
PMD_DRV_LOG(ERR, "Failed to execute command of"
|
|
" OP_DISABLE_VLAN_STRIPPING");
|
|
|
|
return ret;
|
|
}
|
|
|
|
#define VIRTCHNL_VERSION_MAJOR_START 1
|
|
#define VIRTCHNL_VERSION_MINOR_START 1
|
|
|
|
/* Check API version with sync wait until version read from admin queue */
|
|
int
|
|
iavf_check_api_version(struct iavf_adapter *adapter)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_version_info version, *pver;
|
|
struct iavf_cmd_info args;
|
|
int err;
|
|
|
|
version.major = VIRTCHNL_VERSION_MAJOR;
|
|
version.minor = VIRTCHNL_VERSION_MINOR;
|
|
|
|
args.ops = VIRTCHNL_OP_VERSION;
|
|
args.in_args = (uint8_t *)&version;
|
|
args.in_args_size = sizeof(version);
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
|
|
err = iavf_execute_vf_cmd(adapter, &args);
|
|
if (err) {
|
|
PMD_INIT_LOG(ERR, "Fail to execute command of OP_VERSION");
|
|
return err;
|
|
}
|
|
|
|
pver = (struct virtchnl_version_info *)args.out_buffer;
|
|
vf->virtchnl_version = *pver;
|
|
|
|
if (vf->virtchnl_version.major < VIRTCHNL_VERSION_MAJOR_START ||
|
|
(vf->virtchnl_version.major == VIRTCHNL_VERSION_MAJOR_START &&
|
|
vf->virtchnl_version.minor < VIRTCHNL_VERSION_MINOR_START)) {
|
|
PMD_INIT_LOG(ERR, "VIRTCHNL API version should not be lower"
|
|
" than (%u.%u) to support Adapative VF",
|
|
VIRTCHNL_VERSION_MAJOR_START,
|
|
VIRTCHNL_VERSION_MAJOR_START);
|
|
return -1;
|
|
} else if (vf->virtchnl_version.major > VIRTCHNL_VERSION_MAJOR ||
|
|
(vf->virtchnl_version.major == VIRTCHNL_VERSION_MAJOR &&
|
|
vf->virtchnl_version.minor > VIRTCHNL_VERSION_MINOR)) {
|
|
PMD_INIT_LOG(ERR, "PF/VF API version mismatch:(%u.%u)-(%u.%u)",
|
|
vf->virtchnl_version.major,
|
|
vf->virtchnl_version.minor,
|
|
VIRTCHNL_VERSION_MAJOR,
|
|
VIRTCHNL_VERSION_MINOR);
|
|
return -1;
|
|
}
|
|
|
|
PMD_DRV_LOG(DEBUG, "Peer is supported PF host");
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
iavf_get_vf_resource(struct iavf_adapter *adapter)
|
|
{
|
|
struct iavf_hw *hw = IAVF_DEV_PRIVATE_TO_HW(adapter);
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct iavf_cmd_info args;
|
|
uint32_t caps, len;
|
|
int err, i;
|
|
|
|
args.ops = VIRTCHNL_OP_GET_VF_RESOURCES;
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
|
|
caps = IAVF_BASIC_OFFLOAD_CAPS | VIRTCHNL_VF_CAP_ADV_LINK_SPEED |
|
|
VIRTCHNL_VF_OFFLOAD_RX_FLEX_DESC |
|
|
VIRTCHNL_VF_OFFLOAD_FDIR_PF |
|
|
VIRTCHNL_VF_OFFLOAD_ADV_RSS_PF |
|
|
VIRTCHNL_VF_OFFLOAD_REQ_QUEUES |
|
|
VIRTCHNL_VF_OFFLOAD_CRC |
|
|
VIRTCHNL_VF_OFFLOAD_VLAN_V2 |
|
|
VIRTCHNL_VF_LARGE_NUM_QPAIRS |
|
|
VIRTCHNL_VF_OFFLOAD_QOS;
|
|
|
|
args.in_args = (uint8_t *)∩︀
|
|
args.in_args_size = sizeof(caps);
|
|
|
|
err = iavf_execute_vf_cmd(adapter, &args);
|
|
|
|
if (err) {
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to execute command of OP_GET_VF_RESOURCE");
|
|
return -1;
|
|
}
|
|
|
|
len = sizeof(struct virtchnl_vf_resource) +
|
|
IAVF_MAX_VF_VSI * sizeof(struct virtchnl_vsi_resource);
|
|
|
|
rte_memcpy(vf->vf_res, args.out_buffer,
|
|
RTE_MIN(args.out_size, len));
|
|
/* parse VF config message back from PF*/
|
|
iavf_vf_parse_hw_config(hw, vf->vf_res);
|
|
for (i = 0; i < vf->vf_res->num_vsis; i++) {
|
|
if (vf->vf_res->vsi_res[i].vsi_type == VIRTCHNL_VSI_SRIOV)
|
|
vf->vsi_res = &vf->vf_res->vsi_res[i];
|
|
}
|
|
|
|
if (!vf->vsi_res) {
|
|
PMD_INIT_LOG(ERR, "no LAN VSI found");
|
|
return -1;
|
|
}
|
|
|
|
vf->vsi.vsi_id = vf->vsi_res->vsi_id;
|
|
vf->vsi.nb_qps = vf->vsi_res->num_queue_pairs;
|
|
vf->vsi.adapter = adapter;
|
|
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
iavf_get_supported_rxdid(struct iavf_adapter *adapter)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct iavf_cmd_info args;
|
|
int ret;
|
|
|
|
args.ops = VIRTCHNL_OP_GET_SUPPORTED_RXDIDS;
|
|
args.in_args = NULL;
|
|
args.in_args_size = 0;
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
|
|
ret = iavf_execute_vf_cmd(adapter, &args);
|
|
if (ret) {
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to execute command of OP_GET_SUPPORTED_RXDIDS");
|
|
return ret;
|
|
}
|
|
|
|
vf->supported_rxdid =
|
|
((struct virtchnl_supported_rxdids *)args.out_buffer)->supported_rxdids;
|
|
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
iavf_config_vlan_strip_v2(struct iavf_adapter *adapter, bool enable)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_vlan_supported_caps *stripping_caps;
|
|
struct virtchnl_vlan_setting vlan_strip;
|
|
struct iavf_cmd_info args;
|
|
uint32_t *ethertype;
|
|
int ret;
|
|
|
|
stripping_caps = &vf->vlan_v2_caps.offloads.stripping_support;
|
|
|
|
if ((stripping_caps->outer & VIRTCHNL_VLAN_ETHERTYPE_8100) &&
|
|
(stripping_caps->outer & VIRTCHNL_VLAN_TOGGLE))
|
|
ethertype = &vlan_strip.outer_ethertype_setting;
|
|
else if ((stripping_caps->inner & VIRTCHNL_VLAN_ETHERTYPE_8100) &&
|
|
(stripping_caps->inner & VIRTCHNL_VLAN_TOGGLE))
|
|
ethertype = &vlan_strip.inner_ethertype_setting;
|
|
else
|
|
return -ENOTSUP;
|
|
|
|
memset(&vlan_strip, 0, sizeof(vlan_strip));
|
|
vlan_strip.vport_id = vf->vsi_res->vsi_id;
|
|
*ethertype = VIRTCHNL_VLAN_ETHERTYPE_8100;
|
|
|
|
args.ops = enable ? VIRTCHNL_OP_ENABLE_VLAN_STRIPPING_V2 :
|
|
VIRTCHNL_OP_DISABLE_VLAN_STRIPPING_V2;
|
|
args.in_args = (uint8_t *)&vlan_strip;
|
|
args.in_args_size = sizeof(vlan_strip);
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
ret = iavf_execute_vf_cmd(adapter, &args);
|
|
if (ret)
|
|
PMD_DRV_LOG(ERR, "fail to execute command %s",
|
|
enable ? "VIRTCHNL_OP_ENABLE_VLAN_STRIPPING_V2" :
|
|
"VIRTCHNL_OP_DISABLE_VLAN_STRIPPING_V2");
|
|
|
|
return ret;
|
|
}
|
|
|
|
int
|
|
iavf_config_vlan_insert_v2(struct iavf_adapter *adapter, bool enable)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_vlan_supported_caps *insertion_caps;
|
|
struct virtchnl_vlan_setting vlan_insert;
|
|
struct iavf_cmd_info args;
|
|
uint32_t *ethertype;
|
|
int ret;
|
|
|
|
insertion_caps = &vf->vlan_v2_caps.offloads.insertion_support;
|
|
|
|
if ((insertion_caps->outer & VIRTCHNL_VLAN_ETHERTYPE_8100) &&
|
|
(insertion_caps->outer & VIRTCHNL_VLAN_TOGGLE))
|
|
ethertype = &vlan_insert.outer_ethertype_setting;
|
|
else if ((insertion_caps->inner & VIRTCHNL_VLAN_ETHERTYPE_8100) &&
|
|
(insertion_caps->inner & VIRTCHNL_VLAN_TOGGLE))
|
|
ethertype = &vlan_insert.inner_ethertype_setting;
|
|
else
|
|
return -ENOTSUP;
|
|
|
|
memset(&vlan_insert, 0, sizeof(vlan_insert));
|
|
vlan_insert.vport_id = vf->vsi_res->vsi_id;
|
|
*ethertype = VIRTCHNL_VLAN_ETHERTYPE_8100;
|
|
|
|
args.ops = enable ? VIRTCHNL_OP_ENABLE_VLAN_INSERTION_V2 :
|
|
VIRTCHNL_OP_DISABLE_VLAN_INSERTION_V2;
|
|
args.in_args = (uint8_t *)&vlan_insert;
|
|
args.in_args_size = sizeof(vlan_insert);
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
ret = iavf_execute_vf_cmd(adapter, &args);
|
|
if (ret)
|
|
PMD_DRV_LOG(ERR, "fail to execute command %s",
|
|
enable ? "VIRTCHNL_OP_ENABLE_VLAN_INSERTION_V2" :
|
|
"VIRTCHNL_OP_DISABLE_VLAN_INSERTION_V2");
|
|
|
|
return ret;
|
|
}
|
|
|
|
int
|
|
iavf_add_del_vlan_v2(struct iavf_adapter *adapter, uint16_t vlanid, bool add)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_vlan_supported_caps *supported_caps;
|
|
struct virtchnl_vlan_filter_list_v2 vlan_filter;
|
|
struct virtchnl_vlan *vlan_setting;
|
|
struct iavf_cmd_info args;
|
|
uint32_t filtering_caps;
|
|
int err;
|
|
|
|
supported_caps = &vf->vlan_v2_caps.filtering.filtering_support;
|
|
if (supported_caps->outer) {
|
|
filtering_caps = supported_caps->outer;
|
|
vlan_setting = &vlan_filter.filters[0].outer;
|
|
} else {
|
|
filtering_caps = supported_caps->inner;
|
|
vlan_setting = &vlan_filter.filters[0].inner;
|
|
}
|
|
|
|
if (!(filtering_caps & VIRTCHNL_VLAN_ETHERTYPE_8100))
|
|
return -ENOTSUP;
|
|
|
|
memset(&vlan_filter, 0, sizeof(vlan_filter));
|
|
vlan_filter.vport_id = vf->vsi_res->vsi_id;
|
|
vlan_filter.num_elements = 1;
|
|
vlan_setting->tpid = RTE_ETHER_TYPE_VLAN;
|
|
vlan_setting->tci = vlanid;
|
|
|
|
args.ops = add ? VIRTCHNL_OP_ADD_VLAN_V2 : VIRTCHNL_OP_DEL_VLAN_V2;
|
|
args.in_args = (uint8_t *)&vlan_filter;
|
|
args.in_args_size = sizeof(vlan_filter);
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
err = iavf_execute_vf_cmd(adapter, &args);
|
|
if (err)
|
|
PMD_DRV_LOG(ERR, "fail to execute command %s",
|
|
add ? "OP_ADD_VLAN_V2" : "OP_DEL_VLAN_V2");
|
|
|
|
return err;
|
|
}
|
|
|
|
int
|
|
iavf_get_vlan_offload_caps_v2(struct iavf_adapter *adapter)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct iavf_cmd_info args;
|
|
int ret;
|
|
|
|
args.ops = VIRTCHNL_OP_GET_OFFLOAD_VLAN_V2_CAPS;
|
|
args.in_args = NULL;
|
|
args.in_args_size = 0;
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
|
|
ret = iavf_execute_vf_cmd(adapter, &args);
|
|
if (ret) {
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to execute command of VIRTCHNL_OP_GET_OFFLOAD_VLAN_V2_CAPS");
|
|
return ret;
|
|
}
|
|
|
|
rte_memcpy(&vf->vlan_v2_caps, vf->aq_resp, sizeof(vf->vlan_v2_caps));
|
|
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
iavf_enable_queues(struct iavf_adapter *adapter)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_queue_select queue_select;
|
|
struct iavf_cmd_info args;
|
|
int err;
|
|
|
|
memset(&queue_select, 0, sizeof(queue_select));
|
|
queue_select.vsi_id = vf->vsi_res->vsi_id;
|
|
|
|
queue_select.rx_queues = BIT(adapter->eth_dev->data->nb_rx_queues) - 1;
|
|
queue_select.tx_queues = BIT(adapter->eth_dev->data->nb_tx_queues) - 1;
|
|
|
|
args.ops = VIRTCHNL_OP_ENABLE_QUEUES;
|
|
args.in_args = (u8 *)&queue_select;
|
|
args.in_args_size = sizeof(queue_select);
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
err = iavf_execute_vf_cmd(adapter, &args);
|
|
if (err) {
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to execute command of OP_ENABLE_QUEUES");
|
|
return err;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
iavf_disable_queues(struct iavf_adapter *adapter)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_queue_select queue_select;
|
|
struct iavf_cmd_info args;
|
|
int err;
|
|
|
|
memset(&queue_select, 0, sizeof(queue_select));
|
|
queue_select.vsi_id = vf->vsi_res->vsi_id;
|
|
|
|
queue_select.rx_queues = BIT(adapter->eth_dev->data->nb_rx_queues) - 1;
|
|
queue_select.tx_queues = BIT(adapter->eth_dev->data->nb_tx_queues) - 1;
|
|
|
|
args.ops = VIRTCHNL_OP_DISABLE_QUEUES;
|
|
args.in_args = (u8 *)&queue_select;
|
|
args.in_args_size = sizeof(queue_select);
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
err = iavf_execute_vf_cmd(adapter, &args);
|
|
if (err) {
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to execute command of OP_DISABLE_QUEUES");
|
|
return err;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
iavf_switch_queue(struct iavf_adapter *adapter, uint16_t qid,
|
|
bool rx, bool on)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_queue_select queue_select;
|
|
struct iavf_cmd_info args;
|
|
int err;
|
|
|
|
memset(&queue_select, 0, sizeof(queue_select));
|
|
queue_select.vsi_id = vf->vsi_res->vsi_id;
|
|
if (rx)
|
|
queue_select.rx_queues |= 1 << qid;
|
|
else
|
|
queue_select.tx_queues |= 1 << qid;
|
|
|
|
if (on)
|
|
args.ops = VIRTCHNL_OP_ENABLE_QUEUES;
|
|
else
|
|
args.ops = VIRTCHNL_OP_DISABLE_QUEUES;
|
|
args.in_args = (u8 *)&queue_select;
|
|
args.in_args_size = sizeof(queue_select);
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
err = iavf_execute_vf_cmd(adapter, &args);
|
|
if (err)
|
|
PMD_DRV_LOG(ERR, "Failed to execute command of %s",
|
|
on ? "OP_ENABLE_QUEUES" : "OP_DISABLE_QUEUES");
|
|
return err;
|
|
}
|
|
|
|
int
|
|
iavf_enable_queues_lv(struct iavf_adapter *adapter)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_del_ena_dis_queues *queue_select;
|
|
struct virtchnl_queue_chunk *queue_chunk;
|
|
struct iavf_cmd_info args;
|
|
int err, len;
|
|
|
|
len = sizeof(struct virtchnl_del_ena_dis_queues) +
|
|
sizeof(struct virtchnl_queue_chunk) *
|
|
(IAVF_RXTX_QUEUE_CHUNKS_NUM - 1);
|
|
queue_select = rte_zmalloc("queue_select", len, 0);
|
|
if (!queue_select)
|
|
return -ENOMEM;
|
|
|
|
queue_chunk = queue_select->chunks.chunks;
|
|
queue_select->chunks.num_chunks = IAVF_RXTX_QUEUE_CHUNKS_NUM;
|
|
queue_select->vport_id = vf->vsi_res->vsi_id;
|
|
|
|
queue_chunk[VIRTCHNL_QUEUE_TYPE_TX].type = VIRTCHNL_QUEUE_TYPE_TX;
|
|
queue_chunk[VIRTCHNL_QUEUE_TYPE_TX].start_queue_id = 0;
|
|
queue_chunk[VIRTCHNL_QUEUE_TYPE_TX].num_queues =
|
|
adapter->eth_dev->data->nb_tx_queues;
|
|
|
|
queue_chunk[VIRTCHNL_QUEUE_TYPE_RX].type = VIRTCHNL_QUEUE_TYPE_RX;
|
|
queue_chunk[VIRTCHNL_QUEUE_TYPE_RX].start_queue_id = 0;
|
|
queue_chunk[VIRTCHNL_QUEUE_TYPE_RX].num_queues =
|
|
adapter->eth_dev->data->nb_rx_queues;
|
|
|
|
args.ops = VIRTCHNL_OP_ENABLE_QUEUES_V2;
|
|
args.in_args = (u8 *)queue_select;
|
|
args.in_args_size = len;
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
err = iavf_execute_vf_cmd(adapter, &args);
|
|
if (err)
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to execute command of OP_ENABLE_QUEUES_V2");
|
|
|
|
rte_free(queue_select);
|
|
return err;
|
|
}
|
|
|
|
int
|
|
iavf_disable_queues_lv(struct iavf_adapter *adapter)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_del_ena_dis_queues *queue_select;
|
|
struct virtchnl_queue_chunk *queue_chunk;
|
|
struct iavf_cmd_info args;
|
|
int err, len;
|
|
|
|
len = sizeof(struct virtchnl_del_ena_dis_queues) +
|
|
sizeof(struct virtchnl_queue_chunk) *
|
|
(IAVF_RXTX_QUEUE_CHUNKS_NUM - 1);
|
|
queue_select = rte_zmalloc("queue_select", len, 0);
|
|
if (!queue_select)
|
|
return -ENOMEM;
|
|
|
|
queue_chunk = queue_select->chunks.chunks;
|
|
queue_select->chunks.num_chunks = IAVF_RXTX_QUEUE_CHUNKS_NUM;
|
|
queue_select->vport_id = vf->vsi_res->vsi_id;
|
|
|
|
queue_chunk[VIRTCHNL_QUEUE_TYPE_TX].type = VIRTCHNL_QUEUE_TYPE_TX;
|
|
queue_chunk[VIRTCHNL_QUEUE_TYPE_TX].start_queue_id = 0;
|
|
queue_chunk[VIRTCHNL_QUEUE_TYPE_TX].num_queues =
|
|
adapter->eth_dev->data->nb_tx_queues;
|
|
|
|
queue_chunk[VIRTCHNL_QUEUE_TYPE_RX].type = VIRTCHNL_QUEUE_TYPE_RX;
|
|
queue_chunk[VIRTCHNL_QUEUE_TYPE_RX].start_queue_id = 0;
|
|
queue_chunk[VIRTCHNL_QUEUE_TYPE_RX].num_queues =
|
|
adapter->eth_dev->data->nb_rx_queues;
|
|
|
|
args.ops = VIRTCHNL_OP_DISABLE_QUEUES_V2;
|
|
args.in_args = (u8 *)queue_select;
|
|
args.in_args_size = len;
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
err = iavf_execute_vf_cmd(adapter, &args);
|
|
if (err)
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to execute command of OP_DISABLE_QUEUES_V2");
|
|
|
|
rte_free(queue_select);
|
|
return err;
|
|
}
|
|
|
|
int
|
|
iavf_switch_queue_lv(struct iavf_adapter *adapter, uint16_t qid,
|
|
bool rx, bool on)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_del_ena_dis_queues *queue_select;
|
|
struct virtchnl_queue_chunk *queue_chunk;
|
|
struct iavf_cmd_info args;
|
|
int err, len;
|
|
|
|
len = sizeof(struct virtchnl_del_ena_dis_queues);
|
|
queue_select = rte_zmalloc("queue_select", len, 0);
|
|
if (!queue_select)
|
|
return -ENOMEM;
|
|
|
|
queue_chunk = queue_select->chunks.chunks;
|
|
queue_select->chunks.num_chunks = 1;
|
|
queue_select->vport_id = vf->vsi_res->vsi_id;
|
|
|
|
if (rx) {
|
|
queue_chunk->type = VIRTCHNL_QUEUE_TYPE_RX;
|
|
queue_chunk->start_queue_id = qid;
|
|
queue_chunk->num_queues = 1;
|
|
} else {
|
|
queue_chunk->type = VIRTCHNL_QUEUE_TYPE_TX;
|
|
queue_chunk->start_queue_id = qid;
|
|
queue_chunk->num_queues = 1;
|
|
}
|
|
|
|
if (on)
|
|
args.ops = VIRTCHNL_OP_ENABLE_QUEUES_V2;
|
|
else
|
|
args.ops = VIRTCHNL_OP_DISABLE_QUEUES_V2;
|
|
args.in_args = (u8 *)queue_select;
|
|
args.in_args_size = len;
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
err = iavf_execute_vf_cmd(adapter, &args);
|
|
if (err)
|
|
PMD_DRV_LOG(ERR, "Failed to execute command of %s",
|
|
on ? "OP_ENABLE_QUEUES_V2" : "OP_DISABLE_QUEUES_V2");
|
|
|
|
rte_free(queue_select);
|
|
return err;
|
|
}
|
|
|
|
int
|
|
iavf_configure_rss_lut(struct iavf_adapter *adapter)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_rss_lut *rss_lut;
|
|
struct iavf_cmd_info args;
|
|
int len, err = 0;
|
|
|
|
len = sizeof(*rss_lut) + vf->vf_res->rss_lut_size - 1;
|
|
rss_lut = rte_zmalloc("rss_lut", len, 0);
|
|
if (!rss_lut)
|
|
return -ENOMEM;
|
|
|
|
rss_lut->vsi_id = vf->vsi_res->vsi_id;
|
|
rss_lut->lut_entries = vf->vf_res->rss_lut_size;
|
|
rte_memcpy(rss_lut->lut, vf->rss_lut, vf->vf_res->rss_lut_size);
|
|
|
|
args.ops = VIRTCHNL_OP_CONFIG_RSS_LUT;
|
|
args.in_args = (u8 *)rss_lut;
|
|
args.in_args_size = len;
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
|
|
err = iavf_execute_vf_cmd(adapter, &args);
|
|
if (err)
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to execute command of OP_CONFIG_RSS_LUT");
|
|
|
|
rte_free(rss_lut);
|
|
return err;
|
|
}
|
|
|
|
int
|
|
iavf_configure_rss_key(struct iavf_adapter *adapter)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_rss_key *rss_key;
|
|
struct iavf_cmd_info args;
|
|
int len, err = 0;
|
|
|
|
len = sizeof(*rss_key) + vf->vf_res->rss_key_size - 1;
|
|
rss_key = rte_zmalloc("rss_key", len, 0);
|
|
if (!rss_key)
|
|
return -ENOMEM;
|
|
|
|
rss_key->vsi_id = vf->vsi_res->vsi_id;
|
|
rss_key->key_len = vf->vf_res->rss_key_size;
|
|
rte_memcpy(rss_key->key, vf->rss_key, vf->vf_res->rss_key_size);
|
|
|
|
args.ops = VIRTCHNL_OP_CONFIG_RSS_KEY;
|
|
args.in_args = (u8 *)rss_key;
|
|
args.in_args_size = len;
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
|
|
err = iavf_execute_vf_cmd(adapter, &args);
|
|
if (err)
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to execute command of OP_CONFIG_RSS_KEY");
|
|
|
|
rte_free(rss_key);
|
|
return err;
|
|
}
|
|
|
|
int
|
|
iavf_configure_queues(struct iavf_adapter *adapter,
|
|
uint16_t num_queue_pairs, uint16_t index)
|
|
{
|
|
struct iavf_rx_queue **rxq =
|
|
(struct iavf_rx_queue **)adapter->eth_dev->data->rx_queues;
|
|
struct iavf_tx_queue **txq =
|
|
(struct iavf_tx_queue **)adapter->eth_dev->data->tx_queues;
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_vsi_queue_config_info *vc_config;
|
|
struct virtchnl_queue_pair_info *vc_qp;
|
|
struct iavf_cmd_info args;
|
|
uint16_t i, size;
|
|
int err;
|
|
|
|
size = sizeof(*vc_config) +
|
|
sizeof(vc_config->qpair[0]) * num_queue_pairs;
|
|
vc_config = rte_zmalloc("cfg_queue", size, 0);
|
|
if (!vc_config)
|
|
return -ENOMEM;
|
|
|
|
vc_config->vsi_id = vf->vsi_res->vsi_id;
|
|
vc_config->num_queue_pairs = num_queue_pairs;
|
|
|
|
for (i = index, vc_qp = vc_config->qpair;
|
|
i < index + num_queue_pairs;
|
|
i++, vc_qp++) {
|
|
vc_qp->txq.vsi_id = vf->vsi_res->vsi_id;
|
|
vc_qp->txq.queue_id = i;
|
|
|
|
/* Virtchnnl configure tx queues by pairs */
|
|
if (i < adapter->eth_dev->data->nb_tx_queues) {
|
|
vc_qp->txq.ring_len = txq[i]->nb_tx_desc;
|
|
vc_qp->txq.dma_ring_addr = txq[i]->tx_ring_phys_addr;
|
|
}
|
|
|
|
vc_qp->rxq.vsi_id = vf->vsi_res->vsi_id;
|
|
vc_qp->rxq.queue_id = i;
|
|
vc_qp->rxq.max_pkt_size = vf->max_pkt_len;
|
|
|
|
if (i >= adapter->eth_dev->data->nb_rx_queues)
|
|
continue;
|
|
|
|
/* Virtchnnl configure rx queues by pairs */
|
|
vc_qp->rxq.ring_len = rxq[i]->nb_rx_desc;
|
|
vc_qp->rxq.dma_ring_addr = rxq[i]->rx_ring_phys_addr;
|
|
vc_qp->rxq.databuffer_size = rxq[i]->rx_buf_len;
|
|
vc_qp->rxq.crc_disable = rxq[i]->crc_len != 0 ? 1 : 0;
|
|
#ifndef RTE_LIBRTE_IAVF_16BYTE_RX_DESC
|
|
if (vf->vf_res->vf_cap_flags &
|
|
VIRTCHNL_VF_OFFLOAD_RX_FLEX_DESC &&
|
|
vf->supported_rxdid & BIT(rxq[i]->rxdid)) {
|
|
vc_qp->rxq.rxdid = rxq[i]->rxdid;
|
|
PMD_DRV_LOG(NOTICE, "request RXDID[%d] in Queue[%d]",
|
|
vc_qp->rxq.rxdid, i);
|
|
} else {
|
|
PMD_DRV_LOG(NOTICE, "RXDID[%d] is not supported, "
|
|
"request default RXDID[%d] in Queue[%d]",
|
|
rxq[i]->rxdid, IAVF_RXDID_LEGACY_1, i);
|
|
vc_qp->rxq.rxdid = IAVF_RXDID_LEGACY_1;
|
|
}
|
|
#else
|
|
if (vf->vf_res->vf_cap_flags &
|
|
VIRTCHNL_VF_OFFLOAD_RX_FLEX_DESC &&
|
|
vf->supported_rxdid & BIT(IAVF_RXDID_LEGACY_0)) {
|
|
vc_qp->rxq.rxdid = IAVF_RXDID_LEGACY_0;
|
|
PMD_DRV_LOG(NOTICE, "request RXDID[%d] in Queue[%d]",
|
|
vc_qp->rxq.rxdid, i);
|
|
} else {
|
|
PMD_DRV_LOG(ERR, "RXDID[%d] is not supported",
|
|
IAVF_RXDID_LEGACY_0);
|
|
return -1;
|
|
}
|
|
#endif
|
|
}
|
|
|
|
memset(&args, 0, sizeof(args));
|
|
args.ops = VIRTCHNL_OP_CONFIG_VSI_QUEUES;
|
|
args.in_args = (uint8_t *)vc_config;
|
|
args.in_args_size = size;
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
|
|
err = iavf_execute_vf_cmd(adapter, &args);
|
|
if (err)
|
|
PMD_DRV_LOG(ERR, "Failed to execute command of"
|
|
" VIRTCHNL_OP_CONFIG_VSI_QUEUES");
|
|
|
|
rte_free(vc_config);
|
|
return err;
|
|
}
|
|
|
|
int
|
|
iavf_config_irq_map(struct iavf_adapter *adapter)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_irq_map_info *map_info;
|
|
struct virtchnl_vector_map *vecmap;
|
|
struct iavf_cmd_info args;
|
|
int len, i, err;
|
|
|
|
len = sizeof(struct virtchnl_irq_map_info) +
|
|
sizeof(struct virtchnl_vector_map) * vf->nb_msix;
|
|
|
|
map_info = rte_zmalloc("map_info", len, 0);
|
|
if (!map_info)
|
|
return -ENOMEM;
|
|
|
|
map_info->num_vectors = vf->nb_msix;
|
|
for (i = 0; i < adapter->eth_dev->data->nb_rx_queues; i++) {
|
|
vecmap =
|
|
&map_info->vecmap[vf->qv_map[i].vector_id - vf->msix_base];
|
|
vecmap->vsi_id = vf->vsi_res->vsi_id;
|
|
vecmap->rxitr_idx = IAVF_ITR_INDEX_DEFAULT;
|
|
vecmap->vector_id = vf->qv_map[i].vector_id;
|
|
vecmap->txq_map = 0;
|
|
vecmap->rxq_map |= 1 << vf->qv_map[i].queue_id;
|
|
}
|
|
|
|
args.ops = VIRTCHNL_OP_CONFIG_IRQ_MAP;
|
|
args.in_args = (u8 *)map_info;
|
|
args.in_args_size = len;
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
err = iavf_execute_vf_cmd(adapter, &args);
|
|
if (err)
|
|
PMD_DRV_LOG(ERR, "fail to execute command OP_CONFIG_IRQ_MAP");
|
|
|
|
rte_free(map_info);
|
|
return err;
|
|
}
|
|
|
|
int
|
|
iavf_config_irq_map_lv(struct iavf_adapter *adapter, uint16_t num,
|
|
uint16_t index)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_queue_vector_maps *map_info;
|
|
struct virtchnl_queue_vector *qv_maps;
|
|
struct iavf_cmd_info args;
|
|
int len, i, err;
|
|
int count = 0;
|
|
|
|
len = sizeof(struct virtchnl_queue_vector_maps) +
|
|
sizeof(struct virtchnl_queue_vector) * (num - 1);
|
|
|
|
map_info = rte_zmalloc("map_info", len, 0);
|
|
if (!map_info)
|
|
return -ENOMEM;
|
|
|
|
map_info->vport_id = vf->vsi_res->vsi_id;
|
|
map_info->num_qv_maps = num;
|
|
for (i = index; i < index + map_info->num_qv_maps; i++) {
|
|
qv_maps = &map_info->qv_maps[count++];
|
|
qv_maps->itr_idx = VIRTCHNL_ITR_IDX_0;
|
|
qv_maps->queue_type = VIRTCHNL_QUEUE_TYPE_RX;
|
|
qv_maps->queue_id = vf->qv_map[i].queue_id;
|
|
qv_maps->vector_id = vf->qv_map[i].vector_id;
|
|
}
|
|
|
|
args.ops = VIRTCHNL_OP_MAP_QUEUE_VECTOR;
|
|
args.in_args = (u8 *)map_info;
|
|
args.in_args_size = len;
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
err = iavf_execute_vf_cmd(adapter, &args);
|
|
if (err)
|
|
PMD_DRV_LOG(ERR, "fail to execute command OP_MAP_QUEUE_VECTOR");
|
|
|
|
rte_free(map_info);
|
|
return err;
|
|
}
|
|
|
|
void
|
|
iavf_add_del_all_mac_addr(struct iavf_adapter *adapter, bool add)
|
|
{
|
|
struct virtchnl_ether_addr_list *list;
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct rte_ether_addr *addr;
|
|
struct iavf_cmd_info args;
|
|
int len, err, i, j;
|
|
int next_begin = 0;
|
|
int begin = 0;
|
|
|
|
do {
|
|
j = 0;
|
|
len = sizeof(struct virtchnl_ether_addr_list);
|
|
for (i = begin; i < IAVF_NUM_MACADDR_MAX; i++, next_begin++) {
|
|
addr = &adapter->eth_dev->data->mac_addrs[i];
|
|
if (rte_is_zero_ether_addr(addr))
|
|
continue;
|
|
len += sizeof(struct virtchnl_ether_addr);
|
|
if (len >= IAVF_AQ_BUF_SZ) {
|
|
next_begin = i + 1;
|
|
break;
|
|
}
|
|
}
|
|
|
|
list = rte_zmalloc("iavf_del_mac_buffer", len, 0);
|
|
if (!list) {
|
|
PMD_DRV_LOG(ERR, "fail to allocate memory");
|
|
return;
|
|
}
|
|
|
|
for (i = begin; i < next_begin; i++) {
|
|
addr = &adapter->eth_dev->data->mac_addrs[i];
|
|
if (rte_is_zero_ether_addr(addr))
|
|
continue;
|
|
rte_memcpy(list->list[j].addr, addr->addr_bytes,
|
|
sizeof(addr->addr_bytes));
|
|
list->list[j].type = (j == 0 ?
|
|
VIRTCHNL_ETHER_ADDR_PRIMARY :
|
|
VIRTCHNL_ETHER_ADDR_EXTRA);
|
|
PMD_DRV_LOG(DEBUG, "add/rm mac:" RTE_ETHER_ADDR_PRT_FMT,
|
|
RTE_ETHER_ADDR_BYTES(addr));
|
|
j++;
|
|
}
|
|
list->vsi_id = vf->vsi_res->vsi_id;
|
|
list->num_elements = j;
|
|
args.ops = add ? VIRTCHNL_OP_ADD_ETH_ADDR :
|
|
VIRTCHNL_OP_DEL_ETH_ADDR;
|
|
args.in_args = (uint8_t *)list;
|
|
args.in_args_size = len;
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
err = iavf_execute_vf_cmd(adapter, &args);
|
|
if (err)
|
|
PMD_DRV_LOG(ERR, "fail to execute command %s",
|
|
add ? "OP_ADD_ETHER_ADDRESS" :
|
|
"OP_DEL_ETHER_ADDRESS");
|
|
rte_free(list);
|
|
begin = next_begin;
|
|
} while (begin < IAVF_NUM_MACADDR_MAX);
|
|
}
|
|
|
|
int
|
|
iavf_query_stats(struct iavf_adapter *adapter,
|
|
struct virtchnl_eth_stats **pstats)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_queue_select q_stats;
|
|
struct iavf_cmd_info args;
|
|
int err;
|
|
|
|
memset(&q_stats, 0, sizeof(q_stats));
|
|
q_stats.vsi_id = vf->vsi_res->vsi_id;
|
|
args.ops = VIRTCHNL_OP_GET_STATS;
|
|
args.in_args = (uint8_t *)&q_stats;
|
|
args.in_args_size = sizeof(q_stats);
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
|
|
err = iavf_execute_vf_cmd(adapter, &args);
|
|
if (err) {
|
|
PMD_DRV_LOG(ERR, "fail to execute command OP_GET_STATS");
|
|
*pstats = NULL;
|
|
return err;
|
|
}
|
|
*pstats = (struct virtchnl_eth_stats *)args.out_buffer;
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
iavf_config_promisc(struct iavf_adapter *adapter,
|
|
bool enable_unicast,
|
|
bool enable_multicast)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_promisc_info promisc;
|
|
struct iavf_cmd_info args;
|
|
int err;
|
|
|
|
promisc.flags = 0;
|
|
promisc.vsi_id = vf->vsi_res->vsi_id;
|
|
|
|
if (enable_unicast)
|
|
promisc.flags |= FLAG_VF_UNICAST_PROMISC;
|
|
|
|
if (enable_multicast)
|
|
promisc.flags |= FLAG_VF_MULTICAST_PROMISC;
|
|
|
|
args.ops = VIRTCHNL_OP_CONFIG_PROMISCUOUS_MODE;
|
|
args.in_args = (uint8_t *)&promisc;
|
|
args.in_args_size = sizeof(promisc);
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
|
|
err = iavf_execute_vf_cmd(adapter, &args);
|
|
|
|
if (err) {
|
|
PMD_DRV_LOG(ERR,
|
|
"fail to execute command CONFIG_PROMISCUOUS_MODE");
|
|
|
|
if (err == -ENOTSUP)
|
|
return err;
|
|
|
|
return -EAGAIN;
|
|
}
|
|
|
|
vf->promisc_unicast_enabled = enable_unicast;
|
|
vf->promisc_multicast_enabled = enable_multicast;
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
iavf_add_del_eth_addr(struct iavf_adapter *adapter, struct rte_ether_addr *addr,
|
|
bool add, uint8_t type)
|
|
{
|
|
struct virtchnl_ether_addr_list *list;
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
uint8_t cmd_buffer[sizeof(struct virtchnl_ether_addr_list) +
|
|
sizeof(struct virtchnl_ether_addr)];
|
|
struct iavf_cmd_info args;
|
|
int err;
|
|
|
|
list = (struct virtchnl_ether_addr_list *)cmd_buffer;
|
|
list->vsi_id = vf->vsi_res->vsi_id;
|
|
list->num_elements = 1;
|
|
list->list[0].type = type;
|
|
rte_memcpy(list->list[0].addr, addr->addr_bytes,
|
|
sizeof(addr->addr_bytes));
|
|
|
|
args.ops = add ? VIRTCHNL_OP_ADD_ETH_ADDR : VIRTCHNL_OP_DEL_ETH_ADDR;
|
|
args.in_args = cmd_buffer;
|
|
args.in_args_size = sizeof(cmd_buffer);
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
err = iavf_execute_vf_cmd(adapter, &args);
|
|
if (err)
|
|
PMD_DRV_LOG(ERR, "fail to execute command %s",
|
|
add ? "OP_ADD_ETH_ADDR" : "OP_DEL_ETH_ADDR");
|
|
return err;
|
|
}
|
|
|
|
int
|
|
iavf_add_del_vlan(struct iavf_adapter *adapter, uint16_t vlanid, bool add)
|
|
{
|
|
struct virtchnl_vlan_filter_list *vlan_list;
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
uint8_t cmd_buffer[sizeof(struct virtchnl_vlan_filter_list) +
|
|
sizeof(uint16_t)];
|
|
struct iavf_cmd_info args;
|
|
int err;
|
|
|
|
vlan_list = (struct virtchnl_vlan_filter_list *)cmd_buffer;
|
|
vlan_list->vsi_id = vf->vsi_res->vsi_id;
|
|
vlan_list->num_elements = 1;
|
|
vlan_list->vlan_id[0] = vlanid;
|
|
|
|
args.ops = add ? VIRTCHNL_OP_ADD_VLAN : VIRTCHNL_OP_DEL_VLAN;
|
|
args.in_args = cmd_buffer;
|
|
args.in_args_size = sizeof(cmd_buffer);
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
err = iavf_execute_vf_cmd(adapter, &args);
|
|
if (err)
|
|
PMD_DRV_LOG(ERR, "fail to execute command %s",
|
|
add ? "OP_ADD_VLAN" : "OP_DEL_VLAN");
|
|
|
|
return err;
|
|
}
|
|
|
|
int
|
|
iavf_fdir_add(struct iavf_adapter *adapter,
|
|
struct iavf_fdir_conf *filter)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_fdir_add *fdir_ret;
|
|
|
|
struct iavf_cmd_info args;
|
|
int err;
|
|
|
|
filter->add_fltr.vsi_id = vf->vsi_res->vsi_id;
|
|
filter->add_fltr.validate_only = 0;
|
|
|
|
args.ops = VIRTCHNL_OP_ADD_FDIR_FILTER;
|
|
args.in_args = (uint8_t *)(&filter->add_fltr);
|
|
args.in_args_size = sizeof(*(&filter->add_fltr));
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
|
|
err = iavf_execute_vf_cmd(adapter, &args);
|
|
if (err) {
|
|
PMD_DRV_LOG(ERR, "fail to execute command OP_ADD_FDIR_FILTER");
|
|
return err;
|
|
}
|
|
|
|
fdir_ret = (struct virtchnl_fdir_add *)args.out_buffer;
|
|
filter->flow_id = fdir_ret->flow_id;
|
|
|
|
if (fdir_ret->status == VIRTCHNL_FDIR_SUCCESS) {
|
|
PMD_DRV_LOG(INFO,
|
|
"Succeed in adding rule request by PF");
|
|
} else if (fdir_ret->status == VIRTCHNL_FDIR_FAILURE_RULE_NORESOURCE) {
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to add rule request due to no hw resource");
|
|
return -1;
|
|
} else if (fdir_ret->status == VIRTCHNL_FDIR_FAILURE_RULE_EXIST) {
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to add rule request due to the rule is already existed");
|
|
return -1;
|
|
} else if (fdir_ret->status == VIRTCHNL_FDIR_FAILURE_RULE_CONFLICT) {
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to add rule request due to the rule is conflict with existing rule");
|
|
return -1;
|
|
} else if (fdir_ret->status == VIRTCHNL_FDIR_FAILURE_RULE_INVALID) {
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to add rule request due to the hw doesn't support");
|
|
return -1;
|
|
} else if (fdir_ret->status == VIRTCHNL_FDIR_FAILURE_RULE_TIMEOUT) {
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to add rule request due to time out for programming");
|
|
return -1;
|
|
} else {
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to add rule request due to other reasons");
|
|
return -1;
|
|
}
|
|
|
|
return 0;
|
|
};
|
|
|
|
int
|
|
iavf_fdir_del(struct iavf_adapter *adapter,
|
|
struct iavf_fdir_conf *filter)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_fdir_del *fdir_ret;
|
|
|
|
struct iavf_cmd_info args;
|
|
int err;
|
|
|
|
filter->del_fltr.vsi_id = vf->vsi_res->vsi_id;
|
|
filter->del_fltr.flow_id = filter->flow_id;
|
|
|
|
args.ops = VIRTCHNL_OP_DEL_FDIR_FILTER;
|
|
args.in_args = (uint8_t *)(&filter->del_fltr);
|
|
args.in_args_size = sizeof(filter->del_fltr);
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
|
|
err = iavf_execute_vf_cmd(adapter, &args);
|
|
if (err) {
|
|
PMD_DRV_LOG(ERR, "fail to execute command OP_DEL_FDIR_FILTER");
|
|
return err;
|
|
}
|
|
|
|
fdir_ret = (struct virtchnl_fdir_del *)args.out_buffer;
|
|
|
|
if (fdir_ret->status == VIRTCHNL_FDIR_SUCCESS) {
|
|
PMD_DRV_LOG(INFO,
|
|
"Succeed in deleting rule request by PF");
|
|
} else if (fdir_ret->status == VIRTCHNL_FDIR_FAILURE_RULE_NONEXIST) {
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to delete rule request due to this rule doesn't exist");
|
|
return -1;
|
|
} else if (fdir_ret->status == VIRTCHNL_FDIR_FAILURE_RULE_TIMEOUT) {
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to delete rule request due to time out for programming");
|
|
return -1;
|
|
} else {
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to delete rule request due to other reasons");
|
|
return -1;
|
|
}
|
|
|
|
return 0;
|
|
};
|
|
|
|
int
|
|
iavf_fdir_check(struct iavf_adapter *adapter,
|
|
struct iavf_fdir_conf *filter)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_fdir_add *fdir_ret;
|
|
|
|
struct iavf_cmd_info args;
|
|
int err;
|
|
|
|
filter->add_fltr.vsi_id = vf->vsi_res->vsi_id;
|
|
filter->add_fltr.validate_only = 1;
|
|
|
|
args.ops = VIRTCHNL_OP_ADD_FDIR_FILTER;
|
|
args.in_args = (uint8_t *)(&filter->add_fltr);
|
|
args.in_args_size = sizeof(*(&filter->add_fltr));
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
|
|
err = iavf_execute_vf_cmd(adapter, &args);
|
|
if (err) {
|
|
PMD_DRV_LOG(ERR, "fail to check flow direcotor rule");
|
|
return err;
|
|
}
|
|
|
|
fdir_ret = (struct virtchnl_fdir_add *)args.out_buffer;
|
|
|
|
if (fdir_ret->status == VIRTCHNL_FDIR_SUCCESS) {
|
|
PMD_DRV_LOG(INFO,
|
|
"Succeed in checking rule request by PF");
|
|
} else if (fdir_ret->status == VIRTCHNL_FDIR_FAILURE_RULE_INVALID) {
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to check rule request due to parameters validation"
|
|
" or HW doesn't support");
|
|
return -1;
|
|
} else {
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to check rule request due to other reasons");
|
|
return -1;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
iavf_add_del_rss_cfg(struct iavf_adapter *adapter,
|
|
struct virtchnl_rss_cfg *rss_cfg, bool add)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct iavf_cmd_info args;
|
|
int err;
|
|
|
|
memset(&args, 0, sizeof(args));
|
|
args.ops = add ? VIRTCHNL_OP_ADD_RSS_CFG :
|
|
VIRTCHNL_OP_DEL_RSS_CFG;
|
|
args.in_args = (u8 *)rss_cfg;
|
|
args.in_args_size = sizeof(*rss_cfg);
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
|
|
err = iavf_execute_vf_cmd(adapter, &args);
|
|
if (err)
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to execute command of %s",
|
|
add ? "OP_ADD_RSS_CFG" :
|
|
"OP_DEL_RSS_INPUT_CFG");
|
|
|
|
return err;
|
|
}
|
|
|
|
int
|
|
iavf_get_hena_caps(struct iavf_adapter *adapter, uint64_t *caps)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct iavf_cmd_info args;
|
|
int err;
|
|
|
|
args.ops = VIRTCHNL_OP_GET_RSS_HENA_CAPS;
|
|
args.in_args = NULL;
|
|
args.in_args_size = 0;
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
|
|
err = iavf_execute_vf_cmd(adapter, &args);
|
|
if (err) {
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to execute command of OP_GET_RSS_HENA_CAPS");
|
|
return err;
|
|
}
|
|
|
|
*caps = ((struct virtchnl_rss_hena *)args.out_buffer)->hena;
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
iavf_set_hena(struct iavf_adapter *adapter, uint64_t hena)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_rss_hena vrh;
|
|
struct iavf_cmd_info args;
|
|
int err;
|
|
|
|
vrh.hena = hena;
|
|
args.ops = VIRTCHNL_OP_SET_RSS_HENA;
|
|
args.in_args = (u8 *)&vrh;
|
|
args.in_args_size = sizeof(vrh);
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
|
|
err = iavf_execute_vf_cmd(adapter, &args);
|
|
if (err)
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to execute command of OP_SET_RSS_HENA");
|
|
|
|
return err;
|
|
}
|
|
|
|
int
|
|
iavf_get_qos_cap(struct iavf_adapter *adapter)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct iavf_cmd_info args;
|
|
uint32_t len;
|
|
int err;
|
|
|
|
args.ops = VIRTCHNL_OP_GET_QOS_CAPS;
|
|
args.in_args = NULL;
|
|
args.in_args_size = 0;
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
err = iavf_execute_vf_cmd(adapter, &args);
|
|
|
|
if (err) {
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to execute command of OP_GET_VF_RESOURCE");
|
|
return -1;
|
|
}
|
|
|
|
len = sizeof(struct virtchnl_qos_cap_list) +
|
|
IAVF_MAX_TRAFFIC_CLASS * sizeof(struct virtchnl_qos_cap_elem);
|
|
|
|
rte_memcpy(vf->qos_cap, args.out_buffer,
|
|
RTE_MIN(args.out_size, len));
|
|
|
|
return 0;
|
|
}
|
|
|
|
int iavf_set_q_tc_map(struct rte_eth_dev *dev,
|
|
struct virtchnl_queue_tc_mapping *q_tc_mapping, uint16_t size)
|
|
{
|
|
struct iavf_adapter *adapter =
|
|
IAVF_DEV_PRIVATE_TO_ADAPTER(dev->data->dev_private);
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(dev->data->dev_private);
|
|
struct iavf_cmd_info args;
|
|
int err;
|
|
|
|
memset(&args, 0, sizeof(args));
|
|
args.ops = VIRTCHNL_OP_CONFIG_QUEUE_TC_MAP;
|
|
args.in_args = (uint8_t *)q_tc_mapping;
|
|
args.in_args_size = size;
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
|
|
err = iavf_execute_vf_cmd(adapter, &args);
|
|
if (err)
|
|
PMD_DRV_LOG(ERR, "Failed to execute command of"
|
|
" VIRTCHNL_OP_CONFIG_TC_MAP");
|
|
return err;
|
|
}
|
|
|
|
int
|
|
iavf_add_del_mc_addr_list(struct iavf_adapter *adapter,
|
|
struct rte_ether_addr *mc_addrs,
|
|
uint32_t mc_addrs_num, bool add)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
uint8_t cmd_buffer[sizeof(struct virtchnl_ether_addr_list) +
|
|
(IAVF_NUM_MACADDR_MAX * sizeof(struct virtchnl_ether_addr))];
|
|
struct virtchnl_ether_addr_list *list;
|
|
struct iavf_cmd_info args;
|
|
uint32_t i;
|
|
int err;
|
|
|
|
if (mc_addrs == NULL || mc_addrs_num == 0)
|
|
return 0;
|
|
|
|
list = (struct virtchnl_ether_addr_list *)cmd_buffer;
|
|
list->vsi_id = vf->vsi_res->vsi_id;
|
|
list->num_elements = mc_addrs_num;
|
|
|
|
for (i = 0; i < mc_addrs_num; i++) {
|
|
if (!IAVF_IS_MULTICAST(mc_addrs[i].addr_bytes)) {
|
|
PMD_DRV_LOG(ERR, "Invalid mac:" RTE_ETHER_ADDR_PRT_FMT,
|
|
RTE_ETHER_ADDR_BYTES(&mc_addrs[i]));
|
|
return -EINVAL;
|
|
}
|
|
|
|
memcpy(list->list[i].addr, mc_addrs[i].addr_bytes,
|
|
sizeof(list->list[i].addr));
|
|
list->list[i].type = VIRTCHNL_ETHER_ADDR_EXTRA;
|
|
}
|
|
|
|
args.ops = add ? VIRTCHNL_OP_ADD_ETH_ADDR : VIRTCHNL_OP_DEL_ETH_ADDR;
|
|
args.in_args = cmd_buffer;
|
|
args.in_args_size = sizeof(struct virtchnl_ether_addr_list) +
|
|
i * sizeof(struct virtchnl_ether_addr);
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
err = iavf_execute_vf_cmd(adapter, &args);
|
|
|
|
if (err) {
|
|
PMD_DRV_LOG(ERR, "fail to execute command %s",
|
|
add ? "OP_ADD_ETH_ADDR" : "OP_DEL_ETH_ADDR");
|
|
return err;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
iavf_request_queues(struct iavf_adapter *adapter, uint16_t num)
|
|
{
|
|
struct rte_eth_dev *dev = adapter->eth_dev;
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct rte_pci_device *pci_dev = RTE_ETH_DEV_TO_PCI(dev);
|
|
struct virtchnl_vf_res_request vfres;
|
|
struct iavf_cmd_info args;
|
|
uint16_t num_queue_pairs;
|
|
int err;
|
|
|
|
if (!(vf->vf_res->vf_cap_flags &
|
|
VIRTCHNL_VF_OFFLOAD_REQ_QUEUES)) {
|
|
PMD_DRV_LOG(ERR, "request queues not supported");
|
|
return -1;
|
|
}
|
|
|
|
if (num == 0) {
|
|
PMD_DRV_LOG(ERR, "queue number cannot be zero");
|
|
return -1;
|
|
}
|
|
vfres.num_queue_pairs = num;
|
|
|
|
args.ops = VIRTCHNL_OP_REQUEST_QUEUES;
|
|
args.in_args = (u8 *)&vfres;
|
|
args.in_args_size = sizeof(vfres);
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
|
|
/*
|
|
* disable interrupt to avoid the admin queue message to be read
|
|
* before iavf_read_msg_from_pf.
|
|
*/
|
|
rte_intr_disable(&pci_dev->intr_handle);
|
|
err = iavf_execute_vf_cmd(adapter, &args);
|
|
rte_intr_enable(&pci_dev->intr_handle);
|
|
if (err) {
|
|
PMD_DRV_LOG(ERR, "fail to execute command OP_REQUEST_QUEUES");
|
|
return err;
|
|
}
|
|
|
|
/* request queues succeeded, vf is resetting */
|
|
if (vf->vf_reset) {
|
|
PMD_DRV_LOG(INFO, "vf is resetting");
|
|
return 0;
|
|
}
|
|
|
|
/* request additional queues failed, return available number */
|
|
num_queue_pairs =
|
|
((struct virtchnl_vf_res_request *)args.out_buffer)->num_queue_pairs;
|
|
PMD_DRV_LOG(ERR, "request queues failed, only %u queues "
|
|
"available", num_queue_pairs);
|
|
|
|
return -1;
|
|
}
|
|
|
|
int
|
|
iavf_get_max_rss_queue_region(struct iavf_adapter *adapter)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct iavf_cmd_info args;
|
|
uint16_t qregion_width;
|
|
int err;
|
|
|
|
args.ops = VIRTCHNL_OP_GET_MAX_RSS_QREGION;
|
|
args.in_args = NULL;
|
|
args.in_args_size = 0;
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
|
|
err = iavf_execute_vf_cmd(adapter, &args);
|
|
if (err) {
|
|
PMD_DRV_LOG(ERR, "Failed to execute command of VIRTCHNL_OP_GET_MAX_RSS_QREGION");
|
|
return err;
|
|
}
|
|
|
|
qregion_width =
|
|
((struct virtchnl_max_rss_qregion *)args.out_buffer)->qregion_width;
|
|
|
|
vf->max_rss_qregion = (uint16_t)(1 << qregion_width);
|
|
|
|
return 0;
|
|
}
|