7b902af499
Support flow subscribption create/destroy/validation flow rule for AVF. For examples: testpmd> flow create 0 ingress pattern eth / ipv4 / udp src is 11 / end actions represented_port port_id 1 / end testpmd> flow validate 1 ingress pattern eth / ipv4 / tcp src is 22 / end actions represented_port port_id 1 / end testpmd> flow destroy 1 rule 0 VF subscribes to a rule, which means the packets will be sent to VF instead of PF, and only VF will receive the packets. It is allowed multiple VF subscribe to same rule, the packets will be replicated and received by each VF. PF will destroy all subscriptions during VF reset. Signed-off-by: Jie Wang <jie1x.wang@intel.com> Acked-by: Qi Zhang <qi.z.zhang@intel.com>
2076 lines
56 KiB
C
2076 lines
56 KiB
C
/* SPDX-License-Identifier: BSD-3-Clause
|
|
* Copyright(c) 2017 Intel Corporation
|
|
*/
|
|
|
|
#include <stdio.h>
|
|
#include <errno.h>
|
|
#include <stdint.h>
|
|
#include <string.h>
|
|
#include <unistd.h>
|
|
#include <stdarg.h>
|
|
#include <inttypes.h>
|
|
#include <rte_byteorder.h>
|
|
#include <rte_common.h>
|
|
|
|
#include <rte_debug.h>
|
|
#include <rte_alarm.h>
|
|
#include <rte_atomic.h>
|
|
#include <rte_eal.h>
|
|
#include <rte_ether.h>
|
|
#include <ethdev_driver.h>
|
|
#include <ethdev_pci.h>
|
|
#include <dev_driver.h>
|
|
|
|
#include "iavf.h"
|
|
#include "iavf_rxtx.h"
|
|
|
|
#define MAX_TRY_TIMES 2000
|
|
#define ASQ_DELAY_MS 1
|
|
|
|
static uint32_t
|
|
iavf_convert_link_speed(enum virtchnl_link_speed virt_link_speed)
|
|
{
|
|
uint32_t speed;
|
|
|
|
switch (virt_link_speed) {
|
|
case VIRTCHNL_LINK_SPEED_100MB:
|
|
speed = 100;
|
|
break;
|
|
case VIRTCHNL_LINK_SPEED_1GB:
|
|
speed = 1000;
|
|
break;
|
|
case VIRTCHNL_LINK_SPEED_10GB:
|
|
speed = 10000;
|
|
break;
|
|
case VIRTCHNL_LINK_SPEED_40GB:
|
|
speed = 40000;
|
|
break;
|
|
case VIRTCHNL_LINK_SPEED_20GB:
|
|
speed = 20000;
|
|
break;
|
|
case VIRTCHNL_LINK_SPEED_25GB:
|
|
speed = 25000;
|
|
break;
|
|
case VIRTCHNL_LINK_SPEED_2_5GB:
|
|
speed = 2500;
|
|
break;
|
|
case VIRTCHNL_LINK_SPEED_5GB:
|
|
speed = 5000;
|
|
break;
|
|
default:
|
|
speed = 0;
|
|
break;
|
|
}
|
|
|
|
return speed;
|
|
}
|
|
|
|
/* Read data in admin queue to get msg from pf driver */
|
|
static enum iavf_aq_result
|
|
iavf_read_msg_from_pf(struct iavf_adapter *adapter, uint16_t buf_len,
|
|
uint8_t *buf)
|
|
{
|
|
struct iavf_hw *hw = IAVF_DEV_PRIVATE_TO_HW(adapter);
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct iavf_arq_event_info event;
|
|
enum iavf_aq_result result = IAVF_MSG_NON;
|
|
enum virtchnl_ops opcode;
|
|
int ret;
|
|
|
|
event.buf_len = buf_len;
|
|
event.msg_buf = buf;
|
|
ret = iavf_clean_arq_element(hw, &event, NULL);
|
|
/* Can't read any msg from adminQ */
|
|
if (ret) {
|
|
PMD_DRV_LOG(DEBUG, "Can't read msg from AQ");
|
|
if (ret != IAVF_ERR_ADMIN_QUEUE_NO_WORK)
|
|
result = IAVF_MSG_ERR;
|
|
return result;
|
|
}
|
|
|
|
opcode = (enum virtchnl_ops)rte_le_to_cpu_32(event.desc.cookie_high);
|
|
vf->cmd_retval = (enum virtchnl_status_code)rte_le_to_cpu_32(
|
|
event.desc.cookie_low);
|
|
|
|
PMD_DRV_LOG(DEBUG, "AQ from pf carries opcode %u, retval %d",
|
|
opcode, vf->cmd_retval);
|
|
|
|
if (opcode == VIRTCHNL_OP_EVENT) {
|
|
struct virtchnl_pf_event *vpe =
|
|
(struct virtchnl_pf_event *)event.msg_buf;
|
|
|
|
result = IAVF_MSG_SYS;
|
|
switch (vpe->event) {
|
|
case VIRTCHNL_EVENT_LINK_CHANGE:
|
|
vf->link_up =
|
|
vpe->event_data.link_event.link_status;
|
|
if (vf->vf_res->vf_cap_flags &
|
|
VIRTCHNL_VF_CAP_ADV_LINK_SPEED) {
|
|
vf->link_speed =
|
|
vpe->event_data.link_event_adv.link_speed;
|
|
} else {
|
|
enum virtchnl_link_speed speed;
|
|
speed = vpe->event_data.link_event.link_speed;
|
|
vf->link_speed = iavf_convert_link_speed(speed);
|
|
}
|
|
iavf_dev_link_update(vf->eth_dev, 0);
|
|
PMD_DRV_LOG(INFO, "Link status update:%s",
|
|
vf->link_up ? "up" : "down");
|
|
break;
|
|
case VIRTCHNL_EVENT_RESET_IMPENDING:
|
|
vf->vf_reset = true;
|
|
PMD_DRV_LOG(INFO, "VF is resetting");
|
|
break;
|
|
case VIRTCHNL_EVENT_PF_DRIVER_CLOSE:
|
|
vf->dev_closed = true;
|
|
PMD_DRV_LOG(INFO, "PF driver closed");
|
|
break;
|
|
default:
|
|
PMD_DRV_LOG(ERR, "%s: Unknown event %d from pf",
|
|
__func__, vpe->event);
|
|
}
|
|
} else {
|
|
/* async reply msg on command issued by vf previously */
|
|
result = IAVF_MSG_CMD;
|
|
if (opcode != vf->pend_cmd) {
|
|
PMD_DRV_LOG(WARNING, "command mismatch, expect %u, get %u",
|
|
vf->pend_cmd, opcode);
|
|
result = IAVF_MSG_ERR;
|
|
}
|
|
}
|
|
|
|
return result;
|
|
}
|
|
|
|
static int
|
|
iavf_execute_vf_cmd(struct iavf_adapter *adapter, struct iavf_cmd_info *args,
|
|
int async)
|
|
{
|
|
struct iavf_hw *hw = IAVF_DEV_PRIVATE_TO_HW(adapter);
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
enum iavf_aq_result result;
|
|
enum iavf_status ret;
|
|
int err = 0;
|
|
int i = 0;
|
|
|
|
if (vf->vf_reset)
|
|
return -EIO;
|
|
|
|
|
|
if (async) {
|
|
if (_atomic_set_async_response_cmd(vf, args->ops))
|
|
return -1;
|
|
} else {
|
|
if (_atomic_set_cmd(vf, args->ops))
|
|
return -1;
|
|
}
|
|
|
|
ret = iavf_aq_send_msg_to_pf(hw, args->ops, IAVF_SUCCESS,
|
|
args->in_args, args->in_args_size, NULL);
|
|
if (ret) {
|
|
PMD_DRV_LOG(ERR, "fail to send cmd %d", args->ops);
|
|
_clear_cmd(vf);
|
|
return err;
|
|
}
|
|
|
|
switch (args->ops) {
|
|
case VIRTCHNL_OP_RESET_VF:
|
|
/*no need to wait for response */
|
|
_clear_cmd(vf);
|
|
break;
|
|
case VIRTCHNL_OP_VERSION:
|
|
case VIRTCHNL_OP_GET_VF_RESOURCES:
|
|
case VIRTCHNL_OP_GET_SUPPORTED_RXDIDS:
|
|
case VIRTCHNL_OP_GET_OFFLOAD_VLAN_V2_CAPS:
|
|
/* for init virtchnl ops, need to poll the response */
|
|
do {
|
|
result = iavf_read_msg_from_pf(adapter, args->out_size,
|
|
args->out_buffer);
|
|
if (result == IAVF_MSG_CMD)
|
|
break;
|
|
iavf_msec_delay(ASQ_DELAY_MS);
|
|
} while (i++ < MAX_TRY_TIMES);
|
|
if (i >= MAX_TRY_TIMES ||
|
|
vf->cmd_retval != VIRTCHNL_STATUS_SUCCESS) {
|
|
err = -1;
|
|
PMD_DRV_LOG(ERR, "No response or return failure (%d)"
|
|
" for cmd %d", vf->cmd_retval, args->ops);
|
|
}
|
|
_clear_cmd(vf);
|
|
break;
|
|
case VIRTCHNL_OP_REQUEST_QUEUES:
|
|
/*
|
|
* ignore async reply, only wait for system message,
|
|
* vf_reset = true if get VIRTCHNL_EVENT_RESET_IMPENDING,
|
|
* if not, means request queues failed.
|
|
*/
|
|
do {
|
|
result = iavf_read_msg_from_pf(adapter, args->out_size,
|
|
args->out_buffer);
|
|
if (result == IAVF_MSG_SYS && vf->vf_reset) {
|
|
break;
|
|
} else if (result == IAVF_MSG_CMD ||
|
|
result == IAVF_MSG_ERR) {
|
|
err = -1;
|
|
break;
|
|
}
|
|
iavf_msec_delay(ASQ_DELAY_MS);
|
|
/* If don't read msg or read sys event, continue */
|
|
} while (i++ < MAX_TRY_TIMES);
|
|
if (i >= MAX_TRY_TIMES ||
|
|
vf->cmd_retval != VIRTCHNL_STATUS_SUCCESS) {
|
|
err = -1;
|
|
PMD_DRV_LOG(ERR, "No response or return failure (%d)"
|
|
" for cmd %d", vf->cmd_retval, args->ops);
|
|
}
|
|
_clear_cmd(vf);
|
|
break;
|
|
default:
|
|
/* For other virtchnl ops in running time,
|
|
* wait for the cmd done flag.
|
|
*/
|
|
do {
|
|
if (vf->pend_cmd == VIRTCHNL_OP_UNKNOWN)
|
|
break;
|
|
iavf_msec_delay(ASQ_DELAY_MS);
|
|
/* If don't read msg or read sys event, continue */
|
|
} while (i++ < MAX_TRY_TIMES);
|
|
|
|
if (i >= MAX_TRY_TIMES) {
|
|
PMD_DRV_LOG(ERR, "No response for cmd %d", args->ops);
|
|
_clear_cmd(vf);
|
|
err = -EIO;
|
|
} else if (vf->cmd_retval ==
|
|
VIRTCHNL_STATUS_ERR_NOT_SUPPORTED) {
|
|
PMD_DRV_LOG(ERR, "Cmd %d not supported", args->ops);
|
|
err = -ENOTSUP;
|
|
} else if (vf->cmd_retval != VIRTCHNL_STATUS_SUCCESS) {
|
|
PMD_DRV_LOG(ERR, "Return failure %d for cmd %d",
|
|
vf->cmd_retval, args->ops);
|
|
err = -EINVAL;
|
|
}
|
|
break;
|
|
}
|
|
|
|
return err;
|
|
}
|
|
|
|
static void
|
|
iavf_handle_pf_event_msg(struct rte_eth_dev *dev, uint8_t *msg,
|
|
uint16_t msglen)
|
|
{
|
|
struct iavf_adapter *adapter =
|
|
IAVF_DEV_PRIVATE_TO_ADAPTER(dev->data->dev_private);
|
|
struct iavf_info *vf = &adapter->vf;
|
|
struct virtchnl_pf_event *pf_msg =
|
|
(struct virtchnl_pf_event *)msg;
|
|
|
|
if (adapter->closed) {
|
|
PMD_DRV_LOG(DEBUG, "Port closed");
|
|
return;
|
|
}
|
|
|
|
if (msglen < sizeof(struct virtchnl_pf_event)) {
|
|
PMD_DRV_LOG(DEBUG, "Error event");
|
|
return;
|
|
}
|
|
switch (pf_msg->event) {
|
|
case VIRTCHNL_EVENT_RESET_IMPENDING:
|
|
PMD_DRV_LOG(DEBUG, "VIRTCHNL_EVENT_RESET_IMPENDING event");
|
|
vf->vf_reset = true;
|
|
rte_eth_dev_callback_process(dev, RTE_ETH_EVENT_INTR_RESET,
|
|
NULL);
|
|
break;
|
|
case VIRTCHNL_EVENT_LINK_CHANGE:
|
|
PMD_DRV_LOG(DEBUG, "VIRTCHNL_EVENT_LINK_CHANGE event");
|
|
vf->link_up = pf_msg->event_data.link_event.link_status;
|
|
if (vf->vf_res->vf_cap_flags & VIRTCHNL_VF_CAP_ADV_LINK_SPEED) {
|
|
vf->link_speed =
|
|
pf_msg->event_data.link_event_adv.link_speed;
|
|
} else {
|
|
enum virtchnl_link_speed speed;
|
|
speed = pf_msg->event_data.link_event.link_speed;
|
|
vf->link_speed = iavf_convert_link_speed(speed);
|
|
}
|
|
iavf_dev_link_update(dev, 0);
|
|
rte_eth_dev_callback_process(dev, RTE_ETH_EVENT_INTR_LSC, NULL);
|
|
break;
|
|
case VIRTCHNL_EVENT_PF_DRIVER_CLOSE:
|
|
PMD_DRV_LOG(DEBUG, "VIRTCHNL_EVENT_PF_DRIVER_CLOSE event");
|
|
break;
|
|
default:
|
|
PMD_DRV_LOG(ERR, " unknown event received %u", pf_msg->event);
|
|
break;
|
|
}
|
|
}
|
|
|
|
void
|
|
iavf_handle_virtchnl_msg(struct rte_eth_dev *dev)
|
|
{
|
|
struct iavf_hw *hw = IAVF_DEV_PRIVATE_TO_HW(dev->data->dev_private);
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(dev->data->dev_private);
|
|
struct iavf_arq_event_info info;
|
|
uint16_t pending, aq_opc;
|
|
enum virtchnl_ops msg_opc;
|
|
enum iavf_status msg_ret;
|
|
int ret;
|
|
|
|
info.buf_len = IAVF_AQ_BUF_SZ;
|
|
if (!vf->aq_resp) {
|
|
PMD_DRV_LOG(ERR, "Buffer for adminq resp should not be NULL");
|
|
return;
|
|
}
|
|
info.msg_buf = vf->aq_resp;
|
|
|
|
pending = 1;
|
|
while (pending) {
|
|
ret = iavf_clean_arq_element(hw, &info, &pending);
|
|
|
|
if (ret != IAVF_SUCCESS) {
|
|
PMD_DRV_LOG(INFO, "Failed to read msg from AdminQ,"
|
|
"ret: %d", ret);
|
|
break;
|
|
}
|
|
aq_opc = rte_le_to_cpu_16(info.desc.opcode);
|
|
/* For the message sent from pf to vf, opcode is stored in
|
|
* cookie_high of struct iavf_aq_desc, while return error code
|
|
* are stored in cookie_low, Which is done by PF driver.
|
|
*/
|
|
msg_opc = (enum virtchnl_ops)rte_le_to_cpu_32(
|
|
info.desc.cookie_high);
|
|
msg_ret = (enum iavf_status)rte_le_to_cpu_32(
|
|
info.desc.cookie_low);
|
|
switch (aq_opc) {
|
|
case iavf_aqc_opc_send_msg_to_vf:
|
|
if (msg_opc == VIRTCHNL_OP_EVENT) {
|
|
iavf_handle_pf_event_msg(dev, info.msg_buf,
|
|
info.msg_len);
|
|
} else {
|
|
/* check for inline IPsec events */
|
|
struct inline_ipsec_msg *imsg =
|
|
(struct inline_ipsec_msg *)info.msg_buf;
|
|
struct rte_eth_event_ipsec_desc desc;
|
|
if (msg_opc ==
|
|
VIRTCHNL_OP_INLINE_IPSEC_CRYPTO &&
|
|
imsg->ipsec_opcode ==
|
|
INLINE_IPSEC_OP_EVENT) {
|
|
struct virtchnl_ipsec_event *ev =
|
|
imsg->ipsec_data.event;
|
|
desc.subtype =
|
|
RTE_ETH_EVENT_IPSEC_UNKNOWN;
|
|
desc.metadata = ev->ipsec_event_data;
|
|
rte_eth_dev_callback_process(dev,
|
|
RTE_ETH_EVENT_IPSEC,
|
|
&desc);
|
|
return;
|
|
}
|
|
|
|
/* read message and it's expected one */
|
|
if (msg_opc == vf->pend_cmd) {
|
|
uint32_t cmd_count =
|
|
__atomic_sub_fetch(&vf->pend_cmd_count,
|
|
1, __ATOMIC_RELAXED);
|
|
if (cmd_count == 0)
|
|
_notify_cmd(vf, msg_ret);
|
|
} else {
|
|
PMD_DRV_LOG(ERR,
|
|
"command mismatch, expect %u, get %u",
|
|
vf->pend_cmd, msg_opc);
|
|
}
|
|
PMD_DRV_LOG(DEBUG,
|
|
"adminq response is received, opcode = %d",
|
|
msg_opc);
|
|
}
|
|
break;
|
|
default:
|
|
PMD_DRV_LOG(DEBUG, "Request %u is not supported yet",
|
|
aq_opc);
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
|
|
int
|
|
iavf_enable_vlan_strip(struct iavf_adapter *adapter)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct iavf_cmd_info args;
|
|
int ret;
|
|
|
|
memset(&args, 0, sizeof(args));
|
|
args.ops = VIRTCHNL_OP_ENABLE_VLAN_STRIPPING;
|
|
args.in_args = NULL;
|
|
args.in_args_size = 0;
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
ret = iavf_execute_vf_cmd(adapter, &args, 0);
|
|
if (ret)
|
|
PMD_DRV_LOG(ERR, "Failed to execute command of"
|
|
" OP_ENABLE_VLAN_STRIPPING");
|
|
|
|
return ret;
|
|
}
|
|
|
|
int
|
|
iavf_disable_vlan_strip(struct iavf_adapter *adapter)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct iavf_cmd_info args;
|
|
int ret;
|
|
|
|
memset(&args, 0, sizeof(args));
|
|
args.ops = VIRTCHNL_OP_DISABLE_VLAN_STRIPPING;
|
|
args.in_args = NULL;
|
|
args.in_args_size = 0;
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
ret = iavf_execute_vf_cmd(adapter, &args, 0);
|
|
if (ret)
|
|
PMD_DRV_LOG(ERR, "Failed to execute command of"
|
|
" OP_DISABLE_VLAN_STRIPPING");
|
|
|
|
return ret;
|
|
}
|
|
|
|
#define VIRTCHNL_VERSION_MAJOR_START 1
|
|
#define VIRTCHNL_VERSION_MINOR_START 1
|
|
|
|
/* Check API version with sync wait until version read from admin queue */
|
|
int
|
|
iavf_check_api_version(struct iavf_adapter *adapter)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_version_info version, *pver;
|
|
struct iavf_cmd_info args;
|
|
int err;
|
|
|
|
version.major = VIRTCHNL_VERSION_MAJOR;
|
|
version.minor = VIRTCHNL_VERSION_MINOR;
|
|
|
|
args.ops = VIRTCHNL_OP_VERSION;
|
|
args.in_args = (uint8_t *)&version;
|
|
args.in_args_size = sizeof(version);
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
|
|
err = iavf_execute_vf_cmd(adapter, &args, 0);
|
|
if (err) {
|
|
PMD_INIT_LOG(ERR, "Fail to execute command of OP_VERSION");
|
|
return err;
|
|
}
|
|
|
|
pver = (struct virtchnl_version_info *)args.out_buffer;
|
|
vf->virtchnl_version = *pver;
|
|
|
|
if (vf->virtchnl_version.major < VIRTCHNL_VERSION_MAJOR_START ||
|
|
(vf->virtchnl_version.major == VIRTCHNL_VERSION_MAJOR_START &&
|
|
vf->virtchnl_version.minor < VIRTCHNL_VERSION_MINOR_START)) {
|
|
PMD_INIT_LOG(ERR, "VIRTCHNL API version should not be lower"
|
|
" than (%u.%u) to support Adaptive VF",
|
|
VIRTCHNL_VERSION_MAJOR_START,
|
|
VIRTCHNL_VERSION_MAJOR_START);
|
|
return -1;
|
|
} else if (vf->virtchnl_version.major > VIRTCHNL_VERSION_MAJOR ||
|
|
(vf->virtchnl_version.major == VIRTCHNL_VERSION_MAJOR &&
|
|
vf->virtchnl_version.minor > VIRTCHNL_VERSION_MINOR)) {
|
|
PMD_INIT_LOG(ERR, "PF/VF API version mismatch:(%u.%u)-(%u.%u)",
|
|
vf->virtchnl_version.major,
|
|
vf->virtchnl_version.minor,
|
|
VIRTCHNL_VERSION_MAJOR,
|
|
VIRTCHNL_VERSION_MINOR);
|
|
return -1;
|
|
}
|
|
|
|
PMD_DRV_LOG(DEBUG, "Peer is supported PF host");
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
iavf_get_vf_resource(struct iavf_adapter *adapter)
|
|
{
|
|
struct iavf_hw *hw = IAVF_DEV_PRIVATE_TO_HW(adapter);
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct iavf_cmd_info args;
|
|
uint32_t caps, len;
|
|
int err, i;
|
|
|
|
args.ops = VIRTCHNL_OP_GET_VF_RESOURCES;
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
|
|
caps = IAVF_BASIC_OFFLOAD_CAPS | VIRTCHNL_VF_CAP_ADV_LINK_SPEED |
|
|
VIRTCHNL_VF_OFFLOAD_RX_FLEX_DESC |
|
|
VIRTCHNL_VF_OFFLOAD_FDIR_PF |
|
|
VIRTCHNL_VF_OFFLOAD_ADV_RSS_PF |
|
|
VIRTCHNL_VF_OFFLOAD_FSUB_PF |
|
|
VIRTCHNL_VF_OFFLOAD_REQ_QUEUES |
|
|
VIRTCHNL_VF_OFFLOAD_CRC |
|
|
VIRTCHNL_VF_OFFLOAD_VLAN_V2 |
|
|
VIRTCHNL_VF_LARGE_NUM_QPAIRS |
|
|
VIRTCHNL_VF_OFFLOAD_QOS |
|
|
VIRTCHNL_VF_OFFLOAD_INLINE_IPSEC_CRYPTO |
|
|
VIRTCHNL_VF_CAP_PTP;
|
|
|
|
args.in_args = (uint8_t *)∩︀
|
|
args.in_args_size = sizeof(caps);
|
|
|
|
err = iavf_execute_vf_cmd(adapter, &args, 0);
|
|
|
|
if (err) {
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to execute command of OP_GET_VF_RESOURCE");
|
|
return -1;
|
|
}
|
|
|
|
len = sizeof(struct virtchnl_vf_resource) +
|
|
IAVF_MAX_VF_VSI * sizeof(struct virtchnl_vsi_resource);
|
|
|
|
rte_memcpy(vf->vf_res, args.out_buffer,
|
|
RTE_MIN(args.out_size, len));
|
|
/* parse VF config message back from PF*/
|
|
iavf_vf_parse_hw_config(hw, vf->vf_res);
|
|
for (i = 0; i < vf->vf_res->num_vsis; i++) {
|
|
if (vf->vf_res->vsi_res[i].vsi_type == VIRTCHNL_VSI_SRIOV)
|
|
vf->vsi_res = &vf->vf_res->vsi_res[i];
|
|
}
|
|
|
|
if (!vf->vsi_res) {
|
|
PMD_INIT_LOG(ERR, "no LAN VSI found");
|
|
return -1;
|
|
}
|
|
|
|
vf->vsi.vsi_id = vf->vsi_res->vsi_id;
|
|
vf->vsi.nb_qps = vf->vsi_res->num_queue_pairs;
|
|
vf->vsi.adapter = adapter;
|
|
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
iavf_get_supported_rxdid(struct iavf_adapter *adapter)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct iavf_cmd_info args;
|
|
int ret;
|
|
|
|
args.ops = VIRTCHNL_OP_GET_SUPPORTED_RXDIDS;
|
|
args.in_args = NULL;
|
|
args.in_args_size = 0;
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
|
|
ret = iavf_execute_vf_cmd(adapter, &args, 0);
|
|
if (ret) {
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to execute command of OP_GET_SUPPORTED_RXDIDS");
|
|
return ret;
|
|
}
|
|
|
|
vf->supported_rxdid =
|
|
((struct virtchnl_supported_rxdids *)args.out_buffer)->supported_rxdids;
|
|
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
iavf_config_vlan_strip_v2(struct iavf_adapter *adapter, bool enable)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_vlan_supported_caps *stripping_caps;
|
|
struct virtchnl_vlan_setting vlan_strip;
|
|
struct iavf_cmd_info args;
|
|
uint32_t *ethertype;
|
|
int ret;
|
|
|
|
stripping_caps = &vf->vlan_v2_caps.offloads.stripping_support;
|
|
|
|
if ((stripping_caps->outer & VIRTCHNL_VLAN_ETHERTYPE_8100) &&
|
|
(stripping_caps->outer & VIRTCHNL_VLAN_TOGGLE))
|
|
ethertype = &vlan_strip.outer_ethertype_setting;
|
|
else if ((stripping_caps->inner & VIRTCHNL_VLAN_ETHERTYPE_8100) &&
|
|
(stripping_caps->inner & VIRTCHNL_VLAN_TOGGLE))
|
|
ethertype = &vlan_strip.inner_ethertype_setting;
|
|
else
|
|
return -ENOTSUP;
|
|
|
|
memset(&vlan_strip, 0, sizeof(vlan_strip));
|
|
vlan_strip.vport_id = vf->vsi_res->vsi_id;
|
|
*ethertype = VIRTCHNL_VLAN_ETHERTYPE_8100;
|
|
|
|
args.ops = enable ? VIRTCHNL_OP_ENABLE_VLAN_STRIPPING_V2 :
|
|
VIRTCHNL_OP_DISABLE_VLAN_STRIPPING_V2;
|
|
args.in_args = (uint8_t *)&vlan_strip;
|
|
args.in_args_size = sizeof(vlan_strip);
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
ret = iavf_execute_vf_cmd(adapter, &args, 0);
|
|
if (ret)
|
|
PMD_DRV_LOG(ERR, "fail to execute command %s",
|
|
enable ? "VIRTCHNL_OP_ENABLE_VLAN_STRIPPING_V2" :
|
|
"VIRTCHNL_OP_DISABLE_VLAN_STRIPPING_V2");
|
|
|
|
return ret;
|
|
}
|
|
|
|
int
|
|
iavf_config_vlan_insert_v2(struct iavf_adapter *adapter, bool enable)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_vlan_supported_caps *insertion_caps;
|
|
struct virtchnl_vlan_setting vlan_insert;
|
|
struct iavf_cmd_info args;
|
|
uint32_t *ethertype;
|
|
int ret;
|
|
|
|
insertion_caps = &vf->vlan_v2_caps.offloads.insertion_support;
|
|
|
|
if ((insertion_caps->outer & VIRTCHNL_VLAN_ETHERTYPE_8100) &&
|
|
(insertion_caps->outer & VIRTCHNL_VLAN_TOGGLE))
|
|
ethertype = &vlan_insert.outer_ethertype_setting;
|
|
else if ((insertion_caps->inner & VIRTCHNL_VLAN_ETHERTYPE_8100) &&
|
|
(insertion_caps->inner & VIRTCHNL_VLAN_TOGGLE))
|
|
ethertype = &vlan_insert.inner_ethertype_setting;
|
|
else
|
|
return -ENOTSUP;
|
|
|
|
memset(&vlan_insert, 0, sizeof(vlan_insert));
|
|
vlan_insert.vport_id = vf->vsi_res->vsi_id;
|
|
*ethertype = VIRTCHNL_VLAN_ETHERTYPE_8100;
|
|
|
|
args.ops = enable ? VIRTCHNL_OP_ENABLE_VLAN_INSERTION_V2 :
|
|
VIRTCHNL_OP_DISABLE_VLAN_INSERTION_V2;
|
|
args.in_args = (uint8_t *)&vlan_insert;
|
|
args.in_args_size = sizeof(vlan_insert);
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
ret = iavf_execute_vf_cmd(adapter, &args, 0);
|
|
if (ret)
|
|
PMD_DRV_LOG(ERR, "fail to execute command %s",
|
|
enable ? "VIRTCHNL_OP_ENABLE_VLAN_INSERTION_V2" :
|
|
"VIRTCHNL_OP_DISABLE_VLAN_INSERTION_V2");
|
|
|
|
return ret;
|
|
}
|
|
|
|
int
|
|
iavf_add_del_vlan_v2(struct iavf_adapter *adapter, uint16_t vlanid, bool add)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_vlan_supported_caps *supported_caps;
|
|
struct virtchnl_vlan_filter_list_v2 vlan_filter;
|
|
struct virtchnl_vlan *vlan_setting;
|
|
struct iavf_cmd_info args;
|
|
uint32_t filtering_caps;
|
|
int err;
|
|
|
|
supported_caps = &vf->vlan_v2_caps.filtering.filtering_support;
|
|
if (supported_caps->outer) {
|
|
filtering_caps = supported_caps->outer;
|
|
vlan_setting = &vlan_filter.filters[0].outer;
|
|
} else {
|
|
filtering_caps = supported_caps->inner;
|
|
vlan_setting = &vlan_filter.filters[0].inner;
|
|
}
|
|
|
|
if (!(filtering_caps & VIRTCHNL_VLAN_ETHERTYPE_8100))
|
|
return -ENOTSUP;
|
|
|
|
memset(&vlan_filter, 0, sizeof(vlan_filter));
|
|
vlan_filter.vport_id = vf->vsi_res->vsi_id;
|
|
vlan_filter.num_elements = 1;
|
|
vlan_setting->tpid = RTE_ETHER_TYPE_VLAN;
|
|
vlan_setting->tci = vlanid;
|
|
|
|
args.ops = add ? VIRTCHNL_OP_ADD_VLAN_V2 : VIRTCHNL_OP_DEL_VLAN_V2;
|
|
args.in_args = (uint8_t *)&vlan_filter;
|
|
args.in_args_size = sizeof(vlan_filter);
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
err = iavf_execute_vf_cmd(adapter, &args, 0);
|
|
if (err)
|
|
PMD_DRV_LOG(ERR, "fail to execute command %s",
|
|
add ? "OP_ADD_VLAN_V2" : "OP_DEL_VLAN_V2");
|
|
|
|
return err;
|
|
}
|
|
|
|
int
|
|
iavf_get_vlan_offload_caps_v2(struct iavf_adapter *adapter)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct iavf_cmd_info args;
|
|
int ret;
|
|
|
|
args.ops = VIRTCHNL_OP_GET_OFFLOAD_VLAN_V2_CAPS;
|
|
args.in_args = NULL;
|
|
args.in_args_size = 0;
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
|
|
ret = iavf_execute_vf_cmd(adapter, &args, 0);
|
|
if (ret) {
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to execute command of VIRTCHNL_OP_GET_OFFLOAD_VLAN_V2_CAPS");
|
|
return ret;
|
|
}
|
|
|
|
rte_memcpy(&vf->vlan_v2_caps, vf->aq_resp, sizeof(vf->vlan_v2_caps));
|
|
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
iavf_enable_queues(struct iavf_adapter *adapter)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_queue_select queue_select;
|
|
struct iavf_cmd_info args;
|
|
int err;
|
|
|
|
memset(&queue_select, 0, sizeof(queue_select));
|
|
queue_select.vsi_id = vf->vsi_res->vsi_id;
|
|
|
|
queue_select.rx_queues = BIT(adapter->dev_data->nb_rx_queues) - 1;
|
|
queue_select.tx_queues = BIT(adapter->dev_data->nb_tx_queues) - 1;
|
|
|
|
args.ops = VIRTCHNL_OP_ENABLE_QUEUES;
|
|
args.in_args = (u8 *)&queue_select;
|
|
args.in_args_size = sizeof(queue_select);
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
err = iavf_execute_vf_cmd(adapter, &args, 0);
|
|
if (err) {
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to execute command of OP_ENABLE_QUEUES");
|
|
return err;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
iavf_disable_queues(struct iavf_adapter *adapter)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_queue_select queue_select;
|
|
struct iavf_cmd_info args;
|
|
int err;
|
|
|
|
memset(&queue_select, 0, sizeof(queue_select));
|
|
queue_select.vsi_id = vf->vsi_res->vsi_id;
|
|
|
|
queue_select.rx_queues = BIT(adapter->dev_data->nb_rx_queues) - 1;
|
|
queue_select.tx_queues = BIT(adapter->dev_data->nb_tx_queues) - 1;
|
|
|
|
args.ops = VIRTCHNL_OP_DISABLE_QUEUES;
|
|
args.in_args = (u8 *)&queue_select;
|
|
args.in_args_size = sizeof(queue_select);
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
err = iavf_execute_vf_cmd(adapter, &args, 0);
|
|
if (err) {
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to execute command of OP_DISABLE_QUEUES");
|
|
return err;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
iavf_switch_queue(struct iavf_adapter *adapter, uint16_t qid,
|
|
bool rx, bool on)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_queue_select queue_select;
|
|
struct iavf_cmd_info args;
|
|
int err;
|
|
|
|
if (adapter->closed)
|
|
return -EIO;
|
|
|
|
memset(&queue_select, 0, sizeof(queue_select));
|
|
queue_select.vsi_id = vf->vsi_res->vsi_id;
|
|
if (rx)
|
|
queue_select.rx_queues |= 1 << qid;
|
|
else
|
|
queue_select.tx_queues |= 1 << qid;
|
|
|
|
if (on)
|
|
args.ops = VIRTCHNL_OP_ENABLE_QUEUES;
|
|
else
|
|
args.ops = VIRTCHNL_OP_DISABLE_QUEUES;
|
|
args.in_args = (u8 *)&queue_select;
|
|
args.in_args_size = sizeof(queue_select);
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
err = iavf_execute_vf_cmd(adapter, &args, 0);
|
|
if (err)
|
|
PMD_DRV_LOG(ERR, "Failed to execute command of %s",
|
|
on ? "OP_ENABLE_QUEUES" : "OP_DISABLE_QUEUES");
|
|
return err;
|
|
}
|
|
|
|
int
|
|
iavf_enable_queues_lv(struct iavf_adapter *adapter)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_del_ena_dis_queues *queue_select;
|
|
struct virtchnl_queue_chunk *queue_chunk;
|
|
struct iavf_cmd_info args;
|
|
int err, len;
|
|
|
|
len = sizeof(struct virtchnl_del_ena_dis_queues) +
|
|
sizeof(struct virtchnl_queue_chunk) *
|
|
(IAVF_RXTX_QUEUE_CHUNKS_NUM - 1);
|
|
queue_select = rte_zmalloc("queue_select", len, 0);
|
|
if (!queue_select)
|
|
return -ENOMEM;
|
|
|
|
queue_chunk = queue_select->chunks.chunks;
|
|
queue_select->chunks.num_chunks = IAVF_RXTX_QUEUE_CHUNKS_NUM;
|
|
queue_select->vport_id = vf->vsi_res->vsi_id;
|
|
|
|
queue_chunk[VIRTCHNL_QUEUE_TYPE_TX].type = VIRTCHNL_QUEUE_TYPE_TX;
|
|
queue_chunk[VIRTCHNL_QUEUE_TYPE_TX].start_queue_id = 0;
|
|
queue_chunk[VIRTCHNL_QUEUE_TYPE_TX].num_queues =
|
|
adapter->dev_data->nb_tx_queues;
|
|
|
|
queue_chunk[VIRTCHNL_QUEUE_TYPE_RX].type = VIRTCHNL_QUEUE_TYPE_RX;
|
|
queue_chunk[VIRTCHNL_QUEUE_TYPE_RX].start_queue_id = 0;
|
|
queue_chunk[VIRTCHNL_QUEUE_TYPE_RX].num_queues =
|
|
adapter->dev_data->nb_rx_queues;
|
|
|
|
args.ops = VIRTCHNL_OP_ENABLE_QUEUES_V2;
|
|
args.in_args = (u8 *)queue_select;
|
|
args.in_args_size = len;
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
err = iavf_execute_vf_cmd(adapter, &args, 0);
|
|
if (err)
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to execute command of OP_ENABLE_QUEUES_V2");
|
|
|
|
rte_free(queue_select);
|
|
return err;
|
|
}
|
|
|
|
int
|
|
iavf_disable_queues_lv(struct iavf_adapter *adapter)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_del_ena_dis_queues *queue_select;
|
|
struct virtchnl_queue_chunk *queue_chunk;
|
|
struct iavf_cmd_info args;
|
|
int err, len;
|
|
|
|
len = sizeof(struct virtchnl_del_ena_dis_queues) +
|
|
sizeof(struct virtchnl_queue_chunk) *
|
|
(IAVF_RXTX_QUEUE_CHUNKS_NUM - 1);
|
|
queue_select = rte_zmalloc("queue_select", len, 0);
|
|
if (!queue_select)
|
|
return -ENOMEM;
|
|
|
|
queue_chunk = queue_select->chunks.chunks;
|
|
queue_select->chunks.num_chunks = IAVF_RXTX_QUEUE_CHUNKS_NUM;
|
|
queue_select->vport_id = vf->vsi_res->vsi_id;
|
|
|
|
queue_chunk[VIRTCHNL_QUEUE_TYPE_TX].type = VIRTCHNL_QUEUE_TYPE_TX;
|
|
queue_chunk[VIRTCHNL_QUEUE_TYPE_TX].start_queue_id = 0;
|
|
queue_chunk[VIRTCHNL_QUEUE_TYPE_TX].num_queues =
|
|
adapter->dev_data->nb_tx_queues;
|
|
|
|
queue_chunk[VIRTCHNL_QUEUE_TYPE_RX].type = VIRTCHNL_QUEUE_TYPE_RX;
|
|
queue_chunk[VIRTCHNL_QUEUE_TYPE_RX].start_queue_id = 0;
|
|
queue_chunk[VIRTCHNL_QUEUE_TYPE_RX].num_queues =
|
|
adapter->dev_data->nb_rx_queues;
|
|
|
|
args.ops = VIRTCHNL_OP_DISABLE_QUEUES_V2;
|
|
args.in_args = (u8 *)queue_select;
|
|
args.in_args_size = len;
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
err = iavf_execute_vf_cmd(adapter, &args, 0);
|
|
if (err)
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to execute command of OP_DISABLE_QUEUES_V2");
|
|
|
|
rte_free(queue_select);
|
|
return err;
|
|
}
|
|
|
|
int
|
|
iavf_switch_queue_lv(struct iavf_adapter *adapter, uint16_t qid,
|
|
bool rx, bool on)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_del_ena_dis_queues *queue_select;
|
|
struct virtchnl_queue_chunk *queue_chunk;
|
|
struct iavf_cmd_info args;
|
|
int err, len;
|
|
|
|
len = sizeof(struct virtchnl_del_ena_dis_queues);
|
|
queue_select = rte_zmalloc("queue_select", len, 0);
|
|
if (!queue_select)
|
|
return -ENOMEM;
|
|
|
|
queue_chunk = queue_select->chunks.chunks;
|
|
queue_select->chunks.num_chunks = 1;
|
|
queue_select->vport_id = vf->vsi_res->vsi_id;
|
|
|
|
if (rx) {
|
|
queue_chunk->type = VIRTCHNL_QUEUE_TYPE_RX;
|
|
queue_chunk->start_queue_id = qid;
|
|
queue_chunk->num_queues = 1;
|
|
} else {
|
|
queue_chunk->type = VIRTCHNL_QUEUE_TYPE_TX;
|
|
queue_chunk->start_queue_id = qid;
|
|
queue_chunk->num_queues = 1;
|
|
}
|
|
|
|
if (on)
|
|
args.ops = VIRTCHNL_OP_ENABLE_QUEUES_V2;
|
|
else
|
|
args.ops = VIRTCHNL_OP_DISABLE_QUEUES_V2;
|
|
args.in_args = (u8 *)queue_select;
|
|
args.in_args_size = len;
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
err = iavf_execute_vf_cmd(adapter, &args, 0);
|
|
if (err)
|
|
PMD_DRV_LOG(ERR, "Failed to execute command of %s",
|
|
on ? "OP_ENABLE_QUEUES_V2" : "OP_DISABLE_QUEUES_V2");
|
|
|
|
rte_free(queue_select);
|
|
return err;
|
|
}
|
|
|
|
int
|
|
iavf_configure_rss_lut(struct iavf_adapter *adapter)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_rss_lut *rss_lut;
|
|
struct iavf_cmd_info args;
|
|
int len, err = 0;
|
|
|
|
len = sizeof(*rss_lut) + vf->vf_res->rss_lut_size - 1;
|
|
rss_lut = rte_zmalloc("rss_lut", len, 0);
|
|
if (!rss_lut)
|
|
return -ENOMEM;
|
|
|
|
rss_lut->vsi_id = vf->vsi_res->vsi_id;
|
|
rss_lut->lut_entries = vf->vf_res->rss_lut_size;
|
|
rte_memcpy(rss_lut->lut, vf->rss_lut, vf->vf_res->rss_lut_size);
|
|
|
|
args.ops = VIRTCHNL_OP_CONFIG_RSS_LUT;
|
|
args.in_args = (u8 *)rss_lut;
|
|
args.in_args_size = len;
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
|
|
err = iavf_execute_vf_cmd(adapter, &args, 0);
|
|
if (err)
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to execute command of OP_CONFIG_RSS_LUT");
|
|
|
|
rte_free(rss_lut);
|
|
return err;
|
|
}
|
|
|
|
int
|
|
iavf_configure_rss_key(struct iavf_adapter *adapter)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_rss_key *rss_key;
|
|
struct iavf_cmd_info args;
|
|
int len, err = 0;
|
|
|
|
len = sizeof(*rss_key) + vf->vf_res->rss_key_size - 1;
|
|
rss_key = rte_zmalloc("rss_key", len, 0);
|
|
if (!rss_key)
|
|
return -ENOMEM;
|
|
|
|
rss_key->vsi_id = vf->vsi_res->vsi_id;
|
|
rss_key->key_len = vf->vf_res->rss_key_size;
|
|
rte_memcpy(rss_key->key, vf->rss_key, vf->vf_res->rss_key_size);
|
|
|
|
args.ops = VIRTCHNL_OP_CONFIG_RSS_KEY;
|
|
args.in_args = (u8 *)rss_key;
|
|
args.in_args_size = len;
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
|
|
err = iavf_execute_vf_cmd(adapter, &args, 0);
|
|
if (err)
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to execute command of OP_CONFIG_RSS_KEY");
|
|
|
|
rte_free(rss_key);
|
|
return err;
|
|
}
|
|
|
|
int
|
|
iavf_configure_queues(struct iavf_adapter *adapter,
|
|
uint16_t num_queue_pairs, uint16_t index)
|
|
{
|
|
struct iavf_rx_queue **rxq =
|
|
(struct iavf_rx_queue **)adapter->dev_data->rx_queues;
|
|
struct iavf_tx_queue **txq =
|
|
(struct iavf_tx_queue **)adapter->dev_data->tx_queues;
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_vsi_queue_config_info *vc_config;
|
|
struct virtchnl_queue_pair_info *vc_qp;
|
|
struct iavf_cmd_info args;
|
|
uint16_t i, size;
|
|
int err;
|
|
|
|
size = sizeof(*vc_config) +
|
|
sizeof(vc_config->qpair[0]) * num_queue_pairs;
|
|
vc_config = rte_zmalloc("cfg_queue", size, 0);
|
|
if (!vc_config)
|
|
return -ENOMEM;
|
|
|
|
vc_config->vsi_id = vf->vsi_res->vsi_id;
|
|
vc_config->num_queue_pairs = num_queue_pairs;
|
|
|
|
for (i = index, vc_qp = vc_config->qpair;
|
|
i < index + num_queue_pairs;
|
|
i++, vc_qp++) {
|
|
vc_qp->txq.vsi_id = vf->vsi_res->vsi_id;
|
|
vc_qp->txq.queue_id = i;
|
|
|
|
/* Virtchnnl configure tx queues by pairs */
|
|
if (i < adapter->dev_data->nb_tx_queues) {
|
|
vc_qp->txq.ring_len = txq[i]->nb_tx_desc;
|
|
vc_qp->txq.dma_ring_addr = txq[i]->tx_ring_phys_addr;
|
|
}
|
|
|
|
vc_qp->rxq.vsi_id = vf->vsi_res->vsi_id;
|
|
vc_qp->rxq.queue_id = i;
|
|
vc_qp->rxq.max_pkt_size = vf->max_pkt_len;
|
|
|
|
if (i >= adapter->dev_data->nb_rx_queues)
|
|
continue;
|
|
|
|
/* Virtchnnl configure rx queues by pairs */
|
|
vc_qp->rxq.ring_len = rxq[i]->nb_rx_desc;
|
|
vc_qp->rxq.dma_ring_addr = rxq[i]->rx_ring_phys_addr;
|
|
vc_qp->rxq.databuffer_size = rxq[i]->rx_buf_len;
|
|
vc_qp->rxq.crc_disable = rxq[i]->crc_len != 0 ? 1 : 0;
|
|
#ifndef RTE_LIBRTE_IAVF_16BYTE_RX_DESC
|
|
if (vf->vf_res->vf_cap_flags &
|
|
VIRTCHNL_VF_OFFLOAD_RX_FLEX_DESC) {
|
|
if (vf->supported_rxdid & BIT(rxq[i]->rxdid)) {
|
|
vc_qp->rxq.rxdid = rxq[i]->rxdid;
|
|
PMD_DRV_LOG(NOTICE, "request RXDID[%d] in Queue[%d]",
|
|
vc_qp->rxq.rxdid, i);
|
|
} else {
|
|
PMD_DRV_LOG(NOTICE, "RXDID[%d] is not supported, "
|
|
"request default RXDID[%d] in Queue[%d]",
|
|
rxq[i]->rxdid, IAVF_RXDID_LEGACY_1, i);
|
|
vc_qp->rxq.rxdid = IAVF_RXDID_LEGACY_1;
|
|
}
|
|
|
|
if (vf->vf_res->vf_cap_flags & VIRTCHNL_VF_CAP_PTP &&
|
|
vf->ptp_caps & VIRTCHNL_1588_PTP_CAP_RX_TSTAMP)
|
|
vc_qp->rxq.flags |= VIRTCHNL_PTP_RX_TSTAMP;
|
|
}
|
|
#else
|
|
if (vf->vf_res->vf_cap_flags &
|
|
VIRTCHNL_VF_OFFLOAD_RX_FLEX_DESC &&
|
|
vf->supported_rxdid & BIT(IAVF_RXDID_LEGACY_0)) {
|
|
vc_qp->rxq.rxdid = IAVF_RXDID_LEGACY_0;
|
|
PMD_DRV_LOG(NOTICE, "request RXDID[%d] in Queue[%d]",
|
|
vc_qp->rxq.rxdid, i);
|
|
} else {
|
|
PMD_DRV_LOG(ERR, "RXDID[%d] is not supported",
|
|
IAVF_RXDID_LEGACY_0);
|
|
return -1;
|
|
}
|
|
#endif
|
|
}
|
|
|
|
memset(&args, 0, sizeof(args));
|
|
args.ops = VIRTCHNL_OP_CONFIG_VSI_QUEUES;
|
|
args.in_args = (uint8_t *)vc_config;
|
|
args.in_args_size = size;
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
|
|
err = iavf_execute_vf_cmd(adapter, &args, 0);
|
|
if (err)
|
|
PMD_DRV_LOG(ERR, "Failed to execute command of"
|
|
" VIRTCHNL_OP_CONFIG_VSI_QUEUES");
|
|
|
|
rte_free(vc_config);
|
|
return err;
|
|
}
|
|
|
|
int
|
|
iavf_config_irq_map(struct iavf_adapter *adapter)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_irq_map_info *map_info;
|
|
struct virtchnl_vector_map *vecmap;
|
|
struct iavf_cmd_info args;
|
|
int len, i, err;
|
|
|
|
len = sizeof(struct virtchnl_irq_map_info) +
|
|
sizeof(struct virtchnl_vector_map) * vf->nb_msix;
|
|
|
|
map_info = rte_zmalloc("map_info", len, 0);
|
|
if (!map_info)
|
|
return -ENOMEM;
|
|
|
|
map_info->num_vectors = vf->nb_msix;
|
|
for (i = 0; i < adapter->dev_data->nb_rx_queues; i++) {
|
|
vecmap =
|
|
&map_info->vecmap[vf->qv_map[i].vector_id - vf->msix_base];
|
|
vecmap->vsi_id = vf->vsi_res->vsi_id;
|
|
vecmap->rxitr_idx = IAVF_ITR_INDEX_DEFAULT;
|
|
vecmap->vector_id = vf->qv_map[i].vector_id;
|
|
vecmap->txq_map = 0;
|
|
vecmap->rxq_map |= 1 << vf->qv_map[i].queue_id;
|
|
}
|
|
|
|
args.ops = VIRTCHNL_OP_CONFIG_IRQ_MAP;
|
|
args.in_args = (u8 *)map_info;
|
|
args.in_args_size = len;
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
err = iavf_execute_vf_cmd(adapter, &args, 0);
|
|
if (err)
|
|
PMD_DRV_LOG(ERR, "fail to execute command OP_CONFIG_IRQ_MAP");
|
|
|
|
rte_free(map_info);
|
|
return err;
|
|
}
|
|
|
|
int
|
|
iavf_config_irq_map_lv(struct iavf_adapter *adapter, uint16_t num,
|
|
uint16_t index)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_queue_vector_maps *map_info;
|
|
struct virtchnl_queue_vector *qv_maps;
|
|
struct iavf_cmd_info args;
|
|
int len, i, err;
|
|
int count = 0;
|
|
|
|
len = sizeof(struct virtchnl_queue_vector_maps) +
|
|
sizeof(struct virtchnl_queue_vector) * (num - 1);
|
|
|
|
map_info = rte_zmalloc("map_info", len, 0);
|
|
if (!map_info)
|
|
return -ENOMEM;
|
|
|
|
map_info->vport_id = vf->vsi_res->vsi_id;
|
|
map_info->num_qv_maps = num;
|
|
for (i = index; i < index + map_info->num_qv_maps; i++) {
|
|
qv_maps = &map_info->qv_maps[count++];
|
|
qv_maps->itr_idx = VIRTCHNL_ITR_IDX_0;
|
|
qv_maps->queue_type = VIRTCHNL_QUEUE_TYPE_RX;
|
|
qv_maps->queue_id = vf->qv_map[i].queue_id;
|
|
qv_maps->vector_id = vf->qv_map[i].vector_id;
|
|
}
|
|
|
|
args.ops = VIRTCHNL_OP_MAP_QUEUE_VECTOR;
|
|
args.in_args = (u8 *)map_info;
|
|
args.in_args_size = len;
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
err = iavf_execute_vf_cmd(adapter, &args, 0);
|
|
if (err)
|
|
PMD_DRV_LOG(ERR, "fail to execute command OP_MAP_QUEUE_VECTOR");
|
|
|
|
rte_free(map_info);
|
|
return err;
|
|
}
|
|
|
|
void
|
|
iavf_add_del_all_mac_addr(struct iavf_adapter *adapter, bool add)
|
|
{
|
|
struct virtchnl_ether_addr_list *list;
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct rte_ether_addr *addr;
|
|
struct iavf_cmd_info args;
|
|
int len, err, i, j;
|
|
int next_begin = 0;
|
|
int begin = 0;
|
|
|
|
do {
|
|
j = 0;
|
|
len = sizeof(struct virtchnl_ether_addr_list);
|
|
for (i = begin; i < IAVF_NUM_MACADDR_MAX; i++, next_begin++) {
|
|
addr = &adapter->dev_data->mac_addrs[i];
|
|
if (rte_is_zero_ether_addr(addr))
|
|
continue;
|
|
len += sizeof(struct virtchnl_ether_addr);
|
|
if (len >= IAVF_AQ_BUF_SZ) {
|
|
next_begin = i + 1;
|
|
break;
|
|
}
|
|
}
|
|
|
|
list = rte_zmalloc("iavf_del_mac_buffer", len, 0);
|
|
if (!list) {
|
|
PMD_DRV_LOG(ERR, "fail to allocate memory");
|
|
return;
|
|
}
|
|
|
|
for (i = begin; i < next_begin; i++) {
|
|
addr = &adapter->dev_data->mac_addrs[i];
|
|
if (rte_is_zero_ether_addr(addr))
|
|
continue;
|
|
rte_memcpy(list->list[j].addr, addr->addr_bytes,
|
|
sizeof(addr->addr_bytes));
|
|
list->list[j].type = (j == 0 ?
|
|
VIRTCHNL_ETHER_ADDR_PRIMARY :
|
|
VIRTCHNL_ETHER_ADDR_EXTRA);
|
|
PMD_DRV_LOG(DEBUG, "add/rm mac:" RTE_ETHER_ADDR_PRT_FMT,
|
|
RTE_ETHER_ADDR_BYTES(addr));
|
|
j++;
|
|
}
|
|
list->vsi_id = vf->vsi_res->vsi_id;
|
|
list->num_elements = j;
|
|
args.ops = add ? VIRTCHNL_OP_ADD_ETH_ADDR :
|
|
VIRTCHNL_OP_DEL_ETH_ADDR;
|
|
args.in_args = (uint8_t *)list;
|
|
args.in_args_size = len;
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
err = iavf_execute_vf_cmd(adapter, &args, 0);
|
|
if (err)
|
|
PMD_DRV_LOG(ERR, "fail to execute command %s",
|
|
add ? "OP_ADD_ETHER_ADDRESS" :
|
|
"OP_DEL_ETHER_ADDRESS");
|
|
rte_free(list);
|
|
begin = next_begin;
|
|
} while (begin < IAVF_NUM_MACADDR_MAX);
|
|
}
|
|
|
|
int
|
|
iavf_query_stats(struct iavf_adapter *adapter,
|
|
struct virtchnl_eth_stats **pstats)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_queue_select q_stats;
|
|
struct iavf_cmd_info args;
|
|
int err;
|
|
|
|
if (adapter->closed)
|
|
return -EIO;
|
|
|
|
memset(&q_stats, 0, sizeof(q_stats));
|
|
q_stats.vsi_id = vf->vsi_res->vsi_id;
|
|
args.ops = VIRTCHNL_OP_GET_STATS;
|
|
args.in_args = (uint8_t *)&q_stats;
|
|
args.in_args_size = sizeof(q_stats);
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
|
|
err = iavf_execute_vf_cmd(adapter, &args, 0);
|
|
if (err) {
|
|
PMD_DRV_LOG(ERR, "fail to execute command OP_GET_STATS");
|
|
*pstats = NULL;
|
|
return err;
|
|
}
|
|
*pstats = (struct virtchnl_eth_stats *)args.out_buffer;
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
iavf_config_promisc(struct iavf_adapter *adapter,
|
|
bool enable_unicast,
|
|
bool enable_multicast)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_promisc_info promisc;
|
|
struct iavf_cmd_info args;
|
|
int err;
|
|
|
|
if (adapter->closed)
|
|
return -EIO;
|
|
|
|
promisc.flags = 0;
|
|
promisc.vsi_id = vf->vsi_res->vsi_id;
|
|
|
|
if (enable_unicast)
|
|
promisc.flags |= FLAG_VF_UNICAST_PROMISC;
|
|
|
|
if (enable_multicast)
|
|
promisc.flags |= FLAG_VF_MULTICAST_PROMISC;
|
|
|
|
args.ops = VIRTCHNL_OP_CONFIG_PROMISCUOUS_MODE;
|
|
args.in_args = (uint8_t *)&promisc;
|
|
args.in_args_size = sizeof(promisc);
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
|
|
err = iavf_execute_vf_cmd(adapter, &args, 0);
|
|
|
|
if (err) {
|
|
PMD_DRV_LOG(ERR,
|
|
"fail to execute command CONFIG_PROMISCUOUS_MODE");
|
|
|
|
if (err == -ENOTSUP)
|
|
return err;
|
|
|
|
return -EAGAIN;
|
|
}
|
|
|
|
vf->promisc_unicast_enabled = enable_unicast;
|
|
vf->promisc_multicast_enabled = enable_multicast;
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
iavf_add_del_eth_addr(struct iavf_adapter *adapter, struct rte_ether_addr *addr,
|
|
bool add, uint8_t type)
|
|
{
|
|
struct virtchnl_ether_addr_list *list;
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
uint8_t cmd_buffer[sizeof(struct virtchnl_ether_addr_list) +
|
|
sizeof(struct virtchnl_ether_addr)];
|
|
struct iavf_cmd_info args;
|
|
int err;
|
|
|
|
if (adapter->closed)
|
|
return -EIO;
|
|
|
|
list = (struct virtchnl_ether_addr_list *)cmd_buffer;
|
|
list->vsi_id = vf->vsi_res->vsi_id;
|
|
list->num_elements = 1;
|
|
list->list[0].type = type;
|
|
rte_memcpy(list->list[0].addr, addr->addr_bytes,
|
|
sizeof(addr->addr_bytes));
|
|
|
|
args.ops = add ? VIRTCHNL_OP_ADD_ETH_ADDR : VIRTCHNL_OP_DEL_ETH_ADDR;
|
|
args.in_args = cmd_buffer;
|
|
args.in_args_size = sizeof(cmd_buffer);
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
err = iavf_execute_vf_cmd(adapter, &args, 0);
|
|
if (err)
|
|
PMD_DRV_LOG(ERR, "fail to execute command %s",
|
|
add ? "OP_ADD_ETH_ADDR" : "OP_DEL_ETH_ADDR");
|
|
return err;
|
|
}
|
|
|
|
int
|
|
iavf_add_del_vlan(struct iavf_adapter *adapter, uint16_t vlanid, bool add)
|
|
{
|
|
struct virtchnl_vlan_filter_list *vlan_list;
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
uint8_t cmd_buffer[sizeof(struct virtchnl_vlan_filter_list) +
|
|
sizeof(uint16_t)];
|
|
struct iavf_cmd_info args;
|
|
int err;
|
|
|
|
vlan_list = (struct virtchnl_vlan_filter_list *)cmd_buffer;
|
|
vlan_list->vsi_id = vf->vsi_res->vsi_id;
|
|
vlan_list->num_elements = 1;
|
|
vlan_list->vlan_id[0] = vlanid;
|
|
|
|
args.ops = add ? VIRTCHNL_OP_ADD_VLAN : VIRTCHNL_OP_DEL_VLAN;
|
|
args.in_args = cmd_buffer;
|
|
args.in_args_size = sizeof(cmd_buffer);
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
err = iavf_execute_vf_cmd(adapter, &args, 0);
|
|
if (err)
|
|
PMD_DRV_LOG(ERR, "fail to execute command %s",
|
|
add ? "OP_ADD_VLAN" : "OP_DEL_VLAN");
|
|
|
|
return err;
|
|
}
|
|
|
|
int
|
|
iavf_fdir_add(struct iavf_adapter *adapter,
|
|
struct iavf_fdir_conf *filter)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_fdir_add *fdir_ret;
|
|
|
|
struct iavf_cmd_info args;
|
|
int err;
|
|
|
|
filter->add_fltr.vsi_id = vf->vsi_res->vsi_id;
|
|
filter->add_fltr.validate_only = 0;
|
|
|
|
args.ops = VIRTCHNL_OP_ADD_FDIR_FILTER;
|
|
args.in_args = (uint8_t *)(&filter->add_fltr);
|
|
args.in_args_size = sizeof(*(&filter->add_fltr));
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
|
|
err = iavf_execute_vf_cmd(adapter, &args, 0);
|
|
if (err) {
|
|
PMD_DRV_LOG(ERR, "fail to execute command OP_ADD_FDIR_FILTER");
|
|
return err;
|
|
}
|
|
|
|
fdir_ret = (struct virtchnl_fdir_add *)args.out_buffer;
|
|
filter->flow_id = fdir_ret->flow_id;
|
|
|
|
if (fdir_ret->status == VIRTCHNL_FDIR_SUCCESS) {
|
|
PMD_DRV_LOG(INFO,
|
|
"Succeed in adding rule request by PF");
|
|
} else if (fdir_ret->status == VIRTCHNL_FDIR_FAILURE_RULE_NORESOURCE) {
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to add rule request due to no hw resource");
|
|
return -1;
|
|
} else if (fdir_ret->status == VIRTCHNL_FDIR_FAILURE_RULE_EXIST) {
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to add rule request due to the rule is already existed");
|
|
return -1;
|
|
} else if (fdir_ret->status == VIRTCHNL_FDIR_FAILURE_RULE_CONFLICT) {
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to add rule request due to the rule is conflict with existing rule");
|
|
return -1;
|
|
} else if (fdir_ret->status == VIRTCHNL_FDIR_FAILURE_RULE_INVALID) {
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to add rule request due to the hw doesn't support");
|
|
return -1;
|
|
} else if (fdir_ret->status == VIRTCHNL_FDIR_FAILURE_RULE_TIMEOUT) {
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to add rule request due to time out for programming");
|
|
return -1;
|
|
} else {
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to add rule request due to other reasons");
|
|
return -1;
|
|
}
|
|
|
|
return 0;
|
|
};
|
|
|
|
int
|
|
iavf_fdir_del(struct iavf_adapter *adapter,
|
|
struct iavf_fdir_conf *filter)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_fdir_del *fdir_ret;
|
|
|
|
struct iavf_cmd_info args;
|
|
int err;
|
|
|
|
filter->del_fltr.vsi_id = vf->vsi_res->vsi_id;
|
|
filter->del_fltr.flow_id = filter->flow_id;
|
|
|
|
args.ops = VIRTCHNL_OP_DEL_FDIR_FILTER;
|
|
args.in_args = (uint8_t *)(&filter->del_fltr);
|
|
args.in_args_size = sizeof(filter->del_fltr);
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
|
|
err = iavf_execute_vf_cmd(adapter, &args, 0);
|
|
if (err) {
|
|
PMD_DRV_LOG(ERR, "fail to execute command OP_DEL_FDIR_FILTER");
|
|
return err;
|
|
}
|
|
|
|
fdir_ret = (struct virtchnl_fdir_del *)args.out_buffer;
|
|
|
|
if (fdir_ret->status == VIRTCHNL_FDIR_SUCCESS) {
|
|
PMD_DRV_LOG(INFO,
|
|
"Succeed in deleting rule request by PF");
|
|
} else if (fdir_ret->status == VIRTCHNL_FDIR_FAILURE_RULE_NONEXIST) {
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to delete rule request due to this rule doesn't exist");
|
|
return -1;
|
|
} else if (fdir_ret->status == VIRTCHNL_FDIR_FAILURE_RULE_TIMEOUT) {
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to delete rule request due to time out for programming");
|
|
return -1;
|
|
} else {
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to delete rule request due to other reasons");
|
|
return -1;
|
|
}
|
|
|
|
return 0;
|
|
};
|
|
|
|
int
|
|
iavf_fdir_check(struct iavf_adapter *adapter,
|
|
struct iavf_fdir_conf *filter)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_fdir_add *fdir_ret;
|
|
|
|
struct iavf_cmd_info args;
|
|
int err;
|
|
|
|
filter->add_fltr.vsi_id = vf->vsi_res->vsi_id;
|
|
filter->add_fltr.validate_only = 1;
|
|
|
|
args.ops = VIRTCHNL_OP_ADD_FDIR_FILTER;
|
|
args.in_args = (uint8_t *)(&filter->add_fltr);
|
|
args.in_args_size = sizeof(*(&filter->add_fltr));
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
|
|
err = iavf_execute_vf_cmd(adapter, &args, 0);
|
|
if (err) {
|
|
PMD_DRV_LOG(ERR, "fail to check flow director rule");
|
|
return err;
|
|
}
|
|
|
|
fdir_ret = (struct virtchnl_fdir_add *)args.out_buffer;
|
|
|
|
if (fdir_ret->status == VIRTCHNL_FDIR_SUCCESS) {
|
|
PMD_DRV_LOG(INFO,
|
|
"Succeed in checking rule request by PF");
|
|
} else if (fdir_ret->status == VIRTCHNL_FDIR_FAILURE_RULE_INVALID) {
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to check rule request due to parameters validation"
|
|
" or HW doesn't support");
|
|
return -1;
|
|
} else {
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to check rule request due to other reasons");
|
|
return -1;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
iavf_flow_sub(struct iavf_adapter *adapter, struct iavf_fsub_conf *filter)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_flow_sub *fsub_cfg;
|
|
struct iavf_cmd_info args;
|
|
int err;
|
|
|
|
filter->sub_fltr.vsi_id = vf->vsi_res->vsi_id;
|
|
filter->sub_fltr.validate_only = 0;
|
|
|
|
memset(&args, 0, sizeof(args));
|
|
args.ops = VIRTCHNL_OP_FLOW_SUBSCRIBE;
|
|
args.in_args = (uint8_t *)(&filter->sub_fltr);
|
|
args.in_args_size = sizeof(*(&filter->sub_fltr));
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
|
|
err = iavf_execute_vf_cmd(adapter, &args, 0);
|
|
if (err)
|
|
PMD_DRV_LOG(ERR, "Failed to execute command of "
|
|
"OP_FLOW_SUBSCRIBE");
|
|
|
|
fsub_cfg = (struct virtchnl_flow_sub *)args.out_buffer;
|
|
filter->flow_id = fsub_cfg->flow_id;
|
|
|
|
if (fsub_cfg->status == VIRTCHNL_FSUB_SUCCESS) {
|
|
PMD_DRV_LOG(INFO, "Succeed in adding rule request by PF");
|
|
} else if (fsub_cfg->status == VIRTCHNL_FSUB_FAILURE_RULE_NORESOURCE) {
|
|
PMD_DRV_LOG(ERR, "Failed to add rule request due to no hw "
|
|
"resource");
|
|
err = -1;
|
|
} else if (fsub_cfg->status == VIRTCHNL_FSUB_FAILURE_RULE_EXIST) {
|
|
PMD_DRV_LOG(ERR, "Failed to add rule request due to the rule "
|
|
"is already existed");
|
|
err = -1;
|
|
} else if (fsub_cfg->status == VIRTCHNL_FSUB_FAILURE_RULE_INVALID) {
|
|
PMD_DRV_LOG(ERR, "Failed to add rule request due to the hw "
|
|
"doesn't support");
|
|
err = -1;
|
|
} else {
|
|
PMD_DRV_LOG(ERR, "Failed to add rule request due to other "
|
|
"reasons");
|
|
err = -1;
|
|
}
|
|
|
|
return err;
|
|
}
|
|
|
|
int
|
|
iavf_flow_unsub(struct iavf_adapter *adapter, struct iavf_fsub_conf *filter)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_flow_unsub *unsub_cfg;
|
|
struct iavf_cmd_info args;
|
|
int err;
|
|
|
|
filter->unsub_fltr.vsi_id = vf->vsi_res->vsi_id;
|
|
filter->unsub_fltr.flow_id = filter->flow_id;
|
|
|
|
memset(&args, 0, sizeof(args));
|
|
args.ops = VIRTCHNL_OP_FLOW_UNSUBSCRIBE;
|
|
args.in_args = (uint8_t *)(&filter->unsub_fltr);
|
|
args.in_args_size = sizeof(filter->unsub_fltr);
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
|
|
err = iavf_execute_vf_cmd(adapter, &args, 0);
|
|
if (err)
|
|
PMD_DRV_LOG(ERR, "Failed to execute command of "
|
|
"OP_FLOW_UNSUBSCRIBE");
|
|
|
|
unsub_cfg = (struct virtchnl_flow_unsub *)args.out_buffer;
|
|
|
|
if (unsub_cfg->status == VIRTCHNL_FSUB_SUCCESS) {
|
|
PMD_DRV_LOG(INFO, "Succeed in deleting rule request by PF");
|
|
} else if (unsub_cfg->status == VIRTCHNL_FSUB_FAILURE_RULE_NONEXIST) {
|
|
PMD_DRV_LOG(ERR, "Failed to delete rule request due to this "
|
|
"rule doesn't exist");
|
|
err = -1;
|
|
} else {
|
|
PMD_DRV_LOG(ERR, "Failed to delete rule request due to other "
|
|
"reasons");
|
|
err = -1;
|
|
}
|
|
|
|
return err;
|
|
}
|
|
|
|
int
|
|
iavf_flow_sub_check(struct iavf_adapter *adapter,
|
|
struct iavf_fsub_conf *filter)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_flow_sub *fsub_cfg;
|
|
|
|
struct iavf_cmd_info args;
|
|
int err;
|
|
|
|
filter->sub_fltr.vsi_id = vf->vsi_res->vsi_id;
|
|
filter->sub_fltr.validate_only = 1;
|
|
|
|
args.ops = VIRTCHNL_OP_FLOW_SUBSCRIBE;
|
|
args.in_args = (uint8_t *)(&filter->sub_fltr);
|
|
args.in_args_size = sizeof(*(&filter->sub_fltr));
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
|
|
err = iavf_execute_vf_cmd(adapter, &args, 0);
|
|
if (err) {
|
|
PMD_DRV_LOG(ERR, "fail to check flow director rule");
|
|
return err;
|
|
}
|
|
|
|
fsub_cfg = (struct virtchnl_flow_sub *)args.out_buffer;
|
|
|
|
if (fsub_cfg->status == VIRTCHNL_FSUB_SUCCESS) {
|
|
PMD_DRV_LOG(INFO, "Succeed in checking rule request by PF");
|
|
} else if (fsub_cfg->status == VIRTCHNL_FSUB_FAILURE_RULE_INVALID) {
|
|
PMD_DRV_LOG(ERR, "Failed to check rule request due to "
|
|
"parameters validation or HW doesn't "
|
|
"support");
|
|
err = -1;
|
|
} else {
|
|
PMD_DRV_LOG(ERR, "Failed to check rule request due to other "
|
|
"reasons");
|
|
err = -1;
|
|
}
|
|
|
|
return err;
|
|
}
|
|
|
|
int
|
|
iavf_add_del_rss_cfg(struct iavf_adapter *adapter,
|
|
struct virtchnl_rss_cfg *rss_cfg, bool add)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct iavf_cmd_info args;
|
|
int err;
|
|
|
|
memset(&args, 0, sizeof(args));
|
|
args.ops = add ? VIRTCHNL_OP_ADD_RSS_CFG :
|
|
VIRTCHNL_OP_DEL_RSS_CFG;
|
|
args.in_args = (u8 *)rss_cfg;
|
|
args.in_args_size = sizeof(*rss_cfg);
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
|
|
err = iavf_execute_vf_cmd(adapter, &args, 0);
|
|
if (err)
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to execute command of %s",
|
|
add ? "OP_ADD_RSS_CFG" :
|
|
"OP_DEL_RSS_INPUT_CFG");
|
|
|
|
return err;
|
|
}
|
|
|
|
int
|
|
iavf_get_hena_caps(struct iavf_adapter *adapter, uint64_t *caps)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct iavf_cmd_info args;
|
|
int err;
|
|
|
|
args.ops = VIRTCHNL_OP_GET_RSS_HENA_CAPS;
|
|
args.in_args = NULL;
|
|
args.in_args_size = 0;
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
|
|
err = iavf_execute_vf_cmd(adapter, &args, 0);
|
|
if (err) {
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to execute command of OP_GET_RSS_HENA_CAPS");
|
|
return err;
|
|
}
|
|
|
|
*caps = ((struct virtchnl_rss_hena *)args.out_buffer)->hena;
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
iavf_set_hena(struct iavf_adapter *adapter, uint64_t hena)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_rss_hena vrh;
|
|
struct iavf_cmd_info args;
|
|
int err;
|
|
|
|
vrh.hena = hena;
|
|
args.ops = VIRTCHNL_OP_SET_RSS_HENA;
|
|
args.in_args = (u8 *)&vrh;
|
|
args.in_args_size = sizeof(vrh);
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
|
|
err = iavf_execute_vf_cmd(adapter, &args, 0);
|
|
if (err)
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to execute command of OP_SET_RSS_HENA");
|
|
|
|
return err;
|
|
}
|
|
|
|
int
|
|
iavf_get_qos_cap(struct iavf_adapter *adapter)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct iavf_cmd_info args;
|
|
uint32_t len;
|
|
int err;
|
|
|
|
args.ops = VIRTCHNL_OP_GET_QOS_CAPS;
|
|
args.in_args = NULL;
|
|
args.in_args_size = 0;
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
err = iavf_execute_vf_cmd(adapter, &args, 0);
|
|
|
|
if (err) {
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to execute command of OP_GET_VF_RESOURCE");
|
|
return -1;
|
|
}
|
|
|
|
len = sizeof(struct virtchnl_qos_cap_list) +
|
|
IAVF_MAX_TRAFFIC_CLASS * sizeof(struct virtchnl_qos_cap_elem);
|
|
|
|
rte_memcpy(vf->qos_cap, args.out_buffer,
|
|
RTE_MIN(args.out_size, len));
|
|
|
|
return 0;
|
|
}
|
|
|
|
int iavf_set_q_tc_map(struct rte_eth_dev *dev,
|
|
struct virtchnl_queue_tc_mapping *q_tc_mapping, uint16_t size)
|
|
{
|
|
struct iavf_adapter *adapter =
|
|
IAVF_DEV_PRIVATE_TO_ADAPTER(dev->data->dev_private);
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(dev->data->dev_private);
|
|
struct iavf_cmd_info args;
|
|
int err;
|
|
|
|
memset(&args, 0, sizeof(args));
|
|
args.ops = VIRTCHNL_OP_CONFIG_QUEUE_TC_MAP;
|
|
args.in_args = (uint8_t *)q_tc_mapping;
|
|
args.in_args_size = size;
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
|
|
err = iavf_execute_vf_cmd(adapter, &args, 0);
|
|
if (err)
|
|
PMD_DRV_LOG(ERR, "Failed to execute command of"
|
|
" VIRTCHNL_OP_CONFIG_TC_MAP");
|
|
return err;
|
|
}
|
|
|
|
int iavf_set_q_bw(struct rte_eth_dev *dev,
|
|
struct virtchnl_queues_bw_cfg *q_bw, uint16_t size)
|
|
{
|
|
struct iavf_adapter *adapter =
|
|
IAVF_DEV_PRIVATE_TO_ADAPTER(dev->data->dev_private);
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(dev->data->dev_private);
|
|
struct iavf_cmd_info args;
|
|
int err;
|
|
|
|
memset(&args, 0, sizeof(args));
|
|
args.ops = VIRTCHNL_OP_CONFIG_QUEUE_BW;
|
|
args.in_args = (uint8_t *)q_bw;
|
|
args.in_args_size = size;
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
|
|
err = iavf_execute_vf_cmd(adapter, &args, 0);
|
|
if (err)
|
|
PMD_DRV_LOG(ERR, "Failed to execute command of"
|
|
" VIRTCHNL_OP_CONFIG_QUEUE_BW");
|
|
return err;
|
|
}
|
|
|
|
int
|
|
iavf_add_del_mc_addr_list(struct iavf_adapter *adapter,
|
|
struct rte_ether_addr *mc_addrs,
|
|
uint32_t mc_addrs_num, bool add)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
uint8_t cmd_buffer[sizeof(struct virtchnl_ether_addr_list) +
|
|
(IAVF_NUM_MACADDR_MAX * sizeof(struct virtchnl_ether_addr))];
|
|
struct virtchnl_ether_addr_list *list;
|
|
struct iavf_cmd_info args;
|
|
uint32_t i;
|
|
int err;
|
|
|
|
if (mc_addrs == NULL || mc_addrs_num == 0)
|
|
return 0;
|
|
|
|
list = (struct virtchnl_ether_addr_list *)cmd_buffer;
|
|
list->vsi_id = vf->vsi_res->vsi_id;
|
|
list->num_elements = mc_addrs_num;
|
|
|
|
for (i = 0; i < mc_addrs_num; i++) {
|
|
if (!IAVF_IS_MULTICAST(mc_addrs[i].addr_bytes)) {
|
|
PMD_DRV_LOG(ERR, "Invalid mac:" RTE_ETHER_ADDR_PRT_FMT,
|
|
RTE_ETHER_ADDR_BYTES(&mc_addrs[i]));
|
|
return -EINVAL;
|
|
}
|
|
|
|
memcpy(list->list[i].addr, mc_addrs[i].addr_bytes,
|
|
sizeof(list->list[i].addr));
|
|
list->list[i].type = VIRTCHNL_ETHER_ADDR_EXTRA;
|
|
}
|
|
|
|
args.ops = add ? VIRTCHNL_OP_ADD_ETH_ADDR : VIRTCHNL_OP_DEL_ETH_ADDR;
|
|
args.in_args = cmd_buffer;
|
|
args.in_args_size = sizeof(struct virtchnl_ether_addr_list) +
|
|
i * sizeof(struct virtchnl_ether_addr);
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
err = iavf_execute_vf_cmd(adapter, &args, 0);
|
|
|
|
if (err) {
|
|
PMD_DRV_LOG(ERR, "fail to execute command %s",
|
|
add ? "OP_ADD_ETH_ADDR" : "OP_DEL_ETH_ADDR");
|
|
return err;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
iavf_request_queues(struct rte_eth_dev *dev, uint16_t num)
|
|
{
|
|
struct iavf_adapter *adapter =
|
|
IAVF_DEV_PRIVATE_TO_ADAPTER(dev->data->dev_private);
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct rte_pci_device *pci_dev = RTE_ETH_DEV_TO_PCI(dev);
|
|
struct virtchnl_vf_res_request vfres;
|
|
struct iavf_cmd_info args;
|
|
uint16_t num_queue_pairs;
|
|
int err;
|
|
|
|
if (!(vf->vf_res->vf_cap_flags &
|
|
VIRTCHNL_VF_OFFLOAD_REQ_QUEUES)) {
|
|
PMD_DRV_LOG(ERR, "request queues not supported");
|
|
return -1;
|
|
}
|
|
|
|
if (num == 0) {
|
|
PMD_DRV_LOG(ERR, "queue number cannot be zero");
|
|
return -1;
|
|
}
|
|
vfres.num_queue_pairs = num;
|
|
|
|
args.ops = VIRTCHNL_OP_REQUEST_QUEUES;
|
|
args.in_args = (u8 *)&vfres;
|
|
args.in_args_size = sizeof(vfres);
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
|
|
if (vf->vf_res->vf_cap_flags & VIRTCHNL_VF_OFFLOAD_WB_ON_ITR) {
|
|
/* disable interrupt to avoid the admin queue message to be read
|
|
* before iavf_read_msg_from_pf.
|
|
*/
|
|
rte_intr_disable(pci_dev->intr_handle);
|
|
err = iavf_execute_vf_cmd(adapter, &args, 0);
|
|
rte_intr_enable(pci_dev->intr_handle);
|
|
} else {
|
|
rte_eal_alarm_cancel(iavf_dev_alarm_handler, dev);
|
|
err = iavf_execute_vf_cmd(adapter, &args, 0);
|
|
rte_eal_alarm_set(IAVF_ALARM_INTERVAL,
|
|
iavf_dev_alarm_handler, dev);
|
|
}
|
|
|
|
if (err) {
|
|
PMD_DRV_LOG(ERR, "fail to execute command OP_REQUEST_QUEUES");
|
|
return err;
|
|
}
|
|
|
|
/* request queues succeeded, vf is resetting */
|
|
if (vf->vf_reset) {
|
|
PMD_DRV_LOG(INFO, "vf is resetting");
|
|
return 0;
|
|
}
|
|
|
|
/* request additional queues failed, return available number */
|
|
num_queue_pairs =
|
|
((struct virtchnl_vf_res_request *)args.out_buffer)->num_queue_pairs;
|
|
PMD_DRV_LOG(ERR, "request queues failed, only %u queues "
|
|
"available", num_queue_pairs);
|
|
|
|
return -1;
|
|
}
|
|
|
|
int
|
|
iavf_get_max_rss_queue_region(struct iavf_adapter *adapter)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct iavf_cmd_info args;
|
|
uint16_t qregion_width;
|
|
int err;
|
|
|
|
args.ops = VIRTCHNL_OP_GET_MAX_RSS_QREGION;
|
|
args.in_args = NULL;
|
|
args.in_args_size = 0;
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
|
|
err = iavf_execute_vf_cmd(adapter, &args, 0);
|
|
if (err) {
|
|
PMD_DRV_LOG(ERR, "Failed to execute command of VIRTCHNL_OP_GET_MAX_RSS_QREGION");
|
|
return err;
|
|
}
|
|
|
|
qregion_width =
|
|
((struct virtchnl_max_rss_qregion *)args.out_buffer)->qregion_width;
|
|
|
|
vf->max_rss_qregion = (uint16_t)(1 << qregion_width);
|
|
|
|
return 0;
|
|
}
|
|
|
|
|
|
|
|
int
|
|
iavf_ipsec_crypto_request(struct iavf_adapter *adapter,
|
|
uint8_t *msg, size_t msg_len,
|
|
uint8_t *resp_msg, size_t resp_msg_len)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct iavf_cmd_info args;
|
|
int err;
|
|
|
|
args.ops = VIRTCHNL_OP_INLINE_IPSEC_CRYPTO;
|
|
args.in_args = msg;
|
|
args.in_args_size = msg_len;
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
|
|
err = iavf_execute_vf_cmd(adapter, &args, 1);
|
|
if (err) {
|
|
PMD_DRV_LOG(ERR, "fail to execute command %s",
|
|
"OP_INLINE_IPSEC_CRYPTO");
|
|
return err;
|
|
}
|
|
|
|
memcpy(resp_msg, args.out_buffer, resp_msg_len);
|
|
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
iavf_set_vf_quanta_size(struct iavf_adapter *adapter, u16 start_queue_id, u16 num_queues)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct iavf_cmd_info args;
|
|
struct virtchnl_quanta_cfg q_quanta;
|
|
int err;
|
|
|
|
if (adapter->devargs.quanta_size == 0)
|
|
return 0;
|
|
|
|
q_quanta.quanta_size = adapter->devargs.quanta_size;
|
|
q_quanta.queue_select.type = VIRTCHNL_QUEUE_TYPE_TX;
|
|
q_quanta.queue_select.start_queue_id = start_queue_id;
|
|
q_quanta.queue_select.num_queues = num_queues;
|
|
|
|
args.ops = VIRTCHNL_OP_CONFIG_QUANTA;
|
|
args.in_args = (uint8_t *)&q_quanta;
|
|
args.in_args_size = sizeof(q_quanta);
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
|
|
err = iavf_execute_vf_cmd(adapter, &args, 0);
|
|
if (err) {
|
|
PMD_DRV_LOG(ERR, "Failed to execute command VIRTCHNL_OP_CONFIG_QUANTA");
|
|
return err;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
iavf_get_ptp_cap(struct iavf_adapter *adapter)
|
|
{
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_ptp_caps ptp_caps;
|
|
struct iavf_cmd_info args;
|
|
int err;
|
|
|
|
ptp_caps.caps = VIRTCHNL_1588_PTP_CAP_RX_TSTAMP |
|
|
VIRTCHNL_1588_PTP_CAP_READ_PHC;
|
|
|
|
args.ops = VIRTCHNL_OP_1588_PTP_GET_CAPS;
|
|
args.in_args = (uint8_t *)&ptp_caps;
|
|
args.in_args_size = sizeof(ptp_caps);
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
|
|
err = iavf_execute_vf_cmd(adapter, &args, 0);
|
|
if (err) {
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to execute command of OP_1588_PTP_GET_CAPS");
|
|
return err;
|
|
}
|
|
|
|
vf->ptp_caps = ((struct virtchnl_ptp_caps *)args.out_buffer)->caps;
|
|
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
iavf_get_phc_time(struct iavf_rx_queue *rxq)
|
|
{
|
|
struct iavf_adapter *adapter = rxq->vsi->adapter;
|
|
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
|
|
struct virtchnl_phc_time phc_time;
|
|
struct iavf_cmd_info args;
|
|
int err = 0;
|
|
|
|
args.ops = VIRTCHNL_OP_1588_PTP_GET_TIME;
|
|
args.in_args = (uint8_t *)&phc_time;
|
|
args.in_args_size = sizeof(phc_time);
|
|
args.out_buffer = vf->aq_resp;
|
|
args.out_size = IAVF_AQ_BUF_SZ;
|
|
|
|
rte_spinlock_lock(&vf->phc_time_aq_lock);
|
|
err = iavf_execute_vf_cmd(adapter, &args, 0);
|
|
if (err) {
|
|
PMD_DRV_LOG(ERR,
|
|
"Failed to execute command of VIRTCHNL_OP_1588_PTP_GET_TIME");
|
|
goto out;
|
|
}
|
|
rxq->phc_time = ((struct virtchnl_phc_time *)args.out_buffer)->time;
|
|
|
|
out:
|
|
rte_spinlock_unlock(&vf->phc_time_aq_lock);
|
|
return err;
|
|
}
|