numam-dpdk/drivers/net/atlantic/atl_ethdev.c
Jerin Jacob 9c99878aa1 log: introduce logtype register macro
Introduce the RTE_LOG_REGISTER macro to avoid the code duplication
in the logtype registration process.

It is a wrapper macro for declaring the logtype, registering it and
setting its level in the constructor context.

Signed-off-by: Jerin Jacob <jerinj@marvell.com>
Acked-by: Adam Dybkowski <adamx.dybkowski@intel.com>
Acked-by: Sachin Saxena <sachin.saxena@nxp.com>
Acked-by: Akhil Goyal <akhil.goyal@nxp.com>
2020-07-03 15:52:51 +02:00

1931 lines
50 KiB
C

/* SPDX-License-Identifier: BSD-3-Clause
* Copyright(c) 2018 Aquantia Corporation
*/
#include <rte_string_fns.h>
#include <rte_ethdev_pci.h>
#include <rte_alarm.h>
#include "atl_ethdev.h"
#include "atl_common.h"
#include "atl_hw_regs.h"
#include "atl_logs.h"
#include "hw_atl/hw_atl_llh.h"
#include "hw_atl/hw_atl_b0.h"
#include "hw_atl/hw_atl_b0_internal.h"
static int eth_atl_dev_init(struct rte_eth_dev *eth_dev);
static int eth_atl_dev_uninit(struct rte_eth_dev *eth_dev);
static int atl_dev_configure(struct rte_eth_dev *dev);
static int atl_dev_start(struct rte_eth_dev *dev);
static void atl_dev_stop(struct rte_eth_dev *dev);
static int atl_dev_set_link_up(struct rte_eth_dev *dev);
static int atl_dev_set_link_down(struct rte_eth_dev *dev);
static void atl_dev_close(struct rte_eth_dev *dev);
static int atl_dev_reset(struct rte_eth_dev *dev);
static int atl_dev_promiscuous_enable(struct rte_eth_dev *dev);
static int atl_dev_promiscuous_disable(struct rte_eth_dev *dev);
static int atl_dev_allmulticast_enable(struct rte_eth_dev *dev);
static int atl_dev_allmulticast_disable(struct rte_eth_dev *dev);
static int atl_dev_link_update(struct rte_eth_dev *dev, int wait);
static int atl_dev_xstats_get_names(struct rte_eth_dev *dev __rte_unused,
struct rte_eth_xstat_name *xstats_names,
unsigned int size);
static int atl_dev_stats_get(struct rte_eth_dev *dev,
struct rte_eth_stats *stats);
static int atl_dev_xstats_get(struct rte_eth_dev *dev,
struct rte_eth_xstat *stats, unsigned int n);
static int atl_dev_stats_reset(struct rte_eth_dev *dev);
static int atl_fw_version_get(struct rte_eth_dev *dev, char *fw_version,
size_t fw_size);
static const uint32_t *atl_dev_supported_ptypes_get(struct rte_eth_dev *dev);
static int atl_dev_mtu_set(struct rte_eth_dev *dev, uint16_t mtu);
/* VLAN stuff */
static int atl_vlan_filter_set(struct rte_eth_dev *dev,
uint16_t vlan_id, int on);
static int atl_vlan_offload_set(struct rte_eth_dev *dev, int mask);
static void atl_vlan_strip_queue_set(struct rte_eth_dev *dev,
uint16_t queue_id, int on);
static int atl_vlan_tpid_set(struct rte_eth_dev *dev,
enum rte_vlan_type vlan_type, uint16_t tpid);
/* EEPROM */
static int atl_dev_get_eeprom_length(struct rte_eth_dev *dev);
static int atl_dev_get_eeprom(struct rte_eth_dev *dev,
struct rte_dev_eeprom_info *eeprom);
static int atl_dev_set_eeprom(struct rte_eth_dev *dev,
struct rte_dev_eeprom_info *eeprom);
/* Regs */
static int atl_dev_get_regs(struct rte_eth_dev *dev,
struct rte_dev_reg_info *regs);
/* Flow control */
static int atl_flow_ctrl_get(struct rte_eth_dev *dev,
struct rte_eth_fc_conf *fc_conf);
static int atl_flow_ctrl_set(struct rte_eth_dev *dev,
struct rte_eth_fc_conf *fc_conf);
static void atl_dev_link_status_print(struct rte_eth_dev *dev);
/* Interrupts */
static int atl_dev_rxq_interrupt_setup(struct rte_eth_dev *dev);
static int atl_dev_lsc_interrupt_setup(struct rte_eth_dev *dev, uint8_t on);
static int atl_dev_interrupt_get_status(struct rte_eth_dev *dev);
static int atl_dev_interrupt_action(struct rte_eth_dev *dev,
struct rte_intr_handle *handle);
static void atl_dev_interrupt_handler(void *param);
static int atl_add_mac_addr(struct rte_eth_dev *dev,
struct rte_ether_addr *mac_addr,
uint32_t index, uint32_t pool);
static void atl_remove_mac_addr(struct rte_eth_dev *dev, uint32_t index);
static int atl_set_default_mac_addr(struct rte_eth_dev *dev,
struct rte_ether_addr *mac_addr);
static int atl_dev_set_mc_addr_list(struct rte_eth_dev *dev,
struct rte_ether_addr *mc_addr_set,
uint32_t nb_mc_addr);
/* RSS */
static int atl_reta_update(struct rte_eth_dev *dev,
struct rte_eth_rss_reta_entry64 *reta_conf,
uint16_t reta_size);
static int atl_reta_query(struct rte_eth_dev *dev,
struct rte_eth_rss_reta_entry64 *reta_conf,
uint16_t reta_size);
static int atl_rss_hash_update(struct rte_eth_dev *dev,
struct rte_eth_rss_conf *rss_conf);
static int atl_rss_hash_conf_get(struct rte_eth_dev *dev,
struct rte_eth_rss_conf *rss_conf);
static int eth_atl_pci_probe(struct rte_pci_driver *pci_drv __rte_unused,
struct rte_pci_device *pci_dev);
static int eth_atl_pci_remove(struct rte_pci_device *pci_dev);
static int atl_dev_info_get(struct rte_eth_dev *dev,
struct rte_eth_dev_info *dev_info);
/*
* The set of PCI devices this driver supports
*/
static const struct rte_pci_id pci_id_atl_map[] = {
{ RTE_PCI_DEVICE(PCI_VENDOR_ID_AQUANTIA, AQ_DEVICE_ID_0001) },
{ RTE_PCI_DEVICE(PCI_VENDOR_ID_AQUANTIA, AQ_DEVICE_ID_D100) },
{ RTE_PCI_DEVICE(PCI_VENDOR_ID_AQUANTIA, AQ_DEVICE_ID_D107) },
{ RTE_PCI_DEVICE(PCI_VENDOR_ID_AQUANTIA, AQ_DEVICE_ID_D108) },
{ RTE_PCI_DEVICE(PCI_VENDOR_ID_AQUANTIA, AQ_DEVICE_ID_D109) },
{ RTE_PCI_DEVICE(PCI_VENDOR_ID_AQUANTIA, AQ_DEVICE_ID_AQC100) },
{ RTE_PCI_DEVICE(PCI_VENDOR_ID_AQUANTIA, AQ_DEVICE_ID_AQC107) },
{ RTE_PCI_DEVICE(PCI_VENDOR_ID_AQUANTIA, AQ_DEVICE_ID_AQC108) },
{ RTE_PCI_DEVICE(PCI_VENDOR_ID_AQUANTIA, AQ_DEVICE_ID_AQC109) },
{ RTE_PCI_DEVICE(PCI_VENDOR_ID_AQUANTIA, AQ_DEVICE_ID_AQC111) },
{ RTE_PCI_DEVICE(PCI_VENDOR_ID_AQUANTIA, AQ_DEVICE_ID_AQC112) },
{ RTE_PCI_DEVICE(PCI_VENDOR_ID_AQUANTIA, AQ_DEVICE_ID_AQC100S) },
{ RTE_PCI_DEVICE(PCI_VENDOR_ID_AQUANTIA, AQ_DEVICE_ID_AQC107S) },
{ RTE_PCI_DEVICE(PCI_VENDOR_ID_AQUANTIA, AQ_DEVICE_ID_AQC108S) },
{ RTE_PCI_DEVICE(PCI_VENDOR_ID_AQUANTIA, AQ_DEVICE_ID_AQC109S) },
{ RTE_PCI_DEVICE(PCI_VENDOR_ID_AQUANTIA, AQ_DEVICE_ID_AQC111S) },
{ RTE_PCI_DEVICE(PCI_VENDOR_ID_AQUANTIA, AQ_DEVICE_ID_AQC112S) },
{ RTE_PCI_DEVICE(PCI_VENDOR_ID_AQUANTIA, AQ_DEVICE_ID_AQC111E) },
{ RTE_PCI_DEVICE(PCI_VENDOR_ID_AQUANTIA, AQ_DEVICE_ID_AQC112E) },
{ .vendor_id = 0, /* sentinel */ },
};
static struct rte_pci_driver rte_atl_pmd = {
.id_table = pci_id_atl_map,
.drv_flags = RTE_PCI_DRV_NEED_MAPPING | RTE_PCI_DRV_INTR_LSC,
.probe = eth_atl_pci_probe,
.remove = eth_atl_pci_remove,
};
#define ATL_RX_OFFLOADS (DEV_RX_OFFLOAD_VLAN_STRIP \
| DEV_RX_OFFLOAD_IPV4_CKSUM \
| DEV_RX_OFFLOAD_UDP_CKSUM \
| DEV_RX_OFFLOAD_TCP_CKSUM \
| DEV_RX_OFFLOAD_JUMBO_FRAME \
| DEV_RX_OFFLOAD_MACSEC_STRIP \
| DEV_RX_OFFLOAD_VLAN_FILTER)
#define ATL_TX_OFFLOADS (DEV_TX_OFFLOAD_VLAN_INSERT \
| DEV_TX_OFFLOAD_IPV4_CKSUM \
| DEV_TX_OFFLOAD_UDP_CKSUM \
| DEV_TX_OFFLOAD_TCP_CKSUM \
| DEV_TX_OFFLOAD_TCP_TSO \
| DEV_TX_OFFLOAD_MACSEC_INSERT \
| DEV_TX_OFFLOAD_MULTI_SEGS)
#define SFP_EEPROM_SIZE 0x100
static const struct rte_eth_desc_lim rx_desc_lim = {
.nb_max = ATL_MAX_RING_DESC,
.nb_min = ATL_MIN_RING_DESC,
.nb_align = ATL_RXD_ALIGN,
};
static const struct rte_eth_desc_lim tx_desc_lim = {
.nb_max = ATL_MAX_RING_DESC,
.nb_min = ATL_MIN_RING_DESC,
.nb_align = ATL_TXD_ALIGN,
.nb_seg_max = ATL_TX_MAX_SEG,
.nb_mtu_seg_max = ATL_TX_MAX_SEG,
};
enum atl_xstats_type {
XSTATS_TYPE_MSM = 0,
XSTATS_TYPE_MACSEC,
};
#define ATL_XSTATS_FIELD(name) { \
#name, \
offsetof(struct aq_stats_s, name), \
XSTATS_TYPE_MSM \
}
#define ATL_MACSEC_XSTATS_FIELD(name) { \
#name, \
offsetof(struct macsec_stats, name), \
XSTATS_TYPE_MACSEC \
}
struct atl_xstats_tbl_s {
const char *name;
unsigned int offset;
enum atl_xstats_type type;
};
static struct atl_xstats_tbl_s atl_xstats_tbl[] = {
ATL_XSTATS_FIELD(uprc),
ATL_XSTATS_FIELD(mprc),
ATL_XSTATS_FIELD(bprc),
ATL_XSTATS_FIELD(erpt),
ATL_XSTATS_FIELD(uptc),
ATL_XSTATS_FIELD(mptc),
ATL_XSTATS_FIELD(bptc),
ATL_XSTATS_FIELD(erpr),
ATL_XSTATS_FIELD(ubrc),
ATL_XSTATS_FIELD(ubtc),
ATL_XSTATS_FIELD(mbrc),
ATL_XSTATS_FIELD(mbtc),
ATL_XSTATS_FIELD(bbrc),
ATL_XSTATS_FIELD(bbtc),
/* Ingress Common Counters */
ATL_MACSEC_XSTATS_FIELD(in_ctl_pkts),
ATL_MACSEC_XSTATS_FIELD(in_tagged_miss_pkts),
ATL_MACSEC_XSTATS_FIELD(in_untagged_miss_pkts),
ATL_MACSEC_XSTATS_FIELD(in_notag_pkts),
ATL_MACSEC_XSTATS_FIELD(in_untagged_pkts),
ATL_MACSEC_XSTATS_FIELD(in_bad_tag_pkts),
ATL_MACSEC_XSTATS_FIELD(in_no_sci_pkts),
ATL_MACSEC_XSTATS_FIELD(in_unknown_sci_pkts),
/* Ingress SA Counters */
ATL_MACSEC_XSTATS_FIELD(in_untagged_hit_pkts),
ATL_MACSEC_XSTATS_FIELD(in_not_using_sa),
ATL_MACSEC_XSTATS_FIELD(in_unused_sa),
ATL_MACSEC_XSTATS_FIELD(in_not_valid_pkts),
ATL_MACSEC_XSTATS_FIELD(in_invalid_pkts),
ATL_MACSEC_XSTATS_FIELD(in_ok_pkts),
ATL_MACSEC_XSTATS_FIELD(in_unchecked_pkts),
ATL_MACSEC_XSTATS_FIELD(in_validated_octets),
ATL_MACSEC_XSTATS_FIELD(in_decrypted_octets),
/* Egress Common Counters */
ATL_MACSEC_XSTATS_FIELD(out_ctl_pkts),
ATL_MACSEC_XSTATS_FIELD(out_unknown_sa_pkts),
ATL_MACSEC_XSTATS_FIELD(out_untagged_pkts),
ATL_MACSEC_XSTATS_FIELD(out_too_long),
/* Egress SC Counters */
ATL_MACSEC_XSTATS_FIELD(out_sc_protected_pkts),
ATL_MACSEC_XSTATS_FIELD(out_sc_encrypted_pkts),
/* Egress SA Counters */
ATL_MACSEC_XSTATS_FIELD(out_sa_hit_drop_redirect),
ATL_MACSEC_XSTATS_FIELD(out_sa_protected2_pkts),
ATL_MACSEC_XSTATS_FIELD(out_sa_protected_pkts),
ATL_MACSEC_XSTATS_FIELD(out_sa_encrypted_pkts),
};
static const struct eth_dev_ops atl_eth_dev_ops = {
.dev_configure = atl_dev_configure,
.dev_start = atl_dev_start,
.dev_stop = atl_dev_stop,
.dev_set_link_up = atl_dev_set_link_up,
.dev_set_link_down = atl_dev_set_link_down,
.dev_close = atl_dev_close,
.dev_reset = atl_dev_reset,
/* PROMISC */
.promiscuous_enable = atl_dev_promiscuous_enable,
.promiscuous_disable = atl_dev_promiscuous_disable,
.allmulticast_enable = atl_dev_allmulticast_enable,
.allmulticast_disable = atl_dev_allmulticast_disable,
/* Link */
.link_update = atl_dev_link_update,
.get_reg = atl_dev_get_regs,
/* Stats */
.stats_get = atl_dev_stats_get,
.xstats_get = atl_dev_xstats_get,
.xstats_get_names = atl_dev_xstats_get_names,
.stats_reset = atl_dev_stats_reset,
.xstats_reset = atl_dev_stats_reset,
.fw_version_get = atl_fw_version_get,
.dev_infos_get = atl_dev_info_get,
.dev_supported_ptypes_get = atl_dev_supported_ptypes_get,
.mtu_set = atl_dev_mtu_set,
/* VLAN */
.vlan_filter_set = atl_vlan_filter_set,
.vlan_offload_set = atl_vlan_offload_set,
.vlan_tpid_set = atl_vlan_tpid_set,
.vlan_strip_queue_set = atl_vlan_strip_queue_set,
/* Queue Control */
.rx_queue_start = atl_rx_queue_start,
.rx_queue_stop = atl_rx_queue_stop,
.rx_queue_setup = atl_rx_queue_setup,
.rx_queue_release = atl_rx_queue_release,
.tx_queue_start = atl_tx_queue_start,
.tx_queue_stop = atl_tx_queue_stop,
.tx_queue_setup = atl_tx_queue_setup,
.tx_queue_release = atl_tx_queue_release,
.rx_queue_intr_enable = atl_dev_rx_queue_intr_enable,
.rx_queue_intr_disable = atl_dev_rx_queue_intr_disable,
.rx_queue_count = atl_rx_queue_count,
.rx_descriptor_status = atl_dev_rx_descriptor_status,
.tx_descriptor_status = atl_dev_tx_descriptor_status,
/* EEPROM */
.get_eeprom_length = atl_dev_get_eeprom_length,
.get_eeprom = atl_dev_get_eeprom,
.set_eeprom = atl_dev_set_eeprom,
/* Flow Control */
.flow_ctrl_get = atl_flow_ctrl_get,
.flow_ctrl_set = atl_flow_ctrl_set,
/* MAC */
.mac_addr_add = atl_add_mac_addr,
.mac_addr_remove = atl_remove_mac_addr,
.mac_addr_set = atl_set_default_mac_addr,
.set_mc_addr_list = atl_dev_set_mc_addr_list,
.rxq_info_get = atl_rxq_info_get,
.txq_info_get = atl_txq_info_get,
.reta_update = atl_reta_update,
.reta_query = atl_reta_query,
.rss_hash_update = atl_rss_hash_update,
.rss_hash_conf_get = atl_rss_hash_conf_get,
};
static inline int32_t
atl_reset_hw(struct aq_hw_s *hw)
{
return hw_atl_b0_hw_reset(hw);
}
static inline void
atl_enable_intr(struct rte_eth_dev *dev)
{
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
hw_atl_itr_irq_msk_setlsw_set(hw, 0xffffffff);
}
static void
atl_disable_intr(struct aq_hw_s *hw)
{
PMD_INIT_FUNC_TRACE();
hw_atl_itr_irq_msk_clearlsw_set(hw, 0xffffffff);
}
static int
eth_atl_dev_init(struct rte_eth_dev *eth_dev)
{
struct atl_adapter *adapter = eth_dev->data->dev_private;
struct rte_pci_device *pci_dev = RTE_ETH_DEV_TO_PCI(eth_dev);
struct rte_intr_handle *intr_handle = &pci_dev->intr_handle;
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(eth_dev->data->dev_private);
int err = 0;
PMD_INIT_FUNC_TRACE();
eth_dev->dev_ops = &atl_eth_dev_ops;
eth_dev->rx_pkt_burst = &atl_recv_pkts;
eth_dev->tx_pkt_burst = &atl_xmit_pkts;
eth_dev->tx_pkt_prepare = &atl_prep_pkts;
/* For secondary processes, the primary process has done all the work */
if (rte_eal_process_type() != RTE_PROC_PRIMARY)
return 0;
/* Vendor and Device ID need to be set before init of shared code */
hw->device_id = pci_dev->id.device_id;
hw->vendor_id = pci_dev->id.vendor_id;
hw->mmio = (void *)pci_dev->mem_resource[0].addr;
/* Hardware configuration - hardcode */
adapter->hw_cfg.is_lro = false;
adapter->hw_cfg.wol = false;
adapter->hw_cfg.is_rss = false;
adapter->hw_cfg.num_rss_queues = HW_ATL_B0_RSS_MAX;
adapter->hw_cfg.link_speed_msk = AQ_NIC_RATE_10G |
AQ_NIC_RATE_5G |
AQ_NIC_RATE_2G5 |
AQ_NIC_RATE_1G |
AQ_NIC_RATE_100M;
adapter->hw_cfg.flow_control = (AQ_NIC_FC_RX | AQ_NIC_FC_TX);
adapter->hw_cfg.aq_rss.indirection_table_size =
HW_ATL_B0_RSS_REDIRECTION_MAX;
hw->aq_nic_cfg = &adapter->hw_cfg;
pthread_mutex_init(&hw->mbox_mutex, NULL);
/* disable interrupt */
atl_disable_intr(hw);
/* Allocate memory for storing MAC addresses */
eth_dev->data->mac_addrs = rte_zmalloc("atlantic",
RTE_ETHER_ADDR_LEN, 0);
if (eth_dev->data->mac_addrs == NULL) {
PMD_INIT_LOG(ERR, "MAC Malloc failed");
return -ENOMEM;
}
err = hw_atl_utils_initfw(hw, &hw->aq_fw_ops);
if (err)
return err;
/* Copy the permanent MAC address */
if (hw->aq_fw_ops->get_mac_permanent(hw,
eth_dev->data->mac_addrs->addr_bytes) != 0)
return -EINVAL;
/* Reset the hw statistics */
atl_dev_stats_reset(eth_dev);
rte_intr_callback_register(intr_handle,
atl_dev_interrupt_handler, eth_dev);
/* enable uio/vfio intr/eventfd mapping */
rte_intr_enable(intr_handle);
/* enable support intr */
atl_enable_intr(eth_dev);
return err;
}
static int
eth_atl_dev_uninit(struct rte_eth_dev *eth_dev)
{
struct rte_pci_device *pci_dev = RTE_ETH_DEV_TO_PCI(eth_dev);
struct rte_intr_handle *intr_handle = &pci_dev->intr_handle;
struct aq_hw_s *hw;
PMD_INIT_FUNC_TRACE();
if (rte_eal_process_type() != RTE_PROC_PRIMARY)
return -EPERM;
hw = ATL_DEV_PRIVATE_TO_HW(eth_dev->data->dev_private);
if (hw->adapter_stopped == 0)
atl_dev_close(eth_dev);
eth_dev->dev_ops = NULL;
eth_dev->rx_pkt_burst = NULL;
eth_dev->tx_pkt_burst = NULL;
/* disable uio intr before callback unregister */
rte_intr_disable(intr_handle);
rte_intr_callback_unregister(intr_handle,
atl_dev_interrupt_handler, eth_dev);
rte_free(eth_dev->data->mac_addrs);
eth_dev->data->mac_addrs = NULL;
pthread_mutex_destroy(&hw->mbox_mutex);
return 0;
}
static int
eth_atl_pci_probe(struct rte_pci_driver *pci_drv __rte_unused,
struct rte_pci_device *pci_dev)
{
return rte_eth_dev_pci_generic_probe(pci_dev,
sizeof(struct atl_adapter), eth_atl_dev_init);
}
static int
eth_atl_pci_remove(struct rte_pci_device *pci_dev)
{
return rte_eth_dev_pci_generic_remove(pci_dev, eth_atl_dev_uninit);
}
static int
atl_dev_configure(struct rte_eth_dev *dev)
{
struct atl_interrupt *intr =
ATL_DEV_PRIVATE_TO_INTR(dev->data->dev_private);
PMD_INIT_FUNC_TRACE();
/* set flag to update link status after init */
intr->flags |= ATL_FLAG_NEED_LINK_UPDATE;
return 0;
}
/*
* Configure device link speed and setup link.
* It returns 0 on success.
*/
static int
atl_dev_start(struct rte_eth_dev *dev)
{
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
struct rte_pci_device *pci_dev = RTE_ETH_DEV_TO_PCI(dev);
struct rte_intr_handle *intr_handle = &pci_dev->intr_handle;
uint32_t intr_vector = 0;
int status;
int err;
PMD_INIT_FUNC_TRACE();
/* set adapter started */
hw->adapter_stopped = 0;
if (dev->data->dev_conf.link_speeds & ETH_LINK_SPEED_FIXED) {
PMD_INIT_LOG(ERR,
"Invalid link_speeds for port %u, fix speed not supported",
dev->data->port_id);
return -EINVAL;
}
/* disable uio/vfio intr/eventfd mapping */
rte_intr_disable(intr_handle);
/* reinitialize adapter
* this calls reset and start
*/
status = atl_reset_hw(hw);
if (status != 0)
return -EIO;
err = hw_atl_b0_hw_init(hw, dev->data->mac_addrs->addr_bytes);
hw_atl_b0_hw_start(hw);
/* check and configure queue intr-vector mapping */
if ((rte_intr_cap_multiple(intr_handle) ||
!RTE_ETH_DEV_SRIOV(dev).active) &&
dev->data->dev_conf.intr_conf.rxq != 0) {
intr_vector = dev->data->nb_rx_queues;
if (intr_vector > ATL_MAX_INTR_QUEUE_NUM) {
PMD_INIT_LOG(ERR, "At most %d intr queues supported",
ATL_MAX_INTR_QUEUE_NUM);
return -ENOTSUP;
}
if (rte_intr_efd_enable(intr_handle, intr_vector)) {
PMD_INIT_LOG(ERR, "rte_intr_efd_enable failed");
return -1;
}
}
if (rte_intr_dp_is_en(intr_handle) && !intr_handle->intr_vec) {
intr_handle->intr_vec = rte_zmalloc("intr_vec",
dev->data->nb_rx_queues * sizeof(int), 0);
if (intr_handle->intr_vec == NULL) {
PMD_INIT_LOG(ERR, "Failed to allocate %d rx_queues"
" intr_vec", dev->data->nb_rx_queues);
return -ENOMEM;
}
}
/* initialize transmission unit */
atl_tx_init(dev);
/* This can fail when allocating mbufs for descriptor rings */
err = atl_rx_init(dev);
if (err) {
PMD_INIT_LOG(ERR, "Unable to initialize RX hardware");
goto error;
}
PMD_INIT_LOG(DEBUG, "FW version: %u.%u.%u",
hw->fw_ver_actual >> 24,
(hw->fw_ver_actual >> 16) & 0xFF,
hw->fw_ver_actual & 0xFFFF);
PMD_INIT_LOG(DEBUG, "Driver version: %s", ATL_PMD_DRIVER_VERSION);
err = atl_start_queues(dev);
if (err < 0) {
PMD_INIT_LOG(ERR, "Unable to start rxtx queues");
goto error;
}
err = atl_dev_set_link_up(dev);
err = hw->aq_fw_ops->update_link_status(hw);
if (err)
goto error;
dev->data->dev_link.link_status = hw->aq_link_status.mbps != 0;
if (rte_intr_allow_others(intr_handle)) {
/* check if lsc interrupt is enabled */
if (dev->data->dev_conf.intr_conf.lsc != 0)
atl_dev_lsc_interrupt_setup(dev, true);
else
atl_dev_lsc_interrupt_setup(dev, false);
} else {
rte_intr_callback_unregister(intr_handle,
atl_dev_interrupt_handler, dev);
if (dev->data->dev_conf.intr_conf.lsc != 0)
PMD_INIT_LOG(INFO, "lsc won't enable because of"
" no intr multiplex");
}
/* check if rxq interrupt is enabled */
if (dev->data->dev_conf.intr_conf.rxq != 0 &&
rte_intr_dp_is_en(intr_handle))
atl_dev_rxq_interrupt_setup(dev);
/* enable uio/vfio intr/eventfd mapping */
rte_intr_enable(intr_handle);
/* resume enabled intr since hw reset */
atl_enable_intr(dev);
return 0;
error:
atl_stop_queues(dev);
return -EIO;
}
/*
* Stop device: disable rx and tx functions to allow for reconfiguring.
*/
static void
atl_dev_stop(struct rte_eth_dev *dev)
{
struct rte_eth_link link;
struct aq_hw_s *hw =
ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
struct rte_pci_device *pci_dev = RTE_ETH_DEV_TO_PCI(dev);
struct rte_intr_handle *intr_handle = &pci_dev->intr_handle;
PMD_INIT_FUNC_TRACE();
/* disable interrupts */
atl_disable_intr(hw);
/* reset the NIC */
atl_reset_hw(hw);
hw->adapter_stopped = 1;
atl_stop_queues(dev);
/* Clear stored conf */
dev->data->scattered_rx = 0;
dev->data->lro = 0;
/* Clear recorded link status */
memset(&link, 0, sizeof(link));
rte_eth_linkstatus_set(dev, &link);
if (!rte_intr_allow_others(intr_handle))
/* resume to the default handler */
rte_intr_callback_register(intr_handle,
atl_dev_interrupt_handler,
(void *)dev);
/* Clean datapath event and queue/vec mapping */
rte_intr_efd_disable(intr_handle);
if (intr_handle->intr_vec != NULL) {
rte_free(intr_handle->intr_vec);
intr_handle->intr_vec = NULL;
}
}
/*
* Set device link up: enable tx.
*/
static int
atl_dev_set_link_up(struct rte_eth_dev *dev)
{
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
uint32_t link_speeds = dev->data->dev_conf.link_speeds;
uint32_t speed_mask = 0;
if (link_speeds == ETH_LINK_SPEED_AUTONEG) {
speed_mask = hw->aq_nic_cfg->link_speed_msk;
} else {
if (link_speeds & ETH_LINK_SPEED_10G)
speed_mask |= AQ_NIC_RATE_10G;
if (link_speeds & ETH_LINK_SPEED_5G)
speed_mask |= AQ_NIC_RATE_5G;
if (link_speeds & ETH_LINK_SPEED_1G)
speed_mask |= AQ_NIC_RATE_1G;
if (link_speeds & ETH_LINK_SPEED_2_5G)
speed_mask |= AQ_NIC_RATE_2G5;
if (link_speeds & ETH_LINK_SPEED_100M)
speed_mask |= AQ_NIC_RATE_100M;
}
return hw->aq_fw_ops->set_link_speed(hw, speed_mask);
}
/*
* Set device link down: disable tx.
*/
static int
atl_dev_set_link_down(struct rte_eth_dev *dev)
{
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
return hw->aq_fw_ops->set_link_speed(hw, 0);
}
/*
* Reset and stop device.
*/
static void
atl_dev_close(struct rte_eth_dev *dev)
{
PMD_INIT_FUNC_TRACE();
atl_dev_stop(dev);
atl_free_queues(dev);
}
static int
atl_dev_reset(struct rte_eth_dev *dev)
{
int ret;
ret = eth_atl_dev_uninit(dev);
if (ret)
return ret;
ret = eth_atl_dev_init(dev);
return ret;
}
static int
atl_dev_configure_macsec(struct rte_eth_dev *dev)
{
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
struct aq_hw_cfg_s *cf = ATL_DEV_PRIVATE_TO_CFG(dev->data->dev_private);
struct aq_macsec_config *aqcfg = &cf->aq_macsec;
struct macsec_msg_fw_request msg_macsec;
struct macsec_msg_fw_response response;
if (!aqcfg->common.macsec_enabled ||
hw->aq_fw_ops->send_macsec_req == NULL)
return 0;
memset(&msg_macsec, 0, sizeof(msg_macsec));
/* Creating set of sc/sa structures from parameters provided by DPDK */
/* Configure macsec */
msg_macsec.msg_type = macsec_cfg_msg;
msg_macsec.cfg.enabled = aqcfg->common.macsec_enabled;
msg_macsec.cfg.interrupts_enabled = 1;
hw->aq_fw_ops->send_macsec_req(hw, &msg_macsec, &response);
if (response.result)
return -1;
memset(&msg_macsec, 0, sizeof(msg_macsec));
/* Configure TX SC */
msg_macsec.msg_type = macsec_add_tx_sc_msg;
msg_macsec.txsc.index = 0; /* TXSC always one (??) */
msg_macsec.txsc.protect = aqcfg->common.encryption_enabled;
/* MAC addr for TX */
msg_macsec.txsc.mac_sa[0] = rte_bswap32(aqcfg->txsc.mac[1]);
msg_macsec.txsc.mac_sa[1] = rte_bswap32(aqcfg->txsc.mac[0]);
msg_macsec.txsc.sa_mask = 0x3f;
msg_macsec.txsc.da_mask = 0;
msg_macsec.txsc.tci = 0x0B;
msg_macsec.txsc.curr_an = 0; /* SA index which currently used */
/*
* Creating SCI (Secure Channel Identifier).
* SCI constructed from Source MAC and Port identifier
*/
uint32_t sci_hi_part = (msg_macsec.txsc.mac_sa[1] << 16) |
(msg_macsec.txsc.mac_sa[0] >> 16);
uint32_t sci_low_part = (msg_macsec.txsc.mac_sa[0] << 16);
uint32_t port_identifier = 1;
msg_macsec.txsc.sci[1] = sci_hi_part;
msg_macsec.txsc.sci[0] = sci_low_part | port_identifier;
hw->aq_fw_ops->send_macsec_req(hw, &msg_macsec, &response);
if (response.result)
return -1;
memset(&msg_macsec, 0, sizeof(msg_macsec));
/* Configure RX SC */
msg_macsec.msg_type = macsec_add_rx_sc_msg;
msg_macsec.rxsc.index = aqcfg->rxsc.pi;
msg_macsec.rxsc.replay_protect =
aqcfg->common.replay_protection_enabled;
msg_macsec.rxsc.anti_replay_window = 0;
/* MAC addr for RX */
msg_macsec.rxsc.mac_da[0] = rte_bswap32(aqcfg->rxsc.mac[1]);
msg_macsec.rxsc.mac_da[1] = rte_bswap32(aqcfg->rxsc.mac[0]);
msg_macsec.rxsc.da_mask = 0;//0x3f;
msg_macsec.rxsc.sa_mask = 0;
hw->aq_fw_ops->send_macsec_req(hw, &msg_macsec, &response);
if (response.result)
return -1;
memset(&msg_macsec, 0, sizeof(msg_macsec));
/* Configure RX SC */
msg_macsec.msg_type = macsec_add_tx_sa_msg;
msg_macsec.txsa.index = aqcfg->txsa.idx;
msg_macsec.txsa.next_pn = aqcfg->txsa.pn;
msg_macsec.txsa.key[0] = rte_bswap32(aqcfg->txsa.key[3]);
msg_macsec.txsa.key[1] = rte_bswap32(aqcfg->txsa.key[2]);
msg_macsec.txsa.key[2] = rte_bswap32(aqcfg->txsa.key[1]);
msg_macsec.txsa.key[3] = rte_bswap32(aqcfg->txsa.key[0]);
hw->aq_fw_ops->send_macsec_req(hw, &msg_macsec, &response);
if (response.result)
return -1;
memset(&msg_macsec, 0, sizeof(msg_macsec));
/* Configure RX SA */
msg_macsec.msg_type = macsec_add_rx_sa_msg;
msg_macsec.rxsa.index = aqcfg->rxsa.idx;
msg_macsec.rxsa.next_pn = aqcfg->rxsa.pn;
msg_macsec.rxsa.key[0] = rte_bswap32(aqcfg->rxsa.key[3]);
msg_macsec.rxsa.key[1] = rte_bswap32(aqcfg->rxsa.key[2]);
msg_macsec.rxsa.key[2] = rte_bswap32(aqcfg->rxsa.key[1]);
msg_macsec.rxsa.key[3] = rte_bswap32(aqcfg->rxsa.key[0]);
hw->aq_fw_ops->send_macsec_req(hw, &msg_macsec, &response);
if (response.result)
return -1;
return 0;
}
int atl_macsec_enable(struct rte_eth_dev *dev,
uint8_t encr, uint8_t repl_prot)
{
struct aq_hw_cfg_s *cfg =
ATL_DEV_PRIVATE_TO_CFG(dev->data->dev_private);
cfg->aq_macsec.common.macsec_enabled = 1;
cfg->aq_macsec.common.encryption_enabled = encr;
cfg->aq_macsec.common.replay_protection_enabled = repl_prot;
return 0;
}
int atl_macsec_disable(struct rte_eth_dev *dev)
{
struct aq_hw_cfg_s *cfg =
ATL_DEV_PRIVATE_TO_CFG(dev->data->dev_private);
cfg->aq_macsec.common.macsec_enabled = 0;
return 0;
}
int atl_macsec_config_txsc(struct rte_eth_dev *dev, uint8_t *mac)
{
struct aq_hw_cfg_s *cfg =
ATL_DEV_PRIVATE_TO_CFG(dev->data->dev_private);
memset(&cfg->aq_macsec.txsc.mac, 0, sizeof(cfg->aq_macsec.txsc.mac));
memcpy((uint8_t *)&cfg->aq_macsec.txsc.mac + 2, mac,
RTE_ETHER_ADDR_LEN);
return 0;
}
int atl_macsec_config_rxsc(struct rte_eth_dev *dev,
uint8_t *mac, uint16_t pi)
{
struct aq_hw_cfg_s *cfg =
ATL_DEV_PRIVATE_TO_CFG(dev->data->dev_private);
memset(&cfg->aq_macsec.rxsc.mac, 0, sizeof(cfg->aq_macsec.rxsc.mac));
memcpy((uint8_t *)&cfg->aq_macsec.rxsc.mac + 2, mac,
RTE_ETHER_ADDR_LEN);
cfg->aq_macsec.rxsc.pi = pi;
return 0;
}
int atl_macsec_select_txsa(struct rte_eth_dev *dev,
uint8_t idx, uint8_t an,
uint32_t pn, uint8_t *key)
{
struct aq_hw_cfg_s *cfg =
ATL_DEV_PRIVATE_TO_CFG(dev->data->dev_private);
cfg->aq_macsec.txsa.idx = idx;
cfg->aq_macsec.txsa.pn = pn;
cfg->aq_macsec.txsa.an = an;
memcpy(&cfg->aq_macsec.txsa.key, key, 16);
return 0;
}
int atl_macsec_select_rxsa(struct rte_eth_dev *dev,
uint8_t idx, uint8_t an,
uint32_t pn, uint8_t *key)
{
struct aq_hw_cfg_s *cfg =
ATL_DEV_PRIVATE_TO_CFG(dev->data->dev_private);
cfg->aq_macsec.rxsa.idx = idx;
cfg->aq_macsec.rxsa.pn = pn;
cfg->aq_macsec.rxsa.an = an;
memcpy(&cfg->aq_macsec.rxsa.key, key, 16);
return 0;
}
static int
atl_dev_stats_get(struct rte_eth_dev *dev, struct rte_eth_stats *stats)
{
struct atl_adapter *adapter = ATL_DEV_TO_ADAPTER(dev);
struct aq_hw_s *hw = &adapter->hw;
struct atl_sw_stats *swstats = &adapter->sw_stats;
unsigned int i;
hw->aq_fw_ops->update_stats(hw);
/* Fill out the rte_eth_stats statistics structure */
stats->ipackets = hw->curr_stats.dma_pkt_rc;
stats->ibytes = hw->curr_stats.dma_oct_rc;
stats->imissed = hw->curr_stats.dpc;
stats->ierrors = hw->curr_stats.erpt;
stats->opackets = hw->curr_stats.dma_pkt_tc;
stats->obytes = hw->curr_stats.dma_oct_tc;
stats->oerrors = 0;
stats->rx_nombuf = swstats->rx_nombuf;
for (i = 0; i < RTE_ETHDEV_QUEUE_STAT_CNTRS; i++) {
stats->q_ipackets[i] = swstats->q_ipackets[i];
stats->q_opackets[i] = swstats->q_opackets[i];
stats->q_ibytes[i] = swstats->q_ibytes[i];
stats->q_obytes[i] = swstats->q_obytes[i];
stats->q_errors[i] = swstats->q_errors[i];
}
return 0;
}
static int
atl_dev_stats_reset(struct rte_eth_dev *dev)
{
struct atl_adapter *adapter = ATL_DEV_TO_ADAPTER(dev);
struct aq_hw_s *hw = &adapter->hw;
hw->aq_fw_ops->update_stats(hw);
/* Reset software totals */
memset(&hw->curr_stats, 0, sizeof(hw->curr_stats));
memset(&adapter->sw_stats, 0, sizeof(adapter->sw_stats));
return 0;
}
static int
atl_dev_xstats_get_count(struct rte_eth_dev *dev)
{
struct atl_adapter *adapter =
(struct atl_adapter *)dev->data->dev_private;
struct aq_hw_s *hw = &adapter->hw;
unsigned int i, count = 0;
for (i = 0; i < RTE_DIM(atl_xstats_tbl); i++) {
if (atl_xstats_tbl[i].type == XSTATS_TYPE_MACSEC &&
((hw->caps_lo & BIT(CAPS_LO_MACSEC)) == 0))
continue;
count++;
}
return count;
}
static int
atl_dev_xstats_get_names(struct rte_eth_dev *dev __rte_unused,
struct rte_eth_xstat_name *xstats_names,
unsigned int size)
{
unsigned int i;
unsigned int count = atl_dev_xstats_get_count(dev);
if (xstats_names) {
for (i = 0; i < size && i < count; i++) {
snprintf(xstats_names[i].name,
RTE_ETH_XSTATS_NAME_SIZE, "%s",
atl_xstats_tbl[i].name);
}
}
return count;
}
static int
atl_dev_xstats_get(struct rte_eth_dev *dev, struct rte_eth_xstat *stats,
unsigned int n)
{
struct atl_adapter *adapter = dev->data->dev_private;
struct aq_hw_s *hw = &adapter->hw;
struct get_stats req = { 0 };
struct macsec_msg_fw_request msg = { 0 };
struct macsec_msg_fw_response resp = { 0 };
int err = -1;
unsigned int i;
unsigned int count = atl_dev_xstats_get_count(dev);
if (!stats)
return count;
if (hw->aq_fw_ops->send_macsec_req != NULL) {
req.ingress_sa_index = 0xff;
req.egress_sc_index = 0xff;
req.egress_sa_index = 0xff;
msg.msg_type = macsec_get_stats_msg;
msg.stats = req;
err = hw->aq_fw_ops->send_macsec_req(hw, &msg, &resp);
}
for (i = 0; i < n && i < count; i++) {
stats[i].id = i;
switch (atl_xstats_tbl[i].type) {
case XSTATS_TYPE_MSM:
stats[i].value = *(u64 *)((uint8_t *)&hw->curr_stats +
atl_xstats_tbl[i].offset);
break;
case XSTATS_TYPE_MACSEC:
if (!err) {
stats[i].value =
*(u64 *)((uint8_t *)&resp.stats +
atl_xstats_tbl[i].offset);
}
break;
}
}
return i;
}
static int
atl_fw_version_get(struct rte_eth_dev *dev, char *fw_version, size_t fw_size)
{
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
uint32_t fw_ver = 0;
unsigned int ret = 0;
ret = hw_atl_utils_get_fw_version(hw, &fw_ver);
if (ret)
return -EIO;
ret = snprintf(fw_version, fw_size, "%u.%u.%u", fw_ver >> 24,
(fw_ver >> 16) & 0xFFU, fw_ver & 0xFFFFU);
ret += 1; /* add string null-terminator */
if (fw_size < ret)
return ret;
return 0;
}
static int
atl_dev_info_get(struct rte_eth_dev *dev, struct rte_eth_dev_info *dev_info)
{
struct rte_pci_device *pci_dev = RTE_ETH_DEV_TO_PCI(dev);
dev_info->max_rx_queues = AQ_HW_MAX_RX_QUEUES;
dev_info->max_tx_queues = AQ_HW_MAX_TX_QUEUES;
dev_info->min_rx_bufsize = 1024;
dev_info->max_rx_pktlen = HW_ATL_B0_MTU_JUMBO;
dev_info->max_mac_addrs = HW_ATL_B0_MAC_MAX;
dev_info->max_vfs = pci_dev->max_vfs;
dev_info->max_hash_mac_addrs = 0;
dev_info->max_vmdq_pools = 0;
dev_info->vmdq_queue_num = 0;
dev_info->rx_offload_capa = ATL_RX_OFFLOADS;
dev_info->tx_offload_capa = ATL_TX_OFFLOADS;
dev_info->default_rxconf = (struct rte_eth_rxconf) {
.rx_free_thresh = ATL_DEFAULT_RX_FREE_THRESH,
};
dev_info->default_txconf = (struct rte_eth_txconf) {
.tx_free_thresh = ATL_DEFAULT_TX_FREE_THRESH,
};
dev_info->rx_desc_lim = rx_desc_lim;
dev_info->tx_desc_lim = tx_desc_lim;
dev_info->hash_key_size = HW_ATL_B0_RSS_HASHKEY_BITS / 8;
dev_info->reta_size = HW_ATL_B0_RSS_REDIRECTION_MAX;
dev_info->flow_type_rss_offloads = ATL_RSS_OFFLOAD_ALL;
dev_info->speed_capa = ETH_LINK_SPEED_1G | ETH_LINK_SPEED_10G;
dev_info->speed_capa |= ETH_LINK_SPEED_100M;
dev_info->speed_capa |= ETH_LINK_SPEED_2_5G;
dev_info->speed_capa |= ETH_LINK_SPEED_5G;
return 0;
}
static const uint32_t *
atl_dev_supported_ptypes_get(struct rte_eth_dev *dev)
{
static const uint32_t ptypes[] = {
RTE_PTYPE_L2_ETHER,
RTE_PTYPE_L2_ETHER_ARP,
RTE_PTYPE_L2_ETHER_VLAN,
RTE_PTYPE_L3_IPV4,
RTE_PTYPE_L3_IPV6,
RTE_PTYPE_L4_TCP,
RTE_PTYPE_L4_UDP,
RTE_PTYPE_L4_SCTP,
RTE_PTYPE_L4_ICMP,
RTE_PTYPE_UNKNOWN
};
if (dev->rx_pkt_burst == atl_recv_pkts)
return ptypes;
return NULL;
}
static void
atl_dev_delayed_handler(void *param)
{
struct rte_eth_dev *dev = (struct rte_eth_dev *)param;
atl_dev_configure_macsec(dev);
}
/* return 0 means link status changed, -1 means not changed */
static int
atl_dev_link_update(struct rte_eth_dev *dev, int wait __rte_unused)
{
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
struct rte_eth_link link, old;
u32 fc = AQ_NIC_FC_OFF;
int err = 0;
link.link_status = ETH_LINK_DOWN;
link.link_speed = 0;
link.link_duplex = ETH_LINK_FULL_DUPLEX;
link.link_autoneg = hw->is_autoneg ? ETH_LINK_AUTONEG : ETH_LINK_FIXED;
memset(&old, 0, sizeof(old));
/* load old link status */
rte_eth_linkstatus_get(dev, &old);
/* read current link status */
err = hw->aq_fw_ops->update_link_status(hw);
if (err)
return 0;
if (hw->aq_link_status.mbps == 0) {
/* write default (down) link status */
rte_eth_linkstatus_set(dev, &link);
if (link.link_status == old.link_status)
return -1;
return 0;
}
link.link_status = ETH_LINK_UP;
link.link_duplex = ETH_LINK_FULL_DUPLEX;
link.link_speed = hw->aq_link_status.mbps;
rte_eth_linkstatus_set(dev, &link);
if (link.link_status == old.link_status)
return -1;
/* Driver has to update flow control settings on RX block
* on any link event.
* We should query FW whether it negotiated FC.
*/
if (hw->aq_fw_ops->get_flow_control) {
hw->aq_fw_ops->get_flow_control(hw, &fc);
hw_atl_b0_set_fc(hw, fc, 0U);
}
if (rte_eal_alarm_set(1000 * 1000,
atl_dev_delayed_handler, (void *)dev) < 0)
PMD_DRV_LOG(ERR, "rte_eal_alarm_set fail");
return 0;
}
static int
atl_dev_promiscuous_enable(struct rte_eth_dev *dev)
{
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
hw_atl_rpfl2promiscuous_mode_en_set(hw, true);
return 0;
}
static int
atl_dev_promiscuous_disable(struct rte_eth_dev *dev)
{
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
hw_atl_rpfl2promiscuous_mode_en_set(hw, false);
return 0;
}
static int
atl_dev_allmulticast_enable(struct rte_eth_dev *dev)
{
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
hw_atl_rpfl2_accept_all_mc_packets_set(hw, true);
return 0;
}
static int
atl_dev_allmulticast_disable(struct rte_eth_dev *dev)
{
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
if (dev->data->promiscuous == 1)
return 0; /* must remain in all_multicast mode */
hw_atl_rpfl2_accept_all_mc_packets_set(hw, false);
return 0;
}
/**
* It clears the interrupt causes and enables the interrupt.
* It will be called once only during nic initialized.
*
* @param dev
* Pointer to struct rte_eth_dev.
* @param on
* Enable or Disable.
*
* @return
* - On success, zero.
* - On failure, a negative value.
*/
static int
atl_dev_lsc_interrupt_setup(struct rte_eth_dev *dev, uint8_t on __rte_unused)
{
atl_dev_link_status_print(dev);
return 0;
}
static int
atl_dev_rxq_interrupt_setup(struct rte_eth_dev *dev __rte_unused)
{
return 0;
}
static int
atl_dev_interrupt_get_status(struct rte_eth_dev *dev)
{
struct atl_interrupt *intr =
ATL_DEV_PRIVATE_TO_INTR(dev->data->dev_private);
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
u64 cause = 0;
hw_atl_b0_hw_irq_read(hw, &cause);
atl_disable_intr(hw);
if (cause & BIT(ATL_IRQ_CAUSE_LINK))
intr->flags |= ATL_FLAG_NEED_LINK_UPDATE;
return 0;
}
/**
* It gets and then prints the link status.
*
* @param dev
* Pointer to struct rte_eth_dev.
*
* @return
* - On success, zero.
* - On failure, a negative value.
*/
static void
atl_dev_link_status_print(struct rte_eth_dev *dev)
{
struct rte_eth_link link;
memset(&link, 0, sizeof(link));
rte_eth_linkstatus_get(dev, &link);
if (link.link_status) {
PMD_DRV_LOG(INFO, "Port %d: Link Up - speed %u Mbps - %s",
(int)(dev->data->port_id),
(unsigned int)link.link_speed,
link.link_duplex == ETH_LINK_FULL_DUPLEX ?
"full-duplex" : "half-duplex");
} else {
PMD_DRV_LOG(INFO, " Port %d: Link Down",
(int)(dev->data->port_id));
}
#ifdef DEBUG
{
struct rte_pci_device *pci_dev = RTE_ETH_DEV_TO_PCI(dev);
PMD_DRV_LOG(DEBUG, "PCI Address: " PCI_PRI_FMT,
pci_dev->addr.domain,
pci_dev->addr.bus,
pci_dev->addr.devid,
pci_dev->addr.function);
}
#endif
PMD_DRV_LOG(INFO, "Link speed:%d", link.link_speed);
}
/*
* It executes link_update after knowing an interrupt occurred.
*
* @param dev
* Pointer to struct rte_eth_dev.
*
* @return
* - On success, zero.
* - On failure, a negative value.
*/
static int
atl_dev_interrupt_action(struct rte_eth_dev *dev,
struct rte_intr_handle *intr_handle)
{
struct atl_interrupt *intr =
ATL_DEV_PRIVATE_TO_INTR(dev->data->dev_private);
struct atl_adapter *adapter = dev->data->dev_private;
struct aq_hw_s *hw = &adapter->hw;
if (!(intr->flags & ATL_FLAG_NEED_LINK_UPDATE))
goto done;
intr->flags &= ~ATL_FLAG_NEED_LINK_UPDATE;
/* Notify userapp if link status changed */
if (!atl_dev_link_update(dev, 0)) {
atl_dev_link_status_print(dev);
_rte_eth_dev_callback_process(dev,
RTE_ETH_EVENT_INTR_LSC, NULL);
} else {
if (hw->aq_fw_ops->send_macsec_req == NULL)
goto done;
/* Check macsec Keys expired */
struct get_stats req = { 0 };
struct macsec_msg_fw_request msg = { 0 };
struct macsec_msg_fw_response resp = { 0 };
req.ingress_sa_index = 0x0;
req.egress_sc_index = 0x0;
req.egress_sa_index = 0x0;
msg.msg_type = macsec_get_stats_msg;
msg.stats = req;
int err = hw->aq_fw_ops->send_macsec_req(hw, &msg, &resp);
if (err) {
PMD_DRV_LOG(ERR, "send_macsec_req fail");
goto done;
}
if (resp.stats.egress_threshold_expired ||
resp.stats.ingress_threshold_expired ||
resp.stats.egress_expired ||
resp.stats.ingress_expired) {
PMD_DRV_LOG(INFO, "RTE_ETH_EVENT_MACSEC");
_rte_eth_dev_callback_process(dev,
RTE_ETH_EVENT_MACSEC, NULL);
}
}
done:
atl_enable_intr(dev);
rte_intr_ack(intr_handle);
return 0;
}
/**
* Interrupt handler triggered by NIC for handling
* specific interrupt.
*
* @param handle
* Pointer to interrupt handle.
* @param param
* The address of parameter (struct rte_eth_dev *) regsitered before.
*
* @return
* void
*/
static void
atl_dev_interrupt_handler(void *param)
{
struct rte_eth_dev *dev = (struct rte_eth_dev *)param;
atl_dev_interrupt_get_status(dev);
atl_dev_interrupt_action(dev, dev->intr_handle);
}
static int
atl_dev_get_eeprom_length(struct rte_eth_dev *dev __rte_unused)
{
return SFP_EEPROM_SIZE;
}
int atl_dev_get_eeprom(struct rte_eth_dev *dev,
struct rte_dev_eeprom_info *eeprom)
{
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
uint32_t dev_addr = SMBUS_DEVICE_ID;
if (hw->aq_fw_ops->get_eeprom == NULL)
return -ENOTSUP;
if (eeprom->length + eeprom->offset > SFP_EEPROM_SIZE ||
eeprom->data == NULL)
return -EINVAL;
if (eeprom->magic > 0x7F)
return -EINVAL;
if (eeprom->magic)
dev_addr = eeprom->magic;
return hw->aq_fw_ops->get_eeprom(hw, dev_addr, eeprom->data,
eeprom->length, eeprom->offset);
}
int atl_dev_set_eeprom(struct rte_eth_dev *dev,
struct rte_dev_eeprom_info *eeprom)
{
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
uint32_t dev_addr = SMBUS_DEVICE_ID;
if (hw->aq_fw_ops->set_eeprom == NULL)
return -ENOTSUP;
if (eeprom->length + eeprom->offset > SFP_EEPROM_SIZE ||
eeprom->data == NULL)
return -EINVAL;
if (eeprom->magic > 0x7F)
return -EINVAL;
if (eeprom->magic)
dev_addr = eeprom->magic;
return hw->aq_fw_ops->set_eeprom(hw, dev_addr, eeprom->data,
eeprom->length, eeprom->offset);
}
static int
atl_dev_get_regs(struct rte_eth_dev *dev, struct rte_dev_reg_info *regs)
{
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
u32 mif_id;
int err;
if (regs->data == NULL) {
regs->length = hw_atl_utils_hw_get_reg_length();
regs->width = sizeof(u32);
return 0;
}
/* Only full register dump is supported */
if (regs->length && regs->length != hw_atl_utils_hw_get_reg_length())
return -ENOTSUP;
err = hw_atl_utils_hw_get_regs(hw, regs->data);
/* Device version */
mif_id = hw_atl_reg_glb_mif_id_get(hw);
regs->version = mif_id & 0xFFU;
return err;
}
static int
atl_flow_ctrl_get(struct rte_eth_dev *dev, struct rte_eth_fc_conf *fc_conf)
{
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
u32 fc = AQ_NIC_FC_OFF;
if (hw->aq_fw_ops->get_flow_control == NULL)
return -ENOTSUP;
hw->aq_fw_ops->get_flow_control(hw, &fc);
if (fc == AQ_NIC_FC_OFF)
fc_conf->mode = RTE_FC_NONE;
else if ((fc & AQ_NIC_FC_RX) && (fc & AQ_NIC_FC_TX))
fc_conf->mode = RTE_FC_FULL;
else if (fc & AQ_NIC_FC_RX)
fc_conf->mode = RTE_FC_RX_PAUSE;
else if (fc & AQ_NIC_FC_TX)
fc_conf->mode = RTE_FC_TX_PAUSE;
return 0;
}
static int
atl_flow_ctrl_set(struct rte_eth_dev *dev, struct rte_eth_fc_conf *fc_conf)
{
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
uint32_t old_flow_control = hw->aq_nic_cfg->flow_control;
if (hw->aq_fw_ops->set_flow_control == NULL)
return -ENOTSUP;
if (fc_conf->mode == RTE_FC_NONE)
hw->aq_nic_cfg->flow_control = AQ_NIC_FC_OFF;
else if (fc_conf->mode == RTE_FC_RX_PAUSE)
hw->aq_nic_cfg->flow_control = AQ_NIC_FC_RX;
else if (fc_conf->mode == RTE_FC_TX_PAUSE)
hw->aq_nic_cfg->flow_control = AQ_NIC_FC_TX;
else if (fc_conf->mode == RTE_FC_FULL)
hw->aq_nic_cfg->flow_control = (AQ_NIC_FC_RX | AQ_NIC_FC_TX);
if (old_flow_control != hw->aq_nic_cfg->flow_control)
return hw->aq_fw_ops->set_flow_control(hw);
return 0;
}
static int
atl_update_mac_addr(struct rte_eth_dev *dev, uint32_t index,
u8 *mac_addr, bool enable)
{
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
unsigned int h = 0U;
unsigned int l = 0U;
int err;
if (mac_addr) {
h = (mac_addr[0] << 8) | (mac_addr[1]);
l = (mac_addr[2] << 24) | (mac_addr[3] << 16) |
(mac_addr[4] << 8) | mac_addr[5];
}
hw_atl_rpfl2_uc_flr_en_set(hw, 0U, index);
hw_atl_rpfl2unicast_dest_addresslsw_set(hw, l, index);
hw_atl_rpfl2unicast_dest_addressmsw_set(hw, h, index);
if (enable)
hw_atl_rpfl2_uc_flr_en_set(hw, 1U, index);
err = aq_hw_err_from_flags(hw);
return err;
}
static int
atl_add_mac_addr(struct rte_eth_dev *dev, struct rte_ether_addr *mac_addr,
uint32_t index __rte_unused, uint32_t pool __rte_unused)
{
if (rte_is_zero_ether_addr(mac_addr)) {
PMD_DRV_LOG(ERR, "Invalid Ethernet Address");
return -EINVAL;
}
return atl_update_mac_addr(dev, index, (u8 *)mac_addr, true);
}
static void
atl_remove_mac_addr(struct rte_eth_dev *dev, uint32_t index)
{
atl_update_mac_addr(dev, index, NULL, false);
}
static int
atl_set_default_mac_addr(struct rte_eth_dev *dev, struct rte_ether_addr *addr)
{
atl_remove_mac_addr(dev, 0);
atl_add_mac_addr(dev, addr, 0, 0);
return 0;
}
static int
atl_dev_mtu_set(struct rte_eth_dev *dev, uint16_t mtu)
{
struct rte_eth_dev_info dev_info;
int ret;
uint32_t frame_size = mtu + RTE_ETHER_HDR_LEN + RTE_ETHER_CRC_LEN;
ret = atl_dev_info_get(dev, &dev_info);
if (ret != 0)
return ret;
if (mtu < RTE_ETHER_MIN_MTU || frame_size > dev_info.max_rx_pktlen)
return -EINVAL;
/* update max frame size */
dev->data->dev_conf.rxmode.max_rx_pkt_len = frame_size;
return 0;
}
static int
atl_vlan_filter_set(struct rte_eth_dev *dev, uint16_t vlan_id, int on)
{
struct aq_hw_cfg_s *cfg =
ATL_DEV_PRIVATE_TO_CFG(dev->data->dev_private);
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
int err = 0;
int i = 0;
PMD_INIT_FUNC_TRACE();
for (i = 0; i < HW_ATL_B0_MAX_VLAN_IDS; i++) {
if (cfg->vlan_filter[i] == vlan_id) {
if (!on) {
/* Disable VLAN filter. */
hw_atl_rpf_vlan_flr_en_set(hw, 0U, i);
/* Clear VLAN filter entry */
cfg->vlan_filter[i] = 0;
}
break;
}
}
/* VLAN_ID was not found. So, nothing to delete. */
if (i == HW_ATL_B0_MAX_VLAN_IDS && !on)
goto exit;
/* VLAN_ID already exist, or already removed above. Nothing to do. */
if (i != HW_ATL_B0_MAX_VLAN_IDS)
goto exit;
/* Try to found free VLAN filter to add new VLAN_ID */
for (i = 0; i < HW_ATL_B0_MAX_VLAN_IDS; i++) {
if (cfg->vlan_filter[i] == 0)
break;
}
if (i == HW_ATL_B0_MAX_VLAN_IDS) {
/* We have no free VLAN filter to add new VLAN_ID*/
err = -ENOMEM;
goto exit;
}
cfg->vlan_filter[i] = vlan_id;
hw_atl_rpf_vlan_flr_act_set(hw, 1U, i);
hw_atl_rpf_vlan_id_flr_set(hw, vlan_id, i);
hw_atl_rpf_vlan_flr_en_set(hw, 1U, i);
exit:
/* Enable VLAN promisc mode if vlan_filter empty */
for (i = 0; i < HW_ATL_B0_MAX_VLAN_IDS; i++) {
if (cfg->vlan_filter[i] != 0)
break;
}
hw_atl_rpf_vlan_prom_mode_en_set(hw, i == HW_ATL_B0_MAX_VLAN_IDS);
return err;
}
static int
atl_enable_vlan_filter(struct rte_eth_dev *dev, int en)
{
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
struct aq_hw_cfg_s *cfg =
ATL_DEV_PRIVATE_TO_CFG(dev->data->dev_private);
int i;
PMD_INIT_FUNC_TRACE();
for (i = 0; i < HW_ATL_B0_MAX_VLAN_IDS; i++) {
if (cfg->vlan_filter[i])
hw_atl_rpf_vlan_flr_en_set(hw, en, i);
}
return 0;
}
static int
atl_vlan_offload_set(struct rte_eth_dev *dev, int mask)
{
struct aq_hw_cfg_s *cfg =
ATL_DEV_PRIVATE_TO_CFG(dev->data->dev_private);
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
int ret = 0;
int i;
PMD_INIT_FUNC_TRACE();
ret = atl_enable_vlan_filter(dev, mask & ETH_VLAN_FILTER_MASK);
cfg->vlan_strip = !!(mask & ETH_VLAN_STRIP_MASK);
for (i = 0; i < dev->data->nb_rx_queues; i++)
hw_atl_rpo_rx_desc_vlan_stripping_set(hw, cfg->vlan_strip, i);
if (mask & ETH_VLAN_EXTEND_MASK)
ret = -ENOTSUP;
return ret;
}
static int
atl_vlan_tpid_set(struct rte_eth_dev *dev, enum rte_vlan_type vlan_type,
uint16_t tpid)
{
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
int err = 0;
PMD_INIT_FUNC_TRACE();
switch (vlan_type) {
case ETH_VLAN_TYPE_INNER:
hw_atl_rpf_vlan_inner_etht_set(hw, tpid);
break;
case ETH_VLAN_TYPE_OUTER:
hw_atl_rpf_vlan_outer_etht_set(hw, tpid);
break;
default:
PMD_DRV_LOG(ERR, "Unsupported VLAN type");
err = -ENOTSUP;
}
return err;
}
static void
atl_vlan_strip_queue_set(struct rte_eth_dev *dev, uint16_t queue_id, int on)
{
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
PMD_INIT_FUNC_TRACE();
if (queue_id > dev->data->nb_rx_queues) {
PMD_DRV_LOG(ERR, "Invalid queue id");
return;
}
hw_atl_rpo_rx_desc_vlan_stripping_set(hw, on, queue_id);
}
static int
atl_dev_set_mc_addr_list(struct rte_eth_dev *dev,
struct rte_ether_addr *mc_addr_set,
uint32_t nb_mc_addr)
{
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
u32 i;
if (nb_mc_addr > AQ_HW_MULTICAST_ADDRESS_MAX - HW_ATL_B0_MAC_MIN)
return -EINVAL;
/* Update whole uc filters table */
for (i = 0; i < AQ_HW_MULTICAST_ADDRESS_MAX - HW_ATL_B0_MAC_MIN; i++) {
u8 *mac_addr = NULL;
u32 l = 0, h = 0;
if (i < nb_mc_addr) {
mac_addr = mc_addr_set[i].addr_bytes;
l = (mac_addr[2] << 24) | (mac_addr[3] << 16) |
(mac_addr[4] << 8) | mac_addr[5];
h = (mac_addr[0] << 8) | mac_addr[1];
}
hw_atl_rpfl2_uc_flr_en_set(hw, 0U, HW_ATL_B0_MAC_MIN + i);
hw_atl_rpfl2unicast_dest_addresslsw_set(hw, l,
HW_ATL_B0_MAC_MIN + i);
hw_atl_rpfl2unicast_dest_addressmsw_set(hw, h,
HW_ATL_B0_MAC_MIN + i);
hw_atl_rpfl2_uc_flr_en_set(hw, !!mac_addr,
HW_ATL_B0_MAC_MIN + i);
}
return 0;
}
static int
atl_reta_update(struct rte_eth_dev *dev,
struct rte_eth_rss_reta_entry64 *reta_conf,
uint16_t reta_size)
{
int i;
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
struct aq_hw_cfg_s *cf = ATL_DEV_PRIVATE_TO_CFG(dev->data->dev_private);
for (i = 0; i < reta_size && i < cf->aq_rss.indirection_table_size; i++)
cf->aq_rss.indirection_table[i] = min(reta_conf->reta[i],
dev->data->nb_rx_queues - 1);
hw_atl_b0_hw_rss_set(hw, &cf->aq_rss);
return 0;
}
static int
atl_reta_query(struct rte_eth_dev *dev,
struct rte_eth_rss_reta_entry64 *reta_conf,
uint16_t reta_size)
{
int i;
struct aq_hw_cfg_s *cf = ATL_DEV_PRIVATE_TO_CFG(dev->data->dev_private);
for (i = 0; i < reta_size && i < cf->aq_rss.indirection_table_size; i++)
reta_conf->reta[i] = cf->aq_rss.indirection_table[i];
reta_conf->mask = ~0U;
return 0;
}
static int
atl_rss_hash_update(struct rte_eth_dev *dev,
struct rte_eth_rss_conf *rss_conf)
{
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
struct aq_hw_cfg_s *cfg =
ATL_DEV_PRIVATE_TO_CFG(dev->data->dev_private);
static u8 def_rss_key[40] = {
0x1e, 0xad, 0x71, 0x87, 0x65, 0xfc, 0x26, 0x7d,
0x0d, 0x45, 0x67, 0x74, 0xcd, 0x06, 0x1a, 0x18,
0xb6, 0xc1, 0xf0, 0xc7, 0xbb, 0x18, 0xbe, 0xf8,
0x19, 0x13, 0x4b, 0xa9, 0xd0, 0x3e, 0xfe, 0x70,
0x25, 0x03, 0xab, 0x50, 0x6a, 0x8b, 0x82, 0x0c
};
cfg->is_rss = !!rss_conf->rss_hf;
if (rss_conf->rss_key) {
memcpy(cfg->aq_rss.hash_secret_key, rss_conf->rss_key,
rss_conf->rss_key_len);
cfg->aq_rss.hash_secret_key_size = rss_conf->rss_key_len;
} else {
memcpy(cfg->aq_rss.hash_secret_key, def_rss_key,
sizeof(def_rss_key));
cfg->aq_rss.hash_secret_key_size = sizeof(def_rss_key);
}
hw_atl_b0_hw_rss_set(hw, &cfg->aq_rss);
hw_atl_b0_hw_rss_hash_set(hw, &cfg->aq_rss);
return 0;
}
static int
atl_rss_hash_conf_get(struct rte_eth_dev *dev,
struct rte_eth_rss_conf *rss_conf)
{
struct aq_hw_cfg_s *cfg =
ATL_DEV_PRIVATE_TO_CFG(dev->data->dev_private);
rss_conf->rss_hf = cfg->is_rss ? ATL_RSS_OFFLOAD_ALL : 0;
if (rss_conf->rss_key) {
rss_conf->rss_key_len = cfg->aq_rss.hash_secret_key_size;
memcpy(rss_conf->rss_key, cfg->aq_rss.hash_secret_key,
rss_conf->rss_key_len);
}
return 0;
}
static bool
is_device_supported(struct rte_eth_dev *dev, struct rte_pci_driver *drv)
{
if (strcmp(dev->device->driver->name, drv->driver.name))
return false;
return true;
}
bool
is_atlantic_supported(struct rte_eth_dev *dev)
{
return is_device_supported(dev, &rte_atl_pmd);
}
RTE_PMD_REGISTER_PCI(net_atlantic, rte_atl_pmd);
RTE_PMD_REGISTER_PCI_TABLE(net_atlantic, pci_id_atl_map);
RTE_PMD_REGISTER_KMOD_DEP(net_atlantic, "* igb_uio | uio_pci_generic");
RTE_LOG_REGISTER(atl_logtype_init, pmd.net.atlantic.init, NOTICE);
RTE_LOG_REGISTER(atl_logtype_driver, pmd.net.atlantic.driver, NOTICE);