9c99878aa1
Introduce the RTE_LOG_REGISTER macro to avoid the code duplication in the logtype registration process. It is a wrapper macro for declaring the logtype, registering it and setting its level in the constructor context. Signed-off-by: Jerin Jacob <jerinj@marvell.com> Acked-by: Adam Dybkowski <adamx.dybkowski@intel.com> Acked-by: Sachin Saxena <sachin.saxena@nxp.com> Acked-by: Akhil Goyal <akhil.goyal@nxp.com>
1931 lines
50 KiB
C
1931 lines
50 KiB
C
/* SPDX-License-Identifier: BSD-3-Clause
|
|
* Copyright(c) 2018 Aquantia Corporation
|
|
*/
|
|
|
|
#include <rte_string_fns.h>
|
|
#include <rte_ethdev_pci.h>
|
|
#include <rte_alarm.h>
|
|
|
|
#include "atl_ethdev.h"
|
|
#include "atl_common.h"
|
|
#include "atl_hw_regs.h"
|
|
#include "atl_logs.h"
|
|
#include "hw_atl/hw_atl_llh.h"
|
|
#include "hw_atl/hw_atl_b0.h"
|
|
#include "hw_atl/hw_atl_b0_internal.h"
|
|
|
|
static int eth_atl_dev_init(struct rte_eth_dev *eth_dev);
|
|
static int eth_atl_dev_uninit(struct rte_eth_dev *eth_dev);
|
|
|
|
static int atl_dev_configure(struct rte_eth_dev *dev);
|
|
static int atl_dev_start(struct rte_eth_dev *dev);
|
|
static void atl_dev_stop(struct rte_eth_dev *dev);
|
|
static int atl_dev_set_link_up(struct rte_eth_dev *dev);
|
|
static int atl_dev_set_link_down(struct rte_eth_dev *dev);
|
|
static void atl_dev_close(struct rte_eth_dev *dev);
|
|
static int atl_dev_reset(struct rte_eth_dev *dev);
|
|
static int atl_dev_promiscuous_enable(struct rte_eth_dev *dev);
|
|
static int atl_dev_promiscuous_disable(struct rte_eth_dev *dev);
|
|
static int atl_dev_allmulticast_enable(struct rte_eth_dev *dev);
|
|
static int atl_dev_allmulticast_disable(struct rte_eth_dev *dev);
|
|
static int atl_dev_link_update(struct rte_eth_dev *dev, int wait);
|
|
|
|
static int atl_dev_xstats_get_names(struct rte_eth_dev *dev __rte_unused,
|
|
struct rte_eth_xstat_name *xstats_names,
|
|
unsigned int size);
|
|
|
|
static int atl_dev_stats_get(struct rte_eth_dev *dev,
|
|
struct rte_eth_stats *stats);
|
|
|
|
static int atl_dev_xstats_get(struct rte_eth_dev *dev,
|
|
struct rte_eth_xstat *stats, unsigned int n);
|
|
|
|
static int atl_dev_stats_reset(struct rte_eth_dev *dev);
|
|
|
|
static int atl_fw_version_get(struct rte_eth_dev *dev, char *fw_version,
|
|
size_t fw_size);
|
|
|
|
static const uint32_t *atl_dev_supported_ptypes_get(struct rte_eth_dev *dev);
|
|
|
|
static int atl_dev_mtu_set(struct rte_eth_dev *dev, uint16_t mtu);
|
|
|
|
/* VLAN stuff */
|
|
static int atl_vlan_filter_set(struct rte_eth_dev *dev,
|
|
uint16_t vlan_id, int on);
|
|
|
|
static int atl_vlan_offload_set(struct rte_eth_dev *dev, int mask);
|
|
|
|
static void atl_vlan_strip_queue_set(struct rte_eth_dev *dev,
|
|
uint16_t queue_id, int on);
|
|
|
|
static int atl_vlan_tpid_set(struct rte_eth_dev *dev,
|
|
enum rte_vlan_type vlan_type, uint16_t tpid);
|
|
|
|
/* EEPROM */
|
|
static int atl_dev_get_eeprom_length(struct rte_eth_dev *dev);
|
|
static int atl_dev_get_eeprom(struct rte_eth_dev *dev,
|
|
struct rte_dev_eeprom_info *eeprom);
|
|
static int atl_dev_set_eeprom(struct rte_eth_dev *dev,
|
|
struct rte_dev_eeprom_info *eeprom);
|
|
|
|
/* Regs */
|
|
static int atl_dev_get_regs(struct rte_eth_dev *dev,
|
|
struct rte_dev_reg_info *regs);
|
|
|
|
/* Flow control */
|
|
static int atl_flow_ctrl_get(struct rte_eth_dev *dev,
|
|
struct rte_eth_fc_conf *fc_conf);
|
|
static int atl_flow_ctrl_set(struct rte_eth_dev *dev,
|
|
struct rte_eth_fc_conf *fc_conf);
|
|
|
|
static void atl_dev_link_status_print(struct rte_eth_dev *dev);
|
|
|
|
/* Interrupts */
|
|
static int atl_dev_rxq_interrupt_setup(struct rte_eth_dev *dev);
|
|
static int atl_dev_lsc_interrupt_setup(struct rte_eth_dev *dev, uint8_t on);
|
|
static int atl_dev_interrupt_get_status(struct rte_eth_dev *dev);
|
|
static int atl_dev_interrupt_action(struct rte_eth_dev *dev,
|
|
struct rte_intr_handle *handle);
|
|
static void atl_dev_interrupt_handler(void *param);
|
|
|
|
|
|
static int atl_add_mac_addr(struct rte_eth_dev *dev,
|
|
struct rte_ether_addr *mac_addr,
|
|
uint32_t index, uint32_t pool);
|
|
static void atl_remove_mac_addr(struct rte_eth_dev *dev, uint32_t index);
|
|
static int atl_set_default_mac_addr(struct rte_eth_dev *dev,
|
|
struct rte_ether_addr *mac_addr);
|
|
|
|
static int atl_dev_set_mc_addr_list(struct rte_eth_dev *dev,
|
|
struct rte_ether_addr *mc_addr_set,
|
|
uint32_t nb_mc_addr);
|
|
|
|
/* RSS */
|
|
static int atl_reta_update(struct rte_eth_dev *dev,
|
|
struct rte_eth_rss_reta_entry64 *reta_conf,
|
|
uint16_t reta_size);
|
|
static int atl_reta_query(struct rte_eth_dev *dev,
|
|
struct rte_eth_rss_reta_entry64 *reta_conf,
|
|
uint16_t reta_size);
|
|
static int atl_rss_hash_update(struct rte_eth_dev *dev,
|
|
struct rte_eth_rss_conf *rss_conf);
|
|
static int atl_rss_hash_conf_get(struct rte_eth_dev *dev,
|
|
struct rte_eth_rss_conf *rss_conf);
|
|
|
|
|
|
static int eth_atl_pci_probe(struct rte_pci_driver *pci_drv __rte_unused,
|
|
struct rte_pci_device *pci_dev);
|
|
static int eth_atl_pci_remove(struct rte_pci_device *pci_dev);
|
|
|
|
static int atl_dev_info_get(struct rte_eth_dev *dev,
|
|
struct rte_eth_dev_info *dev_info);
|
|
|
|
/*
|
|
* The set of PCI devices this driver supports
|
|
*/
|
|
static const struct rte_pci_id pci_id_atl_map[] = {
|
|
{ RTE_PCI_DEVICE(PCI_VENDOR_ID_AQUANTIA, AQ_DEVICE_ID_0001) },
|
|
{ RTE_PCI_DEVICE(PCI_VENDOR_ID_AQUANTIA, AQ_DEVICE_ID_D100) },
|
|
{ RTE_PCI_DEVICE(PCI_VENDOR_ID_AQUANTIA, AQ_DEVICE_ID_D107) },
|
|
{ RTE_PCI_DEVICE(PCI_VENDOR_ID_AQUANTIA, AQ_DEVICE_ID_D108) },
|
|
{ RTE_PCI_DEVICE(PCI_VENDOR_ID_AQUANTIA, AQ_DEVICE_ID_D109) },
|
|
|
|
{ RTE_PCI_DEVICE(PCI_VENDOR_ID_AQUANTIA, AQ_DEVICE_ID_AQC100) },
|
|
{ RTE_PCI_DEVICE(PCI_VENDOR_ID_AQUANTIA, AQ_DEVICE_ID_AQC107) },
|
|
{ RTE_PCI_DEVICE(PCI_VENDOR_ID_AQUANTIA, AQ_DEVICE_ID_AQC108) },
|
|
{ RTE_PCI_DEVICE(PCI_VENDOR_ID_AQUANTIA, AQ_DEVICE_ID_AQC109) },
|
|
{ RTE_PCI_DEVICE(PCI_VENDOR_ID_AQUANTIA, AQ_DEVICE_ID_AQC111) },
|
|
{ RTE_PCI_DEVICE(PCI_VENDOR_ID_AQUANTIA, AQ_DEVICE_ID_AQC112) },
|
|
|
|
{ RTE_PCI_DEVICE(PCI_VENDOR_ID_AQUANTIA, AQ_DEVICE_ID_AQC100S) },
|
|
{ RTE_PCI_DEVICE(PCI_VENDOR_ID_AQUANTIA, AQ_DEVICE_ID_AQC107S) },
|
|
{ RTE_PCI_DEVICE(PCI_VENDOR_ID_AQUANTIA, AQ_DEVICE_ID_AQC108S) },
|
|
{ RTE_PCI_DEVICE(PCI_VENDOR_ID_AQUANTIA, AQ_DEVICE_ID_AQC109S) },
|
|
{ RTE_PCI_DEVICE(PCI_VENDOR_ID_AQUANTIA, AQ_DEVICE_ID_AQC111S) },
|
|
{ RTE_PCI_DEVICE(PCI_VENDOR_ID_AQUANTIA, AQ_DEVICE_ID_AQC112S) },
|
|
|
|
{ RTE_PCI_DEVICE(PCI_VENDOR_ID_AQUANTIA, AQ_DEVICE_ID_AQC111E) },
|
|
{ RTE_PCI_DEVICE(PCI_VENDOR_ID_AQUANTIA, AQ_DEVICE_ID_AQC112E) },
|
|
{ .vendor_id = 0, /* sentinel */ },
|
|
};
|
|
|
|
static struct rte_pci_driver rte_atl_pmd = {
|
|
.id_table = pci_id_atl_map,
|
|
.drv_flags = RTE_PCI_DRV_NEED_MAPPING | RTE_PCI_DRV_INTR_LSC,
|
|
.probe = eth_atl_pci_probe,
|
|
.remove = eth_atl_pci_remove,
|
|
};
|
|
|
|
#define ATL_RX_OFFLOADS (DEV_RX_OFFLOAD_VLAN_STRIP \
|
|
| DEV_RX_OFFLOAD_IPV4_CKSUM \
|
|
| DEV_RX_OFFLOAD_UDP_CKSUM \
|
|
| DEV_RX_OFFLOAD_TCP_CKSUM \
|
|
| DEV_RX_OFFLOAD_JUMBO_FRAME \
|
|
| DEV_RX_OFFLOAD_MACSEC_STRIP \
|
|
| DEV_RX_OFFLOAD_VLAN_FILTER)
|
|
|
|
#define ATL_TX_OFFLOADS (DEV_TX_OFFLOAD_VLAN_INSERT \
|
|
| DEV_TX_OFFLOAD_IPV4_CKSUM \
|
|
| DEV_TX_OFFLOAD_UDP_CKSUM \
|
|
| DEV_TX_OFFLOAD_TCP_CKSUM \
|
|
| DEV_TX_OFFLOAD_TCP_TSO \
|
|
| DEV_TX_OFFLOAD_MACSEC_INSERT \
|
|
| DEV_TX_OFFLOAD_MULTI_SEGS)
|
|
|
|
#define SFP_EEPROM_SIZE 0x100
|
|
|
|
static const struct rte_eth_desc_lim rx_desc_lim = {
|
|
.nb_max = ATL_MAX_RING_DESC,
|
|
.nb_min = ATL_MIN_RING_DESC,
|
|
.nb_align = ATL_RXD_ALIGN,
|
|
};
|
|
|
|
static const struct rte_eth_desc_lim tx_desc_lim = {
|
|
.nb_max = ATL_MAX_RING_DESC,
|
|
.nb_min = ATL_MIN_RING_DESC,
|
|
.nb_align = ATL_TXD_ALIGN,
|
|
.nb_seg_max = ATL_TX_MAX_SEG,
|
|
.nb_mtu_seg_max = ATL_TX_MAX_SEG,
|
|
};
|
|
|
|
enum atl_xstats_type {
|
|
XSTATS_TYPE_MSM = 0,
|
|
XSTATS_TYPE_MACSEC,
|
|
};
|
|
|
|
#define ATL_XSTATS_FIELD(name) { \
|
|
#name, \
|
|
offsetof(struct aq_stats_s, name), \
|
|
XSTATS_TYPE_MSM \
|
|
}
|
|
|
|
#define ATL_MACSEC_XSTATS_FIELD(name) { \
|
|
#name, \
|
|
offsetof(struct macsec_stats, name), \
|
|
XSTATS_TYPE_MACSEC \
|
|
}
|
|
|
|
struct atl_xstats_tbl_s {
|
|
const char *name;
|
|
unsigned int offset;
|
|
enum atl_xstats_type type;
|
|
};
|
|
|
|
static struct atl_xstats_tbl_s atl_xstats_tbl[] = {
|
|
ATL_XSTATS_FIELD(uprc),
|
|
ATL_XSTATS_FIELD(mprc),
|
|
ATL_XSTATS_FIELD(bprc),
|
|
ATL_XSTATS_FIELD(erpt),
|
|
ATL_XSTATS_FIELD(uptc),
|
|
ATL_XSTATS_FIELD(mptc),
|
|
ATL_XSTATS_FIELD(bptc),
|
|
ATL_XSTATS_FIELD(erpr),
|
|
ATL_XSTATS_FIELD(ubrc),
|
|
ATL_XSTATS_FIELD(ubtc),
|
|
ATL_XSTATS_FIELD(mbrc),
|
|
ATL_XSTATS_FIELD(mbtc),
|
|
ATL_XSTATS_FIELD(bbrc),
|
|
ATL_XSTATS_FIELD(bbtc),
|
|
/* Ingress Common Counters */
|
|
ATL_MACSEC_XSTATS_FIELD(in_ctl_pkts),
|
|
ATL_MACSEC_XSTATS_FIELD(in_tagged_miss_pkts),
|
|
ATL_MACSEC_XSTATS_FIELD(in_untagged_miss_pkts),
|
|
ATL_MACSEC_XSTATS_FIELD(in_notag_pkts),
|
|
ATL_MACSEC_XSTATS_FIELD(in_untagged_pkts),
|
|
ATL_MACSEC_XSTATS_FIELD(in_bad_tag_pkts),
|
|
ATL_MACSEC_XSTATS_FIELD(in_no_sci_pkts),
|
|
ATL_MACSEC_XSTATS_FIELD(in_unknown_sci_pkts),
|
|
/* Ingress SA Counters */
|
|
ATL_MACSEC_XSTATS_FIELD(in_untagged_hit_pkts),
|
|
ATL_MACSEC_XSTATS_FIELD(in_not_using_sa),
|
|
ATL_MACSEC_XSTATS_FIELD(in_unused_sa),
|
|
ATL_MACSEC_XSTATS_FIELD(in_not_valid_pkts),
|
|
ATL_MACSEC_XSTATS_FIELD(in_invalid_pkts),
|
|
ATL_MACSEC_XSTATS_FIELD(in_ok_pkts),
|
|
ATL_MACSEC_XSTATS_FIELD(in_unchecked_pkts),
|
|
ATL_MACSEC_XSTATS_FIELD(in_validated_octets),
|
|
ATL_MACSEC_XSTATS_FIELD(in_decrypted_octets),
|
|
/* Egress Common Counters */
|
|
ATL_MACSEC_XSTATS_FIELD(out_ctl_pkts),
|
|
ATL_MACSEC_XSTATS_FIELD(out_unknown_sa_pkts),
|
|
ATL_MACSEC_XSTATS_FIELD(out_untagged_pkts),
|
|
ATL_MACSEC_XSTATS_FIELD(out_too_long),
|
|
/* Egress SC Counters */
|
|
ATL_MACSEC_XSTATS_FIELD(out_sc_protected_pkts),
|
|
ATL_MACSEC_XSTATS_FIELD(out_sc_encrypted_pkts),
|
|
/* Egress SA Counters */
|
|
ATL_MACSEC_XSTATS_FIELD(out_sa_hit_drop_redirect),
|
|
ATL_MACSEC_XSTATS_FIELD(out_sa_protected2_pkts),
|
|
ATL_MACSEC_XSTATS_FIELD(out_sa_protected_pkts),
|
|
ATL_MACSEC_XSTATS_FIELD(out_sa_encrypted_pkts),
|
|
};
|
|
|
|
static const struct eth_dev_ops atl_eth_dev_ops = {
|
|
.dev_configure = atl_dev_configure,
|
|
.dev_start = atl_dev_start,
|
|
.dev_stop = atl_dev_stop,
|
|
.dev_set_link_up = atl_dev_set_link_up,
|
|
.dev_set_link_down = atl_dev_set_link_down,
|
|
.dev_close = atl_dev_close,
|
|
.dev_reset = atl_dev_reset,
|
|
|
|
/* PROMISC */
|
|
.promiscuous_enable = atl_dev_promiscuous_enable,
|
|
.promiscuous_disable = atl_dev_promiscuous_disable,
|
|
.allmulticast_enable = atl_dev_allmulticast_enable,
|
|
.allmulticast_disable = atl_dev_allmulticast_disable,
|
|
|
|
/* Link */
|
|
.link_update = atl_dev_link_update,
|
|
|
|
.get_reg = atl_dev_get_regs,
|
|
|
|
/* Stats */
|
|
.stats_get = atl_dev_stats_get,
|
|
.xstats_get = atl_dev_xstats_get,
|
|
.xstats_get_names = atl_dev_xstats_get_names,
|
|
.stats_reset = atl_dev_stats_reset,
|
|
.xstats_reset = atl_dev_stats_reset,
|
|
|
|
.fw_version_get = atl_fw_version_get,
|
|
.dev_infos_get = atl_dev_info_get,
|
|
.dev_supported_ptypes_get = atl_dev_supported_ptypes_get,
|
|
|
|
.mtu_set = atl_dev_mtu_set,
|
|
|
|
/* VLAN */
|
|
.vlan_filter_set = atl_vlan_filter_set,
|
|
.vlan_offload_set = atl_vlan_offload_set,
|
|
.vlan_tpid_set = atl_vlan_tpid_set,
|
|
.vlan_strip_queue_set = atl_vlan_strip_queue_set,
|
|
|
|
/* Queue Control */
|
|
.rx_queue_start = atl_rx_queue_start,
|
|
.rx_queue_stop = atl_rx_queue_stop,
|
|
.rx_queue_setup = atl_rx_queue_setup,
|
|
.rx_queue_release = atl_rx_queue_release,
|
|
|
|
.tx_queue_start = atl_tx_queue_start,
|
|
.tx_queue_stop = atl_tx_queue_stop,
|
|
.tx_queue_setup = atl_tx_queue_setup,
|
|
.tx_queue_release = atl_tx_queue_release,
|
|
|
|
.rx_queue_intr_enable = atl_dev_rx_queue_intr_enable,
|
|
.rx_queue_intr_disable = atl_dev_rx_queue_intr_disable,
|
|
|
|
.rx_queue_count = atl_rx_queue_count,
|
|
.rx_descriptor_status = atl_dev_rx_descriptor_status,
|
|
.tx_descriptor_status = atl_dev_tx_descriptor_status,
|
|
|
|
/* EEPROM */
|
|
.get_eeprom_length = atl_dev_get_eeprom_length,
|
|
.get_eeprom = atl_dev_get_eeprom,
|
|
.set_eeprom = atl_dev_set_eeprom,
|
|
|
|
/* Flow Control */
|
|
.flow_ctrl_get = atl_flow_ctrl_get,
|
|
.flow_ctrl_set = atl_flow_ctrl_set,
|
|
|
|
/* MAC */
|
|
.mac_addr_add = atl_add_mac_addr,
|
|
.mac_addr_remove = atl_remove_mac_addr,
|
|
.mac_addr_set = atl_set_default_mac_addr,
|
|
.set_mc_addr_list = atl_dev_set_mc_addr_list,
|
|
.rxq_info_get = atl_rxq_info_get,
|
|
.txq_info_get = atl_txq_info_get,
|
|
|
|
.reta_update = atl_reta_update,
|
|
.reta_query = atl_reta_query,
|
|
.rss_hash_update = atl_rss_hash_update,
|
|
.rss_hash_conf_get = atl_rss_hash_conf_get,
|
|
};
|
|
|
|
static inline int32_t
|
|
atl_reset_hw(struct aq_hw_s *hw)
|
|
{
|
|
return hw_atl_b0_hw_reset(hw);
|
|
}
|
|
|
|
static inline void
|
|
atl_enable_intr(struct rte_eth_dev *dev)
|
|
{
|
|
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
|
|
|
|
hw_atl_itr_irq_msk_setlsw_set(hw, 0xffffffff);
|
|
}
|
|
|
|
static void
|
|
atl_disable_intr(struct aq_hw_s *hw)
|
|
{
|
|
PMD_INIT_FUNC_TRACE();
|
|
hw_atl_itr_irq_msk_clearlsw_set(hw, 0xffffffff);
|
|
}
|
|
|
|
static int
|
|
eth_atl_dev_init(struct rte_eth_dev *eth_dev)
|
|
{
|
|
struct atl_adapter *adapter = eth_dev->data->dev_private;
|
|
struct rte_pci_device *pci_dev = RTE_ETH_DEV_TO_PCI(eth_dev);
|
|
struct rte_intr_handle *intr_handle = &pci_dev->intr_handle;
|
|
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(eth_dev->data->dev_private);
|
|
int err = 0;
|
|
|
|
PMD_INIT_FUNC_TRACE();
|
|
|
|
eth_dev->dev_ops = &atl_eth_dev_ops;
|
|
eth_dev->rx_pkt_burst = &atl_recv_pkts;
|
|
eth_dev->tx_pkt_burst = &atl_xmit_pkts;
|
|
eth_dev->tx_pkt_prepare = &atl_prep_pkts;
|
|
|
|
/* For secondary processes, the primary process has done all the work */
|
|
if (rte_eal_process_type() != RTE_PROC_PRIMARY)
|
|
return 0;
|
|
|
|
/* Vendor and Device ID need to be set before init of shared code */
|
|
hw->device_id = pci_dev->id.device_id;
|
|
hw->vendor_id = pci_dev->id.vendor_id;
|
|
hw->mmio = (void *)pci_dev->mem_resource[0].addr;
|
|
|
|
/* Hardware configuration - hardcode */
|
|
adapter->hw_cfg.is_lro = false;
|
|
adapter->hw_cfg.wol = false;
|
|
adapter->hw_cfg.is_rss = false;
|
|
adapter->hw_cfg.num_rss_queues = HW_ATL_B0_RSS_MAX;
|
|
|
|
adapter->hw_cfg.link_speed_msk = AQ_NIC_RATE_10G |
|
|
AQ_NIC_RATE_5G |
|
|
AQ_NIC_RATE_2G5 |
|
|
AQ_NIC_RATE_1G |
|
|
AQ_NIC_RATE_100M;
|
|
|
|
adapter->hw_cfg.flow_control = (AQ_NIC_FC_RX | AQ_NIC_FC_TX);
|
|
adapter->hw_cfg.aq_rss.indirection_table_size =
|
|
HW_ATL_B0_RSS_REDIRECTION_MAX;
|
|
|
|
hw->aq_nic_cfg = &adapter->hw_cfg;
|
|
|
|
pthread_mutex_init(&hw->mbox_mutex, NULL);
|
|
|
|
/* disable interrupt */
|
|
atl_disable_intr(hw);
|
|
|
|
/* Allocate memory for storing MAC addresses */
|
|
eth_dev->data->mac_addrs = rte_zmalloc("atlantic",
|
|
RTE_ETHER_ADDR_LEN, 0);
|
|
if (eth_dev->data->mac_addrs == NULL) {
|
|
PMD_INIT_LOG(ERR, "MAC Malloc failed");
|
|
return -ENOMEM;
|
|
}
|
|
|
|
err = hw_atl_utils_initfw(hw, &hw->aq_fw_ops);
|
|
if (err)
|
|
return err;
|
|
|
|
/* Copy the permanent MAC address */
|
|
if (hw->aq_fw_ops->get_mac_permanent(hw,
|
|
eth_dev->data->mac_addrs->addr_bytes) != 0)
|
|
return -EINVAL;
|
|
|
|
/* Reset the hw statistics */
|
|
atl_dev_stats_reset(eth_dev);
|
|
|
|
rte_intr_callback_register(intr_handle,
|
|
atl_dev_interrupt_handler, eth_dev);
|
|
|
|
/* enable uio/vfio intr/eventfd mapping */
|
|
rte_intr_enable(intr_handle);
|
|
|
|
/* enable support intr */
|
|
atl_enable_intr(eth_dev);
|
|
|
|
return err;
|
|
}
|
|
|
|
static int
|
|
eth_atl_dev_uninit(struct rte_eth_dev *eth_dev)
|
|
{
|
|
struct rte_pci_device *pci_dev = RTE_ETH_DEV_TO_PCI(eth_dev);
|
|
struct rte_intr_handle *intr_handle = &pci_dev->intr_handle;
|
|
struct aq_hw_s *hw;
|
|
|
|
PMD_INIT_FUNC_TRACE();
|
|
|
|
if (rte_eal_process_type() != RTE_PROC_PRIMARY)
|
|
return -EPERM;
|
|
|
|
hw = ATL_DEV_PRIVATE_TO_HW(eth_dev->data->dev_private);
|
|
|
|
if (hw->adapter_stopped == 0)
|
|
atl_dev_close(eth_dev);
|
|
|
|
eth_dev->dev_ops = NULL;
|
|
eth_dev->rx_pkt_burst = NULL;
|
|
eth_dev->tx_pkt_burst = NULL;
|
|
|
|
/* disable uio intr before callback unregister */
|
|
rte_intr_disable(intr_handle);
|
|
rte_intr_callback_unregister(intr_handle,
|
|
atl_dev_interrupt_handler, eth_dev);
|
|
|
|
rte_free(eth_dev->data->mac_addrs);
|
|
eth_dev->data->mac_addrs = NULL;
|
|
|
|
pthread_mutex_destroy(&hw->mbox_mutex);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
eth_atl_pci_probe(struct rte_pci_driver *pci_drv __rte_unused,
|
|
struct rte_pci_device *pci_dev)
|
|
{
|
|
return rte_eth_dev_pci_generic_probe(pci_dev,
|
|
sizeof(struct atl_adapter), eth_atl_dev_init);
|
|
}
|
|
|
|
static int
|
|
eth_atl_pci_remove(struct rte_pci_device *pci_dev)
|
|
{
|
|
return rte_eth_dev_pci_generic_remove(pci_dev, eth_atl_dev_uninit);
|
|
}
|
|
|
|
static int
|
|
atl_dev_configure(struct rte_eth_dev *dev)
|
|
{
|
|
struct atl_interrupt *intr =
|
|
ATL_DEV_PRIVATE_TO_INTR(dev->data->dev_private);
|
|
|
|
PMD_INIT_FUNC_TRACE();
|
|
|
|
/* set flag to update link status after init */
|
|
intr->flags |= ATL_FLAG_NEED_LINK_UPDATE;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* Configure device link speed and setup link.
|
|
* It returns 0 on success.
|
|
*/
|
|
static int
|
|
atl_dev_start(struct rte_eth_dev *dev)
|
|
{
|
|
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
|
|
struct rte_pci_device *pci_dev = RTE_ETH_DEV_TO_PCI(dev);
|
|
struct rte_intr_handle *intr_handle = &pci_dev->intr_handle;
|
|
uint32_t intr_vector = 0;
|
|
int status;
|
|
int err;
|
|
|
|
PMD_INIT_FUNC_TRACE();
|
|
|
|
/* set adapter started */
|
|
hw->adapter_stopped = 0;
|
|
|
|
if (dev->data->dev_conf.link_speeds & ETH_LINK_SPEED_FIXED) {
|
|
PMD_INIT_LOG(ERR,
|
|
"Invalid link_speeds for port %u, fix speed not supported",
|
|
dev->data->port_id);
|
|
return -EINVAL;
|
|
}
|
|
|
|
/* disable uio/vfio intr/eventfd mapping */
|
|
rte_intr_disable(intr_handle);
|
|
|
|
/* reinitialize adapter
|
|
* this calls reset and start
|
|
*/
|
|
status = atl_reset_hw(hw);
|
|
if (status != 0)
|
|
return -EIO;
|
|
|
|
err = hw_atl_b0_hw_init(hw, dev->data->mac_addrs->addr_bytes);
|
|
|
|
hw_atl_b0_hw_start(hw);
|
|
/* check and configure queue intr-vector mapping */
|
|
if ((rte_intr_cap_multiple(intr_handle) ||
|
|
!RTE_ETH_DEV_SRIOV(dev).active) &&
|
|
dev->data->dev_conf.intr_conf.rxq != 0) {
|
|
intr_vector = dev->data->nb_rx_queues;
|
|
if (intr_vector > ATL_MAX_INTR_QUEUE_NUM) {
|
|
PMD_INIT_LOG(ERR, "At most %d intr queues supported",
|
|
ATL_MAX_INTR_QUEUE_NUM);
|
|
return -ENOTSUP;
|
|
}
|
|
if (rte_intr_efd_enable(intr_handle, intr_vector)) {
|
|
PMD_INIT_LOG(ERR, "rte_intr_efd_enable failed");
|
|
return -1;
|
|
}
|
|
}
|
|
|
|
if (rte_intr_dp_is_en(intr_handle) && !intr_handle->intr_vec) {
|
|
intr_handle->intr_vec = rte_zmalloc("intr_vec",
|
|
dev->data->nb_rx_queues * sizeof(int), 0);
|
|
if (intr_handle->intr_vec == NULL) {
|
|
PMD_INIT_LOG(ERR, "Failed to allocate %d rx_queues"
|
|
" intr_vec", dev->data->nb_rx_queues);
|
|
return -ENOMEM;
|
|
}
|
|
}
|
|
|
|
/* initialize transmission unit */
|
|
atl_tx_init(dev);
|
|
|
|
/* This can fail when allocating mbufs for descriptor rings */
|
|
err = atl_rx_init(dev);
|
|
if (err) {
|
|
PMD_INIT_LOG(ERR, "Unable to initialize RX hardware");
|
|
goto error;
|
|
}
|
|
|
|
PMD_INIT_LOG(DEBUG, "FW version: %u.%u.%u",
|
|
hw->fw_ver_actual >> 24,
|
|
(hw->fw_ver_actual >> 16) & 0xFF,
|
|
hw->fw_ver_actual & 0xFFFF);
|
|
PMD_INIT_LOG(DEBUG, "Driver version: %s", ATL_PMD_DRIVER_VERSION);
|
|
|
|
err = atl_start_queues(dev);
|
|
if (err < 0) {
|
|
PMD_INIT_LOG(ERR, "Unable to start rxtx queues");
|
|
goto error;
|
|
}
|
|
|
|
err = atl_dev_set_link_up(dev);
|
|
|
|
err = hw->aq_fw_ops->update_link_status(hw);
|
|
|
|
if (err)
|
|
goto error;
|
|
|
|
dev->data->dev_link.link_status = hw->aq_link_status.mbps != 0;
|
|
|
|
if (rte_intr_allow_others(intr_handle)) {
|
|
/* check if lsc interrupt is enabled */
|
|
if (dev->data->dev_conf.intr_conf.lsc != 0)
|
|
atl_dev_lsc_interrupt_setup(dev, true);
|
|
else
|
|
atl_dev_lsc_interrupt_setup(dev, false);
|
|
} else {
|
|
rte_intr_callback_unregister(intr_handle,
|
|
atl_dev_interrupt_handler, dev);
|
|
if (dev->data->dev_conf.intr_conf.lsc != 0)
|
|
PMD_INIT_LOG(INFO, "lsc won't enable because of"
|
|
" no intr multiplex");
|
|
}
|
|
|
|
/* check if rxq interrupt is enabled */
|
|
if (dev->data->dev_conf.intr_conf.rxq != 0 &&
|
|
rte_intr_dp_is_en(intr_handle))
|
|
atl_dev_rxq_interrupt_setup(dev);
|
|
|
|
/* enable uio/vfio intr/eventfd mapping */
|
|
rte_intr_enable(intr_handle);
|
|
|
|
/* resume enabled intr since hw reset */
|
|
atl_enable_intr(dev);
|
|
|
|
return 0;
|
|
|
|
error:
|
|
atl_stop_queues(dev);
|
|
return -EIO;
|
|
}
|
|
|
|
/*
|
|
* Stop device: disable rx and tx functions to allow for reconfiguring.
|
|
*/
|
|
static void
|
|
atl_dev_stop(struct rte_eth_dev *dev)
|
|
{
|
|
struct rte_eth_link link;
|
|
struct aq_hw_s *hw =
|
|
ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
|
|
struct rte_pci_device *pci_dev = RTE_ETH_DEV_TO_PCI(dev);
|
|
struct rte_intr_handle *intr_handle = &pci_dev->intr_handle;
|
|
|
|
PMD_INIT_FUNC_TRACE();
|
|
|
|
/* disable interrupts */
|
|
atl_disable_intr(hw);
|
|
|
|
/* reset the NIC */
|
|
atl_reset_hw(hw);
|
|
hw->adapter_stopped = 1;
|
|
|
|
atl_stop_queues(dev);
|
|
|
|
/* Clear stored conf */
|
|
dev->data->scattered_rx = 0;
|
|
dev->data->lro = 0;
|
|
|
|
/* Clear recorded link status */
|
|
memset(&link, 0, sizeof(link));
|
|
rte_eth_linkstatus_set(dev, &link);
|
|
|
|
if (!rte_intr_allow_others(intr_handle))
|
|
/* resume to the default handler */
|
|
rte_intr_callback_register(intr_handle,
|
|
atl_dev_interrupt_handler,
|
|
(void *)dev);
|
|
|
|
/* Clean datapath event and queue/vec mapping */
|
|
rte_intr_efd_disable(intr_handle);
|
|
if (intr_handle->intr_vec != NULL) {
|
|
rte_free(intr_handle->intr_vec);
|
|
intr_handle->intr_vec = NULL;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Set device link up: enable tx.
|
|
*/
|
|
static int
|
|
atl_dev_set_link_up(struct rte_eth_dev *dev)
|
|
{
|
|
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
|
|
uint32_t link_speeds = dev->data->dev_conf.link_speeds;
|
|
uint32_t speed_mask = 0;
|
|
|
|
if (link_speeds == ETH_LINK_SPEED_AUTONEG) {
|
|
speed_mask = hw->aq_nic_cfg->link_speed_msk;
|
|
} else {
|
|
if (link_speeds & ETH_LINK_SPEED_10G)
|
|
speed_mask |= AQ_NIC_RATE_10G;
|
|
if (link_speeds & ETH_LINK_SPEED_5G)
|
|
speed_mask |= AQ_NIC_RATE_5G;
|
|
if (link_speeds & ETH_LINK_SPEED_1G)
|
|
speed_mask |= AQ_NIC_RATE_1G;
|
|
if (link_speeds & ETH_LINK_SPEED_2_5G)
|
|
speed_mask |= AQ_NIC_RATE_2G5;
|
|
if (link_speeds & ETH_LINK_SPEED_100M)
|
|
speed_mask |= AQ_NIC_RATE_100M;
|
|
}
|
|
|
|
return hw->aq_fw_ops->set_link_speed(hw, speed_mask);
|
|
}
|
|
|
|
/*
|
|
* Set device link down: disable tx.
|
|
*/
|
|
static int
|
|
atl_dev_set_link_down(struct rte_eth_dev *dev)
|
|
{
|
|
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
|
|
|
|
return hw->aq_fw_ops->set_link_speed(hw, 0);
|
|
}
|
|
|
|
/*
|
|
* Reset and stop device.
|
|
*/
|
|
static void
|
|
atl_dev_close(struct rte_eth_dev *dev)
|
|
{
|
|
PMD_INIT_FUNC_TRACE();
|
|
|
|
atl_dev_stop(dev);
|
|
|
|
atl_free_queues(dev);
|
|
}
|
|
|
|
static int
|
|
atl_dev_reset(struct rte_eth_dev *dev)
|
|
{
|
|
int ret;
|
|
|
|
ret = eth_atl_dev_uninit(dev);
|
|
if (ret)
|
|
return ret;
|
|
|
|
ret = eth_atl_dev_init(dev);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int
|
|
atl_dev_configure_macsec(struct rte_eth_dev *dev)
|
|
{
|
|
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
|
|
struct aq_hw_cfg_s *cf = ATL_DEV_PRIVATE_TO_CFG(dev->data->dev_private);
|
|
struct aq_macsec_config *aqcfg = &cf->aq_macsec;
|
|
struct macsec_msg_fw_request msg_macsec;
|
|
struct macsec_msg_fw_response response;
|
|
|
|
if (!aqcfg->common.macsec_enabled ||
|
|
hw->aq_fw_ops->send_macsec_req == NULL)
|
|
return 0;
|
|
|
|
memset(&msg_macsec, 0, sizeof(msg_macsec));
|
|
|
|
/* Creating set of sc/sa structures from parameters provided by DPDK */
|
|
|
|
/* Configure macsec */
|
|
msg_macsec.msg_type = macsec_cfg_msg;
|
|
msg_macsec.cfg.enabled = aqcfg->common.macsec_enabled;
|
|
msg_macsec.cfg.interrupts_enabled = 1;
|
|
|
|
hw->aq_fw_ops->send_macsec_req(hw, &msg_macsec, &response);
|
|
|
|
if (response.result)
|
|
return -1;
|
|
|
|
memset(&msg_macsec, 0, sizeof(msg_macsec));
|
|
|
|
/* Configure TX SC */
|
|
|
|
msg_macsec.msg_type = macsec_add_tx_sc_msg;
|
|
msg_macsec.txsc.index = 0; /* TXSC always one (??) */
|
|
msg_macsec.txsc.protect = aqcfg->common.encryption_enabled;
|
|
|
|
/* MAC addr for TX */
|
|
msg_macsec.txsc.mac_sa[0] = rte_bswap32(aqcfg->txsc.mac[1]);
|
|
msg_macsec.txsc.mac_sa[1] = rte_bswap32(aqcfg->txsc.mac[0]);
|
|
msg_macsec.txsc.sa_mask = 0x3f;
|
|
|
|
msg_macsec.txsc.da_mask = 0;
|
|
msg_macsec.txsc.tci = 0x0B;
|
|
msg_macsec.txsc.curr_an = 0; /* SA index which currently used */
|
|
|
|
/*
|
|
* Creating SCI (Secure Channel Identifier).
|
|
* SCI constructed from Source MAC and Port identifier
|
|
*/
|
|
uint32_t sci_hi_part = (msg_macsec.txsc.mac_sa[1] << 16) |
|
|
(msg_macsec.txsc.mac_sa[0] >> 16);
|
|
uint32_t sci_low_part = (msg_macsec.txsc.mac_sa[0] << 16);
|
|
|
|
uint32_t port_identifier = 1;
|
|
|
|
msg_macsec.txsc.sci[1] = sci_hi_part;
|
|
msg_macsec.txsc.sci[0] = sci_low_part | port_identifier;
|
|
|
|
hw->aq_fw_ops->send_macsec_req(hw, &msg_macsec, &response);
|
|
|
|
if (response.result)
|
|
return -1;
|
|
|
|
memset(&msg_macsec, 0, sizeof(msg_macsec));
|
|
|
|
/* Configure RX SC */
|
|
|
|
msg_macsec.msg_type = macsec_add_rx_sc_msg;
|
|
msg_macsec.rxsc.index = aqcfg->rxsc.pi;
|
|
msg_macsec.rxsc.replay_protect =
|
|
aqcfg->common.replay_protection_enabled;
|
|
msg_macsec.rxsc.anti_replay_window = 0;
|
|
|
|
/* MAC addr for RX */
|
|
msg_macsec.rxsc.mac_da[0] = rte_bswap32(aqcfg->rxsc.mac[1]);
|
|
msg_macsec.rxsc.mac_da[1] = rte_bswap32(aqcfg->rxsc.mac[0]);
|
|
msg_macsec.rxsc.da_mask = 0;//0x3f;
|
|
|
|
msg_macsec.rxsc.sa_mask = 0;
|
|
|
|
hw->aq_fw_ops->send_macsec_req(hw, &msg_macsec, &response);
|
|
|
|
if (response.result)
|
|
return -1;
|
|
|
|
memset(&msg_macsec, 0, sizeof(msg_macsec));
|
|
|
|
/* Configure RX SC */
|
|
|
|
msg_macsec.msg_type = macsec_add_tx_sa_msg;
|
|
msg_macsec.txsa.index = aqcfg->txsa.idx;
|
|
msg_macsec.txsa.next_pn = aqcfg->txsa.pn;
|
|
|
|
msg_macsec.txsa.key[0] = rte_bswap32(aqcfg->txsa.key[3]);
|
|
msg_macsec.txsa.key[1] = rte_bswap32(aqcfg->txsa.key[2]);
|
|
msg_macsec.txsa.key[2] = rte_bswap32(aqcfg->txsa.key[1]);
|
|
msg_macsec.txsa.key[3] = rte_bswap32(aqcfg->txsa.key[0]);
|
|
|
|
hw->aq_fw_ops->send_macsec_req(hw, &msg_macsec, &response);
|
|
|
|
if (response.result)
|
|
return -1;
|
|
|
|
memset(&msg_macsec, 0, sizeof(msg_macsec));
|
|
|
|
/* Configure RX SA */
|
|
|
|
msg_macsec.msg_type = macsec_add_rx_sa_msg;
|
|
msg_macsec.rxsa.index = aqcfg->rxsa.idx;
|
|
msg_macsec.rxsa.next_pn = aqcfg->rxsa.pn;
|
|
|
|
msg_macsec.rxsa.key[0] = rte_bswap32(aqcfg->rxsa.key[3]);
|
|
msg_macsec.rxsa.key[1] = rte_bswap32(aqcfg->rxsa.key[2]);
|
|
msg_macsec.rxsa.key[2] = rte_bswap32(aqcfg->rxsa.key[1]);
|
|
msg_macsec.rxsa.key[3] = rte_bswap32(aqcfg->rxsa.key[0]);
|
|
|
|
hw->aq_fw_ops->send_macsec_req(hw, &msg_macsec, &response);
|
|
|
|
if (response.result)
|
|
return -1;
|
|
|
|
return 0;
|
|
}
|
|
|
|
int atl_macsec_enable(struct rte_eth_dev *dev,
|
|
uint8_t encr, uint8_t repl_prot)
|
|
{
|
|
struct aq_hw_cfg_s *cfg =
|
|
ATL_DEV_PRIVATE_TO_CFG(dev->data->dev_private);
|
|
|
|
cfg->aq_macsec.common.macsec_enabled = 1;
|
|
cfg->aq_macsec.common.encryption_enabled = encr;
|
|
cfg->aq_macsec.common.replay_protection_enabled = repl_prot;
|
|
|
|
return 0;
|
|
}
|
|
|
|
int atl_macsec_disable(struct rte_eth_dev *dev)
|
|
{
|
|
struct aq_hw_cfg_s *cfg =
|
|
ATL_DEV_PRIVATE_TO_CFG(dev->data->dev_private);
|
|
|
|
cfg->aq_macsec.common.macsec_enabled = 0;
|
|
|
|
return 0;
|
|
}
|
|
|
|
int atl_macsec_config_txsc(struct rte_eth_dev *dev, uint8_t *mac)
|
|
{
|
|
struct aq_hw_cfg_s *cfg =
|
|
ATL_DEV_PRIVATE_TO_CFG(dev->data->dev_private);
|
|
|
|
memset(&cfg->aq_macsec.txsc.mac, 0, sizeof(cfg->aq_macsec.txsc.mac));
|
|
memcpy((uint8_t *)&cfg->aq_macsec.txsc.mac + 2, mac,
|
|
RTE_ETHER_ADDR_LEN);
|
|
|
|
return 0;
|
|
}
|
|
|
|
int atl_macsec_config_rxsc(struct rte_eth_dev *dev,
|
|
uint8_t *mac, uint16_t pi)
|
|
{
|
|
struct aq_hw_cfg_s *cfg =
|
|
ATL_DEV_PRIVATE_TO_CFG(dev->data->dev_private);
|
|
|
|
memset(&cfg->aq_macsec.rxsc.mac, 0, sizeof(cfg->aq_macsec.rxsc.mac));
|
|
memcpy((uint8_t *)&cfg->aq_macsec.rxsc.mac + 2, mac,
|
|
RTE_ETHER_ADDR_LEN);
|
|
cfg->aq_macsec.rxsc.pi = pi;
|
|
|
|
return 0;
|
|
}
|
|
|
|
int atl_macsec_select_txsa(struct rte_eth_dev *dev,
|
|
uint8_t idx, uint8_t an,
|
|
uint32_t pn, uint8_t *key)
|
|
{
|
|
struct aq_hw_cfg_s *cfg =
|
|
ATL_DEV_PRIVATE_TO_CFG(dev->data->dev_private);
|
|
|
|
cfg->aq_macsec.txsa.idx = idx;
|
|
cfg->aq_macsec.txsa.pn = pn;
|
|
cfg->aq_macsec.txsa.an = an;
|
|
|
|
memcpy(&cfg->aq_macsec.txsa.key, key, 16);
|
|
return 0;
|
|
}
|
|
|
|
int atl_macsec_select_rxsa(struct rte_eth_dev *dev,
|
|
uint8_t idx, uint8_t an,
|
|
uint32_t pn, uint8_t *key)
|
|
{
|
|
struct aq_hw_cfg_s *cfg =
|
|
ATL_DEV_PRIVATE_TO_CFG(dev->data->dev_private);
|
|
|
|
cfg->aq_macsec.rxsa.idx = idx;
|
|
cfg->aq_macsec.rxsa.pn = pn;
|
|
cfg->aq_macsec.rxsa.an = an;
|
|
|
|
memcpy(&cfg->aq_macsec.rxsa.key, key, 16);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
atl_dev_stats_get(struct rte_eth_dev *dev, struct rte_eth_stats *stats)
|
|
{
|
|
struct atl_adapter *adapter = ATL_DEV_TO_ADAPTER(dev);
|
|
struct aq_hw_s *hw = &adapter->hw;
|
|
struct atl_sw_stats *swstats = &adapter->sw_stats;
|
|
unsigned int i;
|
|
|
|
hw->aq_fw_ops->update_stats(hw);
|
|
|
|
/* Fill out the rte_eth_stats statistics structure */
|
|
stats->ipackets = hw->curr_stats.dma_pkt_rc;
|
|
stats->ibytes = hw->curr_stats.dma_oct_rc;
|
|
stats->imissed = hw->curr_stats.dpc;
|
|
stats->ierrors = hw->curr_stats.erpt;
|
|
|
|
stats->opackets = hw->curr_stats.dma_pkt_tc;
|
|
stats->obytes = hw->curr_stats.dma_oct_tc;
|
|
stats->oerrors = 0;
|
|
|
|
stats->rx_nombuf = swstats->rx_nombuf;
|
|
|
|
for (i = 0; i < RTE_ETHDEV_QUEUE_STAT_CNTRS; i++) {
|
|
stats->q_ipackets[i] = swstats->q_ipackets[i];
|
|
stats->q_opackets[i] = swstats->q_opackets[i];
|
|
stats->q_ibytes[i] = swstats->q_ibytes[i];
|
|
stats->q_obytes[i] = swstats->q_obytes[i];
|
|
stats->q_errors[i] = swstats->q_errors[i];
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
atl_dev_stats_reset(struct rte_eth_dev *dev)
|
|
{
|
|
struct atl_adapter *adapter = ATL_DEV_TO_ADAPTER(dev);
|
|
struct aq_hw_s *hw = &adapter->hw;
|
|
|
|
hw->aq_fw_ops->update_stats(hw);
|
|
|
|
/* Reset software totals */
|
|
memset(&hw->curr_stats, 0, sizeof(hw->curr_stats));
|
|
|
|
memset(&adapter->sw_stats, 0, sizeof(adapter->sw_stats));
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
atl_dev_xstats_get_count(struct rte_eth_dev *dev)
|
|
{
|
|
struct atl_adapter *adapter =
|
|
(struct atl_adapter *)dev->data->dev_private;
|
|
|
|
struct aq_hw_s *hw = &adapter->hw;
|
|
unsigned int i, count = 0;
|
|
|
|
for (i = 0; i < RTE_DIM(atl_xstats_tbl); i++) {
|
|
if (atl_xstats_tbl[i].type == XSTATS_TYPE_MACSEC &&
|
|
((hw->caps_lo & BIT(CAPS_LO_MACSEC)) == 0))
|
|
continue;
|
|
|
|
count++;
|
|
}
|
|
|
|
return count;
|
|
}
|
|
|
|
static int
|
|
atl_dev_xstats_get_names(struct rte_eth_dev *dev __rte_unused,
|
|
struct rte_eth_xstat_name *xstats_names,
|
|
unsigned int size)
|
|
{
|
|
unsigned int i;
|
|
unsigned int count = atl_dev_xstats_get_count(dev);
|
|
|
|
if (xstats_names) {
|
|
for (i = 0; i < size && i < count; i++) {
|
|
snprintf(xstats_names[i].name,
|
|
RTE_ETH_XSTATS_NAME_SIZE, "%s",
|
|
atl_xstats_tbl[i].name);
|
|
}
|
|
}
|
|
|
|
return count;
|
|
}
|
|
|
|
static int
|
|
atl_dev_xstats_get(struct rte_eth_dev *dev, struct rte_eth_xstat *stats,
|
|
unsigned int n)
|
|
{
|
|
struct atl_adapter *adapter = dev->data->dev_private;
|
|
struct aq_hw_s *hw = &adapter->hw;
|
|
struct get_stats req = { 0 };
|
|
struct macsec_msg_fw_request msg = { 0 };
|
|
struct macsec_msg_fw_response resp = { 0 };
|
|
int err = -1;
|
|
unsigned int i;
|
|
unsigned int count = atl_dev_xstats_get_count(dev);
|
|
|
|
if (!stats)
|
|
return count;
|
|
|
|
if (hw->aq_fw_ops->send_macsec_req != NULL) {
|
|
req.ingress_sa_index = 0xff;
|
|
req.egress_sc_index = 0xff;
|
|
req.egress_sa_index = 0xff;
|
|
|
|
msg.msg_type = macsec_get_stats_msg;
|
|
msg.stats = req;
|
|
|
|
err = hw->aq_fw_ops->send_macsec_req(hw, &msg, &resp);
|
|
}
|
|
|
|
for (i = 0; i < n && i < count; i++) {
|
|
stats[i].id = i;
|
|
|
|
switch (atl_xstats_tbl[i].type) {
|
|
case XSTATS_TYPE_MSM:
|
|
stats[i].value = *(u64 *)((uint8_t *)&hw->curr_stats +
|
|
atl_xstats_tbl[i].offset);
|
|
break;
|
|
case XSTATS_TYPE_MACSEC:
|
|
if (!err) {
|
|
stats[i].value =
|
|
*(u64 *)((uint8_t *)&resp.stats +
|
|
atl_xstats_tbl[i].offset);
|
|
}
|
|
break;
|
|
}
|
|
}
|
|
|
|
return i;
|
|
}
|
|
|
|
static int
|
|
atl_fw_version_get(struct rte_eth_dev *dev, char *fw_version, size_t fw_size)
|
|
{
|
|
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
|
|
uint32_t fw_ver = 0;
|
|
unsigned int ret = 0;
|
|
|
|
ret = hw_atl_utils_get_fw_version(hw, &fw_ver);
|
|
if (ret)
|
|
return -EIO;
|
|
|
|
ret = snprintf(fw_version, fw_size, "%u.%u.%u", fw_ver >> 24,
|
|
(fw_ver >> 16) & 0xFFU, fw_ver & 0xFFFFU);
|
|
|
|
ret += 1; /* add string null-terminator */
|
|
|
|
if (fw_size < ret)
|
|
return ret;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
atl_dev_info_get(struct rte_eth_dev *dev, struct rte_eth_dev_info *dev_info)
|
|
{
|
|
struct rte_pci_device *pci_dev = RTE_ETH_DEV_TO_PCI(dev);
|
|
|
|
dev_info->max_rx_queues = AQ_HW_MAX_RX_QUEUES;
|
|
dev_info->max_tx_queues = AQ_HW_MAX_TX_QUEUES;
|
|
|
|
dev_info->min_rx_bufsize = 1024;
|
|
dev_info->max_rx_pktlen = HW_ATL_B0_MTU_JUMBO;
|
|
dev_info->max_mac_addrs = HW_ATL_B0_MAC_MAX;
|
|
dev_info->max_vfs = pci_dev->max_vfs;
|
|
|
|
dev_info->max_hash_mac_addrs = 0;
|
|
dev_info->max_vmdq_pools = 0;
|
|
dev_info->vmdq_queue_num = 0;
|
|
|
|
dev_info->rx_offload_capa = ATL_RX_OFFLOADS;
|
|
|
|
dev_info->tx_offload_capa = ATL_TX_OFFLOADS;
|
|
|
|
|
|
dev_info->default_rxconf = (struct rte_eth_rxconf) {
|
|
.rx_free_thresh = ATL_DEFAULT_RX_FREE_THRESH,
|
|
};
|
|
|
|
dev_info->default_txconf = (struct rte_eth_txconf) {
|
|
.tx_free_thresh = ATL_DEFAULT_TX_FREE_THRESH,
|
|
};
|
|
|
|
dev_info->rx_desc_lim = rx_desc_lim;
|
|
dev_info->tx_desc_lim = tx_desc_lim;
|
|
|
|
dev_info->hash_key_size = HW_ATL_B0_RSS_HASHKEY_BITS / 8;
|
|
dev_info->reta_size = HW_ATL_B0_RSS_REDIRECTION_MAX;
|
|
dev_info->flow_type_rss_offloads = ATL_RSS_OFFLOAD_ALL;
|
|
|
|
dev_info->speed_capa = ETH_LINK_SPEED_1G | ETH_LINK_SPEED_10G;
|
|
dev_info->speed_capa |= ETH_LINK_SPEED_100M;
|
|
dev_info->speed_capa |= ETH_LINK_SPEED_2_5G;
|
|
dev_info->speed_capa |= ETH_LINK_SPEED_5G;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static const uint32_t *
|
|
atl_dev_supported_ptypes_get(struct rte_eth_dev *dev)
|
|
{
|
|
static const uint32_t ptypes[] = {
|
|
RTE_PTYPE_L2_ETHER,
|
|
RTE_PTYPE_L2_ETHER_ARP,
|
|
RTE_PTYPE_L2_ETHER_VLAN,
|
|
RTE_PTYPE_L3_IPV4,
|
|
RTE_PTYPE_L3_IPV6,
|
|
RTE_PTYPE_L4_TCP,
|
|
RTE_PTYPE_L4_UDP,
|
|
RTE_PTYPE_L4_SCTP,
|
|
RTE_PTYPE_L4_ICMP,
|
|
RTE_PTYPE_UNKNOWN
|
|
};
|
|
|
|
if (dev->rx_pkt_burst == atl_recv_pkts)
|
|
return ptypes;
|
|
|
|
return NULL;
|
|
}
|
|
|
|
static void
|
|
atl_dev_delayed_handler(void *param)
|
|
{
|
|
struct rte_eth_dev *dev = (struct rte_eth_dev *)param;
|
|
|
|
atl_dev_configure_macsec(dev);
|
|
}
|
|
|
|
|
|
/* return 0 means link status changed, -1 means not changed */
|
|
static int
|
|
atl_dev_link_update(struct rte_eth_dev *dev, int wait __rte_unused)
|
|
{
|
|
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
|
|
struct rte_eth_link link, old;
|
|
u32 fc = AQ_NIC_FC_OFF;
|
|
int err = 0;
|
|
|
|
link.link_status = ETH_LINK_DOWN;
|
|
link.link_speed = 0;
|
|
link.link_duplex = ETH_LINK_FULL_DUPLEX;
|
|
link.link_autoneg = hw->is_autoneg ? ETH_LINK_AUTONEG : ETH_LINK_FIXED;
|
|
memset(&old, 0, sizeof(old));
|
|
|
|
/* load old link status */
|
|
rte_eth_linkstatus_get(dev, &old);
|
|
|
|
/* read current link status */
|
|
err = hw->aq_fw_ops->update_link_status(hw);
|
|
|
|
if (err)
|
|
return 0;
|
|
|
|
if (hw->aq_link_status.mbps == 0) {
|
|
/* write default (down) link status */
|
|
rte_eth_linkstatus_set(dev, &link);
|
|
if (link.link_status == old.link_status)
|
|
return -1;
|
|
return 0;
|
|
}
|
|
|
|
link.link_status = ETH_LINK_UP;
|
|
link.link_duplex = ETH_LINK_FULL_DUPLEX;
|
|
link.link_speed = hw->aq_link_status.mbps;
|
|
|
|
rte_eth_linkstatus_set(dev, &link);
|
|
|
|
if (link.link_status == old.link_status)
|
|
return -1;
|
|
|
|
/* Driver has to update flow control settings on RX block
|
|
* on any link event.
|
|
* We should query FW whether it negotiated FC.
|
|
*/
|
|
if (hw->aq_fw_ops->get_flow_control) {
|
|
hw->aq_fw_ops->get_flow_control(hw, &fc);
|
|
hw_atl_b0_set_fc(hw, fc, 0U);
|
|
}
|
|
|
|
if (rte_eal_alarm_set(1000 * 1000,
|
|
atl_dev_delayed_handler, (void *)dev) < 0)
|
|
PMD_DRV_LOG(ERR, "rte_eal_alarm_set fail");
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
atl_dev_promiscuous_enable(struct rte_eth_dev *dev)
|
|
{
|
|
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
|
|
|
|
hw_atl_rpfl2promiscuous_mode_en_set(hw, true);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
atl_dev_promiscuous_disable(struct rte_eth_dev *dev)
|
|
{
|
|
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
|
|
|
|
hw_atl_rpfl2promiscuous_mode_en_set(hw, false);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
atl_dev_allmulticast_enable(struct rte_eth_dev *dev)
|
|
{
|
|
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
|
|
|
|
hw_atl_rpfl2_accept_all_mc_packets_set(hw, true);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
atl_dev_allmulticast_disable(struct rte_eth_dev *dev)
|
|
{
|
|
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
|
|
|
|
if (dev->data->promiscuous == 1)
|
|
return 0; /* must remain in all_multicast mode */
|
|
|
|
hw_atl_rpfl2_accept_all_mc_packets_set(hw, false);
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* It clears the interrupt causes and enables the interrupt.
|
|
* It will be called once only during nic initialized.
|
|
*
|
|
* @param dev
|
|
* Pointer to struct rte_eth_dev.
|
|
* @param on
|
|
* Enable or Disable.
|
|
*
|
|
* @return
|
|
* - On success, zero.
|
|
* - On failure, a negative value.
|
|
*/
|
|
|
|
static int
|
|
atl_dev_lsc_interrupt_setup(struct rte_eth_dev *dev, uint8_t on __rte_unused)
|
|
{
|
|
atl_dev_link_status_print(dev);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
atl_dev_rxq_interrupt_setup(struct rte_eth_dev *dev __rte_unused)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
|
|
static int
|
|
atl_dev_interrupt_get_status(struct rte_eth_dev *dev)
|
|
{
|
|
struct atl_interrupt *intr =
|
|
ATL_DEV_PRIVATE_TO_INTR(dev->data->dev_private);
|
|
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
|
|
u64 cause = 0;
|
|
|
|
hw_atl_b0_hw_irq_read(hw, &cause);
|
|
|
|
atl_disable_intr(hw);
|
|
|
|
if (cause & BIT(ATL_IRQ_CAUSE_LINK))
|
|
intr->flags |= ATL_FLAG_NEED_LINK_UPDATE;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* It gets and then prints the link status.
|
|
*
|
|
* @param dev
|
|
* Pointer to struct rte_eth_dev.
|
|
*
|
|
* @return
|
|
* - On success, zero.
|
|
* - On failure, a negative value.
|
|
*/
|
|
static void
|
|
atl_dev_link_status_print(struct rte_eth_dev *dev)
|
|
{
|
|
struct rte_eth_link link;
|
|
|
|
memset(&link, 0, sizeof(link));
|
|
rte_eth_linkstatus_get(dev, &link);
|
|
if (link.link_status) {
|
|
PMD_DRV_LOG(INFO, "Port %d: Link Up - speed %u Mbps - %s",
|
|
(int)(dev->data->port_id),
|
|
(unsigned int)link.link_speed,
|
|
link.link_duplex == ETH_LINK_FULL_DUPLEX ?
|
|
"full-duplex" : "half-duplex");
|
|
} else {
|
|
PMD_DRV_LOG(INFO, " Port %d: Link Down",
|
|
(int)(dev->data->port_id));
|
|
}
|
|
|
|
|
|
#ifdef DEBUG
|
|
{
|
|
struct rte_pci_device *pci_dev = RTE_ETH_DEV_TO_PCI(dev);
|
|
|
|
PMD_DRV_LOG(DEBUG, "PCI Address: " PCI_PRI_FMT,
|
|
pci_dev->addr.domain,
|
|
pci_dev->addr.bus,
|
|
pci_dev->addr.devid,
|
|
pci_dev->addr.function);
|
|
}
|
|
#endif
|
|
|
|
PMD_DRV_LOG(INFO, "Link speed:%d", link.link_speed);
|
|
}
|
|
|
|
/*
|
|
* It executes link_update after knowing an interrupt occurred.
|
|
*
|
|
* @param dev
|
|
* Pointer to struct rte_eth_dev.
|
|
*
|
|
* @return
|
|
* - On success, zero.
|
|
* - On failure, a negative value.
|
|
*/
|
|
static int
|
|
atl_dev_interrupt_action(struct rte_eth_dev *dev,
|
|
struct rte_intr_handle *intr_handle)
|
|
{
|
|
struct atl_interrupt *intr =
|
|
ATL_DEV_PRIVATE_TO_INTR(dev->data->dev_private);
|
|
struct atl_adapter *adapter = dev->data->dev_private;
|
|
struct aq_hw_s *hw = &adapter->hw;
|
|
|
|
if (!(intr->flags & ATL_FLAG_NEED_LINK_UPDATE))
|
|
goto done;
|
|
|
|
intr->flags &= ~ATL_FLAG_NEED_LINK_UPDATE;
|
|
|
|
/* Notify userapp if link status changed */
|
|
if (!atl_dev_link_update(dev, 0)) {
|
|
atl_dev_link_status_print(dev);
|
|
_rte_eth_dev_callback_process(dev,
|
|
RTE_ETH_EVENT_INTR_LSC, NULL);
|
|
} else {
|
|
if (hw->aq_fw_ops->send_macsec_req == NULL)
|
|
goto done;
|
|
|
|
/* Check macsec Keys expired */
|
|
struct get_stats req = { 0 };
|
|
struct macsec_msg_fw_request msg = { 0 };
|
|
struct macsec_msg_fw_response resp = { 0 };
|
|
|
|
req.ingress_sa_index = 0x0;
|
|
req.egress_sc_index = 0x0;
|
|
req.egress_sa_index = 0x0;
|
|
msg.msg_type = macsec_get_stats_msg;
|
|
msg.stats = req;
|
|
|
|
int err = hw->aq_fw_ops->send_macsec_req(hw, &msg, &resp);
|
|
if (err) {
|
|
PMD_DRV_LOG(ERR, "send_macsec_req fail");
|
|
goto done;
|
|
}
|
|
if (resp.stats.egress_threshold_expired ||
|
|
resp.stats.ingress_threshold_expired ||
|
|
resp.stats.egress_expired ||
|
|
resp.stats.ingress_expired) {
|
|
PMD_DRV_LOG(INFO, "RTE_ETH_EVENT_MACSEC");
|
|
_rte_eth_dev_callback_process(dev,
|
|
RTE_ETH_EVENT_MACSEC, NULL);
|
|
}
|
|
}
|
|
done:
|
|
atl_enable_intr(dev);
|
|
rte_intr_ack(intr_handle);
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* Interrupt handler triggered by NIC for handling
|
|
* specific interrupt.
|
|
*
|
|
* @param handle
|
|
* Pointer to interrupt handle.
|
|
* @param param
|
|
* The address of parameter (struct rte_eth_dev *) regsitered before.
|
|
*
|
|
* @return
|
|
* void
|
|
*/
|
|
static void
|
|
atl_dev_interrupt_handler(void *param)
|
|
{
|
|
struct rte_eth_dev *dev = (struct rte_eth_dev *)param;
|
|
|
|
atl_dev_interrupt_get_status(dev);
|
|
atl_dev_interrupt_action(dev, dev->intr_handle);
|
|
}
|
|
|
|
|
|
static int
|
|
atl_dev_get_eeprom_length(struct rte_eth_dev *dev __rte_unused)
|
|
{
|
|
return SFP_EEPROM_SIZE;
|
|
}
|
|
|
|
int atl_dev_get_eeprom(struct rte_eth_dev *dev,
|
|
struct rte_dev_eeprom_info *eeprom)
|
|
{
|
|
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
|
|
uint32_t dev_addr = SMBUS_DEVICE_ID;
|
|
|
|
if (hw->aq_fw_ops->get_eeprom == NULL)
|
|
return -ENOTSUP;
|
|
|
|
if (eeprom->length + eeprom->offset > SFP_EEPROM_SIZE ||
|
|
eeprom->data == NULL)
|
|
return -EINVAL;
|
|
|
|
if (eeprom->magic > 0x7F)
|
|
return -EINVAL;
|
|
|
|
if (eeprom->magic)
|
|
dev_addr = eeprom->magic;
|
|
|
|
return hw->aq_fw_ops->get_eeprom(hw, dev_addr, eeprom->data,
|
|
eeprom->length, eeprom->offset);
|
|
}
|
|
|
|
int atl_dev_set_eeprom(struct rte_eth_dev *dev,
|
|
struct rte_dev_eeprom_info *eeprom)
|
|
{
|
|
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
|
|
uint32_t dev_addr = SMBUS_DEVICE_ID;
|
|
|
|
if (hw->aq_fw_ops->set_eeprom == NULL)
|
|
return -ENOTSUP;
|
|
|
|
if (eeprom->length + eeprom->offset > SFP_EEPROM_SIZE ||
|
|
eeprom->data == NULL)
|
|
return -EINVAL;
|
|
|
|
if (eeprom->magic > 0x7F)
|
|
return -EINVAL;
|
|
|
|
if (eeprom->magic)
|
|
dev_addr = eeprom->magic;
|
|
|
|
return hw->aq_fw_ops->set_eeprom(hw, dev_addr, eeprom->data,
|
|
eeprom->length, eeprom->offset);
|
|
}
|
|
|
|
static int
|
|
atl_dev_get_regs(struct rte_eth_dev *dev, struct rte_dev_reg_info *regs)
|
|
{
|
|
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
|
|
u32 mif_id;
|
|
int err;
|
|
|
|
if (regs->data == NULL) {
|
|
regs->length = hw_atl_utils_hw_get_reg_length();
|
|
regs->width = sizeof(u32);
|
|
return 0;
|
|
}
|
|
|
|
/* Only full register dump is supported */
|
|
if (regs->length && regs->length != hw_atl_utils_hw_get_reg_length())
|
|
return -ENOTSUP;
|
|
|
|
err = hw_atl_utils_hw_get_regs(hw, regs->data);
|
|
|
|
/* Device version */
|
|
mif_id = hw_atl_reg_glb_mif_id_get(hw);
|
|
regs->version = mif_id & 0xFFU;
|
|
|
|
return err;
|
|
}
|
|
|
|
static int
|
|
atl_flow_ctrl_get(struct rte_eth_dev *dev, struct rte_eth_fc_conf *fc_conf)
|
|
{
|
|
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
|
|
u32 fc = AQ_NIC_FC_OFF;
|
|
|
|
if (hw->aq_fw_ops->get_flow_control == NULL)
|
|
return -ENOTSUP;
|
|
|
|
hw->aq_fw_ops->get_flow_control(hw, &fc);
|
|
|
|
if (fc == AQ_NIC_FC_OFF)
|
|
fc_conf->mode = RTE_FC_NONE;
|
|
else if ((fc & AQ_NIC_FC_RX) && (fc & AQ_NIC_FC_TX))
|
|
fc_conf->mode = RTE_FC_FULL;
|
|
else if (fc & AQ_NIC_FC_RX)
|
|
fc_conf->mode = RTE_FC_RX_PAUSE;
|
|
else if (fc & AQ_NIC_FC_TX)
|
|
fc_conf->mode = RTE_FC_TX_PAUSE;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
atl_flow_ctrl_set(struct rte_eth_dev *dev, struct rte_eth_fc_conf *fc_conf)
|
|
{
|
|
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
|
|
uint32_t old_flow_control = hw->aq_nic_cfg->flow_control;
|
|
|
|
|
|
if (hw->aq_fw_ops->set_flow_control == NULL)
|
|
return -ENOTSUP;
|
|
|
|
if (fc_conf->mode == RTE_FC_NONE)
|
|
hw->aq_nic_cfg->flow_control = AQ_NIC_FC_OFF;
|
|
else if (fc_conf->mode == RTE_FC_RX_PAUSE)
|
|
hw->aq_nic_cfg->flow_control = AQ_NIC_FC_RX;
|
|
else if (fc_conf->mode == RTE_FC_TX_PAUSE)
|
|
hw->aq_nic_cfg->flow_control = AQ_NIC_FC_TX;
|
|
else if (fc_conf->mode == RTE_FC_FULL)
|
|
hw->aq_nic_cfg->flow_control = (AQ_NIC_FC_RX | AQ_NIC_FC_TX);
|
|
|
|
if (old_flow_control != hw->aq_nic_cfg->flow_control)
|
|
return hw->aq_fw_ops->set_flow_control(hw);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
atl_update_mac_addr(struct rte_eth_dev *dev, uint32_t index,
|
|
u8 *mac_addr, bool enable)
|
|
{
|
|
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
|
|
unsigned int h = 0U;
|
|
unsigned int l = 0U;
|
|
int err;
|
|
|
|
if (mac_addr) {
|
|
h = (mac_addr[0] << 8) | (mac_addr[1]);
|
|
l = (mac_addr[2] << 24) | (mac_addr[3] << 16) |
|
|
(mac_addr[4] << 8) | mac_addr[5];
|
|
}
|
|
|
|
hw_atl_rpfl2_uc_flr_en_set(hw, 0U, index);
|
|
hw_atl_rpfl2unicast_dest_addresslsw_set(hw, l, index);
|
|
hw_atl_rpfl2unicast_dest_addressmsw_set(hw, h, index);
|
|
|
|
if (enable)
|
|
hw_atl_rpfl2_uc_flr_en_set(hw, 1U, index);
|
|
|
|
err = aq_hw_err_from_flags(hw);
|
|
|
|
return err;
|
|
}
|
|
|
|
static int
|
|
atl_add_mac_addr(struct rte_eth_dev *dev, struct rte_ether_addr *mac_addr,
|
|
uint32_t index __rte_unused, uint32_t pool __rte_unused)
|
|
{
|
|
if (rte_is_zero_ether_addr(mac_addr)) {
|
|
PMD_DRV_LOG(ERR, "Invalid Ethernet Address");
|
|
return -EINVAL;
|
|
}
|
|
|
|
return atl_update_mac_addr(dev, index, (u8 *)mac_addr, true);
|
|
}
|
|
|
|
static void
|
|
atl_remove_mac_addr(struct rte_eth_dev *dev, uint32_t index)
|
|
{
|
|
atl_update_mac_addr(dev, index, NULL, false);
|
|
}
|
|
|
|
static int
|
|
atl_set_default_mac_addr(struct rte_eth_dev *dev, struct rte_ether_addr *addr)
|
|
{
|
|
atl_remove_mac_addr(dev, 0);
|
|
atl_add_mac_addr(dev, addr, 0, 0);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
atl_dev_mtu_set(struct rte_eth_dev *dev, uint16_t mtu)
|
|
{
|
|
struct rte_eth_dev_info dev_info;
|
|
int ret;
|
|
uint32_t frame_size = mtu + RTE_ETHER_HDR_LEN + RTE_ETHER_CRC_LEN;
|
|
|
|
ret = atl_dev_info_get(dev, &dev_info);
|
|
if (ret != 0)
|
|
return ret;
|
|
|
|
if (mtu < RTE_ETHER_MIN_MTU || frame_size > dev_info.max_rx_pktlen)
|
|
return -EINVAL;
|
|
|
|
/* update max frame size */
|
|
dev->data->dev_conf.rxmode.max_rx_pkt_len = frame_size;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
atl_vlan_filter_set(struct rte_eth_dev *dev, uint16_t vlan_id, int on)
|
|
{
|
|
struct aq_hw_cfg_s *cfg =
|
|
ATL_DEV_PRIVATE_TO_CFG(dev->data->dev_private);
|
|
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
|
|
int err = 0;
|
|
int i = 0;
|
|
|
|
PMD_INIT_FUNC_TRACE();
|
|
|
|
for (i = 0; i < HW_ATL_B0_MAX_VLAN_IDS; i++) {
|
|
if (cfg->vlan_filter[i] == vlan_id) {
|
|
if (!on) {
|
|
/* Disable VLAN filter. */
|
|
hw_atl_rpf_vlan_flr_en_set(hw, 0U, i);
|
|
|
|
/* Clear VLAN filter entry */
|
|
cfg->vlan_filter[i] = 0;
|
|
}
|
|
break;
|
|
}
|
|
}
|
|
|
|
/* VLAN_ID was not found. So, nothing to delete. */
|
|
if (i == HW_ATL_B0_MAX_VLAN_IDS && !on)
|
|
goto exit;
|
|
|
|
/* VLAN_ID already exist, or already removed above. Nothing to do. */
|
|
if (i != HW_ATL_B0_MAX_VLAN_IDS)
|
|
goto exit;
|
|
|
|
/* Try to found free VLAN filter to add new VLAN_ID */
|
|
for (i = 0; i < HW_ATL_B0_MAX_VLAN_IDS; i++) {
|
|
if (cfg->vlan_filter[i] == 0)
|
|
break;
|
|
}
|
|
|
|
if (i == HW_ATL_B0_MAX_VLAN_IDS) {
|
|
/* We have no free VLAN filter to add new VLAN_ID*/
|
|
err = -ENOMEM;
|
|
goto exit;
|
|
}
|
|
|
|
cfg->vlan_filter[i] = vlan_id;
|
|
hw_atl_rpf_vlan_flr_act_set(hw, 1U, i);
|
|
hw_atl_rpf_vlan_id_flr_set(hw, vlan_id, i);
|
|
hw_atl_rpf_vlan_flr_en_set(hw, 1U, i);
|
|
|
|
exit:
|
|
/* Enable VLAN promisc mode if vlan_filter empty */
|
|
for (i = 0; i < HW_ATL_B0_MAX_VLAN_IDS; i++) {
|
|
if (cfg->vlan_filter[i] != 0)
|
|
break;
|
|
}
|
|
|
|
hw_atl_rpf_vlan_prom_mode_en_set(hw, i == HW_ATL_B0_MAX_VLAN_IDS);
|
|
|
|
return err;
|
|
}
|
|
|
|
static int
|
|
atl_enable_vlan_filter(struct rte_eth_dev *dev, int en)
|
|
{
|
|
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
|
|
struct aq_hw_cfg_s *cfg =
|
|
ATL_DEV_PRIVATE_TO_CFG(dev->data->dev_private);
|
|
int i;
|
|
|
|
PMD_INIT_FUNC_TRACE();
|
|
|
|
for (i = 0; i < HW_ATL_B0_MAX_VLAN_IDS; i++) {
|
|
if (cfg->vlan_filter[i])
|
|
hw_atl_rpf_vlan_flr_en_set(hw, en, i);
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
atl_vlan_offload_set(struct rte_eth_dev *dev, int mask)
|
|
{
|
|
struct aq_hw_cfg_s *cfg =
|
|
ATL_DEV_PRIVATE_TO_CFG(dev->data->dev_private);
|
|
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
|
|
int ret = 0;
|
|
int i;
|
|
|
|
PMD_INIT_FUNC_TRACE();
|
|
|
|
ret = atl_enable_vlan_filter(dev, mask & ETH_VLAN_FILTER_MASK);
|
|
|
|
cfg->vlan_strip = !!(mask & ETH_VLAN_STRIP_MASK);
|
|
|
|
for (i = 0; i < dev->data->nb_rx_queues; i++)
|
|
hw_atl_rpo_rx_desc_vlan_stripping_set(hw, cfg->vlan_strip, i);
|
|
|
|
if (mask & ETH_VLAN_EXTEND_MASK)
|
|
ret = -ENOTSUP;
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int
|
|
atl_vlan_tpid_set(struct rte_eth_dev *dev, enum rte_vlan_type vlan_type,
|
|
uint16_t tpid)
|
|
{
|
|
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
|
|
int err = 0;
|
|
|
|
PMD_INIT_FUNC_TRACE();
|
|
|
|
switch (vlan_type) {
|
|
case ETH_VLAN_TYPE_INNER:
|
|
hw_atl_rpf_vlan_inner_etht_set(hw, tpid);
|
|
break;
|
|
case ETH_VLAN_TYPE_OUTER:
|
|
hw_atl_rpf_vlan_outer_etht_set(hw, tpid);
|
|
break;
|
|
default:
|
|
PMD_DRV_LOG(ERR, "Unsupported VLAN type");
|
|
err = -ENOTSUP;
|
|
}
|
|
|
|
return err;
|
|
}
|
|
|
|
static void
|
|
atl_vlan_strip_queue_set(struct rte_eth_dev *dev, uint16_t queue_id, int on)
|
|
{
|
|
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
|
|
|
|
PMD_INIT_FUNC_TRACE();
|
|
|
|
if (queue_id > dev->data->nb_rx_queues) {
|
|
PMD_DRV_LOG(ERR, "Invalid queue id");
|
|
return;
|
|
}
|
|
|
|
hw_atl_rpo_rx_desc_vlan_stripping_set(hw, on, queue_id);
|
|
}
|
|
|
|
static int
|
|
atl_dev_set_mc_addr_list(struct rte_eth_dev *dev,
|
|
struct rte_ether_addr *mc_addr_set,
|
|
uint32_t nb_mc_addr)
|
|
{
|
|
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
|
|
u32 i;
|
|
|
|
if (nb_mc_addr > AQ_HW_MULTICAST_ADDRESS_MAX - HW_ATL_B0_MAC_MIN)
|
|
return -EINVAL;
|
|
|
|
/* Update whole uc filters table */
|
|
for (i = 0; i < AQ_HW_MULTICAST_ADDRESS_MAX - HW_ATL_B0_MAC_MIN; i++) {
|
|
u8 *mac_addr = NULL;
|
|
u32 l = 0, h = 0;
|
|
|
|
if (i < nb_mc_addr) {
|
|
mac_addr = mc_addr_set[i].addr_bytes;
|
|
l = (mac_addr[2] << 24) | (mac_addr[3] << 16) |
|
|
(mac_addr[4] << 8) | mac_addr[5];
|
|
h = (mac_addr[0] << 8) | mac_addr[1];
|
|
}
|
|
|
|
hw_atl_rpfl2_uc_flr_en_set(hw, 0U, HW_ATL_B0_MAC_MIN + i);
|
|
hw_atl_rpfl2unicast_dest_addresslsw_set(hw, l,
|
|
HW_ATL_B0_MAC_MIN + i);
|
|
hw_atl_rpfl2unicast_dest_addressmsw_set(hw, h,
|
|
HW_ATL_B0_MAC_MIN + i);
|
|
hw_atl_rpfl2_uc_flr_en_set(hw, !!mac_addr,
|
|
HW_ATL_B0_MAC_MIN + i);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
atl_reta_update(struct rte_eth_dev *dev,
|
|
struct rte_eth_rss_reta_entry64 *reta_conf,
|
|
uint16_t reta_size)
|
|
{
|
|
int i;
|
|
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
|
|
struct aq_hw_cfg_s *cf = ATL_DEV_PRIVATE_TO_CFG(dev->data->dev_private);
|
|
|
|
for (i = 0; i < reta_size && i < cf->aq_rss.indirection_table_size; i++)
|
|
cf->aq_rss.indirection_table[i] = min(reta_conf->reta[i],
|
|
dev->data->nb_rx_queues - 1);
|
|
|
|
hw_atl_b0_hw_rss_set(hw, &cf->aq_rss);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
atl_reta_query(struct rte_eth_dev *dev,
|
|
struct rte_eth_rss_reta_entry64 *reta_conf,
|
|
uint16_t reta_size)
|
|
{
|
|
int i;
|
|
struct aq_hw_cfg_s *cf = ATL_DEV_PRIVATE_TO_CFG(dev->data->dev_private);
|
|
|
|
for (i = 0; i < reta_size && i < cf->aq_rss.indirection_table_size; i++)
|
|
reta_conf->reta[i] = cf->aq_rss.indirection_table[i];
|
|
reta_conf->mask = ~0U;
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
atl_rss_hash_update(struct rte_eth_dev *dev,
|
|
struct rte_eth_rss_conf *rss_conf)
|
|
{
|
|
struct aq_hw_s *hw = ATL_DEV_PRIVATE_TO_HW(dev->data->dev_private);
|
|
struct aq_hw_cfg_s *cfg =
|
|
ATL_DEV_PRIVATE_TO_CFG(dev->data->dev_private);
|
|
static u8 def_rss_key[40] = {
|
|
0x1e, 0xad, 0x71, 0x87, 0x65, 0xfc, 0x26, 0x7d,
|
|
0x0d, 0x45, 0x67, 0x74, 0xcd, 0x06, 0x1a, 0x18,
|
|
0xb6, 0xc1, 0xf0, 0xc7, 0xbb, 0x18, 0xbe, 0xf8,
|
|
0x19, 0x13, 0x4b, 0xa9, 0xd0, 0x3e, 0xfe, 0x70,
|
|
0x25, 0x03, 0xab, 0x50, 0x6a, 0x8b, 0x82, 0x0c
|
|
};
|
|
|
|
cfg->is_rss = !!rss_conf->rss_hf;
|
|
if (rss_conf->rss_key) {
|
|
memcpy(cfg->aq_rss.hash_secret_key, rss_conf->rss_key,
|
|
rss_conf->rss_key_len);
|
|
cfg->aq_rss.hash_secret_key_size = rss_conf->rss_key_len;
|
|
} else {
|
|
memcpy(cfg->aq_rss.hash_secret_key, def_rss_key,
|
|
sizeof(def_rss_key));
|
|
cfg->aq_rss.hash_secret_key_size = sizeof(def_rss_key);
|
|
}
|
|
|
|
hw_atl_b0_hw_rss_set(hw, &cfg->aq_rss);
|
|
hw_atl_b0_hw_rss_hash_set(hw, &cfg->aq_rss);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
atl_rss_hash_conf_get(struct rte_eth_dev *dev,
|
|
struct rte_eth_rss_conf *rss_conf)
|
|
{
|
|
struct aq_hw_cfg_s *cfg =
|
|
ATL_DEV_PRIVATE_TO_CFG(dev->data->dev_private);
|
|
|
|
rss_conf->rss_hf = cfg->is_rss ? ATL_RSS_OFFLOAD_ALL : 0;
|
|
if (rss_conf->rss_key) {
|
|
rss_conf->rss_key_len = cfg->aq_rss.hash_secret_key_size;
|
|
memcpy(rss_conf->rss_key, cfg->aq_rss.hash_secret_key,
|
|
rss_conf->rss_key_len);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static bool
|
|
is_device_supported(struct rte_eth_dev *dev, struct rte_pci_driver *drv)
|
|
{
|
|
if (strcmp(dev->device->driver->name, drv->driver.name))
|
|
return false;
|
|
|
|
return true;
|
|
}
|
|
|
|
bool
|
|
is_atlantic_supported(struct rte_eth_dev *dev)
|
|
{
|
|
return is_device_supported(dev, &rte_atl_pmd);
|
|
}
|
|
|
|
RTE_PMD_REGISTER_PCI(net_atlantic, rte_atl_pmd);
|
|
RTE_PMD_REGISTER_PCI_TABLE(net_atlantic, pci_id_atl_map);
|
|
RTE_PMD_REGISTER_KMOD_DEP(net_atlantic, "* igb_uio | uio_pci_generic");
|
|
RTE_LOG_REGISTER(atl_logtype_init, pmd.net.atlantic.init, NOTICE);
|
|
RTE_LOG_REGISTER(atl_logtype_driver, pmd.net.atlantic.driver, NOTICE);
|