numam-dpdk/drivers/net/ice/ice_ethdev.h
Olivier Matz 6d13ea8e8e net: add rte prefix to ether structures
Add 'rte_' prefix to structures:
- rename struct ether_addr as struct rte_ether_addr.
- rename struct ether_hdr as struct rte_ether_hdr.
- rename struct vlan_hdr as struct rte_vlan_hdr.
- rename struct vxlan_hdr as struct rte_vxlan_hdr.
- rename struct vxlan_gpe_hdr as struct rte_vxlan_gpe_hdr.

Do not update the command line library to avoid adding a dependency to
librte_net.

Signed-off-by: Olivier Matz <olivier.matz@6wind.com>
Reviewed-by: Stephen Hemminger <stephen@networkplumber.org>
Reviewed-by: Maxime Coquelin <maxime.coquelin@redhat.com>
Reviewed-by: Ferruh Yigit <ferruh.yigit@intel.com>
2019-05-24 13:34:45 +02:00

369 lines
12 KiB
C

/* SPDX-License-Identifier: BSD-3-Clause
* Copyright(c) 2018 Intel Corporation
*/
#ifndef _ICE_ETHDEV_H_
#define _ICE_ETHDEV_H_
#include <rte_kvargs.h>
#include <rte_ethdev_driver.h>
#include "base/ice_common.h"
#include "base/ice_adminq_cmd.h"
#define ICE_VLAN_TAG_SIZE 4
#define ICE_ADMINQ_LEN 32
#define ICE_SBIOQ_LEN 32
#define ICE_MAILBOXQ_LEN 32
#define ICE_ADMINQ_BUF_SZ 4096
#define ICE_SBIOQ_BUF_SZ 4096
#define ICE_MAILBOXQ_BUF_SZ 4096
/* Number of queues per TC should be one of 1, 2, 4, 8, 16, 32, 64 */
#define ICE_MAX_Q_PER_TC 64
#define ICE_NUM_DESC_DEFAULT 512
#define ICE_BUF_SIZE_MIN 1024
#define ICE_FRAME_SIZE_MAX 9728
#define ICE_QUEUE_BASE_ADDR_UNIT 128
/* number of VSIs and queue default setting */
#define ICE_MAX_QP_NUM_PER_VF 16
#define ICE_DEFAULT_QP_NUM_FDIR 1
#define ICE_UINT32_BIT_SIZE (CHAR_BIT * sizeof(uint32_t))
#define ICE_VFTA_SIZE (4096 / ICE_UINT32_BIT_SIZE)
/* Maximun number of MAC addresses */
#define ICE_NUM_MACADDR_MAX 64
/* Maximum number of VFs */
#define ICE_MAX_VF 128
#define ICE_MAX_INTR_QUEUE_NUM 256
#define ICE_MISC_VEC_ID RTE_INTR_VEC_ZERO_OFFSET
#define ICE_RX_VEC_ID RTE_INTR_VEC_RXTX_OFFSET
#define ICE_MAX_PKT_TYPE 1024
/**
* vlan_id is a 12 bit number.
* The VFTA array is actually a 4096 bit array, 128 of 32bit elements.
* 2^5 = 32. The val of lower 5 bits specifies the bit in the 32bit element.
* The higher 7 bit val specifies VFTA array index.
*/
#define ICE_VFTA_BIT(vlan_id) (1 << ((vlan_id) & 0x1F))
#define ICE_VFTA_IDX(vlan_id) ((vlan_id) >> 5)
/* Default TC traffic in case DCB is not enabled */
#define ICE_DEFAULT_TCMAP 0x1
#define ICE_FDIR_QUEUE_ID 0
/* Always assign pool 0 to main VSI, VMDQ will start from 1 */
#define ICE_VMDQ_POOL_BASE 1
#define ICE_DEFAULT_RX_FREE_THRESH 32
#define ICE_DEFAULT_RX_PTHRESH 8
#define ICE_DEFAULT_RX_HTHRESH 8
#define ICE_DEFAULT_RX_WTHRESH 0
#define ICE_DEFAULT_TX_FREE_THRESH 32
#define ICE_DEFAULT_TX_PTHRESH 32
#define ICE_DEFAULT_TX_HTHRESH 0
#define ICE_DEFAULT_TX_WTHRESH 0
#define ICE_DEFAULT_TX_RSBIT_THRESH 32
/* Bit shift and mask */
#define ICE_4_BIT_WIDTH (CHAR_BIT / 2)
#define ICE_4_BIT_MASK RTE_LEN2MASK(ICE_4_BIT_WIDTH, uint8_t)
#define ICE_8_BIT_WIDTH CHAR_BIT
#define ICE_8_BIT_MASK UINT8_MAX
#define ICE_16_BIT_WIDTH (CHAR_BIT * 2)
#define ICE_16_BIT_MASK UINT16_MAX
#define ICE_32_BIT_WIDTH (CHAR_BIT * 4)
#define ICE_32_BIT_MASK UINT32_MAX
#define ICE_40_BIT_WIDTH (CHAR_BIT * 5)
#define ICE_40_BIT_MASK RTE_LEN2MASK(ICE_40_BIT_WIDTH, uint64_t)
#define ICE_48_BIT_WIDTH (CHAR_BIT * 6)
#define ICE_48_BIT_MASK RTE_LEN2MASK(ICE_48_BIT_WIDTH, uint64_t)
#define ICE_FLAG_RSS BIT_ULL(0)
#define ICE_FLAG_DCB BIT_ULL(1)
#define ICE_FLAG_VMDQ BIT_ULL(2)
#define ICE_FLAG_SRIOV BIT_ULL(3)
#define ICE_FLAG_HEADER_SPLIT_DISABLED BIT_ULL(4)
#define ICE_FLAG_HEADER_SPLIT_ENABLED BIT_ULL(5)
#define ICE_FLAG_FDIR BIT_ULL(6)
#define ICE_FLAG_VXLAN BIT_ULL(7)
#define ICE_FLAG_RSS_AQ_CAPABLE BIT_ULL(8)
#define ICE_FLAG_VF_MAC_BY_PF BIT_ULL(9)
#define ICE_FLAG_ALL (ICE_FLAG_RSS | \
ICE_FLAG_DCB | \
ICE_FLAG_VMDQ | \
ICE_FLAG_SRIOV | \
ICE_FLAG_HEADER_SPLIT_DISABLED | \
ICE_FLAG_HEADER_SPLIT_ENABLED | \
ICE_FLAG_FDIR | \
ICE_FLAG_VXLAN | \
ICE_FLAG_RSS_AQ_CAPABLE | \
ICE_FLAG_VF_MAC_BY_PF)
#define ICE_RSS_OFFLOAD_ALL ( \
ETH_RSS_FRAG_IPV4 | \
ETH_RSS_NONFRAG_IPV4_TCP | \
ETH_RSS_NONFRAG_IPV4_UDP | \
ETH_RSS_NONFRAG_IPV4_SCTP | \
ETH_RSS_NONFRAG_IPV4_OTHER | \
ETH_RSS_FRAG_IPV6 | \
ETH_RSS_NONFRAG_IPV6_TCP | \
ETH_RSS_NONFRAG_IPV6_UDP | \
ETH_RSS_NONFRAG_IPV6_SCTP | \
ETH_RSS_NONFRAG_IPV6_OTHER | \
ETH_RSS_L2_PAYLOAD)
/**
* The overhead from MTU to max frame size.
* Considering QinQ packet, the VLAN tag needs to be counted twice.
*/
#define ICE_ETH_OVERHEAD \
(ETHER_HDR_LEN + ETHER_CRC_LEN + ICE_VLAN_TAG_SIZE * 2)
struct ice_adapter;
/**
* MAC filter structure
*/
struct ice_mac_filter_info {
struct rte_ether_addr mac_addr;
};
TAILQ_HEAD(ice_mac_filter_list, ice_mac_filter);
/* MAC filter list structure */
struct ice_mac_filter {
TAILQ_ENTRY(ice_mac_filter) next;
struct ice_mac_filter_info mac_info;
};
/**
* VLAN filter structure
*/
struct ice_vlan_filter_info {
uint16_t vlan_id;
};
TAILQ_HEAD(ice_vlan_filter_list, ice_vlan_filter);
/* VLAN filter list structure */
struct ice_vlan_filter {
TAILQ_ENTRY(ice_vlan_filter) next;
struct ice_vlan_filter_info vlan_info;
};
struct pool_entry {
LIST_ENTRY(pool_entry) next;
uint16_t base;
uint16_t len;
};
LIST_HEAD(res_list, pool_entry);
struct ice_res_pool_info {
uint32_t base; /* Resource start index */
uint32_t num_alloc; /* Allocated resource number */
uint32_t num_free; /* Total available resource number */
struct res_list alloc_list; /* Allocated resource list */
struct res_list free_list; /* Available resource list */
};
TAILQ_HEAD(ice_vsi_list_head, ice_vsi_list);
struct ice_vsi;
/* VSI list structure */
struct ice_vsi_list {
TAILQ_ENTRY(ice_vsi_list) list;
struct ice_vsi *vsi;
};
struct ice_rx_queue;
struct ice_tx_queue;
/**
* Structure that defines a VSI, associated with a adapter.
*/
struct ice_vsi {
struct ice_adapter *adapter; /* Backreference to associated adapter */
struct ice_aqc_vsi_props info; /* VSI properties */
/**
* When drivers loaded, only a default main VSI exists. In case new VSI
* needs to add, HW needs to know the layout that VSIs are organized.
* Besides that, VSI isan element and can't switch packets, which needs
* to add new component VEB to perform switching. So, a new VSI needs
* to specify the the uplink VSI (Parent VSI) before created. The
* uplink VSI will check whether it had a VEB to switch packets. If no,
* it will try to create one. Then, uplink VSI will move the new VSI
* into its' sib_vsi_list to manage all the downlink VSI.
* sib_vsi_list: the VSI list that shared the same uplink VSI.
* parent_vsi : the uplink VSI. It's NULL for main VSI.
* veb : the VEB associates with the VSI.
*/
struct ice_vsi_list sib_vsi_list; /* sibling vsi list */
struct ice_vsi *parent_vsi;
enum ice_vsi_type type; /* VSI types */
uint16_t vlan_num; /* Total VLAN number */
uint16_t mac_num; /* Total mac number */
struct ice_mac_filter_list mac_list; /* macvlan filter list */
struct ice_vlan_filter_list vlan_list; /* vlan filter list */
uint16_t nb_qps; /* Number of queue pairs VSI can occupy */
uint16_t nb_used_qps; /* Number of queue pairs VSI uses */
uint16_t max_macaddrs; /* Maximum number of MAC addresses */
uint16_t base_queue; /* The first queue index of this VSI */
uint16_t vsi_id; /* Hardware Id */
uint16_t idx; /* vsi_handle: SW index in hw->vsi_ctx */
/* VF number to which the VSI connects, valid when VSI is VF type */
uint8_t vf_num;
uint16_t msix_intr; /* The MSIX interrupt binds to VSI */
uint16_t nb_msix; /* The max number of msix vector */
uint8_t enabled_tc; /* The traffic class enabled */
uint8_t vlan_anti_spoof_on; /* The VLAN anti-spoofing enabled */
uint8_t vlan_filter_on; /* The VLAN filter enabled */
/* information about rss configuration */
u32 rss_key_size;
u32 rss_lut_size;
uint8_t *rss_lut;
uint8_t *rss_key;
struct ice_eth_stats eth_stats_offset;
struct ice_eth_stats eth_stats;
bool offset_loaded;
};
struct ice_pf {
struct ice_adapter *adapter; /* The adapter this PF associate to */
struct ice_vsi *main_vsi; /* pointer to main VSI structure */
/* Used for next free software vsi idx.
* To save the effort, we don't recycle the index.
* Suppose the indexes are more than enough.
*/
uint16_t next_vsi_idx;
uint16_t vsis_allocated;
uint16_t vsis_unallocated;
struct ice_res_pool_info qp_pool; /*Queue pair pool */
struct ice_res_pool_info msix_pool; /* MSIX interrupt pool */
struct rte_eth_dev_data *dev_data; /* Pointer to the device data */
struct rte_ether_addr dev_addr; /* PF device mac address */
uint64_t flags; /* PF feature flags */
uint16_t hash_lut_size; /* The size of hash lookup table */
uint16_t lan_nb_qp_max;
uint16_t lan_nb_qps; /* The number of queue pairs of LAN */
struct ice_hw_port_stats stats_offset;
struct ice_hw_port_stats stats;
/* internal packet statistics, it should be excluded from the total */
struct ice_eth_stats internal_stats_offset;
struct ice_eth_stats internal_stats;
bool offset_loaded;
bool adapter_stopped;
};
/**
* Structure to store private data for each PF/VF instance.
*/
struct ice_adapter {
/* Common for both PF and VF */
struct ice_hw hw;
struct rte_eth_dev *eth_dev;
struct ice_pf pf;
bool rx_bulk_alloc_allowed;
bool tx_simple_allowed;
/* ptype mapping table */
uint32_t ptype_tbl[ICE_MAX_PKT_TYPE] __rte_cache_min_aligned;
bool is_safe_mode;
};
struct ice_vsi_vlan_pvid_info {
uint16_t on; /* Enable or disable pvid */
union {
uint16_t pvid; /* Valid in case 'on' is set to set pvid */
struct {
/* Valid in case 'on' is cleared. 'tagged' will reject
* tagged packets, while 'untagged' will reject
* untagged packets.
*/
uint8_t tagged;
uint8_t untagged;
} reject;
} config;
};
#define ICE_DEV_TO_PCI(eth_dev) \
RTE_DEV_TO_PCI((eth_dev)->device)
/* ICE_DEV_PRIVATE_TO */
#define ICE_DEV_PRIVATE_TO_PF(adapter) \
(&((struct ice_adapter *)adapter)->pf)
#define ICE_DEV_PRIVATE_TO_HW(adapter) \
(&((struct ice_adapter *)adapter)->hw)
#define ICE_DEV_PRIVATE_TO_ADAPTER(adapter) \
((struct ice_adapter *)adapter)
/* ICE_VSI_TO */
#define ICE_VSI_TO_HW(vsi) \
(&(((struct ice_vsi *)vsi)->adapter->hw))
#define ICE_VSI_TO_PF(vsi) \
(&(((struct ice_vsi *)vsi)->adapter->pf))
#define ICE_VSI_TO_ETH_DEV(vsi) \
(((struct ice_vsi *)vsi)->adapter->eth_dev)
/* ICE_PF_TO */
#define ICE_PF_TO_HW(pf) \
(&(((struct ice_pf *)pf)->adapter->hw))
#define ICE_PF_TO_ADAPTER(pf) \
((struct ice_adapter *)(pf)->adapter)
#define ICE_PF_TO_ETH_DEV(pf) \
(((struct ice_pf *)pf)->adapter->eth_dev)
static inline int
ice_align_floor(int n)
{
if (n == 0)
return 0;
return 1 << (sizeof(n) * CHAR_BIT - 1 - __builtin_clz(n));
}
#define ICE_PHY_TYPE_SUPPORT_50G(phy_type) \
(((phy_type) & ICE_PHY_TYPE_LOW_50GBASE_CR2) || \
((phy_type) & ICE_PHY_TYPE_LOW_50GBASE_SR2) || \
((phy_type) & ICE_PHY_TYPE_LOW_50GBASE_LR2) || \
((phy_type) & ICE_PHY_TYPE_LOW_50GBASE_KR2) || \
((phy_type) & ICE_PHY_TYPE_LOW_50G_LAUI2_AOC_ACC) || \
((phy_type) & ICE_PHY_TYPE_LOW_50G_LAUI2) || \
((phy_type) & ICE_PHY_TYPE_LOW_50G_AUI2_AOC_ACC) || \
((phy_type) & ICE_PHY_TYPE_LOW_50G_AUI2) || \
((phy_type) & ICE_PHY_TYPE_LOW_50GBASE_CP) || \
((phy_type) & ICE_PHY_TYPE_LOW_50GBASE_SR) || \
((phy_type) & ICE_PHY_TYPE_LOW_50GBASE_FR) || \
((phy_type) & ICE_PHY_TYPE_LOW_50GBASE_LR) || \
((phy_type) & ICE_PHY_TYPE_LOW_50GBASE_KR_PAM4) || \
((phy_type) & ICE_PHY_TYPE_LOW_50G_AUI1_AOC_ACC) || \
((phy_type) & ICE_PHY_TYPE_LOW_50G_AUI1))
#define ICE_PHY_TYPE_SUPPORT_100G_LOW(phy_type) \
(((phy_type) & ICE_PHY_TYPE_LOW_100GBASE_CR4) || \
((phy_type) & ICE_PHY_TYPE_LOW_100GBASE_SR4) || \
((phy_type) & ICE_PHY_TYPE_LOW_100GBASE_LR4) || \
((phy_type) & ICE_PHY_TYPE_LOW_100GBASE_KR4) || \
((phy_type) & ICE_PHY_TYPE_LOW_100G_CAUI4_AOC_ACC) || \
((phy_type) & ICE_PHY_TYPE_LOW_100G_CAUI4) || \
((phy_type) & ICE_PHY_TYPE_LOW_100G_AUI4_AOC_ACC) || \
((phy_type) & ICE_PHY_TYPE_LOW_100G_AUI4) || \
((phy_type) & ICE_PHY_TYPE_LOW_100GBASE_CR_PAM4) || \
((phy_type) & ICE_PHY_TYPE_LOW_100GBASE_KR_PAM4) || \
((phy_type) & ICE_PHY_TYPE_LOW_100GBASE_CP2) || \
((phy_type) & ICE_PHY_TYPE_LOW_100GBASE_SR2) || \
((phy_type) & ICE_PHY_TYPE_LOW_100GBASE_DR))
#define ICE_PHY_TYPE_SUPPORT_100G_HIGH(phy_type) \
(((phy_type) & ICE_PHY_TYPE_HIGH_100GBASE_KR2_PAM4) || \
((phy_type) & ICE_PHY_TYPE_HIGH_100G_CAUI2_AOC_ACC) || \
((phy_type) & ICE_PHY_TYPE_HIGH_100G_CAUI2) || \
((phy_type) & ICE_PHY_TYPE_HIGH_100G_AUI2_AOC_ACC) || \
((phy_type) & ICE_PHY_TYPE_HIGH_100G_AUI2))
#endif /* _ICE_ETHDEV_H_ */