/* SPDX-License-Identifier: BSD-3-Clause * Copyright (c) 2016 - 2018 Cavium Inc. * All rights reserved. * www.cavium.com */ #ifndef _QEDE_RXTX_H_ #define _QEDE_RXTX_H_ #include "qede_ethdev.h" /* Ring Descriptors */ #define RX_RING_SIZE_POW 16 /* 64K */ #define RX_RING_SIZE (1ULL << RX_RING_SIZE_POW) #define NUM_RX_BDS_MAX (RX_RING_SIZE - 1) #define NUM_RX_BDS_MIN 128 #define NUM_RX_BDS_DEF NUM_RX_BDS_MAX #define NUM_RX_BDS(q) (q->nb_rx_desc - 1) #define TX_RING_SIZE_POW 16 /* 64K */ #define TX_RING_SIZE (1ULL << TX_RING_SIZE_POW) #define NUM_TX_BDS_MAX (TX_RING_SIZE - 1) #define NUM_TX_BDS_MIN 128 #define NUM_TX_BDS_DEF NUM_TX_BDS_MAX #define NUM_TX_BDS(q) (q->nb_tx_desc - 1) #define TX_CONS(txq) (txq->sw_tx_cons & NUM_TX_BDS(txq)) #define TX_PROD(txq) (txq->sw_tx_prod & NUM_TX_BDS(txq)) #define QEDE_DEFAULT_TX_FREE_THRESH 32 #define QEDE_CSUM_ERROR (1 << 0) #define QEDE_CSUM_UNNECESSARY (1 << 1) #define QEDE_TUNN_CSUM_UNNECESSARY (1 << 2) #define QEDE_BD_SET_ADDR_LEN(bd, maddr, len) \ do { \ (bd)->addr.hi = rte_cpu_to_le_32(U64_HI(maddr)); \ (bd)->addr.lo = rte_cpu_to_le_32(U64_LO(maddr)); \ (bd)->nbytes = rte_cpu_to_le_16(len); \ } while (0) #define CQE_HAS_VLAN(flags) \ ((flags) & (PARSING_AND_ERR_FLAGS_TAG8021QEXIST_MASK \ << PARSING_AND_ERR_FLAGS_TAG8021QEXIST_SHIFT)) #define CQE_HAS_OUTER_VLAN(flags) \ ((flags) & (PARSING_AND_ERR_FLAGS_TUNNEL8021QTAGEXIST_MASK \ << PARSING_AND_ERR_FLAGS_TUNNEL8021QTAGEXIST_SHIFT)) #define QEDE_MIN_RX_BUFF_SIZE (1024) #define QEDE_VLAN_TAG_SIZE (4) #define QEDE_LLC_SNAP_HDR_LEN (8) /* Max supported alignment is 256 (8 shift) * minimal alignment shift 6 is optimal for 57xxx HW performance */ #define QEDE_L1_CACHE_SHIFT 6 #define QEDE_RX_ALIGN_SHIFT (RTE_MAX(6, RTE_MIN(8, QEDE_L1_CACHE_SHIFT))) #define QEDE_FW_RX_ALIGN_END (1UL << QEDE_RX_ALIGN_SHIFT) #define QEDE_CEIL_TO_CACHE_LINE_SIZE(n) (((n) + (QEDE_FW_RX_ALIGN_END - 1)) & \ ~(QEDE_FW_RX_ALIGN_END - 1)) #define QEDE_FLOOR_TO_CACHE_LINE_SIZE(n) RTE_ALIGN_FLOOR(n, \ QEDE_FW_RX_ALIGN_END) /* Note: QEDE_LLC_SNAP_HDR_LEN is optional, * +2 is for padding in front of L2 header */ #define QEDE_ETH_OVERHEAD (((2 * QEDE_VLAN_TAG_SIZE)) \ + (QEDE_LLC_SNAP_HDR_LEN) + 2) #define QEDE_MAX_ETHER_HDR_LEN (RTE_ETHER_HDR_LEN + QEDE_ETH_OVERHEAD) #define QEDE_ETH_MAX_LEN (RTE_ETHER_MTU + QEDE_MAX_ETHER_HDR_LEN) #define QEDE_RSS_OFFLOAD_ALL (ETH_RSS_IPV4 |\ ETH_RSS_NONFRAG_IPV4_TCP |\ ETH_RSS_NONFRAG_IPV4_UDP |\ ETH_RSS_IPV6 |\ ETH_RSS_NONFRAG_IPV6_TCP |\ ETH_RSS_NONFRAG_IPV6_UDP |\ ETH_RSS_VXLAN |\ ETH_RSS_GENEVE) #define QEDE_RXTX_MAX(qdev) \ (RTE_MAX(qdev->num_rx_queues, qdev->num_tx_queues)) /* Macros for non-tunnel packet types lkup table */ #define QEDE_PKT_TYPE_UNKNOWN 0x0 #define QEDE_PKT_TYPE_MAX 0x3f #define QEDE_PKT_TYPE_IPV4 0x1 #define QEDE_PKT_TYPE_IPV6 0x2 #define QEDE_PKT_TYPE_IPV4_TCP 0x5 #define QEDE_PKT_TYPE_IPV6_TCP 0x6 #define QEDE_PKT_TYPE_IPV4_UDP 0x9 #define QEDE_PKT_TYPE_IPV6_UDP 0xa /* For frag pkts, corresponding IP bits is set */ #define QEDE_PKT_TYPE_IPV4_FRAG 0x11 #define QEDE_PKT_TYPE_IPV6_FRAG 0x12 #define QEDE_PKT_TYPE_IPV4_VLAN 0x21 #define QEDE_PKT_TYPE_IPV6_VLAN 0x22 #define QEDE_PKT_TYPE_IPV4_TCP_VLAN 0x25 #define QEDE_PKT_TYPE_IPV6_TCP_VLAN 0x26 #define QEDE_PKT_TYPE_IPV4_UDP_VLAN 0x29 #define QEDE_PKT_TYPE_IPV6_UDP_VLAN 0x2a #define QEDE_PKT_TYPE_IPV4_VLAN_FRAG 0x31 #define QEDE_PKT_TYPE_IPV6_VLAN_FRAG 0x32 /* Macros for tunneled packets with next protocol lkup table */ #define QEDE_PKT_TYPE_TUNN_GENEVE 0x1 #define QEDE_PKT_TYPE_TUNN_GRE 0x2 #define QEDE_PKT_TYPE_TUNN_VXLAN 0x3 /* Bit 2 is don't care bit */ #define QEDE_PKT_TYPE_TUNN_L2_TENID_NOEXIST_GENEVE 0x9 #define QEDE_PKT_TYPE_TUNN_L2_TENID_NOEXIST_GRE 0xa #define QEDE_PKT_TYPE_TUNN_L2_TENID_NOEXIST_VXLAN 0xb #define QEDE_PKT_TYPE_TUNN_L2_TENID_EXIST_GENEVE 0xd #define QEDE_PKT_TYPE_TUNN_L2_TENID_EXIST_GRE 0xe #define QEDE_PKT_TYPE_TUNN_L2_TENID_EXIST_VXLAN 0xf #define QEDE_PKT_TYPE_TUNN_IPV4_TENID_NOEXIST_GENEVE 0x11 #define QEDE_PKT_TYPE_TUNN_IPV4_TENID_NOEXIST_GRE 0x12 #define QEDE_PKT_TYPE_TUNN_IPV4_TENID_NOEXIST_VXLAN 0x13 #define QEDE_PKT_TYPE_TUNN_IPV4_TENID_EXIST_GENEVE 0x15 #define QEDE_PKT_TYPE_TUNN_IPV4_TENID_EXIST_GRE 0x16 #define QEDE_PKT_TYPE_TUNN_IPV4_TENID_EXIST_VXLAN 0x17 #define QEDE_PKT_TYPE_TUNN_IPV6_TENID_NOEXIST_GENEVE 0x19 #define QEDE_PKT_TYPE_TUNN_IPV6_TENID_NOEXIST_GRE 0x1a #define QEDE_PKT_TYPE_TUNN_IPV6_TENID_NOEXIST_VXLAN 0x1b #define QEDE_PKT_TYPE_TUNN_IPV6_TENID_EXIST_GENEVE 0x1d #define QEDE_PKT_TYPE_TUNN_IPV6_TENID_EXIST_GRE 0x1e #define QEDE_PKT_TYPE_TUNN_IPV6_TENID_EXIST_VXLAN 0x1f #define QEDE_PKT_TYPE_TUNN_MAX_TYPE 0x20 /* 2^5 */ #define QEDE_TX_CSUM_OFFLOAD_MASK (PKT_TX_IP_CKSUM | \ PKT_TX_TCP_CKSUM | \ PKT_TX_UDP_CKSUM | \ PKT_TX_OUTER_IP_CKSUM | \ PKT_TX_TCP_SEG | \ PKT_TX_IPV4 | \ PKT_TX_IPV6) #define QEDE_TX_OFFLOAD_MASK (QEDE_TX_CSUM_OFFLOAD_MASK | \ PKT_TX_VLAN_PKT | \ PKT_TX_TUNNEL_MASK) #define QEDE_TX_OFFLOAD_NOTSUP_MASK \ (PKT_TX_OFFLOAD_MASK ^ QEDE_TX_OFFLOAD_MASK) /* TPA related structures */ struct qede_agg_info { struct rte_mbuf *tpa_head; /* Pointer to first TPA segment */ struct rte_mbuf *tpa_tail; /* Pointer to last TPA segment */ }; /* * Structure associated with each RX queue. */ struct qede_rx_queue { /* Always keep qdev as first member */ struct qede_dev *qdev; struct rte_mempool *mb_pool; struct ecore_chain rx_bd_ring; struct ecore_chain rx_comp_ring; uint16_t *hw_cons_ptr; void OSAL_IOMEM *hw_rxq_prod_addr; struct rte_mbuf **sw_rx_ring; struct ecore_sb_info *sb_info; uint16_t sw_rx_cons; uint16_t sw_rx_prod; uint16_t nb_rx_desc; uint16_t queue_id; uint16_t port_id; uint16_t rx_buf_size; uint16_t rx_alloc_count; uint16_t unused; uint64_t rcv_pkts; uint64_t rx_segs; uint64_t rx_hw_errors; uint64_t rx_alloc_errors; struct qede_agg_info tpa_info[ETH_TPA_MAX_AGGS_NUM]; void *handle; }; union db_prod { struct eth_db_data data; uint32_t raw; }; struct qede_tx_queue { /* Always keep qdev as first member */ struct qede_dev *qdev; struct ecore_chain tx_pbl; struct rte_mbuf **sw_tx_ring; uint16_t nb_tx_desc; uint16_t nb_tx_avail; uint16_t tx_free_thresh; uint16_t queue_id; uint16_t *hw_cons_ptr; uint16_t sw_tx_cons; uint16_t sw_tx_prod; void OSAL_IOMEM *doorbell_addr; volatile union db_prod tx_db; uint16_t port_id; uint64_t xmit_pkts; bool is_legacy; void *handle; }; struct qede_fastpath { struct ecore_sb_info *sb_info; struct qede_rx_queue *rxq; struct qede_tx_queue *txq; }; /* This structure holds the inforation of fast path queues * belonging to individual engines in CMT mode. */ struct qede_fastpath_cmt { /* Always keep this a first element */ struct qede_dev *qdev; /* fastpath info of engine 0 */ struct qede_fastpath *fp0; /* fastpath info of engine 1 */ struct qede_fastpath *fp1; }; /* * RX/TX function prototypes */ int qede_rx_queue_setup(struct rte_eth_dev *dev, uint16_t queue_idx, uint16_t nb_desc, unsigned int socket_id, const struct rte_eth_rxconf *rx_conf, struct rte_mempool *mp); int qede_tx_queue_setup(struct rte_eth_dev *dev, uint16_t queue_idx, uint16_t nb_desc, unsigned int socket_id, const struct rte_eth_txconf *tx_conf); void qede_rx_queue_release(void *rx_queue); void qede_tx_queue_release(void *tx_queue); uint16_t qede_xmit_pkts(void *p_txq, struct rte_mbuf **tx_pkts, uint16_t nb_pkts); uint16_t qede_xmit_pkts_cmt(void *p_txq, struct rte_mbuf **tx_pkts, uint16_t nb_pkts); uint16_t qede_xmit_pkts_regular(void *p_txq, struct rte_mbuf **tx_pkts, uint16_t nb_pkts); uint16_t qede_xmit_prep_pkts(void *p_txq, struct rte_mbuf **tx_pkts, uint16_t nb_pkts); uint16_t qede_recv_pkts(void *p_rxq, struct rte_mbuf **rx_pkts, uint16_t nb_pkts); uint16_t qede_recv_pkts_cmt(void *p_rxq, struct rte_mbuf **rx_pkts, uint16_t nb_pkts); uint16_t qede_recv_pkts_regular(void *p_rxq, struct rte_mbuf **rx_pkts, uint16_t nb_pkts); uint16_t qede_rxtx_pkts_dummy(void *p_rxq, struct rte_mbuf **pkts, uint16_t nb_pkts); int qede_start_queues(struct rte_eth_dev *eth_dev); void qede_stop_queues(struct rte_eth_dev *eth_dev); int qede_calc_rx_buf_size(struct rte_eth_dev *dev, uint16_t mbufsz, uint16_t max_frame_size); int qede_rx_descriptor_status(void *rxq, uint16_t offset); /* Fastpath resource alloc/dealloc helpers */ int qede_alloc_fp_resc(struct qede_dev *qdev); void qede_dealloc_fp_resc(struct rte_eth_dev *eth_dev); #endif /* _QEDE_RXTX_H_ */