d9237ae227
There are some constellations where Due to lack of resource allocation
in MFW, There would be an insufficient number of L2 queues for all the
VFs.
This introduces a new feature ECORE_VF_L2_QUE which correctly numbers
the number of VF queues. Notice it might be larger than the actual
number of VFs in configuration space, in which case its the ecore
client responsibility not to try activating that many.
As part of the fix, also correct the nubmering of the VF queues. As
their numbering is dependent on the SBs of the PF, which might only be
partially used by L2 [as half would be assigned for RDMA which doesn't
require L2 queues], we make the numbering consecutive with that of the
L2 queues only.
Fixes: ec94dbc573
("qede: add base driver")
Signed-off-by: Rasesh Mody <rasesh.mody@cavium.com>
814 lines
19 KiB
C
814 lines
19 KiB
C
/*
|
|
* Copyright (c) 2016 QLogic Corporation.
|
|
* All rights reserved.
|
|
* www.qlogic.com
|
|
*
|
|
* See LICENSE.qede_pmd for copyright and licensing details.
|
|
*/
|
|
|
|
#ifndef __ECORE_H
|
|
#define __ECORE_H
|
|
|
|
/* @DPDK */
|
|
#include <sys/stat.h>
|
|
#include <fcntl.h>
|
|
#include <unistd.h>
|
|
|
|
#define CONFIG_ECORE_BINARY_FW
|
|
#undef CONFIG_ECORE_ZIPPED_FW
|
|
|
|
#ifdef CONFIG_ECORE_ZIPPED_FW
|
|
#include <zlib.h>
|
|
#endif
|
|
|
|
#include "ecore_hsi_common.h"
|
|
#include "ecore_hsi_debug_tools.h"
|
|
#include "ecore_hsi_init_func.h"
|
|
#include "ecore_hsi_init_tool.h"
|
|
#include "ecore_proto_if.h"
|
|
#include "mcp_public.h"
|
|
|
|
#define MAX_HWFNS_PER_DEVICE (4)
|
|
#define NAME_SIZE 128 /* @DPDK */
|
|
#define VER_SIZE 16
|
|
#define ECORE_WFQ_UNIT 100
|
|
#include "../qede_logs.h" /* @DPDK */
|
|
|
|
#define ISCSI_BDQ_ID(_port_id) (_port_id)
|
|
#define FCOE_BDQ_ID(_port_id) (_port_id + 2)
|
|
/* Constants */
|
|
#define ECORE_WID_SIZE (1024)
|
|
|
|
/* Configurable */
|
|
#define ECORE_PF_DEMS_SIZE (4)
|
|
|
|
/* cau states */
|
|
enum ecore_coalescing_mode {
|
|
ECORE_COAL_MODE_DISABLE,
|
|
ECORE_COAL_MODE_ENABLE
|
|
};
|
|
|
|
enum ecore_nvm_cmd {
|
|
ECORE_PUT_FILE_BEGIN = DRV_MSG_CODE_NVM_PUT_FILE_BEGIN,
|
|
ECORE_PUT_FILE_DATA = DRV_MSG_CODE_NVM_PUT_FILE_DATA,
|
|
ECORE_NVM_READ_NVRAM = DRV_MSG_CODE_NVM_READ_NVRAM,
|
|
ECORE_NVM_WRITE_NVRAM = DRV_MSG_CODE_NVM_WRITE_NVRAM,
|
|
ECORE_NVM_DEL_FILE = DRV_MSG_CODE_NVM_DEL_FILE,
|
|
ECORE_NVM_SET_SECURE_MODE = DRV_MSG_CODE_SET_SECURE_MODE,
|
|
ECORE_PHY_RAW_READ = DRV_MSG_CODE_PHY_RAW_READ,
|
|
ECORE_PHY_RAW_WRITE = DRV_MSG_CODE_PHY_RAW_WRITE,
|
|
ECORE_PHY_CORE_READ = DRV_MSG_CODE_PHY_CORE_READ,
|
|
ECORE_PHY_CORE_WRITE = DRV_MSG_CODE_PHY_CORE_WRITE,
|
|
ECORE_GET_MCP_NVM_RESP = 0xFFFFFF00
|
|
};
|
|
|
|
#ifndef LINUX_REMOVE
|
|
#if !defined(CONFIG_ECORE_L2)
|
|
#define CONFIG_ECORE_L2
|
|
#define CONFIG_ECORE_SRIOV
|
|
#endif
|
|
#endif
|
|
|
|
/* helpers */
|
|
#ifndef __EXTRACT__LINUX__
|
|
#define MASK_FIELD(_name, _value) \
|
|
((_value) &= (_name##_MASK))
|
|
|
|
#define FIELD_VALUE(_name, _value) \
|
|
((_value & _name##_MASK) << _name##_SHIFT)
|
|
|
|
#define SET_FIELD(value, name, flag) \
|
|
do { \
|
|
(value) &= ~(name##_MASK << name##_SHIFT); \
|
|
(value) |= ((((u64)flag) & (u64)name##_MASK) << (name##_SHIFT));\
|
|
} while (0)
|
|
|
|
#define GET_FIELD(value, name) \
|
|
(((value) >> (name##_SHIFT)) & name##_MASK)
|
|
#endif
|
|
|
|
static OSAL_INLINE u32 DB_ADDR(u32 cid, u32 DEMS)
|
|
{
|
|
u32 db_addr = FIELD_VALUE(DB_LEGACY_ADDR_DEMS, DEMS) |
|
|
(cid * ECORE_PF_DEMS_SIZE);
|
|
|
|
return db_addr;
|
|
}
|
|
|
|
static OSAL_INLINE u32 DB_ADDR_VF(u32 cid, u32 DEMS)
|
|
{
|
|
u32 db_addr = FIELD_VALUE(DB_LEGACY_ADDR_DEMS, DEMS) |
|
|
FIELD_VALUE(DB_LEGACY_ADDR_ICID, cid);
|
|
|
|
return db_addr;
|
|
}
|
|
|
|
#define ALIGNED_TYPE_SIZE(type_name, p_hwfn) \
|
|
((sizeof(type_name) + (u32)(1 << (p_hwfn->p_dev->cache_shift)) - 1) & \
|
|
~((1 << (p_hwfn->p_dev->cache_shift)) - 1))
|
|
|
|
#ifndef LINUX_REMOVE
|
|
#ifndef U64_HI
|
|
#define U64_HI(val) ((u32)(((u64)(val)) >> 32))
|
|
#endif
|
|
|
|
#ifndef U64_LO
|
|
#define U64_LO(val) ((u32)(((u64)(val)) & 0xffffffff))
|
|
#endif
|
|
#endif
|
|
|
|
#ifndef __EXTRACT__LINUX__
|
|
enum DP_LEVEL {
|
|
ECORE_LEVEL_VERBOSE = 0x0,
|
|
ECORE_LEVEL_INFO = 0x1,
|
|
ECORE_LEVEL_NOTICE = 0x2,
|
|
ECORE_LEVEL_ERR = 0x3,
|
|
};
|
|
|
|
#define ECORE_LOG_LEVEL_SHIFT (30)
|
|
#define ECORE_LOG_VERBOSE_MASK (0x3fffffff)
|
|
#define ECORE_LOG_INFO_MASK (0x40000000)
|
|
#define ECORE_LOG_NOTICE_MASK (0x80000000)
|
|
|
|
enum DP_MODULE {
|
|
#ifndef LINUX_REMOVE
|
|
ECORE_MSG_DRV = 0x0001,
|
|
ECORE_MSG_PROBE = 0x0002,
|
|
ECORE_MSG_LINK = 0x0004,
|
|
ECORE_MSG_TIMER = 0x0008,
|
|
ECORE_MSG_IFDOWN = 0x0010,
|
|
ECORE_MSG_IFUP = 0x0020,
|
|
ECORE_MSG_RX_ERR = 0x0040,
|
|
ECORE_MSG_TX_ERR = 0x0080,
|
|
ECORE_MSG_TX_QUEUED = 0x0100,
|
|
ECORE_MSG_INTR = 0x0200,
|
|
ECORE_MSG_TX_DONE = 0x0400,
|
|
ECORE_MSG_RX_STATUS = 0x0800,
|
|
ECORE_MSG_PKTDATA = 0x1000,
|
|
ECORE_MSG_HW = 0x2000,
|
|
ECORE_MSG_WOL = 0x4000,
|
|
#endif
|
|
ECORE_MSG_SPQ = 0x10000,
|
|
ECORE_MSG_STATS = 0x20000,
|
|
ECORE_MSG_DCB = 0x40000,
|
|
ECORE_MSG_IOV = 0x80000,
|
|
ECORE_MSG_SP = 0x100000,
|
|
ECORE_MSG_STORAGE = 0x200000,
|
|
ECORE_MSG_OOO = 0x200000,
|
|
ECORE_MSG_CXT = 0x800000,
|
|
ECORE_MSG_LL2 = 0x1000000,
|
|
ECORE_MSG_ILT = 0x2000000,
|
|
ECORE_MSG_RDMA = 0x4000000,
|
|
ECORE_MSG_DEBUG = 0x8000000,
|
|
/* to be added...up to 0x8000000 */
|
|
};
|
|
#endif
|
|
|
|
#define for_each_hwfn(p_dev, i) for (i = 0; i < p_dev->num_hwfns; i++)
|
|
|
|
#define D_TRINE(val, cond1, cond2, true1, true2, def) \
|
|
(val == (cond1) ? true1 : \
|
|
(val == (cond2) ? true2 : def))
|
|
|
|
/* forward */
|
|
struct ecore_ptt_pool;
|
|
struct ecore_spq;
|
|
struct ecore_sb_info;
|
|
struct ecore_sb_attn_info;
|
|
struct ecore_cxt_mngr;
|
|
struct ecore_dma_mem;
|
|
struct ecore_sb_sp_info;
|
|
struct ecore_ll2_info;
|
|
struct ecore_igu_info;
|
|
struct ecore_mcp_info;
|
|
struct ecore_dcbx_info;
|
|
|
|
struct ecore_rt_data {
|
|
u32 *init_val;
|
|
bool *b_valid;
|
|
};
|
|
|
|
enum ecore_tunn_mode {
|
|
ECORE_MODE_L2GENEVE_TUNN,
|
|
ECORE_MODE_IPGENEVE_TUNN,
|
|
ECORE_MODE_L2GRE_TUNN,
|
|
ECORE_MODE_IPGRE_TUNN,
|
|
ECORE_MODE_VXLAN_TUNN,
|
|
};
|
|
|
|
enum ecore_tunn_clss {
|
|
ECORE_TUNN_CLSS_MAC_VLAN,
|
|
ECORE_TUNN_CLSS_MAC_VNI,
|
|
ECORE_TUNN_CLSS_INNER_MAC_VLAN,
|
|
ECORE_TUNN_CLSS_INNER_MAC_VNI,
|
|
ECORE_TUNN_CLSS_MAC_VLAN_DUAL_STAGE,
|
|
MAX_ECORE_TUNN_CLSS,
|
|
};
|
|
|
|
struct ecore_tunn_start_params {
|
|
unsigned long tunn_mode;
|
|
u16 vxlan_udp_port;
|
|
u16 geneve_udp_port;
|
|
u8 update_vxlan_udp_port;
|
|
u8 update_geneve_udp_port;
|
|
u8 tunn_clss_vxlan;
|
|
u8 tunn_clss_l2geneve;
|
|
u8 tunn_clss_ipgeneve;
|
|
u8 tunn_clss_l2gre;
|
|
u8 tunn_clss_ipgre;
|
|
};
|
|
|
|
struct ecore_tunn_update_params {
|
|
unsigned long tunn_mode_update_mask;
|
|
unsigned long tunn_mode;
|
|
u16 vxlan_udp_port;
|
|
u16 geneve_udp_port;
|
|
u8 update_rx_pf_clss;
|
|
u8 update_tx_pf_clss;
|
|
u8 update_vxlan_udp_port;
|
|
u8 update_geneve_udp_port;
|
|
u8 tunn_clss_vxlan;
|
|
u8 tunn_clss_l2geneve;
|
|
u8 tunn_clss_ipgeneve;
|
|
u8 tunn_clss_l2gre;
|
|
u8 tunn_clss_ipgre;
|
|
};
|
|
|
|
/* The PCI personality is not quite synonymous to protocol ID:
|
|
* 1. All personalities need CORE connections
|
|
* 2. The Ethernet personality may support also the RoCE/iWARP protocol
|
|
*/
|
|
enum ecore_pci_personality {
|
|
ECORE_PCI_ETH,
|
|
ECORE_PCI_FCOE,
|
|
ECORE_PCI_ISCSI,
|
|
ECORE_PCI_ETH_ROCE,
|
|
ECORE_PCI_IWARP,
|
|
ECORE_PCI_DEFAULT /* default in shmem */
|
|
};
|
|
|
|
/* All VFs are symmetric, all counters are PF + all VFs */
|
|
struct ecore_qm_iids {
|
|
u32 cids;
|
|
u32 vf_cids;
|
|
u32 tids;
|
|
};
|
|
|
|
#define MAX_PF_PER_PORT 8
|
|
|
|
/* HW / FW resources, output of features supported below, most information
|
|
* is received from MFW.
|
|
*/
|
|
enum ecore_resources {
|
|
ECORE_SB,
|
|
ECORE_L2_QUEUE,
|
|
ECORE_VPORT,
|
|
ECORE_RSS_ENG,
|
|
ECORE_PQ,
|
|
ECORE_RL,
|
|
ECORE_MAC,
|
|
ECORE_VLAN,
|
|
ECORE_RDMA_CNQ_RAM,
|
|
ECORE_ILT,
|
|
ECORE_LL2_QUEUE,
|
|
ECORE_CMDQS_CQS,
|
|
ECORE_RDMA_STATS_QUEUE,
|
|
ECORE_MAX_RESC, /* must be last */
|
|
};
|
|
|
|
/* Features that require resources, given as input to the resource management
|
|
* algorithm, the output are the resources above
|
|
*/
|
|
enum ecore_feature {
|
|
ECORE_PF_L2_QUE,
|
|
ECORE_PF_TC,
|
|
ECORE_VF,
|
|
ECORE_EXTRA_VF_QUE,
|
|
ECORE_VMQ,
|
|
ECORE_RDMA_CNQ,
|
|
ECORE_ISCSI_CQ,
|
|
ECORE_FCOE_CQ,
|
|
ECORE_VF_L2_QUE,
|
|
ECORE_MAX_FEATURES,
|
|
};
|
|
|
|
enum ecore_port_mode {
|
|
ECORE_PORT_MODE_DE_2X40G,
|
|
ECORE_PORT_MODE_DE_2X50G,
|
|
ECORE_PORT_MODE_DE_1X100G,
|
|
ECORE_PORT_MODE_DE_4X10G_F,
|
|
ECORE_PORT_MODE_DE_4X10G_E,
|
|
ECORE_PORT_MODE_DE_4X20G,
|
|
ECORE_PORT_MODE_DE_1X40G,
|
|
ECORE_PORT_MODE_DE_2X25G,
|
|
ECORE_PORT_MODE_DE_1X25G,
|
|
ECORE_PORT_MODE_DE_4X25G,
|
|
ECORE_PORT_MODE_DE_2X10G,
|
|
};
|
|
|
|
enum ecore_dev_cap {
|
|
ECORE_DEV_CAP_ETH,
|
|
ECORE_DEV_CAP_FCOE,
|
|
ECORE_DEV_CAP_ISCSI,
|
|
ECORE_DEV_CAP_ROCE,
|
|
ECORE_DEV_CAP_IWARP
|
|
};
|
|
|
|
#ifndef __EXTRACT__LINUX__
|
|
enum ecore_hw_err_type {
|
|
ECORE_HW_ERR_FAN_FAIL,
|
|
ECORE_HW_ERR_MFW_RESP_FAIL,
|
|
ECORE_HW_ERR_HW_ATTN,
|
|
ECORE_HW_ERR_DMAE_FAIL,
|
|
ECORE_HW_ERR_RAMROD_FAIL,
|
|
ECORE_HW_ERR_FW_ASSERT,
|
|
};
|
|
#endif
|
|
|
|
struct ecore_hw_info {
|
|
/* PCI personality */
|
|
enum ecore_pci_personality personality;
|
|
|
|
/* Resource Allocation scheme results */
|
|
u32 resc_start[ECORE_MAX_RESC];
|
|
u32 resc_num[ECORE_MAX_RESC];
|
|
u32 feat_num[ECORE_MAX_FEATURES];
|
|
|
|
#define RESC_START(_p_hwfn, resc) ((_p_hwfn)->hw_info.resc_start[resc])
|
|
#define RESC_NUM(_p_hwfn, resc) ((_p_hwfn)->hw_info.resc_num[resc])
|
|
#define RESC_END(_p_hwfn, resc) (RESC_START(_p_hwfn, resc) + \
|
|
RESC_NUM(_p_hwfn, resc))
|
|
#define FEAT_NUM(_p_hwfn, resc) ((_p_hwfn)->hw_info.feat_num[resc])
|
|
|
|
/* Amount of traffic classes HW supports */
|
|
u8 num_hw_tc;
|
|
|
|
/* Amount of TCs which should be active according to DCBx or upper layer driver
|
|
* configuration
|
|
*/
|
|
|
|
u8 num_active_tc;
|
|
|
|
/* Traffic class used for tcp out of order traffic */
|
|
u8 ooo_tc;
|
|
|
|
/* The traffic class used by PF for it's offloaded protocol */
|
|
u8 offload_tc;
|
|
|
|
u32 concrete_fid;
|
|
u16 opaque_fid;
|
|
u16 ovlan;
|
|
u32 part_num[4];
|
|
|
|
unsigned char hw_mac_addr[ETH_ALEN];
|
|
u64 node_wwn; /* For FCoE only */
|
|
u64 port_wwn; /* For FCoE only */
|
|
|
|
u16 num_iscsi_conns;
|
|
u16 num_fcoe_conns;
|
|
|
|
struct ecore_igu_info *p_igu_info;
|
|
/* Sriov */
|
|
u8 max_chains_per_vf;
|
|
|
|
u32 port_mode;
|
|
u32 hw_mode;
|
|
unsigned long device_capabilities;
|
|
|
|
/* Default DCBX mode */
|
|
u8 dcbx_mode;
|
|
};
|
|
|
|
struct ecore_hw_cid_data {
|
|
u32 cid;
|
|
bool b_cid_allocated;
|
|
u8 vfid; /* 1-based; 0 signals this is for a PF */
|
|
|
|
/* Additional identifiers */
|
|
u16 opaque_fid;
|
|
u8 vport_id;
|
|
};
|
|
|
|
/* maximun size of read/write commands (HW limit) */
|
|
#define DMAE_MAX_RW_SIZE 0x2000
|
|
|
|
struct ecore_dmae_info {
|
|
/* Mutex for synchronizing access to functions */
|
|
osal_mutex_t mutex;
|
|
|
|
u8 channel;
|
|
|
|
dma_addr_t completion_word_phys_addr;
|
|
|
|
/* The memory location where the DMAE writes the completion
|
|
* value when an operation is finished on this context.
|
|
*/
|
|
u32 *p_completion_word;
|
|
|
|
dma_addr_t intermediate_buffer_phys_addr;
|
|
|
|
/* An intermediate buffer for DMAE operations that use virtual
|
|
* addresses - data is DMA'd to/from this buffer and then
|
|
* memcpy'd to/from the virtual address
|
|
*/
|
|
u32 *p_intermediate_buffer;
|
|
|
|
dma_addr_t dmae_cmd_phys_addr;
|
|
struct dmae_cmd *p_dmae_cmd;
|
|
};
|
|
|
|
struct ecore_wfq_data {
|
|
u32 default_min_speed; /* When wfq feature is not configured */
|
|
u32 min_speed; /* when feature is configured for any 1 vport */
|
|
bool configured;
|
|
};
|
|
|
|
struct ecore_qm_info {
|
|
struct init_qm_pq_params *qm_pq_params;
|
|
struct init_qm_vport_params *qm_vport_params;
|
|
struct init_qm_port_params *qm_port_params;
|
|
u16 start_pq;
|
|
u8 start_vport;
|
|
u8 pure_lb_pq;
|
|
u8 offload_pq;
|
|
u8 pure_ack_pq;
|
|
u8 ooo_pq;
|
|
u8 vf_queues_offset;
|
|
u16 num_pqs;
|
|
u16 num_vf_pqs;
|
|
u8 num_vports;
|
|
u8 max_phys_tcs_per_port;
|
|
bool pf_rl_en;
|
|
bool pf_wfq_en;
|
|
bool vport_rl_en;
|
|
bool vport_wfq_en;
|
|
u8 pf_wfq;
|
|
u32 pf_rl;
|
|
struct ecore_wfq_data *wfq_data;
|
|
u8 num_pf_rls;
|
|
};
|
|
|
|
struct storm_stats {
|
|
u32 address;
|
|
u32 len;
|
|
};
|
|
|
|
struct ecore_fw_data {
|
|
#ifdef CONFIG_ECORE_BINARY_FW
|
|
struct fw_ver_info *fw_ver_info;
|
|
#endif
|
|
const u8 *modes_tree_buf;
|
|
union init_op *init_ops;
|
|
const u32 *arr_data;
|
|
u32 init_ops_size;
|
|
};
|
|
|
|
struct ecore_hwfn {
|
|
struct ecore_dev *p_dev;
|
|
u8 my_id; /* ID inside the PF */
|
|
#define IS_LEAD_HWFN(edev) (!((edev)->my_id))
|
|
u8 rel_pf_id; /* Relative to engine*/
|
|
u8 abs_pf_id;
|
|
#define ECORE_PATH_ID(_p_hwfn) \
|
|
(ECORE_IS_K2((_p_hwfn)->p_dev) ? 0 : ((_p_hwfn)->abs_pf_id & 1))
|
|
u8 port_id;
|
|
bool b_active;
|
|
|
|
u32 dp_module;
|
|
u8 dp_level;
|
|
char name[NAME_SIZE];
|
|
void *dp_ctx;
|
|
|
|
bool first_on_engine;
|
|
bool hw_init_done;
|
|
|
|
u8 num_funcs_on_engine;
|
|
u8 enabled_func_idx;
|
|
|
|
/* BAR access */
|
|
void OSAL_IOMEM *regview;
|
|
void OSAL_IOMEM *doorbells;
|
|
u64 db_phys_addr;
|
|
unsigned long db_size;
|
|
|
|
/* PTT pool */
|
|
struct ecore_ptt_pool *p_ptt_pool;
|
|
|
|
/* HW info */
|
|
struct ecore_hw_info hw_info;
|
|
|
|
/* rt_array (for init-tool) */
|
|
struct ecore_rt_data rt_data;
|
|
|
|
/* SPQ */
|
|
struct ecore_spq *p_spq;
|
|
|
|
/* EQ */
|
|
struct ecore_eq *p_eq;
|
|
|
|
/* Consolidate Q*/
|
|
struct ecore_consq *p_consq;
|
|
|
|
/* Slow-Path definitions */
|
|
osal_dpc_t sp_dpc;
|
|
bool b_sp_dpc_enabled;
|
|
|
|
struct ecore_ptt *p_main_ptt;
|
|
struct ecore_ptt *p_dpc_ptt;
|
|
|
|
struct ecore_sb_sp_info *p_sp_sb;
|
|
struct ecore_sb_attn_info *p_sb_attn;
|
|
|
|
/* Protocol related */
|
|
bool using_ll2;
|
|
struct ecore_ll2_info *p_ll2_info;
|
|
struct ecore_ooo_info *p_ooo_info;
|
|
struct ecore_iscsi_info *p_iscsi_info;
|
|
struct ecore_fcoe_info *p_fcoe_info;
|
|
struct ecore_rdma_info *p_rdma_info;
|
|
struct ecore_pf_params pf_params;
|
|
|
|
bool b_rdma_enabled_in_prs;
|
|
u32 rdma_prs_search_reg;
|
|
|
|
/* Array of sb_info of all status blocks */
|
|
struct ecore_sb_info *sbs_info[MAX_SB_PER_PF_MIMD];
|
|
u16 num_sbs;
|
|
|
|
struct ecore_cxt_mngr *p_cxt_mngr;
|
|
|
|
/* Flag indicating whether interrupts are enabled or not*/
|
|
bool b_int_enabled;
|
|
bool b_int_requested;
|
|
|
|
/* True if the driver requests for the link */
|
|
bool b_drv_link_init;
|
|
|
|
struct ecore_vf_iov *vf_iov_info;
|
|
struct ecore_pf_iov *pf_iov_info;
|
|
struct ecore_mcp_info *mcp_info;
|
|
struct ecore_dcbx_info *p_dcbx_info;
|
|
|
|
struct ecore_hw_cid_data *p_tx_cids;
|
|
struct ecore_hw_cid_data *p_rx_cids;
|
|
|
|
struct ecore_dmae_info dmae_info;
|
|
|
|
/* QM init */
|
|
struct ecore_qm_info qm_info;
|
|
|
|
#ifdef CONFIG_ECORE_ZIPPED_FW
|
|
/* Buffer for unzipping firmware data */
|
|
void *unzip_buf;
|
|
#endif
|
|
|
|
struct dbg_tools_data dbg_info;
|
|
|
|
struct z_stream_s *stream;
|
|
|
|
/* PWM region specific data */
|
|
u32 dpi_size;
|
|
u32 dpi_count;
|
|
u32 dpi_start_offset; /* this is used to
|
|
* calculate th
|
|
* doorbell address
|
|
*/
|
|
|
|
/* If one of the following is set then EDPM shouldn't be used */
|
|
u8 dcbx_no_edpm;
|
|
u8 db_bar_no_edpm;
|
|
};
|
|
|
|
#ifndef __EXTRACT__LINUX__
|
|
enum ecore_mf_mode {
|
|
ECORE_MF_DEFAULT,
|
|
ECORE_MF_OVLAN,
|
|
ECORE_MF_NPAR,
|
|
};
|
|
#endif
|
|
|
|
/* @DPDK */
|
|
struct ecore_dbg_feature {
|
|
u8 *dump_buf;
|
|
u32 buf_size;
|
|
u32 dumped_dwords;
|
|
};
|
|
|
|
enum qed_dbg_features {
|
|
DBG_FEATURE_BUS,
|
|
DBG_FEATURE_GRC,
|
|
DBG_FEATURE_IDLE_CHK,
|
|
DBG_FEATURE_MCP_TRACE,
|
|
DBG_FEATURE_REG_FIFO,
|
|
DBG_FEATURE_PROTECTION_OVERRIDE,
|
|
DBG_FEATURE_NUM
|
|
};
|
|
|
|
struct ecore_dev {
|
|
u32 dp_module;
|
|
u8 dp_level;
|
|
char name[NAME_SIZE];
|
|
void *dp_ctx;
|
|
|
|
u8 type;
|
|
#define ECORE_DEV_TYPE_BB (0 << 0)
|
|
#define ECORE_DEV_TYPE_AH (1 << 0)
|
|
/* Translate type/revision combo into the proper conditions */
|
|
#define ECORE_IS_BB(dev) ((dev)->type == ECORE_DEV_TYPE_BB)
|
|
#define ECORE_IS_BB_A0(dev) (ECORE_IS_BB(dev) && CHIP_REV_IS_A0(dev))
|
|
#ifndef ASIC_ONLY
|
|
#define ECORE_IS_BB_B0(dev) ((ECORE_IS_BB(dev) && CHIP_REV_IS_B0(dev)) || \
|
|
(CHIP_REV_IS_TEDIBEAR(dev)))
|
|
#else
|
|
#define ECORE_IS_BB_B0(dev) (ECORE_IS_BB(dev) && CHIP_REV_IS_B0(dev))
|
|
#endif
|
|
#define ECORE_IS_AH(dev) ((dev)->type == ECORE_DEV_TYPE_AH)
|
|
#define ECORE_IS_K2(dev) ECORE_IS_AH(dev)
|
|
|
|
u16 vendor_id;
|
|
u16 device_id;
|
|
|
|
u16 chip_num;
|
|
#define CHIP_NUM_MASK 0xffff
|
|
#define CHIP_NUM_SHIFT 16
|
|
|
|
u16 chip_rev;
|
|
#define CHIP_REV_MASK 0xf
|
|
#define CHIP_REV_SHIFT 12
|
|
#ifndef ASIC_ONLY
|
|
#define CHIP_REV_IS_TEDIBEAR(_p_dev) ((_p_dev)->chip_rev == 0x5)
|
|
#define CHIP_REV_IS_EMUL_A0(_p_dev) ((_p_dev)->chip_rev == 0xe)
|
|
#define CHIP_REV_IS_EMUL_B0(_p_dev) ((_p_dev)->chip_rev == 0xc)
|
|
#define CHIP_REV_IS_EMUL(_p_dev) (CHIP_REV_IS_EMUL_A0(_p_dev) || \
|
|
CHIP_REV_IS_EMUL_B0(_p_dev))
|
|
#define CHIP_REV_IS_FPGA_A0(_p_dev) ((_p_dev)->chip_rev == 0xf)
|
|
#define CHIP_REV_IS_FPGA_B0(_p_dev) ((_p_dev)->chip_rev == 0xd)
|
|
#define CHIP_REV_IS_FPGA(_p_dev) (CHIP_REV_IS_FPGA_A0(_p_dev) || \
|
|
CHIP_REV_IS_FPGA_B0(_p_dev))
|
|
#define CHIP_REV_IS_SLOW(_p_dev) \
|
|
(CHIP_REV_IS_EMUL(_p_dev) || CHIP_REV_IS_FPGA(_p_dev))
|
|
#define CHIP_REV_IS_A0(_p_dev) \
|
|
(CHIP_REV_IS_EMUL_A0(_p_dev) || \
|
|
CHIP_REV_IS_FPGA_A0(_p_dev) || \
|
|
!(_p_dev)->chip_rev)
|
|
#define CHIP_REV_IS_B0(_p_dev) \
|
|
(CHIP_REV_IS_EMUL_B0(_p_dev) || \
|
|
CHIP_REV_IS_FPGA_B0(_p_dev) || \
|
|
(_p_dev)->chip_rev == 1)
|
|
#define CHIP_REV_IS_ASIC(_p_dev) !CHIP_REV_IS_SLOW(_p_dev)
|
|
#else
|
|
#define CHIP_REV_IS_A0(_p_dev) (!(_p_dev)->chip_rev)
|
|
#define CHIP_REV_IS_B0(_p_dev) ((_p_dev)->chip_rev == 1)
|
|
#endif
|
|
|
|
u16 chip_metal;
|
|
#define CHIP_METAL_MASK 0xff
|
|
#define CHIP_METAL_SHIFT 4
|
|
|
|
u16 chip_bond_id;
|
|
#define CHIP_BOND_ID_MASK 0xf
|
|
#define CHIP_BOND_ID_SHIFT 0
|
|
|
|
u8 num_engines;
|
|
u8 num_ports_in_engines;
|
|
u8 num_funcs_in_port;
|
|
|
|
u8 path_id;
|
|
enum ecore_mf_mode mf_mode;
|
|
#define IS_MF_DEFAULT(_p_hwfn) \
|
|
(((_p_hwfn)->p_dev)->mf_mode == ECORE_MF_DEFAULT)
|
|
#define IS_MF_SI(_p_hwfn) \
|
|
(((_p_hwfn)->p_dev)->mf_mode == ECORE_MF_NPAR)
|
|
#define IS_MF_SD(_p_hwfn) \
|
|
(((_p_hwfn)->p_dev)->mf_mode == ECORE_MF_OVLAN)
|
|
|
|
int pcie_width;
|
|
int pcie_speed;
|
|
u8 ver_str[NAME_SIZE]; /* @DPDK */
|
|
/* Add MF related configuration */
|
|
u8 mcp_rev;
|
|
u8 boot_mode;
|
|
|
|
u8 wol;
|
|
|
|
u32 int_mode;
|
|
enum ecore_coalescing_mode int_coalescing_mode;
|
|
u16 rx_coalesce_usecs;
|
|
u16 tx_coalesce_usecs;
|
|
|
|
/* Start Bar offset of first hwfn */
|
|
void OSAL_IOMEM *regview;
|
|
void OSAL_IOMEM *doorbells;
|
|
u64 db_phys_addr;
|
|
unsigned long db_size;
|
|
|
|
/* PCI */
|
|
u8 cache_shift;
|
|
|
|
/* Init */
|
|
const struct iro *iro_arr;
|
|
#define IRO (p_hwfn->p_dev->iro_arr)
|
|
|
|
/* HW functions */
|
|
u8 num_hwfns;
|
|
struct ecore_hwfn hwfns[MAX_HWFNS_PER_DEVICE];
|
|
|
|
/* SRIOV */
|
|
struct ecore_hw_sriov_info *p_iov_info;
|
|
#define IS_ECORE_SRIOV(p_dev) (!!(p_dev)->p_iov_info)
|
|
unsigned long tunn_mode;
|
|
|
|
bool b_is_vf;
|
|
|
|
u32 drv_type;
|
|
|
|
u32 rdma_max_sge;
|
|
u32 rdma_max_inline;
|
|
u32 rdma_max_srq_sge;
|
|
|
|
struct ecore_eth_stats *reset_stats;
|
|
struct ecore_fw_data *fw_data;
|
|
|
|
u32 mcp_nvm_resp;
|
|
|
|
/* Recovery */
|
|
bool recov_in_prog;
|
|
|
|
/* Indicates whether should prevent attentions from being reasserted */
|
|
|
|
bool attn_clr_en;
|
|
|
|
/* Indicates whether allowing the MFW to collect a crash dump */
|
|
bool mdump_en;
|
|
|
|
/* Indicates if the reg_fifo is checked after any register access */
|
|
bool chk_reg_fifo;
|
|
|
|
#ifndef ASIC_ONLY
|
|
bool b_is_emul_full;
|
|
#endif
|
|
|
|
#ifdef CONFIG_ECORE_BINARY_FW /* @DPDK */
|
|
void *firmware;
|
|
u64 fw_len;
|
|
#endif
|
|
|
|
/* @DPDK */
|
|
struct ecore_dbg_feature dbg_features[DBG_FEATURE_NUM];
|
|
u8 engine_for_debug;
|
|
};
|
|
|
|
#define NUM_OF_VFS(dev) (ECORE_IS_BB(dev) ? MAX_NUM_VFS_BB \
|
|
: MAX_NUM_VFS_K2)
|
|
#define NUM_OF_L2_QUEUES(dev) (ECORE_IS_BB(dev) ? MAX_NUM_L2_QUEUES_BB \
|
|
: MAX_NUM_L2_QUEUES_K2)
|
|
#define NUM_OF_PORTS(dev) (ECORE_IS_BB(dev) ? MAX_NUM_PORTS_BB \
|
|
: MAX_NUM_PORTS_K2)
|
|
#define NUM_OF_SBS(dev) (ECORE_IS_BB(dev) ? MAX_SB_PER_PATH_BB \
|
|
: MAX_SB_PER_PATH_K2)
|
|
#define NUM_OF_ENG_PFS(dev) (ECORE_IS_BB(dev) ? MAX_NUM_PFS_BB \
|
|
: MAX_NUM_PFS_K2)
|
|
|
|
/**
|
|
* @brief ecore_concrete_to_sw_fid - get the sw function id from
|
|
* the concrete value.
|
|
*
|
|
* @param concrete_fid
|
|
*
|
|
* @return OSAL_INLINE u8
|
|
*/
|
|
static OSAL_INLINE u8 ecore_concrete_to_sw_fid(struct ecore_dev *p_dev,
|
|
u32 concrete_fid)
|
|
{
|
|
u8 vfid = GET_FIELD(concrete_fid, PXP_CONCRETE_FID_VFID);
|
|
u8 pfid = GET_FIELD(concrete_fid, PXP_CONCRETE_FID_PFID);
|
|
u8 vf_valid = GET_FIELD(concrete_fid, PXP_CONCRETE_FID_VFVALID);
|
|
u8 sw_fid;
|
|
|
|
if (vf_valid)
|
|
sw_fid = vfid + MAX_NUM_PFS;
|
|
else
|
|
sw_fid = pfid;
|
|
|
|
return sw_fid;
|
|
}
|
|
|
|
#define PURE_LB_TC 8
|
|
#define PKT_LB_TC 9
|
|
|
|
int ecore_configure_vport_wfq(struct ecore_dev *p_dev, u16 vp_id, u32 rate);
|
|
void ecore_configure_vp_wfq_on_link_change(struct ecore_dev *p_dev,
|
|
u32 min_pf_rate);
|
|
|
|
int ecore_configure_pf_max_bandwidth(struct ecore_dev *p_dev, u8 max_bw);
|
|
int ecore_configure_pf_min_bandwidth(struct ecore_dev *p_dev, u8 min_bw);
|
|
void ecore_clean_wfq_db(struct ecore_hwfn *p_hwfn, struct ecore_ptt *p_ptt);
|
|
int ecore_device_num_engines(struct ecore_dev *p_dev);
|
|
int ecore_device_num_ports(struct ecore_dev *p_dev);
|
|
void ecore_set_fw_mac_addr(__le16 *fw_msb, __le16 *fw_mid, __le16 *fw_lsb,
|
|
u8 *mac);
|
|
|
|
#define ECORE_LEADING_HWFN(dev) (&dev->hwfns[0])
|
|
|
|
#endif /* __ECORE_H */
|