2019-09-26 14:02:06 +00:00
|
|
|
/* SPDX-License-Identifier: BSD-3-Clause
|
|
|
|
* Copyright(c) 2018-2019 Hisilicon Limited.
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include <rte_ethdev.h>
|
|
|
|
#include <rte_io.h>
|
|
|
|
#include <rte_malloc.h>
|
|
|
|
|
|
|
|
#include "hns3_ethdev.h"
|
|
|
|
#include "hns3_rxtx.h"
|
|
|
|
#include "hns3_logs.h"
|
2020-04-21 03:32:15 +00:00
|
|
|
#include "hns3_regs.h"
|
2019-09-26 14:02:06 +00:00
|
|
|
|
|
|
|
/* MAC statistics */
|
|
|
|
static const struct hns3_xstats_name_offset hns3_mac_strings[] = {
|
|
|
|
{"mac_tx_mac_pause_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_tx_mac_pause_num)},
|
|
|
|
{"mac_rx_mac_pause_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_rx_mac_pause_num)},
|
|
|
|
{"mac_tx_control_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_tx_ctrl_pkt_num)},
|
|
|
|
{"mac_rx_control_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_rx_ctrl_pkt_num)},
|
|
|
|
{"mac_tx_pfc_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_tx_pfc_pause_pkt_num)},
|
|
|
|
{"mac_tx_pfc_pri0_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_tx_pfc_pri0_pkt_num)},
|
|
|
|
{"mac_tx_pfc_pri1_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_tx_pfc_pri1_pkt_num)},
|
|
|
|
{"mac_tx_pfc_pri2_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_tx_pfc_pri2_pkt_num)},
|
|
|
|
{"mac_tx_pfc_pri3_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_tx_pfc_pri3_pkt_num)},
|
|
|
|
{"mac_tx_pfc_pri4_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_tx_pfc_pri4_pkt_num)},
|
|
|
|
{"mac_tx_pfc_pri5_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_tx_pfc_pri5_pkt_num)},
|
|
|
|
{"mac_tx_pfc_pri6_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_tx_pfc_pri6_pkt_num)},
|
|
|
|
{"mac_tx_pfc_pri7_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_tx_pfc_pri7_pkt_num)},
|
|
|
|
{"mac_rx_pfc_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_rx_pfc_pause_pkt_num)},
|
|
|
|
{"mac_rx_pfc_pri0_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_rx_pfc_pri0_pkt_num)},
|
|
|
|
{"mac_rx_pfc_pri1_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_rx_pfc_pri1_pkt_num)},
|
|
|
|
{"mac_rx_pfc_pri2_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_rx_pfc_pri2_pkt_num)},
|
|
|
|
{"mac_rx_pfc_pri3_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_rx_pfc_pri3_pkt_num)},
|
|
|
|
{"mac_rx_pfc_pri4_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_rx_pfc_pri4_pkt_num)},
|
|
|
|
{"mac_rx_pfc_pri5_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_rx_pfc_pri5_pkt_num)},
|
|
|
|
{"mac_rx_pfc_pri6_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_rx_pfc_pri6_pkt_num)},
|
|
|
|
{"mac_rx_pfc_pri7_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_rx_pfc_pri7_pkt_num)},
|
|
|
|
{"mac_tx_total_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_tx_total_pkt_num)},
|
|
|
|
{"mac_tx_total_oct_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_tx_total_oct_num)},
|
|
|
|
{"mac_tx_good_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_tx_good_pkt_num)},
|
|
|
|
{"mac_tx_bad_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_tx_bad_pkt_num)},
|
|
|
|
{"mac_tx_good_oct_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_tx_good_oct_num)},
|
|
|
|
{"mac_tx_bad_oct_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_tx_bad_oct_num)},
|
|
|
|
{"mac_tx_uni_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_tx_uni_pkt_num)},
|
|
|
|
{"mac_tx_multi_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_tx_multi_pkt_num)},
|
|
|
|
{"mac_tx_broad_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_tx_broad_pkt_num)},
|
|
|
|
{"mac_tx_undersize_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_tx_undersize_pkt_num)},
|
|
|
|
{"mac_tx_oversize_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_tx_oversize_pkt_num)},
|
|
|
|
{"mac_tx_64_oct_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_tx_64_oct_pkt_num)},
|
|
|
|
{"mac_tx_65_127_oct_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_tx_65_127_oct_pkt_num)},
|
|
|
|
{"mac_tx_128_255_oct_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_tx_128_255_oct_pkt_num)},
|
|
|
|
{"mac_tx_256_511_oct_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_tx_256_511_oct_pkt_num)},
|
|
|
|
{"mac_tx_512_1023_oct_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_tx_512_1023_oct_pkt_num)},
|
|
|
|
{"mac_tx_1024_1518_oct_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_tx_1024_1518_oct_pkt_num)},
|
|
|
|
{"mac_tx_1519_2047_oct_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_tx_1519_2047_oct_pkt_num)},
|
|
|
|
{"mac_tx_2048_4095_oct_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_tx_2048_4095_oct_pkt_num)},
|
|
|
|
{"mac_tx_4096_8191_oct_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_tx_4096_8191_oct_pkt_num)},
|
|
|
|
{"mac_tx_8192_9216_oct_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_tx_8192_9216_oct_pkt_num)},
|
|
|
|
{"mac_tx_9217_12287_oct_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_tx_9217_12287_oct_pkt_num)},
|
|
|
|
{"mac_tx_12288_16383_oct_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_tx_12288_16383_oct_pkt_num)},
|
|
|
|
{"mac_tx_1519_max_good_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_tx_1519_max_good_oct_pkt_num)},
|
|
|
|
{"mac_tx_1519_max_bad_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_tx_1519_max_bad_oct_pkt_num)},
|
|
|
|
{"mac_rx_total_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_rx_total_pkt_num)},
|
|
|
|
{"mac_rx_total_oct_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_rx_total_oct_num)},
|
|
|
|
{"mac_rx_good_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_rx_good_pkt_num)},
|
|
|
|
{"mac_rx_bad_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_rx_bad_pkt_num)},
|
|
|
|
{"mac_rx_good_oct_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_rx_good_oct_num)},
|
|
|
|
{"mac_rx_bad_oct_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_rx_bad_oct_num)},
|
|
|
|
{"mac_rx_uni_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_rx_uni_pkt_num)},
|
|
|
|
{"mac_rx_multi_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_rx_multi_pkt_num)},
|
|
|
|
{"mac_rx_broad_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_rx_broad_pkt_num)},
|
|
|
|
{"mac_rx_undersize_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_rx_undersize_pkt_num)},
|
|
|
|
{"mac_rx_oversize_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_rx_oversize_pkt_num)},
|
|
|
|
{"mac_rx_64_oct_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_rx_64_oct_pkt_num)},
|
|
|
|
{"mac_rx_65_127_oct_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_rx_65_127_oct_pkt_num)},
|
|
|
|
{"mac_rx_128_255_oct_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_rx_128_255_oct_pkt_num)},
|
|
|
|
{"mac_rx_256_511_oct_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_rx_256_511_oct_pkt_num)},
|
|
|
|
{"mac_rx_512_1023_oct_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_rx_512_1023_oct_pkt_num)},
|
|
|
|
{"mac_rx_1024_1518_oct_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_rx_1024_1518_oct_pkt_num)},
|
|
|
|
{"mac_rx_1519_2047_oct_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_rx_1519_2047_oct_pkt_num)},
|
|
|
|
{"mac_rx_2048_4095_oct_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_rx_2048_4095_oct_pkt_num)},
|
|
|
|
{"mac_rx_4096_8191_oct_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_rx_4096_8191_oct_pkt_num)},
|
|
|
|
{"mac_rx_8192_9216_oct_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_rx_8192_9216_oct_pkt_num)},
|
|
|
|
{"mac_rx_9217_12287_oct_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_rx_9217_12287_oct_pkt_num)},
|
|
|
|
{"mac_rx_12288_16383_oct_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_rx_12288_16383_oct_pkt_num)},
|
|
|
|
{"mac_rx_1519_max_good_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_rx_1519_max_good_oct_pkt_num)},
|
|
|
|
{"mac_rx_1519_max_bad_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_rx_1519_max_bad_oct_pkt_num)},
|
|
|
|
{"mac_tx_fragment_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_tx_fragment_pkt_num)},
|
|
|
|
{"mac_tx_undermin_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_tx_undermin_pkt_num)},
|
|
|
|
{"mac_tx_jabber_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_tx_jabber_pkt_num)},
|
|
|
|
{"mac_tx_err_all_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_tx_err_all_pkt_num)},
|
|
|
|
{"mac_tx_from_app_good_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_tx_from_app_good_pkt_num)},
|
|
|
|
{"mac_tx_from_app_bad_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_tx_from_app_bad_pkt_num)},
|
|
|
|
{"mac_rx_fragment_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_rx_fragment_pkt_num)},
|
|
|
|
{"mac_rx_undermin_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_rx_undermin_pkt_num)},
|
|
|
|
{"mac_rx_jabber_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_rx_jabber_pkt_num)},
|
|
|
|
{"mac_rx_fcs_err_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_rx_fcs_err_pkt_num)},
|
|
|
|
{"mac_rx_send_app_good_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_rx_send_app_good_pkt_num)},
|
|
|
|
{"mac_rx_send_app_bad_pkt_num",
|
|
|
|
HNS3_MAC_STATS_OFFSET(mac_rx_send_app_bad_pkt_num)}
|
|
|
|
};
|
|
|
|
|
|
|
|
static const struct hns3_xstats_name_offset hns3_error_int_stats_strings[] = {
|
|
|
|
{"MAC_AFIFO_TNL_INT_R",
|
2020-08-25 11:52:59 +00:00
|
|
|
HNS3_ERR_INT_STATS_FIELD_OFFSET(mac_afifo_tnl_int_cnt)},
|
|
|
|
{"PPU_MPF_ABNORMAL_INT_ST2_MSIX",
|
|
|
|
HNS3_ERR_INT_STATS_FIELD_OFFSET(ppu_mpf_abn_int_st2_msix_cnt)},
|
|
|
|
{"SSU_PORT_BASED_ERR_INT_MSIX",
|
|
|
|
HNS3_ERR_INT_STATS_FIELD_OFFSET(ssu_port_based_pf_int_cnt)},
|
2019-09-26 14:02:06 +00:00
|
|
|
{"PPP_PF_ABNORMAL_INT_ST0",
|
2020-08-25 11:52:59 +00:00
|
|
|
HNS3_ERR_INT_STATS_FIELD_OFFSET(ppp_pf_abnormal_int_cnt)},
|
|
|
|
{"PPU_PF_ABNORMAL_INT_ST_MSIX",
|
|
|
|
HNS3_ERR_INT_STATS_FIELD_OFFSET(ppu_pf_abnormal_int_msix_cnt)},
|
|
|
|
{"IMP_TCM_ECC_INT_STS",
|
|
|
|
HNS3_ERR_INT_STATS_FIELD_OFFSET(imp_tcm_ecc_int_cnt)},
|
|
|
|
{"CMDQ_MEM_ECC_INT_STS",
|
|
|
|
HNS3_ERR_INT_STATS_FIELD_OFFSET(cmdq_mem_ecc_int_cnt)},
|
|
|
|
{"IMP_RD_POISON_INT_STS",
|
|
|
|
HNS3_ERR_INT_STATS_FIELD_OFFSET(imp_rd_poison_int_cnt)},
|
|
|
|
{"TQP_INT_ECC_INT_STS",
|
|
|
|
HNS3_ERR_INT_STATS_FIELD_OFFSET(tqp_int_ecc_int_cnt)},
|
|
|
|
{"MSIX_ECC_INT_STS",
|
|
|
|
HNS3_ERR_INT_STATS_FIELD_OFFSET(msix_ecc_int_cnt)},
|
|
|
|
{"SSU_ECC_MULTI_BIT_INT_0",
|
|
|
|
HNS3_ERR_INT_STATS_FIELD_OFFSET(ssu_ecc_multi_bit_int_0_cnt)},
|
|
|
|
{"SSU_ECC_MULTI_BIT_INT_1",
|
|
|
|
HNS3_ERR_INT_STATS_FIELD_OFFSET(ssu_ecc_multi_bit_int_1_cnt)},
|
|
|
|
{"SSU_COMMON_ERR_INT",
|
|
|
|
HNS3_ERR_INT_STATS_FIELD_OFFSET(ssu_common_ecc_int_cnt)},
|
|
|
|
{"IGU_INT_STS",
|
|
|
|
HNS3_ERR_INT_STATS_FIELD_OFFSET(igu_int_cnt)},
|
|
|
|
{"PPP_MPF_ABNORMAL_INT_ST1",
|
|
|
|
HNS3_ERR_INT_STATS_FIELD_OFFSET(ppp_mpf_abnormal_int_st1_cnt)},
|
|
|
|
{"PPP_MPF_ABNORMAL_INT_ST3",
|
|
|
|
HNS3_ERR_INT_STATS_FIELD_OFFSET(ppp_mpf_abnormal_int_st3_cnt)},
|
|
|
|
{"PPU_MPF_ABNORMAL_INT_ST1",
|
|
|
|
HNS3_ERR_INT_STATS_FIELD_OFFSET(ppu_mpf_abnormal_int_st1_cnt)},
|
|
|
|
{"PPU_MPF_ABNORMAL_INT_ST2_RAS",
|
|
|
|
HNS3_ERR_INT_STATS_FIELD_OFFSET(ppu_mpf_abn_int_st2_ras_cnt)},
|
|
|
|
{"PPU_MPF_ABNORMAL_INT_ST3",
|
|
|
|
HNS3_ERR_INT_STATS_FIELD_OFFSET(ppu_mpf_abnormal_int_st3_cnt)},
|
|
|
|
{"TM_SCH_RINT",
|
|
|
|
HNS3_ERR_INT_STATS_FIELD_OFFSET(tm_sch_int_cnt)},
|
|
|
|
{"QCN_FIFO_RINT",
|
|
|
|
HNS3_ERR_INT_STATS_FIELD_OFFSET(qcn_fifo_int_cnt)},
|
|
|
|
{"QCN_ECC_RINT",
|
|
|
|
HNS3_ERR_INT_STATS_FIELD_OFFSET(qcn_ecc_int_cnt)},
|
|
|
|
{"NCSI_ECC_INT_RPT",
|
|
|
|
HNS3_ERR_INT_STATS_FIELD_OFFSET(ncsi_ecc_int_cnt)},
|
|
|
|
{"SSU_PORT_BASED_ERR_INT_RAS",
|
|
|
|
HNS3_ERR_INT_STATS_FIELD_OFFSET(ssu_port_based_err_int_cnt)},
|
|
|
|
{"SSU_FIFO_OVERFLOW_INT",
|
|
|
|
HNS3_ERR_INT_STATS_FIELD_OFFSET(ssu_fifo_overflow_int_cnt)},
|
|
|
|
{"SSU_ETS_TCG_INT",
|
|
|
|
HNS3_ERR_INT_STATS_FIELD_OFFSET(ssu_ets_tcg_int_cnt)},
|
|
|
|
{"IGU_EGU_TNL_INT_STS",
|
|
|
|
HNS3_ERR_INT_STATS_FIELD_OFFSET(igu_egu_tnl_int_cnt)},
|
|
|
|
{"PPU_PF_ABNORMAL_INT_ST_RAS",
|
|
|
|
HNS3_ERR_INT_STATS_FIELD_OFFSET(ppu_pf_abnormal_int_ras_cnt)},
|
2019-09-26 14:02:06 +00:00
|
|
|
};
|
|
|
|
|
|
|
|
/* The statistic of reset */
|
|
|
|
static const struct hns3_xstats_name_offset hns3_reset_stats_strings[] = {
|
|
|
|
{"REQ_RESET_CNT",
|
|
|
|
HNS3_RESET_STATS_FIELD_OFFSET(request_cnt)},
|
|
|
|
{"GLOBAL_RESET_CNT",
|
|
|
|
HNS3_RESET_STATS_FIELD_OFFSET(global_cnt)},
|
|
|
|
{"IMP_RESET_CNT",
|
|
|
|
HNS3_RESET_STATS_FIELD_OFFSET(imp_cnt)},
|
|
|
|
{"RESET_EXEC_CNT",
|
|
|
|
HNS3_RESET_STATS_FIELD_OFFSET(exec_cnt)},
|
|
|
|
{"RESET_SUCCESS_CNT",
|
|
|
|
HNS3_RESET_STATS_FIELD_OFFSET(success_cnt)},
|
|
|
|
{"RESET_FAIL_CNT",
|
|
|
|
HNS3_RESET_STATS_FIELD_OFFSET(fail_cnt)},
|
|
|
|
{"RESET_MERGE_CNT",
|
|
|
|
HNS3_RESET_STATS_FIELD_OFFSET(merge_cnt)}
|
|
|
|
};
|
|
|
|
|
|
|
|
/* The statistic of errors in Rx BD */
|
|
|
|
static const struct hns3_xstats_name_offset hns3_rx_bd_error_strings[] = {
|
2021-01-22 10:18:39 +00:00
|
|
|
{"PKT_LEN_ERRORS",
|
2019-09-26 14:02:06 +00:00
|
|
|
HNS3_RX_BD_ERROR_STATS_FIELD_OFFSET(pkt_len_errors)},
|
2021-01-22 10:18:39 +00:00
|
|
|
{"L2_ERRORS",
|
|
|
|
HNS3_RX_BD_ERROR_STATS_FIELD_OFFSET(l2_errors)}
|
2019-09-26 14:02:06 +00:00
|
|
|
};
|
|
|
|
|
2021-01-22 10:18:39 +00:00
|
|
|
/* The dfx statistic in Rx datapath */
|
|
|
|
static const struct hns3_xstats_name_offset hns3_rxq_dfx_stats_strings[] = {
|
|
|
|
{"L3_CHECKSUM_ERRORS",
|
|
|
|
HNS3_RXQ_DFX_STATS_FIELD_OFFSET(l3_csum_errors)},
|
|
|
|
{"L4_CHECKSUM_ERRORS",
|
|
|
|
HNS3_RXQ_DFX_STATS_FIELD_OFFSET(l4_csum_errors)},
|
|
|
|
{"OL3_CHECKSUM_ERRORS",
|
|
|
|
HNS3_RXQ_DFX_STATS_FIELD_OFFSET(ol3_csum_errors)},
|
|
|
|
{"OL4_CHECKSUM_ERRORS",
|
|
|
|
HNS3_RXQ_DFX_STATS_FIELD_OFFSET(ol4_csum_errors)}
|
|
|
|
};
|
|
|
|
|
|
|
|
/* The dfx statistic in Tx datapath */
|
|
|
|
static const struct hns3_xstats_name_offset hns3_txq_dfx_stats_strings[] = {
|
|
|
|
{"OVER_LENGTH_PKT_CNT",
|
|
|
|
HNS3_TXQ_DFX_STATS_FIELD_OFFSET(over_length_pkt_cnt)},
|
|
|
|
{"EXCEED_LIMITED_BD_PKT_CNT",
|
|
|
|
HNS3_TXQ_DFX_STATS_FIELD_OFFSET(exceed_limit_bd_pkt_cnt)},
|
|
|
|
{"EXCEED_LIMITED_BD_PKT_REASSEMBLE_FAIL_CNT",
|
|
|
|
HNS3_TXQ_DFX_STATS_FIELD_OFFSET(exceed_limit_bd_reassem_fail)},
|
|
|
|
{"UNSUPPORTED_TUNNEL_PKT_CNT",
|
|
|
|
HNS3_TXQ_DFX_STATS_FIELD_OFFSET(unsupported_tunnel_pkt_cnt)},
|
|
|
|
{"QUEUE_FULL_CNT",
|
|
|
|
HNS3_TXQ_DFX_STATS_FIELD_OFFSET(queue_full_cnt)},
|
|
|
|
{"SHORT_PKT_PAD_FAIL_CNT",
|
|
|
|
HNS3_TXQ_DFX_STATS_FIELD_OFFSET(pkt_padding_fail_cnt)}
|
2020-04-29 11:13:23 +00:00
|
|
|
};
|
|
|
|
|
2020-04-21 03:32:15 +00:00
|
|
|
/* The statistic of rx queue */
|
|
|
|
static const struct hns3_xstats_name_offset hns3_rx_queue_strings[] = {
|
|
|
|
{"RX_QUEUE_FBD", HNS3_RING_RX_FBDNUM_REG}
|
|
|
|
};
|
|
|
|
|
|
|
|
/* The statistic of tx queue */
|
|
|
|
static const struct hns3_xstats_name_offset hns3_tx_queue_strings[] = {
|
|
|
|
{"TX_QUEUE_FBD", HNS3_RING_TX_FBDNUM_REG}
|
|
|
|
};
|
|
|
|
|
2019-09-26 14:02:06 +00:00
|
|
|
#define HNS3_NUM_MAC_STATS (sizeof(hns3_mac_strings) / \
|
|
|
|
sizeof(hns3_mac_strings[0]))
|
|
|
|
|
|
|
|
#define HNS3_NUM_ERROR_INT_XSTATS (sizeof(hns3_error_int_stats_strings) / \
|
|
|
|
sizeof(hns3_error_int_stats_strings[0]))
|
|
|
|
|
|
|
|
#define HNS3_NUM_RESET_XSTATS (sizeof(hns3_reset_stats_strings) / \
|
|
|
|
sizeof(hns3_reset_stats_strings[0]))
|
|
|
|
|
|
|
|
#define HNS3_NUM_RX_BD_ERROR_XSTATS (sizeof(hns3_rx_bd_error_strings) / \
|
|
|
|
sizeof(hns3_rx_bd_error_strings[0]))
|
|
|
|
|
2021-01-22 10:18:39 +00:00
|
|
|
#define HNS3_NUM_RXQ_DFX_XSTATS (sizeof(hns3_rxq_dfx_stats_strings) / \
|
|
|
|
sizeof(hns3_rxq_dfx_stats_strings[0]))
|
|
|
|
|
|
|
|
#define HNS3_NUM_TXQ_DFX_XSTATS (sizeof(hns3_txq_dfx_stats_strings) / \
|
|
|
|
sizeof(hns3_txq_dfx_stats_strings[0]))
|
2020-04-29 11:13:23 +00:00
|
|
|
|
2020-04-21 03:32:15 +00:00
|
|
|
#define HNS3_NUM_RX_QUEUE_STATS (sizeof(hns3_rx_queue_strings) / \
|
|
|
|
sizeof(hns3_rx_queue_strings[0]))
|
|
|
|
|
|
|
|
#define HNS3_NUM_TX_QUEUE_STATS (sizeof(hns3_tx_queue_strings) / \
|
|
|
|
sizeof(hns3_tx_queue_strings[0]))
|
|
|
|
|
2019-09-26 14:02:06 +00:00
|
|
|
#define HNS3_FIX_NUM_STATS (HNS3_NUM_MAC_STATS + HNS3_NUM_ERROR_INT_XSTATS + \
|
|
|
|
HNS3_NUM_RESET_XSTATS)
|
|
|
|
|
net/hns3: maximize queue number
The maximum number of queues for hns3 PF and VF driver is 64 based on
hns3 network engine with revision_id equals 0x21. Based on hns3 network
engine with revision_id equals 0x30, the hns3 PF PMD driver can support
up to 1280 queues, and hns3 VF PMD driver can support up to 128 queues.
The following points need to be modified to support maximizing queue
number and maintain better compatibility:
1) Maximizing the number of queues for hns3 PF and VF PMD driver In
current version, VF is not supported when PF is driven by hns3 PMD
driver. If maximum queue numbers allocated to PF PMD driver is less
than total tqps_num allocated to this port, all remaining number of
queues are mapped to VF function, which is unreasonable. So we fix
that all remaining number of queues are mapped to PF function.
Using RTE_LIBRTE_HNS3_MAX_TQP_NUM_PER_PF which comes from
configuration file to limit the queue number allocated to PF device
based on hns3 network engine with revision_id greater than 0x30. And
PF device still keep the maximum 64 queues based on hns3 network
engine with revision_id equals 0x21.
Remove restriction of the macro HNS3_MAX_TQP_NUM_PER_FUNC on the
maximum number of queues in hns3 VF PMD driver and use the value
allocated by hns3 PF kernel netdev driver.
2) According to the queue number allocated to PF device, a variable
array for Rx and Tx queue is dynamically allocated to record the
statistics of Rx and Tx queues during the .dev_init ops
implementation function.
3) Add an extended field in hns3_pf_res_cmd to support the case that
numbers of queue are greater than 1024.
4) Use new base address of Rx or Tx queue if QUEUE_ID of Rx or Tx queue
is greater than 1024.
5) Remove queue id mask and use all bits of actual queue_id as the
queue_id to configure hardware.
6) Currently, 0~9 bits of qset_id in hns3_nq_to_qs_link_cmd used to
record actual qset id and 10 bit as VLD bit are configured to
hardware. So we also need to use 11~15 bits when actual qset_id is
greater than 1024.
7) The number of queue sets based on different network engine are
different. We use it to calculate group number and configure to
hardware in the backpressure configuration.
8) Adding check operations for number of Rx and Tx queue user configured
when mapping queue to tc Rx queue numbers under a single TC must be
less than rss_size_max supported by a single TC. Rx and Tx queue
numbers are allocated to every TC by average. So Rx and Tx queue
numbers must be an integer multiple of 2, or redundant queues are not
available.
9) We can specify which packets enter the queue with a specific queue
number, when creating flow table rules by rte_flow API. Currently,
driver uses 0~9 bits to record the queue_id. So it is necessary to
extend one bit field to record queue_id and configure to hardware, if
the queue_id is greater than 1024.
Signed-off-by: Huisong Li <lihuisong@huawei.com>
Signed-off-by: Wei Hu (Xavier) <xavier.huwei@huawei.com>
2020-09-29 12:01:10 +00:00
|
|
|
static void hns3_tqp_stats_clear(struct hns3_hw *hw);
|
|
|
|
|
2019-09-26 14:02:06 +00:00
|
|
|
/*
|
|
|
|
* Query all the MAC statistics data of Network ICL command ,opcode id: 0x0034.
|
|
|
|
* This command is used before send 'query_mac_stat command', the descriptor
|
|
|
|
* number of 'query_mac_stat command' must match with reg_num in this command.
|
|
|
|
* @praram hw
|
|
|
|
* Pointer to structure hns3_hw.
|
|
|
|
* @return
|
|
|
|
* 0 on success.
|
|
|
|
*/
|
|
|
|
static int
|
|
|
|
hns3_update_mac_stats(struct hns3_hw *hw, const uint32_t desc_num)
|
|
|
|
{
|
|
|
|
uint64_t *data = (uint64_t *)(&hw->mac_stats);
|
|
|
|
struct hns3_cmd_desc *desc;
|
|
|
|
uint64_t *desc_data;
|
|
|
|
uint16_t i, k, n;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
desc = rte_malloc("hns3_mac_desc",
|
|
|
|
desc_num * sizeof(struct hns3_cmd_desc), 0);
|
|
|
|
if (desc == NULL) {
|
|
|
|
hns3_err(hw, "Mac_update_stats alloced desc malloc fail");
|
|
|
|
return -ENOMEM;
|
|
|
|
}
|
|
|
|
|
|
|
|
hns3_cmd_setup_basic_desc(desc, HNS3_OPC_STATS_MAC_ALL, true);
|
|
|
|
ret = hns3_cmd_send(hw, desc, desc_num);
|
|
|
|
if (ret) {
|
|
|
|
hns3_err(hw, "Update complete MAC pkt stats fail : %d", ret);
|
|
|
|
rte_free(desc);
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
for (i = 0; i < desc_num; i++) {
|
|
|
|
/* For special opcode 0034, only the first desc has the head */
|
|
|
|
if (i == 0) {
|
|
|
|
desc_data = (uint64_t *)(&desc[i].data[0]);
|
|
|
|
n = HNS3_RD_FIRST_STATS_NUM;
|
|
|
|
} else {
|
|
|
|
desc_data = (uint64_t *)(&desc[i]);
|
|
|
|
n = HNS3_RD_OTHER_STATS_NUM;
|
|
|
|
}
|
|
|
|
|
|
|
|
for (k = 0; k < n; k++) {
|
|
|
|
*data += rte_le_to_cpu_64(*desc_data);
|
|
|
|
data++;
|
|
|
|
desc_data++;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
rte_free(desc);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Query Mac stat reg num command ,opcode id: 0x0033.
|
|
|
|
* This command is used before send 'query_mac_stat command', the descriptor
|
|
|
|
* number of 'query_mac_stat command' must match with reg_num in this command.
|
|
|
|
* @praram rte_stats
|
|
|
|
* Pointer to structure rte_eth_stats.
|
|
|
|
* @return
|
|
|
|
* 0 on success.
|
|
|
|
*/
|
|
|
|
static int
|
|
|
|
hns3_mac_query_reg_num(struct rte_eth_dev *dev, uint32_t *desc_num)
|
|
|
|
{
|
|
|
|
struct hns3_adapter *hns = dev->data->dev_private;
|
|
|
|
struct hns3_hw *hw = &hns->hw;
|
|
|
|
struct hns3_cmd_desc desc;
|
|
|
|
uint32_t *desc_data;
|
|
|
|
uint32_t reg_num;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
hns3_cmd_setup_basic_desc(&desc, HNS3_OPC_QUERY_MAC_REG_NUM, true);
|
|
|
|
ret = hns3_cmd_send(hw, &desc, 1);
|
|
|
|
if (ret)
|
|
|
|
return ret;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* The num of MAC statistics registers that are provided by IMP in this
|
|
|
|
* version.
|
|
|
|
*/
|
|
|
|
desc_data = (uint32_t *)(&desc.data[0]);
|
|
|
|
reg_num = rte_le_to_cpu_32(*desc_data);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* The descriptor number of 'query_additional_mac_stat command' is
|
|
|
|
* '1 + (reg_num-3)/4 + ((reg_num-3)%4 !=0)';
|
|
|
|
* This value is 83 in this version
|
|
|
|
*/
|
|
|
|
*desc_num = 1 + ((reg_num - 3) >> 2) +
|
|
|
|
(uint32_t)(((reg_num - 3) & 0x3) ? 1 : 0);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
|
|
|
hns3_query_update_mac_stats(struct rte_eth_dev *dev)
|
|
|
|
{
|
|
|
|
struct hns3_adapter *hns = dev->data->dev_private;
|
|
|
|
struct hns3_hw *hw = &hns->hw;
|
|
|
|
uint32_t desc_num;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
ret = hns3_mac_query_reg_num(dev, &desc_num);
|
|
|
|
if (ret == 0)
|
|
|
|
ret = hns3_update_mac_stats(hw, desc_num);
|
|
|
|
else
|
|
|
|
hns3_err(hw, "Query mac reg num fail : %d", ret);
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Get tqp stats from register */
|
|
|
|
static int
|
|
|
|
hns3_update_tqp_stats(struct hns3_hw *hw)
|
|
|
|
{
|
|
|
|
struct hns3_tqp_stats *stats = &hw->tqp_stats;
|
|
|
|
struct hns3_cmd_desc desc;
|
|
|
|
uint64_t cnt;
|
|
|
|
uint16_t i;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
for (i = 0; i < hw->tqps_num; i++) {
|
|
|
|
hns3_cmd_setup_basic_desc(&desc, HNS3_OPC_QUERY_RX_STATUS,
|
|
|
|
true);
|
|
|
|
|
net/hns3: maximize queue number
The maximum number of queues for hns3 PF and VF driver is 64 based on
hns3 network engine with revision_id equals 0x21. Based on hns3 network
engine with revision_id equals 0x30, the hns3 PF PMD driver can support
up to 1280 queues, and hns3 VF PMD driver can support up to 128 queues.
The following points need to be modified to support maximizing queue
number and maintain better compatibility:
1) Maximizing the number of queues for hns3 PF and VF PMD driver In
current version, VF is not supported when PF is driven by hns3 PMD
driver. If maximum queue numbers allocated to PF PMD driver is less
than total tqps_num allocated to this port, all remaining number of
queues are mapped to VF function, which is unreasonable. So we fix
that all remaining number of queues are mapped to PF function.
Using RTE_LIBRTE_HNS3_MAX_TQP_NUM_PER_PF which comes from
configuration file to limit the queue number allocated to PF device
based on hns3 network engine with revision_id greater than 0x30. And
PF device still keep the maximum 64 queues based on hns3 network
engine with revision_id equals 0x21.
Remove restriction of the macro HNS3_MAX_TQP_NUM_PER_FUNC on the
maximum number of queues in hns3 VF PMD driver and use the value
allocated by hns3 PF kernel netdev driver.
2) According to the queue number allocated to PF device, a variable
array for Rx and Tx queue is dynamically allocated to record the
statistics of Rx and Tx queues during the .dev_init ops
implementation function.
3) Add an extended field in hns3_pf_res_cmd to support the case that
numbers of queue are greater than 1024.
4) Use new base address of Rx or Tx queue if QUEUE_ID of Rx or Tx queue
is greater than 1024.
5) Remove queue id mask and use all bits of actual queue_id as the
queue_id to configure hardware.
6) Currently, 0~9 bits of qset_id in hns3_nq_to_qs_link_cmd used to
record actual qset id and 10 bit as VLD bit are configured to
hardware. So we also need to use 11~15 bits when actual qset_id is
greater than 1024.
7) The number of queue sets based on different network engine are
different. We use it to calculate group number and configure to
hardware in the backpressure configuration.
8) Adding check operations for number of Rx and Tx queue user configured
when mapping queue to tc Rx queue numbers under a single TC must be
less than rss_size_max supported by a single TC. Rx and Tx queue
numbers are allocated to every TC by average. So Rx and Tx queue
numbers must be an integer multiple of 2, or redundant queues are not
available.
9) We can specify which packets enter the queue with a specific queue
number, when creating flow table rules by rte_flow API. Currently,
driver uses 0~9 bits to record the queue_id. So it is necessary to
extend one bit field to record queue_id and configure to hardware, if
the queue_id is greater than 1024.
Signed-off-by: Huisong Li <lihuisong@huawei.com>
Signed-off-by: Wei Hu (Xavier) <xavier.huwei@huawei.com>
2020-09-29 12:01:10 +00:00
|
|
|
desc.data[0] = rte_cpu_to_le_32((uint32_t)i);
|
2019-09-26 14:02:06 +00:00
|
|
|
ret = hns3_cmd_send(hw, &desc, 1);
|
|
|
|
if (ret) {
|
2020-11-09 14:28:57 +00:00
|
|
|
hns3_err(hw, "Failed to query RX No.%u queue stat: %d",
|
2019-09-26 14:02:06 +00:00
|
|
|
i, ret);
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
cnt = rte_le_to_cpu_32(desc.data[1]);
|
|
|
|
stats->rcb_rx_ring_pktnum_rcd += cnt;
|
|
|
|
stats->rcb_rx_ring_pktnum[i] += cnt;
|
|
|
|
|
|
|
|
hns3_cmd_setup_basic_desc(&desc, HNS3_OPC_QUERY_TX_STATUS,
|
|
|
|
true);
|
|
|
|
|
net/hns3: maximize queue number
The maximum number of queues for hns3 PF and VF driver is 64 based on
hns3 network engine with revision_id equals 0x21. Based on hns3 network
engine with revision_id equals 0x30, the hns3 PF PMD driver can support
up to 1280 queues, and hns3 VF PMD driver can support up to 128 queues.
The following points need to be modified to support maximizing queue
number and maintain better compatibility:
1) Maximizing the number of queues for hns3 PF and VF PMD driver In
current version, VF is not supported when PF is driven by hns3 PMD
driver. If maximum queue numbers allocated to PF PMD driver is less
than total tqps_num allocated to this port, all remaining number of
queues are mapped to VF function, which is unreasonable. So we fix
that all remaining number of queues are mapped to PF function.
Using RTE_LIBRTE_HNS3_MAX_TQP_NUM_PER_PF which comes from
configuration file to limit the queue number allocated to PF device
based on hns3 network engine with revision_id greater than 0x30. And
PF device still keep the maximum 64 queues based on hns3 network
engine with revision_id equals 0x21.
Remove restriction of the macro HNS3_MAX_TQP_NUM_PER_FUNC on the
maximum number of queues in hns3 VF PMD driver and use the value
allocated by hns3 PF kernel netdev driver.
2) According to the queue number allocated to PF device, a variable
array for Rx and Tx queue is dynamically allocated to record the
statistics of Rx and Tx queues during the .dev_init ops
implementation function.
3) Add an extended field in hns3_pf_res_cmd to support the case that
numbers of queue are greater than 1024.
4) Use new base address of Rx or Tx queue if QUEUE_ID of Rx or Tx queue
is greater than 1024.
5) Remove queue id mask and use all bits of actual queue_id as the
queue_id to configure hardware.
6) Currently, 0~9 bits of qset_id in hns3_nq_to_qs_link_cmd used to
record actual qset id and 10 bit as VLD bit are configured to
hardware. So we also need to use 11~15 bits when actual qset_id is
greater than 1024.
7) The number of queue sets based on different network engine are
different. We use it to calculate group number and configure to
hardware in the backpressure configuration.
8) Adding check operations for number of Rx and Tx queue user configured
when mapping queue to tc Rx queue numbers under a single TC must be
less than rss_size_max supported by a single TC. Rx and Tx queue
numbers are allocated to every TC by average. So Rx and Tx queue
numbers must be an integer multiple of 2, or redundant queues are not
available.
9) We can specify which packets enter the queue with a specific queue
number, when creating flow table rules by rte_flow API. Currently,
driver uses 0~9 bits to record the queue_id. So it is necessary to
extend one bit field to record queue_id and configure to hardware, if
the queue_id is greater than 1024.
Signed-off-by: Huisong Li <lihuisong@huawei.com>
Signed-off-by: Wei Hu (Xavier) <xavier.huwei@huawei.com>
2020-09-29 12:01:10 +00:00
|
|
|
desc.data[0] = rte_cpu_to_le_32((uint32_t)i);
|
2019-09-26 14:02:06 +00:00
|
|
|
ret = hns3_cmd_send(hw, &desc, 1);
|
|
|
|
if (ret) {
|
2020-11-09 14:28:57 +00:00
|
|
|
hns3_err(hw, "Failed to query TX No.%u queue stat: %d",
|
2019-09-26 14:02:06 +00:00
|
|
|
i, ret);
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
cnt = rte_le_to_cpu_32(desc.data[1]);
|
|
|
|
stats->rcb_tx_ring_pktnum_rcd += cnt;
|
|
|
|
stats->rcb_tx_ring_pktnum[i] += cnt;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Query tqp tx queue statistics ,opcode id: 0x0B03.
|
|
|
|
* Query tqp rx queue statistics ,opcode id: 0x0B13.
|
|
|
|
* Get all statistics of a port.
|
|
|
|
* @param eth_dev
|
|
|
|
* Pointer to Ethernet device.
|
|
|
|
* @praram rte_stats
|
|
|
|
* Pointer to structure rte_eth_stats.
|
|
|
|
* @return
|
|
|
|
* 0 on success.
|
|
|
|
*/
|
|
|
|
int
|
|
|
|
hns3_stats_get(struct rte_eth_dev *eth_dev, struct rte_eth_stats *rte_stats)
|
|
|
|
{
|
|
|
|
struct hns3_adapter *hns = eth_dev->data->dev_private;
|
|
|
|
struct hns3_hw *hw = &hns->hw;
|
|
|
|
struct hns3_tqp_stats *stats = &hw->tqp_stats;
|
|
|
|
struct hns3_rx_queue *rxq;
|
2019-10-25 12:37:02 +00:00
|
|
|
struct hns3_tx_queue *txq;
|
2019-09-26 14:02:06 +00:00
|
|
|
uint64_t cnt;
|
|
|
|
uint64_t num;
|
|
|
|
uint16_t i;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
/* Update tqp stats by read register */
|
|
|
|
ret = hns3_update_tqp_stats(hw);
|
|
|
|
if (ret) {
|
|
|
|
hns3_err(hw, "Update tqp stats fail : %d", ret);
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2019-10-25 12:37:02 +00:00
|
|
|
/* Get the error stats of received packets */
|
2019-09-26 14:02:06 +00:00
|
|
|
num = RTE_MIN(RTE_ETHDEV_QUEUE_STAT_CNTRS, eth_dev->data->nb_rx_queues);
|
|
|
|
for (i = 0; i != num; ++i) {
|
|
|
|
rxq = eth_dev->data->rx_queues[i];
|
|
|
|
if (rxq) {
|
2021-01-22 10:18:39 +00:00
|
|
|
cnt = rxq->err_stats.l2_errors +
|
|
|
|
rxq->err_stats.pkt_len_errors;
|
2019-09-26 14:02:06 +00:00
|
|
|
rte_stats->q_errors[i] = cnt;
|
2019-10-25 12:37:02 +00:00
|
|
|
rte_stats->q_ipackets[i] =
|
|
|
|
stats->rcb_rx_ring_pktnum[i] - cnt;
|
2019-09-26 14:02:06 +00:00
|
|
|
rte_stats->ierrors += cnt;
|
|
|
|
}
|
|
|
|
}
|
2019-10-25 12:37:02 +00:00
|
|
|
/* Get the error stats of transmitted packets */
|
|
|
|
num = RTE_MIN(RTE_ETHDEV_QUEUE_STAT_CNTRS, eth_dev->data->nb_tx_queues);
|
|
|
|
for (i = 0; i < num; i++) {
|
|
|
|
txq = eth_dev->data->tx_queues[i];
|
|
|
|
if (txq)
|
|
|
|
rte_stats->q_opackets[i] = stats->rcb_tx_ring_pktnum[i];
|
|
|
|
}
|
|
|
|
|
|
|
|
rte_stats->oerrors = 0;
|
|
|
|
rte_stats->ipackets = stats->rcb_rx_ring_pktnum_rcd -
|
|
|
|
rte_stats->ierrors;
|
|
|
|
rte_stats->opackets = stats->rcb_tx_ring_pktnum_rcd -
|
|
|
|
rte_stats->oerrors;
|
|
|
|
rte_stats->rx_nombuf = eth_dev->data->rx_mbuf_alloc_failed;
|
2019-09-26 14:02:06 +00:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
int
|
|
|
|
hns3_stats_reset(struct rte_eth_dev *eth_dev)
|
|
|
|
{
|
|
|
|
struct hns3_adapter *hns = eth_dev->data->dev_private;
|
|
|
|
struct hns3_hw *hw = &hns->hw;
|
|
|
|
struct hns3_cmd_desc desc_reset;
|
|
|
|
struct hns3_rx_queue *rxq;
|
|
|
|
uint16_t i;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
/*
|
net/hns3: maximize queue number
The maximum number of queues for hns3 PF and VF driver is 64 based on
hns3 network engine with revision_id equals 0x21. Based on hns3 network
engine with revision_id equals 0x30, the hns3 PF PMD driver can support
up to 1280 queues, and hns3 VF PMD driver can support up to 128 queues.
The following points need to be modified to support maximizing queue
number and maintain better compatibility:
1) Maximizing the number of queues for hns3 PF and VF PMD driver In
current version, VF is not supported when PF is driven by hns3 PMD
driver. If maximum queue numbers allocated to PF PMD driver is less
than total tqps_num allocated to this port, all remaining number of
queues are mapped to VF function, which is unreasonable. So we fix
that all remaining number of queues are mapped to PF function.
Using RTE_LIBRTE_HNS3_MAX_TQP_NUM_PER_PF which comes from
configuration file to limit the queue number allocated to PF device
based on hns3 network engine with revision_id greater than 0x30. And
PF device still keep the maximum 64 queues based on hns3 network
engine with revision_id equals 0x21.
Remove restriction of the macro HNS3_MAX_TQP_NUM_PER_FUNC on the
maximum number of queues in hns3 VF PMD driver and use the value
allocated by hns3 PF kernel netdev driver.
2) According to the queue number allocated to PF device, a variable
array for Rx and Tx queue is dynamically allocated to record the
statistics of Rx and Tx queues during the .dev_init ops
implementation function.
3) Add an extended field in hns3_pf_res_cmd to support the case that
numbers of queue are greater than 1024.
4) Use new base address of Rx or Tx queue if QUEUE_ID of Rx or Tx queue
is greater than 1024.
5) Remove queue id mask and use all bits of actual queue_id as the
queue_id to configure hardware.
6) Currently, 0~9 bits of qset_id in hns3_nq_to_qs_link_cmd used to
record actual qset id and 10 bit as VLD bit are configured to
hardware. So we also need to use 11~15 bits when actual qset_id is
greater than 1024.
7) The number of queue sets based on different network engine are
different. We use it to calculate group number and configure to
hardware in the backpressure configuration.
8) Adding check operations for number of Rx and Tx queue user configured
when mapping queue to tc Rx queue numbers under a single TC must be
less than rss_size_max supported by a single TC. Rx and Tx queue
numbers are allocated to every TC by average. So Rx and Tx queue
numbers must be an integer multiple of 2, or redundant queues are not
available.
9) We can specify which packets enter the queue with a specific queue
number, when creating flow table rules by rte_flow API. Currently,
driver uses 0~9 bits to record the queue_id. So it is necessary to
extend one bit field to record queue_id and configure to hardware, if
the queue_id is greater than 1024.
Signed-off-by: Huisong Li <lihuisong@huawei.com>
Signed-off-by: Wei Hu (Xavier) <xavier.huwei@huawei.com>
2020-09-29 12:01:10 +00:00
|
|
|
* Note: Reading hardware statistics of rx/tx queue packet number
|
|
|
|
* will clear them.
|
2019-09-26 14:02:06 +00:00
|
|
|
*/
|
|
|
|
for (i = 0; i < hw->tqps_num; i++) {
|
|
|
|
hns3_cmd_setup_basic_desc(&desc_reset, HNS3_OPC_QUERY_RX_STATUS,
|
|
|
|
true);
|
net/hns3: maximize queue number
The maximum number of queues for hns3 PF and VF driver is 64 based on
hns3 network engine with revision_id equals 0x21. Based on hns3 network
engine with revision_id equals 0x30, the hns3 PF PMD driver can support
up to 1280 queues, and hns3 VF PMD driver can support up to 128 queues.
The following points need to be modified to support maximizing queue
number and maintain better compatibility:
1) Maximizing the number of queues for hns3 PF and VF PMD driver In
current version, VF is not supported when PF is driven by hns3 PMD
driver. If maximum queue numbers allocated to PF PMD driver is less
than total tqps_num allocated to this port, all remaining number of
queues are mapped to VF function, which is unreasonable. So we fix
that all remaining number of queues are mapped to PF function.
Using RTE_LIBRTE_HNS3_MAX_TQP_NUM_PER_PF which comes from
configuration file to limit the queue number allocated to PF device
based on hns3 network engine with revision_id greater than 0x30. And
PF device still keep the maximum 64 queues based on hns3 network
engine with revision_id equals 0x21.
Remove restriction of the macro HNS3_MAX_TQP_NUM_PER_FUNC on the
maximum number of queues in hns3 VF PMD driver and use the value
allocated by hns3 PF kernel netdev driver.
2) According to the queue number allocated to PF device, a variable
array for Rx and Tx queue is dynamically allocated to record the
statistics of Rx and Tx queues during the .dev_init ops
implementation function.
3) Add an extended field in hns3_pf_res_cmd to support the case that
numbers of queue are greater than 1024.
4) Use new base address of Rx or Tx queue if QUEUE_ID of Rx or Tx queue
is greater than 1024.
5) Remove queue id mask and use all bits of actual queue_id as the
queue_id to configure hardware.
6) Currently, 0~9 bits of qset_id in hns3_nq_to_qs_link_cmd used to
record actual qset id and 10 bit as VLD bit are configured to
hardware. So we also need to use 11~15 bits when actual qset_id is
greater than 1024.
7) The number of queue sets based on different network engine are
different. We use it to calculate group number and configure to
hardware in the backpressure configuration.
8) Adding check operations for number of Rx and Tx queue user configured
when mapping queue to tc Rx queue numbers under a single TC must be
less than rss_size_max supported by a single TC. Rx and Tx queue
numbers are allocated to every TC by average. So Rx and Tx queue
numbers must be an integer multiple of 2, or redundant queues are not
available.
9) We can specify which packets enter the queue with a specific queue
number, when creating flow table rules by rte_flow API. Currently,
driver uses 0~9 bits to record the queue_id. So it is necessary to
extend one bit field to record queue_id and configure to hardware, if
the queue_id is greater than 1024.
Signed-off-by: Huisong Li <lihuisong@huawei.com>
Signed-off-by: Wei Hu (Xavier) <xavier.huwei@huawei.com>
2020-09-29 12:01:10 +00:00
|
|
|
desc_reset.data[0] = rte_cpu_to_le_32((uint32_t)i);
|
2019-09-26 14:02:06 +00:00
|
|
|
ret = hns3_cmd_send(hw, &desc_reset, 1);
|
|
|
|
if (ret) {
|
2020-11-09 14:28:57 +00:00
|
|
|
hns3_err(hw, "Failed to reset RX No.%u queue stat: %d",
|
2019-09-26 14:02:06 +00:00
|
|
|
i, ret);
|
2020-04-29 11:13:28 +00:00
|
|
|
return ret;
|
2019-09-26 14:02:06 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
hns3_cmd_setup_basic_desc(&desc_reset, HNS3_OPC_QUERY_TX_STATUS,
|
|
|
|
true);
|
net/hns3: maximize queue number
The maximum number of queues for hns3 PF and VF driver is 64 based on
hns3 network engine with revision_id equals 0x21. Based on hns3 network
engine with revision_id equals 0x30, the hns3 PF PMD driver can support
up to 1280 queues, and hns3 VF PMD driver can support up to 128 queues.
The following points need to be modified to support maximizing queue
number and maintain better compatibility:
1) Maximizing the number of queues for hns3 PF and VF PMD driver In
current version, VF is not supported when PF is driven by hns3 PMD
driver. If maximum queue numbers allocated to PF PMD driver is less
than total tqps_num allocated to this port, all remaining number of
queues are mapped to VF function, which is unreasonable. So we fix
that all remaining number of queues are mapped to PF function.
Using RTE_LIBRTE_HNS3_MAX_TQP_NUM_PER_PF which comes from
configuration file to limit the queue number allocated to PF device
based on hns3 network engine with revision_id greater than 0x30. And
PF device still keep the maximum 64 queues based on hns3 network
engine with revision_id equals 0x21.
Remove restriction of the macro HNS3_MAX_TQP_NUM_PER_FUNC on the
maximum number of queues in hns3 VF PMD driver and use the value
allocated by hns3 PF kernel netdev driver.
2) According to the queue number allocated to PF device, a variable
array for Rx and Tx queue is dynamically allocated to record the
statistics of Rx and Tx queues during the .dev_init ops
implementation function.
3) Add an extended field in hns3_pf_res_cmd to support the case that
numbers of queue are greater than 1024.
4) Use new base address of Rx or Tx queue if QUEUE_ID of Rx or Tx queue
is greater than 1024.
5) Remove queue id mask and use all bits of actual queue_id as the
queue_id to configure hardware.
6) Currently, 0~9 bits of qset_id in hns3_nq_to_qs_link_cmd used to
record actual qset id and 10 bit as VLD bit are configured to
hardware. So we also need to use 11~15 bits when actual qset_id is
greater than 1024.
7) The number of queue sets based on different network engine are
different. We use it to calculate group number and configure to
hardware in the backpressure configuration.
8) Adding check operations for number of Rx and Tx queue user configured
when mapping queue to tc Rx queue numbers under a single TC must be
less than rss_size_max supported by a single TC. Rx and Tx queue
numbers are allocated to every TC by average. So Rx and Tx queue
numbers must be an integer multiple of 2, or redundant queues are not
available.
9) We can specify which packets enter the queue with a specific queue
number, when creating flow table rules by rte_flow API. Currently,
driver uses 0~9 bits to record the queue_id. So it is necessary to
extend one bit field to record queue_id and configure to hardware, if
the queue_id is greater than 1024.
Signed-off-by: Huisong Li <lihuisong@huawei.com>
Signed-off-by: Wei Hu (Xavier) <xavier.huwei@huawei.com>
2020-09-29 12:01:10 +00:00
|
|
|
desc_reset.data[0] = rte_cpu_to_le_32((uint32_t)i);
|
2019-09-26 14:02:06 +00:00
|
|
|
ret = hns3_cmd_send(hw, &desc_reset, 1);
|
|
|
|
if (ret) {
|
2020-11-09 14:28:57 +00:00
|
|
|
hns3_err(hw, "Failed to reset TX No.%u queue stat: %d",
|
2019-09-26 14:02:06 +00:00
|
|
|
i, ret);
|
2020-04-29 11:13:28 +00:00
|
|
|
return ret;
|
2019-09-26 14:02:06 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2021-01-06 03:46:30 +00:00
|
|
|
/*
|
|
|
|
* Clear soft stats of rx error packet which will be dropped
|
|
|
|
* in driver.
|
|
|
|
*/
|
2021-01-22 10:18:39 +00:00
|
|
|
for (i = 0; i < eth_dev->data->nb_rx_queues; i++) {
|
2019-09-26 14:02:06 +00:00
|
|
|
rxq = eth_dev->data->rx_queues[i];
|
|
|
|
if (rxq) {
|
2021-01-22 10:18:39 +00:00
|
|
|
rxq->err_stats.pkt_len_errors = 0;
|
|
|
|
rxq->err_stats.l2_errors = 0;
|
2020-04-29 11:13:23 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
net/hns3: maximize queue number
The maximum number of queues for hns3 PF and VF driver is 64 based on
hns3 network engine with revision_id equals 0x21. Based on hns3 network
engine with revision_id equals 0x30, the hns3 PF PMD driver can support
up to 1280 queues, and hns3 VF PMD driver can support up to 128 queues.
The following points need to be modified to support maximizing queue
number and maintain better compatibility:
1) Maximizing the number of queues for hns3 PF and VF PMD driver In
current version, VF is not supported when PF is driven by hns3 PMD
driver. If maximum queue numbers allocated to PF PMD driver is less
than total tqps_num allocated to this port, all remaining number of
queues are mapped to VF function, which is unreasonable. So we fix
that all remaining number of queues are mapped to PF function.
Using RTE_LIBRTE_HNS3_MAX_TQP_NUM_PER_PF which comes from
configuration file to limit the queue number allocated to PF device
based on hns3 network engine with revision_id greater than 0x30. And
PF device still keep the maximum 64 queues based on hns3 network
engine with revision_id equals 0x21.
Remove restriction of the macro HNS3_MAX_TQP_NUM_PER_FUNC on the
maximum number of queues in hns3 VF PMD driver and use the value
allocated by hns3 PF kernel netdev driver.
2) According to the queue number allocated to PF device, a variable
array for Rx and Tx queue is dynamically allocated to record the
statistics of Rx and Tx queues during the .dev_init ops
implementation function.
3) Add an extended field in hns3_pf_res_cmd to support the case that
numbers of queue are greater than 1024.
4) Use new base address of Rx or Tx queue if QUEUE_ID of Rx or Tx queue
is greater than 1024.
5) Remove queue id mask and use all bits of actual queue_id as the
queue_id to configure hardware.
6) Currently, 0~9 bits of qset_id in hns3_nq_to_qs_link_cmd used to
record actual qset id and 10 bit as VLD bit are configured to
hardware. So we also need to use 11~15 bits when actual qset_id is
greater than 1024.
7) The number of queue sets based on different network engine are
different. We use it to calculate group number and configure to
hardware in the backpressure configuration.
8) Adding check operations for number of Rx and Tx queue user configured
when mapping queue to tc Rx queue numbers under a single TC must be
less than rss_size_max supported by a single TC. Rx and Tx queue
numbers are allocated to every TC by average. So Rx and Tx queue
numbers must be an integer multiple of 2, or redundant queues are not
available.
9) We can specify which packets enter the queue with a specific queue
number, when creating flow table rules by rte_flow API. Currently,
driver uses 0~9 bits to record the queue_id. So it is necessary to
extend one bit field to record queue_id and configure to hardware, if
the queue_id is greater than 1024.
Signed-off-by: Huisong Li <lihuisong@huawei.com>
Signed-off-by: Wei Hu (Xavier) <xavier.huwei@huawei.com>
2020-09-29 12:01:10 +00:00
|
|
|
hns3_tqp_stats_clear(hw);
|
2019-09-26 14:02:06 +00:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2020-04-29 11:13:28 +00:00
|
|
|
static int
|
2019-09-26 14:02:06 +00:00
|
|
|
hns3_mac_stats_reset(__rte_unused struct rte_eth_dev *dev)
|
|
|
|
{
|
|
|
|
struct hns3_adapter *hns = dev->data->dev_private;
|
|
|
|
struct hns3_hw *hw = &hns->hw;
|
|
|
|
struct hns3_mac_stats *mac_stats = &hw->mac_stats;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
ret = hns3_query_update_mac_stats(dev);
|
2020-04-29 11:13:28 +00:00
|
|
|
if (ret) {
|
2019-09-26 14:02:06 +00:00
|
|
|
hns3_err(hw, "Clear Mac stats fail : %d", ret);
|
2020-04-29 11:13:28 +00:00
|
|
|
return ret;
|
|
|
|
}
|
2019-09-26 14:02:06 +00:00
|
|
|
|
|
|
|
memset(mac_stats, 0, sizeof(struct hns3_mac_stats));
|
2020-04-29 11:13:28 +00:00
|
|
|
|
|
|
|
return 0;
|
2019-09-26 14:02:06 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/* This function calculates the number of xstats based on the current config */
|
|
|
|
static int
|
|
|
|
hns3_xstats_calc_num(struct rte_eth_dev *dev)
|
|
|
|
{
|
|
|
|
struct hns3_adapter *hns = dev->data->dev_private;
|
2021-01-22 10:18:39 +00:00
|
|
|
uint16_t nb_rx_q = dev->data->nb_rx_queues;
|
|
|
|
uint16_t nb_tx_q = dev->data->nb_tx_queues;
|
|
|
|
int bderr_stats = nb_rx_q * HNS3_NUM_RX_BD_ERROR_XSTATS;
|
|
|
|
int rx_dfx_stats = nb_rx_q * HNS3_NUM_RXQ_DFX_XSTATS;
|
|
|
|
int tx_dfx_stats = nb_tx_q * HNS3_NUM_TXQ_DFX_XSTATS;
|
|
|
|
int rx_queue_stats = nb_rx_q * HNS3_NUM_RX_QUEUE_STATS;
|
|
|
|
int tx_queue_stats = nb_tx_q * HNS3_NUM_TX_QUEUE_STATS;
|
2019-09-26 14:02:06 +00:00
|
|
|
|
|
|
|
if (hns->is_vf)
|
2021-01-22 10:18:39 +00:00
|
|
|
return bderr_stats + rx_dfx_stats + tx_dfx_stats +
|
|
|
|
rx_queue_stats + tx_queue_stats + HNS3_NUM_RESET_XSTATS;
|
2019-09-26 14:02:06 +00:00
|
|
|
else
|
2021-01-22 10:18:39 +00:00
|
|
|
return bderr_stats + rx_dfx_stats + tx_dfx_stats +
|
|
|
|
rx_queue_stats + tx_queue_stats + HNS3_FIX_NUM_STATS;
|
2020-04-29 11:13:23 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
2021-01-22 10:18:39 +00:00
|
|
|
hns3_queue_stats_get(struct rte_eth_dev *dev, struct rte_eth_xstat *xstats,
|
2020-04-29 11:13:23 +00:00
|
|
|
int *count)
|
|
|
|
{
|
|
|
|
struct hns3_adapter *hns = dev->data->dev_private;
|
|
|
|
struct hns3_hw *hw = &hns->hw;
|
|
|
|
uint32_t reg_offset;
|
|
|
|
uint16_t i, j;
|
|
|
|
|
|
|
|
/* Get rx queue stats */
|
|
|
|
for (j = 0; j < dev->data->nb_rx_queues; j++) {
|
|
|
|
for (i = 0; i < HNS3_NUM_RX_QUEUE_STATS; i++) {
|
net/hns3: maximize queue number
The maximum number of queues for hns3 PF and VF driver is 64 based on
hns3 network engine with revision_id equals 0x21. Based on hns3 network
engine with revision_id equals 0x30, the hns3 PF PMD driver can support
up to 1280 queues, and hns3 VF PMD driver can support up to 128 queues.
The following points need to be modified to support maximizing queue
number and maintain better compatibility:
1) Maximizing the number of queues for hns3 PF and VF PMD driver In
current version, VF is not supported when PF is driven by hns3 PMD
driver. If maximum queue numbers allocated to PF PMD driver is less
than total tqps_num allocated to this port, all remaining number of
queues are mapped to VF function, which is unreasonable. So we fix
that all remaining number of queues are mapped to PF function.
Using RTE_LIBRTE_HNS3_MAX_TQP_NUM_PER_PF which comes from
configuration file to limit the queue number allocated to PF device
based on hns3 network engine with revision_id greater than 0x30. And
PF device still keep the maximum 64 queues based on hns3 network
engine with revision_id equals 0x21.
Remove restriction of the macro HNS3_MAX_TQP_NUM_PER_FUNC on the
maximum number of queues in hns3 VF PMD driver and use the value
allocated by hns3 PF kernel netdev driver.
2) According to the queue number allocated to PF device, a variable
array for Rx and Tx queue is dynamically allocated to record the
statistics of Rx and Tx queues during the .dev_init ops
implementation function.
3) Add an extended field in hns3_pf_res_cmd to support the case that
numbers of queue are greater than 1024.
4) Use new base address of Rx or Tx queue if QUEUE_ID of Rx or Tx queue
is greater than 1024.
5) Remove queue id mask and use all bits of actual queue_id as the
queue_id to configure hardware.
6) Currently, 0~9 bits of qset_id in hns3_nq_to_qs_link_cmd used to
record actual qset id and 10 bit as VLD bit are configured to
hardware. So we also need to use 11~15 bits when actual qset_id is
greater than 1024.
7) The number of queue sets based on different network engine are
different. We use it to calculate group number and configure to
hardware in the backpressure configuration.
8) Adding check operations for number of Rx and Tx queue user configured
when mapping queue to tc Rx queue numbers under a single TC must be
less than rss_size_max supported by a single TC. Rx and Tx queue
numbers are allocated to every TC by average. So Rx and Tx queue
numbers must be an integer multiple of 2, or redundant queues are not
available.
9) We can specify which packets enter the queue with a specific queue
number, when creating flow table rules by rte_flow API. Currently,
driver uses 0~9 bits to record the queue_id. So it is necessary to
extend one bit field to record queue_id and configure to hardware, if
the queue_id is greater than 1024.
Signed-off-by: Huisong Li <lihuisong@huawei.com>
Signed-off-by: Wei Hu (Xavier) <xavier.huwei@huawei.com>
2020-09-29 12:01:10 +00:00
|
|
|
reg_offset = hns3_get_tqp_reg_offset(j);
|
2020-04-29 11:13:23 +00:00
|
|
|
xstats[*count].value = hns3_read_dev(hw,
|
|
|
|
reg_offset + hns3_rx_queue_strings[i].offset);
|
|
|
|
xstats[*count].id = *count;
|
|
|
|
(*count)++;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Get tx queue stats */
|
|
|
|
for (j = 0; j < dev->data->nb_tx_queues; j++) {
|
|
|
|
for (i = 0; i < HNS3_NUM_TX_QUEUE_STATS; i++) {
|
net/hns3: maximize queue number
The maximum number of queues for hns3 PF and VF driver is 64 based on
hns3 network engine with revision_id equals 0x21. Based on hns3 network
engine with revision_id equals 0x30, the hns3 PF PMD driver can support
up to 1280 queues, and hns3 VF PMD driver can support up to 128 queues.
The following points need to be modified to support maximizing queue
number and maintain better compatibility:
1) Maximizing the number of queues for hns3 PF and VF PMD driver In
current version, VF is not supported when PF is driven by hns3 PMD
driver. If maximum queue numbers allocated to PF PMD driver is less
than total tqps_num allocated to this port, all remaining number of
queues are mapped to VF function, which is unreasonable. So we fix
that all remaining number of queues are mapped to PF function.
Using RTE_LIBRTE_HNS3_MAX_TQP_NUM_PER_PF which comes from
configuration file to limit the queue number allocated to PF device
based on hns3 network engine with revision_id greater than 0x30. And
PF device still keep the maximum 64 queues based on hns3 network
engine with revision_id equals 0x21.
Remove restriction of the macro HNS3_MAX_TQP_NUM_PER_FUNC on the
maximum number of queues in hns3 VF PMD driver and use the value
allocated by hns3 PF kernel netdev driver.
2) According to the queue number allocated to PF device, a variable
array for Rx and Tx queue is dynamically allocated to record the
statistics of Rx and Tx queues during the .dev_init ops
implementation function.
3) Add an extended field in hns3_pf_res_cmd to support the case that
numbers of queue are greater than 1024.
4) Use new base address of Rx or Tx queue if QUEUE_ID of Rx or Tx queue
is greater than 1024.
5) Remove queue id mask and use all bits of actual queue_id as the
queue_id to configure hardware.
6) Currently, 0~9 bits of qset_id in hns3_nq_to_qs_link_cmd used to
record actual qset id and 10 bit as VLD bit are configured to
hardware. So we also need to use 11~15 bits when actual qset_id is
greater than 1024.
7) The number of queue sets based on different network engine are
different. We use it to calculate group number and configure to
hardware in the backpressure configuration.
8) Adding check operations for number of Rx and Tx queue user configured
when mapping queue to tc Rx queue numbers under a single TC must be
less than rss_size_max supported by a single TC. Rx and Tx queue
numbers are allocated to every TC by average. So Rx and Tx queue
numbers must be an integer multiple of 2, or redundant queues are not
available.
9) We can specify which packets enter the queue with a specific queue
number, when creating flow table rules by rte_flow API. Currently,
driver uses 0~9 bits to record the queue_id. So it is necessary to
extend one bit field to record queue_id and configure to hardware, if
the queue_id is greater than 1024.
Signed-off-by: Huisong Li <lihuisong@huawei.com>
Signed-off-by: Wei Hu (Xavier) <xavier.huwei@huawei.com>
2020-09-29 12:01:10 +00:00
|
|
|
reg_offset = hns3_get_tqp_reg_offset(j);
|
2020-04-29 11:13:23 +00:00
|
|
|
xstats[*count].value = hns3_read_dev(hw,
|
|
|
|
reg_offset + hns3_tx_queue_strings[i].offset);
|
|
|
|
xstats[*count].id = *count;
|
|
|
|
(*count)++;
|
|
|
|
}
|
|
|
|
}
|
2019-09-26 14:02:06 +00:00
|
|
|
}
|
|
|
|
|
2020-08-25 11:52:59 +00:00
|
|
|
void
|
|
|
|
hns3_error_int_stats_add(struct hns3_adapter *hns, const char *err)
|
|
|
|
{
|
|
|
|
struct hns3_pf *pf = &hns->pf;
|
|
|
|
uint16_t i;
|
|
|
|
char *addr;
|
|
|
|
|
|
|
|
for (i = 0; i < HNS3_NUM_ERROR_INT_XSTATS; i++) {
|
|
|
|
if (strcmp(hns3_error_int_stats_strings[i].name, err) == 0) {
|
|
|
|
addr = (char *)&pf->abn_int_stats +
|
|
|
|
hns3_error_int_stats_strings[i].offset;
|
|
|
|
*(uint64_t *)addr += 1;
|
2020-09-22 12:03:20 +00:00
|
|
|
break;
|
2020-08-25 11:52:59 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2021-01-22 10:18:39 +00:00
|
|
|
static void
|
|
|
|
hns3_rxq_dfx_stats_get(struct rte_eth_dev *dev, struct rte_eth_xstat *xstats,
|
|
|
|
int *count)
|
|
|
|
{
|
|
|
|
struct hns3_rx_dfx_stats *dfx_stats;
|
|
|
|
struct hns3_rx_queue *rxq;
|
|
|
|
uint16_t i, j;
|
|
|
|
char *val;
|
|
|
|
|
|
|
|
for (i = 0; i < dev->data->nb_rx_queues; i++) {
|
|
|
|
rxq = (struct hns3_rx_queue *)dev->data->rx_queues[i];
|
|
|
|
if (rxq == NULL)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
dfx_stats = &rxq->dfx_stats;
|
|
|
|
for (j = 0; j < HNS3_NUM_RXQ_DFX_XSTATS; j++) {
|
|
|
|
val = (char *)dfx_stats +
|
|
|
|
hns3_rxq_dfx_stats_strings[j].offset;
|
|
|
|
xstats[*count].value = *(uint64_t *)val;
|
|
|
|
xstats[*count].id = *count;
|
|
|
|
(*count)++;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
hns3_txq_dfx_stats_get(struct rte_eth_dev *dev, struct rte_eth_xstat *xstats,
|
|
|
|
int *count)
|
|
|
|
{
|
|
|
|
struct hns3_tx_dfx_stats *dfx_stats;
|
|
|
|
struct hns3_tx_queue *txq;
|
|
|
|
uint16_t i, j;
|
|
|
|
char *val;
|
|
|
|
|
|
|
|
for (i = 0; i < dev->data->nb_tx_queues; i++) {
|
|
|
|
txq = (struct hns3_tx_queue *)dev->data->tx_queues[i];
|
|
|
|
if (txq == NULL)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
dfx_stats = &txq->dfx_stats;
|
|
|
|
for (j = 0; j < HNS3_NUM_TXQ_DFX_XSTATS; j++) {
|
|
|
|
val = (char *)dfx_stats +
|
|
|
|
hns3_txq_dfx_stats_strings[j].offset;
|
|
|
|
xstats[*count].value = *(uint64_t *)val;
|
|
|
|
xstats[*count].id = *count;
|
|
|
|
(*count)++;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
hns3_tqp_dfx_stats_get(struct rte_eth_dev *dev, struct rte_eth_xstat *xstats,
|
|
|
|
int *count)
|
|
|
|
{
|
|
|
|
hns3_rxq_dfx_stats_get(dev, xstats, count);
|
|
|
|
hns3_txq_dfx_stats_get(dev, xstats, count);
|
|
|
|
}
|
2019-09-26 14:02:06 +00:00
|
|
|
/*
|
|
|
|
* Retrieve extended(tqp | Mac) statistics of an Ethernet device.
|
|
|
|
* @param dev
|
|
|
|
* Pointer to Ethernet device.
|
|
|
|
* @praram xstats
|
|
|
|
* A pointer to a table of structure of type *rte_eth_xstat*
|
|
|
|
* to be filled with device statistics ids and values.
|
|
|
|
* This parameter can be set to NULL if n is 0.
|
|
|
|
* @param n
|
|
|
|
* The size of the xstats array (number of elements).
|
|
|
|
* @return
|
|
|
|
* 0 on fail, count(The size of the statistics elements) on success.
|
|
|
|
*/
|
|
|
|
int
|
|
|
|
hns3_dev_xstats_get(struct rte_eth_dev *dev, struct rte_eth_xstat *xstats,
|
|
|
|
unsigned int n)
|
|
|
|
{
|
|
|
|
struct hns3_adapter *hns = dev->data->dev_private;
|
|
|
|
struct hns3_pf *pf = &hns->pf;
|
|
|
|
struct hns3_hw *hw = &hns->hw;
|
|
|
|
struct hns3_mac_stats *mac_stats = &hw->mac_stats;
|
|
|
|
struct hns3_reset_stats *reset_stats = &hw->reset.stats;
|
2021-01-22 10:18:39 +00:00
|
|
|
struct hns3_rx_bd_errors_stats *rx_err_stats;
|
2019-09-26 14:02:06 +00:00
|
|
|
struct hns3_rx_queue *rxq;
|
|
|
|
uint16_t i, j;
|
|
|
|
char *addr;
|
|
|
|
int count;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
if (xstats == NULL)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
count = hns3_xstats_calc_num(dev);
|
|
|
|
if ((int)n < count)
|
|
|
|
return count;
|
|
|
|
|
|
|
|
count = 0;
|
|
|
|
|
|
|
|
if (!hns->is_vf) {
|
|
|
|
/* Update Mac stats */
|
|
|
|
ret = hns3_query_update_mac_stats(dev);
|
2021-01-06 03:46:29 +00:00
|
|
|
if (ret < 0) {
|
2019-09-26 14:02:06 +00:00
|
|
|
hns3_err(hw, "Update Mac stats fail : %d", ret);
|
2021-01-06 03:46:29 +00:00
|
|
|
return ret;
|
2019-09-26 14:02:06 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/* Get MAC stats from hw->hw_xstats.mac_stats struct */
|
|
|
|
for (i = 0; i < HNS3_NUM_MAC_STATS; i++) {
|
|
|
|
addr = (char *)mac_stats + hns3_mac_strings[i].offset;
|
|
|
|
xstats[count].value = *(uint64_t *)addr;
|
|
|
|
xstats[count].id = count;
|
|
|
|
count++;
|
|
|
|
}
|
|
|
|
|
|
|
|
for (i = 0; i < HNS3_NUM_ERROR_INT_XSTATS; i++) {
|
|
|
|
addr = (char *)&pf->abn_int_stats +
|
|
|
|
hns3_error_int_stats_strings[i].offset;
|
|
|
|
xstats[count].value = *(uint64_t *)addr;
|
|
|
|
xstats[count].id = count;
|
|
|
|
count++;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Get the reset stat */
|
|
|
|
for (i = 0; i < HNS3_NUM_RESET_XSTATS; i++) {
|
|
|
|
addr = (char *)reset_stats + hns3_reset_stats_strings[i].offset;
|
|
|
|
xstats[count].value = *(uint64_t *)addr;
|
|
|
|
xstats[count].id = count;
|
|
|
|
count++;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Get the Rx BD errors stats */
|
2020-04-21 03:32:15 +00:00
|
|
|
for (j = 0; j < dev->data->nb_rx_queues; j++) {
|
2019-09-26 14:02:06 +00:00
|
|
|
for (i = 0; i < HNS3_NUM_RX_BD_ERROR_XSTATS; i++) {
|
|
|
|
rxq = dev->data->rx_queues[j];
|
2021-01-22 10:18:39 +00:00
|
|
|
if (rxq) {
|
|
|
|
rx_err_stats = &rxq->err_stats;
|
|
|
|
addr = (char *)rx_err_stats +
|
|
|
|
hns3_rx_bd_error_strings[i].offset;
|
|
|
|
xstats[count].value = *(uint64_t *)addr;
|
|
|
|
xstats[count].id = count;
|
|
|
|
count++;
|
|
|
|
}
|
2019-09-26 14:02:06 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2021-01-22 10:18:39 +00:00
|
|
|
hns3_tqp_dfx_stats_get(dev, xstats, &count);
|
|
|
|
hns3_queue_stats_get(dev, xstats, &count);
|
|
|
|
|
|
|
|
return count;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
hns3_tqp_dfx_stats_name_get(struct rte_eth_dev *dev,
|
|
|
|
struct rte_eth_xstat_name *xstats_names,
|
|
|
|
uint32_t *count)
|
|
|
|
{
|
|
|
|
uint16_t i, j;
|
|
|
|
|
|
|
|
for (j = 0; j < dev->data->nb_rx_queues; j++) {
|
|
|
|
for (i = 0; i < HNS3_NUM_RXQ_DFX_XSTATS; i++) {
|
|
|
|
snprintf(xstats_names[*count].name,
|
|
|
|
sizeof(xstats_names[*count].name),
|
|
|
|
"rx_q%u_%s", j,
|
|
|
|
hns3_rxq_dfx_stats_strings[i].name);
|
|
|
|
(*count)++;
|
2020-04-21 03:32:15 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2021-01-22 10:18:39 +00:00
|
|
|
for (j = 0; j < dev->data->nb_tx_queues; j++) {
|
|
|
|
for (i = 0; i < HNS3_NUM_TXQ_DFX_XSTATS; i++) {
|
|
|
|
snprintf(xstats_names[*count].name,
|
|
|
|
sizeof(xstats_names[*count].name),
|
|
|
|
"tx_q%u_%s", j,
|
|
|
|
hns3_txq_dfx_stats_strings[i].name);
|
|
|
|
(*count)++;
|
|
|
|
}
|
|
|
|
}
|
2019-09-26 14:02:06 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Retrieve names of extended statistics of an Ethernet device.
|
|
|
|
*
|
|
|
|
* There is an assumption that 'xstat_names' and 'xstats' arrays are matched
|
|
|
|
* by array index:
|
|
|
|
* xstats_names[i].name => xstats[i].value
|
|
|
|
*
|
|
|
|
* And the array index is same with id field of 'struct rte_eth_xstat':
|
|
|
|
* xstats[i].id == i
|
|
|
|
*
|
|
|
|
* This assumption makes key-value pair matching less flexible but simpler.
|
|
|
|
*
|
|
|
|
* @param dev
|
|
|
|
* Pointer to Ethernet device.
|
|
|
|
* @param xstats_names
|
|
|
|
* An rte_eth_xstat_name array of at least *size* elements to
|
|
|
|
* be filled. If set to NULL, the function returns the required number
|
|
|
|
* of elements.
|
|
|
|
* @param size
|
|
|
|
* The size of the xstats_names array (number of elements).
|
|
|
|
* @return
|
|
|
|
* - A positive value lower or equal to size: success. The return value
|
|
|
|
* is the number of entries filled in the stats table.
|
|
|
|
*/
|
|
|
|
int
|
2020-04-21 03:32:15 +00:00
|
|
|
hns3_dev_xstats_get_names(struct rte_eth_dev *dev,
|
2019-09-26 14:02:06 +00:00
|
|
|
struct rte_eth_xstat_name *xstats_names,
|
|
|
|
__rte_unused unsigned int size)
|
|
|
|
{
|
|
|
|
struct hns3_adapter *hns = dev->data->dev_private;
|
|
|
|
int cnt_stats = hns3_xstats_calc_num(dev);
|
|
|
|
uint32_t count = 0;
|
|
|
|
uint16_t i, j;
|
|
|
|
|
|
|
|
if (xstats_names == NULL)
|
|
|
|
return cnt_stats;
|
|
|
|
|
|
|
|
/* Note: size limited checked in rte_eth_xstats_get_names() */
|
|
|
|
if (!hns->is_vf) {
|
|
|
|
/* Get MAC name from hw->hw_xstats.mac_stats struct */
|
|
|
|
for (i = 0; i < HNS3_NUM_MAC_STATS; i++) {
|
|
|
|
snprintf(xstats_names[count].name,
|
|
|
|
sizeof(xstats_names[count].name),
|
|
|
|
"%s", hns3_mac_strings[i].name);
|
|
|
|
count++;
|
|
|
|
}
|
|
|
|
|
|
|
|
for (i = 0; i < HNS3_NUM_ERROR_INT_XSTATS; i++) {
|
|
|
|
snprintf(xstats_names[count].name,
|
|
|
|
sizeof(xstats_names[count].name),
|
|
|
|
"%s", hns3_error_int_stats_strings[i].name);
|
|
|
|
count++;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
for (i = 0; i < HNS3_NUM_RESET_XSTATS; i++) {
|
|
|
|
snprintf(xstats_names[count].name,
|
|
|
|
sizeof(xstats_names[count].name),
|
|
|
|
"%s", hns3_reset_stats_strings[i].name);
|
|
|
|
count++;
|
|
|
|
}
|
|
|
|
|
|
|
|
for (j = 0; j < dev->data->nb_rx_queues; j++) {
|
|
|
|
for (i = 0; i < HNS3_NUM_RX_BD_ERROR_XSTATS; i++) {
|
|
|
|
snprintf(xstats_names[count].name,
|
|
|
|
sizeof(xstats_names[count].name),
|
2021-01-22 10:18:39 +00:00
|
|
|
"rx_q%u_%s", j,
|
2019-09-26 14:02:06 +00:00
|
|
|
hns3_rx_bd_error_strings[i].name);
|
|
|
|
count++;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2021-01-22 10:18:39 +00:00
|
|
|
hns3_tqp_dfx_stats_name_get(dev, xstats_names, &count);
|
2020-04-29 11:13:23 +00:00
|
|
|
|
2020-04-21 03:32:15 +00:00
|
|
|
for (j = 0; j < dev->data->nb_rx_queues; j++) {
|
|
|
|
for (i = 0; i < HNS3_NUM_RX_QUEUE_STATS; i++) {
|
|
|
|
snprintf(xstats_names[count].name,
|
|
|
|
sizeof(xstats_names[count].name),
|
2021-01-22 10:18:39 +00:00
|
|
|
"rx_q%u_%s", j, hns3_rx_queue_strings[i].name);
|
2020-04-21 03:32:15 +00:00
|
|
|
count++;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
for (j = 0; j < dev->data->nb_tx_queues; j++) {
|
|
|
|
for (i = 0; i < HNS3_NUM_TX_QUEUE_STATS; i++) {
|
|
|
|
snprintf(xstats_names[count].name,
|
|
|
|
sizeof(xstats_names[count].name),
|
2021-01-22 10:18:39 +00:00
|
|
|
"tx_q%u_%s", j, hns3_tx_queue_strings[i].name);
|
2020-04-21 03:32:15 +00:00
|
|
|
count++;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2019-09-26 14:02:06 +00:00
|
|
|
return count;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Retrieve extended statistics of an Ethernet device.
|
|
|
|
*
|
|
|
|
* @param dev
|
|
|
|
* Pointer to Ethernet device.
|
|
|
|
* @param ids
|
|
|
|
* A pointer to an ids array passed by application. This tells which
|
|
|
|
* statistics values function should retrieve. This parameter
|
|
|
|
* can be set to NULL if size is 0. In this case function will retrieve
|
|
|
|
* all avalible statistics.
|
|
|
|
* @param values
|
|
|
|
* A pointer to a table to be filled with device statistics values.
|
|
|
|
* @param size
|
|
|
|
* The size of the ids array (number of elements).
|
|
|
|
* @return
|
|
|
|
* - A positive value lower or equal to size: success. The return value
|
|
|
|
* is the number of entries filled in the stats table.
|
|
|
|
* - A positive value higher than size: error, the given statistics table
|
|
|
|
* is too small. The return value corresponds to the size that should
|
|
|
|
* be given to succeed. The entries in the table are not valid and
|
|
|
|
* shall not be used by the caller.
|
|
|
|
* - 0 on no ids.
|
|
|
|
*/
|
|
|
|
int
|
|
|
|
hns3_dev_xstats_get_by_id(struct rte_eth_dev *dev, const uint64_t *ids,
|
|
|
|
uint64_t *values, uint32_t size)
|
|
|
|
{
|
2020-04-21 03:32:15 +00:00
|
|
|
const uint32_t cnt_stats = hns3_xstats_calc_num(dev);
|
2019-09-26 14:02:06 +00:00
|
|
|
struct hns3_adapter *hns = dev->data->dev_private;
|
2020-04-21 03:32:15 +00:00
|
|
|
struct rte_eth_xstat *values_copy;
|
2019-09-26 14:02:06 +00:00
|
|
|
struct hns3_hw *hw = &hns->hw;
|
2020-04-21 03:32:15 +00:00
|
|
|
uint32_t count_value;
|
2019-09-26 14:02:06 +00:00
|
|
|
uint64_t len;
|
2020-04-21 03:32:15 +00:00
|
|
|
uint32_t i;
|
2019-09-26 14:02:06 +00:00
|
|
|
int ret;
|
|
|
|
|
2021-01-06 03:46:28 +00:00
|
|
|
if (ids == NULL && values == NULL)
|
2019-09-26 14:02:06 +00:00
|
|
|
return cnt_stats;
|
|
|
|
|
2021-01-06 03:46:28 +00:00
|
|
|
if (ids == NULL)
|
|
|
|
if (size < cnt_stats)
|
|
|
|
return cnt_stats;
|
|
|
|
|
2019-09-26 14:02:06 +00:00
|
|
|
/* Update tqp stats by read register */
|
|
|
|
ret = hns3_update_tqp_stats(hw);
|
|
|
|
if (ret) {
|
|
|
|
hns3_err(hw, "Update tqp stats fail : %d", ret);
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2020-04-21 03:32:15 +00:00
|
|
|
len = cnt_stats * sizeof(struct rte_eth_xstat);
|
2019-09-26 14:02:06 +00:00
|
|
|
values_copy = rte_zmalloc("hns3_xstats_values", len, 0);
|
|
|
|
if (values_copy == NULL) {
|
|
|
|
hns3_err(hw, "Failed to allocate %" PRIx64 " bytes needed "
|
|
|
|
"to store statistics values", len);
|
|
|
|
return -ENOMEM;
|
|
|
|
}
|
|
|
|
|
2020-04-21 03:32:15 +00:00
|
|
|
count_value = hns3_dev_xstats_get(dev, values_copy, cnt_stats);
|
|
|
|
if (count_value != cnt_stats) {
|
|
|
|
rte_free(values_copy);
|
|
|
|
return -EINVAL;
|
2019-09-26 14:02:06 +00:00
|
|
|
}
|
|
|
|
|
2021-01-06 03:46:28 +00:00
|
|
|
if (ids == NULL && values != NULL) {
|
|
|
|
for (i = 0; i < cnt_stats; i++)
|
|
|
|
memcpy(&values[i], &values_copy[i].value,
|
|
|
|
sizeof(values[i]));
|
|
|
|
|
|
|
|
rte_free(values_copy);
|
|
|
|
return cnt_stats;
|
|
|
|
}
|
|
|
|
|
2019-09-26 14:02:06 +00:00
|
|
|
for (i = 0; i < size; i++) {
|
|
|
|
if (ids[i] >= cnt_stats) {
|
2020-11-09 14:28:57 +00:00
|
|
|
hns3_err(hw, "ids[%u] (%" PRIx64 ") is invalid, "
|
2019-09-26 14:02:06 +00:00
|
|
|
"should < %u", i, ids[i], cnt_stats);
|
|
|
|
rte_free(values_copy);
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
2020-04-21 03:32:15 +00:00
|
|
|
memcpy(&values[i], &values_copy[ids[i]].value,
|
|
|
|
sizeof(values[i]));
|
2019-09-26 14:02:06 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
rte_free(values_copy);
|
|
|
|
return size;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Retrieve names of extended statistics of an Ethernet device.
|
|
|
|
*
|
|
|
|
* @param dev
|
|
|
|
* Pointer to Ethernet device.
|
|
|
|
* @param xstats_names
|
|
|
|
* An rte_eth_xstat_name array of at least *size* elements to
|
|
|
|
* be filled. If set to NULL, the function returns the required number
|
|
|
|
* of elements.
|
|
|
|
* @param ids
|
|
|
|
* IDs array given by app to retrieve specific statistics
|
|
|
|
* @param size
|
|
|
|
* The size of the xstats_names array (number of elements).
|
|
|
|
* @return
|
|
|
|
* - A positive value lower or equal to size: success. The return value
|
|
|
|
* is the number of entries filled in the stats table.
|
|
|
|
* - A positive value higher than size: error, the given statistics table
|
|
|
|
* is too small. The return value corresponds to the size that should
|
|
|
|
* be given to succeed. The entries in the table are not valid and
|
|
|
|
* shall not be used by the caller.
|
|
|
|
*/
|
|
|
|
int
|
|
|
|
hns3_dev_xstats_get_names_by_id(struct rte_eth_dev *dev,
|
|
|
|
struct rte_eth_xstat_name *xstats_names,
|
|
|
|
const uint64_t *ids, uint32_t size)
|
|
|
|
{
|
2020-04-21 03:32:15 +00:00
|
|
|
const uint32_t cnt_stats = hns3_xstats_calc_num(dev);
|
2019-09-26 14:02:06 +00:00
|
|
|
struct hns3_adapter *hns = dev->data->dev_private;
|
2020-04-21 03:32:15 +00:00
|
|
|
struct rte_eth_xstat_name *names_copy;
|
2019-09-26 14:02:06 +00:00
|
|
|
struct hns3_hw *hw = &hns->hw;
|
|
|
|
uint64_t len;
|
2020-04-21 03:32:15 +00:00
|
|
|
uint32_t i;
|
2019-09-26 14:02:06 +00:00
|
|
|
|
2021-01-06 03:46:28 +00:00
|
|
|
if (xstats_names == NULL)
|
2019-09-26 14:02:06 +00:00
|
|
|
return cnt_stats;
|
|
|
|
|
2021-01-06 03:46:28 +00:00
|
|
|
if (ids == NULL) {
|
|
|
|
if (size < cnt_stats)
|
|
|
|
return cnt_stats;
|
|
|
|
|
|
|
|
return hns3_dev_xstats_get_names(dev, xstats_names, cnt_stats);
|
|
|
|
}
|
|
|
|
|
2019-09-26 14:02:06 +00:00
|
|
|
len = cnt_stats * sizeof(struct rte_eth_xstat_name);
|
2020-04-21 03:32:15 +00:00
|
|
|
names_copy = rte_zmalloc("hns3_xstats_names", len, 0);
|
|
|
|
if (names_copy == NULL) {
|
2019-09-26 14:02:06 +00:00
|
|
|
hns3_err(hw, "Failed to allocate %" PRIx64 " bytes needed "
|
|
|
|
"to store statistics names", len);
|
|
|
|
return -ENOMEM;
|
|
|
|
}
|
|
|
|
|
2020-04-21 03:32:15 +00:00
|
|
|
(void)hns3_dev_xstats_get_names(dev, names_copy, cnt_stats);
|
2019-09-26 14:02:06 +00:00
|
|
|
|
|
|
|
for (i = 0; i < size; i++) {
|
|
|
|
if (ids[i] >= cnt_stats) {
|
2020-11-09 14:28:57 +00:00
|
|
|
hns3_err(hw, "ids[%u] (%" PRIx64 ") is invalid, "
|
2019-09-26 14:02:06 +00:00
|
|
|
"should < %u", i, ids[i], cnt_stats);
|
2020-04-21 03:32:15 +00:00
|
|
|
rte_free(names_copy);
|
2019-09-26 14:02:06 +00:00
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
snprintf(xstats_names[i].name, sizeof(xstats_names[i].name),
|
2020-04-21 03:32:15 +00:00
|
|
|
"%s", names_copy[ids[i]].name);
|
2019-09-26 14:02:06 +00:00
|
|
|
}
|
|
|
|
|
2020-04-21 03:32:15 +00:00
|
|
|
rte_free(names_copy);
|
2019-09-26 14:02:06 +00:00
|
|
|
return size;
|
|
|
|
}
|
|
|
|
|
2021-01-06 03:46:30 +00:00
|
|
|
static void
|
|
|
|
hns3_tqp_dfx_stats_clear(struct rte_eth_dev *dev)
|
|
|
|
{
|
|
|
|
struct hns3_rx_queue *rxq;
|
|
|
|
struct hns3_tx_queue *txq;
|
2021-01-22 10:18:39 +00:00
|
|
|
uint16_t i;
|
2021-01-06 03:46:30 +00:00
|
|
|
|
|
|
|
/* Clear Rx dfx stats */
|
2021-01-22 10:18:39 +00:00
|
|
|
for (i = 0; i < dev->data->nb_rx_queues; i++) {
|
2021-01-06 03:46:30 +00:00
|
|
|
rxq = dev->data->rx_queues[i];
|
2021-01-22 10:18:39 +00:00
|
|
|
if (rxq)
|
|
|
|
memset(&rxq->dfx_stats, 0,
|
|
|
|
sizeof(struct hns3_rx_dfx_stats));
|
2021-01-06 03:46:30 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/* Clear Tx dfx stats */
|
2021-01-22 10:18:39 +00:00
|
|
|
for (i = 0; i < dev->data->nb_tx_queues; i++) {
|
2021-01-06 03:46:30 +00:00
|
|
|
txq = dev->data->tx_queues[i];
|
2021-01-22 10:18:39 +00:00
|
|
|
if (txq)
|
|
|
|
memset(&txq->dfx_stats, 0,
|
|
|
|
sizeof(struct hns3_tx_dfx_stats));
|
2021-01-06 03:46:30 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2019-09-26 14:02:06 +00:00
|
|
|
int
|
|
|
|
hns3_dev_xstats_reset(struct rte_eth_dev *dev)
|
|
|
|
{
|
|
|
|
struct hns3_adapter *hns = dev->data->dev_private;
|
|
|
|
struct hns3_pf *pf = &hns->pf;
|
2020-04-29 11:13:28 +00:00
|
|
|
int ret;
|
2019-09-26 14:02:06 +00:00
|
|
|
|
|
|
|
/* Clear tqp stats */
|
2020-04-29 11:13:28 +00:00
|
|
|
ret = hns3_stats_reset(dev);
|
|
|
|
if (ret)
|
|
|
|
return ret;
|
|
|
|
|
2019-09-26 14:02:06 +00:00
|
|
|
/* Clear reset stats */
|
|
|
|
memset(&hns->hw.reset.stats, 0, sizeof(struct hns3_reset_stats));
|
|
|
|
|
2021-01-06 03:46:30 +00:00
|
|
|
hns3_tqp_dfx_stats_clear(dev);
|
|
|
|
|
2019-09-26 14:02:06 +00:00
|
|
|
if (hns->is_vf)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
/* HW registers are cleared on read */
|
2020-04-29 11:13:28 +00:00
|
|
|
ret = hns3_mac_stats_reset(dev);
|
|
|
|
if (ret)
|
|
|
|
return ret;
|
|
|
|
|
2019-09-26 14:02:06 +00:00
|
|
|
/* Clear error stats */
|
|
|
|
memset(&pf->abn_int_stats, 0, sizeof(struct hns3_err_msix_intr_stats));
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
net/hns3: maximize queue number
The maximum number of queues for hns3 PF and VF driver is 64 based on
hns3 network engine with revision_id equals 0x21. Based on hns3 network
engine with revision_id equals 0x30, the hns3 PF PMD driver can support
up to 1280 queues, and hns3 VF PMD driver can support up to 128 queues.
The following points need to be modified to support maximizing queue
number and maintain better compatibility:
1) Maximizing the number of queues for hns3 PF and VF PMD driver In
current version, VF is not supported when PF is driven by hns3 PMD
driver. If maximum queue numbers allocated to PF PMD driver is less
than total tqps_num allocated to this port, all remaining number of
queues are mapped to VF function, which is unreasonable. So we fix
that all remaining number of queues are mapped to PF function.
Using RTE_LIBRTE_HNS3_MAX_TQP_NUM_PER_PF which comes from
configuration file to limit the queue number allocated to PF device
based on hns3 network engine with revision_id greater than 0x30. And
PF device still keep the maximum 64 queues based on hns3 network
engine with revision_id equals 0x21.
Remove restriction of the macro HNS3_MAX_TQP_NUM_PER_FUNC on the
maximum number of queues in hns3 VF PMD driver and use the value
allocated by hns3 PF kernel netdev driver.
2) According to the queue number allocated to PF device, a variable
array for Rx and Tx queue is dynamically allocated to record the
statistics of Rx and Tx queues during the .dev_init ops
implementation function.
3) Add an extended field in hns3_pf_res_cmd to support the case that
numbers of queue are greater than 1024.
4) Use new base address of Rx or Tx queue if QUEUE_ID of Rx or Tx queue
is greater than 1024.
5) Remove queue id mask and use all bits of actual queue_id as the
queue_id to configure hardware.
6) Currently, 0~9 bits of qset_id in hns3_nq_to_qs_link_cmd used to
record actual qset id and 10 bit as VLD bit are configured to
hardware. So we also need to use 11~15 bits when actual qset_id is
greater than 1024.
7) The number of queue sets based on different network engine are
different. We use it to calculate group number and configure to
hardware in the backpressure configuration.
8) Adding check operations for number of Rx and Tx queue user configured
when mapping queue to tc Rx queue numbers under a single TC must be
less than rss_size_max supported by a single TC. Rx and Tx queue
numbers are allocated to every TC by average. So Rx and Tx queue
numbers must be an integer multiple of 2, or redundant queues are not
available.
9) We can specify which packets enter the queue with a specific queue
number, when creating flow table rules by rte_flow API. Currently,
driver uses 0~9 bits to record the queue_id. So it is necessary to
extend one bit field to record queue_id and configure to hardware, if
the queue_id is greater than 1024.
Signed-off-by: Huisong Li <lihuisong@huawei.com>
Signed-off-by: Wei Hu (Xavier) <xavier.huwei@huawei.com>
2020-09-29 12:01:10 +00:00
|
|
|
|
|
|
|
int
|
|
|
|
hns3_tqp_stats_init(struct hns3_hw *hw)
|
|
|
|
{
|
|
|
|
struct hns3_tqp_stats *tqp_stats = &hw->tqp_stats;
|
|
|
|
|
|
|
|
tqp_stats->rcb_rx_ring_pktnum = rte_zmalloc("hns3_rx_ring_pkt_num",
|
|
|
|
sizeof(uint64_t) * hw->tqps_num, 0);
|
|
|
|
if (tqp_stats->rcb_rx_ring_pktnum == NULL) {
|
|
|
|
hns3_err(hw, "failed to allocate rx_ring pkt_num.");
|
|
|
|
return -ENOMEM;
|
|
|
|
}
|
|
|
|
|
|
|
|
tqp_stats->rcb_tx_ring_pktnum = rte_zmalloc("hns3_tx_ring_pkt_num",
|
|
|
|
sizeof(uint64_t) * hw->tqps_num, 0);
|
|
|
|
if (tqp_stats->rcb_tx_ring_pktnum == NULL) {
|
|
|
|
hns3_err(hw, "failed to allocate tx_ring pkt_num.");
|
|
|
|
rte_free(tqp_stats->rcb_rx_ring_pktnum);
|
|
|
|
tqp_stats->rcb_rx_ring_pktnum = NULL;
|
|
|
|
return -ENOMEM;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
void
|
|
|
|
hns3_tqp_stats_uninit(struct hns3_hw *hw)
|
|
|
|
{
|
|
|
|
struct hns3_tqp_stats *tqp_stats = &hw->tqp_stats;
|
|
|
|
|
|
|
|
rte_free(tqp_stats->rcb_rx_ring_pktnum);
|
|
|
|
tqp_stats->rcb_rx_ring_pktnum = NULL;
|
|
|
|
rte_free(tqp_stats->rcb_tx_ring_pktnum);
|
|
|
|
tqp_stats->rcb_tx_ring_pktnum = NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
hns3_tqp_stats_clear(struct hns3_hw *hw)
|
|
|
|
{
|
|
|
|
struct hns3_tqp_stats *stats = &hw->tqp_stats;
|
|
|
|
|
|
|
|
stats->rcb_rx_ring_pktnum_rcd = 0;
|
|
|
|
stats->rcb_tx_ring_pktnum_rcd = 0;
|
|
|
|
memset(stats->rcb_rx_ring_pktnum, 0, sizeof(uint64_t) * hw->tqps_num);
|
|
|
|
memset(stats->rcb_tx_ring_pktnum, 0, sizeof(uint64_t) * hw->tqps_num);
|
|
|
|
}
|