numam-dpdk/drivers/common/mlx5/mlx5_devx_cmds.h
Xiaoyu Min 4d368e1da3 net/mlx5: support flow counter action for HWS
This commit adds HW steering counter action support.
The pool mechanism is the basic data structure for the HW steering
counter.

The HW steering's counter pool is based on the rte_ring of zero-copy
variation.

There are two global rte_rings:
1. free_list:
     Store the counters indexes, which are ready for use.
2. wait_reset_list:
     Store the counters indexes, which are just freed from the user and
     need to query the hardware counter to get the reset value before
     this counter can be reused again.

The counter pool also supports cache per HW steering's queues, which are
also based on the rte_ring of zero-copy variation.

The cache can be configured in size, preload, threshold, and fetch size,
they are all exposed via device args.

The main operations of the counter pool are as follows:

 - Get one counter from the pool:
   1. The user call _get_* API.
   2. If the cache is enabled, dequeue one counter index from the local
      cache:
      2. A: if the dequeued one from the local cache is still in reset
        status (counter's query_gen_when_free is equal to pool's query
        gen):
        I. Flush all counters in the local cache back to global
           wait_reset_list.
        II. Fetch _fetch_sz_ counters into the cache from the global
            free list.
        III. Fetch one counter from the cache.
   3. If the cache is empty, fetch _fetch_sz_ counters from the global
      free list into the cache and fetch one counter from the cache.
 - Free one counter into the pool:
   1. The user calls _put_* API.
   2. Put the counter into the local cache.
   3. If the local cache is full:
      A: Write back all counters above _threshold_ into the global
         wait_reset_list.
      B: Also, write back this counter into the global wait_reset_list.

When the local cache is disabled, _get_/_put_ cache directly from/into
global list.

Signed-off-by: Xiaoyu Min <jackmin@nvidia.com>
Acked-by: Viacheslav Ovsiienko <viacheslavo@nvidia.com>
2022-10-26 13:33:39 +02:00

814 lines
24 KiB
C

/* SPDX-License-Identifier: BSD-3-Clause
* Copyright 2019 Mellanox Technologies, Ltd
*/
#ifndef RTE_PMD_MLX5_DEVX_CMDS_H_
#define RTE_PMD_MLX5_DEVX_CMDS_H_
#include <rte_compat.h>
#include <rte_bitops.h>
#include "mlx5_glue.h"
#include "mlx5_prm.h"
/* This is limitation of libibverbs: in length variable type is u16. */
#define MLX5_DEVX_MAX_KLM_ENTRIES ((UINT16_MAX - \
MLX5_ST_SZ_DW(create_mkey_in) * 4) / (MLX5_ST_SZ_DW(klm) * 4))
struct mlx5_devx_counter_attr {
uint32_t pd_valid:1;
uint32_t pd:24;
uint32_t bulk_log_max_alloc:1;
union {
uint8_t flow_counter_bulk_log_size;
uint8_t bulk_n_128;
};
};
struct mlx5_devx_mkey_attr {
uint64_t addr;
uint64_t size;
uint32_t umem_id;
uint32_t pd;
uint32_t log_entity_size;
uint32_t pg_access:1;
uint32_t relaxed_ordering_write:1;
uint32_t relaxed_ordering_read:1;
uint32_t umr_en:1;
uint32_t crypto_en:2;
uint32_t set_remote_rw:1;
struct mlx5_klm *klm_array;
int klm_num;
};
/* HCA qos attributes. */
struct mlx5_hca_qos_attr {
uint32_t sup:1; /* Whether QOS is supported. */
uint32_t flow_meter_old:1; /* Flow meter is supported, old version. */
uint32_t packet_pacing:1; /* Packet pacing is supported. */
uint32_t wqe_rate_pp:1; /* Packet pacing WQE rate mode. */
uint32_t flow_meter:1;
/*
* Flow meter is supported, updated version.
* When flow_meter is 1, it indicates that REG_C sharing is supported.
* If flow_meter is 1, flow_meter_old is also 1.
* Using older driver versions, flow_meter_old can be 1
* while flow_meter is 0.
*/
uint32_t flow_meter_aso_sup:1;
/* Whether FLOW_METER_ASO Object is supported. */
uint8_t log_max_flow_meter;
/* Power of the maximum supported meters. */
uint8_t flow_meter_reg_c_ids;
/* Bitmap of the reg_Cs available for flow meter to use. */
uint32_t log_meter_aso_granularity:5;
/* Power of the minimum allocation granularity Object. */
uint32_t log_meter_aso_max_alloc:5;
/* Power of the maximum allocation granularity Object. */
uint32_t log_max_num_meter_aso:5;
/* Power of the maximum number of supported objects. */
};
struct mlx5_hca_vdpa_attr {
uint8_t virtio_queue_type;
uint32_t valid:1;
uint32_t desc_tunnel_offload_type:1;
uint32_t eth_frame_offload_type:1;
uint32_t virtio_version_1_0:1;
uint32_t tso_ipv4:1;
uint32_t tso_ipv6:1;
uint32_t tx_csum:1;
uint32_t rx_csum:1;
uint32_t event_mode:3;
uint32_t log_doorbell_stride:5;
uint32_t log_doorbell_bar_size:5;
uint32_t queue_counters_valid:1;
uint32_t vnet_modify_ext:1;
uint32_t virtio_net_q_addr_modify:1;
uint32_t virtio_q_index_modify:1;
uint32_t max_num_virtio_queues;
struct {
uint32_t a;
uint32_t b;
} umems[3];
uint64_t doorbell_bar_offset;
};
struct mlx5_hca_flow_attr {
uint32_t tunnel_header_0_1;
uint32_t tunnel_header_2_3;
};
/**
* Accumulate port PARSE_GRAPH_NODE capabilities from
* PARSE_GRAPH_NODE Capabilities and HCA Capabilities 2 tables
*/
__extension__
struct mlx5_hca_flex_attr {
uint32_t node_in;
uint32_t node_out;
uint16_t header_length_mode;
uint16_t sample_offset_mode;
uint8_t max_num_arc_in;
uint8_t max_num_arc_out;
uint8_t max_num_sample;
uint8_t max_num_prog_sample:5; /* From HCA CAP 2 */
uint8_t sample_id_in_out:1;
uint16_t max_base_header_length;
uint8_t max_sample_base_offset;
uint16_t max_next_header_offset;
uint8_t header_length_mask_width;
};
/* ISO C restricts enumerator values to range of 'int' */
__extension__
enum {
PARSE_GRAPH_NODE_CAP_SUPPORTED_PROTOCOL_HEAD = RTE_BIT32(1),
PARSE_GRAPH_NODE_CAP_SUPPORTED_PROTOCOL_MAC = RTE_BIT32(2),
PARSE_GRAPH_NODE_CAP_SUPPORTED_PROTOCOL_IP = RTE_BIT32(3),
PARSE_GRAPH_NODE_CAP_SUPPORTED_PROTOCOL_GRE = RTE_BIT32(4),
PARSE_GRAPH_NODE_CAP_SUPPORTED_PROTOCOL_UDP = RTE_BIT32(5),
PARSE_GRAPH_NODE_CAP_SUPPORTED_PROTOCOL_MPLS = RTE_BIT32(6),
PARSE_GRAPH_NODE_CAP_SUPPORTED_PROTOCOL_TCP = RTE_BIT32(7),
PARSE_GRAPH_NODE_CAP_SUPPORTED_PROTOCOL_VXLAN_GRE = RTE_BIT32(8),
PARSE_GRAPH_NODE_CAP_SUPPORTED_PROTOCOL_GENEVE = RTE_BIT32(9),
PARSE_GRAPH_NODE_CAP_SUPPORTED_PROTOCOL_IPSEC_ESP = RTE_BIT32(10),
PARSE_GRAPH_NODE_CAP_SUPPORTED_PROTOCOL_IPV4 = RTE_BIT32(11),
PARSE_GRAPH_NODE_CAP_SUPPORTED_PROTOCOL_IPV6 = RTE_BIT32(12),
PARSE_GRAPH_NODE_CAP_SUPPORTED_PROTOCOL_PROGRAMMABLE = RTE_BIT32(31)
};
enum {
PARSE_GRAPH_NODE_CAP_LENGTH_MODE_FIXED = RTE_BIT32(0),
PARSE_GRAPH_NODE_CAP_LENGTH_MODE_EXPLISIT_FIELD = RTE_BIT32(1),
PARSE_GRAPH_NODE_CAP_LENGTH_MODE_BITMASK_FIELD = RTE_BIT32(2)
};
/*
* DWORD shift is the base for calculating header_length_field_mask
* value in the MLX5_GRAPH_NODE_LEN_FIELD mode.
*/
#define MLX5_PARSE_GRAPH_NODE_HDR_LEN_SHIFT_DWORD 0x02
static inline uint32_t
mlx5_hca_parse_graph_node_base_hdr_len_mask
(const struct mlx5_hca_flex_attr *attr)
{
return (1 << attr->header_length_mask_width) - 1;
}
/* HCA supports this number of time periods for LRO. */
#define MLX5_LRO_NUM_SUPP_PERIODS 4
/* HCA attributes. */
struct mlx5_hca_attr {
uint32_t eswitch_manager:1;
uint32_t flow_counters_dump:1;
uint32_t mem_rq_rmp:1;
uint32_t log_max_rmp:5;
uint32_t log_max_rqt_size:5;
uint32_t parse_graph_flex_node:1;
uint8_t flow_counter_bulk_alloc_bitmap;
uint32_t eth_net_offloads:1;
uint32_t eth_virt:1;
uint32_t wqe_vlan_insert:1;
uint32_t csum_cap:1;
uint32_t vlan_cap:1;
uint32_t wqe_inline_mode:2;
uint32_t vport_inline_mode:3;
uint32_t tunnel_stateless_geneve_rx:1;
uint32_t geneve_max_opt_len:1; /* 0x0: 14DW, 0x1: 63DW */
uint32_t tunnel_stateless_gtp:1;
uint32_t max_lso_cap;
uint32_t scatter_fcs:1;
uint32_t lro_cap:1;
uint32_t tunnel_lro_gre:1;
uint32_t tunnel_lro_vxlan:1;
uint32_t tunnel_stateless_gre:1;
uint32_t tunnel_stateless_vxlan:1;
uint32_t swp:1;
uint32_t swp_csum:1;
uint32_t swp_lso:1;
uint32_t lro_max_msg_sz_mode:2;
uint32_t rq_delay_drop:1;
uint32_t lro_timer_supported_periods[MLX5_LRO_NUM_SUPP_PERIODS];
uint16_t lro_min_mss_size;
uint32_t flex_parser_protocols;
uint32_t max_geneve_tlv_options;
uint32_t max_geneve_tlv_option_data_len;
uint32_t hairpin:1;
uint32_t log_max_hairpin_queues:5;
uint32_t log_max_hairpin_wq_data_sz:5;
uint32_t log_max_hairpin_num_packets:5;
uint32_t hairpin_sq_wqe_bb_size:4;
uint32_t hairpin_sq_wq_in_host_mem:1;
uint32_t hairpin_data_buffer_locked:1;
uint32_t vhca_id:16;
uint32_t relaxed_ordering_write:1;
uint32_t relaxed_ordering_read:1;
uint32_t access_register_user:1;
uint32_t wqe_index_ignore:1;
uint32_t cross_channel:1;
uint32_t non_wire_sq:1; /* SQ with non-wire ops is supported. */
uint32_t log_max_static_sq_wq:5; /* Static WQE size SQ. */
uint32_t num_lag_ports:4; /* Number of ports can be bonded. */
uint32_t dev_freq_khz; /* Timestamp counter frequency, kHz. */
uint32_t scatter_fcs_w_decap_disable:1;
uint32_t flow_hit_aso:1; /* General obj type FLOW_HIT_ASO supported. */
uint32_t roce:1;
uint32_t wait_on_time:1;
uint32_t rq_ts_format:2;
uint32_t sq_ts_format:2;
uint32_t steering_format_version:4;
uint32_t qp_ts_format:2;
uint32_t regexp_params:1;
uint32_t regexp_version:3;
uint32_t reg_c_preserve:1;
uint32_t ct_offload:1; /* General obj type ASO CT offload supported. */
uint32_t crypto:1; /* Crypto engine is supported. */
uint32_t aes_xts:1; /* AES-XTS crypto is supported. */
uint32_t dek:1; /* General obj type DEK is supported. */
uint32_t import_kek:1; /* General obj type IMPORT_KEK supported. */
uint32_t credential:1; /* General obj type CREDENTIAL supported. */
uint32_t crypto_login:1; /* General obj type CRYPTO_LOGIN supported. */
uint32_t regexp_num_of_engines;
uint32_t log_max_ft_sampler_num:8;
uint32_t inner_ipv4_ihl:1;
uint32_t outer_ipv4_ihl:1;
uint32_t geneve_tlv_opt;
uint32_t cqe_compression:1;
uint32_t mini_cqe_resp_flow_tag:1;
uint32_t mini_cqe_resp_l3_l4_tag:1;
uint32_t pkt_integrity_match:1; /* 1 if HW supports integrity item */
struct mlx5_hca_qos_attr qos;
struct mlx5_hca_vdpa_attr vdpa;
struct mlx5_hca_flow_attr flow;
struct mlx5_hca_flex_attr flex;
int log_max_qp_sz;
int log_max_cq_sz;
int log_max_qp;
int log_max_cq;
uint32_t log_max_pd;
uint32_t log_max_mrw_sz;
uint32_t log_max_srq;
uint32_t log_max_srq_sz;
uint32_t rss_ind_tbl_cap;
uint32_t mmo_dma_sq_en:1;
uint32_t mmo_compress_sq_en:1;
uint32_t mmo_decompress_sq_en:1;
uint32_t mmo_dma_qp_en:1;
uint32_t mmo_compress_qp_en:1;
uint32_t mmo_decompress_qp_en:1;
uint32_t mmo_regex_qp_en:1;
uint32_t mmo_regex_sq_en:1;
uint32_t compress_min_block_size:4;
uint32_t log_max_mmo_dma:5;
uint32_t log_max_mmo_compress:5;
uint32_t log_max_mmo_decompress:5;
uint32_t umr_modify_entity_size_disabled:1;
uint32_t umr_indirect_mkey_disabled:1;
uint32_t log_min_stride_wqe_sz:5;
uint32_t esw_mgr_vport_id_valid:1; /* E-Switch Mgr vport ID is valid. */
uint32_t crypto_wrapped_import_method:1;
uint16_t esw_mgr_vport_id; /* E-Switch Mgr vport ID . */
uint16_t max_wqe_sz_sq;
uint32_t set_reg_c:8;
uint32_t nic_flow_table:1;
uint32_t modify_outer_ip_ecn:1;
union {
uint32_t max_flow_counter;
struct {
uint16_t max_flow_counter_15_0;
uint16_t max_flow_counter_31_16;
};
};
uint32_t flow_counter_bulk_log_max_alloc:5;
uint32_t flow_counter_bulk_log_granularity:5;
uint32_t alloc_flow_counter_pd:1;
uint32_t flow_counter_access_aso:1;
uint32_t flow_access_aso_opc_mod:8;
};
/* LAG Context. */
struct mlx5_devx_lag_context {
uint32_t fdb_selection_mode:1;
uint32_t port_select_mode:3;
uint32_t lag_state:3;
uint32_t tx_remap_affinity_1:4;
uint32_t tx_remap_affinity_2:4;
};
struct mlx5_devx_wq_attr {
uint32_t wq_type:4;
uint32_t wq_signature:1;
uint32_t end_padding_mode:2;
uint32_t cd_slave:1;
uint32_t hds_skip_first_sge:1;
uint32_t log2_hds_buf_size:3;
uint32_t page_offset:5;
uint32_t lwm:16;
uint32_t pd:24;
uint32_t uar_page:24;
uint64_t dbr_addr;
uint32_t hw_counter;
uint32_t sw_counter;
uint32_t log_wq_stride:4;
uint32_t log_wq_pg_sz:5;
uint32_t log_wq_sz:5;
uint32_t dbr_umem_valid:1;
uint32_t wq_umem_valid:1;
uint32_t log_hairpin_num_packets:5;
uint32_t log_hairpin_data_sz:5;
uint32_t single_wqe_log_num_of_strides:4;
uint32_t two_byte_shift_en:1;
uint32_t single_stride_log_num_of_bytes:3;
uint32_t dbr_umem_id;
uint32_t wq_umem_id;
uint64_t wq_umem_offset;
};
/* Create RQ attributes structure, used by create RQ operation. */
struct mlx5_devx_create_rq_attr {
uint32_t rlky:1;
uint32_t delay_drop_en:1;
uint32_t scatter_fcs:1;
uint32_t vsd:1;
uint32_t mem_rq_type:4;
uint32_t state:4;
uint32_t flush_in_error_en:1;
uint32_t hairpin:1;
uint32_t hairpin_data_buffer_type:3;
uint32_t ts_format:2;
uint32_t user_index:24;
uint32_t cqn:24;
uint32_t counter_set_id:8;
uint32_t rmpn:24;
struct mlx5_devx_wq_attr wq_attr;
};
/* Modify RQ attributes structure, used by modify RQ operation. */
struct mlx5_devx_modify_rq_attr {
uint32_t rqn:24;
uint32_t rq_state:4; /* Current RQ state. */
uint32_t state:4; /* Required RQ state. */
uint32_t scatter_fcs:1;
uint32_t vsd:1;
uint32_t counter_set_id:8;
uint32_t hairpin_peer_sq:24;
uint32_t hairpin_peer_vhca:16;
uint64_t modify_bitmask;
uint32_t lwm:16; /* Contained WQ lwm. */
};
/* Create RMP attributes structure, used by create RMP operation. */
struct mlx5_devx_create_rmp_attr {
uint32_t rsvd0:8;
uint32_t state:4;
uint32_t rsvd1:20;
uint32_t basic_cyclic_rcv_wqe:1;
uint32_t rsvd4:31;
uint32_t rsvd8[10];
struct mlx5_devx_wq_attr wq_attr;
};
struct mlx5_rx_hash_field_select {
uint32_t l3_prot_type:1;
uint32_t l4_prot_type:1;
uint32_t selected_fields:30;
};
/* TIR attributes structure, used by TIR operations. */
struct mlx5_devx_tir_attr {
uint32_t disp_type:4;
uint32_t lro_timeout_period_usecs:16;
uint32_t lro_enable_mask:4;
uint32_t lro_max_msg_sz:8;
uint32_t inline_rqn:24;
uint32_t rx_hash_symmetric:1;
uint32_t tunneled_offload_en:1;
uint32_t indirect_table:24;
uint32_t rx_hash_fn:4;
uint32_t self_lb_block:2;
uint32_t transport_domain:24;
uint8_t rx_hash_toeplitz_key[MLX5_RSS_HASH_KEY_LEN];
struct mlx5_rx_hash_field_select rx_hash_field_selector_outer;
struct mlx5_rx_hash_field_select rx_hash_field_selector_inner;
};
/* TIR attributes structure, used by TIR modify. */
struct mlx5_devx_modify_tir_attr {
uint32_t tirn:24;
uint64_t modify_bitmask;
struct mlx5_devx_tir_attr tir;
};
/* RQT attributes structure, used by RQT operations. */
struct mlx5_devx_rqt_attr {
uint8_t rq_type;
uint32_t rqt_max_size:16;
uint32_t rqt_actual_size:16;
uint32_t rq_list[];
};
/* TIS attributes structure. */
struct mlx5_devx_tis_attr {
uint32_t strict_lag_tx_port_affinity:1;
uint32_t tls_en:1;
uint32_t lag_tx_port_affinity:4;
uint32_t prio:4;
uint32_t transport_domain:24;
};
/* SQ attributes structure, used by SQ create operation. */
struct mlx5_devx_create_sq_attr {
uint32_t rlky:1;
uint32_t cd_master:1;
uint32_t fre:1;
uint32_t flush_in_error_en:1;
uint32_t allow_multi_pkt_send_wqe:1;
uint32_t min_wqe_inline_mode:3;
uint32_t state:4;
uint32_t reg_umr:1;
uint32_t allow_swp:1;
uint32_t hairpin:1;
uint32_t non_wire:1;
uint32_t static_sq_wq:1;
uint32_t ts_format:2;
uint32_t hairpin_wq_buffer_type:3;
uint32_t user_index:24;
uint32_t cqn:24;
uint32_t packet_pacing_rate_limit_index:16;
uint32_t tis_lst_sz:16;
uint32_t tis_num:24;
struct mlx5_devx_wq_attr wq_attr;
};
/* SQ attributes structure, used by SQ modify operation. */
struct mlx5_devx_modify_sq_attr {
uint32_t sq_state:4;
uint32_t state:4;
uint32_t hairpin_peer_rq:24;
uint32_t hairpin_peer_vhca:16;
};
/* CQ attributes structure, used by CQ operations. */
struct mlx5_devx_cq_attr {
uint32_t q_umem_valid:1;
uint32_t db_umem_valid:1;
uint32_t use_first_only:1;
uint32_t overrun_ignore:1;
uint32_t cqe_comp_en:1;
uint32_t mini_cqe_res_format:2;
uint32_t mini_cqe_res_format_ext:2;
uint32_t log_cq_size:5;
uint32_t log_page_size:5;
uint32_t uar_page_id;
uint32_t q_umem_id;
uint64_t q_umem_offset;
uint32_t db_umem_id;
uint64_t db_umem_offset;
uint32_t eqn;
uint64_t db_addr;
};
/* Virtq attributes structure, used by VIRTQ operations. */
struct mlx5_devx_virtq_attr {
uint16_t hw_available_index;
uint16_t hw_used_index;
uint16_t q_size;
uint32_t pd:24;
uint32_t virtio_version_1_0:1;
uint32_t tso_ipv4:1;
uint32_t tso_ipv6:1;
uint32_t tx_csum:1;
uint32_t rx_csum:1;
uint32_t event_mode:3;
uint32_t state:4;
uint32_t hw_latency_mode:2;
uint32_t hw_max_latency_us:12;
uint32_t hw_max_pending_comp:16;
uint32_t dirty_bitmap_dump_enable:1;
uint32_t dirty_bitmap_mkey;
uint32_t dirty_bitmap_size;
uint32_t mkey;
uint32_t qp_id;
uint32_t queue_index;
uint32_t tis_id;
uint32_t counters_obj_id;
uint64_t dirty_bitmap_addr;
uint64_t mod_fields_bitmap;
uint64_t desc_addr;
uint64_t used_addr;
uint64_t available_addr;
struct {
uint32_t id;
uint32_t size;
uint64_t offset;
} umems[3];
uint8_t error_type;
uint8_t q_type;
};
struct mlx5_devx_qp_attr {
uint32_t pd:24;
uint32_t uar_index:24;
uint32_t cqn:24;
uint32_t log_page_size:5;
uint32_t num_of_receive_wqes:17; /* Must be power of 2. */
uint32_t log_rq_stride:3;
uint32_t num_of_send_wqbbs:17; /* Must be power of 2. */
uint32_t ts_format:2;
uint32_t dbr_umem_valid:1;
uint32_t dbr_umem_id;
uint64_t dbr_address;
uint32_t wq_umem_id;
uint64_t wq_umem_offset;
uint32_t user_index:24;
uint32_t mmo:1;
};
struct mlx5_devx_virtio_q_couners_attr {
uint64_t received_desc;
uint64_t completed_desc;
uint32_t error_cqes;
uint32_t bad_desc_errors;
uint32_t exceed_max_chain;
uint32_t invalid_buffer;
};
/*
* graph flow match sample attributes structure,
* used by flex parser operations.
*/
struct mlx5_devx_match_sample_attr {
uint32_t flow_match_sample_en:1;
uint32_t flow_match_sample_field_offset:16;
uint32_t flow_match_sample_offset_mode:4;
uint32_t flow_match_sample_field_offset_mask;
uint32_t flow_match_sample_field_offset_shift:4;
uint32_t flow_match_sample_field_base_offset:8;
uint32_t flow_match_sample_tunnel_mode:3;
uint32_t flow_match_sample_field_id;
};
/* graph node arc attributes structure, used by flex parser operations. */
struct mlx5_devx_graph_arc_attr {
uint32_t compare_condition_value:16;
uint32_t start_inner_tunnel:1;
uint32_t arc_parse_graph_node:8;
uint32_t parse_graph_node_handle;
};
/* Maximal number of samples per graph node. */
#define MLX5_GRAPH_NODE_SAMPLE_NUM 8
/* Maximal number of input/output arcs per graph node. */
#define MLX5_GRAPH_NODE_ARC_NUM 8
/* parse graph node attributes structure, used by flex parser operations. */
struct mlx5_devx_graph_node_attr {
uint32_t modify_field_select;
uint32_t header_length_mode:4;
uint32_t header_length_base_value:16;
uint32_t header_length_field_shift:4;
uint32_t header_length_field_offset:16;
uint32_t header_length_field_mask;
struct mlx5_devx_match_sample_attr sample[MLX5_GRAPH_NODE_SAMPLE_NUM];
uint32_t next_header_field_offset:16;
uint32_t next_header_field_size:5;
struct mlx5_devx_graph_arc_attr in[MLX5_GRAPH_NODE_ARC_NUM];
struct mlx5_devx_graph_arc_attr out[MLX5_GRAPH_NODE_ARC_NUM];
};
/* Encryption key size is up to 1024 bit, 128 bytes. */
#define MLX5_CRYPTO_KEY_MAX_SIZE 128
struct mlx5_devx_dek_attr {
uint32_t key_size:4;
uint32_t has_keytag:1;
uint32_t key_purpose:4;
uint32_t pd:24;
uint64_t opaque;
uint8_t key[MLX5_CRYPTO_KEY_MAX_SIZE];
};
struct mlx5_devx_import_kek_attr {
uint64_t modify_field_select;
uint32_t state:8;
uint32_t key_size:4;
uint8_t key[MLX5_CRYPTO_KEY_MAX_SIZE];
};
#define MLX5_CRYPTO_CREDENTIAL_SIZE 48
struct mlx5_devx_credential_attr {
uint64_t modify_field_select;
uint32_t state:8;
uint32_t credential_role:8;
uint8_t credential[MLX5_CRYPTO_CREDENTIAL_SIZE];
};
struct mlx5_devx_crypto_login_attr {
uint64_t modify_field_select;
uint32_t credential_pointer:24;
uint32_t session_import_kek_ptr:24;
uint8_t credential[MLX5_CRYPTO_CREDENTIAL_SIZE];
};
/* mlx5_devx_cmds.c */
__rte_internal
struct mlx5_devx_obj *
mlx5_devx_cmd_flow_counter_alloc_general(void *ctx,
struct mlx5_devx_counter_attr *attr);
__rte_internal
struct mlx5_devx_obj *mlx5_devx_cmd_flow_counter_alloc(void *ctx,
uint32_t bulk_sz);
__rte_internal
int mlx5_devx_cmd_destroy(struct mlx5_devx_obj *obj);
__rte_internal
int mlx5_devx_cmd_flow_counter_query(struct mlx5_devx_obj *dcs,
int clear, uint32_t n_counters,
uint64_t *pkts, uint64_t *bytes,
uint32_t mkey, void *addr,
void *cmd_comp,
uint64_t async_id);
__rte_internal
int mlx5_devx_cmd_query_hca_attr(void *ctx,
struct mlx5_hca_attr *attr);
__rte_internal
struct mlx5_devx_obj *mlx5_devx_cmd_mkey_create(void *ctx,
struct mlx5_devx_mkey_attr *attr);
__rte_internal
int mlx5_devx_get_out_command_status(void *out);
__rte_internal
int mlx5_devx_cmd_qp_query_tis_td(void *qp, uint32_t tis_num,
uint32_t *tis_td);
__rte_internal
struct mlx5_devx_obj *mlx5_devx_cmd_create_rq(void *ctx,
struct mlx5_devx_create_rq_attr *rq_attr,
int socket);
__rte_internal
int mlx5_devx_cmd_modify_rq(struct mlx5_devx_obj *rq,
struct mlx5_devx_modify_rq_attr *rq_attr);
__rte_internal
struct mlx5_devx_obj *mlx5_devx_cmd_create_rmp(void *ctx,
struct mlx5_devx_create_rmp_attr *rq_attr, int socket);
__rte_internal
struct mlx5_devx_obj *mlx5_devx_cmd_create_tir(void *ctx,
struct mlx5_devx_tir_attr *tir_attr);
__rte_internal
struct mlx5_devx_obj *mlx5_devx_cmd_create_rqt(void *ctx,
struct mlx5_devx_rqt_attr *rqt_attr);
__rte_internal
struct mlx5_devx_obj *mlx5_devx_cmd_create_sq(void *ctx,
struct mlx5_devx_create_sq_attr *sq_attr);
__rte_internal
int mlx5_devx_cmd_modify_sq(struct mlx5_devx_obj *sq,
struct mlx5_devx_modify_sq_attr *sq_attr);
__rte_internal
struct mlx5_devx_obj *mlx5_devx_cmd_create_tis(void *ctx,
struct mlx5_devx_tis_attr *tis_attr);
__rte_internal
struct mlx5_devx_obj *mlx5_devx_cmd_create_td(void *ctx);
__rte_internal
int mlx5_devx_cmd_flow_dump(void *fdb_domain, void *rx_domain, void *tx_domain,
FILE *file);
__rte_internal
int mlx5_devx_cmd_flow_single_dump(void *rule, FILE *file);
__rte_internal
struct mlx5_devx_obj *mlx5_devx_cmd_create_cq(void *ctx,
struct mlx5_devx_cq_attr *attr);
__rte_internal
struct mlx5_devx_obj *mlx5_devx_cmd_create_virtq(void *ctx,
struct mlx5_devx_virtq_attr *attr);
__rte_internal
int mlx5_devx_cmd_modify_virtq(struct mlx5_devx_obj *virtq_obj,
struct mlx5_devx_virtq_attr *attr);
__rte_internal
int mlx5_devx_cmd_query_virtq(struct mlx5_devx_obj *virtq_obj,
struct mlx5_devx_virtq_attr *attr);
__rte_internal
struct mlx5_devx_obj *mlx5_devx_cmd_create_qp(void *ctx,
struct mlx5_devx_qp_attr *attr);
__rte_internal
int mlx5_devx_cmd_modify_qp_state(struct mlx5_devx_obj *qp,
uint32_t qp_st_mod_op, uint32_t remote_qp_id);
__rte_internal
int mlx5_devx_cmd_modify_rqt(struct mlx5_devx_obj *rqt,
struct mlx5_devx_rqt_attr *rqt_attr);
__rte_internal
int mlx5_devx_cmd_modify_tir(struct mlx5_devx_obj *tir,
struct mlx5_devx_modify_tir_attr *tir_attr);
__rte_internal
int mlx5_devx_cmd_query_parse_samples(struct mlx5_devx_obj *flex_obj,
uint32_t ids[], uint32_t num);
__rte_internal
struct mlx5_devx_obj *
mlx5_devx_cmd_create_flex_parser(void *ctx,
struct mlx5_devx_graph_node_attr *data);
__rte_internal
int mlx5_devx_cmd_register_read(void *ctx, uint16_t reg_id,
uint32_t arg, uint32_t *data, uint32_t dw_cnt);
__rte_internal
int mlx5_devx_cmd_register_write(void *ctx, uint16_t reg_id,
uint32_t arg, uint32_t *data, uint32_t dw_cnt);
__rte_internal
struct mlx5_devx_obj *
mlx5_devx_cmd_create_geneve_tlv_option(void *ctx,
uint16_t class, uint8_t type, uint8_t len);
/**
* Create virtio queue counters object DevX API.
*
* @param[in] ctx
* Device context.
* @return
* The DevX object created, NULL otherwise and rte_errno is set.
*/
__rte_internal
struct mlx5_devx_obj *mlx5_devx_cmd_create_virtio_q_counters(void *ctx);
/**
* Query virtio queue counters object using DevX API.
*
* @param[in] couners_obj
* Pointer to virtq object structure.
* @param [in/out] attr
* Pointer to virtio queue counters attributes structure.
*
* @return
* 0 on success, a negative errno value otherwise and rte_errno is set.
*/
__rte_internal
int mlx5_devx_cmd_query_virtio_q_counters(struct mlx5_devx_obj *couners_obj,
struct mlx5_devx_virtio_q_couners_attr *attr);
__rte_internal
struct mlx5_devx_obj *mlx5_devx_cmd_create_flow_hit_aso_obj(void *ctx,
uint32_t pd);
__rte_internal
struct mlx5_devx_obj *mlx5_devx_cmd_alloc_pd(void *ctx);
__rte_internal
int mlx5_devx_cmd_wq_query(void *wq, uint32_t *counter_set_id);
__rte_internal
struct mlx5_devx_obj *mlx5_devx_cmd_queue_counter_alloc(void *ctx);
__rte_internal
int mlx5_devx_cmd_queue_counter_query(struct mlx5_devx_obj *dcs, int clear,
uint32_t *out_of_buffers);
__rte_internal
struct mlx5_devx_obj *mlx5_devx_cmd_create_conn_track_offload_obj(void *ctx,
uint32_t pd, uint32_t log_obj_size);
/**
* Create general object of type FLOW_METER_ASO using DevX API..
*
* @param[in] ctx
* Device context.
* @param [in] pd
* PD value to associate the FLOW_METER_ASO object with.
* @param [in] log_obj_size
* log_obj_size define to allocate number of 2 * meters
* in one FLOW_METER_ASO object.
*
* @return
* The DevX object created, NULL otherwise and rte_errno is set.
*/
__rte_internal
struct mlx5_devx_obj *mlx5_devx_cmd_create_flow_meter_aso_obj(void *ctx,
uint32_t pd, uint32_t log_obj_size);
__rte_internal
struct mlx5_devx_obj *
mlx5_devx_cmd_create_dek_obj(void *ctx, struct mlx5_devx_dek_attr *attr);
__rte_internal
struct mlx5_devx_obj *
mlx5_devx_cmd_create_import_kek_obj(void *ctx,
struct mlx5_devx_import_kek_attr *attr);
__rte_internal
struct mlx5_devx_obj *
mlx5_devx_cmd_create_credential_obj(void *ctx,
struct mlx5_devx_credential_attr *attr);
__rte_internal
struct mlx5_devx_obj *
mlx5_devx_cmd_create_crypto_login_obj(void *ctx,
struct mlx5_devx_crypto_login_attr *attr);
__rte_internal
int
mlx5_devx_cmd_query_lag(void *ctx,
struct mlx5_devx_lag_context *lag_ctx);
#endif /* RTE_PMD_MLX5_DEVX_CMDS_H_ */