There is a confusion on setting max Rx packet length, this patch aims to clarify it. 'rte_eth_dev_configure()' API accepts max Rx packet size via 'uint32_t max_rx_pkt_len' field of the config struct 'struct rte_eth_conf'. Also 'rte_eth_dev_set_mtu()' API can be used to set the MTU, and result stored into '(struct rte_eth_dev)->data->mtu'. These two APIs are related but they work in a disconnected way, they store the set values in different variables which makes hard to figure out which one to use, also having two different method for a related functionality is confusing for the users. Other issues causing confusion is: * maximum transmission unit (MTU) is payload of the Ethernet frame. And 'max_rx_pkt_len' is the size of the Ethernet frame. Difference is Ethernet frame overhead, and this overhead may be different from device to device based on what device supports, like VLAN and QinQ. * 'max_rx_pkt_len' is only valid when application requested jumbo frame, which adds additional confusion and some APIs and PMDs already discards this documented behavior. * For the jumbo frame enabled case, 'max_rx_pkt_len' is an mandatory field, this adds configuration complexity for application. As solution, both APIs gets MTU as parameter, and both saves the result in same variable '(struct rte_eth_dev)->data->mtu'. For this 'max_rx_pkt_len' updated as 'mtu', and it is always valid independent from jumbo frame. For 'rte_eth_dev_configure()', 'dev->data->dev_conf.rxmode.mtu' is user request and it should be used only within configure function and result should be stored to '(struct rte_eth_dev)->data->mtu'. After that point both application and PMD uses MTU from this variable. When application doesn't provide an MTU during 'rte_eth_dev_configure()' default 'RTE_ETHER_MTU' value is used. Additional clarification done on scattered Rx configuration, in relation to MTU and Rx buffer size. MTU is used to configure the device for physical Rx/Tx size limitation, Rx buffer is where to store Rx packets, many PMDs use mbuf data buffer size as Rx buffer size. PMDs compare MTU against Rx buffer size to decide enabling scattered Rx or not. If scattered Rx is not supported by device, MTU bigger than Rx buffer size should fail. Signed-off-by: Ferruh Yigit <ferruh.yigit@intel.com> Acked-by: Ajit Khaparde <ajit.khaparde@broadcom.com> Acked-by: Somnath Kotur <somnath.kotur@broadcom.com> Acked-by: Huisong Li <lihuisong@huawei.com> Acked-by: Andrew Rybchenko <andrew.rybchenko@oktetlabs.ru> Acked-by: Konstantin Ananyev <konstantin.ananyev@intel.com> Acked-by: Rosen Xu <rosen.xu@intel.com> Acked-by: Hyong Youb Kim <hyonkim@cisco.com>
440 lines
14 KiB
C
440 lines
14 KiB
C
/* SPDX-License-Identifier: BSD-3-Clause
|
|
* Copyright(c) 2010-2014 Intel Corporation
|
|
*/
|
|
|
|
#include <stdint.h>
|
|
#include <memory.h>
|
|
|
|
#include <rte_log.h>
|
|
#include <rte_mbuf.h>
|
|
#include <rte_debug.h>
|
|
#include <rte_ethdev.h>
|
|
#include <rte_mempool.h>
|
|
#include <rte_sched.h>
|
|
#include <rte_cycles.h>
|
|
#include <rte_string_fns.h>
|
|
#include <rte_cfgfile.h>
|
|
|
|
#include "main.h"
|
|
#include "cfg_file.h"
|
|
|
|
uint32_t app_numa_mask = 0;
|
|
static uint32_t app_inited_port_mask = 0;
|
|
|
|
int app_pipe_to_profile[MAX_SCHED_SUBPORTS][MAX_SCHED_PIPES];
|
|
|
|
#define MAX_NAME_LEN 32
|
|
|
|
struct ring_conf ring_conf = {
|
|
.rx_size = APP_RX_DESC_DEFAULT,
|
|
.ring_size = APP_RING_SIZE,
|
|
.tx_size = APP_TX_DESC_DEFAULT,
|
|
};
|
|
|
|
struct burst_conf burst_conf = {
|
|
.rx_burst = MAX_PKT_RX_BURST,
|
|
.ring_burst = PKT_ENQUEUE,
|
|
.qos_dequeue = PKT_DEQUEUE,
|
|
.tx_burst = MAX_PKT_TX_BURST,
|
|
};
|
|
|
|
struct ring_thresh rx_thresh = {
|
|
.pthresh = RX_PTHRESH,
|
|
.hthresh = RX_HTHRESH,
|
|
.wthresh = RX_WTHRESH,
|
|
};
|
|
|
|
struct ring_thresh tx_thresh = {
|
|
.pthresh = TX_PTHRESH,
|
|
.hthresh = TX_HTHRESH,
|
|
.wthresh = TX_WTHRESH,
|
|
};
|
|
|
|
uint32_t nb_pfc;
|
|
const char *cfg_profile = NULL;
|
|
int mp_size = NB_MBUF;
|
|
struct flow_conf qos_conf[MAX_DATA_STREAMS];
|
|
|
|
static struct rte_eth_conf port_conf = {
|
|
.rxmode = {
|
|
.split_hdr_size = 0,
|
|
},
|
|
.txmode = {
|
|
.mq_mode = ETH_DCB_NONE,
|
|
},
|
|
};
|
|
|
|
static int
|
|
app_init_port(uint16_t portid, struct rte_mempool *mp)
|
|
{
|
|
int ret;
|
|
struct rte_eth_link link;
|
|
struct rte_eth_dev_info dev_info;
|
|
struct rte_eth_rxconf rx_conf;
|
|
struct rte_eth_txconf tx_conf;
|
|
uint16_t rx_size;
|
|
uint16_t tx_size;
|
|
struct rte_eth_conf local_port_conf = port_conf;
|
|
char link_status_text[RTE_ETH_LINK_MAX_STR_LEN];
|
|
|
|
/* check if port already initialized (multistream configuration) */
|
|
if (app_inited_port_mask & (1u << portid))
|
|
return 0;
|
|
|
|
rx_conf.rx_thresh.pthresh = rx_thresh.pthresh;
|
|
rx_conf.rx_thresh.hthresh = rx_thresh.hthresh;
|
|
rx_conf.rx_thresh.wthresh = rx_thresh.wthresh;
|
|
rx_conf.rx_free_thresh = 32;
|
|
rx_conf.rx_drop_en = 0;
|
|
rx_conf.rx_deferred_start = 0;
|
|
|
|
tx_conf.tx_thresh.pthresh = tx_thresh.pthresh;
|
|
tx_conf.tx_thresh.hthresh = tx_thresh.hthresh;
|
|
tx_conf.tx_thresh.wthresh = tx_thresh.wthresh;
|
|
tx_conf.tx_free_thresh = 0;
|
|
tx_conf.tx_rs_thresh = 0;
|
|
tx_conf.tx_deferred_start = 0;
|
|
|
|
/* init port */
|
|
RTE_LOG(INFO, APP, "Initializing port %"PRIu16"... ", portid);
|
|
fflush(stdout);
|
|
|
|
ret = rte_eth_dev_info_get(portid, &dev_info);
|
|
if (ret != 0)
|
|
rte_exit(EXIT_FAILURE,
|
|
"Error during getting device (port %u) info: %s\n",
|
|
portid, strerror(-ret));
|
|
|
|
if (dev_info.tx_offload_capa & DEV_TX_OFFLOAD_MBUF_FAST_FREE)
|
|
local_port_conf.txmode.offloads |=
|
|
DEV_TX_OFFLOAD_MBUF_FAST_FREE;
|
|
ret = rte_eth_dev_configure(portid, 1, 1, &local_port_conf);
|
|
if (ret < 0)
|
|
rte_exit(EXIT_FAILURE,
|
|
"Cannot configure device: err=%d, port=%u\n",
|
|
ret, portid);
|
|
|
|
rx_size = ring_conf.rx_size;
|
|
tx_size = ring_conf.tx_size;
|
|
ret = rte_eth_dev_adjust_nb_rx_tx_desc(portid, &rx_size, &tx_size);
|
|
if (ret < 0)
|
|
rte_exit(EXIT_FAILURE,
|
|
"rte_eth_dev_adjust_nb_rx_tx_desc: err=%d,port=%u\n",
|
|
ret, portid);
|
|
ring_conf.rx_size = rx_size;
|
|
ring_conf.tx_size = tx_size;
|
|
|
|
/* init one RX queue */
|
|
fflush(stdout);
|
|
rx_conf.offloads = local_port_conf.rxmode.offloads;
|
|
ret = rte_eth_rx_queue_setup(portid, 0, (uint16_t)ring_conf.rx_size,
|
|
rte_eth_dev_socket_id(portid), &rx_conf, mp);
|
|
if (ret < 0)
|
|
rte_exit(EXIT_FAILURE,
|
|
"rte_eth_tx_queue_setup: err=%d, port=%u\n",
|
|
ret, portid);
|
|
|
|
/* init one TX queue */
|
|
fflush(stdout);
|
|
tx_conf.offloads = local_port_conf.txmode.offloads;
|
|
ret = rte_eth_tx_queue_setup(portid, 0,
|
|
(uint16_t)ring_conf.tx_size, rte_eth_dev_socket_id(portid), &tx_conf);
|
|
if (ret < 0)
|
|
rte_exit(EXIT_FAILURE,
|
|
"rte_eth_tx_queue_setup: err=%d, port=%u queue=%d\n",
|
|
ret, portid, 0);
|
|
|
|
/* Start device */
|
|
ret = rte_eth_dev_start(portid);
|
|
if (ret < 0)
|
|
rte_exit(EXIT_FAILURE,
|
|
"rte_pmd_port_start: err=%d, port=%u\n",
|
|
ret, portid);
|
|
|
|
printf("done: ");
|
|
|
|
/* get link status */
|
|
ret = rte_eth_link_get(portid, &link);
|
|
if (ret < 0)
|
|
rte_exit(EXIT_FAILURE,
|
|
"rte_eth_link_get: err=%d, port=%u: %s\n",
|
|
ret, portid, rte_strerror(-ret));
|
|
|
|
rte_eth_link_to_str(link_status_text, sizeof(link_status_text), &link);
|
|
printf("%s\n", link_status_text);
|
|
|
|
ret = rte_eth_promiscuous_enable(portid);
|
|
if (ret != 0)
|
|
rte_exit(EXIT_FAILURE,
|
|
"rte_eth_promiscuous_enable: err=%s, port=%u\n",
|
|
rte_strerror(-ret), portid);
|
|
|
|
/* mark port as initialized */
|
|
app_inited_port_mask |= 1u << portid;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static struct rte_sched_pipe_params pipe_profiles[MAX_SCHED_PIPE_PROFILES] = {
|
|
{ /* Profile #0 */
|
|
.tb_rate = 305175,
|
|
.tb_size = 1000000,
|
|
|
|
.tc_rate = {305175, 305175, 305175, 305175, 305175, 305175,
|
|
305175, 305175, 305175, 305175, 305175, 305175, 305175},
|
|
.tc_period = 40,
|
|
#ifdef RTE_SCHED_SUBPORT_TC_OV
|
|
.tc_ov_weight = 1,
|
|
#endif
|
|
|
|
.wrr_weights = {1, 1, 1, 1},
|
|
},
|
|
};
|
|
|
|
static struct rte_sched_subport_profile_params
|
|
subport_profile[MAX_SCHED_SUBPORT_PROFILES] = {
|
|
{
|
|
.tb_rate = 1250000000,
|
|
.tb_size = 1000000,
|
|
.tc_rate = {1250000000, 1250000000, 1250000000, 1250000000,
|
|
1250000000, 1250000000, 1250000000, 1250000000, 1250000000,
|
|
1250000000, 1250000000, 1250000000, 1250000000},
|
|
.tc_period = 10,
|
|
},
|
|
};
|
|
|
|
struct rte_sched_subport_params subport_params[MAX_SCHED_SUBPORTS] = {
|
|
{
|
|
.n_pipes_per_subport_enabled = 4096,
|
|
.qsize = {64, 64, 64, 64, 64, 64, 64, 64, 64, 64, 64, 64, 64},
|
|
.pipe_profiles = pipe_profiles,
|
|
.n_pipe_profiles = sizeof(pipe_profiles) /
|
|
sizeof(struct rte_sched_pipe_params),
|
|
.n_max_pipe_profiles = MAX_SCHED_PIPE_PROFILES,
|
|
#ifdef RTE_SCHED_RED
|
|
.red_params = {
|
|
/* Traffic Class 0 Colors Green / Yellow / Red */
|
|
[0][0] = {.min_th = 48, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9},
|
|
[0][1] = {.min_th = 40, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9},
|
|
[0][2] = {.min_th = 32, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9},
|
|
|
|
/* Traffic Class 1 - Colors Green / Yellow / Red */
|
|
[1][0] = {.min_th = 48, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9},
|
|
[1][1] = {.min_th = 40, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9},
|
|
[1][2] = {.min_th = 32, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9},
|
|
|
|
/* Traffic Class 2 - Colors Green / Yellow / Red */
|
|
[2][0] = {.min_th = 48, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9},
|
|
[2][1] = {.min_th = 40, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9},
|
|
[2][2] = {.min_th = 32, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9},
|
|
|
|
/* Traffic Class 3 - Colors Green / Yellow / Red */
|
|
[3][0] = {.min_th = 48, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9},
|
|
[3][1] = {.min_th = 40, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9},
|
|
[3][2] = {.min_th = 32, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9},
|
|
|
|
/* Traffic Class 4 - Colors Green / Yellow / Red */
|
|
[4][0] = {.min_th = 48, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9},
|
|
[4][1] = {.min_th = 40, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9},
|
|
[4][2] = {.min_th = 32, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9},
|
|
|
|
/* Traffic Class 5 - Colors Green / Yellow / Red */
|
|
[5][0] = {.min_th = 48, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9},
|
|
[5][1] = {.min_th = 40, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9},
|
|
[5][2] = {.min_th = 32, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9},
|
|
|
|
/* Traffic Class 6 - Colors Green / Yellow / Red */
|
|
[6][0] = {.min_th = 48, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9},
|
|
[6][1] = {.min_th = 40, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9},
|
|
[6][2] = {.min_th = 32, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9},
|
|
|
|
/* Traffic Class 7 - Colors Green / Yellow / Red */
|
|
[7][0] = {.min_th = 48, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9},
|
|
[7][1] = {.min_th = 40, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9},
|
|
[7][2] = {.min_th = 32, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9},
|
|
|
|
/* Traffic Class 8 - Colors Green / Yellow / Red */
|
|
[8][0] = {.min_th = 48, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9},
|
|
[8][1] = {.min_th = 40, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9},
|
|
[8][2] = {.min_th = 32, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9},
|
|
|
|
/* Traffic Class 9 - Colors Green / Yellow / Red */
|
|
[9][0] = {.min_th = 48, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9},
|
|
[9][1] = {.min_th = 40, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9},
|
|
[9][2] = {.min_th = 32, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9},
|
|
|
|
/* Traffic Class 10 - Colors Green / Yellow / Red */
|
|
[10][0] = {.min_th = 48, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9},
|
|
[10][1] = {.min_th = 40, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9},
|
|
[10][2] = {.min_th = 32, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9},
|
|
|
|
/* Traffic Class 11 - Colors Green / Yellow / Red */
|
|
[11][0] = {.min_th = 48, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9},
|
|
[11][1] = {.min_th = 40, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9},
|
|
[11][2] = {.min_th = 32, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9},
|
|
|
|
/* Traffic Class 12 - Colors Green / Yellow / Red */
|
|
[12][0] = {.min_th = 48, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9},
|
|
[12][1] = {.min_th = 40, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9},
|
|
[12][2] = {.min_th = 32, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9},
|
|
},
|
|
#endif /* RTE_SCHED_RED */
|
|
},
|
|
};
|
|
|
|
struct rte_sched_port_params port_params = {
|
|
.name = "port_scheduler_0",
|
|
.socket = 0, /* computed */
|
|
.rate = 0, /* computed */
|
|
.mtu = 6 + 6 + 4 + 4 + 2 + 1500,
|
|
.frame_overhead = RTE_SCHED_FRAME_OVERHEAD_DEFAULT,
|
|
.n_subports_per_port = 1,
|
|
.n_subport_profiles = 1,
|
|
.subport_profiles = subport_profile,
|
|
.n_max_subport_profiles = MAX_SCHED_SUBPORT_PROFILES,
|
|
.n_pipes_per_subport = MAX_SCHED_PIPES,
|
|
};
|
|
|
|
static struct rte_sched_port *
|
|
app_init_sched_port(uint32_t portid, uint32_t socketid)
|
|
{
|
|
static char port_name[32]; /* static as referenced from global port_params*/
|
|
struct rte_eth_link link;
|
|
struct rte_sched_port *port = NULL;
|
|
uint32_t pipe, subport;
|
|
int err;
|
|
|
|
err = rte_eth_link_get(portid, &link);
|
|
if (err < 0)
|
|
rte_exit(EXIT_FAILURE,
|
|
"rte_eth_link_get: err=%d, port=%u: %s\n",
|
|
err, portid, rte_strerror(-err));
|
|
|
|
port_params.socket = socketid;
|
|
port_params.rate = (uint64_t) link.link_speed * 1000 * 1000 / 8;
|
|
snprintf(port_name, sizeof(port_name), "port_%d", portid);
|
|
port_params.name = port_name;
|
|
|
|
port = rte_sched_port_config(&port_params);
|
|
if (port == NULL){
|
|
rte_exit(EXIT_FAILURE, "Unable to config sched port\n");
|
|
}
|
|
|
|
for (subport = 0; subport < port_params.n_subports_per_port; subport ++) {
|
|
err = rte_sched_subport_config(port, subport,
|
|
&subport_params[subport],
|
|
0);
|
|
if (err) {
|
|
rte_exit(EXIT_FAILURE, "Unable to config sched "
|
|
"subport %u, err=%d\n", subport, err);
|
|
}
|
|
|
|
uint32_t n_pipes_per_subport =
|
|
subport_params[subport].n_pipes_per_subport_enabled;
|
|
|
|
for (pipe = 0; pipe < n_pipes_per_subport; pipe++) {
|
|
if (app_pipe_to_profile[subport][pipe] != -1) {
|
|
err = rte_sched_pipe_config(port, subport, pipe,
|
|
app_pipe_to_profile[subport][pipe]);
|
|
if (err) {
|
|
rte_exit(EXIT_FAILURE, "Unable to config sched pipe %u "
|
|
"for profile %d, err=%d\n", pipe,
|
|
app_pipe_to_profile[subport][pipe], err);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
return port;
|
|
}
|
|
|
|
static int
|
|
app_load_cfg_profile(const char *profile)
|
|
{
|
|
if (profile == NULL)
|
|
return 0;
|
|
struct rte_cfgfile *file = rte_cfgfile_load(profile, 0);
|
|
if (file == NULL)
|
|
rte_exit(EXIT_FAILURE, "Cannot load configuration profile %s\n", profile);
|
|
|
|
cfg_load_port(file, &port_params);
|
|
cfg_load_subport(file, subport_params);
|
|
cfg_load_subport_profile(file, subport_profile);
|
|
cfg_load_pipe(file, pipe_profiles);
|
|
|
|
rte_cfgfile_close(file);
|
|
|
|
return 0;
|
|
}
|
|
|
|
int app_init(void)
|
|
{
|
|
uint32_t i;
|
|
char ring_name[MAX_NAME_LEN];
|
|
char pool_name[MAX_NAME_LEN];
|
|
|
|
if (rte_eth_dev_count_avail() == 0)
|
|
rte_exit(EXIT_FAILURE, "No Ethernet port - bye\n");
|
|
|
|
/* load configuration profile */
|
|
if (app_load_cfg_profile(cfg_profile) != 0)
|
|
rte_exit(EXIT_FAILURE, "Invalid configuration profile\n");
|
|
|
|
/* Initialize each active flow */
|
|
for(i = 0; i < nb_pfc; i++) {
|
|
uint32_t socket = rte_lcore_to_socket_id(qos_conf[i].rx_core);
|
|
struct rte_ring *ring;
|
|
|
|
snprintf(ring_name, MAX_NAME_LEN, "ring-%u-%u", i, qos_conf[i].rx_core);
|
|
ring = rte_ring_lookup(ring_name);
|
|
if (ring == NULL)
|
|
qos_conf[i].rx_ring = rte_ring_create(ring_name, ring_conf.ring_size,
|
|
socket, RING_F_SP_ENQ | RING_F_SC_DEQ);
|
|
else
|
|
qos_conf[i].rx_ring = ring;
|
|
|
|
snprintf(ring_name, MAX_NAME_LEN, "ring-%u-%u", i, qos_conf[i].tx_core);
|
|
ring = rte_ring_lookup(ring_name);
|
|
if (ring == NULL)
|
|
qos_conf[i].tx_ring = rte_ring_create(ring_name, ring_conf.ring_size,
|
|
socket, RING_F_SP_ENQ | RING_F_SC_DEQ);
|
|
else
|
|
qos_conf[i].tx_ring = ring;
|
|
|
|
|
|
/* create the mbuf pools for each RX Port */
|
|
snprintf(pool_name, MAX_NAME_LEN, "mbuf_pool%u", i);
|
|
qos_conf[i].mbuf_pool = rte_pktmbuf_pool_create(pool_name,
|
|
mp_size, burst_conf.rx_burst * 4, 0,
|
|
RTE_MBUF_DEFAULT_BUF_SIZE,
|
|
rte_eth_dev_socket_id(qos_conf[i].rx_port));
|
|
if (qos_conf[i].mbuf_pool == NULL)
|
|
rte_exit(EXIT_FAILURE, "Cannot init mbuf pool for socket %u\n", i);
|
|
|
|
app_init_port(qos_conf[i].rx_port, qos_conf[i].mbuf_pool);
|
|
app_init_port(qos_conf[i].tx_port, qos_conf[i].mbuf_pool);
|
|
|
|
qos_conf[i].sched_port = app_init_sched_port(qos_conf[i].tx_port, socket);
|
|
}
|
|
|
|
RTE_LOG(INFO, APP, "time stamp clock running at %" PRIu64 " Hz\n",
|
|
rte_get_timer_hz());
|
|
|
|
RTE_LOG(INFO, APP, "Ring sizes: NIC RX = %u, Mempool = %d SW queue = %u,"
|
|
"NIC TX = %u\n", ring_conf.rx_size, mp_size, ring_conf.ring_size,
|
|
ring_conf.tx_size);
|
|
|
|
RTE_LOG(INFO, APP, "Burst sizes: RX read = %hu, RX write = %hu,\n"
|
|
" Worker read/QoS enqueue = %hu,\n"
|
|
" QoS dequeue = %hu, Worker write = %hu\n",
|
|
burst_conf.rx_burst, burst_conf.ring_burst, burst_conf.ring_burst,
|
|
burst_conf.qos_dequeue, burst_conf.tx_burst);
|
|
|
|
RTE_LOG(INFO, APP, "NIC thresholds RX (p = %hhu, h = %hhu, w = %hhu),"
|
|
"TX (p = %hhu, h = %hhu, w = %hhu)\n",
|
|
rx_thresh.pthresh, rx_thresh.hthresh, rx_thresh.wthresh,
|
|
tx_thresh.pthresh, tx_thresh.hthresh, tx_thresh.wthresh);
|
|
|
|
return 0;
|
|
}
|