numam-dpdk/lib/port/rte_port_ring.c
Bruce Richardson 99a2dd955f lib: remove librte_ prefix from directory names
There is no reason for the DPDK libraries to all have 'librte_' prefix on
the directory names. This prefix makes the directory names longer and also
makes it awkward to add features referring to individual libraries in the
build - should the lib names be specified with or without the prefix.
Therefore, we can just remove the library prefix and use the library's
unique name as the directory name, i.e. 'eal' rather than 'librte_eal'

Signed-off-by: Bruce Richardson <bruce.richardson@intel.com>
2021-04-21 14:04:09 +02:00

788 lines
18 KiB
C

/* SPDX-License-Identifier: BSD-3-Clause
* Copyright(c) 2010-2014 Intel Corporation
*/
#include <string.h>
#include <stdint.h>
#include <rte_mbuf.h>
#include <rte_ring.h>
#include <rte_malloc.h>
#include "rte_port_ring.h"
/*
* Port RING Reader
*/
#ifdef RTE_PORT_STATS_COLLECT
#define RTE_PORT_RING_READER_STATS_PKTS_IN_ADD(port, val) \
port->stats.n_pkts_in += val
#define RTE_PORT_RING_READER_STATS_PKTS_DROP_ADD(port, val) \
port->stats.n_pkts_drop += val
#else
#define RTE_PORT_RING_READER_STATS_PKTS_IN_ADD(port, val)
#define RTE_PORT_RING_READER_STATS_PKTS_DROP_ADD(port, val)
#endif
struct rte_port_ring_reader {
struct rte_port_in_stats stats;
struct rte_ring *ring;
};
static void *
rte_port_ring_reader_create_internal(void *params, int socket_id,
uint32_t is_multi)
{
struct rte_port_ring_reader_params *conf =
params;
struct rte_port_ring_reader *port;
/* Check input parameters */
if ((conf == NULL) ||
(conf->ring == NULL) ||
(rte_ring_is_cons_single(conf->ring) && is_multi) ||
(!rte_ring_is_cons_single(conf->ring) && !is_multi)) {
RTE_LOG(ERR, PORT, "%s: Invalid Parameters\n", __func__);
return NULL;
}
/* Memory allocation */
port = rte_zmalloc_socket("PORT", sizeof(*port),
RTE_CACHE_LINE_SIZE, socket_id);
if (port == NULL) {
RTE_LOG(ERR, PORT, "%s: Failed to allocate port\n", __func__);
return NULL;
}
/* Initialization */
port->ring = conf->ring;
return port;
}
static void *
rte_port_ring_reader_create(void *params, int socket_id)
{
return rte_port_ring_reader_create_internal(params, socket_id, 0);
}
static void *
rte_port_ring_multi_reader_create(void *params, int socket_id)
{
return rte_port_ring_reader_create_internal(params, socket_id, 1);
}
static int
rte_port_ring_reader_rx(void *port, struct rte_mbuf **pkts, uint32_t n_pkts)
{
struct rte_port_ring_reader *p = port;
uint32_t nb_rx;
nb_rx = rte_ring_sc_dequeue_burst(p->ring, (void **) pkts,
n_pkts, NULL);
RTE_PORT_RING_READER_STATS_PKTS_IN_ADD(p, nb_rx);
return nb_rx;
}
static int
rte_port_ring_multi_reader_rx(void *port, struct rte_mbuf **pkts,
uint32_t n_pkts)
{
struct rte_port_ring_reader *p = port;
uint32_t nb_rx;
nb_rx = rte_ring_mc_dequeue_burst(p->ring, (void **) pkts,
n_pkts, NULL);
RTE_PORT_RING_READER_STATS_PKTS_IN_ADD(p, nb_rx);
return nb_rx;
}
static int
rte_port_ring_reader_free(void *port)
{
if (port == NULL) {
RTE_LOG(ERR, PORT, "%s: port is NULL\n", __func__);
return -EINVAL;
}
rte_free(port);
return 0;
}
static int
rte_port_ring_reader_stats_read(void *port,
struct rte_port_in_stats *stats, int clear)
{
struct rte_port_ring_reader *p =
port;
if (stats != NULL)
memcpy(stats, &p->stats, sizeof(p->stats));
if (clear)
memset(&p->stats, 0, sizeof(p->stats));
return 0;
}
/*
* Port RING Writer
*/
#ifdef RTE_PORT_STATS_COLLECT
#define RTE_PORT_RING_WRITER_STATS_PKTS_IN_ADD(port, val) \
port->stats.n_pkts_in += val
#define RTE_PORT_RING_WRITER_STATS_PKTS_DROP_ADD(port, val) \
port->stats.n_pkts_drop += val
#else
#define RTE_PORT_RING_WRITER_STATS_PKTS_IN_ADD(port, val)
#define RTE_PORT_RING_WRITER_STATS_PKTS_DROP_ADD(port, val)
#endif
struct rte_port_ring_writer {
struct rte_port_out_stats stats;
struct rte_mbuf *tx_buf[2 * RTE_PORT_IN_BURST_SIZE_MAX];
struct rte_ring *ring;
uint32_t tx_burst_sz;
uint32_t tx_buf_count;
uint64_t bsz_mask;
uint32_t is_multi;
};
static void *
rte_port_ring_writer_create_internal(void *params, int socket_id,
uint32_t is_multi)
{
struct rte_port_ring_writer_params *conf =
params;
struct rte_port_ring_writer *port;
/* Check input parameters */
if ((conf == NULL) ||
(conf->ring == NULL) ||
(rte_ring_is_prod_single(conf->ring) && is_multi) ||
(!rte_ring_is_prod_single(conf->ring) && !is_multi) ||
(conf->tx_burst_sz > RTE_PORT_IN_BURST_SIZE_MAX)) {
RTE_LOG(ERR, PORT, "%s: Invalid Parameters\n", __func__);
return NULL;
}
/* Memory allocation */
port = rte_zmalloc_socket("PORT", sizeof(*port),
RTE_CACHE_LINE_SIZE, socket_id);
if (port == NULL) {
RTE_LOG(ERR, PORT, "%s: Failed to allocate port\n", __func__);
return NULL;
}
/* Initialization */
port->ring = conf->ring;
port->tx_burst_sz = conf->tx_burst_sz;
port->tx_buf_count = 0;
port->bsz_mask = 1LLU << (conf->tx_burst_sz - 1);
port->is_multi = is_multi;
return port;
}
static void *
rte_port_ring_writer_create(void *params, int socket_id)
{
return rte_port_ring_writer_create_internal(params, socket_id, 0);
}
static void *
rte_port_ring_multi_writer_create(void *params, int socket_id)
{
return rte_port_ring_writer_create_internal(params, socket_id, 1);
}
static inline void
send_burst(struct rte_port_ring_writer *p)
{
uint32_t nb_tx;
nb_tx = rte_ring_sp_enqueue_burst(p->ring, (void **)p->tx_buf,
p->tx_buf_count, NULL);
RTE_PORT_RING_WRITER_STATS_PKTS_DROP_ADD(p, p->tx_buf_count - nb_tx);
for ( ; nb_tx < p->tx_buf_count; nb_tx++)
rte_pktmbuf_free(p->tx_buf[nb_tx]);
p->tx_buf_count = 0;
}
static inline void
send_burst_mp(struct rte_port_ring_writer *p)
{
uint32_t nb_tx;
nb_tx = rte_ring_mp_enqueue_burst(p->ring, (void **)p->tx_buf,
p->tx_buf_count, NULL);
RTE_PORT_RING_WRITER_STATS_PKTS_DROP_ADD(p, p->tx_buf_count - nb_tx);
for ( ; nb_tx < p->tx_buf_count; nb_tx++)
rte_pktmbuf_free(p->tx_buf[nb_tx]);
p->tx_buf_count = 0;
}
static int
rte_port_ring_writer_tx(void *port, struct rte_mbuf *pkt)
{
struct rte_port_ring_writer *p = port;
p->tx_buf[p->tx_buf_count++] = pkt;
RTE_PORT_RING_WRITER_STATS_PKTS_IN_ADD(p, 1);
if (p->tx_buf_count >= p->tx_burst_sz)
send_burst(p);
return 0;
}
static int
rte_port_ring_multi_writer_tx(void *port, struct rte_mbuf *pkt)
{
struct rte_port_ring_writer *p = port;
p->tx_buf[p->tx_buf_count++] = pkt;
RTE_PORT_RING_WRITER_STATS_PKTS_IN_ADD(p, 1);
if (p->tx_buf_count >= p->tx_burst_sz)
send_burst_mp(p);
return 0;
}
static __rte_always_inline int
rte_port_ring_writer_tx_bulk_internal(void *port,
struct rte_mbuf **pkts,
uint64_t pkts_mask,
uint32_t is_multi)
{
struct rte_port_ring_writer *p =
port;
uint64_t bsz_mask = p->bsz_mask;
uint32_t tx_buf_count = p->tx_buf_count;
uint64_t expr = (pkts_mask & (pkts_mask + 1)) |
((pkts_mask & bsz_mask) ^ bsz_mask);
if (expr == 0) {
uint64_t n_pkts = __builtin_popcountll(pkts_mask);
uint32_t n_pkts_ok;
if (tx_buf_count) {
if (is_multi)
send_burst_mp(p);
else
send_burst(p);
}
RTE_PORT_RING_WRITER_STATS_PKTS_IN_ADD(p, n_pkts);
if (is_multi)
n_pkts_ok = rte_ring_mp_enqueue_burst(p->ring,
(void **)pkts, n_pkts, NULL);
else
n_pkts_ok = rte_ring_sp_enqueue_burst(p->ring,
(void **)pkts, n_pkts, NULL);
RTE_PORT_RING_WRITER_STATS_PKTS_DROP_ADD(p, n_pkts - n_pkts_ok);
for ( ; n_pkts_ok < n_pkts; n_pkts_ok++) {
struct rte_mbuf *pkt = pkts[n_pkts_ok];
rte_pktmbuf_free(pkt);
}
} else {
for ( ; pkts_mask; ) {
uint32_t pkt_index = __builtin_ctzll(pkts_mask);
uint64_t pkt_mask = 1LLU << pkt_index;
struct rte_mbuf *pkt = pkts[pkt_index];
p->tx_buf[tx_buf_count++] = pkt;
RTE_PORT_RING_WRITER_STATS_PKTS_IN_ADD(p, 1);
pkts_mask &= ~pkt_mask;
}
p->tx_buf_count = tx_buf_count;
if (tx_buf_count >= p->tx_burst_sz) {
if (is_multi)
send_burst_mp(p);
else
send_burst(p);
}
}
return 0;
}
static int
rte_port_ring_writer_tx_bulk(void *port,
struct rte_mbuf **pkts,
uint64_t pkts_mask)
{
return rte_port_ring_writer_tx_bulk_internal(port, pkts, pkts_mask, 0);
}
static int
rte_port_ring_multi_writer_tx_bulk(void *port,
struct rte_mbuf **pkts,
uint64_t pkts_mask)
{
return rte_port_ring_writer_tx_bulk_internal(port, pkts, pkts_mask, 1);
}
static int
rte_port_ring_writer_flush(void *port)
{
struct rte_port_ring_writer *p = port;
if (p->tx_buf_count > 0)
send_burst(p);
return 0;
}
static int
rte_port_ring_multi_writer_flush(void *port)
{
struct rte_port_ring_writer *p = port;
if (p->tx_buf_count > 0)
send_burst_mp(p);
return 0;
}
static int
rte_port_ring_writer_free(void *port)
{
struct rte_port_ring_writer *p = port;
if (port == NULL) {
RTE_LOG(ERR, PORT, "%s: Port is NULL\n", __func__);
return -EINVAL;
}
if (p->is_multi)
rte_port_ring_multi_writer_flush(port);
else
rte_port_ring_writer_flush(port);
rte_free(port);
return 0;
}
static int
rte_port_ring_writer_stats_read(void *port,
struct rte_port_out_stats *stats, int clear)
{
struct rte_port_ring_writer *p =
port;
if (stats != NULL)
memcpy(stats, &p->stats, sizeof(p->stats));
if (clear)
memset(&p->stats, 0, sizeof(p->stats));
return 0;
}
/*
* Port RING Writer Nodrop
*/
#ifdef RTE_PORT_STATS_COLLECT
#define RTE_PORT_RING_WRITER_NODROP_STATS_PKTS_IN_ADD(port, val) \
port->stats.n_pkts_in += val
#define RTE_PORT_RING_WRITER_NODROP_STATS_PKTS_DROP_ADD(port, val) \
port->stats.n_pkts_drop += val
#else
#define RTE_PORT_RING_WRITER_NODROP_STATS_PKTS_IN_ADD(port, val)
#define RTE_PORT_RING_WRITER_NODROP_STATS_PKTS_DROP_ADD(port, val)
#endif
struct rte_port_ring_writer_nodrop {
struct rte_port_out_stats stats;
struct rte_mbuf *tx_buf[2 * RTE_PORT_IN_BURST_SIZE_MAX];
struct rte_ring *ring;
uint32_t tx_burst_sz;
uint32_t tx_buf_count;
uint64_t bsz_mask;
uint64_t n_retries;
uint32_t is_multi;
};
static void *
rte_port_ring_writer_nodrop_create_internal(void *params, int socket_id,
uint32_t is_multi)
{
struct rte_port_ring_writer_nodrop_params *conf =
params;
struct rte_port_ring_writer_nodrop *port;
/* Check input parameters */
if ((conf == NULL) ||
(conf->ring == NULL) ||
(rte_ring_is_prod_single(conf->ring) && is_multi) ||
(!rte_ring_is_prod_single(conf->ring) && !is_multi) ||
(conf->tx_burst_sz > RTE_PORT_IN_BURST_SIZE_MAX)) {
RTE_LOG(ERR, PORT, "%s: Invalid Parameters\n", __func__);
return NULL;
}
/* Memory allocation */
port = rte_zmalloc_socket("PORT", sizeof(*port),
RTE_CACHE_LINE_SIZE, socket_id);
if (port == NULL) {
RTE_LOG(ERR, PORT, "%s: Failed to allocate port\n", __func__);
return NULL;
}
/* Initialization */
port->ring = conf->ring;
port->tx_burst_sz = conf->tx_burst_sz;
port->tx_buf_count = 0;
port->bsz_mask = 1LLU << (conf->tx_burst_sz - 1);
port->is_multi = is_multi;
/*
* When n_retries is 0 it means that we should wait for every packet to
* send no matter how many retries should it take. To limit number of
* branches in fast path, we use UINT64_MAX instead of branching.
*/
port->n_retries = (conf->n_retries == 0) ? UINT64_MAX : conf->n_retries;
return port;
}
static void *
rte_port_ring_writer_nodrop_create(void *params, int socket_id)
{
return rte_port_ring_writer_nodrop_create_internal(params, socket_id, 0);
}
static void *
rte_port_ring_multi_writer_nodrop_create(void *params, int socket_id)
{
return rte_port_ring_writer_nodrop_create_internal(params, socket_id, 1);
}
static inline void
send_burst_nodrop(struct rte_port_ring_writer_nodrop *p)
{
uint32_t nb_tx = 0, i;
nb_tx = rte_ring_sp_enqueue_burst(p->ring, (void **)p->tx_buf,
p->tx_buf_count, NULL);
/* We sent all the packets in a first try */
if (nb_tx >= p->tx_buf_count) {
p->tx_buf_count = 0;
return;
}
for (i = 0; i < p->n_retries; i++) {
nb_tx += rte_ring_sp_enqueue_burst(p->ring,
(void **) (p->tx_buf + nb_tx),
p->tx_buf_count - nb_tx, NULL);
/* We sent all the packets in more than one try */
if (nb_tx >= p->tx_buf_count) {
p->tx_buf_count = 0;
return;
}
}
/* We didn't send the packets in maximum allowed attempts */
RTE_PORT_RING_WRITER_NODROP_STATS_PKTS_DROP_ADD(p, p->tx_buf_count - nb_tx);
for ( ; nb_tx < p->tx_buf_count; nb_tx++)
rte_pktmbuf_free(p->tx_buf[nb_tx]);
p->tx_buf_count = 0;
}
static inline void
send_burst_mp_nodrop(struct rte_port_ring_writer_nodrop *p)
{
uint32_t nb_tx = 0, i;
nb_tx = rte_ring_mp_enqueue_burst(p->ring, (void **)p->tx_buf,
p->tx_buf_count, NULL);
/* We sent all the packets in a first try */
if (nb_tx >= p->tx_buf_count) {
p->tx_buf_count = 0;
return;
}
for (i = 0; i < p->n_retries; i++) {
nb_tx += rte_ring_mp_enqueue_burst(p->ring,
(void **) (p->tx_buf + nb_tx),
p->tx_buf_count - nb_tx, NULL);
/* We sent all the packets in more than one try */
if (nb_tx >= p->tx_buf_count) {
p->tx_buf_count = 0;
return;
}
}
/* We didn't send the packets in maximum allowed attempts */
RTE_PORT_RING_WRITER_NODROP_STATS_PKTS_DROP_ADD(p, p->tx_buf_count - nb_tx);
for ( ; nb_tx < p->tx_buf_count; nb_tx++)
rte_pktmbuf_free(p->tx_buf[nb_tx]);
p->tx_buf_count = 0;
}
static int
rte_port_ring_writer_nodrop_tx(void *port, struct rte_mbuf *pkt)
{
struct rte_port_ring_writer_nodrop *p =
port;
p->tx_buf[p->tx_buf_count++] = pkt;
RTE_PORT_RING_WRITER_NODROP_STATS_PKTS_IN_ADD(p, 1);
if (p->tx_buf_count >= p->tx_burst_sz)
send_burst_nodrop(p);
return 0;
}
static int
rte_port_ring_multi_writer_nodrop_tx(void *port, struct rte_mbuf *pkt)
{
struct rte_port_ring_writer_nodrop *p =
port;
p->tx_buf[p->tx_buf_count++] = pkt;
RTE_PORT_RING_WRITER_NODROP_STATS_PKTS_IN_ADD(p, 1);
if (p->tx_buf_count >= p->tx_burst_sz)
send_burst_mp_nodrop(p);
return 0;
}
static __rte_always_inline int
rte_port_ring_writer_nodrop_tx_bulk_internal(void *port,
struct rte_mbuf **pkts,
uint64_t pkts_mask,
uint32_t is_multi)
{
struct rte_port_ring_writer_nodrop *p =
port;
uint64_t bsz_mask = p->bsz_mask;
uint32_t tx_buf_count = p->tx_buf_count;
uint64_t expr = (pkts_mask & (pkts_mask + 1)) |
((pkts_mask & bsz_mask) ^ bsz_mask);
if (expr == 0) {
uint64_t n_pkts = __builtin_popcountll(pkts_mask);
uint32_t n_pkts_ok;
if (tx_buf_count) {
if (is_multi)
send_burst_mp_nodrop(p);
else
send_burst_nodrop(p);
}
RTE_PORT_RING_WRITER_NODROP_STATS_PKTS_IN_ADD(p, n_pkts);
if (is_multi)
n_pkts_ok =
rte_ring_mp_enqueue_burst(p->ring,
(void **)pkts, n_pkts, NULL);
else
n_pkts_ok =
rte_ring_sp_enqueue_burst(p->ring,
(void **)pkts, n_pkts, NULL);
if (n_pkts_ok >= n_pkts)
return 0;
/*
* If we didn't manage to send all packets in single burst, move
* remaining packets to the buffer and call send burst.
*/
for (; n_pkts_ok < n_pkts; n_pkts_ok++) {
struct rte_mbuf *pkt = pkts[n_pkts_ok];
p->tx_buf[p->tx_buf_count++] = pkt;
}
if (is_multi)
send_burst_mp_nodrop(p);
else
send_burst_nodrop(p);
} else {
for ( ; pkts_mask; ) {
uint32_t pkt_index = __builtin_ctzll(pkts_mask);
uint64_t pkt_mask = 1LLU << pkt_index;
struct rte_mbuf *pkt = pkts[pkt_index];
p->tx_buf[tx_buf_count++] = pkt;
RTE_PORT_RING_WRITER_NODROP_STATS_PKTS_IN_ADD(p, 1);
pkts_mask &= ~pkt_mask;
}
p->tx_buf_count = tx_buf_count;
if (tx_buf_count >= p->tx_burst_sz) {
if (is_multi)
send_burst_mp_nodrop(p);
else
send_burst_nodrop(p);
}
}
return 0;
}
static int
rte_port_ring_writer_nodrop_tx_bulk(void *port,
struct rte_mbuf **pkts,
uint64_t pkts_mask)
{
return
rte_port_ring_writer_nodrop_tx_bulk_internal(port, pkts, pkts_mask, 0);
}
static int
rte_port_ring_multi_writer_nodrop_tx_bulk(void *port,
struct rte_mbuf **pkts,
uint64_t pkts_mask)
{
return
rte_port_ring_writer_nodrop_tx_bulk_internal(port, pkts, pkts_mask, 1);
}
static int
rte_port_ring_writer_nodrop_flush(void *port)
{
struct rte_port_ring_writer_nodrop *p =
port;
if (p->tx_buf_count > 0)
send_burst_nodrop(p);
return 0;
}
static int
rte_port_ring_multi_writer_nodrop_flush(void *port)
{
struct rte_port_ring_writer_nodrop *p =
port;
if (p->tx_buf_count > 0)
send_burst_mp_nodrop(p);
return 0;
}
static int
rte_port_ring_writer_nodrop_free(void *port)
{
struct rte_port_ring_writer_nodrop *p =
port;
if (port == NULL) {
RTE_LOG(ERR, PORT, "%s: Port is NULL\n", __func__);
return -EINVAL;
}
if (p->is_multi)
rte_port_ring_multi_writer_nodrop_flush(port);
else
rte_port_ring_writer_nodrop_flush(port);
rte_free(port);
return 0;
}
static int
rte_port_ring_writer_nodrop_stats_read(void *port,
struct rte_port_out_stats *stats, int clear)
{
struct rte_port_ring_writer_nodrop *p =
port;
if (stats != NULL)
memcpy(stats, &p->stats, sizeof(p->stats));
if (clear)
memset(&p->stats, 0, sizeof(p->stats));
return 0;
}
/*
* Summary of port operations
*/
struct rte_port_in_ops rte_port_ring_reader_ops = {
.f_create = rte_port_ring_reader_create,
.f_free = rte_port_ring_reader_free,
.f_rx = rte_port_ring_reader_rx,
.f_stats = rte_port_ring_reader_stats_read,
};
struct rte_port_out_ops rte_port_ring_writer_ops = {
.f_create = rte_port_ring_writer_create,
.f_free = rte_port_ring_writer_free,
.f_tx = rte_port_ring_writer_tx,
.f_tx_bulk = rte_port_ring_writer_tx_bulk,
.f_flush = rte_port_ring_writer_flush,
.f_stats = rte_port_ring_writer_stats_read,
};
struct rte_port_out_ops rte_port_ring_writer_nodrop_ops = {
.f_create = rte_port_ring_writer_nodrop_create,
.f_free = rte_port_ring_writer_nodrop_free,
.f_tx = rte_port_ring_writer_nodrop_tx,
.f_tx_bulk = rte_port_ring_writer_nodrop_tx_bulk,
.f_flush = rte_port_ring_writer_nodrop_flush,
.f_stats = rte_port_ring_writer_nodrop_stats_read,
};
struct rte_port_in_ops rte_port_ring_multi_reader_ops = {
.f_create = rte_port_ring_multi_reader_create,
.f_free = rte_port_ring_reader_free,
.f_rx = rte_port_ring_multi_reader_rx,
.f_stats = rte_port_ring_reader_stats_read,
};
struct rte_port_out_ops rte_port_ring_multi_writer_ops = {
.f_create = rte_port_ring_multi_writer_create,
.f_free = rte_port_ring_writer_free,
.f_tx = rte_port_ring_multi_writer_tx,
.f_tx_bulk = rte_port_ring_multi_writer_tx_bulk,
.f_flush = rte_port_ring_multi_writer_flush,
.f_stats = rte_port_ring_writer_stats_read,
};
struct rte_port_out_ops rte_port_ring_multi_writer_nodrop_ops = {
.f_create = rte_port_ring_multi_writer_nodrop_create,
.f_free = rte_port_ring_writer_nodrop_free,
.f_tx = rte_port_ring_multi_writer_nodrop_tx,
.f_tx_bulk = rte_port_ring_multi_writer_nodrop_tx_bulk,
.f_flush = rte_port_ring_multi_writer_nodrop_flush,
.f_stats = rte_port_ring_writer_nodrop_stats_read,
};