numam-dpdk/lib/ring/rte_ring.c
Sean Morrissey 30a1de105a lib: remove unneeded header includes
These header includes have been flagged by the iwyu_tool
and removed.

Signed-off-by: Sean Morrissey <sean.morrissey@intel.com>
2022-02-22 13:10:39 +01:00

422 lines
9.9 KiB
C

/* SPDX-License-Identifier: BSD-3-Clause
*
* Copyright (c) 2010-2015 Intel Corporation
* Copyright (c) 2007,2008 Kip Macy kmacy@freebsd.org
* All rights reserved.
* Derived from FreeBSD's bufring.h
* Used as BSD-3 Licensed with permission from Kip Macy.
*/
#include <stdio.h>
#include <string.h>
#include <stdint.h>
#include <inttypes.h>
#include <errno.h>
#include <sys/queue.h>
#include <rte_common.h>
#include <rte_log.h>
#include <rte_memzone.h>
#include <rte_malloc.h>
#include <rte_eal_memconfig.h>
#include <rte_errno.h>
#include <rte_string_fns.h>
#include <rte_tailq.h>
#include "rte_ring.h"
#include "rte_ring_elem.h"
TAILQ_HEAD(rte_ring_list, rte_tailq_entry);
static struct rte_tailq_elem rte_ring_tailq = {
.name = RTE_TAILQ_RING_NAME,
};
EAL_REGISTER_TAILQ(rte_ring_tailq)
/* mask of all valid flag values to ring_create() */
#define RING_F_MASK (RING_F_SP_ENQ | RING_F_SC_DEQ | RING_F_EXACT_SZ | \
RING_F_MP_RTS_ENQ | RING_F_MC_RTS_DEQ | \
RING_F_MP_HTS_ENQ | RING_F_MC_HTS_DEQ)
/* true if x is a power of 2 */
#define POWEROF2(x) ((((x)-1) & (x)) == 0)
/* by default set head/tail distance as 1/8 of ring capacity */
#define HTD_MAX_DEF 8
/* return the size of memory occupied by a ring */
ssize_t
rte_ring_get_memsize_elem(unsigned int esize, unsigned int count)
{
ssize_t sz;
/* Check if element size is a multiple of 4B */
if (esize % 4 != 0) {
RTE_LOG(ERR, RING, "element size is not a multiple of 4\n");
return -EINVAL;
}
/* count must be a power of 2 */
if ((!POWEROF2(count)) || (count > RTE_RING_SZ_MASK )) {
RTE_LOG(ERR, RING,
"Requested number of elements is invalid, must be power of 2, and not exceed %u\n",
RTE_RING_SZ_MASK);
return -EINVAL;
}
sz = sizeof(struct rte_ring) + (ssize_t)count * esize;
sz = RTE_ALIGN(sz, RTE_CACHE_LINE_SIZE);
return sz;
}
/* return the size of memory occupied by a ring */
ssize_t
rte_ring_get_memsize(unsigned int count)
{
return rte_ring_get_memsize_elem(sizeof(void *), count);
}
/*
* internal helper function to reset prod/cons head-tail values.
*/
static void
reset_headtail(void *p)
{
struct rte_ring_headtail *ht;
struct rte_ring_hts_headtail *ht_hts;
struct rte_ring_rts_headtail *ht_rts;
ht = p;
ht_hts = p;
ht_rts = p;
switch (ht->sync_type) {
case RTE_RING_SYNC_MT:
case RTE_RING_SYNC_ST:
ht->head = 0;
ht->tail = 0;
break;
case RTE_RING_SYNC_MT_RTS:
ht_rts->head.raw = 0;
ht_rts->tail.raw = 0;
break;
case RTE_RING_SYNC_MT_HTS:
ht_hts->ht.raw = 0;
break;
default:
/* unknown sync mode */
RTE_ASSERT(0);
}
}
void
rte_ring_reset(struct rte_ring *r)
{
reset_headtail(&r->prod);
reset_headtail(&r->cons);
}
/*
* helper function, calculates sync_type values for prod and cons
* based on input flags. Returns zero at success or negative
* errno value otherwise.
*/
static int
get_sync_type(uint32_t flags, enum rte_ring_sync_type *prod_st,
enum rte_ring_sync_type *cons_st)
{
static const uint32_t prod_st_flags =
(RING_F_SP_ENQ | RING_F_MP_RTS_ENQ | RING_F_MP_HTS_ENQ);
static const uint32_t cons_st_flags =
(RING_F_SC_DEQ | RING_F_MC_RTS_DEQ | RING_F_MC_HTS_DEQ);
switch (flags & prod_st_flags) {
case 0:
*prod_st = RTE_RING_SYNC_MT;
break;
case RING_F_SP_ENQ:
*prod_st = RTE_RING_SYNC_ST;
break;
case RING_F_MP_RTS_ENQ:
*prod_st = RTE_RING_SYNC_MT_RTS;
break;
case RING_F_MP_HTS_ENQ:
*prod_st = RTE_RING_SYNC_MT_HTS;
break;
default:
return -EINVAL;
}
switch (flags & cons_st_flags) {
case 0:
*cons_st = RTE_RING_SYNC_MT;
break;
case RING_F_SC_DEQ:
*cons_st = RTE_RING_SYNC_ST;
break;
case RING_F_MC_RTS_DEQ:
*cons_st = RTE_RING_SYNC_MT_RTS;
break;
case RING_F_MC_HTS_DEQ:
*cons_st = RTE_RING_SYNC_MT_HTS;
break;
default:
return -EINVAL;
}
return 0;
}
int
rte_ring_init(struct rte_ring *r, const char *name, unsigned int count,
unsigned int flags)
{
int ret;
/* compilation-time checks */
RTE_BUILD_BUG_ON((sizeof(struct rte_ring) &
RTE_CACHE_LINE_MASK) != 0);
RTE_BUILD_BUG_ON((offsetof(struct rte_ring, cons) &
RTE_CACHE_LINE_MASK) != 0);
RTE_BUILD_BUG_ON((offsetof(struct rte_ring, prod) &
RTE_CACHE_LINE_MASK) != 0);
RTE_BUILD_BUG_ON(offsetof(struct rte_ring_headtail, sync_type) !=
offsetof(struct rte_ring_hts_headtail, sync_type));
RTE_BUILD_BUG_ON(offsetof(struct rte_ring_headtail, tail) !=
offsetof(struct rte_ring_hts_headtail, ht.pos.tail));
RTE_BUILD_BUG_ON(offsetof(struct rte_ring_headtail, sync_type) !=
offsetof(struct rte_ring_rts_headtail, sync_type));
RTE_BUILD_BUG_ON(offsetof(struct rte_ring_headtail, tail) !=
offsetof(struct rte_ring_rts_headtail, tail.val.pos));
/* future proof flags, only allow supported values */
if (flags & ~RING_F_MASK) {
RTE_LOG(ERR, RING,
"Unsupported flags requested %#x\n", flags);
return -EINVAL;
}
/* init the ring structure */
memset(r, 0, sizeof(*r));
ret = strlcpy(r->name, name, sizeof(r->name));
if (ret < 0 || ret >= (int)sizeof(r->name))
return -ENAMETOOLONG;
r->flags = flags;
ret = get_sync_type(flags, &r->prod.sync_type, &r->cons.sync_type);
if (ret != 0)
return ret;
if (flags & RING_F_EXACT_SZ) {
r->size = rte_align32pow2(count + 1);
r->mask = r->size - 1;
r->capacity = count;
} else {
if ((!POWEROF2(count)) || (count > RTE_RING_SZ_MASK)) {
RTE_LOG(ERR, RING,
"Requested size is invalid, must be power of 2, and not exceed the size limit %u\n",
RTE_RING_SZ_MASK);
return -EINVAL;
}
r->size = count;
r->mask = count - 1;
r->capacity = r->mask;
}
/* set default values for head-tail distance */
if (flags & RING_F_MP_RTS_ENQ)
rte_ring_set_prod_htd_max(r, r->capacity / HTD_MAX_DEF);
if (flags & RING_F_MC_RTS_DEQ)
rte_ring_set_cons_htd_max(r, r->capacity / HTD_MAX_DEF);
return 0;
}
/* create the ring for a given element size */
struct rte_ring *
rte_ring_create_elem(const char *name, unsigned int esize, unsigned int count,
int socket_id, unsigned int flags)
{
char mz_name[RTE_MEMZONE_NAMESIZE];
struct rte_ring *r;
struct rte_tailq_entry *te;
const struct rte_memzone *mz;
ssize_t ring_size;
int mz_flags = 0;
struct rte_ring_list* ring_list = NULL;
const unsigned int requested_count = count;
int ret;
ring_list = RTE_TAILQ_CAST(rte_ring_tailq.head, rte_ring_list);
/* for an exact size ring, round up from count to a power of two */
if (flags & RING_F_EXACT_SZ)
count = rte_align32pow2(count + 1);
ring_size = rte_ring_get_memsize_elem(esize, count);
if (ring_size < 0) {
rte_errno = -ring_size;
return NULL;
}
ret = snprintf(mz_name, sizeof(mz_name), "%s%s",
RTE_RING_MZ_PREFIX, name);
if (ret < 0 || ret >= (int)sizeof(mz_name)) {
rte_errno = ENAMETOOLONG;
return NULL;
}
te = rte_zmalloc("RING_TAILQ_ENTRY", sizeof(*te), 0);
if (te == NULL) {
RTE_LOG(ERR, RING, "Cannot reserve memory for tailq\n");
rte_errno = ENOMEM;
return NULL;
}
rte_mcfg_tailq_write_lock();
/* reserve a memory zone for this ring. If we can't get rte_config or
* we are secondary process, the memzone_reserve function will set
* rte_errno for us appropriately - hence no check in this this function */
mz = rte_memzone_reserve_aligned(mz_name, ring_size, socket_id,
mz_flags, __alignof__(*r));
if (mz != NULL) {
r = mz->addr;
/* no need to check return value here, we already checked the
* arguments above */
rte_ring_init(r, name, requested_count, flags);
te->data = (void *) r;
r->memzone = mz;
TAILQ_INSERT_TAIL(ring_list, te, next);
} else {
r = NULL;
RTE_LOG(ERR, RING, "Cannot reserve memory\n");
rte_free(te);
}
rte_mcfg_tailq_write_unlock();
return r;
}
/* create the ring */
struct rte_ring *
rte_ring_create(const char *name, unsigned int count, int socket_id,
unsigned int flags)
{
return rte_ring_create_elem(name, sizeof(void *), count, socket_id,
flags);
}
/* free the ring */
void
rte_ring_free(struct rte_ring *r)
{
struct rte_ring_list *ring_list = NULL;
struct rte_tailq_entry *te;
if (r == NULL)
return;
/*
* Ring was not created with rte_ring_create,
* therefore, there is no memzone to free.
*/
if (r->memzone == NULL) {
RTE_LOG(ERR, RING,
"Cannot free ring, not created with rte_ring_create()\n");
return;
}
if (rte_memzone_free(r->memzone) != 0) {
RTE_LOG(ERR, RING, "Cannot free memory\n");
return;
}
ring_list = RTE_TAILQ_CAST(rte_ring_tailq.head, rte_ring_list);
rte_mcfg_tailq_write_lock();
/* find out tailq entry */
TAILQ_FOREACH(te, ring_list, next) {
if (te->data == (void *) r)
break;
}
if (te == NULL) {
rte_mcfg_tailq_write_unlock();
return;
}
TAILQ_REMOVE(ring_list, te, next);
rte_mcfg_tailq_write_unlock();
rte_free(te);
}
/* dump the status of the ring on the console */
void
rte_ring_dump(FILE *f, const struct rte_ring *r)
{
fprintf(f, "ring <%s>@%p\n", r->name, r);
fprintf(f, " flags=%x\n", r->flags);
fprintf(f, " size=%"PRIu32"\n", r->size);
fprintf(f, " capacity=%"PRIu32"\n", r->capacity);
fprintf(f, " ct=%"PRIu32"\n", r->cons.tail);
fprintf(f, " ch=%"PRIu32"\n", r->cons.head);
fprintf(f, " pt=%"PRIu32"\n", r->prod.tail);
fprintf(f, " ph=%"PRIu32"\n", r->prod.head);
fprintf(f, " used=%u\n", rte_ring_count(r));
fprintf(f, " avail=%u\n", rte_ring_free_count(r));
}
/* dump the status of all rings on the console */
void
rte_ring_list_dump(FILE *f)
{
const struct rte_tailq_entry *te;
struct rte_ring_list *ring_list;
ring_list = RTE_TAILQ_CAST(rte_ring_tailq.head, rte_ring_list);
rte_mcfg_tailq_read_lock();
TAILQ_FOREACH(te, ring_list, next) {
rte_ring_dump(f, (struct rte_ring *) te->data);
}
rte_mcfg_tailq_read_unlock();
}
/* search a ring from its name */
struct rte_ring *
rte_ring_lookup(const char *name)
{
struct rte_tailq_entry *te;
struct rte_ring *r = NULL;
struct rte_ring_list *ring_list;
ring_list = RTE_TAILQ_CAST(rte_ring_tailq.head, rte_ring_list);
rte_mcfg_tailq_read_lock();
TAILQ_FOREACH(te, ring_list, next) {
r = (struct rte_ring *) te->data;
if (strncmp(name, r->name, RTE_RING_NAMESIZE) == 0)
break;
}
rte_mcfg_tailq_read_unlock();
if (te == NULL) {
rte_errno = ENOENT;
return NULL;
}
return r;
}