numam-dpdk/app/test/test_func_reentrancy.c
Bruce Richardson a9de470cc7 test: move to app directory
Since all other apps have been moved to the "app" folder, the autotest app
remains alone in the test folder. Rather than having an entire top-level
folder for this, we can move it back to where it all started in early
versions of DPDK - the "app/" folder.

This move has a couple of advantages:
* This reduces clutter at the top level of the project, due to one less
  folder.
* It eliminates the separate build task necessary for building the
  autotests using make "make test-build" which means that developers are
  less likely to miss something in their own compilation tests
* It re-aligns the final location of the test binary in the app folder when
  building with make with it's location in the source tree.

For meson builds, the autotest app is different from the other apps in that
it needs a series of different test cases defined for it for use by "meson
test". Therefore, it does not get built as part of the main loop in the
app folder, but gets built separately at the end.

Signed-off-by: Bruce Richardson <bruce.richardson@intel.com>
2019-02-26 15:29:27 +01:00

499 lines
11 KiB
C

/* SPDX-License-Identifier: BSD-3-Clause
* Copyright(c) 2010-2014 Intel Corporation
*/
#include <string.h>
#include <stdio.h>
#include <stdlib.h>
#include <stdint.h>
#include <inttypes.h>
#include <stdarg.h>
#include <errno.h>
#include <sys/queue.h>
#include <rte_common.h>
#include <rte_log.h>
#include <rte_debug.h>
#include <rte_memory.h>
#include <rte_launch.h>
#include <rte_cycles.h>
#include <rte_eal.h>
#include <rte_per_lcore.h>
#include <rte_lcore.h>
#include <rte_atomic.h>
#include <rte_branch_prediction.h>
#include <rte_ring.h>
#include <rte_mempool.h>
#include <rte_spinlock.h>
#include <rte_malloc.h>
#ifdef RTE_LIBRTE_HASH
#include <rte_hash.h>
#include <rte_fbk_hash.h>
#include <rte_jhash.h>
#endif /* RTE_LIBRTE_HASH */
#ifdef RTE_LIBRTE_LPM
#include <rte_lpm.h>
#endif /* RTE_LIBRTE_LPM */
#include <rte_string_fns.h>
#include "test.h"
typedef int (*case_func_t)(void* arg);
typedef void (*case_clean_t)(unsigned lcore_id);
#define MAX_STRING_SIZE (256)
#define MAX_ITER_MULTI (16)
#define MAX_ITER_ONCE (4)
#define MAX_LPM_ITER_TIMES (6)
#define MEMPOOL_ELT_SIZE (sizeof(uint32_t))
#define MEMPOOL_SIZE (4)
#define MAX_LCORES (RTE_MAX_MEMZONE / (MAX_ITER_MULTI * 4U))
static rte_atomic32_t obj_count = RTE_ATOMIC32_INIT(0);
static rte_atomic32_t synchro = RTE_ATOMIC32_INIT(0);
#define WAIT_SYNCHRO_FOR_SLAVES() do{ \
if (lcore_self != rte_get_master_lcore()) \
while (rte_atomic32_read(&synchro) == 0); \
} while(0)
/*
* rte_eal_init only init once
*/
static int
test_eal_init_once(__attribute__((unused)) void *arg)
{
unsigned lcore_self = rte_lcore_id();
WAIT_SYNCHRO_FOR_SLAVES();
rte_atomic32_set(&obj_count, 1); /* silent the check in the caller */
if (rte_eal_init(0, NULL) != -1)
return -1;
return 0;
}
/*
* ring create/lookup reentrancy test
*/
static void
ring_clean(unsigned int lcore_id)
{
struct rte_ring *rp;
char ring_name[MAX_STRING_SIZE];
int i;
for (i = 0; i < MAX_ITER_MULTI; i++) {
snprintf(ring_name, sizeof(ring_name),
"fr_test_%d_%d", lcore_id, i);
rp = rte_ring_lookup(ring_name);
if (rp != NULL)
rte_ring_free(rp);
}
}
static int
ring_create_lookup(__attribute__((unused)) void *arg)
{
unsigned lcore_self = rte_lcore_id();
struct rte_ring * rp;
char ring_name[MAX_STRING_SIZE];
int i;
WAIT_SYNCHRO_FOR_SLAVES();
/* create the same ring simultaneously on all threads */
for (i = 0; i < MAX_ITER_ONCE; i++) {
rp = rte_ring_create("fr_test_once", 4096, SOCKET_ID_ANY, 0);
if (rp != NULL)
rte_atomic32_inc(&obj_count);
}
/* create/lookup new ring several times */
for (i = 0; i < MAX_ITER_MULTI; i++) {
snprintf(ring_name, sizeof(ring_name), "fr_test_%d_%d", lcore_self, i);
rp = rte_ring_create(ring_name, 4096, SOCKET_ID_ANY, 0);
if (NULL == rp)
return -1;
if (rte_ring_lookup(ring_name) != rp)
return -1;
/* verify all ring created successful */
if (rte_ring_lookup(ring_name) == NULL)
return -1;
}
return 0;
}
static void
my_obj_init(struct rte_mempool *mp, __attribute__((unused)) void *arg,
void *obj, unsigned i)
{
uint32_t *objnum = obj;
memset(obj, 0, mp->elt_size);
*objnum = i;
}
static void
mempool_clean(unsigned int lcore_id)
{
struct rte_mempool *mp;
char mempool_name[MAX_STRING_SIZE];
int i;
/* verify all ring created successful */
for (i = 0; i < MAX_ITER_MULTI; i++) {
snprintf(mempool_name, sizeof(mempool_name), "fr_test_%d_%d",
lcore_id, i);
mp = rte_mempool_lookup(mempool_name);
if (mp != NULL)
rte_mempool_free(mp);
}
}
static int
mempool_create_lookup(__attribute__((unused)) void *arg)
{
unsigned lcore_self = rte_lcore_id();
struct rte_mempool * mp;
char mempool_name[MAX_STRING_SIZE];
int i;
WAIT_SYNCHRO_FOR_SLAVES();
/* create the same mempool simultaneously on all threads */
for (i = 0; i < MAX_ITER_ONCE; i++) {
mp = rte_mempool_create("fr_test_once", MEMPOOL_SIZE,
MEMPOOL_ELT_SIZE, 0, 0,
NULL, NULL,
my_obj_init, NULL,
SOCKET_ID_ANY, 0);
if (mp != NULL)
rte_atomic32_inc(&obj_count);
}
/* create/lookup new ring several times */
for (i = 0; i < MAX_ITER_MULTI; i++) {
snprintf(mempool_name, sizeof(mempool_name), "fr_test_%d_%d", lcore_self, i);
mp = rte_mempool_create(mempool_name, MEMPOOL_SIZE,
MEMPOOL_ELT_SIZE, 0, 0,
NULL, NULL,
my_obj_init, NULL,
SOCKET_ID_ANY, 0);
if (NULL == mp)
return -1;
if (rte_mempool_lookup(mempool_name) != mp)
return -1;
/* verify all ring created successful */
if (rte_mempool_lookup(mempool_name) == NULL)
return -1;
}
return 0;
}
#ifdef RTE_LIBRTE_HASH
static void
hash_clean(unsigned lcore_id)
{
char hash_name[MAX_STRING_SIZE];
struct rte_hash *handle;
int i;
for (i = 0; i < MAX_ITER_MULTI; i++) {
snprintf(hash_name, sizeof(hash_name), "fr_test_%d_%d", lcore_id, i);
if ((handle = rte_hash_find_existing(hash_name)) != NULL)
rte_hash_free(handle);
}
}
static int
hash_create_free(__attribute__((unused)) void *arg)
{
unsigned lcore_self = rte_lcore_id();
struct rte_hash *handle;
char hash_name[MAX_STRING_SIZE];
int i;
struct rte_hash_parameters hash_params = {
.name = NULL,
.entries = 16,
.key_len = 4,
.hash_func = (rte_hash_function)rte_jhash_32b,
.hash_func_init_val = 0,
.socket_id = 0,
};
WAIT_SYNCHRO_FOR_SLAVES();
/* create the same hash simultaneously on all threads */
hash_params.name = "fr_test_once";
for (i = 0; i < MAX_ITER_ONCE; i++) {
handle = rte_hash_create(&hash_params);
if (handle != NULL)
rte_atomic32_inc(&obj_count);
}
/* create mutiple times simultaneously */
for (i = 0; i < MAX_ITER_MULTI; i++) {
snprintf(hash_name, sizeof(hash_name), "fr_test_%d_%d", lcore_self, i);
hash_params.name = hash_name;
handle = rte_hash_create(&hash_params);
if (NULL == handle)
return -1;
/* verify correct existing and then free all */
if (handle != rte_hash_find_existing(hash_name))
return -1;
rte_hash_free(handle);
/* verify free correct */
if (NULL != rte_hash_find_existing(hash_name))
return -1;
}
return 0;
}
static void
fbk_clean(unsigned lcore_id)
{
char fbk_name[MAX_STRING_SIZE];
struct rte_fbk_hash_table *handle;
int i;
for (i = 0; i < MAX_ITER_MULTI; i++) {
snprintf(fbk_name, sizeof(fbk_name), "fr_test_%d_%d", lcore_id, i);
if ((handle = rte_fbk_hash_find_existing(fbk_name)) != NULL)
rte_fbk_hash_free(handle);
}
}
static int
fbk_create_free(__attribute__((unused)) void *arg)
{
unsigned lcore_self = rte_lcore_id();
struct rte_fbk_hash_table *handle;
char fbk_name[MAX_STRING_SIZE];
int i;
struct rte_fbk_hash_params fbk_params = {
.name = NULL,
.entries = 4,
.entries_per_bucket = 4,
.socket_id = 0,
.hash_func = rte_jhash_1word,
.init_val = RTE_FBK_HASH_INIT_VAL_DEFAULT,
};
WAIT_SYNCHRO_FOR_SLAVES();
/* create the same fbk hash table simultaneously on all threads */
fbk_params.name = "fr_test_once";
for (i = 0; i < MAX_ITER_ONCE; i++) {
handle = rte_fbk_hash_create(&fbk_params);
if (handle != NULL)
rte_atomic32_inc(&obj_count);
}
/* create mutiple fbk tables simultaneously */
for (i = 0; i < MAX_ITER_MULTI; i++) {
snprintf(fbk_name, sizeof(fbk_name), "fr_test_%d_%d", lcore_self, i);
fbk_params.name = fbk_name;
handle = rte_fbk_hash_create(&fbk_params);
if (NULL == handle)
return -1;
/* verify correct existing and then free all */
if (handle != rte_fbk_hash_find_existing(fbk_name))
return -1;
rte_fbk_hash_free(handle);
/* verify free correct */
if (NULL != rte_fbk_hash_find_existing(fbk_name))
return -1;
}
return 0;
}
#endif /* RTE_LIBRTE_HASH */
#ifdef RTE_LIBRTE_LPM
static void
lpm_clean(unsigned int lcore_id)
{
char lpm_name[MAX_STRING_SIZE];
struct rte_lpm *lpm;
int i;
for (i = 0; i < MAX_LPM_ITER_TIMES; i++) {
snprintf(lpm_name, sizeof(lpm_name), "fr_test_%d_%d", lcore_id, i);
if ((lpm = rte_lpm_find_existing(lpm_name)) != NULL)
rte_lpm_free(lpm);
}
}
static int
lpm_create_free(__attribute__((unused)) void *arg)
{
unsigned lcore_self = rte_lcore_id();
struct rte_lpm *lpm;
struct rte_lpm_config config;
config.max_rules = 4;
config.number_tbl8s = 256;
config.flags = 0;
char lpm_name[MAX_STRING_SIZE];
int i;
WAIT_SYNCHRO_FOR_SLAVES();
/* create the same lpm simultaneously on all threads */
for (i = 0; i < MAX_ITER_ONCE; i++) {
lpm = rte_lpm_create("fr_test_once", SOCKET_ID_ANY, &config);
if (lpm != NULL)
rte_atomic32_inc(&obj_count);
}
/* create mutiple fbk tables simultaneously */
for (i = 0; i < MAX_LPM_ITER_TIMES; i++) {
snprintf(lpm_name, sizeof(lpm_name), "fr_test_%d_%d", lcore_self, i);
lpm = rte_lpm_create(lpm_name, SOCKET_ID_ANY, &config);
if (NULL == lpm)
return -1;
/* verify correct existing and then free all */
if (lpm != rte_lpm_find_existing(lpm_name))
return -1;
rte_lpm_free(lpm);
/* verify free correct */
if (NULL != rte_lpm_find_existing(lpm_name))
return -1;
}
return 0;
}
#endif /* RTE_LIBRTE_LPM */
struct test_case{
case_func_t func;
void* arg;
case_clean_t clean;
char name[MAX_STRING_SIZE];
};
/* All test cases in the test suite */
struct test_case test_cases[] = {
{ test_eal_init_once, NULL, NULL, "eal init once" },
{ ring_create_lookup, NULL, ring_clean, "ring create/lookup" },
{ mempool_create_lookup, NULL, mempool_clean,
"mempool create/lookup" },
#ifdef RTE_LIBRTE_HASH
{ hash_create_free, NULL, hash_clean, "hash create/free" },
{ fbk_create_free, NULL, fbk_clean, "fbk create/free" },
#endif /* RTE_LIBRTE_HASH */
#ifdef RTE_LIBRTE_LPM
{ lpm_create_free, NULL, lpm_clean, "lpm create/free" },
#endif /* RTE_LIBRTE_LPM */
};
/**
* launch test case in two separate thread
*/
static int
launch_test(struct test_case *pt_case)
{
int ret = 0;
unsigned lcore_id;
unsigned cores_save = rte_lcore_count();
unsigned cores = RTE_MIN(cores_save, MAX_LCORES);
unsigned count;
if (pt_case->func == NULL)
return -1;
rte_atomic32_set(&obj_count, 0);
rte_atomic32_set(&synchro, 0);
RTE_LCORE_FOREACH_SLAVE(lcore_id) {
if (cores == 1)
break;
cores--;
rte_eal_remote_launch(pt_case->func, pt_case->arg, lcore_id);
}
rte_atomic32_set(&synchro, 1);
if (pt_case->func(pt_case->arg) < 0)
ret = -1;
cores = cores_save;
RTE_LCORE_FOREACH_SLAVE(lcore_id) {
if (cores == 1)
break;
cores--;
if (rte_eal_wait_lcore(lcore_id) < 0)
ret = -1;
if (pt_case->clean != NULL)
pt_case->clean(lcore_id);
}
count = rte_atomic32_read(&obj_count);
if (count != 1) {
printf("%s: common object allocated %d times (should be 1)\n",
pt_case->name, count);
ret = -1;
}
return ret;
}
/**
* Main entry of func_reentrancy test
*/
static int
test_func_reentrancy(void)
{
uint32_t case_id;
struct test_case *pt_case = NULL;
if (rte_lcore_count() <= 1) {
printf("Not enough lcore for testing\n");
return -1;
}
else if (rte_lcore_count() > MAX_LCORES)
printf("Too many lcores, some cores will be disabled\n");
for (case_id = 0; case_id < sizeof(test_cases)/sizeof(struct test_case); case_id ++) {
pt_case = &test_cases[case_id];
if (pt_case->func == NULL)
continue;
if (launch_test(pt_case) < 0) {
printf("Func-ReEnt CASE %"PRIu32": %s FAIL\n", case_id, pt_case->name);
return -1;
}
printf("Func-ReEnt CASE %"PRIu32": %s PASS\n", case_id, pt_case->name);
}
return 0;
}
REGISTER_TEST_COMMAND(func_reentrancy_autotest, test_func_reentrancy);