a986c2b797
IOVA mode in DPDK is either PA or VA. The new build option enable_iova_as_pa configures the mode to PA at compile time. By default, this option is enabled. If the option is disabled, only drivers which support it are enabled. Supported driver can set the flag pmd_supports_disable_iova_as_pa in its build file. mbuf structure holds the physical (PA) and virtual address (VA). If IOVA as PA is disabled at compile time, PA field (buf_iova) of mbuf is redundant as it is the same as VA and is replaced by a dummy field. Signed-off-by: Shijith Thotton <sthotton@marvell.com> Acked-by: Olivier Matz <olivier.matz@6wind.com>
2979 lines
78 KiB
C
2979 lines
78 KiB
C
/* SPDX-License-Identifier: BSD-3-Clause
|
|
* Copyright(c) 2010-2014 Intel Corporation
|
|
*/
|
|
|
|
#include "test.h"
|
|
|
|
#include <string.h>
|
|
#include <stdarg.h>
|
|
#include <stdio.h>
|
|
#include <stdlib.h>
|
|
#include <stdint.h>
|
|
#include <inttypes.h>
|
|
#include <errno.h>
|
|
#include <sys/queue.h>
|
|
|
|
#include <rte_common.h>
|
|
#include <rte_errno.h>
|
|
#include <rte_debug.h>
|
|
#include <rte_log.h>
|
|
#include <rte_memory.h>
|
|
#include <rte_memcpy.h>
|
|
#include <rte_launch.h>
|
|
#include <rte_eal.h>
|
|
#include <rte_per_lcore.h>
|
|
#include <rte_lcore.h>
|
|
#include <rte_branch_prediction.h>
|
|
#include <rte_ring.h>
|
|
#include <rte_mempool.h>
|
|
#include <rte_mbuf.h>
|
|
#include <rte_random.h>
|
|
#include <rte_cycles.h>
|
|
#include <rte_malloc.h>
|
|
#include <rte_ether.h>
|
|
#include <rte_ip.h>
|
|
#include <rte_tcp.h>
|
|
#include <rte_mbuf_dyn.h>
|
|
|
|
#define MEMPOOL_CACHE_SIZE 32
|
|
#define MBUF_DATA_SIZE 2048
|
|
#define NB_MBUF 128
|
|
#define MBUF_TEST_DATA_LEN 1464
|
|
#define MBUF_TEST_DATA_LEN2 50
|
|
#define MBUF_TEST_DATA_LEN3 256
|
|
#define MBUF_TEST_HDR1_LEN 20
|
|
#define MBUF_TEST_HDR2_LEN 30
|
|
#define MBUF_TEST_ALL_HDRS_LEN (MBUF_TEST_HDR1_LEN+MBUF_TEST_HDR2_LEN)
|
|
#define MBUF_TEST_SEG_SIZE 64
|
|
#define MBUF_TEST_BURST 8
|
|
#define EXT_BUF_TEST_DATA_LEN 1024
|
|
#define MBUF_MAX_SEG 16
|
|
#define MBUF_NO_HEADER 0
|
|
#define MBUF_HEADER 1
|
|
#define MBUF_NEG_TEST_READ 2
|
|
#define VAL_NAME(flag) { flag, #flag }
|
|
|
|
/* chain length in bulk test */
|
|
#define CHAIN_LEN 16
|
|
|
|
/* size of private data for mbuf in pktmbuf_pool2 */
|
|
#define MBUF2_PRIV_SIZE 128
|
|
|
|
#define REFCNT_MAX_ITER 64
|
|
#define REFCNT_MAX_TIMEOUT 10
|
|
#define REFCNT_MAX_REF (RTE_MAX_LCORE)
|
|
#define REFCNT_MBUF_NUM 64
|
|
#define REFCNT_RING_SIZE (REFCNT_MBUF_NUM * REFCNT_MAX_REF)
|
|
|
|
#define MAGIC_DATA 0x42424242
|
|
|
|
#define MAKE_STRING(x) # x
|
|
|
|
#ifdef RTE_MBUF_REFCNT_ATOMIC
|
|
|
|
static volatile uint32_t refcnt_stop_workers;
|
|
static unsigned refcnt_lcore[RTE_MAX_LCORE];
|
|
|
|
#endif
|
|
|
|
/*
|
|
* MBUF
|
|
* ====
|
|
*
|
|
* #. Allocate a mbuf pool.
|
|
*
|
|
* - The pool contains NB_MBUF elements, where each mbuf is MBUF_SIZE
|
|
* bytes long.
|
|
*
|
|
* #. Test multiple allocations of mbufs from this pool.
|
|
*
|
|
* - Allocate NB_MBUF and store pointers in a table.
|
|
* - If an allocation fails, return an error.
|
|
* - Free all these mbufs.
|
|
* - Repeat the same test to check that mbufs were freed correctly.
|
|
*
|
|
* #. Test data manipulation in pktmbuf.
|
|
*
|
|
* - Alloc an mbuf.
|
|
* - Append data using rte_pktmbuf_append().
|
|
* - Test for error in rte_pktmbuf_append() when len is too large.
|
|
* - Trim data at the end of mbuf using rte_pktmbuf_trim().
|
|
* - Test for error in rte_pktmbuf_trim() when len is too large.
|
|
* - Prepend a header using rte_pktmbuf_prepend().
|
|
* - Test for error in rte_pktmbuf_prepend() when len is too large.
|
|
* - Remove data at the beginning of mbuf using rte_pktmbuf_adj().
|
|
* - Test for error in rte_pktmbuf_adj() when len is too large.
|
|
* - Check that appended data is not corrupt.
|
|
* - Free the mbuf.
|
|
* - Between all these tests, check data_len and pkt_len, and
|
|
* that the mbuf is contiguous.
|
|
* - Repeat the test to check that allocation operations
|
|
* reinitialize the mbuf correctly.
|
|
*
|
|
* #. Test packet cloning
|
|
* - Clone a mbuf and verify the data
|
|
* - Clone the cloned mbuf and verify the data
|
|
* - Attach a mbuf to another that does not have the same priv_size.
|
|
*/
|
|
|
|
#define GOTO_FAIL(str, ...) do { \
|
|
printf("mbuf test FAILED (l.%d): <" str ">\n", \
|
|
__LINE__, ##__VA_ARGS__); \
|
|
goto fail; \
|
|
} while(0)
|
|
|
|
/*
|
|
* test data manipulation in mbuf with non-ascii data
|
|
*/
|
|
static int
|
|
test_pktmbuf_with_non_ascii_data(struct rte_mempool *pktmbuf_pool)
|
|
{
|
|
struct rte_mbuf *m = NULL;
|
|
char *data;
|
|
|
|
m = rte_pktmbuf_alloc(pktmbuf_pool);
|
|
if (m == NULL)
|
|
GOTO_FAIL("Cannot allocate mbuf");
|
|
if (rte_pktmbuf_pkt_len(m) != 0)
|
|
GOTO_FAIL("Bad length");
|
|
|
|
data = rte_pktmbuf_append(m, MBUF_TEST_DATA_LEN);
|
|
if (data == NULL)
|
|
GOTO_FAIL("Cannot append data");
|
|
if (rte_pktmbuf_pkt_len(m) != MBUF_TEST_DATA_LEN)
|
|
GOTO_FAIL("Bad pkt length");
|
|
if (rte_pktmbuf_data_len(m) != MBUF_TEST_DATA_LEN)
|
|
GOTO_FAIL("Bad data length");
|
|
memset(data, 0xff, rte_pktmbuf_pkt_len(m));
|
|
if (!rte_pktmbuf_is_contiguous(m))
|
|
GOTO_FAIL("Buffer should be continuous");
|
|
rte_pktmbuf_dump(stdout, m, MBUF_TEST_DATA_LEN);
|
|
|
|
rte_pktmbuf_free(m);
|
|
|
|
return 0;
|
|
|
|
fail:
|
|
if(m) {
|
|
rte_pktmbuf_free(m);
|
|
}
|
|
return -1;
|
|
}
|
|
|
|
/*
|
|
* test data manipulation in mbuf
|
|
*/
|
|
static int
|
|
test_one_pktmbuf(struct rte_mempool *pktmbuf_pool)
|
|
{
|
|
struct rte_mbuf *m = NULL;
|
|
char *data, *data2, *hdr;
|
|
unsigned i;
|
|
|
|
printf("Test pktmbuf API\n");
|
|
|
|
/* alloc a mbuf */
|
|
|
|
m = rte_pktmbuf_alloc(pktmbuf_pool);
|
|
if (m == NULL)
|
|
GOTO_FAIL("Cannot allocate mbuf");
|
|
if (rte_pktmbuf_pkt_len(m) != 0)
|
|
GOTO_FAIL("Bad length");
|
|
|
|
rte_pktmbuf_dump(stdout, m, 0);
|
|
|
|
/* append data */
|
|
|
|
data = rte_pktmbuf_append(m, MBUF_TEST_DATA_LEN);
|
|
if (data == NULL)
|
|
GOTO_FAIL("Cannot append data");
|
|
if (rte_pktmbuf_pkt_len(m) != MBUF_TEST_DATA_LEN)
|
|
GOTO_FAIL("Bad pkt length");
|
|
if (rte_pktmbuf_data_len(m) != MBUF_TEST_DATA_LEN)
|
|
GOTO_FAIL("Bad data length");
|
|
memset(data, 0x66, rte_pktmbuf_pkt_len(m));
|
|
if (!rte_pktmbuf_is_contiguous(m))
|
|
GOTO_FAIL("Buffer should be continuous");
|
|
rte_pktmbuf_dump(stdout, m, MBUF_TEST_DATA_LEN);
|
|
rte_pktmbuf_dump(stdout, m, 2*MBUF_TEST_DATA_LEN);
|
|
|
|
/* this append should fail */
|
|
|
|
data2 = rte_pktmbuf_append(m, (uint16_t)(rte_pktmbuf_tailroom(m) + 1));
|
|
if (data2 != NULL)
|
|
GOTO_FAIL("Append should not succeed");
|
|
|
|
/* append some more data */
|
|
|
|
data2 = rte_pktmbuf_append(m, MBUF_TEST_DATA_LEN2);
|
|
if (data2 == NULL)
|
|
GOTO_FAIL("Cannot append data");
|
|
if (rte_pktmbuf_pkt_len(m) != MBUF_TEST_DATA_LEN + MBUF_TEST_DATA_LEN2)
|
|
GOTO_FAIL("Bad pkt length");
|
|
if (rte_pktmbuf_data_len(m) != MBUF_TEST_DATA_LEN + MBUF_TEST_DATA_LEN2)
|
|
GOTO_FAIL("Bad data length");
|
|
if (!rte_pktmbuf_is_contiguous(m))
|
|
GOTO_FAIL("Buffer should be continuous");
|
|
|
|
/* trim data at the end of mbuf */
|
|
|
|
if (rte_pktmbuf_trim(m, MBUF_TEST_DATA_LEN2) < 0)
|
|
GOTO_FAIL("Cannot trim data");
|
|
if (rte_pktmbuf_pkt_len(m) != MBUF_TEST_DATA_LEN)
|
|
GOTO_FAIL("Bad pkt length");
|
|
if (rte_pktmbuf_data_len(m) != MBUF_TEST_DATA_LEN)
|
|
GOTO_FAIL("Bad data length");
|
|
if (!rte_pktmbuf_is_contiguous(m))
|
|
GOTO_FAIL("Buffer should be continuous");
|
|
|
|
/* this trim should fail */
|
|
|
|
if (rte_pktmbuf_trim(m, (uint16_t)(rte_pktmbuf_data_len(m) + 1)) == 0)
|
|
GOTO_FAIL("trim should not succeed");
|
|
|
|
/* prepend one header */
|
|
|
|
hdr = rte_pktmbuf_prepend(m, MBUF_TEST_HDR1_LEN);
|
|
if (hdr == NULL)
|
|
GOTO_FAIL("Cannot prepend");
|
|
if (data - hdr != MBUF_TEST_HDR1_LEN)
|
|
GOTO_FAIL("Prepend failed");
|
|
if (rte_pktmbuf_pkt_len(m) != MBUF_TEST_DATA_LEN + MBUF_TEST_HDR1_LEN)
|
|
GOTO_FAIL("Bad pkt length");
|
|
if (rte_pktmbuf_data_len(m) != MBUF_TEST_DATA_LEN + MBUF_TEST_HDR1_LEN)
|
|
GOTO_FAIL("Bad data length");
|
|
if (!rte_pktmbuf_is_contiguous(m))
|
|
GOTO_FAIL("Buffer should be continuous");
|
|
memset(hdr, 0x55, MBUF_TEST_HDR1_LEN);
|
|
|
|
/* prepend another header */
|
|
|
|
hdr = rte_pktmbuf_prepend(m, MBUF_TEST_HDR2_LEN);
|
|
if (hdr == NULL)
|
|
GOTO_FAIL("Cannot prepend");
|
|
if (data - hdr != MBUF_TEST_ALL_HDRS_LEN)
|
|
GOTO_FAIL("Prepend failed");
|
|
if (rte_pktmbuf_pkt_len(m) != MBUF_TEST_DATA_LEN + MBUF_TEST_ALL_HDRS_LEN)
|
|
GOTO_FAIL("Bad pkt length");
|
|
if (rte_pktmbuf_data_len(m) != MBUF_TEST_DATA_LEN + MBUF_TEST_ALL_HDRS_LEN)
|
|
GOTO_FAIL("Bad data length");
|
|
if (!rte_pktmbuf_is_contiguous(m))
|
|
GOTO_FAIL("Buffer should be continuous");
|
|
memset(hdr, 0x55, MBUF_TEST_HDR2_LEN);
|
|
|
|
rte_mbuf_sanity_check(m, 1);
|
|
rte_mbuf_sanity_check(m, 0);
|
|
rte_pktmbuf_dump(stdout, m, 0);
|
|
|
|
/* this prepend should fail */
|
|
|
|
hdr = rte_pktmbuf_prepend(m, (uint16_t)(rte_pktmbuf_headroom(m) + 1));
|
|
if (hdr != NULL)
|
|
GOTO_FAIL("prepend should not succeed");
|
|
|
|
/* remove data at beginning of mbuf (adj) */
|
|
|
|
if (data != rte_pktmbuf_adj(m, MBUF_TEST_ALL_HDRS_LEN))
|
|
GOTO_FAIL("rte_pktmbuf_adj failed");
|
|
if (rte_pktmbuf_pkt_len(m) != MBUF_TEST_DATA_LEN)
|
|
GOTO_FAIL("Bad pkt length");
|
|
if (rte_pktmbuf_data_len(m) != MBUF_TEST_DATA_LEN)
|
|
GOTO_FAIL("Bad data length");
|
|
if (!rte_pktmbuf_is_contiguous(m))
|
|
GOTO_FAIL("Buffer should be continuous");
|
|
|
|
/* this adj should fail */
|
|
|
|
if (rte_pktmbuf_adj(m, (uint16_t)(rte_pktmbuf_data_len(m) + 1)) != NULL)
|
|
GOTO_FAIL("rte_pktmbuf_adj should not succeed");
|
|
|
|
/* check data */
|
|
|
|
if (!rte_pktmbuf_is_contiguous(m))
|
|
GOTO_FAIL("Buffer should be continuous");
|
|
|
|
for (i=0; i<MBUF_TEST_DATA_LEN; i++) {
|
|
if (data[i] != 0x66)
|
|
GOTO_FAIL("Data corrupted at offset %u", i);
|
|
}
|
|
|
|
/* free mbuf */
|
|
|
|
rte_pktmbuf_free(m);
|
|
m = NULL;
|
|
return 0;
|
|
|
|
fail:
|
|
rte_pktmbuf_free(m);
|
|
return -1;
|
|
}
|
|
|
|
static uint16_t
|
|
testclone_refcnt_read(struct rte_mbuf *m)
|
|
{
|
|
return RTE_MBUF_HAS_PINNED_EXTBUF(m) ?
|
|
rte_mbuf_ext_refcnt_read(m->shinfo) :
|
|
rte_mbuf_refcnt_read(m);
|
|
}
|
|
|
|
static int
|
|
testclone_testupdate_testdetach(struct rte_mempool *pktmbuf_pool,
|
|
struct rte_mempool *clone_pool)
|
|
{
|
|
struct rte_mbuf *m = NULL;
|
|
struct rte_mbuf *clone = NULL;
|
|
struct rte_mbuf *clone2 = NULL;
|
|
unaligned_uint32_t *data;
|
|
|
|
/* alloc a mbuf */
|
|
m = rte_pktmbuf_alloc(pktmbuf_pool);
|
|
if (m == NULL)
|
|
GOTO_FAIL("ooops not allocating mbuf");
|
|
|
|
if (rte_pktmbuf_pkt_len(m) != 0)
|
|
GOTO_FAIL("Bad length");
|
|
|
|
rte_pktmbuf_append(m, sizeof(uint32_t));
|
|
data = rte_pktmbuf_mtod(m, unaligned_uint32_t *);
|
|
*data = MAGIC_DATA;
|
|
|
|
/* clone the allocated mbuf */
|
|
clone = rte_pktmbuf_clone(m, clone_pool);
|
|
if (clone == NULL)
|
|
GOTO_FAIL("cannot clone data\n");
|
|
|
|
data = rte_pktmbuf_mtod(clone, unaligned_uint32_t *);
|
|
if (*data != MAGIC_DATA)
|
|
GOTO_FAIL("invalid data in clone\n");
|
|
|
|
if (testclone_refcnt_read(m) != 2)
|
|
GOTO_FAIL("invalid refcnt in m\n");
|
|
|
|
/* free the clone */
|
|
rte_pktmbuf_free(clone);
|
|
clone = NULL;
|
|
|
|
/* same test with a chained mbuf */
|
|
m->next = rte_pktmbuf_alloc(pktmbuf_pool);
|
|
if (m->next == NULL)
|
|
GOTO_FAIL("Next Pkt Null\n");
|
|
m->nb_segs = 2;
|
|
|
|
rte_pktmbuf_append(m->next, sizeof(uint32_t));
|
|
m->pkt_len = 2 * sizeof(uint32_t);
|
|
|
|
data = rte_pktmbuf_mtod(m->next, unaligned_uint32_t *);
|
|
*data = MAGIC_DATA;
|
|
|
|
clone = rte_pktmbuf_clone(m, clone_pool);
|
|
if (clone == NULL)
|
|
GOTO_FAIL("cannot clone data\n");
|
|
|
|
data = rte_pktmbuf_mtod(clone, unaligned_uint32_t *);
|
|
if (*data != MAGIC_DATA)
|
|
GOTO_FAIL("invalid data in clone\n");
|
|
|
|
data = rte_pktmbuf_mtod(clone->next, unaligned_uint32_t *);
|
|
if (*data != MAGIC_DATA)
|
|
GOTO_FAIL("invalid data in clone->next\n");
|
|
|
|
if (testclone_refcnt_read(m) != 2)
|
|
GOTO_FAIL("invalid refcnt in m\n");
|
|
|
|
if (testclone_refcnt_read(m->next) != 2)
|
|
GOTO_FAIL("invalid refcnt in m->next\n");
|
|
|
|
/* try to clone the clone */
|
|
|
|
clone2 = rte_pktmbuf_clone(clone, clone_pool);
|
|
if (clone2 == NULL)
|
|
GOTO_FAIL("cannot clone the clone\n");
|
|
|
|
data = rte_pktmbuf_mtod(clone2, unaligned_uint32_t *);
|
|
if (*data != MAGIC_DATA)
|
|
GOTO_FAIL("invalid data in clone2\n");
|
|
|
|
data = rte_pktmbuf_mtod(clone2->next, unaligned_uint32_t *);
|
|
if (*data != MAGIC_DATA)
|
|
GOTO_FAIL("invalid data in clone2->next\n");
|
|
|
|
if (testclone_refcnt_read(m) != 3)
|
|
GOTO_FAIL("invalid refcnt in m\n");
|
|
|
|
if (testclone_refcnt_read(m->next) != 3)
|
|
GOTO_FAIL("invalid refcnt in m->next\n");
|
|
|
|
/* free mbuf */
|
|
rte_pktmbuf_free(m);
|
|
rte_pktmbuf_free(clone);
|
|
rte_pktmbuf_free(clone2);
|
|
|
|
m = NULL;
|
|
clone = NULL;
|
|
clone2 = NULL;
|
|
printf("%s ok\n", __func__);
|
|
return 0;
|
|
|
|
fail:
|
|
rte_pktmbuf_free(m);
|
|
rte_pktmbuf_free(clone);
|
|
rte_pktmbuf_free(clone2);
|
|
return -1;
|
|
}
|
|
|
|
static int
|
|
test_pktmbuf_copy(struct rte_mempool *pktmbuf_pool,
|
|
struct rte_mempool *clone_pool)
|
|
{
|
|
struct rte_mbuf *m = NULL;
|
|
struct rte_mbuf *copy = NULL;
|
|
struct rte_mbuf *copy2 = NULL;
|
|
struct rte_mbuf *clone = NULL;
|
|
unaligned_uint32_t *data;
|
|
|
|
/* alloc a mbuf */
|
|
m = rte_pktmbuf_alloc(pktmbuf_pool);
|
|
if (m == NULL)
|
|
GOTO_FAIL("ooops not allocating mbuf");
|
|
|
|
if (rte_pktmbuf_pkt_len(m) != 0)
|
|
GOTO_FAIL("Bad length");
|
|
|
|
rte_pktmbuf_append(m, sizeof(uint32_t));
|
|
data = rte_pktmbuf_mtod(m, unaligned_uint32_t *);
|
|
*data = MAGIC_DATA;
|
|
|
|
/* copy the allocated mbuf */
|
|
copy = rte_pktmbuf_copy(m, pktmbuf_pool, 0, UINT32_MAX);
|
|
if (copy == NULL)
|
|
GOTO_FAIL("cannot copy data\n");
|
|
|
|
if (rte_pktmbuf_pkt_len(copy) != sizeof(uint32_t))
|
|
GOTO_FAIL("copy length incorrect\n");
|
|
|
|
if (rte_pktmbuf_data_len(copy) != sizeof(uint32_t))
|
|
GOTO_FAIL("copy data length incorrect\n");
|
|
|
|
data = rte_pktmbuf_mtod(copy, unaligned_uint32_t *);
|
|
if (*data != MAGIC_DATA)
|
|
GOTO_FAIL("invalid data in copy\n");
|
|
|
|
/* free the copy */
|
|
rte_pktmbuf_free(copy);
|
|
copy = NULL;
|
|
|
|
/* same test with a cloned mbuf */
|
|
clone = rte_pktmbuf_clone(m, clone_pool);
|
|
if (clone == NULL)
|
|
GOTO_FAIL("cannot clone data\n");
|
|
|
|
if ((!RTE_MBUF_HAS_PINNED_EXTBUF(m) &&
|
|
!RTE_MBUF_CLONED(clone)) ||
|
|
(RTE_MBUF_HAS_PINNED_EXTBUF(m) &&
|
|
!RTE_MBUF_HAS_EXTBUF(clone)))
|
|
GOTO_FAIL("clone did not give a cloned mbuf\n");
|
|
|
|
copy = rte_pktmbuf_copy(clone, pktmbuf_pool, 0, UINT32_MAX);
|
|
if (copy == NULL)
|
|
GOTO_FAIL("cannot copy cloned mbuf\n");
|
|
|
|
if (RTE_MBUF_CLONED(copy))
|
|
GOTO_FAIL("copy of clone is cloned?\n");
|
|
|
|
if (rte_pktmbuf_pkt_len(copy) != sizeof(uint32_t))
|
|
GOTO_FAIL("copy clone length incorrect\n");
|
|
|
|
if (rte_pktmbuf_data_len(copy) != sizeof(uint32_t))
|
|
GOTO_FAIL("copy clone data length incorrect\n");
|
|
|
|
data = rte_pktmbuf_mtod(copy, unaligned_uint32_t *);
|
|
if (*data != MAGIC_DATA)
|
|
GOTO_FAIL("invalid data in clone copy\n");
|
|
rte_pktmbuf_free(clone);
|
|
rte_pktmbuf_free(copy);
|
|
copy = NULL;
|
|
clone = NULL;
|
|
|
|
|
|
/* same test with a chained mbuf */
|
|
m->next = rte_pktmbuf_alloc(pktmbuf_pool);
|
|
if (m->next == NULL)
|
|
GOTO_FAIL("Next Pkt Null\n");
|
|
m->nb_segs = 2;
|
|
|
|
rte_pktmbuf_append(m->next, sizeof(uint32_t));
|
|
m->pkt_len = 2 * sizeof(uint32_t);
|
|
data = rte_pktmbuf_mtod(m->next, unaligned_uint32_t *);
|
|
*data = MAGIC_DATA + 1;
|
|
|
|
copy = rte_pktmbuf_copy(m, pktmbuf_pool, 0, UINT32_MAX);
|
|
if (copy == NULL)
|
|
GOTO_FAIL("cannot copy data\n");
|
|
|
|
if (rte_pktmbuf_pkt_len(copy) != 2 * sizeof(uint32_t))
|
|
GOTO_FAIL("chain copy length incorrect\n");
|
|
|
|
if (rte_pktmbuf_data_len(copy) != 2 * sizeof(uint32_t))
|
|
GOTO_FAIL("chain copy data length incorrect\n");
|
|
|
|
data = rte_pktmbuf_mtod(copy, unaligned_uint32_t *);
|
|
if (data[0] != MAGIC_DATA || data[1] != MAGIC_DATA + 1)
|
|
GOTO_FAIL("invalid data in copy\n");
|
|
|
|
rte_pktmbuf_free(copy2);
|
|
|
|
/* test offset copy */
|
|
copy2 = rte_pktmbuf_copy(copy, pktmbuf_pool,
|
|
sizeof(uint32_t), UINT32_MAX);
|
|
if (copy2 == NULL)
|
|
GOTO_FAIL("cannot copy the copy\n");
|
|
|
|
if (rte_pktmbuf_pkt_len(copy2) != sizeof(uint32_t))
|
|
GOTO_FAIL("copy with offset, length incorrect\n");
|
|
|
|
if (rte_pktmbuf_data_len(copy2) != sizeof(uint32_t))
|
|
GOTO_FAIL("copy with offset, data length incorrect\n");
|
|
|
|
data = rte_pktmbuf_mtod(copy2, unaligned_uint32_t *);
|
|
if (data[0] != MAGIC_DATA + 1)
|
|
GOTO_FAIL("copy with offset, invalid data\n");
|
|
|
|
rte_pktmbuf_free(copy2);
|
|
|
|
/* test truncation copy */
|
|
copy2 = rte_pktmbuf_copy(copy, pktmbuf_pool,
|
|
0, sizeof(uint32_t));
|
|
if (copy2 == NULL)
|
|
GOTO_FAIL("cannot copy the copy\n");
|
|
|
|
if (rte_pktmbuf_pkt_len(copy2) != sizeof(uint32_t))
|
|
GOTO_FAIL("copy with truncate, length incorrect\n");
|
|
|
|
if (rte_pktmbuf_data_len(copy2) != sizeof(uint32_t))
|
|
GOTO_FAIL("copy with truncate, data length incorrect\n");
|
|
|
|
data = rte_pktmbuf_mtod(copy2, unaligned_uint32_t *);
|
|
if (data[0] != MAGIC_DATA)
|
|
GOTO_FAIL("copy with truncate, invalid data\n");
|
|
|
|
/* free mbuf */
|
|
rte_pktmbuf_free(m);
|
|
rte_pktmbuf_free(copy);
|
|
rte_pktmbuf_free(copy2);
|
|
|
|
m = NULL;
|
|
copy = NULL;
|
|
copy2 = NULL;
|
|
printf("%s ok\n", __func__);
|
|
return 0;
|
|
|
|
fail:
|
|
rte_pktmbuf_free(m);
|
|
rte_pktmbuf_free(copy);
|
|
rte_pktmbuf_free(copy2);
|
|
return -1;
|
|
}
|
|
|
|
static int
|
|
test_attach_from_different_pool(struct rte_mempool *pktmbuf_pool,
|
|
struct rte_mempool *pktmbuf_pool2)
|
|
{
|
|
struct rte_mbuf *m = NULL;
|
|
struct rte_mbuf *clone = NULL;
|
|
struct rte_mbuf *clone2 = NULL;
|
|
char *data, *c_data, *c_data2;
|
|
|
|
/* alloc a mbuf */
|
|
m = rte_pktmbuf_alloc(pktmbuf_pool);
|
|
if (m == NULL)
|
|
GOTO_FAIL("cannot allocate mbuf");
|
|
|
|
if (rte_pktmbuf_pkt_len(m) != 0)
|
|
GOTO_FAIL("Bad length");
|
|
|
|
data = rte_pktmbuf_mtod(m, char *);
|
|
|
|
/* allocate a new mbuf from the second pool, and attach it to the first
|
|
* mbuf */
|
|
clone = rte_pktmbuf_alloc(pktmbuf_pool2);
|
|
if (clone == NULL)
|
|
GOTO_FAIL("cannot allocate mbuf from second pool\n");
|
|
|
|
/* check data room size and priv size, and erase priv */
|
|
if (rte_pktmbuf_data_room_size(clone->pool) != 0)
|
|
GOTO_FAIL("data room size should be 0\n");
|
|
if (rte_pktmbuf_priv_size(clone->pool) != MBUF2_PRIV_SIZE)
|
|
GOTO_FAIL("data room size should be %d\n", MBUF2_PRIV_SIZE);
|
|
memset(clone + 1, 0, MBUF2_PRIV_SIZE);
|
|
|
|
/* save data pointer to compare it after detach() */
|
|
c_data = rte_pktmbuf_mtod(clone, char *);
|
|
if (c_data != (char *)clone + sizeof(*clone) + MBUF2_PRIV_SIZE)
|
|
GOTO_FAIL("bad data pointer in clone");
|
|
if (rte_pktmbuf_headroom(clone) != 0)
|
|
GOTO_FAIL("bad headroom in clone");
|
|
|
|
rte_pktmbuf_attach(clone, m);
|
|
|
|
if (rte_pktmbuf_mtod(clone, char *) != data)
|
|
GOTO_FAIL("clone was not attached properly\n");
|
|
if (rte_pktmbuf_headroom(clone) != RTE_PKTMBUF_HEADROOM)
|
|
GOTO_FAIL("bad headroom in clone after attach");
|
|
if (rte_mbuf_refcnt_read(m) != 2)
|
|
GOTO_FAIL("invalid refcnt in m\n");
|
|
|
|
/* allocate a new mbuf from the second pool, and attach it to the first
|
|
* cloned mbuf */
|
|
clone2 = rte_pktmbuf_alloc(pktmbuf_pool2);
|
|
if (clone2 == NULL)
|
|
GOTO_FAIL("cannot allocate clone2 from second pool\n");
|
|
|
|
/* check data room size and priv size, and erase priv */
|
|
if (rte_pktmbuf_data_room_size(clone2->pool) != 0)
|
|
GOTO_FAIL("data room size should be 0\n");
|
|
if (rte_pktmbuf_priv_size(clone2->pool) != MBUF2_PRIV_SIZE)
|
|
GOTO_FAIL("data room size should be %d\n", MBUF2_PRIV_SIZE);
|
|
memset(clone2 + 1, 0, MBUF2_PRIV_SIZE);
|
|
|
|
/* save data pointer to compare it after detach() */
|
|
c_data2 = rte_pktmbuf_mtod(clone2, char *);
|
|
if (c_data2 != (char *)clone2 + sizeof(*clone2) + MBUF2_PRIV_SIZE)
|
|
GOTO_FAIL("bad data pointer in clone2");
|
|
if (rte_pktmbuf_headroom(clone2) != 0)
|
|
GOTO_FAIL("bad headroom in clone2");
|
|
|
|
rte_pktmbuf_attach(clone2, clone);
|
|
|
|
if (rte_pktmbuf_mtod(clone2, char *) != data)
|
|
GOTO_FAIL("clone2 was not attached properly\n");
|
|
if (rte_pktmbuf_headroom(clone2) != RTE_PKTMBUF_HEADROOM)
|
|
GOTO_FAIL("bad headroom in clone2 after attach");
|
|
if (rte_mbuf_refcnt_read(m) != 3)
|
|
GOTO_FAIL("invalid refcnt in m\n");
|
|
|
|
/* detach the clones */
|
|
rte_pktmbuf_detach(clone);
|
|
if (c_data != rte_pktmbuf_mtod(clone, char *))
|
|
GOTO_FAIL("clone was not detached properly\n");
|
|
if (rte_mbuf_refcnt_read(m) != 2)
|
|
GOTO_FAIL("invalid refcnt in m\n");
|
|
|
|
rte_pktmbuf_detach(clone2);
|
|
if (c_data2 != rte_pktmbuf_mtod(clone2, char *))
|
|
GOTO_FAIL("clone2 was not detached properly\n");
|
|
if (rte_mbuf_refcnt_read(m) != 1)
|
|
GOTO_FAIL("invalid refcnt in m\n");
|
|
|
|
/* free the clones and the initial mbuf */
|
|
rte_pktmbuf_free(clone2);
|
|
rte_pktmbuf_free(clone);
|
|
rte_pktmbuf_free(m);
|
|
printf("%s ok\n", __func__);
|
|
return 0;
|
|
|
|
fail:
|
|
rte_pktmbuf_free(m);
|
|
rte_pktmbuf_free(clone);
|
|
rte_pktmbuf_free(clone2);
|
|
return -1;
|
|
}
|
|
|
|
/*
|
|
* test allocation and free of mbufs
|
|
*/
|
|
static int
|
|
test_pktmbuf_pool(struct rte_mempool *pktmbuf_pool)
|
|
{
|
|
unsigned i;
|
|
struct rte_mbuf *m[NB_MBUF];
|
|
int ret = 0;
|
|
|
|
for (i=0; i<NB_MBUF; i++)
|
|
m[i] = NULL;
|
|
|
|
/* alloc NB_MBUF mbufs */
|
|
for (i=0; i<NB_MBUF; i++) {
|
|
m[i] = rte_pktmbuf_alloc(pktmbuf_pool);
|
|
if (m[i] == NULL) {
|
|
printf("rte_pktmbuf_alloc() failed (%u)\n", i);
|
|
ret = -1;
|
|
}
|
|
}
|
|
struct rte_mbuf *extra = NULL;
|
|
extra = rte_pktmbuf_alloc(pktmbuf_pool);
|
|
if(extra != NULL) {
|
|
printf("Error pool not empty");
|
|
ret = -1;
|
|
}
|
|
extra = rte_pktmbuf_clone(m[0], pktmbuf_pool);
|
|
if(extra != NULL) {
|
|
printf("Error pool not empty");
|
|
ret = -1;
|
|
}
|
|
/* free them */
|
|
for (i=0; i<NB_MBUF; i++) {
|
|
rte_pktmbuf_free(m[i]);
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
/*
|
|
* test bulk allocation and bulk free of mbufs
|
|
*/
|
|
static int
|
|
test_pktmbuf_pool_bulk(void)
|
|
{
|
|
struct rte_mempool *pool = NULL;
|
|
struct rte_mempool *pool2 = NULL;
|
|
unsigned int i;
|
|
struct rte_mbuf *m;
|
|
struct rte_mbuf *mbufs[NB_MBUF];
|
|
int ret = 0;
|
|
|
|
/* We cannot use the preallocated mbuf pools because their caches
|
|
* prevent us from bulk allocating all objects in them.
|
|
* So we create our own mbuf pools without caches.
|
|
*/
|
|
printf("Create mbuf pools for bulk allocation.\n");
|
|
pool = rte_pktmbuf_pool_create("test_pktmbuf_bulk",
|
|
NB_MBUF, 0, 0, MBUF_DATA_SIZE, SOCKET_ID_ANY);
|
|
if (pool == NULL) {
|
|
printf("rte_pktmbuf_pool_create() failed. rte_errno %d\n",
|
|
rte_errno);
|
|
goto err;
|
|
}
|
|
pool2 = rte_pktmbuf_pool_create("test_pktmbuf_bulk2",
|
|
NB_MBUF, 0, 0, MBUF_DATA_SIZE, SOCKET_ID_ANY);
|
|
if (pool2 == NULL) {
|
|
printf("rte_pktmbuf_pool_create() failed. rte_errno %d\n",
|
|
rte_errno);
|
|
goto err;
|
|
}
|
|
|
|
/* Preconditions: Mempools must be full. */
|
|
if (!(rte_mempool_full(pool) && rte_mempool_full(pool2))) {
|
|
printf("Test precondition failed: mempools not full\n");
|
|
goto err;
|
|
}
|
|
if (!(rte_mempool_avail_count(pool) == NB_MBUF &&
|
|
rte_mempool_avail_count(pool2) == NB_MBUF)) {
|
|
printf("Test precondition failed: mempools: %u+%u != %u+%u",
|
|
rte_mempool_avail_count(pool),
|
|
rte_mempool_avail_count(pool2),
|
|
NB_MBUF, NB_MBUF);
|
|
goto err;
|
|
}
|
|
|
|
printf("Test single bulk alloc, followed by multiple bulk free.\n");
|
|
|
|
/* Bulk allocate all mbufs in the pool, in one go. */
|
|
ret = rte_pktmbuf_alloc_bulk(pool, mbufs, NB_MBUF);
|
|
if (ret != 0) {
|
|
printf("rte_pktmbuf_alloc_bulk() failed: %d\n", ret);
|
|
goto err;
|
|
}
|
|
/* Test that they have been removed from the pool. */
|
|
if (!rte_mempool_empty(pool)) {
|
|
printf("mempool not empty\n");
|
|
goto err;
|
|
}
|
|
/* Bulk free all mbufs, in four steps. */
|
|
RTE_BUILD_BUG_ON(NB_MBUF % 4 != 0);
|
|
for (i = 0; i < NB_MBUF; i += NB_MBUF / 4) {
|
|
rte_pktmbuf_free_bulk(&mbufs[i], NB_MBUF / 4);
|
|
/* Test that they have been returned to the pool. */
|
|
if (rte_mempool_avail_count(pool) != i + NB_MBUF / 4) {
|
|
printf("mempool avail count incorrect\n");
|
|
goto err;
|
|
}
|
|
}
|
|
|
|
printf("Test multiple bulk alloc, followed by single bulk free.\n");
|
|
|
|
/* Bulk allocate all mbufs in the pool, in four steps. */
|
|
for (i = 0; i < NB_MBUF; i += NB_MBUF / 4) {
|
|
ret = rte_pktmbuf_alloc_bulk(pool, &mbufs[i], NB_MBUF / 4);
|
|
if (ret != 0) {
|
|
printf("rte_pktmbuf_alloc_bulk() failed: %d\n", ret);
|
|
goto err;
|
|
}
|
|
}
|
|
/* Test that they have been removed from the pool. */
|
|
if (!rte_mempool_empty(pool)) {
|
|
printf("mempool not empty\n");
|
|
goto err;
|
|
}
|
|
/* Bulk free all mbufs, in one go. */
|
|
rte_pktmbuf_free_bulk(mbufs, NB_MBUF);
|
|
/* Test that they have been returned to the pool. */
|
|
if (!rte_mempool_full(pool)) {
|
|
printf("mempool not full\n");
|
|
goto err;
|
|
}
|
|
|
|
printf("Test bulk free of single long chain.\n");
|
|
|
|
/* Bulk allocate all mbufs in the pool, in one go. */
|
|
ret = rte_pktmbuf_alloc_bulk(pool, mbufs, NB_MBUF);
|
|
if (ret != 0) {
|
|
printf("rte_pktmbuf_alloc_bulk() failed: %d\n", ret);
|
|
goto err;
|
|
}
|
|
/* Create a long mbuf chain. */
|
|
for (i = 1; i < NB_MBUF; i++) {
|
|
ret = rte_pktmbuf_chain(mbufs[0], mbufs[i]);
|
|
if (ret != 0) {
|
|
printf("rte_pktmbuf_chain() failed: %d\n", ret);
|
|
goto err;
|
|
}
|
|
mbufs[i] = NULL;
|
|
}
|
|
/* Free the mbuf chain containing all the mbufs. */
|
|
rte_pktmbuf_free_bulk(mbufs, 1);
|
|
/* Test that they have been returned to the pool. */
|
|
if (!rte_mempool_full(pool)) {
|
|
printf("mempool not full\n");
|
|
goto err;
|
|
}
|
|
|
|
printf("Test bulk free of multiple chains using multiple pools.\n");
|
|
|
|
/* Create mbuf chains containing mbufs from different pools. */
|
|
RTE_BUILD_BUG_ON(CHAIN_LEN % 2 != 0);
|
|
RTE_BUILD_BUG_ON(NB_MBUF % (CHAIN_LEN / 2) != 0);
|
|
for (i = 0; i < NB_MBUF * 2; i++) {
|
|
m = rte_pktmbuf_alloc((i & 4) ? pool2 : pool);
|
|
if (m == NULL) {
|
|
printf("rte_pktmbuf_alloc() failed (%u)\n", i);
|
|
goto err;
|
|
}
|
|
if ((i % CHAIN_LEN) == 0)
|
|
mbufs[i / CHAIN_LEN] = m;
|
|
else
|
|
rte_pktmbuf_chain(mbufs[i / CHAIN_LEN], m);
|
|
}
|
|
/* Test that both pools have been emptied. */
|
|
if (!(rte_mempool_empty(pool) && rte_mempool_empty(pool2))) {
|
|
printf("mempools not empty\n");
|
|
goto err;
|
|
}
|
|
/* Free one mbuf chain. */
|
|
rte_pktmbuf_free_bulk(mbufs, 1);
|
|
/* Test that the segments have been returned to the pools. */
|
|
if (!(rte_mempool_avail_count(pool) == CHAIN_LEN / 2 &&
|
|
rte_mempool_avail_count(pool2) == CHAIN_LEN / 2)) {
|
|
printf("all segments of first mbuf have not been returned\n");
|
|
goto err;
|
|
}
|
|
/* Free the remaining mbuf chains. */
|
|
rte_pktmbuf_free_bulk(&mbufs[1], NB_MBUF * 2 / CHAIN_LEN - 1);
|
|
/* Test that they have been returned to the pools. */
|
|
if (!(rte_mempool_full(pool) && rte_mempool_full(pool2))) {
|
|
printf("mempools not full\n");
|
|
goto err;
|
|
}
|
|
|
|
ret = 0;
|
|
goto done;
|
|
|
|
err:
|
|
ret = -1;
|
|
|
|
done:
|
|
printf("Free mbuf pools for bulk allocation.\n");
|
|
rte_mempool_free(pool);
|
|
rte_mempool_free(pool2);
|
|
return ret;
|
|
}
|
|
|
|
/*
|
|
* test that the pointer to the data on a packet mbuf is set properly
|
|
*/
|
|
static int
|
|
test_pktmbuf_pool_ptr(struct rte_mempool *pktmbuf_pool)
|
|
{
|
|
unsigned i;
|
|
struct rte_mbuf *m[NB_MBUF];
|
|
int ret = 0;
|
|
|
|
for (i=0; i<NB_MBUF; i++)
|
|
m[i] = NULL;
|
|
|
|
/* alloc NB_MBUF mbufs */
|
|
for (i=0; i<NB_MBUF; i++) {
|
|
m[i] = rte_pktmbuf_alloc(pktmbuf_pool);
|
|
if (m[i] == NULL) {
|
|
printf("rte_pktmbuf_alloc() failed (%u)\n", i);
|
|
ret = -1;
|
|
break;
|
|
}
|
|
m[i]->data_off += 64;
|
|
}
|
|
|
|
/* free them */
|
|
for (i=0; i<NB_MBUF; i++) {
|
|
rte_pktmbuf_free(m[i]);
|
|
}
|
|
|
|
for (i=0; i<NB_MBUF; i++)
|
|
m[i] = NULL;
|
|
|
|
/* alloc NB_MBUF mbufs */
|
|
for (i=0; i<NB_MBUF; i++) {
|
|
m[i] = rte_pktmbuf_alloc(pktmbuf_pool);
|
|
if (m[i] == NULL) {
|
|
printf("rte_pktmbuf_alloc() failed (%u)\n", i);
|
|
ret = -1;
|
|
break;
|
|
}
|
|
if (m[i]->data_off != RTE_PKTMBUF_HEADROOM) {
|
|
printf("invalid data_off\n");
|
|
ret = -1;
|
|
}
|
|
}
|
|
|
|
/* free them */
|
|
for (i=0; i<NB_MBUF; i++) {
|
|
rte_pktmbuf_free(m[i]);
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int
|
|
test_pktmbuf_free_segment(struct rte_mempool *pktmbuf_pool)
|
|
{
|
|
unsigned i;
|
|
struct rte_mbuf *m[NB_MBUF];
|
|
int ret = 0;
|
|
|
|
for (i=0; i<NB_MBUF; i++)
|
|
m[i] = NULL;
|
|
|
|
/* alloc NB_MBUF mbufs */
|
|
for (i=0; i<NB_MBUF; i++) {
|
|
m[i] = rte_pktmbuf_alloc(pktmbuf_pool);
|
|
if (m[i] == NULL) {
|
|
printf("rte_pktmbuf_alloc() failed (%u)\n", i);
|
|
ret = -1;
|
|
}
|
|
}
|
|
|
|
/* free them */
|
|
for (i=0; i<NB_MBUF; i++) {
|
|
if (m[i] != NULL) {
|
|
struct rte_mbuf *mb, *mt;
|
|
|
|
mb = m[i];
|
|
while(mb != NULL) {
|
|
mt = mb;
|
|
mb = mb->next;
|
|
rte_pktmbuf_free_seg(mt);
|
|
}
|
|
}
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
/*
|
|
* Stress test for rte_mbuf atomic refcnt.
|
|
* Implies that RTE_MBUF_REFCNT_ATOMIC is defined.
|
|
* For more efficiency, recommended to run with RTE_LIBRTE_MBUF_DEBUG defined.
|
|
*/
|
|
|
|
#ifdef RTE_MBUF_REFCNT_ATOMIC
|
|
|
|
static int
|
|
test_refcnt_worker(void *arg)
|
|
{
|
|
unsigned lcore, free;
|
|
void *mp = 0;
|
|
struct rte_ring *refcnt_mbuf_ring = arg;
|
|
|
|
lcore = rte_lcore_id();
|
|
printf("%s started at lcore %u\n", __func__, lcore);
|
|
|
|
free = 0;
|
|
while (refcnt_stop_workers == 0) {
|
|
if (rte_ring_dequeue(refcnt_mbuf_ring, &mp) == 0) {
|
|
free++;
|
|
rte_pktmbuf_free(mp);
|
|
}
|
|
}
|
|
|
|
refcnt_lcore[lcore] += free;
|
|
printf("%s finished at lcore %u, "
|
|
"number of freed mbufs: %u\n",
|
|
__func__, lcore, free);
|
|
return 0;
|
|
}
|
|
|
|
static void
|
|
test_refcnt_iter(unsigned int lcore, unsigned int iter,
|
|
struct rte_mempool *refcnt_pool,
|
|
struct rte_ring *refcnt_mbuf_ring)
|
|
{
|
|
uint16_t ref;
|
|
unsigned i, n, tref, wn;
|
|
struct rte_mbuf *m;
|
|
|
|
tref = 0;
|
|
|
|
/* For each mbuf in the pool:
|
|
* - allocate mbuf,
|
|
* - increment it's reference up to N+1,
|
|
* - enqueue it N times into the ring for worker cores to free.
|
|
*/
|
|
for (i = 0, n = rte_mempool_avail_count(refcnt_pool);
|
|
i != n && (m = rte_pktmbuf_alloc(refcnt_pool)) != NULL;
|
|
i++) {
|
|
ref = RTE_MAX(rte_rand() % REFCNT_MAX_REF, 1UL);
|
|
tref += ref;
|
|
if ((ref & 1) != 0) {
|
|
rte_pktmbuf_refcnt_update(m, ref);
|
|
while (ref-- != 0)
|
|
rte_ring_enqueue(refcnt_mbuf_ring, m);
|
|
} else {
|
|
while (ref-- != 0) {
|
|
rte_pktmbuf_refcnt_update(m, 1);
|
|
rte_ring_enqueue(refcnt_mbuf_ring, m);
|
|
}
|
|
}
|
|
rte_pktmbuf_free(m);
|
|
}
|
|
|
|
if (i != n)
|
|
rte_panic("(lcore=%u, iter=%u): was able to allocate only "
|
|
"%u from %u mbufs\n", lcore, iter, i, n);
|
|
|
|
/* wait till worker lcores will consume all mbufs */
|
|
while (!rte_ring_empty(refcnt_mbuf_ring))
|
|
;
|
|
|
|
/* check that all mbufs are back into mempool by now */
|
|
for (wn = 0; wn != REFCNT_MAX_TIMEOUT; wn++) {
|
|
if ((i = rte_mempool_avail_count(refcnt_pool)) == n) {
|
|
refcnt_lcore[lcore] += tref;
|
|
printf("%s(lcore=%u, iter=%u) completed, "
|
|
"%u references processed\n",
|
|
__func__, lcore, iter, tref);
|
|
return;
|
|
}
|
|
rte_delay_ms(100);
|
|
}
|
|
|
|
rte_panic("(lcore=%u, iter=%u): after %us only "
|
|
"%u of %u mbufs left free\n", lcore, iter, wn, i, n);
|
|
}
|
|
|
|
static int
|
|
test_refcnt_main(struct rte_mempool *refcnt_pool,
|
|
struct rte_ring *refcnt_mbuf_ring)
|
|
{
|
|
unsigned i, lcore;
|
|
|
|
lcore = rte_lcore_id();
|
|
printf("%s started at lcore %u\n", __func__, lcore);
|
|
|
|
for (i = 0; i != REFCNT_MAX_ITER; i++)
|
|
test_refcnt_iter(lcore, i, refcnt_pool, refcnt_mbuf_ring);
|
|
|
|
refcnt_stop_workers = 1;
|
|
rte_wmb();
|
|
|
|
printf("%s finished at lcore %u\n", __func__, lcore);
|
|
return 0;
|
|
}
|
|
|
|
#endif
|
|
|
|
static int
|
|
test_refcnt_mbuf(void)
|
|
{
|
|
#ifdef RTE_MBUF_REFCNT_ATOMIC
|
|
unsigned int main_lcore, worker, tref;
|
|
int ret = -1;
|
|
struct rte_mempool *refcnt_pool = NULL;
|
|
struct rte_ring *refcnt_mbuf_ring = NULL;
|
|
|
|
if (rte_lcore_count() < 2) {
|
|
printf("Not enough cores for test_refcnt_mbuf, expecting at least 2\n");
|
|
return TEST_SKIPPED;
|
|
}
|
|
|
|
printf("starting %s, at %u lcores\n", __func__, rte_lcore_count());
|
|
|
|
/* create refcnt pool & ring if they don't exist */
|
|
|
|
refcnt_pool = rte_pktmbuf_pool_create(MAKE_STRING(refcnt_pool),
|
|
REFCNT_MBUF_NUM, 0, 0, 0,
|
|
SOCKET_ID_ANY);
|
|
if (refcnt_pool == NULL) {
|
|
printf("%s: cannot allocate " MAKE_STRING(refcnt_pool) "\n",
|
|
__func__);
|
|
return -1;
|
|
}
|
|
|
|
refcnt_mbuf_ring = rte_ring_create("refcnt_mbuf_ring",
|
|
rte_align32pow2(REFCNT_RING_SIZE), SOCKET_ID_ANY,
|
|
RING_F_SP_ENQ);
|
|
if (refcnt_mbuf_ring == NULL) {
|
|
printf("%s: cannot allocate " MAKE_STRING(refcnt_mbuf_ring)
|
|
"\n", __func__);
|
|
goto err;
|
|
}
|
|
|
|
refcnt_stop_workers = 0;
|
|
memset(refcnt_lcore, 0, sizeof (refcnt_lcore));
|
|
|
|
rte_eal_mp_remote_launch(test_refcnt_worker, refcnt_mbuf_ring, SKIP_MAIN);
|
|
|
|
test_refcnt_main(refcnt_pool, refcnt_mbuf_ring);
|
|
|
|
rte_eal_mp_wait_lcore();
|
|
|
|
/* check that we processed all references */
|
|
tref = 0;
|
|
main_lcore = rte_get_main_lcore();
|
|
|
|
RTE_LCORE_FOREACH_WORKER(worker)
|
|
tref += refcnt_lcore[worker];
|
|
|
|
if (tref != refcnt_lcore[main_lcore])
|
|
rte_panic("referenced mbufs: %u, freed mbufs: %u\n",
|
|
tref, refcnt_lcore[main_lcore]);
|
|
|
|
rte_mempool_dump(stdout, refcnt_pool);
|
|
rte_ring_dump(stdout, refcnt_mbuf_ring);
|
|
|
|
ret = 0;
|
|
|
|
err:
|
|
rte_mempool_free(refcnt_pool);
|
|
rte_ring_free(refcnt_mbuf_ring);
|
|
return ret;
|
|
#else
|
|
return 0;
|
|
#endif
|
|
}
|
|
|
|
#ifdef RTE_EXEC_ENV_WINDOWS
|
|
static int
|
|
test_failing_mbuf_sanity_check(struct rte_mempool *pktmbuf_pool)
|
|
{
|
|
RTE_SET_USED(pktmbuf_pool);
|
|
return TEST_SKIPPED;
|
|
}
|
|
#else
|
|
|
|
#include <unistd.h>
|
|
#include <sys/resource.h>
|
|
#include <sys/time.h>
|
|
#include <sys/wait.h>
|
|
|
|
/* use fork() to test mbuf errors panic */
|
|
static int
|
|
verify_mbuf_check_panics(struct rte_mbuf *buf)
|
|
{
|
|
int pid;
|
|
int status;
|
|
|
|
pid = fork();
|
|
|
|
if (pid == 0) {
|
|
struct rlimit rl;
|
|
|
|
/* No need to generate a coredump when panicking. */
|
|
rl.rlim_cur = rl.rlim_max = 0;
|
|
setrlimit(RLIMIT_CORE, &rl);
|
|
rte_mbuf_sanity_check(buf, 1); /* should panic */
|
|
exit(0); /* return normally if it doesn't panic */
|
|
} else if (pid < 0) {
|
|
printf("Fork Failed\n");
|
|
return -1;
|
|
}
|
|
wait(&status);
|
|
if(status == 0)
|
|
return -1;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
test_failing_mbuf_sanity_check(struct rte_mempool *pktmbuf_pool)
|
|
{
|
|
struct rte_mbuf *buf;
|
|
struct rte_mbuf badbuf;
|
|
|
|
printf("Checking rte_mbuf_sanity_check for failure conditions\n");
|
|
|
|
/* get a good mbuf to use to make copies */
|
|
buf = rte_pktmbuf_alloc(pktmbuf_pool);
|
|
if (buf == NULL)
|
|
return -1;
|
|
|
|
printf("Checking good mbuf initially\n");
|
|
if (verify_mbuf_check_panics(buf) != -1)
|
|
return -1;
|
|
|
|
printf("Now checking for error conditions\n");
|
|
|
|
if (verify_mbuf_check_panics(NULL)) {
|
|
printf("Error with NULL mbuf test\n");
|
|
return -1;
|
|
}
|
|
|
|
badbuf = *buf;
|
|
badbuf.pool = NULL;
|
|
if (verify_mbuf_check_panics(&badbuf)) {
|
|
printf("Error with bad-pool mbuf test\n");
|
|
return -1;
|
|
}
|
|
|
|
if (RTE_IOVA_AS_PA) {
|
|
badbuf = *buf;
|
|
rte_mbuf_iova_set(&badbuf, 0);
|
|
if (verify_mbuf_check_panics(&badbuf)) {
|
|
printf("Error with bad-physaddr mbuf test\n");
|
|
return -1;
|
|
}
|
|
}
|
|
|
|
badbuf = *buf;
|
|
badbuf.buf_addr = NULL;
|
|
if (verify_mbuf_check_panics(&badbuf)) {
|
|
printf("Error with bad-addr mbuf test\n");
|
|
return -1;
|
|
}
|
|
|
|
badbuf = *buf;
|
|
badbuf.refcnt = 0;
|
|
if (verify_mbuf_check_panics(&badbuf)) {
|
|
printf("Error with bad-refcnt(0) mbuf test\n");
|
|
return -1;
|
|
}
|
|
|
|
badbuf = *buf;
|
|
badbuf.refcnt = UINT16_MAX;
|
|
if (verify_mbuf_check_panics(&badbuf)) {
|
|
printf("Error with bad-refcnt(MAX) mbuf test\n");
|
|
return -1;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
#endif /* !RTE_EXEC_ENV_WINDOWS */
|
|
|
|
static int
|
|
test_mbuf_linearize(struct rte_mempool *pktmbuf_pool, int pkt_len,
|
|
int nb_segs)
|
|
{
|
|
|
|
struct rte_mbuf *m = NULL, *mbuf = NULL;
|
|
uint8_t *data;
|
|
int data_len = 0;
|
|
int remain;
|
|
int seg, seg_len;
|
|
int i;
|
|
|
|
if (pkt_len < 1) {
|
|
printf("Packet size must be 1 or more (is %d)\n", pkt_len);
|
|
return -1;
|
|
}
|
|
|
|
if (nb_segs < 1) {
|
|
printf("Number of segments must be 1 or more (is %d)\n",
|
|
nb_segs);
|
|
return -1;
|
|
}
|
|
|
|
seg_len = pkt_len / nb_segs;
|
|
if (seg_len == 0)
|
|
seg_len = 1;
|
|
|
|
remain = pkt_len;
|
|
|
|
/* Create chained mbuf_src and fill it generated data */
|
|
for (seg = 0; remain > 0; seg++) {
|
|
|
|
m = rte_pktmbuf_alloc(pktmbuf_pool);
|
|
if (m == NULL) {
|
|
printf("Cannot create segment for source mbuf");
|
|
goto fail;
|
|
}
|
|
|
|
/* Make sure if tailroom is zeroed */
|
|
memset(rte_pktmbuf_mtod(m, uint8_t *), 0,
|
|
rte_pktmbuf_tailroom(m));
|
|
|
|
data_len = remain;
|
|
if (data_len > seg_len)
|
|
data_len = seg_len;
|
|
|
|
data = (uint8_t *)rte_pktmbuf_append(m, data_len);
|
|
if (data == NULL) {
|
|
printf("Cannot append %d bytes to the mbuf\n",
|
|
data_len);
|
|
goto fail;
|
|
}
|
|
|
|
for (i = 0; i < data_len; i++)
|
|
data[i] = (seg * seg_len + i) % 0x0ff;
|
|
|
|
if (seg == 0)
|
|
mbuf = m;
|
|
else
|
|
rte_pktmbuf_chain(mbuf, m);
|
|
|
|
remain -= data_len;
|
|
}
|
|
|
|
/* Create destination buffer to store coalesced data */
|
|
if (rte_pktmbuf_linearize(mbuf)) {
|
|
printf("Mbuf linearization failed\n");
|
|
goto fail;
|
|
}
|
|
|
|
if (!rte_pktmbuf_is_contiguous(mbuf)) {
|
|
printf("Source buffer should be contiguous after "
|
|
"linearization\n");
|
|
goto fail;
|
|
}
|
|
|
|
data = rte_pktmbuf_mtod(mbuf, uint8_t *);
|
|
|
|
for (i = 0; i < pkt_len; i++)
|
|
if (data[i] != (i % 0x0ff)) {
|
|
printf("Incorrect data in linearized mbuf\n");
|
|
goto fail;
|
|
}
|
|
|
|
rte_pktmbuf_free(mbuf);
|
|
return 0;
|
|
|
|
fail:
|
|
rte_pktmbuf_free(mbuf);
|
|
return -1;
|
|
}
|
|
|
|
static int
|
|
test_mbuf_linearize_check(struct rte_mempool *pktmbuf_pool)
|
|
{
|
|
struct test_mbuf_array {
|
|
int size;
|
|
int nb_segs;
|
|
} mbuf_array[] = {
|
|
{ 128, 1 },
|
|
{ 64, 64 },
|
|
{ 512, 10 },
|
|
{ 250, 11 },
|
|
{ 123, 8 },
|
|
};
|
|
unsigned int i;
|
|
|
|
printf("Test mbuf linearize API\n");
|
|
|
|
for (i = 0; i < RTE_DIM(mbuf_array); i++)
|
|
if (test_mbuf_linearize(pktmbuf_pool, mbuf_array[i].size,
|
|
mbuf_array[i].nb_segs)) {
|
|
printf("Test failed for %d, %d\n", mbuf_array[i].size,
|
|
mbuf_array[i].nb_segs);
|
|
return -1;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* Helper function for test_tx_ofload
|
|
*/
|
|
static inline void
|
|
set_tx_offload(struct rte_mbuf *mb, uint64_t il2, uint64_t il3, uint64_t il4,
|
|
uint64_t tso, uint64_t ol3, uint64_t ol2)
|
|
{
|
|
mb->l2_len = il2;
|
|
mb->l3_len = il3;
|
|
mb->l4_len = il4;
|
|
mb->tso_segsz = tso;
|
|
mb->outer_l3_len = ol3;
|
|
mb->outer_l2_len = ol2;
|
|
}
|
|
|
|
static int
|
|
test_tx_offload(void)
|
|
{
|
|
struct rte_mbuf *mb;
|
|
uint64_t tm, v1, v2;
|
|
size_t sz;
|
|
uint32_t i;
|
|
|
|
static volatile struct {
|
|
uint16_t l2;
|
|
uint16_t l3;
|
|
uint16_t l4;
|
|
uint16_t tso;
|
|
} txof;
|
|
|
|
const uint32_t num = 0x10000;
|
|
|
|
txof.l2 = rte_rand() % (1 << RTE_MBUF_L2_LEN_BITS);
|
|
txof.l3 = rte_rand() % (1 << RTE_MBUF_L3_LEN_BITS);
|
|
txof.l4 = rte_rand() % (1 << RTE_MBUF_L4_LEN_BITS);
|
|
txof.tso = rte_rand() % (1 << RTE_MBUF_TSO_SEGSZ_BITS);
|
|
|
|
printf("%s started, tx_offload = {\n"
|
|
"\tl2_len=%#hx,\n"
|
|
"\tl3_len=%#hx,\n"
|
|
"\tl4_len=%#hx,\n"
|
|
"\ttso_segsz=%#hx,\n"
|
|
"\touter_l3_len=%#x,\n"
|
|
"\touter_l2_len=%#x,\n"
|
|
"};\n",
|
|
__func__,
|
|
txof.l2, txof.l3, txof.l4, txof.tso, txof.l3, txof.l2);
|
|
|
|
sz = sizeof(*mb) * num;
|
|
mb = rte_zmalloc(NULL, sz, RTE_CACHE_LINE_SIZE);
|
|
if (mb == NULL) {
|
|
printf("%s failed, out of memory\n", __func__);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
memset(mb, 0, sz);
|
|
tm = rte_rdtsc_precise();
|
|
|
|
for (i = 0; i != num; i++)
|
|
set_tx_offload(mb + i, txof.l2, txof.l3, txof.l4,
|
|
txof.tso, txof.l3, txof.l2);
|
|
|
|
tm = rte_rdtsc_precise() - tm;
|
|
printf("%s set tx_offload by bit-fields: %u iterations, %"
|
|
PRIu64 " cycles, %#Lf cycles/iter\n",
|
|
__func__, num, tm, (long double)tm / num);
|
|
|
|
v1 = mb[rte_rand() % num].tx_offload;
|
|
|
|
memset(mb, 0, sz);
|
|
tm = rte_rdtsc_precise();
|
|
|
|
for (i = 0; i != num; i++)
|
|
mb[i].tx_offload = rte_mbuf_tx_offload(txof.l2, txof.l3,
|
|
txof.l4, txof.tso, txof.l3, txof.l2, 0);
|
|
|
|
tm = rte_rdtsc_precise() - tm;
|
|
printf("%s set raw tx_offload: %u iterations, %"
|
|
PRIu64 " cycles, %#Lf cycles/iter\n",
|
|
__func__, num, tm, (long double)tm / num);
|
|
|
|
v2 = mb[rte_rand() % num].tx_offload;
|
|
|
|
rte_free(mb);
|
|
|
|
printf("%s finished\n"
|
|
"expected tx_offload value: 0x%" PRIx64 ";\n"
|
|
"rte_mbuf_tx_offload value: 0x%" PRIx64 ";\n",
|
|
__func__, v1, v2);
|
|
|
|
return (v1 == v2) ? 0 : -EINVAL;
|
|
}
|
|
|
|
static int
|
|
test_get_rx_ol_flag_list(void)
|
|
{
|
|
int len = 6, ret = 0;
|
|
char buf[256] = "";
|
|
int buflen = 0;
|
|
|
|
/* Test case to check with null buffer */
|
|
ret = rte_get_rx_ol_flag_list(0, NULL, 0);
|
|
if (ret != -1)
|
|
GOTO_FAIL("%s expected: -1, received = %d\n", __func__, ret);
|
|
|
|
/* Test case to check with zero buffer len */
|
|
ret = rte_get_rx_ol_flag_list(RTE_MBUF_F_RX_L4_CKSUM_MASK, buf, 0);
|
|
if (ret != -1)
|
|
GOTO_FAIL("%s expected: -1, received = %d\n", __func__, ret);
|
|
|
|
buflen = strlen(buf);
|
|
if (buflen != 0)
|
|
GOTO_FAIL("%s buffer should be empty, received = %d\n",
|
|
__func__, buflen);
|
|
|
|
/* Test case to check with reduced buffer len */
|
|
ret = rte_get_rx_ol_flag_list(0, buf, len);
|
|
if (ret != -1)
|
|
GOTO_FAIL("%s expected: -1, received = %d\n", __func__, ret);
|
|
|
|
buflen = strlen(buf);
|
|
if (buflen != (len - 1))
|
|
GOTO_FAIL("%s invalid buffer length retrieved, expected: %d,"
|
|
"received = %d\n", __func__,
|
|
(len - 1), buflen);
|
|
|
|
/* Test case to check with zero mask value */
|
|
ret = rte_get_rx_ol_flag_list(0, buf, sizeof(buf));
|
|
if (ret != 0)
|
|
GOTO_FAIL("%s expected: 0, received = %d\n", __func__, ret);
|
|
|
|
buflen = strlen(buf);
|
|
if (buflen == 0)
|
|
GOTO_FAIL("%s expected: %s, received length = 0\n", __func__,
|
|
"non-zero, buffer should not be empty");
|
|
|
|
/* Test case to check with valid mask value */
|
|
ret = rte_get_rx_ol_flag_list(RTE_MBUF_F_RX_SEC_OFFLOAD, buf,
|
|
sizeof(buf));
|
|
if (ret != 0)
|
|
GOTO_FAIL("%s expected: 0, received = %d\n", __func__, ret);
|
|
|
|
buflen = strlen(buf);
|
|
if (buflen == 0)
|
|
GOTO_FAIL("%s expected: %s, received length = 0\n", __func__,
|
|
"non-zero, buffer should not be empty");
|
|
|
|
return 0;
|
|
fail:
|
|
return -1;
|
|
}
|
|
|
|
static int
|
|
test_get_tx_ol_flag_list(void)
|
|
{
|
|
int len = 6, ret = 0;
|
|
char buf[256] = "";
|
|
int buflen = 0;
|
|
|
|
/* Test case to check with null buffer */
|
|
ret = rte_get_tx_ol_flag_list(0, NULL, 0);
|
|
if (ret != -1)
|
|
GOTO_FAIL("%s expected: -1, received = %d\n", __func__, ret);
|
|
|
|
/* Test case to check with zero buffer len */
|
|
ret = rte_get_tx_ol_flag_list(RTE_MBUF_F_TX_IP_CKSUM, buf, 0);
|
|
if (ret != -1)
|
|
GOTO_FAIL("%s expected: -1, received = %d\n", __func__, ret);
|
|
|
|
buflen = strlen(buf);
|
|
if (buflen != 0) {
|
|
GOTO_FAIL("%s buffer should be empty, received = %d\n",
|
|
__func__, buflen);
|
|
}
|
|
|
|
/* Test case to check with reduced buffer len */
|
|
ret = rte_get_tx_ol_flag_list(0, buf, len);
|
|
if (ret != -1)
|
|
GOTO_FAIL("%s expected: -1, received = %d\n", __func__, ret);
|
|
|
|
buflen = strlen(buf);
|
|
if (buflen != (len - 1))
|
|
GOTO_FAIL("%s invalid buffer length retrieved, expected: %d,"
|
|
"received = %d\n", __func__,
|
|
(len - 1), buflen);
|
|
|
|
/* Test case to check with zero mask value */
|
|
ret = rte_get_tx_ol_flag_list(0, buf, sizeof(buf));
|
|
if (ret != 0)
|
|
GOTO_FAIL("%s expected: 0, received = %d\n", __func__, ret);
|
|
|
|
buflen = strlen(buf);
|
|
if (buflen == 0)
|
|
GOTO_FAIL("%s expected: %s, received length = 0\n", __func__,
|
|
"non-zero, buffer should not be empty");
|
|
|
|
/* Test case to check with valid mask value */
|
|
ret = rte_get_tx_ol_flag_list(RTE_MBUF_F_TX_UDP_CKSUM, buf,
|
|
sizeof(buf));
|
|
if (ret != 0)
|
|
GOTO_FAIL("%s expected: 0, received = %d\n", __func__, ret);
|
|
|
|
buflen = strlen(buf);
|
|
if (buflen == 0)
|
|
GOTO_FAIL("%s expected: %s, received length = 0\n", __func__,
|
|
"non-zero, buffer should not be empty");
|
|
|
|
return 0;
|
|
fail:
|
|
return -1;
|
|
|
|
}
|
|
|
|
struct flag_name {
|
|
uint64_t flag;
|
|
const char *name;
|
|
};
|
|
|
|
static int
|
|
test_get_rx_ol_flag_name(void)
|
|
{
|
|
uint16_t i;
|
|
const char *flag_str = NULL;
|
|
const struct flag_name rx_flags[] = {
|
|
VAL_NAME(RTE_MBUF_F_RX_VLAN),
|
|
VAL_NAME(RTE_MBUF_F_RX_RSS_HASH),
|
|
VAL_NAME(RTE_MBUF_F_RX_FDIR),
|
|
VAL_NAME(RTE_MBUF_F_RX_L4_CKSUM_BAD),
|
|
VAL_NAME(RTE_MBUF_F_RX_L4_CKSUM_GOOD),
|
|
VAL_NAME(RTE_MBUF_F_RX_L4_CKSUM_NONE),
|
|
VAL_NAME(RTE_MBUF_F_RX_IP_CKSUM_BAD),
|
|
VAL_NAME(RTE_MBUF_F_RX_IP_CKSUM_GOOD),
|
|
VAL_NAME(RTE_MBUF_F_RX_IP_CKSUM_NONE),
|
|
VAL_NAME(RTE_MBUF_F_RX_OUTER_IP_CKSUM_BAD),
|
|
VAL_NAME(RTE_MBUF_F_RX_VLAN_STRIPPED),
|
|
VAL_NAME(RTE_MBUF_F_RX_IEEE1588_PTP),
|
|
VAL_NAME(RTE_MBUF_F_RX_IEEE1588_TMST),
|
|
VAL_NAME(RTE_MBUF_F_RX_FDIR_ID),
|
|
VAL_NAME(RTE_MBUF_F_RX_FDIR_FLX),
|
|
VAL_NAME(RTE_MBUF_F_RX_QINQ_STRIPPED),
|
|
VAL_NAME(RTE_MBUF_F_RX_LRO),
|
|
VAL_NAME(RTE_MBUF_F_RX_SEC_OFFLOAD),
|
|
VAL_NAME(RTE_MBUF_F_RX_SEC_OFFLOAD_FAILED),
|
|
VAL_NAME(RTE_MBUF_F_RX_OUTER_L4_CKSUM_BAD),
|
|
VAL_NAME(RTE_MBUF_F_RX_OUTER_L4_CKSUM_GOOD),
|
|
VAL_NAME(RTE_MBUF_F_RX_OUTER_L4_CKSUM_INVALID),
|
|
};
|
|
|
|
/* Test case to check with valid flag */
|
|
for (i = 0; i < RTE_DIM(rx_flags); i++) {
|
|
flag_str = rte_get_rx_ol_flag_name(rx_flags[i].flag);
|
|
if (flag_str == NULL)
|
|
GOTO_FAIL("%s: Expected flagname = %s; received null\n",
|
|
__func__, rx_flags[i].name);
|
|
if (strcmp(flag_str, rx_flags[i].name) != 0)
|
|
GOTO_FAIL("%s: Expected flagname = %s; received = %s\n",
|
|
__func__, rx_flags[i].name, flag_str);
|
|
}
|
|
/* Test case to check with invalid flag */
|
|
flag_str = rte_get_rx_ol_flag_name(0);
|
|
if (flag_str != NULL) {
|
|
GOTO_FAIL("%s: Expected flag name = null; received = %s\n",
|
|
__func__, flag_str);
|
|
}
|
|
|
|
return 0;
|
|
fail:
|
|
return -1;
|
|
}
|
|
|
|
static int
|
|
test_get_tx_ol_flag_name(void)
|
|
{
|
|
uint16_t i;
|
|
const char *flag_str = NULL;
|
|
const struct flag_name tx_flags[] = {
|
|
VAL_NAME(RTE_MBUF_F_TX_VLAN),
|
|
VAL_NAME(RTE_MBUF_F_TX_IP_CKSUM),
|
|
VAL_NAME(RTE_MBUF_F_TX_TCP_CKSUM),
|
|
VAL_NAME(RTE_MBUF_F_TX_SCTP_CKSUM),
|
|
VAL_NAME(RTE_MBUF_F_TX_UDP_CKSUM),
|
|
VAL_NAME(RTE_MBUF_F_TX_IEEE1588_TMST),
|
|
VAL_NAME(RTE_MBUF_F_TX_TCP_SEG),
|
|
VAL_NAME(RTE_MBUF_F_TX_IPV4),
|
|
VAL_NAME(RTE_MBUF_F_TX_IPV6),
|
|
VAL_NAME(RTE_MBUF_F_TX_OUTER_IP_CKSUM),
|
|
VAL_NAME(RTE_MBUF_F_TX_OUTER_IPV4),
|
|
VAL_NAME(RTE_MBUF_F_TX_OUTER_IPV6),
|
|
VAL_NAME(RTE_MBUF_F_TX_TUNNEL_VXLAN),
|
|
VAL_NAME(RTE_MBUF_F_TX_TUNNEL_GRE),
|
|
VAL_NAME(RTE_MBUF_F_TX_TUNNEL_IPIP),
|
|
VAL_NAME(RTE_MBUF_F_TX_TUNNEL_GENEVE),
|
|
VAL_NAME(RTE_MBUF_F_TX_TUNNEL_MPLSINUDP),
|
|
VAL_NAME(RTE_MBUF_F_TX_TUNNEL_VXLAN_GPE),
|
|
VAL_NAME(RTE_MBUF_F_TX_TUNNEL_IP),
|
|
VAL_NAME(RTE_MBUF_F_TX_TUNNEL_UDP),
|
|
VAL_NAME(RTE_MBUF_F_TX_QINQ),
|
|
VAL_NAME(RTE_MBUF_F_TX_MACSEC),
|
|
VAL_NAME(RTE_MBUF_F_TX_SEC_OFFLOAD),
|
|
VAL_NAME(RTE_MBUF_F_TX_UDP_SEG),
|
|
VAL_NAME(RTE_MBUF_F_TX_OUTER_UDP_CKSUM),
|
|
};
|
|
|
|
/* Test case to check with valid flag */
|
|
for (i = 0; i < RTE_DIM(tx_flags); i++) {
|
|
flag_str = rte_get_tx_ol_flag_name(tx_flags[i].flag);
|
|
if (flag_str == NULL)
|
|
GOTO_FAIL("%s: Expected flagname = %s; received null\n",
|
|
__func__, tx_flags[i].name);
|
|
if (strcmp(flag_str, tx_flags[i].name) != 0)
|
|
GOTO_FAIL("%s: Expected flagname = %s; received = %s\n",
|
|
__func__, tx_flags[i].name, flag_str);
|
|
}
|
|
/* Test case to check with invalid flag */
|
|
flag_str = rte_get_tx_ol_flag_name(0);
|
|
if (flag_str != NULL) {
|
|
GOTO_FAIL("%s: Expected flag name = null; received = %s\n",
|
|
__func__, flag_str);
|
|
}
|
|
|
|
return 0;
|
|
fail:
|
|
return -1;
|
|
|
|
}
|
|
|
|
static int
|
|
test_mbuf_validate_tx_offload(const char *test_name,
|
|
struct rte_mempool *pktmbuf_pool,
|
|
uint64_t ol_flags,
|
|
uint16_t segsize,
|
|
int expected_retval)
|
|
{
|
|
struct rte_mbuf *m = NULL;
|
|
int ret = 0;
|
|
|
|
/* alloc a mbuf and do sanity check */
|
|
m = rte_pktmbuf_alloc(pktmbuf_pool);
|
|
if (m == NULL)
|
|
GOTO_FAIL("%s: mbuf allocation failed!\n", __func__);
|
|
if (rte_pktmbuf_pkt_len(m) != 0)
|
|
GOTO_FAIL("%s: Bad packet length\n", __func__);
|
|
rte_mbuf_sanity_check(m, 0);
|
|
m->ol_flags = ol_flags;
|
|
m->tso_segsz = segsize;
|
|
ret = rte_validate_tx_offload(m);
|
|
if (ret != expected_retval)
|
|
GOTO_FAIL("%s(%s): expected ret val: %d; received: %d\n",
|
|
__func__, test_name, expected_retval, ret);
|
|
rte_pktmbuf_free(m);
|
|
m = NULL;
|
|
return 0;
|
|
fail:
|
|
if (m) {
|
|
rte_pktmbuf_free(m);
|
|
m = NULL;
|
|
}
|
|
return -1;
|
|
}
|
|
|
|
static int
|
|
test_mbuf_validate_tx_offload_one(struct rte_mempool *pktmbuf_pool)
|
|
{
|
|
/* test to validate tx offload flags */
|
|
uint64_t ol_flags = 0;
|
|
|
|
/* test to validate if IP checksum is counted only for IPV4 packet */
|
|
/* set both IP checksum and IPV6 flags */
|
|
ol_flags |= RTE_MBUF_F_TX_IP_CKSUM;
|
|
ol_flags |= RTE_MBUF_F_TX_IPV6;
|
|
if (test_mbuf_validate_tx_offload("MBUF_TEST_IP_CKSUM_IPV6_SET",
|
|
pktmbuf_pool,
|
|
ol_flags, 0, -EINVAL) < 0)
|
|
GOTO_FAIL("%s failed: IP cksum is set incorrect.\n", __func__);
|
|
/* resetting ol_flags for next testcase */
|
|
ol_flags = 0;
|
|
|
|
/* test to validate if IP type is set when required */
|
|
ol_flags |= RTE_MBUF_F_TX_L4_MASK;
|
|
if (test_mbuf_validate_tx_offload("MBUF_TEST_IP_TYPE_NOT_SET",
|
|
pktmbuf_pool,
|
|
ol_flags, 0, -EINVAL) < 0)
|
|
GOTO_FAIL("%s failed: IP type is not set.\n", __func__);
|
|
|
|
/* test if IP type is set when TCP SEG is on */
|
|
ol_flags |= RTE_MBUF_F_TX_TCP_SEG;
|
|
if (test_mbuf_validate_tx_offload("MBUF_TEST_IP_TYPE_NOT_SET",
|
|
pktmbuf_pool,
|
|
ol_flags, 0, -EINVAL) < 0)
|
|
GOTO_FAIL("%s failed: IP type is not set.\n", __func__);
|
|
|
|
ol_flags = 0;
|
|
/* test to confirm IP type (IPV4/IPV6) is set */
|
|
ol_flags = RTE_MBUF_F_TX_L4_MASK;
|
|
ol_flags |= RTE_MBUF_F_TX_IPV6;
|
|
if (test_mbuf_validate_tx_offload("MBUF_TEST_IP_TYPE_SET",
|
|
pktmbuf_pool,
|
|
ol_flags, 0, 0) < 0)
|
|
GOTO_FAIL("%s failed: tx offload flag error.\n", __func__);
|
|
|
|
ol_flags = 0;
|
|
/* test to check TSO segment size is non-zero */
|
|
ol_flags |= RTE_MBUF_F_TX_IPV4;
|
|
ol_flags |= RTE_MBUF_F_TX_TCP_SEG;
|
|
/* set 0 tso segment size */
|
|
if (test_mbuf_validate_tx_offload("MBUF_TEST_NULL_TSO_SEGSZ",
|
|
pktmbuf_pool,
|
|
ol_flags, 0, -EINVAL) < 0)
|
|
GOTO_FAIL("%s failed: tso segment size is null.\n", __func__);
|
|
|
|
/* retain IPV4 and RTE_MBUF_F_TX_TCP_SEG mask */
|
|
/* set valid tso segment size but IP CKSUM not set */
|
|
if (test_mbuf_validate_tx_offload("MBUF_TEST_TSO_IP_CKSUM_NOT_SET",
|
|
pktmbuf_pool,
|
|
ol_flags, 512, -EINVAL) < 0)
|
|
GOTO_FAIL("%s failed: IP CKSUM is not set.\n", __func__);
|
|
|
|
/* test to validate if IP checksum is set for TSO capability */
|
|
/* retain IPV4, TCP_SEG, tso_seg size */
|
|
ol_flags |= RTE_MBUF_F_TX_IP_CKSUM;
|
|
if (test_mbuf_validate_tx_offload("MBUF_TEST_TSO_IP_CKSUM_SET",
|
|
pktmbuf_pool,
|
|
ol_flags, 512, 0) < 0)
|
|
GOTO_FAIL("%s failed: tx offload flag error.\n", __func__);
|
|
|
|
/* test to confirm TSO for IPV6 type */
|
|
ol_flags = 0;
|
|
ol_flags |= RTE_MBUF_F_TX_IPV6;
|
|
ol_flags |= RTE_MBUF_F_TX_TCP_SEG;
|
|
if (test_mbuf_validate_tx_offload("MBUF_TEST_TSO_IPV6_SET",
|
|
pktmbuf_pool,
|
|
ol_flags, 512, 0) < 0)
|
|
GOTO_FAIL("%s failed: TSO req not met.\n", __func__);
|
|
|
|
ol_flags = 0;
|
|
/* test if outer IP checksum set for non outer IPv4 packet */
|
|
ol_flags |= RTE_MBUF_F_TX_IPV6;
|
|
ol_flags |= RTE_MBUF_F_TX_OUTER_IP_CKSUM;
|
|
if (test_mbuf_validate_tx_offload("MBUF_TEST_OUTER_IPV4_NOT_SET",
|
|
pktmbuf_pool,
|
|
ol_flags, 512, -EINVAL) < 0)
|
|
GOTO_FAIL("%s failed: Outer IP cksum set.\n", __func__);
|
|
|
|
ol_flags = 0;
|
|
/* test to confirm outer IP checksum is set for outer IPV4 packet */
|
|
ol_flags |= RTE_MBUF_F_TX_OUTER_IP_CKSUM;
|
|
ol_flags |= RTE_MBUF_F_TX_OUTER_IPV4;
|
|
if (test_mbuf_validate_tx_offload("MBUF_TEST_OUTER_IPV4_SET",
|
|
pktmbuf_pool,
|
|
ol_flags, 512, 0) < 0)
|
|
GOTO_FAIL("%s failed: tx offload flag error.\n", __func__);
|
|
|
|
ol_flags = 0;
|
|
/* test to confirm if packets with no TX_OFFLOAD_MASK are skipped */
|
|
if (test_mbuf_validate_tx_offload("MBUF_TEST_OL_MASK_NOT_SET",
|
|
pktmbuf_pool,
|
|
ol_flags, 512, 0) < 0)
|
|
GOTO_FAIL("%s failed: tx offload flag error.\n", __func__);
|
|
return 0;
|
|
fail:
|
|
return -1;
|
|
}
|
|
|
|
/*
|
|
* Test for allocating a bulk of mbufs
|
|
* define an array with positive sizes for mbufs allocations.
|
|
*/
|
|
static int
|
|
test_pktmbuf_alloc_bulk(struct rte_mempool *pktmbuf_pool)
|
|
{
|
|
int ret = 0;
|
|
unsigned int idx, loop;
|
|
unsigned int alloc_counts[] = {
|
|
0,
|
|
MEMPOOL_CACHE_SIZE - 1,
|
|
MEMPOOL_CACHE_SIZE + 1,
|
|
MEMPOOL_CACHE_SIZE * 1.5,
|
|
MEMPOOL_CACHE_SIZE * 2,
|
|
MEMPOOL_CACHE_SIZE * 2 - 1,
|
|
MEMPOOL_CACHE_SIZE * 2 + 1,
|
|
MEMPOOL_CACHE_SIZE,
|
|
};
|
|
|
|
/* allocate a large array of mbuf pointers */
|
|
struct rte_mbuf *mbufs[NB_MBUF] = { 0 };
|
|
for (idx = 0; idx < RTE_DIM(alloc_counts); idx++) {
|
|
ret = rte_pktmbuf_alloc_bulk(pktmbuf_pool, mbufs,
|
|
alloc_counts[idx]);
|
|
if (ret == 0) {
|
|
for (loop = 0; loop < alloc_counts[idx] &&
|
|
mbufs[loop] != NULL; loop++)
|
|
rte_pktmbuf_free(mbufs[loop]);
|
|
} else if (ret != 0) {
|
|
printf("%s: Bulk alloc failed count(%u); ret val(%d)\n",
|
|
__func__, alloc_counts[idx], ret);
|
|
return -1;
|
|
}
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* Negative testing for allocating a bulk of mbufs
|
|
*/
|
|
static int
|
|
test_neg_pktmbuf_alloc_bulk(struct rte_mempool *pktmbuf_pool)
|
|
{
|
|
int ret = 0;
|
|
unsigned int idx, loop;
|
|
unsigned int neg_alloc_counts[] = {
|
|
MEMPOOL_CACHE_SIZE - NB_MBUF,
|
|
NB_MBUF + 1,
|
|
NB_MBUF * 8,
|
|
UINT_MAX
|
|
};
|
|
struct rte_mbuf *mbufs[NB_MBUF * 8] = { 0 };
|
|
|
|
for (idx = 0; idx < RTE_DIM(neg_alloc_counts); idx++) {
|
|
ret = rte_pktmbuf_alloc_bulk(pktmbuf_pool, mbufs,
|
|
neg_alloc_counts[idx]);
|
|
if (ret == 0) {
|
|
printf("%s: Bulk alloc must fail! count(%u); ret(%d)\n",
|
|
__func__, neg_alloc_counts[idx], ret);
|
|
for (loop = 0; loop < neg_alloc_counts[idx] &&
|
|
mbufs[loop] != NULL; loop++)
|
|
rte_pktmbuf_free(mbufs[loop]);
|
|
return -1;
|
|
}
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* Test to read mbuf packet using rte_pktmbuf_read
|
|
*/
|
|
static int
|
|
test_pktmbuf_read(struct rte_mempool *pktmbuf_pool)
|
|
{
|
|
struct rte_mbuf *m = NULL;
|
|
char *data = NULL;
|
|
const char *data_copy = NULL;
|
|
int off;
|
|
|
|
/* alloc a mbuf */
|
|
m = rte_pktmbuf_alloc(pktmbuf_pool);
|
|
if (m == NULL)
|
|
GOTO_FAIL("%s: mbuf allocation failed!\n", __func__);
|
|
if (rte_pktmbuf_pkt_len(m) != 0)
|
|
GOTO_FAIL("%s: Bad packet length\n", __func__);
|
|
rte_mbuf_sanity_check(m, 0);
|
|
|
|
data = rte_pktmbuf_append(m, MBUF_TEST_DATA_LEN2);
|
|
if (data == NULL)
|
|
GOTO_FAIL("%s: Cannot append data\n", __func__);
|
|
if (rte_pktmbuf_pkt_len(m) != MBUF_TEST_DATA_LEN2)
|
|
GOTO_FAIL("%s: Bad packet length\n", __func__);
|
|
memset(data, 0xfe, MBUF_TEST_DATA_LEN2);
|
|
|
|
/* read the data from mbuf */
|
|
data_copy = rte_pktmbuf_read(m, 0, MBUF_TEST_DATA_LEN2, NULL);
|
|
if (data_copy == NULL)
|
|
GOTO_FAIL("%s: Error in reading data!\n", __func__);
|
|
for (off = 0; off < MBUF_TEST_DATA_LEN2; off++) {
|
|
if (data_copy[off] != (char)0xfe)
|
|
GOTO_FAIL("Data corrupted at offset %u", off);
|
|
}
|
|
rte_pktmbuf_free(m);
|
|
m = NULL;
|
|
|
|
return 0;
|
|
fail:
|
|
if (m) {
|
|
rte_pktmbuf_free(m);
|
|
m = NULL;
|
|
}
|
|
return -1;
|
|
}
|
|
|
|
/*
|
|
* Test to read mbuf packet data from offset
|
|
*/
|
|
static int
|
|
test_pktmbuf_read_from_offset(struct rte_mempool *pktmbuf_pool)
|
|
{
|
|
struct rte_mbuf *m = NULL;
|
|
struct ether_hdr *hdr = NULL;
|
|
char *data = NULL;
|
|
const char *data_copy = NULL;
|
|
unsigned int off;
|
|
unsigned int hdr_len = sizeof(struct rte_ether_hdr);
|
|
|
|
/* alloc a mbuf */
|
|
m = rte_pktmbuf_alloc(pktmbuf_pool);
|
|
if (m == NULL)
|
|
GOTO_FAIL("%s: mbuf allocation failed!\n", __func__);
|
|
|
|
if (rte_pktmbuf_pkt_len(m) != 0)
|
|
GOTO_FAIL("%s: Bad packet length\n", __func__);
|
|
rte_mbuf_sanity_check(m, 0);
|
|
|
|
/* prepend an ethernet header */
|
|
hdr = (struct ether_hdr *)rte_pktmbuf_prepend(m, hdr_len);
|
|
if (hdr == NULL)
|
|
GOTO_FAIL("%s: Cannot prepend header\n", __func__);
|
|
if (rte_pktmbuf_pkt_len(m) != hdr_len)
|
|
GOTO_FAIL("%s: Bad pkt length", __func__);
|
|
if (rte_pktmbuf_data_len(m) != hdr_len)
|
|
GOTO_FAIL("%s: Bad data length", __func__);
|
|
memset(hdr, 0xde, hdr_len);
|
|
|
|
/* read mbuf header info from 0 offset */
|
|
data_copy = rte_pktmbuf_read(m, 0, hdr_len, NULL);
|
|
if (data_copy == NULL)
|
|
GOTO_FAIL("%s: Error in reading header!\n", __func__);
|
|
for (off = 0; off < hdr_len; off++) {
|
|
if (data_copy[off] != (char)0xde)
|
|
GOTO_FAIL("Header info corrupted at offset %u", off);
|
|
}
|
|
|
|
/* append sample data after ethernet header */
|
|
data = rte_pktmbuf_append(m, MBUF_TEST_DATA_LEN2);
|
|
if (data == NULL)
|
|
GOTO_FAIL("%s: Cannot append data\n", __func__);
|
|
if (rte_pktmbuf_pkt_len(m) != hdr_len + MBUF_TEST_DATA_LEN2)
|
|
GOTO_FAIL("%s: Bad packet length\n", __func__);
|
|
if (rte_pktmbuf_data_len(m) != hdr_len + MBUF_TEST_DATA_LEN2)
|
|
GOTO_FAIL("%s: Bad data length\n", __func__);
|
|
memset(data, 0xcc, MBUF_TEST_DATA_LEN2);
|
|
|
|
/* read mbuf data after header info */
|
|
data_copy = rte_pktmbuf_read(m, hdr_len, MBUF_TEST_DATA_LEN2, NULL);
|
|
if (data_copy == NULL)
|
|
GOTO_FAIL("%s: Error in reading header data!\n", __func__);
|
|
for (off = 0; off < MBUF_TEST_DATA_LEN2; off++) {
|
|
if (data_copy[off] != (char)0xcc)
|
|
GOTO_FAIL("Data corrupted at offset %u", off);
|
|
}
|
|
|
|
/* partial reading of mbuf data */
|
|
data_copy = rte_pktmbuf_read(m, hdr_len + 5, MBUF_TEST_DATA_LEN2 - 5,
|
|
NULL);
|
|
if (data_copy == NULL)
|
|
GOTO_FAIL("%s: Error in reading packet data!\n", __func__);
|
|
for (off = 0; off < MBUF_TEST_DATA_LEN2 - 5; off++) {
|
|
if (data_copy[off] != (char)0xcc)
|
|
GOTO_FAIL("Data corrupted at offset %u", off);
|
|
}
|
|
|
|
/* read length greater than mbuf data_len */
|
|
if (rte_pktmbuf_read(m, hdr_len, rte_pktmbuf_data_len(m) + 1,
|
|
NULL) != NULL)
|
|
GOTO_FAIL("%s: Requested len is larger than mbuf data len!\n",
|
|
__func__);
|
|
|
|
/* read length greater than mbuf pkt_len */
|
|
if (rte_pktmbuf_read(m, hdr_len, rte_pktmbuf_pkt_len(m) + 1,
|
|
NULL) != NULL)
|
|
GOTO_FAIL("%s: Requested len is larger than mbuf pkt len!\n",
|
|
__func__);
|
|
|
|
/* read data of zero len from valid offset */
|
|
data_copy = rte_pktmbuf_read(m, hdr_len, 0, NULL);
|
|
if (data_copy == NULL)
|
|
GOTO_FAIL("%s: Error in reading packet data!\n", __func__);
|
|
for (off = 0; off < MBUF_TEST_DATA_LEN2; off++) {
|
|
if (data_copy[off] != (char)0xcc)
|
|
GOTO_FAIL("Data corrupted at offset %u", off);
|
|
}
|
|
|
|
/* read data of zero length from zero offset */
|
|
data_copy = rte_pktmbuf_read(m, 0, 0, NULL);
|
|
if (data_copy == NULL)
|
|
GOTO_FAIL("%s: Error in reading packet data!\n", __func__);
|
|
/* check if the received address is the beginning of header info */
|
|
if (hdr != (const struct ether_hdr *)data_copy)
|
|
GOTO_FAIL("%s: Corrupted data address!\n", __func__);
|
|
|
|
/* read data of max length from valid offset */
|
|
data_copy = rte_pktmbuf_read(m, hdr_len, UINT_MAX, NULL);
|
|
if (data_copy == NULL)
|
|
GOTO_FAIL("%s: Error in reading packet data!\n", __func__);
|
|
/* check if the received address is the beginning of data segment */
|
|
if (data_copy != data)
|
|
GOTO_FAIL("%s: Corrupted data address!\n", __func__);
|
|
|
|
/* try to read from mbuf with max size offset */
|
|
data_copy = rte_pktmbuf_read(m, UINT_MAX, 0, NULL);
|
|
if (data_copy != NULL)
|
|
GOTO_FAIL("%s: Error in reading packet data!\n", __func__);
|
|
|
|
/* try to read from mbuf with max size offset and len */
|
|
data_copy = rte_pktmbuf_read(m, UINT_MAX, UINT_MAX, NULL);
|
|
if (data_copy != NULL)
|
|
GOTO_FAIL("%s: Error in reading packet data!\n", __func__);
|
|
|
|
rte_pktmbuf_dump(stdout, m, rte_pktmbuf_pkt_len(m));
|
|
|
|
rte_pktmbuf_free(m);
|
|
m = NULL;
|
|
|
|
return 0;
|
|
fail:
|
|
if (m) {
|
|
rte_pktmbuf_free(m);
|
|
m = NULL;
|
|
}
|
|
return -1;
|
|
}
|
|
|
|
struct test_case {
|
|
unsigned int seg_count;
|
|
unsigned int flags;
|
|
uint32_t read_off;
|
|
uint32_t read_len;
|
|
unsigned int seg_lengths[MBUF_MAX_SEG];
|
|
};
|
|
|
|
/* create a mbuf with different sized segments
|
|
* and fill with data [0x00 0x01 0x02 ...]
|
|
*/
|
|
static struct rte_mbuf *
|
|
create_packet(struct rte_mempool *pktmbuf_pool,
|
|
struct test_case *test_data)
|
|
{
|
|
uint16_t i, ret, seg, seg_len = 0;
|
|
uint32_t last_index = 0;
|
|
unsigned int seg_lengths[MBUF_MAX_SEG];
|
|
unsigned int hdr_len;
|
|
struct rte_mbuf *pkt = NULL;
|
|
struct rte_mbuf *pkt_seg = NULL;
|
|
char *hdr = NULL;
|
|
char *data = NULL;
|
|
|
|
memcpy(seg_lengths, test_data->seg_lengths,
|
|
sizeof(unsigned int)*test_data->seg_count);
|
|
for (seg = 0; seg < test_data->seg_count; seg++) {
|
|
hdr_len = 0;
|
|
seg_len = seg_lengths[seg];
|
|
pkt_seg = rte_pktmbuf_alloc(pktmbuf_pool);
|
|
if (pkt_seg == NULL)
|
|
GOTO_FAIL("%s: mbuf allocation failed!\n", __func__);
|
|
if (rte_pktmbuf_pkt_len(pkt_seg) != 0)
|
|
GOTO_FAIL("%s: Bad packet length\n", __func__);
|
|
rte_mbuf_sanity_check(pkt_seg, 0);
|
|
/* Add header only for the first segment */
|
|
if (test_data->flags == MBUF_HEADER && seg == 0) {
|
|
hdr_len = sizeof(struct rte_ether_hdr);
|
|
/* prepend a header and fill with dummy data */
|
|
hdr = (char *)rte_pktmbuf_prepend(pkt_seg, hdr_len);
|
|
if (hdr == NULL)
|
|
GOTO_FAIL("%s: Cannot prepend header\n",
|
|
__func__);
|
|
if (rte_pktmbuf_pkt_len(pkt_seg) != hdr_len)
|
|
GOTO_FAIL("%s: Bad pkt length", __func__);
|
|
if (rte_pktmbuf_data_len(pkt_seg) != hdr_len)
|
|
GOTO_FAIL("%s: Bad data length", __func__);
|
|
for (i = 0; i < hdr_len; i++)
|
|
hdr[i] = (last_index + i) % 0xffff;
|
|
last_index += hdr_len;
|
|
}
|
|
/* skip appending segment with 0 length */
|
|
if (seg_len == 0)
|
|
continue;
|
|
data = rte_pktmbuf_append(pkt_seg, seg_len);
|
|
if (data == NULL)
|
|
GOTO_FAIL("%s: Cannot append data segment\n", __func__);
|
|
if (rte_pktmbuf_pkt_len(pkt_seg) != hdr_len + seg_len)
|
|
GOTO_FAIL("%s: Bad packet segment length: %d\n",
|
|
__func__, rte_pktmbuf_pkt_len(pkt_seg));
|
|
if (rte_pktmbuf_data_len(pkt_seg) != hdr_len + seg_len)
|
|
GOTO_FAIL("%s: Bad data length\n", __func__);
|
|
for (i = 0; i < seg_len; i++)
|
|
data[i] = (last_index + i) % 0xffff;
|
|
/* to fill continuous data from one seg to another */
|
|
last_index += i;
|
|
/* create chained mbufs */
|
|
if (seg == 0)
|
|
pkt = pkt_seg;
|
|
else {
|
|
ret = rte_pktmbuf_chain(pkt, pkt_seg);
|
|
if (ret != 0)
|
|
GOTO_FAIL("%s:FAIL: Chained mbuf creation %d\n",
|
|
__func__, ret);
|
|
}
|
|
|
|
pkt_seg = pkt_seg->next;
|
|
}
|
|
return pkt;
|
|
fail:
|
|
if (pkt != NULL) {
|
|
rte_pktmbuf_free(pkt);
|
|
pkt = NULL;
|
|
}
|
|
if (pkt_seg != NULL) {
|
|
rte_pktmbuf_free(pkt_seg);
|
|
pkt_seg = NULL;
|
|
}
|
|
return NULL;
|
|
}
|
|
|
|
static int
|
|
test_pktmbuf_read_from_chain(struct rte_mempool *pktmbuf_pool)
|
|
{
|
|
struct rte_mbuf *m;
|
|
struct test_case test_cases[] = {
|
|
{
|
|
.seg_lengths = { 100, 100, 100 },
|
|
.seg_count = 3,
|
|
.flags = MBUF_NO_HEADER,
|
|
.read_off = 0,
|
|
.read_len = 300
|
|
},
|
|
{
|
|
.seg_lengths = { 100, 125, 150 },
|
|
.seg_count = 3,
|
|
.flags = MBUF_NO_HEADER,
|
|
.read_off = 99,
|
|
.read_len = 201
|
|
},
|
|
{
|
|
.seg_lengths = { 100, 100 },
|
|
.seg_count = 2,
|
|
.flags = MBUF_NO_HEADER,
|
|
.read_off = 0,
|
|
.read_len = 100
|
|
},
|
|
{
|
|
.seg_lengths = { 100, 200 },
|
|
.seg_count = 2,
|
|
.flags = MBUF_HEADER,
|
|
.read_off = sizeof(struct rte_ether_hdr),
|
|
.read_len = 150
|
|
},
|
|
{
|
|
.seg_lengths = { 1000, 100 },
|
|
.seg_count = 2,
|
|
.flags = MBUF_NO_HEADER,
|
|
.read_off = 0,
|
|
.read_len = 1000
|
|
},
|
|
{
|
|
.seg_lengths = { 1024, 0, 100 },
|
|
.seg_count = 3,
|
|
.flags = MBUF_NO_HEADER,
|
|
.read_off = 100,
|
|
.read_len = 1001
|
|
},
|
|
{
|
|
.seg_lengths = { 1000, 1, 1000 },
|
|
.seg_count = 3,
|
|
.flags = MBUF_NO_HEADER,
|
|
.read_off = 1000,
|
|
.read_len = 2
|
|
},
|
|
{
|
|
.seg_lengths = { MBUF_TEST_DATA_LEN,
|
|
MBUF_TEST_DATA_LEN2,
|
|
MBUF_TEST_DATA_LEN3, 800, 10 },
|
|
.seg_count = 5,
|
|
.flags = MBUF_NEG_TEST_READ,
|
|
.read_off = 1000,
|
|
.read_len = MBUF_DATA_SIZE
|
|
},
|
|
};
|
|
|
|
uint32_t i, pos;
|
|
const char *data_copy = NULL;
|
|
char data_buf[MBUF_DATA_SIZE];
|
|
|
|
memset(data_buf, 0, MBUF_DATA_SIZE);
|
|
|
|
for (i = 0; i < RTE_DIM(test_cases); i++) {
|
|
m = create_packet(pktmbuf_pool, &test_cases[i]);
|
|
if (m == NULL)
|
|
GOTO_FAIL("%s: mbuf allocation failed!\n", __func__);
|
|
|
|
data_copy = rte_pktmbuf_read(m, test_cases[i].read_off,
|
|
test_cases[i].read_len, data_buf);
|
|
if (test_cases[i].flags == MBUF_NEG_TEST_READ) {
|
|
if (data_copy != NULL)
|
|
GOTO_FAIL("%s: mbuf data read should fail!\n",
|
|
__func__);
|
|
else {
|
|
rte_pktmbuf_free(m);
|
|
m = NULL;
|
|
continue;
|
|
}
|
|
}
|
|
if (data_copy == NULL)
|
|
GOTO_FAIL("%s: Error in reading packet data!\n",
|
|
__func__);
|
|
for (pos = 0; pos < test_cases[i].read_len; pos++) {
|
|
if (data_copy[pos] !=
|
|
(char)((test_cases[i].read_off + pos)
|
|
% 0xffff))
|
|
GOTO_FAIL("Data corrupted at offset %u is %2X",
|
|
pos, data_copy[pos]);
|
|
}
|
|
rte_pktmbuf_dump(stdout, m, rte_pktmbuf_pkt_len(m));
|
|
rte_pktmbuf_free(m);
|
|
m = NULL;
|
|
}
|
|
return 0;
|
|
|
|
fail:
|
|
if (m != NULL) {
|
|
rte_pktmbuf_free(m);
|
|
m = NULL;
|
|
}
|
|
return -1;
|
|
}
|
|
|
|
/* Define a free call back function to be used for external buffer */
|
|
static void
|
|
ext_buf_free_callback_fn(void *addr, void *opaque)
|
|
{
|
|
bool *freed = opaque;
|
|
|
|
if (addr == NULL) {
|
|
printf("External buffer address is invalid\n");
|
|
return;
|
|
}
|
|
rte_free(addr);
|
|
*freed = true;
|
|
printf("External buffer freed via callback\n");
|
|
}
|
|
|
|
/*
|
|
* Test to initialize shared data in external buffer before attaching to mbuf
|
|
* - Allocate mbuf with no data.
|
|
* - Allocate external buffer with size should be large enough to accommodate
|
|
* rte_mbuf_ext_shared_info.
|
|
* - Invoke pktmbuf_ext_shinfo_init_helper to initialize shared data.
|
|
* - Invoke rte_pktmbuf_attach_extbuf to attach external buffer to the mbuf.
|
|
* - Clone another mbuf and attach the same external buffer to it.
|
|
* - Invoke rte_pktmbuf_detach_extbuf to detach the external buffer from mbuf.
|
|
*/
|
|
static int
|
|
test_pktmbuf_ext_shinfo_init_helper(struct rte_mempool *pktmbuf_pool)
|
|
{
|
|
struct rte_mbuf *m = NULL;
|
|
struct rte_mbuf *clone = NULL;
|
|
struct rte_mbuf_ext_shared_info *ret_shinfo = NULL;
|
|
rte_iova_t buf_iova;
|
|
void *ext_buf_addr = NULL;
|
|
uint16_t buf_len = EXT_BUF_TEST_DATA_LEN +
|
|
sizeof(struct rte_mbuf_ext_shared_info);
|
|
bool freed = false;
|
|
|
|
/* alloc a mbuf */
|
|
m = rte_pktmbuf_alloc(pktmbuf_pool);
|
|
if (m == NULL)
|
|
GOTO_FAIL("%s: mbuf allocation failed!\n", __func__);
|
|
if (rte_pktmbuf_pkt_len(m) != 0)
|
|
GOTO_FAIL("%s: Bad packet length\n", __func__);
|
|
rte_mbuf_sanity_check(m, 0);
|
|
|
|
ext_buf_addr = rte_malloc("External buffer", buf_len,
|
|
RTE_CACHE_LINE_SIZE);
|
|
if (ext_buf_addr == NULL)
|
|
GOTO_FAIL("%s: External buffer allocation failed\n", __func__);
|
|
|
|
ret_shinfo = rte_pktmbuf_ext_shinfo_init_helper(ext_buf_addr, &buf_len,
|
|
ext_buf_free_callback_fn, &freed);
|
|
if (ret_shinfo == NULL)
|
|
GOTO_FAIL("%s: Shared info initialization failed!\n", __func__);
|
|
|
|
if (rte_mbuf_ext_refcnt_read(ret_shinfo) != 1)
|
|
GOTO_FAIL("%s: External refcount is not 1\n", __func__);
|
|
|
|
if (rte_mbuf_refcnt_read(m) != 1)
|
|
GOTO_FAIL("%s: Invalid refcnt in mbuf\n", __func__);
|
|
|
|
buf_iova = rte_mem_virt2iova(ext_buf_addr);
|
|
rte_pktmbuf_attach_extbuf(m, ext_buf_addr, buf_iova, buf_len,
|
|
ret_shinfo);
|
|
if (m->ol_flags != RTE_MBUF_F_EXTERNAL)
|
|
GOTO_FAIL("%s: External buffer is not attached to mbuf\n",
|
|
__func__);
|
|
|
|
/* allocate one more mbuf */
|
|
clone = rte_pktmbuf_clone(m, pktmbuf_pool);
|
|
if (clone == NULL)
|
|
GOTO_FAIL("%s: mbuf clone allocation failed!\n", __func__);
|
|
if (rte_pktmbuf_pkt_len(clone) != 0)
|
|
GOTO_FAIL("%s: Bad packet length\n", __func__);
|
|
|
|
/* attach the same external buffer to the cloned mbuf */
|
|
rte_pktmbuf_attach_extbuf(clone, ext_buf_addr, buf_iova, buf_len,
|
|
ret_shinfo);
|
|
if (clone->ol_flags != RTE_MBUF_F_EXTERNAL)
|
|
GOTO_FAIL("%s: External buffer is not attached to mbuf\n",
|
|
__func__);
|
|
|
|
if (rte_mbuf_ext_refcnt_read(ret_shinfo) != 2)
|
|
GOTO_FAIL("%s: Invalid ext_buf ref_cnt\n", __func__);
|
|
if (freed)
|
|
GOTO_FAIL("%s: extbuf should not be freed\n", __func__);
|
|
|
|
/* test to manually update ext_buf_ref_cnt from 2 to 3*/
|
|
rte_mbuf_ext_refcnt_update(ret_shinfo, 1);
|
|
if (rte_mbuf_ext_refcnt_read(ret_shinfo) != 3)
|
|
GOTO_FAIL("%s: Update ext_buf ref_cnt failed\n", __func__);
|
|
if (freed)
|
|
GOTO_FAIL("%s: extbuf should not be freed\n", __func__);
|
|
|
|
/* reset the ext_refcnt before freeing the external buffer */
|
|
rte_mbuf_ext_refcnt_set(ret_shinfo, 2);
|
|
if (rte_mbuf_ext_refcnt_read(ret_shinfo) != 2)
|
|
GOTO_FAIL("%s: set ext_buf ref_cnt failed\n", __func__);
|
|
if (freed)
|
|
GOTO_FAIL("%s: extbuf should not be freed\n", __func__);
|
|
|
|
/* detach the external buffer from mbufs */
|
|
rte_pktmbuf_detach_extbuf(m);
|
|
/* check if ref cnt is decremented */
|
|
if (rte_mbuf_ext_refcnt_read(ret_shinfo) != 1)
|
|
GOTO_FAIL("%s: Invalid ext_buf ref_cnt\n", __func__);
|
|
if (freed)
|
|
GOTO_FAIL("%s: extbuf should not be freed\n", __func__);
|
|
|
|
rte_pktmbuf_detach_extbuf(clone);
|
|
if (!freed)
|
|
GOTO_FAIL("%s: extbuf should be freed\n", __func__);
|
|
freed = false;
|
|
|
|
rte_pktmbuf_free(m);
|
|
m = NULL;
|
|
rte_pktmbuf_free(clone);
|
|
clone = NULL;
|
|
|
|
return 0;
|
|
|
|
fail:
|
|
if (m) {
|
|
rte_pktmbuf_free(m);
|
|
m = NULL;
|
|
}
|
|
if (clone) {
|
|
rte_pktmbuf_free(clone);
|
|
clone = NULL;
|
|
}
|
|
if (ext_buf_addr != NULL) {
|
|
rte_free(ext_buf_addr);
|
|
ext_buf_addr = NULL;
|
|
}
|
|
return -1;
|
|
}
|
|
|
|
/*
|
|
* Test the mbuf pool with pinned external data buffers
|
|
* - Allocate memory zone for external buffer
|
|
* - Create the mbuf pool with pinned external buffer
|
|
* - Check the created pool with relevant mbuf pool unit tests
|
|
*/
|
|
static int
|
|
test_pktmbuf_ext_pinned_buffer(struct rte_mempool *std_pool)
|
|
{
|
|
|
|
struct rte_pktmbuf_extmem ext_mem;
|
|
struct rte_mempool *pinned_pool = NULL;
|
|
const struct rte_memzone *mz = NULL;
|
|
|
|
printf("Test mbuf pool with external pinned data buffers\n");
|
|
|
|
/* Allocate memzone for the external data buffer */
|
|
mz = rte_memzone_reserve("pinned_pool",
|
|
NB_MBUF * MBUF_DATA_SIZE,
|
|
SOCKET_ID_ANY,
|
|
RTE_MEMZONE_2MB | RTE_MEMZONE_SIZE_HINT_ONLY);
|
|
if (mz == NULL)
|
|
GOTO_FAIL("%s: Memzone allocation failed\n", __func__);
|
|
|
|
/* Create the mbuf pool with pinned external data buffer */
|
|
ext_mem.buf_ptr = mz->addr;
|
|
ext_mem.buf_iova = mz->iova;
|
|
ext_mem.buf_len = mz->len;
|
|
ext_mem.elt_size = MBUF_DATA_SIZE;
|
|
|
|
pinned_pool = rte_pktmbuf_pool_create_extbuf("test_pinned_pool",
|
|
NB_MBUF, MEMPOOL_CACHE_SIZE, 0,
|
|
MBUF_DATA_SIZE, SOCKET_ID_ANY,
|
|
&ext_mem, 1);
|
|
if (pinned_pool == NULL)
|
|
GOTO_FAIL("%s: Mbuf pool with pinned external"
|
|
" buffer creation failed\n", __func__);
|
|
/* test multiple mbuf alloc */
|
|
if (test_pktmbuf_pool(pinned_pool) < 0)
|
|
GOTO_FAIL("%s: test_mbuf_pool(pinned) failed\n",
|
|
__func__);
|
|
|
|
/* do it another time to check that all mbufs were freed */
|
|
if (test_pktmbuf_pool(pinned_pool) < 0)
|
|
GOTO_FAIL("%s: test_mbuf_pool(pinned) failed (2)\n",
|
|
__func__);
|
|
|
|
/* test that the data pointer on a packet mbuf is set properly */
|
|
if (test_pktmbuf_pool_ptr(pinned_pool) < 0)
|
|
GOTO_FAIL("%s: test_pktmbuf_pool_ptr(pinned) failed\n",
|
|
__func__);
|
|
|
|
/* test data manipulation in mbuf with non-ascii data */
|
|
if (test_pktmbuf_with_non_ascii_data(pinned_pool) < 0)
|
|
GOTO_FAIL("%s: test_pktmbuf_with_non_ascii_data(pinned)"
|
|
" failed\n", __func__);
|
|
|
|
/* test free pktmbuf segment one by one */
|
|
if (test_pktmbuf_free_segment(pinned_pool) < 0)
|
|
GOTO_FAIL("%s: test_pktmbuf_free_segment(pinned) failed\n",
|
|
__func__);
|
|
|
|
if (testclone_testupdate_testdetach(pinned_pool, std_pool) < 0)
|
|
GOTO_FAIL("%s: testclone_and_testupdate(pinned) failed\n",
|
|
__func__);
|
|
|
|
if (test_pktmbuf_copy(pinned_pool, std_pool) < 0)
|
|
GOTO_FAIL("%s: test_pktmbuf_copy(pinned) failed\n",
|
|
__func__);
|
|
|
|
if (test_failing_mbuf_sanity_check(pinned_pool) < 0)
|
|
GOTO_FAIL("%s: test_failing_mbuf_sanity_check(pinned)"
|
|
" failed\n", __func__);
|
|
|
|
if (test_mbuf_linearize_check(pinned_pool) < 0)
|
|
GOTO_FAIL("%s: test_mbuf_linearize_check(pinned) failed\n",
|
|
__func__);
|
|
|
|
/* test for allocating a bulk of mbufs with various sizes */
|
|
if (test_pktmbuf_alloc_bulk(pinned_pool) < 0)
|
|
GOTO_FAIL("%s: test_rte_pktmbuf_alloc_bulk(pinned) failed\n",
|
|
__func__);
|
|
|
|
/* test for allocating a bulk of mbufs with various sizes */
|
|
if (test_neg_pktmbuf_alloc_bulk(pinned_pool) < 0)
|
|
GOTO_FAIL("%s: test_neg_rte_pktmbuf_alloc_bulk(pinned)"
|
|
" failed\n", __func__);
|
|
|
|
/* test to read mbuf packet */
|
|
if (test_pktmbuf_read(pinned_pool) < 0)
|
|
GOTO_FAIL("%s: test_rte_pktmbuf_read(pinned) failed\n",
|
|
__func__);
|
|
|
|
/* test to read mbuf packet from offset */
|
|
if (test_pktmbuf_read_from_offset(pinned_pool) < 0)
|
|
GOTO_FAIL("%s: test_rte_pktmbuf_read_from_offset(pinned)"
|
|
" failed\n", __func__);
|
|
|
|
/* test to read data from chain of mbufs with data segments */
|
|
if (test_pktmbuf_read_from_chain(pinned_pool) < 0)
|
|
GOTO_FAIL("%s: test_rte_pktmbuf_read_from_chain(pinned)"
|
|
" failed\n", __func__);
|
|
|
|
RTE_SET_USED(std_pool);
|
|
rte_mempool_free(pinned_pool);
|
|
rte_memzone_free(mz);
|
|
return 0;
|
|
|
|
fail:
|
|
rte_mempool_free(pinned_pool);
|
|
rte_memzone_free(mz);
|
|
return -1;
|
|
}
|
|
|
|
static int
|
|
test_mbuf_dyn(struct rte_mempool *pktmbuf_pool)
|
|
{
|
|
const struct rte_mbuf_dynfield dynfield = {
|
|
.name = "test-dynfield",
|
|
.size = sizeof(uint8_t),
|
|
.align = __alignof__(uint8_t),
|
|
.flags = 0,
|
|
};
|
|
const struct rte_mbuf_dynfield dynfield2 = {
|
|
.name = "test-dynfield2",
|
|
.size = sizeof(uint16_t),
|
|
.align = __alignof__(uint16_t),
|
|
.flags = 0,
|
|
};
|
|
const struct rte_mbuf_dynfield dynfield3 = {
|
|
.name = "test-dynfield3",
|
|
.size = sizeof(uint8_t),
|
|
.align = __alignof__(uint8_t),
|
|
.flags = 0,
|
|
};
|
|
const struct rte_mbuf_dynfield dynfield_fail_big = {
|
|
.name = "test-dynfield-fail-big",
|
|
.size = 256,
|
|
.align = 1,
|
|
.flags = 0,
|
|
};
|
|
const struct rte_mbuf_dynfield dynfield_fail_align = {
|
|
.name = "test-dynfield-fail-align",
|
|
.size = 1,
|
|
.align = 3,
|
|
.flags = 0,
|
|
};
|
|
const struct rte_mbuf_dynfield dynfield_fail_flag = {
|
|
.name = "test-dynfield",
|
|
.size = sizeof(uint8_t),
|
|
.align = __alignof__(uint8_t),
|
|
.flags = 1,
|
|
};
|
|
const struct rte_mbuf_dynflag dynflag_fail_flag = {
|
|
.name = "test-dynflag",
|
|
.flags = 1,
|
|
};
|
|
const struct rte_mbuf_dynflag dynflag = {
|
|
.name = "test-dynflag",
|
|
.flags = 0,
|
|
};
|
|
const struct rte_mbuf_dynflag dynflag2 = {
|
|
.name = "test-dynflag2",
|
|
.flags = 0,
|
|
};
|
|
const struct rte_mbuf_dynflag dynflag3 = {
|
|
.name = "test-dynflag3",
|
|
.flags = 0,
|
|
};
|
|
struct rte_mbuf *m = NULL;
|
|
int offset, offset2, offset3;
|
|
int flag, flag2, flag3;
|
|
int ret;
|
|
|
|
printf("Test mbuf dynamic fields and flags\n");
|
|
rte_mbuf_dyn_dump(stdout);
|
|
|
|
offset = rte_mbuf_dynfield_register(&dynfield);
|
|
if (offset == -1)
|
|
GOTO_FAIL("failed to register dynamic field, offset=%d: %s",
|
|
offset, strerror(errno));
|
|
|
|
ret = rte_mbuf_dynfield_register(&dynfield);
|
|
if (ret != offset)
|
|
GOTO_FAIL("failed to lookup dynamic field, ret=%d: %s",
|
|
ret, strerror(errno));
|
|
|
|
offset2 = rte_mbuf_dynfield_register(&dynfield2);
|
|
if (offset2 == -1 || offset2 == offset || (offset2 & 1))
|
|
GOTO_FAIL("failed to register dynamic field 2, offset2=%d: %s",
|
|
offset2, strerror(errno));
|
|
|
|
offset3 = rte_mbuf_dynfield_register_offset(&dynfield3,
|
|
offsetof(struct rte_mbuf, dynfield1[1]));
|
|
if (offset3 != offsetof(struct rte_mbuf, dynfield1[1])) {
|
|
if (rte_errno == EBUSY)
|
|
printf("mbuf test error skipped: dynfield is busy\n");
|
|
else
|
|
GOTO_FAIL("failed to register dynamic field 3, offset="
|
|
"%d: %s", offset3, strerror(errno));
|
|
}
|
|
|
|
printf("dynfield: offset=%d, offset2=%d, offset3=%d\n",
|
|
offset, offset2, offset3);
|
|
|
|
ret = rte_mbuf_dynfield_register(&dynfield_fail_big);
|
|
if (ret != -1)
|
|
GOTO_FAIL("dynamic field creation should fail (too big)");
|
|
|
|
ret = rte_mbuf_dynfield_register(&dynfield_fail_align);
|
|
if (ret != -1)
|
|
GOTO_FAIL("dynamic field creation should fail (bad alignment)");
|
|
|
|
ret = rte_mbuf_dynfield_register_offset(&dynfield_fail_align,
|
|
offsetof(struct rte_mbuf, ol_flags));
|
|
if (ret != -1)
|
|
GOTO_FAIL("dynamic field creation should fail (not avail)");
|
|
|
|
ret = rte_mbuf_dynfield_register(&dynfield_fail_flag);
|
|
if (ret != -1)
|
|
GOTO_FAIL("dynamic field creation should fail (invalid flag)");
|
|
|
|
ret = rte_mbuf_dynflag_register(&dynflag_fail_flag);
|
|
if (ret != -1)
|
|
GOTO_FAIL("dynamic flag creation should fail (invalid flag)");
|
|
|
|
flag = rte_mbuf_dynflag_register(&dynflag);
|
|
if (flag == -1)
|
|
GOTO_FAIL("failed to register dynamic flag, flag=%d: %s",
|
|
flag, strerror(errno));
|
|
|
|
ret = rte_mbuf_dynflag_register(&dynflag);
|
|
if (ret != flag)
|
|
GOTO_FAIL("failed to lookup dynamic flag, ret=%d: %s",
|
|
ret, strerror(errno));
|
|
|
|
flag2 = rte_mbuf_dynflag_register(&dynflag2);
|
|
if (flag2 == -1 || flag2 == flag)
|
|
GOTO_FAIL("failed to register dynamic flag 2, flag2=%d: %s",
|
|
flag2, strerror(errno));
|
|
|
|
flag3 = rte_mbuf_dynflag_register_bitnum(&dynflag3,
|
|
rte_bsf64(RTE_MBUF_F_LAST_FREE));
|
|
if ((uint32_t)flag3 != rte_bsf64(RTE_MBUF_F_LAST_FREE))
|
|
GOTO_FAIL("failed to register dynamic flag 3, flag3=%d: %s",
|
|
flag3, strerror(errno));
|
|
|
|
printf("dynflag: flag=%d, flag2=%d, flag3=%d\n", flag, flag2, flag3);
|
|
|
|
/* set, get dynamic field */
|
|
m = rte_pktmbuf_alloc(pktmbuf_pool);
|
|
if (m == NULL)
|
|
GOTO_FAIL("Cannot allocate mbuf");
|
|
|
|
*RTE_MBUF_DYNFIELD(m, offset, uint8_t *) = 1;
|
|
if (*RTE_MBUF_DYNFIELD(m, offset, uint8_t *) != 1)
|
|
GOTO_FAIL("failed to read dynamic field");
|
|
*RTE_MBUF_DYNFIELD(m, offset2, uint16_t *) = 1000;
|
|
if (*RTE_MBUF_DYNFIELD(m, offset2, uint16_t *) != 1000)
|
|
GOTO_FAIL("failed to read dynamic field");
|
|
|
|
/* set a dynamic flag */
|
|
m->ol_flags |= (1ULL << flag);
|
|
|
|
rte_mbuf_dyn_dump(stdout);
|
|
rte_pktmbuf_free(m);
|
|
return 0;
|
|
fail:
|
|
rte_pktmbuf_free(m);
|
|
return -1;
|
|
}
|
|
|
|
/* check that m->nb_segs and m->next are reset on mbuf free */
|
|
static int
|
|
test_nb_segs_and_next_reset(void)
|
|
{
|
|
struct rte_mbuf *m0 = NULL, *m1 = NULL, *m2 = NULL;
|
|
struct rte_mempool *pool = NULL;
|
|
|
|
pool = rte_pktmbuf_pool_create("test_mbuf_reset",
|
|
3, 0, 0, MBUF_DATA_SIZE, SOCKET_ID_ANY);
|
|
if (pool == NULL)
|
|
GOTO_FAIL("Failed to create mbuf pool");
|
|
|
|
/* alloc mbufs */
|
|
m0 = rte_pktmbuf_alloc(pool);
|
|
m1 = rte_pktmbuf_alloc(pool);
|
|
m2 = rte_pktmbuf_alloc(pool);
|
|
if (m0 == NULL || m1 == NULL || m2 == NULL)
|
|
GOTO_FAIL("Failed to allocate mbuf");
|
|
|
|
/* append data in all of them */
|
|
if (rte_pktmbuf_append(m0, 500) == NULL ||
|
|
rte_pktmbuf_append(m1, 500) == NULL ||
|
|
rte_pktmbuf_append(m2, 500) == NULL)
|
|
GOTO_FAIL("Failed to append data in mbuf");
|
|
|
|
/* chain them in one mbuf m0 */
|
|
rte_pktmbuf_chain(m1, m2);
|
|
rte_pktmbuf_chain(m0, m1);
|
|
if (m0->nb_segs != 3 || m0->next != m1 || m1->next != m2 ||
|
|
m2->next != NULL) {
|
|
m1 = m2 = NULL;
|
|
GOTO_FAIL("Failed to chain mbufs");
|
|
}
|
|
|
|
/* split m0 chain in two, between m1 and m2 */
|
|
m0->nb_segs = 2;
|
|
m1->next = NULL;
|
|
m2->nb_segs = 1;
|
|
|
|
/* free the 2 mbuf chains m0 and m2 */
|
|
rte_pktmbuf_free(m0);
|
|
rte_pktmbuf_free(m2);
|
|
|
|
/* realloc the 3 mbufs */
|
|
m0 = rte_mbuf_raw_alloc(pool);
|
|
m1 = rte_mbuf_raw_alloc(pool);
|
|
m2 = rte_mbuf_raw_alloc(pool);
|
|
if (m0 == NULL || m1 == NULL || m2 == NULL)
|
|
GOTO_FAIL("Failed to reallocate mbuf");
|
|
|
|
/* ensure that m->next and m->nb_segs are reset allocated mbufs */
|
|
if (m0->nb_segs != 1 || m0->next != NULL ||
|
|
m1->nb_segs != 1 || m1->next != NULL ||
|
|
m2->nb_segs != 1 || m2->next != NULL)
|
|
GOTO_FAIL("nb_segs or next was not reset properly");
|
|
|
|
return 0;
|
|
|
|
fail:
|
|
rte_mempool_free(pool);
|
|
return -1;
|
|
}
|
|
|
|
static int
|
|
test_mbuf(void)
|
|
{
|
|
int ret = -1;
|
|
struct rte_mempool *pktmbuf_pool = NULL;
|
|
struct rte_mempool *pktmbuf_pool2 = NULL;
|
|
|
|
|
|
RTE_BUILD_BUG_ON(sizeof(struct rte_mbuf) != RTE_CACHE_LINE_MIN_SIZE * 2);
|
|
|
|
/* create pktmbuf pool if it does not exist */
|
|
pktmbuf_pool = rte_pktmbuf_pool_create("test_pktmbuf_pool",
|
|
NB_MBUF, MEMPOOL_CACHE_SIZE, 0, MBUF_DATA_SIZE,
|
|
SOCKET_ID_ANY);
|
|
|
|
if (pktmbuf_pool == NULL) {
|
|
printf("cannot allocate mbuf pool\n");
|
|
goto err;
|
|
}
|
|
|
|
/* test registration of dynamic fields and flags */
|
|
if (test_mbuf_dyn(pktmbuf_pool) < 0) {
|
|
printf("mbuf dynflag test failed\n");
|
|
goto err;
|
|
}
|
|
|
|
/* create a specific pktmbuf pool with a priv_size != 0 and no data
|
|
* room size */
|
|
pktmbuf_pool2 = rte_pktmbuf_pool_create("test_pktmbuf_pool2",
|
|
NB_MBUF, MEMPOOL_CACHE_SIZE, MBUF2_PRIV_SIZE, 0,
|
|
SOCKET_ID_ANY);
|
|
|
|
if (pktmbuf_pool2 == NULL) {
|
|
printf("cannot allocate mbuf pool\n");
|
|
goto err;
|
|
}
|
|
|
|
/* test multiple mbuf alloc */
|
|
if (test_pktmbuf_pool(pktmbuf_pool) < 0) {
|
|
printf("test_mbuf_pool() failed\n");
|
|
goto err;
|
|
}
|
|
|
|
/* do it another time to check that all mbufs were freed */
|
|
if (test_pktmbuf_pool(pktmbuf_pool) < 0) {
|
|
printf("test_mbuf_pool() failed (2)\n");
|
|
goto err;
|
|
}
|
|
|
|
/* test bulk mbuf alloc and free */
|
|
if (test_pktmbuf_pool_bulk() < 0) {
|
|
printf("test_pktmbuf_pool_bulk() failed\n");
|
|
goto err;
|
|
}
|
|
|
|
/* test that the pointer to the data on a packet mbuf is set properly */
|
|
if (test_pktmbuf_pool_ptr(pktmbuf_pool) < 0) {
|
|
printf("test_pktmbuf_pool_ptr() failed\n");
|
|
goto err;
|
|
}
|
|
|
|
/* test data manipulation in mbuf */
|
|
if (test_one_pktmbuf(pktmbuf_pool) < 0) {
|
|
printf("test_one_mbuf() failed\n");
|
|
goto err;
|
|
}
|
|
|
|
|
|
/*
|
|
* do it another time, to check that allocation reinitialize
|
|
* the mbuf correctly
|
|
*/
|
|
if (test_one_pktmbuf(pktmbuf_pool) < 0) {
|
|
printf("test_one_mbuf() failed (2)\n");
|
|
goto err;
|
|
}
|
|
|
|
if (test_pktmbuf_with_non_ascii_data(pktmbuf_pool) < 0) {
|
|
printf("test_pktmbuf_with_non_ascii_data() failed\n");
|
|
goto err;
|
|
}
|
|
|
|
/* test free pktmbuf segment one by one */
|
|
if (test_pktmbuf_free_segment(pktmbuf_pool) < 0) {
|
|
printf("test_pktmbuf_free_segment() failed.\n");
|
|
goto err;
|
|
}
|
|
|
|
if (testclone_testupdate_testdetach(pktmbuf_pool, pktmbuf_pool) < 0) {
|
|
printf("testclone_and_testupdate() failed \n");
|
|
goto err;
|
|
}
|
|
|
|
if (test_pktmbuf_copy(pktmbuf_pool, pktmbuf_pool) < 0) {
|
|
printf("test_pktmbuf_copy() failed\n");
|
|
goto err;
|
|
}
|
|
|
|
if (test_attach_from_different_pool(pktmbuf_pool, pktmbuf_pool2) < 0) {
|
|
printf("test_attach_from_different_pool() failed\n");
|
|
goto err;
|
|
}
|
|
|
|
if (test_refcnt_mbuf() < 0) {
|
|
printf("test_refcnt_mbuf() failed \n");
|
|
goto err;
|
|
}
|
|
|
|
if (test_failing_mbuf_sanity_check(pktmbuf_pool) < 0) {
|
|
printf("test_failing_mbuf_sanity_check() failed\n");
|
|
goto err;
|
|
}
|
|
|
|
if (test_mbuf_linearize_check(pktmbuf_pool) < 0) {
|
|
printf("test_mbuf_linearize_check() failed\n");
|
|
goto err;
|
|
}
|
|
|
|
if (test_tx_offload() < 0) {
|
|
printf("test_tx_offload() failed\n");
|
|
goto err;
|
|
}
|
|
|
|
if (test_get_rx_ol_flag_list() < 0) {
|
|
printf("test_rte_get_rx_ol_flag_list() failed\n");
|
|
goto err;
|
|
}
|
|
|
|
if (test_get_tx_ol_flag_list() < 0) {
|
|
printf("test_rte_get_tx_ol_flag_list() failed\n");
|
|
goto err;
|
|
}
|
|
|
|
if (test_get_rx_ol_flag_name() < 0) {
|
|
printf("test_rte_get_rx_ol_flag_name() failed\n");
|
|
goto err;
|
|
}
|
|
|
|
if (test_get_tx_ol_flag_name() < 0) {
|
|
printf("test_rte_get_tx_ol_flag_name() failed\n");
|
|
goto err;
|
|
}
|
|
|
|
if (test_mbuf_validate_tx_offload_one(pktmbuf_pool) < 0) {
|
|
printf("test_mbuf_validate_tx_offload_one() failed\n");
|
|
goto err;
|
|
}
|
|
|
|
/* test for allocating a bulk of mbufs with various sizes */
|
|
if (test_pktmbuf_alloc_bulk(pktmbuf_pool) < 0) {
|
|
printf("test_rte_pktmbuf_alloc_bulk() failed\n");
|
|
goto err;
|
|
}
|
|
|
|
/* test for allocating a bulk of mbufs with various sizes */
|
|
if (test_neg_pktmbuf_alloc_bulk(pktmbuf_pool) < 0) {
|
|
printf("test_neg_rte_pktmbuf_alloc_bulk() failed\n");
|
|
goto err;
|
|
}
|
|
|
|
/* test to read mbuf packet */
|
|
if (test_pktmbuf_read(pktmbuf_pool) < 0) {
|
|
printf("test_rte_pktmbuf_read() failed\n");
|
|
goto err;
|
|
}
|
|
|
|
/* test to read mbuf packet from offset */
|
|
if (test_pktmbuf_read_from_offset(pktmbuf_pool) < 0) {
|
|
printf("test_rte_pktmbuf_read_from_offset() failed\n");
|
|
goto err;
|
|
}
|
|
|
|
/* test to read data from chain of mbufs with data segments */
|
|
if (test_pktmbuf_read_from_chain(pktmbuf_pool) < 0) {
|
|
printf("test_rte_pktmbuf_read_from_chain() failed\n");
|
|
goto err;
|
|
}
|
|
|
|
/* test to initialize shared info. at the end of external buffer */
|
|
if (test_pktmbuf_ext_shinfo_init_helper(pktmbuf_pool) < 0) {
|
|
printf("test_pktmbuf_ext_shinfo_init_helper() failed\n");
|
|
goto err;
|
|
}
|
|
|
|
/* test the mbuf pool with pinned external data buffers */
|
|
if (test_pktmbuf_ext_pinned_buffer(pktmbuf_pool) < 0) {
|
|
printf("test_pktmbuf_ext_pinned_buffer() failed\n");
|
|
goto err;
|
|
}
|
|
|
|
/* test reset of m->nb_segs and m->next on mbuf free */
|
|
if (test_nb_segs_and_next_reset() < 0) {
|
|
printf("test_nb_segs_and_next_reset() failed\n");
|
|
goto err;
|
|
}
|
|
|
|
ret = 0;
|
|
err:
|
|
rte_mempool_free(pktmbuf_pool);
|
|
rte_mempool_free(pktmbuf_pool2);
|
|
return ret;
|
|
}
|
|
#undef GOTO_FAIL
|
|
|
|
REGISTER_TEST_COMMAND(mbuf_autotest, test_mbuf);
|