a9de470cc7
Since all other apps have been moved to the "app" folder, the autotest app remains alone in the test folder. Rather than having an entire top-level folder for this, we can move it back to where it all started in early versions of DPDK - the "app/" folder. This move has a couple of advantages: * This reduces clutter at the top level of the project, due to one less folder. * It eliminates the separate build task necessary for building the autotests using make "make test-build" which means that developers are less likely to miss something in their own compilation tests * It re-aligns the final location of the test binary in the app folder when building with make with it's location in the source tree. For meson builds, the autotest app is different from the other apps in that it needs a series of different test cases defined for it for use by "meson test". Therefore, it does not get built as part of the main loop in the app folder, but gets built separately at the end. Signed-off-by: Bruce Richardson <bruce.richardson@intel.com>
166 lines
4.5 KiB
C
166 lines
4.5 KiB
C
/* SPDX-License-Identifier: BSD-3-Clause
|
|
* Copyright(c) 2015 Intel Corporation
|
|
*/
|
|
|
|
|
|
#include <stdio.h>
|
|
#include <inttypes.h>
|
|
#include <rte_ring.h>
|
|
#include <rte_cycles.h>
|
|
#include <rte_launch.h>
|
|
#include <rte_ethdev.h>
|
|
#include <rte_eth_ring.h>
|
|
#include <rte_bus_vdev.h>
|
|
|
|
#include "test.h"
|
|
|
|
#define RING_NAME "RING_PERF"
|
|
#define RING_SIZE 4096
|
|
#define MAX_BURST 32
|
|
|
|
/*
|
|
* the sizes to enqueue and dequeue in testing
|
|
* (marked volatile so they won't be seen as compile-time constants)
|
|
*/
|
|
static const volatile unsigned bulk_sizes[] = { 1, 8, 32 };
|
|
|
|
/* The ring structure used for tests */
|
|
static struct rte_ring *r;
|
|
static uint16_t ring_ethdev_port;
|
|
|
|
/* Get cycle counts for dequeuing from an empty ring. Should be 2 or 3 cycles */
|
|
static void
|
|
test_empty_dequeue(void)
|
|
{
|
|
const unsigned iter_shift = 26;
|
|
const unsigned iterations = 1 << iter_shift;
|
|
unsigned i = 0;
|
|
void *burst[MAX_BURST];
|
|
|
|
const uint64_t sc_start = rte_rdtsc();
|
|
for (i = 0; i < iterations; i++)
|
|
rte_ring_sc_dequeue_bulk(r, burst, bulk_sizes[0], NULL);
|
|
const uint64_t sc_end = rte_rdtsc();
|
|
|
|
const uint64_t eth_start = rte_rdtsc();
|
|
for (i = 0; i < iterations; i++)
|
|
rte_eth_rx_burst(ring_ethdev_port, 0, (void *)burst,
|
|
bulk_sizes[0]);
|
|
const uint64_t eth_end = rte_rdtsc();
|
|
|
|
printf("Ring empty dequeue : %.1F\n",
|
|
(double)(sc_end - sc_start) / iterations);
|
|
printf("Ethdev empty dequeue: %.1F\n",
|
|
(double)(eth_end - eth_start) / iterations);
|
|
}
|
|
|
|
/*
|
|
* Test function that determines how long an enqueue + dequeue of a single item
|
|
* takes on a single lcore. Result is for comparison with the bulk enq+deq.
|
|
*/
|
|
static void
|
|
test_single_enqueue_dequeue(void)
|
|
{
|
|
const unsigned iter_shift = 24;
|
|
const unsigned iterations = 1 << iter_shift;
|
|
unsigned i = 0;
|
|
void *burst = NULL;
|
|
struct rte_mbuf *mburst[1] = { NULL };
|
|
|
|
const uint64_t sc_start = rte_rdtsc_precise();
|
|
rte_compiler_barrier();
|
|
for (i = 0; i < iterations; i++) {
|
|
rte_ring_enqueue_bulk(r, &burst, 1, NULL);
|
|
rte_ring_dequeue_bulk(r, &burst, 1, NULL);
|
|
}
|
|
const uint64_t sc_end = rte_rdtsc_precise();
|
|
rte_compiler_barrier();
|
|
|
|
const uint64_t eth_start = rte_rdtsc_precise();
|
|
rte_compiler_barrier();
|
|
for (i = 0; i < iterations; i++) {
|
|
rte_eth_tx_burst(ring_ethdev_port, 0, mburst, 1);
|
|
rte_eth_rx_burst(ring_ethdev_port, 0, mburst, 1);
|
|
}
|
|
const uint64_t eth_end = rte_rdtsc_precise();
|
|
rte_compiler_barrier();
|
|
|
|
printf("Ring single enq/dequeue : %"PRIu64"\n",
|
|
(sc_end-sc_start) >> iter_shift);
|
|
printf("Ethdev single enq/dequeue: %"PRIu64"\n",
|
|
(eth_end-eth_start) >> iter_shift);
|
|
}
|
|
|
|
/* Times enqueue and dequeue on a single lcore */
|
|
static void
|
|
test_bulk_enqueue_dequeue(void)
|
|
{
|
|
const unsigned iter_shift = 23;
|
|
const unsigned iterations = 1 << iter_shift;
|
|
unsigned sz, i = 0;
|
|
struct rte_mbuf *burst[MAX_BURST] = {0};
|
|
|
|
for (sz = 0; sz < sizeof(bulk_sizes)/sizeof(bulk_sizes[0]); sz++) {
|
|
const uint64_t sc_start = rte_rdtsc();
|
|
for (i = 0; i < iterations; i++) {
|
|
rte_ring_sp_enqueue_bulk(r, (void *)burst,
|
|
bulk_sizes[sz], NULL);
|
|
rte_ring_sc_dequeue_bulk(r, (void *)burst,
|
|
bulk_sizes[sz], NULL);
|
|
}
|
|
const uint64_t sc_end = rte_rdtsc();
|
|
|
|
const uint64_t eth_start = rte_rdtsc_precise();
|
|
rte_compiler_barrier();
|
|
for (i = 0; i < iterations; i++) {
|
|
rte_eth_tx_burst(ring_ethdev_port, 0, burst, bulk_sizes[sz]);
|
|
rte_eth_rx_burst(ring_ethdev_port, 0, burst, bulk_sizes[sz]);
|
|
}
|
|
const uint64_t eth_end = rte_rdtsc_precise();
|
|
rte_compiler_barrier();
|
|
|
|
double sc_avg = ((double)(sc_end-sc_start) /
|
|
(iterations * bulk_sizes[sz]));
|
|
double eth_avg = ((double)(eth_end-eth_start) /
|
|
(iterations * bulk_sizes[sz]));
|
|
|
|
printf("ring bulk enq/deq (size: %u) : %.1F\n", bulk_sizes[sz],
|
|
sc_avg);
|
|
printf("ethdev bulk enq/deq (size:%u): %.1F\n", bulk_sizes[sz],
|
|
eth_avg);
|
|
|
|
printf("\n");
|
|
}
|
|
}
|
|
|
|
static int
|
|
test_ring_pmd_perf(void)
|
|
{
|
|
char name[RTE_ETH_NAME_MAX_LEN];
|
|
|
|
r = rte_ring_create(RING_NAME, RING_SIZE, rte_socket_id(),
|
|
RING_F_SP_ENQ|RING_F_SC_DEQ);
|
|
if (r == NULL && (r = rte_ring_lookup(RING_NAME)) == NULL)
|
|
return -1;
|
|
|
|
ring_ethdev_port = rte_eth_from_ring(r);
|
|
|
|
printf("\n### Testing const single element enq/deq ###\n");
|
|
test_single_enqueue_dequeue();
|
|
|
|
printf("\n### Testing empty dequeue ###\n");
|
|
test_empty_dequeue();
|
|
|
|
printf("\n### Testing using a single lcore ###\n");
|
|
test_bulk_enqueue_dequeue();
|
|
|
|
/* release port and ring resources */
|
|
rte_eth_dev_stop(ring_ethdev_port);
|
|
rte_eth_dev_get_name_by_port(ring_ethdev_port, name);
|
|
rte_vdev_uninit(name);
|
|
rte_ring_free(r);
|
|
return 0;
|
|
}
|
|
|
|
REGISTER_TEST_COMMAND(ring_pmd_perf_autotest, test_ring_pmd_perf);
|