examples/ip_pipeline: rework config file syntax

New syntax of config files is needed for ip_pipeline example
enhancements. Some old files are temporarily disabled in the Makefile.
It is part of a bigger change.

Signed-off-by: Pawel Wodkowski <pawelx.wodkowski@intel.com>
Acked-by: Cristian Dumitrescu <cristian.dumitrescu@intel.com>
This commit is contained in:
Pawel Wodkowski 2015-07-07 10:09:25 +02:00 committed by Thomas Monjalon
parent d27f716ad7
commit 7f64b9c004
11 changed files with 4730 additions and 853 deletions

View File

@ -43,20 +43,9 @@ APP = ip_pipeline
# all source are stored in SRCS-y
SRCS-$(CONFIG_RTE_LIBRTE_PIPELINE) := main.c
SRCS-$(CONFIG_RTE_LIBRTE_PIPELINE) += config.c
SRCS-$(CONFIG_RTE_LIBRTE_PIPELINE) += init.c
SRCS-$(CONFIG_RTE_LIBRTE_PIPELINE) += cmdline.c
SRCS-$(CONFIG_RTE_LIBRTE_PIPELINE) += pipeline_rx.c
SRCS-$(CONFIG_RTE_LIBRTE_PIPELINE) += pipeline_tx.c
SRCS-$(CONFIG_RTE_LIBRTE_PIPELINE) += pipeline_flow_classification.c
SRCS-$(CONFIG_RTE_LIBRTE_PIPELINE) += pipeline_routing.c
SRCS-$(CONFIG_RTE_LIBRTE_PIPELINE) += pipeline_passthrough.c
SRCS-$(CONFIG_RTE_LIBRTE_PIPELINE) += pipeline_ipv4_frag.c
SRCS-$(CONFIG_RTE_LIBRTE_PIPELINE) += pipeline_ipv4_ras.c
ifeq ($(CONFIG_RTE_LIBRTE_ACL),y)
SRCS-$(CONFIG_RTE_LIBRTE_PIPELINE) += pipeline_firewall.c
endif
SRCS-$(CONFIG_RTE_LIBRTE_PIPELINE) += config_parse.c
SRCS-$(CONFIG_RTE_LIBRTE_PIPELINE) += config_parse_tm.c
SRCS-$(CONFIG_RTE_LIBRTE_PIPELINE) += cpu_core_map.c
CFLAGS += -O3
CFLAGS += $(WERROR_FLAGS)

909
examples/ip_pipeline/app.h Normal file
View File

@ -0,0 +1,909 @@
/*-
* BSD LICENSE
*
* Copyright(c) 2010-2015 Intel Corporation. All rights reserved.
* All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
*
* * Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* * Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in
* the documentation and/or other materials provided with the
* distribution.
* * Neither the name of Intel Corporation nor the names of its
* contributors may be used to endorse or promote products derived
* from this software without specific prior written permission.
*
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
* OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*/
#ifndef __INCLUDE_APP_H__
#define __INCLUDE_APP_H__
#include <stdint.h>
#include <string.h>
#include <rte_common.h>
#include <rte_mempool.h>
#include <rte_ring.h>
#include <rte_sched.h>
#include <cmdline_parse.h>
#include <rte_ethdev.h>
#include "cpu_core_map.h"
#include "pipeline.h"
#define APP_PARAM_NAME_SIZE PIPELINE_NAME_SIZE
struct app_mempool_params {
char *name;
uint32_t parsed;
uint32_t buffer_size;
uint32_t pool_size;
uint32_t cache_size;
uint32_t cpu_socket_id;
};
struct app_link_params {
char *name;
uint32_t parsed;
uint32_t pmd_id; /* Generated based on port mask */
uint32_t arp_q; /* 0 = Disabled (packets go to default queue 0) */
uint32_t tcp_syn_local_q; /* 0 = Disabled (pkts go to default queue) */
uint32_t ip_local_q; /* 0 = Disabled (pkts go to default queue 0) */
uint32_t tcp_local_q; /* 0 = Disabled (pkts go to default queue 0) */
uint32_t udp_local_q; /* 0 = Disabled (pkts go to default queue 0) */
uint32_t sctp_local_q; /* 0 = Disabled (pkts go to default queue 0) */
uint32_t state; /* DOWN = 0, UP = 1 */
uint32_t ip; /* 0 = Invalid */
uint32_t depth; /* Valid only when IP is valid */
uint64_t mac_addr; /* Read from HW */
struct rte_eth_conf conf;
uint8_t promisc;
};
struct app_pktq_hwq_in_params {
char *name;
uint32_t parsed;
uint32_t mempool_id; /* Position in the app->mempool_params */
uint32_t size;
uint32_t burst;
struct rte_eth_rxconf conf;
};
struct app_pktq_hwq_out_params {
char *name;
uint32_t parsed;
uint32_t size;
uint32_t burst;
uint32_t dropless;
uint64_t n_retries;
struct rte_eth_txconf conf;
};
struct app_pktq_swq_params {
char *name;
uint32_t parsed;
uint32_t size;
uint32_t burst_read;
uint32_t burst_write;
uint32_t dropless;
uint64_t n_retries;
uint32_t cpu_socket_id;
};
#ifndef APP_FILE_NAME_SIZE
#define APP_FILE_NAME_SIZE 256
#endif
#ifndef APP_MAX_SCHED_SUBPORTS
#define APP_MAX_SCHED_SUBPORTS 8
#endif
#ifndef APP_MAX_SCHED_PIPES
#define APP_MAX_SCHED_PIPES 4096
#endif
struct app_pktq_tm_params {
char *name;
uint32_t parsed;
const char *file_name;
struct rte_sched_port_params sched_port_params;
struct rte_sched_subport_params
sched_subport_params[APP_MAX_SCHED_SUBPORTS];
struct rte_sched_pipe_params
sched_pipe_profiles[RTE_SCHED_PIPE_PROFILES_PER_PORT];
int sched_pipe_to_profile[APP_MAX_SCHED_SUBPORTS * APP_MAX_SCHED_PIPES];
uint32_t burst_read;
uint32_t burst_write;
};
struct app_pktq_source_params {
char *name;
uint32_t parsed;
uint32_t mempool_id; /* Position in the app->mempool_params array */
uint32_t burst;
};
struct app_pktq_sink_params {
char *name;
uint8_t parsed;
};
struct app_msgq_params {
char *name;
uint32_t parsed;
uint32_t size;
uint32_t cpu_socket_id;
};
enum app_pktq_in_type {
APP_PKTQ_IN_HWQ,
APP_PKTQ_IN_SWQ,
APP_PKTQ_IN_TM,
APP_PKTQ_IN_SOURCE,
};
struct app_pktq_in_params {
enum app_pktq_in_type type;
uint32_t id; /* Position in the appropriate app array */
};
enum app_pktq_out_type {
APP_PKTQ_OUT_HWQ,
APP_PKTQ_OUT_SWQ,
APP_PKTQ_OUT_TM,
APP_PKTQ_OUT_SINK,
};
struct app_pktq_out_params {
enum app_pktq_out_type type;
uint32_t id; /* Position in the appropriate app array */
};
#ifndef APP_PIPELINE_TYPE_SIZE
#define APP_PIPELINE_TYPE_SIZE 64
#endif
#define APP_MAX_PIPELINE_PKTQ_IN PIPELINE_MAX_PORT_IN
#define APP_MAX_PIPELINE_PKTQ_OUT PIPELINE_MAX_PORT_OUT
#define APP_MAX_PIPELINE_MSGQ_IN PIPELINE_MAX_MSGQ_IN
#define APP_MAX_PIPELINE_MSGQ_OUT PIPELINE_MAX_MSGQ_OUT
#define APP_MAX_PIPELINE_ARGS PIPELINE_MAX_ARGS
struct app_pipeline_params {
char *name;
uint8_t parsed;
char type[APP_PIPELINE_TYPE_SIZE];
uint32_t socket_id;
uint32_t core_id;
uint32_t hyper_th_id;
struct app_pktq_in_params pktq_in[APP_MAX_PIPELINE_PKTQ_IN];
struct app_pktq_out_params pktq_out[APP_MAX_PIPELINE_PKTQ_OUT];
uint32_t msgq_in[APP_MAX_PIPELINE_MSGQ_IN];
uint32_t msgq_out[APP_MAX_PIPELINE_MSGQ_OUT];
uint32_t n_pktq_in;
uint32_t n_pktq_out;
uint32_t n_msgq_in;
uint32_t n_msgq_out;
uint32_t timer_period;
char *args_name[APP_MAX_PIPELINE_ARGS];
char *args_value[APP_MAX_PIPELINE_ARGS];
uint32_t n_args;
};
struct app_pipeline_data {
void *be;
void *fe;
uint64_t timer_period;
};
struct app_thread_pipeline_data {
void *be;
pipeline_be_op_run f_run;
pipeline_be_op_timer f_timer;
uint64_t timer_period;
uint64_t deadline;
};
#ifndef APP_MAX_THREAD_PIPELINES
#define APP_MAX_THREAD_PIPELINES 16
#endif
struct app_thread_data {
struct app_thread_pipeline_data regular[APP_MAX_THREAD_PIPELINES];
struct app_thread_pipeline_data custom[APP_MAX_THREAD_PIPELINES];
uint32_t n_regular;
uint32_t n_custom;
uint64_t deadline;
};
struct app_eal_params {
/* Map lcore set to physical cpu set */
char *coremap;
/* Core ID that is used as master */
uint32_t master_lcore_present;
uint32_t master_lcore;
/* Number of memory channels */
uint32_t channels_present;
uint32_t channels;
/* Memory to allocate (see also --socket-mem) */
uint32_t memory_present;
uint32_t memory;
/* Force number of memory ranks (don't detect) */
uint32_t ranks_present;
uint32_t ranks;
/* Add a PCI device in black list. */
char *pci_blacklist;
/* Add a PCI device in white list. */
char *pci_whitelist;
/* Add a virtual device. */
char *vdev;
/* Use VMware TSC map instead of native RDTSC */
uint32_t vmware_tsc_map_present;
int vmware_tsc_map;
/* Type of this process (primary|secondary|auto) */
char *proc_type;
/* Set syslog facility */
char *syslog;
/* Set default log level */
uint32_t log_level_present;
uint32_t log_level;
/* Display version information on startup */
uint32_t version_present;
int version;
/* This help */
uint32_t help_present;
int help;
/* Use malloc instead of hugetlbfs */
uint32_t no_huge_present;
int no_huge;
/* Disable PCI */
uint32_t no_pci_present;
int no_pci;
/* Disable HPET */
uint32_t no_hpet_present;
int no_hpet;
/* No shared config (mmap'd files) */
uint32_t no_shconf_present;
int no_shconf;
/* Add driver */
char *add_driver;
/* Memory to allocate on sockets (comma separated values)*/
char *socket_mem;
/* Directory where hugetlbfs is mounted */
char *huge_dir;
/* Prefix for hugepage filenames */
char *file_prefix;
/* Base virtual address */
char *base_virtaddr;
/* Create /dev/uioX (usually done by hotplug) */
uint32_t create_uio_dev_present;
int create_uio_dev;
/* Interrupt mode for VFIO (legacy|msi|msix) */
char *vfio_intr;
/* Support running on Xen dom0 without hugetlbfs */
uint32_t xen_dom0_present;
int xen_dom0;
};
#ifndef APP_APPNAME_SIZE
#define APP_APPNAME_SIZE 256
#endif
#ifndef APP_MAX_MEMPOOLS
#define APP_MAX_MEMPOOLS 8
#endif
#ifndef APP_MAX_LINKS
#define APP_MAX_LINKS 16
#endif
#ifndef APP_LINK_MAX_HWQ_IN
#define APP_LINK_MAX_HWQ_IN 64
#endif
#ifndef APP_LINK_MAX_HWQ_OUT
#define APP_LINK_MAX_HWQ_OUT 64
#endif
#define APP_MAX_HWQ_IN (APP_MAX_LINKS * APP_LINK_MAX_HWQ_IN)
#define APP_MAX_HWQ_OUT (APP_MAX_LINKS * APP_LINK_MAX_HWQ_OUT)
#ifndef APP_MAX_PKTQ_SWQ
#define APP_MAX_PKTQ_SWQ 256
#endif
#define APP_MAX_PKTQ_TM APP_MAX_LINKS
#ifndef APP_MAX_PKTQ_SOURCE
#define APP_MAX_PKTQ_SOURCE 16
#endif
#ifndef APP_MAX_PKTQ_SINK
#define APP_MAX_PKTQ_SINK 16
#endif
#ifndef APP_MAX_MSGQ
#define APP_MAX_MSGQ 64
#endif
#ifndef APP_MAX_PIPELINES
#define APP_MAX_PIPELINES 64
#endif
#ifndef APP_EAL_ARGC
#define APP_EAL_ARGC 32
#endif
#ifndef APP_MAX_PIPELINE_TYPES
#define APP_MAX_PIPELINE_TYPES 64
#endif
#ifndef APP_MAX_THREADS
#define APP_MAX_THREADS RTE_MAX_LCORE
#endif
#ifndef APP_MAX_CMDS
#define APP_MAX_CMDS 64
#endif
struct app_params {
/* Config */
char app_name[APP_APPNAME_SIZE];
const char *config_file;
const char *script_file;
uint64_t port_mask;
uint32_t log_level;
struct app_eal_params eal_params;
struct app_mempool_params mempool_params[APP_MAX_MEMPOOLS];
struct app_link_params link_params[APP_MAX_LINKS];
struct app_pktq_hwq_in_params hwq_in_params[APP_MAX_HWQ_IN];
struct app_pktq_hwq_out_params hwq_out_params[APP_MAX_HWQ_OUT];
struct app_pktq_swq_params swq_params[APP_MAX_PKTQ_SWQ];
struct app_pktq_tm_params tm_params[APP_MAX_PKTQ_TM];
struct app_pktq_source_params source_params[APP_MAX_PKTQ_SOURCE];
struct app_pktq_sink_params sink_params[APP_MAX_PKTQ_SINK];
struct app_msgq_params msgq_params[APP_MAX_MSGQ];
struct app_pipeline_params pipeline_params[APP_MAX_PIPELINES];
uint32_t n_mempools;
uint32_t n_links;
uint32_t n_pktq_hwq_in;
uint32_t n_pktq_hwq_out;
uint32_t n_pktq_swq;
uint32_t n_pktq_tm;
uint32_t n_pktq_source;
uint32_t n_pktq_sink;
uint32_t n_msgq;
uint32_t n_pipelines;
/* Init */
char *eal_argv[1 + APP_EAL_ARGC];
struct cpu_core_map *core_map;
uint64_t core_mask;
struct rte_mempool *mempool[APP_MAX_MEMPOOLS];
struct rte_ring *swq[APP_MAX_PKTQ_SWQ];
struct rte_sched_port *tm[APP_MAX_PKTQ_TM];
struct rte_ring *msgq[APP_MAX_MSGQ];
struct pipeline_type pipeline_type[APP_MAX_PIPELINE_TYPES];
struct app_pipeline_data pipeline_data[APP_MAX_PIPELINES];
struct app_thread_data thread_data[APP_MAX_THREADS];
cmdline_parse_ctx_t cmds[APP_MAX_CMDS + 1];
int eal_argc;
uint32_t n_pipeline_types;
uint32_t n_cmds;
};
#define APP_PARAM_VALID(obj) ((obj)->name != NULL)
#define APP_PARAM_COUNT(obj_array, n_objs) \
{ \
size_t i; \
\
n_objs = 0; \
for (i = 0; i < RTE_DIM(obj_array); i++) \
if (APP_PARAM_VALID(&((obj_array)[i]))) \
n_objs++; \
}
#define APP_PARAM_FIND(obj_array, key) \
({ \
ssize_t obj_idx; \
const ssize_t obj_count = RTE_DIM(obj_array); \
\
for (obj_idx = 0; obj_idx < obj_count; obj_idx++) { \
if (!APP_PARAM_VALID(&((obj_array)[obj_idx]))) \
continue; \
\
if (strcmp(key, (obj_array)[obj_idx].name) == 0) \
break; \
} \
obj_idx < obj_count ? obj_idx : -ENOENT; \
})
#define APP_PARAM_FIND_BY_ID(obj_array, prefix, id, obj) \
do { \
char name[APP_PARAM_NAME_SIZE]; \
ssize_t pos; \
\
sprintf(name, prefix "%" PRIu32, id); \
pos = APP_PARAM_FIND(obj_array, name); \
obj = (pos < 0) ? NULL : &((obj_array)[pos]); \
} while (0)
#define APP_PARAM_GET_ID(obj, prefix, id) \
do \
sscanf(obj->name, prefix "%" SCNu32, &id); \
while (0) \
#define APP_PARAM_ADD(obj_array, obj_name) \
({ \
ssize_t obj_idx; \
const ssize_t obj_count = RTE_DIM(obj_array); \
\
obj_idx = APP_PARAM_FIND(obj_array, obj_name); \
if (obj_idx < 0) { \
for (obj_idx = 0; obj_idx < obj_count; obj_idx++) { \
if (!APP_PARAM_VALID(&((obj_array)[obj_idx]))) \
break; \
} \
\
if (obj_idx < obj_count) { \
(obj_array)[obj_idx].name = strdup(obj_name); \
if ((obj_array)[obj_idx].name == NULL) \
obj_idx = -EINVAL; \
} else \
obj_idx = -ENOMEM; \
} \
obj_idx; \
})
#define APP_CHECK(exp, fmt, ...) \
do { \
if (!(exp)) { \
fprintf(stderr, fmt "\n", ## __VA_ARGS__); \
abort(); \
} \
} while (0)
enum app_log_level {
APP_LOG_LEVEL_HIGH = 1,
APP_LOG_LEVEL_LOW,
APP_LOG_LEVELS
};
#define APP_LOG(app, level, fmt, ...) \
do { \
if (app->log_level >= APP_LOG_LEVEL_ ## level) \
fprintf(stdout, "[APP] " fmt "\n", ## __VA_ARGS__); \
} while (0)
static inline uint32_t
app_link_get_n_rxq(struct app_params *app, struct app_link_params *link)
{
uint32_t n_rxq = 0, link_id, i;
uint32_t n_pktq_hwq_in = RTE_MIN(app->n_pktq_hwq_in,
RTE_DIM(app->hwq_in_params));
APP_PARAM_GET_ID(link, "LINK", link_id);
for (i = 0; i < n_pktq_hwq_in; i++) {
struct app_pktq_hwq_in_params *p = &app->hwq_in_params[i];
uint32_t rxq_link_id, rxq_queue_id;
sscanf(p->name, "RXQ%" SCNu32 ".%" SCNu32,
&rxq_link_id, &rxq_queue_id);
if (rxq_link_id == link_id)
n_rxq++;
}
return n_rxq;
}
static inline uint32_t
app_link_get_n_txq(struct app_params *app, struct app_link_params *link)
{
uint32_t n_txq = 0, link_id, i;
uint32_t n_pktq_hwq_out = RTE_MIN(app->n_pktq_hwq_out,
RTE_DIM(app->hwq_out_params));
APP_PARAM_GET_ID(link, "LINK", link_id);
for (i = 0; i < n_pktq_hwq_out; i++) {
struct app_pktq_hwq_out_params *p = &app->hwq_out_params[i];
uint32_t txq_link_id, txq_queue_id;
sscanf(p->name, "TXQ%" SCNu32 ".%" SCNu32,
&txq_link_id, &txq_queue_id);
if (txq_link_id == link_id)
n_txq++;
}
return n_txq;
}
static inline uint32_t
app_rxq_get_readers(struct app_params *app, struct app_pktq_hwq_in_params *rxq)
{
uint32_t pos = rxq - app->hwq_in_params;
uint32_t n_pipelines = RTE_MIN(app->n_pipelines,
RTE_DIM(app->pipeline_params));
uint32_t n_readers = 0, i;
for (i = 0; i < n_pipelines; i++) {
struct app_pipeline_params *p = &app->pipeline_params[i];
uint32_t n_pktq_in = RTE_MIN(p->n_pktq_in, RTE_DIM(p->pktq_in));
uint32_t j;
for (j = 0; j < n_pktq_in; j++) {
struct app_pktq_in_params *pktq = &p->pktq_in[j];
if ((pktq->type == APP_PKTQ_IN_HWQ) &&
(pktq->id == pos))
n_readers++;
}
}
return n_readers;
}
static inline uint32_t
app_swq_get_readers(struct app_params *app, struct app_pktq_swq_params *swq)
{
uint32_t pos = swq - app->swq_params;
uint32_t n_pipelines = RTE_MIN(app->n_pipelines,
RTE_DIM(app->pipeline_params));
uint32_t n_readers = 0, i;
for (i = 0; i < n_pipelines; i++) {
struct app_pipeline_params *p = &app->pipeline_params[i];
uint32_t n_pktq_in = RTE_MIN(p->n_pktq_in, RTE_DIM(p->pktq_in));
uint32_t j;
for (j = 0; j < n_pktq_in; j++) {
struct app_pktq_in_params *pktq = &p->pktq_in[j];
if ((pktq->type == APP_PKTQ_IN_SWQ) &&
(pktq->id == pos))
n_readers++;
}
}
return n_readers;
}
static inline uint32_t
app_tm_get_readers(struct app_params *app, struct app_pktq_tm_params *tm)
{
uint32_t pos = tm - app->tm_params;
uint32_t n_pipelines = RTE_MIN(app->n_pipelines,
RTE_DIM(app->pipeline_params));
uint32_t n_readers = 0, i;
for (i = 0; i < n_pipelines; i++) {
struct app_pipeline_params *p = &app->pipeline_params[i];
uint32_t n_pktq_in = RTE_MIN(p->n_pktq_in, RTE_DIM(p->pktq_in));
uint32_t j;
for (j = 0; j < n_pktq_in; j++) {
struct app_pktq_in_params *pktq = &p->pktq_in[j];
if ((pktq->type == APP_PKTQ_IN_TM) &&
(pktq->id == pos))
n_readers++;
}
}
return n_readers;
}
static inline uint32_t
app_source_get_readers(struct app_params *app,
struct app_pktq_source_params *source)
{
uint32_t pos = source - app->source_params;
uint32_t n_pipelines = RTE_MIN(app->n_pipelines,
RTE_DIM(app->pipeline_params));
uint32_t n_readers = 0, i;
for (i = 0; i < n_pipelines; i++) {
struct app_pipeline_params *p = &app->pipeline_params[i];
uint32_t n_pktq_in = RTE_MIN(p->n_pktq_in, RTE_DIM(p->pktq_in));
uint32_t j;
for (j = 0; j < n_pktq_in; j++) {
struct app_pktq_in_params *pktq = &p->pktq_in[j];
if ((pktq->type == APP_PKTQ_IN_SOURCE) &&
(pktq->id == pos))
n_readers++;
}
}
return n_readers;
}
static inline uint32_t
app_msgq_get_readers(struct app_params *app, struct app_msgq_params *msgq)
{
uint32_t pos = msgq - app->msgq_params;
uint32_t n_pipelines = RTE_MIN(app->n_pipelines,
RTE_DIM(app->pipeline_params));
uint32_t n_readers = 0, i;
for (i = 0; i < n_pipelines; i++) {
struct app_pipeline_params *p = &app->pipeline_params[i];
uint32_t n_msgq_in = RTE_MIN(p->n_msgq_in, RTE_DIM(p->msgq_in));
uint32_t j;
for (j = 0; j < n_msgq_in; j++)
if (p->msgq_in[j] == pos)
n_readers++;
}
return n_readers;
}
static inline uint32_t
app_txq_get_writers(struct app_params *app, struct app_pktq_hwq_out_params *txq)
{
uint32_t pos = txq - app->hwq_out_params;
uint32_t n_pipelines = RTE_MIN(app->n_pipelines,
RTE_DIM(app->pipeline_params));
uint32_t n_writers = 0, i;
for (i = 0; i < n_pipelines; i++) {
struct app_pipeline_params *p = &app->pipeline_params[i];
uint32_t n_pktq_out = RTE_MIN(p->n_pktq_out,
RTE_DIM(p->pktq_out));
uint32_t j;
for (j = 0; j < n_pktq_out; j++) {
struct app_pktq_out_params *pktq = &p->pktq_out[j];
if ((pktq->type == APP_PKTQ_OUT_HWQ) &&
(pktq->id == pos))
n_writers++;
}
}
return n_writers;
}
static inline uint32_t
app_swq_get_writers(struct app_params *app, struct app_pktq_swq_params *swq)
{
uint32_t pos = swq - app->swq_params;
uint32_t n_pipelines = RTE_MIN(app->n_pipelines,
RTE_DIM(app->pipeline_params));
uint32_t n_writers = 0, i;
for (i = 0; i < n_pipelines; i++) {
struct app_pipeline_params *p = &app->pipeline_params[i];
uint32_t n_pktq_out = RTE_MIN(p->n_pktq_out,
RTE_DIM(p->pktq_out));
uint32_t j;
for (j = 0; j < n_pktq_out; j++) {
struct app_pktq_out_params *pktq = &p->pktq_out[j];
if ((pktq->type == APP_PKTQ_OUT_SWQ) &&
(pktq->id == pos))
n_writers++;
}
}
return n_writers;
}
static inline uint32_t
app_tm_get_writers(struct app_params *app, struct app_pktq_tm_params *tm)
{
uint32_t pos = tm - app->tm_params;
uint32_t n_pipelines = RTE_MIN(app->n_pipelines,
RTE_DIM(app->pipeline_params));
uint32_t n_writers = 0, i;
for (i = 0; i < n_pipelines; i++) {
struct app_pipeline_params *p = &app->pipeline_params[i];
uint32_t n_pktq_out = RTE_MIN(p->n_pktq_out,
RTE_DIM(p->pktq_out));
uint32_t j;
for (j = 0; j < n_pktq_out; j++) {
struct app_pktq_out_params *pktq = &p->pktq_out[j];
if ((pktq->type == APP_PKTQ_OUT_TM) &&
(pktq->id == pos))
n_writers++;
}
}
return n_writers;
}
static inline uint32_t
app_sink_get_writers(struct app_params *app, struct app_pktq_sink_params *sink)
{
uint32_t pos = sink - app->sink_params;
uint32_t n_pipelines = RTE_MIN(app->n_pipelines,
RTE_DIM(app->pipeline_params));
uint32_t n_writers = 0, i;
for (i = 0; i < n_pipelines; i++) {
struct app_pipeline_params *p = &app->pipeline_params[i];
uint32_t n_pktq_out = RTE_MIN(p->n_pktq_out,
RTE_DIM(p->pktq_out));
uint32_t j;
for (j = 0; j < n_pktq_out; j++) {
struct app_pktq_out_params *pktq = &p->pktq_out[j];
if ((pktq->type == APP_PKTQ_OUT_SINK) &&
(pktq->id == pos))
n_writers++;
}
}
return n_writers;
}
static inline uint32_t
app_msgq_get_writers(struct app_params *app, struct app_msgq_params *msgq)
{
uint32_t pos = msgq - app->msgq_params;
uint32_t n_pipelines = RTE_MIN(app->n_pipelines,
RTE_DIM(app->pipeline_params));
uint32_t n_writers = 0, i;
for (i = 0; i < n_pipelines; i++) {
struct app_pipeline_params *p = &app->pipeline_params[i];
uint32_t n_msgq_out = RTE_MIN(p->n_msgq_out,
RTE_DIM(p->msgq_out));
uint32_t j;
for (j = 0; j < n_msgq_out; j++)
if (p->msgq_out[j] == pos)
n_writers++;
}
return n_writers;
}
static inline struct app_link_params *
app_get_link_for_rxq(struct app_params *app, struct app_pktq_hwq_in_params *p)
{
char link_name[APP_PARAM_NAME_SIZE];
ssize_t link_param_idx;
uint32_t rxq_link_id, rxq_queue_id;
sscanf(p->name, "RXQ%" SCNu32 ".%" SCNu32,
&rxq_link_id, &rxq_queue_id);
sprintf(link_name, "LINK%" PRIu32, rxq_link_id);
link_param_idx = APP_PARAM_FIND(app->link_params, link_name);
APP_CHECK((link_param_idx >= 0),
"Cannot find %s for %s", link_name, p->name);
return &app->link_params[link_param_idx];
}
static inline struct app_link_params *
app_get_link_for_txq(struct app_params *app, struct app_pktq_hwq_out_params *p)
{
char link_name[APP_PARAM_NAME_SIZE];
ssize_t link_param_idx;
uint32_t txq_link_id, txq_queue_id;
sscanf(p->name, "TXQ%" SCNu32 ".%" SCNu32,
&txq_link_id, &txq_queue_id);
sprintf(link_name, "LINK%" PRIu32, txq_link_id);
link_param_idx = APP_PARAM_FIND(app->link_params, link_name);
APP_CHECK((link_param_idx >= 0),
"Cannot find %s for %s", link_name, p->name);
return &app->link_params[link_param_idx];
}
static inline struct app_link_params *
app_get_link_for_tm(struct app_params *app, struct app_pktq_tm_params *p_tm)
{
char link_name[APP_PARAM_NAME_SIZE];
uint32_t link_id;
ssize_t link_param_idx;
sscanf(p_tm->name, "TM%" PRIu32, &link_id);
sprintf(link_name, "LINK%" PRIu32, link_id);
link_param_idx = APP_PARAM_FIND(app->link_params, link_name);
APP_CHECK((link_param_idx >= 0),
"Cannot find %s for %s", link_name, p_tm->name);
return &app->link_params[link_param_idx];
}
int app_config_init(struct app_params *app);
int app_config_args(struct app_params *app,
int argc, char **argv);
int app_config_parse(struct app_params *app,
const char *file_name);
int app_config_parse_tm(struct app_params *app);
void app_config_save(struct app_params *app,
const char *file_name);
int app_config_check(struct app_params *app);
int app_init(struct app_params *app);
int app_thread(void *arg);
int app_pipeline_type_register(struct app_params *app,
struct pipeline_type *ptype);
struct pipeline_type *app_pipeline_type_find(struct app_params *app,
char *name);
void app_link_up_internal(struct app_params *app,
struct app_link_params *cp);
void app_link_down_internal(struct app_params *app,
struct app_link_params *cp);
#endif

View File

@ -1,419 +0,0 @@
/*-
* BSD LICENSE
*
* Copyright(c) 2010-2014 Intel Corporation. All rights reserved.
* All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
*
* * Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* * Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in
* the documentation and/or other materials provided with the
* distribution.
* * Neither the name of Intel Corporation nor the names of its
* contributors may be used to endorse or promote products derived
* from this software without specific prior written permission.
*
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
* OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*/
#include <stdio.h>
#include <stdlib.h>
#include <stdint.h>
#include <inttypes.h>
#include <sys/types.h>
#include <string.h>
#include <sys/queue.h>
#include <stdarg.h>
#include <errno.h>
#include <getopt.h>
#include <rte_common.h>
#include <rte_byteorder.h>
#include <rte_log.h>
#include <rte_memory.h>
#include <rte_memcpy.h>
#include <rte_memzone.h>
#include <rte_eal.h>
#include <rte_per_lcore.h>
#include <rte_launch.h>
#include <rte_atomic.h>
#include <rte_cycles.h>
#include <rte_prefetch.h>
#include <rte_lcore.h>
#include <rte_per_lcore.h>
#include <rte_branch_prediction.h>
#include <rte_interrupts.h>
#include <rte_pci.h>
#include <rte_random.h>
#include <rte_debug.h>
#include <rte_ether.h>
#include <rte_ethdev.h>
#include <rte_ring.h>
#include <rte_mempool.h>
#include <rte_mbuf.h>
#include <rte_ip.h>
#include <rte_tcp.h>
#include <rte_lpm.h>
#include <rte_lpm6.h>
#include <rte_string_fns.h>
#include <rte_cfgfile.h>
#include "main.h"
struct app_params app;
static const char usage[] =
"Usage: %s EAL_OPTIONS-- -p PORT_MASK [-f CONFIG_FILE]\n";
void
app_print_usage(char *prgname)
{
printf(usage, prgname);
}
const char *
app_core_type_id_to_string(enum app_core_type id)
{
switch (id) {
case APP_CORE_NONE: return "NONE";
case APP_CORE_MASTER: return "MASTER";
case APP_CORE_RX: return "RX";
case APP_CORE_TX: return "TX";
case APP_CORE_PT: return "PT";
case APP_CORE_FC: return "FC";
case APP_CORE_FW: return "FW";
case APP_CORE_RT: return "RT";
case APP_CORE_TM: return "TM";
case APP_CORE_IPV4_FRAG: return "IPV4_FRAG";
case APP_CORE_IPV4_RAS: return "IPV4_RAS";
default: return NULL;
}
}
int
app_core_type_string_to_id(const char *string, enum app_core_type *id)
{
if (strcmp(string, "NONE") == 0) {
*id = APP_CORE_NONE;
return 0;
}
if (strcmp(string, "MASTER") == 0) {
*id = APP_CORE_MASTER;
return 0;
}
if (strcmp(string, "RX") == 0) {
*id = APP_CORE_RX;
return 0;
}
if (strcmp(string, "TX") == 0) {
*id = APP_CORE_TX;
return 0;
}
if (strcmp(string, "PT") == 0) {
*id = APP_CORE_PT;
return 0;
}
if (strcmp(string, "FC") == 0) {
*id = APP_CORE_FC;
return 0;
}
if (strcmp(string, "FW") == 0) {
*id = APP_CORE_FW;
return 0;
}
if (strcmp(string, "RT") == 0) {
*id = APP_CORE_RT;
return 0;
}
if (strcmp(string, "TM") == 0) {
*id = APP_CORE_TM;
return 0;
}
if (strcmp(string, "IPV4_FRAG") == 0) {
*id = APP_CORE_IPV4_FRAG;
return 0;
}
if (strcmp(string, "IPV4_RAS") == 0) {
*id = APP_CORE_IPV4_RAS;
return 0;
}
return -1;
}
static uint64_t
app_get_core_mask(void)
{
uint64_t core_mask = 0;
uint32_t i;
for (i = 0; i < RTE_MAX_LCORE; i++) {
if (rte_lcore_is_enabled(i) == 0)
continue;
core_mask |= 1LLU << i;
}
return core_mask;
}
static int
app_install_coremask(uint64_t core_mask)
{
uint32_t n_cores, i;
for (n_cores = 0, i = 0; i < RTE_MAX_LCORE; i++)
if (app.cores[i].core_type != APP_CORE_NONE)
n_cores++;
if (n_cores != app.n_cores) {
rte_panic("Number of cores in COREMASK should be %u instead "
"of %u\n", n_cores, app.n_cores);
return -1;
}
for (i = 0; i < RTE_MAX_LCORE; i++) {
uint32_t core_id;
if (app.cores[i].core_type == APP_CORE_NONE)
continue;
core_id = __builtin_ctzll(core_mask);
core_mask &= ~(1LLU << core_id);
app.cores[i].core_id = core_id;
}
return 0;
}
static int
app_install_cfgfile(const char *file_name)
{
struct rte_cfgfile *file;
uint32_t n_cores, i;
memset(app.cores, 0, sizeof(app.cores));
if (file_name[0] == '\0')
return -1;
file = rte_cfgfile_load(file_name, 0);
if (file == NULL) {
rte_panic("Config file %s not found\n", file_name);
return -1;
}
n_cores = (uint32_t) rte_cfgfile_num_sections(file, "core",
strnlen("core", 5));
if (n_cores < app.n_cores) {
rte_panic("Config file parse error: not enough cores specified "
"(%u cores missing)\n", app.n_cores - n_cores);
return -1;
}
if (n_cores > app.n_cores) {
rte_panic("Config file parse error: too many cores specified "
"(%u cores too many)\n", n_cores - app.n_cores);
return -1;
}
for (i = 0; i < n_cores; i++) {
struct app_core_params *p = &app.cores[i];
char section_name[16];
const char *entry;
uint32_t j;
/* [core X] */
snprintf(section_name, sizeof(section_name), "core %u", i);
if (!rte_cfgfile_has_section(file, section_name)) {
rte_panic("Config file parse error: core IDs are not "
"sequential (core %u missing)\n", i);
return -1;
}
/* type */
entry = rte_cfgfile_get_entry(file, section_name, "type");
if (!entry) {
rte_panic("Config file parse error: core %u type not "
"defined\n", i);
return -1;
}
if ((app_core_type_string_to_id(entry, &p->core_type) != 0) ||
(p->core_type == APP_CORE_NONE)) {
rte_panic("Config file parse error: core %u type "
"error\n", i);
return -1;
}
/* queues in */
entry = rte_cfgfile_get_entry(file, section_name, "queues in");
if (!entry) {
rte_panic("Config file parse error: core %u queues in "
"not defined\n", i);
return -1;
}
for (j = 0; (j < APP_MAX_SWQ_PER_CORE) && (entry != NULL);
j++) {
char *next;
p->swq_in[j] = (uint32_t) strtol(entry, &next, 10);
if (next == entry)
break;
entry = next;
}
if ((j != APP_MAX_SWQ_PER_CORE) || (*entry != '\0')) {
rte_panic("Config file parse error: core %u queues in "
"error\n", i);
return -1;
}
/* queues out */
entry = rte_cfgfile_get_entry(file, section_name, "queues out");
if (!entry) {
rte_panic("Config file parse error: core %u queues out "
"not defined\n", i);
return -1;
}
for (j = 0; (j < APP_MAX_SWQ_PER_CORE) && (entry != NULL);
j++) {
char *next;
p->swq_out[j] = (uint32_t) strtol(entry, &next, 10);
if (next == entry)
break;
entry = next;
}
if ((j != APP_MAX_SWQ_PER_CORE) || (*entry != '\0')) {
rte_panic("Config file parse error: core %u queues out "
"error\n", i);
return -1;
}
}
rte_cfgfile_close(file);
return 0;
}
void app_cores_config_print(void)
{
uint32_t i;
for (i = 0; i < RTE_MAX_LCORE; i++) {
struct app_core_params *p = &app.cores[i];
uint32_t j;
if (app.cores[i].core_type == APP_CORE_NONE)
continue;
printf("---> core %u: id = %u type = %6s [", i, p->core_id,
app_core_type_id_to_string(p->core_type));
for (j = 0; j < APP_MAX_SWQ_PER_CORE; j++)
printf("%2d ", (int) p->swq_in[j]);
printf("] [");
for (j = 0; j < APP_MAX_SWQ_PER_CORE; j++)
printf("%2d ", (int) p->swq_out[j]);
printf("]\n");
}
}
static int
app_install_port_mask(const char *arg)
{
char *end = NULL;
uint64_t port_mask;
uint32_t i;
if (arg[0] == '\0')
return -1;
port_mask = strtoul(arg, &end, 16);
if ((end == NULL) || (*end != '\0'))
return -2;
if (port_mask == 0)
return -3;
app.n_ports = 0;
for (i = 0; i < 64; i++) {
if ((port_mask & (1LLU << i)) == 0)
continue;
if (app.n_ports >= APP_MAX_PORTS)
return -4;
app.ports[app.n_ports] = i;
app.n_ports++;
}
if (!rte_is_power_of_2(app.n_ports))
return -5;
return 0;
}
int
app_parse_args(int argc, char **argv)
{
int opt, ret;
char **argvopt;
int option_index;
char *prgname = argv[0];
static struct option lgopts[] = {
{NULL, 0, 0, 0}
};
uint64_t core_mask = app_get_core_mask();
app.n_cores = __builtin_popcountll(core_mask);
argvopt = argv;
while ((opt = getopt_long(argc, argvopt, "p:f:", lgopts,
&option_index)) != EOF) {
switch (opt) {
case 'p':
if (app_install_port_mask(optarg) != 0)
rte_panic("PORT_MASK should specify a number "
"of ports that is power of 2 less or "
"equal to %u\n", APP_MAX_PORTS);
break;
case 'f':
app_install_cfgfile(optarg);
break;
default:
return -1;
}
}
app_install_coremask(core_mask);
app_cores_config_print();
if (optind >= 0)
argv[optind - 1] = prgname;
ret = optind - 1;
optind = 0; /* reset getopt lib */
return ret;
}

File diff suppressed because it is too large Load Diff

View File

@ -0,0 +1,448 @@
/*-
* BSD LICENSE
*
* Copyright(c) 2010-2015 Intel Corporation. All rights reserved.
* All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
*
* * Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* * Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in
* the documentation and/or other materials provided with the
* distribution.
* * Neither the name of Intel Corporation nor the names of its
* contributors may be used to endorse or promote products derived
* from this software without specific prior written permission.
*
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
* OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*/
#include <stdint.h>
#include <stdlib.h>
#include <stdio.h>
#include <ctype.h>
#include <getopt.h>
#include <errno.h>
#include <stdarg.h>
#include <string.h>
#include <libgen.h>
#include <unistd.h>
#include <rte_errno.h>
#include <rte_cfgfile.h>
#include <rte_string_fns.h>
#include "app.h"
static int
tm_cfgfile_load_sched_port(
struct rte_cfgfile *file,
struct rte_sched_port_params *port_params)
{
const char *entry;
int j;
entry = rte_cfgfile_get_entry(file, "port", "frame overhead");
if (entry)
port_params->frame_overhead = (uint32_t)atoi(entry);
entry = rte_cfgfile_get_entry(file, "port", "mtu");
if (entry)
port_params->mtu = (uint32_t)atoi(entry);
entry = rte_cfgfile_get_entry(file,
"port",
"number of subports per port");
if (entry)
port_params->n_subports_per_port = (uint32_t) atoi(entry);
entry = rte_cfgfile_get_entry(file,
"port",
"number of pipes per subport");
if (entry)
port_params->n_pipes_per_subport = (uint32_t) atoi(entry);
entry = rte_cfgfile_get_entry(file, "port", "queue sizes");
if (entry) {
char *next;
for (j = 0; j < RTE_SCHED_TRAFFIC_CLASSES_PER_PIPE; j++) {
port_params->qsize[j] = (uint16_t)
strtol(entry, &next, 10);
if (next == NULL)
break;
entry = next;
}
}
#ifdef RTE_SCHED_RED
for (j = 0; j < RTE_SCHED_TRAFFIC_CLASSES_PER_PIPE; j++) {
char str[32];
/* Parse WRED min thresholds */
snprintf(str, sizeof(str), "tc %" PRId32 " wred min", j);
entry = rte_cfgfile_get_entry(file, "red", str);
if (entry) {
char *next;
int k;
/* for each packet colour (green, yellow, red) */
for (k = 0; k < e_RTE_METER_COLORS; k++) {
port_params->red_params[j][k].min_th
= (uint16_t)strtol(entry, &next, 10);
if (next == NULL)
break;
entry = next;
}
}
/* Parse WRED max thresholds */
snprintf(str, sizeof(str), "tc %" PRId32 " wred max", j);
entry = rte_cfgfile_get_entry(file, "red", str);
if (entry) {
char *next;
int k;
/* for each packet colour (green, yellow, red) */
for (k = 0; k < e_RTE_METER_COLORS; k++) {
port_params->red_params[j][k].max_th
= (uint16_t)strtol(entry, &next, 10);
if (next == NULL)
break;
entry = next;
}
}
/* Parse WRED inverse mark probabilities */
snprintf(str, sizeof(str), "tc %" PRId32 " wred inv prob", j);
entry = rte_cfgfile_get_entry(file, "red", str);
if (entry) {
char *next;
int k;
/* for each packet colour (green, yellow, red) */
for (k = 0; k < e_RTE_METER_COLORS; k++) {
port_params->red_params[j][k].maxp_inv
= (uint8_t)strtol(entry, &next, 10);
if (next == NULL)
break;
entry = next;
}
}
/* Parse WRED EWMA filter weights */
snprintf(str, sizeof(str), "tc %" PRId32 " wred weight", j);
entry = rte_cfgfile_get_entry(file, "red", str);
if (entry) {
char *next;
int k;
/* for each packet colour (green, yellow, red) */
for (k = 0; k < e_RTE_METER_COLORS; k++) {
port_params->red_params[j][k].wq_log2
= (uint8_t)strtol(entry, &next, 10);
if (next == NULL)
break;
entry = next;
}
}
}
#endif /* RTE_SCHED_RED */
return 0;
}
static int
tm_cfgfile_load_sched_pipe(
struct rte_cfgfile *file,
struct rte_sched_port_params *port_params,
struct rte_sched_pipe_params *pipe_params)
{
int i, j;
char *next;
const char *entry;
int profiles;
profiles = rte_cfgfile_num_sections(file,
"pipe profile", sizeof("pipe profile") - 1);
port_params->n_pipe_profiles = profiles;
for (j = 0; j < profiles; j++) {
char pipe_name[32];
snprintf(pipe_name, sizeof(pipe_name),
"pipe profile %" PRId32, j);
entry = rte_cfgfile_get_entry(file, pipe_name, "tb rate");
if (entry)
pipe_params[j].tb_rate = (uint32_t) atoi(entry);
entry = rte_cfgfile_get_entry(file, pipe_name, "tb size");
if (entry)
pipe_params[j].tb_size = (uint32_t) atoi(entry);
entry = rte_cfgfile_get_entry(file, pipe_name, "tc period");
if (entry)
pipe_params[j].tc_period = (uint32_t) atoi(entry);
entry = rte_cfgfile_get_entry(file, pipe_name, "tc 0 rate");
if (entry)
pipe_params[j].tc_rate[0] = (uint32_t) atoi(entry);
entry = rte_cfgfile_get_entry(file, pipe_name, "tc 1 rate");
if (entry)
pipe_params[j].tc_rate[1] = (uint32_t) atoi(entry);
entry = rte_cfgfile_get_entry(file, pipe_name, "tc 2 rate");
if (entry)
pipe_params[j].tc_rate[2] = (uint32_t) atoi(entry);
entry = rte_cfgfile_get_entry(file, pipe_name, "tc 3 rate");
if (entry)
pipe_params[j].tc_rate[3] = (uint32_t) atoi(entry);
#ifdef RTE_SCHED_SUBPORT_TC_OV
entry = rte_cfgfile_get_entry(file, pipe_name,
"tc 3 oversubscription weight");
if (entry)
pipe_params[j].tc_ov_weight = (uint8_t)atoi(entry);
#endif
entry = rte_cfgfile_get_entry(file,
pipe_name,
"tc 0 wrr weights");
if (entry)
for (i = 0; i < RTE_SCHED_QUEUES_PER_TRAFFIC_CLASS; i++) {
pipe_params[j].wrr_weights[RTE_SCHED_TRAFFIC_CLASSES_PER_PIPE*0 + i] =
(uint8_t) strtol(entry, &next, 10);
if (next == NULL)
break;
entry = next;
}
entry = rte_cfgfile_get_entry(file, pipe_name, "tc 1 wrr weights");
if (entry)
for (i = 0; i < RTE_SCHED_QUEUES_PER_TRAFFIC_CLASS; i++) {
pipe_params[j].wrr_weights[RTE_SCHED_TRAFFIC_CLASSES_PER_PIPE*1 + i] =
(uint8_t) strtol(entry, &next, 10);
if (next == NULL)
break;
entry = next;
}
entry = rte_cfgfile_get_entry(file, pipe_name, "tc 2 wrr weights");
if (entry)
for (i = 0; i < RTE_SCHED_QUEUES_PER_TRAFFIC_CLASS; i++) {
pipe_params[j].wrr_weights[RTE_SCHED_TRAFFIC_CLASSES_PER_PIPE*2 + i] =
(uint8_t) strtol(entry, &next, 10);
if (next == NULL)
break;
entry = next;
}
entry = rte_cfgfile_get_entry(file, pipe_name, "tc 3 wrr weights");
if (entry)
for (i = 0; i < RTE_SCHED_QUEUES_PER_TRAFFIC_CLASS; i++) {
pipe_params[j].wrr_weights[RTE_SCHED_TRAFFIC_CLASSES_PER_PIPE*3 + i] =
(uint8_t) strtol(entry, &next, 10);
if (next == NULL)
break;
entry = next;
}
}
return 0;
}
static int
tm_cfgfile_load_sched_subport(
struct rte_cfgfile *file,
struct rte_sched_subport_params *subport_params,
int *pipe_to_profile)
{
const char *entry;
int i, j, k;
for (i = 0; i < APP_MAX_SCHED_SUBPORTS; i++) {
char sec_name[CFG_NAME_LEN];
snprintf(sec_name, sizeof(sec_name),
"subport %" PRId32, i);
if (rte_cfgfile_has_section(file, sec_name)) {
entry = rte_cfgfile_get_entry(file,
sec_name,
"tb rate");
if (entry)
subport_params[i].tb_rate =
(uint32_t) atoi(entry);
entry = rte_cfgfile_get_entry(file,
sec_name,
"tb size");
if (entry)
subport_params[i].tb_size =
(uint32_t) atoi(entry);
entry = rte_cfgfile_get_entry(file,
sec_name,
"tc period");
if (entry)
subport_params[i].tc_period =
(uint32_t) atoi(entry);
entry = rte_cfgfile_get_entry(file,
sec_name,
"tc 0 rate");
if (entry)
subport_params[i].tc_rate[0] =
(uint32_t) atoi(entry);
entry = rte_cfgfile_get_entry(file,
sec_name,
"tc 1 rate");
if (entry)
subport_params[i].tc_rate[1] =
(uint32_t) atoi(entry);
entry = rte_cfgfile_get_entry(file,
sec_name,
"tc 2 rate");
if (entry)
subport_params[i].tc_rate[2] =
(uint32_t) atoi(entry);
entry = rte_cfgfile_get_entry(file,
sec_name,
"tc 3 rate");
if (entry)
subport_params[i].tc_rate[3] =
(uint32_t) atoi(entry);
int n_entries = rte_cfgfile_section_num_entries(file,
sec_name);
struct rte_cfgfile_entry entries[n_entries];
rte_cfgfile_section_entries(file,
sec_name,
entries,
n_entries);
for (j = 0; j < n_entries; j++)
if (strncmp("pipe",
entries[j].name,
sizeof("pipe") - 1) == 0) {
int profile;
char *tokens[2] = {NULL, NULL};
int n_tokens;
int begin, end;
char name[CFG_NAME_LEN];
profile = atoi(entries[j].value);
strncpy(name,
entries[j].name,
sizeof(name));
n_tokens = rte_strsplit(
&name[sizeof("pipe")],
strnlen(name, CFG_NAME_LEN),
tokens, 2, '-');
begin = atoi(tokens[0]);
if (n_tokens == 2)
end = atoi(tokens[1]);
else
end = begin;
if ((end >= APP_MAX_SCHED_PIPES) ||
(begin > end))
return -1;
for (k = begin; k <= end; k++) {
char profile_name[CFG_NAME_LEN];
snprintf(profile_name,
sizeof(profile_name),
"pipe profile %" PRId32,
profile);
if (rte_cfgfile_has_section(file, profile_name))
pipe_to_profile[i * APP_MAX_SCHED_PIPES + k] = profile;
else
rte_exit(EXIT_FAILURE,
"Wrong pipe profile %s\n",
entries[j].value);
}
}
}
}
return 0;
}
static int
tm_cfgfile_load(struct app_pktq_tm_params *tm)
{
struct rte_cfgfile *file;
uint32_t i;
memset(tm->sched_subport_params, 0, sizeof(tm->sched_subport_params));
memset(tm->sched_pipe_profiles, 0, sizeof(tm->sched_pipe_profiles));
memset(&tm->sched_port_params, 0, sizeof(tm->sched_pipe_profiles));
for (i = 0; i < APP_MAX_SCHED_SUBPORTS * APP_MAX_SCHED_PIPES; i++)
tm->sched_pipe_to_profile[i] = -1;
tm->sched_port_params.pipe_profiles = &tm->sched_pipe_profiles[0];
if (tm->file_name[0] == '\0')
return -1;
file = rte_cfgfile_load(tm->file_name, 0);
if (file == NULL)
return -1;
tm_cfgfile_load_sched_port(file,
&tm->sched_port_params);
tm_cfgfile_load_sched_subport(file,
tm->sched_subport_params,
tm->sched_pipe_to_profile);
tm_cfgfile_load_sched_pipe(file,
&tm->sched_port_params,
tm->sched_pipe_profiles);
rte_cfgfile_close(file);
return 0;
}
int
app_config_parse_tm(struct app_params *app)
{
uint32_t i;
for (i = 0; i < RTE_DIM(app->tm_params); i++) {
struct app_pktq_tm_params *p = &app->tm_params[i];
int status;
if (!APP_PARAM_VALID(p))
break;
status = tm_cfgfile_load(p);
APP_CHECK(status == 0,
"Parse error for %s configuration file \"%s\"\n",
p->name,
p->file_name);
}
return 0;
}

View File

@ -0,0 +1,493 @@
/*-
* BSD LICENSE
*
* Copyright(c) 2010-2015 Intel Corporation. All rights reserved.
* All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
*
* * Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* * Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in
* the documentation and/or other materials provided with the
* distribution.
* * Neither the name of Intel Corporation nor the names of its
* contributors may be used to endorse or promote products derived
* from this software without specific prior written permission.
*
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
* OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*/
#include <inttypes.h>
#include <stdlib.h>
#include <stdio.h>
#include <string.h>
#include <rte_lcore.h>
#include "cpu_core_map.h"
struct cpu_core_map {
uint32_t n_max_sockets;
uint32_t n_max_cores_per_socket;
uint32_t n_max_ht_per_core;
uint32_t n_sockets;
uint32_t n_cores_per_socket;
uint32_t n_ht_per_core;
int map[0];
};
static inline uint32_t
cpu_core_map_pos(struct cpu_core_map *map,
uint32_t socket_id,
uint32_t core_id,
uint32_t ht_id)
{
return (socket_id * map->n_max_cores_per_socket + core_id) *
map->n_max_ht_per_core + ht_id;
}
static int
cpu_core_map_compute_eal(struct cpu_core_map *map);
static int
cpu_core_map_compute_linux(struct cpu_core_map *map);
static int
cpu_core_map_compute_and_check(struct cpu_core_map *map);
struct cpu_core_map *
cpu_core_map_init(uint32_t n_max_sockets,
uint32_t n_max_cores_per_socket,
uint32_t n_max_ht_per_core,
uint32_t eal_initialized)
{
uint32_t map_size, map_mem_size, i;
struct cpu_core_map *map;
int status;
/* Check input arguments */
if ((n_max_sockets == 0) ||
(n_max_cores_per_socket == 0) ||
(n_max_ht_per_core == 0))
return NULL;
/* Memory allocation */
map_size = n_max_sockets * n_max_cores_per_socket * n_max_ht_per_core;
map_mem_size = sizeof(struct cpu_core_map) + map_size * sizeof(int);
map = (struct cpu_core_map *) malloc(map_mem_size);
if (map == NULL)
return NULL;
/* Initialization */
map->n_max_sockets = n_max_sockets;
map->n_max_cores_per_socket = n_max_cores_per_socket;
map->n_max_ht_per_core = n_max_ht_per_core;
map->n_sockets = 0;
map->n_cores_per_socket = 0;
map->n_ht_per_core = 0;
for (i = 0; i < map_size; i++)
map->map[i] = -1;
status = (eal_initialized) ?
cpu_core_map_compute_eal(map) :
cpu_core_map_compute_linux(map);
if (status) {
free(map);
return NULL;
}
status = cpu_core_map_compute_and_check(map);
if (status) {
free(map);
return NULL;
}
return map;
}
int
cpu_core_map_compute_eal(struct cpu_core_map *map)
{
uint32_t socket_id, core_id, ht_id;
/* Compute map */
for (socket_id = 0; socket_id < map->n_max_sockets; socket_id++) {
uint32_t n_detected, core_id_contig;
int lcore_id;
n_detected = 0;
for (lcore_id = 0; lcore_id < RTE_MAX_LCORE; lcore_id++) {
struct lcore_config *p = &lcore_config[lcore_id];
if ((p->detected) && (p->socket_id == socket_id))
n_detected++;
}
core_id_contig = 0;
for (core_id = 0; n_detected ; core_id++) {
ht_id = 0;
for (lcore_id = 0;
lcore_id < RTE_MAX_LCORE;
lcore_id++) {
struct lcore_config *p =
&lcore_config[lcore_id];
if ((p->detected) &&
(p->socket_id == socket_id) &&
(p->core_id == core_id)) {
uint32_t pos = cpu_core_map_pos(map,
socket_id,
core_id_contig,
ht_id);
map->map[pos] = lcore_id;
ht_id++;
n_detected--;
}
}
if (ht_id) {
core_id_contig++;
if (core_id_contig ==
map->n_max_cores_per_socket)
return -1;
}
}
}
return 0;
}
int
cpu_core_map_compute_and_check(struct cpu_core_map *map)
{
uint32_t socket_id, core_id, ht_id;
/* Compute n_ht_per_core, n_cores_per_socket, n_sockets */
for (ht_id = 0; ht_id < map->n_max_ht_per_core; ht_id++) {
if (map->map[ht_id] == -1)
break;
map->n_ht_per_core++;
}
if (map->n_ht_per_core == 0)
return -1;
for (core_id = 0; core_id < map->n_max_cores_per_socket; core_id++) {
uint32_t pos = core_id * map->n_max_ht_per_core;
if (map->map[pos] == -1)
break;
map->n_cores_per_socket++;
}
if (map->n_cores_per_socket == 0)
return -1;
for (socket_id = 0; socket_id < map->n_max_sockets; socket_id++) {
uint32_t pos = socket_id * map->n_max_cores_per_socket *
map->n_max_ht_per_core;
if (map->map[pos] == -1)
break;
map->n_sockets++;
}
if (map->n_sockets == 0)
return -1;
/* Check that each socket has exactly the same number of cores
and that each core has exactly the same number of hyper-threads */
for (socket_id = 0; socket_id < map->n_sockets; socket_id++) {
for (core_id = 0; core_id < map->n_cores_per_socket; core_id++)
for (ht_id = 0;
ht_id < map->n_max_ht_per_core;
ht_id++) {
uint32_t pos = (socket_id *
map->n_max_cores_per_socket + core_id) *
map->n_max_ht_per_core + ht_id;
if (((ht_id < map->n_ht_per_core) &&
(map->map[pos] == -1)) ||
((ht_id >= map->n_ht_per_core) &&
(map->map[pos] != -1)))
return -1;
}
for ( ; core_id < map->n_max_cores_per_socket; core_id++)
for (ht_id = 0;
ht_id < map->n_max_ht_per_core;
ht_id++) {
uint32_t pos = cpu_core_map_pos(map,
socket_id,
core_id,
ht_id);
if (map->map[pos] != -1)
return -1;
}
}
return 0;
}
#define FILE_LINUX_CPU_N_LCORES \
"/sys/devices/system/cpu/present"
static int
cpu_core_map_get_n_lcores_linux(void)
{
char buffer[64], *string;
FILE *fd;
fd = fopen(FILE_LINUX_CPU_N_LCORES, "r");
if (fd == NULL)
return -1;
if (fgets(buffer, sizeof(buffer), fd) == NULL) {
fclose(fd);
return -1;
}
fclose(fd);
string = index(buffer, '-');
if (string == NULL)
return -1;
return (atoi(++string) + 1);
}
#define FILE_LINUX_CPU_CORE_ID \
"/sys/devices/system/cpu/cpu%" PRIu32 "/topology/core_id"
static int
cpu_core_map_get_core_id_linux(int lcore_id)
{
char buffer[64];
FILE *fd;
int core_id;
snprintf(buffer, sizeof(buffer), FILE_LINUX_CPU_CORE_ID, lcore_id);
fd = fopen(buffer, "r");
if (fd == NULL)
return -1;
if (fgets(buffer, sizeof(buffer), fd) == NULL) {
fclose(fd);
return -1;
}
fclose(fd);
core_id = atoi(buffer);
return core_id;
}
#define FILE_LINUX_CPU_SOCKET_ID \
"/sys/devices/system/cpu/cpu%" PRIu32 "/topology/physical_package_id"
static int
cpu_core_map_get_socket_id_linux(int lcore_id)
{
char buffer[64];
FILE *fd;
int socket_id;
snprintf(buffer, sizeof(buffer), FILE_LINUX_CPU_SOCKET_ID, lcore_id);
fd = fopen(buffer, "r");
if (fd == NULL)
return -1;
if (fgets(buffer, sizeof(buffer), fd) == NULL) {
fclose(fd);
return -1;
}
fclose(fd);
socket_id = atoi(buffer);
return socket_id;
}
int
cpu_core_map_compute_linux(struct cpu_core_map *map)
{
uint32_t socket_id, core_id, ht_id;
int n_lcores;
n_lcores = cpu_core_map_get_n_lcores_linux();
if (n_lcores <= 0)
return -1;
/* Compute map */
for (socket_id = 0; socket_id < map->n_max_sockets; socket_id++) {
uint32_t n_detected, core_id_contig;
int lcore_id;
n_detected = 0;
for (lcore_id = 0; lcore_id < n_lcores; lcore_id++) {
int lcore_socket_id =
cpu_core_map_get_socket_id_linux(lcore_id);
if (lcore_socket_id < 0)
return -1;
if (((uint32_t) lcore_socket_id) == socket_id)
n_detected++;
}
core_id_contig = 0;
for (core_id = 0; n_detected ; core_id++) {
ht_id = 0;
for (lcore_id = 0; lcore_id < n_lcores; lcore_id++) {
int lcore_socket_id =
cpu_core_map_get_socket_id_linux(
lcore_id);
if (lcore_socket_id < 0)
return -1;
int lcore_core_id =
cpu_core_map_get_core_id_linux(
lcore_id);
if (lcore_core_id < 0)
return -1;
if (((uint32_t) lcore_socket_id == socket_id) &&
((uint32_t) lcore_core_id == core_id)) {
uint32_t pos = cpu_core_map_pos(map,
socket_id,
core_id_contig,
ht_id);
map->map[pos] = lcore_id;
ht_id++;
n_detected--;
}
}
if (ht_id) {
core_id_contig++;
if (core_id_contig ==
map->n_max_cores_per_socket)
return -1;
}
}
}
return 0;
}
void
cpu_core_map_print(struct cpu_core_map *map)
{
uint32_t socket_id, core_id, ht_id;
if (map == NULL)
return;
for (socket_id = 0; socket_id < map->n_sockets; socket_id++) {
printf("Socket %" PRIu32 ":\n", socket_id);
for (core_id = 0;
core_id < map->n_cores_per_socket;
core_id++) {
printf("[%" PRIu32 "] = [", core_id);
for (ht_id = 0; ht_id < map->n_ht_per_core; ht_id++) {
int lcore_id = cpu_core_map_get_lcore_id(map,
socket_id,
core_id,
ht_id);
uint32_t core_id_noncontig =
cpu_core_map_get_core_id_linux(
lcore_id);
printf(" %" PRId32 " (%" PRIu32 ") ",
lcore_id,
core_id_noncontig);
}
printf("]\n");
}
}
}
uint32_t
cpu_core_map_get_n_sockets(struct cpu_core_map *map)
{
if (map == NULL)
return 0;
return map->n_sockets;
}
uint32_t
cpu_core_map_get_n_cores_per_socket(struct cpu_core_map *map)
{
if (map == NULL)
return 0;
return map->n_cores_per_socket;
}
uint32_t
cpu_core_map_get_n_ht_per_core(struct cpu_core_map *map)
{
if (map == NULL)
return 0;
return map->n_ht_per_core;
}
int
cpu_core_map_get_lcore_id(struct cpu_core_map *map,
uint32_t socket_id,
uint32_t core_id,
uint32_t ht_id)
{
uint32_t pos;
if ((map == NULL) ||
(socket_id >= map->n_sockets) ||
(core_id >= map->n_cores_per_socket) ||
(ht_id >= map->n_ht_per_core))
return -1;
pos = cpu_core_map_pos(map, socket_id, core_id, ht_id);
return map->map[pos];
}
void
cpu_core_map_free(struct cpu_core_map *map)
{
if (map)
free(map);
}

View File

@ -0,0 +1,69 @@
/*-
* BSD LICENSE
*
* Copyright(c) 2010-2015 Intel Corporation. All rights reserved.
* All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
*
* * Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* * Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in
* the documentation and/or other materials provided with the
* distribution.
* * Neither the name of Intel Corporation nor the names of its
* contributors may be used to endorse or promote products derived
* from this software without specific prior written permission.
*
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
* OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*/
#ifndef __INCLUDE_CPU_CORE_MAP_H__
#define __INCLUDE_CPU_CORE_MAP_H__
#include <stdio.h>
#include <rte_lcore.h>
struct cpu_core_map;
struct cpu_core_map *
cpu_core_map_init(uint32_t n_max_sockets,
uint32_t n_max_cores_per_socket,
uint32_t n_max_ht_per_core,
uint32_t eal_initialized);
uint32_t
cpu_core_map_get_n_sockets(struct cpu_core_map *map);
uint32_t
cpu_core_map_get_n_cores_per_socket(struct cpu_core_map *map);
uint32_t
cpu_core_map_get_n_ht_per_core(struct cpu_core_map *map);
int
cpu_core_map_get_lcore_id(struct cpu_core_map *map,
uint32_t socket_id,
uint32_t core_id,
uint32_t ht_id);
void cpu_core_map_print(struct cpu_core_map *map);
void
cpu_core_map_free(struct cpu_core_map *map);
#endif

View File

@ -1,7 +1,7 @@
/*-
* BSD LICENSE
*
* Copyright(c) 2010-2014 Intel Corporation. All rights reserved.
* Copyright(c) 2010-2015 Intel Corporation. All rights reserved.
* All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
@ -31,135 +31,21 @@
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*/
#include <stdio.h>
#include <stdlib.h>
#include <stdint.h>
#include <inttypes.h>
#include <sys/types.h>
#include <string.h>
#include <sys/queue.h>
#include <stdarg.h>
#include <errno.h>
#include <getopt.h>
#include <unistd.h>
#include "app.h"
#include <rte_common.h>
#include <rte_byteorder.h>
#include <rte_log.h>
#include <rte_memory.h>
#include <rte_memcpy.h>
#include <rte_memzone.h>
#include <rte_eal.h>
#include <rte_per_lcore.h>
#include <rte_launch.h>
#include <rte_atomic.h>
#include <rte_cycles.h>
#include <rte_prefetch.h>
#include <rte_lcore.h>
#include <rte_per_lcore.h>
#include <rte_branch_prediction.h>
#include <rte_interrupts.h>
#include <rte_pci.h>
#include <rte_random.h>
#include <rte_debug.h>
#include <rte_ether.h>
#include <rte_ethdev.h>
#include <rte_ring.h>
#include <rte_mempool.h>
#include <rte_mbuf.h>
#include <rte_ip.h>
#include <rte_tcp.h>
#include <rte_lpm.h>
#include <rte_lpm6.h>
#include "main.h"
static struct app_params app;
int
main(int argc, char **argv)
{
int ret;
rte_openlog_stream(stderr);
/* Init EAL */
ret = rte_eal_init(argc, argv);
if (ret < 0)
return -1;
argc -= ret;
argv += ret;
/* Config */
app_config_init(&app);
/* Parse application arguments (after the EAL ones) */
ret = app_parse_args(argc, argv);
if (ret < 0) {
app_print_usage(argv[0]);
return -1;
}
app_config_args(&app, argc, argv);
/* Init */
app_init();
/* Launch per-lcore init on every lcore */
rte_eal_mp_remote_launch(app_lcore_main_loop, NULL, CALL_MASTER);
app_config_parse(&app, app.config_file);
return 0;
}
int
app_lcore_main_loop(__attribute__((unused)) void *arg)
{
uint32_t core_id, i;
core_id = rte_lcore_id();
for (i = 0; i < app.n_cores; i++) {
struct app_core_params *p = &app.cores[i];
if (p->core_id != core_id)
continue;
switch (p->core_type) {
case APP_CORE_MASTER:
app_ping();
app_main_loop_cmdline();
return 0;
case APP_CORE_RX:
app_main_loop_pipeline_rx();
/* app_main_loop_rx(); */
return 0;
case APP_CORE_TX:
app_main_loop_pipeline_tx();
/* app_main_loop_tx(); */
return 0;
case APP_CORE_PT:
/* app_main_loop_pipeline_passthrough(); */
app_main_loop_passthrough();
return 0;
case APP_CORE_FC:
app_main_loop_pipeline_flow_classification();
return 0;
case APP_CORE_FW:
case APP_CORE_RT:
app_main_loop_pipeline_routing();
return 0;
#ifdef RTE_LIBRTE_ACL
app_main_loop_pipeline_firewall();
return 0;
#else
rte_exit(EXIT_FAILURE, "ACL not present in build\n");
#endif
case APP_CORE_IPV4_FRAG:
app_main_loop_pipeline_ipv4_frag();
return 0;
case APP_CORE_IPV4_RAS:
app_main_loop_pipeline_ipv4_ras();
return 0;
default:
rte_panic("%s: Invalid core type for core %u\n",
__func__, i);
}
}
rte_panic("%s: Algorithmic error\n", __func__);
return -1;
}

View File

@ -1,298 +0,0 @@
/*-
* BSD LICENSE
*
* Copyright(c) 2010-2014 Intel Corporation. All rights reserved.
* All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
*
* * Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* * Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in
* the documentation and/or other materials provided with the
* distribution.
* * Neither the name of Intel Corporation nor the names of its
* contributors may be used to endorse or promote products derived
* from this software without specific prior written permission.
*
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
* OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*/
#ifndef _MAIN_H_
#define _MAIN_H_
#include <stdint.h>
#include <rte_mempool.h>
#include <rte_mbuf.h>
#include <rte_ring.h>
#include <rte_ethdev.h>
#ifdef RTE_LIBRTE_ACL
#include <rte_table_acl.h>
#endif
struct app_flow_key {
union {
struct {
uint8_t ttl; /* needs to be set to 0 */
uint8_t proto;
uint16_t header_checksum; /* needs to be set to 0 */
uint32_t ip_src;
};
uint64_t slab0;
};
union {
struct {
uint32_t ip_dst;
uint16_t port_src;
uint16_t port_dst;
};
uint64_t slab1;
};
} __attribute__((__packed__));
struct app_arp_key {
uint32_t nh_ip;
uint32_t nh_iface;
} __attribute__((__packed__));
struct app_pkt_metadata {
uint32_t signature;
uint8_t reserved1[28];
struct app_flow_key flow_key;
struct app_arp_key arp_key;
struct ether_addr nh_arp;
uint8_t reserved3[2];
} __attribute__((__packed__));
#ifndef APP_MBUF_ARRAY_SIZE
#define APP_MBUF_ARRAY_SIZE 256
#endif
struct app_mbuf_array {
struct rte_mbuf *array[APP_MBUF_ARRAY_SIZE];
uint32_t n_mbufs;
};
#ifndef APP_MAX_PORTS
#define APP_MAX_PORTS 4
#endif
#ifndef APP_MAX_SWQ_PER_CORE
#define APP_MAX_SWQ_PER_CORE 8
#endif
#define APP_SWQ_INVALID ((uint32_t)(-1))
#define APP_SWQ_IN_REQ (APP_MAX_SWQ_PER_CORE - 1)
#define APP_SWQ_OUT_RESP (APP_MAX_SWQ_PER_CORE - 1)
enum app_core_type {
APP_CORE_NONE = 0, /* Unused */
APP_CORE_MASTER, /* Management */
APP_CORE_RX, /* Reception */
APP_CORE_TX, /* Transmission */
APP_CORE_PT, /* Pass-through */
APP_CORE_FC, /* Flow Classification */
APP_CORE_FW, /* Firewall */
APP_CORE_RT, /* Routing */
APP_CORE_TM, /* Traffic Management */
APP_CORE_IPV4_FRAG,/* IPv4 Fragmentation */
APP_CORE_IPV4_RAS, /* IPv4 Reassembly */
};
struct app_core_params {
uint32_t core_id;
enum app_core_type core_type;
/* SWQ map */
uint32_t swq_in[APP_MAX_SWQ_PER_CORE];
uint32_t swq_out[APP_MAX_SWQ_PER_CORE];
} __rte_cache_aligned;
struct app_params {
/* CPU cores */
struct app_core_params cores[RTE_MAX_LCORE];
uint32_t n_cores;
/* Ports*/
uint32_t ports[APP_MAX_PORTS];
uint32_t n_ports;
uint32_t rsz_hwq_rx;
uint32_t rsz_hwq_tx;
uint32_t bsz_hwq_rd;
uint32_t bsz_hwq_wr;
struct rte_eth_conf port_conf;
struct rte_eth_rxconf rx_conf;
struct rte_eth_txconf tx_conf;
/* SW Queues (SWQs) */
struct rte_ring **rings;
uint32_t rsz_swq;
uint32_t bsz_swq_rd;
uint32_t bsz_swq_wr;
/* Buffer pool */
struct rte_mempool *pool;
struct rte_mempool *indirect_pool;
uint32_t pool_buffer_size;
uint32_t pool_size;
uint32_t pool_cache_size;
/* Message buffer pool */
struct rte_mempool *msg_pool;
uint32_t msg_pool_buffer_size;
uint32_t msg_pool_size;
uint32_t msg_pool_cache_size;
/* Rule tables */
uint32_t max_arp_rules;
uint32_t max_routing_rules;
uint32_t max_firewall_rules;
uint32_t max_flow_rules;
/* Processing */
uint32_t ether_hdr_pop_push;
} __rte_cache_aligned;
extern struct app_params app;
const char *app_core_type_id_to_string(enum app_core_type id);
int app_core_type_string_to_id(const char *string, enum app_core_type *id);
void app_cores_config_print(void);
void app_check_core_params(void);
struct app_core_params *app_get_core_params(uint32_t core_id);
uint32_t app_get_first_core_id(enum app_core_type core_type);
struct rte_ring *app_get_ring_req(uint32_t core_id);
struct rte_ring *app_get_ring_resp(uint32_t core_id);
int app_parse_args(int argc, char **argv);
void app_print_usage(char *prgname);
void app_init(void);
void app_ping(void);
int app_lcore_main_loop(void *arg);
/* Hash functions */
uint64_t test_hash(void *key, uint32_t key_size, uint64_t seed);
uint32_t rte_jhash2_16(uint32_t *k, uint32_t initval);
#if defined(__x86_64__)
uint32_t rte_aeshash_16(uint64_t *k, uint64_t seed);
uint32_t rte_crchash_16(uint64_t *k, uint64_t seed);
#endif
/* I/O with no pipeline */
void app_main_loop_rx(void);
void app_main_loop_tx(void);
void app_main_loop_passthrough(void);
/* Pipeline */
void app_main_loop_pipeline_rx(void);
void app_main_loop_pipeline_rx_frag(void);
void app_main_loop_pipeline_tx(void);
void app_main_loop_pipeline_tx_ras(void);
void app_main_loop_pipeline_flow_classification(void);
void app_main_loop_pipeline_firewall(void);
void app_main_loop_pipeline_routing(void);
void app_main_loop_pipeline_passthrough(void);
void app_main_loop_pipeline_ipv4_frag(void);
void app_main_loop_pipeline_ipv4_ras(void);
/* Command Line Interface (CLI) */
void app_main_loop_cmdline(void);
/* Messages */
enum app_msg_req_type {
APP_MSG_REQ_PING,
APP_MSG_REQ_FC_ADD,
APP_MSG_REQ_FC_DEL,
APP_MSG_REQ_FC_ADD_ALL,
APP_MSG_REQ_FW_ADD,
APP_MSG_REQ_FW_DEL,
APP_MSG_REQ_RT_ADD,
APP_MSG_REQ_RT_DEL,
APP_MSG_REQ_ARP_ADD,
APP_MSG_REQ_ARP_DEL,
APP_MSG_REQ_RX_PORT_ENABLE,
APP_MSG_REQ_RX_PORT_DISABLE,
};
struct app_msg_req {
enum app_msg_req_type type;
union {
struct {
uint32_t ip;
uint8_t depth;
uint8_t port;
uint32_t nh_ip;
} routing_add;
struct {
uint32_t ip;
uint8_t depth;
} routing_del;
struct {
uint8_t out_iface;
uint32_t nh_ip;
struct ether_addr nh_arp;
} arp_add;
struct {
uint8_t out_iface;
uint32_t nh_ip;
} arp_del;
struct {
union {
uint8_t key_raw[16];
struct app_flow_key key;
};
uint8_t port;
} flow_classif_add;
struct {
union {
uint8_t key_raw[16];
struct app_flow_key key;
};
} flow_classif_del;
#ifdef RTE_LIBRTE_ACL
struct {
struct rte_table_acl_rule_add_params add_params;
uint8_t port;
} firewall_add;
struct {
struct rte_table_acl_rule_delete_params delete_params;
} firewall_del;
#endif
struct {
uint8_t port;
} rx_up;
struct {
uint8_t port;
} rx_down;
};
};
struct app_msg_resp {
int result;
};
#define APP_FLUSH 0xFF
#endif /* _MAIN_H_ */

View File

@ -0,0 +1,87 @@
/*-
* BSD LICENSE
*
* Copyright(c) 2010-2015 Intel Corporation. All rights reserved.
* All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
*
* * Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* * Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in
* the documentation and/or other materials provided with the
* distribution.
* * Neither the name of Intel Corporation nor the names of its
* contributors may be used to endorse or promote products derived
* from this software without specific prior written permission.
*
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
* OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*/
#ifndef __INCLUDE_PIPELINE_H__
#define __INCLUDE_PIPELINE_H__
#include <cmdline_parse.h>
#include "pipeline_be.h"
/*
* Pipeline type front-end operations
*/
typedef void* (*pipeline_fe_op_init)(struct pipeline_params *params, void *arg);
typedef int (*pipeline_fe_op_free)(void *pipeline);
struct pipeline_fe_ops {
pipeline_fe_op_init f_init;
pipeline_fe_op_free f_free;
cmdline_parse_ctx_t *cmds;
};
/*
* Pipeline type
*/
struct pipeline_type {
const char *name;
/* pipeline back-end */
struct pipeline_be_ops *be_ops;
/* pipeline front-end */
struct pipeline_fe_ops *fe_ops;
};
static inline uint32_t
pipeline_type_cmds_count(struct pipeline_type *ptype)
{
cmdline_parse_ctx_t *cmds;
uint32_t n_cmds;
if (ptype->fe_ops == NULL)
return 0;
cmds = ptype->fe_ops->cmds;
if (cmds == NULL)
return 0;
for (n_cmds = 0; cmds[n_cmds]; n_cmds++);
return n_cmds;
}
#endif

View File

@ -0,0 +1,256 @@
/*-
* BSD LICENSE
*
* Copyright(c) 2010-2015 Intel Corporation. All rights reserved.
* All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
*
* * Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* * Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in
* the documentation and/or other materials provided with the
* distribution.
* * Neither the name of Intel Corporation nor the names of its
* contributors may be used to endorse or promote products derived
* from this software without specific prior written permission.
*
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
* OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*/
#ifndef __INCLUDE_PIPELINE_BE_H__
#define __INCLUDE_PIPELINE_BE_H__
#include <rte_port_ethdev.h>
#include <rte_port_ring.h>
#include <rte_port_frag.h>
#include <rte_port_ras.h>
#include <rte_port_sched.h>
#include <rte_port_source_sink.h>
#include <rte_pipeline.h>
enum pipeline_port_in_type {
PIPELINE_PORT_IN_ETHDEV_READER,
PIPELINE_PORT_IN_RING_READER,
PIPELINE_PORT_IN_RING_READER_IPV4_FRAG,
PIPELINE_PORT_IN_RING_READER_IPV6_FRAG,
PIPELINE_PORT_IN_SCHED_READER,
PIPELINE_PORT_IN_SOURCE,
};
struct pipeline_port_in_params {
enum pipeline_port_in_type type;
union {
struct rte_port_ethdev_reader_params ethdev;
struct rte_port_ring_reader_params ring;
struct rte_port_ring_reader_ipv4_frag_params ring_ipv4_frag;
struct rte_port_ring_reader_ipv6_frag_params ring_ipv6_frag;
struct rte_port_sched_reader_params sched;
struct rte_port_source_params source;
} params;
uint32_t burst_size;
};
static inline void *
pipeline_port_in_params_convert(struct pipeline_port_in_params *p)
{
switch (p->type) {
case PIPELINE_PORT_IN_ETHDEV_READER:
return (void *) &p->params.ethdev;
case PIPELINE_PORT_IN_RING_READER:
return (void *) &p->params.ring;
case PIPELINE_PORT_IN_RING_READER_IPV4_FRAG:
return (void *) &p->params.ring_ipv4_frag;
case PIPELINE_PORT_IN_RING_READER_IPV6_FRAG:
return (void *) &p->params.ring_ipv6_frag;
case PIPELINE_PORT_IN_SCHED_READER:
return (void *) &p->params.sched;
case PIPELINE_PORT_IN_SOURCE:
return (void *) &p->params.source;
default:
return NULL;
}
}
static inline struct rte_port_in_ops *
pipeline_port_in_params_get_ops(struct pipeline_port_in_params *p)
{
switch (p->type) {
case PIPELINE_PORT_IN_ETHDEV_READER:
return &rte_port_ethdev_reader_ops;
case PIPELINE_PORT_IN_RING_READER:
return &rte_port_ring_reader_ops;
case PIPELINE_PORT_IN_RING_READER_IPV4_FRAG:
return &rte_port_ring_reader_ipv4_frag_ops;
case PIPELINE_PORT_IN_RING_READER_IPV6_FRAG:
return &rte_port_ring_reader_ipv6_frag_ops;
case PIPELINE_PORT_IN_SCHED_READER:
return &rte_port_sched_reader_ops;
case PIPELINE_PORT_IN_SOURCE:
return &rte_port_source_ops;
default:
return NULL;
}
}
enum pipeline_port_out_type {
PIPELINE_PORT_OUT_ETHDEV_WRITER,
PIPELINE_PORT_OUT_ETHDEV_WRITER_NODROP,
PIPELINE_PORT_OUT_RING_WRITER,
PIPELINE_PORT_OUT_RING_WRITER_NODROP,
PIPELINE_PORT_OUT_RING_WRITER_IPV4_RAS,
PIPELINE_PORT_OUT_RING_WRITER_IPV6_RAS,
PIPELINE_PORT_OUT_SCHED_WRITER,
PIPELINE_PORT_OUT_SINK,
};
struct pipeline_port_out_params {
enum pipeline_port_out_type type;
union {
struct rte_port_ethdev_writer_params ethdev;
struct rte_port_ethdev_writer_nodrop_params ethdev_nodrop;
struct rte_port_ring_writer_params ring;
struct rte_port_ring_writer_nodrop_params ring_nodrop;
struct rte_port_ring_writer_ipv4_ras_params ring_ipv4_ras;
struct rte_port_ring_writer_ipv6_ras_params ring_ipv6_ras;
struct rte_port_sched_writer_params sched;
} params;
};
static inline void *
pipeline_port_out_params_convert(struct pipeline_port_out_params *p)
{
switch (p->type) {
case PIPELINE_PORT_OUT_ETHDEV_WRITER:
return (void *) &p->params.ethdev;
case PIPELINE_PORT_OUT_ETHDEV_WRITER_NODROP:
return (void *) &p->params.ethdev_nodrop;
case PIPELINE_PORT_OUT_RING_WRITER:
return (void *) &p->params.ring;
case PIPELINE_PORT_OUT_RING_WRITER_NODROP:
return (void *) &p->params.ring_nodrop;
case PIPELINE_PORT_OUT_RING_WRITER_IPV4_RAS:
return (void *) &p->params.ring_ipv4_ras;
case PIPELINE_PORT_OUT_RING_WRITER_IPV6_RAS:
return (void *) &p->params.ring_ipv6_ras;
case PIPELINE_PORT_OUT_SCHED_WRITER:
return (void *) &p->params.sched;
case PIPELINE_PORT_OUT_SINK:
default:
return NULL;
}
}
static inline void *
pipeline_port_out_params_get_ops(struct pipeline_port_out_params *p)
{
switch (p->type) {
case PIPELINE_PORT_OUT_ETHDEV_WRITER:
return &rte_port_ethdev_writer_ops;
case PIPELINE_PORT_OUT_ETHDEV_WRITER_NODROP:
return &rte_port_ethdev_writer_nodrop_ops;
case PIPELINE_PORT_OUT_RING_WRITER:
return &rte_port_ring_writer_ops;
case PIPELINE_PORT_OUT_RING_WRITER_NODROP:
return &rte_port_ring_writer_nodrop_ops;
case PIPELINE_PORT_OUT_RING_WRITER_IPV4_RAS:
return &rte_port_ring_writer_ipv4_ras_ops;
case PIPELINE_PORT_OUT_RING_WRITER_IPV6_RAS:
return &rte_port_ring_writer_ipv6_ras_ops;
case PIPELINE_PORT_OUT_SCHED_WRITER:
return &rte_port_sched_writer_ops;
case PIPELINE_PORT_OUT_SINK:
return &rte_port_sink_ops;
default:
return NULL;
}
}
#ifndef PIPELINE_NAME_SIZE
#define PIPELINE_NAME_SIZE 32
#endif
#ifndef PIPELINE_MAX_PORT_IN
#define PIPELINE_MAX_PORT_IN 16
#endif
#ifndef PIPELINE_MAX_PORT_OUT
#define PIPELINE_MAX_PORT_OUT 16
#endif
#ifndef PIPELINE_MAX_TABLES
#define PIPELINE_MAX_TABLES 16
#endif
#ifndef PIPELINE_MAX_MSGQ_IN
#define PIPELINE_MAX_MSGQ_IN 16
#endif
#ifndef PIPELINE_MAX_MSGQ_OUT
#define PIPELINE_MAX_MSGQ_OUT 16
#endif
#ifndef PIPELINE_MAX_ARGS
#define PIPELINE_MAX_ARGS 32
#endif
struct pipeline_params {
char name[PIPELINE_NAME_SIZE];
struct pipeline_port_in_params port_in[PIPELINE_MAX_PORT_IN];
struct pipeline_port_out_params port_out[PIPELINE_MAX_PORT_OUT];
struct rte_ring *msgq_in[PIPELINE_MAX_MSGQ_IN];
struct rte_ring *msgq_out[PIPELINE_MAX_MSGQ_OUT];
uint32_t n_ports_in;
uint32_t n_ports_out;
uint32_t n_msgq;
int socket_id;
char *args_name[PIPELINE_MAX_ARGS];
char *args_value[PIPELINE_MAX_ARGS];
uint32_t n_args;
uint32_t log_level;
};
/*
* Pipeline type back-end operations
*/
typedef void* (*pipeline_be_op_init)(struct pipeline_params *params,
void *arg);
typedef int (*pipeline_be_op_free)(void *pipeline);
typedef int (*pipeline_be_op_run)(void *pipeline);
typedef int (*pipeline_be_op_timer)(void *pipeline);
typedef int (*pipeline_be_op_track)(void *pipeline,
uint32_t port_in,
uint32_t *port_out);
struct pipeline_be_ops {
pipeline_be_op_init f_init;
pipeline_be_op_free f_free;
pipeline_be_op_run f_run;
pipeline_be_op_timer f_timer;
pipeline_be_op_track f_track;
};
#endif