numam-dpdk/drivers/event/opdl/opdl_evdev.c
Timothy McDaniel 75d113136f eventdev: express DLB/DLB2 PMD constraints
This commit implements the eventdev ABI changes required by
the DLB/DLB2 PMDs.  Several data structures and constants are modified
or added in this patch, thereby requiring modifications to the
dependent apps and examples.

The DLB/DLB2 hardware does not conform exactly to the eventdev interface.
1) It has a limit on the number of queues that may be linked to a port.
2) Some ports a further restricted to a maximum of 1 linked queue.
3) DLB does not have the ability to carry the flow_id as part
   of the event (QE) payload. Note that the DLB2 hardware is capable of
   carrying the flow_id.

Following is a detailed description of the changes that have been made.

1) Add new fields to the rte_event_dev_info struct. These fields allow
the device to advertise its capabilities so that applications can take
the appropriate actions based on those capabilities.

    struct rte_event_dev_info {
	uint32_t max_event_port_links;
	/**< Maximum number of queues that can be linked to a single event
	 * port by this device.
	 */

	uint8_t max_single_link_event_port_queue_pairs;
	/**< Maximum number of event ports and queues that are optimized for
	 * (and only capable of) single-link configurations supported by this
	 * device. These ports and queues are not accounted for in
	 * max_event_ports or max_event_queues.
	 */
    }

2) Add a new field to the rte_event_dev_config struct. This field allows
the application to specify how many of its ports are limited to a single
link, or will be used in single link mode.

    /** Event device configuration structure */
    struct rte_event_dev_config {
	uint8_t nb_single_link_event_port_queues;
	/**< Number of event ports and queues that will be singly-linked to
	 * each other. These are a subset of the overall event ports and
	 * queues; this value cannot exceed *nb_event_ports* or
	 * *nb_event_queues*. If the device has ports and queues that are
	 * optimized for single-link usage, this field is a hint for how many
	 * to allocate; otherwise, regular event ports and queues can be used.
	 */
    }

3) Replace the dedicated implicit_release_disabled field with a bit field
of explicit port capabilities. The implicit_release_disable functionality
is assigned to one bit, and a port-is-single-link-only  attribute is
assigned to other, with the remaining bits available for future assignment.

	* Event port configuration bitmap flags */
	#define RTE_EVENT_PORT_CFG_DISABLE_IMPL_REL    (1ULL << 0)
	/**< Configure the port not to release outstanding events in
	 * rte_event_dev_dequeue_burst(). If set, all events received through
	 * the port must be explicitly released with RTE_EVENT_OP_RELEASE or
	 * RTE_EVENT_OP_FORWARD. Must be unset if the device is not
	 * RTE_EVENT_DEV_CAP_IMPLICIT_RELEASE_DISABLE capable.
	 */
	#define RTE_EVENT_PORT_CFG_SINGLE_LINK         (1ULL << 1)

	/**< This event port links only to a single event queue.
	 *
	 *  @see rte_event_port_setup(), rte_event_port_link()
	 */

	#define RTE_EVENT_PORT_ATTR_IMPLICIT_RELEASE_DISABLE 3
	/**
	 * The implicit release disable attribute of the port
	 */

	struct rte_event_port_conf {
		uint32_t event_port_cfg;
		/**< Port cfg flags(EVENT_PORT_CFG_) */
	}

This patch also removes the depreciation notice and announce
the new eventdev ABI changes in release note.

Signed-off-by: Timothy McDaniel <timothy.mcdaniel@intel.com>
Acked-by: Harry van Haaren <harry.van.haaren@intel.com>
Acked-by: Pavan Nikhilesh <pbhagavatula@marvell.com>
Acked-by: Jerin Jacob <jerinj@marvell.com>
2020-10-15 23:16:07 +02:00

763 lines
16 KiB
C

/* SPDX-License-Identifier: BSD-3-Clause
* Copyright(c) 2017 Intel Corporation
*/
#include <inttypes.h>
#include <string.h>
#include <rte_bus_vdev.h>
#include <rte_lcore.h>
#include <rte_memzone.h>
#include <rte_kvargs.h>
#include <rte_errno.h>
#include <rte_cycles.h>
#include "opdl_evdev.h"
#include "opdl_ring.h"
#include "opdl_log.h"
#define EVENTDEV_NAME_OPDL_PMD event_opdl
#define NUMA_NODE_ARG "numa_node"
#define DO_VALIDATION_ARG "do_validation"
#define DO_TEST_ARG "self_test"
static void
opdl_info_get(struct rte_eventdev *dev, struct rte_event_dev_info *info);
uint16_t
opdl_event_enqueue_burst(void *port,
const struct rte_event ev[],
uint16_t num)
{
struct opdl_port *p = port;
if (unlikely(!p->opdl->data->dev_started))
return 0;
/* either rx_enqueue or disclaim*/
return p->enq(p, ev, num);
}
uint16_t
opdl_event_enqueue(void *port, const struct rte_event *ev)
{
struct opdl_port *p = port;
if (unlikely(!p->opdl->data->dev_started))
return 0;
return p->enq(p, ev, 1);
}
uint16_t
opdl_event_dequeue_burst(void *port,
struct rte_event *ev,
uint16_t num,
uint64_t wait)
{
struct opdl_port *p = (void *)port;
RTE_SET_USED(wait);
if (unlikely(!p->opdl->data->dev_started))
return 0;
/* This function pointer can point to tx_dequeue or claim*/
return p->deq(p, ev, num);
}
uint16_t
opdl_event_dequeue(void *port,
struct rte_event *ev,
uint64_t wait)
{
struct opdl_port *p = (void *)port;
if (unlikely(!p->opdl->data->dev_started))
return 0;
RTE_SET_USED(wait);
return p->deq(p, ev, 1);
}
static int
opdl_port_link(struct rte_eventdev *dev,
void *port,
const uint8_t queues[],
const uint8_t priorities[],
uint16_t num)
{
struct opdl_port *p = port;
RTE_SET_USED(priorities);
RTE_SET_USED(dev);
if (unlikely(dev->data->dev_started)) {
PMD_DRV_LOG(ERR, "DEV_ID:[%02d] : "
"Attempt to link queue (%u) to port %d while device started\n",
dev->data->dev_id,
queues[0],
p->id);
rte_errno = EINVAL;
return 0;
}
/* Max of 1 queue per port */
if (num > 1) {
PMD_DRV_LOG(ERR, "DEV_ID:[%02d] : "
"Attempt to link more than one queue (%u) to port %d requested\n",
dev->data->dev_id,
num,
p->id);
rte_errno = EDQUOT;
return 0;
}
if (!p->configured) {
PMD_DRV_LOG(ERR, "DEV_ID:[%02d] : "
"port %d not configured, cannot link to %u\n",
dev->data->dev_id,
p->id,
queues[0]);
rte_errno = EINVAL;
return 0;
}
if (p->external_qid != OPDL_INVALID_QID) {
PMD_DRV_LOG(ERR, "DEV_ID:[%02d] : "
"port %d already linked to queue %u, cannot link to %u\n",
dev->data->dev_id,
p->id,
p->external_qid,
queues[0]);
rte_errno = EINVAL;
return 0;
}
p->external_qid = queues[0];
return 1;
}
static int
opdl_port_unlink(struct rte_eventdev *dev,
void *port,
uint8_t queues[],
uint16_t nb_unlinks)
{
struct opdl_port *p = port;
RTE_SET_USED(queues);
RTE_SET_USED(nb_unlinks);
if (unlikely(dev->data->dev_started)) {
PMD_DRV_LOG(ERR, "DEV_ID:[%02d] : "
"Attempt to unlink queue (%u) to port %d while device started\n",
dev->data->dev_id,
queues[0],
p->id);
rte_errno = EINVAL;
return 0;
}
RTE_SET_USED(nb_unlinks);
/* Port Stuff */
p->queue_id = OPDL_INVALID_QID;
p->p_type = OPDL_INVALID_PORT;
p->external_qid = OPDL_INVALID_QID;
/* always unlink 0 queue due to statice pipeline */
return 0;
}
static int
opdl_port_setup(struct rte_eventdev *dev,
uint8_t port_id,
const struct rte_event_port_conf *conf)
{
struct opdl_evdev *device = opdl_pmd_priv(dev);
struct opdl_port *p = &device->ports[port_id];
RTE_SET_USED(conf);
/* Check if port already configured */
if (p->configured) {
PMD_DRV_LOG(ERR, "DEV_ID:[%02d] : "
"Attempt to setup port %d which is already setup\n",
dev->data->dev_id,
p->id);
return -EDQUOT;
}
*p = (struct opdl_port){0}; /* zero entire structure */
p->id = port_id;
p->opdl = device;
p->queue_id = OPDL_INVALID_QID;
p->external_qid = OPDL_INVALID_QID;
dev->data->ports[port_id] = p;
rte_smp_wmb();
p->configured = 1;
device->nb_ports++;
return 0;
}
static void
opdl_port_release(void *port)
{
struct opdl_port *p = (void *)port;
if (p == NULL ||
p->opdl->data->dev_started) {
return;
}
p->configured = 0;
p->initialized = 0;
}
static void
opdl_port_def_conf(struct rte_eventdev *dev, uint8_t port_id,
struct rte_event_port_conf *port_conf)
{
RTE_SET_USED(dev);
RTE_SET_USED(port_id);
port_conf->new_event_threshold = MAX_OPDL_CONS_Q_DEPTH;
port_conf->dequeue_depth = MAX_OPDL_CONS_Q_DEPTH;
port_conf->enqueue_depth = MAX_OPDL_CONS_Q_DEPTH;
}
static int
opdl_queue_setup(struct rte_eventdev *dev,
uint8_t queue_id,
const struct rte_event_queue_conf *conf)
{
enum queue_type type;
struct opdl_evdev *device = opdl_pmd_priv(dev);
/* Extra sanity check, probably not needed */
if (queue_id == OPDL_INVALID_QID) {
PMD_DRV_LOG(ERR, "DEV_ID:[%02d] : "
"Invalid queue id %u requested\n",
dev->data->dev_id,
queue_id);
return -EINVAL;
}
if (device->nb_q_md > device->max_queue_nb) {
PMD_DRV_LOG(ERR, "DEV_ID:[%02d] : "
"Max number of queues %u exceeded by request %u\n",
dev->data->dev_id,
device->max_queue_nb,
device->nb_q_md);
return -EINVAL;
}
if (RTE_EVENT_QUEUE_CFG_ALL_TYPES
& conf->event_queue_cfg) {
PMD_DRV_LOG(ERR, "DEV_ID:[%02d] : "
"QUEUE_CFG_ALL_TYPES not supported\n",
dev->data->dev_id);
return -ENOTSUP;
} else if (RTE_EVENT_QUEUE_CFG_SINGLE_LINK
& conf->event_queue_cfg) {
type = OPDL_Q_TYPE_SINGLE_LINK;
} else {
switch (conf->schedule_type) {
case RTE_SCHED_TYPE_ORDERED:
type = OPDL_Q_TYPE_ORDERED;
break;
case RTE_SCHED_TYPE_ATOMIC:
type = OPDL_Q_TYPE_ATOMIC;
break;
case RTE_SCHED_TYPE_PARALLEL:
type = OPDL_Q_TYPE_ORDERED;
break;
default:
PMD_DRV_LOG(ERR, "DEV_ID:[%02d] : "
"Unknown queue type %d requested\n",
dev->data->dev_id,
conf->event_queue_cfg);
return -EINVAL;
}
}
/* Check if queue id has been setup already */
uint32_t i;
for (i = 0; i < device->nb_q_md; i++) {
if (device->q_md[i].ext_id == queue_id) {
PMD_DRV_LOG(ERR, "DEV_ID:[%02d] : "
"queue id %u already setup\n",
dev->data->dev_id,
queue_id);
return -EINVAL;
}
}
device->q_md[device->nb_q_md].ext_id = queue_id;
device->q_md[device->nb_q_md].type = type;
device->q_md[device->nb_q_md].setup = 1;
device->nb_q_md++;
return 1;
}
static void
opdl_queue_release(struct rte_eventdev *dev, uint8_t queue_id)
{
struct opdl_evdev *device = opdl_pmd_priv(dev);
RTE_SET_USED(queue_id);
if (device->data->dev_started)
return;
}
static void
opdl_queue_def_conf(struct rte_eventdev *dev,
uint8_t queue_id,
struct rte_event_queue_conf *conf)
{
RTE_SET_USED(dev);
RTE_SET_USED(queue_id);
static const struct rte_event_queue_conf default_conf = {
.nb_atomic_flows = 1024,
.nb_atomic_order_sequences = 1,
.event_queue_cfg = 0,
.schedule_type = RTE_SCHED_TYPE_ORDERED,
.priority = RTE_EVENT_DEV_PRIORITY_NORMAL,
};
*conf = default_conf;
}
static int
opdl_dev_configure(const struct rte_eventdev *dev)
{
struct opdl_evdev *opdl = opdl_pmd_priv(dev);
const struct rte_eventdev_data *data = dev->data;
const struct rte_event_dev_config *conf = &data->dev_conf;
opdl->max_queue_nb = conf->nb_event_queues;
opdl->max_port_nb = conf->nb_event_ports;
opdl->nb_events_limit = conf->nb_events_limit;
if (conf->event_dev_cfg & RTE_EVENT_DEV_CFG_PER_DEQUEUE_TIMEOUT) {
PMD_DRV_LOG(ERR, "DEV_ID:[%02d] : "
"DEQUEUE_TIMEOUT not supported\n",
dev->data->dev_id);
return -ENOTSUP;
}
return 0;
}
static void
opdl_info_get(struct rte_eventdev *dev, struct rte_event_dev_info *info)
{
RTE_SET_USED(dev);
static const struct rte_event_dev_info evdev_opdl_info = {
.driver_name = OPDL_PMD_NAME,
.max_event_queues = RTE_EVENT_MAX_QUEUES_PER_DEV,
.max_event_queue_flows = OPDL_QID_NUM_FIDS,
.max_event_queue_priority_levels = OPDL_Q_PRIORITY_MAX,
.max_event_priority_levels = OPDL_IQS_MAX,
.max_event_ports = OPDL_PORTS_MAX,
.max_event_port_dequeue_depth = MAX_OPDL_CONS_Q_DEPTH,
.max_event_port_enqueue_depth = MAX_OPDL_CONS_Q_DEPTH,
.max_num_events = OPDL_INFLIGHT_EVENTS_TOTAL,
.event_dev_cap = RTE_EVENT_DEV_CAP_BURST_MODE |
RTE_EVENT_DEV_CAP_CARRY_FLOW_ID,
};
*info = evdev_opdl_info;
}
static void
opdl_dump(struct rte_eventdev *dev, FILE *f)
{
struct opdl_evdev *device = opdl_pmd_priv(dev);
if (!device->do_validation)
return;
fprintf(f,
"\n\n -- RING STATISTICS --\n");
uint32_t i;
for (i = 0; i < device->nb_opdls; i++)
opdl_ring_dump(device->opdl[i], f);
fprintf(f,
"\n\n -- PORT STATISTICS --\n"
"Type Port Index Port Id Queue Id Av. Req Size "
"Av. Grant Size Av. Cycles PP"
" Empty DEQs Non Empty DEQs Pkts Processed\n");
for (i = 0; i < device->max_port_nb; i++) {
char queue_id[64];
char total_cyc[64];
const char *p_type;
uint64_t cne, cpg;
struct opdl_port *port = &device->ports[i];
if (port->initialized) {
cne = port->port_stat[claim_non_empty];
cpg = port->port_stat[claim_pkts_granted];
if (port->p_type == OPDL_REGULAR_PORT)
p_type = "REG";
else if (port->p_type == OPDL_PURE_RX_PORT)
p_type = " RX";
else if (port->p_type == OPDL_PURE_TX_PORT)
p_type = " TX";
else if (port->p_type == OPDL_ASYNC_PORT)
p_type = "SYNC";
else
p_type = "????";
snprintf(queue_id, sizeof(queue_id), "%02u",
port->external_qid);
if (port->p_type == OPDL_REGULAR_PORT ||
port->p_type == OPDL_ASYNC_PORT)
snprintf(total_cyc, sizeof(total_cyc),
" %'16"PRIu64"",
(cpg != 0 ?
port->port_stat[total_cycles] / cpg
: 0));
else
snprintf(total_cyc, sizeof(total_cyc),
" ----");
fprintf(f,
"%4s %10u %8u %9s %'16"PRIu64" %'16"PRIu64" %s "
"%'16"PRIu64" %'16"PRIu64" %'16"PRIu64"\n",
p_type,
i,
port->id,
(port->external_qid == OPDL_INVALID_QID ? "---"
: queue_id),
(cne != 0 ?
port->port_stat[claim_pkts_requested] / cne
: 0),
(cne != 0 ?
port->port_stat[claim_pkts_granted] / cne
: 0),
total_cyc,
port->port_stat[claim_empty],
port->port_stat[claim_non_empty],
port->port_stat[claim_pkts_granted]);
}
}
fprintf(f, "\n");
}
static void
opdl_stop(struct rte_eventdev *dev)
{
struct opdl_evdev *device = opdl_pmd_priv(dev);
opdl_xstats_uninit(dev);
destroy_queues_and_rings(dev);
device->started = 0;
rte_smp_wmb();
}
static int
opdl_start(struct rte_eventdev *dev)
{
int err = 0;
if (!err)
err = create_queues_and_rings(dev);
if (!err)
err = assign_internal_queue_ids(dev);
if (!err)
err = initialise_queue_zero_ports(dev);
if (!err)
err = initialise_all_other_ports(dev);
if (!err)
err = check_queues_linked(dev);
if (!err)
err = opdl_add_event_handlers(dev);
if (!err)
err = build_all_dependencies(dev);
if (!err) {
opdl_xstats_init(dev);
struct opdl_evdev *device = opdl_pmd_priv(dev);
PMD_DRV_LOG(INFO, "DEV_ID:[%02d] : "
"SUCCESS : Created %u total queues (%u ex, %u in),"
" %u opdls, %u event_dev ports, %u input ports",
opdl_pmd_dev_id(device),
device->nb_queues,
(device->nb_queues - device->nb_opdls),
device->nb_opdls,
device->nb_opdls,
device->nb_ports,
device->queue[0].nb_ports);
} else
opdl_stop(dev);
return err;
}
static int
opdl_close(struct rte_eventdev *dev)
{
struct opdl_evdev *device = opdl_pmd_priv(dev);
uint32_t i;
for (i = 0; i < device->max_port_nb; i++) {
memset(&device->ports[i],
0,
sizeof(struct opdl_port));
}
memset(&device->s_md,
0x0,
sizeof(struct opdl_stage_meta_data)*OPDL_PORTS_MAX);
memset(&device->q_md,
0xFF,
sizeof(struct opdl_queue_meta_data)*OPDL_MAX_QUEUES);
memset(device->q_map_ex_to_in,
0,
sizeof(uint8_t)*OPDL_INVALID_QID);
opdl_xstats_uninit(dev);
device->max_port_nb = 0;
device->max_queue_nb = 0;
device->nb_opdls = 0;
device->nb_queues = 0;
device->nb_ports = 0;
device->nb_q_md = 0;
dev->data->nb_queues = 0;
dev->data->nb_ports = 0;
return 0;
}
static int
assign_numa_node(const char *key __rte_unused, const char *value, void *opaque)
{
int *socket_id = opaque;
*socket_id = atoi(value);
if (*socket_id >= RTE_MAX_NUMA_NODES)
return -1;
return 0;
}
static int
set_do_validation(const char *key __rte_unused, const char *value, void *opaque)
{
int *do_val = opaque;
*do_val = atoi(value);
if (*do_val != 0)
*do_val = 1;
return 0;
}
static int
set_do_test(const char *key __rte_unused, const char *value, void *opaque)
{
int *do_test = opaque;
*do_test = atoi(value);
if (*do_test != 0)
*do_test = 1;
return 0;
}
static int
opdl_probe(struct rte_vdev_device *vdev)
{
static struct rte_eventdev_ops evdev_opdl_ops = {
.dev_configure = opdl_dev_configure,
.dev_infos_get = opdl_info_get,
.dev_close = opdl_close,
.dev_start = opdl_start,
.dev_stop = opdl_stop,
.dump = opdl_dump,
.queue_def_conf = opdl_queue_def_conf,
.queue_setup = opdl_queue_setup,
.queue_release = opdl_queue_release,
.port_def_conf = opdl_port_def_conf,
.port_setup = opdl_port_setup,
.port_release = opdl_port_release,
.port_link = opdl_port_link,
.port_unlink = opdl_port_unlink,
.xstats_get = opdl_xstats_get,
.xstats_get_names = opdl_xstats_get_names,
.xstats_get_by_name = opdl_xstats_get_by_name,
.xstats_reset = opdl_xstats_reset,
};
static const char *const args[] = {
NUMA_NODE_ARG,
DO_VALIDATION_ARG,
DO_TEST_ARG,
NULL
};
const char *name;
const char *params;
struct rte_eventdev *dev;
struct opdl_evdev *opdl;
int socket_id = rte_socket_id();
int do_validation = 0;
int do_test = 0;
int str_len;
int test_result = 0;
name = rte_vdev_device_name(vdev);
params = rte_vdev_device_args(vdev);
if (params != NULL && params[0] != '\0') {
struct rte_kvargs *kvlist = rte_kvargs_parse(params, args);
if (!kvlist) {
PMD_DRV_LOG(INFO,
"Ignoring unsupported parameters when creating device '%s'\n",
name);
} else {
int ret = rte_kvargs_process(kvlist, NUMA_NODE_ARG,
assign_numa_node, &socket_id);
if (ret != 0) {
PMD_DRV_LOG(ERR,
"%s: Error parsing numa node parameter",
name);
rte_kvargs_free(kvlist);
return ret;
}
ret = rte_kvargs_process(kvlist, DO_VALIDATION_ARG,
set_do_validation, &do_validation);
if (ret != 0) {
PMD_DRV_LOG(ERR,
"%s: Error parsing do validation parameter",
name);
rte_kvargs_free(kvlist);
return ret;
}
ret = rte_kvargs_process(kvlist, DO_TEST_ARG,
set_do_test, &do_test);
if (ret != 0) {
PMD_DRV_LOG(ERR,
"%s: Error parsing do test parameter",
name);
rte_kvargs_free(kvlist);
return ret;
}
rte_kvargs_free(kvlist);
}
}
dev = rte_event_pmd_vdev_init(name,
sizeof(struct opdl_evdev), socket_id);
if (dev == NULL) {
PMD_DRV_LOG(ERR, "eventdev vdev init() failed");
return -EFAULT;
}
PMD_DRV_LOG(INFO, "DEV_ID:[%02d] : "
"Success - creating eventdev device %s, numa_node:[%d], do_valdation:[%s]"
" , self_test:[%s]\n",
dev->data->dev_id,
name,
socket_id,
(do_validation ? "true" : "false"),
(do_test ? "true" : "false"));
dev->dev_ops = &evdev_opdl_ops;
dev->enqueue = opdl_event_enqueue;
dev->enqueue_burst = opdl_event_enqueue_burst;
dev->enqueue_new_burst = opdl_event_enqueue_burst;
dev->enqueue_forward_burst = opdl_event_enqueue_burst;
dev->dequeue = opdl_event_dequeue;
dev->dequeue_burst = opdl_event_dequeue_burst;
if (rte_eal_process_type() != RTE_PROC_PRIMARY)
return 0;
opdl = dev->data->dev_private;
opdl->data = dev->data;
opdl->socket = socket_id;
opdl->do_validation = do_validation;
opdl->do_test = do_test;
str_len = strlen(name);
memcpy(opdl->service_name, name, str_len);
if (do_test == 1)
test_result = opdl_selftest();
return test_result;
}
static int
opdl_remove(struct rte_vdev_device *vdev)
{
const char *name;
name = rte_vdev_device_name(vdev);
if (name == NULL)
return -EINVAL;
PMD_DRV_LOG(INFO, "Closing eventdev opdl device %s\n", name);
return rte_event_pmd_vdev_uninit(name);
}
static struct rte_vdev_driver evdev_opdl_pmd_drv = {
.probe = opdl_probe,
.remove = opdl_remove
};
RTE_LOG_REGISTER(opdl_logtype_driver, pmd.event.opdl.driver, INFO);
RTE_PMD_REGISTER_VDEV(EVENTDEV_NAME_OPDL_PMD, evdev_opdl_pmd_drv);
RTE_PMD_REGISTER_PARAM_STRING(event_opdl, NUMA_NODE_ARG "=<int>"
DO_VALIDATION_ARG "=<int>" DO_TEST_ARG "=<int>");