numam-dpdk/drivers/net/softnic/rte_eth_softnic.c
Ferruh Yigit ffc905f3b8 ethdev: separate driver APIs
Create a rte_ethdev_driver.h file and move PMD specific APIs here.
Drivers updated to include this new header file.

There is no update in header content and since ethdev.h included by
ethdev_driver.h, nothing changed from driver point of view, only
logically grouping of APIs. From applications point of view they can't
access to driver specific APIs anymore and they shouldn't.

More PMD specific data structures still remain in ethdev.h because of
inline functions in header use them. Those will be handled separately.

Signed-off-by: Ferruh Yigit <ferruh.yigit@intel.com>
Acked-by: Shreyansh Jain <shreyansh.jain@nxp.com>
Acked-by: Andrew Rybchenko <arybchenko@solarflare.com>
Acked-by: Thomas Monjalon <thomas@monjalon.net>
2018-01-22 01:26:49 +01:00

823 lines
19 KiB
C

/* SPDX-License-Identifier: BSD-3-Clause
* Copyright(c) 2017 Intel Corporation
*/
#include <stdint.h>
#include <stdlib.h>
#include <string.h>
#include <rte_ethdev_driver.h>
#include <rte_ethdev_vdev.h>
#include <rte_malloc.h>
#include <rte_bus_vdev.h>
#include <rte_kvargs.h>
#include <rte_errno.h>
#include <rte_ring.h>
#include <rte_sched.h>
#include <rte_tm_driver.h>
#include "rte_eth_softnic.h"
#include "rte_eth_softnic_internals.h"
#define DEV_HARD(p) \
(&rte_eth_devices[p->hard.port_id])
#define PMD_PARAM_SOFT_TM "soft_tm"
#define PMD_PARAM_SOFT_TM_RATE "soft_tm_rate"
#define PMD_PARAM_SOFT_TM_NB_QUEUES "soft_tm_nb_queues"
#define PMD_PARAM_SOFT_TM_QSIZE0 "soft_tm_qsize0"
#define PMD_PARAM_SOFT_TM_QSIZE1 "soft_tm_qsize1"
#define PMD_PARAM_SOFT_TM_QSIZE2 "soft_tm_qsize2"
#define PMD_PARAM_SOFT_TM_QSIZE3 "soft_tm_qsize3"
#define PMD_PARAM_SOFT_TM_ENQ_BSZ "soft_tm_enq_bsz"
#define PMD_PARAM_SOFT_TM_DEQ_BSZ "soft_tm_deq_bsz"
#define PMD_PARAM_HARD_NAME "hard_name"
#define PMD_PARAM_HARD_TX_QUEUE_ID "hard_tx_queue_id"
static const char *pmd_valid_args[] = {
PMD_PARAM_SOFT_TM,
PMD_PARAM_SOFT_TM_RATE,
PMD_PARAM_SOFT_TM_NB_QUEUES,
PMD_PARAM_SOFT_TM_QSIZE0,
PMD_PARAM_SOFT_TM_QSIZE1,
PMD_PARAM_SOFT_TM_QSIZE2,
PMD_PARAM_SOFT_TM_QSIZE3,
PMD_PARAM_SOFT_TM_ENQ_BSZ,
PMD_PARAM_SOFT_TM_DEQ_BSZ,
PMD_PARAM_HARD_NAME,
PMD_PARAM_HARD_TX_QUEUE_ID,
NULL
};
static const struct rte_eth_dev_info pmd_dev_info = {
.min_rx_bufsize = 0,
.max_rx_pktlen = UINT32_MAX,
.max_rx_queues = UINT16_MAX,
.max_tx_queues = UINT16_MAX,
.rx_desc_lim = {
.nb_max = UINT16_MAX,
.nb_min = 0,
.nb_align = 1,
},
.tx_desc_lim = {
.nb_max = UINT16_MAX,
.nb_min = 0,
.nb_align = 1,
},
};
static void
pmd_dev_infos_get(struct rte_eth_dev *dev __rte_unused,
struct rte_eth_dev_info *dev_info)
{
memcpy(dev_info, &pmd_dev_info, sizeof(*dev_info));
}
static int
pmd_dev_configure(struct rte_eth_dev *dev)
{
struct pmd_internals *p = dev->data->dev_private;
struct rte_eth_dev *hard_dev = DEV_HARD(p);
if (dev->data->nb_rx_queues > hard_dev->data->nb_rx_queues)
return -1;
if (p->params.hard.tx_queue_id >= hard_dev->data->nb_tx_queues)
return -1;
return 0;
}
static int
pmd_rx_queue_setup(struct rte_eth_dev *dev,
uint16_t rx_queue_id,
uint16_t nb_rx_desc __rte_unused,
unsigned int socket_id,
const struct rte_eth_rxconf *rx_conf __rte_unused,
struct rte_mempool *mb_pool __rte_unused)
{
struct pmd_internals *p = dev->data->dev_private;
if (p->params.soft.intrusive == 0) {
struct pmd_rx_queue *rxq;
rxq = rte_zmalloc_socket(p->params.soft.name,
sizeof(struct pmd_rx_queue), 0, socket_id);
if (rxq == NULL)
return -ENOMEM;
rxq->hard.port_id = p->hard.port_id;
rxq->hard.rx_queue_id = rx_queue_id;
dev->data->rx_queues[rx_queue_id] = rxq;
} else {
struct rte_eth_dev *hard_dev = DEV_HARD(p);
void *rxq = hard_dev->data->rx_queues[rx_queue_id];
if (rxq == NULL)
return -1;
dev->data->rx_queues[rx_queue_id] = rxq;
}
return 0;
}
static int
pmd_tx_queue_setup(struct rte_eth_dev *dev,
uint16_t tx_queue_id,
uint16_t nb_tx_desc,
unsigned int socket_id,
const struct rte_eth_txconf *tx_conf __rte_unused)
{
uint32_t size = RTE_ETH_NAME_MAX_LEN + strlen("_txq") + 4;
char name[size];
struct rte_ring *r;
snprintf(name, sizeof(name), "%s_txq%04x",
dev->data->name, tx_queue_id);
r = rte_ring_create(name, nb_tx_desc, socket_id,
RING_F_SP_ENQ | RING_F_SC_DEQ);
if (r == NULL)
return -1;
dev->data->tx_queues[tx_queue_id] = r;
return 0;
}
static int
pmd_dev_start(struct rte_eth_dev *dev)
{
struct pmd_internals *p = dev->data->dev_private;
if (tm_used(dev)) {
int status = tm_start(p);
if (status)
return status;
}
dev->data->dev_link.link_status = ETH_LINK_UP;
if (p->params.soft.intrusive) {
struct rte_eth_dev *hard_dev = DEV_HARD(p);
/* The hard_dev->rx_pkt_burst should be stable by now */
dev->rx_pkt_burst = hard_dev->rx_pkt_burst;
}
return 0;
}
static void
pmd_dev_stop(struct rte_eth_dev *dev)
{
struct pmd_internals *p = dev->data->dev_private;
dev->data->dev_link.link_status = ETH_LINK_DOWN;
if (tm_used(dev))
tm_stop(p);
}
static void
pmd_dev_close(struct rte_eth_dev *dev)
{
uint32_t i;
/* TX queues */
for (i = 0; i < dev->data->nb_tx_queues; i++)
rte_ring_free((struct rte_ring *)dev->data->tx_queues[i]);
}
static int
pmd_link_update(struct rte_eth_dev *dev __rte_unused,
int wait_to_complete __rte_unused)
{
return 0;
}
static int
pmd_tm_ops_get(struct rte_eth_dev *dev, void *arg)
{
*(const struct rte_tm_ops **)arg =
(tm_enabled(dev)) ? &pmd_tm_ops : NULL;
return 0;
}
static const struct eth_dev_ops pmd_ops = {
.dev_configure = pmd_dev_configure,
.dev_start = pmd_dev_start,
.dev_stop = pmd_dev_stop,
.dev_close = pmd_dev_close,
.link_update = pmd_link_update,
.dev_infos_get = pmd_dev_infos_get,
.rx_queue_setup = pmd_rx_queue_setup,
.tx_queue_setup = pmd_tx_queue_setup,
.tm_ops_get = pmd_tm_ops_get,
};
static uint16_t
pmd_rx_pkt_burst(void *rxq,
struct rte_mbuf **rx_pkts,
uint16_t nb_pkts)
{
struct pmd_rx_queue *rx_queue = rxq;
return rte_eth_rx_burst(rx_queue->hard.port_id,
rx_queue->hard.rx_queue_id,
rx_pkts,
nb_pkts);
}
static uint16_t
pmd_tx_pkt_burst(void *txq,
struct rte_mbuf **tx_pkts,
uint16_t nb_pkts)
{
return (uint16_t)rte_ring_enqueue_burst(txq,
(void **)tx_pkts,
nb_pkts,
NULL);
}
static __rte_always_inline int
run_default(struct rte_eth_dev *dev)
{
struct pmd_internals *p = dev->data->dev_private;
/* Persistent context: Read Only (update not required) */
struct rte_mbuf **pkts = p->soft.def.pkts;
uint16_t nb_tx_queues = dev->data->nb_tx_queues;
/* Persistent context: Read - Write (update required) */
uint32_t txq_pos = p->soft.def.txq_pos;
uint32_t pkts_len = p->soft.def.pkts_len;
uint32_t flush_count = p->soft.def.flush_count;
/* Not part of the persistent context */
uint32_t pos;
uint16_t i;
/* Soft device TXQ read, Hard device TXQ write */
for (i = 0; i < nb_tx_queues; i++) {
struct rte_ring *txq = dev->data->tx_queues[txq_pos];
/* Read soft device TXQ burst to packet enqueue buffer */
pkts_len += rte_ring_sc_dequeue_burst(txq,
(void **)&pkts[pkts_len],
DEFAULT_BURST_SIZE,
NULL);
/* Increment soft device TXQ */
txq_pos++;
if (txq_pos >= nb_tx_queues)
txq_pos = 0;
/* Hard device TXQ write when complete burst is available */
if (pkts_len >= DEFAULT_BURST_SIZE) {
for (pos = 0; pos < pkts_len; )
pos += rte_eth_tx_burst(p->hard.port_id,
p->params.hard.tx_queue_id,
&pkts[pos],
(uint16_t)(pkts_len - pos));
pkts_len = 0;
flush_count = 0;
break;
}
}
if (flush_count >= FLUSH_COUNT_THRESHOLD) {
for (pos = 0; pos < pkts_len; )
pos += rte_eth_tx_burst(p->hard.port_id,
p->params.hard.tx_queue_id,
&pkts[pos],
(uint16_t)(pkts_len - pos));
pkts_len = 0;
flush_count = 0;
}
p->soft.def.txq_pos = txq_pos;
p->soft.def.pkts_len = pkts_len;
p->soft.def.flush_count = flush_count + 1;
return 0;
}
static __rte_always_inline int
run_tm(struct rte_eth_dev *dev)
{
struct pmd_internals *p = dev->data->dev_private;
/* Persistent context: Read Only (update not required) */
struct rte_sched_port *sched = p->soft.tm.sched;
struct rte_mbuf **pkts_enq = p->soft.tm.pkts_enq;
struct rte_mbuf **pkts_deq = p->soft.tm.pkts_deq;
uint32_t enq_bsz = p->params.soft.tm.enq_bsz;
uint32_t deq_bsz = p->params.soft.tm.deq_bsz;
uint16_t nb_tx_queues = dev->data->nb_tx_queues;
/* Persistent context: Read - Write (update required) */
uint32_t txq_pos = p->soft.tm.txq_pos;
uint32_t pkts_enq_len = p->soft.tm.pkts_enq_len;
uint32_t flush_count = p->soft.tm.flush_count;
/* Not part of the persistent context */
uint32_t pkts_deq_len, pos;
uint16_t i;
/* Soft device TXQ read, TM enqueue */
for (i = 0; i < nb_tx_queues; i++) {
struct rte_ring *txq = dev->data->tx_queues[txq_pos];
/* Read TXQ burst to packet enqueue buffer */
pkts_enq_len += rte_ring_sc_dequeue_burst(txq,
(void **)&pkts_enq[pkts_enq_len],
enq_bsz,
NULL);
/* Increment TXQ */
txq_pos++;
if (txq_pos >= nb_tx_queues)
txq_pos = 0;
/* TM enqueue when complete burst is available */
if (pkts_enq_len >= enq_bsz) {
rte_sched_port_enqueue(sched, pkts_enq, pkts_enq_len);
pkts_enq_len = 0;
flush_count = 0;
break;
}
}
if (flush_count >= FLUSH_COUNT_THRESHOLD) {
if (pkts_enq_len)
rte_sched_port_enqueue(sched, pkts_enq, pkts_enq_len);
pkts_enq_len = 0;
flush_count = 0;
}
p->soft.tm.txq_pos = txq_pos;
p->soft.tm.pkts_enq_len = pkts_enq_len;
p->soft.tm.flush_count = flush_count + 1;
/* TM dequeue, Hard device TXQ write */
pkts_deq_len = rte_sched_port_dequeue(sched, pkts_deq, deq_bsz);
for (pos = 0; pos < pkts_deq_len; )
pos += rte_eth_tx_burst(p->hard.port_id,
p->params.hard.tx_queue_id,
&pkts_deq[pos],
(uint16_t)(pkts_deq_len - pos));
return 0;
}
int
rte_pmd_softnic_run(uint16_t port_id)
{
struct rte_eth_dev *dev = &rte_eth_devices[port_id];
#ifdef RTE_LIBRTE_ETHDEV_DEBUG
RTE_ETH_VALID_PORTID_OR_ERR_RET(port_id, 0);
#endif
return (tm_used(dev)) ? run_tm(dev) : run_default(dev);
}
static struct ether_addr eth_addr = { .addr_bytes = {0} };
static uint32_t
eth_dev_speed_max_mbps(uint32_t speed_capa)
{
uint32_t rate_mbps[32] = {
ETH_SPEED_NUM_NONE,
ETH_SPEED_NUM_10M,
ETH_SPEED_NUM_10M,
ETH_SPEED_NUM_100M,
ETH_SPEED_NUM_100M,
ETH_SPEED_NUM_1G,
ETH_SPEED_NUM_2_5G,
ETH_SPEED_NUM_5G,
ETH_SPEED_NUM_10G,
ETH_SPEED_NUM_20G,
ETH_SPEED_NUM_25G,
ETH_SPEED_NUM_40G,
ETH_SPEED_NUM_50G,
ETH_SPEED_NUM_56G,
ETH_SPEED_NUM_100G,
};
uint32_t pos = (speed_capa) ? (31 - __builtin_clz(speed_capa)) : 0;
return rate_mbps[pos];
}
static int
default_init(struct pmd_internals *p,
struct pmd_params *params,
int numa_node)
{
p->soft.def.pkts = rte_zmalloc_socket(params->soft.name,
2 * DEFAULT_BURST_SIZE * sizeof(struct rte_mbuf *),
0,
numa_node);
if (p->soft.def.pkts == NULL)
return -ENOMEM;
return 0;
}
static void
default_free(struct pmd_internals *p)
{
rte_free(p->soft.def.pkts);
}
static void *
pmd_init(struct pmd_params *params, int numa_node)
{
struct pmd_internals *p;
int status;
p = rte_zmalloc_socket(params->soft.name,
sizeof(struct pmd_internals),
0,
numa_node);
if (p == NULL)
return NULL;
memcpy(&p->params, params, sizeof(p->params));
rte_eth_dev_get_port_by_name(params->hard.name, &p->hard.port_id);
/* Default */
status = default_init(p, params, numa_node);
if (status) {
free(p->params.hard.name);
rte_free(p);
return NULL;
}
/* Traffic Management (TM)*/
if (params->soft.flags & PMD_FEATURE_TM) {
status = tm_init(p, params, numa_node);
if (status) {
default_free(p);
free(p->params.hard.name);
rte_free(p);
return NULL;
}
}
return p;
}
static void
pmd_free(struct pmd_internals *p)
{
if (p->params.soft.flags & PMD_FEATURE_TM)
tm_free(p);
default_free(p);
free(p->params.hard.name);
rte_free(p);
}
static int
pmd_ethdev_register(struct rte_vdev_device *vdev,
struct pmd_params *params,
void *dev_private)
{
struct rte_eth_dev_info hard_info;
struct rte_eth_dev *soft_dev;
uint32_t hard_speed;
int numa_node;
uint16_t hard_port_id;
rte_eth_dev_get_port_by_name(params->hard.name, &hard_port_id);
rte_eth_dev_info_get(hard_port_id, &hard_info);
hard_speed = eth_dev_speed_max_mbps(hard_info.speed_capa);
numa_node = rte_eth_dev_socket_id(hard_port_id);
/* Ethdev entry allocation */
soft_dev = rte_eth_dev_allocate(params->soft.name);
if (!soft_dev)
return -ENOMEM;
/* dev */
soft_dev->rx_pkt_burst = (params->soft.intrusive) ?
NULL : /* set up later */
pmd_rx_pkt_burst;
soft_dev->tx_pkt_burst = pmd_tx_pkt_burst;
soft_dev->tx_pkt_prepare = NULL;
soft_dev->dev_ops = &pmd_ops;
soft_dev->device = &vdev->device;
/* dev->data */
soft_dev->data->dev_private = dev_private;
soft_dev->data->dev_link.link_speed = hard_speed;
soft_dev->data->dev_link.link_duplex = ETH_LINK_FULL_DUPLEX;
soft_dev->data->dev_link.link_autoneg = ETH_LINK_AUTONEG;
soft_dev->data->dev_link.link_status = ETH_LINK_DOWN;
soft_dev->data->mac_addrs = &eth_addr;
soft_dev->data->promiscuous = 1;
soft_dev->data->kdrv = RTE_KDRV_NONE;
soft_dev->data->numa_node = numa_node;
return 0;
}
static int
get_string(const char *key __rte_unused, const char *value, void *extra_args)
{
if (!value || !extra_args)
return -EINVAL;
*(char **)extra_args = strdup(value);
if (!*(char **)extra_args)
return -ENOMEM;
return 0;
}
static int
get_uint32(const char *key __rte_unused, const char *value, void *extra_args)
{
if (!value || !extra_args)
return -EINVAL;
*(uint32_t *)extra_args = strtoull(value, NULL, 0);
return 0;
}
static int
pmd_parse_args(struct pmd_params *p, const char *name, const char *params)
{
struct rte_kvargs *kvlist;
int i, ret;
kvlist = rte_kvargs_parse(params, pmd_valid_args);
if (kvlist == NULL)
return -EINVAL;
/* Set default values */
memset(p, 0, sizeof(*p));
p->soft.name = name;
p->soft.intrusive = INTRUSIVE;
p->soft.tm.rate = 0;
p->soft.tm.nb_queues = SOFTNIC_SOFT_TM_NB_QUEUES;
for (i = 0; i < RTE_SCHED_TRAFFIC_CLASSES_PER_PIPE; i++)
p->soft.tm.qsize[i] = SOFTNIC_SOFT_TM_QUEUE_SIZE;
p->soft.tm.enq_bsz = SOFTNIC_SOFT_TM_ENQ_BSZ;
p->soft.tm.deq_bsz = SOFTNIC_SOFT_TM_DEQ_BSZ;
p->hard.tx_queue_id = SOFTNIC_HARD_TX_QUEUE_ID;
/* SOFT: TM (optional) */
if (rte_kvargs_count(kvlist, PMD_PARAM_SOFT_TM) == 1) {
char *s;
ret = rte_kvargs_process(kvlist, PMD_PARAM_SOFT_TM,
&get_string, &s);
if (ret < 0)
goto out_free;
if (strcmp(s, "on") == 0)
p->soft.flags |= PMD_FEATURE_TM;
else if (strcmp(s, "off") == 0)
p->soft.flags &= ~PMD_FEATURE_TM;
else
ret = -EINVAL;
free(s);
if (ret)
goto out_free;
}
/* SOFT: TM rate (measured in bytes/second) (optional) */
if (rte_kvargs_count(kvlist, PMD_PARAM_SOFT_TM_RATE) == 1) {
ret = rte_kvargs_process(kvlist, PMD_PARAM_SOFT_TM_RATE,
&get_uint32, &p->soft.tm.rate);
if (ret < 0)
goto out_free;
p->soft.flags |= PMD_FEATURE_TM;
}
/* SOFT: TM number of queues (optional) */
if (rte_kvargs_count(kvlist, PMD_PARAM_SOFT_TM_NB_QUEUES) == 1) {
ret = rte_kvargs_process(kvlist, PMD_PARAM_SOFT_TM_NB_QUEUES,
&get_uint32, &p->soft.tm.nb_queues);
if (ret < 0)
goto out_free;
p->soft.flags |= PMD_FEATURE_TM;
}
/* SOFT: TM queue size 0 .. 3 (optional) */
if (rte_kvargs_count(kvlist, PMD_PARAM_SOFT_TM_QSIZE0) == 1) {
uint32_t qsize;
ret = rte_kvargs_process(kvlist, PMD_PARAM_SOFT_TM_QSIZE0,
&get_uint32, &qsize);
if (ret < 0)
goto out_free;
p->soft.tm.qsize[0] = (uint16_t)qsize;
p->soft.flags |= PMD_FEATURE_TM;
}
if (rte_kvargs_count(kvlist, PMD_PARAM_SOFT_TM_QSIZE1) == 1) {
uint32_t qsize;
ret = rte_kvargs_process(kvlist, PMD_PARAM_SOFT_TM_QSIZE1,
&get_uint32, &qsize);
if (ret < 0)
goto out_free;
p->soft.tm.qsize[1] = (uint16_t)qsize;
p->soft.flags |= PMD_FEATURE_TM;
}
if (rte_kvargs_count(kvlist, PMD_PARAM_SOFT_TM_QSIZE2) == 1) {
uint32_t qsize;
ret = rte_kvargs_process(kvlist, PMD_PARAM_SOFT_TM_QSIZE2,
&get_uint32, &qsize);
if (ret < 0)
goto out_free;
p->soft.tm.qsize[2] = (uint16_t)qsize;
p->soft.flags |= PMD_FEATURE_TM;
}
if (rte_kvargs_count(kvlist, PMD_PARAM_SOFT_TM_QSIZE3) == 1) {
uint32_t qsize;
ret = rte_kvargs_process(kvlist, PMD_PARAM_SOFT_TM_QSIZE3,
&get_uint32, &qsize);
if (ret < 0)
goto out_free;
p->soft.tm.qsize[3] = (uint16_t)qsize;
p->soft.flags |= PMD_FEATURE_TM;
}
/* SOFT: TM enqueue burst size (optional) */
if (rte_kvargs_count(kvlist, PMD_PARAM_SOFT_TM_ENQ_BSZ) == 1) {
ret = rte_kvargs_process(kvlist, PMD_PARAM_SOFT_TM_ENQ_BSZ,
&get_uint32, &p->soft.tm.enq_bsz);
if (ret < 0)
goto out_free;
p->soft.flags |= PMD_FEATURE_TM;
}
/* SOFT: TM dequeue burst size (optional) */
if (rte_kvargs_count(kvlist, PMD_PARAM_SOFT_TM_DEQ_BSZ) == 1) {
ret = rte_kvargs_process(kvlist, PMD_PARAM_SOFT_TM_DEQ_BSZ,
&get_uint32, &p->soft.tm.deq_bsz);
if (ret < 0)
goto out_free;
p->soft.flags |= PMD_FEATURE_TM;
}
/* HARD: name (mandatory) */
if (rte_kvargs_count(kvlist, PMD_PARAM_HARD_NAME) == 1) {
ret = rte_kvargs_process(kvlist, PMD_PARAM_HARD_NAME,
&get_string, &p->hard.name);
if (ret < 0)
goto out_free;
} else {
ret = -EINVAL;
goto out_free;
}
/* HARD: tx_queue_id (optional) */
if (rte_kvargs_count(kvlist, PMD_PARAM_HARD_TX_QUEUE_ID) == 1) {
ret = rte_kvargs_process(kvlist, PMD_PARAM_HARD_TX_QUEUE_ID,
&get_uint32, &p->hard.tx_queue_id);
if (ret < 0)
goto out_free;
}
out_free:
rte_kvargs_free(kvlist);
return ret;
}
static int
pmd_probe(struct rte_vdev_device *vdev)
{
struct pmd_params p;
const char *params;
int status;
struct rte_eth_dev_info hard_info;
uint32_t hard_speed;
uint16_t hard_port_id;
int numa_node;
void *dev_private;
RTE_LOG(INFO, PMD,
"Probing device \"%s\"\n",
rte_vdev_device_name(vdev));
/* Parse input arguments */
params = rte_vdev_device_args(vdev);
if (!params)
return -EINVAL;
status = pmd_parse_args(&p, rte_vdev_device_name(vdev), params);
if (status)
return status;
/* Check input arguments */
if (rte_eth_dev_get_port_by_name(p.hard.name, &hard_port_id))
return -EINVAL;
rte_eth_dev_info_get(hard_port_id, &hard_info);
hard_speed = eth_dev_speed_max_mbps(hard_info.speed_capa);
numa_node = rte_eth_dev_socket_id(hard_port_id);
if (p.hard.tx_queue_id >= hard_info.max_tx_queues)
return -EINVAL;
if (p.soft.flags & PMD_FEATURE_TM) {
status = tm_params_check(&p, hard_speed);
if (status)
return status;
}
/* Allocate and initialize soft ethdev private data */
dev_private = pmd_init(&p, numa_node);
if (dev_private == NULL)
return -ENOMEM;
/* Register soft ethdev */
RTE_LOG(INFO, PMD,
"Creating soft ethdev \"%s\" for hard ethdev \"%s\"\n",
p.soft.name, p.hard.name);
status = pmd_ethdev_register(vdev, &p, dev_private);
if (status) {
pmd_free(dev_private);
return status;
}
return 0;
}
static int
pmd_remove(struct rte_vdev_device *vdev)
{
struct rte_eth_dev *dev = NULL;
struct pmd_internals *p;
if (!vdev)
return -EINVAL;
RTE_LOG(INFO, PMD, "Removing device \"%s\"\n",
rte_vdev_device_name(vdev));
/* Find the ethdev entry */
dev = rte_eth_dev_allocated(rte_vdev_device_name(vdev));
if (dev == NULL)
return -ENODEV;
p = dev->data->dev_private;
/* Free device data structures*/
pmd_free(p);
rte_free(dev->data);
rte_eth_dev_release_port(dev);
return 0;
}
static struct rte_vdev_driver pmd_softnic_drv = {
.probe = pmd_probe,
.remove = pmd_remove,
};
RTE_PMD_REGISTER_VDEV(net_softnic, pmd_softnic_drv);
RTE_PMD_REGISTER_PARAM_STRING(net_softnic,
PMD_PARAM_SOFT_TM "=on|off "
PMD_PARAM_SOFT_TM_RATE "=<int> "
PMD_PARAM_SOFT_TM_NB_QUEUES "=<int> "
PMD_PARAM_SOFT_TM_QSIZE0 "=<int> "
PMD_PARAM_SOFT_TM_QSIZE1 "=<int> "
PMD_PARAM_SOFT_TM_QSIZE2 "=<int> "
PMD_PARAM_SOFT_TM_QSIZE3 "=<int> "
PMD_PARAM_SOFT_TM_ENQ_BSZ "=<int> "
PMD_PARAM_SOFT_TM_DEQ_BSZ "=<int> "
PMD_PARAM_HARD_NAME "=<string> "
PMD_PARAM_HARD_TX_QUEUE_ID "=<int>");