numam-dpdk/drivers/net/ark/ark_ethdev.c
Ferruh Yigit cbfc6111b5 ethdev: move inline device operations
This patch is a preparation to hide the 'struct eth_dev_ops' from
applications by moving some device operations from 'struct eth_dev_ops'
to 'struct rte_eth_dev'.

Mentioned ethdev APIs are in the data path and implemented as inline
because of performance reasons.

Exposing 'struct eth_dev_ops' to applications is bad because it is a
contract between ethdev and PMDs, not really needs to be known by
applications, also changes in the struct causing ABI breakages which
shouldn't.

To be able to both keep APIs inline and hide the 'struct eth_dev_ops',
moving device operations used in ethdev inline APIs to 'struct
rte_eth_dev' to the same level with Rx/Tx burst functions.

The list of dev_ops moved:
eth_rx_queue_count_t       rx_queue_count;
eth_rx_descriptor_done_t   rx_descriptor_done;
eth_rx_descriptor_status_t rx_descriptor_status;
eth_tx_descriptor_status_t tx_descriptor_status;

Signed-off-by: Ferruh Yigit <ferruh.yigit@intel.com>
Reviewed-by: Andrew Rybchenko <arybchenko@solarflare.com>
Acked-by: David Marchand <david.marchand@redhat.com>
Acked-by: Sachin Saxena <sachin.saxena@nxp.com>
2020-09-18 18:55:08 +02:00

1015 lines
26 KiB
C

/* SPDX-License-Identifier: BSD-3-Clause
* Copyright (c) 2015-2018 Atomic Rules LLC
*/
#include <unistd.h>
#include <sys/stat.h>
#include <dlfcn.h>
#include <rte_bus_pci.h>
#include <rte_ethdev_pci.h>
#include <rte_kvargs.h>
#include "ark_global.h"
#include "ark_logs.h"
#include "ark_ethdev_tx.h"
#include "ark_ethdev_rx.h"
#include "ark_mpu.h"
#include "ark_ddm.h"
#include "ark_udm.h"
#include "ark_rqp.h"
#include "ark_pktdir.h"
#include "ark_pktgen.h"
#include "ark_pktchkr.h"
/* Internal prototypes */
static int eth_ark_check_args(struct ark_adapter *ark, const char *params);
static int eth_ark_dev_init(struct rte_eth_dev *dev);
static int ark_config_device(struct rte_eth_dev *dev);
static int eth_ark_dev_uninit(struct rte_eth_dev *eth_dev);
static int eth_ark_dev_configure(struct rte_eth_dev *dev);
static int eth_ark_dev_start(struct rte_eth_dev *dev);
static void eth_ark_dev_stop(struct rte_eth_dev *dev);
static void eth_ark_dev_close(struct rte_eth_dev *dev);
static int eth_ark_dev_info_get(struct rte_eth_dev *dev,
struct rte_eth_dev_info *dev_info);
static int eth_ark_dev_link_update(struct rte_eth_dev *dev,
int wait_to_complete);
static int eth_ark_dev_set_link_up(struct rte_eth_dev *dev);
static int eth_ark_dev_set_link_down(struct rte_eth_dev *dev);
static int eth_ark_dev_stats_get(struct rte_eth_dev *dev,
struct rte_eth_stats *stats);
static int eth_ark_dev_stats_reset(struct rte_eth_dev *dev);
static int eth_ark_set_default_mac_addr(struct rte_eth_dev *dev,
struct rte_ether_addr *mac_addr);
static int eth_ark_macaddr_add(struct rte_eth_dev *dev,
struct rte_ether_addr *mac_addr,
uint32_t index,
uint32_t pool);
static void eth_ark_macaddr_remove(struct rte_eth_dev *dev,
uint32_t index);
static int eth_ark_set_mtu(struct rte_eth_dev *dev, uint16_t size);
/*
* The packet generator is a functional block used to generate packet
* patterns for testing. It is not intended for nominal use.
*/
#define ARK_PKTGEN_ARG "Pkt_gen"
/*
* The packet checker is a functional block used to verify packet
* patterns for testing. It is not intended for nominal use.
*/
#define ARK_PKTCHKR_ARG "Pkt_chkr"
/*
* The packet director is used to select the internal ingress and
* egress packets paths during testing. It is not intended for
* nominal use.
*/
#define ARK_PKTDIR_ARG "Pkt_dir"
/* Devinfo configurations */
#define ARK_RX_MAX_QUEUE (4096 * 4)
#define ARK_RX_MIN_QUEUE (512)
#define ARK_RX_MAX_PKT_LEN ((16 * 1024) - 128)
#define ARK_RX_MIN_BUFSIZE (1024)
#define ARK_TX_MAX_QUEUE (4096 * 4)
#define ARK_TX_MIN_QUEUE (256)
static const char * const valid_arguments[] = {
ARK_PKTGEN_ARG,
ARK_PKTCHKR_ARG,
ARK_PKTDIR_ARG,
NULL
};
static const struct rte_pci_id pci_id_ark_map[] = {
{RTE_PCI_DEVICE(0x1d6c, 0x100d)},
{RTE_PCI_DEVICE(0x1d6c, 0x100e)},
{.vendor_id = 0, /* sentinel */ },
};
static int
eth_ark_pci_probe(struct rte_pci_driver *pci_drv __rte_unused,
struct rte_pci_device *pci_dev)
{
struct rte_eth_dev *eth_dev;
int ret;
eth_dev = rte_eth_dev_pci_allocate(pci_dev, sizeof(struct ark_adapter));
if (eth_dev == NULL)
return -ENOMEM;
ret = eth_ark_dev_init(eth_dev);
if (ret)
rte_eth_dev_pci_release(eth_dev);
return ret;
}
static int
eth_ark_pci_remove(struct rte_pci_device *pci_dev)
{
return rte_eth_dev_pci_generic_remove(pci_dev, eth_ark_dev_uninit);
}
static struct rte_pci_driver rte_ark_pmd = {
.id_table = pci_id_ark_map,
.drv_flags = RTE_PCI_DRV_NEED_MAPPING | RTE_PCI_DRV_INTR_LSC,
.probe = eth_ark_pci_probe,
.remove = eth_ark_pci_remove,
};
static const struct eth_dev_ops ark_eth_dev_ops = {
.dev_configure = eth_ark_dev_configure,
.dev_start = eth_ark_dev_start,
.dev_stop = eth_ark_dev_stop,
.dev_close = eth_ark_dev_close,
.dev_infos_get = eth_ark_dev_info_get,
.rx_queue_setup = eth_ark_dev_rx_queue_setup,
.tx_queue_setup = eth_ark_tx_queue_setup,
.link_update = eth_ark_dev_link_update,
.dev_set_link_up = eth_ark_dev_set_link_up,
.dev_set_link_down = eth_ark_dev_set_link_down,
.rx_queue_start = eth_ark_rx_start_queue,
.rx_queue_stop = eth_ark_rx_stop_queue,
.tx_queue_start = eth_ark_tx_queue_start,
.tx_queue_stop = eth_ark_tx_queue_stop,
.stats_get = eth_ark_dev_stats_get,
.stats_reset = eth_ark_dev_stats_reset,
.mac_addr_add = eth_ark_macaddr_add,
.mac_addr_remove = eth_ark_macaddr_remove,
.mac_addr_set = eth_ark_set_default_mac_addr,
.mtu_set = eth_ark_set_mtu,
};
static int
check_for_ext(struct ark_adapter *ark)
{
int found = 0;
/* Get the env */
const char *dllpath = getenv("ARK_EXT_PATH");
if (dllpath == NULL) {
ARK_PMD_LOG(DEBUG, "EXT NO dll path specified\n");
return 0;
}
ARK_PMD_LOG(NOTICE, "EXT found dll path at %s\n", dllpath);
/* Open and load the .so */
ark->d_handle = dlopen(dllpath, RTLD_LOCAL | RTLD_LAZY);
if (ark->d_handle == NULL) {
ARK_PMD_LOG(ERR, "Could not load user extension %s\n",
dllpath);
return -1;
}
ARK_PMD_LOG(DEBUG, "SUCCESS: loaded user extension %s\n",
dllpath);
/* Get the entry points */
ark->user_ext.dev_init =
(void *(*)(struct rte_eth_dev *, void *, int))
dlsym(ark->d_handle, "dev_init");
ARK_PMD_LOG(DEBUG, "device ext init pointer = %p\n",
ark->user_ext.dev_init);
ark->user_ext.dev_get_port_count =
(int (*)(struct rte_eth_dev *, void *))
dlsym(ark->d_handle, "dev_get_port_count");
ark->user_ext.dev_uninit =
(void (*)(struct rte_eth_dev *, void *))
dlsym(ark->d_handle, "dev_uninit");
ark->user_ext.dev_configure =
(int (*)(struct rte_eth_dev *, void *))
dlsym(ark->d_handle, "dev_configure");
ark->user_ext.dev_start =
(int (*)(struct rte_eth_dev *, void *))
dlsym(ark->d_handle, "dev_start");
ark->user_ext.dev_stop =
(void (*)(struct rte_eth_dev *, void *))
dlsym(ark->d_handle, "dev_stop");
ark->user_ext.dev_close =
(void (*)(struct rte_eth_dev *, void *))
dlsym(ark->d_handle, "dev_close");
ark->user_ext.link_update =
(int (*)(struct rte_eth_dev *, int, void *))
dlsym(ark->d_handle, "link_update");
ark->user_ext.dev_set_link_up =
(int (*)(struct rte_eth_dev *, void *))
dlsym(ark->d_handle, "dev_set_link_up");
ark->user_ext.dev_set_link_down =
(int (*)(struct rte_eth_dev *, void *))
dlsym(ark->d_handle, "dev_set_link_down");
ark->user_ext.stats_get =
(int (*)(struct rte_eth_dev *, struct rte_eth_stats *,
void *))
dlsym(ark->d_handle, "stats_get");
ark->user_ext.stats_reset =
(void (*)(struct rte_eth_dev *, void *))
dlsym(ark->d_handle, "stats_reset");
ark->user_ext.mac_addr_add =
(void (*)(struct rte_eth_dev *, struct rte_ether_addr *,
uint32_t, uint32_t, void *))
dlsym(ark->d_handle, "mac_addr_add");
ark->user_ext.mac_addr_remove =
(void (*)(struct rte_eth_dev *, uint32_t, void *))
dlsym(ark->d_handle, "mac_addr_remove");
ark->user_ext.mac_addr_set =
(void (*)(struct rte_eth_dev *, struct rte_ether_addr *,
void *))
dlsym(ark->d_handle, "mac_addr_set");
ark->user_ext.set_mtu =
(int (*)(struct rte_eth_dev *, uint16_t,
void *))
dlsym(ark->d_handle, "set_mtu");
return found;
}
static int
eth_ark_dev_init(struct rte_eth_dev *dev)
{
struct ark_adapter *ark = dev->data->dev_private;
struct rte_pci_device *pci_dev;
int ret;
int port_count = 1;
int p;
ark->eth_dev = dev;
ARK_PMD_LOG(DEBUG, "\n");
/* Check to see if there is an extension that we need to load */
ret = check_for_ext(ark);
if (ret)
return ret;
pci_dev = RTE_ETH_DEV_TO_PCI(dev);
rte_eth_copy_pci_info(dev, pci_dev);
/* Use dummy function until setup */
dev->rx_pkt_burst = &eth_ark_recv_pkts_noop;
dev->tx_pkt_burst = &eth_ark_xmit_pkts_noop;
/* Let rte_eth_dev_close() release the port resources */
dev->data->dev_flags |= RTE_ETH_DEV_CLOSE_REMOVE;
ark->bar0 = (uint8_t *)pci_dev->mem_resource[0].addr;
ark->a_bar = (uint8_t *)pci_dev->mem_resource[2].addr;
ark->sysctrl.v = (void *)&ark->bar0[ARK_SYSCTRL_BASE];
ark->mpurx.v = (void *)&ark->bar0[ARK_MPU_RX_BASE];
ark->udm.v = (void *)&ark->bar0[ARK_UDM_BASE];
ark->mputx.v = (void *)&ark->bar0[ARK_MPU_TX_BASE];
ark->ddm.v = (void *)&ark->bar0[ARK_DDM_BASE];
ark->cmac.v = (void *)&ark->bar0[ARK_CMAC_BASE];
ark->external.v = (void *)&ark->bar0[ARK_EXTERNAL_BASE];
ark->pktdir.v = (void *)&ark->bar0[ARK_PKTDIR_BASE];
ark->pktgen.v = (void *)&ark->bar0[ARK_PKTGEN_BASE];
ark->pktchkr.v = (void *)&ark->bar0[ARK_PKTCHKR_BASE];
ark->rqpacing =
(struct ark_rqpace_t *)(ark->bar0 + ARK_RCPACING_BASE);
ark->started = 0;
ARK_PMD_LOG(INFO, "Sys Ctrl Const = 0x%x HW Commit_ID: %08x\n",
ark->sysctrl.t32[4],
rte_be_to_cpu_32(ark->sysctrl.t32[0x20 / 4]));
ARK_PMD_LOG(NOTICE, "Arkville HW Commit_ID: %08x\n",
rte_be_to_cpu_32(ark->sysctrl.t32[0x20 / 4]));
/* If HW sanity test fails, return an error */
if (ark->sysctrl.t32[4] != 0xcafef00d) {
ARK_PMD_LOG(ERR,
"HW Sanity test has failed, expected constant"
" 0x%x, read 0x%x (%s)\n",
0xcafef00d,
ark->sysctrl.t32[4], __func__);
return -1;
}
if (ark->sysctrl.t32[3] != 0) {
if (ark_rqp_lasped(ark->rqpacing)) {
ARK_PMD_LOG(ERR, "Arkville Evaluation System - "
"Timer has Expired\n");
return -1;
}
ARK_PMD_LOG(WARNING, "Arkville Evaluation System - "
"Timer is Running\n");
}
ARK_PMD_LOG(DEBUG,
"HW Sanity test has PASSED, expected constant"
" 0x%x, read 0x%x (%s)\n",
0xcafef00d, ark->sysctrl.t32[4], __func__);
/* We are a single function multi-port device. */
ret = ark_config_device(dev);
if (ret)
return -1;
dev->dev_ops = &ark_eth_dev_ops;
dev->rx_queue_count = eth_ark_dev_rx_queue_count;
dev->data->mac_addrs = rte_zmalloc("ark", RTE_ETHER_ADDR_LEN, 0);
if (!dev->data->mac_addrs) {
ARK_PMD_LOG(ERR,
"Failed to allocated memory for storing mac address"
);
}
if (ark->user_ext.dev_init) {
ark->user_data[dev->data->port_id] =
ark->user_ext.dev_init(dev, ark->a_bar, 0);
if (!ark->user_data[dev->data->port_id]) {
ARK_PMD_LOG(WARNING,
"Failed to initialize PMD extension!"
" continuing without it\n");
memset(&ark->user_ext, 0, sizeof(struct ark_user_ext));
dlclose(ark->d_handle);
}
}
if (pci_dev->device.devargs)
ret = eth_ark_check_args(ark, pci_dev->device.devargs->args);
else
ARK_PMD_LOG(INFO, "No Device args found\n");
if (ret)
goto error;
/*
* We will create additional devices based on the number of requested
* ports
*/
if (ark->user_ext.dev_get_port_count)
port_count =
ark->user_ext.dev_get_port_count(dev,
ark->user_data[dev->data->port_id]);
ark->num_ports = port_count;
for (p = 0; p < port_count; p++) {
struct rte_eth_dev *eth_dev;
char name[RTE_ETH_NAME_MAX_LEN];
snprintf(name, sizeof(name), "arketh%d",
dev->data->port_id + p);
if (p == 0) {
/* First port is already allocated by DPDK */
eth_dev = ark->eth_dev;
rte_eth_dev_probing_finish(eth_dev);
continue;
}
/* reserve an ethdev entry */
eth_dev = rte_eth_dev_allocate(name);
if (!eth_dev) {
ARK_PMD_LOG(ERR,
"Could not allocate eth_dev for port %d\n",
p);
goto error;
}
eth_dev->device = &pci_dev->device;
eth_dev->data->dev_private = ark;
eth_dev->dev_ops = ark->eth_dev->dev_ops;
eth_dev->tx_pkt_burst = ark->eth_dev->tx_pkt_burst;
eth_dev->rx_pkt_burst = ark->eth_dev->rx_pkt_burst;
rte_eth_copy_pci_info(eth_dev, pci_dev);
eth_dev->data->mac_addrs = rte_zmalloc(name,
RTE_ETHER_ADDR_LEN, 0);
if (!eth_dev->data->mac_addrs) {
ARK_PMD_LOG(ERR,
"Memory allocation for MAC failed!"
" Exiting.\n");
goto error;
}
if (ark->user_ext.dev_init) {
ark->user_data[eth_dev->data->port_id] =
ark->user_ext.dev_init(dev, ark->a_bar, p);
}
rte_eth_dev_probing_finish(eth_dev);
}
return ret;
error:
rte_free(dev->data->mac_addrs);
dev->data->mac_addrs = NULL;
return -1;
}
/*
*Initial device configuration when device is opened
* setup the DDM, and UDM
* Called once per PCIE device
*/
static int
ark_config_device(struct rte_eth_dev *dev)
{
struct ark_adapter *ark = dev->data->dev_private;
uint16_t num_q, i;
struct ark_mpu_t *mpu;
/*
* Make sure that the packet director, generator and checker are in a
* known state
*/
ark->start_pg = 0;
ark->pg = ark_pktgen_init(ark->pktgen.v, 0, 1);
if (ark->pg == NULL)
return -1;
ark_pktgen_reset(ark->pg);
ark->pc = ark_pktchkr_init(ark->pktchkr.v, 0, 1);
if (ark->pc == NULL)
return -1;
ark_pktchkr_stop(ark->pc);
ark->pd = ark_pktdir_init(ark->pktdir.v);
if (ark->pd == NULL)
return -1;
/* Verify HW */
if (ark_udm_verify(ark->udm.v))
return -1;
if (ark_ddm_verify(ark->ddm.v))
return -1;
/* UDM */
if (ark_udm_reset(ark->udm.v)) {
ARK_PMD_LOG(ERR, "Unable to stop and reset UDM\n");
return -1;
}
/* Keep in reset until the MPU are cleared */
/* MPU reset */
mpu = ark->mpurx.v;
num_q = ark_api_num_queues(mpu);
ark->rx_queues = num_q;
for (i = 0; i < num_q; i++) {
ark_mpu_reset(mpu);
mpu = RTE_PTR_ADD(mpu, ARK_MPU_QOFFSET);
}
ark_udm_stop(ark->udm.v, 0);
ark_udm_configure(ark->udm.v,
RTE_PKTMBUF_HEADROOM,
RTE_MBUF_DEFAULT_DATAROOM,
ARK_RX_WRITE_TIME_NS);
ark_udm_stats_reset(ark->udm.v);
ark_udm_stop(ark->udm.v, 0);
/* TX -- DDM */
if (ark_ddm_stop(ark->ddm.v, 1))
ARK_PMD_LOG(ERR, "Unable to stop DDM\n");
mpu = ark->mputx.v;
num_q = ark_api_num_queues(mpu);
ark->tx_queues = num_q;
for (i = 0; i < num_q; i++) {
ark_mpu_reset(mpu);
mpu = RTE_PTR_ADD(mpu, ARK_MPU_QOFFSET);
}
ark_ddm_reset(ark->ddm.v);
ark_ddm_stats_reset(ark->ddm.v);
ark_ddm_stop(ark->ddm.v, 0);
ark_rqp_stats_reset(ark->rqpacing);
return 0;
}
static int
eth_ark_dev_uninit(struct rte_eth_dev *dev)
{
struct ark_adapter *ark = dev->data->dev_private;
if (rte_eal_process_type() != RTE_PROC_PRIMARY)
return 0;
if (ark->user_ext.dev_uninit)
ark->user_ext.dev_uninit(dev,
ark->user_data[dev->data->port_id]);
ark_pktgen_uninit(ark->pg);
ark_pktchkr_uninit(ark->pc);
dev->dev_ops = NULL;
dev->rx_pkt_burst = NULL;
dev->tx_pkt_burst = NULL;
return 0;
}
static int
eth_ark_dev_configure(struct rte_eth_dev *dev)
{
struct ark_adapter *ark = dev->data->dev_private;
eth_ark_dev_set_link_up(dev);
if (ark->user_ext.dev_configure)
return ark->user_ext.dev_configure(dev,
ark->user_data[dev->data->port_id]);
return 0;
}
static void *
delay_pg_start(void *arg)
{
struct ark_adapter *ark = (struct ark_adapter *)arg;
/* This function is used exclusively for regression testing, We
* perform a blind sleep here to ensure that the external test
* application has time to setup the test before we generate packets
*/
usleep(100000);
ark_pktgen_run(ark->pg);
return NULL;
}
static int
eth_ark_dev_start(struct rte_eth_dev *dev)
{
struct ark_adapter *ark = dev->data->dev_private;
int i;
/* RX Side */
/* start UDM */
ark_udm_start(ark->udm.v);
for (i = 0; i < dev->data->nb_rx_queues; i++)
eth_ark_rx_start_queue(dev, i);
/* TX Side */
for (i = 0; i < dev->data->nb_tx_queues; i++)
eth_ark_tx_queue_start(dev, i);
/* start DDM */
ark_ddm_start(ark->ddm.v);
ark->started = 1;
/* set xmit and receive function */
dev->rx_pkt_burst = &eth_ark_recv_pkts;
dev->tx_pkt_burst = &eth_ark_xmit_pkts;
if (ark->start_pg)
ark_pktchkr_run(ark->pc);
if (ark->start_pg && (dev->data->port_id == 0)) {
pthread_t thread;
/* Delay packet generatpr start allow the hardware to be ready
* This is only used for sanity checking with internal generator
*/
if (pthread_create(&thread, NULL, delay_pg_start, ark)) {
ARK_PMD_LOG(ERR, "Could not create pktgen "
"starter thread\n");
return -1;
}
}
if (ark->user_ext.dev_start)
ark->user_ext.dev_start(dev,
ark->user_data[dev->data->port_id]);
return 0;
}
static void
eth_ark_dev_stop(struct rte_eth_dev *dev)
{
uint16_t i;
int status;
struct ark_adapter *ark = dev->data->dev_private;
struct ark_mpu_t *mpu;
if (ark->started == 0)
return;
ark->started = 0;
/* Stop the extension first */
if (ark->user_ext.dev_stop)
ark->user_ext.dev_stop(dev,
ark->user_data[dev->data->port_id]);
/* Stop the packet generator */
if (ark->start_pg)
ark_pktgen_pause(ark->pg);
dev->rx_pkt_burst = &eth_ark_recv_pkts_noop;
dev->tx_pkt_burst = &eth_ark_xmit_pkts_noop;
/* STOP TX Side */
for (i = 0; i < dev->data->nb_tx_queues; i++) {
status = eth_ark_tx_queue_stop(dev, i);
if (status != 0) {
uint16_t port = dev->data->port_id;
ARK_PMD_LOG(ERR,
"tx_queue stop anomaly"
" port %u, queue %u\n",
port, i);
}
}
/* Stop DDM */
/* Wait up to 0.1 second. each stop is up to 1000 * 10 useconds */
for (i = 0; i < 10; i++) {
status = ark_ddm_stop(ark->ddm.v, 1);
if (status == 0)
break;
}
if (status || i != 0) {
ARK_PMD_LOG(ERR, "DDM stop anomaly. status:"
" %d iter: %u. (%s)\n",
status,
i,
__func__);
ark_ddm_dump(ark->ddm.v, "Stop anomaly");
mpu = ark->mputx.v;
for (i = 0; i < ark->tx_queues; i++) {
ark_mpu_dump(mpu, "DDM failure dump", i);
mpu = RTE_PTR_ADD(mpu, ARK_MPU_QOFFSET);
}
}
/* STOP RX Side */
/* Stop UDM multiple tries attempted */
for (i = 0; i < 10; i++) {
status = ark_udm_stop(ark->udm.v, 1);
if (status == 0)
break;
}
if (status || i != 0) {
ARK_PMD_LOG(ERR, "UDM stop anomaly. status %d iter: %u. (%s)\n",
status, i, __func__);
ark_udm_dump(ark->udm.v, "Stop anomaly");
mpu = ark->mpurx.v;
for (i = 0; i < ark->rx_queues; i++) {
ark_mpu_dump(mpu, "UDM Stop anomaly", i);
mpu = RTE_PTR_ADD(mpu, ARK_MPU_QOFFSET);
}
}
ark_udm_dump_stats(ark->udm.v, "Post stop");
ark_udm_dump_perf(ark->udm.v, "Post stop");
for (i = 0; i < dev->data->nb_rx_queues; i++)
eth_ark_rx_dump_queue(dev, i, __func__);
/* Stop the packet checker if it is running */
if (ark->start_pg) {
ark_pktchkr_dump_stats(ark->pc);
ark_pktchkr_stop(ark->pc);
}
}
static void
eth_ark_dev_close(struct rte_eth_dev *dev)
{
struct ark_adapter *ark = dev->data->dev_private;
uint16_t i;
if (ark->user_ext.dev_close)
ark->user_ext.dev_close(dev,
ark->user_data[dev->data->port_id]);
eth_ark_dev_stop(dev);
eth_ark_udm_force_close(dev);
/*
* TODO This should only be called once for the device during shutdown
*/
ark_rqp_dump(ark->rqpacing);
for (i = 0; i < dev->data->nb_tx_queues; i++) {
eth_ark_tx_queue_release(dev->data->tx_queues[i]);
dev->data->tx_queues[i] = 0;
}
for (i = 0; i < dev->data->nb_rx_queues; i++) {
eth_ark_dev_rx_queue_release(dev->data->rx_queues[i]);
dev->data->rx_queues[i] = 0;
}
rte_free(dev->data->mac_addrs);
dev->data->mac_addrs = 0;
}
static int
eth_ark_dev_info_get(struct rte_eth_dev *dev,
struct rte_eth_dev_info *dev_info)
{
struct ark_adapter *ark = dev->data->dev_private;
struct ark_mpu_t *tx_mpu = RTE_PTR_ADD(ark->bar0, ARK_MPU_TX_BASE);
struct ark_mpu_t *rx_mpu = RTE_PTR_ADD(ark->bar0, ARK_MPU_RX_BASE);
uint16_t ports = ark->num_ports;
dev_info->max_rx_pktlen = ARK_RX_MAX_PKT_LEN;
dev_info->min_rx_bufsize = ARK_RX_MIN_BUFSIZE;
dev_info->max_rx_queues = ark_api_num_queues_per_port(rx_mpu, ports);
dev_info->max_tx_queues = ark_api_num_queues_per_port(tx_mpu, ports);
dev_info->rx_desc_lim = (struct rte_eth_desc_lim) {
.nb_max = ARK_RX_MAX_QUEUE,
.nb_min = ARK_RX_MIN_QUEUE,
.nb_align = ARK_RX_MIN_QUEUE}; /* power of 2 */
dev_info->tx_desc_lim = (struct rte_eth_desc_lim) {
.nb_max = ARK_TX_MAX_QUEUE,
.nb_min = ARK_TX_MIN_QUEUE,
.nb_align = ARK_TX_MIN_QUEUE}; /* power of 2 */
/* ARK PMD supports all line rates, how do we indicate that here ?? */
dev_info->speed_capa = (ETH_LINK_SPEED_1G |
ETH_LINK_SPEED_10G |
ETH_LINK_SPEED_25G |
ETH_LINK_SPEED_40G |
ETH_LINK_SPEED_50G |
ETH_LINK_SPEED_100G);
return 0;
}
static int
eth_ark_dev_link_update(struct rte_eth_dev *dev, int wait_to_complete)
{
ARK_PMD_LOG(DEBUG, "link status = %d\n",
dev->data->dev_link.link_status);
struct ark_adapter *ark = dev->data->dev_private;
if (ark->user_ext.link_update) {
return ark->user_ext.link_update
(dev, wait_to_complete,
ark->user_data[dev->data->port_id]);
}
return 0;
}
static int
eth_ark_dev_set_link_up(struct rte_eth_dev *dev)
{
dev->data->dev_link.link_status = 1;
struct ark_adapter *ark = dev->data->dev_private;
if (ark->user_ext.dev_set_link_up)
return ark->user_ext.dev_set_link_up(dev,
ark->user_data[dev->data->port_id]);
return 0;
}
static int
eth_ark_dev_set_link_down(struct rte_eth_dev *dev)
{
dev->data->dev_link.link_status = 0;
struct ark_adapter *ark = dev->data->dev_private;
if (ark->user_ext.dev_set_link_down)
return ark->user_ext.dev_set_link_down(dev,
ark->user_data[dev->data->port_id]);
return 0;
}
static int
eth_ark_dev_stats_get(struct rte_eth_dev *dev, struct rte_eth_stats *stats)
{
uint16_t i;
struct ark_adapter *ark = dev->data->dev_private;
stats->ipackets = 0;
stats->ibytes = 0;
stats->opackets = 0;
stats->obytes = 0;
stats->imissed = 0;
stats->oerrors = 0;
for (i = 0; i < dev->data->nb_tx_queues; i++)
eth_tx_queue_stats_get(dev->data->tx_queues[i], stats);
for (i = 0; i < dev->data->nb_rx_queues; i++)
eth_rx_queue_stats_get(dev->data->rx_queues[i], stats);
if (ark->user_ext.stats_get)
return ark->user_ext.stats_get(dev, stats,
ark->user_data[dev->data->port_id]);
return 0;
}
static int
eth_ark_dev_stats_reset(struct rte_eth_dev *dev)
{
uint16_t i;
struct ark_adapter *ark = dev->data->dev_private;
for (i = 0; i < dev->data->nb_tx_queues; i++)
eth_tx_queue_stats_reset(dev->data->tx_queues[i]);
for (i = 0; i < dev->data->nb_rx_queues; i++)
eth_rx_queue_stats_reset(dev->data->rx_queues[i]);
if (ark->user_ext.stats_reset)
ark->user_ext.stats_reset(dev,
ark->user_data[dev->data->port_id]);
return 0;
}
static int
eth_ark_macaddr_add(struct rte_eth_dev *dev,
struct rte_ether_addr *mac_addr,
uint32_t index,
uint32_t pool)
{
struct ark_adapter *ark = dev->data->dev_private;
if (ark->user_ext.mac_addr_add) {
ark->user_ext.mac_addr_add(dev,
mac_addr,
index,
pool,
ark->user_data[dev->data->port_id]);
return 0;
}
return -ENOTSUP;
}
static void
eth_ark_macaddr_remove(struct rte_eth_dev *dev, uint32_t index)
{
struct ark_adapter *ark = dev->data->dev_private;
if (ark->user_ext.mac_addr_remove)
ark->user_ext.mac_addr_remove(dev, index,
ark->user_data[dev->data->port_id]);
}
static int
eth_ark_set_default_mac_addr(struct rte_eth_dev *dev,
struct rte_ether_addr *mac_addr)
{
struct ark_adapter *ark = dev->data->dev_private;
if (ark->user_ext.mac_addr_set) {
ark->user_ext.mac_addr_set(dev, mac_addr,
ark->user_data[dev->data->port_id]);
return 0;
}
return -ENOTSUP;
}
static int
eth_ark_set_mtu(struct rte_eth_dev *dev, uint16_t size)
{
struct ark_adapter *ark = dev->data->dev_private;
if (ark->user_ext.set_mtu)
return ark->user_ext.set_mtu(dev, size,
ark->user_data[dev->data->port_id]);
return -ENOTSUP;
}
static inline int
process_pktdir_arg(const char *key, const char *value,
void *extra_args)
{
ARK_PMD_LOG(DEBUG, "key = %s, value = %s\n",
key, value);
struct ark_adapter *ark =
(struct ark_adapter *)extra_args;
ark->pkt_dir_v = strtol(value, NULL, 16);
ARK_PMD_LOG(DEBUG, "pkt_dir_v = 0x%x\n", ark->pkt_dir_v);
return 0;
}
static inline int
process_file_args(const char *key, const char *value, void *extra_args)
{
ARK_PMD_LOG(DEBUG, "key = %s, value = %s\n",
key, value);
char *args = (char *)extra_args;
/* Open the configuration file */
FILE *file = fopen(value, "r");
char line[ARK_MAX_ARG_LEN];
int size = 0;
int first = 1;
if (file == NULL) {
ARK_PMD_LOG(ERR, "Unable to open "
"config file %s\n", value);
return -1;
}
while (fgets(line, sizeof(line), file)) {
size += strlen(line);
if (size >= ARK_MAX_ARG_LEN) {
ARK_PMD_LOG(ERR, "Unable to parse file %s args, "
"parameter list is too long\n", value);
fclose(file);
return -1;
}
if (first) {
strncpy(args, line, ARK_MAX_ARG_LEN);
first = 0;
} else {
strncat(args, line, ARK_MAX_ARG_LEN);
}
}
ARK_PMD_LOG(DEBUG, "file = %s\n", args);
fclose(file);
return 0;
}
static int
eth_ark_check_args(struct ark_adapter *ark, const char *params)
{
struct rte_kvargs *kvlist;
unsigned int k_idx;
struct rte_kvargs_pair *pair = NULL;
int ret = -1;
kvlist = rte_kvargs_parse(params, valid_arguments);
if (kvlist == NULL)
return 0;
ark->pkt_gen_args[0] = 0;
ark->pkt_chkr_args[0] = 0;
for (k_idx = 0; k_idx < kvlist->count; k_idx++) {
pair = &kvlist->pairs[k_idx];
ARK_PMD_LOG(DEBUG, "**** Arg passed to PMD = %s:%s\n",
pair->key,
pair->value);
}
if (rte_kvargs_process(kvlist,
ARK_PKTDIR_ARG,
&process_pktdir_arg,
ark) != 0) {
ARK_PMD_LOG(ERR, "Unable to parse arg %s\n", ARK_PKTDIR_ARG);
goto free_kvlist;
}
if (rte_kvargs_process(kvlist,
ARK_PKTGEN_ARG,
&process_file_args,
ark->pkt_gen_args) != 0) {
ARK_PMD_LOG(ERR, "Unable to parse arg %s\n", ARK_PKTGEN_ARG);
goto free_kvlist;
}
if (rte_kvargs_process(kvlist,
ARK_PKTCHKR_ARG,
&process_file_args,
ark->pkt_chkr_args) != 0) {
ARK_PMD_LOG(ERR, "Unable to parse arg %s\n", ARK_PKTCHKR_ARG);
goto free_kvlist;
}
ARK_PMD_LOG(INFO, "packet director set to 0x%x\n", ark->pkt_dir_v);
/* Setup the packet director */
ark_pktdir_setup(ark->pd, ark->pkt_dir_v);
/* Setup the packet generator */
if (ark->pkt_gen_args[0]) {
ARK_PMD_LOG(DEBUG, "Setting up the packet generator\n");
ark_pktgen_parse(ark->pkt_gen_args);
ark_pktgen_reset(ark->pg);
ark_pktgen_setup(ark->pg);
ark->start_pg = 1;
}
/* Setup the packet checker */
if (ark->pkt_chkr_args[0]) {
ark_pktchkr_parse(ark->pkt_chkr_args);
ark_pktchkr_setup(ark->pc);
}
ret = 0;
free_kvlist:
rte_kvargs_free(kvlist);
return ret;
}
RTE_PMD_REGISTER_PCI(net_ark, rte_ark_pmd);
RTE_PMD_REGISTER_KMOD_DEP(net_ark, "* igb_uio | uio_pci_generic ");
RTE_PMD_REGISTER_PCI_TABLE(net_ark, pci_id_ark_map);
RTE_PMD_REGISTER_PARAM_STRING(net_ark,
ARK_PKTGEN_ARG "=<filename> "
ARK_PKTCHKR_ARG "=<filename> "
ARK_PKTDIR_ARG "=<bitmap>");
RTE_LOG_REGISTER(ark_logtype, pmd.net.ark, NOTICE);