numam-dpdk/drivers/net/ark/ark_ethdev.c
Matan Azrad d5b0924ba6 ethdev: add return value to stats get dev op
The stats_get dev op API doesn't include return value, so PMD cannot
return an error in case of failure at stats getting process time.

Since PCI devices can be removed and there is a time between the
physical removal to the RMV interrupt, the user may get invalid stats
without any indication.

This patch changes the stats_get API return value to be int instead of
void.

All the net PMDs stats_get dev ops are adjusted by this patch.

Signed-off-by: Matan Azrad <matan@mellanox.com>
Reviewed-by: Ferruh Yigit <ferruh.yigit@intel.com>
2017-10-12 01:52:49 +01:00

1048 lines
28 KiB
C

/*-
* BSD LICENSE
*
* Copyright (c) 2015-2017 Atomic Rules LLC
* All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
*
* * Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* * Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in
* the documentation and/or other materials provided with the
* distribution.
* * Neither the name of copyright holder nor the names of its
* contributors may be used to endorse or promote products derived
* from this software without specific prior written permission.
*
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
* OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*/
#include <unistd.h>
#include <sys/stat.h>
#include <dlfcn.h>
#include <rte_ethdev_pci.h>
#include <rte_kvargs.h>
#include "ark_global.h"
#include "ark_logs.h"
#include "ark_ethdev_tx.h"
#include "ark_ethdev_rx.h"
#include "ark_mpu.h"
#include "ark_ddm.h"
#include "ark_udm.h"
#include "ark_rqp.h"
#include "ark_pktdir.h"
#include "ark_pktgen.h"
#include "ark_pktchkr.h"
/* Internal prototypes */
static int eth_ark_check_args(struct ark_adapter *ark, const char *params);
static int eth_ark_dev_init(struct rte_eth_dev *dev);
static int ark_config_device(struct rte_eth_dev *dev);
static int eth_ark_dev_uninit(struct rte_eth_dev *eth_dev);
static int eth_ark_dev_configure(struct rte_eth_dev *dev);
static int eth_ark_dev_start(struct rte_eth_dev *dev);
static void eth_ark_dev_stop(struct rte_eth_dev *dev);
static void eth_ark_dev_close(struct rte_eth_dev *dev);
static void eth_ark_dev_info_get(struct rte_eth_dev *dev,
struct rte_eth_dev_info *dev_info);
static int eth_ark_dev_link_update(struct rte_eth_dev *dev,
int wait_to_complete);
static int eth_ark_dev_set_link_up(struct rte_eth_dev *dev);
static int eth_ark_dev_set_link_down(struct rte_eth_dev *dev);
static int eth_ark_dev_stats_get(struct rte_eth_dev *dev,
struct rte_eth_stats *stats);
static void eth_ark_dev_stats_reset(struct rte_eth_dev *dev);
static void eth_ark_set_default_mac_addr(struct rte_eth_dev *dev,
struct ether_addr *mac_addr);
static int eth_ark_macaddr_add(struct rte_eth_dev *dev,
struct ether_addr *mac_addr,
uint32_t index,
uint32_t pool);
static void eth_ark_macaddr_remove(struct rte_eth_dev *dev,
uint32_t index);
static int eth_ark_set_mtu(struct rte_eth_dev *dev, uint16_t size);
/*
* The packet generator is a functional block used to generate packet
* patterns for testing. It is not intended for nominal use.
*/
#define ARK_PKTGEN_ARG "Pkt_gen"
/*
* The packet checker is a functional block used to verify packet
* patterns for testing. It is not intended for nominal use.
*/
#define ARK_PKTCHKR_ARG "Pkt_chkr"
/*
* The packet director is used to select the internal ingress and
* egress packets paths during testing. It is not intended for
* nominal use.
*/
#define ARK_PKTDIR_ARG "Pkt_dir"
/* Devinfo configurations */
#define ARK_RX_MAX_QUEUE (4096 * 4)
#define ARK_RX_MIN_QUEUE (512)
#define ARK_RX_MAX_PKT_LEN ((16 * 1024) - 128)
#define ARK_RX_MIN_BUFSIZE (1024)
#define ARK_TX_MAX_QUEUE (4096 * 4)
#define ARK_TX_MIN_QUEUE (256)
static const char * const valid_arguments[] = {
ARK_PKTGEN_ARG,
ARK_PKTCHKR_ARG,
ARK_PKTDIR_ARG,
NULL
};
static const struct rte_pci_id pci_id_ark_map[] = {
{RTE_PCI_DEVICE(0x1d6c, 0x100d)},
{RTE_PCI_DEVICE(0x1d6c, 0x100e)},
{.vendor_id = 0, /* sentinel */ },
};
static int
eth_ark_pci_probe(struct rte_pci_driver *pci_drv __rte_unused,
struct rte_pci_device *pci_dev)
{
struct rte_eth_dev *eth_dev;
int ret;
eth_dev = rte_eth_dev_pci_allocate(pci_dev, sizeof(struct ark_adapter));
if (eth_dev == NULL)
return -ENOMEM;
ret = eth_ark_dev_init(eth_dev);
if (ret)
rte_eth_dev_pci_release(eth_dev);
return ret;
}
static int
eth_ark_pci_remove(struct rte_pci_device *pci_dev)
{
return rte_eth_dev_pci_generic_remove(pci_dev, eth_ark_dev_uninit);
}
static struct rte_pci_driver rte_ark_pmd = {
.id_table = pci_id_ark_map,
.drv_flags = RTE_PCI_DRV_NEED_MAPPING | RTE_PCI_DRV_INTR_LSC,
.probe = eth_ark_pci_probe,
.remove = eth_ark_pci_remove,
};
static const struct eth_dev_ops ark_eth_dev_ops = {
.dev_configure = eth_ark_dev_configure,
.dev_start = eth_ark_dev_start,
.dev_stop = eth_ark_dev_stop,
.dev_close = eth_ark_dev_close,
.dev_infos_get = eth_ark_dev_info_get,
.rx_queue_setup = eth_ark_dev_rx_queue_setup,
.rx_queue_count = eth_ark_dev_rx_queue_count,
.tx_queue_setup = eth_ark_tx_queue_setup,
.link_update = eth_ark_dev_link_update,
.dev_set_link_up = eth_ark_dev_set_link_up,
.dev_set_link_down = eth_ark_dev_set_link_down,
.rx_queue_start = eth_ark_rx_start_queue,
.rx_queue_stop = eth_ark_rx_stop_queue,
.tx_queue_start = eth_ark_tx_queue_start,
.tx_queue_stop = eth_ark_tx_queue_stop,
.stats_get = eth_ark_dev_stats_get,
.stats_reset = eth_ark_dev_stats_reset,
.mac_addr_add = eth_ark_macaddr_add,
.mac_addr_remove = eth_ark_macaddr_remove,
.mac_addr_set = eth_ark_set_default_mac_addr,
.mtu_set = eth_ark_set_mtu,
};
static int
check_for_ext(struct ark_adapter *ark)
{
int found = 0;
/* Get the env */
const char *dllpath = getenv("ARK_EXT_PATH");
if (dllpath == NULL) {
PMD_DEBUG_LOG(DEBUG, "ARK EXT NO dll path specified\n");
return 0;
}
PMD_DRV_LOG(INFO, "ARK EXT found dll path at %s\n", dllpath);
/* Open and load the .so */
ark->d_handle = dlopen(dllpath, RTLD_LOCAL | RTLD_LAZY);
if (ark->d_handle == NULL) {
PMD_DRV_LOG(ERR, "Could not load user extension %s\n",
dllpath);
return -1;
}
PMD_DRV_LOG(INFO, "SUCCESS: loaded user extension %s\n",
dllpath);
/* Get the entry points */
ark->user_ext.dev_init =
(void *(*)(struct rte_eth_dev *, void *, int))
dlsym(ark->d_handle, "dev_init");
PMD_DEBUG_LOG(DEBUG, "device ext init pointer = %p\n",
ark->user_ext.dev_init);
ark->user_ext.dev_get_port_count =
(int (*)(struct rte_eth_dev *, void *))
dlsym(ark->d_handle, "dev_get_port_count");
ark->user_ext.dev_uninit =
(void (*)(struct rte_eth_dev *, void *))
dlsym(ark->d_handle, "dev_uninit");
ark->user_ext.dev_configure =
(int (*)(struct rte_eth_dev *, void *))
dlsym(ark->d_handle, "dev_configure");
ark->user_ext.dev_start =
(int (*)(struct rte_eth_dev *, void *))
dlsym(ark->d_handle, "dev_start");
ark->user_ext.dev_stop =
(void (*)(struct rte_eth_dev *, void *))
dlsym(ark->d_handle, "dev_stop");
ark->user_ext.dev_close =
(void (*)(struct rte_eth_dev *, void *))
dlsym(ark->d_handle, "dev_close");
ark->user_ext.link_update =
(int (*)(struct rte_eth_dev *, int, void *))
dlsym(ark->d_handle, "link_update");
ark->user_ext.dev_set_link_up =
(int (*)(struct rte_eth_dev *, void *))
dlsym(ark->d_handle, "dev_set_link_up");
ark->user_ext.dev_set_link_down =
(int (*)(struct rte_eth_dev *, void *))
dlsym(ark->d_handle, "dev_set_link_down");
ark->user_ext.stats_get =
(int (*)(struct rte_eth_dev *, struct rte_eth_stats *,
void *))
dlsym(ark->d_handle, "stats_get");
ark->user_ext.stats_reset =
(void (*)(struct rte_eth_dev *, void *))
dlsym(ark->d_handle, "stats_reset");
ark->user_ext.mac_addr_add =
(void (*)(struct rte_eth_dev *, struct ether_addr *, uint32_t,
uint32_t, void *))
dlsym(ark->d_handle, "mac_addr_add");
ark->user_ext.mac_addr_remove =
(void (*)(struct rte_eth_dev *, uint32_t, void *))
dlsym(ark->d_handle, "mac_addr_remove");
ark->user_ext.mac_addr_set =
(void (*)(struct rte_eth_dev *, struct ether_addr *,
void *))
dlsym(ark->d_handle, "mac_addr_set");
ark->user_ext.set_mtu =
(int (*)(struct rte_eth_dev *, uint16_t,
void *))
dlsym(ark->d_handle, "set_mtu");
return found;
}
static int
eth_ark_dev_init(struct rte_eth_dev *dev)
{
struct ark_adapter *ark =
(struct ark_adapter *)dev->data->dev_private;
struct rte_pci_device *pci_dev;
int ret;
int port_count = 1;
int p;
ark->eth_dev = dev;
PMD_FUNC_LOG(DEBUG, "\n");
/* Check to see if there is an extension that we need to load */
ret = check_for_ext(ark);
if (ret)
return ret;
pci_dev = RTE_ETH_DEV_TO_PCI(dev);
rte_eth_copy_pci_info(dev, pci_dev);
/* Use dummy function until setup */
dev->rx_pkt_burst = &eth_ark_recv_pkts_noop;
dev->tx_pkt_burst = &eth_ark_xmit_pkts_noop;
ark->bar0 = (uint8_t *)pci_dev->mem_resource[0].addr;
ark->a_bar = (uint8_t *)pci_dev->mem_resource[2].addr;
ark->sysctrl.v = (void *)&ark->bar0[ARK_SYSCTRL_BASE];
ark->mpurx.v = (void *)&ark->bar0[ARK_MPU_RX_BASE];
ark->udm.v = (void *)&ark->bar0[ARK_UDM_BASE];
ark->mputx.v = (void *)&ark->bar0[ARK_MPU_TX_BASE];
ark->ddm.v = (void *)&ark->bar0[ARK_DDM_BASE];
ark->cmac.v = (void *)&ark->bar0[ARK_CMAC_BASE];
ark->external.v = (void *)&ark->bar0[ARK_EXTERNAL_BASE];
ark->pktdir.v = (void *)&ark->bar0[ARK_PKTDIR_BASE];
ark->pktgen.v = (void *)&ark->bar0[ARK_PKTGEN_BASE];
ark->pktchkr.v = (void *)&ark->bar0[ARK_PKTCHKR_BASE];
ark->rqpacing =
(struct ark_rqpace_t *)(ark->bar0 + ARK_RCPACING_BASE);
ark->started = 0;
PMD_DEBUG_LOG(INFO, "Sys Ctrl Const = 0x%x HW Commit_ID: %08x\n",
ark->sysctrl.t32[4],
rte_be_to_cpu_32(ark->sysctrl.t32[0x20 / 4]));
PMD_DRV_LOG(INFO, "Arkville HW Commit_ID: %08x\n",
rte_be_to_cpu_32(ark->sysctrl.t32[0x20 / 4]));
/* If HW sanity test fails, return an error */
if (ark->sysctrl.t32[4] != 0xcafef00d) {
PMD_DRV_LOG(ERR,
"HW Sanity test has failed, expected constant"
" 0x%x, read 0x%x (%s)\n",
0xcafef00d,
ark->sysctrl.t32[4], __func__);
return -1;
}
if (ark->sysctrl.t32[3] != 0) {
if (ark_rqp_lasped(ark->rqpacing)) {
PMD_DRV_LOG(ERR, "Arkville Evaluation System - "
"Timer has Expired\n");
return -1;
}
PMD_DRV_LOG(WARNING, "Arkville Evaluation System - "
"Timer is Running\n");
}
PMD_DRV_LOG(INFO,
"HW Sanity test has PASSED, expected constant"
" 0x%x, read 0x%x (%s)\n",
0xcafef00d, ark->sysctrl.t32[4], __func__);
/* We are a single function multi-port device. */
ret = ark_config_device(dev);
dev->dev_ops = &ark_eth_dev_ops;
dev->data->dev_flags |= RTE_ETH_DEV_DETACHABLE;
dev->data->mac_addrs = rte_zmalloc("ark", ETHER_ADDR_LEN, 0);
if (!dev->data->mac_addrs) {
PMD_DRV_LOG(ERR,
"Failed to allocated memory for storing mac address"
);
}
if (ark->user_ext.dev_init) {
ark->user_data[dev->data->port_id] =
ark->user_ext.dev_init(dev, ark->a_bar, 0);
if (!ark->user_data[dev->data->port_id]) {
PMD_DRV_LOG(INFO,
"Failed to initialize PMD extension!"
" continuing without it\n");
memset(&ark->user_ext, 0, sizeof(struct ark_user_ext));
dlclose(ark->d_handle);
}
}
if (pci_dev->device.devargs)
ret = eth_ark_check_args(ark, pci_dev->device.devargs->args);
else
PMD_DRV_LOG(INFO, "No Device args found\n");
if (ret)
goto error;
/*
* We will create additional devices based on the number of requested
* ports
*/
if (ark->user_ext.dev_get_port_count)
port_count =
ark->user_ext.dev_get_port_count(dev,
ark->user_data[dev->data->port_id]);
ark->num_ports = port_count;
for (p = 0; p < port_count; p++) {
struct rte_eth_dev *eth_dev;
char name[RTE_ETH_NAME_MAX_LEN];
snprintf(name, sizeof(name), "arketh%d",
dev->data->port_id + p);
if (p == 0) {
/* First port is already allocated by DPDK */
eth_dev = ark->eth_dev;
continue;
}
/* reserve an ethdev entry */
eth_dev = rte_eth_dev_allocate(name);
if (!eth_dev) {
PMD_DRV_LOG(ERR,
"Could not allocate eth_dev for port %d\n",
p);
goto error;
}
eth_dev->device = &pci_dev->device;
eth_dev->data->dev_private = ark;
eth_dev->dev_ops = ark->eth_dev->dev_ops;
eth_dev->tx_pkt_burst = ark->eth_dev->tx_pkt_burst;
eth_dev->rx_pkt_burst = ark->eth_dev->rx_pkt_burst;
rte_eth_copy_pci_info(eth_dev, pci_dev);
eth_dev->data->mac_addrs = rte_zmalloc(name, ETHER_ADDR_LEN, 0);
if (!eth_dev->data->mac_addrs) {
PMD_DRV_LOG(ERR,
"Memory allocation for MAC failed!"
" Exiting.\n");
goto error;
}
if (ark->user_ext.dev_init) {
ark->user_data[eth_dev->data->port_id] =
ark->user_ext.dev_init(dev, ark->a_bar, p);
}
}
return ret;
error:
if (dev->data->mac_addrs)
rte_free(dev->data->mac_addrs);
return -1;
}
/*
*Initial device configuration when device is opened
* setup the DDM, and UDM
* Called once per PCIE device
*/
static int
ark_config_device(struct rte_eth_dev *dev)
{
struct ark_adapter *ark =
(struct ark_adapter *)dev->data->dev_private;
uint16_t num_q, i;
struct ark_mpu_t *mpu;
/*
* Make sure that the packet director, generator and checker are in a
* known state
*/
ark->start_pg = 0;
ark->pg = ark_pktgen_init(ark->pktgen.v, 0, 1);
if (ark->pg == NULL)
return -1;
ark_pktgen_reset(ark->pg);
ark->pc = ark_pktchkr_init(ark->pktchkr.v, 0, 1);
if (ark->pc == NULL)
return -1;
ark_pktchkr_stop(ark->pc);
ark->pd = ark_pktdir_init(ark->pktdir.v);
if (ark->pd == NULL)
return -1;
/* Verify HW */
if (ark_udm_verify(ark->udm.v))
return -1;
if (ark_ddm_verify(ark->ddm.v))
return -1;
/* UDM */
if (ark_udm_reset(ark->udm.v)) {
PMD_DRV_LOG(ERR, "Unable to stop and reset UDM\n");
return -1;
}
/* Keep in reset until the MPU are cleared */
/* MPU reset */
mpu = ark->mpurx.v;
num_q = ark_api_num_queues(mpu);
ark->rx_queues = num_q;
for (i = 0; i < num_q; i++) {
ark_mpu_reset(mpu);
mpu = RTE_PTR_ADD(mpu, ARK_MPU_QOFFSET);
}
ark_udm_stop(ark->udm.v, 0);
ark_udm_configure(ark->udm.v,
RTE_PKTMBUF_HEADROOM,
RTE_MBUF_DEFAULT_DATAROOM,
ARK_RX_WRITE_TIME_NS);
ark_udm_stats_reset(ark->udm.v);
ark_udm_stop(ark->udm.v, 0);
/* TX -- DDM */
if (ark_ddm_stop(ark->ddm.v, 1))
PMD_DRV_LOG(ERR, "Unable to stop DDM\n");
mpu = ark->mputx.v;
num_q = ark_api_num_queues(mpu);
ark->tx_queues = num_q;
for (i = 0; i < num_q; i++) {
ark_mpu_reset(mpu);
mpu = RTE_PTR_ADD(mpu, ARK_MPU_QOFFSET);
}
ark_ddm_reset(ark->ddm.v);
ark_ddm_stats_reset(ark->ddm.v);
ark_ddm_stop(ark->ddm.v, 0);
ark_rqp_stats_reset(ark->rqpacing);
return 0;
}
static int
eth_ark_dev_uninit(struct rte_eth_dev *dev)
{
struct ark_adapter *ark =
(struct ark_adapter *)dev->data->dev_private;
if (rte_eal_process_type() != RTE_PROC_PRIMARY)
return 0;
if (ark->user_ext.dev_uninit)
ark->user_ext.dev_uninit(dev,
ark->user_data[dev->data->port_id]);
ark_pktgen_uninit(ark->pg);
ark_pktchkr_uninit(ark->pc);
dev->dev_ops = NULL;
dev->rx_pkt_burst = NULL;
dev->tx_pkt_burst = NULL;
rte_free(dev->data->mac_addrs);
return 0;
}
static int
eth_ark_dev_configure(struct rte_eth_dev *dev)
{
PMD_FUNC_LOG(DEBUG, "\n");
struct ark_adapter *ark =
(struct ark_adapter *)dev->data->dev_private;
eth_ark_dev_set_link_up(dev);
if (ark->user_ext.dev_configure)
return ark->user_ext.dev_configure(dev,
ark->user_data[dev->data->port_id]);
return 0;
}
static void *
delay_pg_start(void *arg)
{
struct ark_adapter *ark = (struct ark_adapter *)arg;
/* This function is used exclusively for regression testing, We
* perform a blind sleep here to ensure that the external test
* application has time to setup the test before we generate packets
*/
usleep(100000);
ark_pktgen_run(ark->pg);
return NULL;
}
static int
eth_ark_dev_start(struct rte_eth_dev *dev)
{
struct ark_adapter *ark =
(struct ark_adapter *)dev->data->dev_private;
int i;
PMD_FUNC_LOG(DEBUG, "\n");
/* RX Side */
/* start UDM */
ark_udm_start(ark->udm.v);
for (i = 0; i < dev->data->nb_rx_queues; i++)
eth_ark_rx_start_queue(dev, i);
/* TX Side */
for (i = 0; i < dev->data->nb_tx_queues; i++)
eth_ark_tx_queue_start(dev, i);
/* start DDM */
ark_ddm_start(ark->ddm.v);
ark->started = 1;
/* set xmit and receive function */
dev->rx_pkt_burst = &eth_ark_recv_pkts;
dev->tx_pkt_burst = &eth_ark_xmit_pkts;
if (ark->start_pg)
ark_pktchkr_run(ark->pc);
if (ark->start_pg && (dev->data->port_id == 0)) {
pthread_t thread;
/* Delay packet generatpr start allow the hardware to be ready
* This is only used for sanity checking with internal generator
*/
if (pthread_create(&thread, NULL, delay_pg_start, ark)) {
PMD_DRV_LOG(ERR, "Could not create pktgen "
"starter thread\n");
return -1;
}
}
if (ark->user_ext.dev_start)
ark->user_ext.dev_start(dev,
ark->user_data[dev->data->port_id]);
return 0;
}
static void
eth_ark_dev_stop(struct rte_eth_dev *dev)
{
uint16_t i;
int status;
struct ark_adapter *ark =
(struct ark_adapter *)dev->data->dev_private;
struct ark_mpu_t *mpu;
PMD_FUNC_LOG(DEBUG, "\n");
if (ark->started == 0)
return;
ark->started = 0;
/* Stop the extension first */
if (ark->user_ext.dev_stop)
ark->user_ext.dev_stop(dev,
ark->user_data[dev->data->port_id]);
/* Stop the packet generator */
if (ark->start_pg)
ark_pktgen_pause(ark->pg);
dev->rx_pkt_burst = &eth_ark_recv_pkts_noop;
dev->tx_pkt_burst = &eth_ark_xmit_pkts_noop;
/* STOP TX Side */
for (i = 0; i < dev->data->nb_tx_queues; i++) {
status = eth_ark_tx_queue_stop(dev, i);
if (status != 0) {
uint16_t port = dev->data->port_id;
PMD_DRV_LOG(ERR,
"tx_queue stop anomaly"
" port %u, queue %u\n",
port, i);
}
}
/* Stop DDM */
/* Wait up to 0.1 second. each stop is up to 1000 * 10 useconds */
for (i = 0; i < 10; i++) {
status = ark_ddm_stop(ark->ddm.v, 1);
if (status == 0)
break;
}
if (status || i != 0) {
PMD_DRV_LOG(ERR, "DDM stop anomaly. status:"
" %d iter: %u. (%s)\n",
status,
i,
__func__);
ark_ddm_dump(ark->ddm.v, "Stop anomaly");
mpu = ark->mputx.v;
for (i = 0; i < ark->tx_queues; i++) {
ark_mpu_dump(mpu, "DDM failure dump", i);
mpu = RTE_PTR_ADD(mpu, ARK_MPU_QOFFSET);
}
}
/* STOP RX Side */
/* Stop UDM multiple tries attempted */
for (i = 0; i < 10; i++) {
status = ark_udm_stop(ark->udm.v, 1);
if (status == 0)
break;
}
if (status || i != 0) {
PMD_DRV_LOG(ERR, "UDM stop anomaly. status %d iter: %u. (%s)\n",
status, i, __func__);
ark_udm_dump(ark->udm.v, "Stop anomaly");
mpu = ark->mpurx.v;
for (i = 0; i < ark->rx_queues; i++) {
ark_mpu_dump(mpu, "UDM Stop anomaly", i);
mpu = RTE_PTR_ADD(mpu, ARK_MPU_QOFFSET);
}
}
ark_udm_dump_stats(ark->udm.v, "Post stop");
ark_udm_dump_perf(ark->udm.v, "Post stop");
for (i = 0; i < dev->data->nb_rx_queues; i++)
eth_ark_rx_dump_queue(dev, i, __func__);
/* Stop the packet checker if it is running */
if (ark->start_pg) {
ark_pktchkr_dump_stats(ark->pc);
ark_pktchkr_stop(ark->pc);
}
}
static void
eth_ark_dev_close(struct rte_eth_dev *dev)
{
struct ark_adapter *ark =
(struct ark_adapter *)dev->data->dev_private;
uint16_t i;
if (ark->user_ext.dev_close)
ark->user_ext.dev_close(dev,
ark->user_data[dev->data->port_id]);
eth_ark_dev_stop(dev);
eth_ark_udm_force_close(dev);
/*
* TODO This should only be called once for the device during shutdown
*/
ark_rqp_dump(ark->rqpacing);
for (i = 0; i < dev->data->nb_tx_queues; i++) {
eth_ark_tx_queue_release(dev->data->tx_queues[i]);
dev->data->tx_queues[i] = 0;
}
for (i = 0; i < dev->data->nb_rx_queues; i++) {
eth_ark_dev_rx_queue_release(dev->data->rx_queues[i]);
dev->data->rx_queues[i] = 0;
}
}
static void
eth_ark_dev_info_get(struct rte_eth_dev *dev,
struct rte_eth_dev_info *dev_info)
{
struct ark_adapter *ark =
(struct ark_adapter *)dev->data->dev_private;
struct ark_mpu_t *tx_mpu = RTE_PTR_ADD(ark->bar0, ARK_MPU_TX_BASE);
struct ark_mpu_t *rx_mpu = RTE_PTR_ADD(ark->bar0, ARK_MPU_RX_BASE);
uint16_t ports = ark->num_ports;
dev_info->max_rx_pktlen = ARK_RX_MAX_PKT_LEN;
dev_info->min_rx_bufsize = ARK_RX_MIN_BUFSIZE;
dev_info->max_rx_queues = ark_api_num_queues_per_port(rx_mpu, ports);
dev_info->max_tx_queues = ark_api_num_queues_per_port(tx_mpu, ports);
dev_info->rx_desc_lim = (struct rte_eth_desc_lim) {
.nb_max = ARK_RX_MAX_QUEUE,
.nb_min = ARK_RX_MIN_QUEUE,
.nb_align = ARK_RX_MIN_QUEUE}; /* power of 2 */
dev_info->tx_desc_lim = (struct rte_eth_desc_lim) {
.nb_max = ARK_TX_MAX_QUEUE,
.nb_min = ARK_TX_MIN_QUEUE,
.nb_align = ARK_TX_MIN_QUEUE}; /* power of 2 */
/* ARK PMD supports all line rates, how do we indicate that here ?? */
dev_info->speed_capa = (ETH_LINK_SPEED_1G |
ETH_LINK_SPEED_10G |
ETH_LINK_SPEED_25G |
ETH_LINK_SPEED_40G |
ETH_LINK_SPEED_50G |
ETH_LINK_SPEED_100G);
dev_info->pci_dev = RTE_ETH_DEV_TO_PCI(dev);
}
static int
eth_ark_dev_link_update(struct rte_eth_dev *dev, int wait_to_complete)
{
PMD_DEBUG_LOG(DEBUG, "link status = %d\n",
dev->data->dev_link.link_status);
struct ark_adapter *ark =
(struct ark_adapter *)dev->data->dev_private;
if (ark->user_ext.link_update) {
return ark->user_ext.link_update
(dev, wait_to_complete,
ark->user_data[dev->data->port_id]);
}
return 0;
}
static int
eth_ark_dev_set_link_up(struct rte_eth_dev *dev)
{
dev->data->dev_link.link_status = 1;
struct ark_adapter *ark =
(struct ark_adapter *)dev->data->dev_private;
if (ark->user_ext.dev_set_link_up)
return ark->user_ext.dev_set_link_up(dev,
ark->user_data[dev->data->port_id]);
return 0;
}
static int
eth_ark_dev_set_link_down(struct rte_eth_dev *dev)
{
dev->data->dev_link.link_status = 0;
struct ark_adapter *ark =
(struct ark_adapter *)dev->data->dev_private;
if (ark->user_ext.dev_set_link_down)
return ark->user_ext.dev_set_link_down(dev,
ark->user_data[dev->data->port_id]);
return 0;
}
static int
eth_ark_dev_stats_get(struct rte_eth_dev *dev, struct rte_eth_stats *stats)
{
uint16_t i;
struct ark_adapter *ark =
(struct ark_adapter *)dev->data->dev_private;
stats->ipackets = 0;
stats->ibytes = 0;
stats->opackets = 0;
stats->obytes = 0;
stats->imissed = 0;
stats->oerrors = 0;
for (i = 0; i < dev->data->nb_tx_queues; i++)
eth_tx_queue_stats_get(dev->data->tx_queues[i], stats);
for (i = 0; i < dev->data->nb_rx_queues; i++)
eth_rx_queue_stats_get(dev->data->rx_queues[i], stats);
if (ark->user_ext.stats_get)
return ark->user_ext.stats_get(dev, stats,
ark->user_data[dev->data->port_id]);
return 0;
}
static void
eth_ark_dev_stats_reset(struct rte_eth_dev *dev)
{
uint16_t i;
struct ark_adapter *ark =
(struct ark_adapter *)dev->data->dev_private;
for (i = 0; i < dev->data->nb_tx_queues; i++)
eth_tx_queue_stats_reset(dev->data->tx_queues[i]);
for (i = 0; i < dev->data->nb_rx_queues; i++)
eth_rx_queue_stats_reset(dev->data->rx_queues[i]);
if (ark->user_ext.stats_reset)
ark->user_ext.stats_reset(dev,
ark->user_data[dev->data->port_id]);
}
static int
eth_ark_macaddr_add(struct rte_eth_dev *dev,
struct ether_addr *mac_addr,
uint32_t index,
uint32_t pool)
{
struct ark_adapter *ark =
(struct ark_adapter *)dev->data->dev_private;
if (ark->user_ext.mac_addr_add) {
ark->user_ext.mac_addr_add(dev,
mac_addr,
index,
pool,
ark->user_data[dev->data->port_id]);
return 0;
}
return -ENOTSUP;
}
static void
eth_ark_macaddr_remove(struct rte_eth_dev *dev, uint32_t index)
{
struct ark_adapter *ark =
(struct ark_adapter *)dev->data->dev_private;
if (ark->user_ext.mac_addr_remove)
ark->user_ext.mac_addr_remove(dev, index,
ark->user_data[dev->data->port_id]);
}
static void
eth_ark_set_default_mac_addr(struct rte_eth_dev *dev,
struct ether_addr *mac_addr)
{
struct ark_adapter *ark =
(struct ark_adapter *)dev->data->dev_private;
if (ark->user_ext.mac_addr_set)
ark->user_ext.mac_addr_set(dev, mac_addr,
ark->user_data[dev->data->port_id]);
}
static int
eth_ark_set_mtu(struct rte_eth_dev *dev, uint16_t size)
{
struct ark_adapter *ark =
(struct ark_adapter *)dev->data->dev_private;
if (ark->user_ext.set_mtu)
return ark->user_ext.set_mtu(dev, size,
ark->user_data[dev->data->port_id]);
return -ENOTSUP;
}
static inline int
process_pktdir_arg(const char *key, const char *value,
void *extra_args)
{
PMD_FUNC_LOG(DEBUG, "key = %s, value = %s\n",
key, value);
struct ark_adapter *ark =
(struct ark_adapter *)extra_args;
ark->pkt_dir_v = strtol(value, NULL, 16);
PMD_FUNC_LOG(DEBUG, "pkt_dir_v = 0x%x\n", ark->pkt_dir_v);
return 0;
}
static inline int
process_file_args(const char *key, const char *value, void *extra_args)
{
PMD_FUNC_LOG(DEBUG, "key = %s, value = %s\n",
key, value);
char *args = (char *)extra_args;
/* Open the configuration file */
FILE *file = fopen(value, "r");
char line[ARK_MAX_ARG_LEN];
int size = 0;
int first = 1;
if (file == NULL) {
PMD_DRV_LOG(ERR, "Unable to open "
"config file %s\n", value);
return -1;
}
while (fgets(line, sizeof(line), file)) {
size += strlen(line);
if (size >= ARK_MAX_ARG_LEN) {
PMD_DRV_LOG(ERR, "Unable to parse file %s args, "
"parameter list is too long\n", value);
fclose(file);
return -1;
}
if (first) {
strncpy(args, line, ARK_MAX_ARG_LEN);
first = 0;
} else {
strncat(args, line, ARK_MAX_ARG_LEN);
}
}
PMD_FUNC_LOG(DEBUG, "file = %s\n", args);
fclose(file);
return 0;
}
static int
eth_ark_check_args(struct ark_adapter *ark, const char *params)
{
struct rte_kvargs *kvlist;
unsigned int k_idx;
struct rte_kvargs_pair *pair = NULL;
int ret = -1;
kvlist = rte_kvargs_parse(params, valid_arguments);
if (kvlist == NULL)
return 0;
ark->pkt_gen_args[0] = 0;
ark->pkt_chkr_args[0] = 0;
for (k_idx = 0; k_idx < kvlist->count; k_idx++) {
pair = &kvlist->pairs[k_idx];
PMD_FUNC_LOG(DEBUG, "**** Arg passed to PMD = %s:%s\n",
pair->key,
pair->value);
}
if (rte_kvargs_process(kvlist,
ARK_PKTDIR_ARG,
&process_pktdir_arg,
ark) != 0) {
PMD_DRV_LOG(ERR, "Unable to parse arg %s\n", ARK_PKTDIR_ARG);
goto free_kvlist;
}
if (rte_kvargs_process(kvlist,
ARK_PKTGEN_ARG,
&process_file_args,
ark->pkt_gen_args) != 0) {
PMD_DRV_LOG(ERR, "Unable to parse arg %s\n", ARK_PKTGEN_ARG);
goto free_kvlist;
}
if (rte_kvargs_process(kvlist,
ARK_PKTCHKR_ARG,
&process_file_args,
ark->pkt_chkr_args) != 0) {
PMD_DRV_LOG(ERR, "Unable to parse arg %s\n", ARK_PKTCHKR_ARG);
goto free_kvlist;
}
PMD_DRV_LOG(INFO, "packet director set to 0x%x\n", ark->pkt_dir_v);
/* Setup the packet director */
ark_pktdir_setup(ark->pd, ark->pkt_dir_v);
/* Setup the packet generator */
if (ark->pkt_gen_args[0]) {
PMD_DRV_LOG(INFO, "Setting up the packet generator\n");
ark_pktgen_parse(ark->pkt_gen_args);
ark_pktgen_reset(ark->pg);
ark_pktgen_setup(ark->pg);
ark->start_pg = 1;
}
/* Setup the packet checker */
if (ark->pkt_chkr_args[0]) {
ark_pktchkr_parse(ark->pkt_chkr_args);
ark_pktchkr_setup(ark->pc);
}
ret = 0;
free_kvlist:
rte_kvargs_free(kvlist);
return ret;
}
RTE_PMD_REGISTER_PCI(net_ark, rte_ark_pmd);
RTE_PMD_REGISTER_KMOD_DEP(net_ark, "* igb_uio | uio_pci_generic ");
RTE_PMD_REGISTER_PCI_TABLE(net_ark, pci_id_ark_map);
RTE_PMD_REGISTER_PARAM_STRING(net_ark,
ARK_PKTGEN_ARG "=<filename> "
ARK_PKTCHKR_ARG "=<filename> "
ARK_PKTDIR_ARG "=<bitmap>");