f30e69b41f
Queue stats are stored in 'struct rte_eth_stats' as array and array size is defined by 'RTE_ETHDEV_QUEUE_STAT_CNTRS' compile time flag. As a result of technical board discussion, decided to remove the queue statistics from 'struct rte_eth_stats' in the long term. Instead PMDs should represent the queue statistics via xstats, this gives more flexibility on the number of the queues supported. Currently queue stats in the xstats are filled by ethdev layer, using some basic stats, when queue stats removed from basic stats the responsibility to fill the relevant xstats will be pushed to the PMDs. During the switch period, temporary 'RTE_ETH_DEV_AUTOFILL_QUEUE_XSTATS' device flag is created. Initially all PMDs using xstats set this flag. The PMDs implemented queue stats in the xstats should clear the flag. When all PMDs switch to the xstats for the queue stats, queue stats related fields from 'struct rte_eth_stats' will be removed, as well as 'RTE_ETH_DEV_AUTOFILL_QUEUE_XSTATS' flag. Later 'RTE_ETHDEV_QUEUE_STAT_CNTRS' compile time flag also can be removed. Signed-off-by: Ferruh Yigit <ferruh.yigit@intel.com> Acked-by: Haiyue Wang <haiyue.wang@intel.com> Acked-by: Xiao Wang <xiao.w.wang@intel.com> Acked-by: Thomas Monjalon <thomas@monjalon.net>
1017 lines
26 KiB
C
1017 lines
26 KiB
C
/* SPDX-License-Identifier: BSD-3-Clause
|
|
* Copyright (c) 2015-2018 Atomic Rules LLC
|
|
*/
|
|
|
|
#include <unistd.h>
|
|
#include <sys/stat.h>
|
|
#include <dlfcn.h>
|
|
|
|
#include <rte_bus_pci.h>
|
|
#include <rte_ethdev_pci.h>
|
|
#include <rte_kvargs.h>
|
|
|
|
#include "ark_global.h"
|
|
#include "ark_logs.h"
|
|
#include "ark_ethdev_tx.h"
|
|
#include "ark_ethdev_rx.h"
|
|
#include "ark_mpu.h"
|
|
#include "ark_ddm.h"
|
|
#include "ark_udm.h"
|
|
#include "ark_rqp.h"
|
|
#include "ark_pktdir.h"
|
|
#include "ark_pktgen.h"
|
|
#include "ark_pktchkr.h"
|
|
|
|
/* Internal prototypes */
|
|
static int eth_ark_check_args(struct ark_adapter *ark, const char *params);
|
|
static int eth_ark_dev_init(struct rte_eth_dev *dev);
|
|
static int ark_config_device(struct rte_eth_dev *dev);
|
|
static int eth_ark_dev_uninit(struct rte_eth_dev *eth_dev);
|
|
static int eth_ark_dev_configure(struct rte_eth_dev *dev);
|
|
static int eth_ark_dev_start(struct rte_eth_dev *dev);
|
|
static int eth_ark_dev_stop(struct rte_eth_dev *dev);
|
|
static int eth_ark_dev_close(struct rte_eth_dev *dev);
|
|
static int eth_ark_dev_info_get(struct rte_eth_dev *dev,
|
|
struct rte_eth_dev_info *dev_info);
|
|
static int eth_ark_dev_link_update(struct rte_eth_dev *dev,
|
|
int wait_to_complete);
|
|
static int eth_ark_dev_set_link_up(struct rte_eth_dev *dev);
|
|
static int eth_ark_dev_set_link_down(struct rte_eth_dev *dev);
|
|
static int eth_ark_dev_stats_get(struct rte_eth_dev *dev,
|
|
struct rte_eth_stats *stats);
|
|
static int eth_ark_dev_stats_reset(struct rte_eth_dev *dev);
|
|
static int eth_ark_set_default_mac_addr(struct rte_eth_dev *dev,
|
|
struct rte_ether_addr *mac_addr);
|
|
static int eth_ark_macaddr_add(struct rte_eth_dev *dev,
|
|
struct rte_ether_addr *mac_addr,
|
|
uint32_t index,
|
|
uint32_t pool);
|
|
static void eth_ark_macaddr_remove(struct rte_eth_dev *dev,
|
|
uint32_t index);
|
|
static int eth_ark_set_mtu(struct rte_eth_dev *dev, uint16_t size);
|
|
|
|
/*
|
|
* The packet generator is a functional block used to generate packet
|
|
* patterns for testing. It is not intended for nominal use.
|
|
*/
|
|
#define ARK_PKTGEN_ARG "Pkt_gen"
|
|
|
|
/*
|
|
* The packet checker is a functional block used to verify packet
|
|
* patterns for testing. It is not intended for nominal use.
|
|
*/
|
|
#define ARK_PKTCHKR_ARG "Pkt_chkr"
|
|
|
|
/*
|
|
* The packet director is used to select the internal ingress and
|
|
* egress packets paths during testing. It is not intended for
|
|
* nominal use.
|
|
*/
|
|
#define ARK_PKTDIR_ARG "Pkt_dir"
|
|
|
|
/* Devinfo configurations */
|
|
#define ARK_RX_MAX_QUEUE (4096 * 4)
|
|
#define ARK_RX_MIN_QUEUE (512)
|
|
#define ARK_RX_MAX_PKT_LEN ((16 * 1024) - 128)
|
|
#define ARK_RX_MIN_BUFSIZE (1024)
|
|
|
|
#define ARK_TX_MAX_QUEUE (4096 * 4)
|
|
#define ARK_TX_MIN_QUEUE (256)
|
|
|
|
static const char * const valid_arguments[] = {
|
|
ARK_PKTGEN_ARG,
|
|
ARK_PKTCHKR_ARG,
|
|
ARK_PKTDIR_ARG,
|
|
NULL
|
|
};
|
|
|
|
static const struct rte_pci_id pci_id_ark_map[] = {
|
|
{RTE_PCI_DEVICE(0x1d6c, 0x100d)},
|
|
{RTE_PCI_DEVICE(0x1d6c, 0x100e)},
|
|
{.vendor_id = 0, /* sentinel */ },
|
|
};
|
|
|
|
static int
|
|
eth_ark_pci_probe(struct rte_pci_driver *pci_drv __rte_unused,
|
|
struct rte_pci_device *pci_dev)
|
|
{
|
|
struct rte_eth_dev *eth_dev;
|
|
int ret;
|
|
|
|
eth_dev = rte_eth_dev_pci_allocate(pci_dev, sizeof(struct ark_adapter));
|
|
|
|
if (eth_dev == NULL)
|
|
return -ENOMEM;
|
|
|
|
ret = eth_ark_dev_init(eth_dev);
|
|
if (ret)
|
|
rte_eth_dev_release_port(eth_dev);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int
|
|
eth_ark_pci_remove(struct rte_pci_device *pci_dev)
|
|
{
|
|
return rte_eth_dev_pci_generic_remove(pci_dev, eth_ark_dev_uninit);
|
|
}
|
|
|
|
static struct rte_pci_driver rte_ark_pmd = {
|
|
.id_table = pci_id_ark_map,
|
|
.drv_flags = RTE_PCI_DRV_NEED_MAPPING | RTE_PCI_DRV_INTR_LSC,
|
|
.probe = eth_ark_pci_probe,
|
|
.remove = eth_ark_pci_remove,
|
|
};
|
|
|
|
static const struct eth_dev_ops ark_eth_dev_ops = {
|
|
.dev_configure = eth_ark_dev_configure,
|
|
.dev_start = eth_ark_dev_start,
|
|
.dev_stop = eth_ark_dev_stop,
|
|
.dev_close = eth_ark_dev_close,
|
|
|
|
.dev_infos_get = eth_ark_dev_info_get,
|
|
|
|
.rx_queue_setup = eth_ark_dev_rx_queue_setup,
|
|
.tx_queue_setup = eth_ark_tx_queue_setup,
|
|
|
|
.link_update = eth_ark_dev_link_update,
|
|
.dev_set_link_up = eth_ark_dev_set_link_up,
|
|
.dev_set_link_down = eth_ark_dev_set_link_down,
|
|
|
|
.rx_queue_start = eth_ark_rx_start_queue,
|
|
.rx_queue_stop = eth_ark_rx_stop_queue,
|
|
|
|
.tx_queue_start = eth_ark_tx_queue_start,
|
|
.tx_queue_stop = eth_ark_tx_queue_stop,
|
|
|
|
.stats_get = eth_ark_dev_stats_get,
|
|
.stats_reset = eth_ark_dev_stats_reset,
|
|
|
|
.mac_addr_add = eth_ark_macaddr_add,
|
|
.mac_addr_remove = eth_ark_macaddr_remove,
|
|
.mac_addr_set = eth_ark_set_default_mac_addr,
|
|
|
|
.mtu_set = eth_ark_set_mtu,
|
|
};
|
|
|
|
static int
|
|
check_for_ext(struct ark_adapter *ark)
|
|
{
|
|
int found = 0;
|
|
|
|
/* Get the env */
|
|
const char *dllpath = getenv("ARK_EXT_PATH");
|
|
|
|
if (dllpath == NULL) {
|
|
ARK_PMD_LOG(DEBUG, "EXT NO dll path specified\n");
|
|
return 0;
|
|
}
|
|
ARK_PMD_LOG(NOTICE, "EXT found dll path at %s\n", dllpath);
|
|
|
|
/* Open and load the .so */
|
|
ark->d_handle = dlopen(dllpath, RTLD_LOCAL | RTLD_LAZY);
|
|
if (ark->d_handle == NULL) {
|
|
ARK_PMD_LOG(ERR, "Could not load user extension %s\n",
|
|
dllpath);
|
|
return -1;
|
|
}
|
|
ARK_PMD_LOG(DEBUG, "SUCCESS: loaded user extension %s\n",
|
|
dllpath);
|
|
|
|
/* Get the entry points */
|
|
ark->user_ext.dev_init =
|
|
(void *(*)(struct rte_eth_dev *, void *, int))
|
|
dlsym(ark->d_handle, "dev_init");
|
|
ARK_PMD_LOG(DEBUG, "device ext init pointer = %p\n",
|
|
ark->user_ext.dev_init);
|
|
ark->user_ext.dev_get_port_count =
|
|
(int (*)(struct rte_eth_dev *, void *))
|
|
dlsym(ark->d_handle, "dev_get_port_count");
|
|
ark->user_ext.dev_uninit =
|
|
(void (*)(struct rte_eth_dev *, void *))
|
|
dlsym(ark->d_handle, "dev_uninit");
|
|
ark->user_ext.dev_configure =
|
|
(int (*)(struct rte_eth_dev *, void *))
|
|
dlsym(ark->d_handle, "dev_configure");
|
|
ark->user_ext.dev_start =
|
|
(int (*)(struct rte_eth_dev *, void *))
|
|
dlsym(ark->d_handle, "dev_start");
|
|
ark->user_ext.dev_stop =
|
|
(void (*)(struct rte_eth_dev *, void *))
|
|
dlsym(ark->d_handle, "dev_stop");
|
|
ark->user_ext.dev_close =
|
|
(void (*)(struct rte_eth_dev *, void *))
|
|
dlsym(ark->d_handle, "dev_close");
|
|
ark->user_ext.link_update =
|
|
(int (*)(struct rte_eth_dev *, int, void *))
|
|
dlsym(ark->d_handle, "link_update");
|
|
ark->user_ext.dev_set_link_up =
|
|
(int (*)(struct rte_eth_dev *, void *))
|
|
dlsym(ark->d_handle, "dev_set_link_up");
|
|
ark->user_ext.dev_set_link_down =
|
|
(int (*)(struct rte_eth_dev *, void *))
|
|
dlsym(ark->d_handle, "dev_set_link_down");
|
|
ark->user_ext.stats_get =
|
|
(int (*)(struct rte_eth_dev *, struct rte_eth_stats *,
|
|
void *))
|
|
dlsym(ark->d_handle, "stats_get");
|
|
ark->user_ext.stats_reset =
|
|
(void (*)(struct rte_eth_dev *, void *))
|
|
dlsym(ark->d_handle, "stats_reset");
|
|
ark->user_ext.mac_addr_add =
|
|
(void (*)(struct rte_eth_dev *, struct rte_ether_addr *,
|
|
uint32_t, uint32_t, void *))
|
|
dlsym(ark->d_handle, "mac_addr_add");
|
|
ark->user_ext.mac_addr_remove =
|
|
(void (*)(struct rte_eth_dev *, uint32_t, void *))
|
|
dlsym(ark->d_handle, "mac_addr_remove");
|
|
ark->user_ext.mac_addr_set =
|
|
(void (*)(struct rte_eth_dev *, struct rte_ether_addr *,
|
|
void *))
|
|
dlsym(ark->d_handle, "mac_addr_set");
|
|
ark->user_ext.set_mtu =
|
|
(int (*)(struct rte_eth_dev *, uint16_t,
|
|
void *))
|
|
dlsym(ark->d_handle, "set_mtu");
|
|
|
|
return found;
|
|
}
|
|
|
|
static int
|
|
eth_ark_dev_init(struct rte_eth_dev *dev)
|
|
{
|
|
struct ark_adapter *ark = dev->data->dev_private;
|
|
struct rte_pci_device *pci_dev;
|
|
int ret;
|
|
int port_count = 1;
|
|
int p;
|
|
|
|
ark->eth_dev = dev;
|
|
|
|
ARK_PMD_LOG(DEBUG, "\n");
|
|
|
|
/* Check to see if there is an extension that we need to load */
|
|
ret = check_for_ext(ark);
|
|
if (ret)
|
|
return ret;
|
|
pci_dev = RTE_ETH_DEV_TO_PCI(dev);
|
|
rte_eth_copy_pci_info(dev, pci_dev);
|
|
dev->data->dev_flags |= RTE_ETH_DEV_AUTOFILL_QUEUE_XSTATS;
|
|
|
|
/* Use dummy function until setup */
|
|
dev->rx_pkt_burst = ð_ark_recv_pkts_noop;
|
|
dev->tx_pkt_burst = ð_ark_xmit_pkts_noop;
|
|
|
|
ark->bar0 = (uint8_t *)pci_dev->mem_resource[0].addr;
|
|
ark->a_bar = (uint8_t *)pci_dev->mem_resource[2].addr;
|
|
|
|
ark->sysctrl.v = (void *)&ark->bar0[ARK_SYSCTRL_BASE];
|
|
ark->mpurx.v = (void *)&ark->bar0[ARK_MPU_RX_BASE];
|
|
ark->udm.v = (void *)&ark->bar0[ARK_UDM_BASE];
|
|
ark->mputx.v = (void *)&ark->bar0[ARK_MPU_TX_BASE];
|
|
ark->ddm.v = (void *)&ark->bar0[ARK_DDM_BASE];
|
|
ark->cmac.v = (void *)&ark->bar0[ARK_CMAC_BASE];
|
|
ark->external.v = (void *)&ark->bar0[ARK_EXTERNAL_BASE];
|
|
ark->pktdir.v = (void *)&ark->bar0[ARK_PKTDIR_BASE];
|
|
ark->pktgen.v = (void *)&ark->bar0[ARK_PKTGEN_BASE];
|
|
ark->pktchkr.v = (void *)&ark->bar0[ARK_PKTCHKR_BASE];
|
|
|
|
ark->rqpacing =
|
|
(struct ark_rqpace_t *)(ark->bar0 + ARK_RCPACING_BASE);
|
|
ark->started = 0;
|
|
|
|
ARK_PMD_LOG(INFO, "Sys Ctrl Const = 0x%x HW Commit_ID: %08x\n",
|
|
ark->sysctrl.t32[4],
|
|
rte_be_to_cpu_32(ark->sysctrl.t32[0x20 / 4]));
|
|
ARK_PMD_LOG(NOTICE, "Arkville HW Commit_ID: %08x\n",
|
|
rte_be_to_cpu_32(ark->sysctrl.t32[0x20 / 4]));
|
|
|
|
/* If HW sanity test fails, return an error */
|
|
if (ark->sysctrl.t32[4] != 0xcafef00d) {
|
|
ARK_PMD_LOG(ERR,
|
|
"HW Sanity test has failed, expected constant"
|
|
" 0x%x, read 0x%x (%s)\n",
|
|
0xcafef00d,
|
|
ark->sysctrl.t32[4], __func__);
|
|
return -1;
|
|
}
|
|
if (ark->sysctrl.t32[3] != 0) {
|
|
if (ark_rqp_lasped(ark->rqpacing)) {
|
|
ARK_PMD_LOG(ERR, "Arkville Evaluation System - "
|
|
"Timer has Expired\n");
|
|
return -1;
|
|
}
|
|
ARK_PMD_LOG(WARNING, "Arkville Evaluation System - "
|
|
"Timer is Running\n");
|
|
}
|
|
|
|
ARK_PMD_LOG(DEBUG,
|
|
"HW Sanity test has PASSED, expected constant"
|
|
" 0x%x, read 0x%x (%s)\n",
|
|
0xcafef00d, ark->sysctrl.t32[4], __func__);
|
|
|
|
/* We are a single function multi-port device. */
|
|
ret = ark_config_device(dev);
|
|
if (ret)
|
|
return -1;
|
|
|
|
dev->dev_ops = &ark_eth_dev_ops;
|
|
dev->rx_queue_count = eth_ark_dev_rx_queue_count;
|
|
|
|
dev->data->mac_addrs = rte_zmalloc("ark", RTE_ETHER_ADDR_LEN, 0);
|
|
if (!dev->data->mac_addrs) {
|
|
ARK_PMD_LOG(ERR,
|
|
"Failed to allocated memory for storing mac address"
|
|
);
|
|
}
|
|
|
|
if (ark->user_ext.dev_init) {
|
|
ark->user_data[dev->data->port_id] =
|
|
ark->user_ext.dev_init(dev, ark->a_bar, 0);
|
|
if (!ark->user_data[dev->data->port_id]) {
|
|
ARK_PMD_LOG(WARNING,
|
|
"Failed to initialize PMD extension!"
|
|
" continuing without it\n");
|
|
memset(&ark->user_ext, 0, sizeof(struct ark_user_ext));
|
|
dlclose(ark->d_handle);
|
|
}
|
|
}
|
|
|
|
if (pci_dev->device.devargs)
|
|
ret = eth_ark_check_args(ark, pci_dev->device.devargs->args);
|
|
else
|
|
ARK_PMD_LOG(INFO, "No Device args found\n");
|
|
|
|
if (ret)
|
|
goto error;
|
|
/*
|
|
* We will create additional devices based on the number of requested
|
|
* ports
|
|
*/
|
|
if (ark->user_ext.dev_get_port_count)
|
|
port_count =
|
|
ark->user_ext.dev_get_port_count(dev,
|
|
ark->user_data[dev->data->port_id]);
|
|
ark->num_ports = port_count;
|
|
|
|
for (p = 0; p < port_count; p++) {
|
|
struct rte_eth_dev *eth_dev;
|
|
char name[RTE_ETH_NAME_MAX_LEN];
|
|
|
|
snprintf(name, sizeof(name), "arketh%d",
|
|
dev->data->port_id + p);
|
|
|
|
if (p == 0) {
|
|
/* First port is already allocated by DPDK */
|
|
eth_dev = ark->eth_dev;
|
|
rte_eth_dev_probing_finish(eth_dev);
|
|
continue;
|
|
}
|
|
|
|
/* reserve an ethdev entry */
|
|
eth_dev = rte_eth_dev_allocate(name);
|
|
if (!eth_dev) {
|
|
ARK_PMD_LOG(ERR,
|
|
"Could not allocate eth_dev for port %d\n",
|
|
p);
|
|
goto error;
|
|
}
|
|
|
|
eth_dev->device = &pci_dev->device;
|
|
eth_dev->data->dev_private = ark;
|
|
eth_dev->dev_ops = ark->eth_dev->dev_ops;
|
|
eth_dev->tx_pkt_burst = ark->eth_dev->tx_pkt_burst;
|
|
eth_dev->rx_pkt_burst = ark->eth_dev->rx_pkt_burst;
|
|
|
|
rte_eth_copy_pci_info(eth_dev, pci_dev);
|
|
eth_dev->data->dev_flags |= RTE_ETH_DEV_AUTOFILL_QUEUE_XSTATS;
|
|
|
|
eth_dev->data->mac_addrs = rte_zmalloc(name,
|
|
RTE_ETHER_ADDR_LEN, 0);
|
|
if (!eth_dev->data->mac_addrs) {
|
|
ARK_PMD_LOG(ERR,
|
|
"Memory allocation for MAC failed!"
|
|
" Exiting.\n");
|
|
goto error;
|
|
}
|
|
|
|
if (ark->user_ext.dev_init) {
|
|
ark->user_data[eth_dev->data->port_id] =
|
|
ark->user_ext.dev_init(dev, ark->a_bar, p);
|
|
}
|
|
|
|
rte_eth_dev_probing_finish(eth_dev);
|
|
}
|
|
|
|
return ret;
|
|
|
|
error:
|
|
rte_free(dev->data->mac_addrs);
|
|
dev->data->mac_addrs = NULL;
|
|
return -1;
|
|
}
|
|
|
|
/*
|
|
*Initial device configuration when device is opened
|
|
* setup the DDM, and UDM
|
|
* Called once per PCIE device
|
|
*/
|
|
static int
|
|
ark_config_device(struct rte_eth_dev *dev)
|
|
{
|
|
struct ark_adapter *ark = dev->data->dev_private;
|
|
uint16_t num_q, i;
|
|
struct ark_mpu_t *mpu;
|
|
|
|
/*
|
|
* Make sure that the packet director, generator and checker are in a
|
|
* known state
|
|
*/
|
|
ark->start_pg = 0;
|
|
ark->pg = ark_pktgen_init(ark->pktgen.v, 0, 1);
|
|
if (ark->pg == NULL)
|
|
return -1;
|
|
ark_pktgen_reset(ark->pg);
|
|
ark->pc = ark_pktchkr_init(ark->pktchkr.v, 0, 1);
|
|
if (ark->pc == NULL)
|
|
return -1;
|
|
ark_pktchkr_stop(ark->pc);
|
|
ark->pd = ark_pktdir_init(ark->pktdir.v);
|
|
if (ark->pd == NULL)
|
|
return -1;
|
|
|
|
/* Verify HW */
|
|
if (ark_udm_verify(ark->udm.v))
|
|
return -1;
|
|
if (ark_ddm_verify(ark->ddm.v))
|
|
return -1;
|
|
|
|
/* UDM */
|
|
if (ark_udm_reset(ark->udm.v)) {
|
|
ARK_PMD_LOG(ERR, "Unable to stop and reset UDM\n");
|
|
return -1;
|
|
}
|
|
/* Keep in reset until the MPU are cleared */
|
|
|
|
/* MPU reset */
|
|
mpu = ark->mpurx.v;
|
|
num_q = ark_api_num_queues(mpu);
|
|
ark->rx_queues = num_q;
|
|
for (i = 0; i < num_q; i++) {
|
|
ark_mpu_reset(mpu);
|
|
mpu = RTE_PTR_ADD(mpu, ARK_MPU_QOFFSET);
|
|
}
|
|
|
|
ark_udm_stop(ark->udm.v, 0);
|
|
ark_udm_configure(ark->udm.v,
|
|
RTE_PKTMBUF_HEADROOM,
|
|
RTE_MBUF_DEFAULT_DATAROOM,
|
|
ARK_RX_WRITE_TIME_NS);
|
|
ark_udm_stats_reset(ark->udm.v);
|
|
ark_udm_stop(ark->udm.v, 0);
|
|
|
|
/* TX -- DDM */
|
|
if (ark_ddm_stop(ark->ddm.v, 1))
|
|
ARK_PMD_LOG(ERR, "Unable to stop DDM\n");
|
|
|
|
mpu = ark->mputx.v;
|
|
num_q = ark_api_num_queues(mpu);
|
|
ark->tx_queues = num_q;
|
|
for (i = 0; i < num_q; i++) {
|
|
ark_mpu_reset(mpu);
|
|
mpu = RTE_PTR_ADD(mpu, ARK_MPU_QOFFSET);
|
|
}
|
|
|
|
ark_ddm_reset(ark->ddm.v);
|
|
ark_ddm_stats_reset(ark->ddm.v);
|
|
|
|
ark_ddm_stop(ark->ddm.v, 0);
|
|
ark_rqp_stats_reset(ark->rqpacing);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
eth_ark_dev_uninit(struct rte_eth_dev *dev)
|
|
{
|
|
struct ark_adapter *ark = dev->data->dev_private;
|
|
|
|
if (rte_eal_process_type() != RTE_PROC_PRIMARY)
|
|
return 0;
|
|
|
|
if (ark->user_ext.dev_uninit)
|
|
ark->user_ext.dev_uninit(dev,
|
|
ark->user_data[dev->data->port_id]);
|
|
|
|
ark_pktgen_uninit(ark->pg);
|
|
ark_pktchkr_uninit(ark->pc);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
eth_ark_dev_configure(struct rte_eth_dev *dev)
|
|
{
|
|
struct ark_adapter *ark = dev->data->dev_private;
|
|
|
|
eth_ark_dev_set_link_up(dev);
|
|
if (ark->user_ext.dev_configure)
|
|
return ark->user_ext.dev_configure(dev,
|
|
ark->user_data[dev->data->port_id]);
|
|
return 0;
|
|
}
|
|
|
|
static void *
|
|
delay_pg_start(void *arg)
|
|
{
|
|
struct ark_adapter *ark = (struct ark_adapter *)arg;
|
|
|
|
/* This function is used exclusively for regression testing, We
|
|
* perform a blind sleep here to ensure that the external test
|
|
* application has time to setup the test before we generate packets
|
|
*/
|
|
usleep(100000);
|
|
ark_pktgen_run(ark->pg);
|
|
return NULL;
|
|
}
|
|
|
|
static int
|
|
eth_ark_dev_start(struct rte_eth_dev *dev)
|
|
{
|
|
struct ark_adapter *ark = dev->data->dev_private;
|
|
int i;
|
|
|
|
/* RX Side */
|
|
/* start UDM */
|
|
ark_udm_start(ark->udm.v);
|
|
|
|
for (i = 0; i < dev->data->nb_rx_queues; i++)
|
|
eth_ark_rx_start_queue(dev, i);
|
|
|
|
/* TX Side */
|
|
for (i = 0; i < dev->data->nb_tx_queues; i++)
|
|
eth_ark_tx_queue_start(dev, i);
|
|
|
|
/* start DDM */
|
|
ark_ddm_start(ark->ddm.v);
|
|
|
|
ark->started = 1;
|
|
/* set xmit and receive function */
|
|
dev->rx_pkt_burst = ð_ark_recv_pkts;
|
|
dev->tx_pkt_burst = ð_ark_xmit_pkts;
|
|
|
|
if (ark->start_pg)
|
|
ark_pktchkr_run(ark->pc);
|
|
|
|
if (ark->start_pg && (dev->data->port_id == 0)) {
|
|
pthread_t thread;
|
|
|
|
/* Delay packet generatpr start allow the hardware to be ready
|
|
* This is only used for sanity checking with internal generator
|
|
*/
|
|
if (pthread_create(&thread, NULL, delay_pg_start, ark)) {
|
|
ARK_PMD_LOG(ERR, "Could not create pktgen "
|
|
"starter thread\n");
|
|
return -1;
|
|
}
|
|
}
|
|
|
|
if (ark->user_ext.dev_start)
|
|
ark->user_ext.dev_start(dev,
|
|
ark->user_data[dev->data->port_id]);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
eth_ark_dev_stop(struct rte_eth_dev *dev)
|
|
{
|
|
uint16_t i;
|
|
int status;
|
|
struct ark_adapter *ark = dev->data->dev_private;
|
|
struct ark_mpu_t *mpu;
|
|
|
|
if (ark->started == 0)
|
|
return 0;
|
|
ark->started = 0;
|
|
dev->data->dev_started = 0;
|
|
|
|
/* Stop the extension first */
|
|
if (ark->user_ext.dev_stop)
|
|
ark->user_ext.dev_stop(dev,
|
|
ark->user_data[dev->data->port_id]);
|
|
|
|
/* Stop the packet generator */
|
|
if (ark->start_pg)
|
|
ark_pktgen_pause(ark->pg);
|
|
|
|
dev->rx_pkt_burst = ð_ark_recv_pkts_noop;
|
|
dev->tx_pkt_burst = ð_ark_xmit_pkts_noop;
|
|
|
|
/* STOP TX Side */
|
|
for (i = 0; i < dev->data->nb_tx_queues; i++) {
|
|
status = eth_ark_tx_queue_stop(dev, i);
|
|
if (status != 0) {
|
|
uint16_t port = dev->data->port_id;
|
|
ARK_PMD_LOG(ERR,
|
|
"tx_queue stop anomaly"
|
|
" port %u, queue %u\n",
|
|
port, i);
|
|
}
|
|
}
|
|
|
|
/* Stop DDM */
|
|
/* Wait up to 0.1 second. each stop is up to 1000 * 10 useconds */
|
|
for (i = 0; i < 10; i++) {
|
|
status = ark_ddm_stop(ark->ddm.v, 1);
|
|
if (status == 0)
|
|
break;
|
|
}
|
|
if (status || i != 0) {
|
|
ARK_PMD_LOG(ERR, "DDM stop anomaly. status:"
|
|
" %d iter: %u. (%s)\n",
|
|
status,
|
|
i,
|
|
__func__);
|
|
ark_ddm_dump(ark->ddm.v, "Stop anomaly");
|
|
|
|
mpu = ark->mputx.v;
|
|
for (i = 0; i < ark->tx_queues; i++) {
|
|
ark_mpu_dump(mpu, "DDM failure dump", i);
|
|
mpu = RTE_PTR_ADD(mpu, ARK_MPU_QOFFSET);
|
|
}
|
|
}
|
|
|
|
/* STOP RX Side */
|
|
/* Stop UDM multiple tries attempted */
|
|
for (i = 0; i < 10; i++) {
|
|
status = ark_udm_stop(ark->udm.v, 1);
|
|
if (status == 0)
|
|
break;
|
|
}
|
|
if (status || i != 0) {
|
|
ARK_PMD_LOG(ERR, "UDM stop anomaly. status %d iter: %u. (%s)\n",
|
|
status, i, __func__);
|
|
ark_udm_dump(ark->udm.v, "Stop anomaly");
|
|
|
|
mpu = ark->mpurx.v;
|
|
for (i = 0; i < ark->rx_queues; i++) {
|
|
ark_mpu_dump(mpu, "UDM Stop anomaly", i);
|
|
mpu = RTE_PTR_ADD(mpu, ARK_MPU_QOFFSET);
|
|
}
|
|
}
|
|
|
|
ark_udm_dump_stats(ark->udm.v, "Post stop");
|
|
ark_udm_dump_perf(ark->udm.v, "Post stop");
|
|
|
|
for (i = 0; i < dev->data->nb_rx_queues; i++)
|
|
eth_ark_rx_dump_queue(dev, i, __func__);
|
|
|
|
/* Stop the packet checker if it is running */
|
|
if (ark->start_pg) {
|
|
ark_pktchkr_dump_stats(ark->pc);
|
|
ark_pktchkr_stop(ark->pc);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
eth_ark_dev_close(struct rte_eth_dev *dev)
|
|
{
|
|
struct ark_adapter *ark = dev->data->dev_private;
|
|
uint16_t i;
|
|
|
|
if (rte_eal_process_type() != RTE_PROC_PRIMARY)
|
|
return 0;
|
|
|
|
if (ark->user_ext.dev_close)
|
|
ark->user_ext.dev_close(dev,
|
|
ark->user_data[dev->data->port_id]);
|
|
|
|
eth_ark_dev_stop(dev);
|
|
eth_ark_udm_force_close(dev);
|
|
|
|
/*
|
|
* TODO This should only be called once for the device during shutdown
|
|
*/
|
|
ark_rqp_dump(ark->rqpacing);
|
|
|
|
for (i = 0; i < dev->data->nb_tx_queues; i++) {
|
|
eth_ark_tx_queue_release(dev->data->tx_queues[i]);
|
|
dev->data->tx_queues[i] = 0;
|
|
}
|
|
|
|
for (i = 0; i < dev->data->nb_rx_queues; i++) {
|
|
eth_ark_dev_rx_queue_release(dev->data->rx_queues[i]);
|
|
dev->data->rx_queues[i] = 0;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
eth_ark_dev_info_get(struct rte_eth_dev *dev,
|
|
struct rte_eth_dev_info *dev_info)
|
|
{
|
|
struct ark_adapter *ark = dev->data->dev_private;
|
|
struct ark_mpu_t *tx_mpu = RTE_PTR_ADD(ark->bar0, ARK_MPU_TX_BASE);
|
|
struct ark_mpu_t *rx_mpu = RTE_PTR_ADD(ark->bar0, ARK_MPU_RX_BASE);
|
|
uint16_t ports = ark->num_ports;
|
|
|
|
dev_info->max_rx_pktlen = ARK_RX_MAX_PKT_LEN;
|
|
dev_info->min_rx_bufsize = ARK_RX_MIN_BUFSIZE;
|
|
|
|
dev_info->max_rx_queues = ark_api_num_queues_per_port(rx_mpu, ports);
|
|
dev_info->max_tx_queues = ark_api_num_queues_per_port(tx_mpu, ports);
|
|
|
|
dev_info->rx_desc_lim = (struct rte_eth_desc_lim) {
|
|
.nb_max = ARK_RX_MAX_QUEUE,
|
|
.nb_min = ARK_RX_MIN_QUEUE,
|
|
.nb_align = ARK_RX_MIN_QUEUE}; /* power of 2 */
|
|
|
|
dev_info->tx_desc_lim = (struct rte_eth_desc_lim) {
|
|
.nb_max = ARK_TX_MAX_QUEUE,
|
|
.nb_min = ARK_TX_MIN_QUEUE,
|
|
.nb_align = ARK_TX_MIN_QUEUE}; /* power of 2 */
|
|
|
|
/* ARK PMD supports all line rates, how do we indicate that here ?? */
|
|
dev_info->speed_capa = (ETH_LINK_SPEED_1G |
|
|
ETH_LINK_SPEED_10G |
|
|
ETH_LINK_SPEED_25G |
|
|
ETH_LINK_SPEED_40G |
|
|
ETH_LINK_SPEED_50G |
|
|
ETH_LINK_SPEED_100G);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
eth_ark_dev_link_update(struct rte_eth_dev *dev, int wait_to_complete)
|
|
{
|
|
ARK_PMD_LOG(DEBUG, "link status = %d\n",
|
|
dev->data->dev_link.link_status);
|
|
struct ark_adapter *ark = dev->data->dev_private;
|
|
|
|
if (ark->user_ext.link_update) {
|
|
return ark->user_ext.link_update
|
|
(dev, wait_to_complete,
|
|
ark->user_data[dev->data->port_id]);
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
eth_ark_dev_set_link_up(struct rte_eth_dev *dev)
|
|
{
|
|
dev->data->dev_link.link_status = 1;
|
|
struct ark_adapter *ark = dev->data->dev_private;
|
|
|
|
if (ark->user_ext.dev_set_link_up)
|
|
return ark->user_ext.dev_set_link_up(dev,
|
|
ark->user_data[dev->data->port_id]);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
eth_ark_dev_set_link_down(struct rte_eth_dev *dev)
|
|
{
|
|
dev->data->dev_link.link_status = 0;
|
|
struct ark_adapter *ark = dev->data->dev_private;
|
|
|
|
if (ark->user_ext.dev_set_link_down)
|
|
return ark->user_ext.dev_set_link_down(dev,
|
|
ark->user_data[dev->data->port_id]);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
eth_ark_dev_stats_get(struct rte_eth_dev *dev, struct rte_eth_stats *stats)
|
|
{
|
|
uint16_t i;
|
|
struct ark_adapter *ark = dev->data->dev_private;
|
|
|
|
stats->ipackets = 0;
|
|
stats->ibytes = 0;
|
|
stats->opackets = 0;
|
|
stats->obytes = 0;
|
|
stats->imissed = 0;
|
|
stats->oerrors = 0;
|
|
|
|
for (i = 0; i < dev->data->nb_tx_queues; i++)
|
|
eth_tx_queue_stats_get(dev->data->tx_queues[i], stats);
|
|
for (i = 0; i < dev->data->nb_rx_queues; i++)
|
|
eth_rx_queue_stats_get(dev->data->rx_queues[i], stats);
|
|
if (ark->user_ext.stats_get)
|
|
return ark->user_ext.stats_get(dev, stats,
|
|
ark->user_data[dev->data->port_id]);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
eth_ark_dev_stats_reset(struct rte_eth_dev *dev)
|
|
{
|
|
uint16_t i;
|
|
struct ark_adapter *ark = dev->data->dev_private;
|
|
|
|
for (i = 0; i < dev->data->nb_tx_queues; i++)
|
|
eth_tx_queue_stats_reset(dev->data->tx_queues[i]);
|
|
for (i = 0; i < dev->data->nb_rx_queues; i++)
|
|
eth_rx_queue_stats_reset(dev->data->rx_queues[i]);
|
|
if (ark->user_ext.stats_reset)
|
|
ark->user_ext.stats_reset(dev,
|
|
ark->user_data[dev->data->port_id]);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
eth_ark_macaddr_add(struct rte_eth_dev *dev,
|
|
struct rte_ether_addr *mac_addr,
|
|
uint32_t index,
|
|
uint32_t pool)
|
|
{
|
|
struct ark_adapter *ark = dev->data->dev_private;
|
|
|
|
if (ark->user_ext.mac_addr_add) {
|
|
ark->user_ext.mac_addr_add(dev,
|
|
mac_addr,
|
|
index,
|
|
pool,
|
|
ark->user_data[dev->data->port_id]);
|
|
return 0;
|
|
}
|
|
return -ENOTSUP;
|
|
}
|
|
|
|
static void
|
|
eth_ark_macaddr_remove(struct rte_eth_dev *dev, uint32_t index)
|
|
{
|
|
struct ark_adapter *ark = dev->data->dev_private;
|
|
|
|
if (ark->user_ext.mac_addr_remove)
|
|
ark->user_ext.mac_addr_remove(dev, index,
|
|
ark->user_data[dev->data->port_id]);
|
|
}
|
|
|
|
static int
|
|
eth_ark_set_default_mac_addr(struct rte_eth_dev *dev,
|
|
struct rte_ether_addr *mac_addr)
|
|
{
|
|
struct ark_adapter *ark = dev->data->dev_private;
|
|
|
|
if (ark->user_ext.mac_addr_set) {
|
|
ark->user_ext.mac_addr_set(dev, mac_addr,
|
|
ark->user_data[dev->data->port_id]);
|
|
return 0;
|
|
}
|
|
return -ENOTSUP;
|
|
}
|
|
|
|
static int
|
|
eth_ark_set_mtu(struct rte_eth_dev *dev, uint16_t size)
|
|
{
|
|
struct ark_adapter *ark = dev->data->dev_private;
|
|
|
|
if (ark->user_ext.set_mtu)
|
|
return ark->user_ext.set_mtu(dev, size,
|
|
ark->user_data[dev->data->port_id]);
|
|
|
|
return -ENOTSUP;
|
|
}
|
|
|
|
static inline int
|
|
process_pktdir_arg(const char *key, const char *value,
|
|
void *extra_args)
|
|
{
|
|
ARK_PMD_LOG(DEBUG, "key = %s, value = %s\n",
|
|
key, value);
|
|
struct ark_adapter *ark =
|
|
(struct ark_adapter *)extra_args;
|
|
|
|
ark->pkt_dir_v = strtol(value, NULL, 16);
|
|
ARK_PMD_LOG(DEBUG, "pkt_dir_v = 0x%x\n", ark->pkt_dir_v);
|
|
return 0;
|
|
}
|
|
|
|
static inline int
|
|
process_file_args(const char *key, const char *value, void *extra_args)
|
|
{
|
|
ARK_PMD_LOG(DEBUG, "key = %s, value = %s\n",
|
|
key, value);
|
|
char *args = (char *)extra_args;
|
|
|
|
/* Open the configuration file */
|
|
FILE *file = fopen(value, "r");
|
|
char line[ARK_MAX_ARG_LEN];
|
|
int size = 0;
|
|
int first = 1;
|
|
|
|
if (file == NULL) {
|
|
ARK_PMD_LOG(ERR, "Unable to open "
|
|
"config file %s\n", value);
|
|
return -1;
|
|
}
|
|
|
|
while (fgets(line, sizeof(line), file)) {
|
|
size += strlen(line);
|
|
if (size >= ARK_MAX_ARG_LEN) {
|
|
ARK_PMD_LOG(ERR, "Unable to parse file %s args, "
|
|
"parameter list is too long\n", value);
|
|
fclose(file);
|
|
return -1;
|
|
}
|
|
if (first) {
|
|
strncpy(args, line, ARK_MAX_ARG_LEN);
|
|
first = 0;
|
|
} else {
|
|
strncat(args, line, ARK_MAX_ARG_LEN);
|
|
}
|
|
}
|
|
ARK_PMD_LOG(DEBUG, "file = %s\n", args);
|
|
fclose(file);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
eth_ark_check_args(struct ark_adapter *ark, const char *params)
|
|
{
|
|
struct rte_kvargs *kvlist;
|
|
unsigned int k_idx;
|
|
struct rte_kvargs_pair *pair = NULL;
|
|
int ret = -1;
|
|
|
|
kvlist = rte_kvargs_parse(params, valid_arguments);
|
|
if (kvlist == NULL)
|
|
return 0;
|
|
|
|
ark->pkt_gen_args[0] = 0;
|
|
ark->pkt_chkr_args[0] = 0;
|
|
|
|
for (k_idx = 0; k_idx < kvlist->count; k_idx++) {
|
|
pair = &kvlist->pairs[k_idx];
|
|
ARK_PMD_LOG(DEBUG, "**** Arg passed to PMD = %s:%s\n",
|
|
pair->key,
|
|
pair->value);
|
|
}
|
|
|
|
if (rte_kvargs_process(kvlist,
|
|
ARK_PKTDIR_ARG,
|
|
&process_pktdir_arg,
|
|
ark) != 0) {
|
|
ARK_PMD_LOG(ERR, "Unable to parse arg %s\n", ARK_PKTDIR_ARG);
|
|
goto free_kvlist;
|
|
}
|
|
|
|
if (rte_kvargs_process(kvlist,
|
|
ARK_PKTGEN_ARG,
|
|
&process_file_args,
|
|
ark->pkt_gen_args) != 0) {
|
|
ARK_PMD_LOG(ERR, "Unable to parse arg %s\n", ARK_PKTGEN_ARG);
|
|
goto free_kvlist;
|
|
}
|
|
|
|
if (rte_kvargs_process(kvlist,
|
|
ARK_PKTCHKR_ARG,
|
|
&process_file_args,
|
|
ark->pkt_chkr_args) != 0) {
|
|
ARK_PMD_LOG(ERR, "Unable to parse arg %s\n", ARK_PKTCHKR_ARG);
|
|
goto free_kvlist;
|
|
}
|
|
|
|
ARK_PMD_LOG(INFO, "packet director set to 0x%x\n", ark->pkt_dir_v);
|
|
/* Setup the packet director */
|
|
ark_pktdir_setup(ark->pd, ark->pkt_dir_v);
|
|
|
|
/* Setup the packet generator */
|
|
if (ark->pkt_gen_args[0]) {
|
|
ARK_PMD_LOG(DEBUG, "Setting up the packet generator\n");
|
|
ark_pktgen_parse(ark->pkt_gen_args);
|
|
ark_pktgen_reset(ark->pg);
|
|
ark_pktgen_setup(ark->pg);
|
|
ark->start_pg = 1;
|
|
}
|
|
|
|
/* Setup the packet checker */
|
|
if (ark->pkt_chkr_args[0]) {
|
|
ark_pktchkr_parse(ark->pkt_chkr_args);
|
|
ark_pktchkr_setup(ark->pc);
|
|
}
|
|
|
|
ret = 0;
|
|
|
|
free_kvlist:
|
|
rte_kvargs_free(kvlist);
|
|
|
|
return ret;
|
|
}
|
|
|
|
RTE_PMD_REGISTER_PCI(net_ark, rte_ark_pmd);
|
|
RTE_PMD_REGISTER_KMOD_DEP(net_ark, "* igb_uio | uio_pci_generic ");
|
|
RTE_PMD_REGISTER_PCI_TABLE(net_ark, pci_id_ark_map);
|
|
RTE_PMD_REGISTER_PARAM_STRING(net_ark,
|
|
ARK_PKTGEN_ARG "=<filename> "
|
|
ARK_PKTCHKR_ARG "=<filename> "
|
|
ARK_PKTDIR_ARG "=<bitmap>");
|
|
RTE_LOG_REGISTER(ark_logtype, pmd.net.ark, NOTICE);
|