numam-dpdk/drivers/net/dpaa2/dpaa2_mux.c
Thomas Monjalon b7ed955a20 ethdev: deprecate legacy filter API
As stated in the deprecation notice from December 2016,
"the legacy filter API, including rte_eth_dev_filter_supported(),
rte_eth_dev_filter_ctrl() as well as filter types MACVLAN, ETHERTYPE,
FLEXIBLE, SYN, NTUPLE, TUNNEL, FDIR, HASH and L2_TUNNEL, is superseded
by the generic flow API (rte_flow)".

After a long wait of more than two years, the legacy filter API
is marked as deprecated, while still tested with testpmd and
the tep_termination example.

The next step will be to announce a deadline for complete removal.
As preparation of the removal of rte_eth_ctrl.h,
RTE_ETH_FLOW_*, RTE_TUNNEL_TYPE_* and RTE_ETH_HASH_FUNCTION_* definitions
are moved to rte_ethdev.h and rte_flow.h.

Signed-off-by: Thomas Monjalon <thomas@monjalon.net>
Acked-by: Shahaf Shuler <shahafs@mellanox.com>
Acked-by: Andrew Rybchenko <arybchenko@solarflare.com>
Acked-by: Jerin Jacob <jerinj@marvell.com>
Acked-by: Stephen Hemminger <stephen@networkplumber.org>
Acked-by: Hemant Agrawal <hemant.agrawal@nxp.com>
Acked-by: Adrien Mazarguil <adrien.mazarguil@6wind.com>
Acked-by: Ajit Khaparde <ajit.khaparde@broadcom.com>
2019-04-19 14:51:54 +02:00

222 lines
5.8 KiB
C

/* SPDX-License-Identifier: BSD-3-Clause
* Copyright 2018 NXP
*/
#include <sys/queue.h>
#include <stdio.h>
#include <errno.h>
#include <stdint.h>
#include <string.h>
#include <unistd.h>
#include <stdarg.h>
#include <rte_ethdev.h>
#include <rte_log.h>
#include <rte_malloc.h>
#include <rte_flow_driver.h>
#include <rte_tailq.h>
#include <rte_fslmc.h>
#include <fsl_dpdmux.h>
#include <fsl_dpkg.h>
#include <dpaa2_ethdev.h>
#include <dpaa2_pmd_logs.h>
struct dpaa2_dpdmux_dev {
TAILQ_ENTRY(dpaa2_dpdmux_dev) next;
/**< Pointer to Next device instance */
struct fsl_mc_io dpdmux; /** handle to DPDMUX portal object */
uint16_t token;
uint32_t dpdmux_id; /*HW ID for DPDMUX object */
uint8_t num_ifs; /* Number of interfaces in DPDMUX */
};
struct rte_flow {
struct dpdmux_rule_cfg rule;
};
TAILQ_HEAD(dpdmux_dev_list, dpaa2_dpdmux_dev);
static struct dpdmux_dev_list dpdmux_dev_list =
TAILQ_HEAD_INITIALIZER(dpdmux_dev_list); /*!< DPDMUX device list */
static struct dpaa2_dpdmux_dev *get_dpdmux_from_id(uint32_t dpdmux_id)
{
struct dpaa2_dpdmux_dev *dpdmux_dev = NULL;
/* Get DPBP dev handle from list using index */
TAILQ_FOREACH(dpdmux_dev, &dpdmux_dev_list, next) {
if (dpdmux_dev->dpdmux_id == dpdmux_id)
break;
}
return dpdmux_dev;
}
struct rte_flow *
rte_pmd_dpaa2_mux_flow_create(uint32_t dpdmux_id,
struct rte_flow_item *pattern[],
struct rte_flow_action *actions[])
{
struct dpaa2_dpdmux_dev *dpdmux_dev;
struct dpkg_profile_cfg kg_cfg;
const struct rte_flow_item_ipv4 *spec;
const struct rte_flow_action_vf *vf_conf;
struct dpdmux_cls_action dpdmux_action;
struct rte_flow *flow = NULL;
void *key_iova, *mask_iova, *key_cfg_iova = NULL;
int ret;
if (pattern[0]->type != RTE_FLOW_ITEM_TYPE_IPV4) {
DPAA2_PMD_ERR("Not supported pattern type: %d",
pattern[0]->type);
return NULL;
}
/* Find the DPDMUX from dpdmux_id in our list */
dpdmux_dev = get_dpdmux_from_id(dpdmux_id);
if (!dpdmux_dev) {
DPAA2_PMD_ERR("Invalid dpdmux_id: %d", dpdmux_id);
return NULL;
}
key_cfg_iova = rte_zmalloc(NULL, DIST_PARAM_IOVA_SIZE,
RTE_CACHE_LINE_SIZE);
if (!key_cfg_iova) {
DPAA2_PMD_ERR("Unable to allocate flow-dist parameters");
return NULL;
}
/* Currently taking only IP protocol as an extract type.
* This can be exended to other fields using pattern->type.
*/
memset(&kg_cfg, 0, sizeof(struct dpkg_profile_cfg));
kg_cfg.extracts[0].extract.from_hdr.prot = NET_PROT_IP;
kg_cfg.extracts[0].extract.from_hdr.field = NH_FLD_IP_PROTO;
kg_cfg.extracts[0].type = DPKG_EXTRACT_FROM_HDR;
kg_cfg.extracts[0].extract.from_hdr.type = DPKG_FULL_FIELD;
kg_cfg.num_extracts = 1;
ret = dpkg_prepare_key_cfg(&kg_cfg, key_cfg_iova);
if (ret) {
DPAA2_PMD_ERR("dpkg_prepare_key_cfg failed: err(%d)", ret);
goto creation_error;
}
ret = dpdmux_set_custom_key(&dpdmux_dev->dpdmux, CMD_PRI_LOW,
dpdmux_dev->token,
(uint64_t)(DPAA2_VADDR_TO_IOVA(key_cfg_iova)));
if (ret) {
DPAA2_PMD_ERR("dpdmux_set_custom_key failed: err(%d)", ret);
goto creation_error;
}
/* As now our key extract parameters are set, let us configure
* the rule.
*/
flow = rte_zmalloc(NULL, sizeof(struct rte_flow) +
(2 * DIST_PARAM_IOVA_SIZE), RTE_CACHE_LINE_SIZE);
if (!flow) {
DPAA2_PMD_ERR(
"Memory allocation failure for rule configration\n");
goto creation_error;
}
key_iova = (void *)((size_t)flow + sizeof(struct rte_flow));
mask_iova = (void *)((size_t)key_iova + DIST_PARAM_IOVA_SIZE);
spec = (const struct rte_flow_item_ipv4 *)pattern[0]->spec;
memcpy(key_iova, (const void *)&spec->hdr.next_proto_id,
sizeof(uint8_t));
memcpy(mask_iova, pattern[0]->mask, sizeof(uint8_t));
flow->rule.key_iova = (uint64_t)(DPAA2_VADDR_TO_IOVA(key_iova));
flow->rule.mask_iova = (uint64_t)(DPAA2_VADDR_TO_IOVA(mask_iova));
flow->rule.key_size = sizeof(uint8_t);
vf_conf = (const struct rte_flow_action_vf *)(actions[0]->conf);
if (vf_conf->id == 0 || vf_conf->id > dpdmux_dev->num_ifs) {
DPAA2_PMD_ERR("Invalid destination id\n");
goto creation_error;
}
dpdmux_action.dest_if = vf_conf->id;
ret = dpdmux_add_custom_cls_entry(&dpdmux_dev->dpdmux, CMD_PRI_LOW,
dpdmux_dev->token, &flow->rule,
&dpdmux_action);
if (ret) {
DPAA2_PMD_ERR("dpdmux_add_custom_cls_entry failed: err(%d)",
ret);
goto creation_error;
}
return flow;
creation_error:
rte_free((void *)key_cfg_iova);
rte_free((void *)flow);
return NULL;
}
static int
dpaa2_create_dpdmux_device(int vdev_fd __rte_unused,
struct vfio_device_info *obj_info __rte_unused,
int dpdmux_id)
{
struct dpaa2_dpdmux_dev *dpdmux_dev;
struct dpdmux_attr attr;
int ret;
PMD_INIT_FUNC_TRACE();
/* Allocate DPAA2 dpdmux handle */
dpdmux_dev = rte_malloc(NULL, sizeof(struct dpaa2_dpdmux_dev), 0);
if (!dpdmux_dev) {
DPAA2_PMD_ERR("Memory allocation failed for DPDMUX Device");
return -1;
}
/* Open the dpdmux object */
dpdmux_dev->dpdmux.regs = rte_mcp_ptr_list[MC_PORTAL_INDEX];
ret = dpdmux_open(&dpdmux_dev->dpdmux, CMD_PRI_LOW, dpdmux_id,
&dpdmux_dev->token);
if (ret) {
DPAA2_PMD_ERR("Unable to open dpdmux object: err(%d)", ret);
goto init_err;
}
ret = dpdmux_get_attributes(&dpdmux_dev->dpdmux, CMD_PRI_LOW,
dpdmux_dev->token, &attr);
if (ret) {
DPAA2_PMD_ERR("Unable to get dpdmux attr: err(%d)", ret);
goto init_err;
}
ret = dpdmux_if_set_default(&dpdmux_dev->dpdmux, CMD_PRI_LOW,
dpdmux_dev->token, 1);
if (ret) {
DPAA2_PMD_ERR("setting default interface failed in %s",
__func__);
goto init_err;
}
dpdmux_dev->dpdmux_id = dpdmux_id;
dpdmux_dev->num_ifs = attr.num_ifs;
TAILQ_INSERT_TAIL(&dpdmux_dev_list, dpdmux_dev, next);
return 0;
init_err:
if (dpdmux_dev)
rte_free(dpdmux_dev);
return -1;
}
static struct rte_dpaa2_object rte_dpaa2_dpdmux_obj = {
.dev_type = DPAA2_MUX,
.create = dpaa2_create_dpdmux_device,
};
RTE_PMD_REGISTER_DPAA2_OBJECT(dpdmux, rte_dpaa2_dpdmux_obj);