aa07bf8fa7
The function names for converting between RQ indexes known to the RTE code and internal RQ indexes for primary Start of Packet (SOP) queues and spill-over (Data) queues was unclear and confusing. Clarify with more explicit function names. Signed-off-by: John Daley <johndale@cisco.com> Reviewed-by: Nelson Escobar <neescoba@cisco.com>
520 lines
15 KiB
C
520 lines
15 KiB
C
/*
|
|
* Copyright 2008-2014 Cisco Systems, Inc. All rights reserved.
|
|
* Copyright 2007 Nuova Systems, Inc. All rights reserved.
|
|
*
|
|
* Copyright (c) 2014, Cisco Systems, Inc.
|
|
* All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
*
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
*
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in
|
|
* the documentation and/or other materials provided with the
|
|
* distribution.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
|
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
|
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS
|
|
* FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE
|
|
* COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT,
|
|
* INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING,
|
|
* BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
|
|
* LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
|
|
* CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
|
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN
|
|
* ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
|
|
* POSSIBILITY OF SUCH DAMAGE.
|
|
*
|
|
*/
|
|
|
|
#include <libgen.h>
|
|
|
|
#include <rte_ethdev.h>
|
|
#include <rte_malloc.h>
|
|
#include <rte_hash.h>
|
|
#include <rte_byteorder.h>
|
|
#include <rte_ip.h>
|
|
#include <rte_tcp.h>
|
|
#include <rte_udp.h>
|
|
#include <rte_sctp.h>
|
|
#include <rte_eth_ctrl.h>
|
|
|
|
#include "enic_compat.h"
|
|
#include "enic.h"
|
|
#include "wq_enet_desc.h"
|
|
#include "rq_enet_desc.h"
|
|
#include "cq_enet_desc.h"
|
|
#include "vnic_enet.h"
|
|
#include "vnic_dev.h"
|
|
#include "vnic_wq.h"
|
|
#include "vnic_rq.h"
|
|
#include "vnic_cq.h"
|
|
#include "vnic_intr.h"
|
|
#include "vnic_nic.h"
|
|
|
|
#ifdef RTE_MACHINE_CPUFLAG_SSE4_2
|
|
#include <rte_hash_crc.h>
|
|
#define DEFAULT_HASH_FUNC rte_hash_crc
|
|
#else
|
|
#include <rte_jhash.h>
|
|
#define DEFAULT_HASH_FUNC rte_jhash
|
|
#endif
|
|
|
|
#define ENICPMD_CLSF_HASH_ENTRIES ENICPMD_FDIR_MAX
|
|
|
|
void enic_fdir_stats_get(struct enic *enic, struct rte_eth_fdir_stats *stats)
|
|
{
|
|
*stats = enic->fdir.stats;
|
|
}
|
|
|
|
void enic_fdir_info_get(struct enic *enic, struct rte_eth_fdir_info *info)
|
|
{
|
|
info->mode = (enum rte_fdir_mode)enic->fdir.modes;
|
|
info->flow_types_mask[0] = enic->fdir.types_mask;
|
|
}
|
|
|
|
void enic_fdir_info(struct enic *enic)
|
|
{
|
|
enic->fdir.modes = (u32)RTE_FDIR_MODE_PERFECT;
|
|
enic->fdir.types_mask = 1 << RTE_ETH_FLOW_NONFRAG_IPV4_UDP |
|
|
1 << RTE_ETH_FLOW_NONFRAG_IPV4_TCP;
|
|
if (enic->adv_filters) {
|
|
enic->fdir.types_mask |= 1 << RTE_ETH_FLOW_NONFRAG_IPV4_OTHER |
|
|
1 << RTE_ETH_FLOW_NONFRAG_IPV4_SCTP |
|
|
1 << RTE_ETH_FLOW_NONFRAG_IPV6_UDP |
|
|
1 << RTE_ETH_FLOW_NONFRAG_IPV6_TCP |
|
|
1 << RTE_ETH_FLOW_NONFRAG_IPV6_SCTP |
|
|
1 << RTE_ETH_FLOW_NONFRAG_IPV6_OTHER;
|
|
enic->fdir.copy_fltr_fn = copy_fltr_v2;
|
|
} else {
|
|
enic->fdir.copy_fltr_fn = copy_fltr_v1;
|
|
}
|
|
}
|
|
|
|
static void
|
|
enic_set_layer(struct filter_generic_1 *gp, unsigned int flag,
|
|
enum filter_generic_1_layer layer, void *mask, void *val,
|
|
unsigned int len)
|
|
{
|
|
gp->mask_flags |= flag;
|
|
gp->val_flags |= gp->mask_flags;
|
|
memcpy(gp->layer[layer].mask, mask, len);
|
|
memcpy(gp->layer[layer].val, val, len);
|
|
}
|
|
|
|
/* Copy Flow Director filter to a VIC ipv4 filter (for Cisco VICs
|
|
* without advanced filter support.
|
|
*/
|
|
void
|
|
copy_fltr_v1(struct filter_v2 *fltr, struct rte_eth_fdir_input *input,
|
|
__rte_unused struct rte_eth_fdir_masks *masks)
|
|
{
|
|
fltr->type = FILTER_IPV4_5TUPLE;
|
|
fltr->u.ipv4.src_addr = rte_be_to_cpu_32(
|
|
input->flow.ip4_flow.src_ip);
|
|
fltr->u.ipv4.dst_addr = rte_be_to_cpu_32(
|
|
input->flow.ip4_flow.dst_ip);
|
|
fltr->u.ipv4.src_port = rte_be_to_cpu_16(
|
|
input->flow.udp4_flow.src_port);
|
|
fltr->u.ipv4.dst_port = rte_be_to_cpu_16(
|
|
input->flow.udp4_flow.dst_port);
|
|
|
|
if (input->flow_type == RTE_ETH_FLOW_NONFRAG_IPV4_TCP)
|
|
fltr->u.ipv4.protocol = PROTO_TCP;
|
|
else
|
|
fltr->u.ipv4.protocol = PROTO_UDP;
|
|
|
|
fltr->u.ipv4.flags = FILTER_FIELDS_IPV4_5TUPLE;
|
|
}
|
|
|
|
/* Copy Flow Director filter to a VIC generic filter (requires advanced
|
|
* filter support.
|
|
*/
|
|
void
|
|
copy_fltr_v2(struct filter_v2 *fltr, struct rte_eth_fdir_input *input,
|
|
struct rte_eth_fdir_masks *masks)
|
|
{
|
|
struct filter_generic_1 *gp = &fltr->u.generic_1;
|
|
int i;
|
|
|
|
fltr->type = FILTER_DPDK_1;
|
|
memset(gp, 0, sizeof(*gp));
|
|
|
|
if (input->flow_type == RTE_ETH_FLOW_NONFRAG_IPV4_UDP) {
|
|
struct udp_hdr udp_mask, udp_val;
|
|
memset(&udp_mask, 0, sizeof(udp_mask));
|
|
memset(&udp_val, 0, sizeof(udp_val));
|
|
|
|
if (input->flow.udp4_flow.src_port) {
|
|
udp_mask.src_port = masks->src_port_mask;
|
|
udp_val.src_port = input->flow.udp4_flow.src_port;
|
|
}
|
|
if (input->flow.udp4_flow.dst_port) {
|
|
udp_mask.src_port = masks->dst_port_mask;
|
|
udp_val.dst_port = input->flow.udp4_flow.dst_port;
|
|
}
|
|
|
|
enic_set_layer(gp, FILTER_GENERIC_1_UDP, FILTER_GENERIC_1_L4,
|
|
&udp_mask, &udp_val, sizeof(struct udp_hdr));
|
|
} else if (input->flow_type == RTE_ETH_FLOW_NONFRAG_IPV4_TCP) {
|
|
struct tcp_hdr tcp_mask, tcp_val;
|
|
memset(&tcp_mask, 0, sizeof(tcp_mask));
|
|
memset(&tcp_val, 0, sizeof(tcp_val));
|
|
|
|
if (input->flow.tcp4_flow.src_port) {
|
|
tcp_mask.src_port = masks->src_port_mask;
|
|
tcp_val.src_port = input->flow.tcp4_flow.src_port;
|
|
}
|
|
if (input->flow.tcp4_flow.dst_port) {
|
|
tcp_mask.dst_port = masks->dst_port_mask;
|
|
tcp_val.dst_port = input->flow.tcp4_flow.dst_port;
|
|
}
|
|
|
|
enic_set_layer(gp, FILTER_GENERIC_1_TCP, FILTER_GENERIC_1_L4,
|
|
&tcp_mask, &tcp_val, sizeof(struct tcp_hdr));
|
|
} else if (input->flow_type == RTE_ETH_FLOW_NONFRAG_IPV4_SCTP) {
|
|
struct sctp_hdr sctp_mask, sctp_val;
|
|
memset(&sctp_mask, 0, sizeof(sctp_mask));
|
|
memset(&sctp_val, 0, sizeof(sctp_val));
|
|
|
|
if (input->flow.sctp4_flow.src_port) {
|
|
sctp_mask.src_port = masks->src_port_mask;
|
|
sctp_val.src_port = input->flow.sctp4_flow.src_port;
|
|
}
|
|
if (input->flow.sctp4_flow.dst_port) {
|
|
sctp_mask.dst_port = masks->dst_port_mask;
|
|
sctp_val.dst_port = input->flow.sctp4_flow.dst_port;
|
|
}
|
|
if (input->flow.sctp4_flow.verify_tag) {
|
|
sctp_mask.tag = 0xffffffff;
|
|
sctp_val.tag = input->flow.sctp4_flow.verify_tag;
|
|
}
|
|
|
|
/* v4 proto should be 132, override ip4_flow.proto */
|
|
input->flow.ip4_flow.proto = 132;
|
|
|
|
enic_set_layer(gp, 0, FILTER_GENERIC_1_L4, &sctp_mask,
|
|
&sctp_val, sizeof(struct sctp_hdr));
|
|
}
|
|
|
|
if (input->flow_type == RTE_ETH_FLOW_NONFRAG_IPV4_UDP ||
|
|
input->flow_type == RTE_ETH_FLOW_NONFRAG_IPV4_TCP ||
|
|
input->flow_type == RTE_ETH_FLOW_NONFRAG_IPV4_SCTP ||
|
|
input->flow_type == RTE_ETH_FLOW_NONFRAG_IPV4_OTHER) {
|
|
struct ipv4_hdr ip4_mask, ip4_val;
|
|
memset(&ip4_mask, 0, sizeof(struct ipv4_hdr));
|
|
memset(&ip4_val, 0, sizeof(struct ipv4_hdr));
|
|
|
|
if (input->flow.ip4_flow.tos) {
|
|
ip4_mask.type_of_service = 0xff;
|
|
ip4_val.type_of_service = input->flow.ip4_flow.tos;
|
|
}
|
|
if (input->flow.ip4_flow.ttl) {
|
|
ip4_mask.time_to_live = 0xff;
|
|
ip4_val.time_to_live = input->flow.ip4_flow.ttl;
|
|
}
|
|
if (input->flow.ip4_flow.proto) {
|
|
ip4_mask.next_proto_id = 0xff;
|
|
ip4_val.next_proto_id = input->flow.ip4_flow.proto;
|
|
}
|
|
if (input->flow.ip4_flow.src_ip) {
|
|
ip4_mask.src_addr = masks->ipv4_mask.src_ip;
|
|
ip4_val.src_addr = input->flow.ip4_flow.src_ip;
|
|
}
|
|
if (input->flow.ip4_flow.dst_ip) {
|
|
ip4_mask.dst_addr = masks->ipv4_mask.dst_ip;
|
|
ip4_val.dst_addr = input->flow.ip4_flow.dst_ip;
|
|
}
|
|
|
|
enic_set_layer(gp, FILTER_GENERIC_1_IPV4, FILTER_GENERIC_1_L3,
|
|
&ip4_mask, &ip4_val, sizeof(struct ipv4_hdr));
|
|
}
|
|
|
|
if (input->flow_type == RTE_ETH_FLOW_NONFRAG_IPV6_UDP) {
|
|
struct udp_hdr udp_mask, udp_val;
|
|
memset(&udp_mask, 0, sizeof(udp_mask));
|
|
memset(&udp_val, 0, sizeof(udp_val));
|
|
|
|
if (input->flow.udp6_flow.src_port) {
|
|
udp_mask.src_port = masks->src_port_mask;
|
|
udp_val.src_port = input->flow.udp6_flow.src_port;
|
|
}
|
|
if (input->flow.udp6_flow.dst_port) {
|
|
udp_mask.dst_port = masks->dst_port_mask;
|
|
udp_val.dst_port = input->flow.udp6_flow.dst_port;
|
|
}
|
|
enic_set_layer(gp, FILTER_GENERIC_1_UDP, FILTER_GENERIC_1_L4,
|
|
&udp_mask, &udp_val, sizeof(struct udp_hdr));
|
|
} else if (input->flow_type == RTE_ETH_FLOW_NONFRAG_IPV6_TCP) {
|
|
struct tcp_hdr tcp_mask, tcp_val;
|
|
memset(&tcp_mask, 0, sizeof(tcp_mask));
|
|
memset(&tcp_val, 0, sizeof(tcp_val));
|
|
|
|
if (input->flow.tcp6_flow.src_port) {
|
|
tcp_mask.src_port = masks->src_port_mask;
|
|
tcp_val.src_port = input->flow.tcp6_flow.src_port;
|
|
}
|
|
if (input->flow.tcp6_flow.dst_port) {
|
|
tcp_mask.dst_port = masks->dst_port_mask;
|
|
tcp_val.dst_port = input->flow.tcp6_flow.dst_port;
|
|
}
|
|
enic_set_layer(gp, FILTER_GENERIC_1_TCP, FILTER_GENERIC_1_L4,
|
|
&tcp_mask, &tcp_val, sizeof(struct tcp_hdr));
|
|
} else if (input->flow_type == RTE_ETH_FLOW_NONFRAG_IPV6_SCTP) {
|
|
struct sctp_hdr sctp_mask, sctp_val;
|
|
memset(&sctp_mask, 0, sizeof(sctp_mask));
|
|
memset(&sctp_val, 0, sizeof(sctp_val));
|
|
|
|
if (input->flow.sctp6_flow.src_port) {
|
|
sctp_mask.src_port = masks->src_port_mask;
|
|
sctp_val.src_port = input->flow.sctp6_flow.src_port;
|
|
}
|
|
if (input->flow.sctp6_flow.dst_port) {
|
|
sctp_mask.dst_port = masks->dst_port_mask;
|
|
sctp_val.dst_port = input->flow.sctp6_flow.dst_port;
|
|
}
|
|
if (input->flow.sctp6_flow.verify_tag) {
|
|
sctp_mask.tag = 0xffffffff;
|
|
sctp_val.tag = input->flow.sctp6_flow.verify_tag;
|
|
}
|
|
|
|
/* v4 proto should be 132, override ipv6_flow.proto */
|
|
input->flow.ipv6_flow.proto = 132;
|
|
|
|
enic_set_layer(gp, 0, FILTER_GENERIC_1_L4, &sctp_mask,
|
|
&sctp_val, sizeof(struct sctp_hdr));
|
|
}
|
|
|
|
if (input->flow_type == RTE_ETH_FLOW_NONFRAG_IPV6_UDP ||
|
|
input->flow_type == RTE_ETH_FLOW_NONFRAG_IPV6_TCP ||
|
|
input->flow_type == RTE_ETH_FLOW_NONFRAG_IPV6_SCTP ||
|
|
input->flow_type == RTE_ETH_FLOW_NONFRAG_IPV6_OTHER) {
|
|
struct ipv6_hdr ipv6_mask, ipv6_val;
|
|
memset(&ipv6_mask, 0, sizeof(struct ipv6_hdr));
|
|
memset(&ipv6_val, 0, sizeof(struct ipv6_hdr));
|
|
|
|
if (input->flow.ipv6_flow.proto) {
|
|
ipv6_mask.proto = 0xff;
|
|
ipv6_val.proto = input->flow.ipv6_flow.proto;
|
|
}
|
|
for (i = 0; i < 4; i++) {
|
|
*(uint32_t *)&ipv6_mask.src_addr[i * 4] =
|
|
masks->ipv6_mask.src_ip[i];
|
|
*(uint32_t *)&ipv6_val.src_addr[i * 4] =
|
|
input->flow.ipv6_flow.src_ip[i];
|
|
}
|
|
for (i = 0; i < 4; i++) {
|
|
*(uint32_t *)&ipv6_mask.dst_addr[i * 4] =
|
|
masks->ipv6_mask.src_ip[i];
|
|
*(uint32_t *)&ipv6_val.dst_addr[i * 4] =
|
|
input->flow.ipv6_flow.dst_ip[i];
|
|
}
|
|
if (input->flow.ipv6_flow.tc) {
|
|
ipv6_mask.vtc_flow = 0x00ff0000;
|
|
ipv6_val.vtc_flow = input->flow.ipv6_flow.tc << 16;
|
|
}
|
|
if (input->flow.ipv6_flow.hop_limits) {
|
|
ipv6_mask.hop_limits = 0xff;
|
|
ipv6_val.hop_limits = input->flow.ipv6_flow.hop_limits;
|
|
}
|
|
|
|
enic_set_layer(gp, FILTER_GENERIC_1_IPV6, FILTER_GENERIC_1_L3,
|
|
&ipv6_mask, &ipv6_val, sizeof(struct ipv6_hdr));
|
|
}
|
|
}
|
|
|
|
int enic_fdir_del_fltr(struct enic *enic, struct rte_eth_fdir_filter *params)
|
|
{
|
|
int32_t pos;
|
|
struct enic_fdir_node *key;
|
|
/* See if the key is in the table */
|
|
pos = rte_hash_del_key(enic->fdir.hash, params);
|
|
switch (pos) {
|
|
case -EINVAL:
|
|
case -ENOENT:
|
|
enic->fdir.stats.f_remove++;
|
|
return -EINVAL;
|
|
default:
|
|
/* The entry is present in the table */
|
|
key = enic->fdir.nodes[pos];
|
|
|
|
/* Delete the filter */
|
|
vnic_dev_classifier(enic->vdev, CLSF_DEL,
|
|
&key->fltr_id, NULL);
|
|
rte_free(key);
|
|
enic->fdir.nodes[pos] = NULL;
|
|
enic->fdir.stats.free++;
|
|
enic->fdir.stats.remove++;
|
|
break;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
int enic_fdir_add_fltr(struct enic *enic, struct rte_eth_fdir_filter *params)
|
|
{
|
|
struct enic_fdir_node *key;
|
|
struct filter_v2 fltr;
|
|
int32_t pos;
|
|
u8 do_free = 0;
|
|
u16 old_fltr_id = 0;
|
|
u32 flowtype_supported;
|
|
u16 flex_bytes;
|
|
u16 queue;
|
|
|
|
memset(&fltr, 0, sizeof(fltr));
|
|
flowtype_supported = enic->fdir.types_mask
|
|
& (1 << params->input.flow_type);
|
|
|
|
flex_bytes = ((params->input.flow_ext.flexbytes[1] << 8 & 0xFF00) |
|
|
(params->input.flow_ext.flexbytes[0] & 0xFF));
|
|
|
|
if (!enic->fdir.hash ||
|
|
(params->input.flow_ext.vlan_tci & 0xFFF) ||
|
|
!flowtype_supported || flex_bytes ||
|
|
params->action.behavior /* drop */) {
|
|
enic->fdir.stats.f_add++;
|
|
return -ENOTSUP;
|
|
}
|
|
|
|
/* Get the enicpmd RQ from the DPDK Rx queue */
|
|
queue = enic_rte_rq_idx_to_sop_idx(params->action.rx_queue);
|
|
|
|
if (!enic->rq[queue].in_use)
|
|
return -EINVAL;
|
|
|
|
/* See if the key is already there in the table */
|
|
pos = rte_hash_del_key(enic->fdir.hash, params);
|
|
switch (pos) {
|
|
case -EINVAL:
|
|
enic->fdir.stats.f_add++;
|
|
return -EINVAL;
|
|
case -ENOENT:
|
|
/* Add a new classifier entry */
|
|
if (!enic->fdir.stats.free) {
|
|
enic->fdir.stats.f_add++;
|
|
return -ENOSPC;
|
|
}
|
|
key = rte_zmalloc("enic_fdir_node",
|
|
sizeof(struct enic_fdir_node), 0);
|
|
if (!key) {
|
|
enic->fdir.stats.f_add++;
|
|
return -ENOMEM;
|
|
}
|
|
break;
|
|
default:
|
|
/* The entry is already present in the table.
|
|
* Check if there is a change in queue
|
|
*/
|
|
key = enic->fdir.nodes[pos];
|
|
enic->fdir.nodes[pos] = NULL;
|
|
if (unlikely(key->rq_index == queue)) {
|
|
/* Nothing to be done */
|
|
enic->fdir.stats.f_add++;
|
|
pos = rte_hash_add_key(enic->fdir.hash, params);
|
|
if (pos < 0) {
|
|
dev_err(enic, "Add hash key failed\n");
|
|
return pos;
|
|
}
|
|
enic->fdir.nodes[pos] = key;
|
|
dev_warning(enic,
|
|
"FDIR rule is already present\n");
|
|
return 0;
|
|
}
|
|
|
|
if (likely(enic->fdir.stats.free)) {
|
|
/* Add the filter and then delete the old one.
|
|
* This is to avoid packets from going into the
|
|
* default queue during the window between
|
|
* delete and add
|
|
*/
|
|
do_free = 1;
|
|
old_fltr_id = key->fltr_id;
|
|
} else {
|
|
/* No free slots in the classifier.
|
|
* Delete the filter and add the modified one later
|
|
*/
|
|
vnic_dev_classifier(enic->vdev, CLSF_DEL,
|
|
&key->fltr_id, NULL);
|
|
enic->fdir.stats.free++;
|
|
}
|
|
|
|
break;
|
|
}
|
|
|
|
key->filter = *params;
|
|
key->rq_index = queue;
|
|
|
|
enic->fdir.copy_fltr_fn(&fltr, ¶ms->input,
|
|
&enic->rte_dev->data->dev_conf.fdir_conf.mask);
|
|
|
|
if (!vnic_dev_classifier(enic->vdev, CLSF_ADD, &queue, &fltr)) {
|
|
key->fltr_id = queue;
|
|
} else {
|
|
dev_err(enic, "Add classifier entry failed\n");
|
|
enic->fdir.stats.f_add++;
|
|
rte_free(key);
|
|
return -1;
|
|
}
|
|
|
|
if (do_free)
|
|
vnic_dev_classifier(enic->vdev, CLSF_DEL, &old_fltr_id, NULL);
|
|
else{
|
|
enic->fdir.stats.free--;
|
|
enic->fdir.stats.add++;
|
|
}
|
|
|
|
pos = rte_hash_add_key(enic->fdir.hash, params);
|
|
if (pos < 0) {
|
|
enic->fdir.stats.f_add++;
|
|
dev_err(enic, "Add hash key failed\n");
|
|
return pos;
|
|
}
|
|
|
|
enic->fdir.nodes[pos] = key;
|
|
return 0;
|
|
}
|
|
|
|
void enic_clsf_destroy(struct enic *enic)
|
|
{
|
|
u32 index;
|
|
struct enic_fdir_node *key;
|
|
/* delete classifier entries */
|
|
for (index = 0; index < ENICPMD_FDIR_MAX; index++) {
|
|
key = enic->fdir.nodes[index];
|
|
if (key) {
|
|
vnic_dev_classifier(enic->vdev, CLSF_DEL,
|
|
&key->fltr_id, NULL);
|
|
rte_free(key);
|
|
enic->fdir.nodes[index] = NULL;
|
|
}
|
|
}
|
|
|
|
if (enic->fdir.hash) {
|
|
rte_hash_free(enic->fdir.hash);
|
|
enic->fdir.hash = NULL;
|
|
}
|
|
}
|
|
|
|
int enic_clsf_init(struct enic *enic)
|
|
{
|
|
char clsf_name[RTE_HASH_NAMESIZE];
|
|
struct rte_hash_parameters hash_params = {
|
|
.name = clsf_name,
|
|
.entries = ENICPMD_CLSF_HASH_ENTRIES,
|
|
.key_len = sizeof(struct rte_eth_fdir_filter),
|
|
.hash_func = DEFAULT_HASH_FUNC,
|
|
.hash_func_init_val = 0,
|
|
.socket_id = SOCKET_ID_ANY,
|
|
};
|
|
snprintf(clsf_name, RTE_HASH_NAMESIZE, "enic_clsf_%s", enic->bdf_name);
|
|
enic->fdir.hash = rte_hash_create(&hash_params);
|
|
memset(&enic->fdir.stats, 0, sizeof(enic->fdir.stats));
|
|
enic->fdir.stats.free = ENICPMD_FDIR_MAX;
|
|
return NULL == enic->fdir.hash;
|
|
}
|