71e9b334a2
To avoid name collisions, add cxgbe_ prefix to some global functions. Also, make some local functions static in cxgbe_filter.c. Fixes:ee61f5113b
("net/cxgbe: parse and validate flows") Fixes:9eb2c9a480
("net/cxgbe: implement flow create operation") Fixes:3a381a4116
("net/cxgbe: query firmware for HASH filter resources") Fixes:af44a57798
("net/cxgbe: support to offload flows to HASH region") Fixes:41dc98b082
("net/cxgbe: support to delete flows in HASH region") Fixes:23af667f15
("net/cxgbe: add API to program hardware layer 2 table") Cc: stable@dpdk.org Signed-off-by: Rahul Lakkireddy <rahul.lakkireddy@chelsio.com>
230 lines
5.4 KiB
C
230 lines
5.4 KiB
C
/* SPDX-License-Identifier: BSD-3-Clause
|
|
* Copyright(c) 2018 Chelsio Communications.
|
|
* All rights reserved.
|
|
*/
|
|
|
|
#include "base/common.h"
|
|
#include "l2t.h"
|
|
|
|
/**
|
|
* cxgbe_l2t_release - Release associated L2T entry
|
|
* @e: L2T entry to release
|
|
*
|
|
* Releases ref count and frees up an L2T entry from L2T table
|
|
*/
|
|
void cxgbe_l2t_release(struct l2t_entry *e)
|
|
{
|
|
if (rte_atomic32_read(&e->refcnt) != 0)
|
|
rte_atomic32_dec(&e->refcnt);
|
|
}
|
|
|
|
/**
|
|
* Process a CPL_L2T_WRITE_RPL. Note that the TID in the reply is really
|
|
* the L2T index it refers to.
|
|
*/
|
|
void cxgbe_do_l2t_write_rpl(struct adapter *adap,
|
|
const struct cpl_l2t_write_rpl *rpl)
|
|
{
|
|
struct l2t_data *d = adap->l2t;
|
|
unsigned int tid = GET_TID(rpl);
|
|
unsigned int l2t_idx = tid % L2T_SIZE;
|
|
|
|
if (unlikely(rpl->status != CPL_ERR_NONE)) {
|
|
dev_err(adap,
|
|
"Unexpected L2T_WRITE_RPL status %u for entry %u\n",
|
|
rpl->status, l2t_idx);
|
|
return;
|
|
}
|
|
|
|
if (tid & F_SYNC_WR) {
|
|
struct l2t_entry *e = &d->l2tab[l2t_idx - d->l2t_start];
|
|
|
|
t4_os_lock(&e->lock);
|
|
if (e->state != L2T_STATE_SWITCHING)
|
|
e->state = L2T_STATE_VALID;
|
|
t4_os_unlock(&e->lock);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Write an L2T entry. Must be called with the entry locked.
|
|
* The write may be synchronous or asynchronous.
|
|
*/
|
|
static int write_l2e(struct rte_eth_dev *dev, struct l2t_entry *e, int sync,
|
|
bool loopback, bool arpmiss)
|
|
{
|
|
struct adapter *adap = ethdev2adap(dev);
|
|
struct l2t_data *d = adap->l2t;
|
|
struct rte_mbuf *mbuf;
|
|
struct cpl_l2t_write_req *req;
|
|
struct sge_ctrl_txq *ctrlq;
|
|
unsigned int l2t_idx = e->idx + d->l2t_start;
|
|
unsigned int port_id = ethdev2pinfo(dev)->port_id;
|
|
|
|
ctrlq = &adap->sge.ctrlq[port_id];
|
|
mbuf = rte_pktmbuf_alloc(ctrlq->mb_pool);
|
|
if (!mbuf)
|
|
return -ENOMEM;
|
|
|
|
mbuf->data_len = sizeof(*req);
|
|
mbuf->pkt_len = mbuf->data_len;
|
|
|
|
req = rte_pktmbuf_mtod(mbuf, struct cpl_l2t_write_req *);
|
|
INIT_TP_WR(req, 0);
|
|
|
|
OPCODE_TID(req) =
|
|
cpu_to_be32(MK_OPCODE_TID(CPL_L2T_WRITE_REQ,
|
|
l2t_idx | V_SYNC_WR(sync) |
|
|
V_TID_QID(adap->sge.fw_evtq.abs_id)));
|
|
req->params = cpu_to_be16(V_L2T_W_PORT(e->lport) |
|
|
V_L2T_W_LPBK(loopback) |
|
|
V_L2T_W_ARPMISS(arpmiss) |
|
|
V_L2T_W_NOREPLY(!sync));
|
|
req->l2t_idx = cpu_to_be16(l2t_idx);
|
|
req->vlan = cpu_to_be16(e->vlan);
|
|
rte_memcpy(req->dst_mac, e->dmac, RTE_ETHER_ADDR_LEN);
|
|
|
|
if (loopback)
|
|
memset(req->dst_mac, 0, RTE_ETHER_ADDR_LEN);
|
|
|
|
t4_mgmt_tx(ctrlq, mbuf);
|
|
|
|
if (sync && e->state != L2T_STATE_SWITCHING)
|
|
e->state = L2T_STATE_SYNC_WRITE;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* find_or_alloc_l2e - Find/Allocate a free L2T entry
|
|
* @d: L2T table
|
|
* @vlan: VLAN id to compare/add
|
|
* @port: port id to compare/add
|
|
* @dmac: Destination MAC address to compare/add
|
|
* Returns pointer to the L2T entry found/created
|
|
*
|
|
* Finds/Allocates an L2T entry to be used by switching rule of a filter.
|
|
*/
|
|
static struct l2t_entry *find_or_alloc_l2e(struct l2t_data *d, u16 vlan,
|
|
u8 port, u8 *dmac)
|
|
{
|
|
struct l2t_entry *end, *e;
|
|
struct l2t_entry *first_free = NULL;
|
|
|
|
for (e = &d->l2tab[0], end = &d->l2tab[d->l2t_size]; e != end; ++e) {
|
|
if (rte_atomic32_read(&e->refcnt) == 0) {
|
|
if (!first_free)
|
|
first_free = e;
|
|
} else {
|
|
if (e->state == L2T_STATE_SWITCHING) {
|
|
if ((!memcmp(e->dmac, dmac, RTE_ETHER_ADDR_LEN)) &&
|
|
e->vlan == vlan && e->lport == port)
|
|
goto exists;
|
|
}
|
|
}
|
|
}
|
|
|
|
if (first_free) {
|
|
e = first_free;
|
|
goto found;
|
|
}
|
|
|
|
return NULL;
|
|
|
|
found:
|
|
e->state = L2T_STATE_UNUSED;
|
|
|
|
exists:
|
|
return e;
|
|
}
|
|
|
|
static struct l2t_entry *t4_l2t_alloc_switching(struct rte_eth_dev *dev,
|
|
u16 vlan, u8 port,
|
|
u8 *eth_addr)
|
|
{
|
|
struct adapter *adap = ethdev2adap(dev);
|
|
struct l2t_data *d = adap->l2t;
|
|
struct l2t_entry *e;
|
|
int ret = 0;
|
|
|
|
t4_os_write_lock(&d->lock);
|
|
e = find_or_alloc_l2e(d, vlan, port, eth_addr);
|
|
if (e) {
|
|
t4_os_lock(&e->lock);
|
|
if (!rte_atomic32_read(&e->refcnt)) {
|
|
e->state = L2T_STATE_SWITCHING;
|
|
e->vlan = vlan;
|
|
e->lport = port;
|
|
rte_memcpy(e->dmac, eth_addr, RTE_ETHER_ADDR_LEN);
|
|
rte_atomic32_set(&e->refcnt, 1);
|
|
ret = write_l2e(dev, e, 0, !L2T_LPBK, !L2T_ARPMISS);
|
|
if (ret < 0)
|
|
dev_debug(adap, "Failed to write L2T entry: %d",
|
|
ret);
|
|
} else {
|
|
rte_atomic32_inc(&e->refcnt);
|
|
}
|
|
t4_os_unlock(&e->lock);
|
|
}
|
|
t4_os_write_unlock(&d->lock);
|
|
|
|
return ret ? NULL : e;
|
|
}
|
|
|
|
/**
|
|
* cxgbe_l2t_alloc_switching - Allocate a L2T entry for switching rule
|
|
* @dev: rte_eth_dev pointer
|
|
* @vlan: VLAN Id
|
|
* @port: Associated port
|
|
* @dmac: Destination MAC address to add to L2T
|
|
* Returns pointer to the allocated l2t entry
|
|
*
|
|
* Allocates a L2T entry for use by switching rule of a filter
|
|
*/
|
|
struct l2t_entry *cxgbe_l2t_alloc_switching(struct rte_eth_dev *dev, u16 vlan,
|
|
u8 port, u8 *dmac)
|
|
{
|
|
return t4_l2t_alloc_switching(dev, vlan, port, dmac);
|
|
}
|
|
|
|
/**
|
|
* Initialize L2 Table
|
|
*/
|
|
struct l2t_data *t4_init_l2t(unsigned int l2t_start, unsigned int l2t_end)
|
|
{
|
|
unsigned int l2t_size;
|
|
unsigned int i;
|
|
struct l2t_data *d;
|
|
|
|
if (l2t_start >= l2t_end || l2t_end >= L2T_SIZE)
|
|
return NULL;
|
|
l2t_size = l2t_end - l2t_start + 1;
|
|
|
|
d = t4_os_alloc(sizeof(*d) + l2t_size * sizeof(struct l2t_entry));
|
|
if (!d)
|
|
return NULL;
|
|
|
|
d->l2t_start = l2t_start;
|
|
d->l2t_size = l2t_size;
|
|
|
|
t4_os_rwlock_init(&d->lock);
|
|
|
|
for (i = 0; i < d->l2t_size; ++i) {
|
|
d->l2tab[i].idx = i;
|
|
d->l2tab[i].state = L2T_STATE_UNUSED;
|
|
t4_os_lock_init(&d->l2tab[i].lock);
|
|
rte_atomic32_set(&d->l2tab[i].refcnt, 0);
|
|
}
|
|
|
|
return d;
|
|
}
|
|
|
|
/**
|
|
* Cleanup L2 Table
|
|
*/
|
|
void t4_cleanup_l2t(struct adapter *adap)
|
|
{
|
|
if (adap->l2t)
|
|
t4_os_free(adap->l2t);
|
|
}
|