c3d30e9d75
Disable CQ_DISABLED error interrupt in NIX_LF_ERR_INT
to fix spurious interrupts in event dev mode. Also skip
configuring RSS when RQ count is '0' because
RSS table initialization is done incorrectly due to
divide-by-zero error and it is leading to RQ_OOR error
in NIX_LF_ERR_INT.
Fixes: 83ce2880e2
("net/octeontx2: support RSS")
Cc: stable@dpdk.org
Signed-off-by: Nithin Dabilpuram <ndabilpuram@marvell.com>
Acked-by: Jerin Jacob <jerinj@marvell.com>
495 lines
13 KiB
C
495 lines
13 KiB
C
/* SPDX-License-Identifier: BSD-3-Clause
|
|
* Copyright(C) 2019 Marvell International Ltd.
|
|
*/
|
|
|
|
#include <inttypes.h>
|
|
|
|
#include <rte_bus_pci.h>
|
|
#include <rte_malloc.h>
|
|
|
|
#include "otx2_ethdev.h"
|
|
|
|
static void
|
|
nix_lf_err_irq(void *param)
|
|
{
|
|
struct rte_eth_dev *eth_dev = (struct rte_eth_dev *)param;
|
|
struct otx2_eth_dev *dev = otx2_eth_pmd_priv(eth_dev);
|
|
uint64_t intr;
|
|
|
|
intr = otx2_read64(dev->base + NIX_LF_ERR_INT);
|
|
if (intr == 0)
|
|
return;
|
|
|
|
otx2_err("Err_intr=0x%" PRIx64 " pf=%d, vf=%d", intr, dev->pf, dev->vf);
|
|
|
|
/* Clear interrupt */
|
|
otx2_write64(intr, dev->base + NIX_LF_ERR_INT);
|
|
|
|
/* Dump registers to std out */
|
|
otx2_nix_reg_dump(dev, NULL);
|
|
otx2_nix_queues_ctx_dump(eth_dev);
|
|
}
|
|
|
|
static int
|
|
nix_lf_register_err_irq(struct rte_eth_dev *eth_dev)
|
|
{
|
|
struct rte_pci_device *pci_dev = RTE_ETH_DEV_TO_PCI(eth_dev);
|
|
struct rte_intr_handle *handle = &pci_dev->intr_handle;
|
|
struct otx2_eth_dev *dev = otx2_eth_pmd_priv(eth_dev);
|
|
int rc, vec;
|
|
|
|
vec = dev->nix_msixoff + NIX_LF_INT_VEC_ERR_INT;
|
|
|
|
/* Clear err interrupt */
|
|
otx2_nix_err_intr_enb_dis(eth_dev, false);
|
|
/* Set used interrupt vectors */
|
|
rc = otx2_register_irq(handle, nix_lf_err_irq, eth_dev, vec);
|
|
/* Enable all dev interrupt except for RQ_DISABLED */
|
|
otx2_nix_err_intr_enb_dis(eth_dev, true);
|
|
|
|
return rc;
|
|
}
|
|
|
|
static void
|
|
nix_lf_unregister_err_irq(struct rte_eth_dev *eth_dev)
|
|
{
|
|
struct rte_pci_device *pci_dev = RTE_ETH_DEV_TO_PCI(eth_dev);
|
|
struct rte_intr_handle *handle = &pci_dev->intr_handle;
|
|
struct otx2_eth_dev *dev = otx2_eth_pmd_priv(eth_dev);
|
|
int vec;
|
|
|
|
vec = dev->nix_msixoff + NIX_LF_INT_VEC_ERR_INT;
|
|
|
|
/* Clear err interrupt */
|
|
otx2_nix_err_intr_enb_dis(eth_dev, false);
|
|
otx2_unregister_irq(handle, nix_lf_err_irq, eth_dev, vec);
|
|
}
|
|
|
|
static void
|
|
nix_lf_ras_irq(void *param)
|
|
{
|
|
struct rte_eth_dev *eth_dev = (struct rte_eth_dev *)param;
|
|
struct otx2_eth_dev *dev = otx2_eth_pmd_priv(eth_dev);
|
|
uint64_t intr;
|
|
|
|
intr = otx2_read64(dev->base + NIX_LF_RAS);
|
|
if (intr == 0)
|
|
return;
|
|
|
|
otx2_err("Ras_intr=0x%" PRIx64 " pf=%d, vf=%d", intr, dev->pf, dev->vf);
|
|
|
|
/* Clear interrupt */
|
|
otx2_write64(intr, dev->base + NIX_LF_RAS);
|
|
|
|
/* Dump registers to std out */
|
|
otx2_nix_reg_dump(dev, NULL);
|
|
otx2_nix_queues_ctx_dump(eth_dev);
|
|
}
|
|
|
|
static int
|
|
nix_lf_register_ras_irq(struct rte_eth_dev *eth_dev)
|
|
{
|
|
struct rte_pci_device *pci_dev = RTE_ETH_DEV_TO_PCI(eth_dev);
|
|
struct rte_intr_handle *handle = &pci_dev->intr_handle;
|
|
struct otx2_eth_dev *dev = otx2_eth_pmd_priv(eth_dev);
|
|
int rc, vec;
|
|
|
|
vec = dev->nix_msixoff + NIX_LF_INT_VEC_POISON;
|
|
|
|
/* Clear err interrupt */
|
|
otx2_nix_ras_intr_enb_dis(eth_dev, false);
|
|
/* Set used interrupt vectors */
|
|
rc = otx2_register_irq(handle, nix_lf_ras_irq, eth_dev, vec);
|
|
/* Enable dev interrupt */
|
|
otx2_nix_ras_intr_enb_dis(eth_dev, true);
|
|
|
|
return rc;
|
|
}
|
|
|
|
static void
|
|
nix_lf_unregister_ras_irq(struct rte_eth_dev *eth_dev)
|
|
{
|
|
struct rte_pci_device *pci_dev = RTE_ETH_DEV_TO_PCI(eth_dev);
|
|
struct rte_intr_handle *handle = &pci_dev->intr_handle;
|
|
struct otx2_eth_dev *dev = otx2_eth_pmd_priv(eth_dev);
|
|
int vec;
|
|
|
|
vec = dev->nix_msixoff + NIX_LF_INT_VEC_POISON;
|
|
|
|
/* Clear err interrupt */
|
|
otx2_nix_ras_intr_enb_dis(eth_dev, false);
|
|
otx2_unregister_irq(handle, nix_lf_ras_irq, eth_dev, vec);
|
|
}
|
|
|
|
static inline uint8_t
|
|
nix_lf_q_irq_get_and_clear(struct otx2_eth_dev *dev, uint16_t q,
|
|
uint32_t off, uint64_t mask)
|
|
{
|
|
uint64_t reg, wdata;
|
|
uint8_t qint;
|
|
|
|
wdata = (uint64_t)q << 44;
|
|
reg = otx2_atomic64_add_nosync(wdata, (int64_t *)(dev->base + off));
|
|
|
|
if (reg & BIT_ULL(42) /* OP_ERR */) {
|
|
otx2_err("Failed execute irq get off=0x%x", off);
|
|
return 0;
|
|
}
|
|
|
|
qint = reg & 0xff;
|
|
wdata &= mask;
|
|
otx2_write64(wdata | qint, dev->base + off);
|
|
|
|
return qint;
|
|
}
|
|
|
|
static inline uint8_t
|
|
nix_lf_rq_irq_get_and_clear(struct otx2_eth_dev *dev, uint16_t rq)
|
|
{
|
|
return nix_lf_q_irq_get_and_clear(dev, rq, NIX_LF_RQ_OP_INT, ~0xff00);
|
|
}
|
|
|
|
static inline uint8_t
|
|
nix_lf_cq_irq_get_and_clear(struct otx2_eth_dev *dev, uint16_t cq)
|
|
{
|
|
return nix_lf_q_irq_get_and_clear(dev, cq, NIX_LF_CQ_OP_INT, ~0xff00);
|
|
}
|
|
|
|
static inline uint8_t
|
|
nix_lf_sq_irq_get_and_clear(struct otx2_eth_dev *dev, uint16_t sq)
|
|
{
|
|
return nix_lf_q_irq_get_and_clear(dev, sq, NIX_LF_SQ_OP_INT, ~0x1ff00);
|
|
}
|
|
|
|
static inline void
|
|
nix_lf_sq_debug_reg(struct otx2_eth_dev *dev, uint32_t off)
|
|
{
|
|
uint64_t reg;
|
|
|
|
reg = otx2_read64(dev->base + off);
|
|
if (reg & BIT_ULL(44))
|
|
otx2_err("SQ=%d err_code=0x%x",
|
|
(int)((reg >> 8) & 0xfffff), (uint8_t)(reg & 0xff));
|
|
}
|
|
|
|
static void
|
|
nix_lf_cq_irq(void *param)
|
|
{
|
|
struct otx2_qint *cint = (struct otx2_qint *)param;
|
|
struct rte_eth_dev *eth_dev = cint->eth_dev;
|
|
struct otx2_eth_dev *dev;
|
|
|
|
dev = otx2_eth_pmd_priv(eth_dev);
|
|
/* Clear interrupt */
|
|
otx2_write64(BIT_ULL(0), dev->base + NIX_LF_CINTX_INT(cint->qintx));
|
|
}
|
|
|
|
static void
|
|
nix_lf_q_irq(void *param)
|
|
{
|
|
struct otx2_qint *qint = (struct otx2_qint *)param;
|
|
struct rte_eth_dev *eth_dev = qint->eth_dev;
|
|
struct otx2_eth_dev *dev = otx2_eth_pmd_priv(eth_dev);
|
|
uint8_t irq, qintx = qint->qintx;
|
|
int q, cq, rq, sq;
|
|
uint64_t intr;
|
|
|
|
intr = otx2_read64(dev->base + NIX_LF_QINTX_INT(qintx));
|
|
if (intr == 0)
|
|
return;
|
|
|
|
otx2_err("Queue_intr=0x%" PRIx64 " qintx=%d pf=%d, vf=%d",
|
|
intr, qintx, dev->pf, dev->vf);
|
|
|
|
/* Handle RQ interrupts */
|
|
for (q = 0; q < eth_dev->data->nb_rx_queues; q++) {
|
|
rq = q % dev->qints;
|
|
irq = nix_lf_rq_irq_get_and_clear(dev, rq);
|
|
|
|
if (irq & BIT_ULL(NIX_RQINT_DROP))
|
|
otx2_err("RQ=%d NIX_RQINT_DROP", rq);
|
|
|
|
if (irq & BIT_ULL(NIX_RQINT_RED))
|
|
otx2_err("RQ=%d NIX_RQINT_RED", rq);
|
|
}
|
|
|
|
/* Handle CQ interrupts */
|
|
for (q = 0; q < eth_dev->data->nb_rx_queues; q++) {
|
|
cq = q % dev->qints;
|
|
irq = nix_lf_cq_irq_get_and_clear(dev, cq);
|
|
|
|
if (irq & BIT_ULL(NIX_CQERRINT_DOOR_ERR))
|
|
otx2_err("CQ=%d NIX_CQERRINT_DOOR_ERR", cq);
|
|
|
|
if (irq & BIT_ULL(NIX_CQERRINT_WR_FULL))
|
|
otx2_err("CQ=%d NIX_CQERRINT_WR_FULL", cq);
|
|
|
|
if (irq & BIT_ULL(NIX_CQERRINT_CQE_FAULT))
|
|
otx2_err("CQ=%d NIX_CQERRINT_CQE_FAULT", cq);
|
|
}
|
|
|
|
/* Handle SQ interrupts */
|
|
for (q = 0; q < eth_dev->data->nb_tx_queues; q++) {
|
|
sq = q % dev->qints;
|
|
irq = nix_lf_sq_irq_get_and_clear(dev, sq);
|
|
|
|
if (irq & BIT_ULL(NIX_SQINT_LMT_ERR)) {
|
|
otx2_err("SQ=%d NIX_SQINT_LMT_ERR", sq);
|
|
nix_lf_sq_debug_reg(dev, NIX_LF_SQ_OP_ERR_DBG);
|
|
}
|
|
if (irq & BIT_ULL(NIX_SQINT_MNQ_ERR)) {
|
|
otx2_err("SQ=%d NIX_SQINT_MNQ_ERR", sq);
|
|
nix_lf_sq_debug_reg(dev, NIX_LF_MNQ_ERR_DBG);
|
|
}
|
|
if (irq & BIT_ULL(NIX_SQINT_SEND_ERR)) {
|
|
otx2_err("SQ=%d NIX_SQINT_SEND_ERR", sq);
|
|
nix_lf_sq_debug_reg(dev, NIX_LF_SEND_ERR_DBG);
|
|
}
|
|
if (irq & BIT_ULL(NIX_SQINT_SQB_ALLOC_FAIL)) {
|
|
otx2_err("SQ=%d NIX_SQINT_SQB_ALLOC_FAIL", sq);
|
|
nix_lf_sq_debug_reg(dev, NIX_LF_SEND_ERR_DBG);
|
|
}
|
|
}
|
|
|
|
/* Clear interrupt */
|
|
otx2_write64(intr, dev->base + NIX_LF_QINTX_INT(qintx));
|
|
|
|
/* Dump registers to std out */
|
|
otx2_nix_reg_dump(dev, NULL);
|
|
otx2_nix_queues_ctx_dump(eth_dev);
|
|
}
|
|
|
|
int
|
|
oxt2_nix_register_queue_irqs(struct rte_eth_dev *eth_dev)
|
|
{
|
|
struct rte_pci_device *pci_dev = RTE_ETH_DEV_TO_PCI(eth_dev);
|
|
struct rte_intr_handle *handle = &pci_dev->intr_handle;
|
|
struct otx2_eth_dev *dev = otx2_eth_pmd_priv(eth_dev);
|
|
int vec, q, sqs, rqs, qs, rc = 0;
|
|
|
|
/* Figure out max qintx required */
|
|
rqs = RTE_MIN(dev->qints, eth_dev->data->nb_rx_queues);
|
|
sqs = RTE_MIN(dev->qints, eth_dev->data->nb_tx_queues);
|
|
qs = RTE_MAX(rqs, sqs);
|
|
|
|
dev->configured_qints = qs;
|
|
|
|
for (q = 0; q < qs; q++) {
|
|
vec = dev->nix_msixoff + NIX_LF_INT_VEC_QINT_START + q;
|
|
|
|
/* Clear QINT CNT */
|
|
otx2_write64(0, dev->base + NIX_LF_QINTX_CNT(q));
|
|
|
|
/* Clear interrupt */
|
|
otx2_write64(~0ull, dev->base + NIX_LF_QINTX_ENA_W1C(q));
|
|
|
|
dev->qints_mem[q].eth_dev = eth_dev;
|
|
dev->qints_mem[q].qintx = q;
|
|
|
|
/* Sync qints_mem update */
|
|
rte_smp_wmb();
|
|
|
|
/* Register queue irq vector */
|
|
rc = otx2_register_irq(handle, nix_lf_q_irq,
|
|
&dev->qints_mem[q], vec);
|
|
if (rc)
|
|
break;
|
|
|
|
otx2_write64(0, dev->base + NIX_LF_QINTX_CNT(q));
|
|
otx2_write64(0, dev->base + NIX_LF_QINTX_INT(q));
|
|
/* Enable QINT interrupt */
|
|
otx2_write64(~0ull, dev->base + NIX_LF_QINTX_ENA_W1S(q));
|
|
}
|
|
|
|
return rc;
|
|
}
|
|
|
|
void
|
|
oxt2_nix_unregister_queue_irqs(struct rte_eth_dev *eth_dev)
|
|
{
|
|
struct rte_pci_device *pci_dev = RTE_ETH_DEV_TO_PCI(eth_dev);
|
|
struct rte_intr_handle *handle = &pci_dev->intr_handle;
|
|
struct otx2_eth_dev *dev = otx2_eth_pmd_priv(eth_dev);
|
|
int vec, q;
|
|
|
|
for (q = 0; q < dev->configured_qints; q++) {
|
|
vec = dev->nix_msixoff + NIX_LF_INT_VEC_QINT_START + q;
|
|
|
|
/* Clear QINT CNT */
|
|
otx2_write64(0, dev->base + NIX_LF_QINTX_CNT(q));
|
|
otx2_write64(0, dev->base + NIX_LF_QINTX_INT(q));
|
|
|
|
/* Clear interrupt */
|
|
otx2_write64(~0ull, dev->base + NIX_LF_QINTX_ENA_W1C(q));
|
|
|
|
/* Unregister queue irq vector */
|
|
otx2_unregister_irq(handle, nix_lf_q_irq,
|
|
&dev->qints_mem[q], vec);
|
|
}
|
|
}
|
|
|
|
int
|
|
oxt2_nix_register_cq_irqs(struct rte_eth_dev *eth_dev)
|
|
{
|
|
struct rte_pci_device *pci_dev = RTE_ETH_DEV_TO_PCI(eth_dev);
|
|
struct rte_intr_handle *handle = &pci_dev->intr_handle;
|
|
struct otx2_eth_dev *dev = otx2_eth_pmd_priv(eth_dev);
|
|
uint8_t rc = 0, vec, q;
|
|
|
|
dev->configured_cints = RTE_MIN(dev->cints,
|
|
eth_dev->data->nb_rx_queues);
|
|
|
|
for (q = 0; q < dev->configured_cints; q++) {
|
|
vec = dev->nix_msixoff + NIX_LF_INT_VEC_CINT_START + q;
|
|
|
|
/* Clear CINT CNT */
|
|
otx2_write64(0, dev->base + NIX_LF_CINTX_CNT(q));
|
|
|
|
/* Clear interrupt */
|
|
otx2_write64(BIT_ULL(0), dev->base + NIX_LF_CINTX_ENA_W1C(q));
|
|
|
|
dev->cints_mem[q].eth_dev = eth_dev;
|
|
dev->cints_mem[q].qintx = q;
|
|
|
|
/* Sync cints_mem update */
|
|
rte_smp_wmb();
|
|
|
|
/* Register queue irq vector */
|
|
rc = otx2_register_irq(handle, nix_lf_cq_irq,
|
|
&dev->cints_mem[q], vec);
|
|
if (rc) {
|
|
otx2_err("Fail to register CQ irq, rc=%d", rc);
|
|
return rc;
|
|
}
|
|
|
|
if (!handle->intr_vec) {
|
|
handle->intr_vec = rte_zmalloc("intr_vec",
|
|
dev->configured_cints *
|
|
sizeof(int), 0);
|
|
if (!handle->intr_vec) {
|
|
otx2_err("Failed to allocate %d rx intr_vec",
|
|
dev->configured_cints);
|
|
return -ENOMEM;
|
|
}
|
|
}
|
|
/* VFIO vector zero is resereved for misc interrupt so
|
|
* doing required adjustment. (b13bfab4cd)
|
|
*/
|
|
handle->intr_vec[q] = RTE_INTR_VEC_RXTX_OFFSET + vec;
|
|
|
|
/* Configure CQE interrupt coalescing parameters */
|
|
otx2_write64(((CQ_CQE_THRESH_DEFAULT) |
|
|
(CQ_CQE_THRESH_DEFAULT << 32) |
|
|
(CQ_TIMER_THRESH_DEFAULT << 48)),
|
|
dev->base + NIX_LF_CINTX_WAIT((q)));
|
|
|
|
/* Keeping the CQ interrupt disabled as the rx interrupt
|
|
* feature needs to be enabled/disabled on demand.
|
|
*/
|
|
}
|
|
|
|
return rc;
|
|
}
|
|
|
|
void
|
|
oxt2_nix_unregister_cq_irqs(struct rte_eth_dev *eth_dev)
|
|
{
|
|
struct rte_pci_device *pci_dev = RTE_ETH_DEV_TO_PCI(eth_dev);
|
|
struct rte_intr_handle *handle = &pci_dev->intr_handle;
|
|
struct otx2_eth_dev *dev = otx2_eth_pmd_priv(eth_dev);
|
|
int vec, q;
|
|
|
|
for (q = 0; q < dev->configured_cints; q++) {
|
|
vec = dev->nix_msixoff + NIX_LF_INT_VEC_CINT_START + q;
|
|
|
|
/* Clear CINT CNT */
|
|
otx2_write64(0, dev->base + NIX_LF_CINTX_CNT(q));
|
|
|
|
/* Clear interrupt */
|
|
otx2_write64(BIT_ULL(0), dev->base + NIX_LF_CINTX_ENA_W1C(q));
|
|
|
|
/* Unregister queue irq vector */
|
|
otx2_unregister_irq(handle, nix_lf_cq_irq,
|
|
&dev->cints_mem[q], vec);
|
|
}
|
|
}
|
|
|
|
int
|
|
otx2_nix_register_irqs(struct rte_eth_dev *eth_dev)
|
|
{
|
|
struct otx2_eth_dev *dev = otx2_eth_pmd_priv(eth_dev);
|
|
int rc;
|
|
|
|
if (dev->nix_msixoff == MSIX_VECTOR_INVALID) {
|
|
otx2_err("Invalid NIXLF MSIX vector offset vector: 0x%x",
|
|
dev->nix_msixoff);
|
|
return -EINVAL;
|
|
}
|
|
|
|
/* Register lf err interrupt */
|
|
rc = nix_lf_register_err_irq(eth_dev);
|
|
/* Register RAS interrupt */
|
|
rc |= nix_lf_register_ras_irq(eth_dev);
|
|
|
|
return rc;
|
|
}
|
|
|
|
void
|
|
otx2_nix_unregister_irqs(struct rte_eth_dev *eth_dev)
|
|
{
|
|
nix_lf_unregister_err_irq(eth_dev);
|
|
nix_lf_unregister_ras_irq(eth_dev);
|
|
}
|
|
|
|
int
|
|
otx2_nix_rx_queue_intr_enable(struct rte_eth_dev *eth_dev,
|
|
uint16_t rx_queue_id)
|
|
{
|
|
struct otx2_eth_dev *dev = otx2_eth_pmd_priv(eth_dev);
|
|
|
|
/* Enable CINT interrupt */
|
|
otx2_write64(BIT_ULL(0), dev->base +
|
|
NIX_LF_CINTX_ENA_W1S(rx_queue_id));
|
|
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
otx2_nix_rx_queue_intr_disable(struct rte_eth_dev *eth_dev,
|
|
uint16_t rx_queue_id)
|
|
{
|
|
struct otx2_eth_dev *dev = otx2_eth_pmd_priv(eth_dev);
|
|
|
|
/* Clear and disable CINT interrupt */
|
|
otx2_write64(BIT_ULL(0), dev->base +
|
|
NIX_LF_CINTX_ENA_W1C(rx_queue_id));
|
|
|
|
return 0;
|
|
}
|
|
|
|
void
|
|
otx2_nix_err_intr_enb_dis(struct rte_eth_dev *eth_dev, bool enb)
|
|
{
|
|
struct otx2_eth_dev *dev = otx2_eth_pmd_priv(eth_dev);
|
|
|
|
/* Enable all nix lf error interrupts except
|
|
* RQ_DISABLED and CQ_DISABLED.
|
|
*/
|
|
if (enb)
|
|
otx2_write64(~(BIT_ULL(11) | BIT_ULL(24)),
|
|
dev->base + NIX_LF_ERR_INT_ENA_W1S);
|
|
else
|
|
otx2_write64(~0ull, dev->base + NIX_LF_ERR_INT_ENA_W1C);
|
|
}
|
|
|
|
void
|
|
otx2_nix_ras_intr_enb_dis(struct rte_eth_dev *eth_dev, bool enb)
|
|
{
|
|
struct otx2_eth_dev *dev = otx2_eth_pmd_priv(eth_dev);
|
|
|
|
if (enb)
|
|
otx2_write64(~0ull, dev->base + NIX_LF_RAS_ENA_W1S);
|
|
else
|
|
otx2_write64(~0ull, dev->base + NIX_LF_RAS_ENA_W1C);
|
|
}
|