numam-dpdk/drivers/net/mlx4/mlx4_intr.c
Harman Kalra d61138d4f0 drivers: remove direct access to interrupt handle
Removing direct access to interrupt handle structure fields,
rather use respective get set APIs for the same.
Making changes to all the drivers access the interrupt handle fields.

Signed-off-by: Harman Kalra <hkalra@marvell.com>
Acked-by: Hyong Youb Kim <hyonkim@cisco.com>
Signed-off-by: David Marchand <david.marchand@redhat.com>
Tested-by: Raslan Darawsheh <rasland@nvidia.com>
2021-10-25 21:20:12 +02:00

422 lines
9.5 KiB
C

/* SPDX-License-Identifier: BSD-3-Clause
* Copyright 2017 6WIND S.A.
* Copyright 2017 Mellanox Technologies, Ltd
*/
/**
* @file
* Interrupts handling for mlx4 driver.
*/
#include <errno.h>
#include <stdint.h>
#include <stdlib.h>
/* Verbs headers do not support -pedantic. */
#ifdef PEDANTIC
#pragma GCC diagnostic ignored "-Wpedantic"
#endif
#include <infiniband/verbs.h>
#ifdef PEDANTIC
#pragma GCC diagnostic error "-Wpedantic"
#endif
#include <rte_alarm.h>
#include <rte_errno.h>
#include <ethdev_driver.h>
#include <rte_io.h>
#include <rte_interrupts.h>
#include "mlx4.h"
#include "mlx4_glue.h"
#include "mlx4_rxtx.h"
#include "mlx4_utils.h"
static int mlx4_link_status_check(struct mlx4_priv *priv);
/**
* Clean up Rx interrupts handler.
*
* @param priv
* Pointer to private structure.
*/
static void
mlx4_rx_intr_vec_disable(struct mlx4_priv *priv)
{
struct rte_intr_handle *intr_handle = priv->intr_handle;
rte_intr_free_epoll_fd(intr_handle);
rte_intr_vec_list_free(intr_handle);
rte_intr_nb_efd_set(intr_handle, 0);
}
/**
* Allocate queue vector and fill epoll fd list for Rx interrupts.
*
* @param priv
* Pointer to private structure.
*
* @return
* 0 on success, negative errno value otherwise and rte_errno is set.
*/
static int
mlx4_rx_intr_vec_enable(struct mlx4_priv *priv)
{
unsigned int i;
unsigned int rxqs_n = ETH_DEV(priv)->data->nb_rx_queues;
unsigned int n = RTE_MIN(rxqs_n, (uint32_t)RTE_MAX_RXTX_INTR_VEC_ID);
unsigned int count = 0;
struct rte_intr_handle *intr_handle = priv->intr_handle;
mlx4_rx_intr_vec_disable(priv);
if (rte_intr_vec_list_alloc(intr_handle, NULL, n)) {
rte_errno = ENOMEM;
ERROR("failed to allocate memory for interrupt vector,"
" Rx interrupts will not be supported");
return -rte_errno;
}
for (i = 0; i != n; ++i) {
struct rxq *rxq = ETH_DEV(priv)->data->rx_queues[i];
/* Skip queues that cannot request interrupts. */
if (!rxq || !rxq->channel) {
/* Use invalid intr_vec[] index to disable entry. */
if (rte_intr_vec_list_index_set(intr_handle, i,
RTE_INTR_VEC_RXTX_OFFSET + RTE_MAX_RXTX_INTR_VEC_ID))
return -rte_errno;
continue;
}
if (count >= RTE_MAX_RXTX_INTR_VEC_ID) {
rte_errno = E2BIG;
ERROR("too many Rx queues for interrupt vector size"
" (%d), Rx interrupts cannot be enabled",
RTE_MAX_RXTX_INTR_VEC_ID);
mlx4_rx_intr_vec_disable(priv);
return -rte_errno;
}
if (rte_intr_vec_list_index_set(intr_handle, i,
RTE_INTR_VEC_RXTX_OFFSET + count))
return -rte_errno;
if (rte_intr_efds_index_set(intr_handle, i,
rxq->channel->fd))
return -rte_errno;
count++;
}
if (!count)
mlx4_rx_intr_vec_disable(priv);
else if (rte_intr_nb_efd_set(intr_handle, count))
return -rte_errno;
return 0;
}
/**
* Process scheduled link status check.
*
* If LSC interrupts are requested, process related callback.
*
* @param priv
* Pointer to private structure.
*/
static void
mlx4_link_status_alarm(struct mlx4_priv *priv)
{
const struct rte_eth_intr_conf *const intr_conf =
&ETH_DEV(priv)->data->dev_conf.intr_conf;
MLX4_ASSERT(priv->intr_alarm == 1);
priv->intr_alarm = 0;
if (intr_conf->lsc && !mlx4_link_status_check(priv))
rte_eth_dev_callback_process(ETH_DEV(priv),
RTE_ETH_EVENT_INTR_LSC,
NULL);
}
/**
* Check link status.
*
* In case of inconsistency, another check is scheduled.
*
* @param priv
* Pointer to private structure.
*
* @return
* 0 on success (link status is consistent), negative errno value
* otherwise and rte_errno is set.
*/
static int
mlx4_link_status_check(struct mlx4_priv *priv)
{
struct rte_eth_link *link = &ETH_DEV(priv)->data->dev_link;
int ret = mlx4_link_update(ETH_DEV(priv), 0);
if (ret)
return ret;
if ((!link->link_speed && link->link_status) ||
(link->link_speed && !link->link_status)) {
if (!priv->intr_alarm) {
/* Inconsistent status, check again later. */
ret = rte_eal_alarm_set(MLX4_INTR_ALARM_TIMEOUT,
(void (*)(void *))
mlx4_link_status_alarm,
priv);
if (ret)
return ret;
priv->intr_alarm = 1;
}
rte_errno = EINPROGRESS;
return -rte_errno;
}
return 0;
}
/**
* Handle interrupts from the NIC.
*
* @param priv
* Pointer to private structure.
*/
static void
mlx4_interrupt_handler(struct mlx4_priv *priv)
{
enum { LSC, RMV, };
static const enum rte_eth_event_type type[] = {
[LSC] = RTE_ETH_EVENT_INTR_LSC,
[RMV] = RTE_ETH_EVENT_INTR_RMV,
};
uint32_t caught[RTE_DIM(type)] = { 0 };
struct ibv_async_event event;
const struct rte_eth_intr_conf *const intr_conf =
&ETH_DEV(priv)->data->dev_conf.intr_conf;
unsigned int i;
/* Read all message and acknowledge them. */
while (!mlx4_glue->get_async_event(priv->ctx, &event)) {
switch (event.event_type) {
case IBV_EVENT_PORT_ACTIVE:
case IBV_EVENT_PORT_ERR:
if (intr_conf->lsc && !mlx4_link_status_check(priv))
++caught[LSC];
break;
case IBV_EVENT_DEVICE_FATAL:
if (intr_conf->rmv)
++caught[RMV];
break;
default:
DEBUG("event type %d on physical port %d not handled",
event.event_type, event.element.port_num);
}
mlx4_glue->ack_async_event(&event);
}
for (i = 0; i != RTE_DIM(caught); ++i)
if (caught[i])
rte_eth_dev_callback_process(ETH_DEV(priv), type[i],
NULL);
}
/**
* MLX4 CQ notification .
*
* @param rxq
* Pointer to receive queue structure.
* @param solicited
* Is request solicited or not.
*/
static void
mlx4_arm_cq(struct rxq *rxq, int solicited)
{
struct mlx4_cq *cq = &rxq->mcq;
uint64_t doorbell;
uint32_t sn = cq->arm_sn & MLX4_CQ_DB_GEQ_N_MASK;
uint32_t ci = cq->cons_index & MLX4_CQ_DB_CI_MASK;
uint32_t cmd = solicited ? MLX4_CQ_DB_REQ_NOT_SOL : MLX4_CQ_DB_REQ_NOT;
*cq->arm_db = rte_cpu_to_be_32(sn << 28 | cmd | ci);
/*
* Make sure that the doorbell record in host memory is
* written before ringing the doorbell via PCI MMIO.
*/
rte_wmb();
doorbell = sn << 28 | cmd | cq->cqn;
doorbell <<= 32;
doorbell |= ci;
rte_write64(rte_cpu_to_be_64(doorbell), cq->cq_db_reg);
}
/**
* Uninstall interrupt handler.
*
* @param priv
* Pointer to private structure.
*
* @return
* 0 on success, negative errno value otherwise and rte_errno is set.
*/
int
mlx4_intr_uninstall(struct mlx4_priv *priv)
{
int err = rte_errno; /* Make sure rte_errno remains unchanged. */
if (rte_intr_fd_get(priv->intr_handle) != -1) {
rte_intr_callback_unregister(priv->intr_handle,
(void (*)(void *))
mlx4_interrupt_handler,
priv);
if (rte_intr_fd_set(priv->intr_handle, -1))
return -rte_errno;
}
rte_eal_alarm_cancel((void (*)(void *))mlx4_link_status_alarm, priv);
priv->intr_alarm = 0;
mlx4_rxq_intr_disable(priv);
rte_errno = err;
return 0;
}
/**
* Install interrupt handler.
*
* @param priv
* Pointer to private structure.
*
* @return
* 0 on success, negative errno value otherwise and rte_errno is set.
*/
int
mlx4_intr_install(struct mlx4_priv *priv)
{
const struct rte_eth_intr_conf *const intr_conf =
&ETH_DEV(priv)->data->dev_conf.intr_conf;
int rc;
mlx4_intr_uninstall(priv);
if (intr_conf->lsc | intr_conf->rmv) {
if (rte_intr_fd_set(priv->intr_handle, priv->ctx->async_fd))
return -rte_errno;
rc = rte_intr_callback_register(priv->intr_handle,
(void (*)(void *))
mlx4_interrupt_handler,
priv);
if (rc < 0) {
rte_errno = -rc;
goto error;
}
}
return 0;
error:
mlx4_intr_uninstall(priv);
return -rte_errno;
}
/**
* DPDK callback for Rx queue interrupt disable.
*
* @param dev
* Pointer to Ethernet device structure.
* @param idx
* Rx queue index.
*
* @return
* 0 on success, negative errno value otherwise and rte_errno is set.
*/
int
mlx4_rx_intr_disable(struct rte_eth_dev *dev, uint16_t idx)
{
struct rxq *rxq = dev->data->rx_queues[idx];
struct ibv_cq *ev_cq;
void *ev_ctx;
int ret;
if (!rxq || !rxq->channel) {
ret = EINVAL;
} else {
ret = mlx4_glue->get_cq_event(rxq->cq->channel, &ev_cq,
&ev_ctx);
/** For non-zero ret save the errno (may be EAGAIN
* which means the get_cq_event function was called before
* receiving one).
*/
if (ret)
ret = errno;
else if (ev_cq != rxq->cq)
ret = EINVAL;
}
if (ret) {
rte_errno = ret;
if (ret != EAGAIN)
WARN("unable to disable interrupt on rx queue %d",
idx);
} else {
rxq->mcq.arm_sn++;
mlx4_glue->ack_cq_events(rxq->cq, 1);
}
return -ret;
}
/**
* DPDK callback for Rx queue interrupt enable.
*
* @param dev
* Pointer to Ethernet device structure.
* @param idx
* Rx queue index.
*
* @return
* 0 on success, negative errno value otherwise and rte_errno is set.
*/
int
mlx4_rx_intr_enable(struct rte_eth_dev *dev, uint16_t idx)
{
struct rxq *rxq = dev->data->rx_queues[idx];
int ret = 0;
if (!rxq || !rxq->channel) {
ret = EINVAL;
rte_errno = ret;
WARN("unable to arm interrupt on rx queue %d", idx);
} else {
mlx4_arm_cq(rxq, 0);
}
return -ret;
}
/**
* Enable datapath interrupts.
*
* @param priv
* Pointer to private structure.
*
* @return
* 0 on success, negative errno value otherwise and rte_errno is set.
*/
int
mlx4_rxq_intr_enable(struct mlx4_priv *priv)
{
const struct rte_eth_intr_conf *const intr_conf =
&ETH_DEV(priv)->data->dev_conf.intr_conf;
if (intr_conf->rxq && mlx4_rx_intr_vec_enable(priv) < 0)
goto error;
return 0;
error:
return -rte_errno;
}
/**
* Disable datapath interrupts, keeping other interrupts intact.
*
* @param priv
* Pointer to private structure.
*/
void
mlx4_rxq_intr_disable(struct mlx4_priv *priv)
{
int err = rte_errno; /* Make sure rte_errno remains unchanged. */
mlx4_rx_intr_vec_disable(priv);
rte_errno = err;
}