numam-dpdk/drivers/net/mlx4/mlx4_intr.c
Bruce Richardson df96fd0d73 ethdev: make driver-only headers private
The rte_ethdev_driver.h, rte_ethdev_vdev.h and rte_ethdev_pci.h files are
for drivers only and should be a private to DPDK and not installed.

Signed-off-by: Bruce Richardson <bruce.richardson@intel.com>
Reviewed-by: Maxime Coquelin <maxime.coquelin@redhat.com>
Acked-by: Thomas Monjalon <thomas@monjalon.net>
Acked-by: Steven Webster <steven.webster@windriver.com>
2021-01-29 20:59:09 +01:00

413 lines
9.3 KiB
C

/* SPDX-License-Identifier: BSD-3-Clause
* Copyright 2017 6WIND S.A.
* Copyright 2017 Mellanox Technologies, Ltd
*/
/**
* @file
* Interrupts handling for mlx4 driver.
*/
#include <errno.h>
#include <stdint.h>
#include <stdlib.h>
/* Verbs headers do not support -pedantic. */
#ifdef PEDANTIC
#pragma GCC diagnostic ignored "-Wpedantic"
#endif
#include <infiniband/verbs.h>
#ifdef PEDANTIC
#pragma GCC diagnostic error "-Wpedantic"
#endif
#include <rte_alarm.h>
#include <rte_errno.h>
#include <ethdev_driver.h>
#include <rte_io.h>
#include <rte_interrupts.h>
#include "mlx4.h"
#include "mlx4_glue.h"
#include "mlx4_rxtx.h"
#include "mlx4_utils.h"
static int mlx4_link_status_check(struct mlx4_priv *priv);
/**
* Clean up Rx interrupts handler.
*
* @param priv
* Pointer to private structure.
*/
static void
mlx4_rx_intr_vec_disable(struct mlx4_priv *priv)
{
struct rte_intr_handle *intr_handle = &priv->intr_handle;
rte_intr_free_epoll_fd(intr_handle);
free(intr_handle->intr_vec);
intr_handle->nb_efd = 0;
intr_handle->intr_vec = NULL;
}
/**
* Allocate queue vector and fill epoll fd list for Rx interrupts.
*
* @param priv
* Pointer to private structure.
*
* @return
* 0 on success, negative errno value otherwise and rte_errno is set.
*/
static int
mlx4_rx_intr_vec_enable(struct mlx4_priv *priv)
{
unsigned int i;
unsigned int rxqs_n = ETH_DEV(priv)->data->nb_rx_queues;
unsigned int n = RTE_MIN(rxqs_n, (uint32_t)RTE_MAX_RXTX_INTR_VEC_ID);
unsigned int count = 0;
struct rte_intr_handle *intr_handle = &priv->intr_handle;
mlx4_rx_intr_vec_disable(priv);
intr_handle->intr_vec = malloc(n * sizeof(intr_handle->intr_vec[0]));
if (intr_handle->intr_vec == NULL) {
rte_errno = ENOMEM;
ERROR("failed to allocate memory for interrupt vector,"
" Rx interrupts will not be supported");
return -rte_errno;
}
for (i = 0; i != n; ++i) {
struct rxq *rxq = ETH_DEV(priv)->data->rx_queues[i];
/* Skip queues that cannot request interrupts. */
if (!rxq || !rxq->channel) {
/* Use invalid intr_vec[] index to disable entry. */
intr_handle->intr_vec[i] =
RTE_INTR_VEC_RXTX_OFFSET +
RTE_MAX_RXTX_INTR_VEC_ID;
continue;
}
if (count >= RTE_MAX_RXTX_INTR_VEC_ID) {
rte_errno = E2BIG;
ERROR("too many Rx queues for interrupt vector size"
" (%d), Rx interrupts cannot be enabled",
RTE_MAX_RXTX_INTR_VEC_ID);
mlx4_rx_intr_vec_disable(priv);
return -rte_errno;
}
intr_handle->intr_vec[i] = RTE_INTR_VEC_RXTX_OFFSET + count;
intr_handle->efds[count] = rxq->channel->fd;
count++;
}
if (!count)
mlx4_rx_intr_vec_disable(priv);
else
intr_handle->nb_efd = count;
return 0;
}
/**
* Process scheduled link status check.
*
* If LSC interrupts are requested, process related callback.
*
* @param priv
* Pointer to private structure.
*/
static void
mlx4_link_status_alarm(struct mlx4_priv *priv)
{
const struct rte_intr_conf *const intr_conf =
&ETH_DEV(priv)->data->dev_conf.intr_conf;
MLX4_ASSERT(priv->intr_alarm == 1);
priv->intr_alarm = 0;
if (intr_conf->lsc && !mlx4_link_status_check(priv))
rte_eth_dev_callback_process(ETH_DEV(priv),
RTE_ETH_EVENT_INTR_LSC,
NULL);
}
/**
* Check link status.
*
* In case of inconsistency, another check is scheduled.
*
* @param priv
* Pointer to private structure.
*
* @return
* 0 on success (link status is consistent), negative errno value
* otherwise and rte_errno is set.
*/
static int
mlx4_link_status_check(struct mlx4_priv *priv)
{
struct rte_eth_link *link = &ETH_DEV(priv)->data->dev_link;
int ret = mlx4_link_update(ETH_DEV(priv), 0);
if (ret)
return ret;
if ((!link->link_speed && link->link_status) ||
(link->link_speed && !link->link_status)) {
if (!priv->intr_alarm) {
/* Inconsistent status, check again later. */
ret = rte_eal_alarm_set(MLX4_INTR_ALARM_TIMEOUT,
(void (*)(void *))
mlx4_link_status_alarm,
priv);
if (ret)
return ret;
priv->intr_alarm = 1;
}
rte_errno = EINPROGRESS;
return -rte_errno;
}
return 0;
}
/**
* Handle interrupts from the NIC.
*
* @param priv
* Pointer to private structure.
*/
static void
mlx4_interrupt_handler(struct mlx4_priv *priv)
{
enum { LSC, RMV, };
static const enum rte_eth_event_type type[] = {
[LSC] = RTE_ETH_EVENT_INTR_LSC,
[RMV] = RTE_ETH_EVENT_INTR_RMV,
};
uint32_t caught[RTE_DIM(type)] = { 0 };
struct ibv_async_event event;
const struct rte_intr_conf *const intr_conf =
&ETH_DEV(priv)->data->dev_conf.intr_conf;
unsigned int i;
/* Read all message and acknowledge them. */
while (!mlx4_glue->get_async_event(priv->ctx, &event)) {
switch (event.event_type) {
case IBV_EVENT_PORT_ACTIVE:
case IBV_EVENT_PORT_ERR:
if (intr_conf->lsc && !mlx4_link_status_check(priv))
++caught[LSC];
break;
case IBV_EVENT_DEVICE_FATAL:
if (intr_conf->rmv)
++caught[RMV];
break;
default:
DEBUG("event type %d on physical port %d not handled",
event.event_type, event.element.port_num);
}
mlx4_glue->ack_async_event(&event);
}
for (i = 0; i != RTE_DIM(caught); ++i)
if (caught[i])
rte_eth_dev_callback_process(ETH_DEV(priv), type[i],
NULL);
}
/**
* MLX4 CQ notification .
*
* @param rxq
* Pointer to receive queue structure.
* @param solicited
* Is request solicited or not.
*/
static void
mlx4_arm_cq(struct rxq *rxq, int solicited)
{
struct mlx4_cq *cq = &rxq->mcq;
uint64_t doorbell;
uint32_t sn = cq->arm_sn & MLX4_CQ_DB_GEQ_N_MASK;
uint32_t ci = cq->cons_index & MLX4_CQ_DB_CI_MASK;
uint32_t cmd = solicited ? MLX4_CQ_DB_REQ_NOT_SOL : MLX4_CQ_DB_REQ_NOT;
*cq->arm_db = rte_cpu_to_be_32(sn << 28 | cmd | ci);
/*
* Make sure that the doorbell record in host memory is
* written before ringing the doorbell via PCI MMIO.
*/
rte_wmb();
doorbell = sn << 28 | cmd | cq->cqn;
doorbell <<= 32;
doorbell |= ci;
rte_write64(rte_cpu_to_be_64(doorbell), cq->cq_db_reg);
}
/**
* Uninstall interrupt handler.
*
* @param priv
* Pointer to private structure.
*
* @return
* 0 on success, negative errno value otherwise and rte_errno is set.
*/
int
mlx4_intr_uninstall(struct mlx4_priv *priv)
{
int err = rte_errno; /* Make sure rte_errno remains unchanged. */
if (priv->intr_handle.fd != -1) {
rte_intr_callback_unregister(&priv->intr_handle,
(void (*)(void *))
mlx4_interrupt_handler,
priv);
priv->intr_handle.fd = -1;
}
rte_eal_alarm_cancel((void (*)(void *))mlx4_link_status_alarm, priv);
priv->intr_alarm = 0;
mlx4_rxq_intr_disable(priv);
rte_errno = err;
return 0;
}
/**
* Install interrupt handler.
*
* @param priv
* Pointer to private structure.
*
* @return
* 0 on success, negative errno value otherwise and rte_errno is set.
*/
int
mlx4_intr_install(struct mlx4_priv *priv)
{
const struct rte_intr_conf *const intr_conf =
&ETH_DEV(priv)->data->dev_conf.intr_conf;
int rc;
mlx4_intr_uninstall(priv);
if (intr_conf->lsc | intr_conf->rmv) {
priv->intr_handle.fd = priv->ctx->async_fd;
rc = rte_intr_callback_register(&priv->intr_handle,
(void (*)(void *))
mlx4_interrupt_handler,
priv);
if (rc < 0) {
rte_errno = -rc;
goto error;
}
}
return 0;
error:
mlx4_intr_uninstall(priv);
return -rte_errno;
}
/**
* DPDK callback for Rx queue interrupt disable.
*
* @param dev
* Pointer to Ethernet device structure.
* @param idx
* Rx queue index.
*
* @return
* 0 on success, negative errno value otherwise and rte_errno is set.
*/
int
mlx4_rx_intr_disable(struct rte_eth_dev *dev, uint16_t idx)
{
struct rxq *rxq = dev->data->rx_queues[idx];
struct ibv_cq *ev_cq;
void *ev_ctx;
int ret;
if (!rxq || !rxq->channel) {
ret = EINVAL;
} else {
ret = mlx4_glue->get_cq_event(rxq->cq->channel, &ev_cq,
&ev_ctx);
/** For non-zero ret save the errno (may be EAGAIN
* which means the get_cq_event function was called before
* receiving one).
*/
if (ret)
ret = errno;
else if (ev_cq != rxq->cq)
ret = EINVAL;
}
if (ret) {
rte_errno = ret;
if (ret != EAGAIN)
WARN("unable to disable interrupt on rx queue %d",
idx);
} else {
rxq->mcq.arm_sn++;
mlx4_glue->ack_cq_events(rxq->cq, 1);
}
return -ret;
}
/**
* DPDK callback for Rx queue interrupt enable.
*
* @param dev
* Pointer to Ethernet device structure.
* @param idx
* Rx queue index.
*
* @return
* 0 on success, negative errno value otherwise and rte_errno is set.
*/
int
mlx4_rx_intr_enable(struct rte_eth_dev *dev, uint16_t idx)
{
struct rxq *rxq = dev->data->rx_queues[idx];
int ret = 0;
if (!rxq || !rxq->channel) {
ret = EINVAL;
rte_errno = ret;
WARN("unable to arm interrupt on rx queue %d", idx);
} else {
mlx4_arm_cq(rxq, 0);
}
return -ret;
}
/**
* Enable datapath interrupts.
*
* @param priv
* Pointer to private structure.
*
* @return
* 0 on success, negative errno value otherwise and rte_errno is set.
*/
int
mlx4_rxq_intr_enable(struct mlx4_priv *priv)
{
const struct rte_intr_conf *const intr_conf =
&ETH_DEV(priv)->data->dev_conf.intr_conf;
if (intr_conf->rxq && mlx4_rx_intr_vec_enable(priv) < 0)
goto error;
return 0;
error:
return -rte_errno;
}
/**
* Disable datapath interrupts, keeping other interrupts intact.
*
* @param priv
* Pointer to private structure.
*/
void
mlx4_rxq_intr_disable(struct mlx4_priv *priv)
{
int err = rte_errno; /* Make sure rte_errno remains unchanged. */
mlx4_rx_intr_vec_disable(priv);
rte_errno = err;
}