36c3c4df13
At device probe, the fslmc bus driver calls rte_vfio_get_group_fd() to
get a fd associated to a vfio group. This function first checks if the
group is already opened, else it opens /dev/vfio/%u, and increases the
number of active groups in default_vfio_cfg (which references the
default vfio container).
When adding the first group to a vfio_cfg, the caller is supposed to
pick an IOMMU type and set up DMA mappings for container, as it's done
by pci bus, but it is not done here. Instead, a new container is created
and used.
This prevents the pci bus driver, which uses the default_vfio_cfg
container, to configure the container because
default_vfio_cfg->active_group > 1.
This patch fixes the issue by always creating a new container (and its
associated vfio_cfg) and binding the group to it.
Fixes: a69f793002
("bus/fslmc: support multi VFIO group")
Cc: stable@dpdk.org
Signed-off-by: Romain Delhomel <romain.delhomel@6wind.com>
Acked-by: Hemant Agrawal <hemant.agrawal@nxp.com>
1077 lines
26 KiB
C
1077 lines
26 KiB
C
/* SPDX-License-Identifier: BSD-3-Clause
|
|
*
|
|
* Copyright (c) 2015-2016 Freescale Semiconductor, Inc. All rights reserved.
|
|
* Copyright 2016-2021 NXP
|
|
*
|
|
*/
|
|
|
|
#include <unistd.h>
|
|
#include <stdio.h>
|
|
#include <sys/types.h>
|
|
#include <string.h>
|
|
#include <stdlib.h>
|
|
#include <fcntl.h>
|
|
#include <errno.h>
|
|
#include <sys/ioctl.h>
|
|
#include <sys/stat.h>
|
|
#include <sys/mman.h>
|
|
#include <sys/vfs.h>
|
|
#include <libgen.h>
|
|
#include <dirent.h>
|
|
#include <sys/eventfd.h>
|
|
|
|
#include <eal_filesystem.h>
|
|
#include <rte_mbuf.h>
|
|
#include <ethdev_driver.h>
|
|
#include <rte_malloc.h>
|
|
#include <rte_memcpy.h>
|
|
#include <rte_string_fns.h>
|
|
#include <rte_cycles.h>
|
|
#include <rte_kvargs.h>
|
|
#include <rte_dev.h>
|
|
#include <rte_bus.h>
|
|
#include <rte_eal_memconfig.h>
|
|
|
|
#include "rte_fslmc.h"
|
|
#include "fslmc_vfio.h"
|
|
#include "fslmc_logs.h"
|
|
#include <mc/fsl_dpmng.h>
|
|
|
|
#include "portal/dpaa2_hw_pvt.h"
|
|
#include "portal/dpaa2_hw_dpio.h"
|
|
|
|
#define FSLMC_CONTAINER_MAX_LEN 8 /**< Of the format dprc.XX */
|
|
|
|
/* Number of VFIO containers & groups with in */
|
|
static struct fslmc_vfio_group vfio_group;
|
|
static struct fslmc_vfio_container vfio_container;
|
|
static int container_device_fd;
|
|
char *fslmc_container;
|
|
static int fslmc_iommu_type;
|
|
static uint32_t *msi_intr_vaddr;
|
|
void *(*rte_mcp_ptr_list);
|
|
|
|
void *
|
|
dpaa2_get_mcp_ptr(int portal_idx)
|
|
{
|
|
if (rte_mcp_ptr_list)
|
|
return rte_mcp_ptr_list[portal_idx];
|
|
else
|
|
return NULL;
|
|
}
|
|
|
|
static struct rte_dpaa2_object_list dpaa2_obj_list =
|
|
TAILQ_HEAD_INITIALIZER(dpaa2_obj_list);
|
|
|
|
/*register a fslmc bus based dpaa2 driver */
|
|
void
|
|
rte_fslmc_object_register(struct rte_dpaa2_object *object)
|
|
{
|
|
RTE_VERIFY(object);
|
|
|
|
TAILQ_INSERT_TAIL(&dpaa2_obj_list, object, next);
|
|
}
|
|
|
|
int
|
|
fslmc_get_container_group(int *groupid)
|
|
{
|
|
int ret;
|
|
char *container;
|
|
|
|
if (!fslmc_container) {
|
|
container = getenv("DPRC");
|
|
if (container == NULL) {
|
|
DPAA2_BUS_DEBUG("DPAA2: DPRC not available");
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (strlen(container) >= FSLMC_CONTAINER_MAX_LEN) {
|
|
DPAA2_BUS_ERR("Invalid container name: %s", container);
|
|
return -1;
|
|
}
|
|
|
|
fslmc_container = strdup(container);
|
|
if (!fslmc_container) {
|
|
DPAA2_BUS_ERR("Mem alloc failure; Container name");
|
|
return -ENOMEM;
|
|
}
|
|
}
|
|
|
|
fslmc_iommu_type = (rte_vfio_noiommu_is_enabled() == 1) ?
|
|
RTE_VFIO_NOIOMMU : VFIO_TYPE1_IOMMU;
|
|
|
|
/* get group number */
|
|
ret = rte_vfio_get_group_num(SYSFS_FSL_MC_DEVICES,
|
|
fslmc_container, groupid);
|
|
if (ret <= 0) {
|
|
DPAA2_BUS_ERR("Unable to find %s IOMMU group", fslmc_container);
|
|
return -1;
|
|
}
|
|
|
|
DPAA2_BUS_DEBUG("Container: %s has VFIO iommu group id = %d",
|
|
fslmc_container, *groupid);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
vfio_connect_container(void)
|
|
{
|
|
int fd, ret;
|
|
|
|
if (vfio_container.used) {
|
|
DPAA2_BUS_DEBUG("No container available");
|
|
return -1;
|
|
}
|
|
|
|
/* Try connecting to vfio container if already created */
|
|
if (!ioctl(vfio_group.fd, VFIO_GROUP_SET_CONTAINER,
|
|
&vfio_container.fd)) {
|
|
DPAA2_BUS_DEBUG(
|
|
"Container pre-exists with FD[0x%x] for this group",
|
|
vfio_container.fd);
|
|
vfio_group.container = &vfio_container;
|
|
return 0;
|
|
}
|
|
|
|
/* Opens main vfio file descriptor which represents the "container" */
|
|
fd = rte_vfio_get_container_fd();
|
|
if (fd < 0) {
|
|
DPAA2_BUS_ERR("Failed to open VFIO container");
|
|
return -errno;
|
|
}
|
|
|
|
/* Check whether support for SMMU type IOMMU present or not */
|
|
if (ioctl(fd, VFIO_CHECK_EXTENSION, fslmc_iommu_type)) {
|
|
/* Connect group to container */
|
|
ret = ioctl(vfio_group.fd, VFIO_GROUP_SET_CONTAINER, &fd);
|
|
if (ret) {
|
|
DPAA2_BUS_ERR("Failed to setup group container");
|
|
close(fd);
|
|
return -errno;
|
|
}
|
|
|
|
ret = ioctl(fd, VFIO_SET_IOMMU, fslmc_iommu_type);
|
|
if (ret) {
|
|
DPAA2_BUS_ERR("Failed to setup VFIO iommu");
|
|
close(fd);
|
|
return -errno;
|
|
}
|
|
} else {
|
|
DPAA2_BUS_ERR("No supported IOMMU available");
|
|
close(fd);
|
|
return -EINVAL;
|
|
}
|
|
|
|
vfio_container.used = 1;
|
|
vfio_container.fd = fd;
|
|
vfio_container.group = &vfio_group;
|
|
vfio_group.container = &vfio_container;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int vfio_map_irq_region(struct fslmc_vfio_group *group)
|
|
{
|
|
int ret;
|
|
unsigned long *vaddr = NULL;
|
|
struct vfio_iommu_type1_dma_map map = {
|
|
.argsz = sizeof(map),
|
|
.flags = VFIO_DMA_MAP_FLAG_READ | VFIO_DMA_MAP_FLAG_WRITE,
|
|
.vaddr = 0x6030000,
|
|
.iova = 0x6030000,
|
|
.size = 0x1000,
|
|
};
|
|
|
|
vaddr = (unsigned long *)mmap(NULL, 0x1000, PROT_WRITE |
|
|
PROT_READ, MAP_SHARED, container_device_fd, 0x6030000);
|
|
if (vaddr == MAP_FAILED) {
|
|
DPAA2_BUS_INFO("Unable to map region (errno = %d)", errno);
|
|
return -errno;
|
|
}
|
|
|
|
msi_intr_vaddr = (uint32_t *)((char *)(vaddr) + 64);
|
|
map.vaddr = (unsigned long)vaddr;
|
|
ret = ioctl(group->container->fd, VFIO_IOMMU_MAP_DMA, &map);
|
|
if (ret == 0)
|
|
return 0;
|
|
|
|
DPAA2_BUS_ERR("Unable to map DMA address (errno = %d)", errno);
|
|
return -errno;
|
|
}
|
|
|
|
static int fslmc_map_dma(uint64_t vaddr, rte_iova_t iovaddr, size_t len);
|
|
static int fslmc_unmap_dma(uint64_t vaddr, rte_iova_t iovaddr, size_t len);
|
|
|
|
static void
|
|
fslmc_memevent_cb(enum rte_mem_event type, const void *addr, size_t len,
|
|
void *arg __rte_unused)
|
|
{
|
|
struct rte_memseg_list *msl;
|
|
struct rte_memseg *ms;
|
|
size_t cur_len = 0, map_len = 0;
|
|
uint64_t virt_addr;
|
|
rte_iova_t iova_addr;
|
|
int ret;
|
|
|
|
msl = rte_mem_virt2memseg_list(addr);
|
|
|
|
while (cur_len < len) {
|
|
const void *va = RTE_PTR_ADD(addr, cur_len);
|
|
|
|
ms = rte_mem_virt2memseg(va, msl);
|
|
iova_addr = ms->iova;
|
|
virt_addr = ms->addr_64;
|
|
map_len = ms->len;
|
|
|
|
DPAA2_BUS_DEBUG("Request for %s, va=%p, "
|
|
"virt_addr=0x%" PRIx64 ", "
|
|
"iova=0x%" PRIx64 ", map_len=%zu",
|
|
type == RTE_MEM_EVENT_ALLOC ?
|
|
"alloc" : "dealloc",
|
|
va, virt_addr, iova_addr, map_len);
|
|
|
|
/* iova_addr may be set to RTE_BAD_IOVA */
|
|
if (iova_addr == RTE_BAD_IOVA) {
|
|
DPAA2_BUS_DEBUG("Segment has invalid iova, skipping\n");
|
|
cur_len += map_len;
|
|
continue;
|
|
}
|
|
|
|
if (type == RTE_MEM_EVENT_ALLOC)
|
|
ret = fslmc_map_dma(virt_addr, iova_addr, map_len);
|
|
else
|
|
ret = fslmc_unmap_dma(virt_addr, iova_addr, map_len);
|
|
|
|
if (ret != 0) {
|
|
DPAA2_BUS_ERR("DMA Mapping/Unmapping failed. "
|
|
"Map=%d, addr=%p, len=%zu, err:(%d)",
|
|
type, va, map_len, ret);
|
|
return;
|
|
}
|
|
|
|
cur_len += map_len;
|
|
}
|
|
|
|
if (type == RTE_MEM_EVENT_ALLOC)
|
|
DPAA2_BUS_DEBUG("Total Mapped: addr=%p, len=%zu",
|
|
addr, len);
|
|
else
|
|
DPAA2_BUS_DEBUG("Total Unmapped: addr=%p, len=%zu",
|
|
addr, len);
|
|
}
|
|
|
|
static int
|
|
fslmc_map_dma(uint64_t vaddr, rte_iova_t iovaddr __rte_unused, size_t len)
|
|
{
|
|
struct fslmc_vfio_group *group;
|
|
struct vfio_iommu_type1_dma_map dma_map = {
|
|
.argsz = sizeof(struct vfio_iommu_type1_dma_map),
|
|
.flags = VFIO_DMA_MAP_FLAG_READ | VFIO_DMA_MAP_FLAG_WRITE,
|
|
};
|
|
int ret;
|
|
|
|
if (fslmc_iommu_type == RTE_VFIO_NOIOMMU) {
|
|
DPAA2_BUS_DEBUG("Running in NOIOMMU mode");
|
|
return 0;
|
|
}
|
|
|
|
dma_map.size = len;
|
|
dma_map.vaddr = vaddr;
|
|
|
|
#ifdef RTE_LIBRTE_DPAA2_USE_PHYS_IOVA
|
|
dma_map.iova = iovaddr;
|
|
#else
|
|
dma_map.iova = dma_map.vaddr;
|
|
#endif
|
|
|
|
/* SET DMA MAP for IOMMU */
|
|
group = &vfio_group;
|
|
|
|
if (!group->container) {
|
|
DPAA2_BUS_ERR("Container is not connected ");
|
|
return -1;
|
|
}
|
|
|
|
DPAA2_BUS_DEBUG("--> Map address: 0x%"PRIx64", size: %"PRIu64"",
|
|
(uint64_t)dma_map.vaddr, (uint64_t)dma_map.size);
|
|
ret = ioctl(group->container->fd, VFIO_IOMMU_MAP_DMA, &dma_map);
|
|
if (ret) {
|
|
DPAA2_BUS_ERR("VFIO_IOMMU_MAP_DMA API(errno = %d)",
|
|
errno);
|
|
return -1;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
fslmc_unmap_dma(uint64_t vaddr, uint64_t iovaddr __rte_unused, size_t len)
|
|
{
|
|
struct fslmc_vfio_group *group;
|
|
struct vfio_iommu_type1_dma_unmap dma_unmap = {
|
|
.argsz = sizeof(struct vfio_iommu_type1_dma_unmap),
|
|
.flags = 0,
|
|
};
|
|
int ret;
|
|
|
|
if (fslmc_iommu_type == RTE_VFIO_NOIOMMU) {
|
|
DPAA2_BUS_DEBUG("Running in NOIOMMU mode");
|
|
return 0;
|
|
}
|
|
|
|
dma_unmap.size = len;
|
|
dma_unmap.iova = vaddr;
|
|
|
|
/* SET DMA MAP for IOMMU */
|
|
group = &vfio_group;
|
|
|
|
if (!group->container) {
|
|
DPAA2_BUS_ERR("Container is not connected ");
|
|
return -1;
|
|
}
|
|
|
|
DPAA2_BUS_DEBUG("--> Unmap address: 0x%"PRIx64", size: %"PRIu64"",
|
|
(uint64_t)dma_unmap.iova, (uint64_t)dma_unmap.size);
|
|
ret = ioctl(group->container->fd, VFIO_IOMMU_UNMAP_DMA, &dma_unmap);
|
|
if (ret) {
|
|
DPAA2_BUS_ERR("VFIO_IOMMU_UNMAP_DMA API(errno = %d)",
|
|
errno);
|
|
return -1;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
fslmc_dmamap_seg(const struct rte_memseg_list *msl __rte_unused,
|
|
const struct rte_memseg *ms, void *arg)
|
|
{
|
|
int *n_segs = arg;
|
|
int ret;
|
|
|
|
/* if IOVA address is invalid, skip */
|
|
if (ms->iova == RTE_BAD_IOVA)
|
|
return 0;
|
|
|
|
ret = fslmc_map_dma(ms->addr_64, ms->iova, ms->len);
|
|
if (ret)
|
|
DPAA2_BUS_ERR("Unable to VFIO map (addr=%p, len=%zu)",
|
|
ms->addr, ms->len);
|
|
else
|
|
(*n_segs)++;
|
|
|
|
return ret;
|
|
}
|
|
|
|
int
|
|
rte_fslmc_vfio_mem_dmamap(uint64_t vaddr, uint64_t iova, uint64_t size)
|
|
{
|
|
int ret;
|
|
struct fslmc_vfio_group *group;
|
|
struct vfio_iommu_type1_dma_map dma_map = {
|
|
.argsz = sizeof(struct vfio_iommu_type1_dma_map),
|
|
.flags = VFIO_DMA_MAP_FLAG_READ | VFIO_DMA_MAP_FLAG_WRITE,
|
|
};
|
|
|
|
if (fslmc_iommu_type == RTE_VFIO_NOIOMMU) {
|
|
DPAA2_BUS_DEBUG("Running in NOIOMMU mode");
|
|
return 0;
|
|
}
|
|
|
|
/* SET DMA MAP for IOMMU */
|
|
group = &vfio_group;
|
|
if (!group->container) {
|
|
DPAA2_BUS_ERR("Container is not connected");
|
|
return -1;
|
|
}
|
|
|
|
dma_map.size = size;
|
|
dma_map.vaddr = vaddr;
|
|
dma_map.iova = iova;
|
|
|
|
DPAA2_BUS_DEBUG("VFIOdmamap 0x%"PRIx64":0x%"PRIx64",size 0x%"PRIx64"\n",
|
|
(uint64_t)dma_map.vaddr, (uint64_t)dma_map.iova,
|
|
(uint64_t)dma_map.size);
|
|
ret = ioctl(group->container->fd, VFIO_IOMMU_MAP_DMA,
|
|
&dma_map);
|
|
if (ret) {
|
|
printf("Unable to map DMA address (errno = %d)\n",
|
|
errno);
|
|
return ret;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
int rte_fslmc_vfio_dmamap(void)
|
|
{
|
|
int i = 0, ret;
|
|
|
|
/* Lock before parsing and registering callback to memory subsystem */
|
|
rte_mcfg_mem_read_lock();
|
|
|
|
if (rte_memseg_walk(fslmc_dmamap_seg, &i) < 0) {
|
|
rte_mcfg_mem_read_unlock();
|
|
return -1;
|
|
}
|
|
|
|
ret = rte_mem_event_callback_register("fslmc_memevent_clb",
|
|
fslmc_memevent_cb, NULL);
|
|
if (ret && rte_errno == ENOTSUP)
|
|
DPAA2_BUS_DEBUG("Memory event callbacks not supported");
|
|
else if (ret)
|
|
DPAA2_BUS_DEBUG("Unable to install memory handler");
|
|
else
|
|
DPAA2_BUS_DEBUG("Installed memory callback handler");
|
|
|
|
DPAA2_BUS_DEBUG("Total %d segments found.", i);
|
|
|
|
/* TODO - This is a W.A. as VFIO currently does not add the mapping of
|
|
* the interrupt region to SMMU. This should be removed once the
|
|
* support is added in the Kernel.
|
|
*/
|
|
vfio_map_irq_region(&vfio_group);
|
|
|
|
/* Existing segments have been mapped and memory callback for hotplug
|
|
* has been installed.
|
|
*/
|
|
rte_mcfg_mem_read_unlock();
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
fslmc_vfio_setup_device(const char *sysfs_base, const char *dev_addr,
|
|
int *vfio_dev_fd, struct vfio_device_info *device_info)
|
|
{
|
|
struct vfio_group_status group_status = {
|
|
.argsz = sizeof(group_status)
|
|
};
|
|
int vfio_group_fd, vfio_container_fd, iommu_group_no, ret;
|
|
|
|
/* get group number */
|
|
ret = rte_vfio_get_group_num(sysfs_base, dev_addr, &iommu_group_no);
|
|
if (ret < 0)
|
|
return -1;
|
|
|
|
/* get the actual group fd */
|
|
vfio_group_fd = rte_vfio_get_group_fd(iommu_group_no);
|
|
if (vfio_group_fd < 0 && vfio_group_fd != -ENOENT)
|
|
return -1;
|
|
|
|
/*
|
|
* if vfio_group_fd == -ENOENT, that means the device
|
|
* isn't managed by VFIO
|
|
*/
|
|
if (vfio_group_fd == -ENOENT) {
|
|
RTE_LOG(WARNING, EAL, " %s not managed by VFIO driver, skipping\n",
|
|
dev_addr);
|
|
return 1;
|
|
}
|
|
|
|
/* Opens main vfio file descriptor which represents the "container" */
|
|
vfio_container_fd = rte_vfio_get_container_fd();
|
|
if (vfio_container_fd < 0) {
|
|
DPAA2_BUS_ERR("Failed to open VFIO container");
|
|
return -errno;
|
|
}
|
|
|
|
/* check if the group is viable */
|
|
ret = ioctl(vfio_group_fd, VFIO_GROUP_GET_STATUS, &group_status);
|
|
if (ret) {
|
|
DPAA2_BUS_ERR(" %s cannot get group status, "
|
|
"error %i (%s)\n", dev_addr,
|
|
errno, strerror(errno));
|
|
close(vfio_group_fd);
|
|
rte_vfio_clear_group(vfio_group_fd);
|
|
return -1;
|
|
} else if (!(group_status.flags & VFIO_GROUP_FLAGS_VIABLE)) {
|
|
DPAA2_BUS_ERR(" %s VFIO group is not viable!\n", dev_addr);
|
|
close(vfio_group_fd);
|
|
rte_vfio_clear_group(vfio_group_fd);
|
|
return -1;
|
|
}
|
|
/* At this point, we know that this group is viable (meaning,
|
|
* all devices are either bound to VFIO or not bound to anything)
|
|
*/
|
|
|
|
/* check if group does not have a container yet */
|
|
if (!(group_status.flags & VFIO_GROUP_FLAGS_CONTAINER_SET)) {
|
|
|
|
/* add group to a container */
|
|
ret = ioctl(vfio_group_fd, VFIO_GROUP_SET_CONTAINER,
|
|
&vfio_container_fd);
|
|
if (ret) {
|
|
DPAA2_BUS_ERR(" %s cannot add VFIO group to container, "
|
|
"error %i (%s)\n", dev_addr,
|
|
errno, strerror(errno));
|
|
close(vfio_group_fd);
|
|
close(vfio_container_fd);
|
|
rte_vfio_clear_group(vfio_group_fd);
|
|
return -1;
|
|
}
|
|
|
|
/*
|
|
* set an IOMMU type for container
|
|
*
|
|
*/
|
|
if (ioctl(vfio_container_fd, VFIO_CHECK_EXTENSION,
|
|
fslmc_iommu_type)) {
|
|
ret = ioctl(vfio_container_fd, VFIO_SET_IOMMU,
|
|
fslmc_iommu_type);
|
|
if (ret) {
|
|
DPAA2_BUS_ERR("Failed to setup VFIO iommu");
|
|
close(vfio_group_fd);
|
|
close(vfio_container_fd);
|
|
return -errno;
|
|
}
|
|
} else {
|
|
DPAA2_BUS_ERR("No supported IOMMU available");
|
|
close(vfio_group_fd);
|
|
close(vfio_container_fd);
|
|
return -EINVAL;
|
|
}
|
|
}
|
|
|
|
/* get a file descriptor for the device */
|
|
*vfio_dev_fd = ioctl(vfio_group_fd, VFIO_GROUP_GET_DEVICE_FD, dev_addr);
|
|
if (*vfio_dev_fd < 0) {
|
|
/* if we cannot get a device fd, this implies a problem with
|
|
* the VFIO group or the container not having IOMMU configured.
|
|
*/
|
|
|
|
DPAA2_BUS_WARN("Getting a vfio_dev_fd for %s failed", dev_addr);
|
|
close(vfio_group_fd);
|
|
close(vfio_container_fd);
|
|
rte_vfio_clear_group(vfio_group_fd);
|
|
return -1;
|
|
}
|
|
|
|
/* test and setup the device */
|
|
ret = ioctl(*vfio_dev_fd, VFIO_DEVICE_GET_INFO, device_info);
|
|
if (ret) {
|
|
DPAA2_BUS_ERR(" %s cannot get device info, error %i (%s)",
|
|
dev_addr, errno, strerror(errno));
|
|
close(*vfio_dev_fd);
|
|
close(vfio_group_fd);
|
|
close(vfio_container_fd);
|
|
rte_vfio_clear_group(vfio_group_fd);
|
|
return -1;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static intptr_t vfio_map_mcp_obj(const char *mcp_obj)
|
|
{
|
|
intptr_t v_addr = (intptr_t)MAP_FAILED;
|
|
int32_t ret, mc_fd;
|
|
struct vfio_group_status status = { .argsz = sizeof(status) };
|
|
|
|
struct vfio_device_info d_info = { .argsz = sizeof(d_info) };
|
|
struct vfio_region_info reg_info = { .argsz = sizeof(reg_info) };
|
|
|
|
fslmc_vfio_setup_device(SYSFS_FSL_MC_DEVICES, mcp_obj,
|
|
&mc_fd, &d_info);
|
|
|
|
/* getting device region info*/
|
|
ret = ioctl(mc_fd, VFIO_DEVICE_GET_REGION_INFO, ®_info);
|
|
if (ret < 0) {
|
|
DPAA2_BUS_ERR("Error in VFIO getting REGION_INFO");
|
|
goto MC_FAILURE;
|
|
}
|
|
|
|
v_addr = (size_t)mmap(NULL, reg_info.size,
|
|
PROT_WRITE | PROT_READ, MAP_SHARED,
|
|
mc_fd, reg_info.offset);
|
|
|
|
MC_FAILURE:
|
|
close(mc_fd);
|
|
|
|
return v_addr;
|
|
}
|
|
|
|
#define IRQ_SET_BUF_LEN (sizeof(struct vfio_irq_set) + sizeof(int))
|
|
|
|
int rte_dpaa2_intr_enable(struct rte_intr_handle *intr_handle, int index)
|
|
{
|
|
int len, ret;
|
|
char irq_set_buf[IRQ_SET_BUF_LEN];
|
|
struct vfio_irq_set *irq_set;
|
|
int *fd_ptr, vfio_dev_fd;
|
|
|
|
len = sizeof(irq_set_buf);
|
|
|
|
irq_set = (struct vfio_irq_set *)irq_set_buf;
|
|
irq_set->argsz = len;
|
|
irq_set->count = 1;
|
|
irq_set->flags =
|
|
VFIO_IRQ_SET_DATA_EVENTFD | VFIO_IRQ_SET_ACTION_TRIGGER;
|
|
irq_set->index = index;
|
|
irq_set->start = 0;
|
|
fd_ptr = (int *)&irq_set->data;
|
|
*fd_ptr = rte_intr_fd_get(intr_handle);
|
|
|
|
vfio_dev_fd = rte_intr_dev_fd_get(intr_handle);
|
|
ret = ioctl(vfio_dev_fd, VFIO_DEVICE_SET_IRQS, irq_set);
|
|
if (ret) {
|
|
DPAA2_BUS_ERR("Error:dpaa2 SET IRQs fd=%d, err = %d(%s)",
|
|
rte_intr_fd_get(intr_handle), errno,
|
|
strerror(errno));
|
|
return ret;
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
int rte_dpaa2_intr_disable(struct rte_intr_handle *intr_handle, int index)
|
|
{
|
|
struct vfio_irq_set *irq_set;
|
|
char irq_set_buf[IRQ_SET_BUF_LEN];
|
|
int len, ret, vfio_dev_fd;
|
|
|
|
len = sizeof(struct vfio_irq_set);
|
|
|
|
irq_set = (struct vfio_irq_set *)irq_set_buf;
|
|
irq_set->argsz = len;
|
|
irq_set->flags = VFIO_IRQ_SET_DATA_NONE | VFIO_IRQ_SET_ACTION_TRIGGER;
|
|
irq_set->index = index;
|
|
irq_set->start = 0;
|
|
irq_set->count = 0;
|
|
|
|
vfio_dev_fd = rte_intr_dev_fd_get(intr_handle);
|
|
ret = ioctl(vfio_dev_fd, VFIO_DEVICE_SET_IRQS, irq_set);
|
|
if (ret)
|
|
DPAA2_BUS_ERR(
|
|
"Error disabling dpaa2 interrupts for fd %d",
|
|
rte_intr_fd_get(intr_handle));
|
|
|
|
return ret;
|
|
}
|
|
|
|
/* set up interrupt support (but not enable interrupts) */
|
|
int
|
|
rte_dpaa2_vfio_setup_intr(struct rte_intr_handle *intr_handle,
|
|
int vfio_dev_fd,
|
|
int num_irqs)
|
|
{
|
|
int i, ret;
|
|
|
|
/* start from MSI-X interrupt type */
|
|
for (i = 0; i < num_irqs; i++) {
|
|
struct vfio_irq_info irq_info = { .argsz = sizeof(irq_info) };
|
|
int fd = -1;
|
|
|
|
irq_info.index = i;
|
|
|
|
ret = ioctl(vfio_dev_fd, VFIO_DEVICE_GET_IRQ_INFO, &irq_info);
|
|
if (ret < 0) {
|
|
DPAA2_BUS_ERR("Cannot get IRQ(%d) info, error %i (%s)",
|
|
i, errno, strerror(errno));
|
|
return -1;
|
|
}
|
|
|
|
/* if this vector cannot be used with eventfd,
|
|
* fail if we explicitly
|
|
* specified interrupt type, otherwise continue
|
|
*/
|
|
if ((irq_info.flags & VFIO_IRQ_INFO_EVENTFD) == 0)
|
|
continue;
|
|
|
|
/* set up an eventfd for interrupts */
|
|
fd = eventfd(0, EFD_NONBLOCK | EFD_CLOEXEC);
|
|
if (fd < 0) {
|
|
DPAA2_BUS_ERR("Cannot set up eventfd, error %i (%s)",
|
|
errno, strerror(errno));
|
|
return -1;
|
|
}
|
|
|
|
if (rte_intr_fd_set(intr_handle, fd))
|
|
return -rte_errno;
|
|
|
|
if (rte_intr_type_set(intr_handle, RTE_INTR_HANDLE_VFIO_MSI))
|
|
return -rte_errno;
|
|
|
|
if (rte_intr_dev_fd_set(intr_handle, vfio_dev_fd))
|
|
return -rte_errno;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* if we're here, we haven't found a suitable interrupt vector */
|
|
return -1;
|
|
}
|
|
|
|
/*
|
|
* fslmc_process_iodevices for processing only IO (ETH, CRYPTO, and possibly
|
|
* EVENT) devices.
|
|
*/
|
|
static int
|
|
fslmc_process_iodevices(struct rte_dpaa2_device *dev)
|
|
{
|
|
int dev_fd;
|
|
struct vfio_device_info device_info = { .argsz = sizeof(device_info) };
|
|
struct rte_dpaa2_object *object = NULL;
|
|
|
|
fslmc_vfio_setup_device(SYSFS_FSL_MC_DEVICES, dev->device.name,
|
|
&dev_fd, &device_info);
|
|
|
|
switch (dev->dev_type) {
|
|
case DPAA2_ETH:
|
|
rte_dpaa2_vfio_setup_intr(dev->intr_handle, dev_fd,
|
|
device_info.num_irqs);
|
|
break;
|
|
case DPAA2_CON:
|
|
case DPAA2_IO:
|
|
case DPAA2_CI:
|
|
case DPAA2_BPOOL:
|
|
case DPAA2_DPRTC:
|
|
case DPAA2_MUX:
|
|
case DPAA2_DPRC:
|
|
TAILQ_FOREACH(object, &dpaa2_obj_list, next) {
|
|
if (dev->dev_type == object->dev_type)
|
|
object->create(dev_fd, &device_info,
|
|
dev->object_id);
|
|
else
|
|
continue;
|
|
}
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
DPAA2_BUS_LOG(DEBUG, "Device (%s) abstracted from VFIO",
|
|
dev->device.name);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
fslmc_process_mcp(struct rte_dpaa2_device *dev)
|
|
{
|
|
int ret;
|
|
intptr_t v_addr;
|
|
struct fsl_mc_io dpmng = {0};
|
|
struct mc_version mc_ver_info = {0};
|
|
|
|
rte_mcp_ptr_list = malloc(sizeof(void *) * (MC_PORTAL_INDEX + 1));
|
|
if (!rte_mcp_ptr_list) {
|
|
DPAA2_BUS_ERR("Unable to allocate MC portal memory");
|
|
ret = -ENOMEM;
|
|
goto cleanup;
|
|
}
|
|
|
|
v_addr = vfio_map_mcp_obj(dev->device.name);
|
|
if (v_addr == (intptr_t)MAP_FAILED) {
|
|
DPAA2_BUS_ERR("Error mapping region (errno = %d)", errno);
|
|
ret = -1;
|
|
goto cleanup;
|
|
}
|
|
|
|
/* check the MC version compatibility */
|
|
dpmng.regs = (void *)v_addr;
|
|
|
|
/* In case of secondary processes, MC version check is no longer
|
|
* required.
|
|
*/
|
|
if (rte_eal_process_type() == RTE_PROC_SECONDARY) {
|
|
rte_mcp_ptr_list[MC_PORTAL_INDEX] = (void *)v_addr;
|
|
return 0;
|
|
}
|
|
|
|
if (mc_get_version(&dpmng, CMD_PRI_LOW, &mc_ver_info)) {
|
|
DPAA2_BUS_ERR("Unable to obtain MC version");
|
|
ret = -1;
|
|
goto cleanup;
|
|
}
|
|
|
|
if ((mc_ver_info.major != MC_VER_MAJOR) ||
|
|
(mc_ver_info.minor < MC_VER_MINOR)) {
|
|
DPAA2_BUS_ERR("DPAA2 MC version not compatible!"
|
|
" Expected %d.%d.x, Detected %d.%d.%d",
|
|
MC_VER_MAJOR, MC_VER_MINOR,
|
|
mc_ver_info.major, mc_ver_info.minor,
|
|
mc_ver_info.revision);
|
|
ret = -1;
|
|
goto cleanup;
|
|
}
|
|
rte_mcp_ptr_list[MC_PORTAL_INDEX] = (void *)v_addr;
|
|
|
|
return 0;
|
|
|
|
cleanup:
|
|
if (rte_mcp_ptr_list) {
|
|
free(rte_mcp_ptr_list);
|
|
rte_mcp_ptr_list = NULL;
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
int
|
|
fslmc_vfio_process_group(void)
|
|
{
|
|
int ret;
|
|
int found_mportal = 0;
|
|
struct rte_dpaa2_device *dev, *dev_temp;
|
|
bool is_dpmcp_in_blocklist = false, is_dpio_in_blocklist = false;
|
|
int dpmcp_count = 0, dpio_count = 0, current_device;
|
|
|
|
RTE_TAILQ_FOREACH_SAFE(dev, &rte_fslmc_bus.device_list, next,
|
|
dev_temp) {
|
|
if (dev->dev_type == DPAA2_MPORTAL) {
|
|
dpmcp_count++;
|
|
if (dev->device.devargs &&
|
|
dev->device.devargs->policy == RTE_DEV_BLOCKED)
|
|
is_dpmcp_in_blocklist = true;
|
|
}
|
|
if (dev->dev_type == DPAA2_IO) {
|
|
dpio_count++;
|
|
if (dev->device.devargs &&
|
|
dev->device.devargs->policy == RTE_DEV_BLOCKED)
|
|
is_dpio_in_blocklist = true;
|
|
}
|
|
}
|
|
|
|
/* Search the MCP as that should be initialized first. */
|
|
current_device = 0;
|
|
RTE_TAILQ_FOREACH_SAFE(dev, &rte_fslmc_bus.device_list, next,
|
|
dev_temp) {
|
|
if (dev->dev_type == DPAA2_MPORTAL) {
|
|
current_device++;
|
|
if (dev->device.devargs &&
|
|
dev->device.devargs->policy == RTE_DEV_BLOCKED) {
|
|
DPAA2_BUS_LOG(DEBUG, "%s Blocked, skipping",
|
|
dev->device.name);
|
|
TAILQ_REMOVE(&rte_fslmc_bus.device_list,
|
|
dev, next);
|
|
continue;
|
|
}
|
|
|
|
if (rte_eal_process_type() == RTE_PROC_SECONDARY &&
|
|
!is_dpmcp_in_blocklist) {
|
|
if (dpmcp_count == 1 ||
|
|
current_device != dpmcp_count) {
|
|
TAILQ_REMOVE(&rte_fslmc_bus.device_list,
|
|
dev, next);
|
|
continue;
|
|
}
|
|
}
|
|
|
|
if (!found_mportal) {
|
|
ret = fslmc_process_mcp(dev);
|
|
if (ret) {
|
|
DPAA2_BUS_ERR("Unable to map MC Portal");
|
|
return -1;
|
|
}
|
|
found_mportal = 1;
|
|
}
|
|
|
|
TAILQ_REMOVE(&rte_fslmc_bus.device_list, dev, next);
|
|
free(dev);
|
|
dev = NULL;
|
|
/* Ideally there is only a single dpmcp, but in case
|
|
* multiple exists, looping on remaining devices.
|
|
*/
|
|
}
|
|
}
|
|
|
|
/* Cannot continue if there is not even a single mportal */
|
|
if (!found_mportal) {
|
|
DPAA2_BUS_ERR("No MC Portal device found. Not continuing");
|
|
return -1;
|
|
}
|
|
|
|
/* Search for DPRC device next as it updates endpoint of
|
|
* other devices.
|
|
*/
|
|
current_device = 0;
|
|
RTE_TAILQ_FOREACH_SAFE(dev, &rte_fslmc_bus.device_list, next, dev_temp) {
|
|
if (dev->dev_type == DPAA2_DPRC) {
|
|
ret = fslmc_process_iodevices(dev);
|
|
if (ret) {
|
|
DPAA2_BUS_ERR("Unable to process dprc");
|
|
return -1;
|
|
}
|
|
TAILQ_REMOVE(&rte_fslmc_bus.device_list, dev, next);
|
|
}
|
|
}
|
|
|
|
current_device = 0;
|
|
RTE_TAILQ_FOREACH_SAFE(dev, &rte_fslmc_bus.device_list, next,
|
|
dev_temp) {
|
|
if (dev->dev_type == DPAA2_IO)
|
|
current_device++;
|
|
if (dev->device.devargs &&
|
|
dev->device.devargs->policy == RTE_DEV_BLOCKED) {
|
|
DPAA2_BUS_LOG(DEBUG, "%s Blocked, skipping",
|
|
dev->device.name);
|
|
TAILQ_REMOVE(&rte_fslmc_bus.device_list, dev, next);
|
|
continue;
|
|
}
|
|
if (rte_eal_process_type() == RTE_PROC_SECONDARY &&
|
|
dev->dev_type != DPAA2_ETH &&
|
|
dev->dev_type != DPAA2_CRYPTO &&
|
|
dev->dev_type != DPAA2_QDMA &&
|
|
dev->dev_type != DPAA2_IO) {
|
|
TAILQ_REMOVE(&rte_fslmc_bus.device_list, dev, next);
|
|
continue;
|
|
}
|
|
switch (dev->dev_type) {
|
|
case DPAA2_ETH:
|
|
case DPAA2_CRYPTO:
|
|
case DPAA2_QDMA:
|
|
ret = fslmc_process_iodevices(dev);
|
|
if (ret) {
|
|
DPAA2_BUS_DEBUG("Dev (%s) init failed",
|
|
dev->device.name);
|
|
return ret;
|
|
}
|
|
break;
|
|
case DPAA2_CON:
|
|
case DPAA2_CI:
|
|
case DPAA2_BPOOL:
|
|
case DPAA2_DPRTC:
|
|
case DPAA2_MUX:
|
|
/* IN case of secondary processes, all control objects
|
|
* like dpbp, dpcon, dpci are not initialized/required
|
|
* - all of these are assumed to be initialized and made
|
|
* available by primary.
|
|
*/
|
|
if (rte_eal_process_type() == RTE_PROC_SECONDARY)
|
|
continue;
|
|
|
|
/* Call the object creation routine and remove the
|
|
* device entry from device list
|
|
*/
|
|
ret = fslmc_process_iodevices(dev);
|
|
if (ret) {
|
|
DPAA2_BUS_DEBUG("Dev (%s) init failed",
|
|
dev->device.name);
|
|
return -1;
|
|
}
|
|
|
|
break;
|
|
case DPAA2_IO:
|
|
if (!is_dpio_in_blocklist && dpio_count > 1) {
|
|
if (rte_eal_process_type() == RTE_PROC_SECONDARY
|
|
&& current_device != dpio_count) {
|
|
TAILQ_REMOVE(&rte_fslmc_bus.device_list,
|
|
dev, next);
|
|
break;
|
|
}
|
|
if (rte_eal_process_type() == RTE_PROC_PRIMARY
|
|
&& current_device == dpio_count) {
|
|
TAILQ_REMOVE(&rte_fslmc_bus.device_list,
|
|
dev, next);
|
|
break;
|
|
}
|
|
}
|
|
|
|
ret = fslmc_process_iodevices(dev);
|
|
if (ret) {
|
|
DPAA2_BUS_DEBUG("Dev (%s) init failed",
|
|
dev->device.name);
|
|
return -1;
|
|
}
|
|
|
|
break;
|
|
case DPAA2_UNKNOWN:
|
|
default:
|
|
/* Unknown - ignore */
|
|
DPAA2_BUS_DEBUG("Found unknown device (%s)",
|
|
dev->device.name);
|
|
TAILQ_REMOVE(&rte_fslmc_bus.device_list, dev, next);
|
|
free(dev);
|
|
dev = NULL;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
fslmc_vfio_setup_group(void)
|
|
{
|
|
int groupid;
|
|
int ret;
|
|
int vfio_container_fd;
|
|
struct vfio_group_status status = { .argsz = sizeof(status) };
|
|
|
|
/* if already done once */
|
|
if (container_device_fd)
|
|
return 0;
|
|
|
|
ret = fslmc_get_container_group(&groupid);
|
|
if (ret)
|
|
return ret;
|
|
|
|
/* In case this group was already opened, continue without any
|
|
* processing.
|
|
*/
|
|
if (vfio_group.groupid == groupid) {
|
|
DPAA2_BUS_ERR("groupid already exists %d", groupid);
|
|
return 0;
|
|
}
|
|
|
|
ret = rte_vfio_container_create();
|
|
if (ret < 0) {
|
|
DPAA2_BUS_ERR("Failed to open VFIO container");
|
|
return ret;
|
|
}
|
|
vfio_container_fd = ret;
|
|
|
|
/* Get the actual group fd */
|
|
ret = rte_vfio_container_group_bind(vfio_container_fd, groupid);
|
|
if (ret < 0)
|
|
return ret;
|
|
vfio_group.fd = ret;
|
|
|
|
/* Check group viability */
|
|
ret = ioctl(vfio_group.fd, VFIO_GROUP_GET_STATUS, &status);
|
|
if (ret) {
|
|
DPAA2_BUS_ERR("VFIO error getting group status");
|
|
close(vfio_group.fd);
|
|
rte_vfio_clear_group(vfio_group.fd);
|
|
return ret;
|
|
}
|
|
|
|
if (!(status.flags & VFIO_GROUP_FLAGS_VIABLE)) {
|
|
DPAA2_BUS_ERR("VFIO group not viable");
|
|
close(vfio_group.fd);
|
|
rte_vfio_clear_group(vfio_group.fd);
|
|
return -EPERM;
|
|
}
|
|
/* Since Group is VIABLE, Store the groupid */
|
|
vfio_group.groupid = groupid;
|
|
|
|
/* check if group does not have a container yet */
|
|
if (!(status.flags & VFIO_GROUP_FLAGS_CONTAINER_SET)) {
|
|
/* Now connect this IOMMU group to given container */
|
|
ret = vfio_connect_container();
|
|
if (ret) {
|
|
DPAA2_BUS_ERR(
|
|
"Error connecting container with groupid %d",
|
|
groupid);
|
|
close(vfio_group.fd);
|
|
rte_vfio_clear_group(vfio_group.fd);
|
|
return ret;
|
|
}
|
|
}
|
|
|
|
/* Get Device information */
|
|
ret = ioctl(vfio_group.fd, VFIO_GROUP_GET_DEVICE_FD, fslmc_container);
|
|
if (ret < 0) {
|
|
DPAA2_BUS_ERR("Error getting device %s fd from group %d",
|
|
fslmc_container, vfio_group.groupid);
|
|
close(vfio_group.fd);
|
|
rte_vfio_clear_group(vfio_group.fd);
|
|
return ret;
|
|
}
|
|
container_device_fd = ret;
|
|
DPAA2_BUS_DEBUG("VFIO Container FD is [0x%X]",
|
|
container_device_fd);
|
|
|
|
return 0;
|
|
}
|