vdpa/ifc: support virtio block device

Re-use the vdpa/ifc code, distinguish blk and net device by pci_device_id.
Blk and net device are implemented with proper feature and ops.

Signed-off-by: Andy Pei <andy.pei@intel.com>
Reviewed-by: Chenbo Xia <chenbo.xia@intel.com>
This commit is contained in:
Andy Pei 2022-05-24 10:48:05 +08:00 committed by Maxime Coquelin
parent 9e89b06d02
commit a60b747d0a
2 changed files with 102 additions and 15 deletions

View File

@ -5,12 +5,19 @@
#ifndef _IFCVF_H_
#define _IFCVF_H_
#include <linux/virtio_blk.h>
#include "ifcvf_osdep.h"
#define IFCVF_VENDOR_ID 0x1AF4
#define IFCVF_DEVICE_ID 0x1041
#define IFCVF_SUBSYS_VENDOR_ID 0x8086
#define IFCVF_SUBSYS_DEVICE_ID 0x001A
#define IFCVF_NET 0
#define IFCVF_BLK 1
#define IFCVF_VENDOR_ID 0x1AF4
#define IFCVF_NET_DEVICE_ID 0x1041
#define IFCVF_BLK_MODERN_DEVICE_ID 0x1042
#define IFCVF_BLK_TRANSITIONAL_DEVICE_ID 0x1001
#define IFCVF_SUBSYS_VENDOR_ID 0x8086
#define IFCVF_SUBSYS_DEVICE_ID 0x001A
#define IFCVF_BLK_DEVICE_ID 0x0002
#define IFCVF_MAX_QUEUES 1
@ -126,13 +133,18 @@ struct ifcvf_hw {
u8 notify_region;
u32 notify_off_multiplier;
struct ifcvf_pci_common_cfg *common_cfg;
struct ifcvf_net_config *dev_cfg;
union {
struct ifcvf_net_config *net_cfg;
struct virtio_blk_config *blk_cfg;
void *dev_cfg;
};
u8 *isr;
u16 *notify_base;
u16 *notify_addr[IFCVF_MAX_QUEUES * 2];
u8 *lm_cfg;
struct vring_info vring[IFCVF_MAX_QUEUES * 2];
u8 nr_vring;
int device_type;
struct ifcvf_pci_mem_resource mem_resource[IFCVF_PCI_MAX_RESOURCE];
};

View File

@ -75,6 +75,12 @@ struct internal_list {
struct ifcvf_internal *internal;
};
/* vdpa device info includes device features and devcic operation. */
struct rte_vdpa_dev_info {
uint64_t features;
struct rte_vdpa_dev_ops *ops;
};
TAILQ_HEAD(internal_list_head, internal_list);
static struct internal_list_head internal_list =
TAILQ_HEAD_INITIALIZER(internal_list);
@ -1138,7 +1144,7 @@ exit:
return 0;
}
static struct rte_vdpa_dev_ops ifcvf_ops = {
static struct rte_vdpa_dev_ops ifcvf_net_ops = {
.get_queue_num = ifcvf_get_queue_num,
.get_features = ifcvf_get_vdpa_features,
.get_protocol_features = ifcvf_get_protocol_features,
@ -1167,6 +1173,48 @@ open_int(const char *key __rte_unused, const char *value, void *extra_args)
return 0;
}
static int16_t
ifcvf_pci_get_device_type(struct rte_pci_device *pci_dev)
{
uint16_t pci_device_id = pci_dev->id.device_id;
uint16_t device_id;
if (pci_device_id < 0x1000 || pci_device_id > 0x107f) {
DRV_LOG(ERR, "Probe device is not a virtio device\n");
return -1;
}
if (pci_device_id < 0x1040) {
/* Transitional devices: use the PCI subsystem device id as
* virtio device id, same as legacy driver always did.
*/
device_id = pci_dev->id.subsystem_device_id;
} else {
/* Modern devices: simply use PCI device id,
* but start from 0x1040.
*/
device_id = pci_device_id - 0x1040;
}
return device_id;
}
struct rte_vdpa_dev_info dev_info[] = {
{
.features = (1ULL << VIRTIO_NET_F_GUEST_ANNOUNCE) |
(1ULL << VIRTIO_NET_F_CTRL_VQ) |
(1ULL << VIRTIO_NET_F_STATUS) |
(1ULL << VHOST_USER_F_PROTOCOL_FEATURES) |
(1ULL << VHOST_F_LOG_ALL),
.ops = &ifcvf_net_ops,
},
{
.features = (1ULL << VHOST_USER_F_PROTOCOL_FEATURES) |
(1ULL << VHOST_F_LOG_ALL),
.ops = NULL,
},
};
static int
ifcvf_pci_probe(struct rte_pci_driver *pci_drv __rte_unused,
struct rte_pci_device *pci_dev)
@ -1178,6 +1226,7 @@ ifcvf_pci_probe(struct rte_pci_driver *pci_drv __rte_unused,
int sw_fallback_lm = 0;
struct rte_kvargs *kvlist = NULL;
int ret = 0;
int16_t device_id;
if (rte_eal_process_type() != RTE_PROC_PRIMARY)
return 0;
@ -1227,13 +1276,24 @@ ifcvf_pci_probe(struct rte_pci_driver *pci_drv __rte_unused,
internal->configured = 0;
internal->max_queues = IFCVF_MAX_QUEUES;
features = ifcvf_get_features(&internal->hw);
internal->features = (features &
~(1ULL << VIRTIO_F_IOMMU_PLATFORM)) |
(1ULL << VIRTIO_NET_F_GUEST_ANNOUNCE) |
(1ULL << VIRTIO_NET_F_CTRL_VQ) |
(1ULL << VIRTIO_NET_F_STATUS) |
(1ULL << VHOST_USER_F_PROTOCOL_FEATURES) |
(1ULL << VHOST_F_LOG_ALL);
device_id = ifcvf_pci_get_device_type(pci_dev);
if (device_id < 0) {
DRV_LOG(ERR, "failed to get device %s type", pci_dev->name);
goto error;
}
if (device_id == VIRTIO_ID_NET) {
internal->hw.device_type = IFCVF_NET;
internal->features = features &
~(1ULL << VIRTIO_F_IOMMU_PLATFORM);
internal->features |= dev_info[IFCVF_NET].features;
} else if (device_id == VIRTIO_ID_BLOCK) {
internal->hw.device_type = IFCVF_BLK;
internal->features = features &
~(1ULL << VIRTIO_F_IOMMU_PLATFORM);
internal->features |= dev_info[IFCVF_BLK].features;
}
list->internal = internal;
@ -1245,7 +1305,8 @@ ifcvf_pci_probe(struct rte_pci_driver *pci_drv __rte_unused,
}
internal->sw_lm = sw_fallback_lm;
internal->vdev = rte_vdpa_register_device(&pci_dev->device, &ifcvf_ops);
internal->vdev = rte_vdpa_register_device(&pci_dev->device,
dev_info[internal->hw.device_type].ops);
if (internal->vdev == NULL) {
DRV_LOG(ERR, "failed to register device %s", pci_dev->name);
goto error;
@ -1308,11 +1369,25 @@ ifcvf_pci_remove(struct rte_pci_device *pci_dev)
static const struct rte_pci_id pci_id_ifcvf_map[] = {
{ .class_id = RTE_CLASS_ANY_ID,
.vendor_id = IFCVF_VENDOR_ID,
.device_id = IFCVF_DEVICE_ID,
.device_id = IFCVF_NET_DEVICE_ID,
.subsystem_vendor_id = IFCVF_SUBSYS_VENDOR_ID,
.subsystem_device_id = IFCVF_SUBSYS_DEVICE_ID,
},
{ .class_id = RTE_CLASS_ANY_ID,
.vendor_id = IFCVF_VENDOR_ID,
.device_id = IFCVF_BLK_TRANSITIONAL_DEVICE_ID,
.subsystem_vendor_id = IFCVF_SUBSYS_VENDOR_ID,
.subsystem_device_id = IFCVF_BLK_DEVICE_ID,
},
{ .class_id = RTE_CLASS_ANY_ID,
.vendor_id = IFCVF_VENDOR_ID,
.device_id = IFCVF_BLK_MODERN_DEVICE_ID,
.subsystem_vendor_id = IFCVF_SUBSYS_VENDOR_ID,
.subsystem_device_id = IFCVF_BLK_DEVICE_ID,
},
{ .vendor_id = 0, /* sentinel */
},
};