Added virtio_pci_dev_init, an equivalent to virtio_user_dev_init. It takes an opaque pci context pointer that is provided via virtio PCI enumerate API. virtio_pci will do only PCI-related initialization. The virtio_dev creation and SCSI-specific logic now belong to upper layers Change-Id: I7ad450591f893c4ab953c09cfb6441b186736751 Signed-off-by: Dariusz Stojaczyk <dariuszx.stojaczyk@intel.com> Reviewed-on: https://review.gerrithub.io/388302 Tested-by: SPDK Automated Test System <sys_sgsw@intel.com> Reviewed-by: Daniel Verkamp <daniel.verkamp@intel.com> Reviewed-by: Jim Harris <james.r.harris@intel.com>
455 lines
12 KiB
C
455 lines
12 KiB
C
/*-
|
|
* BSD LICENSE
|
|
*
|
|
* Copyright(c) 2010-2016 Intel Corporation. All rights reserved.
|
|
* All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
*
|
|
* * Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* * Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in
|
|
* the documentation and/or other materials provided with the
|
|
* distribution.
|
|
* * Neither the name of Intel Corporation nor the names of its
|
|
* contributors may be used to endorse or promote products derived
|
|
* from this software without specific prior written permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
|
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
|
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
|
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
|
* OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
|
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
|
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
|
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
|
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
|
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
*/
|
|
|
|
#include "spdk/stdinc.h"
|
|
|
|
#include <sys/eventfd.h>
|
|
|
|
#include <linux/virtio_scsi.h>
|
|
|
|
#include <rte_config.h>
|
|
#include <rte_malloc.h>
|
|
#include <rte_alarm.h>
|
|
|
|
#include "virtio_user/vhost.h"
|
|
#include "virtio.h"
|
|
|
|
#include "spdk/string.h"
|
|
|
|
static int
|
|
virtio_user_create_queue(struct virtio_dev *vdev, uint32_t queue_sel)
|
|
{
|
|
struct virtio_user_dev *dev = vdev->ctx;
|
|
|
|
/* Of all per virtqueue MSGs, make sure VHOST_SET_VRING_CALL come
|
|
* firstly because vhost depends on this msg to allocate virtqueue
|
|
* pair.
|
|
*/
|
|
struct vhost_vring_file file;
|
|
|
|
file.index = queue_sel;
|
|
file.fd = dev->callfds[queue_sel];
|
|
dev->ops->send_request(dev, VHOST_USER_SET_VRING_CALL, &file);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
virtio_user_kick_queue(struct virtio_dev *vdev, uint32_t queue_sel)
|
|
{
|
|
struct virtio_user_dev *dev = vdev->ctx;
|
|
struct vhost_vring_file file;
|
|
struct vhost_vring_state state;
|
|
struct vring *vring = &dev->vrings[queue_sel];
|
|
struct vhost_vring_addr addr = {
|
|
.index = queue_sel,
|
|
.desc_user_addr = (uint64_t)(uintptr_t)vring->desc,
|
|
.avail_user_addr = (uint64_t)(uintptr_t)vring->avail,
|
|
.used_user_addr = (uint64_t)(uintptr_t)vring->used,
|
|
.log_guest_addr = 0,
|
|
.flags = 0, /* disable log */
|
|
};
|
|
|
|
state.index = queue_sel;
|
|
state.num = vring->num;
|
|
dev->ops->send_request(dev, VHOST_USER_SET_VRING_NUM, &state);
|
|
|
|
state.index = queue_sel;
|
|
state.num = 0; /* no reservation */
|
|
dev->ops->send_request(dev, VHOST_USER_SET_VRING_BASE, &state);
|
|
|
|
dev->ops->send_request(dev, VHOST_USER_SET_VRING_ADDR, &addr);
|
|
|
|
/* Of all per virtqueue MSGs, make sure VHOST_USER_SET_VRING_KICK comes
|
|
* lastly because vhost depends on this msg to judge if
|
|
* virtio is ready.
|
|
*/
|
|
file.index = queue_sel;
|
|
file.fd = dev->kickfds[queue_sel];
|
|
dev->ops->send_request(dev, VHOST_USER_SET_VRING_KICK, &file);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
virtio_user_stop_queue(struct virtio_dev *vdev, uint32_t queue_sel)
|
|
{
|
|
struct virtio_user_dev *dev = vdev->ctx;
|
|
struct vhost_vring_state state;
|
|
|
|
state.index = queue_sel;
|
|
state.num = 0;
|
|
dev->ops->send_request(dev, VHOST_USER_GET_VRING_BASE, &state);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
virtio_user_queue_setup(struct virtio_dev *vdev,
|
|
int (*fn)(struct virtio_dev *, uint32_t))
|
|
{
|
|
uint32_t i;
|
|
|
|
for (i = 0; i < vdev->max_queues; ++i) {
|
|
if (fn(vdev, i) < 0) {
|
|
SPDK_ERRLOG("setup tx vq fails: %"PRIu32".\n", i);
|
|
return -1;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
virtio_user_start_device(struct virtio_dev *vdev)
|
|
{
|
|
struct virtio_user_dev *dev = vdev->ctx;
|
|
int ret;
|
|
|
|
/* tell vhost to create queues */
|
|
if (virtio_user_queue_setup(vdev, virtio_user_create_queue) < 0)
|
|
return -1;
|
|
|
|
/* share memory regions */
|
|
ret = dev->ops->send_request(dev, VHOST_USER_SET_MEM_TABLE, NULL);
|
|
if (ret < 0)
|
|
return -1;
|
|
|
|
/* kick queues */
|
|
if (virtio_user_queue_setup(vdev, virtio_user_kick_queue) < 0)
|
|
return -1;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int virtio_user_stop_device(struct virtio_dev *vdev)
|
|
{
|
|
return virtio_user_queue_setup(vdev, virtio_user_stop_queue);
|
|
}
|
|
|
|
static int
|
|
virtio_user_dev_setup(struct virtio_dev *vdev)
|
|
{
|
|
struct virtio_user_dev *dev = vdev->ctx;
|
|
uint16_t i;
|
|
|
|
dev->vhostfd = -1;
|
|
|
|
for (i = 0; i < SPDK_VIRTIO_MAX_VIRTQUEUES; ++i) {
|
|
dev->callfds[i] = -1;
|
|
dev->kickfds[i] = -1;
|
|
}
|
|
|
|
dev->ops = &ops_user;
|
|
|
|
if (dev->ops->setup(dev) < 0)
|
|
return -1;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void
|
|
virtio_user_read_dev_config(struct virtio_dev *vdev, size_t offset,
|
|
void *dst, int length)
|
|
{
|
|
SPDK_ERRLOG("not supported offset=%zu, len=%d\n", offset, length);
|
|
}
|
|
|
|
static void
|
|
virtio_user_write_dev_config(struct virtio_dev *vdev, size_t offset,
|
|
const void *src, int length)
|
|
{
|
|
SPDK_ERRLOG("not supported offset=%zu, len=%d\n", offset, length);
|
|
}
|
|
|
|
static void
|
|
virtio_user_set_status(struct virtio_dev *vdev, uint8_t status)
|
|
{
|
|
struct virtio_user_dev *dev = vdev->ctx;
|
|
|
|
if (status & VIRTIO_CONFIG_S_DRIVER_OK) {
|
|
virtio_user_start_device(vdev);
|
|
} else if (status == VIRTIO_CONFIG_S_RESET &&
|
|
(dev->status & VIRTIO_CONFIG_S_DRIVER_OK)) {
|
|
virtio_user_stop_device(vdev);
|
|
}
|
|
dev->status = status;
|
|
}
|
|
|
|
static uint8_t
|
|
virtio_user_get_status(struct virtio_dev *vdev)
|
|
{
|
|
struct virtio_user_dev *dev = vdev->ctx;
|
|
|
|
return dev->status;
|
|
}
|
|
|
|
static uint64_t
|
|
virtio_user_get_features(struct virtio_dev *vdev)
|
|
{
|
|
struct virtio_user_dev *dev = vdev->ctx;
|
|
char err_str[64];
|
|
uint64_t features;
|
|
|
|
if (dev->ops->send_request(dev, VHOST_USER_GET_FEATURES, &features) < 0) {
|
|
spdk_strerror_r(errno, err_str, sizeof(err_str));
|
|
SPDK_ERRLOG("get_features failed: %s\n", err_str);
|
|
return 0;
|
|
}
|
|
|
|
return features;
|
|
}
|
|
|
|
static int
|
|
virtio_user_set_features(struct virtio_dev *vdev, uint64_t features)
|
|
{
|
|
struct virtio_user_dev *dev = vdev->ctx;
|
|
int ret;
|
|
|
|
ret = dev->ops->send_request(dev, VHOST_USER_SET_FEATURES, &features);
|
|
if (ret < 0) {
|
|
return -1;
|
|
}
|
|
|
|
vdev->negotiated_features = features;
|
|
vdev->modern = virtio_dev_has_feature(vdev, VIRTIO_F_VERSION_1);
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* This function is to get the queue size, aka, number of descs, of a specified
|
|
* queue. Different with the VHOST_USER_GET_QUEUE_NUM, which is used to get the
|
|
* max supported queues.
|
|
*/
|
|
static uint16_t
|
|
virtio_user_get_queue_num(struct virtio_dev *vdev, uint16_t queue_id)
|
|
{
|
|
struct virtio_user_dev *dev = vdev->ctx;
|
|
|
|
/* Currently, each queue has same queue size */
|
|
return dev->queue_size;
|
|
}
|
|
|
|
static int
|
|
virtio_user_setup_queue(struct virtio_dev *vdev, struct virtqueue *vq)
|
|
{
|
|
struct virtio_user_dev *dev = vdev->ctx;
|
|
uint16_t queue_idx = vq->vq_queue_index;
|
|
uint64_t desc_addr, avail_addr, used_addr;
|
|
char err_str[64];
|
|
int callfd;
|
|
int kickfd;
|
|
|
|
if (dev->callfds[queue_idx] != -1 || dev->kickfds[queue_idx] != -1) {
|
|
SPDK_ERRLOG("queue %"PRIu16" already exists\n", queue_idx);
|
|
return -1;
|
|
}
|
|
|
|
/* May use invalid flag, but some backend uses kickfd and
|
|
* callfd as criteria to judge if dev is alive. so finally we
|
|
* use real event_fd.
|
|
*/
|
|
callfd = eventfd(0, EFD_CLOEXEC | EFD_NONBLOCK);
|
|
if (callfd < 0) {
|
|
spdk_strerror_r(errno, err_str, sizeof(err_str));
|
|
SPDK_ERRLOG("callfd error, %s\n", err_str);
|
|
return -1;
|
|
}
|
|
|
|
kickfd = eventfd(0, EFD_CLOEXEC | EFD_NONBLOCK);
|
|
if (kickfd < 0) {
|
|
spdk_strerror_r(errno, err_str, sizeof(err_str));
|
|
SPDK_ERRLOG("kickfd error, %s\n", err_str);
|
|
close(callfd);
|
|
return -1;
|
|
}
|
|
|
|
dev->callfds[queue_idx] = callfd;
|
|
dev->kickfds[queue_idx] = kickfd;
|
|
|
|
desc_addr = (uintptr_t)vq->vq_ring_virt_mem;
|
|
avail_addr = desc_addr + vq->vq_nentries * sizeof(struct vring_desc);
|
|
used_addr = RTE_ALIGN_CEIL(avail_addr + offsetof(struct vring_avail,
|
|
ring[vq->vq_nentries]),
|
|
VIRTIO_PCI_VRING_ALIGN);
|
|
|
|
dev->vrings[queue_idx].num = vq->vq_nentries;
|
|
dev->vrings[queue_idx].desc = (void *)(uintptr_t)desc_addr;
|
|
dev->vrings[queue_idx].avail = (void *)(uintptr_t)avail_addr;
|
|
dev->vrings[queue_idx].used = (void *)(uintptr_t)used_addr;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void
|
|
virtio_user_del_queue(struct virtio_dev *vdev, struct virtqueue *vq)
|
|
{
|
|
/* For legacy devices, write 0 to VIRTIO_PCI_QUEUE_PFN port, QEMU
|
|
* correspondingly stops the ioeventfds, and reset the status of
|
|
* the device.
|
|
* For modern devices, set queue desc, avail, used in PCI bar to 0,
|
|
* not see any more behavior in QEMU.
|
|
*
|
|
* Here we just care about what information to deliver to vhost-user
|
|
* or vhost-kernel. So we just close ioeventfd for now.
|
|
*/
|
|
struct virtio_user_dev *dev = vdev->ctx;
|
|
|
|
close(dev->callfds[vq->vq_queue_index]);
|
|
close(dev->kickfds[vq->vq_queue_index]);
|
|
dev->callfds[vq->vq_queue_index] = -1;
|
|
dev->kickfds[vq->vq_queue_index] = -1;
|
|
}
|
|
|
|
static void
|
|
virtio_user_notify_queue(struct virtio_dev *vdev, struct virtqueue *vq)
|
|
{
|
|
uint64_t buf = 1;
|
|
struct virtio_user_dev *dev = vdev->ctx;
|
|
char err_str[64];
|
|
|
|
if (write(dev->kickfds[vq->vq_queue_index], &buf, sizeof(buf)) < 0) {
|
|
spdk_strerror_r(errno, err_str, sizeof(err_str));
|
|
SPDK_ERRLOG("failed to kick backend: %s.\n", err_str);
|
|
}
|
|
}
|
|
|
|
static void
|
|
virtio_user_destroy(struct virtio_dev *vdev)
|
|
{
|
|
struct virtio_user_dev *dev = vdev->ctx;
|
|
|
|
close(dev->vhostfd);
|
|
free(vdev->name);
|
|
free(dev);
|
|
}
|
|
|
|
static void
|
|
virtio_user_dump_json_config(struct virtio_dev *vdev, struct spdk_json_write_ctx *w)
|
|
{
|
|
struct virtio_user_dev *dev = vdev->ctx;
|
|
|
|
spdk_json_write_name(w, "type");
|
|
spdk_json_write_string(w, "user");
|
|
|
|
spdk_json_write_name(w, "socket");
|
|
spdk_json_write_string(w, dev->path);
|
|
}
|
|
|
|
static const struct virtio_dev_ops virtio_user_ops = {
|
|
.read_dev_cfg = virtio_user_read_dev_config,
|
|
.write_dev_cfg = virtio_user_write_dev_config,
|
|
.get_status = virtio_user_get_status,
|
|
.set_status = virtio_user_set_status,
|
|
.get_features = virtio_user_get_features,
|
|
.set_features = virtio_user_set_features,
|
|
.destruct_dev = virtio_user_destroy,
|
|
.get_queue_num = virtio_user_get_queue_num,
|
|
.setup_queue = virtio_user_setup_queue,
|
|
.del_queue = virtio_user_del_queue,
|
|
.notify_queue = virtio_user_notify_queue,
|
|
.dump_json_config = virtio_user_dump_json_config,
|
|
};
|
|
|
|
int
|
|
virtio_user_dev_init(struct virtio_dev *vdev, const char *name, const char *path,
|
|
uint16_t requested_queues, uint32_t queue_size, uint16_t fixed_queue_num)
|
|
{
|
|
struct virtio_user_dev *dev;
|
|
uint64_t max_queues;
|
|
char err_str[64];
|
|
int rc;
|
|
|
|
if (name == NULL) {
|
|
SPDK_ERRLOG("No name gived for controller: %s\n", path);
|
|
return -1;
|
|
} else if (requested_queues == 0) {
|
|
SPDK_ERRLOG("Can't create controller with no queues: %s\n", path);
|
|
return -1;
|
|
}
|
|
|
|
dev = calloc(1, sizeof(*dev));
|
|
if (dev == NULL) {
|
|
return -1;
|
|
}
|
|
|
|
rc = virtio_dev_construct(vdev, &virtio_user_ops, dev);
|
|
if (rc != 0) {
|
|
SPDK_ERRLOG("Failed to init device: %s\n", path);
|
|
free(dev);
|
|
return -1;
|
|
}
|
|
|
|
vdev->is_hw = 0;
|
|
vdev->name = strdup(name);
|
|
if (!vdev->name) {
|
|
SPDK_ERRLOG("Failed to reserve memory for controller name: %s\n", path);
|
|
goto err;
|
|
}
|
|
|
|
snprintf(dev->path, PATH_MAX, "%s", path);
|
|
dev->queue_size = queue_size;
|
|
|
|
if (virtio_user_dev_setup(vdev) < 0) {
|
|
SPDK_ERRLOG("backend set up fails\n");
|
|
goto err;
|
|
}
|
|
|
|
if (dev->ops->send_request(dev, VHOST_USER_GET_QUEUE_NUM, &max_queues) < 0) {
|
|
spdk_strerror_r(errno, err_str, sizeof(err_str));
|
|
SPDK_ERRLOG("get_queue_num fails: %s\n", err_str);
|
|
goto err;
|
|
}
|
|
|
|
if (requested_queues > max_queues) {
|
|
SPDK_ERRLOG("requested %"PRIu16" request queues but only %"PRIu64" available\n",
|
|
requested_queues, max_queues);
|
|
goto err;
|
|
}
|
|
|
|
vdev->max_queues = fixed_queue_num + requested_queues;
|
|
|
|
if (dev->ops->send_request(dev, VHOST_USER_SET_OWNER, NULL) < 0) {
|
|
spdk_strerror_r(errno, err_str, sizeof(err_str));
|
|
SPDK_ERRLOG("set_owner fails: %s\n", err_str);
|
|
goto err;
|
|
}
|
|
|
|
TAILQ_INSERT_TAIL(&g_virtio_driver.init_ctrlrs, vdev, tailq);
|
|
return 0;
|
|
|
|
err:
|
|
virtio_dev_destruct(vdev);
|
|
return -1;
|
|
}
|