/* SPDX-License-Identifier: BSD-3-Clause * Copyright 2019 Mellanox Technologies, Ltd */ #include #include #include "mlx5_vdpa_utils.h" #include "mlx5_vdpa.h" int mlx5_vdpa_logging_enable(struct mlx5_vdpa_priv *priv, int enable) { struct mlx5_devx_virtq_attr attr = { .type = MLX5_VIRTQ_MODIFY_TYPE_DIRTY_BITMAP_DUMP_ENABLE, .dirty_bitmap_dump_enable = enable, }; struct mlx5_vdpa_virtq *virtq; SLIST_FOREACH(virtq, &priv->virtq_list, next) { attr.queue_index = virtq->index; if (mlx5_devx_cmd_modify_virtq(virtq->virtq, &attr)) { DRV_LOG(ERR, "Failed to modify virtq %d logging.", virtq->index); return -1; } } return 0; } int mlx5_vdpa_dirty_bitmap_set(struct mlx5_vdpa_priv *priv, uint64_t log_base, uint64_t log_size) { struct mlx5_devx_mkey_attr mkey_attr = { .addr = (uintptr_t)log_base, .size = log_size, .pd = priv->pdn, .pg_access = 1, .klm_array = NULL, .klm_num = 0, }; struct mlx5_devx_virtq_attr attr = { .type = MLX5_VIRTQ_MODIFY_TYPE_DIRTY_BITMAP_PARAMS, .dirty_bitmap_addr = log_base, .dirty_bitmap_size = log_size, }; struct mlx5_vdpa_query_mr *mr = rte_malloc(__func__, sizeof(*mr), 0); struct mlx5_vdpa_virtq *virtq; if (!mr) { DRV_LOG(ERR, "Failed to allocate mem for lm mr."); return -1; } mr->umem = mlx5_glue->devx_umem_reg(priv->ctx, (void *)(uintptr_t)log_base, log_size, IBV_ACCESS_LOCAL_WRITE); if (!mr->umem) { DRV_LOG(ERR, "Failed to register umem for lm mr."); goto err; } mkey_attr.umem_id = mr->umem->umem_id; mr->mkey = mlx5_devx_cmd_mkey_create(priv->ctx, &mkey_attr); if (!mr->mkey) { DRV_LOG(ERR, "Failed to create Mkey for lm."); goto err; } attr.dirty_bitmap_mkey = mr->mkey->id; SLIST_FOREACH(virtq, &priv->virtq_list, next) { attr.queue_index = virtq->index; if (mlx5_devx_cmd_modify_virtq(virtq->virtq, &attr)) { DRV_LOG(ERR, "Failed to modify virtq %d for lm.", virtq->index); goto err; } } mr->is_indirect = 0; SLIST_INSERT_HEAD(&priv->mr_list, mr, next); return 0; err: if (mr->mkey) mlx5_devx_cmd_destroy(mr->mkey); if (mr->umem) mlx5_glue->devx_umem_dereg(mr->umem); rte_free(mr); return -1; } #define MLX5_VDPA_USED_RING_LEN(size) \ ((size) * sizeof(struct vring_used_elem) + sizeof(uint16_t) * 3) int mlx5_vdpa_lm_log(struct mlx5_vdpa_priv *priv) { struct mlx5_devx_virtq_attr attr = {0}; struct mlx5_vdpa_virtq *virtq; uint64_t features; int ret = rte_vhost_get_negotiated_features(priv->vid, &features); if (ret) { DRV_LOG(ERR, "Failed to get negotiated features."); return -1; } if (!RTE_VHOST_NEED_LOG(features)) return 0; SLIST_FOREACH(virtq, &priv->virtq_list, next) { ret = mlx5_vdpa_virtq_modify(virtq, 0); if (ret) return -1; if (mlx5_devx_cmd_query_virtq(virtq->virtq, &attr)) { DRV_LOG(ERR, "Failed to query virtq %d.", virtq->index); return -1; } DRV_LOG(INFO, "Query vid %d vring %d: hw_available_idx=%d, " "hw_used_index=%d", priv->vid, virtq->index, attr.hw_available_index, attr.hw_used_index); ret = rte_vhost_set_vring_base(priv->vid, virtq->index, attr.hw_available_index, attr.hw_used_index); if (ret) { DRV_LOG(ERR, "Failed to set virtq %d base.", virtq->index); return -1; } rte_vhost_log_used_vring(priv->vid, virtq->index, 0, MLX5_VDPA_USED_RING_LEN(virtq->vq_size)); } return 0; }