numam-spdk/lib/nvme/nvme_ns_ocssd_cmd.c
Darek Stojaczyk 8c4ed83b49 vtophys: add length parameter to the vtophys function
This follows the same trend as the mem_map APIs.

Currently, most of the spdk_vtophys() callers manually
detect physically noncontiguous buffers to split them
into multiple physically contiguous chunks. This patch
is a first step towards encapsulating most of that logic
in a single place - in spdk_vtophys() itself.

This patch doesn't change any functionality on its own,
it only extends the API.

Change-Id: I16faa9dea270c370f2a814cd399f59055b5ccc3d
Signed-off-by: Darek Stojaczyk <dariusz.stojaczyk@intel.com>
Reviewed-on: https://review.gerrithub.io/c/438449
Tested-by: SPDK CI Jenkins <sys_sgci@intel.com>
Chandler-Test-Pool: SPDK Automated Test System <sys_sgsw@intel.com>
Reviewed-by: Ben Walker <benjamin.walker@intel.com>
Reviewed-by: wuzhouhui <wuzhouhui@kingsoft.com>
Reviewed-by: Shuhei Matsumoto <shuhei.matsumoto.xt@hitachi.com>
2019-01-09 08:24:59 +00:00

233 lines
6.9 KiB
C

/*-
* BSD LICENSE
*
* Copyright (c) Intel Corporation.
* All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
*
* * Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* * Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in
* the documentation and/or other materials provided with the
* distribution.
* * Neither the name of Intel Corporation nor the names of its
* contributors may be used to endorse or promote products derived
* from this software without specific prior written permission.
*
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
* OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*/
#include "spdk/nvme_ocssd.h"
#include "nvme_internal.h"
int
spdk_nvme_ocssd_ns_cmd_vector_reset(struct spdk_nvme_ns *ns,
struct spdk_nvme_qpair *qpair,
uint64_t *lba_list, uint32_t num_lbas,
struct spdk_ocssd_chunk_information_entry *chunk_info,
spdk_nvme_cmd_cb cb_fn, void *cb_arg)
{
struct nvme_request *req;
struct spdk_nvme_cmd *cmd;
if (!lba_list || (num_lbas == 0) ||
(num_lbas > SPDK_NVME_OCSSD_MAX_LBAL_ENTRIES)) {
return -EINVAL;
}
req = nvme_allocate_request_null(qpair, cb_fn, cb_arg);
if (req == NULL) {
return -ENOMEM;
}
cmd = &req->cmd;
cmd->opc = SPDK_OCSSD_OPC_VECTOR_RESET;
cmd->nsid = ns->id;
if (chunk_info != NULL) {
cmd->mptr = spdk_vtophys(chunk_info, NULL);
}
/*
* Dword 10 and 11 store a pointer to the list of logical block addresses.
* If there is a single entry in the LBA list, the logical block
* address should be stored instead.
*/
if (num_lbas == 1) {
*(uint64_t *)&cmd->cdw10 = *lba_list;
} else {
*(uint64_t *)&cmd->cdw10 = spdk_vtophys(lba_list, NULL);
}
cmd->cdw12 = num_lbas - 1;
return nvme_qpair_submit_request(qpair, req);
}
static int
_nvme_ocssd_ns_cmd_vector_rw_with_md(struct spdk_nvme_ns *ns,
struct spdk_nvme_qpair *qpair,
void *buffer, void *metadata,
uint64_t *lba_list, uint32_t num_lbas,
spdk_nvme_cmd_cb cb_fn, void *cb_arg,
enum spdk_ocssd_io_opcode opc,
uint32_t io_flags)
{
struct nvme_request *req;
struct spdk_nvme_cmd *cmd;
struct nvme_payload payload;
uint32_t valid_flags = SPDK_OCSSD_IO_FLAGS_LIMITED_RETRY;
if (io_flags & ~valid_flags) {
return -EINVAL;
}
if (!buffer || !lba_list || (num_lbas == 0) ||
(num_lbas > SPDK_NVME_OCSSD_MAX_LBAL_ENTRIES)) {
return -EINVAL;
}
payload = NVME_PAYLOAD_CONTIG(buffer, metadata);
req = nvme_allocate_request(qpair, &payload, num_lbas * ns->sector_size, cb_fn, cb_arg);
if (req == NULL) {
return -ENOMEM;
}
cmd = &req->cmd;
cmd->opc = opc;
cmd->nsid = ns->id;
/*
* Dword 10 and 11 store a pointer to the list of logical block addresses.
* If there is a single entry in the LBA list, the logical block
* address should be stored instead.
*/
if (num_lbas == 1) {
*(uint64_t *)&cmd->cdw10 = *lba_list;
} else {
*(uint64_t *)&cmd->cdw10 = spdk_vtophys(lba_list, NULL);
}
cmd->cdw12 = num_lbas - 1;
cmd->cdw12 |= io_flags;
return nvme_qpair_submit_request(qpair, req);
}
int
spdk_nvme_ocssd_ns_cmd_vector_write_with_md(struct spdk_nvme_ns *ns,
struct spdk_nvme_qpair *qpair,
void *buffer, void *metadata,
uint64_t *lba_list, uint32_t num_lbas,
spdk_nvme_cmd_cb cb_fn, void *cb_arg,
uint32_t io_flags)
{
return _nvme_ocssd_ns_cmd_vector_rw_with_md(ns, qpair, buffer, metadata, lba_list,
num_lbas, cb_fn, cb_arg, SPDK_OCSSD_OPC_VECTOR_WRITE, io_flags);
}
int
spdk_nvme_ocssd_ns_cmd_vector_write(struct spdk_nvme_ns *ns,
struct spdk_nvme_qpair *qpair,
void *buffer,
uint64_t *lba_list, uint32_t num_lbas,
spdk_nvme_cmd_cb cb_fn, void *cb_arg,
uint32_t io_flags)
{
return _nvme_ocssd_ns_cmd_vector_rw_with_md(ns, qpair, buffer, NULL, lba_list,
num_lbas, cb_fn, cb_arg, SPDK_OCSSD_OPC_VECTOR_WRITE, io_flags);
}
int
spdk_nvme_ocssd_ns_cmd_vector_read_with_md(struct spdk_nvme_ns *ns,
struct spdk_nvme_qpair *qpair,
void *buffer, void *metadata,
uint64_t *lba_list, uint32_t num_lbas,
spdk_nvme_cmd_cb cb_fn, void *cb_arg,
uint32_t io_flags)
{
return _nvme_ocssd_ns_cmd_vector_rw_with_md(ns, qpair, buffer, metadata, lba_list,
num_lbas, cb_fn, cb_arg, SPDK_OCSSD_OPC_VECTOR_READ, io_flags);
}
int
spdk_nvme_ocssd_ns_cmd_vector_read(struct spdk_nvme_ns *ns,
struct spdk_nvme_qpair *qpair,
void *buffer,
uint64_t *lba_list, uint32_t num_lbas,
spdk_nvme_cmd_cb cb_fn, void *cb_arg,
uint32_t io_flags)
{
return _nvme_ocssd_ns_cmd_vector_rw_with_md(ns, qpair, buffer, NULL, lba_list,
num_lbas, cb_fn, cb_arg, SPDK_OCSSD_OPC_VECTOR_READ, io_flags);
}
int
spdk_nvme_ocssd_ns_cmd_vector_copy(struct spdk_nvme_ns *ns,
struct spdk_nvme_qpair *qpair,
uint64_t *dst_lba_list,
uint64_t *src_lba_list,
uint32_t num_lbas,
spdk_nvme_cmd_cb cb_fn, void *cb_arg,
uint32_t io_flags)
{
struct nvme_request *req;
struct spdk_nvme_cmd *cmd;
uint32_t valid_flags = SPDK_OCSSD_IO_FLAGS_LIMITED_RETRY;
if (io_flags & ~valid_flags) {
return -EINVAL;
}
if (!dst_lba_list || !src_lba_list || (num_lbas == 0) ||
(num_lbas > SPDK_NVME_OCSSD_MAX_LBAL_ENTRIES)) {
return -EINVAL;
}
req = nvme_allocate_request_null(qpair, cb_fn, cb_arg);
if (req == NULL) {
return -ENOMEM;
}
cmd = &req->cmd;
cmd->opc = SPDK_OCSSD_OPC_VECTOR_COPY;
cmd->nsid = ns->id;
/*
* Dword 10 and 11 store a pointer to the list of source logical
* block addresses.
* Dword 14 and 15 store a pointer to the list of destination logical
* block addresses.
* If there is a single entry in the LBA list, the logical block
* address should be stored instead.
*/
if (num_lbas == 1) {
*(uint64_t *)&cmd->cdw10 = *src_lba_list;
*(uint64_t *)&cmd->cdw14 = *dst_lba_list;
} else {
*(uint64_t *)&cmd->cdw10 = spdk_vtophys(src_lba_list, NULL);
*(uint64_t *)&cmd->cdw14 = spdk_vtophys(dst_lba_list, NULL);
}
cmd->cdw12 = num_lbas - 1;
cmd->cdw12 |= io_flags;
return nvme_qpair_submit_request(qpair, req);
}