numam-spdk/lib/nvme/nvme_ns_ocssd_cmd.c
Changpeng Liu 68d3bb2de4 nvme: save separate metadata size to nvme request
Previously the SPDK NVMe driver always set PSDT to 01b for hardware SGLs
which is aligned to the Linux NVMe driver, for this case the metadata length
is not required when filling the NVMe command fields.  There is no alignment
nor granularity requirement for Data Blocks for PSDT 01b case.  And if the
drive reported that it needs dword alignment with SGL, for this case, when
using spearate metadata, it needs a length parameter to fill the SGL descriptor.

Change-Id: I56ffaada775fe66de7637dae15b509ee9556e80a
Signed-off-by: Changpeng Liu <changpeng.liu@intel.com>
Reviewed-on: https://review.spdk.io/gerrit/c/spdk/spdk/+/1351
Tested-by: SPDK CI Jenkins <sys_sgci@intel.com>
Reviewed-by: Shuhei Matsumoto <shuhei.matsumoto.xt@hitachi.com>
Reviewed-by: Jim Harris <james.r.harris@intel.com>
Reviewed-by: Aleksey Marchuk <alexeymar@mellanox.com>
2020-04-02 07:02:38 +00:00

234 lines
7.0 KiB
C

/*-
* BSD LICENSE
*
* Copyright (c) Intel Corporation.
* All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
*
* * Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* * Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in
* the documentation and/or other materials provided with the
* distribution.
* * Neither the name of Intel Corporation nor the names of its
* contributors may be used to endorse or promote products derived
* from this software without specific prior written permission.
*
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
* OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*/
#include "spdk/nvme_ocssd.h"
#include "nvme_internal.h"
int
spdk_nvme_ocssd_ns_cmd_vector_reset(struct spdk_nvme_ns *ns,
struct spdk_nvme_qpair *qpair,
uint64_t *lba_list, uint32_t num_lbas,
struct spdk_ocssd_chunk_information_entry *chunk_info,
spdk_nvme_cmd_cb cb_fn, void *cb_arg)
{
struct nvme_request *req;
struct spdk_nvme_cmd *cmd;
if (!lba_list || (num_lbas == 0) ||
(num_lbas > SPDK_NVME_OCSSD_MAX_LBAL_ENTRIES)) {
return -EINVAL;
}
req = nvme_allocate_request_null(qpair, cb_fn, cb_arg);
if (req == NULL) {
return -ENOMEM;
}
cmd = &req->cmd;
cmd->opc = SPDK_OCSSD_OPC_VECTOR_RESET;
cmd->nsid = ns->id;
if (chunk_info != NULL) {
cmd->mptr = spdk_vtophys(chunk_info, NULL);
}
/*
* Dword 10 and 11 store a pointer to the list of logical block addresses.
* If there is a single entry in the LBA list, the logical block
* address should be stored instead.
*/
if (num_lbas == 1) {
*(uint64_t *)&cmd->cdw10 = *lba_list;
} else {
*(uint64_t *)&cmd->cdw10 = spdk_vtophys(lba_list, NULL);
}
cmd->cdw12 = num_lbas - 1;
return nvme_qpair_submit_request(qpair, req);
}
static int
_nvme_ocssd_ns_cmd_vector_rw_with_md(struct spdk_nvme_ns *ns,
struct spdk_nvme_qpair *qpair,
void *buffer, void *metadata,
uint64_t *lba_list, uint32_t num_lbas,
spdk_nvme_cmd_cb cb_fn, void *cb_arg,
enum spdk_ocssd_io_opcode opc,
uint32_t io_flags)
{
struct nvme_request *req;
struct spdk_nvme_cmd *cmd;
struct nvme_payload payload;
uint32_t valid_flags = SPDK_OCSSD_IO_FLAGS_LIMITED_RETRY;
if (io_flags & ~valid_flags) {
return -EINVAL;
}
if (!buffer || !lba_list || (num_lbas == 0) ||
(num_lbas > SPDK_NVME_OCSSD_MAX_LBAL_ENTRIES)) {
return -EINVAL;
}
payload = NVME_PAYLOAD_CONTIG(buffer, metadata);
req = nvme_allocate_request(qpair, &payload, num_lbas * ns->sector_size, num_lbas * ns->md_size,
cb_fn, cb_arg);
if (req == NULL) {
return -ENOMEM;
}
cmd = &req->cmd;
cmd->opc = opc;
cmd->nsid = ns->id;
/*
* Dword 10 and 11 store a pointer to the list of logical block addresses.
* If there is a single entry in the LBA list, the logical block
* address should be stored instead.
*/
if (num_lbas == 1) {
*(uint64_t *)&cmd->cdw10 = *lba_list;
} else {
*(uint64_t *)&cmd->cdw10 = spdk_vtophys(lba_list, NULL);
}
cmd->cdw12 = num_lbas - 1;
cmd->cdw12 |= io_flags;
return nvme_qpair_submit_request(qpair, req);
}
int
spdk_nvme_ocssd_ns_cmd_vector_write_with_md(struct spdk_nvme_ns *ns,
struct spdk_nvme_qpair *qpair,
void *buffer, void *metadata,
uint64_t *lba_list, uint32_t num_lbas,
spdk_nvme_cmd_cb cb_fn, void *cb_arg,
uint32_t io_flags)
{
return _nvme_ocssd_ns_cmd_vector_rw_with_md(ns, qpair, buffer, metadata, lba_list,
num_lbas, cb_fn, cb_arg, SPDK_OCSSD_OPC_VECTOR_WRITE, io_flags);
}
int
spdk_nvme_ocssd_ns_cmd_vector_write(struct spdk_nvme_ns *ns,
struct spdk_nvme_qpair *qpair,
void *buffer,
uint64_t *lba_list, uint32_t num_lbas,
spdk_nvme_cmd_cb cb_fn, void *cb_arg,
uint32_t io_flags)
{
return _nvme_ocssd_ns_cmd_vector_rw_with_md(ns, qpair, buffer, NULL, lba_list,
num_lbas, cb_fn, cb_arg, SPDK_OCSSD_OPC_VECTOR_WRITE, io_flags);
}
int
spdk_nvme_ocssd_ns_cmd_vector_read_with_md(struct spdk_nvme_ns *ns,
struct spdk_nvme_qpair *qpair,
void *buffer, void *metadata,
uint64_t *lba_list, uint32_t num_lbas,
spdk_nvme_cmd_cb cb_fn, void *cb_arg,
uint32_t io_flags)
{
return _nvme_ocssd_ns_cmd_vector_rw_with_md(ns, qpair, buffer, metadata, lba_list,
num_lbas, cb_fn, cb_arg, SPDK_OCSSD_OPC_VECTOR_READ, io_flags);
}
int
spdk_nvme_ocssd_ns_cmd_vector_read(struct spdk_nvme_ns *ns,
struct spdk_nvme_qpair *qpair,
void *buffer,
uint64_t *lba_list, uint32_t num_lbas,
spdk_nvme_cmd_cb cb_fn, void *cb_arg,
uint32_t io_flags)
{
return _nvme_ocssd_ns_cmd_vector_rw_with_md(ns, qpair, buffer, NULL, lba_list,
num_lbas, cb_fn, cb_arg, SPDK_OCSSD_OPC_VECTOR_READ, io_flags);
}
int
spdk_nvme_ocssd_ns_cmd_vector_copy(struct spdk_nvme_ns *ns,
struct spdk_nvme_qpair *qpair,
uint64_t *dst_lba_list,
uint64_t *src_lba_list,
uint32_t num_lbas,
spdk_nvme_cmd_cb cb_fn, void *cb_arg,
uint32_t io_flags)
{
struct nvme_request *req;
struct spdk_nvme_cmd *cmd;
uint32_t valid_flags = SPDK_OCSSD_IO_FLAGS_LIMITED_RETRY;
if (io_flags & ~valid_flags) {
return -EINVAL;
}
if (!dst_lba_list || !src_lba_list || (num_lbas == 0) ||
(num_lbas > SPDK_NVME_OCSSD_MAX_LBAL_ENTRIES)) {
return -EINVAL;
}
req = nvme_allocate_request_null(qpair, cb_fn, cb_arg);
if (req == NULL) {
return -ENOMEM;
}
cmd = &req->cmd;
cmd->opc = SPDK_OCSSD_OPC_VECTOR_COPY;
cmd->nsid = ns->id;
/*
* Dword 10 and 11 store a pointer to the list of source logical
* block addresses.
* Dword 14 and 15 store a pointer to the list of destination logical
* block addresses.
* If there is a single entry in the LBA list, the logical block
* address should be stored instead.
*/
if (num_lbas == 1) {
*(uint64_t *)&cmd->cdw10 = *src_lba_list;
*(uint64_t *)&cmd->cdw14 = *dst_lba_list;
} else {
*(uint64_t *)&cmd->cdw10 = spdk_vtophys(src_lba_list, NULL);
*(uint64_t *)&cmd->cdw14 = spdk_vtophys(dst_lba_list, NULL);
}
cmd->cdw12 = num_lbas - 1;
cmd->cdw12 |= io_flags;
return nvme_qpair_submit_request(qpair, req);
}