2020-09-25 12:10:48 +00:00
|
|
|
/*-
|
|
|
|
* BSD LICENSE
|
|
|
|
*
|
|
|
|
* Copyright (c) 2020, Western Digital Corporation. All rights reserved.
|
2020-11-23 17:11:12 +00:00
|
|
|
* Copyright (c) 2021 Mellanox Technologies LTD. All rights reserved.
|
2020-09-25 12:10:48 +00:00
|
|
|
*
|
|
|
|
* Redistribution and use in source and binary forms, with or without
|
|
|
|
* modification, are permitted provided that the following conditions
|
|
|
|
* are met:
|
|
|
|
*
|
|
|
|
* * Redistributions of source code must retain the above copyright
|
|
|
|
* notice, this list of conditions and the following disclaimer.
|
|
|
|
* * Redistributions in binary form must reproduce the above copyright
|
|
|
|
* notice, this list of conditions and the following disclaimer in
|
|
|
|
* the documentation and/or other materials provided with the
|
|
|
|
* distribution.
|
|
|
|
* * Neither the name of Intel Corporation nor the names of its
|
|
|
|
* contributors may be used to endorse or promote products derived
|
|
|
|
* from this software without specific prior written permission.
|
|
|
|
*
|
|
|
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
|
|
|
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
|
|
|
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
|
|
|
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
|
|
|
* OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
|
|
|
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
|
|
|
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
|
|
|
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
|
|
|
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
|
|
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
|
|
|
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
|
|
*/
|
|
|
|
|
2020-10-13 11:46:16 +00:00
|
|
|
#include "spdk/nvme_zns.h"
|
2020-09-25 12:10:48 +00:00
|
|
|
#include "nvme_internal.h"
|
|
|
|
|
|
|
|
const struct spdk_nvme_zns_ns_data *
|
|
|
|
spdk_nvme_zns_ns_get_data(struct spdk_nvme_ns *ns)
|
|
|
|
{
|
2020-11-23 17:11:12 +00:00
|
|
|
return ns->nsdata_zns;
|
2020-09-25 12:10:48 +00:00
|
|
|
}
|
|
|
|
|
2020-10-20 09:29:01 +00:00
|
|
|
uint64_t
|
2021-02-15 13:50:14 +00:00
|
|
|
spdk_nvme_zns_ns_get_zone_size_sectors(struct spdk_nvme_ns *ns)
|
2020-10-20 09:29:01 +00:00
|
|
|
{
|
|
|
|
const struct spdk_nvme_zns_ns_data *nsdata_zns = spdk_nvme_zns_ns_get_data(ns);
|
|
|
|
const struct spdk_nvme_ns_data *nsdata = spdk_nvme_ns_get_data(ns);
|
|
|
|
|
2021-02-15 13:50:14 +00:00
|
|
|
return nsdata_zns->lbafe[nsdata->flbas.format].zsze;
|
|
|
|
}
|
|
|
|
|
|
|
|
uint64_t
|
|
|
|
spdk_nvme_zns_ns_get_zone_size(struct spdk_nvme_ns *ns)
|
|
|
|
{
|
|
|
|
return spdk_nvme_zns_ns_get_zone_size_sectors(ns) * spdk_nvme_ns_get_sector_size(ns);
|
2020-10-20 09:29:01 +00:00
|
|
|
}
|
|
|
|
|
2020-10-20 09:30:12 +00:00
|
|
|
uint64_t
|
|
|
|
spdk_nvme_zns_ns_get_num_zones(struct spdk_nvme_ns *ns)
|
|
|
|
{
|
2021-02-15 13:55:53 +00:00
|
|
|
return spdk_nvme_ns_get_num_sectors(ns) / spdk_nvme_zns_ns_get_zone_size_sectors(ns);
|
2020-10-20 09:30:12 +00:00
|
|
|
}
|
|
|
|
|
2021-02-16 12:40:55 +00:00
|
|
|
uint32_t
|
|
|
|
spdk_nvme_zns_ns_get_max_open_zones(struct spdk_nvme_ns *ns)
|
|
|
|
{
|
|
|
|
const struct spdk_nvme_zns_ns_data *nsdata_zns = spdk_nvme_zns_ns_get_data(ns);
|
|
|
|
|
|
|
|
return nsdata_zns->mor + 1;
|
|
|
|
}
|
|
|
|
|
2021-02-16 12:42:38 +00:00
|
|
|
uint32_t
|
|
|
|
spdk_nvme_zns_ns_get_max_active_zones(struct spdk_nvme_ns *ns)
|
|
|
|
{
|
|
|
|
const struct spdk_nvme_zns_ns_data *nsdata_zns = spdk_nvme_zns_ns_get_data(ns);
|
|
|
|
|
|
|
|
return nsdata_zns->mar + 1;
|
|
|
|
}
|
|
|
|
|
2020-09-25 12:10:48 +00:00
|
|
|
const struct spdk_nvme_zns_ctrlr_data *
|
|
|
|
spdk_nvme_zns_ctrlr_get_data(struct spdk_nvme_ctrlr *ctrlr)
|
|
|
|
{
|
|
|
|
return ctrlr->cdata_zns;
|
|
|
|
}
|
2020-10-13 12:29:54 +00:00
|
|
|
|
2020-11-17 15:44:04 +00:00
|
|
|
uint32_t
|
|
|
|
spdk_nvme_zns_ctrlr_get_max_zone_append_size(const struct spdk_nvme_ctrlr *ctrlr)
|
|
|
|
{
|
|
|
|
return ctrlr->max_zone_append_size;
|
|
|
|
}
|
|
|
|
|
2021-01-19 11:10:56 +00:00
|
|
|
int
|
|
|
|
spdk_nvme_zns_zone_append(struct spdk_nvme_ns *ns, struct spdk_nvme_qpair *qpair,
|
|
|
|
void *buffer, uint64_t zslba,
|
|
|
|
uint32_t lba_count, spdk_nvme_cmd_cb cb_fn, void *cb_arg,
|
|
|
|
uint32_t io_flags)
|
|
|
|
{
|
|
|
|
return nvme_ns_cmd_zone_append_with_md(ns, qpair, buffer, NULL, zslba, lba_count,
|
|
|
|
cb_fn, cb_arg, io_flags, 0, 0);
|
|
|
|
}
|
|
|
|
|
|
|
|
int
|
|
|
|
spdk_nvme_zns_zone_append_with_md(struct spdk_nvme_ns *ns, struct spdk_nvme_qpair *qpair,
|
|
|
|
void *buffer, void *metadata, uint64_t zslba,
|
|
|
|
uint32_t lba_count, spdk_nvme_cmd_cb cb_fn, void *cb_arg,
|
|
|
|
uint32_t io_flags, uint16_t apptag_mask, uint16_t apptag)
|
|
|
|
{
|
|
|
|
return nvme_ns_cmd_zone_append_with_md(ns, qpair, buffer, metadata, zslba, lba_count,
|
|
|
|
cb_fn, cb_arg, io_flags, apptag_mask, apptag);
|
|
|
|
}
|
|
|
|
|
nvme: add support for ZNS zone append vector variant
We already have support for spdk_nvme_zns_zone_append(),
add support for spdk_nvme_zns_zone_appendv() (zone append with
NVME_PAYLOAD_TYPE_SGL).
_nvme_ns_cmd_rw() currently performs verification of the SGL,
if the parameter check_sgl is set. This parameter is set for all
calls with payload of type NVME_PAYLOAD_TYPE_SGL.
In order to be able to perform the same check_sgl verfication on
zone append vectors, we need to refactor _nvme_ns_cmd_rw() a bit.
Setting check_sgl ensures that _nvme_ns_cmd_split_request_sgl() or
_nvme_ns_cmd_split_request_prp() gets called.
These functions will split an oversized I/O into several different
requests. However, they also iterate the SGE entries, verifies that
the total payload size, total SGE entries is not too many, and that
buffers are properly aligned. A proper request will not get split.
For zone append, splitting a request into several is not allowed,
however, we still want the verification part to be done, such that
(e.g.) a non first/last SGE which is not page aligned, will cause
the whole request to be rejected.
(In the case of spdk_nvme_ns_cmd_write(), a non first/last SGE which
is not page aligned will instead cause the request to be split.)
An alternative would be to try to rip out the verification part from
_nvme_ns_cmd_split_request_sgl() and _nvme_ns_cmd_split_request_prp().
However, that is non-trivial, and would most likely end up with a lot
of duplicated code, which would easily get out of sync.
Signed-off-by: Niklas Cassel <niklas.cassel@wdc.com>
Change-Id: I2728acdcadeb70b1f0ed628704df19e75d14dcca
Reviewed-on: https://review.spdk.io/gerrit/c/spdk/spdk/+/6248
Tested-by: SPDK CI Jenkins <sys_sgci@intel.com>
Community-CI: Mellanox Build Bot
Reviewed-by: Ben Walker <benjamin.walker@intel.com>
Reviewed-by: Shuhei Matsumoto <shuhei.matsumoto.xt@hitachi.com>
Reviewed-by: Jim Harris <james.r.harris@intel.com>
2021-01-26 09:42:58 +00:00
|
|
|
int
|
|
|
|
spdk_nvme_zns_zone_appendv(struct spdk_nvme_ns *ns, struct spdk_nvme_qpair *qpair,
|
|
|
|
uint64_t zslba, uint32_t lba_count,
|
|
|
|
spdk_nvme_cmd_cb cb_fn, void *cb_arg, uint32_t io_flags,
|
|
|
|
spdk_nvme_req_reset_sgl_cb reset_sgl_fn,
|
|
|
|
spdk_nvme_req_next_sge_cb next_sge_fn)
|
|
|
|
{
|
|
|
|
return nvme_ns_cmd_zone_appendv_with_md(ns, qpair, zslba, lba_count, cb_fn, cb_arg,
|
|
|
|
io_flags, reset_sgl_fn, next_sge_fn,
|
|
|
|
NULL, 0, 0);
|
|
|
|
}
|
|
|
|
|
|
|
|
int
|
|
|
|
spdk_nvme_zns_zone_appendv_with_md(struct spdk_nvme_ns *ns, struct spdk_nvme_qpair *qpair,
|
|
|
|
uint64_t zslba, uint32_t lba_count,
|
|
|
|
spdk_nvme_cmd_cb cb_fn, void *cb_arg, uint32_t io_flags,
|
|
|
|
spdk_nvme_req_reset_sgl_cb reset_sgl_fn,
|
|
|
|
spdk_nvme_req_next_sge_cb next_sge_fn, void *metadata,
|
|
|
|
uint16_t apptag_mask, uint16_t apptag)
|
|
|
|
{
|
|
|
|
return nvme_ns_cmd_zone_appendv_with_md(ns, qpair, zslba, lba_count, cb_fn, cb_arg,
|
|
|
|
io_flags, reset_sgl_fn, next_sge_fn,
|
|
|
|
metadata, apptag_mask, apptag);
|
|
|
|
}
|
|
|
|
|
2020-10-13 12:33:38 +00:00
|
|
|
static int
|
|
|
|
nvme_zns_zone_mgmt_recv(struct spdk_nvme_ns *ns, struct spdk_nvme_qpair *qpair,
|
|
|
|
void *payload, uint32_t payload_size, uint64_t slba,
|
|
|
|
uint8_t zone_recv_action, uint8_t zra_spec_field, bool zra_spec_feats,
|
|
|
|
spdk_nvme_cmd_cb cb_fn, void *cb_arg)
|
|
|
|
{
|
|
|
|
struct nvme_request *req;
|
|
|
|
struct spdk_nvme_cmd *cmd;
|
|
|
|
|
|
|
|
req = nvme_allocate_request_user_copy(qpair, payload, payload_size, cb_fn, cb_arg, false);
|
|
|
|
if (req == NULL) {
|
|
|
|
return -ENOMEM;
|
|
|
|
}
|
|
|
|
|
|
|
|
cmd = &req->cmd;
|
|
|
|
cmd->opc = SPDK_NVME_OPC_ZONE_MGMT_RECV;
|
|
|
|
cmd->nsid = ns->id;
|
|
|
|
|
|
|
|
*(uint64_t *)&cmd->cdw10 = slba;
|
2021-01-15 07:45:39 +00:00
|
|
|
cmd->cdw12 = spdk_nvme_bytes_to_numd(payload_size);
|
2020-10-13 12:33:38 +00:00
|
|
|
cmd->cdw13 = zone_recv_action | zra_spec_field << 8 | zra_spec_feats << 16;
|
|
|
|
|
|
|
|
return nvme_qpair_submit_request(qpair, req);
|
|
|
|
}
|
|
|
|
|
|
|
|
int
|
|
|
|
spdk_nvme_zns_report_zones(struct spdk_nvme_ns *ns, struct spdk_nvme_qpair *qpair,
|
|
|
|
void *payload, uint32_t payload_size, uint64_t slba,
|
|
|
|
enum spdk_nvme_zns_zra_report_opts report_opts, bool partial_report,
|
|
|
|
spdk_nvme_cmd_cb cb_fn, void *cb_arg)
|
|
|
|
{
|
|
|
|
return nvme_zns_zone_mgmt_recv(ns, qpair, payload, payload_size, slba,
|
|
|
|
SPDK_NVME_ZONE_REPORT, report_opts, partial_report,
|
|
|
|
cb_fn, cb_arg);
|
|
|
|
}
|
|
|
|
|
2021-09-14 09:01:28 +00:00
|
|
|
int
|
|
|
|
spdk_nvme_zns_ext_report_zones(struct spdk_nvme_ns *ns, struct spdk_nvme_qpair *qpair,
|
|
|
|
void *payload, uint32_t payload_size, uint64_t slba,
|
|
|
|
enum spdk_nvme_zns_zra_report_opts report_opts, bool partial_report,
|
|
|
|
spdk_nvme_cmd_cb cb_fn, void *cb_arg)
|
|
|
|
{
|
|
|
|
return nvme_zns_zone_mgmt_recv(ns, qpair, payload, payload_size, slba,
|
|
|
|
SPDK_NVME_ZONE_EXTENDED_REPORT, report_opts, partial_report,
|
|
|
|
cb_fn, cb_arg);
|
|
|
|
}
|
|
|
|
|
2020-10-13 12:29:54 +00:00
|
|
|
static int
|
|
|
|
nvme_zns_zone_mgmt_send(struct spdk_nvme_ns *ns, struct spdk_nvme_qpair *qpair,
|
|
|
|
uint64_t slba, bool select_all, uint8_t zone_send_action,
|
|
|
|
spdk_nvme_cmd_cb cb_fn, void *cb_arg)
|
|
|
|
{
|
|
|
|
struct nvme_request *req;
|
|
|
|
struct spdk_nvme_cmd *cmd;
|
|
|
|
|
|
|
|
req = nvme_allocate_request_null(qpair, cb_fn, cb_arg);
|
|
|
|
if (req == NULL) {
|
|
|
|
return -ENOMEM;
|
|
|
|
}
|
|
|
|
|
|
|
|
cmd = &req->cmd;
|
|
|
|
cmd->opc = SPDK_NVME_OPC_ZONE_MGMT_SEND;
|
|
|
|
cmd->nsid = ns->id;
|
|
|
|
|
|
|
|
if (!select_all) {
|
|
|
|
*(uint64_t *)&cmd->cdw10 = slba;
|
|
|
|
}
|
|
|
|
|
|
|
|
cmd->cdw13 = zone_send_action | select_all << 8;
|
|
|
|
|
|
|
|
return nvme_qpair_submit_request(qpair, req);
|
|
|
|
}
|
|
|
|
|
|
|
|
int
|
|
|
|
spdk_nvme_zns_close_zone(struct spdk_nvme_ns *ns, struct spdk_nvme_qpair *qpair, uint64_t slba,
|
|
|
|
bool select_all, spdk_nvme_cmd_cb cb_fn, void *cb_arg)
|
|
|
|
{
|
|
|
|
return nvme_zns_zone_mgmt_send(ns, qpair, slba, select_all, SPDK_NVME_ZONE_CLOSE,
|
|
|
|
cb_fn, cb_arg);
|
|
|
|
}
|
|
|
|
|
|
|
|
int
|
|
|
|
spdk_nvme_zns_finish_zone(struct spdk_nvme_ns *ns, struct spdk_nvme_qpair *qpair, uint64_t slba,
|
|
|
|
bool select_all, spdk_nvme_cmd_cb cb_fn, void *cb_arg)
|
|
|
|
{
|
|
|
|
return nvme_zns_zone_mgmt_send(ns, qpair, slba, select_all, SPDK_NVME_ZONE_FINISH,
|
|
|
|
cb_fn, cb_arg);
|
|
|
|
}
|
|
|
|
|
|
|
|
int
|
|
|
|
spdk_nvme_zns_open_zone(struct spdk_nvme_ns *ns, struct spdk_nvme_qpair *qpair, uint64_t slba,
|
|
|
|
bool select_all, spdk_nvme_cmd_cb cb_fn, void *cb_arg)
|
|
|
|
{
|
|
|
|
return nvme_zns_zone_mgmt_send(ns, qpair, slba, select_all, SPDK_NVME_ZONE_OPEN,
|
|
|
|
cb_fn, cb_arg);
|
|
|
|
}
|
|
|
|
|
|
|
|
int
|
|
|
|
spdk_nvme_zns_reset_zone(struct spdk_nvme_ns *ns, struct spdk_nvme_qpair *qpair, uint64_t slba,
|
|
|
|
bool select_all, spdk_nvme_cmd_cb cb_fn, void *cb_arg)
|
|
|
|
{
|
|
|
|
return nvme_zns_zone_mgmt_send(ns, qpair, slba, select_all, SPDK_NVME_ZONE_RESET,
|
|
|
|
cb_fn, cb_arg);
|
|
|
|
}
|
|
|
|
|
|
|
|
int
|
|
|
|
spdk_nvme_zns_offline_zone(struct spdk_nvme_ns *ns, struct spdk_nvme_qpair *qpair, uint64_t slba,
|
|
|
|
bool select_all, spdk_nvme_cmd_cb cb_fn, void *cb_arg)
|
|
|
|
{
|
|
|
|
return nvme_zns_zone_mgmt_send(ns, qpair, slba, select_all, SPDK_NVME_ZONE_OFFLINE,
|
|
|
|
cb_fn, cb_arg);
|
|
|
|
}
|
2021-09-14 09:01:28 +00:00
|
|
|
|
|
|
|
int
|
|
|
|
spdk_nvme_zns_set_zone_desc_ext(struct spdk_nvme_ns *ns, struct spdk_nvme_qpair *qpair,
|
|
|
|
uint64_t slba, void *buffer, uint32_t payload_size,
|
|
|
|
spdk_nvme_cmd_cb cb_fn, void *cb_arg)
|
|
|
|
{
|
|
|
|
struct nvme_request *req;
|
|
|
|
struct spdk_nvme_cmd *cmd;
|
|
|
|
|
|
|
|
if (payload_size == 0) {
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (buffer == NULL) {
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
req = nvme_allocate_request_user_copy(qpair, buffer, payload_size, cb_fn, cb_arg, true);
|
|
|
|
if (req == NULL) {
|
|
|
|
return -ENOMEM;
|
|
|
|
}
|
|
|
|
|
|
|
|
cmd = &req->cmd;
|
|
|
|
cmd->opc = SPDK_NVME_OPC_ZONE_MGMT_SEND;
|
|
|
|
cmd->nsid = ns->id;
|
|
|
|
|
|
|
|
*(uint64_t *)&cmd->cdw10 = slba;
|
|
|
|
|
|
|
|
cmd->cdw13 = SPDK_NVME_ZONE_SET_ZDE;
|
|
|
|
|
|
|
|
return nvme_qpair_submit_request(qpair, req);
|
|
|
|
}
|