e47e16d3a4
FTL working on top of zoned bdev doesn't need physical page address (PPA) anymore. ftl_ppa was replaced with ftl_addr which represents zoned device addressing schema. Change-Id: Ied5750a7ab2f4ce42067ff3e69c1f26f85f5022a Signed-off-by: Wojciech Malikowski <wojciech.malikowski@intel.com> Reviewed-on: https://review.gerrithub.io/c/spdk/spdk/+/467633 Tested-by: SPDK CI Jenkins <sys_sgci@intel.com> Reviewed-by: Konrad Sztyber <konrad.sztyber@intel.com> Reviewed-by: Jim Harris <james.r.harris@intel.com> Reviewed-by: Shuhei Matsumoto <shuhei.matsumoto.xt@hitachi.com> Community-CI: Broadcom SPDK FC-NVMe CI <spdk-ci.pdl@broadcom.com>
172 lines
5.5 KiB
C
172 lines
5.5 KiB
C
/*-
|
|
* BSD LICENSE
|
|
*
|
|
* Copyright (c) Intel Corporation.
|
|
* All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
*
|
|
* * Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* * Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in
|
|
* the documentation and/or other materials provided with the
|
|
* distribution.
|
|
* * Neither the name of Intel Corporation nor the names of its
|
|
* contributors may be used to endorse or promote products derived
|
|
* from this software without specific prior written permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
|
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
|
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
|
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
|
* OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
|
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
|
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
|
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
|
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
|
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
*/
|
|
|
|
#ifndef FTL_RWB_H
|
|
#define FTL_RWB_H
|
|
|
|
#include "spdk/stdinc.h"
|
|
#include "spdk/queue.h"
|
|
|
|
#include "ftl_io.h"
|
|
#include "ftl_addr.h"
|
|
#include "ftl_trace.h"
|
|
|
|
struct ftl_rwb;
|
|
struct ftl_rwb_batch;
|
|
struct ftl_band;
|
|
struct spdk_ftl_conf;
|
|
|
|
enum ftl_rwb_entry_type {
|
|
FTL_RWB_TYPE_INTERNAL,
|
|
FTL_RWB_TYPE_USER,
|
|
FTL_RWB_TYPE_MAX
|
|
};
|
|
|
|
/* Write buffer entry */
|
|
struct ftl_rwb_entry {
|
|
/* Owner rwb */
|
|
struct ftl_rwb *rwb;
|
|
|
|
/* Batch containing the entry */
|
|
struct ftl_rwb_batch *batch;
|
|
|
|
/* Logical address */
|
|
uint64_t lba;
|
|
|
|
/* Physical address */
|
|
struct ftl_addr addr;
|
|
|
|
/* Band the data is moved from (only valid when relocating data) */
|
|
struct ftl_band *band;
|
|
|
|
/* Position within the rwb's buffer */
|
|
unsigned int pos;
|
|
|
|
/* Data pointer */
|
|
void *data;
|
|
|
|
/* Metadata pointer */
|
|
void *md;
|
|
|
|
/* Data/state lock */
|
|
pthread_spinlock_t lock;
|
|
|
|
/* Flags */
|
|
unsigned int flags;
|
|
|
|
/* Indicates whether the entry is part of cache and is assigned a physical address */
|
|
bool valid;
|
|
|
|
/* Trace group id */
|
|
uint64_t trace;
|
|
|
|
/* Batch list entry */
|
|
LIST_ENTRY(ftl_rwb_entry) list_entry;
|
|
};
|
|
|
|
struct ftl_rwb *ftl_rwb_init(const struct spdk_ftl_conf *conf, size_t xfer_size,
|
|
size_t md_size, size_t num_punits);
|
|
size_t ftl_rwb_get_active_batches(const struct ftl_rwb *rwb);
|
|
void ftl_rwb_free(struct ftl_rwb *rwb);
|
|
void ftl_rwb_batch_release(struct ftl_rwb_batch *batch);
|
|
void ftl_rwb_push(struct ftl_rwb_entry *entry);
|
|
size_t ftl_rwb_entry_cnt(const struct ftl_rwb *rwb);
|
|
void ftl_rwb_set_limits(struct ftl_rwb *rwb, const size_t limit[FTL_RWB_TYPE_MAX]);
|
|
void ftl_rwb_get_limits(struct ftl_rwb *rwb, size_t limit[FTL_RWB_TYPE_MAX]);
|
|
size_t ftl_rwb_num_acquired(struct ftl_rwb *rwb, enum ftl_rwb_entry_type type);
|
|
size_t ftl_rwb_num_batches(const struct ftl_rwb *rwb);
|
|
size_t ftl_rwb_size(const struct ftl_rwb *rwb);
|
|
struct ftl_rwb_entry *ftl_rwb_acquire(struct ftl_rwb *rwb, enum ftl_rwb_entry_type type);
|
|
struct ftl_rwb_batch *ftl_rwb_pop(struct ftl_rwb *rwb);
|
|
struct ftl_rwb_batch *ftl_rwb_first_batch(struct ftl_rwb *rwb);
|
|
struct ftl_rwb_batch *ftl_rwb_next_batch(struct ftl_rwb_batch *batch);
|
|
int ftl_rwb_batch_empty(struct ftl_rwb_batch *batch);
|
|
struct ftl_rwb_entry *ftl_rwb_entry_from_offset(struct ftl_rwb *rwb, size_t offset);
|
|
size_t ftl_rwb_batch_get_offset(const struct ftl_rwb_batch *batch);
|
|
void ftl_rwb_batch_revert(struct ftl_rwb_batch *batch);
|
|
struct ftl_rwb_entry *ftl_rwb_batch_first_entry(struct ftl_rwb_batch *batch);
|
|
void *ftl_rwb_batch_get_data(struct ftl_rwb_batch *batch);
|
|
void *ftl_rwb_batch_get_md(struct ftl_rwb_batch *batch);
|
|
void ftl_rwb_disable_interleaving(struct ftl_rwb *rwb);
|
|
unsigned int ftl_rwb_num_pending(struct ftl_rwb *rwb);
|
|
|
|
static inline void
|
|
_ftl_rwb_entry_set_valid(struct ftl_rwb_entry *entry, bool valid)
|
|
{
|
|
__atomic_store_n(&entry->valid, valid, __ATOMIC_SEQ_CST);
|
|
}
|
|
|
|
static inline void
|
|
ftl_rwb_entry_set_valid(struct ftl_rwb_entry *entry)
|
|
{
|
|
_ftl_rwb_entry_set_valid(entry, true);
|
|
}
|
|
|
|
static inline void
|
|
ftl_rwb_entry_invalidate(struct ftl_rwb_entry *entry)
|
|
{
|
|
_ftl_rwb_entry_set_valid(entry, false);
|
|
}
|
|
|
|
static inline int
|
|
ftl_rwb_entry_valid(struct ftl_rwb_entry *entry)
|
|
{
|
|
return __atomic_load_n(&entry->valid, __ATOMIC_SEQ_CST);
|
|
}
|
|
|
|
static inline enum ftl_rwb_entry_type
|
|
ftl_rwb_type_from_flags(int flags) {
|
|
return (flags & FTL_IO_INTERNAL) ? FTL_RWB_TYPE_INTERNAL : FTL_RWB_TYPE_USER;
|
|
}
|
|
|
|
static inline enum ftl_rwb_entry_type
|
|
ftl_rwb_entry_type(const struct ftl_rwb_entry *entry) {
|
|
return ftl_rwb_type_from_flags(entry->flags);
|
|
}
|
|
|
|
static inline int
|
|
ftl_rwb_entry_internal(const struct ftl_rwb_entry *entry)
|
|
{
|
|
return ftl_rwb_entry_type(entry) == FTL_RWB_TYPE_INTERNAL;
|
|
}
|
|
|
|
#define ftl_rwb_foreach(entry, batch) \
|
|
for (entry = ftl_rwb_batch_first_entry(batch); \
|
|
entry; entry = LIST_NEXT(entry, list_entry))
|
|
|
|
#define ftl_rwb_foreach_batch(batch, rwb) \
|
|
for (batch = ftl_rwb_first_batch(rwb); batch; \
|
|
batch = ftl_rwb_next_batch(batch))
|
|
|
|
#endif /* FTL_RWB_H */
|