931e3d51d9
As specified in the doc of spdk_for_each_channel:
```
Once 'fn' has been called on each channel, 'cpl' will be called
on the thread that spdk_for_each_channel was initially called from.
```
Fixes: ff87d29cc3
("io_channel: Add mechanism to call a function on
each channel")
Change-Id: Ic60b061ec402672f510d99697943e96ff9a73417
Signed-off-by: Dariusz Stojaczyk <dariuszx.stojaczyk@intel.com>
Reviewed-on: https://review.gerrithub.io/369719
Reviewed-by: Ben Walker <benjamin.walker@intel.com>
Reviewed-by: Daniel Verkamp <daniel.verkamp@intel.com>
Tested-by: SPDK Automated Test System <sys_sgsw@intel.com>
406 lines
9.3 KiB
C
406 lines
9.3 KiB
C
/*-
|
|
* BSD LICENSE
|
|
*
|
|
* Copyright (c) Intel Corporation.
|
|
* All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
*
|
|
* * Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* * Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in
|
|
* the documentation and/or other materials provided with the
|
|
* distribution.
|
|
* * Neither the name of Intel Corporation nor the names of its
|
|
* contributors may be used to endorse or promote products derived
|
|
* from this software without specific prior written permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
|
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
|
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
|
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
|
* OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
|
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
|
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
|
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
|
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
|
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
*/
|
|
|
|
#include "spdk/stdinc.h"
|
|
|
|
#include "spdk/io_channel.h"
|
|
#include "spdk/log.h"
|
|
|
|
static pthread_mutex_t g_devlist_mutex = PTHREAD_MUTEX_INITIALIZER;
|
|
|
|
struct io_device {
|
|
void *io_device_ctx;
|
|
spdk_io_channel_create_cb create_cb;
|
|
spdk_io_channel_destroy_cb destroy_cb;
|
|
uint32_t ctx_size;
|
|
TAILQ_ENTRY(io_device) tailq;
|
|
};
|
|
|
|
static TAILQ_HEAD(, io_device) g_io_devices = TAILQ_HEAD_INITIALIZER(g_io_devices);
|
|
|
|
struct spdk_io_channel {
|
|
struct spdk_thread *thread;
|
|
void *io_device;
|
|
uint32_t ref;
|
|
TAILQ_ENTRY(spdk_io_channel) tailq;
|
|
spdk_io_channel_destroy_cb destroy_cb;
|
|
|
|
/*
|
|
* Modules will allocate extra memory off the end of this structure
|
|
* to store references to hardware-specific references (i.e. NVMe queue
|
|
* pairs, or references to child device spdk_io_channels (i.e.
|
|
* virtual bdevs).
|
|
*/
|
|
};
|
|
|
|
struct spdk_thread {
|
|
pthread_t thread_id;
|
|
spdk_thread_pass_msg fn;
|
|
void *thread_ctx;
|
|
TAILQ_HEAD(, spdk_io_channel) io_channels;
|
|
TAILQ_ENTRY(spdk_thread) tailq;
|
|
};
|
|
|
|
static TAILQ_HEAD(, spdk_thread) g_threads = TAILQ_HEAD_INITIALIZER(g_threads);
|
|
|
|
static struct spdk_thread *
|
|
_get_thread(void)
|
|
{
|
|
pthread_t thread_id;
|
|
struct spdk_thread *thread;
|
|
|
|
thread_id = pthread_self();
|
|
|
|
thread = NULL;
|
|
TAILQ_FOREACH(thread, &g_threads, tailq) {
|
|
if (thread->thread_id == thread_id) {
|
|
return thread;
|
|
}
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
struct spdk_thread *
|
|
spdk_allocate_thread(spdk_thread_pass_msg fn, void *thread_ctx)
|
|
{
|
|
struct spdk_thread *thread;
|
|
|
|
pthread_mutex_lock(&g_devlist_mutex);
|
|
|
|
thread = _get_thread();
|
|
if (thread) {
|
|
SPDK_ERRLOG("Double allocated SPDK thread\n");
|
|
pthread_mutex_unlock(&g_devlist_mutex);
|
|
return NULL;
|
|
}
|
|
|
|
thread = calloc(1, sizeof(*thread));
|
|
if (!thread) {
|
|
SPDK_ERRLOG("Unable to allocate memory for thread\n");
|
|
pthread_mutex_unlock(&g_devlist_mutex);
|
|
return NULL;
|
|
}
|
|
|
|
thread->thread_id = pthread_self();
|
|
thread->fn = fn;
|
|
thread->thread_ctx = thread_ctx;
|
|
TAILQ_INIT(&thread->io_channels);
|
|
TAILQ_INSERT_TAIL(&g_threads, thread, tailq);
|
|
|
|
pthread_mutex_unlock(&g_devlist_mutex);
|
|
|
|
return thread;
|
|
}
|
|
|
|
void
|
|
spdk_free_thread(void)
|
|
{
|
|
struct spdk_thread *thread;
|
|
|
|
pthread_mutex_lock(&g_devlist_mutex);
|
|
|
|
thread = _get_thread();
|
|
if (!thread) {
|
|
SPDK_ERRLOG("No thread allocated\n");
|
|
pthread_mutex_unlock(&g_devlist_mutex);
|
|
return;
|
|
}
|
|
|
|
TAILQ_REMOVE(&g_threads, thread, tailq);
|
|
free(thread);
|
|
|
|
pthread_mutex_unlock(&g_devlist_mutex);
|
|
}
|
|
|
|
struct spdk_thread *
|
|
spdk_get_thread(void)
|
|
{
|
|
struct spdk_thread *thread;
|
|
|
|
pthread_mutex_lock(&g_devlist_mutex);
|
|
|
|
thread = _get_thread();
|
|
if (!thread) {
|
|
SPDK_ERRLOG("No thread allocated\n");
|
|
}
|
|
|
|
pthread_mutex_unlock(&g_devlist_mutex);
|
|
|
|
return thread;
|
|
}
|
|
|
|
void
|
|
spdk_thread_send_msg(const struct spdk_thread *thread, spdk_thread_fn fn, void *ctx)
|
|
{
|
|
thread->fn(fn, ctx, thread->thread_ctx);
|
|
}
|
|
|
|
void
|
|
spdk_io_device_register(void *io_device, spdk_io_channel_create_cb create_cb,
|
|
spdk_io_channel_destroy_cb destroy_cb, uint32_t ctx_size)
|
|
{
|
|
struct io_device *dev, *tmp;
|
|
|
|
dev = calloc(1, sizeof(struct io_device));
|
|
if (dev == NULL) {
|
|
SPDK_ERRLOG("could not allocate io_device\n");
|
|
return;
|
|
}
|
|
|
|
dev->io_device_ctx = io_device;
|
|
dev->create_cb = create_cb;
|
|
dev->destroy_cb = destroy_cb;
|
|
dev->ctx_size = ctx_size;
|
|
|
|
pthread_mutex_lock(&g_devlist_mutex);
|
|
TAILQ_FOREACH(tmp, &g_io_devices, tailq) {
|
|
if (tmp->io_device_ctx == io_device) {
|
|
SPDK_ERRLOG("io_device %p already registered\n", io_device);
|
|
free(dev);
|
|
pthread_mutex_unlock(&g_devlist_mutex);
|
|
return;
|
|
}
|
|
}
|
|
TAILQ_INSERT_TAIL(&g_io_devices, dev, tailq);
|
|
pthread_mutex_unlock(&g_devlist_mutex);
|
|
}
|
|
|
|
void
|
|
spdk_io_device_unregister(void *io_device)
|
|
{
|
|
struct io_device *dev;
|
|
|
|
pthread_mutex_lock(&g_devlist_mutex);
|
|
TAILQ_FOREACH(dev, &g_io_devices, tailq) {
|
|
if (dev->io_device_ctx == io_device) {
|
|
TAILQ_REMOVE(&g_io_devices, dev, tailq);
|
|
free(dev);
|
|
pthread_mutex_unlock(&g_devlist_mutex);
|
|
return;
|
|
}
|
|
}
|
|
SPDK_ERRLOG("io_device %p not found\n", io_device);
|
|
pthread_mutex_unlock(&g_devlist_mutex);
|
|
}
|
|
|
|
struct spdk_io_channel *
|
|
spdk_get_io_channel(void *io_device)
|
|
{
|
|
struct spdk_io_channel *ch;
|
|
struct spdk_thread *thread;
|
|
struct io_device *dev;
|
|
int rc;
|
|
|
|
pthread_mutex_lock(&g_devlist_mutex);
|
|
TAILQ_FOREACH(dev, &g_io_devices, tailq) {
|
|
if (dev->io_device_ctx == io_device) {
|
|
break;
|
|
}
|
|
}
|
|
if (dev == NULL) {
|
|
SPDK_ERRLOG("could not find io_device %p\n", io_device);
|
|
pthread_mutex_unlock(&g_devlist_mutex);
|
|
return NULL;
|
|
}
|
|
|
|
thread = _get_thread();
|
|
if (!thread) {
|
|
SPDK_ERRLOG("No thread allocated\n");
|
|
pthread_mutex_unlock(&g_devlist_mutex);
|
|
return NULL;
|
|
}
|
|
|
|
pthread_mutex_unlock(&g_devlist_mutex);
|
|
|
|
TAILQ_FOREACH(ch, &thread->io_channels, tailq) {
|
|
if (ch->io_device == io_device) {
|
|
ch->ref++;
|
|
/*
|
|
* An I/O channel already exists for this device on this
|
|
* thread, so return it.
|
|
*/
|
|
return ch;
|
|
}
|
|
}
|
|
|
|
ch = calloc(1, sizeof(*ch) + dev->ctx_size);
|
|
if (ch == NULL) {
|
|
SPDK_ERRLOG("could not calloc spdk_io_channel\n");
|
|
return NULL;
|
|
}
|
|
rc = dev->create_cb(io_device, (uint8_t *)ch + sizeof(*ch));
|
|
if (rc == -1) {
|
|
free(ch);
|
|
return NULL;
|
|
}
|
|
|
|
ch->io_device = io_device;
|
|
ch->destroy_cb = dev->destroy_cb;
|
|
ch->thread = thread;
|
|
ch->ref = 1;
|
|
TAILQ_INSERT_TAIL(&thread->io_channels, ch, tailq);
|
|
return ch;
|
|
}
|
|
|
|
static void
|
|
_spdk_put_io_channel(void *arg)
|
|
{
|
|
struct spdk_io_channel *ch = arg;
|
|
|
|
if (ch->ref == 0) {
|
|
SPDK_ERRLOG("ref already zero\n");
|
|
return;
|
|
}
|
|
|
|
ch->ref--;
|
|
|
|
if (ch->ref == 0) {
|
|
TAILQ_REMOVE(&ch->thread->io_channels, ch, tailq);
|
|
ch->destroy_cb(ch->io_device, (uint8_t *)ch + sizeof(*ch));
|
|
free(ch);
|
|
}
|
|
}
|
|
|
|
void
|
|
spdk_put_io_channel(struct spdk_io_channel *ch)
|
|
{
|
|
spdk_thread_send_msg(ch->thread, _spdk_put_io_channel, ch);
|
|
}
|
|
|
|
void *
|
|
spdk_io_channel_get_ctx(struct spdk_io_channel *ch)
|
|
{
|
|
return (uint8_t *)ch + sizeof(*ch);
|
|
}
|
|
|
|
struct spdk_thread *
|
|
spdk_io_channel_get_thread(struct spdk_io_channel *ch)
|
|
{
|
|
return ch->thread;
|
|
}
|
|
|
|
struct call_channel {
|
|
void *io_device;
|
|
spdk_channel_msg fn;
|
|
void *ctx;
|
|
|
|
struct spdk_thread *cur_thread;
|
|
struct spdk_io_channel *cur_ch;
|
|
|
|
struct spdk_thread *orig_thread;
|
|
spdk_channel_for_each_cpl cpl;
|
|
};
|
|
|
|
static void
|
|
_call_completion(void *ctx)
|
|
{
|
|
struct call_channel *ch_ctx = ctx;
|
|
|
|
ch_ctx->cpl(ch_ctx->io_device, ch_ctx->ctx);
|
|
free(ch_ctx);
|
|
}
|
|
|
|
static void
|
|
_call_channel(void *ctx)
|
|
{
|
|
struct call_channel *ch_ctx = ctx;
|
|
struct spdk_thread *thread;
|
|
struct spdk_io_channel *ch;
|
|
|
|
thread = ch_ctx->cur_thread;
|
|
ch = ch_ctx->cur_ch;
|
|
|
|
ch_ctx->fn(ch_ctx->io_device, ch, ch_ctx->ctx);
|
|
|
|
pthread_mutex_lock(&g_devlist_mutex);
|
|
thread = TAILQ_NEXT(thread, tailq);
|
|
while (thread) {
|
|
TAILQ_FOREACH(ch, &thread->io_channels, tailq) {
|
|
if (ch->io_device == ch_ctx->io_device) {
|
|
ch_ctx->cur_thread = thread;
|
|
ch_ctx->cur_ch = ch;
|
|
pthread_mutex_unlock(&g_devlist_mutex);
|
|
spdk_thread_send_msg(thread, _call_channel, ch_ctx);
|
|
return;
|
|
}
|
|
}
|
|
thread = TAILQ_NEXT(thread, tailq);
|
|
}
|
|
|
|
pthread_mutex_unlock(&g_devlist_mutex);
|
|
|
|
spdk_thread_send_msg(ch_ctx->orig_thread, _call_completion, ch_ctx);
|
|
}
|
|
|
|
void
|
|
spdk_for_each_channel(void *io_device, spdk_channel_msg fn, void *ctx,
|
|
spdk_channel_for_each_cpl cpl)
|
|
{
|
|
struct spdk_thread *thread;
|
|
struct spdk_io_channel *ch;
|
|
struct call_channel *ch_ctx;
|
|
|
|
ch_ctx = calloc(1, sizeof(*ch_ctx));
|
|
if (!ch_ctx) {
|
|
SPDK_ERRLOG("Unable to allocate context\n");
|
|
return;
|
|
}
|
|
|
|
ch_ctx->io_device = io_device;
|
|
ch_ctx->fn = fn;
|
|
ch_ctx->ctx = ctx;
|
|
ch_ctx->cpl = cpl;
|
|
|
|
pthread_mutex_lock(&g_devlist_mutex);
|
|
ch_ctx->orig_thread = _get_thread();
|
|
|
|
TAILQ_FOREACH(thread, &g_threads, tailq) {
|
|
TAILQ_FOREACH(ch, &thread->io_channels, tailq) {
|
|
if (ch->io_device == io_device) {
|
|
ch_ctx->cur_thread = thread;
|
|
ch_ctx->cur_ch = ch;
|
|
pthread_mutex_unlock(&g_devlist_mutex);
|
|
spdk_thread_send_msg(thread, _call_channel, ch_ctx);
|
|
return;
|
|
}
|
|
}
|
|
}
|
|
|
|
free(ch_ctx);
|
|
|
|
pthread_mutex_unlock(&g_devlist_mutex);
|
|
|
|
cpl(io_device, ctx);
|
|
}
|