env_dpdk/memory: handle spdk_mem_map_notify_walk() failures
If during mem_map creation we managed to register only a part of the total registered memory and then failed in the middle, we will now undo our registrations and gracefully return an error code leading to a spdk_mem_map_alloc() failure(). Change-Id: Id549f35109bd381318adc676392d9ee26d035359 Signed-off-by: Darek Stojaczyk <dariusz.stojaczyk@intel.com> Reviewed-on: https://review.gerrithub.io/423538 Chandler-Test-Pool: SPDK Automated Test System <sys_sgsw@intel.com> Tested-by: SPDK CI Jenkins <sys_sgci@intel.com> Reviewed-by: Ben Walker <benjamin.walker@intel.com> Reviewed-by: Jim Harris <james.r.harris@intel.com>
This commit is contained in:
parent
598d3cfbd3
commit
803449ca7b
@ -95,27 +95,18 @@ static pthread_mutex_t g_spdk_mem_map_mutex = PTHREAD_MUTEX_INITIALIZER;
|
||||
* Walk the currently registered memory via the main memory registration map
|
||||
* and call the new map's notify callback for each virtually contiguous region.
|
||||
*/
|
||||
static void
|
||||
static int
|
||||
spdk_mem_map_notify_walk(struct spdk_mem_map *map, enum spdk_mem_map_notify_action action)
|
||||
{
|
||||
size_t idx_256tb;
|
||||
uint64_t idx_1gb;
|
||||
uint64_t contig_start = 0;
|
||||
uint64_t contig_end = 0;
|
||||
|
||||
#define END_RANGE() \
|
||||
do { \
|
||||
if (contig_start != 0) { \
|
||||
/* End of of a virtually contiguous range */ \
|
||||
map->ops.notify_cb(map->cb_ctx, map, action, \
|
||||
(void *)contig_start, \
|
||||
contig_end - contig_start + 2 * 1024 * 1024); \
|
||||
} \
|
||||
contig_start = 0; \
|
||||
} while (0)
|
||||
|
||||
struct map_1gb *map_1gb;
|
||||
int rc;
|
||||
|
||||
if (!g_mem_reg_map) {
|
||||
return;
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
/* Hold the memory registration map mutex so no new registrations can be added while we are looping. */
|
||||
@ -124,11 +115,20 @@ spdk_mem_map_notify_walk(struct spdk_mem_map *map, enum spdk_mem_map_notify_acti
|
||||
for (idx_256tb = 0;
|
||||
idx_256tb < sizeof(g_mem_reg_map->map_256tb.map) / sizeof(g_mem_reg_map->map_256tb.map[0]);
|
||||
idx_256tb++) {
|
||||
const struct map_1gb *map_1gb = g_mem_reg_map->map_256tb.map[idx_256tb];
|
||||
uint64_t idx_1gb;
|
||||
map_1gb = g_mem_reg_map->map_256tb.map[idx_256tb];
|
||||
|
||||
if (!map_1gb) {
|
||||
END_RANGE();
|
||||
if (contig_start != 0) {
|
||||
/* End of of a virtually contiguous range */
|
||||
rc = map->ops.notify_cb(map->cb_ctx, map, action,
|
||||
(void *)contig_start,
|
||||
contig_end - contig_start + VALUE_2MB);
|
||||
/* Don't bother handling unregister failures. It can't be any worse */
|
||||
if (rc != 0 && action == SPDK_MEM_MAP_NOTIFY_REGISTER) {
|
||||
goto err_unregister;
|
||||
}
|
||||
}
|
||||
contig_start = 0;
|
||||
continue;
|
||||
}
|
||||
|
||||
@ -140,20 +140,84 @@ spdk_mem_map_notify_walk(struct spdk_mem_map *map, enum spdk_mem_map_notify_acti
|
||||
if (contig_start == 0) {
|
||||
contig_start = vaddr;
|
||||
}
|
||||
|
||||
contig_end = vaddr;
|
||||
} else {
|
||||
END_RANGE();
|
||||
if (contig_start != 0) {
|
||||
/* End of of a virtually contiguous range */
|
||||
rc = map->ops.notify_cb(map->cb_ctx, map, action,
|
||||
(void *)contig_start,
|
||||
contig_end - contig_start + VALUE_2MB);
|
||||
/* Don't bother handling unregister failures. It can't be any worse */
|
||||
if (rc != 0 && action == SPDK_MEM_MAP_NOTIFY_REGISTER) {
|
||||
goto err_unregister;
|
||||
}
|
||||
}
|
||||
contig_start = 0;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
pthread_mutex_unlock(&g_mem_reg_map->mutex);
|
||||
return 0;
|
||||
|
||||
err_unregister:
|
||||
/* Unwind to the first empty translation so we don't unregister
|
||||
* a region that just failed to register.
|
||||
*/
|
||||
idx_256tb = MAP_256TB_IDX((contig_start >> SHIFT_2MB) - 1);
|
||||
idx_1gb = MAP_1GB_IDX((contig_start >> SHIFT_2MB) - 1);
|
||||
contig_start = 0;
|
||||
contig_end = 0;
|
||||
|
||||
/* Unregister any memory we managed to register before the failure */
|
||||
for (; idx_256tb < SIZE_MAX; idx_256tb--) {
|
||||
map_1gb = g_mem_reg_map->map_256tb.map[idx_256tb];
|
||||
|
||||
if (!map_1gb) {
|
||||
if (contig_end != 0) {
|
||||
/* End of of a virtually contiguous range */
|
||||
map->ops.notify_cb(map->cb_ctx, map,
|
||||
SPDK_MEM_MAP_NOTIFY_UNREGISTER,
|
||||
(void *)contig_start,
|
||||
contig_end - contig_start + VALUE_2MB);
|
||||
}
|
||||
contig_end = 0;
|
||||
continue;
|
||||
}
|
||||
|
||||
for (; idx_1gb < UINT64_MAX; idx_1gb--) {
|
||||
if (map_1gb->map[idx_1gb].translation_2mb != 0) {
|
||||
/* Rebuild the virtual address from the indexes */
|
||||
uint64_t vaddr = (idx_256tb << SHIFT_1GB) | (idx_1gb << SHIFT_2MB);
|
||||
|
||||
if (contig_end == 0) {
|
||||
contig_end = vaddr;
|
||||
}
|
||||
contig_start = vaddr;
|
||||
} else {
|
||||
if (contig_end != 0) {
|
||||
/* End of of a virtually contiguous range */
|
||||
map->ops.notify_cb(map->cb_ctx, map,
|
||||
SPDK_MEM_MAP_NOTIFY_UNREGISTER,
|
||||
(void *)contig_start,
|
||||
contig_end - contig_start + VALUE_2MB);
|
||||
}
|
||||
contig_end = 0;
|
||||
}
|
||||
}
|
||||
idx_1gb = sizeof(map_1gb->map) / sizeof(map_1gb->map[0]) - 1;
|
||||
}
|
||||
|
||||
pthread_mutex_unlock(&g_mem_reg_map->mutex);
|
||||
return rc;
|
||||
}
|
||||
|
||||
struct spdk_mem_map *
|
||||
spdk_mem_map_alloc(uint64_t default_translation, const struct spdk_mem_map_ops *ops, void *cb_ctx)
|
||||
{
|
||||
struct spdk_mem_map *map;
|
||||
int rc;
|
||||
|
||||
map = calloc(1, sizeof(*map));
|
||||
if (map == NULL) {
|
||||
@ -174,7 +238,14 @@ spdk_mem_map_alloc(uint64_t default_translation, const struct spdk_mem_map_ops *
|
||||
pthread_mutex_lock(&g_spdk_mem_map_mutex);
|
||||
|
||||
if (ops && ops->notify_cb) {
|
||||
spdk_mem_map_notify_walk(map, SPDK_MEM_MAP_NOTIFY_REGISTER);
|
||||
rc = spdk_mem_map_notify_walk(map, SPDK_MEM_MAP_NOTIFY_REGISTER);
|
||||
if (rc != 0) {
|
||||
pthread_mutex_unlock(&g_spdk_mem_map_mutex);
|
||||
DEBUG_PRINT("Initial mem_map notify failed\n");
|
||||
pthread_mutex_destroy(&map->mutex);
|
||||
free(map);
|
||||
return NULL;
|
||||
}
|
||||
TAILQ_INSERT_TAIL(&g_spdk_mem_maps, map, tailq);
|
||||
}
|
||||
|
||||
|
81
test/env/vtophys/vtophys.c
vendored
81
test/env/vtophys/vtophys.c
vendored
@ -35,6 +35,11 @@
|
||||
|
||||
#include "spdk/env.h"
|
||||
|
||||
#define DEFAULT_TRANSLATION 0xDEADBEEF0BADF00DULL
|
||||
|
||||
static struct spdk_mem_map *g_mem_map;
|
||||
static void *g_last_registered_vaddr;
|
||||
|
||||
static int
|
||||
vtophys_negative_test(void)
|
||||
{
|
||||
@ -120,6 +125,8 @@ test_map_notify(void *cb_ctx, struct spdk_mem_map *map,
|
||||
switch (action) {
|
||||
case SPDK_MEM_MAP_NOTIFY_REGISTER:
|
||||
action_str = "register";
|
||||
g_last_registered_vaddr = vaddr;
|
||||
spdk_mem_map_set_translation(map, (uint64_t)vaddr, size, (uint64_t)vaddr);
|
||||
break;
|
||||
case SPDK_MEM_MAP_NOTIFY_UNREGISTER:
|
||||
action_str = "unregister";
|
||||
@ -130,27 +137,93 @@ test_map_notify(void *cb_ctx, struct spdk_mem_map *map,
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int
|
||||
test_map_notify_fail(void *cb_ctx, struct spdk_mem_map *map,
|
||||
enum spdk_mem_map_notify_action action,
|
||||
void *vaddr, size_t size)
|
||||
{
|
||||
|
||||
switch (action) {
|
||||
case SPDK_MEM_MAP_NOTIFY_REGISTER:
|
||||
if (vaddr == g_last_registered_vaddr) {
|
||||
/* Test the error handling */
|
||||
spdk_mem_map_clear_translation(g_mem_map, (uint64_t)vaddr, size);
|
||||
return -1;
|
||||
}
|
||||
break;
|
||||
case SPDK_MEM_MAP_NOTIFY_UNREGISTER:
|
||||
/* Clear the same region in the other mem_map to be able to
|
||||
* verify that there was no memory left still registered after
|
||||
* the mem_map creation failure.
|
||||
*/
|
||||
spdk_mem_map_clear_translation(g_mem_map, (uint64_t)vaddr, size);
|
||||
break;
|
||||
}
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int
|
||||
test_map_notify_verify(void *cb_ctx, struct spdk_mem_map *map,
|
||||
enum spdk_mem_map_notify_action action,
|
||||
void *vaddr, size_t size)
|
||||
{
|
||||
uint64_t reg, reg_size;
|
||||
|
||||
switch (action) {
|
||||
case SPDK_MEM_MAP_NOTIFY_REGISTER:
|
||||
reg = spdk_mem_map_translate(g_mem_map, (uint64_t)vaddr, ®_size);
|
||||
if (reg != DEFAULT_TRANSLATION) {
|
||||
return -1;
|
||||
}
|
||||
break;
|
||||
case SPDK_MEM_MAP_NOTIFY_UNREGISTER:
|
||||
break;
|
||||
}
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int
|
||||
mem_map_test(void)
|
||||
{
|
||||
struct spdk_mem_map *map;
|
||||
uint64_t default_translation = 0xDEADBEEF0BADF00D;
|
||||
const struct spdk_mem_map_ops test_map_ops = {
|
||||
struct spdk_mem_map_ops test_map_ops = {
|
||||
.notify_cb = test_map_notify,
|
||||
.are_contiguous = NULL
|
||||
};
|
||||
|
||||
g_mem_map = spdk_mem_map_alloc(DEFAULT_TRANSLATION, &test_map_ops, NULL);
|
||||
if (g_mem_map == NULL) {
|
||||
return 1;
|
||||
}
|
||||
|
||||
map = spdk_mem_map_alloc(default_translation, &test_map_ops, NULL);
|
||||
test_map_ops.notify_cb = test_map_notify_fail;
|
||||
map = spdk_mem_map_alloc(DEFAULT_TRANSLATION, &test_map_ops, NULL);
|
||||
if (map != NULL) {
|
||||
printf("Err: successfully created incomplete mem_map\n");
|
||||
spdk_mem_map_free(&map);
|
||||
spdk_mem_map_free(&g_mem_map);
|
||||
return 1;
|
||||
}
|
||||
|
||||
/* Register another map to walk through all memory regions
|
||||
* again and verify that all of them were unregistered by
|
||||
* the failed mem_map alloc above.
|
||||
*/
|
||||
test_map_ops.notify_cb = test_map_notify_verify;
|
||||
map = spdk_mem_map_alloc(DEFAULT_TRANSLATION, &test_map_ops, NULL);
|
||||
if (map == NULL) {
|
||||
printf("Err: failed mem_map creation leaked memory registrations\n");
|
||||
spdk_mem_map_free(&g_mem_map);
|
||||
return 1;
|
||||
}
|
||||
|
||||
spdk_mem_map_free(&map);
|
||||
spdk_mem_map_free(&g_mem_map);
|
||||
return 0;
|
||||
}
|
||||
|
||||
|
||||
int
|
||||
main(int argc, char **argv)
|
||||
{
|
||||
|
Loading…
x
Reference in New Issue
Block a user