7bc528bac9
For indexed pool with local cache, when a new trunk is allocated,
half of the trunk's index was fetched to the local cache. In case
of local cache size was less then half of the trunk size, memory
overlap happened.
This commit adds the check of the fetch size, if local cache size
is less than fetch size, adjust the fetch size to be local cache
size.
Fixes: d15c0946be
("net/mlx5: add indexed pool local cache")
Cc: stable@dpdk.org
Signed-off-by: Suanming Mou <suanmingm@nvidia.com>
Acked-by: Viacheslav Ovsiienko <viacheslavo@nvidia.com>
1189 lines
31 KiB
C
1189 lines
31 KiB
C
/* SPDX-License-Identifier: BSD-3-Clause
|
|
* Copyright 2019 Mellanox Technologies, Ltd
|
|
*/
|
|
|
|
#include <rte_malloc.h>
|
|
|
|
#include <mlx5_malloc.h>
|
|
|
|
#include "mlx5_utils.h"
|
|
|
|
/********************* Indexed pool **********************/
|
|
|
|
static inline void
|
|
mlx5_ipool_lock(struct mlx5_indexed_pool *pool)
|
|
{
|
|
if (pool->cfg.need_lock)
|
|
rte_spinlock_lock(&pool->rsz_lock);
|
|
}
|
|
|
|
static inline void
|
|
mlx5_ipool_unlock(struct mlx5_indexed_pool *pool)
|
|
{
|
|
if (pool->cfg.need_lock)
|
|
rte_spinlock_unlock(&pool->rsz_lock);
|
|
}
|
|
|
|
static inline uint32_t
|
|
mlx5_trunk_idx_get(struct mlx5_indexed_pool *pool, uint32_t entry_idx)
|
|
{
|
|
struct mlx5_indexed_pool_config *cfg = &pool->cfg;
|
|
uint32_t trunk_idx = 0;
|
|
uint32_t i;
|
|
|
|
if (!cfg->grow_trunk)
|
|
return entry_idx / cfg->trunk_size;
|
|
if (entry_idx >= pool->grow_tbl[cfg->grow_trunk - 1]) {
|
|
trunk_idx = (entry_idx - pool->grow_tbl[cfg->grow_trunk - 1]) /
|
|
(cfg->trunk_size << (cfg->grow_shift *
|
|
cfg->grow_trunk)) + cfg->grow_trunk;
|
|
} else {
|
|
for (i = 0; i < cfg->grow_trunk; i++) {
|
|
if (entry_idx < pool->grow_tbl[i])
|
|
break;
|
|
}
|
|
trunk_idx = i;
|
|
}
|
|
return trunk_idx;
|
|
}
|
|
|
|
static inline uint32_t
|
|
mlx5_trunk_size_get(struct mlx5_indexed_pool *pool, uint32_t trunk_idx)
|
|
{
|
|
struct mlx5_indexed_pool_config *cfg = &pool->cfg;
|
|
|
|
return cfg->trunk_size << (cfg->grow_shift *
|
|
(trunk_idx > cfg->grow_trunk ? cfg->grow_trunk : trunk_idx));
|
|
}
|
|
|
|
static inline uint32_t
|
|
mlx5_trunk_idx_offset_get(struct mlx5_indexed_pool *pool, uint32_t trunk_idx)
|
|
{
|
|
struct mlx5_indexed_pool_config *cfg = &pool->cfg;
|
|
uint32_t offset = 0;
|
|
|
|
if (!trunk_idx)
|
|
return 0;
|
|
if (!cfg->grow_trunk)
|
|
return cfg->trunk_size * trunk_idx;
|
|
if (trunk_idx < cfg->grow_trunk)
|
|
offset = pool->grow_tbl[trunk_idx - 1];
|
|
else
|
|
offset = pool->grow_tbl[cfg->grow_trunk - 1] +
|
|
(cfg->trunk_size << (cfg->grow_shift *
|
|
cfg->grow_trunk)) * (trunk_idx - cfg->grow_trunk);
|
|
return offset;
|
|
}
|
|
|
|
struct mlx5_indexed_pool *
|
|
mlx5_ipool_create(struct mlx5_indexed_pool_config *cfg)
|
|
{
|
|
struct mlx5_indexed_pool *pool;
|
|
uint32_t i;
|
|
|
|
if (!cfg || (!cfg->malloc ^ !cfg->free) ||
|
|
(cfg->per_core_cache && cfg->release_mem_en) ||
|
|
(cfg->trunk_size && ((cfg->trunk_size & (cfg->trunk_size - 1)) ||
|
|
((__builtin_ffs(cfg->trunk_size) + TRUNK_IDX_BITS) > 32))))
|
|
return NULL;
|
|
pool = mlx5_malloc(MLX5_MEM_ZERO, sizeof(*pool) + cfg->grow_trunk *
|
|
sizeof(pool->grow_tbl[0]), RTE_CACHE_LINE_SIZE,
|
|
SOCKET_ID_ANY);
|
|
if (!pool)
|
|
return NULL;
|
|
pool->cfg = *cfg;
|
|
if (!pool->cfg.trunk_size)
|
|
pool->cfg.trunk_size = MLX5_IPOOL_DEFAULT_TRUNK_SIZE;
|
|
if (!cfg->malloc && !cfg->free) {
|
|
pool->cfg.malloc = mlx5_malloc;
|
|
pool->cfg.free = mlx5_free;
|
|
}
|
|
if (pool->cfg.need_lock)
|
|
rte_spinlock_init(&pool->rsz_lock);
|
|
/*
|
|
* Initialize the dynamic grow trunk size lookup table to have a quick
|
|
* lookup for the trunk entry index offset.
|
|
*/
|
|
for (i = 0; i < cfg->grow_trunk; i++) {
|
|
pool->grow_tbl[i] = cfg->trunk_size << (cfg->grow_shift * i);
|
|
if (i > 0)
|
|
pool->grow_tbl[i] += pool->grow_tbl[i - 1];
|
|
}
|
|
if (!pool->cfg.max_idx)
|
|
pool->cfg.max_idx =
|
|
mlx5_trunk_idx_offset_get(pool, TRUNK_MAX_IDX + 1);
|
|
if (!cfg->per_core_cache)
|
|
pool->free_list = TRUNK_INVALID;
|
|
rte_spinlock_init(&pool->lcore_lock);
|
|
return pool;
|
|
}
|
|
|
|
static int
|
|
mlx5_ipool_grow(struct mlx5_indexed_pool *pool)
|
|
{
|
|
struct mlx5_indexed_trunk *trunk;
|
|
struct mlx5_indexed_trunk **trunk_tmp;
|
|
struct mlx5_indexed_trunk **p;
|
|
size_t trunk_size = 0;
|
|
size_t data_size;
|
|
size_t bmp_size;
|
|
uint32_t idx, cur_max_idx, i;
|
|
|
|
cur_max_idx = mlx5_trunk_idx_offset_get(pool, pool->n_trunk_valid);
|
|
if (pool->n_trunk_valid == TRUNK_MAX_IDX ||
|
|
cur_max_idx >= pool->cfg.max_idx)
|
|
return -ENOMEM;
|
|
if (pool->n_trunk_valid == pool->n_trunk) {
|
|
/* No free trunk flags, expand trunk list. */
|
|
int n_grow = pool->n_trunk_valid ? pool->n_trunk :
|
|
RTE_CACHE_LINE_SIZE / sizeof(void *);
|
|
|
|
p = pool->cfg.malloc(0, (pool->n_trunk_valid + n_grow) *
|
|
sizeof(struct mlx5_indexed_trunk *),
|
|
RTE_CACHE_LINE_SIZE, rte_socket_id());
|
|
if (!p)
|
|
return -ENOMEM;
|
|
if (pool->trunks)
|
|
memcpy(p, pool->trunks, pool->n_trunk_valid *
|
|
sizeof(struct mlx5_indexed_trunk *));
|
|
memset(RTE_PTR_ADD(p, pool->n_trunk_valid * sizeof(void *)), 0,
|
|
n_grow * sizeof(void *));
|
|
trunk_tmp = pool->trunks;
|
|
pool->trunks = p;
|
|
if (trunk_tmp)
|
|
pool->cfg.free(trunk_tmp);
|
|
pool->n_trunk += n_grow;
|
|
}
|
|
if (!pool->cfg.release_mem_en) {
|
|
idx = pool->n_trunk_valid;
|
|
} else {
|
|
/* Find the first available slot in trunk list */
|
|
for (idx = 0; idx < pool->n_trunk; idx++)
|
|
if (pool->trunks[idx] == NULL)
|
|
break;
|
|
}
|
|
trunk_size += sizeof(*trunk);
|
|
data_size = mlx5_trunk_size_get(pool, idx);
|
|
bmp_size = rte_bitmap_get_memory_footprint(data_size);
|
|
/* rte_bitmap requires memory cacheline aligned. */
|
|
trunk_size += RTE_CACHE_LINE_ROUNDUP(data_size * pool->cfg.size);
|
|
trunk_size += bmp_size;
|
|
trunk = pool->cfg.malloc(0, trunk_size,
|
|
RTE_CACHE_LINE_SIZE, rte_socket_id());
|
|
if (!trunk)
|
|
return -ENOMEM;
|
|
pool->trunks[idx] = trunk;
|
|
trunk->idx = idx;
|
|
trunk->free = data_size;
|
|
trunk->prev = TRUNK_INVALID;
|
|
trunk->next = TRUNK_INVALID;
|
|
MLX5_ASSERT(pool->free_list == TRUNK_INVALID);
|
|
pool->free_list = idx;
|
|
/* Mark all entries as available. */
|
|
trunk->bmp = rte_bitmap_init_with_all_set(data_size, &trunk->data
|
|
[RTE_CACHE_LINE_ROUNDUP(data_size * pool->cfg.size)],
|
|
bmp_size);
|
|
/* Clear the overhead bits in the trunk if it happens. */
|
|
if (cur_max_idx + data_size > pool->cfg.max_idx) {
|
|
for (i = pool->cfg.max_idx - cur_max_idx; i < data_size; i++)
|
|
rte_bitmap_clear(trunk->bmp, i);
|
|
}
|
|
MLX5_ASSERT(trunk->bmp);
|
|
pool->n_trunk_valid++;
|
|
#ifdef POOL_DEBUG
|
|
pool->trunk_new++;
|
|
pool->trunk_avail++;
|
|
#endif
|
|
return 0;
|
|
}
|
|
|
|
static inline struct mlx5_indexed_cache *
|
|
mlx5_ipool_update_global_cache(struct mlx5_indexed_pool *pool, int cidx)
|
|
{
|
|
struct mlx5_indexed_cache *gc, *lc, *olc = NULL;
|
|
|
|
lc = pool->cache[cidx]->lc;
|
|
gc = __atomic_load_n(&pool->gc, __ATOMIC_RELAXED);
|
|
if (gc && lc != gc) {
|
|
mlx5_ipool_lock(pool);
|
|
if (lc && !(--lc->ref_cnt))
|
|
olc = lc;
|
|
lc = pool->gc;
|
|
lc->ref_cnt++;
|
|
pool->cache[cidx]->lc = lc;
|
|
mlx5_ipool_unlock(pool);
|
|
if (olc)
|
|
pool->cfg.free(olc);
|
|
}
|
|
return lc;
|
|
}
|
|
|
|
static uint32_t
|
|
mlx5_ipool_allocate_from_global(struct mlx5_indexed_pool *pool, int cidx)
|
|
{
|
|
struct mlx5_indexed_trunk *trunk;
|
|
struct mlx5_indexed_cache *p, *lc, *olc = NULL;
|
|
size_t trunk_size = 0;
|
|
size_t data_size;
|
|
uint32_t cur_max_idx, trunk_idx, trunk_n;
|
|
uint32_t fetch_size, ts_idx, i;
|
|
int n_grow;
|
|
|
|
check_again:
|
|
p = NULL;
|
|
fetch_size = 0;
|
|
/*
|
|
* Fetch new index from global if possible. First round local
|
|
* cache will be NULL.
|
|
*/
|
|
lc = pool->cache[cidx]->lc;
|
|
mlx5_ipool_lock(pool);
|
|
/* Try to update local cache first. */
|
|
if (likely(pool->gc)) {
|
|
if (lc != pool->gc) {
|
|
if (lc && !(--lc->ref_cnt))
|
|
olc = lc;
|
|
lc = pool->gc;
|
|
lc->ref_cnt++;
|
|
pool->cache[cidx]->lc = lc;
|
|
}
|
|
if (lc->len) {
|
|
/* Use the updated local cache to fetch index. */
|
|
fetch_size = pool->cfg.per_core_cache >> 2;
|
|
if (lc->len < fetch_size)
|
|
fetch_size = lc->len;
|
|
lc->len -= fetch_size;
|
|
memcpy(pool->cache[cidx]->idx, &lc->idx[lc->len],
|
|
sizeof(uint32_t) * fetch_size);
|
|
}
|
|
}
|
|
mlx5_ipool_unlock(pool);
|
|
if (unlikely(olc)) {
|
|
pool->cfg.free(olc);
|
|
olc = NULL;
|
|
}
|
|
if (fetch_size) {
|
|
pool->cache[cidx]->len = fetch_size - 1;
|
|
return pool->cache[cidx]->idx[pool->cache[cidx]->len];
|
|
}
|
|
trunk_idx = lc ? __atomic_load_n(&lc->n_trunk_valid,
|
|
__ATOMIC_ACQUIRE) : 0;
|
|
trunk_n = lc ? lc->n_trunk : 0;
|
|
cur_max_idx = mlx5_trunk_idx_offset_get(pool, trunk_idx);
|
|
/* Check if index reach maximum. */
|
|
if (trunk_idx == TRUNK_MAX_IDX ||
|
|
cur_max_idx >= pool->cfg.max_idx)
|
|
return 0;
|
|
/* No enough space in trunk array, resize the trunks array. */
|
|
if (trunk_idx == trunk_n) {
|
|
n_grow = trunk_idx ? trunk_idx :
|
|
RTE_CACHE_LINE_SIZE / sizeof(void *);
|
|
cur_max_idx = mlx5_trunk_idx_offset_get(pool, trunk_n + n_grow);
|
|
/* Resize the trunk array. */
|
|
p = pool->cfg.malloc(0, ((trunk_idx + n_grow) *
|
|
sizeof(struct mlx5_indexed_trunk *)) +
|
|
(cur_max_idx * sizeof(uint32_t)) + sizeof(*p),
|
|
RTE_CACHE_LINE_SIZE, rte_socket_id());
|
|
if (!p)
|
|
return 0;
|
|
p->trunks = (struct mlx5_indexed_trunk **)&p->idx[cur_max_idx];
|
|
if (lc)
|
|
memcpy(p->trunks, lc->trunks, trunk_idx *
|
|
sizeof(struct mlx5_indexed_trunk *));
|
|
#ifdef RTE_LIBRTE_MLX5_DEBUG
|
|
memset(RTE_PTR_ADD(p->trunks, trunk_idx * sizeof(void *)), 0,
|
|
n_grow * sizeof(void *));
|
|
#endif
|
|
p->n_trunk_valid = trunk_idx;
|
|
p->n_trunk = trunk_n + n_grow;
|
|
p->len = 0;
|
|
}
|
|
/* Prepare the new trunk. */
|
|
trunk_size = sizeof(*trunk);
|
|
data_size = mlx5_trunk_size_get(pool, trunk_idx);
|
|
trunk_size += RTE_CACHE_LINE_ROUNDUP(data_size * pool->cfg.size);
|
|
trunk = pool->cfg.malloc(0, trunk_size,
|
|
RTE_CACHE_LINE_SIZE, rte_socket_id());
|
|
if (unlikely(!trunk)) {
|
|
pool->cfg.free(p);
|
|
return 0;
|
|
}
|
|
trunk->idx = trunk_idx;
|
|
trunk->free = data_size;
|
|
mlx5_ipool_lock(pool);
|
|
/*
|
|
* Double check if trunks has been updated or have available index.
|
|
* During the new trunk allocate, index may still be flushed to the
|
|
* global cache. So also need to check the pool->gc->len.
|
|
*/
|
|
if (pool->gc && (lc != pool->gc ||
|
|
lc->n_trunk_valid != trunk_idx ||
|
|
pool->gc->len)) {
|
|
mlx5_ipool_unlock(pool);
|
|
if (p)
|
|
pool->cfg.free(p);
|
|
pool->cfg.free(trunk);
|
|
goto check_again;
|
|
}
|
|
/* Resize the trunk array and update local cache first. */
|
|
if (p) {
|
|
if (lc && !(--lc->ref_cnt))
|
|
olc = lc;
|
|
lc = p;
|
|
lc->ref_cnt = 1;
|
|
pool->cache[cidx]->lc = lc;
|
|
__atomic_store_n(&pool->gc, p, __ATOMIC_RELAXED);
|
|
}
|
|
/* Add trunk to trunks array. */
|
|
lc->trunks[trunk_idx] = trunk;
|
|
__atomic_fetch_add(&lc->n_trunk_valid, 1, __ATOMIC_RELAXED);
|
|
/* Enqueue half of the index to global. */
|
|
ts_idx = mlx5_trunk_idx_offset_get(pool, trunk_idx) + 1;
|
|
fetch_size = trunk->free >> 1;
|
|
if (fetch_size > pool->cfg.per_core_cache)
|
|
fetch_size = trunk->free - pool->cfg.per_core_cache;
|
|
for (i = 0; i < fetch_size; i++)
|
|
lc->idx[i] = ts_idx + i;
|
|
lc->len = fetch_size;
|
|
mlx5_ipool_unlock(pool);
|
|
/* Copy left half - 1 to local cache index array. */
|
|
pool->cache[cidx]->len = trunk->free - fetch_size - 1;
|
|
ts_idx += fetch_size;
|
|
for (i = 0; i < pool->cache[cidx]->len; i++)
|
|
pool->cache[cidx]->idx[i] = ts_idx + i;
|
|
if (olc)
|
|
pool->cfg.free(olc);
|
|
return ts_idx + i;
|
|
}
|
|
|
|
static void *
|
|
_mlx5_ipool_get_cache(struct mlx5_indexed_pool *pool, int cidx, uint32_t idx)
|
|
{
|
|
struct mlx5_indexed_trunk *trunk;
|
|
struct mlx5_indexed_cache *lc;
|
|
uint32_t trunk_idx;
|
|
uint32_t entry_idx;
|
|
|
|
MLX5_ASSERT(idx);
|
|
if (unlikely(!pool->cache[cidx])) {
|
|
pool->cache[cidx] = pool->cfg.malloc(MLX5_MEM_ZERO,
|
|
sizeof(struct mlx5_ipool_per_lcore) +
|
|
(pool->cfg.per_core_cache * sizeof(uint32_t)),
|
|
RTE_CACHE_LINE_SIZE, SOCKET_ID_ANY);
|
|
if (!pool->cache[cidx]) {
|
|
DRV_LOG(ERR, "Ipool cache%d allocate failed\n", cidx);
|
|
return NULL;
|
|
}
|
|
}
|
|
lc = mlx5_ipool_update_global_cache(pool, cidx);
|
|
idx -= 1;
|
|
trunk_idx = mlx5_trunk_idx_get(pool, idx);
|
|
trunk = lc->trunks[trunk_idx];
|
|
MLX5_ASSERT(trunk);
|
|
entry_idx = idx - mlx5_trunk_idx_offset_get(pool, trunk_idx);
|
|
return &trunk->data[entry_idx * pool->cfg.size];
|
|
}
|
|
|
|
static void *
|
|
mlx5_ipool_get_cache(struct mlx5_indexed_pool *pool, uint32_t idx)
|
|
{
|
|
void *entry;
|
|
int cidx;
|
|
|
|
cidx = rte_lcore_index(rte_lcore_id());
|
|
if (unlikely(cidx == -1)) {
|
|
cidx = RTE_MAX_LCORE;
|
|
rte_spinlock_lock(&pool->lcore_lock);
|
|
}
|
|
entry = _mlx5_ipool_get_cache(pool, cidx, idx);
|
|
if (unlikely(cidx == RTE_MAX_LCORE))
|
|
rte_spinlock_unlock(&pool->lcore_lock);
|
|
return entry;
|
|
}
|
|
|
|
|
|
static void *
|
|
_mlx5_ipool_malloc_cache(struct mlx5_indexed_pool *pool, int cidx,
|
|
uint32_t *idx)
|
|
{
|
|
if (unlikely(!pool->cache[cidx])) {
|
|
pool->cache[cidx] = pool->cfg.malloc(MLX5_MEM_ZERO,
|
|
sizeof(struct mlx5_ipool_per_lcore) +
|
|
(pool->cfg.per_core_cache * sizeof(uint32_t)),
|
|
RTE_CACHE_LINE_SIZE, SOCKET_ID_ANY);
|
|
if (!pool->cache[cidx]) {
|
|
DRV_LOG(ERR, "Ipool cache%d allocate failed\n", cidx);
|
|
return NULL;
|
|
}
|
|
} else if (pool->cache[cidx]->len) {
|
|
pool->cache[cidx]->len--;
|
|
*idx = pool->cache[cidx]->idx[pool->cache[cidx]->len];
|
|
return _mlx5_ipool_get_cache(pool, cidx, *idx);
|
|
}
|
|
/* Not enough idx in global cache. Keep fetching from global. */
|
|
*idx = mlx5_ipool_allocate_from_global(pool, cidx);
|
|
if (unlikely(!(*idx)))
|
|
return NULL;
|
|
return _mlx5_ipool_get_cache(pool, cidx, *idx);
|
|
}
|
|
|
|
static void *
|
|
mlx5_ipool_malloc_cache(struct mlx5_indexed_pool *pool, uint32_t *idx)
|
|
{
|
|
void *entry;
|
|
int cidx;
|
|
|
|
cidx = rte_lcore_index(rte_lcore_id());
|
|
if (unlikely(cidx == -1)) {
|
|
cidx = RTE_MAX_LCORE;
|
|
rte_spinlock_lock(&pool->lcore_lock);
|
|
}
|
|
entry = _mlx5_ipool_malloc_cache(pool, cidx, idx);
|
|
if (unlikely(cidx == RTE_MAX_LCORE))
|
|
rte_spinlock_unlock(&pool->lcore_lock);
|
|
return entry;
|
|
}
|
|
|
|
static void
|
|
_mlx5_ipool_free_cache(struct mlx5_indexed_pool *pool, int cidx, uint32_t idx)
|
|
{
|
|
struct mlx5_ipool_per_lcore *ilc;
|
|
struct mlx5_indexed_cache *gc, *olc = NULL;
|
|
uint32_t reclaim_num = 0;
|
|
|
|
MLX5_ASSERT(idx);
|
|
/*
|
|
* When index was allocated on core A but freed on core B. In this
|
|
* case check if local cache on core B was allocated before.
|
|
*/
|
|
if (unlikely(!pool->cache[cidx])) {
|
|
pool->cache[cidx] = pool->cfg.malloc(MLX5_MEM_ZERO,
|
|
sizeof(struct mlx5_ipool_per_lcore) +
|
|
(pool->cfg.per_core_cache * sizeof(uint32_t)),
|
|
RTE_CACHE_LINE_SIZE, SOCKET_ID_ANY);
|
|
if (!pool->cache[cidx]) {
|
|
DRV_LOG(ERR, "Ipool cache%d allocate failed\n", cidx);
|
|
return;
|
|
}
|
|
}
|
|
/* Try to enqueue to local index cache. */
|
|
if (pool->cache[cidx]->len < pool->cfg.per_core_cache) {
|
|
pool->cache[cidx]->idx[pool->cache[cidx]->len] = idx;
|
|
pool->cache[cidx]->len++;
|
|
return;
|
|
}
|
|
ilc = pool->cache[cidx];
|
|
reclaim_num = pool->cfg.per_core_cache >> 2;
|
|
ilc->len -= reclaim_num;
|
|
/* Local index cache full, try with global index cache. */
|
|
mlx5_ipool_lock(pool);
|
|
gc = pool->gc;
|
|
if (ilc->lc != gc) {
|
|
if (!(--ilc->lc->ref_cnt))
|
|
olc = ilc->lc;
|
|
gc->ref_cnt++;
|
|
ilc->lc = gc;
|
|
}
|
|
memcpy(&gc->idx[gc->len], &ilc->idx[ilc->len],
|
|
reclaim_num * sizeof(uint32_t));
|
|
gc->len += reclaim_num;
|
|
mlx5_ipool_unlock(pool);
|
|
if (olc)
|
|
pool->cfg.free(olc);
|
|
pool->cache[cidx]->idx[pool->cache[cidx]->len] = idx;
|
|
pool->cache[cidx]->len++;
|
|
}
|
|
|
|
static void
|
|
mlx5_ipool_free_cache(struct mlx5_indexed_pool *pool, uint32_t idx)
|
|
{
|
|
int cidx;
|
|
|
|
cidx = rte_lcore_index(rte_lcore_id());
|
|
if (unlikely(cidx == -1)) {
|
|
cidx = RTE_MAX_LCORE;
|
|
rte_spinlock_lock(&pool->lcore_lock);
|
|
}
|
|
_mlx5_ipool_free_cache(pool, cidx, idx);
|
|
if (unlikely(cidx == RTE_MAX_LCORE))
|
|
rte_spinlock_unlock(&pool->lcore_lock);
|
|
}
|
|
|
|
void *
|
|
mlx5_ipool_malloc(struct mlx5_indexed_pool *pool, uint32_t *idx)
|
|
{
|
|
struct mlx5_indexed_trunk *trunk;
|
|
uint64_t slab = 0;
|
|
uint32_t iidx = 0;
|
|
void *p;
|
|
|
|
if (pool->cfg.per_core_cache)
|
|
return mlx5_ipool_malloc_cache(pool, idx);
|
|
mlx5_ipool_lock(pool);
|
|
if (pool->free_list == TRUNK_INVALID) {
|
|
/* If no available trunks, grow new. */
|
|
if (mlx5_ipool_grow(pool)) {
|
|
mlx5_ipool_unlock(pool);
|
|
return NULL;
|
|
}
|
|
}
|
|
MLX5_ASSERT(pool->free_list != TRUNK_INVALID);
|
|
trunk = pool->trunks[pool->free_list];
|
|
MLX5_ASSERT(trunk->free);
|
|
if (!rte_bitmap_scan(trunk->bmp, &iidx, &slab)) {
|
|
mlx5_ipool_unlock(pool);
|
|
return NULL;
|
|
}
|
|
MLX5_ASSERT(slab);
|
|
iidx += __builtin_ctzll(slab);
|
|
MLX5_ASSERT(iidx != UINT32_MAX);
|
|
MLX5_ASSERT(iidx < mlx5_trunk_size_get(pool, trunk->idx));
|
|
rte_bitmap_clear(trunk->bmp, iidx);
|
|
p = &trunk->data[iidx * pool->cfg.size];
|
|
/*
|
|
* The ipool index should grow continually from small to big,
|
|
* some features as metering only accept limited bits of index.
|
|
* Random index with MSB set may be rejected.
|
|
*/
|
|
iidx += mlx5_trunk_idx_offset_get(pool, trunk->idx);
|
|
iidx += 1; /* non-zero index. */
|
|
trunk->free--;
|
|
#ifdef POOL_DEBUG
|
|
pool->n_entry++;
|
|
#endif
|
|
if (!trunk->free) {
|
|
/* Full trunk will be removed from free list in imalloc. */
|
|
MLX5_ASSERT(pool->free_list == trunk->idx);
|
|
pool->free_list = trunk->next;
|
|
if (trunk->next != TRUNK_INVALID)
|
|
pool->trunks[trunk->next]->prev = TRUNK_INVALID;
|
|
trunk->prev = TRUNK_INVALID;
|
|
trunk->next = TRUNK_INVALID;
|
|
#ifdef POOL_DEBUG
|
|
pool->trunk_empty++;
|
|
pool->trunk_avail--;
|
|
#endif
|
|
}
|
|
*idx = iidx;
|
|
mlx5_ipool_unlock(pool);
|
|
return p;
|
|
}
|
|
|
|
void *
|
|
mlx5_ipool_zmalloc(struct mlx5_indexed_pool *pool, uint32_t *idx)
|
|
{
|
|
void *entry = mlx5_ipool_malloc(pool, idx);
|
|
|
|
if (entry && pool->cfg.size)
|
|
memset(entry, 0, pool->cfg.size);
|
|
return entry;
|
|
}
|
|
|
|
void
|
|
mlx5_ipool_free(struct mlx5_indexed_pool *pool, uint32_t idx)
|
|
{
|
|
struct mlx5_indexed_trunk *trunk;
|
|
uint32_t trunk_idx;
|
|
uint32_t entry_idx;
|
|
|
|
if (!idx)
|
|
return;
|
|
if (pool->cfg.per_core_cache) {
|
|
mlx5_ipool_free_cache(pool, idx);
|
|
return;
|
|
}
|
|
idx -= 1;
|
|
mlx5_ipool_lock(pool);
|
|
trunk_idx = mlx5_trunk_idx_get(pool, idx);
|
|
if ((!pool->cfg.release_mem_en && trunk_idx >= pool->n_trunk_valid) ||
|
|
(pool->cfg.release_mem_en && trunk_idx >= pool->n_trunk))
|
|
goto out;
|
|
trunk = pool->trunks[trunk_idx];
|
|
if (!trunk)
|
|
goto out;
|
|
entry_idx = idx - mlx5_trunk_idx_offset_get(pool, trunk->idx);
|
|
if (trunk_idx != trunk->idx ||
|
|
rte_bitmap_get(trunk->bmp, entry_idx))
|
|
goto out;
|
|
rte_bitmap_set(trunk->bmp, entry_idx);
|
|
trunk->free++;
|
|
if (pool->cfg.release_mem_en && trunk->free == mlx5_trunk_size_get
|
|
(pool, trunk->idx)) {
|
|
if (pool->free_list == trunk->idx)
|
|
pool->free_list = trunk->next;
|
|
if (trunk->next != TRUNK_INVALID)
|
|
pool->trunks[trunk->next]->prev = trunk->prev;
|
|
if (trunk->prev != TRUNK_INVALID)
|
|
pool->trunks[trunk->prev]->next = trunk->next;
|
|
pool->cfg.free(trunk);
|
|
pool->trunks[trunk_idx] = NULL;
|
|
pool->n_trunk_valid--;
|
|
#ifdef POOL_DEBUG
|
|
pool->trunk_avail--;
|
|
pool->trunk_free++;
|
|
#endif
|
|
if (pool->n_trunk_valid == 0) {
|
|
pool->cfg.free(pool->trunks);
|
|
pool->trunks = NULL;
|
|
pool->n_trunk = 0;
|
|
}
|
|
} else if (trunk->free == 1) {
|
|
/* Put into free trunk list head. */
|
|
MLX5_ASSERT(pool->free_list != trunk->idx);
|
|
trunk->next = pool->free_list;
|
|
trunk->prev = TRUNK_INVALID;
|
|
if (pool->free_list != TRUNK_INVALID)
|
|
pool->trunks[pool->free_list]->prev = trunk->idx;
|
|
pool->free_list = trunk->idx;
|
|
#ifdef POOL_DEBUG
|
|
pool->trunk_empty--;
|
|
pool->trunk_avail++;
|
|
#endif
|
|
}
|
|
#ifdef POOL_DEBUG
|
|
pool->n_entry--;
|
|
#endif
|
|
out:
|
|
mlx5_ipool_unlock(pool);
|
|
}
|
|
|
|
void *
|
|
mlx5_ipool_get(struct mlx5_indexed_pool *pool, uint32_t idx)
|
|
{
|
|
struct mlx5_indexed_trunk *trunk;
|
|
void *p = NULL;
|
|
uint32_t trunk_idx;
|
|
uint32_t entry_idx;
|
|
|
|
if (!idx)
|
|
return NULL;
|
|
if (pool->cfg.per_core_cache)
|
|
return mlx5_ipool_get_cache(pool, idx);
|
|
idx -= 1;
|
|
mlx5_ipool_lock(pool);
|
|
trunk_idx = mlx5_trunk_idx_get(pool, idx);
|
|
if ((!pool->cfg.release_mem_en && trunk_idx >= pool->n_trunk_valid) ||
|
|
(pool->cfg.release_mem_en && trunk_idx >= pool->n_trunk))
|
|
goto out;
|
|
trunk = pool->trunks[trunk_idx];
|
|
if (!trunk)
|
|
goto out;
|
|
entry_idx = idx - mlx5_trunk_idx_offset_get(pool, trunk->idx);
|
|
if (trunk_idx != trunk->idx ||
|
|
rte_bitmap_get(trunk->bmp, entry_idx))
|
|
goto out;
|
|
p = &trunk->data[entry_idx * pool->cfg.size];
|
|
out:
|
|
mlx5_ipool_unlock(pool);
|
|
return p;
|
|
}
|
|
|
|
int
|
|
mlx5_ipool_destroy(struct mlx5_indexed_pool *pool)
|
|
{
|
|
struct mlx5_indexed_trunk **trunks = NULL;
|
|
struct mlx5_indexed_cache *gc = pool->gc;
|
|
uint32_t i, n_trunk_valid = 0;
|
|
|
|
MLX5_ASSERT(pool);
|
|
mlx5_ipool_lock(pool);
|
|
if (pool->cfg.per_core_cache) {
|
|
for (i = 0; i <= RTE_MAX_LCORE; i++) {
|
|
/*
|
|
* Free only old global cache. Pool gc will be
|
|
* freed at last.
|
|
*/
|
|
if (pool->cache[i]) {
|
|
if (pool->cache[i]->lc &&
|
|
pool->cache[i]->lc != pool->gc &&
|
|
(!(--pool->cache[i]->lc->ref_cnt)))
|
|
pool->cfg.free(pool->cache[i]->lc);
|
|
pool->cfg.free(pool->cache[i]);
|
|
}
|
|
}
|
|
if (gc) {
|
|
trunks = gc->trunks;
|
|
n_trunk_valid = gc->n_trunk_valid;
|
|
}
|
|
} else {
|
|
gc = NULL;
|
|
trunks = pool->trunks;
|
|
n_trunk_valid = pool->n_trunk_valid;
|
|
}
|
|
for (i = 0; i < n_trunk_valid; i++) {
|
|
if (trunks[i])
|
|
pool->cfg.free(trunks[i]);
|
|
}
|
|
if (!gc && trunks)
|
|
pool->cfg.free(trunks);
|
|
if (gc)
|
|
pool->cfg.free(gc);
|
|
mlx5_ipool_unlock(pool);
|
|
mlx5_free(pool);
|
|
return 0;
|
|
}
|
|
|
|
void
|
|
mlx5_ipool_flush_cache(struct mlx5_indexed_pool *pool)
|
|
{
|
|
uint32_t i, j;
|
|
struct mlx5_indexed_cache *gc;
|
|
struct rte_bitmap *ibmp;
|
|
uint32_t bmp_num, mem_size;
|
|
|
|
if (!pool->cfg.per_core_cache)
|
|
return;
|
|
gc = pool->gc;
|
|
if (!gc)
|
|
return;
|
|
/* Reset bmp. */
|
|
bmp_num = mlx5_trunk_idx_offset_get(pool, gc->n_trunk_valid);
|
|
mem_size = rte_bitmap_get_memory_footprint(bmp_num);
|
|
pool->bmp_mem = pool->cfg.malloc(MLX5_MEM_ZERO, mem_size,
|
|
RTE_CACHE_LINE_SIZE, rte_socket_id());
|
|
if (!pool->bmp_mem) {
|
|
DRV_LOG(ERR, "Ipool bitmap mem allocate failed.\n");
|
|
return;
|
|
}
|
|
ibmp = rte_bitmap_init_with_all_set(bmp_num, pool->bmp_mem, mem_size);
|
|
if (!ibmp) {
|
|
pool->cfg.free(pool->bmp_mem);
|
|
pool->bmp_mem = NULL;
|
|
DRV_LOG(ERR, "Ipool bitmap create failed.\n");
|
|
return;
|
|
}
|
|
pool->ibmp = ibmp;
|
|
/* Clear global cache. */
|
|
for (i = 0; i < gc->len; i++)
|
|
rte_bitmap_clear(ibmp, gc->idx[i] - 1);
|
|
/* Clear core cache. */
|
|
for (i = 0; i < RTE_MAX_LCORE + 1; i++) {
|
|
struct mlx5_ipool_per_lcore *ilc = pool->cache[i];
|
|
|
|
if (!ilc)
|
|
continue;
|
|
for (j = 0; j < ilc->len; j++)
|
|
rte_bitmap_clear(ibmp, ilc->idx[j] - 1);
|
|
}
|
|
}
|
|
|
|
static void *
|
|
mlx5_ipool_get_next_cache(struct mlx5_indexed_pool *pool, uint32_t *pos)
|
|
{
|
|
struct rte_bitmap *ibmp;
|
|
uint64_t slab = 0;
|
|
uint32_t iidx = *pos;
|
|
|
|
ibmp = pool->ibmp;
|
|
if (!ibmp || !rte_bitmap_scan(ibmp, &iidx, &slab)) {
|
|
if (pool->bmp_mem) {
|
|
pool->cfg.free(pool->bmp_mem);
|
|
pool->bmp_mem = NULL;
|
|
pool->ibmp = NULL;
|
|
}
|
|
return NULL;
|
|
}
|
|
iidx += __builtin_ctzll(slab);
|
|
rte_bitmap_clear(ibmp, iidx);
|
|
iidx++;
|
|
*pos = iidx;
|
|
return mlx5_ipool_get_cache(pool, iidx);
|
|
}
|
|
|
|
void *
|
|
mlx5_ipool_get_next(struct mlx5_indexed_pool *pool, uint32_t *pos)
|
|
{
|
|
uint32_t idx = *pos;
|
|
void *entry;
|
|
|
|
if (pool->cfg.per_core_cache)
|
|
return mlx5_ipool_get_next_cache(pool, pos);
|
|
while (idx <= mlx5_trunk_idx_offset_get(pool, pool->n_trunk)) {
|
|
entry = mlx5_ipool_get(pool, idx);
|
|
if (entry) {
|
|
*pos = idx;
|
|
return entry;
|
|
}
|
|
idx++;
|
|
}
|
|
return NULL;
|
|
}
|
|
|
|
void
|
|
mlx5_ipool_dump(struct mlx5_indexed_pool *pool)
|
|
{
|
|
printf("Pool %s entry size %u, trunks %u, %d entry per trunk, "
|
|
"total: %d\n",
|
|
pool->cfg.type, pool->cfg.size, pool->n_trunk_valid,
|
|
pool->cfg.trunk_size, pool->n_trunk_valid);
|
|
#ifdef POOL_DEBUG
|
|
printf("Pool %s entry %u, trunk alloc %u, empty: %u, "
|
|
"available %u free %u\n",
|
|
pool->cfg.type, pool->n_entry, pool->trunk_new,
|
|
pool->trunk_empty, pool->trunk_avail, pool->trunk_free);
|
|
#endif
|
|
}
|
|
|
|
struct mlx5_l3t_tbl *
|
|
mlx5_l3t_create(enum mlx5_l3t_type type)
|
|
{
|
|
struct mlx5_l3t_tbl *tbl;
|
|
struct mlx5_indexed_pool_config l3t_ip_cfg = {
|
|
.trunk_size = 16,
|
|
.grow_trunk = 6,
|
|
.grow_shift = 1,
|
|
.need_lock = 0,
|
|
.release_mem_en = 1,
|
|
.malloc = mlx5_malloc,
|
|
.free = mlx5_free,
|
|
};
|
|
|
|
if (type >= MLX5_L3T_TYPE_MAX) {
|
|
rte_errno = EINVAL;
|
|
return NULL;
|
|
}
|
|
tbl = mlx5_malloc(MLX5_MEM_ZERO, sizeof(struct mlx5_l3t_tbl), 1,
|
|
SOCKET_ID_ANY);
|
|
if (!tbl) {
|
|
rte_errno = ENOMEM;
|
|
return NULL;
|
|
}
|
|
tbl->type = type;
|
|
switch (type) {
|
|
case MLX5_L3T_TYPE_WORD:
|
|
l3t_ip_cfg.size = sizeof(struct mlx5_l3t_entry_word);
|
|
l3t_ip_cfg.type = "mlx5_l3t_e_tbl_w";
|
|
break;
|
|
case MLX5_L3T_TYPE_DWORD:
|
|
l3t_ip_cfg.size = sizeof(struct mlx5_l3t_entry_dword);
|
|
l3t_ip_cfg.type = "mlx5_l3t_e_tbl_dw";
|
|
break;
|
|
case MLX5_L3T_TYPE_QWORD:
|
|
l3t_ip_cfg.size = sizeof(struct mlx5_l3t_entry_qword);
|
|
l3t_ip_cfg.type = "mlx5_l3t_e_tbl_qw";
|
|
break;
|
|
default:
|
|
l3t_ip_cfg.size = sizeof(struct mlx5_l3t_entry_ptr);
|
|
l3t_ip_cfg.type = "mlx5_l3t_e_tbl_tpr";
|
|
break;
|
|
}
|
|
rte_spinlock_init(&tbl->sl);
|
|
tbl->eip = mlx5_ipool_create(&l3t_ip_cfg);
|
|
if (!tbl->eip) {
|
|
rte_errno = ENOMEM;
|
|
mlx5_free(tbl);
|
|
tbl = NULL;
|
|
}
|
|
return tbl;
|
|
}
|
|
|
|
void
|
|
mlx5_l3t_destroy(struct mlx5_l3t_tbl *tbl)
|
|
{
|
|
struct mlx5_l3t_level_tbl *g_tbl, *m_tbl;
|
|
uint32_t i, j;
|
|
|
|
if (!tbl)
|
|
return;
|
|
g_tbl = tbl->tbl;
|
|
if (g_tbl) {
|
|
for (i = 0; i < MLX5_L3T_GT_SIZE; i++) {
|
|
m_tbl = g_tbl->tbl[i];
|
|
if (!m_tbl)
|
|
continue;
|
|
for (j = 0; j < MLX5_L3T_MT_SIZE; j++) {
|
|
if (!m_tbl->tbl[j])
|
|
continue;
|
|
MLX5_ASSERT(!((struct mlx5_l3t_entry_word *)
|
|
m_tbl->tbl[j])->ref_cnt);
|
|
mlx5_ipool_free(tbl->eip,
|
|
((struct mlx5_l3t_entry_word *)
|
|
m_tbl->tbl[j])->idx);
|
|
m_tbl->tbl[j] = 0;
|
|
if (!(--m_tbl->ref_cnt))
|
|
break;
|
|
}
|
|
MLX5_ASSERT(!m_tbl->ref_cnt);
|
|
mlx5_free(g_tbl->tbl[i]);
|
|
g_tbl->tbl[i] = 0;
|
|
if (!(--g_tbl->ref_cnt))
|
|
break;
|
|
}
|
|
MLX5_ASSERT(!g_tbl->ref_cnt);
|
|
mlx5_free(tbl->tbl);
|
|
tbl->tbl = 0;
|
|
}
|
|
mlx5_ipool_destroy(tbl->eip);
|
|
mlx5_free(tbl);
|
|
}
|
|
|
|
static int32_t
|
|
__l3t_get_entry(struct mlx5_l3t_tbl *tbl, uint32_t idx,
|
|
union mlx5_l3t_data *data)
|
|
{
|
|
struct mlx5_l3t_level_tbl *g_tbl, *m_tbl;
|
|
struct mlx5_l3t_entry_word *w_e_tbl;
|
|
struct mlx5_l3t_entry_dword *dw_e_tbl;
|
|
struct mlx5_l3t_entry_qword *qw_e_tbl;
|
|
struct mlx5_l3t_entry_ptr *ptr_e_tbl;
|
|
void *e_tbl;
|
|
uint32_t entry_idx;
|
|
|
|
g_tbl = tbl->tbl;
|
|
if (!g_tbl)
|
|
return -1;
|
|
m_tbl = g_tbl->tbl[(idx >> MLX5_L3T_GT_OFFSET) & MLX5_L3T_GT_MASK];
|
|
if (!m_tbl)
|
|
return -1;
|
|
e_tbl = m_tbl->tbl[(idx >> MLX5_L3T_MT_OFFSET) & MLX5_L3T_MT_MASK];
|
|
if (!e_tbl)
|
|
return -1;
|
|
entry_idx = idx & MLX5_L3T_ET_MASK;
|
|
switch (tbl->type) {
|
|
case MLX5_L3T_TYPE_WORD:
|
|
w_e_tbl = (struct mlx5_l3t_entry_word *)e_tbl;
|
|
data->word = w_e_tbl->entry[entry_idx].data;
|
|
if (w_e_tbl->entry[entry_idx].data)
|
|
w_e_tbl->entry[entry_idx].ref_cnt++;
|
|
break;
|
|
case MLX5_L3T_TYPE_DWORD:
|
|
dw_e_tbl = (struct mlx5_l3t_entry_dword *)e_tbl;
|
|
data->dword = dw_e_tbl->entry[entry_idx].data;
|
|
if (dw_e_tbl->entry[entry_idx].data)
|
|
dw_e_tbl->entry[entry_idx].ref_cnt++;
|
|
break;
|
|
case MLX5_L3T_TYPE_QWORD:
|
|
qw_e_tbl = (struct mlx5_l3t_entry_qword *)e_tbl;
|
|
data->qword = qw_e_tbl->entry[entry_idx].data;
|
|
if (qw_e_tbl->entry[entry_idx].data)
|
|
qw_e_tbl->entry[entry_idx].ref_cnt++;
|
|
break;
|
|
default:
|
|
ptr_e_tbl = (struct mlx5_l3t_entry_ptr *)e_tbl;
|
|
data->ptr = ptr_e_tbl->entry[entry_idx].data;
|
|
if (ptr_e_tbl->entry[entry_idx].data)
|
|
ptr_e_tbl->entry[entry_idx].ref_cnt++;
|
|
break;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
int32_t
|
|
mlx5_l3t_get_entry(struct mlx5_l3t_tbl *tbl, uint32_t idx,
|
|
union mlx5_l3t_data *data)
|
|
{
|
|
int ret;
|
|
|
|
rte_spinlock_lock(&tbl->sl);
|
|
ret = __l3t_get_entry(tbl, idx, data);
|
|
rte_spinlock_unlock(&tbl->sl);
|
|
return ret;
|
|
}
|
|
|
|
int32_t
|
|
mlx5_l3t_clear_entry(struct mlx5_l3t_tbl *tbl, uint32_t idx)
|
|
{
|
|
struct mlx5_l3t_level_tbl *g_tbl, *m_tbl;
|
|
struct mlx5_l3t_entry_word *w_e_tbl;
|
|
struct mlx5_l3t_entry_dword *dw_e_tbl;
|
|
struct mlx5_l3t_entry_qword *qw_e_tbl;
|
|
struct mlx5_l3t_entry_ptr *ptr_e_tbl;
|
|
void *e_tbl;
|
|
uint32_t entry_idx;
|
|
uint64_t ref_cnt;
|
|
int32_t ret = -1;
|
|
|
|
rte_spinlock_lock(&tbl->sl);
|
|
g_tbl = tbl->tbl;
|
|
if (!g_tbl)
|
|
goto out;
|
|
m_tbl = g_tbl->tbl[(idx >> MLX5_L3T_GT_OFFSET) & MLX5_L3T_GT_MASK];
|
|
if (!m_tbl)
|
|
goto out;
|
|
e_tbl = m_tbl->tbl[(idx >> MLX5_L3T_MT_OFFSET) & MLX5_L3T_MT_MASK];
|
|
if (!e_tbl)
|
|
goto out;
|
|
entry_idx = idx & MLX5_L3T_ET_MASK;
|
|
switch (tbl->type) {
|
|
case MLX5_L3T_TYPE_WORD:
|
|
w_e_tbl = (struct mlx5_l3t_entry_word *)e_tbl;
|
|
MLX5_ASSERT(w_e_tbl->entry[entry_idx].ref_cnt);
|
|
ret = --w_e_tbl->entry[entry_idx].ref_cnt;
|
|
if (ret)
|
|
goto out;
|
|
w_e_tbl->entry[entry_idx].data = 0;
|
|
ref_cnt = --w_e_tbl->ref_cnt;
|
|
break;
|
|
case MLX5_L3T_TYPE_DWORD:
|
|
dw_e_tbl = (struct mlx5_l3t_entry_dword *)e_tbl;
|
|
MLX5_ASSERT(dw_e_tbl->entry[entry_idx].ref_cnt);
|
|
ret = --dw_e_tbl->entry[entry_idx].ref_cnt;
|
|
if (ret)
|
|
goto out;
|
|
dw_e_tbl->entry[entry_idx].data = 0;
|
|
ref_cnt = --dw_e_tbl->ref_cnt;
|
|
break;
|
|
case MLX5_L3T_TYPE_QWORD:
|
|
qw_e_tbl = (struct mlx5_l3t_entry_qword *)e_tbl;
|
|
MLX5_ASSERT(qw_e_tbl->entry[entry_idx].ref_cnt);
|
|
ret = --qw_e_tbl->entry[entry_idx].ref_cnt;
|
|
if (ret)
|
|
goto out;
|
|
qw_e_tbl->entry[entry_idx].data = 0;
|
|
ref_cnt = --qw_e_tbl->ref_cnt;
|
|
break;
|
|
default:
|
|
ptr_e_tbl = (struct mlx5_l3t_entry_ptr *)e_tbl;
|
|
MLX5_ASSERT(ptr_e_tbl->entry[entry_idx].ref_cnt);
|
|
ret = --ptr_e_tbl->entry[entry_idx].ref_cnt;
|
|
if (ret)
|
|
goto out;
|
|
ptr_e_tbl->entry[entry_idx].data = NULL;
|
|
ref_cnt = --ptr_e_tbl->ref_cnt;
|
|
break;
|
|
}
|
|
if (!ref_cnt) {
|
|
mlx5_ipool_free(tbl->eip,
|
|
((struct mlx5_l3t_entry_word *)e_tbl)->idx);
|
|
m_tbl->tbl[(idx >> MLX5_L3T_MT_OFFSET) & MLX5_L3T_MT_MASK] =
|
|
NULL;
|
|
if (!(--m_tbl->ref_cnt)) {
|
|
mlx5_free(m_tbl);
|
|
g_tbl->tbl
|
|
[(idx >> MLX5_L3T_GT_OFFSET) & MLX5_L3T_GT_MASK] = NULL;
|
|
if (!(--g_tbl->ref_cnt)) {
|
|
mlx5_free(g_tbl);
|
|
tbl->tbl = 0;
|
|
}
|
|
}
|
|
}
|
|
out:
|
|
rte_spinlock_unlock(&tbl->sl);
|
|
return ret;
|
|
}
|
|
|
|
static int32_t
|
|
__l3t_set_entry(struct mlx5_l3t_tbl *tbl, uint32_t idx,
|
|
union mlx5_l3t_data *data)
|
|
{
|
|
struct mlx5_l3t_level_tbl *g_tbl, *m_tbl;
|
|
struct mlx5_l3t_entry_word *w_e_tbl;
|
|
struct mlx5_l3t_entry_dword *dw_e_tbl;
|
|
struct mlx5_l3t_entry_qword *qw_e_tbl;
|
|
struct mlx5_l3t_entry_ptr *ptr_e_tbl;
|
|
void *e_tbl;
|
|
uint32_t entry_idx, tbl_idx = 0;
|
|
|
|
/* Check the global table, create it if empty. */
|
|
g_tbl = tbl->tbl;
|
|
if (!g_tbl) {
|
|
g_tbl = mlx5_malloc(MLX5_MEM_ZERO,
|
|
sizeof(struct mlx5_l3t_level_tbl) +
|
|
sizeof(void *) * MLX5_L3T_GT_SIZE, 1,
|
|
SOCKET_ID_ANY);
|
|
if (!g_tbl) {
|
|
rte_errno = ENOMEM;
|
|
return -1;
|
|
}
|
|
tbl->tbl = g_tbl;
|
|
}
|
|
/*
|
|
* Check the middle table, create it if empty. Ref_cnt will be
|
|
* increased if new sub table created.
|
|
*/
|
|
m_tbl = g_tbl->tbl[(idx >> MLX5_L3T_GT_OFFSET) & MLX5_L3T_GT_MASK];
|
|
if (!m_tbl) {
|
|
m_tbl = mlx5_malloc(MLX5_MEM_ZERO,
|
|
sizeof(struct mlx5_l3t_level_tbl) +
|
|
sizeof(void *) * MLX5_L3T_MT_SIZE, 1,
|
|
SOCKET_ID_ANY);
|
|
if (!m_tbl) {
|
|
rte_errno = ENOMEM;
|
|
return -1;
|
|
}
|
|
g_tbl->tbl[(idx >> MLX5_L3T_GT_OFFSET) & MLX5_L3T_GT_MASK] =
|
|
m_tbl;
|
|
g_tbl->ref_cnt++;
|
|
}
|
|
/*
|
|
* Check the entry table, create it if empty. Ref_cnt will be
|
|
* increased if new sub entry table created.
|
|
*/
|
|
e_tbl = m_tbl->tbl[(idx >> MLX5_L3T_MT_OFFSET) & MLX5_L3T_MT_MASK];
|
|
if (!e_tbl) {
|
|
e_tbl = mlx5_ipool_zmalloc(tbl->eip, &tbl_idx);
|
|
if (!e_tbl) {
|
|
rte_errno = ENOMEM;
|
|
return -1;
|
|
}
|
|
((struct mlx5_l3t_entry_word *)e_tbl)->idx = tbl_idx;
|
|
m_tbl->tbl[(idx >> MLX5_L3T_MT_OFFSET) & MLX5_L3T_MT_MASK] =
|
|
e_tbl;
|
|
m_tbl->ref_cnt++;
|
|
}
|
|
entry_idx = idx & MLX5_L3T_ET_MASK;
|
|
switch (tbl->type) {
|
|
case MLX5_L3T_TYPE_WORD:
|
|
w_e_tbl = (struct mlx5_l3t_entry_word *)e_tbl;
|
|
if (w_e_tbl->entry[entry_idx].data) {
|
|
data->word = w_e_tbl->entry[entry_idx].data;
|
|
w_e_tbl->entry[entry_idx].ref_cnt++;
|
|
rte_errno = EEXIST;
|
|
return -1;
|
|
}
|
|
w_e_tbl->entry[entry_idx].data = data->word;
|
|
w_e_tbl->entry[entry_idx].ref_cnt = 1;
|
|
w_e_tbl->ref_cnt++;
|
|
break;
|
|
case MLX5_L3T_TYPE_DWORD:
|
|
dw_e_tbl = (struct mlx5_l3t_entry_dword *)e_tbl;
|
|
if (dw_e_tbl->entry[entry_idx].data) {
|
|
data->dword = dw_e_tbl->entry[entry_idx].data;
|
|
dw_e_tbl->entry[entry_idx].ref_cnt++;
|
|
rte_errno = EEXIST;
|
|
return -1;
|
|
}
|
|
dw_e_tbl->entry[entry_idx].data = data->dword;
|
|
dw_e_tbl->entry[entry_idx].ref_cnt = 1;
|
|
dw_e_tbl->ref_cnt++;
|
|
break;
|
|
case MLX5_L3T_TYPE_QWORD:
|
|
qw_e_tbl = (struct mlx5_l3t_entry_qword *)e_tbl;
|
|
if (qw_e_tbl->entry[entry_idx].data) {
|
|
data->qword = qw_e_tbl->entry[entry_idx].data;
|
|
qw_e_tbl->entry[entry_idx].ref_cnt++;
|
|
rte_errno = EEXIST;
|
|
return -1;
|
|
}
|
|
qw_e_tbl->entry[entry_idx].data = data->qword;
|
|
qw_e_tbl->entry[entry_idx].ref_cnt = 1;
|
|
qw_e_tbl->ref_cnt++;
|
|
break;
|
|
default:
|
|
ptr_e_tbl = (struct mlx5_l3t_entry_ptr *)e_tbl;
|
|
if (ptr_e_tbl->entry[entry_idx].data) {
|
|
data->ptr = ptr_e_tbl->entry[entry_idx].data;
|
|
ptr_e_tbl->entry[entry_idx].ref_cnt++;
|
|
rte_errno = EEXIST;
|
|
return -1;
|
|
}
|
|
ptr_e_tbl->entry[entry_idx].data = data->ptr;
|
|
ptr_e_tbl->entry[entry_idx].ref_cnt = 1;
|
|
ptr_e_tbl->ref_cnt++;
|
|
break;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
int32_t
|
|
mlx5_l3t_set_entry(struct mlx5_l3t_tbl *tbl, uint32_t idx,
|
|
union mlx5_l3t_data *data)
|
|
{
|
|
int ret;
|
|
|
|
rte_spinlock_lock(&tbl->sl);
|
|
ret = __l3t_set_entry(tbl, idx, data);
|
|
rte_spinlock_unlock(&tbl->sl);
|
|
return ret;
|
|
}
|