cb77b060eb
Add RTE_ prefix to macro used to register mempool driver. The old one is still available but deprecated. Signed-off-by: Andrew Rybchenko <andrew.rybchenko@oktetlabs.ru> Acked-by: Olivier Matz <olivier.matz@6wind.com>
90 lines
2.3 KiB
C
90 lines
2.3 KiB
C
/* SPDX-License-Identifier: BSD-3-Clause
|
|
* Copyright(C) 2021 Marvell.
|
|
*/
|
|
|
|
#include <rte_mempool.h>
|
|
|
|
#include "roc_api.h"
|
|
#include "cnxk_mempool.h"
|
|
|
|
static int __rte_hot
|
|
cn9k_mempool_enq(struct rte_mempool *mp, void *const *obj_table, unsigned int n)
|
|
{
|
|
/* Ensure mbuf init changes are written before the free pointers
|
|
* are enqueued to the stack.
|
|
*/
|
|
rte_io_wmb();
|
|
roc_npa_aura_op_bulk_free(mp->pool_id, (const uint64_t *)obj_table, n,
|
|
0);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static inline int __rte_hot
|
|
cn9k_mempool_deq(struct rte_mempool *mp, void **obj_table, unsigned int n)
|
|
{
|
|
unsigned int count;
|
|
|
|
count = roc_npa_aura_op_bulk_alloc(mp->pool_id, (uint64_t *)obj_table,
|
|
n, 0, 1);
|
|
|
|
if (unlikely(count != n)) {
|
|
/* If bulk alloc failed to allocate all pointers, try
|
|
* allocating remaining pointers with the default alloc
|
|
* with retry scheme.
|
|
*/
|
|
if (cnxk_mempool_deq(mp, &obj_table[count], n - count)) {
|
|
cn9k_mempool_enq(mp, obj_table, count);
|
|
return -ENOENT;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
cn9k_mempool_alloc(struct rte_mempool *mp)
|
|
{
|
|
size_t block_size, padding;
|
|
|
|
block_size = mp->elt_size + mp->header_size + mp->trailer_size;
|
|
/* Align header size to ROC_ALIGN */
|
|
if (mp->header_size % ROC_ALIGN != 0) {
|
|
padding = RTE_ALIGN_CEIL(mp->header_size, ROC_ALIGN) -
|
|
mp->header_size;
|
|
mp->header_size += padding;
|
|
block_size += padding;
|
|
}
|
|
|
|
/* Align block size to ROC_ALIGN */
|
|
if (block_size % ROC_ALIGN != 0) {
|
|
padding = RTE_ALIGN_CEIL(block_size, ROC_ALIGN) - block_size;
|
|
mp->trailer_size += padding;
|
|
block_size += padding;
|
|
}
|
|
|
|
/*
|
|
* Marvell CN9k has 8 sets, 41 ways L1D cache, VA<9:7> bits dictate the
|
|
* set selection. Add additional padding to ensure that the element size
|
|
* always occupies odd number of cachelines to ensure even distribution
|
|
* of elements among L1D cache sets.
|
|
*/
|
|
padding = ((block_size / ROC_ALIGN) % 2) ? 0 : ROC_ALIGN;
|
|
mp->trailer_size += padding;
|
|
|
|
return cnxk_mempool_alloc(mp);
|
|
}
|
|
|
|
static struct rte_mempool_ops cn9k_mempool_ops = {
|
|
.name = "cn9k_mempool_ops",
|
|
.alloc = cn9k_mempool_alloc,
|
|
.free = cnxk_mempool_free,
|
|
.enqueue = cn9k_mempool_enq,
|
|
.dequeue = cn9k_mempool_deq,
|
|
.get_count = cnxk_mempool_get_count,
|
|
.calc_mem_size = cnxk_mempool_calc_mem_size,
|
|
.populate = cnxk_mempool_populate,
|
|
};
|
|
|
|
RTE_MEMPOOL_REGISTER_OPS(cn9k_mempool_ops);
|