2018-01-29 13:11:22 +00:00
|
|
|
/* SPDX-License-Identifier: BSD-3-Clause
|
|
|
|
* Copyright(c) 2010-2014 Intel Corporation.
|
|
|
|
* Copyright(c) 2016 6WIND S.A.
|
2012-09-04 12:54:00 +00:00
|
|
|
*/
|
|
|
|
|
mempool: support new allocation methods
If a user has specified that the zone should have contiguous memory,
add a memzone flag to request contiguous memory. Otherwise, account
for the fact that unless we're in IOVA_AS_VA mode, we cannot
guarantee that the pages would be physically contiguous, so we
calculate the memzone size and alignments as if we were getting
the smallest page size available.
However, for the non-IOVA contiguous case, existing mempool size
calculation function doesn't give us expected results, because it
will return memzone sizes aligned to page size (e.g. a 1MB mempool
may use an entire 1GB page), therefore in cases where we weren't
specifically asked to reserve non-contiguous memory, first try
reserving a memzone as IOVA-contiguous, and if that fails, then
try reserving with page-aligned size/alignment.
Signed-off-by: Anatoly Burakov <anatoly.burakov@intel.com>
Acked-by: Olivier Matz <olivier.matz@6wind.com>
Tested-by: Santosh Shukla <santosh.shukla@caviumnetworks.com>
Tested-by: Hemant Agrawal <hemant.agrawal@nxp.com>
Tested-by: Gowrishankar Muthukrishnan <gowrishankar.m@linux.vnet.ibm.com>
2018-04-11 12:29:59 +00:00
|
|
|
#include <stdbool.h>
|
2012-09-04 12:54:00 +00:00
|
|
|
#include <stdio.h>
|
|
|
|
#include <string.h>
|
|
|
|
#include <stdint.h>
|
|
|
|
#include <stdarg.h>
|
2013-03-01 15:10:57 +00:00
|
|
|
#include <unistd.h>
|
2012-09-04 12:54:00 +00:00
|
|
|
#include <inttypes.h>
|
|
|
|
#include <errno.h>
|
|
|
|
#include <sys/queue.h>
|
2016-05-18 11:04:50 +00:00
|
|
|
#include <sys/mman.h>
|
2012-09-04 12:54:00 +00:00
|
|
|
|
|
|
|
#include <rte_common.h>
|
|
|
|
#include <rte_log.h>
|
|
|
|
#include <rte_debug.h>
|
|
|
|
#include <rte_memory.h>
|
|
|
|
#include <rte_memzone.h>
|
2014-06-20 15:42:22 +00:00
|
|
|
#include <rte_malloc.h>
|
2012-09-04 12:54:00 +00:00
|
|
|
#include <rte_atomic.h>
|
|
|
|
#include <rte_launch.h>
|
|
|
|
#include <rte_eal.h>
|
2012-12-19 23:00:00 +00:00
|
|
|
#include <rte_eal_memconfig.h>
|
2012-09-04 12:54:00 +00:00
|
|
|
#include <rte_per_lcore.h>
|
|
|
|
#include <rte_lcore.h>
|
|
|
|
#include <rte_branch_prediction.h>
|
|
|
|
#include <rte_errno.h>
|
|
|
|
#include <rte_string_fns.h>
|
2012-12-19 23:00:00 +00:00
|
|
|
#include <rte_spinlock.h>
|
2019-07-05 13:10:30 +00:00
|
|
|
#include <rte_tailq.h>
|
2012-09-04 12:54:00 +00:00
|
|
|
|
|
|
|
#include "rte_mempool.h"
|
|
|
|
|
2014-06-20 15:42:22 +00:00
|
|
|
TAILQ_HEAD(rte_mempool_list, rte_tailq_entry);
|
2012-09-04 12:54:00 +00:00
|
|
|
|
2015-03-04 21:50:08 +00:00
|
|
|
static struct rte_tailq_elem rte_mempool_tailq = {
|
|
|
|
.name = "RTE_MEMPOOL",
|
|
|
|
};
|
|
|
|
EAL_REGISTER_TAILQ(rte_mempool_tailq)
|
|
|
|
|
2012-12-19 23:00:00 +00:00
|
|
|
#define CACHE_FLUSHTHRESH_MULTIPLIER 1.5
|
2015-05-18 15:35:14 +00:00
|
|
|
#define CALC_CACHE_FLUSHTHRESH(c) \
|
|
|
|
((typeof(c))((c) * CACHE_FLUSHTHRESH_MULTIPLIER))
|
2012-09-04 12:54:00 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* return the greatest common divisor between a and b (fast algorithm)
|
|
|
|
*
|
|
|
|
*/
|
|
|
|
static unsigned get_gcd(unsigned a, unsigned b)
|
|
|
|
{
|
|
|
|
unsigned c;
|
|
|
|
|
|
|
|
if (0 == a)
|
|
|
|
return b;
|
|
|
|
if (0 == b)
|
|
|
|
return a;
|
|
|
|
|
|
|
|
if (a < b) {
|
|
|
|
c = a;
|
|
|
|
a = b;
|
|
|
|
b = c;
|
|
|
|
}
|
|
|
|
|
|
|
|
while (b != 0) {
|
|
|
|
c = a % b;
|
|
|
|
a = b;
|
|
|
|
b = c;
|
|
|
|
}
|
|
|
|
|
|
|
|
return a;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
2014-05-02 23:42:52 +00:00
|
|
|
* Depending on memory configuration, objects addresses are spread
|
2012-09-04 12:54:00 +00:00
|
|
|
* between channels and ranks in RAM: the pool allocator will add
|
|
|
|
* padding between objects. This function return the new size of the
|
|
|
|
* object.
|
|
|
|
*/
|
|
|
|
static unsigned optimize_object_size(unsigned obj_size)
|
|
|
|
{
|
|
|
|
unsigned nrank, nchan;
|
|
|
|
unsigned new_obj_size;
|
|
|
|
|
|
|
|
/* get number of channels */
|
|
|
|
nchan = rte_memory_get_nchannel();
|
|
|
|
if (nchan == 0)
|
2015-10-15 11:49:04 +00:00
|
|
|
nchan = 4;
|
2012-09-04 12:54:00 +00:00
|
|
|
|
|
|
|
nrank = rte_memory_get_nrank();
|
|
|
|
if (nrank == 0)
|
|
|
|
nrank = 1;
|
|
|
|
|
|
|
|
/* process new object size */
|
2015-07-09 08:25:16 +00:00
|
|
|
new_obj_size = (obj_size + RTE_MEMPOOL_ALIGN_MASK) / RTE_MEMPOOL_ALIGN;
|
mem: remove redundant check in optimize_object_size
The second condition of this logical OR:
(get_gcd(new_obj_size, nrank * nchan) != 1 ||
get_gcd(nchan, new_obj_size) != 1)
is redundant with the first condition.
We can show that the first condition is equivalent to its disjunction
with the second condition using these two results:
- R1: For all conditions A and B, if B implies A, then (A || B) is
equivalent to A.
- R2: (get_gcd(nchan, new_obj_size) != 1) implies
(get_gcd(new_obj_size, nrank * nchan) != 1)
We can show R1 with the following truth table (0 is false, 1 is true):
+-----+-----++----------+-----+-------------+
| A | B || (A || B) | A | B implies A |
+-----+-----++----------+-----+-------------+
| 0 | 0 || 0 | 0 | 1 |
| 0 | 1 || 1 | 0 | 0 |
| 1 | 0 || 1 | 1 | 1 |
| 1 | 1 || 1 | 1 | 1 |
+-----+-----++----------+-----+-------------+
Truth table of (A || B) and A
We can show R2 by looking at the code of optimize_object_size and
get_gcd.
We see that:
- S1: (nchan >= 1) and (nrank >= 1).
- S2: get_gcd returns 0 only when both arguments are 0.
Let:
- X be get_gcd(new_obj_size, nrank * nchan).
- Y be get_gcd(nchan, new_obj_size).
Suppose:
- H1: get_gcd returns the greatest common divisor of its arguments.
- H2: (nrank * nchan) does not exceed UINT_MAX.
We prove (Y != 1) implies (X != 1) with the following steps:
- Suppose L0: (Y != 1). We have to show (X != 1).
- By H1, Y is the greatest common divisor of nchan and new_obj_size.
In particular, we have L1: Y divides nchan and new_obj_size.
- By H2, we have L2: nchan divides (nrank * nchan)
- By L1 and L2, we have L3: Y divides (nrank * nchan) and
new_obj_size.
- By H1 and L3, we have L4: (Y <= X).
- By S1 and S2, we have L5: (Y != 0).
- By L0 and L5, we have L6: (Y > 1).
- By L4 and L6, we have (X > 1) and thus (X != 1), which concludes.
R2 was also tested for all values of new_obj_size, nrank, and nchan
between 0 and 2000.
This redundant condition was found using TrustInSoft Analyzer.
Signed-off-by: Julien Cretin <julien.cretin@trust-in-soft.com>
Acked-by: Thomas Monjalon <thomas.monjalon@6wind.com>
2014-05-12 15:35:10 +00:00
|
|
|
while (get_gcd(new_obj_size, nrank * nchan) != 1)
|
2012-09-04 12:54:00 +00:00
|
|
|
new_obj_size++;
|
2015-07-09 08:25:16 +00:00
|
|
|
return new_obj_size * RTE_MEMPOOL_ALIGN;
|
2012-09-04 12:54:00 +00:00
|
|
|
}
|
|
|
|
|
2018-10-02 13:34:40 +00:00
|
|
|
struct pagesz_walk_arg {
|
|
|
|
int socket_id;
|
|
|
|
size_t min;
|
|
|
|
};
|
|
|
|
|
2018-04-11 12:30:05 +00:00
|
|
|
static int
|
mem: replace memseg with memseg lists
Before, we were aggregating multiple pages into one memseg, so the
number of memsegs was small. Now, each page gets its own memseg,
so the list of memsegs is huge. To accommodate the new memseg list
size and to keep the under-the-hood workings sane, the memseg list
is now not just a single list, but multiple lists. To be precise,
each hugepage size available on the system gets one or more memseg
lists, per socket.
In order to support dynamic memory allocation, we reserve all
memory in advance (unless we're in 32-bit legacy mode, in which
case we do not preallocate memory). As in, we do an anonymous
mmap() of the entire maximum size of memory per hugepage size, per
socket (which is limited to either RTE_MAX_MEMSEG_PER_TYPE pages or
RTE_MAX_MEM_MB_PER_TYPE megabytes worth of memory, whichever is the
smaller one), split over multiple lists (which are limited to
either RTE_MAX_MEMSEG_PER_LIST memsegs or RTE_MAX_MEM_MB_PER_LIST
megabytes per list, whichever is the smaller one). There is also
a global limit of CONFIG_RTE_MAX_MEM_MB megabytes, which is mainly
used for 32-bit targets to limit amounts of preallocated memory,
but can be used to place an upper limit on total amount of VA
memory that can be allocated by DPDK application.
So, for each hugepage size, we get (by default) up to 128G worth
of memory, per socket, split into chunks of up to 32G in size.
The address space is claimed at the start, in eal_common_memory.c.
The actual page allocation code is in eal_memalloc.c (Linux-only),
and largely consists of copied EAL memory init code.
Pages in the list are also indexed by address. That is, in order
to figure out where the page belongs, one can simply look at base
address for a memseg list. Similarly, figuring out IOVA address
of a memzone is a matter of finding the right memseg list, getting
offset and dividing by page size to get the appropriate memseg.
This commit also removes rte_eal_dump_physmem_layout() call,
according to deprecation notice [1], and removes that deprecation
notice as well.
On 32-bit targets due to limited VA space, DPDK will no longer
spread memory to different sockets like before. Instead, it will
(by default) allocate all of the memory on socket where master
lcore is. To override this behavior, --socket-mem must be used.
The rest of the changes are really ripple effects from the memseg
change - heap changes, compile fixes, and rewrites to support
fbarray-backed memseg lists. Due to earlier switch to _walk()
functions, most of the changes are simple fixes, however some
of the _walk() calls were switched to memseg list walk, where
it made sense to do so.
Additionally, we are also switching locks from flock() to fcntl().
Down the line, we will be introducing single-file segments option,
and we cannot use flock() locks to lock parts of the file. Therefore,
we will use fcntl() locks for legacy mem as well, in case someone is
unfortunate enough to accidentally start legacy mem primary process
alongside an already working non-legacy mem-based primary process.
[1] http://dpdk.org/dev/patchwork/patch/34002/
Signed-off-by: Anatoly Burakov <anatoly.burakov@intel.com>
Tested-by: Santosh Shukla <santosh.shukla@caviumnetworks.com>
Tested-by: Hemant Agrawal <hemant.agrawal@nxp.com>
Tested-by: Gowrishankar Muthukrishnan <gowrishankar.m@linux.vnet.ibm.com>
2018-04-11 12:30:24 +00:00
|
|
|
find_min_pagesz(const struct rte_memseg_list *msl, void *arg)
|
2018-04-11 12:30:05 +00:00
|
|
|
{
|
2018-10-02 13:34:40 +00:00
|
|
|
struct pagesz_walk_arg *wa = arg;
|
|
|
|
bool valid;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* we need to only look at page sizes available for a particular socket
|
|
|
|
* ID. so, we either need an exact match on socket ID (can match both
|
|
|
|
* native and external memory), or, if SOCKET_ID_ANY was specified as a
|
|
|
|
* socket ID argument, we must only look at native memory and ignore any
|
|
|
|
* page sizes associated with external memory.
|
|
|
|
*/
|
|
|
|
valid = msl->socket_id == wa->socket_id;
|
|
|
|
valid |= wa->socket_id == SOCKET_ID_ANY && msl->external == 0;
|
2018-04-11 12:30:05 +00:00
|
|
|
|
2018-10-02 13:34:40 +00:00
|
|
|
if (valid && msl->page_sz < wa->min)
|
|
|
|
wa->min = msl->page_sz;
|
2018-04-11 12:30:05 +00:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
mempool: support new allocation methods
If a user has specified that the zone should have contiguous memory,
add a memzone flag to request contiguous memory. Otherwise, account
for the fact that unless we're in IOVA_AS_VA mode, we cannot
guarantee that the pages would be physically contiguous, so we
calculate the memzone size and alignments as if we were getting
the smallest page size available.
However, for the non-IOVA contiguous case, existing mempool size
calculation function doesn't give us expected results, because it
will return memzone sizes aligned to page size (e.g. a 1MB mempool
may use an entire 1GB page), therefore in cases where we weren't
specifically asked to reserve non-contiguous memory, first try
reserving a memzone as IOVA-contiguous, and if that fails, then
try reserving with page-aligned size/alignment.
Signed-off-by: Anatoly Burakov <anatoly.burakov@intel.com>
Acked-by: Olivier Matz <olivier.matz@6wind.com>
Tested-by: Santosh Shukla <santosh.shukla@caviumnetworks.com>
Tested-by: Hemant Agrawal <hemant.agrawal@nxp.com>
Tested-by: Gowrishankar Muthukrishnan <gowrishankar.m@linux.vnet.ibm.com>
2018-04-11 12:29:59 +00:00
|
|
|
static size_t
|
2018-10-02 13:34:40 +00:00
|
|
|
get_min_page_size(int socket_id)
|
mempool: support new allocation methods
If a user has specified that the zone should have contiguous memory,
add a memzone flag to request contiguous memory. Otherwise, account
for the fact that unless we're in IOVA_AS_VA mode, we cannot
guarantee that the pages would be physically contiguous, so we
calculate the memzone size and alignments as if we were getting
the smallest page size available.
However, for the non-IOVA contiguous case, existing mempool size
calculation function doesn't give us expected results, because it
will return memzone sizes aligned to page size (e.g. a 1MB mempool
may use an entire 1GB page), therefore in cases where we weren't
specifically asked to reserve non-contiguous memory, first try
reserving a memzone as IOVA-contiguous, and if that fails, then
try reserving with page-aligned size/alignment.
Signed-off-by: Anatoly Burakov <anatoly.burakov@intel.com>
Acked-by: Olivier Matz <olivier.matz@6wind.com>
Tested-by: Santosh Shukla <santosh.shukla@caviumnetworks.com>
Tested-by: Hemant Agrawal <hemant.agrawal@nxp.com>
Tested-by: Gowrishankar Muthukrishnan <gowrishankar.m@linux.vnet.ibm.com>
2018-04-11 12:29:59 +00:00
|
|
|
{
|
2018-10-02 13:34:40 +00:00
|
|
|
struct pagesz_walk_arg wa;
|
|
|
|
|
|
|
|
wa.min = SIZE_MAX;
|
|
|
|
wa.socket_id = socket_id;
|
mempool: support new allocation methods
If a user has specified that the zone should have contiguous memory,
add a memzone flag to request contiguous memory. Otherwise, account
for the fact that unless we're in IOVA_AS_VA mode, we cannot
guarantee that the pages would be physically contiguous, so we
calculate the memzone size and alignments as if we were getting
the smallest page size available.
However, for the non-IOVA contiguous case, existing mempool size
calculation function doesn't give us expected results, because it
will return memzone sizes aligned to page size (e.g. a 1MB mempool
may use an entire 1GB page), therefore in cases where we weren't
specifically asked to reserve non-contiguous memory, first try
reserving a memzone as IOVA-contiguous, and if that fails, then
try reserving with page-aligned size/alignment.
Signed-off-by: Anatoly Burakov <anatoly.burakov@intel.com>
Acked-by: Olivier Matz <olivier.matz@6wind.com>
Tested-by: Santosh Shukla <santosh.shukla@caviumnetworks.com>
Tested-by: Hemant Agrawal <hemant.agrawal@nxp.com>
Tested-by: Gowrishankar Muthukrishnan <gowrishankar.m@linux.vnet.ibm.com>
2018-04-11 12:29:59 +00:00
|
|
|
|
2018-10-02 13:34:40 +00:00
|
|
|
rte_memseg_list_walk(find_min_pagesz, &wa);
|
mempool: support new allocation methods
If a user has specified that the zone should have contiguous memory,
add a memzone flag to request contiguous memory. Otherwise, account
for the fact that unless we're in IOVA_AS_VA mode, we cannot
guarantee that the pages would be physically contiguous, so we
calculate the memzone size and alignments as if we were getting
the smallest page size available.
However, for the non-IOVA contiguous case, existing mempool size
calculation function doesn't give us expected results, because it
will return memzone sizes aligned to page size (e.g. a 1MB mempool
may use an entire 1GB page), therefore in cases where we weren't
specifically asked to reserve non-contiguous memory, first try
reserving a memzone as IOVA-contiguous, and if that fails, then
try reserving with page-aligned size/alignment.
Signed-off-by: Anatoly Burakov <anatoly.burakov@intel.com>
Acked-by: Olivier Matz <olivier.matz@6wind.com>
Tested-by: Santosh Shukla <santosh.shukla@caviumnetworks.com>
Tested-by: Hemant Agrawal <hemant.agrawal@nxp.com>
Tested-by: Gowrishankar Muthukrishnan <gowrishankar.m@linux.vnet.ibm.com>
2018-04-11 12:29:59 +00:00
|
|
|
|
2018-10-02 13:34:40 +00:00
|
|
|
return wa.min == SIZE_MAX ? (size_t) getpagesize() : wa.min;
|
mempool: support new allocation methods
If a user has specified that the zone should have contiguous memory,
add a memzone flag to request contiguous memory. Otherwise, account
for the fact that unless we're in IOVA_AS_VA mode, we cannot
guarantee that the pages would be physically contiguous, so we
calculate the memzone size and alignments as if we were getting
the smallest page size available.
However, for the non-IOVA contiguous case, existing mempool size
calculation function doesn't give us expected results, because it
will return memzone sizes aligned to page size (e.g. a 1MB mempool
may use an entire 1GB page), therefore in cases where we weren't
specifically asked to reserve non-contiguous memory, first try
reserving a memzone as IOVA-contiguous, and if that fails, then
try reserving with page-aligned size/alignment.
Signed-off-by: Anatoly Burakov <anatoly.burakov@intel.com>
Acked-by: Olivier Matz <olivier.matz@6wind.com>
Tested-by: Santosh Shukla <santosh.shukla@caviumnetworks.com>
Tested-by: Hemant Agrawal <hemant.agrawal@nxp.com>
Tested-by: Gowrishankar Muthukrishnan <gowrishankar.m@linux.vnet.ibm.com>
2018-04-11 12:29:59 +00:00
|
|
|
}
|
|
|
|
|
mem: replace memseg with memseg lists
Before, we were aggregating multiple pages into one memseg, so the
number of memsegs was small. Now, each page gets its own memseg,
so the list of memsegs is huge. To accommodate the new memseg list
size and to keep the under-the-hood workings sane, the memseg list
is now not just a single list, but multiple lists. To be precise,
each hugepage size available on the system gets one or more memseg
lists, per socket.
In order to support dynamic memory allocation, we reserve all
memory in advance (unless we're in 32-bit legacy mode, in which
case we do not preallocate memory). As in, we do an anonymous
mmap() of the entire maximum size of memory per hugepage size, per
socket (which is limited to either RTE_MAX_MEMSEG_PER_TYPE pages or
RTE_MAX_MEM_MB_PER_TYPE megabytes worth of memory, whichever is the
smaller one), split over multiple lists (which are limited to
either RTE_MAX_MEMSEG_PER_LIST memsegs or RTE_MAX_MEM_MB_PER_LIST
megabytes per list, whichever is the smaller one). There is also
a global limit of CONFIG_RTE_MAX_MEM_MB megabytes, which is mainly
used for 32-bit targets to limit amounts of preallocated memory,
but can be used to place an upper limit on total amount of VA
memory that can be allocated by DPDK application.
So, for each hugepage size, we get (by default) up to 128G worth
of memory, per socket, split into chunks of up to 32G in size.
The address space is claimed at the start, in eal_common_memory.c.
The actual page allocation code is in eal_memalloc.c (Linux-only),
and largely consists of copied EAL memory init code.
Pages in the list are also indexed by address. That is, in order
to figure out where the page belongs, one can simply look at base
address for a memseg list. Similarly, figuring out IOVA address
of a memzone is a matter of finding the right memseg list, getting
offset and dividing by page size to get the appropriate memseg.
This commit also removes rte_eal_dump_physmem_layout() call,
according to deprecation notice [1], and removes that deprecation
notice as well.
On 32-bit targets due to limited VA space, DPDK will no longer
spread memory to different sockets like before. Instead, it will
(by default) allocate all of the memory on socket where master
lcore is. To override this behavior, --socket-mem must be used.
The rest of the changes are really ripple effects from the memseg
change - heap changes, compile fixes, and rewrites to support
fbarray-backed memseg lists. Due to earlier switch to _walk()
functions, most of the changes are simple fixes, however some
of the _walk() calls were switched to memseg list walk, where
it made sense to do so.
Additionally, we are also switching locks from flock() to fcntl().
Down the line, we will be introducing single-file segments option,
and we cannot use flock() locks to lock parts of the file. Therefore,
we will use fcntl() locks for legacy mem as well, in case someone is
unfortunate enough to accidentally start legacy mem primary process
alongside an already working non-legacy mem-based primary process.
[1] http://dpdk.org/dev/patchwork/patch/34002/
Signed-off-by: Anatoly Burakov <anatoly.burakov@intel.com>
Tested-by: Santosh Shukla <santosh.shukla@caviumnetworks.com>
Tested-by: Hemant Agrawal <hemant.agrawal@nxp.com>
Tested-by: Gowrishankar Muthukrishnan <gowrishankar.m@linux.vnet.ibm.com>
2018-04-11 12:30:24 +00:00
|
|
|
|
2014-02-12 15:32:25 +00:00
|
|
|
static void
|
2018-04-16 13:24:34 +00:00
|
|
|
mempool_add_elem(struct rte_mempool *mp, __rte_unused void *opaque,
|
|
|
|
void *obj, rte_iova_t iova)
|
2014-02-12 15:32:25 +00:00
|
|
|
{
|
2015-06-19 16:16:37 +00:00
|
|
|
struct rte_mempool_objhdr *hdr;
|
2015-06-19 16:16:38 +00:00
|
|
|
struct rte_mempool_objtlr *tlr __rte_unused;
|
2014-02-12 15:32:25 +00:00
|
|
|
|
|
|
|
/* set mempool ptr in header */
|
2015-06-22 18:34:15 +00:00
|
|
|
hdr = RTE_PTR_SUB(obj, sizeof(*hdr));
|
2015-06-19 16:16:37 +00:00
|
|
|
hdr->mp = mp;
|
2017-10-20 12:31:31 +00:00
|
|
|
hdr->iova = iova;
|
2016-05-18 11:04:27 +00:00
|
|
|
STAILQ_INSERT_TAIL(&mp->elt_list, hdr, next);
|
2016-05-18 11:04:36 +00:00
|
|
|
mp->populated_size++;
|
2014-02-12 15:32:25 +00:00
|
|
|
|
|
|
|
#ifdef RTE_LIBRTE_MEMPOOL_DEBUG
|
2015-06-19 16:16:37 +00:00
|
|
|
hdr->cookie = RTE_MEMPOOL_HEADER_COOKIE2;
|
2015-06-19 16:16:38 +00:00
|
|
|
tlr = __mempool_get_trailer(obj);
|
|
|
|
tlr->cookie = RTE_MEMPOOL_TRAILER_COOKIE;
|
2014-02-12 15:32:25 +00:00
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
2016-05-18 11:04:30 +00:00
|
|
|
/* call obj_cb() for each mempool element */
|
|
|
|
uint32_t
|
|
|
|
rte_mempool_obj_iter(struct rte_mempool *mp,
|
|
|
|
rte_mempool_obj_cb_t *obj_cb, void *obj_cb_arg)
|
|
|
|
{
|
|
|
|
struct rte_mempool_objhdr *hdr;
|
|
|
|
void *obj;
|
|
|
|
unsigned n = 0;
|
|
|
|
|
|
|
|
STAILQ_FOREACH(hdr, &mp->elt_list, next) {
|
|
|
|
obj = (char *)hdr + sizeof(*hdr);
|
|
|
|
obj_cb(mp, obj_cb_arg, obj, n);
|
|
|
|
n++;
|
|
|
|
}
|
|
|
|
|
|
|
|
return n;
|
|
|
|
}
|
|
|
|
|
2016-05-18 11:04:37 +00:00
|
|
|
/* call mem_cb() for each mempool memory chunk */
|
|
|
|
uint32_t
|
|
|
|
rte_mempool_mem_iter(struct rte_mempool *mp,
|
|
|
|
rte_mempool_mem_cb_t *mem_cb, void *mem_cb_arg)
|
|
|
|
{
|
|
|
|
struct rte_mempool_memhdr *hdr;
|
|
|
|
unsigned n = 0;
|
|
|
|
|
|
|
|
STAILQ_FOREACH(hdr, &mp->mem_list, next) {
|
|
|
|
mem_cb(mp, mem_cb_arg, hdr, n);
|
|
|
|
n++;
|
|
|
|
}
|
|
|
|
|
|
|
|
return n;
|
|
|
|
}
|
|
|
|
|
2016-05-18 11:04:23 +00:00
|
|
|
/* get the header, trailer and total size of a mempool element. */
|
2014-02-12 15:32:25 +00:00
|
|
|
uint32_t
|
|
|
|
rte_mempool_calc_obj_size(uint32_t elt_size, uint32_t flags,
|
|
|
|
struct rte_mempool_objsz *sz)
|
|
|
|
{
|
|
|
|
struct rte_mempool_objsz lsz;
|
|
|
|
|
|
|
|
sz = (sz != NULL) ? sz : &lsz;
|
|
|
|
|
2016-05-18 11:04:25 +00:00
|
|
|
sz->header_size = sizeof(struct rte_mempool_objhdr);
|
2014-02-12 15:32:25 +00:00
|
|
|
if ((flags & MEMPOOL_F_NO_CACHE_ALIGN) == 0)
|
|
|
|
sz->header_size = RTE_ALIGN_CEIL(sz->header_size,
|
2015-07-09 08:25:16 +00:00
|
|
|
RTE_MEMPOOL_ALIGN);
|
2014-02-12 15:32:25 +00:00
|
|
|
|
2016-07-13 12:30:41 +00:00
|
|
|
#ifdef RTE_LIBRTE_MEMPOOL_DEBUG
|
2016-05-18 11:04:25 +00:00
|
|
|
sz->trailer_size = sizeof(struct rte_mempool_objtlr);
|
2016-07-13 12:30:41 +00:00
|
|
|
#else
|
|
|
|
sz->trailer_size = 0;
|
|
|
|
#endif
|
2016-05-18 11:04:25 +00:00
|
|
|
|
2014-02-12 15:32:25 +00:00
|
|
|
/* element size is 8 bytes-aligned at least */
|
|
|
|
sz->elt_size = RTE_ALIGN_CEIL(elt_size, sizeof(uint64_t));
|
|
|
|
|
|
|
|
/* expand trailer to next cache line */
|
|
|
|
if ((flags & MEMPOOL_F_NO_CACHE_ALIGN) == 0) {
|
|
|
|
sz->total_size = sz->header_size + sz->elt_size +
|
|
|
|
sz->trailer_size;
|
2015-07-09 08:25:16 +00:00
|
|
|
sz->trailer_size += ((RTE_MEMPOOL_ALIGN -
|
|
|
|
(sz->total_size & RTE_MEMPOOL_ALIGN_MASK)) &
|
|
|
|
RTE_MEMPOOL_ALIGN_MASK);
|
2014-02-12 15:32:25 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* increase trailer to add padding between objects in order to
|
2014-05-02 23:42:52 +00:00
|
|
|
* spread them across memory channels/ranks
|
2014-02-12 15:32:25 +00:00
|
|
|
*/
|
|
|
|
if ((flags & MEMPOOL_F_NO_SPREAD) == 0) {
|
|
|
|
unsigned new_size;
|
|
|
|
new_size = optimize_object_size(sz->header_size + sz->elt_size +
|
|
|
|
sz->trailer_size);
|
|
|
|
sz->trailer_size = new_size - sz->header_size - sz->elt_size;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* this is the size of an object, including header and trailer */
|
|
|
|
sz->total_size = sz->header_size + sz->elt_size + sz->trailer_size;
|
|
|
|
|
2015-05-11 14:10:25 +00:00
|
|
|
return sz->total_size;
|
2014-02-12 15:32:25 +00:00
|
|
|
}
|
|
|
|
|
2016-05-18 11:04:39 +00:00
|
|
|
/* free a memchunk allocated with rte_memzone_reserve() */
|
2016-05-18 11:04:41 +00:00
|
|
|
static void
|
2016-05-18 11:04:39 +00:00
|
|
|
rte_mempool_memchunk_mz_free(__rte_unused struct rte_mempool_memhdr *memhdr,
|
|
|
|
void *opaque)
|
|
|
|
{
|
|
|
|
const struct rte_memzone *mz = opaque;
|
|
|
|
rte_memzone_free(mz);
|
|
|
|
}
|
|
|
|
|
2016-05-18 11:04:36 +00:00
|
|
|
/* Free memory chunks used by a mempool. Objects must be in pool */
|
|
|
|
static void
|
|
|
|
rte_mempool_free_memchunks(struct rte_mempool *mp)
|
|
|
|
{
|
|
|
|
struct rte_mempool_memhdr *memhdr;
|
|
|
|
void *elt;
|
|
|
|
|
|
|
|
while (!STAILQ_EMPTY(&mp->elt_list)) {
|
2016-06-22 09:27:27 +00:00
|
|
|
rte_mempool_ops_dequeue_bulk(mp, &elt, 1);
|
2016-05-18 11:04:36 +00:00
|
|
|
(void)elt;
|
|
|
|
STAILQ_REMOVE_HEAD(&mp->elt_list, next);
|
|
|
|
mp->populated_size--;
|
|
|
|
}
|
|
|
|
|
|
|
|
while (!STAILQ_EMPTY(&mp->mem_list)) {
|
|
|
|
memhdr = STAILQ_FIRST(&mp->mem_list);
|
|
|
|
STAILQ_REMOVE_HEAD(&mp->mem_list, next);
|
2016-05-18 11:04:39 +00:00
|
|
|
if (memhdr->free_cb != NULL)
|
|
|
|
memhdr->free_cb(memhdr, memhdr->opaque);
|
2016-05-18 11:04:36 +00:00
|
|
|
rte_free(memhdr);
|
|
|
|
mp->nb_mem_chunks--;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2018-04-16 13:24:32 +00:00
|
|
|
static int
|
|
|
|
mempool_ops_alloc_once(struct rte_mempool *mp)
|
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
/* create the internal ring if not already done */
|
|
|
|
if ((mp->flags & MEMPOOL_F_POOL_CREATED) == 0) {
|
|
|
|
ret = rte_mempool_ops_alloc(mp);
|
|
|
|
if (ret != 0)
|
|
|
|
return ret;
|
|
|
|
mp->flags |= MEMPOOL_F_POOL_CREATED;
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2016-05-18 11:04:36 +00:00
|
|
|
/* Add objects in the pool, using a physically contiguous memory
|
|
|
|
* zone. Return the number of objects added, or a negative value
|
|
|
|
* on error.
|
|
|
|
*/
|
2016-05-18 11:04:51 +00:00
|
|
|
int
|
2017-11-05 22:26:24 +00:00
|
|
|
rte_mempool_populate_iova(struct rte_mempool *mp, char *vaddr,
|
|
|
|
rte_iova_t iova, size_t len, rte_mempool_memchunk_free_cb_t *free_cb,
|
2016-05-18 11:04:39 +00:00
|
|
|
void *opaque)
|
2016-05-18 11:04:36 +00:00
|
|
|
{
|
|
|
|
unsigned i = 0;
|
|
|
|
size_t off;
|
|
|
|
struct rte_mempool_memhdr *memhdr;
|
2016-05-18 11:04:49 +00:00
|
|
|
int ret;
|
|
|
|
|
2018-04-16 13:24:32 +00:00
|
|
|
ret = mempool_ops_alloc_once(mp);
|
|
|
|
if (ret != 0)
|
|
|
|
return ret;
|
2016-05-18 11:04:36 +00:00
|
|
|
|
|
|
|
/* mempool is already populated */
|
|
|
|
if (mp->populated_size >= mp->size)
|
|
|
|
return -ENOSPC;
|
|
|
|
|
|
|
|
memhdr = rte_zmalloc("MEMPOOL_MEMHDR", sizeof(*memhdr), 0);
|
|
|
|
if (memhdr == NULL)
|
|
|
|
return -ENOMEM;
|
|
|
|
|
|
|
|
memhdr->mp = mp;
|
|
|
|
memhdr->addr = vaddr;
|
2017-11-05 22:26:24 +00:00
|
|
|
memhdr->iova = iova;
|
2016-05-18 11:04:36 +00:00
|
|
|
memhdr->len = len;
|
2016-05-18 11:04:39 +00:00
|
|
|
memhdr->free_cb = free_cb;
|
|
|
|
memhdr->opaque = opaque;
|
2016-05-18 11:04:36 +00:00
|
|
|
|
2018-04-16 13:24:35 +00:00
|
|
|
if (mp->flags & MEMPOOL_F_NO_CACHE_ALIGN)
|
2016-05-18 11:04:36 +00:00
|
|
|
off = RTE_PTR_ALIGN_CEIL(vaddr, 8) - vaddr;
|
|
|
|
else
|
2019-11-05 15:37:06 +00:00
|
|
|
off = RTE_PTR_ALIGN_CEIL(vaddr, RTE_MEMPOOL_ALIGN) - vaddr;
|
2016-05-18 11:04:36 +00:00
|
|
|
|
2018-04-16 13:24:34 +00:00
|
|
|
if (off > len) {
|
|
|
|
ret = -EINVAL;
|
|
|
|
goto fail;
|
2016-05-18 11:04:36 +00:00
|
|
|
}
|
|
|
|
|
2018-04-16 13:24:34 +00:00
|
|
|
i = rte_mempool_ops_populate(mp, mp->size - mp->populated_size,
|
|
|
|
(char *)vaddr + off,
|
|
|
|
(iova == RTE_BAD_IOVA) ? RTE_BAD_IOVA : (iova + off),
|
|
|
|
len - off, mempool_add_elem, NULL);
|
|
|
|
|
2016-05-18 11:04:36 +00:00
|
|
|
/* not enough room to store one object */
|
2018-04-16 13:24:30 +00:00
|
|
|
if (i == 0) {
|
|
|
|
ret = -EINVAL;
|
|
|
|
goto fail;
|
|
|
|
}
|
2016-05-18 11:04:36 +00:00
|
|
|
|
|
|
|
STAILQ_INSERT_TAIL(&mp->mem_list, memhdr, next);
|
|
|
|
mp->nb_mem_chunks++;
|
|
|
|
return i;
|
2018-04-16 13:24:30 +00:00
|
|
|
|
|
|
|
fail:
|
|
|
|
rte_free(memhdr);
|
|
|
|
return ret;
|
2016-05-18 11:04:36 +00:00
|
|
|
}
|
|
|
|
|
2019-11-14 13:58:20 +00:00
|
|
|
static rte_iova_t
|
|
|
|
get_iova(void *addr)
|
|
|
|
{
|
|
|
|
struct rte_memseg *ms;
|
|
|
|
|
|
|
|
/* try registered memory first */
|
|
|
|
ms = rte_mem_virt2memseg(addr, NULL);
|
|
|
|
if (ms == NULL || ms->iova == RTE_BAD_IOVA)
|
|
|
|
/* fall back to actual physical address */
|
|
|
|
return rte_mem_virt2iova(addr);
|
|
|
|
return ms->iova + RTE_PTR_DIFF(addr, ms->addr);
|
|
|
|
}
|
|
|
|
|
2016-05-18 11:04:43 +00:00
|
|
|
/* Populate the mempool with a virtual area. Return the number of
|
|
|
|
* objects added, or a negative value on error.
|
|
|
|
*/
|
2016-05-18 11:04:51 +00:00
|
|
|
int
|
2016-05-18 11:04:43 +00:00
|
|
|
rte_mempool_populate_virt(struct rte_mempool *mp, char *addr,
|
|
|
|
size_t len, size_t pg_sz, rte_mempool_memchunk_free_cb_t *free_cb,
|
|
|
|
void *opaque)
|
|
|
|
{
|
2017-10-20 12:31:31 +00:00
|
|
|
rte_iova_t iova;
|
2016-05-18 11:04:43 +00:00
|
|
|
size_t off, phys_len;
|
|
|
|
int ret, cnt = 0;
|
|
|
|
|
2018-04-16 13:24:31 +00:00
|
|
|
if (mp->flags & MEMPOOL_F_NO_IOVA_CONTIG)
|
2017-11-05 22:26:24 +00:00
|
|
|
return rte_mempool_populate_iova(mp, addr, RTE_BAD_IOVA,
|
2016-05-18 11:04:54 +00:00
|
|
|
len, free_cb, opaque);
|
|
|
|
|
2019-11-05 15:37:00 +00:00
|
|
|
for (off = 0; off < len &&
|
2016-05-18 11:04:43 +00:00
|
|
|
mp->populated_size < mp->size; off += phys_len) {
|
|
|
|
|
2019-11-14 13:58:20 +00:00
|
|
|
iova = get_iova(addr + off);
|
2016-05-18 11:04:48 +00:00
|
|
|
|
2016-05-18 11:04:43 +00:00
|
|
|
/* populate with the largest group of contiguous pages */
|
2019-11-05 15:37:00 +00:00
|
|
|
for (phys_len = RTE_MIN(
|
|
|
|
(size_t)(RTE_PTR_ALIGN_CEIL(addr + off + 1, pg_sz) -
|
|
|
|
(addr + off)),
|
|
|
|
len - off);
|
|
|
|
off + phys_len < len;
|
|
|
|
phys_len = RTE_MIN(phys_len + pg_sz, len - off)) {
|
2017-10-20 12:31:31 +00:00
|
|
|
rte_iova_t iova_tmp;
|
2016-05-18 11:04:43 +00:00
|
|
|
|
2019-11-14 13:58:20 +00:00
|
|
|
iova_tmp = get_iova(addr + off + phys_len);
|
2016-05-18 11:04:43 +00:00
|
|
|
|
2019-11-05 15:37:00 +00:00
|
|
|
if (iova_tmp == RTE_BAD_IOVA ||
|
|
|
|
iova_tmp != iova + phys_len)
|
2016-05-18 11:04:43 +00:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
2017-11-05 22:26:24 +00:00
|
|
|
ret = rte_mempool_populate_iova(mp, addr + off, iova,
|
2016-05-18 11:04:43 +00:00
|
|
|
phys_len, free_cb, opaque);
|
|
|
|
if (ret < 0)
|
|
|
|
goto fail;
|
|
|
|
/* no need to call the free callback for next chunks */
|
|
|
|
free_cb = NULL;
|
|
|
|
cnt += ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
return cnt;
|
|
|
|
|
|
|
|
fail:
|
|
|
|
rte_mempool_free_memchunks(mp);
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2019-11-05 15:37:03 +00:00
|
|
|
/* Get the minimal page size used in a mempool before populating it. */
|
|
|
|
int
|
|
|
|
rte_mempool_get_page_size(struct rte_mempool *mp, size_t *pg_sz)
|
|
|
|
{
|
|
|
|
bool need_iova_contig_obj;
|
|
|
|
bool alloc_in_ext_mem;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
/* check if we can retrieve a valid socket ID */
|
|
|
|
ret = rte_malloc_heap_socket_is_external(mp->socket_id);
|
|
|
|
if (ret < 0)
|
|
|
|
return -EINVAL;
|
|
|
|
alloc_in_ext_mem = (ret == 1);
|
|
|
|
need_iova_contig_obj = !(mp->flags & MEMPOOL_F_NO_IOVA_CONTIG);
|
|
|
|
|
|
|
|
if (!need_iova_contig_obj)
|
|
|
|
*pg_sz = 0;
|
|
|
|
else if (rte_eal_has_hugepages() || alloc_in_ext_mem)
|
|
|
|
*pg_sz = get_min_page_size(mp->socket_id);
|
|
|
|
else
|
|
|
|
*pg_sz = getpagesize();
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2016-05-18 11:04:43 +00:00
|
|
|
/* Default function to populate the mempool: allocate memory in memzones,
|
2016-05-18 11:04:41 +00:00
|
|
|
* and populate them. Return the number of objects added, or a negative
|
|
|
|
* value on error.
|
|
|
|
*/
|
2016-05-18 11:04:51 +00:00
|
|
|
int
|
2016-05-18 11:04:43 +00:00
|
|
|
rte_mempool_populate_default(struct rte_mempool *mp)
|
2016-05-18 11:04:41 +00:00
|
|
|
{
|
2017-10-01 09:28:56 +00:00
|
|
|
unsigned int mz_flags = RTE_MEMZONE_1GB|RTE_MEMZONE_SIZE_HINT_ONLY;
|
2016-05-18 11:04:41 +00:00
|
|
|
char mz_name[RTE_MEMZONE_NAMESIZE];
|
|
|
|
const struct rte_memzone *mz;
|
2018-04-16 13:24:33 +00:00
|
|
|
ssize_t mem_size;
|
2019-11-05 15:37:03 +00:00
|
|
|
size_t align, pg_sz, pg_shift = 0;
|
2017-10-20 12:31:31 +00:00
|
|
|
rte_iova_t iova;
|
2016-05-18 11:04:41 +00:00
|
|
|
unsigned mz_id, n;
|
|
|
|
int ret;
|
2019-10-08 09:34:06 +00:00
|
|
|
bool need_iova_contig_obj;
|
2016-05-18 11:04:41 +00:00
|
|
|
|
2018-04-16 13:24:32 +00:00
|
|
|
ret = mempool_ops_alloc_once(mp);
|
|
|
|
if (ret != 0)
|
|
|
|
return ret;
|
|
|
|
|
2016-05-18 11:04:41 +00:00
|
|
|
/* mempool must not be populated */
|
|
|
|
if (mp->nb_mem_chunks != 0)
|
|
|
|
return -EEXIST;
|
|
|
|
|
mempool: support new allocation methods
If a user has specified that the zone should have contiguous memory,
add a memzone flag to request contiguous memory. Otherwise, account
for the fact that unless we're in IOVA_AS_VA mode, we cannot
guarantee that the pages would be physically contiguous, so we
calculate the memzone size and alignments as if we were getting
the smallest page size available.
However, for the non-IOVA contiguous case, existing mempool size
calculation function doesn't give us expected results, because it
will return memzone sizes aligned to page size (e.g. a 1MB mempool
may use an entire 1GB page), therefore in cases where we weren't
specifically asked to reserve non-contiguous memory, first try
reserving a memzone as IOVA-contiguous, and if that fails, then
try reserving with page-aligned size/alignment.
Signed-off-by: Anatoly Burakov <anatoly.burakov@intel.com>
Acked-by: Olivier Matz <olivier.matz@6wind.com>
Tested-by: Santosh Shukla <santosh.shukla@caviumnetworks.com>
Tested-by: Hemant Agrawal <hemant.agrawal@nxp.com>
Tested-by: Gowrishankar Muthukrishnan <gowrishankar.m@linux.vnet.ibm.com>
2018-04-11 12:29:59 +00:00
|
|
|
/*
|
|
|
|
* the following section calculates page shift and page size values.
|
|
|
|
*
|
2018-04-16 13:24:33 +00:00
|
|
|
* these values impact the result of calc_mem_size operation, which
|
mempool: support new allocation methods
If a user has specified that the zone should have contiguous memory,
add a memzone flag to request contiguous memory. Otherwise, account
for the fact that unless we're in IOVA_AS_VA mode, we cannot
guarantee that the pages would be physically contiguous, so we
calculate the memzone size and alignments as if we were getting
the smallest page size available.
However, for the non-IOVA contiguous case, existing mempool size
calculation function doesn't give us expected results, because it
will return memzone sizes aligned to page size (e.g. a 1MB mempool
may use an entire 1GB page), therefore in cases where we weren't
specifically asked to reserve non-contiguous memory, first try
reserving a memzone as IOVA-contiguous, and if that fails, then
try reserving with page-aligned size/alignment.
Signed-off-by: Anatoly Burakov <anatoly.burakov@intel.com>
Acked-by: Olivier Matz <olivier.matz@6wind.com>
Tested-by: Santosh Shukla <santosh.shukla@caviumnetworks.com>
Tested-by: Hemant Agrawal <hemant.agrawal@nxp.com>
Tested-by: Gowrishankar Muthukrishnan <gowrishankar.m@linux.vnet.ibm.com>
2018-04-11 12:29:59 +00:00
|
|
|
* returns the amount of memory that should be allocated to store the
|
|
|
|
* desired number of objects. when not zero, it allocates more memory
|
|
|
|
* for the padding between objects, to ensure that an object does not
|
|
|
|
* cross a page boundary. in other words, page size/shift are to be set
|
|
|
|
* to zero if mempool elements won't care about page boundaries.
|
|
|
|
* there are several considerations for page size and page shift here.
|
|
|
|
*
|
|
|
|
* if we don't need our mempools to have physically contiguous objects,
|
|
|
|
* then just set page shift and page size to 0, because the user has
|
|
|
|
* indicated that there's no need to care about anything.
|
|
|
|
*
|
2019-11-05 15:37:05 +00:00
|
|
|
* if we do need contiguous objects (if a mempool driver has its
|
|
|
|
* own calc_size() method returning min_chunk_size = mem_size),
|
|
|
|
* there is also an option to reserve the entire mempool memory
|
|
|
|
* as one contiguous block of memory.
|
mempool: support new allocation methods
If a user has specified that the zone should have contiguous memory,
add a memzone flag to request contiguous memory. Otherwise, account
for the fact that unless we're in IOVA_AS_VA mode, we cannot
guarantee that the pages would be physically contiguous, so we
calculate the memzone size and alignments as if we were getting
the smallest page size available.
However, for the non-IOVA contiguous case, existing mempool size
calculation function doesn't give us expected results, because it
will return memzone sizes aligned to page size (e.g. a 1MB mempool
may use an entire 1GB page), therefore in cases where we weren't
specifically asked to reserve non-contiguous memory, first try
reserving a memzone as IOVA-contiguous, and if that fails, then
try reserving with page-aligned size/alignment.
Signed-off-by: Anatoly Burakov <anatoly.burakov@intel.com>
Acked-by: Olivier Matz <olivier.matz@6wind.com>
Tested-by: Santosh Shukla <santosh.shukla@caviumnetworks.com>
Tested-by: Hemant Agrawal <hemant.agrawal@nxp.com>
Tested-by: Gowrishankar Muthukrishnan <gowrishankar.m@linux.vnet.ibm.com>
2018-04-11 12:29:59 +00:00
|
|
|
*
|
|
|
|
* if we require contiguous objects, but not necessarily the entire
|
2019-11-05 15:37:05 +00:00
|
|
|
* mempool reserved space to be contiguous, pg_sz will be != 0,
|
|
|
|
* and the default ops->populate() will take care of not placing
|
|
|
|
* objects across pages.
|
mempool: support new allocation methods
If a user has specified that the zone should have contiguous memory,
add a memzone flag to request contiguous memory. Otherwise, account
for the fact that unless we're in IOVA_AS_VA mode, we cannot
guarantee that the pages would be physically contiguous, so we
calculate the memzone size and alignments as if we were getting
the smallest page size available.
However, for the non-IOVA contiguous case, existing mempool size
calculation function doesn't give us expected results, because it
will return memzone sizes aligned to page size (e.g. a 1MB mempool
may use an entire 1GB page), therefore in cases where we weren't
specifically asked to reserve non-contiguous memory, first try
reserving a memzone as IOVA-contiguous, and if that fails, then
try reserving with page-aligned size/alignment.
Signed-off-by: Anatoly Burakov <anatoly.burakov@intel.com>
Acked-by: Olivier Matz <olivier.matz@6wind.com>
Tested-by: Santosh Shukla <santosh.shukla@caviumnetworks.com>
Tested-by: Hemant Agrawal <hemant.agrawal@nxp.com>
Tested-by: Gowrishankar Muthukrishnan <gowrishankar.m@linux.vnet.ibm.com>
2018-04-11 12:29:59 +00:00
|
|
|
*
|
|
|
|
* if our IO addresses are physical, we may get memory from bigger
|
|
|
|
* pages, or we might get memory from smaller pages, and how much of it
|
|
|
|
* we require depends on whether we want bigger or smaller pages.
|
|
|
|
* However, requesting each and every memory size is too much work, so
|
|
|
|
* what we'll do instead is walk through the page sizes available, pick
|
|
|
|
* the smallest one and set up page shift to match that one. We will be
|
|
|
|
* wasting some space this way, but it's much nicer than looping around
|
|
|
|
* trying to reserve each and every page size.
|
|
|
|
*
|
2019-11-05 15:37:01 +00:00
|
|
|
* If we fail to get enough contiguous memory, then we'll go and
|
|
|
|
* reserve space in smaller chunks.
|
mempool: support new allocation methods
If a user has specified that the zone should have contiguous memory,
add a memzone flag to request contiguous memory. Otherwise, account
for the fact that unless we're in IOVA_AS_VA mode, we cannot
guarantee that the pages would be physically contiguous, so we
calculate the memzone size and alignments as if we were getting
the smallest page size available.
However, for the non-IOVA contiguous case, existing mempool size
calculation function doesn't give us expected results, because it
will return memzone sizes aligned to page size (e.g. a 1MB mempool
may use an entire 1GB page), therefore in cases where we weren't
specifically asked to reserve non-contiguous memory, first try
reserving a memzone as IOVA-contiguous, and if that fails, then
try reserving with page-aligned size/alignment.
Signed-off-by: Anatoly Burakov <anatoly.burakov@intel.com>
Acked-by: Olivier Matz <olivier.matz@6wind.com>
Tested-by: Santosh Shukla <santosh.shukla@caviumnetworks.com>
Tested-by: Hemant Agrawal <hemant.agrawal@nxp.com>
Tested-by: Gowrishankar Muthukrishnan <gowrishankar.m@linux.vnet.ibm.com>
2018-04-11 12:29:59 +00:00
|
|
|
*/
|
|
|
|
|
2019-10-08 09:34:06 +00:00
|
|
|
need_iova_contig_obj = !(mp->flags & MEMPOOL_F_NO_IOVA_CONTIG);
|
2019-11-05 15:37:03 +00:00
|
|
|
ret = rte_mempool_get_page_size(mp, &pg_sz);
|
|
|
|
if (ret < 0)
|
|
|
|
return ret;
|
2019-10-08 09:34:06 +00:00
|
|
|
|
2019-11-05 15:37:03 +00:00
|
|
|
if (pg_sz != 0)
|
2016-05-18 11:04:43 +00:00
|
|
|
pg_shift = rte_bsf32(pg_sz);
|
|
|
|
|
2016-05-18 11:04:41 +00:00
|
|
|
for (mz_id = 0, n = mp->size; n > 0; mz_id++, n -= ret) {
|
2018-04-16 13:24:33 +00:00
|
|
|
size_t min_chunk_size;
|
|
|
|
|
2019-11-05 15:37:02 +00:00
|
|
|
mem_size = rte_mempool_ops_calc_mem_size(
|
|
|
|
mp, n, pg_shift, &min_chunk_size, &align);
|
2018-04-16 13:24:33 +00:00
|
|
|
|
|
|
|
if (mem_size < 0) {
|
|
|
|
ret = mem_size;
|
|
|
|
goto fail;
|
|
|
|
}
|
2016-05-18 11:04:41 +00:00
|
|
|
|
|
|
|
ret = snprintf(mz_name, sizeof(mz_name),
|
|
|
|
RTE_MEMPOOL_MZ_FORMAT "_%d", mp->name, mz_id);
|
|
|
|
if (ret < 0 || ret >= (int)sizeof(mz_name)) {
|
|
|
|
ret = -ENAMETOOLONG;
|
|
|
|
goto fail;
|
|
|
|
}
|
|
|
|
|
mempool: support new allocation methods
If a user has specified that the zone should have contiguous memory,
add a memzone flag to request contiguous memory. Otherwise, account
for the fact that unless we're in IOVA_AS_VA mode, we cannot
guarantee that the pages would be physically contiguous, so we
calculate the memzone size and alignments as if we were getting
the smallest page size available.
However, for the non-IOVA contiguous case, existing mempool size
calculation function doesn't give us expected results, because it
will return memzone sizes aligned to page size (e.g. a 1MB mempool
may use an entire 1GB page), therefore in cases where we weren't
specifically asked to reserve non-contiguous memory, first try
reserving a memzone as IOVA-contiguous, and if that fails, then
try reserving with page-aligned size/alignment.
Signed-off-by: Anatoly Burakov <anatoly.burakov@intel.com>
Acked-by: Olivier Matz <olivier.matz@6wind.com>
Tested-by: Santosh Shukla <santosh.shukla@caviumnetworks.com>
Tested-by: Hemant Agrawal <hemant.agrawal@nxp.com>
Tested-by: Gowrishankar Muthukrishnan <gowrishankar.m@linux.vnet.ibm.com>
2018-04-11 12:29:59 +00:00
|
|
|
/* if we're trying to reserve contiguous memory, add appropriate
|
|
|
|
* memzone flag.
|
|
|
|
*/
|
2019-11-05 15:37:02 +00:00
|
|
|
if (min_chunk_size == (size_t)mem_size)
|
|
|
|
mz_flags |= RTE_MEMZONE_IOVA_CONTIG;
|
mempool: support new allocation methods
If a user has specified that the zone should have contiguous memory,
add a memzone flag to request contiguous memory. Otherwise, account
for the fact that unless we're in IOVA_AS_VA mode, we cannot
guarantee that the pages would be physically contiguous, so we
calculate the memzone size and alignments as if we were getting
the smallest page size available.
However, for the non-IOVA contiguous case, existing mempool size
calculation function doesn't give us expected results, because it
will return memzone sizes aligned to page size (e.g. a 1MB mempool
may use an entire 1GB page), therefore in cases where we weren't
specifically asked to reserve non-contiguous memory, first try
reserving a memzone as IOVA-contiguous, and if that fails, then
try reserving with page-aligned size/alignment.
Signed-off-by: Anatoly Burakov <anatoly.burakov@intel.com>
Acked-by: Olivier Matz <olivier.matz@6wind.com>
Tested-by: Santosh Shukla <santosh.shukla@caviumnetworks.com>
Tested-by: Hemant Agrawal <hemant.agrawal@nxp.com>
Tested-by: Gowrishankar Muthukrishnan <gowrishankar.m@linux.vnet.ibm.com>
2018-04-11 12:29:59 +00:00
|
|
|
|
2018-04-16 13:24:33 +00:00
|
|
|
mz = rte_memzone_reserve_aligned(mz_name, mem_size,
|
2019-11-05 15:37:02 +00:00
|
|
|
mp->socket_id, mz_flags, align);
|
mempool: support new allocation methods
If a user has specified that the zone should have contiguous memory,
add a memzone flag to request contiguous memory. Otherwise, account
for the fact that unless we're in IOVA_AS_VA mode, we cannot
guarantee that the pages would be physically contiguous, so we
calculate the memzone size and alignments as if we were getting
the smallest page size available.
However, for the non-IOVA contiguous case, existing mempool size
calculation function doesn't give us expected results, because it
will return memzone sizes aligned to page size (e.g. a 1MB mempool
may use an entire 1GB page), therefore in cases where we weren't
specifically asked to reserve non-contiguous memory, first try
reserving a memzone as IOVA-contiguous, and if that fails, then
try reserving with page-aligned size/alignment.
Signed-off-by: Anatoly Burakov <anatoly.burakov@intel.com>
Acked-by: Olivier Matz <olivier.matz@6wind.com>
Tested-by: Santosh Shukla <santosh.shukla@caviumnetworks.com>
Tested-by: Hemant Agrawal <hemant.agrawal@nxp.com>
Tested-by: Gowrishankar Muthukrishnan <gowrishankar.m@linux.vnet.ibm.com>
2018-04-11 12:29:59 +00:00
|
|
|
|
|
|
|
/* don't try reserving with 0 size if we were asked to reserve
|
|
|
|
* IOVA-contiguous memory.
|
|
|
|
*/
|
2018-04-16 13:24:33 +00:00
|
|
|
if (min_chunk_size < (size_t)mem_size && mz == NULL) {
|
mempool: support new allocation methods
If a user has specified that the zone should have contiguous memory,
add a memzone flag to request contiguous memory. Otherwise, account
for the fact that unless we're in IOVA_AS_VA mode, we cannot
guarantee that the pages would be physically contiguous, so we
calculate the memzone size and alignments as if we were getting
the smallest page size available.
However, for the non-IOVA contiguous case, existing mempool size
calculation function doesn't give us expected results, because it
will return memzone sizes aligned to page size (e.g. a 1MB mempool
may use an entire 1GB page), therefore in cases where we weren't
specifically asked to reserve non-contiguous memory, first try
reserving a memzone as IOVA-contiguous, and if that fails, then
try reserving with page-aligned size/alignment.
Signed-off-by: Anatoly Burakov <anatoly.burakov@intel.com>
Acked-by: Olivier Matz <olivier.matz@6wind.com>
Tested-by: Santosh Shukla <santosh.shukla@caviumnetworks.com>
Tested-by: Hemant Agrawal <hemant.agrawal@nxp.com>
Tested-by: Gowrishankar Muthukrishnan <gowrishankar.m@linux.vnet.ibm.com>
2018-04-11 12:29:59 +00:00
|
|
|
/* not enough memory, retry with the biggest zone we
|
|
|
|
* have
|
|
|
|
*/
|
2016-05-18 11:04:41 +00:00
|
|
|
mz = rte_memzone_reserve_aligned(mz_name, 0,
|
2019-11-05 15:37:02 +00:00
|
|
|
mp->socket_id, mz_flags, align);
|
mempool: support new allocation methods
If a user has specified that the zone should have contiguous memory,
add a memzone flag to request contiguous memory. Otherwise, account
for the fact that unless we're in IOVA_AS_VA mode, we cannot
guarantee that the pages would be physically contiguous, so we
calculate the memzone size and alignments as if we were getting
the smallest page size available.
However, for the non-IOVA contiguous case, existing mempool size
calculation function doesn't give us expected results, because it
will return memzone sizes aligned to page size (e.g. a 1MB mempool
may use an entire 1GB page), therefore in cases where we weren't
specifically asked to reserve non-contiguous memory, first try
reserving a memzone as IOVA-contiguous, and if that fails, then
try reserving with page-aligned size/alignment.
Signed-off-by: Anatoly Burakov <anatoly.burakov@intel.com>
Acked-by: Olivier Matz <olivier.matz@6wind.com>
Tested-by: Santosh Shukla <santosh.shukla@caviumnetworks.com>
Tested-by: Hemant Agrawal <hemant.agrawal@nxp.com>
Tested-by: Gowrishankar Muthukrishnan <gowrishankar.m@linux.vnet.ibm.com>
2018-04-11 12:29:59 +00:00
|
|
|
}
|
2016-05-18 11:04:41 +00:00
|
|
|
if (mz == NULL) {
|
|
|
|
ret = -rte_errno;
|
|
|
|
goto fail;
|
|
|
|
}
|
|
|
|
|
2018-04-16 13:24:33 +00:00
|
|
|
if (mz->len < min_chunk_size) {
|
|
|
|
rte_memzone_free(mz);
|
|
|
|
ret = -ENOMEM;
|
|
|
|
goto fail;
|
|
|
|
}
|
|
|
|
|
2019-10-08 09:34:06 +00:00
|
|
|
if (need_iova_contig_obj)
|
2017-10-20 12:31:31 +00:00
|
|
|
iova = mz->iova;
|
2019-10-08 09:34:06 +00:00
|
|
|
else
|
|
|
|
iova = RTE_BAD_IOVA;
|
2016-05-18 11:04:54 +00:00
|
|
|
|
2019-11-05 15:37:02 +00:00
|
|
|
if (pg_sz == 0 || (mz_flags & RTE_MEMZONE_IOVA_CONTIG))
|
2017-11-05 22:26:24 +00:00
|
|
|
ret = rte_mempool_populate_iova(mp, mz->addr,
|
2017-10-20 12:31:31 +00:00
|
|
|
iova, mz->len,
|
2016-05-18 11:04:43 +00:00
|
|
|
rte_mempool_memchunk_mz_free,
|
|
|
|
(void *)(uintptr_t)mz);
|
|
|
|
else
|
|
|
|
ret = rte_mempool_populate_virt(mp, mz->addr,
|
2019-11-05 15:37:00 +00:00
|
|
|
mz->len, pg_sz,
|
2016-05-18 11:04:43 +00:00
|
|
|
rte_mempool_memchunk_mz_free,
|
|
|
|
(void *)(uintptr_t)mz);
|
2016-11-11 15:47:10 +00:00
|
|
|
if (ret < 0) {
|
|
|
|
rte_memzone_free(mz);
|
2016-05-18 11:04:41 +00:00
|
|
|
goto fail;
|
2016-11-11 15:47:10 +00:00
|
|
|
}
|
2016-05-18 11:04:41 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
return mp->size;
|
|
|
|
|
|
|
|
fail:
|
|
|
|
rte_mempool_free_memchunks(mp);
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2016-05-18 11:04:50 +00:00
|
|
|
/* return the memory size required for mempool objects in anonymous mem */
|
2018-04-16 13:24:33 +00:00
|
|
|
static ssize_t
|
2016-05-18 11:04:50 +00:00
|
|
|
get_anon_size(const struct rte_mempool *mp)
|
|
|
|
{
|
2018-04-16 13:24:33 +00:00
|
|
|
ssize_t size;
|
|
|
|
size_t pg_sz, pg_shift;
|
|
|
|
size_t min_chunk_size;
|
|
|
|
size_t align;
|
2016-05-18 11:04:50 +00:00
|
|
|
|
|
|
|
pg_sz = getpagesize();
|
|
|
|
pg_shift = rte_bsf32(pg_sz);
|
2018-04-16 13:24:33 +00:00
|
|
|
size = rte_mempool_ops_calc_mem_size(mp, mp->size, pg_shift,
|
|
|
|
&min_chunk_size, &align);
|
2016-05-18 11:04:50 +00:00
|
|
|
|
|
|
|
return size;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* unmap a memory zone mapped by rte_mempool_populate_anon() */
|
|
|
|
static void
|
|
|
|
rte_mempool_memchunk_anon_free(struct rte_mempool_memhdr *memhdr,
|
|
|
|
void *opaque)
|
|
|
|
{
|
2018-04-16 13:24:33 +00:00
|
|
|
ssize_t size;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Calculate size since memhdr->len has contiguous chunk length
|
|
|
|
* which may be smaller if anon map is split into many contiguous
|
|
|
|
* chunks. Result must be the same as we calculated on populate.
|
|
|
|
*/
|
|
|
|
size = get_anon_size(memhdr->mp);
|
|
|
|
if (size < 0)
|
|
|
|
return;
|
|
|
|
|
|
|
|
munmap(opaque, size);
|
2016-05-18 11:04:50 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/* populate the mempool with an anonymous mapping */
|
2016-05-18 11:04:51 +00:00
|
|
|
int
|
2016-05-18 11:04:50 +00:00
|
|
|
rte_mempool_populate_anon(struct rte_mempool *mp)
|
|
|
|
{
|
2018-04-16 13:24:33 +00:00
|
|
|
ssize_t size;
|
2016-05-18 11:04:50 +00:00
|
|
|
int ret;
|
|
|
|
char *addr;
|
|
|
|
|
|
|
|
/* mempool is already populated, error */
|
2018-05-14 16:06:15 +00:00
|
|
|
if ((!STAILQ_EMPTY(&mp->mem_list)) || mp->nb_mem_chunks != 0) {
|
2016-05-18 11:04:50 +00:00
|
|
|
rte_errno = EINVAL;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2018-04-16 13:24:32 +00:00
|
|
|
ret = mempool_ops_alloc_once(mp);
|
|
|
|
if (ret != 0)
|
|
|
|
return ret;
|
|
|
|
|
2016-05-18 11:04:50 +00:00
|
|
|
size = get_anon_size(mp);
|
2018-04-16 13:24:33 +00:00
|
|
|
if (size < 0) {
|
|
|
|
rte_errno = -size;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* get chunk of virtually continuous memory */
|
2016-05-18 11:04:50 +00:00
|
|
|
addr = mmap(NULL, size, PROT_READ | PROT_WRITE,
|
|
|
|
MAP_SHARED | MAP_ANONYMOUS, -1, 0);
|
|
|
|
if (addr == MAP_FAILED) {
|
|
|
|
rte_errno = errno;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
/* can't use MMAP_LOCKED, it does not exist on BSD */
|
|
|
|
if (mlock(addr, size) < 0) {
|
|
|
|
rte_errno = errno;
|
|
|
|
munmap(addr, size);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
ret = rte_mempool_populate_virt(mp, addr, size, getpagesize(),
|
|
|
|
rte_mempool_memchunk_anon_free, addr);
|
|
|
|
if (ret == 0)
|
|
|
|
goto fail;
|
|
|
|
|
|
|
|
return mp->populated_size;
|
|
|
|
|
|
|
|
fail:
|
|
|
|
rte_mempool_free_memchunks(mp);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2016-05-18 11:04:45 +00:00
|
|
|
/* free a mempool */
|
2016-05-18 11:04:51 +00:00
|
|
|
void
|
2016-05-18 11:04:45 +00:00
|
|
|
rte_mempool_free(struct rte_mempool *mp)
|
|
|
|
{
|
|
|
|
struct rte_mempool_list *mempool_list = NULL;
|
|
|
|
struct rte_tailq_entry *te;
|
|
|
|
|
|
|
|
if (mp == NULL)
|
|
|
|
return;
|
|
|
|
|
|
|
|
mempool_list = RTE_TAILQ_CAST(rte_mempool_tailq.head, rte_mempool_list);
|
2019-07-05 13:10:28 +00:00
|
|
|
rte_mcfg_tailq_write_lock();
|
2016-05-18 11:04:45 +00:00
|
|
|
/* find out tailq entry */
|
|
|
|
TAILQ_FOREACH(te, mempool_list, next) {
|
|
|
|
if (te->data == (void *)mp)
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (te != NULL) {
|
|
|
|
TAILQ_REMOVE(mempool_list, te, next);
|
|
|
|
rte_free(te);
|
|
|
|
}
|
2019-07-05 13:10:28 +00:00
|
|
|
rte_mcfg_tailq_write_unlock();
|
2016-05-18 11:04:45 +00:00
|
|
|
|
|
|
|
rte_mempool_free_memchunks(mp);
|
2016-06-22 09:27:27 +00:00
|
|
|
rte_mempool_ops_free(mp);
|
2016-05-18 11:04:45 +00:00
|
|
|
rte_memzone_free(mp->mz);
|
|
|
|
}
|
|
|
|
|
mempool: allow user-owned cache
The mempool cache is only available to EAL threads as a per-lcore
resource. Change this so that the user can create and provide their own
cache on mempool get and put operations. This works with non-EAL threads
too. This commit introduces the new API calls:
rte_mempool_cache_create(size, socket_id)
rte_mempool_cache_free(cache)
rte_mempool_cache_flush(cache, mp)
rte_mempool_default_cache(mp, lcore_id)
Changes the API calls:
rte_mempool_generic_put(mp, obj_table, n, cache, flags)
rte_mempool_generic_get(mp, obj_table, n, cache, flags)
The cache-oblivious API calls use the per-lcore default local cache.
Signed-off-by: Lazaros Koromilas <l@nofutznetworks.com>
Acked-by: Olivier Matz <olivier.matz@6wind.com>
2016-06-28 23:47:38 +00:00
|
|
|
static void
|
|
|
|
mempool_cache_init(struct rte_mempool_cache *cache, uint32_t size)
|
|
|
|
{
|
|
|
|
cache->size = size;
|
|
|
|
cache->flushthresh = CALC_CACHE_FLUSHTHRESH(size);
|
|
|
|
cache->len = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Create and initialize a cache for objects that are retrieved from and
|
|
|
|
* returned to an underlying mempool. This structure is identical to the
|
|
|
|
* local_cache[lcore_id] pointed to by the mempool structure.
|
|
|
|
*/
|
|
|
|
struct rte_mempool_cache *
|
|
|
|
rte_mempool_cache_create(uint32_t size, int socket_id)
|
|
|
|
{
|
|
|
|
struct rte_mempool_cache *cache;
|
|
|
|
|
|
|
|
if (size == 0 || size > RTE_MEMPOOL_CACHE_MAX_SIZE) {
|
|
|
|
rte_errno = EINVAL;
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
cache = rte_zmalloc_socket("MEMPOOL_CACHE", sizeof(*cache),
|
|
|
|
RTE_CACHE_LINE_SIZE, socket_id);
|
|
|
|
if (cache == NULL) {
|
|
|
|
RTE_LOG(ERR, MEMPOOL, "Cannot allocate mempool cache.\n");
|
|
|
|
rte_errno = ENOMEM;
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
mempool_cache_init(cache, size);
|
|
|
|
|
|
|
|
return cache;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Free a cache. It's the responsibility of the user to make sure that any
|
|
|
|
* remaining objects in the cache are flushed to the corresponding
|
|
|
|
* mempool.
|
|
|
|
*/
|
|
|
|
void
|
|
|
|
rte_mempool_cache_free(struct rte_mempool_cache *cache)
|
|
|
|
{
|
|
|
|
rte_free(cache);
|
|
|
|
}
|
|
|
|
|
2016-05-18 11:04:46 +00:00
|
|
|
/* create an empty mempool */
|
2016-05-18 11:04:51 +00:00
|
|
|
struct rte_mempool *
|
2016-05-18 11:04:46 +00:00
|
|
|
rte_mempool_create_empty(const char *name, unsigned n, unsigned elt_size,
|
|
|
|
unsigned cache_size, unsigned private_data_size,
|
|
|
|
int socket_id, unsigned flags)
|
2012-09-04 12:54:00 +00:00
|
|
|
{
|
|
|
|
char mz_name[RTE_MEMZONE_NAMESIZE];
|
2015-03-04 21:50:06 +00:00
|
|
|
struct rte_mempool_list *mempool_list;
|
2012-12-19 23:00:00 +00:00
|
|
|
struct rte_mempool *mp = NULL;
|
2016-02-16 14:40:10 +00:00
|
|
|
struct rte_tailq_entry *te = NULL;
|
2016-05-18 11:04:41 +00:00
|
|
|
const struct rte_memzone *mz = NULL;
|
2014-02-12 15:32:25 +00:00
|
|
|
size_t mempool_size;
|
2017-10-01 09:28:56 +00:00
|
|
|
unsigned int mz_flags = RTE_MEMZONE_1GB|RTE_MEMZONE_SIZE_HINT_ONLY;
|
2014-02-12 15:32:25 +00:00
|
|
|
struct rte_mempool_objsz objsz;
|
mempool: allow user-owned cache
The mempool cache is only available to EAL threads as a per-lcore
resource. Change this so that the user can create and provide their own
cache on mempool get and put operations. This works with non-EAL threads
too. This commit introduces the new API calls:
rte_mempool_cache_create(size, socket_id)
rte_mempool_cache_free(cache)
rte_mempool_cache_flush(cache, mp)
rte_mempool_default_cache(mp, lcore_id)
Changes the API calls:
rte_mempool_generic_put(mp, obj_table, n, cache, flags)
rte_mempool_generic_get(mp, obj_table, n, cache, flags)
The cache-oblivious API calls use the per-lcore default local cache.
Signed-off-by: Lazaros Koromilas <l@nofutznetworks.com>
Acked-by: Olivier Matz <olivier.matz@6wind.com>
2016-06-28 23:47:38 +00:00
|
|
|
unsigned lcore_id;
|
2016-05-18 11:04:53 +00:00
|
|
|
int ret;
|
2012-09-04 12:54:00 +00:00
|
|
|
|
|
|
|
/* compilation-time checks */
|
|
|
|
RTE_BUILD_BUG_ON((sizeof(struct rte_mempool) &
|
2014-11-19 12:26:06 +00:00
|
|
|
RTE_CACHE_LINE_MASK) != 0);
|
2012-09-04 12:54:00 +00:00
|
|
|
RTE_BUILD_BUG_ON((sizeof(struct rte_mempool_cache) &
|
2014-11-19 12:26:06 +00:00
|
|
|
RTE_CACHE_LINE_MASK) != 0);
|
2012-09-04 12:54:00 +00:00
|
|
|
#ifdef RTE_LIBRTE_MEMPOOL_DEBUG
|
|
|
|
RTE_BUILD_BUG_ON((sizeof(struct rte_mempool_debug_stats) &
|
2014-11-19 12:26:06 +00:00
|
|
|
RTE_CACHE_LINE_MASK) != 0);
|
2012-09-04 12:54:00 +00:00
|
|
|
RTE_BUILD_BUG_ON((offsetof(struct rte_mempool, stats) &
|
2014-11-19 12:26:06 +00:00
|
|
|
RTE_CACHE_LINE_MASK) != 0);
|
2012-09-04 12:54:00 +00:00
|
|
|
#endif
|
|
|
|
|
2015-03-04 21:50:08 +00:00
|
|
|
mempool_list = RTE_TAILQ_CAST(rte_mempool_tailq.head, rte_mempool_list);
|
2014-06-03 23:42:50 +00:00
|
|
|
|
2018-08-02 00:35:04 +00:00
|
|
|
/* asked for zero items */
|
|
|
|
if (n == 0) {
|
|
|
|
rte_errno = EINVAL;
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2012-09-04 12:54:00 +00:00
|
|
|
/* asked cache too big */
|
2015-05-18 15:35:14 +00:00
|
|
|
if (cache_size > RTE_MEMPOOL_CACHE_MAX_SIZE ||
|
|
|
|
CALC_CACHE_FLUSHTHRESH(cache_size) > n) {
|
2014-02-12 15:32:25 +00:00
|
|
|
rte_errno = EINVAL;
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2012-09-04 12:54:00 +00:00
|
|
|
/* "no cache align" imply "no spread" */
|
|
|
|
if (flags & MEMPOOL_F_NO_CACHE_ALIGN)
|
|
|
|
flags |= MEMPOOL_F_NO_SPREAD;
|
|
|
|
|
2014-02-12 15:32:25 +00:00
|
|
|
/* calculate mempool object sizes. */
|
2013-03-01 15:10:57 +00:00
|
|
|
if (!rte_mempool_calc_obj_size(elt_size, flags, &objsz)) {
|
|
|
|
rte_errno = EINVAL;
|
|
|
|
return NULL;
|
|
|
|
}
|
2014-02-12 15:32:25 +00:00
|
|
|
|
2019-07-05 13:10:29 +00:00
|
|
|
rte_mcfg_mempool_write_lock();
|
2012-12-19 23:00:00 +00:00
|
|
|
|
2012-09-04 12:54:00 +00:00
|
|
|
/*
|
2014-02-12 15:32:25 +00:00
|
|
|
* reserve a memory zone for this mempool: private data is
|
|
|
|
* cache-aligned
|
2012-09-04 12:54:00 +00:00
|
|
|
*/
|
|
|
|
private_data_size = (private_data_size +
|
2015-07-09 08:25:16 +00:00
|
|
|
RTE_MEMPOOL_ALIGN_MASK) & (~RTE_MEMPOOL_ALIGN_MASK);
|
2014-02-12 15:32:25 +00:00
|
|
|
|
2013-03-01 15:10:57 +00:00
|
|
|
|
2014-06-20 15:42:22 +00:00
|
|
|
/* try to allocate tailq entry */
|
|
|
|
te = rte_zmalloc("MEMPOOL_TAILQ_ENTRY", sizeof(*te), 0);
|
|
|
|
if (te == NULL) {
|
|
|
|
RTE_LOG(ERR, MEMPOOL, "Cannot allocate tailq entry!\n");
|
2016-02-16 14:40:10 +00:00
|
|
|
goto exit_unlock;
|
2014-06-20 15:42:22 +00:00
|
|
|
}
|
|
|
|
|
2016-05-18 11:04:36 +00:00
|
|
|
mempool_size = MEMPOOL_HEADER_SIZE(mp, cache_size);
|
2016-04-14 09:42:36 +00:00
|
|
|
mempool_size += private_data_size;
|
2015-07-09 08:25:16 +00:00
|
|
|
mempool_size = RTE_ALIGN_CEIL(mempool_size, RTE_MEMPOOL_ALIGN);
|
2013-03-01 15:10:57 +00:00
|
|
|
|
2016-05-18 11:04:53 +00:00
|
|
|
ret = snprintf(mz_name, sizeof(mz_name), RTE_MEMPOOL_MZ_FORMAT, name);
|
|
|
|
if (ret < 0 || ret >= (int)sizeof(mz_name)) {
|
|
|
|
rte_errno = ENAMETOOLONG;
|
|
|
|
goto exit_unlock;
|
|
|
|
}
|
2012-12-19 23:00:00 +00:00
|
|
|
|
2012-09-04 12:54:00 +00:00
|
|
|
mz = rte_memzone_reserve(mz_name, mempool_size, socket_id, mz_flags);
|
2016-02-16 14:40:10 +00:00
|
|
|
if (mz == NULL)
|
|
|
|
goto exit_unlock;
|
2012-09-04 12:54:00 +00:00
|
|
|
|
|
|
|
/* init the mempool structure */
|
2016-05-18 11:04:46 +00:00
|
|
|
mp = mz->addr;
|
2016-06-09 08:19:47 +00:00
|
|
|
memset(mp, 0, MEMPOOL_HEADER_SIZE(mp, cache_size));
|
2019-04-03 14:45:04 +00:00
|
|
|
ret = strlcpy(mp->name, name, sizeof(mp->name));
|
2016-05-18 11:04:53 +00:00
|
|
|
if (ret < 0 || ret >= (int)sizeof(mp->name)) {
|
|
|
|
rte_errno = ENAMETOOLONG;
|
|
|
|
goto exit_unlock;
|
|
|
|
}
|
2016-05-18 11:04:44 +00:00
|
|
|
mp->mz = mz;
|
2012-09-04 12:54:00 +00:00
|
|
|
mp->size = n;
|
|
|
|
mp->flags = flags;
|
2016-05-18 11:04:46 +00:00
|
|
|
mp->socket_id = socket_id;
|
2014-02-12 15:32:25 +00:00
|
|
|
mp->elt_size = objsz.elt_size;
|
|
|
|
mp->header_size = objsz.header_size;
|
|
|
|
mp->trailer_size = objsz.trailer_size;
|
mempool: allow user-owned cache
The mempool cache is only available to EAL threads as a per-lcore
resource. Change this so that the user can create and provide their own
cache on mempool get and put operations. This works with non-EAL threads
too. This commit introduces the new API calls:
rte_mempool_cache_create(size, socket_id)
rte_mempool_cache_free(cache)
rte_mempool_cache_flush(cache, mp)
rte_mempool_default_cache(mp, lcore_id)
Changes the API calls:
rte_mempool_generic_put(mp, obj_table, n, cache, flags)
rte_mempool_generic_get(mp, obj_table, n, cache, flags)
The cache-oblivious API calls use the per-lcore default local cache.
Signed-off-by: Lazaros Koromilas <l@nofutznetworks.com>
Acked-by: Olivier Matz <olivier.matz@6wind.com>
2016-06-28 23:47:38 +00:00
|
|
|
/* Size of default caches, zero means disabled. */
|
2012-09-04 12:54:00 +00:00
|
|
|
mp->cache_size = cache_size;
|
|
|
|
mp->private_data_size = private_data_size;
|
2016-05-18 11:04:27 +00:00
|
|
|
STAILQ_INIT(&mp->elt_list);
|
2016-05-18 11:04:36 +00:00
|
|
|
STAILQ_INIT(&mp->mem_list);
|
2012-09-04 12:54:00 +00:00
|
|
|
|
2016-04-14 09:42:36 +00:00
|
|
|
/*
|
|
|
|
* local_cache pointer is set even if cache_size is zero.
|
|
|
|
* The local_cache points to just past the elt_pa[] array.
|
|
|
|
*/
|
|
|
|
mp->local_cache = (struct rte_mempool_cache *)
|
2016-05-18 11:04:36 +00:00
|
|
|
RTE_PTR_ADD(mp, MEMPOOL_HEADER_SIZE(mp, 0));
|
2014-02-12 15:32:25 +00:00
|
|
|
|
mempool: allow user-owned cache
The mempool cache is only available to EAL threads as a per-lcore
resource. Change this so that the user can create and provide their own
cache on mempool get and put operations. This works with non-EAL threads
too. This commit introduces the new API calls:
rte_mempool_cache_create(size, socket_id)
rte_mempool_cache_free(cache)
rte_mempool_cache_flush(cache, mp)
rte_mempool_default_cache(mp, lcore_id)
Changes the API calls:
rte_mempool_generic_put(mp, obj_table, n, cache, flags)
rte_mempool_generic_get(mp, obj_table, n, cache, flags)
The cache-oblivious API calls use the per-lcore default local cache.
Signed-off-by: Lazaros Koromilas <l@nofutznetworks.com>
Acked-by: Olivier Matz <olivier.matz@6wind.com>
2016-06-28 23:47:38 +00:00
|
|
|
/* Init all default caches. */
|
|
|
|
if (cache_size != 0) {
|
|
|
|
for (lcore_id = 0; lcore_id < RTE_MAX_LCORE; lcore_id++)
|
|
|
|
mempool_cache_init(&mp->local_cache[lcore_id],
|
|
|
|
cache_size);
|
|
|
|
}
|
|
|
|
|
2016-05-18 11:04:46 +00:00
|
|
|
te->data = mp;
|
2016-06-22 09:27:27 +00:00
|
|
|
|
2019-07-05 13:10:28 +00:00
|
|
|
rte_mcfg_tailq_write_lock();
|
2016-05-18 11:04:46 +00:00
|
|
|
TAILQ_INSERT_TAIL(mempool_list, te, next);
|
2019-07-05 13:10:28 +00:00
|
|
|
rte_mcfg_tailq_write_unlock();
|
2019-07-05 13:10:29 +00:00
|
|
|
rte_mcfg_mempool_write_unlock();
|
2016-05-18 11:04:46 +00:00
|
|
|
|
|
|
|
return mp;
|
|
|
|
|
|
|
|
exit_unlock:
|
2019-07-05 13:10:29 +00:00
|
|
|
rte_mcfg_mempool_write_unlock();
|
2016-05-18 11:04:46 +00:00
|
|
|
rte_free(te);
|
|
|
|
rte_mempool_free(mp);
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* create the mempool */
|
|
|
|
struct rte_mempool *
|
|
|
|
rte_mempool_create(const char *name, unsigned n, unsigned elt_size,
|
|
|
|
unsigned cache_size, unsigned private_data_size,
|
|
|
|
rte_mempool_ctor_t *mp_init, void *mp_init_arg,
|
|
|
|
rte_mempool_obj_cb_t *obj_init, void *obj_init_arg,
|
|
|
|
int socket_id, unsigned flags)
|
|
|
|
{
|
2017-03-31 05:35:35 +00:00
|
|
|
int ret;
|
2016-05-18 11:04:46 +00:00
|
|
|
struct rte_mempool *mp;
|
|
|
|
|
|
|
|
mp = rte_mempool_create_empty(name, n, elt_size, cache_size,
|
|
|
|
private_data_size, socket_id, flags);
|
|
|
|
if (mp == NULL)
|
|
|
|
return NULL;
|
|
|
|
|
2016-06-22 09:27:27 +00:00
|
|
|
/*
|
|
|
|
* Since we have 4 combinations of the SP/SC/MP/MC examine the flags to
|
|
|
|
* set the correct index into the table of ops structs.
|
|
|
|
*/
|
2016-09-08 15:29:57 +00:00
|
|
|
if ((flags & MEMPOOL_F_SP_PUT) && (flags & MEMPOOL_F_SC_GET))
|
2017-03-31 05:35:35 +00:00
|
|
|
ret = rte_mempool_set_ops_byname(mp, "ring_sp_sc", NULL);
|
2016-06-22 09:27:27 +00:00
|
|
|
else if (flags & MEMPOOL_F_SP_PUT)
|
2017-03-31 05:35:35 +00:00
|
|
|
ret = rte_mempool_set_ops_byname(mp, "ring_sp_mc", NULL);
|
2016-06-22 09:27:27 +00:00
|
|
|
else if (flags & MEMPOOL_F_SC_GET)
|
2017-03-31 05:35:35 +00:00
|
|
|
ret = rte_mempool_set_ops_byname(mp, "ring_mp_sc", NULL);
|
2016-06-22 09:27:27 +00:00
|
|
|
else
|
2017-03-31 05:35:35 +00:00
|
|
|
ret = rte_mempool_set_ops_byname(mp, "ring_mp_mc", NULL);
|
|
|
|
|
|
|
|
if (ret)
|
|
|
|
goto fail;
|
2016-06-22 09:27:27 +00:00
|
|
|
|
2016-05-18 11:04:46 +00:00
|
|
|
/* call the mempool priv initializer */
|
2016-05-18 11:04:36 +00:00
|
|
|
if (mp_init)
|
|
|
|
mp_init(mp, mp_init_arg);
|
2014-02-12 15:32:25 +00:00
|
|
|
|
2016-05-18 11:04:46 +00:00
|
|
|
if (rte_mempool_populate_default(mp) < 0)
|
|
|
|
goto fail;
|
|
|
|
|
|
|
|
/* call the object initializers */
|
|
|
|
if (obj_init)
|
|
|
|
rte_mempool_obj_iter(mp, obj_init, obj_init_arg);
|
|
|
|
|
|
|
|
return mp;
|
|
|
|
|
|
|
|
fail:
|
|
|
|
rte_mempool_free(mp);
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2012-09-04 12:54:00 +00:00
|
|
|
/* Return the number of entries in the mempool */
|
2016-06-30 12:49:25 +00:00
|
|
|
unsigned int
|
|
|
|
rte_mempool_avail_count(const struct rte_mempool *mp)
|
2012-09-04 12:54:00 +00:00
|
|
|
{
|
|
|
|
unsigned count;
|
2016-04-14 09:42:36 +00:00
|
|
|
unsigned lcore_id;
|
2012-09-04 12:54:00 +00:00
|
|
|
|
2016-06-22 09:27:27 +00:00
|
|
|
count = rte_mempool_ops_get_count(mp);
|
2012-09-04 12:54:00 +00:00
|
|
|
|
2016-04-14 09:42:36 +00:00
|
|
|
if (mp->cache_size == 0)
|
|
|
|
return count;
|
2012-09-04 12:54:00 +00:00
|
|
|
|
2016-04-14 09:42:36 +00:00
|
|
|
for (lcore_id = 0; lcore_id < RTE_MAX_LCORE; lcore_id++)
|
|
|
|
count += mp->local_cache[lcore_id].len;
|
2012-09-04 12:54:00 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* due to race condition (access to len is not locked), the
|
|
|
|
* total can be greater than size... so fix the result
|
|
|
|
*/
|
|
|
|
if (count > mp->size)
|
|
|
|
return mp->size;
|
|
|
|
return count;
|
|
|
|
}
|
|
|
|
|
2016-06-30 12:49:25 +00:00
|
|
|
/* return the number of entries allocated from the mempool */
|
|
|
|
unsigned int
|
|
|
|
rte_mempool_in_use_count(const struct rte_mempool *mp)
|
|
|
|
{
|
|
|
|
return mp->size - rte_mempool_avail_count(mp);
|
|
|
|
}
|
|
|
|
|
2012-09-04 12:54:00 +00:00
|
|
|
/* dump the cache status */
|
|
|
|
static unsigned
|
2014-05-02 23:42:56 +00:00
|
|
|
rte_mempool_dump_cache(FILE *f, const struct rte_mempool *mp)
|
2012-09-04 12:54:00 +00:00
|
|
|
{
|
|
|
|
unsigned lcore_id;
|
|
|
|
unsigned count = 0;
|
|
|
|
unsigned cache_count;
|
|
|
|
|
mempool: allow user-owned cache
The mempool cache is only available to EAL threads as a per-lcore
resource. Change this so that the user can create and provide their own
cache on mempool get and put operations. This works with non-EAL threads
too. This commit introduces the new API calls:
rte_mempool_cache_create(size, socket_id)
rte_mempool_cache_free(cache)
rte_mempool_cache_flush(cache, mp)
rte_mempool_default_cache(mp, lcore_id)
Changes the API calls:
rte_mempool_generic_put(mp, obj_table, n, cache, flags)
rte_mempool_generic_get(mp, obj_table, n, cache, flags)
The cache-oblivious API calls use the per-lcore default local cache.
Signed-off-by: Lazaros Koromilas <l@nofutznetworks.com>
Acked-by: Olivier Matz <olivier.matz@6wind.com>
2016-06-28 23:47:38 +00:00
|
|
|
fprintf(f, " internal cache infos:\n");
|
2014-05-02 23:42:56 +00:00
|
|
|
fprintf(f, " cache_size=%"PRIu32"\n", mp->cache_size);
|
2016-04-14 09:42:36 +00:00
|
|
|
|
|
|
|
if (mp->cache_size == 0)
|
|
|
|
return count;
|
|
|
|
|
2012-09-04 12:54:00 +00:00
|
|
|
for (lcore_id = 0; lcore_id < RTE_MAX_LCORE; lcore_id++) {
|
|
|
|
cache_count = mp->local_cache[lcore_id].len;
|
mempool: allow user-owned cache
The mempool cache is only available to EAL threads as a per-lcore
resource. Change this so that the user can create and provide their own
cache on mempool get and put operations. This works with non-EAL threads
too. This commit introduces the new API calls:
rte_mempool_cache_create(size, socket_id)
rte_mempool_cache_free(cache)
rte_mempool_cache_flush(cache, mp)
rte_mempool_default_cache(mp, lcore_id)
Changes the API calls:
rte_mempool_generic_put(mp, obj_table, n, cache, flags)
rte_mempool_generic_get(mp, obj_table, n, cache, flags)
The cache-oblivious API calls use the per-lcore default local cache.
Signed-off-by: Lazaros Koromilas <l@nofutznetworks.com>
Acked-by: Olivier Matz <olivier.matz@6wind.com>
2016-06-28 23:47:38 +00:00
|
|
|
fprintf(f, " cache_count[%u]=%"PRIu32"\n",
|
|
|
|
lcore_id, cache_count);
|
2012-09-04 12:54:00 +00:00
|
|
|
count += cache_count;
|
|
|
|
}
|
2014-05-02 23:42:56 +00:00
|
|
|
fprintf(f, " total_cache_count=%u\n", count);
|
2012-09-04 12:54:00 +00:00
|
|
|
return count;
|
|
|
|
}
|
|
|
|
|
|
|
|
#ifndef __INTEL_COMPILER
|
|
|
|
#pragma GCC diagnostic ignored "-Wcast-qual"
|
|
|
|
#endif
|
2014-02-12 15:32:25 +00:00
|
|
|
|
2016-05-18 11:04:24 +00:00
|
|
|
/* check and update cookies or panic (internal) */
|
|
|
|
void rte_mempool_check_cookies(const struct rte_mempool *mp,
|
|
|
|
void * const *obj_table_const, unsigned n, int free)
|
|
|
|
{
|
|
|
|
#ifdef RTE_LIBRTE_MEMPOOL_DEBUG
|
|
|
|
struct rte_mempool_objhdr *hdr;
|
|
|
|
struct rte_mempool_objtlr *tlr;
|
|
|
|
uint64_t cookie;
|
|
|
|
void *tmp;
|
|
|
|
void *obj;
|
|
|
|
void **obj_table;
|
|
|
|
|
|
|
|
/* Force to drop the "const" attribute. This is done only when
|
|
|
|
* DEBUG is enabled */
|
|
|
|
tmp = (void *) obj_table_const;
|
2017-04-07 17:44:47 +00:00
|
|
|
obj_table = tmp;
|
2016-05-18 11:04:24 +00:00
|
|
|
|
|
|
|
while (n--) {
|
|
|
|
obj = obj_table[n];
|
|
|
|
|
|
|
|
if (rte_mempool_from_obj(obj) != mp)
|
|
|
|
rte_panic("MEMPOOL: object is owned by another "
|
|
|
|
"mempool\n");
|
|
|
|
|
|
|
|
hdr = __mempool_get_header(obj);
|
|
|
|
cookie = hdr->cookie;
|
|
|
|
|
|
|
|
if (free == 0) {
|
|
|
|
if (cookie != RTE_MEMPOOL_HEADER_COOKIE1) {
|
|
|
|
RTE_LOG(CRIT, MEMPOOL,
|
|
|
|
"obj=%p, mempool=%p, cookie=%" PRIx64 "\n",
|
|
|
|
obj, (const void *) mp, cookie);
|
|
|
|
rte_panic("MEMPOOL: bad header cookie (put)\n");
|
|
|
|
}
|
|
|
|
hdr->cookie = RTE_MEMPOOL_HEADER_COOKIE2;
|
|
|
|
} else if (free == 1) {
|
|
|
|
if (cookie != RTE_MEMPOOL_HEADER_COOKIE2) {
|
|
|
|
RTE_LOG(CRIT, MEMPOOL,
|
|
|
|
"obj=%p, mempool=%p, cookie=%" PRIx64 "\n",
|
|
|
|
obj, (const void *) mp, cookie);
|
|
|
|
rte_panic("MEMPOOL: bad header cookie (get)\n");
|
|
|
|
}
|
|
|
|
hdr->cookie = RTE_MEMPOOL_HEADER_COOKIE1;
|
|
|
|
} else if (free == 2) {
|
|
|
|
if (cookie != RTE_MEMPOOL_HEADER_COOKIE1 &&
|
|
|
|
cookie != RTE_MEMPOOL_HEADER_COOKIE2) {
|
|
|
|
RTE_LOG(CRIT, MEMPOOL,
|
|
|
|
"obj=%p, mempool=%p, cookie=%" PRIx64 "\n",
|
|
|
|
obj, (const void *) mp, cookie);
|
|
|
|
rte_panic("MEMPOOL: bad header cookie (audit)\n");
|
|
|
|
}
|
|
|
|
}
|
|
|
|
tlr = __mempool_get_trailer(obj);
|
|
|
|
cookie = tlr->cookie;
|
|
|
|
if (cookie != RTE_MEMPOOL_TRAILER_COOKIE) {
|
|
|
|
RTE_LOG(CRIT, MEMPOOL,
|
|
|
|
"obj=%p, mempool=%p, cookie=%" PRIx64 "\n",
|
|
|
|
obj, (const void *) mp, cookie);
|
|
|
|
rte_panic("MEMPOOL: bad trailer cookie\n");
|
|
|
|
}
|
|
|
|
}
|
|
|
|
#else
|
|
|
|
RTE_SET_USED(mp);
|
|
|
|
RTE_SET_USED(obj_table_const);
|
|
|
|
RTE_SET_USED(n);
|
|
|
|
RTE_SET_USED(free);
|
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
2018-04-26 10:59:21 +00:00
|
|
|
void
|
|
|
|
rte_mempool_contig_blocks_check_cookies(const struct rte_mempool *mp,
|
|
|
|
void * const *first_obj_table_const, unsigned int n, int free)
|
|
|
|
{
|
|
|
|
#ifdef RTE_LIBRTE_MEMPOOL_DEBUG
|
|
|
|
struct rte_mempool_info info;
|
|
|
|
const size_t total_elt_sz =
|
|
|
|
mp->header_size + mp->elt_size + mp->trailer_size;
|
|
|
|
unsigned int i, j;
|
|
|
|
|
|
|
|
rte_mempool_ops_get_info(mp, &info);
|
|
|
|
|
|
|
|
for (i = 0; i < n; ++i) {
|
|
|
|
void *first_obj = first_obj_table_const[i];
|
|
|
|
|
|
|
|
for (j = 0; j < info.contig_block_size; ++j) {
|
|
|
|
void *obj;
|
|
|
|
|
|
|
|
obj = (void *)((uintptr_t)first_obj + j * total_elt_sz);
|
|
|
|
rte_mempool_check_cookies(mp, &obj, 1, free);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
#else
|
|
|
|
RTE_SET_USED(mp);
|
|
|
|
RTE_SET_USED(first_obj_table_const);
|
|
|
|
RTE_SET_USED(n);
|
|
|
|
RTE_SET_USED(free);
|
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
2016-05-18 11:04:24 +00:00
|
|
|
#ifdef RTE_LIBRTE_MEMPOOL_DEBUG
|
2012-09-04 12:54:00 +00:00
|
|
|
static void
|
2016-05-18 11:04:31 +00:00
|
|
|
mempool_obj_audit(struct rte_mempool *mp, __rte_unused void *opaque,
|
|
|
|
void *obj, __rte_unused unsigned idx)
|
2012-09-04 12:54:00 +00:00
|
|
|
{
|
2016-05-18 11:04:31 +00:00
|
|
|
__mempool_check_cookies(mp, &obj, 1, 2);
|
2014-02-12 15:32:25 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
2016-05-18 11:04:29 +00:00
|
|
|
mempool_audit_cookies(struct rte_mempool *mp)
|
2014-02-12 15:32:25 +00:00
|
|
|
{
|
2016-05-18 11:04:31 +00:00
|
|
|
unsigned num;
|
2014-02-12 15:32:25 +00:00
|
|
|
|
2016-05-18 11:04:31 +00:00
|
|
|
num = rte_mempool_obj_iter(mp, mempool_obj_audit, NULL);
|
2014-02-12 15:32:25 +00:00
|
|
|
if (num != mp->size) {
|
2016-05-18 11:04:31 +00:00
|
|
|
rte_panic("rte_mempool_obj_iter(mempool=%p, size=%u) "
|
2014-02-12 15:32:25 +00:00
|
|
|
"iterated only over %u elements\n",
|
|
|
|
mp, mp->size, num);
|
2012-09-04 12:54:00 +00:00
|
|
|
}
|
|
|
|
}
|
2016-05-18 11:04:24 +00:00
|
|
|
#else
|
|
|
|
#define mempool_audit_cookies(mp) do {} while(0)
|
|
|
|
#endif
|
2014-02-12 15:32:25 +00:00
|
|
|
|
2012-09-04 12:54:00 +00:00
|
|
|
#ifndef __INTEL_COMPILER
|
|
|
|
#pragma GCC diagnostic error "-Wcast-qual"
|
|
|
|
#endif
|
|
|
|
|
|
|
|
/* check cookies before and after objects */
|
|
|
|
static void
|
|
|
|
mempool_audit_cache(const struct rte_mempool *mp)
|
|
|
|
{
|
|
|
|
/* check cache size consistency */
|
|
|
|
unsigned lcore_id;
|
2016-04-14 09:42:36 +00:00
|
|
|
|
|
|
|
if (mp->cache_size == 0)
|
|
|
|
return;
|
|
|
|
|
2012-09-04 12:54:00 +00:00
|
|
|
for (lcore_id = 0; lcore_id < RTE_MAX_LCORE; lcore_id++) {
|
mempool: allow user-owned cache
The mempool cache is only available to EAL threads as a per-lcore
resource. Change this so that the user can create and provide their own
cache on mempool get and put operations. This works with non-EAL threads
too. This commit introduces the new API calls:
rte_mempool_cache_create(size, socket_id)
rte_mempool_cache_free(cache)
rte_mempool_cache_flush(cache, mp)
rte_mempool_default_cache(mp, lcore_id)
Changes the API calls:
rte_mempool_generic_put(mp, obj_table, n, cache, flags)
rte_mempool_generic_get(mp, obj_table, n, cache, flags)
The cache-oblivious API calls use the per-lcore default local cache.
Signed-off-by: Lazaros Koromilas <l@nofutznetworks.com>
Acked-by: Olivier Matz <olivier.matz@6wind.com>
2016-06-28 23:47:38 +00:00
|
|
|
const struct rte_mempool_cache *cache;
|
|
|
|
cache = &mp->local_cache[lcore_id];
|
|
|
|
if (cache->len > cache->flushthresh) {
|
2012-09-04 12:54:00 +00:00
|
|
|
RTE_LOG(CRIT, MEMPOOL, "badness on cache[%u]\n",
|
|
|
|
lcore_id);
|
|
|
|
rte_panic("MEMPOOL: invalid cache len\n");
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/* check the consistency of mempool (size, cookies, ...) */
|
|
|
|
void
|
2016-05-18 11:04:29 +00:00
|
|
|
rte_mempool_audit(struct rte_mempool *mp)
|
2012-09-04 12:54:00 +00:00
|
|
|
{
|
|
|
|
mempool_audit_cache(mp);
|
|
|
|
mempool_audit_cookies(mp);
|
2012-12-19 23:00:00 +00:00
|
|
|
|
|
|
|
/* For case where mempool DEBUG is not set, and cache size is 0 */
|
|
|
|
RTE_SET_USED(mp);
|
2012-09-04 12:54:00 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/* dump the status of the mempool on the console */
|
|
|
|
void
|
2016-05-18 11:04:29 +00:00
|
|
|
rte_mempool_dump(FILE *f, struct rte_mempool *mp)
|
2012-09-04 12:54:00 +00:00
|
|
|
{
|
|
|
|
#ifdef RTE_LIBRTE_MEMPOOL_DEBUG
|
2018-04-26 10:59:21 +00:00
|
|
|
struct rte_mempool_info info;
|
2012-09-04 12:54:00 +00:00
|
|
|
struct rte_mempool_debug_stats sum;
|
|
|
|
unsigned lcore_id;
|
|
|
|
#endif
|
2016-05-18 11:04:36 +00:00
|
|
|
struct rte_mempool_memhdr *memhdr;
|
2012-09-04 12:54:00 +00:00
|
|
|
unsigned common_count;
|
|
|
|
unsigned cache_count;
|
2016-05-18 11:04:36 +00:00
|
|
|
size_t mem_len = 0;
|
2012-09-04 12:54:00 +00:00
|
|
|
|
2016-04-22 12:21:26 +00:00
|
|
|
RTE_ASSERT(f != NULL);
|
|
|
|
RTE_ASSERT(mp != NULL);
|
2014-09-28 05:28:44 +00:00
|
|
|
|
2014-05-02 23:42:56 +00:00
|
|
|
fprintf(f, "mempool <%s>@%p\n", mp->name, mp);
|
|
|
|
fprintf(f, " flags=%x\n", mp->flags);
|
2016-06-22 09:27:27 +00:00
|
|
|
fprintf(f, " pool=%p\n", mp->pool_data);
|
2017-11-04 01:22:28 +00:00
|
|
|
fprintf(f, " iova=0x%" PRIx64 "\n", mp->mz->iova);
|
2016-05-18 11:04:36 +00:00
|
|
|
fprintf(f, " nb_mem_chunks=%u\n", mp->nb_mem_chunks);
|
2014-05-02 23:42:56 +00:00
|
|
|
fprintf(f, " size=%"PRIu32"\n", mp->size);
|
2016-05-18 11:04:36 +00:00
|
|
|
fprintf(f, " populated_size=%"PRIu32"\n", mp->populated_size);
|
2014-05-02 23:42:56 +00:00
|
|
|
fprintf(f, " header_size=%"PRIu32"\n", mp->header_size);
|
|
|
|
fprintf(f, " elt_size=%"PRIu32"\n", mp->elt_size);
|
|
|
|
fprintf(f, " trailer_size=%"PRIu32"\n", mp->trailer_size);
|
|
|
|
fprintf(f, " total_obj_size=%"PRIu32"\n",
|
2012-09-04 12:54:00 +00:00
|
|
|
mp->header_size + mp->elt_size + mp->trailer_size);
|
|
|
|
|
2014-05-02 23:42:56 +00:00
|
|
|
fprintf(f, " private_data_size=%"PRIu32"\n", mp->private_data_size);
|
2016-05-18 11:04:36 +00:00
|
|
|
|
|
|
|
STAILQ_FOREACH(memhdr, &mp->mem_list, next)
|
|
|
|
mem_len += memhdr->len;
|
|
|
|
if (mem_len != 0) {
|
2014-05-02 23:42:56 +00:00
|
|
|
fprintf(f, " avg bytes/object=%#Lf\n",
|
2016-05-18 11:04:36 +00:00
|
|
|
(long double)mem_len / mp->size);
|
|
|
|
}
|
2014-02-12 15:32:25 +00:00
|
|
|
|
2014-05-02 23:42:56 +00:00
|
|
|
cache_count = rte_mempool_dump_cache(f, mp);
|
2016-06-22 09:27:27 +00:00
|
|
|
common_count = rte_mempool_ops_get_count(mp);
|
2012-09-04 12:54:00 +00:00
|
|
|
if ((cache_count + common_count) > mp->size)
|
|
|
|
common_count = mp->size - cache_count;
|
2014-05-02 23:42:56 +00:00
|
|
|
fprintf(f, " common_pool_count=%u\n", common_count);
|
2012-09-04 12:54:00 +00:00
|
|
|
|
|
|
|
/* sum and dump statistics */
|
|
|
|
#ifdef RTE_LIBRTE_MEMPOOL_DEBUG
|
2018-04-26 10:59:21 +00:00
|
|
|
rte_mempool_ops_get_info(mp, &info);
|
2012-09-04 12:54:00 +00:00
|
|
|
memset(&sum, 0, sizeof(sum));
|
|
|
|
for (lcore_id = 0; lcore_id < RTE_MAX_LCORE; lcore_id++) {
|
|
|
|
sum.put_bulk += mp->stats[lcore_id].put_bulk;
|
|
|
|
sum.put_objs += mp->stats[lcore_id].put_objs;
|
|
|
|
sum.get_success_bulk += mp->stats[lcore_id].get_success_bulk;
|
|
|
|
sum.get_success_objs += mp->stats[lcore_id].get_success_objs;
|
|
|
|
sum.get_fail_bulk += mp->stats[lcore_id].get_fail_bulk;
|
|
|
|
sum.get_fail_objs += mp->stats[lcore_id].get_fail_objs;
|
2018-04-26 10:59:21 +00:00
|
|
|
sum.get_success_blks += mp->stats[lcore_id].get_success_blks;
|
|
|
|
sum.get_fail_blks += mp->stats[lcore_id].get_fail_blks;
|
2012-09-04 12:54:00 +00:00
|
|
|
}
|
2014-05-02 23:42:56 +00:00
|
|
|
fprintf(f, " stats:\n");
|
|
|
|
fprintf(f, " put_bulk=%"PRIu64"\n", sum.put_bulk);
|
|
|
|
fprintf(f, " put_objs=%"PRIu64"\n", sum.put_objs);
|
|
|
|
fprintf(f, " get_success_bulk=%"PRIu64"\n", sum.get_success_bulk);
|
|
|
|
fprintf(f, " get_success_objs=%"PRIu64"\n", sum.get_success_objs);
|
|
|
|
fprintf(f, " get_fail_bulk=%"PRIu64"\n", sum.get_fail_bulk);
|
|
|
|
fprintf(f, " get_fail_objs=%"PRIu64"\n", sum.get_fail_objs);
|
2018-04-26 10:59:21 +00:00
|
|
|
if (info.contig_block_size > 0) {
|
|
|
|
fprintf(f, " get_success_blks=%"PRIu64"\n",
|
|
|
|
sum.get_success_blks);
|
|
|
|
fprintf(f, " get_fail_blks=%"PRIu64"\n", sum.get_fail_blks);
|
|
|
|
}
|
2012-09-04 12:54:00 +00:00
|
|
|
#else
|
2014-05-02 23:42:56 +00:00
|
|
|
fprintf(f, " no statistics available\n");
|
2012-09-04 12:54:00 +00:00
|
|
|
#endif
|
|
|
|
|
|
|
|
rte_mempool_audit(mp);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* dump the status of all mempools on the console */
|
|
|
|
void
|
2014-05-02 23:42:56 +00:00
|
|
|
rte_mempool_list_dump(FILE *f)
|
2012-09-04 12:54:00 +00:00
|
|
|
{
|
2016-05-18 11:04:29 +00:00
|
|
|
struct rte_mempool *mp = NULL;
|
2014-06-20 15:42:22 +00:00
|
|
|
struct rte_tailq_entry *te;
|
2012-12-19 23:00:00 +00:00
|
|
|
struct rte_mempool_list *mempool_list;
|
|
|
|
|
2015-03-04 21:50:08 +00:00
|
|
|
mempool_list = RTE_TAILQ_CAST(rte_mempool_tailq.head, rte_mempool_list);
|
2012-09-04 12:54:00 +00:00
|
|
|
|
2019-07-05 13:10:29 +00:00
|
|
|
rte_mcfg_mempool_read_lock();
|
2012-12-19 23:00:00 +00:00
|
|
|
|
2014-06-20 15:42:22 +00:00
|
|
|
TAILQ_FOREACH(te, mempool_list, next) {
|
|
|
|
mp = (struct rte_mempool *) te->data;
|
2014-05-02 23:42:56 +00:00
|
|
|
rte_mempool_dump(f, mp);
|
2012-09-04 12:54:00 +00:00
|
|
|
}
|
2012-12-19 23:00:00 +00:00
|
|
|
|
2019-07-05 13:10:29 +00:00
|
|
|
rte_mcfg_mempool_read_unlock();
|
2012-09-04 12:54:00 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/* search a mempool from its name */
|
|
|
|
struct rte_mempool *
|
|
|
|
rte_mempool_lookup(const char *name)
|
|
|
|
{
|
|
|
|
struct rte_mempool *mp = NULL;
|
2014-06-20 15:42:22 +00:00
|
|
|
struct rte_tailq_entry *te;
|
2012-12-19 23:00:00 +00:00
|
|
|
struct rte_mempool_list *mempool_list;
|
2012-09-04 12:54:00 +00:00
|
|
|
|
2015-03-04 21:50:08 +00:00
|
|
|
mempool_list = RTE_TAILQ_CAST(rte_mempool_tailq.head, rte_mempool_list);
|
2012-09-04 12:54:00 +00:00
|
|
|
|
2019-07-05 13:10:29 +00:00
|
|
|
rte_mcfg_mempool_read_lock();
|
2012-12-19 23:00:00 +00:00
|
|
|
|
2014-06-20 15:42:22 +00:00
|
|
|
TAILQ_FOREACH(te, mempool_list, next) {
|
|
|
|
mp = (struct rte_mempool *) te->data;
|
2012-09-04 12:54:00 +00:00
|
|
|
if (strncmp(name, mp->name, RTE_MEMPOOL_NAMESIZE) == 0)
|
|
|
|
break;
|
|
|
|
}
|
2014-06-03 23:42:50 +00:00
|
|
|
|
2019-07-05 13:10:29 +00:00
|
|
|
rte_mcfg_mempool_read_unlock();
|
2014-06-03 23:42:50 +00:00
|
|
|
|
2014-06-20 15:42:22 +00:00
|
|
|
if (te == NULL) {
|
2012-09-04 12:54:00 +00:00
|
|
|
rte_errno = ENOENT;
|
2014-06-20 15:42:22 +00:00
|
|
|
return NULL;
|
|
|
|
}
|
2012-09-04 12:54:00 +00:00
|
|
|
|
|
|
|
return mp;
|
|
|
|
}
|
2014-05-02 23:42:54 +00:00
|
|
|
|
2016-05-18 11:04:28 +00:00
|
|
|
void rte_mempool_walk(void (*func)(struct rte_mempool *, void *),
|
2014-05-02 23:42:54 +00:00
|
|
|
void *arg)
|
|
|
|
{
|
2014-06-20 15:42:22 +00:00
|
|
|
struct rte_tailq_entry *te = NULL;
|
2014-05-02 23:42:54 +00:00
|
|
|
struct rte_mempool_list *mempool_list;
|
2016-07-25 19:32:03 +00:00
|
|
|
void *tmp_te;
|
2014-05-02 23:42:54 +00:00
|
|
|
|
2015-03-04 21:50:08 +00:00
|
|
|
mempool_list = RTE_TAILQ_CAST(rte_mempool_tailq.head, rte_mempool_list);
|
2014-05-02 23:42:54 +00:00
|
|
|
|
2019-07-05 13:10:29 +00:00
|
|
|
rte_mcfg_mempool_read_lock();
|
2014-05-02 23:42:54 +00:00
|
|
|
|
2016-07-25 19:32:03 +00:00
|
|
|
TAILQ_FOREACH_SAFE(te, mempool_list, next, tmp_te) {
|
2014-06-20 15:42:22 +00:00
|
|
|
(*func)((struct rte_mempool *) te->data, arg);
|
2014-05-02 23:42:54 +00:00
|
|
|
}
|
|
|
|
|
2019-07-05 13:10:29 +00:00
|
|
|
rte_mcfg_mempool_read_unlock();
|
2014-05-02 23:42:54 +00:00
|
|
|
}
|