b0489e7bca
Problems with lib rte_malloc: 1. Rte_malloc searches a heap's entire free list looking for the best fit, resulting in linear complexity. 2. Heaps store free blocks in a singly-linked list, resulting in linear complexity when rte_free needs to remove an adjacent block. 3. The library inserts and removes free blocks with ad hoc, in-line code, rather than using linked-list functions or macros. 4. The library wastes potential small blocks of size 64 and 128 bytes (plus overhead of 64 bytes) as padding when reusing free blocks or resizing allocated blocks. This patch addresses those problems as follows: 1. Replace single free list with a handful of free lists. Each free list contains blocks of a specified size range, for example: list[0]: (0 , 2^8] list[1]: (2^8 , 2^10] list[2]: (2^10, 2^12] list[3]: (2^12, 2^14] list[4]: (2^14, MAX_SIZE] When allocating a block, start at the first list that can contain a big enough block. Search subsequent lists, if necessary. Terminate the search as soon as we find a block that is big enough. 2. Use doubly-linked lists, so that we can remove free blocks in constant time. 3. Use BSD LIST macros, as defined in sys/queue.h and the QUEUE(3) man page. 4. Change code to utilize small blocks of data size 64 and 128, when splitting larger blocks. Signed-off-by: Robert Sanford <rsanford2@gmail.com> Acked-by: Pablo de Lara <pablo.de.lara.guarch@intel.com>
189 lines
5.9 KiB
C
189 lines
5.9 KiB
C
/*-
|
|
* BSD LICENSE
|
|
*
|
|
* Copyright(c) 2010-2014 Intel Corporation. All rights reserved.
|
|
* All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
*
|
|
* * Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* * Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in
|
|
* the documentation and/or other materials provided with the
|
|
* distribution.
|
|
* * Neither the name of Intel Corporation nor the names of its
|
|
* contributors may be used to endorse or promote products derived
|
|
* from this software without specific prior written permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
|
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
|
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
|
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
|
* OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
|
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
|
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
|
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
|
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
|
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
*/
|
|
|
|
#ifndef MALLOC_ELEM_H_
|
|
#define MALLOC_ELEM_H_
|
|
|
|
/* dummy definition of struct so we can use pointers to it in malloc_elem struct */
|
|
struct malloc_heap;
|
|
|
|
enum elem_state {
|
|
ELEM_FREE = 0,
|
|
ELEM_BUSY,
|
|
ELEM_PAD /* element is a padding-only header */
|
|
};
|
|
|
|
struct malloc_elem {
|
|
struct malloc_heap *heap;
|
|
struct malloc_elem *volatile prev; /* points to prev elem in memzone */
|
|
LIST_ENTRY(malloc_elem) free_list; /* list of free elements in heap */
|
|
const struct rte_memzone *mz;
|
|
volatile enum elem_state state;
|
|
uint32_t pad;
|
|
size_t size;
|
|
#ifdef RTE_LIBRTE_MALLOC_DEBUG
|
|
uint64_t header_cookie; /* Cookie marking start of data */
|
|
/* trailer cookie at start + size */
|
|
#endif
|
|
} __rte_cache_aligned;
|
|
|
|
#ifndef RTE_LIBRTE_MALLOC_DEBUG
|
|
static const unsigned MALLOC_ELEM_TRAILER_LEN = 0;
|
|
|
|
/* dummy function - just check if pointer is non-null */
|
|
static inline int
|
|
malloc_elem_cookies_ok(const struct malloc_elem *elem){ return elem != NULL; }
|
|
|
|
/* dummy function - no header if malloc_debug is not enabled */
|
|
static inline void
|
|
set_header(struct malloc_elem *elem __rte_unused){ }
|
|
|
|
/* dummy function - no trailer if malloc_debug is not enabled */
|
|
static inline void
|
|
set_trailer(struct malloc_elem *elem __rte_unused){ }
|
|
|
|
|
|
#else
|
|
static const unsigned MALLOC_ELEM_TRAILER_LEN = CACHE_LINE_SIZE;
|
|
|
|
#define MALLOC_HEADER_COOKIE 0xbadbadbadadd2e55ULL /**< Header cookie. */
|
|
#define MALLOC_TRAILER_COOKIE 0xadd2e55badbadbadULL /**< Trailer cookie.*/
|
|
|
|
/* define macros to make referencing the header and trailer cookies easier */
|
|
#define MALLOC_ELEM_TRAILER(elem) (*((uint64_t*)RTE_PTR_ADD(elem, \
|
|
elem->size - MALLOC_ELEM_TRAILER_LEN)))
|
|
#define MALLOC_ELEM_HEADER(elem) (elem->header_cookie)
|
|
|
|
static inline void
|
|
set_header(struct malloc_elem *elem)
|
|
{
|
|
if (elem != NULL)
|
|
MALLOC_ELEM_HEADER(elem) = MALLOC_HEADER_COOKIE;
|
|
}
|
|
|
|
static inline void
|
|
set_trailer(struct malloc_elem *elem)
|
|
{
|
|
if (elem != NULL)
|
|
MALLOC_ELEM_TRAILER(elem) = MALLOC_TRAILER_COOKIE;
|
|
}
|
|
|
|
/* check that the header and trailer cookies are set correctly */
|
|
static inline int
|
|
malloc_elem_cookies_ok(const struct malloc_elem *elem)
|
|
{
|
|
return (elem != NULL &&
|
|
MALLOC_ELEM_HEADER(elem) == MALLOC_HEADER_COOKIE &&
|
|
MALLOC_ELEM_TRAILER(elem) == MALLOC_TRAILER_COOKIE);
|
|
}
|
|
|
|
#endif
|
|
|
|
static const unsigned MALLOC_ELEM_HEADER_LEN = sizeof(struct malloc_elem);
|
|
#define MALLOC_ELEM_OVERHEAD (MALLOC_ELEM_HEADER_LEN + MALLOC_ELEM_TRAILER_LEN)
|
|
|
|
/*
|
|
* Given a pointer to the start of a memory block returned by malloc, get
|
|
* the actual malloc_elem header for that block.
|
|
*/
|
|
static inline struct malloc_elem *
|
|
malloc_elem_from_data(const void *data)
|
|
{
|
|
if (data == NULL)
|
|
return NULL;
|
|
|
|
struct malloc_elem *elem = RTE_PTR_SUB(data, MALLOC_ELEM_HEADER_LEN);
|
|
if (!malloc_elem_cookies_ok(elem))
|
|
return NULL;
|
|
return elem->state != ELEM_PAD ? elem: RTE_PTR_SUB(elem, elem->pad);
|
|
}
|
|
|
|
/*
|
|
* initialise a malloc_elem header
|
|
*/
|
|
void
|
|
malloc_elem_init(struct malloc_elem *elem,
|
|
struct malloc_heap *heap,
|
|
const struct rte_memzone *mz,
|
|
size_t size);
|
|
|
|
/*
|
|
* initialise a dummy malloc_elem header for the end-of-memzone marker
|
|
*/
|
|
void
|
|
malloc_elem_mkend(struct malloc_elem *elem,
|
|
struct malloc_elem *prev_free);
|
|
|
|
/*
|
|
* return true if the current malloc_elem can hold a block of data
|
|
* of the requested size and with the requested alignment
|
|
*/
|
|
int
|
|
malloc_elem_can_hold(struct malloc_elem *elem, size_t size, unsigned align);
|
|
|
|
/*
|
|
* reserve a block of data in an existing malloc_elem. If the malloc_elem
|
|
* is much larger than the data block requested, we split the element in two.
|
|
*/
|
|
struct malloc_elem *
|
|
malloc_elem_alloc(struct malloc_elem *elem, size_t size, unsigned align);
|
|
|
|
/*
|
|
* free a malloc_elem block by adding it to the free list. If the
|
|
* blocks either immediately before or immediately after newly freed block
|
|
* are also free, the blocks are merged together.
|
|
*/
|
|
int
|
|
malloc_elem_free(struct malloc_elem *elem);
|
|
|
|
/*
|
|
* attempt to resize a malloc_elem by expanding into any free space
|
|
* immediately after it in memory.
|
|
*/
|
|
int
|
|
malloc_elem_resize(struct malloc_elem *elem, size_t size);
|
|
|
|
/*
|
|
* Given an element size, compute its freelist index.
|
|
*/
|
|
size_t
|
|
malloc_elem_free_list_index(size_t size);
|
|
|
|
/*
|
|
* Add element to its heap's free list.
|
|
*/
|
|
void
|
|
malloc_elem_free_list_insert(struct malloc_elem *elem);
|
|
|
|
#endif /* MALLOC_ELEM_H_ */
|