- Support separate memory pools for malloc() and malloc_dma() - Default to the old behaviour of a shared static pool - Add API for switching the memory pool for DMA allocations - Hide (alloc|free)_memblock() as internal - Uninline (free|malloc)_dma() for size reasons
This adds 47 bytes to malloc. Signed-off-by: Piotr Jaroszyński <p.jaroszyn...@gmail.com> --- src/core/malloc.c | 76 +++++++++++++++++++++++++++++++------------- src/include/gpxe/malloc.h | 36 +++++++++++---------- 2 files changed, 72 insertions(+), 40 deletions(-) diff --git a/src/core/malloc.c b/src/core/malloc.c index df09265..b70e70d 100644 --- a/src/core/malloc.c +++ b/src/core/malloc.c @@ -71,8 +71,12 @@ struct autosized_block { */ #define NOWHERE ( ( void * ) ~( ( intptr_t ) 0 ) ) -/** List of free memory blocks */ -static LIST_HEAD ( free_blocks ); +static struct memory_pool base_pool = { + .free_blocks = LIST_HEAD_INIT ( base_pool.free_blocks ), +}; + +static struct memory_pool * normal_pool = &base_pool; +static struct memory_pool * dma_pool = &base_pool; /** Total amount of free memory */ size_t freemem; @@ -88,28 +92,36 @@ size_t freemem; static char heap[HEAP_SIZE] __attribute__ (( aligned ( __alignof__(void *) ))); -static inline void valgrind_make_blocks_defined ( void ) +static inline void valgrind_make_blocks_defined ( struct memory_pool *pool ) { struct memory_block *block; if (RUNNING_ON_VALGRIND > 0) { - VALGRIND_MAKE_MEM_DEFINED ( &free_blocks, sizeof ( free_blocks ) ); - list_for_each_entry ( block, &free_blocks, list ) + VALGRIND_MAKE_MEM_DEFINED ( &pool->free_blocks, sizeof ( pool->free_blocks ) ); + list_for_each_entry ( block, &pool->free_blocks, list ) VALGRIND_MAKE_MEM_DEFINED ( block, sizeof ( *block ) ); } } -static inline void valgrind_make_blocks_noaccess ( void ) +static inline void valgrind_make_blocks_noaccess ( struct memory_pool *pool ) { struct memory_block *block, *tmp; if ( RUNNING_ON_VALGRIND > 0 ) { - list_for_each_entry_safe ( block, tmp, &free_blocks, list ) + list_for_each_entry_safe ( block, tmp, &pool->free_blocks, list ) VALGRIND_MAKE_MEM_NOACCESS ( block, sizeof ( *block ) ); - VALGRIND_MAKE_MEM_NOACCESS ( &free_blocks, sizeof ( free_blocks ) ); + VALGRIND_MAKE_MEM_NOACCESS ( &pool->free_blocks, sizeof ( pool->free_blocks ) ); } } +struct memory_pool * set_dma_pool(struct memory_pool *new_pool) +{ + struct memory_pool * old_pool = dma_pool; + dma_pool = new_pool; + return old_pool; +} + + /** * Discard some cached data * @@ -128,6 +140,7 @@ static unsigned int discard_cache ( void ) { /** * Allocate a memory block * + * @v pool Memory pool to allocate from * @v size Requested size * @v align Physical alignment * @ret ptr Memory block, or NULL @@ -137,7 +150,7 @@ static unsigned int discard_cache ( void ) { * * @c align must be a power of two. @c size may not be zero. */ -void * alloc_memblock ( size_t size, size_t align ) { +void * alloc_memblock ( struct memory_pool *pool, size_t size, size_t align ) { struct memory_block *block; size_t align_mask; size_t pre_size; @@ -146,7 +159,7 @@ void * alloc_memblock ( size_t size, size_t align ) { struct memory_block *post; struct memory_block *ret = NULL; - valgrind_make_blocks_defined ( ); + valgrind_make_blocks_defined ( pool ); /* Round up size to multiple of MIN_MEMBLOCK_SIZE and * calculate alignment mask. @@ -157,7 +170,7 @@ void * alloc_memblock ( size_t size, size_t align ) { DBG ( "Allocating %#zx (aligned %#zx)\n", size, align ); while ( 1 ) { /* Search through blocks for the first one with enough space */ - list_for_each_entry ( block, &free_blocks, list ) { + list_for_each_entry ( block, &pool->free_blocks, list ) { pre_size = ( - virt_to_phys ( block ) ) & align_mask; post_size = block->size - pre_size - size; if ( post_size >= 0 ) { @@ -216,19 +229,21 @@ void * alloc_memblock ( size_t size, size_t align ) { } cleanup: - valgrind_make_blocks_noaccess ( ); + valgrind_make_blocks_noaccess ( pool ); + return ret; } /** * Free a memory block * + * @v pool Memory pool the block was allocated from * @v ptr Memory allocated by alloc_memblock(), or NULL * @v size Size of the memory * * If @c ptr is NULL, no action is taken. */ -void free_memblock ( void *ptr, size_t size ) { +void free_memblock ( struct memory_pool *pool, void *ptr, size_t size ) { struct memory_block *freeing; struct memory_block *block; ssize_t gap_before; @@ -238,7 +253,7 @@ void free_memblock ( void *ptr, size_t size ) { if ( ! ptr ) return; - valgrind_make_blocks_defined ( ); + valgrind_make_blocks_defined ( pool ); /* Round up size to match actual size that alloc_memblock() * would have used. @@ -250,7 +265,7 @@ void free_memblock ( void *ptr, size_t size ) { DBG ( "Freeing [%p,%p)\n", freeing, ( ( ( void * ) freeing ) + size )); /* Insert/merge into free list */ - list_for_each_entry ( block, &free_blocks, list ) { + list_for_each_entry ( block, &pool->free_blocks, list ) { /* Calculate gaps before and after the "freeing" block */ gap_before = ( ( ( void * ) freeing ) - ( ( ( void * ) block ) + block->size ) ); @@ -289,7 +304,7 @@ void free_memblock ( void *ptr, size_t size ) { /* Update free memory counter */ freemem += size; - valgrind_make_blocks_noaccess ( ); + valgrind_make_blocks_noaccess ( pool ); } /** @@ -326,7 +341,7 @@ void * realloc ( void *old_ptr, size_t new_size ) { if ( new_size ) { new_total_size = ( new_size + offsetof ( struct autosized_block, data ) ); - new_block = alloc_memblock ( new_total_size, 1 ); + new_block = alloc_memblock ( normal_pool, new_total_size, 1 ); if ( ! new_block ) return NULL; VALGRIND_MAKE_MEM_UNDEFINED ( new_block, offsetof ( struct autosized_block, data ) ); @@ -350,7 +365,7 @@ void * realloc ( void *old_ptr, size_t new_size ) { offsetof ( struct autosized_block, data ) ); memcpy ( new_ptr, old_ptr, ( ( old_size < new_size ) ? old_size : new_size ) ); - free_memblock ( old_block, old_total_size ); + free_memblock ( normal_pool, old_block, old_total_size ); VALGRIND_MAKE_MEM_NOACCESS ( old_block, offsetof ( struct autosized_block, data ) ); VALGRIND_FREELIKE_BLOCK ( old_ptr, 0 ); } @@ -405,6 +420,21 @@ void * zalloc ( size_t size ) { return data; } +void free_dma ( void *ptr, size_t size ) { + free_memblock ( dma_pool, ptr, size ); + VALGRIND_FREELIKE_BLOCK ( ptr, 0 ); +} + +void * __malloc malloc_dma ( size_t size, size_t phys_align ) { + void * ptr = alloc_memblock ( dma_pool, size, phys_align ); + if ( ptr && size ) { + VALGRIND_MALLOCLIKE_BLOCK ( ptr, size, 0, 0 ); + /* DMA memory can be modified by hardware w/o valgrind noticing */ + VALGRIND_MAKE_MEM_DEFINED ( ptr, size ); + } + return ptr; +} + /** * Add memory to allocation pool * @@ -416,11 +446,11 @@ void * zalloc ( size_t size ) { * * @c start must be aligned to at least a multiple of sizeof(void*). */ -void mpopulate ( void *start, size_t len ) { +void mpopulate ( struct memory_pool *pool, void *start, size_t len ) { /* Prevent free_memblock() from rounding up len beyond the end * of what we were actually given... */ - free_memblock ( start, ( len & ~( MIN_MEMBLOCK_SIZE - 1 ) ) ); + free_memblock ( pool, start, ( len & ~( MIN_MEMBLOCK_SIZE - 1 ) ) ); } /** @@ -429,7 +459,7 @@ void mpopulate ( void *start, size_t len ) { */ static void init_heap ( void ) { VALGRIND_MAKE_MEM_NOACCESS ( heap, sizeof ( heap ) ); - mpopulate ( heap, sizeof ( heap ) ); + mpopulate ( normal_pool, heap, sizeof ( heap ) ); } /** Memory allocator initialisation function */ @@ -443,11 +473,11 @@ struct init_fn heap_init_fn __init_fn ( INIT_EARLY ) = { * Dump free block list * */ -void mdumpfree ( void ) { +void mdumpfree ( struct memory_pool *pool ) { struct memory_block *block; printf ( "Free block list:\n" ); - list_for_each_entry ( block, &free_blocks, list ) { + list_for_each_entry ( block, &pool->free_blocks, list ) { printf ( "[%p,%p] (size %#zx)\n", block, ( ( ( void * ) block ) + block->size ), block->size ); } diff --git a/src/include/gpxe/malloc.h b/src/include/gpxe/malloc.h index 5effc98..a9fd927 100644 --- a/src/include/gpxe/malloc.h +++ b/src/include/gpxe/malloc.h @@ -19,15 +19,28 @@ FILE_LICENCE ( GPL2_OR_LATER ); */ #include <stdlib.h> #include <gpxe/tables.h> +#include <gpxe/list.h> #include <valgrind/memcheck.h> +/** A memory pool */ +struct memory_pool { + /** List of free memory blocks */ + struct list_head free_blocks; +}; + extern size_t freemem; -extern void * __malloc alloc_memblock ( size_t size, size_t align ); -extern void free_memblock ( void *ptr, size_t size ); -extern void mpopulate ( void *start, size_t len ); -extern void mdumpfree ( void ); +/** + * Set a new memory pool for DMA allocations. + * + * @v new_pool New DMA pool + * @ret old_pool Old DMA pool + */ +extern struct memory_pool *set_dma_pool ( struct memory_pool *new_pool ); + +extern void mpopulate ( struct memory_pool *pool, void *start, size_t len ); +extern void mdumpfree ( struct memory_pool *pool ); /** * Allocate memory for DMA @@ -40,15 +53,7 @@ extern void mdumpfree ( void ); * * @c align must be a power of two. @c size may not be zero. */ -static inline void * __malloc malloc_dma ( size_t size, size_t phys_align ) { - void * ptr = alloc_memblock ( size, phys_align ); - if ( ptr && size ) { - VALGRIND_MALLOCLIKE_BLOCK ( ptr, size, 0, 0 ); - /* DMA memory can be modified by hardware w/o valgrind noticing */ - VALGRIND_MAKE_MEM_DEFINED ( ptr, size ); - } - return ptr; -} +extern void * __malloc malloc_dma ( size_t size, size_t phys_align ); /** * Free memory allocated with malloc_dma() @@ -61,10 +66,7 @@ static inline void * __malloc malloc_dma ( size_t size, size_t phys_align ) { * * If @c ptr is NULL, no action is taken. */ -static inline void free_dma ( void *ptr, size_t size ) { - free_memblock ( ptr, size ); - VALGRIND_FREELIKE_BLOCK ( ptr, 0 ); -} +extern void free_dma ( void *ptr, size_t size ); /** A cache discarder */ struct cache_discarder { -- 1.7.1 _______________________________________________ gPXE-devel mailing list gPXE-devel@etherboot.org http://etherboot.org/mailman/listinfo/gpxe-devel