mm: free delay list when exceeding specified count

Signed-off-by: Xu Xingliang <xuxingliang@xiaomi.com>
This commit is contained in:
Xu Xingliang 2023-08-29 17:52:23 +08:00 committed by Petro Karashchenko
parent 6dabcd567e
commit 6e7115ca09
6 changed files with 316 additions and 86 deletions

View file

@ -25,6 +25,7 @@
#include <nuttx/config.h>
#include <assert.h>
#include <debug.h>
#include <string.h>
#include <stdatomic.h>
#include <stdbool.h>
@ -51,6 +52,11 @@ struct mm_delaynode_s
struct mm_heap_s
{
struct mm_delaynode_s *mm_delaylist[CONFIG_SMP_NCPUS];
#if CONFIG_MM_FREE_DELAYCOUNT_MAX > 0
size_t mm_delaycount[CONFIG_SMP_NCPUS];
#endif
atomic_int aordblks;
atomic_int uordblks;
atomic_int usmblks;
@ -60,6 +66,12 @@ struct mm_heap_s
#endif
};
/****************************************************************************
* Private Function Prototypes
****************************************************************************/
static void mm_delayfree(struct mm_heap_s *heap, void *mem, bool delay);
/****************************************************************************
* Private Functions
****************************************************************************/
@ -77,12 +89,17 @@ static void mm_add_delaylist(struct mm_heap_s *heap, void *mem)
tmp->flink = heap->mm_delaylist[up_cpu_index()];
heap->mm_delaylist[up_cpu_index()] = tmp;
#if CONFIG_MM_FREE_DELAYCOUNT_MAX > 0
heap->mm_delaycount[up_cpu_index()]++;
#endif
up_irq_restore(flags);
#endif
}
static void mm_free_delaylist(struct mm_heap_s *heap)
static bool mm_free_delaylist(struct mm_heap_s *heap, bool force)
{
bool ret = false;
#if defined(CONFIG_BUILD_FLAT) || defined(__KERNEL__)
struct mm_delaynode_s *tmp;
irqstate_t flags;
@ -92,12 +109,26 @@ static void mm_free_delaylist(struct mm_heap_s *heap)
flags = up_irq_save();
tmp = heap->mm_delaylist[up_cpu_index()];
#if CONFIG_MM_FREE_DELAYCOUNT_MAX > 0
if (tmp == NULL ||
(!force &&
heap->mm_delaycount[up_cpu_index()] < CONFIG_MM_FREE_DELAYCOUNT_MAX))
{
up_irq_restore(flags);
return false;
}
heap->mm_delaycount[up_cpu_index()] = 0;
#endif
heap->mm_delaylist[up_cpu_index()] = NULL;
up_irq_restore(flags);
/* Test if the delayed is empty */
ret = tmp != NULL;
while (tmp)
{
void *address;
@ -111,9 +142,51 @@ static void mm_free_delaylist(struct mm_heap_s *heap)
* 'while' condition above.
*/
mm_free(heap, address);
mm_delayfree(heap, address, false);
}
#endif
return ret;
}
/****************************************************************************
* Name: mm_delayfree
*
* Description:
* Delay free memory if `delay` is true, otherwise free it immediately.
*
****************************************************************************/
static void mm_delayfree(struct mm_heap_s *heap, void *mem, bool delay)
{
#if defined(CONFIG_BUILD_FLAT) || defined(__KERNEL__)
/* Check current environment */
if (up_interrupt_context())
{
/* We are in ISR, add to the delay list */
mm_add_delaylist(heap, mem);
}
else
#endif
if (nxsched_gettid() < 0 || delay)
{
/* nxsched_gettid() return -ESRCH, means we are in situations
* during context switching(See nxsched_gettid's comment).
* Then add to the delay list.
*/
mm_add_delaylist(heap, mem);
}
else
{
int size = host_mallocsize(mem);
atomic_fetch_sub(&heap->aordblks, 1);
atomic_fetch_sub(&heap->uordblks, size);
host_free(mem);
}
}
/****************************************************************************
@ -208,34 +281,16 @@ void *mm_malloc(struct mm_heap_s *heap, size_t size)
void mm_free(struct mm_heap_s *heap, void *mem)
{
#if defined(CONFIG_BUILD_FLAT) || defined(__KERNEL__)
/* Check current environment */
minfo("Freeing %p\n", mem);
if (up_interrupt_context())
/* Protect against attempts to free a NULL reference */
if (mem == NULL)
{
/* We are in ISR, add to the delay list */
mm_add_delaylist(heap, mem);
return;
}
else
#endif
if (nxsched_gettid() < 0)
{
/* nxsched_gettid() return -ESRCH, means we are in situations
* during context switching(See nxsched_gettid's comment).
* Then add to the delay list.
*/
mm_add_delaylist(heap, mem);
}
else
{
int size = host_mallocsize(mem);
atomic_fetch_sub(&heap->aordblks, 1);
atomic_fetch_sub(&heap->uordblks, size);
host_free(mem);
}
mm_delayfree(heap, mem, CONFIG_MM_FREE_DELAYCOUNT_MAX > 0);
}
/****************************************************************************
@ -269,7 +324,7 @@ void *mm_realloc(struct mm_heap_s *heap, void *oldmem,
int usmblks;
int newsize;
mm_free_delaylist(heap);
mm_free_delaylist(heap, false);
if (size == 0)
{
@ -295,6 +350,13 @@ void *mm_realloc(struct mm_heap_s *heap, void *oldmem,
}
while (atomic_compare_exchange_weak(&heap->usmblks, &usmblks, uordblks));
#if CONFIG_MM_FREE_DELAYCOUNT_MAX > 0
if (mem == NULL && mm_free_delaylist(heap, true))
{
return mm_realloc(heap, oldmem, size);
}
#endif
return mem;
}
@ -358,7 +420,7 @@ void *mm_memalign(struct mm_heap_s *heap, size_t alignment, size_t size)
int uordblks;
int usmblks;
mm_free_delaylist(heap);
mm_free_delaylist(heap, false);
mem = host_memalign(alignment, size);
if (mem == NULL)
@ -381,6 +443,13 @@ void *mm_memalign(struct mm_heap_s *heap, size_t alignment, size_t size)
}
while (atomic_compare_exchange_weak(&heap->usmblks, &usmblks, uordblks));
#if CONFIG_MM_FREE_DELAYCOUNT_MAX > 0
if (mem == NULL && mm_free_delaylist(heap, true))
{
return mm_memalign(heap, alignment, size);
}
#endif
return mem;
}

View file

@ -358,4 +358,12 @@ config MM_PANIC_ON_FAILURE
default n
depends on DEBUG_MM
config MM_FREE_DELAYCOUNT_MAX
int "Maximum memory nodes can be delayed to free"
default 0
---help---
Set to 0 to disable the delayed free mechanism. Otherwise,
the value decides the maximum number of memory nodes that
will be delayed to free.
source "mm/iob/Kconfig"

View file

@ -260,6 +260,10 @@ struct mm_heap_s
FAR struct mm_delaynode_s *mm_delaylist[CONFIG_SMP_NCPUS];
#if CONFIG_MM_FREE_DELAYCOUNT_MAX > 0
size_t mm_delaycount[CONFIG_SMP_NCPUS];
#endif
/* The is a multiple mempool of the heap */
#if CONFIG_MM_HEAP_MEMPOOL_THRESHOLD != 0
@ -304,4 +308,8 @@ int mm_size2ndx(size_t size);
void mm_foreach(FAR struct mm_heap_s *heap, mm_node_handler_t handler,
FAR void *arg);
/* Functions contained in mm_free.c *****************************************/
void mm_delayfree(FAR struct mm_heap_s *heap, FAR void *mem, bool delay);
#endif /* __MM_MM_HEAP_MM_H */

View file

@ -50,6 +50,10 @@ static void add_delaylist(FAR struct mm_heap_s *heap, FAR void *mem)
tmp->flink = heap->mm_delaylist[up_cpu_index()];
heap->mm_delaylist[up_cpu_index()] = tmp;
#if CONFIG_MM_FREE_DELAYCOUNT_MAX > 0
heap->mm_delaycount[up_cpu_index()]++;
#endif
up_irq_restore(flags);
#endif
}
@ -59,15 +63,14 @@ static void add_delaylist(FAR struct mm_heap_s *heap, FAR void *mem)
****************************************************************************/
/****************************************************************************
* Name: mm_free
* Name: mm_delayfree
*
* Description:
* Returns a chunk of memory to the list of free nodes, merging with
* adjacent free chunks if possible.
* Delay free memory if `delay` is true, otherwise free it immediately.
*
****************************************************************************/
void mm_free(FAR struct mm_heap_s *heap, FAR void *mem)
void mm_delayfree(FAR struct mm_heap_s *heap, FAR void *mem, bool delay)
{
FAR struct mm_freenode_s *node;
FAR struct mm_freenode_s *prev;
@ -75,24 +78,6 @@ void mm_free(FAR struct mm_heap_s *heap, FAR void *mem)
size_t nodesize;
size_t prevsize;
minfo("Freeing %p\n", mem);
/* Protect against attempts to free a NULL reference */
if (!mem)
{
return;
}
DEBUGASSERT(mm_heapmember(heap, mem));
#if CONFIG_MM_HEAP_MEMPOOL_THRESHOLD != 0
if (mempool_multiple_free(heap->mm_mpool, mem) >= 0)
{
return;
}
#endif
if (mm_lock(heap) < 0)
{
/* Meet -ESRCH return, which means we are in situations
@ -110,6 +95,13 @@ void mm_free(FAR struct mm_heap_s *heap, FAR void *mem)
kasan_poison(mem, mm_malloc_size(heap, mem));
if (delay)
{
mm_unlock(heap);
add_delaylist(heap, mem);
return;
}
/* Map the memory chunk into a free node */
node = (FAR struct mm_freenode_s *)((FAR char *)mem - MM_SIZEOF_ALLOCNODE);
@ -202,3 +194,35 @@ void mm_free(FAR struct mm_heap_s *heap, FAR void *mem)
mm_addfreechunk(heap, node);
mm_unlock(heap);
}
/****************************************************************************
* Name: mm_free
*
* Description:
* Returns a chunk of memory to the list of free nodes, merging with
* adjacent free chunks if possible.
*
****************************************************************************/
void mm_free(FAR struct mm_heap_s *heap, FAR void *mem)
{
minfo("Freeing %p\n", mem);
/* Protect against attempts to free a NULL reference */
if (mem == NULL)
{
return;
}
DEBUGASSERT(mm_heapmember(heap, mem));
#if CONFIG_MM_HEAP_MEMPOOL_THRESHOLD != 0
if (mempool_multiple_free(heap->mm_mpool, mem) >= 0)
{
return;
}
#endif
mm_delayfree(heap, mem, CONFIG_MM_FREE_DELAYCOUNT_MAX > 0);
}

View file

@ -39,8 +39,23 @@
* Private Functions
****************************************************************************/
static void free_delaylist(FAR struct mm_heap_s *heap)
/****************************************************************************
* Name: free_delaylist
*
* Description:
* Free the memory in delay list either added because of mm_lock failed or
* added because of CONFIG_MM_FREE_DELAYCOUNT_MAX.
* Set force to true to free all the memory in delay list immediately, set
* to false will only free delaylist when time is up if
* CONFIG_MM_FREE_DELAYCOUNT_MAX is enabled.
*
* Return true if there is memory freed.
*
****************************************************************************/
static bool free_delaylist(FAR struct mm_heap_s *heap, bool force)
{
bool ret = false;
#if defined(CONFIG_BUILD_FLAT) || defined(__KERNEL__)
FAR struct mm_delaynode_s *tmp;
irqstate_t flags;
@ -50,12 +65,27 @@ static void free_delaylist(FAR struct mm_heap_s *heap)
flags = up_irq_save();
tmp = heap->mm_delaylist[up_cpu_index()];
#if CONFIG_MM_FREE_DELAYCOUNT_MAX > 0
if (tmp == NULL ||
(!force &&
heap->mm_delaycount[up_cpu_index()] < CONFIG_MM_FREE_DELAYCOUNT_MAX))
{
up_irq_restore(flags);
return false;
}
heap->mm_delaycount[up_cpu_index()] = 0;
#endif
heap->mm_delaylist[up_cpu_index()] = NULL;
up_irq_restore(flags);
/* Test if the delayed is empty */
ret = tmp != NULL;
while (tmp)
{
FAR void *address;
@ -69,9 +99,11 @@ static void free_delaylist(FAR struct mm_heap_s *heap)
* 'while' condition above.
*/
mm_free(heap, address);
mm_delayfree(heap, address, false);
}
#endif
return ret;
}
#if CONFIG_MM_BACKTRACE >= 0
@ -126,7 +158,7 @@ FAR void *mm_malloc(FAR struct mm_heap_s *heap, size_t size)
/* Free the delay list first */
free_delaylist(heap);
free_delaylist(heap, false);
#if CONFIG_MM_HEAP_MEMPOOL_THRESHOLD != 0
ret = mempool_multiple_alloc(heap->mm_mpool, size);
@ -279,6 +311,16 @@ FAR void *mm_malloc(FAR struct mm_heap_s *heap, size_t size)
minfo("Allocated %p, size %zu\n", ret, alignsize);
#endif
}
#if CONFIG_MM_FREE_DELAYCOUNT_MAX > 0
/* Try again after free delay list */
else if (free_delaylist(heap, true))
{
return mm_malloc(heap, size);
}
#endif
#ifdef CONFIG_DEBUG_MM
else if (MM_INTERNAL_HEAP(heap))
{

View file

@ -103,10 +103,14 @@ struct mm_heap_s
FAR struct mempool_multiple_s *mm_mpool;
#endif
/* Free delay list, for some situation can't do free immdiately */
/* Free delay list, for some situation can't do free immediately */
struct mm_delaynode_s *mm_delaylist[CONFIG_SMP_NCPUS];
#if CONFIG_MM_FREE_DELAYCOUNT_MAX > 0
size_t mm_delaycount[CONFIG_SMP_NCPUS];
#endif
#if defined(CONFIG_FS_PROCFS) && !defined(CONFIG_FS_PROCFS_EXCLUDE_MEMINFO)
struct procfs_meminfo_entry_s mm_procfs;
#endif
@ -129,6 +133,12 @@ struct mm_mallinfo_handler_s
FAR struct mallinfo_task *info;
};
/****************************************************************************
* Private Function Prototypes
****************************************************************************/
static void mm_delayfree(struct mm_heap_s *heap, void *mem, bool delay);
/****************************************************************************
* Private Functions
****************************************************************************/
@ -182,6 +192,10 @@ static void add_delaylist(FAR struct mm_heap_s *heap, FAR void *mem)
tmp->flink = heap->mm_delaylist[up_cpu_index()];
heap->mm_delaylist[up_cpu_index()] = tmp;
#if CONFIG_MM_FREE_DELAYCOUNT_MAX > 0
heap->mm_delaycount[up_cpu_index()]++;
#endif
up_irq_restore(flags);
#endif
}
@ -190,8 +204,9 @@ static void add_delaylist(FAR struct mm_heap_s *heap, FAR void *mem)
* Name: free_delaylist
****************************************************************************/
static void free_delaylist(FAR struct mm_heap_s *heap)
static bool free_delaylist(FAR struct mm_heap_s *heap, bool force)
{
bool ret = false;
#if defined(CONFIG_BUILD_FLAT) || defined(__KERNEL__)
FAR struct mm_delaynode_s *tmp;
irqstate_t flags;
@ -201,12 +216,27 @@ static void free_delaylist(FAR struct mm_heap_s *heap)
flags = up_irq_save();
tmp = heap->mm_delaylist[up_cpu_index()];
#if CONFIG_MM_FREE_DELAYCOUNT_MAX > 0
if (tmp == NULL ||
(!force &&
heap->mm_delaycount[up_cpu_index()] < CONFIG_MM_FREE_DELAYCOUNT_MAX))
{
up_irq_restore(flags);
return false;
}
heap->mm_delaycount[up_cpu_index()] = 0;
#endif
heap->mm_delaylist[up_cpu_index()] = NULL;
up_irq_restore(flags);
/* Test if the delayed is empty */
ret = tmp != NULL;
while (tmp)
{
FAR void *address;
@ -220,9 +250,11 @@ static void free_delaylist(FAR struct mm_heap_s *heap)
* 'while' condition above.
*/
mm_free(heap, address);
mm_delayfree(heap, address, false);
}
#endif
return ret;
}
#if CONFIG_MM_HEAP_MEMPOOL_THRESHOLD != 0 && CONFIG_MM_BACKTRACE >= 0
@ -457,6 +489,50 @@ static void memdump_handler(FAR void *ptr, size_t size, int used,
}
}
/****************************************************************************
* Name: mm_delayfree
*
* Description:
* Delay free memory if `delay` is true, otherwise free it immediately.
*
****************************************************************************/
static void mm_delayfree(FAR struct mm_heap_s *heap, FAR void *mem,
bool delay)
{
if (mm_lock(heap) == 0)
{
#ifdef CONFIG_MM_FILL_ALLOCATIONS
memset(mem, 0x55, mm_malloc_size(heap, mem));
#endif
kasan_poison(mem, mm_malloc_size(heap, mem));
/* Update heap statistics */
heap->mm_curused -= mm_malloc_size(heap, mem);
/* Pass, return to the tlsf pool */
if (delay)
{
add_delaylist(heap, mem);
}
else
{
tlsf_free(heap->mm_tlsf, mem);
}
mm_unlock(heap);
}
else
{
/* Add to the delay list(see the comment in mm_lock) */
add_delaylist(heap, mem);
}
}
/****************************************************************************
* Public Functions
****************************************************************************/
@ -678,18 +754,17 @@ void mm_extend(FAR struct mm_heap_s *heap, FAR void *mem, size_t size,
void mm_free(FAR struct mm_heap_s *heap, FAR void *mem)
{
int ret;
UNUSED(ret);
minfo("Freeing %p\n", mem);
/* Protect against attempts to free a NULL reference */
if (!mem)
if (mem == NULL)
{
return;
}
DEBUGASSERT(mm_heapmember(heap, mem));
#if CONFIG_MM_HEAP_MEMPOOL_THRESHOLD != 0
if (mempool_multiple_free(heap->mm_mpool, mem) >= 0)
{
@ -697,29 +772,7 @@ void mm_free(FAR struct mm_heap_s *heap, FAR void *mem)
}
#endif
if (mm_lock(heap) == 0)
{
#ifdef CONFIG_MM_FILL_ALLOCATIONS
memset(mem, 0x55, mm_malloc_size(heap, mem));
#endif
kasan_poison(mem, mm_malloc_size(heap, mem));
/* Update heap statistics */
heap->mm_curused -= mm_malloc_size(heap, mem);
/* Pass, return to the tlsf pool */
tlsf_free(heap->mm_tlsf, mem);
mm_unlock(heap);
}
else
{
/* Add to the delay list(see the comment in mm_lock) */
add_delaylist(heap, mem);
}
mm_delayfree(heap, mem, CONFIG_MM_FREE_DELAYCOUNT_MAX > 0);
}
/****************************************************************************
@ -1061,7 +1114,7 @@ FAR void *mm_malloc(FAR struct mm_heap_s *heap, size_t size)
/* Free the delay list first */
free_delaylist(heap);
free_delaylist(heap, false);
/* Allocate from the tlsf pool */
@ -1095,6 +1148,15 @@ FAR void *mm_malloc(FAR struct mm_heap_s *heap, size_t size)
#endif
}
#if CONFIG_MM_FREE_DELAYCOUNT_MAX > 0
/* Try again after free delay list */
else if (free_delaylist(heap, true))
{
return mm_malloc(heap, size);
}
#endif
return ret;
}
@ -1126,7 +1188,7 @@ FAR void *mm_memalign(FAR struct mm_heap_s *heap, size_t alignment,
/* Free the delay list first */
free_delaylist(heap);
free_delaylist(heap, false);
/* Allocate from the tlsf pool */
@ -1156,6 +1218,15 @@ FAR void *mm_memalign(FAR struct mm_heap_s *heap, size_t alignment,
kasan_unpoison(ret, mm_malloc_size(heap, ret));
}
#if CONFIG_MM_FREE_DELAYCOUNT_MAX > 0
/* Try again after free delay list */
else if (free_delaylist(heap, true))
{
return mm_memalign(heap, alignment, size);
}
#endif
return ret;
}
@ -1237,7 +1308,7 @@ FAR void *mm_realloc(FAR struct mm_heap_s *heap, FAR void *oldmem,
#else
/* Free the delay list first */
free_delaylist(heap);
free_delaylist(heap, false);
/* Allocate from the tlsf pool */
@ -1268,8 +1339,16 @@ FAR void *mm_realloc(FAR struct mm_heap_s *heap, FAR void *oldmem,
}
#endif
#if CONFIG_MM_FREE_DELAYCOUNT_MAX > 0
/* Try again after free delay list */
if (newmem == NULL && free_delaylist(heap, true))
{
return mm_realloc(heap, oldmem, size);
}
#endif
#endif
return newmem;
}