New debugfs stat counters are added to track the numbers of kmem_cache_alloc() and kmem_cache_free() function calls to get a sense of how the internal debug objects cache management is performing.
Signed-off-by: Waiman Long <long...@redhat.com> --- lib/debugobjects.c | 10 ++++++++++ 1 file changed, 10 insertions(+) diff --git a/lib/debugobjects.c b/lib/debugobjects.c index bb415a7..820e733 100644 --- a/lib/debugobjects.c +++ b/lib/debugobjects.c @@ -58,6 +58,12 @@ struct debug_bucket { = ODEBUG_POOL_MIN_LEVEL; static struct debug_obj_descr *descr_test __read_mostly; +/* + * Track numbers of kmem_cache_alloc and kmem_cache_free done. + */ +static int debug_objects_alloc; +static int debug_objects_freed; + static void free_obj_work(struct work_struct *work); static DECLARE_WORK(debug_obj_work, free_obj_work); @@ -105,6 +111,7 @@ static void fill_pool(void) raw_spin_lock_irqsave(&pool_lock, flags); hlist_add_head(&new->node, &obj_pool); + debug_objects_alloc++; obj_pool_free++; raw_spin_unlock_irqrestore(&pool_lock, flags); } @@ -176,6 +183,7 @@ static void free_obj_work(struct work_struct *work) obj = hlist_entry(obj_pool.first, typeof(*obj), node); hlist_del(&obj->node); obj_pool_free--; + debug_objects_freed++; /* * We release pool_lock across kmem_cache_free() to * avoid contention on pool_lock. @@ -753,6 +761,8 @@ static int debug_stats_show(struct seq_file *m, void *v) seq_printf(m, "pool_min_free :%d\n", obj_pool_min_free); seq_printf(m, "pool_used :%d\n", obj_pool_used); seq_printf(m, "pool_max_used :%d\n", obj_pool_max_used); + seq_printf(m, "objects_alloc :%d\n", debug_objects_alloc); + seq_printf(m, "objects_freed :%d\n", debug_objects_freed); return 0; } -- 1.8.3.1