The shrinker_map may be touched from any cpu (e.g., a bit there may be set by a task running everywhere) but kswapd is always bound to specific node. So allocate shrinker_map from the related NUMA node to respect its NUMA locality. Also, this follows generic way we use for allocation of memcg's per-node data.
Signed-off-by: Kirill Tkhai <ktk...@virtuozzo.com> Signed-off-by: Andrew Morton <a...@linux-foundation.org> Reviewed-by: David Hildenbrand <da...@redhat.com> Reviewed-by: Shakeel Butt <shake...@google.com> Reviewed-by: Roman Gushchin <g...@fb.com> Acked-by: Michal Hocko <mho...@suse.com> Cc: Johannes Weiner <han...@cmpxchg.org> Cc: Vladimir Davydov <vdavydov....@gmail.com> Link: http://lkml.kernel.org/r/fff0e636-4c36-ed10-281c-8cdb0687c...@virtuozzo.com Signed-off-by: Linus Torvalds <torva...@linux-foundation.org> (cherry picked from commit 86daf94efb11d7319fbef5e480018c4807add6ef) taken from vz7 commit bb7cce9fa82d517daaa2f9e7e3faea0f99a8da73 Signed-off-by: Vasily Averin <v...@virtuozzo.com> --- mm/memcontrol.c | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/mm/memcontrol.c b/mm/memcontrol.c index a7d971c42367..43efd7998c81 100644 --- a/mm/memcontrol.c +++ b/mm/memcontrol.c @@ -341,7 +341,7 @@ static int memcg_expand_one_shrinker_map(struct mem_cgroup *memcg, if (!old) return 0; - new = kvmalloc(sizeof(*new) + size, GFP_KERNEL); + new = kvmalloc_node(sizeof(*new) + size, GFP_KERNEL, nid); if (!new) return -ENOMEM; @@ -385,7 +385,7 @@ static int memcg_alloc_shrinker_maps(struct mem_cgroup *memcg) mutex_lock(&memcg_shrinker_map_mutex); size = memcg_shrinker_map_size; for_each_node(nid) { - map = kvzalloc(sizeof(*map) + size, GFP_KERNEL); + map = kvzalloc_node(sizeof(*map) + size, GFP_KERNEL, nid); if (!map) { memcg_free_shrinker_maps(memcg); ret = -ENOMEM; -- 2.25.1 _______________________________________________ Devel mailing list Devel@openvz.org https://lists.openvz.org/mailman/listinfo/devel