Public bug reported:
Issue description: Memory demand increases very fast up to the maximum
available memory, call traces show up in /var/log/syslog, the VM becomes
unresponsive during the memory consumption, but it becomes responsive
right after stress-ng ends its execution, therefore we can say the issue
might be in the Dynamic Memory feature.
Platform: host independent
Distribution name and release:
- Bionic (4.15.0-10-generic)
- Linux Azure kernel (4.13.0-1012-azure)
Repro rate: 100%
VM configuration:
Kernel: 4.13.0-1012-azure
CPU: 8 cores
RAM: 2048 MB
Enable Dynamic Memory: Yes
Minimum RAM: 512 MB
Maximum RAM: 8192 MB
Memory buffer: 20%
Memory weight: 100%
Repro steps:
1. Start the VM with the above configuration.
2. Run stress-ng with the following parameters:
stress-ng -m 16 -vm-bytes 256M -t 200 --backoff 10000000
3. In less than 60 seconds the entire memory is consumed.
Additional info:
1. We also tested this on Xenial with default kernel and the issue does
not occur.
2. The call trace message from syslog can be seen below.
Mar 14 08:42:17 xenial kernel: [ 257.120261] Call Trace:
Mar 14 08:42:17 xenial kernel: [ 257.120267] dump_stack+0x63/0x82
Mar 14 08:42:17 xenial kernel: [ 257.120271] dump_header+0x97/0x225
Mar 14 08:42:17 xenial kernel: [ 257.120276] ?
security_capable_noaudit+0x4b/0x70
Mar 14 08:42:17 xenial kernel: [ 257.120277] oom_kill_process+0x219/0x420
Mar 14 08:42:17 xenial kernel: [ 257.120279] out_of_memory+0x11d/0x4b0
Mar 14 08:42:17 xenial kernel: [ 257.120282]
__alloc_pages_slowpath+0xd32/0xe10
Mar 14 08:42:17 xenial kernel: [ 257.120284]
__alloc_pages_nodemask+0x263/0x280
Mar 14 08:42:17 xenial kernel: [ 257.120289] alloc_pages_vma+0x88/0x1e0
Mar 14 08:42:17 xenial kernel: [ 257.120292] shmem_alloc_page+0x70/0xc0
Mar 14 08:42:17 xenial kernel: [ 257.120295] ? __radix_tree_insert+0x45/0x230
Mar 14 08:42:17 xenial kernel: [ 257.120297]
shmem_alloc_and_acct_page+0x73/0x1b0
Mar 14 08:42:17 xenial kernel: [ 257.120298] ? find_get_entry+0x1e/0xe0
Mar 14 08:42:17 xenial kernel: [ 257.120300] shmem_getpage_gfp+0x18f/0xbf0
Mar 14 08:42:17 xenial kernel: [ 257.120303] shmem_fault+0x9d/0x1e0
Mar 14 08:42:17 xenial kernel: [ 257.120306] ? file_update_time+0x60/0x110
Mar 14 08:42:17 xenial kernel: [ 257.120309] __do_fault+0x24/0xd0
Mar 14 08:42:17 xenial kernel: [ 257.120311] __handle_mm_fault+0x983/0x1080
Mar 14 08:42:17 xenial kernel: [ 257.120314] ? set_next_entity+0xd9/0x220
Mar 14 08:42:17 xenial kernel: [ 257.120316] handle_mm_fault+0xcc/0x1c0
Mar 14 08:42:17 xenial kernel: [ 257.120319] __do_page_fault+0x258/0x4f0
Mar 14 08:42:17 xenial kernel: [ 257.120320] do_page_fault+0x22/0x30
Mar 14 08:42:17 xenial kernel: [ 257.120323] ? page_fault+0x36/0x60
Mar 14 08:42:17 xenial kernel: [ 257.120325] page_fault+0x4c/0x60
Mar 14 08:42:17 xenial kernel: [ 257.120327] RIP: 0033:0x4493ad
Mar 14 08:42:17 xenial kernel: [ 257.120328] RSP: 002b:00007ffd3ff927b0
EFLAGS: 00010286
Mar 14 08:42:17 xenial kernel: [ 257.120329] RAX: 00000000c6a6a9e7 RBX:
00000001a8ae9d07 RCX: 0000000021b83c00
Mar 14 08:42:17 xenial kernel: [ 257.120330] RDX: 00007f2ad911d000 RSI:
00000000271ea9e7 RDI: 00000000271e9660
Mar 14 08:42:17 xenial kernel: [ 257.120331] RBP: 00007f2adab70000 R08:
000000001bda3103 R09: 0000000048373eca
Mar 14 08:42:17 xenial kernel: [ 257.120331] R10: 0000000048373eca R11:
00007f2acab70000 R12: 0000000000000000
Mar 14 08:42:17 xenial kernel: [ 257.120332] R13: 8309310348373eca R14:
56b63c1fe6166568 R15: 00007f2acab70000
Mar 14 08:42:17 xenial kernel: [ 257.120333] Mem-Info:
Mar 14 08:42:17 xenial kernel: [ 257.120337] active_anon:249563
inactive_anon:83607 isolated_anon:65
Mar 14 08:42:17 xenial kernel: [ 257.120337] active_file:54 inactive_file:74
isolated_file:0
Mar 14 08:42:17 xenial kernel: [ 257.120337] unevictable:2561 dirty:0
writeback:68 unstable:0
Mar 14 08:42:17 xenial kernel: [ 257.120337] slab_reclaimable:10707
slab_unreclaimable:16622
Mar 14 08:42:17 xenial kernel: [ 257.120337] mapped:132260 shmem:332303
pagetables:2888 bounce:0
Mar 14 08:42:17 xenial kernel: [ 257.120337] free:12983 free_pcp:493
free_cma:0
Mar 14 08:42:17 xenial kernel: [ 257.120340] Node 0 active_anon:998252kB
inactive_anon:334428kB active_file:216kB inactive_file:296kB
unevictable:10244kB isolated(anon):260kB isolated(file):0kB mapped:529040kB
dirty:0kB writeback:272kB shmem:1329212kB shmem_thp: 0kB shmem_pmdmapped: 0kB
anon_thp: 0kB writeback_tmp:0kB unstable:0kB all_unreclaimable? no
Mar 14 08:42:17 xenial kernel: [ 257.120341] Node 0 DMA free:7404kB min:392kB
low:488kB high:584kB active_anon:5616kB inactive_anon:2384kB active_file:0kB
inactive_file:0kB unevictable:0kB writepending:0kB present:15992kB
managed:15908kB mlocked:0kB kernel_stack:0kB pagetables:64kB bounce:0kB
free_pcp:0kB local_pcp:0kB free_cma:0kB
Mar 14 08:42:17 xenial kernel: [ 257.120345] lowmem_reserve[]: 0 1754 1754
1754 1754
Mar 14 08:42:17 xenial kernel: [ 257.120347] Node 0 DMA32 free:44528kB
min:44660kB low:55824kB high:66988kB active_anon:992772kB
inactive_anon:331896kB active_file:216kB inactive_file:296kB
unevictable:10244kB writepending:600kB present:2080704kB managed:1818152kB
mlocked:10244kB kernel_stack:4512kB pagetables:11488kB bounce:0kB
free_pcp:1972kB local_pcp:236kB free_cma:0kB
Mar 14 08:42:17 xenial kernel: [ 257.120351] lowmem_reserve[]: 0 0 0 0 0
Mar 14 08:42:17 xenial kernel: [ 257.120353] Node 0 DMA: 1*4kB (U) 9*8kB (UME)
10*16kB (UME) 4*32kB (UM) 4*64kB (UM) 5*128kB (UME) 4*256kB (UM) 2*512kB (ME)
2*1024kB (ME) 1*2048kB (E) 0*4096kB = 7404kB
Mar 14 08:42:17 xenial kernel: [ 257.120363] Node 0 DMA32: 12*4kB (UME)
175*8kB (UE) 138*16kB (UE) 91*32kB (UM) 27*64kB (UM) 49*128kB (UM) 19*256kB
(UM) 8*512kB (M) 21*1024kB (M) 0*2048kB 0*4096kB = 45032kB
** Affects: linux (Ubuntu)
Importance: Undecided
Status: New
** Tags: hyper-v
--
You received this bug notification because you are a member of Ubuntu
Bugs, which is subscribed to Ubuntu.
https://bugs.launchpad.net/bugs/1756129
Title:
[Hyper-V] Dynamic Memory HotAdd memory demand increases very fast to
maximum and logs show call trace messages
To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1756129/+subscriptions
--
ubuntu-bugs mailing list
[email protected]
https://lists.ubuntu.com/mailman/listinfo/ubuntu-bugs