% slabtop -s c --once |head -10
Active / Total Objects (% used) : 185625 / 205302 (90.4%)
Active / Total Slabs (% used) : 16756 / 16757 (100.0%)
Active / Total Caches (% used) : 101 / 182 (55.5%)
Active / Total Size (% used) : 856865.31K / 859451.65K (99.7%)
Minimum / Average / Maximum Object : 0.02K / 4.19K / 4096.00K
OBJS ACTIVE USE OBJ SIZE SLABS OBJ/SLAB CACHE SIZE NAME
384 384 100% 2048.00K 384 1 786432K size-2097152
26276 26250 99% 1.00K 6569 4 26276K ext4_inode_cache
181 181 100% 32.12K 181 1 11584K kmem_cache
very different indeed.
(btw size-2097152 sounds like one of the default name used by hugectl (or
hugeadm) from the hugetlbfs tools). is that mounted in your case? and are there
any hugepages reserved? )
Not that I'd know of. But wasn't there a new feature called "transparent hugepage
support" in 6.1?
thp shows in /proc/meminfo as AnonHugePages
]# grep Huge /proc/meminfo
AnonHugePages: 18432 kB
HugePages_Total: 0
HugePages_Free: 0
HugePages_Rsvd: 0
HugePages_Surp: 0
Hugepagesize: 2048 kB
you show check those there too to see where the come from
stijn
Cheers,
Stephan
stijn
On 09/05/2011 05:10 PM, Stephan Wiesand wrote:
Hi Stijn,
On Sep 5, 2011, at 16:24, Stijn De Weirdt wrote:
hi all,
we are having an "issue" with some SL61 nodes. after a reboot, free reports
1.4GB of memory in use, of which 24+163=187MB buffers+cache.
i'm unable to identify what is holding the memory, and i'd like to know if
others see this too and how i could proceed to find the culprit.
yes, we see this as well. On a 48 GB system without users or special processes:
# free -m
total used free shared buffers cached
Mem: 48388 1374 47013 0 30 186
-/+ buffers/cache: 1157 47231
In /proc/meminfo, I find that the difference to what I'd consider reasonable
(and see on a 48GB SL5 system) is due to slabs.
A "slabtop -s c" reveals that it's a "size-2097152" pool accounting for this.
Do you see this as well?
Cheers,
Stephan
(it is a 32core/64GB machine; kernel commandline has crashkernel=128M@16M (but
no difference then eg crashkernel=auto and kdump is off))
many thanks,
stijn
free
# free -m
total used free shared buffers cached
Mem: 64554 1604 62949 0 24 166
-/+ buffers/cache: 1413 63140
Swap: 16394 0 16394
mem sorted top
top - 16:13:52 up 13 min, 1 user, load average: 0.00, 0.01, 0.01
Tasks: 694 total, 1 running, 693 sleeping, 0 stopped, 0 zombie
Cpu(s): 0.0%us, 0.0%sy, 0.0%ni,100.0%id, 0.0%wa, 0.0%hi, 0.0%si, 0.0%st
Mem: 66103768k total, 1643336k used, 64460432k free, 25164k buffers
Swap: 16787916k total, 0k used, 16787916k free, 170552k cached
PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
2788 root 20 0 37988 25m 2876 S 0.0 0.0 0:00.06 pbs_mom
2653 root 20 0 159m 12m 1472 S 0.0 0.0 0:00.19 ncm-cdispd
2643 root 20 0 138m 5604 840 S 0.0 0.0 0:00.00 cdp-listend
3276 root 20 0 120m 4156 3232 S 0.0 0.0 0:00.07 sshd
2620 root 20 0 745m 3788 1764 S 0.0 0.0 0:00.12 automount
3102 nslcd 20 0 427m 2936 488 S 0.0 0.0 0:00.00 nslcd
3301 root 20 0 103m 1688 1336 S 0.0 0.0 0:00.05 bash
3623 root 20 0 13528 1604 844 R 0.3 0.0 0:00.14 top
1 root 20 0 21416 1544 1240 S 0.0 0.0 0:06.23 init
2482 root 20 0 194m 1484 1108 S 0.0 0.0 0:00.14 qlgc_dsc
2325 root 20 0 242m 1412 928 S 0.0 0.0 0:00.04 rsyslogd
2459 rpcuser 20 0 23112 1168 884 S 0.0 0.0 0:00.00 rpc.statd
2606 root 18 -2 10956 1144 412 S 0.0 0.0 0:00.03 udevd
3164 nscd 20 0 583m 1132 788 S 0.0 0.0 0:00.02 nscd
2697 root 20 0 62040 1064 464 S 0.0 0.0 0:00.00 sshd
943 root 16 -4 10960 1052 316 S 0.0 0.0 0:00.12 udevd
2607 root 18 -2 10956 1052 320 S 0.0 0.0 0:00.00 udevd
2723 root 20 0 112m 1012 380 S 0.0 0.0 0:00.00 crond
2707 root 20 0 22488 992 752 S 0.0 0.0 0:00.03 xinetd
2439 rpc 20 0 18940 908 672 S 0.0 0.0 0:00.04 rpcbind
2568 dbus 20 0 23448 876 604 S 0.0 0.0 0:00.01 dbus-daemon
2972 nagios 20 0 37096 796 452 S 0.0 0.0 0:00.00 nrpe