hi stephan

the first lines of slabtop show

Active / Total Objects (% used)    : 238382 / 243482 (97.9%)
 Active / Total Slabs (% used)      : 4987 / 4987 (100.0%)
 Active / Total Caches (% used)     : 64 / 80 (80.0%)
 Active / Total Size (% used)       : 68080.12K / 69386.02K (98.1%)
 Minimum / Average / Maximum Object : 0.01K / 0.28K / 8.00K

is this similar?

(btw size-2097152 sounds like one of the default name used by hugectl (or hugeadm) from the hugetlbfs tools). is that mounted in your case? and are there any hugepages reserved? )


stijn

On 09/05/2011 05:10 PM, Stephan Wiesand wrote:
Hi Stijn,

On Sep 5, 2011, at 16:24, Stijn De Weirdt wrote:

hi all,

we are having an "issue" with some SL61 nodes. after a reboot, free reports 
1.4GB of memory in use, of which 24+163=187MB buffers+cache.

i'm unable to identify what is holding the memory, and i'd like to know if 
others see this too and how i could proceed to find the culprit.

yes, we see this as well. On a 48 GB system without users or special processes:

# free -m
              total       used       free     shared    buffers     cached
Mem:         48388       1374      47013          0         30        186
-/+ buffers/cache:       1157      47231

In /proc/meminfo, I find that the difference to what I'd consider reasonable 
(and see on a 48GB SL5 system) is due to slabs.

A "slabtop -s c" reveals that it's a "size-2097152" pool accounting for this. 
Do you see this as well?

Cheers,
        Stephan


(it is a 32core/64GB machine; kernel commandline has crashkernel=128M@16M (but 
no difference then eg crashkernel=auto and kdump is off))

many thanks,


stijn

free
# free -m
             total       used       free     shared    buffers     cached
Mem:         64554       1604      62949          0         24        166
-/+ buffers/cache:       1413      63140
Swap:        16394          0      16394


mem sorted top

top - 16:13:52 up 13 min,  1 user,  load average: 0.00, 0.01, 0.01
Tasks: 694 total,   1 running, 693 sleeping,   0 stopped,   0 zombie
Cpu(s):  0.0%us,  0.0%sy,  0.0%ni,100.0%id,  0.0%wa,  0.0%hi,  0.0%si, 0.0%st
Mem:  66103768k total,  1643336k used, 64460432k free,    25164k buffers
Swap: 16787916k total,        0k used, 16787916k free,   170552k cached

  PID USER      PR  NI  VIRT  RES  SHR S %CPU %MEM    TIME+  COMMAND
2788 root      20   0 37988  25m 2876 S  0.0  0.0   0:00.06 pbs_mom
2653 root      20   0  159m  12m 1472 S  0.0  0.0   0:00.19 ncm-cdispd
2643 root      20   0  138m 5604  840 S  0.0  0.0   0:00.00 cdp-listend
3276 root      20   0  120m 4156 3232 S  0.0  0.0   0:00.07 sshd
2620 root      20   0  745m 3788 1764 S  0.0  0.0   0:00.12 automount
3102 nslcd     20   0  427m 2936  488 S  0.0  0.0   0:00.00 nslcd
3301 root      20   0  103m 1688 1336 S  0.0  0.0   0:00.05 bash
3623 root      20   0 13528 1604  844 R  0.3  0.0   0:00.14 top
    1 root      20   0 21416 1544 1240 S  0.0  0.0   0:06.23 init
2482 root      20   0  194m 1484 1108 S  0.0  0.0   0:00.14 qlgc_dsc
2325 root      20   0  242m 1412  928 S  0.0  0.0   0:00.04 rsyslogd
2459 rpcuser   20   0 23112 1168  884 S  0.0  0.0   0:00.00 rpc.statd
2606 root      18  -2 10956 1144  412 S  0.0  0.0   0:00.03 udevd
3164 nscd      20   0  583m 1132  788 S  0.0  0.0   0:00.02 nscd
2697 root      20   0 62040 1064  464 S  0.0  0.0   0:00.00 sshd
  943 root      16  -4 10960 1052  316 S  0.0  0.0   0:00.12 udevd
2607 root      18  -2 10956 1052  320 S  0.0  0.0   0:00.00 udevd
2723 root      20   0  112m 1012  380 S  0.0  0.0   0:00.00 crond
2707 root      20   0 22488  992  752 S  0.0  0.0   0:00.03 xinetd
2439 rpc       20   0 18940  908  672 S  0.0  0.0   0:00.04 rpcbind
2568 dbus      20   0 23448  876  604 S  0.0  0.0   0:00.01 dbus-daemon
2972 nagios    20   0 37096  796  452 S  0.0  0.0   0:00.00 nrpe

Reply via email to