Hi there,

I'm having an issue with a process dying (being killed by OpenVZ
limits, presumably), and I can't figure out exactly why it's getting
killed.

Background info:
- kernel 2.6.18-53.1.19.el5.028stab053.14
- CentOS5 host, 2 CentOS5 guests
- The host has 2GB memory, 0.5GB swap
- One guest is only running BIND (plus the usual, sshd, syslogd,
  sendmail, crond services).  Am having no issues with this guest.
  privvmpages is set to provide it up to 256MB memory, and it's using
  about half that.
- The other guest is running postgresql, java, apache, and freeradius.

The problem is that freeradius keeps dying.  Whenever it dies, failcnt
on privvmpages goes up and indeed the maxheld privvmpages value is
above the limit value, so I guess the issue is that OpenVZ thinks that
something is taking too much memory and is killing radiusd (no other
failcnt numbers go up -- only privvmpages).

There doesn't appear to be anything logged in the dmesg output on the
host or the guest to indicate that anything was killed due to a limit
being exceeded (should there be?).

A few strange things:

- Although the maxheld privvmpages value is above the limit, I've
  never seen the held privvmpages value get anywhere near the limit,
  even checking the value only seconds before radiusd gets killed, the
  held privvmpages value is under half the limit, eg. just before
  radiusd is killed:

  2008-07-11 06:39:24:
       uid  resource           held    maxheld    barrier      limit  failcnt
            privvmpages      224497     581366     506368     557056      486

  Then 10 seconds later (radiusd was killed and possibly restarted
  sometime in this interval):

  2008-07-11 06:39:34:
       uid  resource           held    maxheld    barrier      limit  failcnt
            privvmpages      182445     581366     506368     557056      487

  (is there any way to reset the maxheld values without restarting the
  guest?)

- Similarly, the output of free doesn't indicate anything wrong:

  2008-07-11 06:39:24:
               total       used       free     shared    buffers     cached
  Mem:       2071924     898092    1173832          0          0          0
  -/+ buffers/cache:     898092    1173832

  2008-07-11 06:39:34:
               total       used       free     shared    buffers     cached
  Mem:       2071924     729884    1342040          0          0          0
  -/+ buffers/cache:     729884    1342040

- I've found that I can reproduce the issue on demand by sending many
  RADIUS requests to radiusd at once, but watching what radiusd does
  with ltrace -f doesn't show anything out of the ordinary.  I summed
  up all the malloc() requests and saw only 22MB requested.

- Finally, using strace -f to see what radiusd was doing -- there were
  only about 22MB worth of calls to brk() (matching malloc(), as you'd
  expect).  And summing the mmap() length parameters (not counting
  munmap() calls) I only came up with 300MB, well within the free
  memory.

Any ideas on debugging this?

Thanks in advance,
-- 
Geoffrey D. Bennett, RHCE, RHCX               mailto:[EMAIL PROTECTED]
Senior Systems Engineer                          sip:[EMAIL PROTECTED]
NetCraft Australia Pty Ltd        http://www.netcraft.com.au/geoffrey/
_______________________________________________
Users mailing list
Users@openvz.org
https://openvz.org/mailman/listinfo/users

Reply via email to