On Tuesday, 11 November 2014 at 20:20:26 UTC, Rainer Schuetze
wrote:
On 11.11.2014 08:36, Iain Buclaw wrote:
Hi,
I find myself wondering what do people use to inspect the GC
on a
running process?
Last night, I had a look at a couple of vibe.d servers that
had been
running for a little over 100 days. But the same code, but
one used
less (or not at all).
Given that the main app logic is rather simple, and things
that may be
otherwise held in memory (files) are offloaded onto a Redis
server.
I've have thought that it's consumption would have stayed
pretty much
stable. But to my surprise, I found that the server that is
under more
(but not heavy) use is consuming a whooping 200MB.
Initially I had tried to see if this could be shrunk in some
way or
form. Attached gdb to the process, called gc_minimize(),
gc_collect() -
but it didn't have any effect.
The GC allocates pools of memory in increasing size. It starts
with 1 MB, then adds 3MB for every new pool. (Numbers might be
slightly different depending on the druntime version). These
pools are then used to service any allocation request.
gc_minimize can only return memory to the system if all the
allocation in a pool are collected, which is very unlikely.
I'm aware of roughly how the gc grows. But it seems an unlikely
scenario to have 200MB worth of 3MB pools with at least one
object in each.
And if it did get to that state, the next question would be, how?
I could say, I'd expect that if a large number of requests came
in all at once, but I would have been prompted by this in the
network graphs.
When I noticed gc_stats with an informative *experimental*
warning, I
thought "lets just run it anyway and see what happens"...
SEGV. Wonderful.
I suspect calling gc_stats from the debugger is "experimental"
because it returns a struct. With a bit of casting, you might
by able to call "_gc.getStats( *cast(GCStats*)some_mem_adr );"
instead.
No, that is not the reason. More like the iterative scan may be
unsafe. I should have looked closer at the backtrace / memory
location that was violated (I was in a hurry to get the site back
up), but a more likely cause of the SEGV is that one of the pools
in gcx.pooltable[n] or pages in pool.pagetable[n] was pointing to
a free'd, stomped, or null location.
Iain.