Justin Wyllie wrote:
Hi clint
Yes. Linux and this script looks good. We've think that part of the problem
is in the modules Apache is loading so this will be useful.
I also have another couple of questions:
I have found the errant code where our process jumps by 13 Mbs. One part
does something like this:
$file_handle->read($s, $length); #$s is about 1/2 Mb
@data = unpack($format , $s);
##at this point memory usage jumps by 8 Mbs (measured using GTop->size() )
As Clinton said, perl doesn't free the memory back to the OS when you
slurp this file into ram. If you really want to free up the resources
(which will get reused by subsequent requests, they just aren't
available to the OS) you can use $r->child_terminate to make that child
die after handling that request, which will free the resources (and
likely spawn another child in it's place.
Adam