#578: Garbage Collector in 0.6 ---------------------------+------------------------------------------------ Reporter: j...@… | Owner: lsansone...@… Type: defect | Status: new Priority: major | Milestone: MacRuby 0.6 Component: MacRuby | Keywords: ---------------------------+------------------------------------------------ Reporting this as a bug, since i'm fairly sure this is not the way it's intended to work.
Running the following code on a bunch of 600mb - 2gb files will (eventually) cause a SIGSEGV when memory allocation fails. {{{ def get_digest(file) digest = Digest::MD5.new() fil = File.open(file,'r') while((l = fil.read(READ_BUFFER_SIZE))!=nil) digest << l end fil.close() digest.hexdigest end }}} If I run it on a smaller set of files, 1.8gb to be exact, the memory allocated to the process jumps to 700-900mb, and just stays there. I let the process sit there for 30 mins, without noticing any drop in allocated memory. I've tried with various sizes for READ_BUFFER_SIZE, from 16kb to 32mb, and also tried with File.read(), and it all behaves the same way, although the higher read buffer, the faster memory usage shoots up. setting "l=nil" for every read operation and/or "digest=nil" before returning doesn't seem to make a difference. Using Digest::MD5.file(fname).hexdigest results in "writing to non- bytestrings is not supported at this time." I'm not sure if this is a "string cache" thing gone wrong, or it's simply not collecting garbage. -- Ticket URL: <http://www.macruby.org/trac/ticket/578> MacRuby <http://macruby.org/> _______________________________________________ MacRuby-devel mailing list MacRuby-devel@lists.macosforge.org http://lists.macosforge.org/mailman/listinfo.cgi/macruby-devel