On Feb 22, 2010, at 5:37 PM, Kenny Daily wrote: > from pygr import worldbase > all_resources = worldbase.dir("Bio.Test.Annotation") > dbs = filter(lambda x: x.endswith("db"), all_resources) > > for x in dbs: > resource = worldbase(x) > resource.close() > > This uses ~50-100MB per resource loaded, and it accumulates. In my > case, the AnnotationDB's sliceDB is a MySQL table. Any ideas about > where the memory could be going?
Our XMLRPC server loads all the AnnotationDB that it serves (must be at least 100 - 200) with little memory usage, so this horrible memory usage should absolutely not be happening to you. If you tested this with beta1 code, please re-run your example with an official release i.e. 0.8.0 or 0.8.1. I suspect that this may solve your problem -- as I recall we discovered a nasty bug in MySQLdb that uses up massive amounts of memory when you simply ask for an iterator. Starting with 0.8.0 we included a workaround to avoid that problem. -- You received this message because you are subscribed to the Google Groups "pygr-dev" group. To post to this group, send email to pygr-...@googlegroups.com. To unsubscribe from this group, send email to pygr-dev+unsubscr...@googlegroups.com. For more options, visit this group at http://groups.google.com/group/pygr-dev?hl=en.