I tried searching for an older posting on this without any luck, and I'm sure it's been discussed before. The closest I could come up with it this:
https://groups.google.com/forum/?fromgroups=#!topic/sage-support/FvrXRUuhy1Q which pretty much describes the issue I'm encountering. However, the memory question was not addressed, rather, a way to circumvent the issue was provided. I'm doing a similar thing: I have a particular set I can compute for a given dynamical system, then I wish to do this for many dynamical systems just storing the small amount of data that is the result. Each computation takes a small amount of memory, however, even with 16Gb of memory it is quickly running out of memory in some thousands of iterations. My best guess is that Sage/Python is caching/storing information from the previous computations. Is there a way to clear this and essentially have a "clean slate" for the next iteration? (I'd like to be doing millions or billions of such computations...) I'd post my code except that it isn't a nice simple snippet. It involves a couple experimental patches and the computation is actually quite involved. Thanks, Ben -- You received this message because you are subscribed to the Google Groups "sage-support" group. To post to this group, send email to [email protected]. To unsubscribe from this group, send email to [email protected]. Visit this group at http://groups.google.com/group/sage-support?hl=en.
