Hi,

I am wondering about the following phenomenon: I have a list "interrels" of 
polynomials in many variables and a list "potential_sols" of potential 
solutions and I am using the following code to check which are actually 
solutions (roots). I use the following code

sols=[]
for psol in potential_sols:
    for interrel in interrels:
        if interrel(*psol)!=0:
            break
    else:
        print(psol)
        sols.append(psol)

I am not surprised (under my circumstances) that this takes long, though 
actually it wasn't that bad at first. What surprises me is that it uses huge 
amounts of memory. The actual data don't take that much memory and the code 
just linearly runs through all possibilities so it practically shouldn't need 
any extra memory at all (at most another copy of potential_sols). But instead 
it had filled the 4 GB physical memory plus 2 GB swap when I canceled it (of 
course by that time it also was terribly slow, just because of all the 
swapping). This seems to suggest memory leaks.
Is there any way I could check what part of sage (and which data structures) 
use how much memory? Are there any known issues about memory leaks related to 
polynomial rings over finite fields? Is there a garbage collection I have to 
call manually?
Thanks for any help!

Best,
Stefan

-- 
You received this message because you are subscribed to the Google Groups 
"sage-support" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To post to this group, send email to [email protected].
Visit this group at http://groups.google.com/group/sage-support?hl=en.
For more options, visit https://groups.google.com/groups/opt_out.


Reply via email to