So, I have a derby database that could potentially contain millions of
records, but at the moment it has 13K.

I am trying to iterate through that table after calling a basic
"Select * from myTable".  After about 5000 records, I get an out of
memory
exception/error on the server side.  This table has very a couple
small clobs in it (each field contains about only 100 chars at the
moment) and one large clob.  I have no transactions running, the
result set is read only.  I would think that I should be able to loop
through an infinite set of results.  If I do this in embedded mode
it's fine. I have read over 10 million records with no problem. But,
again, in network mode it fails early.

Anyone have any advice?  I would think that a non-scrollable result
set should not keep any of the results in memory (on the server) once
the cursor has passed it.





-- 
"Conscious decisions by conscious minds are what make reality real"

Reply via email to