- Are you calling commit() after each batch?
- What does the memory usage look like when it slows down?
- If the memory usage is running high, trying using jdump to get a heap
dump, and then use Eclipse MAT to analyze the dump
- If the memory usage is fine, try hooking up a profiler likes Netbeans
Profiler and see what is taking the time
On 2012-03-25 07:26, Harish Mahadevan wrote:
Hi All,
I am linking a table of 5 million records from one h2 database to
another h2 database and am trying to copy the 5 million records 300
times. The first 175 times the copy occurs at a constant time interval
of approx 25 times every 40 minutes, but after the new table has
approximately a billion records, the inserts seems to be slowing down,
it almost takes about 24 hours to copy the same 5 million another 25
times. I am using a INSERT INTO with a SELECT sql, thats the reason
behind the linking of the table. There is literally no other activity
on the machine i am using to run this copy. I have dedicated around 8
gigs for this process. Any idea on why this behavior with h2
databases?
regards
Harish
--
You received this message because you are subscribed to the Google Groups "H2
Database" group.
To post to this group, send email to [email protected].
To unsubscribe from this group, send email to
[email protected].
For more options, visit this group at
http://groups.google.com/group/h2-database?hl=en.