Hi, I am having a problem with inserting a large amount of data with my libpqxx program into an initially empty database. It appears to be the EXACT same problem discussed here:
http://archives.postgresql.org/pgsql-bugs/2005-03/msg00183.php In fact my situation is nearly identical, with roughly 5 major tables, with foreign keys between each other. All the tables are being loaded into similtaneously with about 2-3 million rows each. It seems that the problem is caused by the fact that I am using prepared statments, that cause the query planner to choose sequential scans for the foreign key checks due to the table being initially empty. As with the post above, if I dump my connection after about 4000 inserts, and restablish it the inserts speed up by a couple of orders of magnitude and remain realtively constant through the whole insertion. At first I was using straight insert statments, and although they were a bit slower than the prepared statments(after the restablished connection) they never ran into this problem with the database being initially empty. I only changed to the prepared statements because it was suggested in the documentation for advice on bulk data loads =). I can work around this problem, and I am sure somebody is working on fixing this, but I thought it might be good to reaffirm the problem. Thanks, Morgan Kita ---------------------------(end of broadcast)--------------------------- TIP 2: you can get off all lists at once with the unregister command (send "unregister YourEmailAddressHere" to [EMAIL PROTECTED])