connection pooling.
-- Mark Lewis
On Tue, 2005-04-12 at 22:09, Slavisa Garic wrote:
This is a serious problem for me as there are multiple users using our
software on our server and I would want to avoid having connections
open for a long time. In the scenario mentioned below I haven't
Hi,
This looks very interesting. I'll give it a better look and see if the
performance penalties pgpool brings are not substantial in which case
this program could be very helpful,
Thanks for the hint,
Slavisa
On 4/14/05, Richard Huxton dev@archonet.com wrote:
Slavisa Garic wrote
Hi all,
I've just noticed an interesting behaviour with PGSQL. My software is
made up of few different modules that interact through PGSQL database.
Almost every query they do is an individual transaction and there is a
good reason for that. After every query done there is some processing
done by
to connect could mean a loss of
computation results that were gathered over long periods of time.
Thanks for the help by the way :),
Regards,
Slavisa
On 12 Apr 2005 23:27:09 -0400, Greg Stark [EMAIL PROTECTED] wrote:
Tom Lane [EMAIL PROTECTED] writes:
Slavisa Garic [EMAIL PROTECTED] writes
Hi,
I have a quick question. In order to speed up insertion of large number of
rows (100s of thousands) I replaced the INSERT with the COPY. This works
fine but one question popped into my mind. Does copy updates indexes on
that table if there are some defined?
Thanks,
Slavisa
Thanks for the reply and thanks even more for the good one :).
Cheers,
Slavisa
On Fri, 6 Feb 2004, Christopher Kings-Lynne wrote:
I have a quick question. In order to speed up insertion of large number of
rows (100s of thousands) I replaced the INSERT with the COPY. This works
fine but
Hi Everyone,
I am using PostgreSQL 7.3.2 and have used earlier versions (7.1.x onwards)
and with all of them I noticed same problem with INSERTs when there is a
large data set. Just to so you guys can compare time it takes to insert
one row into a table when there are only few rows present and