--- Bruno Wolff III [EMAIL PROTECTED] wrote:
This does not belong on the pgsql-bugs list. The
pgsql-novice or
pgsql-performance lists seem more appropiate. I have
set followups
to the pgsql-novice list.
On Thu, Jun 02, 2005 at 12:05:00 +0100,
Bahadur Singh [EMAIL PROTECTED] wrote:
Hello,
I found situtation that, when I am selecting data
from a table of 200
records, getting slower as I do continous update
to the same existing data.
You need to be vacuuming (and possibly analyzing)
the table more often as the
updates will leave dead rows in the table which will
bloat the table size and
slow down access, particularly sequential scans. If
the updates modify the
data value distributions significantly, then you
will also need to
reanalyze the table to help the planner make good
decisions.
Many thanks for this tip !
But is this good idea to analyse/vacuuming the
database tables while updates are taking place..
Since, I update continuously say (100,000 ) times or
more the same data set.
This is the result of analyze command.
INFO: analyzing public.salesarticle
INFO: salesarticle: scanned 3000 of 20850 pages,
containing 62 live rows and 134938 dead rows; 62 rows
in sample, 431 estimated total rows
Gesamtlaufzeit der Abfrage: 5531 ms.
Total Time Taken : 5531 ms.
Can you suggest me some clever way to so, because I
would prefer to do vaccumming while database is not
loaded with queries/transactions.
Regards
Bahadur
__
Discover Yahoo!
Find restaurants, movies, travel and more fun for the weekend. Check it out!
http://discover.yahoo.com/weekend.html
---(end of broadcast)---
TIP 2: you can get off all lists at once with the unregister command
(send unregister YourEmailAddressHere to [EMAIL PROTECTED])