Hi,

  I am updating a big table (90M records) with data from another rather large 
table (4M entries). Here is my update query:

    update links set target_size = 
    ( select size from articles where articles.article_id = links.article_to)

 I have built all the indexes one might need, increased shared mem buffers to 
400MB, I looked at the query plan and it looks reasonable.
But its taking an eternity to run: I've been running the query for 3 hours now 
on my new Mac laptop, and looking at the activity monitor I see that postrges 
is spending all of this time in disk IO (average CPU load of postgres process 
is about 4-5%).

 However, just looking at the query, postgres could cache the articles table 
and do a single pass over the links table...

 Please let me know if there is a good solution to this.

Thanks!
Pavel Velikhov
Institute of Systems Programming
Russian Academy of Sciences



__________________________________________________
Do You Yahoo!?
Tired of spam?  Yahoo! Mail has the best spam protection around 
http://mail.yahoo.com 

Reply via email to