Re: [HACKERS] Automatic free space map filling
Ühel kenal päeval, R, 2006-03-03 kell 11:37, kirjutas Tom Lane: Alvaro Herrera [EMAIL PROTECTED] writes: So for you it would certainly help a lot to be able to vacuum the first X pages of the big table, stop, release locks, create new transaction, continue with the next X pages, lather, rinse, repeat. This is perfectly doable, it only needs enough motivation from a knowledgeable person. Bruce and I were discussing this the other day; it'd be pretty easy to make plain VACUUM start a fresh transaction immediately after it finishes a scan heap/clean indexes/clean heap cycle. Do you mean the full (scan heap/clean indexes/clean heap) cycle or some smaller cycles inside each step ? If you mean the full cycle, then it is probably not worth it, as even a single 'clean index' pass can take hours on larger tables. The infrastructure for this (in particular, session-level locks that won't be lost by closing the xact) is all there. You'd have to figure out how often to start a new xact ... every cycle is probably too often, at least for smaller maintenance_work_mem settings ... but it'd not be hard or involve any strange changes in system semantics. --- Hannu ---(end of broadcast)--- TIP 1: if posting/reading through Usenet, please send an appropriate subscribe-nomail command to [EMAIL PROTECTED] so that your message can get through to the mailing list cleanly
Re: [HACKERS] Automatic free space map filling
Hannu Krosing [EMAIL PROTECTED] writes: If you mean the full cycle, then it is probably not worth it, as even a single 'clean index' pass can take hours on larger tables. The patch Heikki is working on will probably alleviate that problem, because it will allow vacuum to scan the indexes in physical rather than logical order. regards, tom lane ---(end of broadcast)--- TIP 1: if posting/reading through Usenet, please send an appropriate subscribe-nomail command to [EMAIL PROTECTED] so that your message can get through to the mailing list cleanly
Re: [HACKERS] Automatic free space map filling
On Fri, 2006-04-28 at 15:58 -0400, Bruce Momjian wrote: Tom Lane wrote: Alvaro Herrera [EMAIL PROTECTED] writes: So for you it would certainly help a lot to be able to vacuum the first X pages of the big table, stop, release locks, create new transaction, continue with the next X pages, lather, rinse, repeat. This is perfectly doable, it only needs enough motivation from a knowledgeable person. Bruce and I were discussing this the other day; it'd be pretty easy to make plain VACUUM start a fresh transaction immediately after it finishes a scan heap/clean indexes/clean heap cycle. The infrastructure for this (in particular, session-level locks that won't be lost by closing the xact) is all there. You'd have to figure out how often to start a new xact ... every cycle is probably too often, at least for smaller maintenance_work_mem settings ... but it'd not be hard or involve any strange changes in system semantics. Should this be a TODO? One item of discussion was taht people should just increase their workmem so the job can be done faster in larger batches. Yes, I think it should be a todo item. Csaba's point was that it was the duration a VACUUM transaction was held open that caused problems. Increasing maintenance_work_mem won't help with that problem. This would then allow a VACUUM to progress with a high vacuum_cost_delay without any ill effects elsewhere in the system. -- Simon Riggs EnterpriseDB http://www.enterprisedb.com ---(end of broadcast)--- TIP 6: explain analyze is your friend
Re: [HACKERS] Automatic free space map filling
On Mon, May 01, 2006 at 10:24:50PM +0200, Dawid Kuroczko wrote: VACUUM table WHERE some_col now()-'1 hour'::interval; I.e. Let vacuum run piggyback on some index. This would allow for a quick vacuum of a fraction of a large table. Especially when the table is large, and only some data (new data) are being modified. The vacuum for such a table would: 1. scan the index accoriding to the where criteria and create bitmap of blocks to look at. 2. go through these blocks and vacuum them. Hmm, another perhaps silly idea -- a special index kind for tracking tuple deaths. Ie -- something like whenever tuple is updated/deleted, insert an entry into such index, using last session the tuple is visible for as a key. Then, perhaps, vacuum could scan such an index and find tuples which are candidates for removal. I lack the knowledge of PostgreSQL's internals, so forgive me if I am writing something completely insane. :) There is a TODO to create a 'dead space map' which would cover #2 and probably eliminate any use for #1. -- Jim C. Nasby, Sr. Engineering Consultant [EMAIL PROTECTED] Pervasive Software http://pervasive.comwork: 512-231-6117 vcard: http://jim.nasby.net/pervasive.vcf cell: 512-569-9461 ---(end of broadcast)--- TIP 4: Have you searched our list archives? http://archives.postgresql.org
Re: [HACKERS] Automatic free space map filling
On Fri, Apr 28, 2006 at 03:58:16PM -0400, Bruce Momjian wrote: Tom Lane wrote: Alvaro Herrera [EMAIL PROTECTED] writes: So for you it would certainly help a lot to be able to vacuum the first X pages of the big table, stop, release locks, create new transaction, continue with the next X pages, lather, rinse, repeat. This is perfectly doable, it only needs enough motivation from a knowledgeable person. Bruce and I were discussing this the other day; it'd be pretty easy to make plain VACUUM start a fresh transaction immediately after it finishes a scan heap/clean indexes/clean heap cycle. The infrastructure for this (in particular, session-level locks that won't be lost by closing the xact) is all there. You'd have to figure out how often to start a new xact ... every cycle is probably too often, at least for smaller maintenance_work_mem settings ... but it'd not be hard or involve any strange changes in system semantics. Should this be a TODO? One item of discussion was taht people should just increase their workmem so the job can be done faster in larger batches. Except that wouldn't help when vacuuming a lot of small tables; each one would get it's own transaction. ISTM that tying this directly to maintenance_work_mem is a bit confusing, since the idea is to keep vacuum transaction duration down so that it isn't causing dead tuples to build up itself. It seems like it would be better to have vacuum start a fresh transaction after a certain number of tuples have died. But since there's no way to actually measure that without having row level stats turned on, maybe number of transactions or length of time would be good surrogates. Since it sounds like we'd want the transaction to start only at the start of a clean cycle it could just check the limits at the start of each cycle. That would prevent it from wrapping the vacuum of each small table with a (rather pointless) new transaction. -- Jim C. Nasby, Sr. Engineering Consultant [EMAIL PROTECTED] Pervasive Software http://pervasive.comwork: 512-231-6117 vcard: http://jim.nasby.net/pervasive.vcf cell: 512-569-9461 ---(end of broadcast)--- TIP 6: explain analyze is your friend
Re: [HACKERS] Automatic free space map filling
Jim C. Nasby [EMAIL PROTECTED] writes: Alvaro Herrera [EMAIL PROTECTED] writes: So for you it would certainly help a lot to be able to vacuum the first X pages of the big table, stop, release locks, create new transaction, continue with the next X pages, lather, rinse, repeat. Bruce and I were discussing this the other day; it'd be pretty easy to make plain VACUUM start a fresh transaction immediately after it finishes a scan heap/clean indexes/clean heap cycle. Except that wouldn't help when vacuuming a lot of small tables; each one would get it's own transaction. What's your point? There's only a problem for big tables, and VACUUM already does use a new transaction for each table. regards, tom lane ---(end of broadcast)--- TIP 4: Have you searched our list archives? http://archives.postgresql.org
Re: [HACKERS] Automatic free space map filling
On Mon, May 01, 2006 at 01:19:30PM -0500, Jim C. Nasby wrote: ISTM that tying this directly to maintenance_work_mem is a bit confusing, since the idea is to keep vacuum transaction duration down so that it isn't causing dead tuples to build up itself. It seems like it would be better to have vacuum start a fresh transaction after a certain number of tuples have died. But since there's no way to actually measure that without having row level stats turned on, maybe number of transactions or length of time would be good surrogates. AIUI, vacuum starts a fresh cycle because it's accumulated a certain number of dead tuples to clean up. Isn't that what you're asking for? maintenance_work_mem is the limit on the amount of deleted tuple information that can be stored (amongst other things I'm sure)... Since it sounds like we'd want the transaction to start only at the start of a clean cycle it could just check the limits at the start of each cycle. That would prevent it from wrapping the vacuum of each small table with a (rather pointless) new transaction. Every table has to be in its own transaction since thats the duration of the locks. Vacuum handling multiple tables in one transaction leaves you open to deadlocks. Have a nice day, -- Martijn van Oosterhout kleptog@svana.org http://svana.org/kleptog/ From each according to his ability. To each according to his ability to litigate. signature.asc Description: Digital signature
Re: [HACKERS] Automatic free space map filling
On 5/1/06, Martijn van Oosterhout kleptog@svana.org wrote: On Mon, May 01, 2006 at 01:19:30PM -0500, Jim C. Nasby wrote: ISTM that tying this directly to maintenance_work_mem is a bit confusing, since the idea is to keep vacuum transaction duration down so that it isn't causing dead tuples to build up itself. It seems like it would be better to have vacuum start a fresh transaction after a certain number of tuples have died. But since there's no way to actually measure that without having row level stats turned on, maybe number of transactions or length of time would be good surrogates.AIUI, vacuum starts a fresh cycle because it's accumulated a certainnumber of dead tuples to clean up. Isn't that what you're asking for?maintenance_work_mem is the limit on the amount of deleted tuple information that can be stored (amongst other things I'm sure)...Hmm, one idea, which may (or may not) be interesting for largetable vacuum is allowing a syntax similar to:VACUUM table WHERE some_col now()-'1 hour'::interval; I.e. Let vacuum run piggyback on some index. This would allowfor a quick vacuum of a fraction of a large table. Especially whenthe table is large, and only some data (new data) are being modified. The vacuum for such a table would:1. scan the index accoriding to the where criteria and create bitmap of blocks to look at.2. go through these blocks and vacuum them.Hmm, another perhaps silly idea -- a special index kind for tracking tuple deaths. Ie -- something like whenever tuple is updated/deleted,insert an entry into such index, using last session the tuple is visiblefor as a key. Then, perhaps, vacuum could scan such an index and find tuples which are candidates for removal. I lack the knowledge ofPostgreSQL's internals, so forgive me if I am writing somethingcompletely insane. :) Regards, Dawid
Re: [HACKERS] Automatic free space map filling
Tom Lane wrote: Alvaro Herrera [EMAIL PROTECTED] writes: So for you it would certainly help a lot to be able to vacuum the first X pages of the big table, stop, release locks, create new transaction, continue with the next X pages, lather, rinse, repeat. This is perfectly doable, it only needs enough motivation from a knowledgeable person. Bruce and I were discussing this the other day; it'd be pretty easy to make plain VACUUM start a fresh transaction immediately after it finishes a scan heap/clean indexes/clean heap cycle. The infrastructure for this (in particular, session-level locks that won't be lost by closing the xact) is all there. You'd have to figure out how often to start a new xact ... every cycle is probably too often, at least for smaller maintenance_work_mem settings ... but it'd not be hard or involve any strange changes in system semantics. Should this be a TODO? One item of discussion was taht people should just increase their workmem so the job can be done faster in larger batches. -- Bruce Momjian http://candle.pha.pa.us EnterpriseDBhttp://www.enterprisedb.com + If your life is a hard drive, Christ can be your backup. + ---(end of broadcast)--- TIP 5: don't forget to increase your free space map settings
Re: [PATCHES] [HACKERS] Automatic free space map filling
On Mon, 2006-03-13 at 17:38 +0900, ITAGAKI Takahiro wrote: Simon Riggs [EMAIL PROTECTED] wrote: Zeugswetter Andreas DCP SD [EMAIL PROTECTED] wrote: Ok, we cannot reuse a dead tuple. Maybe we can reuse the space of a dead tuple by reducing the tuple to it's header info. Attached patch realizes the concept of his idea. The dead tuples will be reduced to their headers are done by bgwriter. I'm interested in this patch but you need to say more about it. I get the general idea but it would be useful if you could give a full description of what this patch is trying to do and why. OK, I try to explain the patch. Excuse me for a long writing. OK. I'll take a look at this, thanks. Best Regards, Simon Riggs ---(end of broadcast)--- TIP 2: Don't 'kill -9' the postmaster
Re: [PATCHES] [HACKERS] Automatic free space map filling
Simon Riggs [EMAIL PROTECTED] wrote: Zeugswetter Andreas DCP SD [EMAIL PROTECTED] wrote: Ok, we cannot reuse a dead tuple. Maybe we can reuse the space of a dead tuple by reducing the tuple to it's header info. Attached patch realizes the concept of his idea. The dead tuples will be reduced to their headers are done by bgwriter. I'm interested in this patch but you need to say more about it. I get the general idea but it would be useful if you could give a full description of what this patch is trying to do and why. OK, I try to explain the patch. Excuse me for a long writing. * Purpose The basic idea is just reducing the dead tuple to it's header info, suggested by Andreas. This is a lightweight per-page sweeping to reduce the consumption of free space map and the necessity of VACUUM; i.e, normal VACUUM is still needed occasionally. I think it is useful on heavy-update workloads. It showed 5-10% of performance improvement on DBT-2 after 9 hours running *without* vacuum. I don't know whether it is still effective with well-scheduled vacuum. * Why does bgwriter do vacuum? Sweeping has cost, so non-backend process should do. Also, the page worth vacuum are almost always dirty, because tuples on the page are just updated or deleted. Bgwriter treats dirty pages, so I think it is a good place for sweeping. * Locking We must take super-exclusive-lock of the pages before vacuum. In the patch, bgwriter tries to take exclusive-lock before it writes a page, and does vacuum only if the lock is super-exclusive. Otherwise, it gives up and writes the pages normally. This is an optimistic way, but I assume the possibility is high because the most pages written by bgwriter are least recently used (LRU). * Keep the headers We cannot remove dead tuples completely in per-page sweep, because references to the tuples from indexes still remains. We might keep only line pointers (4 bytes), but it might lead line-pointer-bloat problems, (http://archives.postgresql.org/pgsql-hackers/2006-03/msg00116.php). so the headers (4+32 byte) should be left. * Other twists and GUC variables in the patch - Bgwriter cannot access the catalogs, so I added BM_RELATION hint bit to BufferDesc. Only relation pages will be swept. This is enabled by GUC variable 'bgvacuum_relation'. - I changed bgwriter_lru_maxpages to be adjusted automatically. Backends won't do vacuum not to disturb their processing, so bgwriter should write most of dirty pages. ('bgvacuum_autotune') - After sweepping, the page will be added to free space map. I made a simple replacement algorithm of free space map, that replaces the page with least spaces near the added one. ('bgvacuum_fsm') * Issues - If WAL is produced by sweeping a page, writing the page should be pended for a while, because flushing the WAL is needed before writing the page. - Bgwriter writes pages in 4 contexts, background-writes for LRU, ALL, checkpoint and shutdown. In current patch, pages are swept in 3 contexts except shutdown, but it may be better to do only on LRU. * Related discussions - Real-Time Vacuum Possibility (Rod Taylor) http://archives.postgresql.org/pgsql-hackers/2005-03/msg00518.php | have the bgwriter take a look at the pages it has, and see if it can do | any vacuum work based on pages it is about to send to disk - Pre-allocated free space for row updating (like PCTFREE) (Satoshi Nagayasu) http://archives.postgresql.org/pgsql-hackers/2005-08/msg01135.php | light-weight repairing on a single page is needed to maintain free space - Dead Space Map (Heikki Linnakangas) http://archives.postgresql.org/pgsql-hackers/2006-02/msg01125.php | vacuuming pages one by one as they're written by bgwriter Thank you for reading till the last. I'd like to hear your comments. --- ITAGAKI Takahiro NTT Cyber Space Laboratories ---(end of broadcast)--- TIP 9: In versions below 8.0, the planner will ignore your desire to choose an index scan if your joining column's datatypes do not match
Re: [HACKERS] Automatic free space map filling
Zeugswetter Andreas DCP SD [EMAIL PROTECTED] wrote: Ok, we cannot reuse a dead tuple. Maybe we can reuse the space of a dead tuple by reducing the tuple to it's header info. I was just working about your idea. In my work, bgwriter truncates dead tuples and leaves only their headers. I'll send a concept patch to PATCHES. We must take super-exclusive-lock of pages before vacuum. Bgwriter tries to take exclusive-lock before it writes a page, and does vacuum only if the lock is super-exclusive. Otherwise, it gives up and writes normally. This is an optimistic way, but I assume the possibility is high because the most pages written by bgwriter are least recently used (LRU). Also, I changed bgwriter_lru_maxpages to be adjusted automatically, because backends won't do vacuum not to disturb main transaction processing, so bgwriter should write most of the dirty pages. There are much room for discussion on this idea. Comments are welcome. --- ITAGAKI Takahiro NTT Cyber Space Laboratories ---(end of broadcast)--- TIP 3: Have you checked our extensive FAQ? http://www.postgresql.org/docs/faq
Re: [HACKERS] Automatic free space map filling
But you could do the indexes first and remember how far you can vacuum the heap later. But the indexes _can't_ be done first; you _first_ need to know which tuples are dead, which requires looking at the table itself. If we already had the all tuples visible bitmap I think we could first scan the bitmap and decide whether we can afford to look at the visibility info for each entry in the index. We only collect the ctids before so we don't have the inefficient lookups, but if we can avoid the lookup in most cases it would again be attractive. Andreas ---(end of broadcast)--- TIP 4: Have you searched our list archives? http://archives.postgresql.org
Re: [HACKERS] Automatic free space map filling
Zeugswetter Andreas DCP SD wrote: But you could do the indexes first and remember how far you can vacuum the heap later. But the indexes _can't_ be done first; you _first_ need to know which tuples are dead, which requires looking at the table itself. If we already had the all tuples visible bitmap I think we could first scan the bitmap and decide whether we can afford to look at the visibility info for each entry in the index. We only collect the ctids before so we don't have the inefficient lookups, but if we can avoid the lookup in most cases it would again be attractive. The problem is that index to heap lookups are very slow. -- Bruce Momjian http://candle.pha.pa.us SRA OSS, Inc. http://www.sraoss.com + If your life is a hard drive, Christ can be your backup. + ---(end of broadcast)--- TIP 2: Don't 'kill -9' the postmaster
Re: [HACKERS] Automatic free space map filling
Jim C. Nasby wrote: ... how many pages per bit ... Are we trying to set up a complex solution to a problem that'll be mostly moot once partitioning is easier and partitioned tables are common? In many cases I can think of the bulk of the data would be in old partitions that are practically never written to (so would need no vacuuming and could always use index-only lookups); while the hot parts of large tables would be on partitions that would need frequent vacuuming and wouldn't benefit from index-only lookups. In these cases, 1 bit per partition would work well, and seems a lot easier to keep track of than bits-per-page. ---(end of broadcast)--- TIP 9: In versions below 8.0, the planner will ignore your desire to choose an index scan if your joining column's datatypes do not match
Re: [HACKERS] Automatic free space map filling
Are you running 8.1? If so, you can use autovacuum and set per table thresholds (read vacuum aggressivly) and per table cost delay settings so that the performance impact is minimal. If you have tried 8.1 autovacuum and found it unhelpful, I would be curious to find out why. Yes, I'm running 8.1, and I've set up per table auto-vacuum settings :-) And I lowered the general thresholds too. Generally autovacuum is very useful from my POV, and in particular the per table settings are so. But the problem I have is not the performance impact of the vacuum itself, but the impact of the long running transaction of vacuuming big tables. I do have big tables which are frequently updated and small tables which are basically queue tables, so each inserted row will be updated a few times and then deleted. Those queue tables tend to get huge unvacuumable dead space during any long running transaction, and vacuum on the big tables is such a long running transaction. And I have a few of them, and one is in particular very busy (a task table, all activities go through that one). Now when the queue tables get 1000 times dead space compared to their normal size, I get performance problems. So tweaking vacuum cost delay doesn't buy me anything, as not vacuum per se is the performance problem, it's long run time for big tables is. Cheers, Csaba. ---(end of broadcast)--- TIP 3: Have you checked our extensive FAQ? http://www.postgresql.org/docs/faq
Re: [HACKERS] Automatic free space map filling
Csaba Nagy wrote: Now when the queue tables get 1000 times dead space compared to their normal size, I get performance problems. So tweaking vacuum cost delay doesn't buy me anything, as not vacuum per se is the performance problem, it's long run time for big tables is. So for you it would certainly help a lot to be able to vacuum the first X pages of the big table, stop, release locks, create new transaction, continue with the next X pages, lather, rinse, repeat. This is perfectly doable, it only needs enough motivation from a knowledgeable person. -- Alvaro Herrerahttp://www.CommandPrompt.com/ PostgreSQL Replication, Consulting, Custom Development, 24x7 support ---(end of broadcast)--- TIP 9: In versions below 8.0, the planner will ignore your desire to choose an index scan if your joining column's datatypes do not match
Re: [HACKERS] Automatic free space map filling
Alvaro Herrera wrote: Csaba Nagy wrote: Now when the queue tables get 1000 times dead space compared to their normal size, I get performance problems. So tweaking vacuum cost delay doesn't buy me anything, as not vacuum per se is the performance problem, it's long run time for big tables is. So for you it would certainly help a lot to be able to vacuum the first X pages of the big table, stop, release locks, create new transaction, continue with the next X pages, lather, rinse, repeat. But what about index clearing? When do you scan each index? -- Bruce Momjian http://candle.pha.pa.us SRA OSS, Inc. http://www.sraoss.com + If your life is a hard drive, Christ can be your backup. + ---(end of broadcast)--- TIP 6: explain analyze is your friend
Re: [HACKERS] Automatic free space map filling
Bruce Momjian wrote: Alvaro Herrera wrote: Csaba Nagy wrote: Now when the queue tables get 1000 times dead space compared to their normal size, I get performance problems. So tweaking vacuum cost delay doesn't buy me anything, as not vacuum per se is the performance problem, it's long run time for big tables is. So for you it would certainly help a lot to be able to vacuum the first X pages of the big table, stop, release locks, create new transaction, continue with the next X pages, lather, rinse, repeat. But what about index clearing? When do you scan each index? At the end of each iteration (or earlier, depending on maintenance_work_mem). So for each iteration you would need to scan the indexes. Maybe we could make maintenance_work_mem be the deciding factor; after scanning the indexes, do the release/reacquire locks cycle. -- Alvaro Herrerahttp://www.CommandPrompt.com/ The PostgreSQL Company - Command Prompt, Inc. ---(end of broadcast)--- TIP 2: Don't 'kill -9' the postmaster
Re: [HACKERS] Automatic free space map filling
On Fri, Mar 03, 2006 at 11:40:40AM -0300, Alvaro Herrera wrote: Csaba Nagy wrote: Now when the queue tables get 1000 times dead space compared to their normal size, I get performance problems. So tweaking vacuum cost delay doesn't buy me anything, as not vacuum per se is the performance problem, it's long run time for big tables is. So for you it would certainly help a lot to be able to vacuum the first X pages of the big table, stop, release locks, create new transaction, continue with the next X pages, lather, rinse, repeat. I think the issue is that even for that small section, you still need to scan all the indexes to delete the tuples there. So you actually cause more work because you have to scan the indexes for each portion of the table rather than just at the end. However, if this were combined with some optimistic index deletion code where the tuple was used to find the entry directly rather than via bulkdelete, maybe it'd be doable. More overall I/O due to the index lookups but the transactions become shorter. I say optimistic because if you don't find the tuple the quick way you can always queue it for a bulkdelete later. Hopefully it will be the uncommon case. Have a nice day, -- Martijn van Oosterhout kleptog@svana.org http://svana.org/kleptog/ Patent. n. Genius is 5% inspiration and 95% perspiration. A patent is a tool for doing 5% of the work and then sitting around waiting for someone else to do the other 95% so you can sue them. signature.asc Description: Digital signature
Re: [HACKERS] Automatic free space map filling
Alvaro Herrera wrote: Bruce Momjian wrote: Alvaro Herrera wrote: Csaba Nagy wrote: Now when the queue tables get 1000 times dead space compared to their normal size, I get performance problems. So tweaking vacuum cost delay doesn't buy me anything, as not vacuum per se is the performance problem, it's long run time for big tables is. So for you it would certainly help a lot to be able to vacuum the first X pages of the big table, stop, release locks, create new transaction, continue with the next X pages, lather, rinse, repeat. But what about index clearing? When do you scan each index? At the end of each iteration (or earlier, depending on maintenance_work_mem). So for each iteration you would need to scan the indexes. Maybe we could make maintenance_work_mem be the deciding factor; after scanning the indexes, do the release/reacquire locks cycle. Ewe. How expensive is scanning an index compared to the heap? Does anyone have figure on that in terms of I/O and time? -- Bruce Momjian http://candle.pha.pa.us SRA OSS, Inc. http://www.sraoss.com + If your life is a hard drive, Christ can be your backup. + ---(end of broadcast)--- TIP 2: Don't 'kill -9' the postmaster
Re: [HACKERS] Automatic free space map filling
Ewe. How expensive is scanning an index compared to the heap? Does anyone have figure on that in terms of I/O and time? See this post for an example: http://archives.postgresql.org/pgsql-performance/2006-02/msg00416.php For my 200 million table, scanning the pk index took ~ 4 hours. And then there are some more indexes... So if the index has to be scanned completely, that's still too much. Cheers, Csaba. ---(end of broadcast)--- TIP 9: In versions below 8.0, the planner will ignore your desire to choose an index scan if your joining column's datatypes do not match
Re: [HACKERS] Automatic free space map filling
But what about index clearing? When do you scan each index? At the end of each iteration (or earlier, depending on maintenance_work_mem). So for each iteration you would need to scan the indexes. Maybe we could make maintenance_work_mem be the deciding factor; after scanning the indexes, do the release/reacquire locks cycle. But you could do the indexes first and remember how far you can vacuum the heap later. So you might as well do each index separately first and remember how far you can go with the heap for each one. Then do the heap with a special restriction that comes from what you remembered from the indexes. You can now separate the heap vacuum in arbitrarily large transactions, since the indexes are already taken care of. (You only vacuum to the point of the eldest vacuumed index) Andreas ---(end of broadcast)--- TIP 2: Don't 'kill -9' the postmaster
Re: [HACKERS] Automatic free space map filling
Alvaro Herrera [EMAIL PROTECTED] writes: So for you it would certainly help a lot to be able to vacuum the first X pages of the big table, stop, release locks, create new transaction, continue with the next X pages, lather, rinse, repeat. This is perfectly doable, it only needs enough motivation from a knowledgeable person. Bruce and I were discussing this the other day; it'd be pretty easy to make plain VACUUM start a fresh transaction immediately after it finishes a scan heap/clean indexes/clean heap cycle. The infrastructure for this (in particular, session-level locks that won't be lost by closing the xact) is all there. You'd have to figure out how often to start a new xact ... every cycle is probably too often, at least for smaller maintenance_work_mem settings ... but it'd not be hard or involve any strange changes in system semantics. regards, tom lane ---(end of broadcast)--- TIP 4: Have you searched our list archives? http://archives.postgresql.org
Re: [HACKERS] Automatic free space map filling
Tom Lane wrote: Alvaro Herrera [EMAIL PROTECTED] writes: So for you it would certainly help a lot to be able to vacuum the first X pages of the big table, stop, release locks, create new transaction, continue with the next X pages, lather, rinse, repeat. This is perfectly doable, it only needs enough motivation from a knowledgeable person. Bruce and I were discussing this the other day; it'd be pretty easy to make plain VACUUM start a fresh transaction immediately after it finishes a scan heap/clean indexes/clean heap cycle. The infrastructure for this (in particular, session-level locks that won't be lost by closing the xact) is all there. You'd have to figure out how often to start a new xact ... every cycle is probably too often, at least for smaller maintenance_work_mem settings ... but it'd not be hard or involve any strange changes in system semantics. Oh, reading the original posting, these are cases where maintenance_work_mem is full and we are going to rescan the indexes multiple times anyway for this table. -- Bruce Momjian http://candle.pha.pa.us SRA OSS, Inc. http://www.sraoss.com + If your life is a hard drive, Christ can be your backup. + ---(end of broadcast)--- TIP 1: if posting/reading through Usenet, please send an appropriate subscribe-nomail command to [EMAIL PROTECTED] so that your message can get through to the mailing list cleanly
Re: [HACKERS] Automatic free space map filling
Alvaro Herrera wrote: Csaba Nagy wrote: Now when the queue tables get 1000 times dead space compared to their normal size, I get performance problems. So tweaking vacuum cost delay doesn't buy me anything, as not vacuum per se is the performance problem, it's long run time for big tables is. So for you it would certainly help a lot to be able to vacuum the first X pages of the big table, stop, release locks, create new transaction, continue with the next X pages, lather, rinse, repeat. I got the impression that Csaba is looking more for multiple simultaneous vacuum more than the partial vacuum. Not sure the best way to set this up, but perhaps a flag in the pg_autovacuum table that says vacuum this table even if there is another vacuum running that way you can control things and not have autovacuum firing off lots of vacuums at the same time. Sounds to me that these frequently updated queue tables need to be monitored closely and not ignored for a long period of time because we are vacuuming another table. Has anyone looked more closely at the multiple vacuum patch that was submitted to the patches list a while ago? Matt ---(end of broadcast)--- TIP 9: In versions below 8.0, the planner will ignore your desire to choose an index scan if your joining column's datatypes do not match
Re: [HACKERS] Automatic free space map filling
Matthew T. O'Connor wrote: Alvaro Herrera wrote: Csaba Nagy wrote: Now when the queue tables get 1000 times dead space compared to their normal size, I get performance problems. So tweaking vacuum cost delay doesn't buy me anything, as not vacuum per se is the performance problem, it's long run time for big tables is. So for you it would certainly help a lot to be able to vacuum the first X pages of the big table, stop, release locks, create new transaction, continue with the next X pages, lather, rinse, repeat. I got the impression that Csaba is looking more for multiple simultaneous vacuum more than the partial vacuum. So he rather needs Hannu Krosing's patch for simultaneous vacuum ... -- Alvaro Herrerahttp://www.CommandPrompt.com/ The PostgreSQL Company - Command Prompt, Inc. ---(end of broadcast)--- TIP 6: explain analyze is your friend
Re: [HACKERS] Automatic free space map filling
I got the impression that Csaba is looking more for multiple simultaneous vacuum more than the partial vacuum. So he rather needs Hannu Krosing's patch for simultaneous vacuum ... Well, I guess that would be a good solution to the queue table problem. The problem is that I can't deploy that patch on our production systems without being fairly sure it won't corrupt any data... and I can't rely on non-production testing either. Basically I'm waiting to see Tom saying it will fly :-) Cheers, Csaba. ---(end of broadcast)--- TIP 1: if posting/reading through Usenet, please send an appropriate subscribe-nomail command to [EMAIL PROTECTED] so that your message can get through to the mailing list cleanly
Re: [HACKERS] Automatic free space map filling
Csaba Nagy wrote: So he rather needs Hannu Krosing's patch for simultaneous vacuum ... Well, I guess that would be a good solution to the queue table problem. The problem is that I can't deploy that patch on our production systems without being fairly sure it won't corrupt any data... and I can't rely on non-production testing either. Basically I'm waiting to see Tom saying it will fly :-) That patch is a step forward if it's deemed OK by the powers that be. However, autovacuum would still need to be taught to handle simultaneous vacuums. I suppose that in the interim, you could disable autovacuum for the problematic queue table and have cron issue a manual vacuum command for that table at the required frequency. Anyone up for working on / testing / improving Hannu's patch? I think it's beyond my skill set. Matt ---(end of broadcast)--- TIP 2: Don't 'kill -9' the postmaster
Re: [HACKERS] Automatic free space map filling
Matthew T. O'Connor matthew@zeut.net writes: That patch is a step forward if it's deemed OK by the powers that be. However, autovacuum would still need to be taught to handle simultaneous vacuums. I suppose that in the interim, you could disable autovacuum for the problematic queue table and have cron issue a manual vacuum command for that table at the required frequency. I'm not sure you should think of that as an interim solution. I don't really like the idea of multiple autovacuums running concurrently. ISTM autovac is intended to be something that lurks in the background and doesn't take up an unreasonable percentage of your system bandwidth ... but if there's more than one of them, it's going to be mighty hard to control the overall load penalty. Plus you have to worry about keeping them off each others' backs, ie, not all trying to vac the same table at once. And in a scenario like Csaba's, I think the hotspot tables are just exactly what they'd all try to vacuum. For small hotspot tables I think a scheduled vacuum process is just the thing, whereas autovac is more of a free-lance thing to keep the rest of your DB in line. regards, tom lane ---(end of broadcast)--- TIP 6: explain analyze is your friend
Re: [HACKERS] Automatic free space map filling
Tom Lane wrote: Matthew T. O'Connor matthew@zeut.net writes: That patch is a step forward if it's deemed OK by the powers that be. However, autovacuum would still need to be taught to handle simultaneous vacuums. I suppose that in the interim, you could disable autovacuum for the problematic queue table and have cron issue a manual vacuum command for that table at the required frequency. I'm not sure you should think of that as an interim solution. I don't really like the idea of multiple autovacuums running concurrently. ISTM autovac is intended to be something that lurks in the background and doesn't take up an unreasonable percentage of your system bandwidth ... but if there's more than one of them, it's going to be mighty hard to control the overall load penalty. Plus you have to worry about keeping them off each others' backs, ie, not all trying to vac the same table at once. And in a scenario like Csaba's, I think the hotspot tables are just exactly what they'd all try to vacuum. For small hotspot tables I think a scheduled vacuum process is just the thing, whereas autovac is more of a free-lance thing to keep the rest of your DB in line. While I agree that given the current state of affairs the cron solution is elegant, I personally want autovac to solve all of our vacuuming needs, I really dislike the idea of requiring a cron based solution to solve a fairly typical problem. Besides the cron solution is sloppy, it blindly vacuums whether it's needed or not resulting in a net increase of cycles spent vacuuming. Anyway, I don't know the best way to implement it but I wasn't thinking of just firing off multiple autovac processes. I was envisioning something like an autovacuum master process that launches (forks?) VACUUM commands and has some smarts about how many processes to fire off, or that it would only fire off simultaneous VACUUMS for tables that have been flagged as hot spot tables. I recognize that teaching autovac to handle simultaneous VACUUM's in a sane way will require a quantum leap of complexity but it still seems a better long term solution. I would agree that using cron makes sense if we were seeing lots of different scenarios that we couldn't possibly anticipate, but I don't think that is where we are. BTW, this discussion is only relevant if we allow simultaneous vacuum. Is this something you see as inevitable whether or not you think Hannu's implementation is acceptable. Matt ---(end of broadcast)--- TIP 1: if posting/reading through Usenet, please send an appropriate subscribe-nomail command to [EMAIL PROTECTED] so that your message can get through to the mailing list cleanly
Re: [HACKERS] Automatic free space map filling
On Thu, Mar 02, 2006 at 10:05:28AM -0500, Tom Lane wrote: Hannu Krosing [EMAIL PROTECTED] writes: ??hel kenal p??eval, N, 2006-03-02 kell 09:53, kirjutas Zeugswetter Andreas DCP SD: Ok, we cannot reuse a dead tuple. Maybe we can reuse the space of a dead tuple by reducing the tuple to it's header info. Andreas' idea is possibly doable but I am not sure that I see the point. It does not reduce the need for vacuum nor the I/O load imposed by vacuum. What it does do is bias the system in the direction of allocating an unreasonably large number of tuple line pointers on a page (ie, more than are useful when the page is fully packed with normal tuples). Since we never reclaim such pointers, over time all the pages in a table would tend to develop line-pointer-bloat. I don't know what the net overhead would be, but it'd definitely impose some aggregate inefficiency. What would be involved in reclaiming item pointer space? Is there any reason it's not done today? (I know I've been bit once by this...) -- Jim C. Nasby, Sr. Engineering Consultant [EMAIL PROTECTED] Pervasive Software http://pervasive.comwork: 512-231-6117 vcard: http://jim.nasby.net/pervasive.vcf cell: 512-569-9461 ---(end of broadcast)--- TIP 2: Don't 'kill -9' the postmaster
Re: [HACKERS] Automatic free space map filling
On Thu, Mar 02, 2006 at 03:19:46PM +0100, Martijn van Oosterhout wrote: Note, for this purpose you don't need to keep a bit per page. The OS I/O system will load 64k+ (8+ pages) in one go so one bit per 8 pages would be sufficient. AFAIK that's entirely dependant on the filesystem and how it's created (and possibly the OS as well). So arbitrarily deciding each bit is 8 pages is a bad idea. I could see allowing for a setting that determins how many pages per bit, though, but I think we're also getting ahead of ourselves. -- Jim C. Nasby, Sr. Engineering Consultant [EMAIL PROTECTED] Pervasive Software http://pervasive.comwork: 512-231-6117 vcard: http://jim.nasby.net/pervasive.vcf cell: 512-569-9461 ---(end of broadcast)--- TIP 3: Have you checked our extensive FAQ? http://www.postgresql.org/docs/faq
Re: [HACKERS] Automatic free space map filling
On Fri, Mar 03, 2006 at 04:14:41PM +0100, Csaba Nagy wrote: Ewe. How expensive is scanning an index compared to the heap? Does anyone have figure on that in terms of I/O and time? See this post for an example: http://archives.postgresql.org/pgsql-performance/2006-02/msg00416.php For my 200 million table, scanning the pk index took ~ 4 hours. And then there are some more indexes... So if the index has to be scanned completely, that's still too much. But how does a scan of the index compare to a scan of the table? For example, if indexes are 1/5th the size of the table, you can (theoretically) scan 5 indexes in the same amount of time it takes to scan the heap. That indicates to me that even if we did have to scan all indexes, a dirty page bitmap would still be a win over the current situation. But it appears that it should be safe to do index lookups on indexes that aren't expressions. And I believe that we could take steps down the road to allow for index lookups on indexes that only used functions that were known to be safe. -- Jim C. Nasby, Sr. Engineering Consultant [EMAIL PROTECTED] Pervasive Software http://pervasive.comwork: 512-231-6117 vcard: http://jim.nasby.net/pervasive.vcf cell: 512-569-9461 ---(end of broadcast)--- TIP 9: In versions below 8.0, the planner will ignore your desire to choose an index scan if your joining column's datatypes do not match
Re: [HACKERS] Automatic free space map filling
On Fri, Mar 03, 2006 at 11:37:00AM -0500, Tom Lane wrote: Bruce and I were discussing this the other day; it'd be pretty easy to make plain VACUUM start a fresh transaction immediately after it finishes a scan heap/clean indexes/clean heap cycle. The infrastructure for this (in particular, session-level locks that won't be lost by closing the xact) is all there. You'd have to figure out how often to start a new xact ... every cycle is probably too often, at least for smaller maintenance_work_mem settings ... but it'd not be hard or If maintenance_work_mem is small you're likely to have poor performance anyway; I'm suspicious that the overhead of starting a new xact would be all that important. If you care about performance, you'll probably have increased maintenance_work_mem anyway. -- Jim C. Nasby, Sr. Engineering Consultant [EMAIL PROTECTED] Pervasive Software http://pervasive.comwork: 512-231-6117 vcard: http://jim.nasby.net/pervasive.vcf cell: 512-569-9461 ---(end of broadcast)--- TIP 3: Have you checked our extensive FAQ? http://www.postgresql.org/docs/faq
Re: [HACKERS] Automatic free space map filling
Centuries ago, Nostradamus foresaw when [EMAIL PROTECTED] (Zeugswetter Andreas DCP SD) would write: But what about index clearing? When do you scan each index? At the end of each iteration (or earlier, depending on maintenance_work_mem). So for each iteration you would need to scan the indexes. Maybe we could make maintenance_work_mem be the deciding factor; after scanning the indexes, do the release/reacquire locks cycle. But you could do the indexes first and remember how far you can vacuum the heap later. But the indexes _can't_ be done first; you _first_ need to know which tuples are dead, which requires looking at the table itself. -- select 'cbbrowne' || '@' || 'gmail.com'; http://linuxdatabases.info/info/languages.html Pound for pound, the amoeba is the most vicious animal on earth. ---(end of broadcast)--- TIP 4: Have you searched our list archives? http://archives.postgresql.org
Re: [HACKERS] Automatic free space map filling
I thought we had sufficiently destroyed that reuse a tuple meme yesterday. You can't do that: there are too many aspects of the system design that are predicated on the assumption that dead tuples do not come back to life. You have to do the full vacuuming bit (index entry removal, super-exclusive page locking, etc) before you can remove a dead tuple. One more idea I would like to throw in. Ok, we cannot reuse a dead tuple. Maybe we can reuse the space of a dead tuple by reducing the tuple to it's header info. (If you still wanted to be able to locate index entries fast, you would need to keep indexed columns, but I think we agreed that there is no real use) I think that would be achievable at reasonable cost (since you can avoid one page IO) on the page of the currently active tuple (the first page that is considered). On this page: if freespace available -- use it elsif freespace available after reducing all dead rows -- use the freespace with a new slot else Of course this only works when we still have free slots, but I think that might not really be an issue. Andreas ---(end of broadcast)--- TIP 5: don't forget to increase your free space map settings
Re: [HACKERS] Automatic free space map filling
Ühel kenal päeval, N, 2006-03-02 kell 09:53, kirjutas Zeugswetter Andreas DCP SD: I thought we had sufficiently destroyed that reuse a tuple meme yesterday. You can't do that: there are too many aspects of the system design that are predicated on the assumption that dead tuples do not come back to life. You have to do the full vacuuming bit (index entry removal, super-exclusive page locking, etc) before you can remove a dead tuple. One more idea I would like to throw in. Ok, we cannot reuse a dead tuple. Maybe we can reuse the space of a dead tuple by reducing the tuple to it's header info. (If you still wanted to be able to locate index entries fast, you would need to keep indexed columns, but I think we agreed that there is no real use) I don't even think you need the header, just truncate the slot to be 0-size (the next pointer is the same as this one or make the pointer point to unaligned byte or smth) and detect this condition when accessing tuples. this would add on compare to all accesse to the tuple, but I suspect that mostly it is a noop performance-wise as all data needed is already available in level1 cache. This would decouple declaring a tuple to be dead/reuse data space and final cleanup/free index space. Hannu ---(end of broadcast)--- TIP 3: Have you checked our extensive FAQ? http://www.postgresql.org/docs/faq
Re: [HACKERS] Automatic free space map filling
[sorry to everyone if that mail arrives multiple times, but i had some odd problems with my mail gateway yesterday...] On Wed, Mar 01, 2006 at 12:41:01PM -0500, Tom Lane wrote: Peter Eisentraut [EMAIL PROTECTED] writes: Tom Lane wrote: How does an optimistic FSM entry avoid the need to run vacuum? It ensures that all freed tuples are already in the FSM. That has nothing to do with it, because the space isn't actually free for re-use until vacuum deletes the tuple. But couldn't such an opportunistic approach be used for another lightweight VACUUM mode in such a way, that VACUUM could look at a special Hot Spot queue, which represents potential candidates for freeing? Let's call it a 2-phase VACUUMthis would avoid a constant long running VACUUM run on big tables, e.g. when tuples gets updated (or deleted) frequently. Just an idea... Bernd ---(end of broadcast)--- TIP 4: Have you searched our list archives? http://archives.postgresql.org
Re: [HACKERS] Automatic free space map filling
On Wed, Mar 01, 2006 at 12:41:01PM -0500, Tom Lane wrote: Peter Eisentraut [EMAIL PROTECTED] writes: Tom Lane wrote: How does an optimistic FSM entry avoid the need to run vacuum? It ensures that all freed tuples are already in the FSM. That has nothing to do with it, because the space isn't actually free for re-use until vacuum deletes the tuple. Hmm, but couldn't such an opportunistic approach be used for another leightweight VACUUM mode in such a way, that VACUUM could look at a special Hot Spot queue, which represents potential candidates for freeing? Let's call it a 2-phase VACUUMthis would avoid a long running VACUUM run on big tables, e.g. when tuples gets updated (or deleted) frequently. Just an idea... Bernd ---(end of broadcast)--- TIP 4: Have you searched our list archives? http://archives.postgresql.org
Re: [HACKERS] Automatic free space map filling
Centuries ago, Nostradamus foresaw when [EMAIL PROTECTED] (Tom Lane) would write: I thought we had sufficiently destroyed that reuse a tuple meme yesterday. You can't do that: there are too many aspects of the system design that are predicated on the assumption that dead tuples do not come back to life. This discussion needs to come up again in October when the zombie movies come out :-). That's the other problem: it's not apparent why pushing work from vacuum back into foreground processing is a good idea. Especially not why retail vacuuming of individual tuples will be better than wholesale. What is unclear to me in the discussion is whether or not this is invalidating the item on the TODO list... --- Create a bitmap of pages that need vacuuming Instead of sequentially scanning the entire table, have the background writer or some other process record pages that have expired rows, then VACUUM can look at just those pages rather than the entire table. In the event of a system crash, the bitmap would probably be invalidated. One complexity is that index entries still have to be vacuumed, and doing this without an index scan (by using the heap values to find the index entry) might be slow and unreliable, especially for user-defined index functions. --- It strikes me as a non-starter to draw vacuum work directly into the foreground; there is a *clear* loss in that the death of the tuple can't actually take place at that point, due to MVCC and the fact that it is likely that other transactions will be present, keeping the tuple from being destroyed. But it would *seem* attractive to do what is in the TODO, above. Alas, the user defined index functions make cleanout of indexes much more troublesome :-(. But what's in the TODO is still wholesale, albeit involving more targetted selling than the usual Kirby VACUUM :-). -- select 'cbbrowne' || '@' || 'gmail.com'; http://linuxdatabases.info/info/rdbms.html Rules of the Evil Overlord #140. I will instruct my guards when checking a cell that appears empty to look for the chamber pot. If the chamber pot is still there, then the prisoner has escaped and they may enter and search for clues. If the chamber pot is not there, then either the prisoner is perched above the lintel waiting to strike them with it or else he decided to take it as a souvenir (in which case he is obviously deeply disturbed and poses no threat). Either way, there's no point in entering. http://www.eviloverlord.com/ ---(end of broadcast)--- TIP 3: Have you checked our extensive FAQ? http://www.postgresql.org/docs/faq
Re: [HACKERS] Automatic free space map filling
On Thu, Mar 02, 2006 at 08:33:46AM -0500, Christopher Browne wrote: What is unclear to me in the discussion is whether or not this is invalidating the item on the TODO list... --- Create a bitmap of pages that need vacuuming snip I think this is doable, and not invalidated by anything said so far. All this is changeing is whether to scan the whole table or just the bits changed. Unfortunatly I don't think you can avoid scanning the indexes :(. Note, for this purpose you don't need to keep a bit per page. The OS I/O system will load 64k+ (8+ pages) in one go so one bit per 8 pages would be sufficient. The inverse is keep a list of pages where we know all tuples are visible to everyone. I'm not sure if this can be done race condition free. ISTM it would be possible to get the new Bitmap Index Scans to avoid checking visiblity straight away but wait until it has been AND/OR'd with other bitmaps and only at the end checking visibility. But maybe that already happens... Have a nice day, -- Martijn van Oosterhout kleptog@svana.org http://svana.org/kleptog/ Patent. n. Genius is 5% inspiration and 95% perspiration. A patent is a tool for doing 5% of the work and then sitting around waiting for someone else to do the other 95% so you can sue them. signature.asc Description: Digital signature
Re: [HACKERS] Automatic free space map filling
Hannu Krosing [EMAIL PROTECTED] writes: Ãhel kenal päeval, N, 2006-03-02 kell 09:53, kirjutas Zeugswetter Andreas DCP SD: Ok, we cannot reuse a dead tuple. Maybe we can reuse the space of a dead tuple by reducing the tuple to it's header info. I don't even think you need the header, just truncate the slot to be 0-size I think you must keep the header because the tuple might be part of an update chain (cf vacuuming bugs we repaired just a few months ago). t_ctid is potentially interesting data even in a certainly-dead tuple. Andreas' idea is possibly doable but I am not sure that I see the point. It does not reduce the need for vacuum nor the I/O load imposed by vacuum. What it does do is bias the system in the direction of allocating an unreasonably large number of tuple line pointers on a page (ie, more than are useful when the page is fully packed with normal tuples). Since we never reclaim such pointers, over time all the pages in a table would tend to develop line-pointer-bloat. I don't know what the net overhead would be, but it'd definitely impose some aggregate inefficiency. regards, tom lane ---(end of broadcast)--- TIP 3: Have you checked our extensive FAQ? http://www.postgresql.org/docs/faq
Re: [HACKERS] Automatic free space map filling
Bernd Helmle [EMAIL PROTECTED] writes: But couldn't such an opportunistic approach be used for another lightweight VACUUM mode in such a way, that VACUUM could look at a special Hot Spot queue, which represents potential candidates for freeing? The proposed dirty-page bit map seems a superior solution to that. regards, tom lane ---(end of broadcast)--- TIP 4: Have you searched our list archives? http://archives.postgresql.org
Re: [HACKERS] Automatic free space map filling
Christopher Browne [EMAIL PROTECTED] writes: What is unclear to me in the discussion is whether or not this is invalidating the item on the TODO list... No, I don't think any of this is an argument against the dirty-page-bitmap idea. The amount of foreground effort needed to set a dirty-page bit is minimal (maybe even zero, if we can make the bgwriter do it, though I'm pretty suspicious of that idea because I think it needs to be done immediately when the page is dirtied). I don't see the dirty-page bitmap as changing the way that VACUUM works in any fundamental respect --- it will just allow the vacuum process to skip reading pages that certainly don't need to change. One point that does need to be considered though is what about anti-wraparound processing (ie, replacing old XIDs with FrozenXID before they wrap around)? VACUUM currently is a safe way to handle that, but if its normal mode of operation stops looking at every tuple then we're going to have an issue there. regards, tom lane ---(end of broadcast)--- TIP 9: In versions below 8.0, the planner will ignore your desire to choose an index scan if your joining column's datatypes do not match
Re: [HACKERS] Automatic free space map filling
Tom Lane wrote: Christopher Browne [EMAIL PROTECTED] writes: What is unclear to me in the discussion is whether or not this is invalidating the item on the TODO list... No, I don't think any of this is an argument against the dirty-page-bitmap idea. The amount of foreground effort needed to set a dirty-page bit is minimal (maybe even zero, if we can make the bgwriter do it, though I'm pretty suspicious of that idea because I think it needs to be done immediately when the page is dirtied). I don't see the dirty-page bitmap as changing the way that VACUUM works in any fundamental respect --- it will just allow the vacuum process to skip reading pages that certainly don't need to change. See the email I just posted. I am questioning how big a win it is to skip heap pages if we have to sequentially scan all indexes. One point that does need to be considered though is what about anti-wraparound processing (ie, replacing old XIDs with FrozenXID before they wrap around)? VACUUM currently is a safe way to handle that, but if its normal mode of operation stops looking at every tuple then we're going to have an issue there. We would need to do sequential scan occasionally and somehow track that. -- Bruce Momjian http://candle.pha.pa.us SRA OSS, Inc. http://www.sraoss.com + If your life is a hard drive, Christ can be your backup. + ---(end of broadcast)--- TIP 9: In versions below 8.0, the planner will ignore your desire to choose an index scan if your joining column's datatypes do not match
Re: [HACKERS] Automatic free space map filling
Christopher Browne wrote: What is unclear to me in the discussion is whether or not this is invalidating the item on the TODO list... --- Create a bitmap of pages that need vacuuming Instead of sequentially scanning the entire table, have the background writer or some other process record pages that have expired rows, then VACUUM can look at just those pages rather than the entire table. In the event of a system crash, the bitmap would probably be invalidated. One complexity is that index entries still have to be vacuumed, and doing this without an index scan (by using the heap values to find the index entry) might be slow and unreliable, especially for user-defined index functions. --- It strikes me as a non-starter to draw vacuum work directly into the foreground; there is a *clear* loss in that the death of the tuple can't actually take place at that point, due to MVCC and the fact that it is likely that other transactions will be present, keeping the tuple from being destroyed. But it would *seem* attractive to do what is in the TODO, above. Alas, the user defined index functions make cleanout of indexes much more troublesome :-(. But what's in the TODO is still wholesale, albeit involving more targetted selling than the usual Kirby VACUUM :-). What bothers me about the TODO item is that if we have to sequentially scan indexes, are we really gaining much by not having to sequentially scan the heap? If the heap is large enough to gain from a bitmap, the index is going to be large too. Is disabling per-index cleanout for expression indexes the answer? The entire expression index problem is outlined in this thread: http://archives.postgresql.org/pgsql-hackers/2006-02/msg01127.php I don't think it is a show-stopper because if we fail to find the index that matches the heap, we know we have a problem and can report it and fall back to an index scan. Anyway, as I remember, if you have a 20gig table, a vacuum / sequential scan is painful, but if we have to sequential scan the all indexes, that is probably just as painful. If we can't make headway there and we can't cleanout indexes without an sequential index scan, I think we should just remove the TODO item and give up on improving vacuum performance. For the bitmaps, index-only scans require a bit that says all page tuples are visible while vacuum wants some tuples are expired. DELETE would clear both bits, while INSERT would clear just the first, and update is a mix of INSERT and UPDATE, though perhaps on different pages. -- Bruce Momjian http://candle.pha.pa.us SRA OSS, Inc. http://www.sraoss.com + If your life is a hard drive, Christ can be your backup. + ---(end of broadcast)--- TIP 6: explain analyze is your friend
Re: [HACKERS] Automatic free space map filling
What bothers me about the TODO item is that if we have to sequentially scan indexes, are we really gaining much by not having to sequentially scan the heap? If the heap is large enough to gain from a bitmap, the index is going to be large too. Is disabling per-index cleanout for expression indexes the answer? I guess you're saying that full index scan should only be done when the index is a functional one, and use index lookup for safe indexes ? That would be a huge win for most of my vacuum-problematic tables, as I don't have any functional indexes. But I guess full index scan would still be faster if the percentage of pages changed is more than some threshold. On the other hand it would allow very frequent vacuuming even for huge tables so that situation should not occur. Autovacuum thresholds could be lowered drastically in that case... Anyway, as I remember, if you have a 20gig table, a vacuum / sequential scan is painful, but if we have to sequential scan the all indexes, that is probably just as painful. If we can't make headway there and we can't cleanout indexes without an sequential index scan, I think we should just remove the TODO item and give up on improving vacuum performance. From my POV, there must be a way to speed up vacuums on huge tables and small percentage of to-be-vacuumed tuples... a 200 million rows table with frequent updates of the _same_ record is causing me some pain right now. I would like to have that table vacuumed as often as possible, but right now it only works to do it once per week due to load problems on long-running transactions preventing vacuuming other tables. Cheers, Csaba. ---(end of broadcast)--- TIP 2: Don't 'kill -9' the postmaster
Re: [HACKERS] Automatic free space map filling
Csaba Nagy wrote: What bothers me about the TODO item is that if we have to sequentially scan indexes, are we really gaining much by not having to sequentially scan the heap? If the heap is large enough to gain from a bitmap, the index is going to be large too. Is disabling per-index cleanout for expression indexes the answer? I guess you're saying that full index scan should only be done when the index is a functional one, and use index lookup for safe indexes ? That would be a huge win for most of my vacuum-problematic tables, as I don't have any functional indexes. But I guess full index scan would still be faster if the percentage of pages changed is more than some threshold. On the other hand it would allow very frequent vacuuming even for huge tables so that situation should not occur. Autovacuum thresholds could be lowered drastically in that case... Right. Another idea would be to remove the heap space held by expired rows, but to keep the tid slot in place because it is pointed to by an index. The index entry could be recycled by a later vacuum index scan, or if an index lookup finds such an entry. Because of multiple indexes, I don't think the tid slot can be removed except by sequential index scans of all indexes. There is also the concern that updating the single-page bitmap will cause contention by multiple sessions modifing a table. I am thinking as long as we have to sequential-scan every index, we aren't going to improve vacuum performance dramatically. If the bitmap adds contention, and it is only a marginal improvement, it might not be a win. The bitmap can be a win, but I think we have to think more boldly to ensure it is a win. -- Bruce Momjian http://candle.pha.pa.us SRA OSS, Inc. http://www.sraoss.com + If your life is a hard drive, Christ can be your backup. + ---(end of broadcast)--- TIP 1: if posting/reading through Usenet, please send an appropriate subscribe-nomail command to [EMAIL PROTECTED] so that your message can get through to the mailing list cleanly
Re: [HACKERS] Automatic free space map filling
I think you must keep the header because the tuple might be part of an update chain (cf vacuuming bugs we repaired just a few months ago). t_ctid is potentially interesting data even in a certainly-dead tuple. yes, I'd still want to keep the full header. Andreas' idea is possibly doable but I am not sure that I see the point. It does not reduce the need for vacuum nor the I/O load imposed by vacuum. What it does do is bias the system in the direction of allocating an unreasonably large number of tuple line pointers on a page (ie, more than are useful when the page is fully packed with normal tuples). Since we never reclaim such pointers, over time all the pages in a table would tend to develop line-pointer-bloat. I don't know what the net overhead would be, but it'd definitely impose some aggregate inefficiency. Ok, for vacuum the slot would look like any other dead row and thus be target for removal. Why do we not truncate the line pointer array ? Is it, that vacuum (not the full version) does not move rows to other pages or slots ? Of course vacuum full could do it, but I see your point. Maybe we could impose an upper limit on the number of slots to allow, after which the optimization is turned off. But this starts to sound not so good :-( Andreas ---(end of broadcast)--- TIP 2: Don't 'kill -9' the postmaster
Re: [HACKERS] Automatic free space map filling
Zeugswetter Andreas DCP SD [EMAIL PROTECTED] writes: Why do we not truncate the line pointer array ? Is it, that vacuum (not the full version) does not move rows to other pages or slots ? Of course vacuum full could do it, but I see your point. We can't reassign tuple TIDs safely except in vacuum full. It's possible that a plain vacuum could safely truncate off unused line pointers at the end of the array, but in the absence of a forcing function to make those pointers become unused, I'm not sure it'd help much. regards, tom lane ---(end of broadcast)--- TIP 6: explain analyze is your friend
Re: [HACKERS] Automatic free space map filling
Csaba Nagy wrote From my POV, there must be a way to speed up vacuums on huge tables and small percentage of to-be-vacuumed tuples... a 200 million rows table with frequent updates of the _same_ record is causing me some pain right now. I would like to have that table vacuumed as often as possible, but right now it only works to do it once per week due to load problems on long-running transactions preventing vacuuming other tables. Are you running 8.1? If so, you can use autovacuum and set per table thresholds (read vacuum aggressivly) and per table cost delay settings so that the performance impact is minimal. If you have tried 8.1 autovacuum and found it unhelpful, I would be curious to find out why. Matt ---(end of broadcast)--- TIP 9: In versions below 8.0, the planner will ignore your desire to choose an index scan if your joining column's datatypes do not match
Re: [HACKERS] Automatic free space map filling
Am Montag, 27. Februar 2006 19:42 schrieb Tom Lane: The free-space map is not the hard part of the problem. You still have to VACUUM --- that is, wait until the dead tuple is not only committed dead but is certainly dead to all onlooker transactions, and then remove its index entries as well as the tuple itself. The first part of this makes it impossible for a transaction to be responsible for vacuuming its own detritus. I'm not sure if I made myself clear. The idea is that you fill the free-space map early with opportunitistic entries in the hope that most updates and deletes go through soon. That is, these entries will be invalid for a short time but hopefully by the time another write looks at them, the entries will have become valid. That way you don't actually have to run vacuum on these deleted rows. -- Peter Eisentraut http://developer.postgresql.org/~petere/ ---(end of broadcast)--- TIP 4: Have you searched our list archives? http://archives.postgresql.org
Re: [HACKERS] Automatic free space map filling
Peter Eisentraut [EMAIL PROTECTED] writes: I'm not sure if I made myself clear. The idea is that you fill the free-space map early with opportunitistic entries in the hope that most updates and deletes go through soon. That is, these entries will be invalid for a short time but hopefully by the time another write looks at them, the entries will have become valid. That way you don't actually have to run vacuum on these deleted rows. How does an optimistic FSM entry avoid the need to run vacuum? All that will happen is that some backend will visit the page and not find usable free space. regards, tom lane ---(end of broadcast)--- TIP 4: Have you searched our list archives? http://archives.postgresql.org
Re: [HACKERS] Automatic free space map filling
Tom Lane wrote: Peter Eisentraut [EMAIL PROTECTED] writes: I'm not sure if I made myself clear. The idea is that you fill the free-space map early with opportunitistic entries in the hope that most updates and deletes go through soon. That is, these entries will be invalid for a short time but hopefully by the time another write looks at them, the entries will have become valid. That way you don't actually have to run vacuum on these deleted rows. How does an optimistic FSM entry avoid the need to run vacuum? All that will happen is that some backend will visit the page and not find usable free space. Because the index isn't removed, right? That index thing is what usually kills us. -- Bruce Momjian http://candle.pha.pa.us SRA OSS, Inc. http://www.sraoss.com + If your life is a hard drive, Christ can be your backup. + ---(end of broadcast)--- TIP 3: Have you checked our extensive FAQ? http://www.postgresql.org/docs/faq
Re: [HACKERS] Automatic free space map filling
Tom Lane wrote: How does an optimistic FSM entry avoid the need to run vacuum? It ensures that all freed tuples are already in the FSM. -- Peter Eisentraut http://developer.postgresql.org/~petere/ ---(end of broadcast)--- TIP 3: Have you checked our extensive FAQ? http://www.postgresql.org/docs/faq
Re: [HACKERS] Automatic free space map filling
Peter Eisentraut [EMAIL PROTECTED] writes: Tom Lane wrote: How does an optimistic FSM entry avoid the need to run vacuum? It ensures that all freed tuples are already in the FSM. That has nothing to do with it, because the space isn't actually free for re-use until vacuum deletes the tuple. regards, tom lane ---(end of broadcast)--- TIP 1: if posting/reading through Usenet, please send an appropriate subscribe-nomail command to [EMAIL PROTECTED] so that your message can get through to the mailing list cleanly
Re: [HACKERS] Automatic free space map filling
Tom Lane wrote: Peter Eisentraut [EMAIL PROTECTED] writes: Tom Lane wrote: How does an optimistic FSM entry avoid the need to run vacuum? It ensures that all freed tuples are already in the FSM. That has nothing to do with it, because the space isn't actually free for re-use until vacuum deletes the tuple. I think the idea is a different free space map of sorts, whereby a transaction that obsoletes a tuple puts its block number in that map. A transaction that inserts a new tuple goes to the FSM. If nothing is found, it then goes to the new map. A block returned from that map is then scanned and any tuple that's no longer visible for anyone is reused. The problem with this idea is scanning the block and for each tuple determine if it's alive. Essentially, we would be folding the find dead tuples and compress page logic, which is currently in vacuum, back to insert. IMHO this is unacceptable from a performance PoV. -- Alvaro Herrerahttp://www.CommandPrompt.com/ PostgreSQL Replication, Consulting, Custom Development, 24x7 support ---(end of broadcast)--- TIP 2: Don't 'kill -9' the postmaster
Re: [HACKERS] Automatic free space map filling
Alvaro Herrera [EMAIL PROTECTED] writes: Tom Lane wrote: That has nothing to do with it, because the space isn't actually free for re-use until vacuum deletes the tuple. I think the idea is a different free space map of sorts, whereby a transaction that obsoletes a tuple puts its block number in that map. A transaction that inserts a new tuple goes to the FSM. If nothing is found, it then goes to the new map. A block returned from that map is then scanned and any tuple that's no longer visible for anyone is reused. I thought we had sufficiently destroyed that reuse a tuple meme yesterday. You can't do that: there are too many aspects of the system design that are predicated on the assumption that dead tuples do not come back to life. You have to do the full vacuuming bit (index entry removal, super-exclusive page locking, etc) before you can remove a dead tuple. Essentially, we would be folding the find dead tuples and compress page logic, which is currently in vacuum, back to insert. IMHO this is unacceptable from a performance PoV. That's the other problem: it's not apparent why pushing work from vacuum back into foreground processing is a good idea. Especially not why retail vacuuming of individual tuples will be better than wholesale. regards, tom lane ---(end of broadcast)--- TIP 1: if posting/reading through Usenet, please send an appropriate subscribe-nomail command to [EMAIL PROTECTED] so that your message can get through to the mailing list cleanly
Re: [HACKERS] Automatic free space map filling
On Thu, Mar 02, 2006 at 01:01:21AM -0500, Tom Lane wrote: Essentially, we would be folding the find dead tuples and compress page logic, which is currently in vacuum, back to insert. IMHO this is unacceptable from a performance PoV. That's the other problem: it's not apparent why pushing work from vacuum back into foreground processing is a good idea. Especially not why retail vacuuming of individual tuples will be better than wholesale. The problem is that even with vacuum_cost_delay, vacuum is still very slow and problematic in situations such as a large tables in a heavy transaction environment. Anything that could help reduce the need for 'traditional' vacuuming could well be a win. Even so, I think the most productive path to pursue at this time is a dead-space-map/known-clean-map. Either one is almost guaranteed to provide benefits. Once we know what good they do we can move forward from there with further improvements. -- Jim C. Nasby, Sr. Engineering Consultant [EMAIL PROTECTED] Pervasive Software http://pervasive.comwork: 512-231-6117 vcard: http://jim.nasby.net/pervasive.vcf cell: 512-569-9461 ---(end of broadcast)--- TIP 1: if posting/reading through Usenet, please send an appropriate subscribe-nomail command to [EMAIL PROTECTED] so that your message can get through to the mailing list cleanly
Re: [HACKERS] Automatic free space map filling
Ühel kenal päeval, E, 2006-02-27 kell 19:20, kirjutas Peter Eisentraut: Something came to my mind today, I'm not sure if it's feasible but I would like to know opinions on it. We've seen database applications that PostgreSQL simply could not manage because one would have to vacuum continuously. What's wrong with vacuuminng continuously ? I am running an application, that in fact does vacuum continuously without any ill effects. A case when things become compliacted, is when you have one huge table (say 50.000.000 rows) that is updated at a moderate rate and needs an occasional vacuum + a fast-update table, which needs continuous vacuum. Due to current implementation of vacuum, you have to abandon continuous vacuuming during vacuum of bigtable, but i have written and submitted to patches list a patch which allows vacuums not to block each other out, this is stalled due to Tom's unesyness about its possible hidden effects, but it should be available from patches list to anyone in distress :p Perhaps in those situations one could arrange it that an update (or delete) of a row registers the space in the free space map right away, on the assumption that by the time it is up for reuse, the transaction will likely have committed. Naturally, this would need to be secured in some way, for example a maybe bit in the FSM itself or simply checking that the supposed free space is really free before using it, perhaps combined with a timeout (don't consider until 5 seconds from now). Unfortunately transactions have no knowledge about wallclock time :( I think with applications that have a more or less constant data volume Hannu ---(end of broadcast)--- TIP 5: don't forget to increase your free space map settings
Re: [HACKERS] Automatic free space map filling
Hannu Krosing wrote: Due to current implementation of vacuum, you have to abandon continuous vacuuming during vacuum of bigtable, but i have written and submitted to patches list a patch which allows vacuums not to block each other out, this is stalled due to Tom's unesyness about its possible hidden effects, but it should be available from patches list to anyone in distress :p Do you use it in production? Have you noticed any ill effects? -- Alvaro Herrerahttp://www.CommandPrompt.com/ The PostgreSQL Company - Command Prompt, Inc. ---(end of broadcast)--- TIP 2: Don't 'kill -9' the postmaster
[HACKERS] Automatic free space map filling
Something came to my mind today, I'm not sure if it's feasible but I would like to know opinions on it. We've seen database applications that PostgreSQL simply could not manage because one would have to vacuum continuously. Perhaps in those situations one could arrange it that an update (or delete) of a row registers the space in the free space map right away, on the assumption that by the time it is up for reuse, the transaction will likely have committed. Naturally, this would need to be secured in some way, for example a maybe bit in the FSM itself or simply checking that the supposed free space is really free before using it, perhaps combined with a timeout (don't consider until 5 seconds from now). I think with applications that have a more or less constant data volume but update that data a lot, this could assure constant disk space usage (even if it's only a constant factor above the ideal usage) without any vacuuming. Comments? -- Peter Eisentraut http://developer.postgresql.org/~petere/ ---(end of broadcast)--- TIP 2: Don't 'kill -9' the postmaster
Re: [HACKERS] Automatic free space map filling
Peter Eisentraut [EMAIL PROTECTED] writes: We've seen database applications that PostgreSQL simply could not manage because one would have to vacuum continuously. Perhaps in those situations one could arrange it that an update (or delete) of a row registers the space in the free space map right away, on the assumption that by the time it is up for reuse, the transaction will likely have committed. The free-space map is not the hard part of the problem. You still have to VACUUM --- that is, wait until the dead tuple is not only committed dead but is certainly dead to all onlooker transactions, and then remove its index entries as well as the tuple itself. The first part of this makes it impossible for a transaction to be responsible for vacuuming its own detritus. Naturally, this would need to be secured in some way, The FSM is only a hint anyway --- if it points someone to a page that in reality does not have adequate free space, nothing bad happens except for the wasted cycles to visit the page and find that out. See the loop in RelationGetBufferForTuple(). regards, tom lane ---(end of broadcast)--- TIP 4: Have you searched our list archives? http://archives.postgresql.org