On Tue, Nov 16, 2010 at 11:12 AM, Alvaro Herrera
<alvhe...@alvh.no-ip.org> wrote:
> Magnus was just talking to me about having a better way of controlling
> memory usage on autovacuum.  Instead of each worker using up to
> maintenance_work_mem, which ends up as a disaster when DBA A sets to a
> large value and DBA B raises autovacuum_max_workers, we could simply
> have an "autovacuum_maintenance_memory" setting (name TBD), that defines
> the maximum amount of memory that autovacuum is going to use regardless
> of the number of workers.
>
> So for the initial implementation, we could just have each worker set
> its local maintenance_work_mem to autovacuum_maintenance_memory / max_workers.
> That way there's never excessive memory usage.
>
> This implementation is not ideal, because most of the time they wouldn't
> use that much memory, and so vacuums could be slower.  But I think it's
> better than what we currently have.
>
> Thoughts?

I'm a little skeptical about creating more memory tunables.  DBAs who
are used to previous versions of PG will find that their vacuum is now
really slow, because they adjusted maintenance_work_mem but not this
new parameter.  If we could divide up the vacuum memory intelligently
between the workers in some way, that would be a win.  But just
creating a different variable that controls the same thing in
different units doesn't seem to add much.

-- 
Robert Haas
EnterpriseDB: http://www.enterprisedb.com
The Enterprise PostgreSQL Company

-- 
Sent via pgsql-hackers mailing list (pgsql-hackers@postgresql.org)
To make changes to your subscription:
http://www.postgresql.org/mailpref/pgsql-hackers

Reply via email to