On Fri, May 16, 2008 at 11:50 AM, Roberto Edwins
<[EMAIL PROTECTED]> wrote:
> What are the values that should go in postgresql.conf for a larga
> database, a couple of which tables shall contain over 100 million
> records?

Well, that really kind of depends.  Is this a database to handle lots
of users with small bank style transactions?  Or is it designed to
hold the text type content for millions of users but with only a
handful accessing it at once?  Or maybe it's a reporting database with
lots of statistical data that will be accessed by one or two people at
a time but do huge ugly queries.

It really all kind of depends.

A good idea is to keep from setting max clients any higher than you
have to.  Set shared_buffers around 25% of total memory to start, and
set work_mem to about 8M or so.  Note that it's easy to use up a lot
of memory fast with higher work_mem settings because each sort / hash
agg etc can use up work_mem worth of memory.  100 users times 2 sorts
each would equal 8M*200 or 1600M with a setting of 8M work_mem.

Most importantly, make small incremental changes and benchmark with a
realistic load, it's the only way to be sure.

-- 
Sent via pgsql-admin mailing list (pgsql-admin@postgresql.org)
To make changes to your subscription:
http://www.postgresql.org/mailpref/pgsql-admin

Reply via email to