Tom Lane wrote:

One objection to this is that after moving "off the gold standard" of
1.0 = one page fetch, there is no longer any clear meaning to the
cost estimate units; you're faced with the fact that they're just an
arbitrary scale.  I'm not sure that's such a bad thing, though.

It seems to me the appropriate gold standard is Time, in microseconds
or milliseconds.

The default postgresql.conf can come with a set of hardcoded
values that reasonably approximate some real-world system; and
if that's documented in the file someone reading it can say
 "hey, my CPU's about the same but my disk subsystem is much
 faster, so I know in which direction to change things".
And another person may say "ooh, now I know that my 4GHz
machines should have about twice the number here as my 2GHz
box".

For people who *really* care a lot (HW vendors?), they could
eventually make measurements on their systems.

---------------------------(end of broadcast)---------------------------
TIP 1: if posting/reading through Usenet, please send an appropriate
      subscribe-nomail command to [EMAIL PROTECTED] so that your
      message can get through to the mailing list cleanly

Reply via email to