Eric Davies <[EMAIL PROTECTED]> writes:

>    Back in the days of 7.4.2, we tried storing large blobs  (1GB+) in
>    postgres but found them too slow because the blob was being chopped
>    into 2K rows stored in some other table.
>    However, it has occurred to us that if it was possible to configure
>    the server to split blobs into bigger pieces, say 32K,  our speed
>    problems might diminish correspondingly.
>    Is there a compile time constant or a run time configuration entry
>    that accomplish this?

I *think* the limit would be 8k (the size of a PG page) even if you
could change it.  Upping that would require recompiling with PAGE_SIZE
set larger, which would have a lot of other consequences.

-Doug

---------------------------(end of broadcast)---------------------------
TIP 6: explain analyze is your friend

Reply via email to