Eric Davies <Eric@barrodale.com> writes:
> Back in the days of 7.4.2, we tried storing large blobs (1GB+) in
> postgres but found them too slow because the blob was being chopped
> into 2K rows stored in some other table.
> However, it has occurred to us that if it was possible to configure
> the server to split blobs into bigger pieces, say 32K, our speed
> problems might diminish correspondingly.
> Is there a compile time constant or a run time configuration entry
> that accomplish this?
I *think* the limit would be 8k (the size of a PG page) even if you
could change it. Upping that would require recompiling with PAGE_SIZE
set larger, which would have a lot of other consequences.
-Doug