Centuries ago, Nostradamus foresaw when peter_e@gmx.net (Peter Eisentraut) would write:
> Is there any practical limit on the number of parallel connections that a
> PostgreSQL server can service? We're in the process of setting up a system
> that will require up to 10000 connections open in parallel. The query load
> is not the problem, but we're wondering about the number of connections.
> Does anyone have experience with these kinds of numbers?
We commonly have a thousand connections open, on some servers, and
while it works, we consider there to be something problematic about
it. It tends to lead to using spinlocks a lot.
You might want to look into pgpool:
<http://www2b.biglobe.ne.jp/~caco/pgpool/index-e.html>
Jan Wieck has tried it out with his version of the TPC-W benchmark,
and found that it allowed cutting down on the _true_ number of
connections, and was very helpful in improving performance under
conditions where the application imagined it needed a lot of
connections.
--
(reverse (concatenate 'string "gro.gultn" "@" "enworbbc"))
http://www.ntlug.org/~cbbrowne/spiritual.html
"The last good thing written in C was Franz Schubert's Symphony number
9." -- Erwin Dieterich