You could look at contrib/dblink but if you plan to join between your four
servers a lot it seems like you might want cleaner solution, like, say,
slony-I or pgpool, and maybe some data partitioning on that main table as
well.
On Wed, 21 Feb 2007, Michelle Konzack wrote:
> Hello,
>
> My current PostgreSQL Database has arround 560 GByte without indices.
> If not more then 3-5 user query it it works fast like Google... :-)
>
> But I have over 600 Users now which all query it over the Net and hit
> a heavy performance problem.
>
> The biggest problem is the "main" table which is arround 80% of the DB.
>
> Now I have made a statistic and know, which Datas are most quried and
> like to split the Database (exactly THE MONSTER-TABLE) over at least 4
> Physicaly PostgreSQL servers.
>
> The Database is generaly STATIC an Data will only be added and NEVER
> removed or changed. (If DATE or a ROW change, it will get a FULL NEW
> ROW with version number)
>
> Since I access the PostgreSQL from a pysicaly seperated dedicated
> apache server using php5 I like to know from some Experts how to
> query 4 or more Databases ar once and join the data.
>
> I thinking, putting the main database with most requested data on the
> first Server which will have arround 180 GByte and then the rest of
> "THE" table on the other servers.
>
> Note: "THE" table is a historical timeline...
> and the datasets are not ordered in the table.
>
> Greetings
> Michelle Konzack
> Systemadministrator
> Tamay Dogan Network
> Debian GNU/Linux Consultant
>
>
> --
> Linux-User #280138 with the Linux Counter, http://counter.li.org/
> ##################### Debian GNU/Linux Consultant #####################
> Michelle Konzack Apt. 917 ICQ #328449886
> 50, rue de Soultz MSM LinuxMichi
> 0033/6/61925193 67100 Strasbourg/France IRC #Debian (irc.icq.com)
>