Thread: Quering 4 or more physicale different PostgreSQL server at once

Quering 4 or more physicale different PostgreSQL server at once

From
Michelle Konzack
Date:
Hello,

My current PostgreSQL Database has arround 560 GByte without indices.
If not more then 3-5 user query it it works fast like Google...  :-)

But I have over 600 Users now which all query it over the Net and hit
a heavy performance problem.

The biggest problem is the "main" table which is arround 80% of the DB.

Now I have made a statistic and know, which Datas are most quried and
like to split the Database (exactly THE MONSTER-TABLE) over at least 4
Physicaly PostgreSQL servers.

The Database is generaly STATIC an Data will only be added and NEVER
removed or changed.  (If DATE or a ROW change, it will get a FULL NEW
ROW with version number)

Since I access the PostgreSQL from a pysicaly seperated dedicated
apache server using php5 I like to know from some Experts how to
query 4 or more Databases ar once and join the data.

I thinking, putting the main database with most requested data on the
first Server which will have arround 180 GByte and then the rest of
"THE" table on the other servers.

Note:  "THE" table is a historical timeline...
       and the datasets are not ordered in the table.

Greetings
    Michelle Konzack
    Systemadministrator
    Tamay Dogan Network
    Debian GNU/Linux Consultant


--
Linux-User #280138 with the Linux Counter, http://counter.li.org/
##################### Debian GNU/Linux Consultant #####################
Michelle Konzack   Apt. 917                  ICQ #328449886
                   50, rue de Soultz         MSM LinuxMichi
0033/6/61925193    67100 Strasbourg/France   IRC #Debian (irc.icq.com)

Attachment

Re: Quering 4 or more physicale different PostgreSQL server at once

From
Ben
Date:
You could look at contrib/dblink but if you plan to join between your four
servers a lot it seems like you might want cleaner solution, like, say,
slony-I or pgpool, and maybe some data partitioning on that main table as
well.

On Wed, 21 Feb 2007, Michelle Konzack wrote:

> Hello,
>
> My current PostgreSQL Database has arround 560 GByte without indices.
> If not more then 3-5 user query it it works fast like Google...  :-)
>
> But I have over 600 Users now which all query it over the Net and hit
> a heavy performance problem.
>
> The biggest problem is the "main" table which is arround 80% of the DB.
>
> Now I have made a statistic and know, which Datas are most quried and
> like to split the Database (exactly THE MONSTER-TABLE) over at least 4
> Physicaly PostgreSQL servers.
>
> The Database is generaly STATIC an Data will only be added and NEVER
> removed or changed.  (If DATE or a ROW change, it will get a FULL NEW
> ROW with version number)
>
> Since I access the PostgreSQL from a pysicaly seperated dedicated
> apache server using php5 I like to know from some Experts how to
> query 4 or more Databases ar once and join the data.
>
> I thinking, putting the main database with most requested data on the
> first Server which will have arround 180 GByte and then the rest of
> "THE" table on the other servers.
>
> Note:  "THE" table is a historical timeline...
>       and the datasets are not ordered in the table.
>
> Greetings
>    Michelle Konzack
>    Systemadministrator
>    Tamay Dogan Network
>    Debian GNU/Linux Consultant
>
>
> --
> Linux-User #280138 with the Linux Counter, http://counter.li.org/
> ##################### Debian GNU/Linux Consultant #####################
> Michelle Konzack   Apt. 917                  ICQ #328449886
>                   50, rue de Soultz         MSM LinuxMichi
> 0033/6/61925193    67100 Strasbourg/France   IRC #Debian (irc.icq.com)
>