Re: replication choices - Mailing list pgsql-general

From Bruce Momjian
Subject Re: replication choices
Date
Msg-id 200701252019.l0PKJYW15733@momjian.us
Whole thread Raw
In response to replication choices  (Ben <bench@silentmedia.com>)
Responses Re: replication choices
List pgsql-general
Have you read the 8.2 documentation about this:

    http://www.postgresql.org/docs/8.2/static/high-availability.html

---------------------------------------------------------------------------

Ben wrote:
> Hi guys. I've inherited a system that I'm looking to add replication to.
> It already has some custom replication code, but it's being nice to say
> that code less than good. I'm hoping there's an existing project out there
> that will work much better. Unfortunately, I'm not seeing anything that
> obviously fits my needs, so maybe somebody here can suggest something.
>
> I've got a single cluster in the datacenter and dozens of remote sites.
> Many of these sites are on unreliable connections to the internet, and
> while they're online more often then not, when their network will go down
> isn't known, and even when it's up, the network isn't that fast.
>
> A vast majority of activity occurs at these remote sites, with very little
> at the datacenter cluster. That said, the datacenter cluster needs to keep
> pretty good copies of most (but not all) of the data at each site.
> Obviously the network unrealiability puts a limit on how up to date the
> datacenter can be, but loosing data is considered Bad. So, for instance,
> restoring the daily backup of each site at the datacenter is too
> infrequent.
>
> Each site will replicate to its own schema in the datacenter cluster, so I
> don't *think* I need a multi-master solution.... but at the same time,
> because data will be coming from multiple sites, simply replaying WAL
> files at the datacenter won't work.
>
> In addition, there will be some data changes made at the datacenter that
> will need to replicate to all of the remote sites as soon as they're
> online. It's ok if data being replicated from the datacenter ends up in a
> different schema at the remote sites than the schema which holds the data
> that will be replicated back to the datacenter.
>
> My current best guess of what to do is create a global schema at every
> database, a local schema at each site, and a schema for each site at the
> datacenter. Then I can use Slony to replicate the global schema from the
> datacenter to each site, and again use Slony to replicate the local schema
> from each site to that site's schema in the datacenter. But I'm not too
> familiar with Slony, and from what I understand, using Slony with bad
> networks leads to bad problems. I'm also not sure that Slony supports
> replicating from multiple sources to the same postgres install, even if
> each replication process is writing to a different schema.
>
> Are there any better options? Or is my Slony idea not so bad?
>
> ---------------------------(end of broadcast)---------------------------
> TIP 6: explain analyze is your friend

--
  Bruce Momjian   bruce@momjian.us
  EnterpriseDB    http://www.enterprisedb.com

  + If your life is a hard drive, Christ can be your backup. +

pgsql-general by date:

Previous
From: Ben
Date:
Subject: replication choices
Next
From: "Merlin Moncure"
Date:
Subject: Re: loop plpgsql recordset variable