Re: WIP patch for parallel pg_dump - Mailing list pgsql-hackers

From Andrew Dunstan
Subject Re: WIP patch for parallel pg_dump
Date
Msg-id 4CF83808.2010307@dunslane.net
Whole thread Raw
In response to Re: WIP patch for parallel pg_dump  (Robert Haas <robertmhaas@gmail.com>)
Responses Re: WIP patch for parallel pg_dump  (Robert Haas <robertmhaas@gmail.com>)
List pgsql-hackers

On 12/02/2010 07:13 PM, Robert Haas wrote:
> On Thu, Dec 2, 2010 at 5:32 PM, Tom Lane<tgl@sss.pgh.pa.us>  wrote:
>> Andrew Dunstan<andrew@dunslane.net>  writes:
>>> On 12/02/2010 05:01 PM, Tom Lane wrote:
>>>> In the past, proposals for this have always been rejected on the grounds
>>>> that it's impossible to assure a consistent dump if different
>>>> connections are used to read different tables.  I fail to understand
>>>> why that consideration can be allowed to go by the wayside now.
>>> Well, snapshot cloning should allow that objection to be overcome, no?
>> Possibly, but we need to see that patch first not second.
> Yes, by all means let's allow the perfect to be the enemy of the good.
>

That seems like a bit of an easy shot. Requiring that parallel pg_dump 
produce a dump that is as consistent as non-parallel pg_dump currently 
produces isn't unreasonable. It's not stopping us moving forward, it's 
just not wanting to go backwards.

And it shouldn't be terribly hard. IIRC Joachim has already done some 
work on it.

cheers

andrew


pgsql-hackers by date:

Previous
From: Josh Berkus
Date:
Subject: Re: should we set hint bits without dirtying the page?
Next
From: Bruce Momjian
Date:
Subject: Re: DELETE with LIMIT (or my first hack)