Thread: pg_dump --split patch

pg_dump --split patch

From
Joel Jacobson
Date:
Dear fellow hackers,

Problem: A normal diff of two slightly different schema dump files (pg_dump -s), will not produce a user-friendly diff, as you get all changes in the same file.

Solution: I propose a new option to pg_dump, --split, which dumps each object to a separate file in a user friendly directory structure:

[-f filename] : main dump file, imports each splitted part using \i
[-f filename]-split/[desc]/[tag]/[oid].sql : dump of the oid

Example: If the filename (-f) is "pg.dump", the following directory structure would be created:

$ ./pg_dump -f /crypt/pg.dump --split -F p -s glue
/crypt/pg.dump-split/VIEW/
/crypt/pg.dump-split/TYPE/
/crypt/pg.dump-split/TRIGGER/
/crypt/pg.dump-split/TABLE/
/crypt/pg.dump-split/SEQUENCE/
/crypt/pg.dump-split/SCHEMA/
/crypt/pg.dump-split/PROCEDURAL_LANGUAGE/
/crypt/pg.dump-split/INDEX/
/crypt/pg.dump-split/FUNCTION/
/crypt/pg.dump-split/FK_CONSTRAINT/
/crypt/pg.dump-split/CONSTRAINT/
/crypt/pg.dump-split/AGGREGATE/

In each such directory, one directory per object name is created.
If we would have a function "foobar" with oid "12345" it would be saved to:
/crypt/pg.dump-split/FUNCTION/foobar/12345.sql

In the "pg.dump" plain text file, the files are "linked in" using the "\i" psql command, e.g.:
\i /crypt/pg.dump-split/FUNCTION/foobar/12345.sql

Potential use-case scenarios:

*) Version control your database schema, by exporting it daily (using --split) and commiting the differences.

*) Compare differences of schema dumps created in different points in time. Since objects are stored in separate files, it is easier to see what areas were modified, compared to looking at the diff of two entire schemas.

*) Restore only some objects, based on type (e.g., only the functions) or name (e.g. only fucntions of certain name/names).

I've tested the patch for both the latest HEAD (9.1devel) as well as 8.4.6.

Feedback welcome.

--
Best regards,

Joel Jacobson
Glue Finance
Attachment

Re: pg_dump --split patch

From
Tom Lane
Date:
Joel Jacobson <joel@gluefinance.com> writes:
> Dear fellow hackers,
> Problem: A normal diff of two slightly different schema dump files (pg_dump
> -s), will not produce a user-friendly diff, as you get all changes in the
> same file.

> Solution: I propose a new option to pg_dump, --split, which dumps each
> object to a separate file in a user friendly directory structure:

Um ... how does that solve the claimed problem exactly?

> [-f filename] : main dump file, imports each splitted part using \i
> [-f filename]-split/[desc]/[tag]/[oid].sql : dump of the oid

This particular choice seems remarkably *un* friendly, since two dumps
from different DBs will inevitably not share the same OIDs, making it
practically impossible to compare them even if they are logically
identical.  But even without the choice to use OIDs in the filenames
I'm unconvinced that file-per-object is a good idea in any way shape or
form.
        regards, tom lane


Re: pg_dump --split patch

From
Joel Jacobson
Date:


2010/12/28 Tom Lane <tgl@sss.pgh.pa.us>
Joel Jacobson <joel@gluefinance.com> writes:
> Dear fellow hackers,
> Problem: A normal diff of two slightly different schema dump files (pg_dump
> -s), will not produce a user-friendly diff, as you get all changes in the
> same file.

> Solution: I propose a new option to pg_dump, --split, which dumps each
> object to a separate file in a user friendly directory structure:

Um ... how does that solve the claimed problem exactly?

Because then you can do,
$ diff -r <old schema dump dir> <new schema dump dir>,
instead of,
$ diff <old entire schema dump> <new entire schema dump>
which will nicely reveal each individual object modified, as opposed to a huge global diff of everything
 

> [-f filename] : main dump file, imports each splitted part using \i
> [-f filename]-split/[desc]/[tag]/[oid].sql : dump of the oid

This particular choice seems remarkably *un* friendly, since two dumps
from different DBs will inevitably not share the same OIDs, making it
practically impossible to compare them even if they are logically
identical.  But even without the choice to use OIDs in the filenames
I'm unconvinced that file-per-object is a good idea in any way shape or
form.

Good point!

To compare two different database, perhaps it's possible to use a sequence, 1,2,...,n for each file in each directory, i.e., /[desc]/[tag]/[n], and to sort them by something distinct which will ensure the same numbering between different databases, such as the arguments for functions, or other properties for other kind of objects. Any ideas?

(In my case, I didn't need to compare schemas between different database. I needed to compare two dumps created at different points in time of the same database, which do share the same oids for objects existing in both versions.)


 

                       regards, tom lane



--
Best regards,

Joel Jacobson
Glue Finance

E: jj@gluefinance.com
T: +46 70 360 38 01

Postal address:
Glue Finance AB
Box  549
114 11  Stockholm
Sweden

Visiting address:
Glue Finance AB
Birger Jarlsgatan 14
114 34 Stockholm
Sweden

Re: pg_dump --split patch

From
Tom Lane
Date:
Joel Jacobson <joel@gluefinance.com> writes:
> 2010/12/28 Tom Lane <tgl@sss.pgh.pa.us>
>> Joel Jacobson <joel@gluefinance.com> writes:
>>> Solution: I propose a new option to pg_dump, --split, which dumps each
>>> object to a separate file in a user friendly directory structure:
>> 
>> Um ... how does that solve the claimed problem exactly?

> Because then you can do,
> $ diff -r <old schema dump dir> <new schema dump dir>,
> instead of,
> $ diff <old entire schema dump> <new entire schema dump>
> which will nicely reveal each individual object modified, as opposed to a
> huge global diff of everything

That has at least as many failure modes as the other representation.
        regards, tom lane


Re: pg_dump --split patch

From
Joel Jacobson
Date:
2010/12/28 Tom Lane <tgl@sss.pgh.pa.us>

That has at least as many failure modes as the other representation.


I don't follow, what do you mean with "failure modes"? The oid in the filename? I suggested to use a sequence instead but you didn't comment on that. Are there any other failure modes which could cause a diff -r between two different databases to break?

(This might be a bad idea for some other reason, but I noticed a few other users requesting the same feature when I googled "pg_dump split".)

--
Best regards,

Joel Jacobson
Glue Finance

Re: pg_dump --split patch

From
Andrew Dunstan
Date:

On 12/28/2010 11:59 AM, Joel Jacobson wrote:
> 2010/12/28 Tom Lane <tgl@sss.pgh.pa.us <mailto:tgl@sss.pgh.pa.us>>
>
>
>     That has at least as many failure modes as the other representation.
>
>
> I don't follow, what do you mean with "failure modes"? The oid in the 
> filename? I suggested to use a sequence instead but you didn't comment 
> on that. Are there any other failure modes which could cause a diff -r 
> between two different databases to break?
>
> (This might be a bad idea for some other reason, but I noticed a few 
> other users requesting the same feature when I googled "pg_dump split".)
>
>

A better approach to the problem might be to have a tool which did a 
comparison of structures rather than a textual comparison of dumps. For 
extra credit, such a tool might even try to produce a sync script for 
you ...

Of course, that task might involve more effort than you want to devote 
to it.

cheers

andrew


Re: pg_dump --split patch

From
Tom Lane
Date:
Joel Jacobson <joel@gluefinance.com> writes:
> 2010/12/28 Tom Lane <tgl@sss.pgh.pa.us>
>> That has at least as many failure modes as the other representation.

> I don't follow, what do you mean with "failure modes"? The oid in the
> filename? I suggested to use a sequence instead but you didn't comment on
> that. Are there any other failure modes which could cause a diff -r between
> two different databases to break?

AFAIK the primary failure modes for diff'ing text dumps are

(1) randomly different ordering of objects from one dump to another.
Your initial proposal would avoid that problem as long as the object
OIDs didn't change, but since it falls down completely across a dump and
reload, or delete and recreate, I can't really see that it's a step
forward.  Using a sequence number generated by pg_dump doesn't change
this at all --- the sequence would be just as unpredictable.

(2) randomly different ordering of rows within a table.  Your patch
didn't address that, unless I misunderstood quite a bit.

I think the correct fix for (1) is to improve pg_dump's method for
sorting objects.  It's not that bad now, but it does have issues with
random ordering of similarly-named objects.  IIRC Peter Eisentraut
proposed something for this last winter but it seemed a mite too ugly,
and he got beaten down to just this:

commit 1acc06a1f4ae752793d2199d8d462a6708c8acc2
Author: Peter Eisentraut <peter_e@gmx.net>
Date:   Mon Feb 15 19:59:47 2010 +0000
   When sorting functions in pg_dump, break ties (same name) by number of argum
ents

Maybe you can do better, but I'd suggest going back to reread the
discussion that preceded that patch.

> (This might be a bad idea for some other reason, but I noticed a few other
> users requesting the same feature when I googled "pg_dump split".)

AFAIR what those folk really wanted was a selective dump with more
selectivity knobs than exist now.  I don't think their lives would be
improved by having to root through a twisty little maze of numbered
files to find the object they wanted.
        regards, tom lane


Re: pg_dump --split patch

From
Gurjeet Singh
Date:
On Tue, Dec 28, 2010 at 11:00 AM, Joel Jacobson <joel@gluefinance.com> wrote:
Dear fellow hackers,

Problem: A normal diff of two slightly different schema dump files (pg_dump -s), will not produce a user-friendly diff, as you get all changes in the same file.

Solution: I propose a new option to pg_dump, --split, which dumps each object to a separate file in a user friendly directory structure:

[-f filename] : main dump file, imports each splitted part using \i
[-f filename]-split/[desc]/[tag]/[oid].sql : dump of the oid

Example: If the filename (-f) is "pg.dump", the following directory structure would be created:

$ ./pg_dump -f /crypt/pg.dump --split -F p -s glue
/crypt/pg.dump-split/VIEW/
/crypt/pg.dump-split/TYPE/
/crypt/pg.dump-split/TRIGGER/
/crypt/pg.dump-split/TABLE/
/crypt/pg.dump-split/SEQUENCE/
/crypt/pg.dump-split/SCHEMA/
/crypt/pg.dump-split/PROCEDURAL_LANGUAGE/
/crypt/pg.dump-split/INDEX/
/crypt/pg.dump-split/FUNCTION/
/crypt/pg.dump-split/FK_CONSTRAINT/
/crypt/pg.dump-split/CONSTRAINT/
/crypt/pg.dump-split/AGGREGATE/

In each such directory, one directory per object name is created.
If we would have a function "foobar" with oid "12345" it would be saved to:
/crypt/pg.dump-split/FUNCTION/foobar/12345.sql

In the "pg.dump" plain text file, the files are "linked in" using the "\i" psql command, e.g.:
\i /crypt/pg.dump-split/FUNCTION/foobar/12345.sql

Potential use-case scenarios:

*) Version control your database schema, by exporting it daily (using --split) and commiting the differences.

*) Compare differences of schema dumps created in different points in time. Since objects are stored in separate files, it is easier to see what areas were modified, compared to looking at the diff of two entire schemas.

*) Restore only some objects, based on type (e.g., only the functions) or name (e.g. only fucntions of certain name/names).

I've tested the patch for both the latest HEAD (9.1devel) as well as 8.4.6.


I would suggest the directory structure as:

/crypt/pg.dump-split/schema-name-1/VIEWS/view-name-1.sql
/crypt/pg.dump-split/schema-name-1/TABLES/table-name-1.sql
...
/crypt/pg.dump-split/schema-name-2/VIEWS/view-name-1.sql
/crypt/pg.dump-split/schema-name-2/TABLES/table-name-1.sql

This might n be more amenable to diff'ing the different dumps. Schemas are logical grouping of other objects and hence making that apparent in your dump's hierarchy makes more sense.

Most importantly, as Tom suggested, don't use or rely on OIDs. I think function overloading is the only case where you can have more than one object with the same name under a schema. That can be resolved if you included function signature in filename:

/crypt/pg.dump-split/emp/FUNCTIONS/myfunc-int-char.sql
/crypt/pg.dump-split/emp/FUNCTIONS/myfunc-int-int.sql
/crypt/pg.dump-split/emp/FUNCTIONS/myfunc-int.sql

Regards,
--
gurjeet.singh
@ EnterpriseDB - The Enterprise Postgres Company
http://www.EnterpriseDB.com

singh.gurjeet@{ gmail | yahoo }.com
Twitter/Skype: singh_gurjeet

Mail sent from my BlackLaptop device

Re: pg_dump --split patch

From
Aidan Van Dyk
Date:
On Tue, Dec 28, 2010 at 11:59 AM, Joel Jacobson <joel@gluefinance.com> wrote:

> I don't follow, what do you mean with "failure modes"? The oid in the
> filename? I suggested to use a sequence instead but you didn't comment on
> that. Are there any other failure modes which could cause a diff -r between
> two different databases to break?

Both OID and sequence mean that your likely to get a diff which is
nothing more than complete files removed from 1 side and added to the
othe rside with different names (i.e. oid's don't match, or an
added/removed object changes all following sequence assingments).

If you're going to try and split, I really think the only usefull
filename has to be similar to something like:   <schema>/<type>/<name>/<part>

If you want to use "diff", you pretty much have to make sure that the
*path* will be identical for similary named objects, irrespective of
anything else in the database.  And path has to be encoding aware.

And you want names that glob well, so for instance, you could exclude
*.data (or a schema) from the diff.

a.

--
Aidan Van Dyk                                             Create like a god,
aidan@highrise.ca                                       command like a king,
http://www.highrise.ca/                                   work like a slave.


Re: pg_dump --split patch

From
Joel Jacobson
Date:
2010/12/28 Gurjeet Singh <singh.gurjeet@gmail.com>
I would suggest the directory structure as:

/crypt/pg.dump-split/schema-name-1/VIEWS/view-name-1.sql
/crypt/pg.dump-split/schema-name-1/TABLES/table-name-1.sql
...
/crypt/pg.dump-split/schema-name-2/VIEWS/view-name-1.sql
/crypt/pg.dump-split/schema-name-2/TABLES/table-name-1.sql

This might n be more amenable to diff'ing the different dumps. Schemas are logical grouping of other objects and hence making that apparent in your dump's hierarchy makes more sense.

Thanks Gurjeet and Tom for good feedback!

I've made some changes and attached new patches.
Looks much better now I think!

This is what I've changed,

*) Not using oid anymore in the filename
*) New filename/path structure: [-f filename]-split/[schema]/[desc]/[tag].sql
*) If two objects share the same name tag for the same [schema]/[desc], -2, -3, etc is appended to the name. Example:
~/pg.dump-split/public/FUNCTION/foobar.sql
~/pg.dump-split/public/FUNCTION/foobar-2.sql
~/pg.dump-split/public/FUNCTION/barfoo.sql
~/pg.dump-split/public/FUNCTION/barfoo-2.sql
~/pg.dump-split/public/FUNCTION/barfoo-3.sql

I think you are right about functions (and aggregates) being the only desc-type where two objects can share the same name in the same schema.
This means the problem of dumping objects in different order is a very limited problem, only affecting overloaded functions.

I didn't include the arguments in the file name, as it would lead to very long file names unless truncated, and since the problem is very limited, I think we shouldn't include it. It's cleaner with just the name part of the tag in the file name.

--
Best regards,

Joel Jacobson
Glue Finance
Attachment

Re: pg_dump --split patch

From
Gurjeet Singh
Date:
On Tue, Dec 28, 2010 at 2:39 PM, Joel Jacobson <joel@gluefinance.com> wrote:
2010/12/28 Gurjeet Singh <singh.gurjeet@gmail.com>
I would suggest the directory structure as:

/crypt/pg.dump-split/schema-name-1/VIEWS/view-name-1.sql
/crypt/pg.dump-split/schema-name-1/TABLES/table-name-1.sql
...
/crypt/pg.dump-split/schema-name-2/VIEWS/view-name-1.sql
/crypt/pg.dump-split/schema-name-2/TABLES/table-name-1.sql

This might n be more amenable to diff'ing the different dumps. Schemas are logical grouping of other objects and hence making that apparent in your dump's hierarchy makes more sense.

Thanks Gurjeet and Tom for good feedback!

I've made some changes and attached new patches.
Looks much better now I think!

This is what I've changed,

*) Not using oid anymore in the filename
*) New filename/path structure: [-f filename]-split/[schema]/[desc]/[tag].sql
*) If two objects share the same name tag for the same [schema]/[desc], -2, -3, etc is appended to the name. Example:
~/pg.dump-split/public/FUNCTION/foobar.sql
~/pg.dump-split/public/FUNCTION/foobar-2.sql
~/pg.dump-split/public/FUNCTION/barfoo.sql
~/pg.dump-split/public/FUNCTION/barfoo-2.sql
~/pg.dump-split/public/FUNCTION/barfoo-3.sql

I think you are right about functions (and aggregates) being the only desc-type where two objects can share the same name in the same schema.
This means the problem of dumping objects in different order is a very limited problem, only affecting overloaded functions.

I didn't include the arguments in the file name, as it would lead to very long file names unless truncated, and since the problem is very limited, I think we shouldn't include it. It's cleaner with just the name part of the tag in the file name.


I haven't seen your code yet, but we need to make sure that in case of name collision we emit the object definitions in a sorted order so that the dump is always deterministic: func1(char) should be _always_ dumped before func1(int), that is, output file names are always deterministic.

The problem I see with suffixing a sequence id to the objects with name collision is that one day the dump may name myfunc(int) as myfunc.sql and after an overloaded version is created, say myfunc(char, int), then the same myfunc(int) may be dumped in myfunc-2.sql, which again is non-deterministic.

Also, it is a project policy that we do not introduce new features in back branches, so spending time on an 8.4.6 patch may not be the best use of your time.

Regards,
--
gurjeet.singh
@ EnterpriseDB - The Enterprise Postgres Company
http://www.EnterpriseDB.com

singh.gurjeet@{ gmail | yahoo }.com
Twitter/Skype: singh_gurjeet

Mail sent from my BlackLaptop device

Re: pg_dump --split patch

From
Joel Jacobson
Date:
Sent from my iPhone

On 28 dec 2010, at 21:45, Gurjeet Singh <singh.gurjeet@gmail.com> wrote:

The problem I see with suffixing a sequence id to the objects with name
collision is that one day the dump may name myfunc(int) as myfunc.sql and
after an overloaded version is created, say myfunc(char, int), then the same
myfunc(int) may be dumped in myfunc-2.sql, which again is non-deterministic.


I agree, good point!
Perhaps abbreviations are to prefer, e.g., myfunc_i, myfunc_i_c, etc to
reduce the need of truncating filenames.



Also, it is a project policy that we do not introduce new features in back
branches, so spending time on an 8.4.6 patch may not be the best use of your
time.


My company is using 8.4 and needs this feature, so I'll have to patch it
anyway :)


Regards,
--
gurjeet.singh
@ EnterpriseDB - The Enterprise Postgres Company
http://www.EnterpriseDB.com

singh.gurjeet@{ gmail | yahoo }.com
Twitter/Skype: singh_gurjeet

Mail sent from my BlackLaptop device

Re: pg_dump --split patch

From
Andrew Dunstan
Date:

On 12/28/2010 04:44 PM, Joel Jacobson wrote:
>
>
>
>
>> The problem I see with suffixing a sequence id to the objects with 
>> name collision is that one day the dump may name myfunc(int) as 
>> myfunc.sql and after an overloaded version is created, say 
>> myfunc(char, int), then the same myfunc(int) may be dumped in 
>> myfunc-2.sql, which again is non-deterministic.
>
> I agree, good point!
> Perhaps abbreviations are to prefer, e.g., myfunc_i, myfunc_i_c, etc 
> to reduce the need of truncating filenames.
>
>
>

I think that's just horrible. Does the i stand for integer or inet? And 
it will get *really* ugly for type names with spaces in them ...


cheers

andrew


Re: pg_dump --split patch

From
Joel Jacobson
Date:
2010/12/28 Andrew Dunstan <andrew@dunslane.net>

I think that's just horrible. Does the i stand for integer or inet? And it will get *really* ugly for type names with spaces in them ...


True, true.

But while "c" is too short, I think "character varying" is too long. Is there some convenient lookup table to convert between the long names to the short names?
E.g.,
character varying => varchar
timestamp with time zone => timestamptz
etc.

--
Best regards,

Joel Jacobson
Glue Finance

Re: pg_dump --split patch

From
Tom Lane
Date:
Andrew Dunstan <andrew@dunslane.net> writes:
> On 12/28/2010 04:44 PM, Joel Jacobson wrote:
>> Perhaps abbreviations are to prefer, e.g., myfunc_i, myfunc_i_c, etc 
>> to reduce the need of truncating filenames.

> I think that's just horrible. Does the i stand for integer or inet? And 
> it will get *really* ugly for type names with spaces in them ...

You think spaces are bad, try slashes ;-)

Not to mention the need for including schemas in typenames sometimes.
I think you're going to have a real problem trying to fully describe a
function's signature in a file name of reasonable max length.
        regards, tom lane


Re: pg_dump --split patch

From
Gurjeet Singh
Date:
On Tue, Dec 28, 2010 at 4:57 PM, Andrew Dunstan <andrew@dunslane.net> wrote:


On 12/28/2010 04:44 PM, Joel Jacobson wrote:




The problem I see with suffixing a sequence id to the objects with name collision is that one day the dump may name myfunc(int) as myfunc.sql and after an overloaded version is created, say myfunc(char, int), then the same myfunc(int) may be dumped in myfunc-2.sql, which again is non-deterministic.

I agree, good point!
Perhaps abbreviations are to prefer, e.g., myfunc_i, myfunc_i_c, etc to reduce the need of truncating filenames.




I think that's just horrible. Does the i stand for integer or inet? And it will get *really* ugly for type names with spaces in them ...


Do you mean using data type names in filename is a bad idea, or is abbreviating the type names is a bad idea?

Maybe we can compute a hash based on the type names and use that in the file's name?

Regards,
--
gurjeet.singh
@ EnterpriseDB - The Enterprise Postgres Company
http://www.EnterpriseDB.com

singh.gurjeet@{ gmail | yahoo }.com
Twitter/Skype: singh_gurjeet

Mail sent from my BlackLaptop device

Re: pg_dump --split patch

From
David Wilson
Date:
On Tue, Dec 28, 2010 at 2:39 PM, Joel Jacobson <joel@gluefinance.com> wrote:

I think you are right about functions (and aggregates) being the only desc-type where two objects can share the same name in the same schema.
This means the problem of dumping objects in different order is a very limited problem, only affecting overloaded functions.

I didn't include the arguments in the file name, as it would lead to very long file names unless truncated, and since the problem is very limited, I think we shouldn't include it. It's cleaner with just the name part of the tag in the file name.

Why not place all overloads of a function within the same file? Then, assuming you order them deterministically within that file, we sidestep the file naming issue and maintain useful diff capabilities, since a diff of the function's file will show additions or removals of various overloaded versions.

--
- David T. Wilson
david.t.wilson@gmail.com

Re: pg_dump --split patch

From
Joel Jacobson
Date:
2010/12/29 David Wilson <david.t.wilson@gmail.com>
Why not place all overloads of a function within the same file? Then, assuming you order them deterministically within that file, we sidestep the file naming issue and maintain useful diff capabilities, since a diff of the function's file will show additions or removals of various overloaded versions.


Good suggestion. I agree, trying to put variations of the same function in different files simply becomes too ugly and the problem it solves is not big enough.
Then we just need to make sure pg_dump dumps objects in the same order, and let diff take care of the rest of the challenge. :)

Brb with a new patch.

--
Best regards,

Joel Jacobson
Glue Finance

Re: pg_dump --split patch

From
Tom Lane
Date:
David Wilson <david.t.wilson@gmail.com> writes:
> On Tue, Dec 28, 2010 at 2:39 PM, Joel Jacobson <joel@gluefinance.com> wrote:
>> I didn't include the arguments in the file name, as it would lead to very
>> long file names unless truncated, and since the problem is very limited, I
>> think we shouldn't include it. It's cleaner with just the name part of the
>> tag in the file name.

> Why not place all overloads of a function within the same file? Then,
> assuming you order them deterministically within that file, we sidestep the
> file naming issue and maintain useful diff capabilities, since a diff of the
> function's file will show additions or removals of various overloaded
> versions.

If you've solved the deterministic-ordering problem, then this entire
patch is quite useless.  You can just run a normal dump and diff it.
        regards, tom lane


Re: pg_dump --split patch

From
Joel Jacobson
Date:
2010/12/29 Tom Lane <tgl@sss.pgh.pa.us>

If you've solved the deterministic-ordering problem, then this entire
patch is quite useless.  You can just run a normal dump and diff it.


No, that's only half true.

Diff will do a good job minimizing the "size" of the diff output, yes, but such a diff is still quite useless if you want to quickly grasp the context of the change.

If you have a hundreds of functions, just looking at the changed source code is not enough to figure out which functions were modified, unless you have the brain power to memorize every single line of code and are able to figure out the function name just by looking at the old and new line of codes.

To understand a change to my database functions, I would start by looking at the top-level, only focusing on the names of the functions modified/added/removed.
At this stage, you want as little information as possible about each change, such as only the names of the functions.
To do this, get a list of changes functions, you cannot compare two full schema plain text dumps using diff, as it would only reveal the lines changed, not the name of the functions, unless you are lucky to get the name of the function within the (by default) 3 lines of copied context.

While you could increase the number of copied lines of context to a value which would ensure you would see the name of the function in the diff, that is not feasible if you want to quickly "get a picture" of the code areas modified, since you would then need to read through even more lines of diff output.

For a less database-centric system where you don't have hundreds of stored procedures, I would agree it's not an issue to keep track of changes by diffing entire schema files, but for extremely database-centric systems, such as the one we have developed at my company, it's not possible to "get the whole picture" of a change by analyzing diffs of entire schema dumps.

The patch has been updated:

*) Only spit objects with a namespace (schema) not being null
*) Append all objects of same tag (name) of same type (desc) of same namespace (schema) to the same file (i.e., do not append -2, -3, like before) (Suggested by David Wilson, thanks.)

I also tested to play around with "ORDER BY pronargs" and "ORDER BY pronargs DESC" to the queries in getFuncs() in pg_dump.c, but it had no effect to the order the functions of same name but different number of arguments were dumped.
Perhaps functions are already sorted?
Anyway, it doesn't matter that much, keeping all functions of the same name in the same file is a fair trade-off I think. The main advantage is the ability to quickly get a picture of the names of all changed functions, secondly to optimize the actual diff output.


--
Best regards,

Joel Jacobson
Glue Finance

E: jj@gluefinance.com
T: +46 70 360 38 01

Postal address:
Glue Finance AB
Box  549
114 11  Stockholm
Sweden

Visiting address:
Glue Finance AB
Birger Jarlsgatan 14
114 34 Stockholm
Sweden
Attachment

Re: pg_dump --split patch

From
Andrew Dunstan
Date:

On 12/28/2010 08:18 PM, Joel Jacobson wrote:
> 2010/12/29 Tom Lane <tgl@sss.pgh.pa.us <mailto:tgl@sss.pgh.pa.us>>
>
>
>     If you've solved the deterministic-ordering problem, then this entire
>     patch is quite useless.  You can just run a normal dump and diff it.
>
>
> No, that's only half true.
>
> Diff will do a good job minimizing the "size" of the diff output, yes, 
> but such a diff is still quite useless if you want to quickly grasp 
> the context of the change.
>
> If you have a hundreds of functions, just looking at the changed 
> source code is not enough to figure out which functions were modified, 
> unless you have the brain power to memorize every single line of code 
> and are able to figure out the function name just by looking at the 
> old and new line of codes.
>
>

try:
 diff -F '^CREATE' ...

cheers

andrew


Re: pg_dump --split patch

From
Joel Jacobson
Date:


Sent from my iPhone

On 28 dec 2010, at 21:45, Gurjeet Singh <singh.gurjeet@gmail.com> wrote:

The problem I see with suffixing a sequence id to the objects with name collision is that one day the dump may name myfunc(int) as myfunc.sql and after an overloaded version is created, say myfunc(char, int), then the same myfunc(int) may be dumped in myfunc-2.sql, which again is non-deterministic.

I agree, good point!
Perhaps abbreviations are to prefer, e.g., myfunc_i, myfunc_i_c, etc to reduce the need of truncating filenames.



Also, it is a project policy that we do not introduce new features in back branches, so spending time on an 8.4.6 patch may not be the best use of your time.

My company is using 8.4 and needs this feature, so I'll have to patch it anyway :)


Regards,
--
gurjeet.singh
@ EnterpriseDB - The Enterprise Postgres Company
http://www.EnterpriseDB.com

singh.gurjeet@{ gmail | yahoo }.com
Twitter/Skype: singh_gurjeet

Mail sent from my BlackLaptop device

Re: pg_dump --split patch

From
Joel Jacobson
Date:
2010/12/29 Andrew Dunstan <andrew@dunslane.net>
try:

 diff -F '^CREATE' ...

cheers

andrew

Embarrasing, I'm sure I've done `man diff` before, must have missed that one, wish I'd known about that feature before, would have saved me many hours! :-) Thanks for the tip!

There are some other real-life use-cases where I think splitting would be nice and save a lot of time:

a) if you don't have a perfect 1:1 relationship between all the SPs in your database and your source code repository (handled by your favorite version control system), i.e. if you suspect some SPs in the database might differ compared to the source code files in your repo. In this scenario, it might be simpler to "start over" and continue developing on a repo built from a pg_dump --split export. You would lose all history, but it might still be worth it if the "compare everything in database against source code files in repo"-project would take a lot of man hours.

b) quick branching - perhaps you are a consultant at a company where they don't even have the SPs stored in separate files, they might have been magically installed by some consultant before you without any trace. :-) To get up to speed solving the problem you've been assigned, which in this example involves a lot of SP coding and modifications of existing functions, it would save a lot of time if you had all functions in separate files before you started coding, then you would use git or any other nice version control system to track your changes and figure out what you've done once you get everything to work.

c) automatically saving daily snapshots of your production database schema to your version control system. While the best version control system (git) does not track individual files, many of the ancient ones still very popular ones like svn do so. If every function in the production database schema would be saved automatically to the VCS, you would be guaranteed to have a tack of all deployed changes affecting each function, which is probably a lot fewer changes compared to the entire history for each function, assuming developers commit things while developing and not only when deploying.

d) while pg_dump offers some options to limit the output content, such as -s for "schema only" and -t/-T to limit which tables to dump, it lacks options to export "functions only" or "these functions only". It would require quite a lot of such options to provide the same flexibility as a split dump, highly reducing the need for such options as you could then compose your own restore script based on the dump.

Of course, not all of these scenarios are relevant for everybody.

--
Best regards,

Joel Jacobson
Glue Finance

Re: pg_dump --split patch

From
Aidan Van Dyk
Date:
On Wed, Dec 29, 2010 at 2:27 AM, Joel Jacobson <joel@gluefinance.com> wrote:

<description of split stuff>

So, how different (or not) is this to the "directory" format that was
coming out of the desire of a parallel pg_dump?

a.

--
Aidan Van Dyk                                             Create like a god,
aidan@highrise.ca                                       command like a king,
http://www.highrise.ca/                                   work like a slave.


Re: pg_dump --split patch

From
Joel Jacobson
Date:


2010/12/29 Aidan Van Dyk <aidan@highrise.ca>
On Wed, Dec 29, 2010 at 2:27 AM, Joel Jacobson <joel@gluefinance.com> wrote:

<description of split stuff>

So, how different (or not) is this to the "directory" format that was
coming out of the desire of a parallel pg_dump?

Not sure what format you are referring to? Custom, tar or plain text?
I noticed there are two undocumented formats as well, "append" and "file".
I tried both of these undocumented formats, but it did not procude any directory structure of the dumped objects.

Could you please explain how to use the "directory format" is such a format already exists?
I can't find it in the documentation nor the source code of HEAD.

 

a.

--
Aidan Van Dyk                                             Create like a god,
aidan@highrise.ca                                       command like a king,
http://www.highrise.ca/                                   work like a slave.



--
Best regards,

Joel Jacobson
Glue Finance

E: jj@gluefinance.com
T: +46 70 360 38 01

Postal address:
Glue Finance AB
Box  549
114 11  Stockholm
Sweden

Visiting address:
Glue Finance AB
Birger Jarlsgatan 14
114 34 Stockholm
Sweden

Re: pg_dump --split patch

From
Gurjeet Singh
Date:
On Wed, Dec 29, 2010 at 8:31 AM, Joel Jacobson <joel@gluefinance.com> wrote:


2010/12/29 Aidan Van Dyk <aidan@highrise.ca>

On Wed, Dec 29, 2010 at 2:27 AM, Joel Jacobson <joel@gluefinance.com> wrote:

<description of split stuff>

So, how different (or not) is this to the "directory" format that was
coming out of the desire of a parallel pg_dump?

Not sure what format you are referring to? Custom, tar or plain text?
I noticed there are two undocumented formats as well, "append" and "file".
I tried both of these undocumented formats, but it did not procude any directory structure of the dumped objects.

Could you please explain how to use the "directory format" is such a format already exists?
I can't find it in the documentation nor the source code of HEAD.


It is still being discussed as a patch to pg_dump. Google for "directory archive format for pg_dump", specifically in archives.postgresql.org.

AFAIK, that applies to parallel dumps of data (may help in --schema-only dumps too), and what you are trying is for schema.

Regards
--
gurjeet.singh
@ EnterpriseDB - The Enterprise Postgres Company
http://www.EnterpriseDB.com

singh.gurjeet@{ gmail | yahoo }.com
Twitter/Skype: singh_gurjeet

Mail sent from my BlackLaptop device

Re: pg_dump --split patch

From
Aidan Van Dyk
Date:
On Wed, Dec 29, 2010 at 9:11 AM, Gurjeet Singh <singh.gurjeet@gmail.com> wrote:
> On Wed, Dec 29, 2010 at 8:31 AM, Joel Jacobson <joel@gluefinance.com> wrote:
>>
>>
>> 2010/12/29 Aidan Van Dyk <aidan@highrise.ca>
>>>
>>> On Wed, Dec 29, 2010 at 2:27 AM, Joel Jacobson <joel@gluefinance.com>
>>> wrote:
>>>
>>> <description of split stuff>
>>>
>>> So, how different (or not) is this to the "directory" format that was
>>> coming out of the desire of a parallel pg_dump?
>>
>> Not sure what format you are referring to? Custom, tar or plain text?
>> I noticed there are two undocumented formats as well, "append" and "file".
>> I tried both of these undocumented formats, but it did not procude any
>> directory structure of the dumped objects.
>> Could you please explain how to use the "directory format" is such a
>> format already exists?
>> I can't find it in the documentation nor the source code of HEAD.
>
> It is still being discussed as a patch to pg_dump. Google for "directory
> archive format for pg_dump", specifically in archives.postgresql.org.

Specifically:
Message-ID: <AANLkTimUELTXwRSQDQNwxik_k1y3YcH1u-9NgHZqpi9e@mail.gmail.com>

> AFAIK, that applies to parallel dumps of data (may help in --schema-only
> dumps too), and what you are trying is for schema.

Right, but one of the things it does is break the dump in to parts,
and put them in a directory/file organization.

Both are doing it for different reasons, but doing pretty much the
same thing.  But can the layout/organization of Joachim's patch can be
made "human friendly" in the vein of Joel's vision?

a.

--
Aidan Van Dyk                                             Create like a god,
aidan@highrise.ca                                       command like a king,
http://www.highrise.ca/                                   work like a slave.


Re: pg_dump --split patch

From
Tom Lane
Date:
Aidan Van Dyk <aidan@highrise.ca> writes:
> On Wed, Dec 29, 2010 at 9:11 AM, Gurjeet Singh <singh.gurjeet@gmail.com> wrote:
>> AFAIK, that applies to parallel dumps of data (may help in --schema-only
>> dumps too), and what you are trying is for schema.

> Right, but one of the things it does is break the dump in to parts,
> and put them in a directory/file organization.

> Both are doing it for different reasons, but doing pretty much the
> same thing.  But can the layout/organization of Joachim's patch can be
> made "human friendly" in the vein of Joel's vision?

I think they're fundamentally different things, because the previously
proposed patch is an extension of the machine-readable archive format,
and has to remain so because of the expectation that people will want
to use parallel restore with it.  Joel is arguing for a split-up of
the text dump format.
        regards, tom lane


Re: pg_dump --split patch

From
Joel Jacobson
Date:
2010/12/29 Tom Lane <tgl@sss.pgh.pa.us>
I think they're fundamentally different things, because the previously
proposed patch is an extension of the machine-readable archive format,
and has to remain so because of the expectation that people will want
to use parallel restore with it.  Joel is arguing for a split-up of
the text dump format.


Yes, exactly.

My patch is of course also a lot smaller :-)
pg_dump-directory.diff.........................: 112 853 bytes
pg-dump-split-plain-text-files-9.1devel.patch..:   5 579 bytes

I just tried the pg_dump-directory.diff patch.
The only thing is has in common with my patch is it writes data to different files, and it's only the data which is splitted into different files, the schema appears to go into the single file "TOC".

Example, pg_dump-directory.diff:

$ ./pg_dump -f /crypt/dirpatch -F d -s glue
$ ls -la /crypt/dirpatch/
TOC
(1 file)

$ rm -rf /crypt/dirpatch

$ ./pg_dump -f /crypt/dirpatch -F d glue

$ ls /crypt/dirpatch/
6503.dat
6504.dat
...lots of files...
6871.dat
6872.dat
6873.dat
6874.dat
TOC

Example, pg_dump --split patch:

$ pg_dump -f /crypt/splitpatch -F p --split -s glue

$ ls /crypt/splitpatch*
/crypt/splitpatch (file)
/crypt/splitpatch-split: (directory)
    myschema1
    myschema2
    public
$ ls /crypt/splitpatch-split/public/
AGGREGATE
CONSTRAINT
FK_CONSTRAINT
FUNCTION
INDEX
SEQUENCE
TABLE
TRIGGER
TYPE
VIEW

$ ls /crypt/splitpatch-split/public/FUNCTION/
    myfunc.sql
    otherfunc.sql

$ cat /crypt/splitpatch
--
-- PostgreSQL database dump
--

SET statement_timeout = 0;
SET client_encoding = 'UTF8';
SET standard_conforming_strings = off;
SET check_function_bodies = false;
SET client_min_messages = warning;
SET escape_string_warning = off;
...etc...
\i /crypt/splitpatch-split/public/FUNCTION/myfunc.sql
\i /crypt/splitpatch-split/public/FUNCTION/otherfunc.sql


--
Best regards,

Joel Jacobson
Glue Finance

Re: pg_dump --split patch

From
Dimitri Fontaine
Date:
Joel Jacobson <joel@gluefinance.com> writes:
> Solution: I propose a new option to pg_dump, --split, which dumps each
> object to a separate file in a user friendly directory structure:

Please have a look at getddl:
 https://github.com/dimitri/getddl

Regards,
-- 
Dimitri Fontaine
http://2ndQuadrant.fr     PostgreSQL : Expertise, Formation et Support


Re: pg_dump --split patch

From
Joel Jacobson
Date:
2010/12/29 Dimitri Fontaine <dimitri@2ndquadrant.fr>
Please have a look at getddl:

 https://github.com/dimitri/getddl


Nice! Looks like a nifty tool.
When I tried it, "./getddl.py -f -F /crypt/funcs -d glue", I got the error "No such file or directory: 'sql/schemas.sql'".

While the task of splitting objects into separate files could be solved by an external "wrapper tool" like yours around pg_dump,
I argue it makes more sense of putting the (minimal required) logics into pg_dump, due to a number of reasons, most importantly because it's simplier and less complex, thus less error prone.

My patch is only a few lines of code and doesn't add any logics to pg_dump, it merely reroutes the fwrite() system calls based on the toc entries.

Just the fact you and others had to create own tools to do the splitting shows the feature is important, which I think should be included in the normal pg_dump tool.

Happy new year!

Joel Jacobson
Glue Finance

Re: pg_dump --split patch

From
Robert Treat
Date:


On Thu, Dec 30, 2010 at 2:13 AM, Joel Jacobson <joel@gluefinance.com> wrote:
2010/12/29 Dimitri Fontaine <dimitri@2ndquadrant.fr>

Please have a look at getddl:

 https://github.com/dimitri/getddl


Nice! Looks like a nifty tool.
When I tried it, "./getddl.py -f -F /crypt/funcs -d glue", I got the error "No such file or directory: 'sql/schemas.sql'".

While the task of splitting objects into separate files could be solved by an external "wrapper tool" like yours around pg_dump,
I argue it makes more sense of putting the (minimal required) logics into pg_dump, due to a number of reasons, most importantly because it's simplier and less complex, thus less error prone.

My patch is only a few lines of code and doesn't add any logics to pg_dump, it merely reroutes the fwrite() system calls based on the toc entries.

Just the fact you and others had to create own tools to do the splitting shows the feature is important, which I think should be included in the normal pg_dump tool.

As someone whose own version of "getddl" helped inspire Dimitri to create his own version, I've both enjoyed reading this thread and seeing this wheel reinvented yet again, and wholeheartedly +1 the idea of building this directly into pg_dump. (The only thing better would be to make everything thing sql callable, but that's a problem for another day).  


Robert Treat

Re: pg_dump --split patch

From
Peter Eisentraut
Date:
On tis, 2010-12-28 at 12:33 -0500, Tom Lane wrote:
> (2) randomly different ordering of rows within a table.  Your patch
> didn't address that, unless I misunderstood quite a bit.

This issue here is just comparing schemas, so that part is a separate
problem for someone else.

> I think the correct fix for (1) is to improve pg_dump's method for
> sorting objects.  It's not that bad now, but it does have issues with
> random ordering of similarly-named objects.  IIRC Peter Eisentraut
> proposed something for this last winter but it seemed a mite too ugly,
> and he got beaten down to just this:
> 
> commit 1acc06a1f4ae752793d2199d8d462a6708c8acc2
> Author: Peter Eisentraut <peter_e@gmx.net>
> Date:   Mon Feb 15 19:59:47 2010 +0000
> 
>     When sorting functions in pg_dump, break ties (same name) by
> number of arguments

Yes, that was addressing the same underlying problem.  Frankly, I have
been thinking split files a lot before and since then.  If the files
were appropriately named, it would remove a lot of problems compared to
diffing one even perfectly sorted big dump file.




Re: pg_dump --split patch

From
Peter Eisentraut
Date:
On tis, 2010-12-28 at 20:51 -0500, Andrew Dunstan wrote:
> try:
> 
>   diff -F '^CREATE' ...

This works about 67% of the time and still doesn't actually tell at a
glance what changed.  It will only tell you what the change you are
currently looking at probably belongs to.



Re: pg_dump --split patch

From
Dmitry Koterov
Date:
To me, this is a wonderful feature, thanks! I think many people would be happy if this patch woud be included to the mainstream (and it is quite short and simple).

About name ordering - I think that the problem exists for objects:

1. Stored functions.
2. Foreign keys/triggers (objects which has owning objects).

It is wonderful that you store all functions with the same name to the same file. To order them within this file we may simply compare the first definition line lexicographically (or - first line which differs one function definition from another).

Foreign key/triggers ordering problem is described by me at

The big problem is with triggers: many triggers may have the same name, but be bound to different tables. It would be great to include these triggers to table's definition or, at least, have separated files for each trigger+table pair.



On Wed, Dec 29, 2010 at 6:21 PM, Joel Jacobson <joel@gluefinance.com> wrote:
2010/12/29 Tom Lane <tgl@sss.pgh.pa.us>
I think they're fundamentally different things, because the previously
proposed patch is an extension of the machine-readable archive format,
and has to remain so because of the expectation that people will want
to use parallel restore with it.  Joel is arguing for a split-up of
the text dump format.


Yes, exactly.

My patch is of course also a lot smaller :-)
pg_dump-directory.diff.........................: 112 853 bytes
pg-dump-split-plain-text-files-9.1devel.patch..:   5 579 bytes

I just tried the pg_dump-directory.diff patch.
The only thing is has in common with my patch is it writes data to different files, and it's only the data which is splitted into different files, the schema appears to go into the single file "TOC".

Example, pg_dump-directory.diff:

$ ./pg_dump -f /crypt/dirpatch -F d -s glue
$ ls -la /crypt/dirpatch/
TOC
(1 file)

$ rm -rf /crypt/dirpatch

$ ./pg_dump -f /crypt/dirpatch -F d glue

$ ls /crypt/dirpatch/
6503.dat
6504.dat
...lots of files...
6871.dat
6872.dat
6873.dat
6874.dat
TOC

Example, pg_dump --split patch:

$ pg_dump -f /crypt/splitpatch -F p --split -s glue

$ ls /crypt/splitpatch*
/crypt/splitpatch (file)
/crypt/splitpatch-split: (directory)
    myschema1
    myschema2
    public
$ ls /crypt/splitpatch-split/public/
AGGREGATE
CONSTRAINT
FK_CONSTRAINT
FUNCTION
INDEX
SEQUENCE
TABLE
TRIGGER
TYPE
VIEW

$ ls /crypt/splitpatch-split/public/FUNCTION/
    myfunc.sql
    otherfunc.sql

$ cat /crypt/splitpatch
--
-- PostgreSQL database dump
--

SET statement_timeout = 0;
SET client_encoding = 'UTF8';
SET standard_conforming_strings = off;
SET check_function_bodies = false;
SET client_min_messages = warning;
SET escape_string_warning = off;
...etc...
\i /crypt/splitpatch-split/public/FUNCTION/myfunc.sql
\i /crypt/splitpatch-split/public/FUNCTION/otherfunc.sql


--
Best regards,

Joel Jacobson
Glue Finance

Re: pg_dump --split patch

From
Robert Haas
Date:
On Mon, Jan 3, 2011 at 7:11 AM, Dmitry Koterov <dmitry@koterov.ru> wrote:
> To me, this is a wonderful feature, thanks! I think many people would be
> happy if this patch woud be included to the mainstream (and it is quite
> short and simple).
> About name ordering - I think that the problem exists for objects:
> 1. Stored functions.
> 2. Foreign keys/triggers (objects which has owning objects).
> It is wonderful that you store all functions with the same name to the same
> file. To order them within this file we may simply compare the first
> definition line lexicographically (or - first line which differs one
> function definition from another).
> Foreign key/triggers ordering problem is described by me at
> http://www.mail-archive.com/pgsql-hackers@postgresql.org/msg114586.html
> The big problem is with triggers: many triggers may have the same name, but
> be bound to different tables. It would be great to include these triggers to
> table's definition or, at least, have separated files for each trigger+table
> pair.

Hmm... this sounds awfully specific to your particular environment.
My triggers never have the same names...  and on the flip side, I
don't see why someone couldn't want one function per file rather than
all same-named functions in one file.

I think the problem with this patch is that different people are
likely to want slightly different things, and there may not be any
single format that pleases everyone, and supporting too many variants
will become confusing for users and hard for us to maintain.  We're
going to need to agree on something that won't be perfect for
everyone, but will hopefully be a sufficient improvement for enough
people to be worth doing.

On the specific issue of overloaded functions, I have a feeling that
the only feasible option is going to be to put them all in the same
file.  If you put them in different files, the names will either be
very long (because they'll have to include the argument types) or
fairly incomprehensible (if you did something like hash the argument
types and append 8 hex digits to the function name) or not all that
static (if you use OIDs; or if you number them sequentially, like
foo1.sql, foo2.sql, foo3.sql, then foo3.sql might end up as foo2.sql
on a system where there are only two variants of foo, making diff not
work very well).

-- 
Robert Haas
EnterpriseDB: http://www.enterprisedb.com
The Enterprise PostgreSQL Company


Re: pg_dump --split patch

From
Joel Jacobson
Date:
2011/1/3 Robert Haas <robertmhaas@gmail.com>:
> will become confusing for users and hard for us to maintain.  We're
> going to need to agree on something that won't be perfect for
> everyone, but will hopefully be a sufficient improvement for enough
> people to be worth doing.

Good point.
I think we can at least agree the "bare minimum" is splitting per
namespace, object type and name.

> On the specific issue of overloaded functions, I have a feeling that
> the only feasible option is going to be to put them all in the same
> file.  If you put them in different files, the names will either be
> very long (because they'll have to include the argument types) or
> fairly incomprehensible (if you did something like hash the argument
> types and append 8 hex digits to the function name) or not all that
> static (if you use OIDs; or if you number them sequentially, like
> foo1.sql, foo2.sql, foo3.sql, then foo3.sql might end up as foo2.sql
> on a system where there are only two variants of foo, making diff not
> work very well).

I agree.
Even if the overloaded functions are not written in the same order,
you will quickly and easily note "function(s) of this particular name
has been changed", which should narrow down your
mind-mapping-change-grasping-exercise quite a lot.

--
Best regards,

Joel Jacobson
Glue Finance


Re: pg_dump --split patch

From
Tom Lane
Date:
Robert Haas <robertmhaas@gmail.com> writes:
> On the specific issue of overloaded functions, I have a feeling that
> the only feasible option is going to be to put them all in the same
> file.  If you put them in different files, the names will either be
> very long (because they'll have to include the argument types) or
> fairly incomprehensible (if you did something like hash the argument
> types and append 8 hex digits to the function name) or not all that
> static (if you use OIDs; or if you number them sequentially, like
> foo1.sql, foo2.sql, foo3.sql, then foo3.sql might end up as foo2.sql
> on a system where there are only two variants of foo, making diff not
> work very well).

If you put all the variants in the same file, diff is *still* not going
to work very well.  At least not unless you solve the problems that keep
pg_dump from dumping objects in a consistent order ... and once you do
that, you don't need this patch.

> I think the problem with this patch is that different people are
> likely to want slightly different things, and there may not be any
> single format that pleases everyone, and supporting too many variants
> will become confusing for users and hard for us to maintain.

Yeah, that's exactly it.  I can think of some possible uses for
splitting up pg_dump output, but frankly "to ease diff-ing" is not
one of them.  For that problem, it's nothing but a crude kluge that
only sort-of helps.  If we're to get anywhere on this, we need a
better-defined problem statement that everyone can agree is worth
solving and is well solved with this particular approach.
        regards, tom lane


Re: pg_dump --split patch

From
Joel Jacobson
Date:
2011/1/3 Tom Lane <tgl@sss.pgh.pa.us>:
> pg_dump from dumping objects in a consistent order ... and once you do
> that, you don't need this patch.
> Yeah, that's exactly it.  I can think of some possible uses for
> splitting up pg_dump output, but frankly "to ease diff-ing" is not
> one of them.  For that problem, it's nothing but a crude kluge that
> only sort-of helps.  If we're to get anywhere on this, we need a
> better-defined problem statement that everyone can agree is worth
> solving and is well solved with this particular approach.

The problem statement is only partly diffing, I think the two major
other problems with one-single-giant-schema-file is:

1. Automatically version controlling your production database schema
using a file based approach is not possible.
Splitting the schema would mean it's very simple to setup a cronjob
which automatically commits the schema changes every night. You would
be able to follow the changes by simply looking at the vcs log,
instead of grepping the log files and trying to figure out what
changed.

2. Splitting is a single option which reduces the need for any other
imaginable exclude/include options, such as "only export these
functions" or "do not export views" etc. Tables are of course the most
common thing you want to include/exclude in a dump, but there are
quite a lot of different object types, however, no single object type
is "important enough" to motivate a new pg_dump option to allow a
exclusion/inclusion option, but I think the sum of the need for such
an option for all object types is greather than the extra complexity
introduced by a patch consisting of only a few lines of code.

--
Best regards,

Joel Jacobson
Glue Finance


Re: pg_dump --split patch

From
Robert Haas
Date:
On Mon, Jan 3, 2011 at 1:34 PM, Tom Lane <tgl@sss.pgh.pa.us> wrote:
> Robert Haas <robertmhaas@gmail.com> writes:
>> On the specific issue of overloaded functions, I have a feeling that
>> the only feasible option is going to be to put them all in the same
>> file.  If you put them in different files, the names will either be
>> very long (because they'll have to include the argument types) or
>> fairly incomprehensible (if you did something like hash the argument
>> types and append 8 hex digits to the function name) or not all that
>> static (if you use OIDs; or if you number them sequentially, like
>> foo1.sql, foo2.sql, foo3.sql, then foo3.sql might end up as foo2.sql
>> on a system where there are only two variants of foo, making diff not
>> work very well).
>
> If you put all the variants in the same file, diff is *still* not going
> to work very well.  At least not unless you solve the problems that keep
> pg_dump from dumping objects in a consistent order ... and once you do
> that, you don't need this patch.

That's not really true.  It's a whole lot easier to look a diff of two
100-line files and then repeat that N times than to look at a single
diff of two N*100 line files.  I certainly spend enough of my
patch-review doing "git diff master <some particular source file>",
and then if what's going on isn't clear you can look at just that file
in more detail without worrying about every other source file in the
system.  And I have encountered this problem when comparing database
schemas (and sometimes data) also.  Yes, I've done that using diff.
Yes, it did suck.  Yes, I got it done before my boss fired me.

>> I think the problem with this patch is that different people are
>> likely to want slightly different things, and there may not be any
>> single format that pleases everyone, and supporting too many variants
>> will become confusing for users and hard for us to maintain.
>
> Yeah, that's exactly it.  I can think of some possible uses for
> splitting up pg_dump output, but frankly "to ease diff-ing" is not
> one of them.  For that problem, it's nothing but a crude kluge that
> only sort-of helps.  If we're to get anywhere on this, we need a
> better-defined problem statement that everyone can agree is worth
> solving and is well solved with this particular approach.

I have to admit I'm a bit unsold on the approach as well.  It seems
like you could write a short Perl script which would transform a text
format dump into the proposed format pretty easily, and if you did
that and published the script, then the next poor shmuck who had the
same problem could either use the script as-is or hack it up to meet
some slightly different set of requirements.  Or maybe you'd be better
off basing such a script on the custom or tar format instead, in order
to avoid the problem of misidentifying a line beginning with --- as a
comment when it's really part of a data item.  Or maybe even writing a
whole "schema diff" tool that would take two custom-format dumps as
inputs.

On the other hand, I can certainly think of times when even a pretty
dumb implementation of this would have saved me some time.

--
Robert Haas
EnterpriseDB: http://www.enterprisedb.com
The Enterprise PostgreSQL Company


Re: pg_dump --split patch

From
Tom Lane
Date:
Robert Haas <robertmhaas@gmail.com> writes:
> On Mon, Jan 3, 2011 at 1:34 PM, Tom Lane <tgl@sss.pgh.pa.us> wrote:
>> Yeah, that's exactly it. �I can think of some possible uses for
>> splitting up pg_dump output, but frankly "to ease diff-ing" is not
>> one of them. �For that problem, it's nothing but a crude kluge that
>> only sort-of helps. �If we're to get anywhere on this, we need a
>> better-defined problem statement that everyone can agree is worth
>> solving and is well solved with this particular approach.

> I have to admit I'm a bit unsold on the approach as well.  It seems
> like you could write a short Perl script which would transform a text
> format dump into the proposed format pretty easily, and if you did
> that and published the script, then the next poor shmuck who had the
> same problem could either use the script as-is or hack it up to meet
> some slightly different set of requirements.  Or maybe you'd be better
> off basing such a script on the custom or tar format instead, in order
> to avoid the problem of misidentifying a line beginning with --- as a
> comment when it's really part of a data item.  Or maybe even writing a
> whole "schema diff" tool that would take two custom-format dumps as
> inputs.

> On the other hand, I can certainly think of times when even a pretty
> dumb implementation of this would have saved me some time.

The basic objection that I have to this patch is that it proposes to
institutionalize a pretty dumb implementation.  And, as you mentioned,
once it's in there it'll be more or less set in stone because we aren't
going to want to support umpteen variants.

I like the idea of a postprocessing script a lot better --- it seems
like it wouldn't get in the way of people making their own variants.
And as you say it'd likely be pretty trivial to do.
        regards, tom lane


Re: pg_dump --split patch

From
Joel Jacobson
Date:
> Robert Haas <robertmhaas@gmail.com> writes:
>> I have to admit I'm a bit unsold on the approach as well.  It seems
>> like you could write a short Perl script which would transform a text
>> format dump into the proposed format pretty easily, and if you did
>> that and published the script, then the next poor shmuck who had the
>> same problem could either use the script as-is or hack it up to meet
>> some slightly different set of requirements.  Or maybe you'd be better

That's actually what I first did a couple of months ago,
https://github.com/gluefinance/parse_schema/blob/master/parse_schema.pl

My major concern of parsing the schema file is I would never fully
trust the output from the script, even if the regex is extremely
paranoid and really strict, there is still a risk it contains a bug.

If you cannot trust the output from the schema parse script, it's not
safe to use it to do a partial restore of objects.
Let's say you want to restore only a few functions from your backup
schema. Without --split, you would need to restore them manually
somehow. With the --split option, you could simply restore them from
the indivudual files, at least functions where only the source code
has been modified and not the arguments.

2011/1/3 Tom Lane <tgl@sss.pgh.pa.us>:
> The basic objection that I have to this patch is that it proposes to
> institutionalize a pretty dumb implementation.  And, as you mentioned,

What's dumb about it? It's simple and it works.
Please give me an idea of what a smart implementation would be, and I
will see if I can rewrite the patch.

> once it's in there it'll be more or less set in stone because we aren't
> going to want to support umpteen variants.

Yes, that's why it should be a "bare minimum" solution to the problem,
which is a lot better than no solution at all.

What other variants than the discussion on the path name structure can
you think of?

> I like the idea of a postprocessing script a lot better --- it seems
> like it wouldn't get in the way of people making their own variants.
> And as you say it'd likely be pretty trivial to do.

I'm quite sure such a postprocessing script has been written hundreds
of times over the past years by different postgres users, not even
submitting a question to the mailing list, since it's a quite
managable task for anyone with some regex and scripting experience.

Why not just provide a simple "bare minimum" solution to the "problem"
and let the remaining x % of the users who have a very specific need
write their own specialized script if they need to? It would save a
lot of time for all future users who need to write their own script to
solve the problem or google until they find my script or any other
script solving the same problem. Multiple almost identical solutions
to exactly the same problem is kind of time waste.

--
Best regards,

Joel Jacobson
Glue Finance


Re: pg_dump --split patch

From
Robert Haas
Date:
On Mon, Jan 3, 2011 at 2:46 PM, Joel Jacobson <joel@gluefinance.com> wrote:
> My major concern of parsing the schema file is I would never fully
> trust the output from the script, even if the regex is extremely
> paranoid and really strict, there is still a risk it contains a bug.

That could possibly be resolved by using the custom or tar formats.

> If you cannot trust the output from the schema parse script, it's not
> safe to use it to do a partial restore of objects.
> Let's say you want to restore only a few functions from your backup
> schema.

I don't think this is the right solution to that problem; or at least,
it's not the solution I want personally.  I want something like
--function=glob (and yes I do want that same thing for every other
object type, too, but PARTICULARLY for functions).  This might be a
next-best solution to that problem, but I'm still holding out for the
home run.

-- 
Robert Haas
EnterpriseDB: http://www.enterprisedb.com
The Enterprise PostgreSQL Company


Re: pg_dump --split patch

From
Dimitri Fontaine
Date:
Robert Haas <robertmhaas@gmail.com> writes:
> I have to admit I'm a bit unsold on the approach as well.  It seems
> like you could write a short Perl script which would transform a text
> format dump into the proposed format pretty easily, and if you did
> that and published the script, then the next poor shmuck who had the
> same problem could either use the script as-is or hack it up to meet
> some slightly different set of requirements.  Or maybe you'd be better
> off basing such a script on the custom or tar format instead, in order
> to avoid the problem of misidentifying a line beginning with --- as a
> comment when it's really part of a data item.  Or maybe even writing a
> whole "schema diff" tool that would take two custom-format dumps as
> inputs.
>
> On the other hand, I can certainly think of times when even a pretty
> dumb implementation of this would have saved me some time.

You mean like those:
 https://labs.omniti.com/labs/pgtreats/wiki/getddl https://github.com/dimitri/getddl

In this case hacking the perl version was impossible for me to hack on
so my version is in python, but still is somewhat of a fork.  I even
have an intermediate shell version somewhere that a colleague wrote, but
I much prefer maintaining python code.

Regards,
-- 
Dimitri Fontaine
http://2ndQuadrant.fr     PostgreSQL : Expertise, Formation et Support


Re: pg_dump --split patch

From
Robert Haas
Date:
On Mon, Jan 3, 2011 at 3:15 PM, Dimitri Fontaine <dimitri@2ndquadrant.fr> wrote:
>> On the other hand, I can certainly think of times when even a pretty
>> dumb implementation of this would have saved me some time.
>
> You mean like those:
>
>  https://labs.omniti.com/labs/pgtreats/wiki/getddl
>  https://github.com/dimitri/getddl

That's not exactly what I was trying to do, but it's in the ballpark...

--
Robert Haas
EnterpriseDB: http://www.enterprisedb.com
The Enterprise PostgreSQL Company


Re: pg_dump --split patch

From
Hannu Krosing
Date:
On 28.12.2010 22:44, Joel Jacobson wrote:


Sent from my iPhone

On 28 dec 2010, at 21:45, Gurjeet Singh <singh.gurjeet@gmail.com> wrote:

The problem I see with suffixing a sequence id to the objects with name collision is that one day the dump may name myfunc(int) as myfunc.sql and after an overloaded version is created, say myfunc(char, int), then the same myfunc(int) may be dumped in myfunc-2.sql, which again is non-deterministic.

I agree, good point!
Perhaps abbreviations are to prefer, e.g., myfunc_i, myfunc_i_c, etc to reduce the need of truncating filenames.



Also, it is a project policy that we do not introduce new features in back branches, so spending time on an 8.4.6 patch may not be the best use of your time.

My company is using 8.4 and needs this feature, so I'll have to patch it anyway :)
Start the easy way, by writing a (python|perl) filter for pg_dump -s output

Once this is done, convert it into a patch for pg_dump

--------------------------------------------
Hannu Krosing
http://www.2ndQuadrant.com/books/

Re: pg_dump --split patch

From
Hannu Krosing
Date:
On 28.12.2010 23:51, Tom Lane wrote:
> Andrew Dunstan<andrew@dunslane.net>  writes:
>> On 12/28/2010 04:44 PM, Joel Jacobson wrote:
>>> Perhaps abbreviations are to prefer, e.g., myfunc_i, myfunc_i_c, etc
>>> to reduce the need of truncating filenames.
>> I think that's just horrible. Does the i stand for integer or inet? And
>> it will get *really* ugly for type names with spaces in them ...
> You think spaces are bad, try slashes ;-)
>
> Not to mention the need for including schemas in typenames sometimes.
> I think you're going to have a real problem trying to fully describe a
> function's signature in a file name of reasonable max length.
something like

funcname_<number-of-arguments>_<hash_of_argument_type_list>.sql

seems like a reasonable compromise - you can find the function you are 
looking for without too much searching, even when overloaded and the 
uniqueness is still guaranteed.

--------------------------------------------
Hannu Krosing
http://www.2ndQuadrant.com/books/





Re: pg_dump --split patch

From
Hannu Krosing
Date:
On 28.12.2010 17:00, Joel Jacobson wrote:
> Dear fellow hackers,
>
> Problem: A normal diff of two slightly different schema dump files 
> (pg_dump -s), will not produce a user-friendly diff, as you get all 
> changes in the same file.
>
Another Solution: I have used a python script for spliiting "dump -s" 
output into a directory structure for years

A structure that seems to work well is

1. <database_name>.load.sql - file containing \i for all other files
2. tree of
schema1\    +- <objtype1>\    |    + obj_of_type1_name1.sql    |    + ...    +-<objtype2>\ ...

That is, first have one directory per schema on top level and then one 
for each type of objects
inside have one file per object.
Some things which may have non-unique names, like functions need extra 
"uniquefication", like adding a underscore (or colon) separated list of 
argument types at the end.

Most times you want to  leave out the comment lines with OIDs so that 
you can diff the files against another version

------------------------------------------
Hannu Krosing
http://www.2ndQuadrant.com/books/


Re: pg_dump --split patch

From
Greg Smith
Date:
Joel Jacobson wrote:
> To understand a change to my database functions, I would start by 
> looking at the top-level, only focusing on the names of the functions 
> modified/added/removed.
> At this stage, you want as little information as possible about each 
> change, such as only the names of the functions.
> To do this, get a list of changes functions, you cannot compare two 
> full schema plain text dumps using diff, as it would only reveal the 
> lines changed, not the name of the functions, unless you are lucky to 
> get the name of the function within the (by default) 3 lines of copied 
> context.
>
> While you could increase the number of copied lines of context to a 
> value which would ensure you would see the name of the function in the 
> diff, that is not feasible if you want to quickly "get a picture" of 
> the code areas modified, since you would then need to read through 
> even more lines of diff output.

I can agree on some use cases you've outlined, where there's merit to 
the general idea of your patch.  But as an aside, you really should 
launch an investigation into some better diff tools if this is how 
you're doing this type of work.  Last week I reviewed 3K lines worth of 
changes from two versions of a 12K line schema dump I'd never seen 
before in a couple of hours using kdiff3.  I'd have killed myself before 
finishing if I had to do the same job with traditional diff as you're 
describing it here.

-- 
Greg Smith   2ndQuadrant US    greg@2ndQuadrant.com   Baltimore, MD
PostgreSQL Training, Services and Support        www.2ndQuadrant.us
"PostgreSQL 9.0 High Performance": http://www.2ndQuadrant.com/books



Re: pg_dump --split patch

From
Robert Haas
Date:
On Mon, Jan 3, 2011 at 2:18 PM, Tom Lane <tgl@sss.pgh.pa.us> wrote:
> Robert Haas <robertmhaas@gmail.com> writes:
>> On Mon, Jan 3, 2011 at 1:34 PM, Tom Lane <tgl@sss.pgh.pa.us> wrote:
>>> Yeah, that's exactly it.  I can think of some possible uses for
>>> splitting up pg_dump output, but frankly "to ease diff-ing" is not
>>> one of them.  For that problem, it's nothing but a crude kluge that
>>> only sort-of helps.  If we're to get anywhere on this, we need a
>>> better-defined problem statement that everyone can agree is worth
>>> solving and is well solved with this particular approach.
>
>> I have to admit I'm a bit unsold on the approach as well.  It seems
>> like you could write a short Perl script which would transform a text
>> format dump into the proposed format pretty easily, and if you did
>> that and published the script, then the next poor shmuck who had the
>> same problem could either use the script as-is or hack it up to meet
>> some slightly different set of requirements.  Or maybe you'd be better
>> off basing such a script on the custom or tar format instead, in order
>> to avoid the problem of misidentifying a line beginning with --- as a
>> comment when it's really part of a data item.  Or maybe even writing a
>> whole "schema diff" tool that would take two custom-format dumps as
>> inputs.
>
>> On the other hand, I can certainly think of times when even a pretty
>> dumb implementation of this would have saved me some time.
>
> The basic objection that I have to this patch is that it proposes to
> institutionalize a pretty dumb implementation.  And, as you mentioned,
> once it's in there it'll be more or less set in stone because we aren't
> going to want to support umpteen variants.
>
> I like the idea of a postprocessing script a lot better --- it seems
> like it wouldn't get in the way of people making their own variants.
> And as you say it'd likely be pretty trivial to do.

I notice that this patch is marked as "Needs Review" in the CommitFest
application, but I think it's fair to say that there's no consensus to
commit something along these lines.  Accordingly, I'm going to mark it
"Returned with Feedback".  There is clearly a need for better tooling
in this area, but I think there's a great deal of legitimate doubt
about whether this is the right solution to that problem.

--
Robert Haas
EnterpriseDB: http://www.enterprisedb.com
The Enterprise PostgreSQL Company


Re: pg_dump --split patch

From
Marko Tiikkaja
Date:
Hi,

Now that the (at least as far as I know) last ordering problem in
pg_dump has been solved [1], I'm going to attempt resurrecting this old
thread.

It seemed to me that the biggest objections to this patch in the old
discussions were directed at the implementation, which I have tried to
improve.  The attached patch implements the actual splitting in a new
backup format.

The general output scheme looks like this:
    schemaname/OBJECT_TYPES/object_name.sql,

but there are some exceptions.

Overloaded functions are dumped into the same file.  Object names are
encoded into the POSIX Portable Filename Character Set ([a-z0-9._-]) by
replacing any characters outside that set with an underscore.

Restoring the dump is supported through an index.sql file containing
statements which include (through \i) the actual object files in the
dump directory.

Any thoughts?  Objections on the idea or the implementation?


[1]:
http://git.postgresql.org/gitweb/?p=postgresql.git;a=commit;h=7b583b20b1c95acb621c71251150beef958bb603


Regards,
Marko Tiikkaja

Attachment

Re: pg_dump --split patch

From
Dimitri Fontaine
Date:
Hi,

Marko Tiikkaja <pgmail@joh.to> writes:
> The general output scheme looks like this:
>    schemaname/OBJECT_TYPES/object_name.sql,

I like this feature, I actually did have to code it myself in the past
and several other people did so, so we already have at least 3 copies of
`getddl` variants around. I really think this feature should be shipped
by default with PostgreSQL.

I don't much care for the all uppercase formating of object type
directories in your patch though.

> Overloaded functions are dumped into the same file.  Object names are
> encoded into the POSIX Portable Filename Character Set ([a-z0-9._-]) by
> replacing any characters outside that set with an underscore.

What happens if you have a table foo and another table "FoO"?

> Restoring the dump is supported through an index.sql file containing
> statements which include (through \i) the actual object files in the dump
> directory.

I think we should be using \ir now that we have that.

> Any thoughts?  Objections on the idea or the implementation?

As far as the implementation goes, someone with more experience on the
Archiver Handles should have a look. To me, it looks like you are trying
to shoehorn your feature in the current API and that doesn't feel good.

The holly grail here that we've been speaking about in the past would be
to separate out tooling and formats so that we have:
  pg_dump | pg_restore  pg_export | psql

In that case we would almost certainly need libpgdump to share the code,
and we maybe could implement a binary output option for pg_dump too
(yeah, last time it was proposed we ended up with bytea_output = 'hex').

That libpgdump idea basically means we won't have the --split feature in
9.3, and that's really bad, as we already are some releases late on
delivering that, in my opinion.

Maybe the pg_export and pg_dump tool could share code by just #include
magic rather than a full blown lib in a first incantation?

Regards,
-- 
Dimitri Fontaine
http://2ndQuadrant.fr     PostgreSQL : Expertise, Formation et Support



Re: pg_dump --split patch

From
Marko Tiikkaja
Date:
Hi,

On 16/11/2012 15:52, Dimitri Fontaine wrote:
> Marko Tiikkaja <pgmail@joh.to> writes:
>> The general output scheme looks like this:
>>     schemaname/OBJECT_TYPES/object_name.sql,
>
> I like this feature, I actually did have to code it myself in the past
> and several other people did so, so we already have at least 3 copies of
> `getddl` variants around. I really think this feature should be shipped
> by default with PostgreSQL.
>
> I don't much care for the all uppercase formating of object type
> directories in your patch though.

*shrug*  I have no real preference to one way or the other.

>> Overloaded functions are dumped into the same file.  Object names are
>> encoded into the POSIX Portable Filename Character Set ([a-z0-9._-]) by
>> replacing any characters outside that set with an underscore.
>
> What happens if you have a table foo and another table "FoO"?

They would go to the same file.  If you think there are technical issues 
behind that decision (e.g. the dump would not restore), I would like to 
hear an example case.

On the other hand, some people might find it preferrable to have them in 
different files (for example foo, foo.1, foo.2 etc).  Or some might 
prefer some other naming scheme.  One of the problems with this patch is 
exactly that people prefer different things, and providing switches for 
all of the different options people come up with would mean a lot of 
switches. :-(

>> Restoring the dump is supported through an index.sql file containing
>> statements which include (through \i) the actual object files in the dump
>> directory.
>
> I think we should be using \ir now that we have that.

Good point, will have to get that fixed.

>> Any thoughts?  Objections on the idea or the implementation?
>
> As far as the implementation goes, someone with more experience on the
> Archiver Handles should have a look. To me, it looks like you are trying
> to shoehorn your feature in the current API and that doesn't feel good.

It feels a bit icky to me too, but I didn't feel comfortable with 
putting in a lot of work to refactor the API because of how 
controversial this feature is.

> The holly grail here that we've been speaking about in the past would be
> to separate out tooling and formats so that we have:
>
>     pg_dump | pg_restore
>     pg_export | psql
>
> In that case we would almost certainly need libpgdump to share the code,
> and we maybe could implement a binary output option for pg_dump too
> (yeah, last time it was proposed we ended up with bytea_output = 'hex').

While I agree that this idea - when implemented - would be nicer in 
practically every way, I'm not sure I want to volunteer to do all the 
necessary work.

> That libpgdump idea basically means we won't have the --split feature in
> 9.3, and that's really bad, as we already are some releases late on
> delivering that, in my opinion.
>
> Maybe the pg_export and pg_dump tool could share code by just #include
> magic rather than a full blown lib in a first incantation?

That's one idea..


Regards,
Marko Tiikkaja



Re: pg_dump --split patch

From
Dimitri Fontaine
Date:
Marko Tiikkaja <pgmail@joh.to> writes:
>> What happens if you have a table foo and another table "FoO"?
>
> They would go to the same file.  If you think there are technical issues
> behind that decision (e.g. the dump would not restore), I would like to hear
> an example case.

I didn't try the patch itself yet so I wanted to hear that's something
you did actually try out, that's about it :) As soon as we're past the
initial agreement on this patch landing in (which I really want to see
happen), I'll devote some time on testing it.

> On the other hand, some people might find it preferrable to have them in
> different files (for example foo, foo.1, foo.2 etc).  Or some might prefer
> some other naming scheme.  One of the problems with this patch is exactly
> that people prefer different things, and providing switches for all of the
> different options people come up with would mean a lot of switches. :-(

I think this facility should provide something simple and useful, and
not something tasty. Database backups and exports are not meant to cater
with taste, they are meant to be easy to restore. In that very case, we
want to have a set of properly organized SQL files for doing partial
restores, right?

> It feels a bit icky to me too, but I didn't feel comfortable with putting in
> a lot of work to refactor the API because of how controversial this feature
> is.

+1

>>     pg_dump | pg_restore
>>     pg_export | psql
>
> While I agree that this idea - when implemented - would be nicer in
> practically every way, I'm not sure I want to volunteer to do all the
> necessary work.

What I think needs to happen now is a commiter's buy in that we want to
get there at some point and that your current patch is not painting us
into any corner now. So that we can accept it and have a documented path
forward.

Regards,
--
Dimitri Fontaine                                        06 63 07 10 78
http://2ndQuadrant.fr     PostgreSQL : Expertise, Formation et Support



Re: pg_dump --split patch

From
Dimitri Fontaine
Date:
Dimitri Fontaine <dimitri@2ndQuadrant.fr> writes:
>>>     pg_dump | pg_restore
>>>     pg_export | psql
>>
>> While I agree that this idea - when implemented - would be nicer in
>> practically every way, I'm not sure I want to volunteer to do all the
>> necessary work.
>
> What I think needs to happen now is a commiter's buy in that we want to
> get there at some point and that your current patch is not painting us
> into any corner now. So that we can accept it and have a documented path
> forward.

Just stumbled accross this message while reading some older threads
about the current topic:
 http://archives.postgresql.org/pgsql-hackers/2010-12/msg02496.php

Where Robert Treat said:
> I've both enjoyed reading this thread and seeing this wheel reinvented
> yet again, and wholeheartedly +1 the idea of building this directly
> into pg_dump. (The only thing better would be to make everything thing
> sql callable, but that's a problem for another day).

I know Andrew has been working on his "Retail DDL" project which is
basically a bunch of server-side functions that spits out SQL object
definitions. Andrew, were you able to make progress on that project?

On the other hand, pg_dump -Fs still is something I would like to have
as a complement to Andrew's facility.

Regards,
-- 
Dimitri Fontaine
http://2ndQuadrant.fr     PostgreSQL : Expertise, Formation et Support



Re: pg_dump --split patch

From
Andrew Dunstan
Date:
On 11/19/2012 09:07 AM, Dimitri Fontaine wrote:
> Dimitri Fontaine <dimitri@2ndQuadrant.fr> writes:
>>>>      pg_dump | pg_restore
>>>>      pg_export | psql
>>> While I agree that this idea - when implemented - would be nicer in
>>> practically every way, I'm not sure I want to volunteer to do all the
>>> necessary work.
>> What I think needs to happen now is a commiter's buy in that we want to
>> get there at some point and that your current patch is not painting us
>> into any corner now. So that we can accept it and have a documented path
>> forward.
> Just stumbled accross this message while reading some older threads
> about the current topic:
>
>    http://archives.postgresql.org/pgsql-hackers/2010-12/msg02496.php
>
> Where Robert Treat said:
>> I've both enjoyed reading this thread and seeing this wheel reinvented
>> yet again, and wholeheartedly +1 the idea of building this directly
>> into pg_dump. (The only thing better would be to make everything thing
>> sql callable, but that's a problem for another day).
> I know Andrew has been working on his "Retail DDL" project which is
> basically a bunch of server-side functions that spits out SQL object
> definitions. Andrew, were you able to make progress on that project?
>
> On the other hand, pg_dump -Fs still is something I would like to have
> as a complement to Andrew's facility.
>


No, sorry, it's on hold - I'm finding trouble finding time to work on it.

cheers

andrew



Re: pg_dump --split patch

From
Alvaro Herrera
Date:
Marko Tiikkaja wrote:

> On 16/11/2012 15:52, Dimitri Fontaine wrote:
> >
> >What happens if you have a table foo and another table "FoO"?
>
> They would go to the same file.  If you think there are technical
> issues behind that decision (e.g. the dump would not restore), I
> would like to hear an example case.

create table foo (a int, b text);
create type bar as (stuff foo);
create table "FoO" (more bar);


--
Álvaro Herrera                http://www.2ndQuadrant.com/
PostgreSQL Development, 24x7 Support, Training & Services