Re: pg_dump additional options for performance - Mailing list pgsql-hackers

From Alvaro Herrera
Subject Re: pg_dump additional options for performance
Date
Msg-id 20080226191836.GL5763@alvh.no-ip.org
Whole thread Raw
In response to Re: pg_dump additional options for performance  (Tom Lane <tgl@sss.pgh.pa.us>)
Responses Re: pg_dump additional options for performance  (Simon Riggs <simon@2ndquadrant.com>)
List pgsql-hackers
Tom Lane wrote:
> "Joshua D. Drake" <jd@commandprompt.com> writes:
> > IMO the place to start is COPY which is per my tests, slow. Multi
> > worker connection restore is great and I have proven that with some
> > work it can provide o.k. results but it is certainly not acceptable.
> 
> It was already pointed out to you that we can hope for only incremental
> speedups in COPY per se.  Don't be too quick to dismiss the discussion
> of large-grain parallelism, because I don't see anything else within
> reach that might give integer multiples rather than percentage points.

Well, one idea would be dividing the input file in similarly-sized parts
and giving each one to a different COPY process.  This would help in
cases where you have a single very large table to restore.

Another thing we could do is selective binary output/input for bytea
columns, to avoid the escaping step.

-- 
Alvaro Herrera                                http://www.CommandPrompt.com/
PostgreSQL Replication, Consulting, Custom Development, 24x7 support


pgsql-hackers by date:

Previous
From: Greg Smith
Date:
Subject: Re: pg_dump additional options for performance
Next
From: Simon Riggs
Date:
Subject: Re: pg_dump additional options for performance