dump of 700 GB database - Mailing list pgsql-general

From karsten vennemann
Subject dump of 700 GB database
Date
Msg-id C5A82EFB1B824835BADC754969A234BE@snuggie
Whole thread Raw
Responses Re: dump of 700 GB database  (John R Pierce <pierce@hogranch.com>)
List pgsql-general

I have to write a 700 GB large database to a dump to clean out a lot of dead records on an Ubuntu server with postgres 8.3.8. What is the proper procedure to succeed with this - last time the dump stopped at 3.8 GB size I guess. Should I combine the -Fc option of pg_dump and and the split command ?
I thought something like
"pg_dump -Fc test | split -b 1000m - testdb.dump"
might work ?
Karsten
 
Terra GIS LTD
Seattle, WA, USA 
 

pgsql-general by date:

Previous
From: Pavel Stehule
Date:
Subject: Re: more than 2GB data string save
Next
From: Scott Marlowe
Date:
Subject: Re: more than 2GB data string save