Re: 7.4.6 pg_dump failed - Mailing list pgsql-admin

From Tom Lane
Subject Re: 7.4.6 pg_dump failed
Date
Msg-id 15489.1106756375@sss.pgh.pa.us
Whole thread Raw
In response to 7.4.6 pg_dump failed  (Marty Scholes <marty@outputservices.com>)
List pgsql-admin
Marty Scholes <marty@outputservices.com> writes:
> A pg_dump of one table ran for 28:53:29.50 and produced a 30 GB dump
> before it aborted with:

> pg_dump: dumpClasses(): SQL command failed
> pg_dump: Error message from server: out of memory for query result
> pg_dump: The command was: FETCH 100 FROM _pg_dump_cursor

Even though it says "from server", this is actually an out-of-memory
problem inside pg_dump, or more specifically inside libpq.

> The table contains a text field that could contain several hundred MB of
> data, although always less than 2GB.

"Could contain"?  What's the actual maximum field width, and how often
do very wide values occur?  I don't recall the exact space allocation
algorithms inside libpq, but I'm wondering if it could choke on such a
wide row.

You might have better luck if you didn't use -d.

            regards, tom lane

pgsql-admin by date:

Previous
From: Marty Scholes
Date:
Subject: 7.4.6 pg_dump failed
Next
From: "Haron, Charles"
Date:
Subject: Re: Trouble Escaping Quotes