Re: pg_upgrade: can I use same binary for old & new? - Mailing list pgsql-general

From Adrian Klaver
Subject Re: pg_upgrade: can I use same binary for old & new?
Date
Msg-id 0de4c4cb-7c54-4c4d-a50d-6ef6e005f724@aklaver.com
Whole thread Raw
In response to Re: pg_upgrade: can I use same binary for old & new?  (Pierre Fortin <pf@pfortin.com>)
List pgsql-general
On 7/5/25 12:19, Pierre Fortin wrote:
> On Sat, 05 Jul 2025 14:30:20 -0400 Tom Lane wrote:
> 
> Forgive my ignorance; always trying to learn more... :)
> 
>> pf@pfortin.com writes:
>>> On Sat, 5 Jul 2025 11:11:32 -0700 Adrian Klaver wrote:
>>>> How did you measure above?
>>
>>> # du -sb /var/lib/pgsql/data
>>> 8227910662297   /var/lib/pgsql/data
>>
>> It's likely that there's a deal of bloat in that.  Even if there's not
>> much bloat, this number will include indexes and WAL data that don't
>> appear in pg_dump output.
> 
> Does this imply that on restore, I'll have to re-index everything?

The dump file includes CREATE INDEX commands and per:

https://www.postgresql.org/docs/current/sql-createindex.html

"Creating an index can interfere with regular operation of a database. 
Normally PostgreSQL locks the table to be indexed against writes and 
performs the entire index build with a single scan of the table. Other 
transactions can still read the table, but if they try to insert, 
update, or delete rows in the table they will block until the index 
build is finished. This could have a severe effect if the system is a 
live production database. Very large tables can take many hours to be 
indexed, and even for smaller tables, an index build can lock out 
writers for periods that are unacceptably long for a production system."

Which is why pg_restore:

https://www.postgresql.org/docs/current/app-pgrestore.html

has:

"-j number-of-jobs
--jobs=number-of-jobs

     Run the most time-consuming steps of pg_restore — those that load 
data, create indexes, or create constraints — concurrently, using up to 
number-of-jobs concurrent sessions. This option can dramatically reduce 
the time to restore a large database to a server running on a 
multiprocessor machine. This option is ignored when emitting a script 
rather than connecting directly to a database server."


> 
>>>> What was the pg_dump command?
>>
>>> Didn't try given:
>>> $ df /mnt/db
>>> Filesystem      Size  Used Avail Use% Mounted on
>>> /dev/sdh1        17T   13T  3.0T  82% /mnt/db
>>
>> I'd say give it a try; be sure to use one of the pg_dump modes
>> that compress the data.
> 
> OK...  I failed to mention I have several databases in this cluster; so
> digging into pg_dumpall, I see:
>     --binary-upgrade
>      This option is for use by in-place upgrade utilities. Its use for
>      other purposes is not recommended or supported. The behavior of the
>      option may change in future releases without notice.
> 
> pg_upgrade has --link option; but I'm puzzled by this option in a
> dumpall/restore process. My imagination wonders if this alludes to a way
> to do something like:
>   pg_dumpall --globals-only --roles-only --schema-only ...
> Would restoring this be a way to update only the control structures? Big
> assumption that the actual data remains untouched...
> 
> Inquiring mind...  :)
> 
> Back to my upgrade issue...
> All my DBs are static (only queries once loaded). Assuming the dumpall
> file fits on one of my drives:
>   pg_dumpall -f <path>/PG.backup -v
> appears to be all I need? pg_dump has compression by default; but I don't
> see compression with dumpall other than for TOAST.
> 
> Thanks, You guys are awesome!
>   
>>             regards, tom lane
> 
> 

-- 
Adrian Klaver
adrian.klaver@aklaver.com




pgsql-general by date:

Previous
From: Adrian Klaver
Date:
Subject: Re: pg_upgrade: can I use same binary for old & new?
Next
From: Adrian Klaver
Date:
Subject: Re: pg_upgrade: can I use same binary for old & new?