Re: In-order pg_dump (or in-order COPY TO) - Mailing list pgsql-general
From | Achilleas Mantzios |
---|---|
Subject | Re: In-order pg_dump (or in-order COPY TO) |
Date | |
Msg-id | e48a1211-ca90-4944-958d-d8fb29ab944c@cloud.gatewaynet.com Whole thread Raw |
In response to | Re: In-order pg_dump (or in-order COPY TO) (Dimitrios Apostolou <jimis@gmx.net>) |
List | pgsql-general |
On 9/4/25 19:08, Dimitrios Apostolou wrote: > Hi Álvaro and Greg, > > On Thursday 2025-09-04 14:02, Álvaro Herrera wrote: > >> It's generally considered nowadays that pg_dump is not the best option >> to create backups of very large databases. You may be better served by >> using a binary backup tool -- something like Barman. With current >> Postgres releases you can create incremental backups, which would >> probably be more effective at deduplicating than playing with pg_dump's >> TOC, because it's based on what actually happens to the data. Barman >> provides support for hook scripts, which perhaps can be used to transfer >> the backup files to Borg. (I haven't actually tried to do this, but the >> Barman developers talk about using them to transfer the backups to tape, >> so I imagine getting them to play with Borg it's a Simple Matter of >> Programming.) > > On Wed, 27 Aug 2025, Greg Sabino Mullane wrote: > >> I suggest looking into pgBackRest, and it's block incremental >> feature, which sounds similar to what you are doing. But it also does >> it with parallel processes, and can do things like grab backup files >> from your replicas, plus a lot of other features. > > > if I'm not mistaken, both Barman and pgBackRest are based on physical > dumps of the database (pg_basebackup). At the start of this project I > had evaluated pg_basebackup, but decided logical backup fitted my > needs better. > > + pg_basebackup was slower, measuring speeds of around 10MB/s, because > of issues with 8KB page size and compressed btrfs (see [1]; situation > has been improved both on the postgres side and the kernel side; > I'm not sure how pg_basebackup fares today). > > + pg_basebackup was much bigger, because of including indices etc. As a > result of size and speed, pg_basebackup was also taking a longer time. > > + physical dumps would change a lot during maintenance (vacuum full, > cluster etc) while the data would remain the same. This would > reduce the effect of deduplication and increase size requirements even > further. At that point in time I did not expect logical dumps to > change too, when the data hasn't changed. > > + I use logical dumps as a tool, not only as a backup, to copy the > database to other servers with different postgresql versions. > > + I also use it to verify the VCS-committed SQL schema: doing pg_restore > --data-only on an already created database will fail if the SQL schema > had been modified on the original server without committing the > changes. > > + Finally I don't really need all the advanced features that physical > replication offers, like HA, PITR, load balancing. It's a > non-mission-critical service that can take a little time off in case > of disaster recovery. > > [1] > https://www.postgresql.org/message-id/flat/218fa2e0-bc58-e469-35dd-c5cb35906064%40gmx.net > if that helps , this is a writeup I had some some years ago comparing pgbackrest , barman and probackup : https://severalnines.com/blog/current-state-open-source-backup-management-postgresql/ > > Regards, > Dimitris
pgsql-general by date: