Table of Contents
pg_probackup — manage backup and recovery of PostgreSQL database clusters
pg_probackup
version
pg_probackup
help
[command
]
pg_probackup
init
-B
backup_dir
pg_probackup
add-instance
-B
backup_dir
-D
data_dir
--instance
instance_name
pg_probackup
del-instance
-B
backup_dir
--instance
instance_name
pg_probackup
set-config
-B
backup_dir
--instance
instance_name
[option
...]
pg_probackup
set-backup
-B
backup_dir
--instance
instance_name
-i
backup_id
[option
...]
pg_probackup
show-config
-B
backup_dir
--instance
instance_name
[--format=
]format
pg_probackup
show
-B
backup_dir
[option
...]
pg_probackup
backup
-B
backup_dir
--instance
instance_name
-b
backup_mode
[option
...]
pg_probackup
restore
-B
backup_dir
--instance
instance_name
[option
...]
pg_probackup
checkdb
-B
backup_dir
--instance
instance_name
-D
data_dir
[option
...]
pg_probackup
validate
-B
backup_dir
[option
...]
pg_probackup
merge
-B
backup_dir
--instance
instance_name
-i
backup_id
[option
...]
pg_probackup
delete
-B
backup_dir
--instance
instance_name
{ -i
backup_id
| --delete-wal
| --delete-expired
| --merge-expired
} [option
...]
pg_probackup
archive-push
-B
backup_dir
--instance
instance_name
--wal-file-path
wal_file_path
--wal-file-name
wal_file_name
[option
...]
pg_probackup
archive-get
-B
backup_dir
--instance
instance_name
--wal-file-path
wal_file_path
--wal-file-name
wal_file_name
[option
...]
pg_probackup
catchup
-b
catchup_mode
--source-pgdata
=path_to_pgdata_on_remote_server
--destination-pgdata
=path_to_local_dir
[option
...]
pg_probackup is a utility to manage backup and recovery of PostgreSQL database clusters. It is designed to perform periodic backups of the PostgreSQL instance that enable you to restore the server in case of a failure. pg_probackup supports PostgreSQL 11 or higher.
As compared to other backup solutions, pg_probackup offers the following benefits that can help you implement different backup strategies and deal with large amounts of data:
Incremental backup: with three different incremental modes, you can plan the backup strategy in accordance with your data flow. Incremental backups allow you to save disk space and speed up backup as compared to taking full backups. It is also faster to restore the cluster by applying incremental backups than by replaying WAL files.
Incremental restore: speed up restore from backup by reusing valid unchanged pages available in PGDATA.
Validation: automatic data consistency checks and on-demand backup validation without actual data recovery.
Verification: on-demand verification of PostgreSQL instance with the checkdb command.
Retention: managing WAL archive and backups in accordance with retention policy. You can configure retention policy based on recovery time or the number of backups to keep, as well as specify time to live (TTL) for a particular backup. Expired backups can be merged or deleted.
Parallelization: running backup, restore, merge, delete, validate, and checkdb processes on multiple parallel threads.
Compression: storing backup data in a compressed state to save disk space.
Deduplication: saving disk space by excluding non-data
files (such as _vm
or _fsm
)
from incremental backups if these files have not changed since
they were copied into one of the previous backups in this incremental chain.
Remote operations: backing up PostgreSQL instance located on a remote system or restoring a backup remotely.
Backup from standby: avoiding extra load on master by taking backups from a standby server.
External directories: backing up files and directories
located outside of the PostgreSQL data
directory (PGDATA
), such as scripts, configuration
files, logs, or SQL dump files.
Backup catalog: getting the list of backups and the corresponding meta information in plain text or JSON formats.
Archive catalog: getting the list of all WAL timelines and the corresponding meta information in plain text or JSON formats.
Partial restore: restoring only the specified databases.
Catchup: cloning a PostgreSQL instance for a fallen-behind standby server to “catch up” with master.
To manage backup data, pg_probackup creates a backup catalog. This is a directory that stores all backup files with additional meta information, as well as WAL archives required for point-in-time recovery. You can store backups for different instances in separate subdirectories of a single backup catalog.
Using pg_probackup, you can take full or incremental backups:
FULL backups contain all the data files required to restore the database cluster.
Incremental backups operate at the page level, only storing the data that has changed since the previous backup. It allows you to save disk space and speed up the backup process as compared to taking full backups. It is also faster to restore the cluster by applying incremental backups than by replaying WAL files. pg_probackup supports the following modes of incremental backups:
DELTA backup. In this mode, pg_probackup reads all data files in the data directory and copies only those pages that have changed since the previous backup. This mode can impose read-only I/O pressure equal to a full backup.
PAGE backup. In this mode, pg_probackup scans all WAL files in the archive from the moment the previous full or incremental backup was taken. Newly created backups contain only the pages that were mentioned in WAL records. This requires all the WAL files since the previous backup to be present in the WAL archive. If the size of these files is comparable to the total size of the database cluster files, speedup is smaller, but the backup still takes less space. You have to configure WAL archiving as explained in Setting up continuous WAL archiving to make PAGE backups.
PTRACK backup. In this mode, PostgreSQL tracks page changes on the fly. Continuous archiving is not necessary for it to operate. Each time a relation page is updated, this page is marked in a special PTRACK bitmap. Tracking implies some minor overhead on the database server operation, but speeds up incremental backups significantly.
pg_probackup can take only physical online backups, and online backups require WAL for consistent recovery. So regardless of the chosen backup mode (FULL, PAGE or DELTA), any backup taken with pg_probackup must use one of the following WAL delivery modes:
ARCHIVE. Such backups rely on continuous archiving to ensure consistent recovery. This is the default WAL delivery mode.
STREAM. Such backups include all the files required to restore the cluster to a consistent state at the time the backup was taken. Regardless of continuous archiving having been set up or not, the WAL segments required for consistent recovery are streamed via replication protocol during backup and included into the backup files. That's why such backups are called autonomous, or standalone.
pg_probackup currently has the following limitations:
pg_probackup only supports PostgreSQL 9.5 and higher.
The remote mode is not supported on Windows systems.
On Unix systems, for PostgreSQL 11,
a backup can be made only by the same OS user that has started the PostgreSQL
server. For example, if PostgreSQL server is started by
user postgres
, the backup
command must also be run
by user postgres
. To satisfy this requirement when taking backups in the
remote mode using SSH, you must set
--remote-user
option to postgres
.
For PostgreSQL 9.5, functions
pg_create_restore_point(text)
and
pg_switch_xlog()
can be executed only if
the backup role is a superuser, so backup of a
cluster with low amount of WAL traffic by a non-superuser
role can take longer than the backup of the same cluster by
a superuser role.
The PostgreSQL server from which the backup was taken and the restored server must be compatible by the block_size and wal_block_size parameters and have the same major release number. Depending on cluster configuration, PostgreSQL itself may apply additional restrictions, such as CPU architecture or libc/icu versions.
To quickly get started with pg_probackup, complete the steps below. This will set up FULL and DELTA backups in the remote mode and demonstrate some basic pg_probackup operations. In the following, these terms are used:
backup
— PostgreSQL
role used to connect to the PostgreSQL
cluster.
backupdb
— database used to connect to the
PostgreSQL cluster.
backup_host
— host with the backup catalog.
backup_user
— user on
backup_host
running all pg_probackup
operations.
/mnt/backups
— directory on
backup_host
where the backup catalog is stored.
postgres_host
— host with the
PostgreSQL cluster.
postgres
— user on
postgres_host
under which
PostgreSQL cluster processes are running.
/var/lib/postgresql/16/main
—
PostgreSQL data directory on
postgres_host
.
Install pg_probackup on both backup_host
and postgres_host
.
Set up an SSH connection from backup_host
to postgres_host
.
Configure your database cluster for STREAM backups.
Initialize the backup catalog:
backup_user@backup_host:~$ pg_probackup-16 init -B /mnt/backups INFO: Backup catalog '/mnt/backups' successfully initialized
Add a backup instance called mydb
to the backup catalog:
backup_user@backup_host:~$ pg_probackup-16 add-instance \ -B /mnt/backups \ -D /var/lib/pgpro/std-16/data \ --instance=node \ --remote-host=postgres_host \ --remote-user=postgres INFO: Instance 'node' successfully initialized
Make a FULL backup:
backup_user@backup_host:~$ pg_probackup-16 backup \ -B /mnt/backups \ -b FULL \ --instance=node \ --stream \ --compress-algorithm=zlib \ --remote-host=postgres_host \ --remote-user=postgres \ -U backup \ -d backupdb INFO: Backup start, pg_probackup version: 2.5.15, instance: node, backup ID: SCUN1Q, backup mode: FULL, wal mode: STREAM, remote: true, compress-algorithm: zlib, compress-level: 1 INFO: This PostgreSQL instance was initialized with data block checksums. Data block corruption will be detected INFO: Database backup start INFO: wait for pg_backup_start() INFO: Wait for WAL segment /mnt/backups/backups/node/SCUN1Q/database/pg_wal/000000010000000000000008 to be streamed INFO: PGDATA size: 96MB INFO: Current Start LSN: 0/8000028, TLI: 1 INFO: Start transferring data files INFO: Data files are transferred, time elapsed: 1s INFO: wait for pg_stop_backup() INFO: pg_stop backup() successfully executed INFO: stop_lsn: 0/800BBD0 INFO: Getting the Recovery Time from WAL INFO: Syncing backup files to disk INFO: Backup files are synced, time elapsed: 1s INFO: Validating backup SCUN1Q INFO: Backup SCUN1Q data files are valid INFO: Backup SCUN1Q resident size: 56MB INFO: Backup SCUN1Q completed
List the backups of the instance:
backup_user@backup_host:~$ pg_probackup-16 show \ -B /mnt/backups \ --instance=node ================================================================================================================================ Instance Version ID Recovery Time Mode WAL Mode TLI Time Data WAL Zratio Start LSN Stop LSN Status ================================================================================================================================ node 16 SCUN1Q 2024-05-02 11:17:53+03 FULL STREAM 1/0 12s 40MB 16MB 2.42 0/8000028 0/800BBD0 OK
Make an incremental backup in the DELTA mode:
backup_user@backup_host:~$ pg_probackup-16 backup \ -B /mnt/backups \ -b DELTA \ --instance=node \ --stream \ --compress-algorithm=zlib \ --remote-host=postgres_host \ --remote-user=postgres \ -U backup \ -d backupdb INFO: Backup start, pg_probackup version: 2.5.15, instance: node, backup ID: SCUN22, backup mode: DELTA, wal mode: STREAM, remote: true, compress-algorithm: zlib, compress-level: 1 INFO: This PostgreSQL instance was initialized with data block checksums. Data block corruption will be detected INFO: Database backup start INFO: wait for pg_backup_start() INFO: Parent backup: SCUN1Q INFO: Wait for WAL segment /mnt/backups/backups/node/SCUN22/database/pg_wal/000000010000000000000009 to be streamed INFO: PGDATA size: 96MB INFO: Current Start LSN: 0/9000028, TLI: 1 INFO: Parent Start LSN: 0/8000028, TLI: 1 INFO: Start transferring data files INFO: Data files are transferred, time elapsed: 1s INFO: wait for pg_stop_backup() INFO: pg_stop backup() successfully executed INFO: stop_lsn: 0/9000168 INFO: Getting the Recovery Time from WAL INFO: Syncing backup files to disk INFO: Backup files are synced, time elapsed: 1s INFO: Validating backup SCUN22 INFO: Backup SCUN22 data files are valid INFO: Backup SCUN22 resident size: 34MB INFO: Backup SCUN22 completed
Add or modify some parameters in the pg_probackup configuration file, so that you do not have to specify them each time on the command line:
backup_user@backup_host:~$ pg_probackup-16 set-config \ -B /mnt/backups \ --instance=node \ --remote-host=postgres_host \ --remote-user=postgres \ -U backup \ -d backupdb
Check the configuration of the instance:
backup_user@backup_host:~$ pg_probackup-16 show-config \ -B /mnt/backups \ --instance=node # Backup instance information pgdata = /var/lib/pgpro/std-16/data system-identifier = 7364313570668255886 xlog-seg-size = 16777216 # Connection parameters pgdatabase = backupdb pghost = postgres_host pguser = backup # Replica parameters replica-timeout = 5min # Archive parameters archive-timeout = 5min # Logging parameters log-level-console = INFO log-level-file = OFF log-format-console = PLAIN log-format-file = PLAIN log-filename = pg_probackup.log log-rotation-size = 0TB log-rotation-age = 0d # Retention parameters retention-redundancy = 0 retention-window = 0 wal-depth = 0 # Compression parameters compress-algorithm = none compress-level = 1 # Remote access parameters remote-proto = ssh remote-host = postgres_host remote-user = postgres
Note that the parameters not modified via set-config retain their default values.
Make another incremental backup in the DELTA mode, omitting the parameters stored in the configuration file earlier:
backup_user@backup_host:~$ pg_probackup-16 backup \ -B /mnt/backups \ -b DELTA \ --instance=node \ --stream \ --compress-algorithm=zlib INFO: Backup start, pg_probackup version: 2.5.15, instance: node, backup ID: SCUN2C, backup mode: DELTA, wal mode: STREAM, remote: true, compress-algorithm: zlib, compress-level: 1 INFO: This PostgreSQL instance was initialized with data block checksums. Data block corruption will be detected INFO: Database backup start INFO: wait for pg_backup_start() INFO: Parent backup: SCUN22 INFO: Wait for WAL segment /mnt/backups/backups/node/SCUN2C/database/pg_wal/00000001000000000000000B to be streamed INFO: PGDATA size: 96MB INFO: Current Start LSN: 0/B000028, TLI: 1 INFO: Parent Start LSN: 0/9000028, TLI: 1 INFO: Start transferring data files INFO: Data files are transferred, time elapsed: 0 INFO: wait for pg_stop_backup() INFO: pg_stop backup() successfully executed INFO: stop_lsn: 0/B000168 INFO: Getting the Recovery Time from WAL INFO: Syncing backup files to disk INFO: Backup files are synced, time elapsed: 0 INFO: Validating backup SCUN2C INFO: Backup SCUN2C data files are valid INFO: Backup SCUN2C resident size: 17MB INFO: Backup SCUN2C completed
List the backups of the instance again:
backup_user@backup_host:~$ pg_probackup-16 show \ -B /mnt/backups \ --instance=node =================================================================================================================================== Instance Version ID Recovery Time Mode WAL Mode TLI Time Data WAL Zratio Start LSN Stop LSN Status =================================================================================================================================== node 16 SCUN2C 2024-05-02 11:18:13+03 DELTA STREAM 1/1 10s 1139kB 16MB 1.00 0/B000028 0/B000168 OK node 16 SCUN22 2024-05-02 11:18:04+03 DELTA STREAM 1/1 10s 2357kB 32MB 1.02 0/9000028 0/9000168 OK node 16 SCUN1Q 2024-05-02 11:17:53+03 FULL STREAM 1/0 12s 40MB 16MB 2.42 0/8000028 0/800BBD0 OK
Restore the data from the latest available backup to an arbitrary location:
backup_user@backup_host:~$ pg_probackup-16 restore \ -B /mnt/backups \ -D /var/lib/pgpro/std-16/staging-data \ --instance=node INFO: Validating parents for backup SCUN2C INFO: Validating backup SCUN1Q INFO: Backup SCUN1Q data files are valid INFO: Validating backup SCUN22 INFO: Backup SCUN22 data files are valid INFO: Validating backup SCUN2C INFO: Backup SCUN2C data files are valid INFO: Backup SCUN2C WAL segments are valid INFO: Backup SCUN2C is valid. INFO: Restoring the database from backup SCUN2C on localhost INFO: Start restoring backup files. PGDATA size: 112MB INFO: Backup files are restored. Transfered bytes: 112MB, time elapsed: 0 INFO: Restore incremental ratio (less is better): 100% (112MB/112MB) INFO: Syncing restored files to disk INFO: Restored backup files are synced, time elapsed: 2s INFO: Restore of backup SCUN2C completed.
You may need to use apt-get instead of apt on older systems in the commands below.
Add the pg_probackup repository GPG key
sudo apt install gpg wget wget -qO - https://repo.postgrespro.ru/pg_probackup/keys/GPG-KEY-PG-PROBACKUP | \ sudo tee /etc/apt/trusted.gpg.d/pg_probackup.asc
Setup the binary package repository
. /etc/os-release echo "deb [arch=amd64] https://repo.postgrespro.ru/pg_probackup/deb $VERSION_CODENAME main-$VERSION_CODENAME" | \ sudo tee /etc/apt/sources.list.d/pg_probackup.list
Optionally setup the source package repository for rebuilding the binaries
echo "deb-src [arch=amd64] https://repo.postgrespro.ru/pg_probackup/deb $VERSION_CODENAME main-$VERSION_CODENAME" | \ sudo tee -a /etc/apt/sources.list.d/pg_probackup.list
List the available pg_probackup packages
Using apt:
sudo apt update apt search pg_probackup
Using apt-get:
sudo apt-get update apt-cache search pg_probackup
Install or upgrade a pg_probackup version of your choice
sudo apt install pg-probackup-16
Optionally install the debug package
sudo apt install pg-probackup-16-dbg
Optionally install the source package (provided you have set up the source package repository as described above)
sudo apt install dpkg-dev sudo apt source pg-probackup-16
You may need to use yum instead of dnf on older systems in the commands below.
Install the pg_probackup repository
dnf install https://repo.postgrespro.ru/pg_probackup/keys/pg_probackup-repo-centos.noarch.rpm
List the available pg_probackup packages
dnf search pg_probackup
Install or upgrade a pg_probackup version of your choice
dnf install pg_probackup-16
Optionally install the debug package
dnf install pg_probackup-16-debuginfo
Optionally install the source package for rebuilding the binaries
Using dnf:
dnf install 'dnf-command(download)' dnf download --source pg_probackup-16
Using yum:
yumdownloader --source pg_probackup-16
Setup the repository
On ALT Linux 10:
. /etc/os-release echo "rpm http://repo.postgrespro.ru/pg_probackup/rpm/latest/altlinux-p$VERSION_ID x86_64 vanilla" | \ sudo tee /etc/apt/sources.list.d/pg_probackup.list
On ALT Linux 8 and 9:
. /etc/os-release echo "rpm http://repo.postgrespro.ru/pg_probackup/rpm/latest/altlinux-$VERSION_ID x86_64 vanilla" | \ sudo tee /etc/apt/sources.list.d/pg_probackup.list
List the available pg_probackup packages
sudo apt-get update apt-cache search pg_probackup
Install or upgrade a pg_probackup version of your choice
sudo apt-get install pg_probackup-16
Optionally install the debug package
sudo apt-get install pg_probackup-16-debuginfo
Add the pg_probackup repository GPG key
zypper in -y gpg wget wget -O GPG-KEY-PG_PROBACKUP https://repo.postgrespro.ru/pg_probackup/keys/GPG-KEY-PG_PROBACKUP rpm --import GPG-KEY-PG_PROBACKUP
Setup the repository
zypper in https://repo.postgrespro.ru/pg_probackup/keys/pg_probackup-repo-suse.noarch.rpm
List the available pg_probackup packages
zypper se pg_probackup
Install or upgrade a pg_probackup version of your choice
zypper in pg_probackup-16
Optionally install the source package for rebuilding the binaries
zypper si pg_probackup-16
Once you have pg_probackup installed, complete the following setup:
Initialize the backup catalog.
Add a new backup instance to the backup catalog.
Configure the database cluster to enable pg_probackup backups.
Optionally, configure SSH for running pg_probackup operations in the remote mode.
pg_probackup stores all WAL and backup files in the corresponding subdirectories of the backup catalog.
To initialize the backup catalog, run the following command:
pg_probackup init -B backup_dir
where backup_dir
is the path to the backup
catalog. If the backup_dir
already exists,
it must be empty. Otherwise, pg_probackup returns an error.
The user launching pg_probackup must have full access to
the backup_dir
directory.
pg_probackup creates the backup_dir
backup
catalog, with the following subdirectories:
wal/
— directory for WAL files.
backups/
— directory for backup files.
Once the backup catalog is initialized, you can add a new backup instance.
pg_probackup can store backups for multiple database clusters in a single backup catalog. To set up the required subdirectories, you must add a backup instance to the backup catalog for each database cluster you are going to back up.
To add a new backup instance, run the following command:
pg_probackup add-instance -Bbackup_dir
-Ddata_dir
--instance=instance_name
[remote_options
]
Where:
data_dir
is the data directory of the
cluster you are going to back up. To set up and use
pg_probackup, write access to this directory is required.
instance_name
is the name of the
subdirectories that will store WAL and backup files for this
cluster.
remote_options
are optional parameters that need to be specified only if
data_dir
is located
on a remote system.
pg_probackup creates the instance_name
subdirectories under the backups/
and wal/
directories of
the backup catalog. The
backups/
directory contains
the instance_name
pg_probackup.conf
configuration file that controls
pg_probackup settings for this backup instance. To add
remote_options to the configuration file, use the
set-config command.
For details on how to fine-tune pg_probackup configuration, see the section called “Configuring pg_probackup”.
The user launching pg_probackup must have full access to
backup_dir
directory and at least read-only
access to data_dir
directory. If you
specify the path to the backup catalog in the
BACKUP_PATH
environment variable, you can
omit the corresponding option when running pg_probackup
commands.
For PostgreSQL 11 or higher, it is recommended to use the allow-group-access feature, so that backup can be done by any OS user in the same group as the cluster owner. In this case, the user should have read permissions for the cluster directory.
Although pg_probackup can be used by a superuser, it is
recommended to create a separate role with the minimum
permissions required for the chosen backup strategy. In these
configuration instructions, the backup
role
is used as an example.
For security reasons, it is recommended to run the configuration SQL queries below in a separate database.
postgres=# CREATE DATABASE backupdb; postgres=# \c backupdb
To perform a backup, the following
permissions for role backup
are required
only in the database used for
connection to the PostgreSQL server.
For PostgreSQL versions 11 — 14:
BEGIN; CREATE ROLE backup WITH LOGIN; GRANT USAGE ON SCHEMA pg_catalog TO backup; GRANT EXECUTE ON FUNCTION pg_catalog.current_setting(text) TO backup; GRANT EXECUTE ON FUNCTION pg_catalog.set_config(text, text, boolean) TO backup; GRANT EXECUTE ON FUNCTION pg_catalog.pg_is_in_recovery() TO backup; GRANT EXECUTE ON FUNCTION pg_catalog.pg_start_backup(text, boolean, boolean) TO backup; GRANT EXECUTE ON FUNCTION pg_catalog.pg_stop_backup(boolean, boolean) TO backup; GRANT EXECUTE ON FUNCTION pg_catalog.pg_create_restore_point(text) TO backup; GRANT EXECUTE ON FUNCTION pg_catalog.pg_switch_wal() TO backup; GRANT EXECUTE ON FUNCTION pg_catalog.pg_last_wal_replay_lsn() TO backup; GRANT EXECUTE ON FUNCTION pg_catalog.txid_current() TO backup; GRANT EXECUTE ON FUNCTION pg_catalog.txid_current_snapshot() TO backup; GRANT EXECUTE ON FUNCTION pg_catalog.txid_snapshot_xmax(txid_snapshot) TO backup; GRANT EXECUTE ON FUNCTION pg_catalog.pg_control_checkpoint() TO backup; COMMIT;
For PostgreSQL 15 or higher:
BEGIN; CREATE ROLE backup WITH LOGIN; GRANT USAGE ON SCHEMA pg_catalog TO backup; GRANT EXECUTE ON FUNCTION pg_catalog.current_setting(text) TO backup; GRANT EXECUTE ON FUNCTION pg_catalog.set_config(text, text, boolean) TO backup; GRANT EXECUTE ON FUNCTION pg_catalog.pg_is_in_recovery() TO backup; GRANT EXECUTE ON FUNCTION pg_catalog.pg_backup_start(text, boolean) TO backup; GRANT EXECUTE ON FUNCTION pg_catalog.pg_backup_stop(boolean) TO backup; GRANT EXECUTE ON FUNCTION pg_catalog.pg_create_restore_point(text) TO backup; GRANT EXECUTE ON FUNCTION pg_catalog.pg_switch_wal() TO backup; GRANT EXECUTE ON FUNCTION pg_catalog.pg_last_wal_replay_lsn() TO backup; GRANT EXECUTE ON FUNCTION pg_catalog.txid_current() TO backup; GRANT EXECUTE ON FUNCTION pg_catalog.txid_current_snapshot() TO backup; GRANT EXECUTE ON FUNCTION pg_catalog.txid_snapshot_xmax(txid_snapshot) TO backup; GRANT EXECUTE ON FUNCTION pg_catalog.pg_control_checkpoint() TO backup; COMMIT;
In the
pg_hba.conf
file, allow connection to the database cluster on behalf of the
backup
role.
Since pg_probackup needs to read cluster files directly,
pg_probackup must be started by (or connected to,
if used in the remote mode) the OS user that has read access to all files and
directories inside the data directory (PGDATA
) you are going to
back up.
Depending on whether you plan to take standalone or archive backups, PostgreSQL cluster configuration will differ, as specified in the sections below. To back up the database cluster from a standby server, run pg_probackup in the remote mode, or create PTRACK backups, additional setup is required.
For details, see the sections Setting up STREAM Backups, Setting up continuous WAL archiving, Setting up Backup from Standby, Configuring the Remote Mode, Setting up Partial Restore, and Setting up PTRACK Backups.
To set up the cluster for STREAM backups, complete the following steps:
If the backup
role does not exist, create it with
the REPLICATION
privilege when
Configuring the
Database Cluster:
CREATE ROLE backup WITH LOGIN REPLICATION;
If the backup
role already exists, grant it with the REPLICATION
privilege:
ALTER ROLE backup WITH REPLICATION;
In the
pg_hba.conf
file, allow replication on behalf of the
backup
role.
Make sure the parameter max_wal_senders is set high enough to leave at least one session available for the backup process.
Set the parameter
wal_level
to be higher than minimal
.
If you are planning to take PAGE backups in the STREAM mode or perform PITR with STREAM backups, you still have to configure WAL archiving, as explained in the section Setting up continuous WAL archiving.
Once these steps are complete, you can start taking FULL, PAGE, DELTA, and PTRACK backups in the STREAM WAL mode.
If you are planning to rely on
.pgpass
for authentication when running backup in STREAM mode,
then .pgpass must contain credentials for replication
database,
used to establish connection via replication protocol. Example:
pghost:5432:replication:backup_user:my_strong_password
Making backups in PAGE backup mode, performing PITR and making backups with ARCHIVE WAL delivery mode require continuous WAL archiving to be enabled. To set up continuous archiving in the cluster, complete the following steps:
Make sure the
wal_level
parameter is higher than minimal
.
If you are configuring archiving on master,
archive_mode
must be set to on
or
always
. To perform archiving on standby,
set this parameter to always
.
Set the archive_command parameter, as follows:
archive_command = '"install_dir
/pg_probackup" archive-push -B "backup_dir
" --instance=instance_name
--wal-file-name=%f [remote_options
]'
where install_dir
is the
installation directory of the pg_probackup
version you are going to use, backup_dir
and
instance_name
refer to the already
initialized backup catalog instance for this database cluster,
and remote_options
only need to be specified to archive WAL on a remote host. For details about all
possible archive-push parameters, see the
section archive-push.
Once these steps are complete, you can start making backups in the ARCHIVE WAL mode, backups in the PAGE backup mode, as well as perform PITR.
You can view the current state of the WAL archive using the show command. For details, see the section called “Viewing WAL Archive Information”.
If you are planning to make PAGE backups and/or backups with
ARCHIVE WAL mode from a
standby server that generates a small amount of WAL traffic,
without long waiting for WAL segment to fill up, consider
setting the
archive_timeout
PostgreSQL parameter on
master. The value of this parameter should be slightly
lower than the --archive-timeout
setting (5 minutes by default),
so that there is enough time for the rotated
segment to be streamed to standby and sent to WAL archive before the
backup is aborted because of --archive-timeout
.
Instead of using the archive-push
command provided by pg_probackup, you can use
any other tool to set up continuous archiving as long as it delivers WAL segments into
directory. If compression is used, it should be
backup_dir
/wal/instance_name
gzip
, and .gz
suffix in filename is
mandatory.
Instead of configuring continuous archiving by setting the
archive_mode
and archive_command
parameters, you can opt for using the
pg_receivewal
utility. In this case, pg_receivewal -D
option should point to
directory
directory. pg_probackup supports WAL compression
that can be done by pg_receivewal.
“Zero Data Loss” archive strategy can be
achieved only by using pg_receivewal.
backup_dir
/wal/instance_name
pg_probackup can take backups from a standby server. This requires the following additional setup:
On the standby server, set the
hot_standby
parameter to on
.
On the master server, set the
full_page_writes
parameter to on
.
To perform standalone backups on standby, complete all steps in section Setting up STREAM Backups.
To perform archive backups on standby, complete all steps in section Setting up continuous WAL archiving.
Once these steps are complete, you can start taking FULL, PAGE, DELTA, or PTRACK backups with appropriate WAL delivery mode: ARCHIVE or STREAM, from the standby server.
Backup from the standby server has the following limitations:
If the standby is promoted to the master during backup, the backup fails.
All WAL records required for the backup must contain
sufficient full-page writes. This requires you to enable
full_page_writes
on the master, and not
to use tools like pg_compresslog as
archive_command
to remove full-page writes from WAL files.
Logical verification of a database cluster requires the following
additional setup. Role backup
is used as an
example:
Install the amcheck or amcheck_next extension in every database of the cluster:
CREATE EXTENSION amcheck;
Grant the following permissions to the backup
role in every database of the cluster:
GRANT SELECT ON TABLE pg_catalog.pg_am TO backup; GRANT SELECT ON TABLE pg_catalog.pg_class TO backup; GRANT SELECT ON TABLE pg_catalog.pg_database TO backup; GRANT SELECT ON TABLE pg_catalog.pg_namespace TO backup; GRANT SELECT ON TABLE pg_catalog.pg_extension TO backup; GRANT EXECUTE ON FUNCTION bt_index_check(regclass) TO backup; GRANT EXECUTE ON FUNCTION bt_index_check(regclass, bool) TO backup; GRANT EXECUTE ON FUNCTION bt_index_check(regclass, bool, bool) TO backup;
If you are planning to use partial restore, complete the following additional step:
Grant the read-only access to pg_catalog.pg_database
to the
backup
role only in the database
used for connection to
PostgreSQL server:
GRANT SELECT ON TABLE pg_catalog.pg_database TO backup;
pg_probackup supports the remote mode that allows you to perform backup, restore and WAL archiving operations remotely. In this mode, the backup catalog is stored on a local system, while PostgreSQL instance to backup and/or to restore is located on a remote system. Currently the only supported remote protocol is SSH.
If you are going to use pg_probackup in remote mode via SSH, complete the following steps:
Install pg_probackup on both systems:
backup_host
and
postgres_host
.
For communication between the hosts set up a passwordless
SSH connection between the backup_user
user on
backup_host
and the
postgres
user on
postgres_host
:
[backup_user@backup_host] ssh-copy-id postgres@postgres_host
Where:
backup_host
is the system with
backup catalog.
postgres_host
is the system with the PostgreSQL
cluster.
backup_user
is the OS user on
backup_host
used to run pg_probackup.
postgres
is the user on
postgres_host
under which
PostgreSQL cluster processes are running.
For PostgreSQL 11 or higher a
more secure approach can be used thanks to
allow-group-access feature.
If you are going to rely on
continuous
WAL archiving, set up a passwordless SSH
connection between the postgres
user on
postgres_host
and the backup
user on backup_host
:
[postgres@postgres_host] ssh-copy-id backup_user@backup_host
Make sure pg_probackup on postgres_host
can be located when a connection via SSH is made. For example, for Bash, you can
modify PATH
in ~/.bashrc
of the postgres
user
(above the line in bashrc
that exits the script for non-interactive shells).
Alternatively, for pg_probackup commands, specify the path to the directory
containing the pg_probackup binary on postgres_host
via
the --remote-path option.
pg_probackup in the remote mode via SSH works as follows:
Only the following commands can be launched in the remote mode: add-instance, backup, restore, catchup, archive-push, and archive-get.
Operating in remote mode requires pg_probackup binary to be installed on both local and remote systems. The versions of local and remote binary must be the same.
When started in the remote mode, the main pg_probackup process
on the local system connects to the remote system via SSH and
launches one or more agent processes on the remote system, which are called
remote agents. The number of remote agents
is equal to the -j
/--threads
setting.
The main pg_probackup process uses remote agents to access remote files and transfer data between local and remote systems.
Remote agents try to minimize the network traffic and the number of round-trips between hosts.
The main process is usually started on
backup_host
and connects to
postgres_host
, but in case of
archive-push and
archive-get commands the main process
is started on postgres_host
and connects to
backup_host
.
Once data transfer is complete, remote agents are terminated and SSH connections are closed.
If an error condition is encountered by a remote agent, then all agents are terminated and error details are reported by the main pg_probackup process, which exits with an error.
Compression is always done on
postgres_host
, while decompression is always done on
backup_host
.
You can impose additional restrictions on SSH settings to protect the system in the event of account compromise.
The PTRACK backup mode can be used only for Postgres Pro Standard and Postgres Pro Enterprise installations, or patched vanilla PostgreSQL. Links to PTRACK patches can be found here.
PTRACK versions lower than 2.0 are deprecated and not supported. Postgres Pro Standard and Postgres Pro Enterprise versions starting with 11.9.1 contain PTRACK 2.0. Upgrade your server to avoid issues in backups that you will take in future and be sure to take fresh backups of your clusters with the upgraded PTRACK since the backups taken with PTRACK 1.x might be corrupt.
If you are going to use PTRACK backups, complete the following
additional steps. The role that will perform PTRACK backups
(the backup
role in the examples below) must have
access to all the databases of the cluster.
For PostgreSQL 11 or higher:
Create PTRACK extension:
CREATE EXTENSION ptrack;
To enable tracking page updates, set ptrack.map_size
parameter to a positive integer and restart the server.
For optimal performance, it is recommended to set
ptrack.map_size
to
, where
N
/ 1024N
is the size of the
PostgreSQL cluster, in MB. If you set this
parameter to a lower value, PTRACK is more likely to map several blocks
together, which leads to false-positive results when tracking changed
blocks and increases the incremental backup size as unchanged blocks
can also be copied into the incremental backup.
Setting ptrack.map_size
to a higher value does not
affect PTRACK operation, but it is not recommended to set this parameter
to a value higher than 1024.
If you change the ptrack.map_size
parameter value,
the previously created PTRACK map file is cleared,
and tracking newly changed blocks starts from scratch. Thus, you have
to retake a full backup before taking incremental PTRACK backups after
changing ptrack.map_size
.
To create a backup, run the following command:
pg_probackup backup -Bbackup_dir
--instance=instance_name
-bbackup_mode
Where backup_mode
can take one of the
following values:
FULL
,
DELTA
,
PAGE
, and
PTRACK
.
When restoring a cluster from an incremental backup, pg_probackup relies on the parent full backup and all the incremental backups between them, which is called “the backup chain”. You must create at least one full backup before taking incremental ones.
ARCHIVE is the default WAL delivery mode.
For example, to make a FULL backup in ARCHIVE mode, run:
pg_probackup backup -Bbackup_dir
--instance=instance_name
-b FULL
ARCHIVE backups rely on continuous archiving to get WAL segments required to restore the cluster to a consistent state at the time the backup was taken.
When a backup is taken, pg_probackup
ensures that WAL files containing WAL records between Start
LSN
and Stop LSN
actually exist in
directory. pg_probackup also ensures that WAL records between
backup_dir
/wal/instance_name
Start LSN
and Stop LSN
can be parsed. This precaution
eliminates the risk of silent WAL corruption.
STREAM is the optional WAL delivery mode.
For example, to make a FULL backup in the STREAM mode, add the
--stream
flag to the command from the
previous example:
pg_probackup backup -Bbackup_dir
--instance=instance_name
-b FULL --stream --temp-slot
The optional --temp-slot
flag ensures that
the required segments remain available if the WAL is rotated
before the backup is complete.
Unlike backups in ARCHIVE mode, STREAM backups include all the WAL segments required to restore the cluster to a consistent state at the time the backup was taken.
During backup pg_probackup
streams WAL files containing WAL records between Start LSN
and
Stop LSN
to
directory. To eliminate the risk
of silent WAL corruption, pg_probackup also
checks that WAL records between backup_dir
/backups/instance_name
/backup_id
/database/pg_walStart LSN
and
Stop LSN
can be parsed.
Even if you are using continuous archiving, STREAM backups can still be useful in the following cases:
STREAM backups can be restored on the server that has no file access to WAL archive.
STREAM backups enable you to restore the cluster state at the point in time for which WAL files in archive are no longer available.
Backup in STREAM mode can be taken from a standby of a server that generates small amount of WAL traffic, without long waiting for WAL segment to fill up.
If data checksums are enabled in the database cluster, pg_probackup uses this information to check correctness of data files during backup. While reading each page, pg_probackup checks whether the calculated checksum coincides with the checksum stored in the page header. This guarantees that the PostgreSQL instance and the backup itself have no corrupt pages. Note that pg_probackup reads database files directly from the filesystem, so under heavy write load during backup it can show false-positive checksum mismatches because of partial writes. If a page checksum mismatch occurs, the page is re-read and checksum comparison is repeated.
A page is considered corrupt if checksum comparison has failed more than 100 times. In this case, the backup is aborted.
Even if data checksums are not enabled, pg_probackup always performs sanity checks for page headers.
To back up a directory located outside of the data directory,
use the optional --external-dirs
parameter
that specifies the path to this directory. If you would like
to add more than one external directory, you can provide several paths
separated by colons on Linux systems or semicolons on Windows systems.
For example, to include /etc/dir1
and
/etc/dir2
directories into the full
backup of your instance_name
instance
that will be stored under the backup_dir
directory on Linux, run:
pg_probackup backup -Bbackup_dir
--instance=instance_name
-b FULL --external-dirs=/etc/dir1:/etc/dir2
Similarly, to include C:\dir1
and
C:\dir2
directories into the full backup
on Windows, run:
pg_probackup backup -Bbackup_dir
--instance=instance_name
-b FULL --external-dirs=C:\dir1;C:\dir2
pg_probackup recursively copies the contents of each external directory into a separate subdirectory in the backup catalog. Since external directories included into different backups do not have to be the same, when you are restoring the cluster from an incremental backup, only those directories that belong to this particular backup will be restored. Any external directories stored in the previous backups will be ignored.
To include the same directories into each backup of your
instance, you can specify them in the pg_probackup.conf
configuration file using the
set-config command with the
--external-dirs
option.
To verify that PostgreSQL database cluster is not corrupt, run the following command:
pg_probackup checkdb [-Bbackup_dir
[--instance=instance_name
]] [-Ddata_dir
] [connection_options
]
This command performs physical verification of all data files located in the specified data directory by running page header sanity checks, as well as block-level checksum verification if checksums are enabled. If a corrupt page is detected, checkdb continues cluster verification until all pages in the cluster are validated.
By default, similar page validation
is performed automatically while a backup is taken by
pg_probackup. The checkdb
command enables you to perform such page validation
on demand, without taking any backup copies, even if the cluster
is not backed up using pg_probackup at all.
To perform cluster verification, pg_probackup
needs to connect to the cluster to be verified. In general, it is
enough to specify the backup instance of this cluster for
pg_probackup to determine the required
connection options. However, if -B
and
--instance
options are omitted, you have to provide
connection options and
data_dir
via environment
variables or command-line options.
Physical verification cannot detect logical inconsistencies, missing or nullified blocks and entire files, or similar anomalies. Extensions amcheck and amcheck_next provide a partial solution to these problems.
If you would like, in addition to physical verification, to
verify all indexes in all databases using these extensions, you
can specify the --amcheck
flag when running
the checkdb command:
pg_probackup checkdb -Ddata_dir
--amcheck [connection_options
]
You can skip physical verification by specifying the
--skip-block-validation
flag. In this case,
you can omit backup_dir and
data_dir options, only
connection options are
mandatory:
pg_probackup checkdb --amcheck --skip-block-validation [connection_options
]
Logical verification can be done more thoroughly with the
--heapallindexed
flag by checking that all heap
tuples that should be indexed are actually indexed, but at the
higher cost of CPU, memory, and I/O consumption.
pg_probackup calculates checksums for each file in a backup during the backup process. The process of checking checksums of backup data files is called the backup validation. By default, validation is run immediately after the backup is taken and right before the restore, to detect possible backup corruption.
If you would like to skip backup validation, you can specify the
--no-validate
flag when running
backup and
restore commands.
To ensure that all the required backup files are present and can be used to restore the database cluster, you can run the validate command with the exact recovery target options you are going to use for recovery.
For example, to check that you can restore the database cluster from a backup copy up to transaction ID 4242, run this command:
pg_probackup validate -Bbackup_dir
--instance=instance_name
--recovery-target-xid=4242
If validation completes successfully, pg_probackup displays the corresponding message. If validation fails, you will receive an error message with the exact time, transaction ID, and LSN up to which the recovery is possible.
If you specify backup_id via
-i/--backup-id
option, then only the backup copy
with specified backup ID will be validated. If
backup_id is specified with
recovery target
options, the validate command will check whether it is possible
to restore the specified backup to the specified
recovery target.
For example, to check that you can restore the database cluster
from a backup copy with the SCUN2C
backup ID up to the
specified timestamp, run this command:
pg_probackup validate -Bbackup_dir
--instance=instance_name
-i SCUN2C --recovery-target-time="2024-05-03 11:18:13+03"
If you specify the backup_id
of an incremental backup,
all its parents starting from FULL backup will be
validated.
If you omit all the parameters, all backups are validated.
To restore the database cluster from a backup, run the restore command with at least the following options:
pg_probackup restore -Bbackup_dir
--instance=instance_name
-ibackup_id
Where:
backup_dir
is the backup catalog that
stores all backup files and meta information.
instance_name
is the backup instance
for the cluster to be restored.
backup_id
specifies the backup to
restore the cluster from. If you omit this option,
pg_probackup uses the latest valid backup available for the
specified instance. If you specify an incremental backup to
restore, pg_probackup automatically restores the underlying
full backup and then sequentially applies all the necessary
increments.
Once the restore
command is complete, start
the database service.
If you restore ARCHIVE backups,
perform PITR,
or specify the --restore-as-replica
flag with the
restore
command to set up a standby server,
pg_probackup creates a recovery configuration
file once all data files are copied into the target directory. This file
includes the minimal settings required for recovery, except for the password in the
primary_conninfo
parameter; you have to add the password manually or use
the --primary-conninfo
option, if required.
For PostgreSQL 11,
recovery settings are written into the recovery.conf
file. Starting from PostgreSQL 12,
pg_probackup writes these settings into
the probackup_recovery.conf
file and then includes
it into postgresql.auto.conf
.
If you are restoring a STREAM backup, the restore is complete
at once, with the cluster returned to a self-consistent state at
the point when the backup was taken. For ARCHIVE backups,
PostgreSQL replays all available archived WAL
segments, so the cluster is restored to the latest state possible
within the current timeline. You can change this behavior by using the
recovery target
options with the restore
command,
as explained in the section called “Performing Point-in-Time (PITR) Recovery”.
If the cluster to restore contains tablespaces, pg_probackup
restores them to their original location by default. To restore
tablespaces to a different location, use the
--tablespace-mapping
/-T
option. Otherwise,
restoring the cluster on the same host will fail if tablespaces
are in use, because the backup would have to be written to the
same directories.
When using the --tablespace-mapping
/-T
option, you must provide absolute paths to the old and new
tablespace directories. If a path happens to contain an equals
sign (=
), escape it with a backslash. This option can be
specified multiple times for multiple tablespaces. For example:
pg_probackup restore -Bbackup_dir
--instance=instance_name
-Ddata_dir
-j 4 -ibackup_id
-T tablespace1_dir=tablespace1_newdir
-T tablespace2_dir=tablespace2_newdir
To restore the cluster on a remote host, follow the instructions in the section called “Using pg_probackup in the Remote Mode”.
By default, the restore
command validates the specified backup before restoring the
cluster. If you run regular backup validations and would like
to save time when restoring the cluster, you can specify the
--no-validate
flag to skip validation and
speed up the recovery.
The speed of restore from backup can be significantly improved by replacing only invalid and changed pages in already existing PostgreSQL data directory using incremental restore options with the restore command.
To restore the database cluster from a backup in incremental mode, run the restore command with the following options:
pg_probackup restore -Bbackup_dir
--instance=instance_name
-Ddata_dir
-Iincremental_mode
Where incremental_mode
can take one of the
following values:
CHECKSUM — read all data files in the data directory, validate header and checksum in every page and replace only invalid pages and those with checksum and LSN not matching with corresponding page in backup. This is the simplest, the most fool-proof incremental mode. Recommended to use by default.
LSN — read the pg_control
in the
data directory to obtain redo LSN and redo TLI, which allows you
to determine a point in history(shiftpoint), where data directory
state shifted from target backup chain history. If shiftpoint is not within
reach of backup chain history, then restore is aborted.
If shiftpoint is within reach of backup chain history, then read
all data files in the data directory, validate header and checksum in
every page and replace only invalid pages and those with LSN greater
than shiftpoint.
This mode offers a greater speed up compared to CHECKSUM, but rely
on two conditions to be met. First,
data checksums parameter must be enabled in data directory (to avoid corruption
due to hint bits). This condition will be checked at the start of
incremental restore and the operation will be aborted if checksums are disabled.
Second, the pg_control
file must be
synched with state of data directory. This condition cannot checked
at the start of restore, so it is a user responsibility to ensure
that pg_control
contain valid information.
Therefore it is not recommended to use LSN mode in any situation,
where pg_control cannot be trusted or has been tampered with:
after pg_resetxlog
execution,
after restore from backup without recovery been run, etc.
NONE — regular restore without any incremental optimizations.
Regardless of chosen incremental mode, pg_probackup will check, that postmaster
in given destination directory is not running and system-identifier
is
the same as in the backup.
Suppose you want to return an old master as replica after switchover using incremental restore in LSN mode:
====================================================================================================================================== Instance Version ID Recovery Time Mode WAL Mode TLI Time Data WAL Zratio Start LSN Stop LSN Status ====================================================================================================================================== node 16 SCUN3Y 2024-05-02 11:19:16+03 DELTA STREAM 16/15 7s 92MB 208MB 2.27 0/3C0043A8 0/46159C70 OK node 16 SCUN3M 2024-05-02 11:19:01+03 PTRACK STREAM 15/15 10s 30MB 16MB 2.23 0/32000028 0/32005ED0 OK node 16 SCUN39 2024-05-02 11:18:50+03 PAGE STREAM 15/15 12s 46MB 32MB 1.44 0/2A000028 0/2B0000B8 OK node 16 SCUN2V 2024-05-02 11:18:38+03 FULL STREAM 15/0 11s 154MB 16MB 2.32 0/23000028 0/23000168 OK backup_user@backup_host:~$ pg_probackup-16 restore -B /mnt/backups --instance=node -R -I lsn INFO: Destination directory and tablespace directories are empty, disable incremental restore INFO: Validating parents for backup SCUN3Y INFO: Validating backup SCUN2V INFO: Backup SCUN2V data files are valid INFO: Validating backup SCUN39 INFO: Backup SCUN39 data files are valid INFO: Validating backup SCUN3M INFO: Backup SCUN3M data files are valid INFO: Validating backup SCUN3Y INFO: Backup SCUN3Y data files are valid INFO: Backup SCUN3Y WAL segments are valid INFO: Backup SCUN3Y is valid. INFO: Restoring the database from backup SCUN3Y INFO: Start restoring backup files. PGDATA size: 759MB INFO: Backup files are restored. Transfered bytes: 759MB, time elapsed: 3s INFO: Restore incremental ratio (less is better): 100% (759MB/759MB) INFO: Syncing restored files to disk INFO: Restored backup files are synced, time elapsed: 1s INFO: Restore of backup SCUN3Y completed.
Incremental restore is possible only for backups with
program_version
equal or greater than 2.4.0.
If you have enabled partial restore before taking backups, you can restore only some of the databases using partial restore options with the restore commands.
To restore the specified databases only, run the restore command with the following options:
pg_probackup restore -Bbackup_dir
--instance=instance_name
--db-include=database_name
The --db-include
option can be specified
multiple times. For example, to restore only databases
db1
and db2
, run the
following command:
pg_probackup restore -Bbackup_dir
--instance=instance_name
--db-include=db1 --db-include=db2
To exclude one or more databases from restore, use
the --db-exclude
option:
pg_probackup restore -Bbackup_dir
--instance=instance_name
--db-exclude=database_name
The --db-exclude
option can be specified
multiple times. For example, to exclude the databases
db1
and db2
from
restore, run the following command:
pg_probackup restore -Bbackup_dir
--instance=instance_name
--db-exclude=db1 --db-exclude=db2
Partial restore relies on lax behavior of PostgreSQL recovery process toward truncated files. For recovery to work properly, files of excluded databases are restored as files of zero size. After the PostgreSQL cluster is successfully started, you must drop the excluded databases using DROP DATABASE command.
To decouple a single cluster containing multiple databases into separate clusters with minimal downtime,
you can do partial restore of the cluster as a standby using the --restore-as-replica
option
for specific databases.
The template0
and
template1
databases are always restored.
Due to recovery specifics of PostgreSQL versions earlier than 12,
it is advisable that you set the
hot_standby
parameter to off
when running partial
restore of a PostgreSQL cluster of version earlier than 12.
Otherwise the recovery may fail.
If you have enabled continuous WAL archiving before taking backups, you can restore the cluster to its state at an arbitrary point in time (recovery target) using recovery target options with the restore command.
You can use both STREAM and ARCHIVE backups for point in time
recovery as long as the WAL archive is available at least starting
from the time the backup was taken.
If -i
/--backup-id
option is omitted,
pg_probackup automatically chooses the backup that is the
closest to the specified recovery target and starts the restore
process, otherwise pg_probackup will try to restore
the specified backup to the specified recovery target.
To restore the cluster state at the exact time, specify the
--recovery-target-time
option, in the
timestamp format. For example:
pg_probackup restore -Bbackup_dir
--instance=instance_name
--recovery-target-time="2024-05-03 11:18:13+03"
To restore the cluster state up to a specific transaction
ID, use the --recovery-target-xid
option:
pg_probackup restore -Bbackup_dir
--instance=instance_name
--recovery-target-xid=687
To restore the cluster state up to the specific LSN, use
--recovery-target-lsn
option:
pg_probackup restore -Bbackup_dir
--instance=instance_name
--recovery-target-lsn=16/B374D848
To restore the cluster state up to the specific named restore
point, use --recovery-target-name
option:
pg_probackup restore -Bbackup_dir
--instance=instance_name
--recovery-target-name="before_app_upgrade"
To restore the backup to the latest state available in
the WAL archive, use --recovery-target
option
with latest
value:
pg_probackup restore -Bbackup_dir
--instance=instance_name
--recovery-target="latest"
To restore the cluster to the earliest point of consistency,
use --recovery-target
option with the
immediate
value:
pg_probackup restore -Bbackup_dir
--instance=instance_name
--recovery-target='immediate'
pg_probackup supports the remote mode that allows you to perform
backup
and restore
operations remotely via SSH. In this mode, the backup catalog is
stored on a local system, while PostgreSQL instance to be backed
up is located on a remote system. You must have pg_probackup
installed on both systems.
pg_probackup relies on passwordless SSH connection for communication between the hosts.
In addition to SSH connection, pg_probackup uses a regular connection to the database to manage the remote operation. See the section Configuring the Database Cluster for details of how to set up a database connection.
The typical workflow is as follows:
On your backup host, configure pg_probackup as explained in the section Setup. For the add-instance and set-config commands, make sure to specify remote options that point to the database host with the PostgreSQL instance.
If you would like to take remote backups in PAGE mode, or rely on ARCHIVE WAL delivery mode, or use PITR, configure continuous WAL archiving from the database host to the backup host as explained in the section Setting up continuous WAL archiving. For the archive-push and archive-get commands, you must specify the remote options that point to the backup host with the backup catalog.
Run backup or restore commands with remote options on the backup host. pg_probackup connects to the remote system via SSH and creates a backup locally or restores the previously taken backup on the remote system, respectively.
For example, to create an archive full backup of a
PostgreSQL cluster located on
a remote system with host address 192.168.0.2
on behalf of the postgres
user via SSH connection
through port 2302
, run:
pg_probackup backup -Bbackup_dir
--instance=instance_name
-b FULL --remote-user=postgres --remote-host=192.168.0.2 --remote-port=2302
To restore the latest available backup on a remote system with host address
192.168.0.2
on behalf of the postgres
user via SSH connection through port 2302
, run:
pg_probackup restore -Bbackup_dir
--instance=instance_name
--remote-user=postgres --remote-host=192.168.0.2 --remote-port=2302
Restoring an ARCHIVE backup or performing PITR in the remote mode
require additional information: destination address, port and
username for establishing an SSH connection
from the host with database
to the host with the backup
catalog. This information will be used by the
restore_command to copy WAL segments
from the archive to the PostgreSQL pg_wal
directory.
To solve this problem, you can use Remote WAL Archive Options.
For example, to restore latest backup on remote system using
remote mode through SSH connection to user
postgres
on host with address
192.168.0.2
via port 2302
and user backup
on backup catalog host with
address 192.168.0.3
via port
2303
, run:
pg_probackup restore -Bbackup_dir
--instance=instance_name
--remote-user=postgres --remote-host=192.168.0.2 --remote-port=2302 --archive-host=192.168.0.3 --archive-port=2303 --archive-user=backup
Provided arguments will be used to construct the restore_command:
restore_command = '"install_dir
/pg_probackup" archive-get -B "backup_dir
" --instance=instance_name
--wal-file-path=%p --wal-file-name=%f --remote-host=192.168.0.3 --remote-port=2303 --remote-user=backup'
Alternatively, you can use the --restore-command
option to provide the entire restore_command
:
pg_probackup restore -Bbackup_dir
--instance=instance_name
--remote-user=postgres --remote-host=192.168.0.2 --remote-port=2302 --restore-command='"install_dir
/pg_probackup" archive-get -B "backup_dir
" --instance=instance_name
--wal-file-path=%p --wal-file-name=%f --remote-host=192.168.0.3 --remote-port=2303 --remote-user=backup'
The remote mode is currently unavailable for Windows systems.
backup, restore, merge, delete, catchup, checkdb, and validate processes can be executed on several parallel threads. This can significantly speed up pg_probackup operation given enough resources (CPU cores, disk, and network bandwidth).
Parallel execution is controlled by the
-j/--threads
command-line option. For
example, to create a backup using four parallel threads, run:
pg_probackup backup -Bbackup_dir
--instance=instance_name
-b FULL -j 4
Parallel restore applies only to copying data from the backup catalog to the data directory of the cluster. When PostgreSQL server is started, WAL records need to be replayed, and this cannot be done in parallel.
Once the backup catalog is initialized and a new backup instance
is added, you can use the pg_probackup.conf
configuration file
located in the
directory to fine-tune pg_probackup configuration.
backup_dir
/backups/instance_name
For example, backup and
checkdb commands use a regular
PostgreSQL connection. To avoid specifying
connection options
each time on the command line, you can set them in the
pg_probackup.conf
configuration file using the
set-config command.
It is not recommended
to edit pg_probackup.conf
manually.
Initially, pg_probackup.conf
contains the following settings:
PGDATA
— the path to the data directory of the cluster to
back up.
system-identifier
— the unique identifier of the PostgreSQL
instance.
Additionally, you can define remote, retention, logging, and compression settings using the set-config command:
pg_probackup set-config -Bbackup_dir
--instance=instance_name
[--external-dirs=external_directory_path
] [remote_options
] [connection_options
] [retention_options
] [logging_options
]
To view the current settings, run the following command:
pg_probackup show-config -Bbackup_dir
--instance=instance_name
You can override the settings defined in pg_probackup.conf
when
running pg_probackup commands
via the corresponding environment variables and/or command line
options.
If you define connection settings in the pg_probackup.conf
configuration file, you can omit connection options in all the
subsequent pg_probackup commands. However, if the corresponding
environment variables are set, they get higher priority. The
options provided on the command line overwrite both environment
variables and configuration file settings.
If nothing is given, the default values are taken. By default
pg_probackup tries to use local connection via Unix domain
socket (localhost
on Windows) and tries to get the database name
and the user name from the PGUSER
environment variable or the
current OS user name.
With pg_probackup, you can manage backups from the command line:
To view the list of existing backups for every instance, run the command:
pg_probackup show -B backup_dir
pg_probackup displays the list of all the available backups. For example:
BACKUP INSTANCE 'node' ====================================================================================================================================== Instance Version ID Recovery Time Mode WAL Mode TLI Time Data WAL Zratio Start LSN Stop LSN Status ====================================================================================================================================== node 16 SCUN4E 2024-05-02 11:19:37+03 FULL ARCHIVE 1/0 13s 239MB 16MB 2.31 0/4C000028 0/4D0000B8 OK node 16 SCUN3Y 2024-05-02 11:19:16+03 DELTA STREAM 1/1 7s 92MB 208MB 2.27 0/3C0043A8 0/46159C70 OK node 16 SCUN3M 2024-05-02 11:19:01+03 PTRACK STREAM 1/1 10s 30MB 16MB 2.23 0/32000028 0/32005ED0 OK node 16 SCUN39 2024-05-02 11:18:50+03 PAGE STREAM 1/1 12s 46MB 32MB 1.44 0/2A000028 0/2B0000B8 OK node 16 SCUN2V 2024-05-02 11:18:38+03 FULL STREAM 1/0 11s 154MB 16MB 2.32 0/23000028 0/23000168 OK
For each backup, the following information is provided:
Instance
— the instance name.
Version
— PostgreSQL major version.
ID
— the backup identifier.
Recovery time
— the earliest moment for which you can
restore the state of the database cluster.
Mode
— the method used to take this backup. Possible
values: FULL
, PAGE
, DELTA
, PTRACK
.
WAL Mode
— WAL delivery mode. Possible values: STREAM
and ARCHIVE
.
TLI
— timeline identifiers of the current backup and its
parent.
Time
— the time it took to perform the backup.
Data
— the size of the data files in this backup. This
value does not include the size of WAL files. For
STREAM backups, the total size of the backup can be calculated
as Data
+ WAL
.
WAL
— the uncompressed size of WAL files
that need to be applied during recovery for the backup to reach a consistent state.
Zratio
— compression ratio calculated as
“uncompressed-bytes” / “data-bytes”.
Start LSN
— WAL log sequence number corresponding to the
start of the backup process. REDO point for PostgreSQL
recovery process to start from.
Stop LSN
— WAL log sequence number corresponding to the
end of the backup process. Consistency point for
PostgreSQL recovery process.
Status
— backup status. Possible values:
OK
— the backup is complete and valid.
DONE
— the backup is complete, but was not validated.
RUNNING
— the backup is in progress.
MERGING
— the backup is being merged.
MERGED
— the backup data files were
successfully merged, but its metadata is in the process
of being updated. Only full backups can have this status.
DELETING
— the backup files are being deleted.
CORRUPT
— some of the backup files are corrupt.
ERROR
— the backup was aborted because of an
unexpected error.
ORPHAN
— the backup is invalid because one of its
parent backups is corrupt or missing.
You can restore the cluster from the backup only if the backup
status is OK
or DONE
.
To get more detailed information about the backup, run the show command with the backup ID:
pg_probackup show -Bbackup_dir
--instance=instance_name
-ibackup_id
The sample output is as follows:
#Configuration backup-mode = FULL stream = false compress-alg = zlib compress-level = 1 from-replica = false #Compatibility block-size = 8192 xlog-block-size = 8192 checksum-version = 1 program-version = 2.5.15 server-version = 16 #Result backup info timelineid = 1 start-lsn = 0/4C000028 stop-lsn = 0/4D0000B8 start-time = '2024-05-02 11:19:26+03' end-time = '2024-05-02 11:19:39+03' recovery-xid = 743 recovery-time = '2024-05-02 11:19:37+03' data-bytes = 250827955 wal-bytes = 16777216 uncompressed-bytes = 578216425 pgdata-bytes = 578216107 status = OK primary_conninfo = 'user=backup channel_binding=prefer host=localhost port=5432 sslmode=prefer sslcompression=0 sslcertmode=allow sslsni=1 ssl_min_protocol_version=TLSv1.2 gssencmode=prefer krbsrvname=postgres gssdelegation=0 target_session_attrs=any load_balance_hosts=disable' content-crc = 802820606
Detailed output has additional attributes:
compress-alg
— compression algorithm used during backup. Possible values:
zlib
, pglz
, none
.
compress-level
— compression level used during backup.
from-replica
— was this backup taken on standby? Possible values:
1
, 0
.
block-size
— the block_size
setting of PostgreSQL cluster at the backup start.
checksum-version
— are
data
block checksums enabled in the backed up PostgreSQL cluster? Possible values: 1
, 0
.
program-version
— full version of pg_probackup binary used to create the backup.
start-time
— the backup start time.
end-time
— the backup end time.
expire-time
— the point in time
when a pinned backup can be removed in accordance with retention
policy. This attribute is only available for pinned backups.
uncompressed-bytes
— the size of data files before adding page headers and applying
compression. You can evaluate the effectiveness of compression
by comparing uncompressed-bytes
to data-bytes
if
compression if used.
pgdata-bytes
— the size of PostgreSQL
cluster data files at the time of backup. You can evaluate the
effectiveness of an incremental backup by comparing
pgdata-bytes
to uncompressed-bytes
.
recovery-xid
— transaction ID at the backup end time.
parent-backup-id
— ID of the parent backup. Available only
for incremental backups.
primary_conninfo
— libpq connection parameters
used to connect to the PostgreSQL cluster to take this backup. The
password is not included.
note
— text note attached to backup.
content-crc
— CRC32 checksum of backup_content.control
file.
It is used to detect corruption of backup metainformation.
You can also get the detailed information about the backup in the JSON format:
pg_probackup show -Bbackup_dir
--instance=instance_name
--format=json -i backup_id
The sample output is as follows:
[ { "instance": "node", "backups": [ { "id": "SCUN4E", "backup-mode": "FULL", "wal": "ARCHIVE", "compress-alg": "zlib", "compress-level": 1, "from-replica": "false", "block-size": 8192, "xlog-block-size": 8192, "checksum-version": 1, "program-version": "2.5.15", "server-version": "16", "current-tli": 16, "parent-tli": 2, "start-lsn": "0/4C000028", "stop-lsn": "0/4D0000B8", "start-time": "2024-05-02 11:19:26+03", "end-time": "2024-05-02 11:19:39+03", "recovery-xid": 743, "recovery-time": "2024-05-02 11:19:37+03", "data-bytes": 250827955, "wal-bytes": 16777216, "uncompressed-bytes": 578216425, "pgdata-bytes": 578216107, "primary_conninfo": "user=backup channel_binding=prefer host=localhost port=5432 sslmode=prefer sslcompression=0 sslcertmode=allow sslsni=1 ssl_min_protocol_version=TLSv1.2 gssencmode=prefer krbsrvname=postgres gssdelegation=0 target_session_attrs=any load_balance_hosts=disable", "status": "OK", "content-crc": 802820606 } ] } ]
To view the information about WAL archive for every instance, run the command:
pg_probackup show -Bbackup_dir
[--instance=instance_name
] --archive
pg_probackup displays the list of all the available WAL files grouped by timelines. For example:
ARCHIVE INSTANCE 'node' ================================================================================================================================ TLI Parent TLI Switchpoint Min Segno Max Segno N segments Size Zratio N backups Status ================================================================================================================================ 1 0 0/0 000000010000000000000019 00000001000000000000004D 53 848MB 1.00 5 OK
For each timeline, the following information is provided:
TLI
— timeline identifier.
Parent TLI
— identifier of the timeline from which this timeline branched off.
Switchpoint
— LSN of the moment when the timeline branched
off from its parent timeline.
Min Segno
— the first WAL segment
belonging to the timeline.
Max Segno
— the last WAL segment
belonging to the timeline.
N segments
— number of WAL segments belonging to the
timeline.
Size
— the size that files take on disk.
Zratio
— compression ratio calculated as N segments
*
wal_segment_size
* wal_block_size
/ Size
.
N backups
— number of backups belonging to the timeline.
To get the details about backups, use the JSON format.
Status
— status of the WAL archive for this timeline. Possible
values:
OK
— all WAL segments between Min Segno
and Max Segno
are present.
DEGRADED
— some WAL segments between Min Segno
and Max Segno
are missing. To find out which files are lost,
view this report in the JSON format.
To get more detailed information about the WAL archive in the JSON format, run the command:
pg_probackup show -Bbackup_dir
[--instance=instance_name
] --archive --format=json
The sample output is as follows:
[ { "instance": "node", "timelines": [ { "tli": 1, "parent-tli": 0, "switchpoint": "0/0", "min-segno": "000000010000000000000019", "max-segno": "00000001000000000000004D", "n-segments": 53, "size": 889192448, "zratio": 1.00, "closest-backup-id": "", "status": "OK", "lost-segments": [], "backups": [ { "id": "SCUN4E", "backup-mode": "FULL", "wal": "ARCHIVE", "compress-alg": "zlib", "compress-level": 1, "from-replica": "false", "block-size": 8192, "xlog-block-size": 8192, "checksum-version": 1, "program-version": "2.5.15", "server-version": "16", "current-tli": 1, "parent-tli": 0, "start-lsn": "0/4C000028", "stop-lsn": "0/4D0000B8", "start-time": "2024-05-02 11:19:26+03", "end-time": "2024-05-02 11:19:39+03", "recovery-xid": 743, "recovery-time": "2024-05-02 11:19:37+03", "data-bytes": 250827955, "wal-bytes": 16777216, "uncompressed-bytes": 578216425, "pgdata-bytes": 578216107, "primary_conninfo": "user=backup channel_binding=prefer host=localhost port=5432 sslmode=prefer sslcompression=0 sslcertmode=allow sslsni=1 ssl_min_protocol_version=TLSv1.2 gssencmode=prefer krbsrvname=postgres gssdelegation=0 target_session_attrs=any load_balance_hosts=disable", "status": "OK", "content-crc": 802820606 }, { "id": "SCUN3Y", "parent-backup-id": "SCUN3M", "backup-mode": "DELTA", "wal": "STREAM", "compress-alg": "zlib", "compress-level": 1, "from-replica": "false", "block-size": 8192, "xlog-block-size": 8192, "checksum-version": 1, "program-version": "2.5.15", "server-version": "16", "current-tli": 1, "parent-tli": 1, "start-lsn": "0/3C0043A8", "stop-lsn": "0/46159C70", "start-time": "2024-05-02 11:19:10+03", "end-time": "2024-05-02 11:19:17+03", "recovery-xid": 743, "recovery-time": "2024-05-02 11:19:16+03", "data-bytes": 96029293, "wal-bytes": 218103808, "uncompressed-bytes": 217639806, "pgdata-bytes": 578216107, "primary_conninfo": "user=backup channel_binding=prefer host=localhost port=5432 sslmode=prefer sslcompression=0 sslcertmode=allow sslsni=1 ssl_min_protocol_version=TLSv1.2 gssencmode=prefer krbsrvname=postgres gssdelegation=0 target_session_attrs=any load_balance_hosts=disable", "status": "OK", "content-crc": 3074300814 }, { "id": "SCUN3M", "parent-backup-id": "SCUN39", "backup-mode": "PTRACK", "wal": "STREAM", "compress-alg": "zlib", "compress-level": 1, "from-replica": "false", "block-size": 8192, "xlog-block-size": 8192, "checksum-version": 1, "program-version": "2.5.15", "server-version": "16", "current-tli": 1, "parent-tli": 1, "start-lsn": "0/32000028", "stop-lsn": "0/32005ED0", "start-time": "2024-05-02 11:18:58+03", "end-time": "2024-05-02 11:19:08+03", "recovery-xid": 742, "recovery-time": "2024-05-02 11:19:01+03", "data-bytes": 31205704, "wal-bytes": 16777216, "uncompressed-bytes": 69585790, "pgdata-bytes": 509927595, "primary_conninfo": "user=backup channel_binding=prefer host=localhost port=5432 sslmode=prefer sslcompression=0 sslcertmode=allow sslsni=1 ssl_min_protocol_version=TLSv1.2 gssencmode=prefer krbsrvname=postgres gssdelegation=0 target_session_attrs=any load_balance_hosts=disable", "status": "OK", "content-crc": 3446949708 }, { "id": "SCUN39", "parent-backup-id": "SCUN2V", "backup-mode": "PAGE", "wal": "STREAM", "compress-alg": "pglz", "compress-level": 1, "from-replica": "false", "block-size": 8192, "xlog-block-size": 8192, "checksum-version": 1, "program-version": "2.5.15", "server-version": "16", "current-tli": 1, "parent-tli": 1, "start-lsn": "0/2A000028", "stop-lsn": "0/2B0000B8", "start-time": "2024-05-02 11:18:45+03", "end-time": "2024-05-02 11:18:57+03", "recovery-xid": 741, "recovery-time": "2024-05-02 11:18:50+03", "data-bytes": 48381612, "wal-bytes": 33554432, "uncompressed-bytes": 69569406, "pgdata-bytes": 441639083, "primary_conninfo": "user=backup channel_binding=prefer host=localhost port=5432 sslmode=prefer sslcompression=0 sslcertmode=allow sslsni=1 ssl_min_protocol_version=TLSv1.2 gssencmode=prefer krbsrvname=postgres gssdelegation=0 target_session_attrs=any load_balance_hosts=disable", "status": "OK", "content-crc": 3492989773 }, { "id": "SCUN2V", "backup-mode": "FULL", "wal": "STREAM", "compress-alg": "zlib", "compress-level": 1, "from-replica": "false", "block-size": 8192, "xlog-block-size": 8192, "checksum-version": 1, "program-version": "2.5.15", "server-version": "16", "current-tli": 1, "parent-tli": 0, "start-lsn": "0/23000028", "stop-lsn": "0/23000168", "start-time": "2024-05-02 11:18:31+03", "end-time": "2024-05-02 11:18:42+03", "recovery-xid": 740, "recovery-time": "2024-05-02 11:18:38+03", "data-bytes": 161084290, "wal-bytes": 16777216, "uncompressed-bytes": 373359081, "pgdata-bytes": 373358763, "primary_conninfo": "user=backup channel_binding=prefer host=localhost port=5432 sslmode=prefer sslcompression=0 sslcertmode=allow sslsni=1 ssl_min_protocol_version=TLSv1.2 gssencmode=prefer krbsrvname=postgres gssdelegation=0 target_session_attrs=any load_balance_hosts=disable", "status": "OK", "content-crc": 1621343133 } ] } ] } ]
Most fields are consistent with the plain format, with some exceptions:
The size is in bytes.
The closest-backup-id
attribute
contains the ID of the most recent valid backup that belongs to
one of the previous timelines. You can use this backup to perform
point-in-time recovery to this timeline. If
such a backup does not exist, this string is empty.
The lost-segments
array provides with
information about intervals of missing segments in DEGRADED
timelines. In OK
timelines, the lost-segments
array is empty.
The backups
array lists all backups
belonging to the timeline. If the timeline has no backups, this array is empty.
With pg_probackup, you can configure retention policy to remove redundant backups, clean up unneeded WAL files, as well as pin specific backups to ensure they are kept for the specified time, as explained in the sections below. All these actions can be combined together in any way.
By default, all backup copies created with pg_probackup are stored in the specified backup catalog. To save disk space, you can configure retention policy to remove redundant backup copies.
To configure retention policy, set one or more of the
following variables in the pg_probackup.conf
file via
set-config:
--retention-redundancy=redundancy
Specifies the number of full backup copies to keep in the backup catalog.
--retention-window=window
Defines the earliest point in time for which pg_probackup can
complete the recovery. This option is set in
the number of days from the
current moment. For example, if
retention-window=7
, pg_probackup must
keep at least one backup copy that is older than seven days, with
all the corresponding WAL files, and all the backups that follow.
If both --retention-redundancy
and
--retention-window
options are set, both these
conditions have to be taken into account when purging the backup
catalog. For example, if you set --retention-redundancy=2
and --retention-window=7
,
pg_probackup has to keep two full backup
copies, as well as all the backups required to ensure recoverability
for the last seven days:
pg_probackup set-config -Bbackup_dir
--instance=instance_name
--retention-redundancy=2 --retention-window=7
To clean up the backup catalog in accordance with retention policy, you have to run the delete command with retention flags, as shown below, or use the backup command with these flags to process the outdated backup copies right when the new backup is created.
For example, to remove all backup copies that no longer satisfy the
defined retention policy, run the following command with the
--delete-expired
flag:
pg_probackup delete -Bbackup_dir
--instance=instance_name
--delete-expired
If you would like to also remove the WAL files that are no
longer required for any of the backups, you should also specify the
--delete-wal
flag:
pg_probackup delete -Bbackup_dir
--instance=instance_name
--delete-expired --delete-wal
You can also set or override the current retention policy by
specifying --retention-redundancy
and
--retention-window
options directly when
running delete or backup
commands:
pg_probackup delete -Bbackup_dir
--instance=instance_name
--delete-expired --retention-window=7 --retention-redundancy=2
Since incremental backups require that their parent full
backup and all the preceding incremental backups are
available, if any of such backups expire, they still cannot be
removed while at least one incremental backup in this chain
satisfies the retention policy. To avoid keeping expired
backups that are still required to restore an active
incremental one, you can merge them with this backup using the
--merge-expired
flag when running
backup or
delete commands.
Suppose you have backed up the node
instance in the backup_dir
directory,
with the --retention-window
option set
to 7
, and you have the following backups
available on May 02, 2024:
BACKUP INSTANCE 'node' ===================================================================================================================================== Instance Version ID Recovery Time Mode WAL Mode TLI Time Data WAL Zratio Start LSN Stop LSN Status ===================================================================================================================================== node 16 SCUN6L 2024-05-02 11:20:48+03 FULL ARCHIVE 1/0 5s 296MB 16MB 2.30 0/46000028 0/470000B8 OK node 16 SCQXUI 2024-04-30 11:20:45+03 PAGE ARCHIVE 1/1 5s 6280kB 16MB 1.00 0/44000028 0/450000F0 OK node 16 SCFTUG 2024-04-24 11:20:43+03 DELTA ARCHIVE 1/1 5s 6280kB 16MB 1.00 0/42000028 0/430000B8 OK ----------------------------------------------------------retention window----------------------------------------------------------- node 16 SCDZ6D 2024-04-23 11:20:40+03 PAGE ARCHIVE 1/1 5s 6280kB 16MB 1.00 0/40000028 0/410000B8 OK node 16 SCC4HX 2024-04-22 11:20:24+03 FULL ARCHIVE 1/0 5s 296MB 16MB 2.30 0/3E000028 0/3F0000F0 OK node 16 SC8F5G 2024-04-20 11:20:07+03 FULL ARCHIVE 1/0 5s 296MB 16MB 2.30 0/3C0000D8 0/3D00BB58 OK
Even though SCC4HX
and SCDZ6D
backups are outside the
retention window, they cannot be removed as it invalidates the
succeeding incremental backups SCFTUG
and SCQXUI
that are
still required, so, if you run the
delete command with the
--delete-expired
flag, only the SC8F5G
full
backup will be removed.
With the --merge-expired
option, the SCFTUG
backup is merged with the underlying SCDZ6D
and SCC4HX
backups
and becomes a full one, so there is no need to keep these
expired backups anymore:
pg_probackup delete -Bbackup_dir
--instance=node
--delete-expired --merge-expired pg_probackup show -Bbackup_dir
BACKUP INSTANCE 'node' ===================================================================================================================================== Instance Version ID Recovery Time Mode WAL Mode TLI Time Data WAL Zratio Start LSN Stop LSN Status ===================================================================================================================================== node 16 SCUN6L 2024-05-02 11:20:48+03 FULL ARCHIVE 1/0 5s 296MB 16MB 2.30 0/46000028 0/470000B8 OK node 16 SCQXUI 2024-04-30 11:20:45+03 PAGE ARCHIVE 1/1 5s 6280kB 16MB 1.00 0/44000028 0/450000F0 OK node 16 SCFTUG 2024-04-24 11:20:43+03 FULL ARCHIVE 1/1 5s 296MB 16MB 1.00 0/42000028 0/430000B8 OK
The Time
field for the merged backup displays the time
required for the merge.
If you need to keep certain backups longer than the established retention policy allows, you can pin them for arbitrary time. For example:
pg_probackup set-backup -Bbackup_dir
--instance=instance_name
-ibackup_id
--ttl=30d
This command sets the expiration time of the
specified backup to 30 days starting from the time
indicated in its recovery-time
attribute.
You can also explicitly set the expiration time for a backup
using the --expire-time
option. For example:
pg_probackup set-backup -Bbackup_dir
--instance=instance_name
-ibackup_id
--expire-time="2027-05-02 11:21:00+00"
Alternatively, you can use the --ttl
and
--expire-time
options with the
backup command to pin the newly
created backup:
pg_probackup backup -Bbackup_dir
--instance=instance_name
-b FULL --ttl=30d pg_probackup backup -Bbackup_dir
--instance=instance_name
-b FULL --expire-time="2027-05-02 11:21:00+00"
To check if the backup is pinned, run the show command:
pg_probackup show -Bbackup_dir
--instance=instance_name
-ibackup_id
If the backup is pinned, it has the expire-time
attribute that displays its expiration time:
... recovery-time = '2024-05-02 11:21:00+00' expire-time = '2027-05-02 11:21:00+00' data-bytes = 22288792 ...
You can unpin the backup by setting the --ttl
option to zero:
pg_probackup set-backup -Bbackup_dir
--instance=instance_name
-ibackup_id
--ttl=0
A pinned incremental backup implicitly pins all its parent backups. If you unpin such a backup later, its implicitly pinned parents will also be automatically unpinned.
When continuous
WAL archiving is enabled, archived WAL segments can take a lot
of disk space. Even if you delete old backup copies from time to time,
the --delete-wal
flag can
purge only those WAL segments that do not apply to any of the
remaining backups in the backup catalog. However, if point-in-time
recovery is critical only for the most recent backups, you can
configure WAL archive retention policy to keep WAL archive of
limited depth and win back some more disk space.
To configure WAL archive retention policy, you have to run the
set-config command with the
--wal-depth
option that specifies the number
of backups that can be used for PITR.
This setting applies to all the timelines, so you should be able to perform
PITR for the same number of backups on each timeline, if available.
Pinned backups are
not included into this count: if one of the latest backups
is pinned, pg_probackup ensures that
PITR is possible for one extra backup.
To remove WAL segments that do not satisfy the defined WAL archive
retention policy, you simply have to run the delete
or backup command with the --delete-wal
flag. For archive backups, WAL segments between Start LSN
and Stop LSN
are always kept intact, so such backups
remain valid regardless of the --wal-depth
setting
and can still be restored, if required.
You can also use the --wal-depth
option
with the delete and backup
commands to override the previously defined WAL archive retention
policy and purge old WAL segments on the fly.
Suppose you have backed up the node
instance in the backup_dir
directory and
configured
continuous WAL
archiving:
pg_probackup show -Bbackup_dir
--instance=node
====================================================================================================================================== Instance Version ID Recovery Time Mode WAL Mode TLI Time Data WAL Zratio Start LSN Stop LSN Status ====================================================================================================================================== node 16 SCUN92 2024-05-02 11:22:16+03 DELTA STREAM 1/1 9s 1162kB 32MB 1.08 0/7C000028 0/7C000168 OK node 16 SCUN8N 2024-05-02 11:22:09+03 FULL STREAM 1/0 12s 296MB 16MB 2.30 0/7A000028 0/7A009A08 OK node 16 SCUN8I 2024-05-02 11:21:55+03 DELTA STREAM 1/1 5s 1148kB 32MB 1.01 0/78000028 0/78000168 OK node 16 SCUN86 2024-05-02 11:21:47+03 DELTA STREAM 1/1 11s 120MB 16MB 2.27 0/76000028 0/760001A0 OK node 16 SCUN7I 2024-05-02 11:21:29+03 FULL STREAM 1/0 22s 296MB 288MB 2.30 0/63012FE8 0/74E7ADA0 OK node 16 SCUN71 2024-05-02 11:21:12+03 FULL STREAM 1/0 13s 296MB 272MB 2.30 0/49000028 0/573683B8 OK
You can check the state of the WAL archive by running the
show command with the
--archive
flag:
pg_probackup show -B backup_dir
--instance=node --archive
ARCHIVE INSTANCE 'node' ================================================================================================================================ TLI Parent TLI Switchpoint Min Segno Max Segno N segments Size Zratio N backups Status ================================================================================================================================ 1 0 0/0 000000010000000000000048 00000001000000000000007C 53 848MB 1.00 6 OK
WAL purge without --wal-depth
cannot
achieve much, only one segment is removed:
pg_probackup delete -B backup_dir
--instance=node --delete-wal
ARCHIVE INSTANCE 'node' ================================================================================================================================ TLI Parent TLI Switchpoint Min Segno Max Segno N segments Size Zratio N backups Status ================================================================================================================================ 1 0 0/0 000000010000000000000049 00000001000000000000007C 52 832MB 1.00 6 OK
If you would like, for example, to keep only those WAL
segments that can be applied to the latest valid backup, set the
--wal-depth
option to 1:
pg_probackup delete -B backup_dir
--instance=node --delete-wal --wal-depth=1
ARCHIVE INSTANCE 'node' =============================================================================================================================== TLI Parent TLI Switchpoint Min Segno Max Segno N segments Size Zratio N backups Status =============================================================================================================================== 1 0 0/0 00000001000000000000007C 00000001000000000000007C 1 16MB 1.00 6 OK
Alternatively, you can use the --wal-depth
option with the backup command:
pg_probackup backup -B backup_dir
--instance=node -b DELTA --wal-depth=1 --delete-wal
ARCHIVE INSTANCE 'node' =============================================================================================================================== TLI Parent TLI Switchpoint Min Segno Max Segno N segments Size Zratio N backups Status =============================================================================================================================== 1 0 0/0 00000001000000000000007E 00000001000000000000007E 1 16MB 1.00 7 OK
As you take more and more incremental backups, the total size of the backup catalog can substantially grow. To save disk space, you can merge incremental backups to their parent full backup by running the merge command, specifying the backup ID of the most recent incremental backup you would like to merge:
pg_probackup merge -Bbackup_dir
--instance=instance_name
-ibackup_id
This command merges backups that belong to a common incremental backup chain. If you specify a full backup, it will be merged with its first incremental backup. If you specify an incremental backup, it will be merged to its parent full backup, together with all incremental backups between them. Once the merge is complete, the full backup takes in all the merged data, and the incremental backups are removed as redundant. Thus, the merge operation is virtually equivalent to retaking a full backup and removing all the outdated backups, but it allows you to save much time, especially for large data volumes, as well as I/O and network traffic if you are using pg_probackup in the remote mode.
Before the merge, pg_probackup validates all the affected backups to ensure that they are valid. You can check the current backup status by running the show command with the backup ID:
pg_probackup show -Bbackup_dir
--instance=instance_name
-ibackup_id
If the merge is still in progress, the backup status is
displayed as MERGING
. For full backups,
it can also be shown as MERGED
while the
metadata is being updated at the final stage of the merge.
The merge is idempotent, so you can
restart the merge if it was interrupted.
To delete a backup that is no longer required, run the following command:
pg_probackup delete -Bbackup_dir
--instance=instance_name
-ibackup_id
This command will delete the backup with the specified
backup_id
, together with all the
incremental backups that descend from
backup_id
, if any. This way you can delete
some recent incremental backups, retaining the underlying full
backup and some of the incremental backups that follow it.
To delete obsolete WAL files that are not necessary to restore
any of the remaining backups, use the
--delete-wal
flag:
pg_probackup delete -Bbackup_dir
--instance=instance_name
--delete-wal
To delete backups that are expired according to the current
retention policy, use the --delete-expired
flag:
pg_probackup delete -Bbackup_dir
--instance=instance_name
--delete-expired
Expired backups cannot be removed while at least one
incremental backup that satisfies the retention policy is based
on them. If you would like to minimize the number of backups
still required to keep incremental backups valid, specify the
--merge-expired
flag when running this
command:
pg_probackup delete -Bbackup_dir
--instance=instance_name
--delete-expired --merge-expired
In this case, pg_probackup searches for the oldest incremental backup that satisfies the retention policy and merges this backup with the underlying full and incremental backups that have already expired, thus making it a full backup. Once the merge is complete, the remaining expired backups are deleted.
Before merging or deleting backups, you can run the
delete command with the
--dry-run
flag, which displays the status of
all the available backups according to the current retention
policy, without performing any irreversible actions.
To delete all backups with specific status, use the --status
:
pg_probackup delete -Bbackup_dir
--instance=instance_name
--status=ERROR
Deleting backups by status ignores established retention policies.
pg_probackup can create a copy of a PostgreSQL instance directly, without using the backup catalog. To do this, you can run the catchup command. It can be useful in the following cases:
To add a new standby server.
Usually, pg_basebackup is used to create a copy of a PostgreSQL instance. If the data directory of the destination instance is empty, the catchup command works similarly, but it can be faster if run in parallel mode.
To have a fallen-behind standby server “catch up” with master.
Under write-intensive load, replicas may fail to replay WAL fast enough to keep up with master and hence may lag behind. A usual solution to create a new replica and switch to it requires a lot of extra space and data transfer. The catchup command allows you to update an existing replica much faster by bringing differences from master.
catchup is different from other pg_probackup operations:
The backup catalog is not required.
STREAM WAL delivery mode is only supported.
Copying external directories is not supported.
DDL commands CREATE TABLESPACE/DROP TABLESPACE cannot be run simultaneously with catchup.
catchup takes configuration files, such as
postgresql.conf
, postgresql.auto.conf
,
or pg_hba.conf
, from the source server and overwrites them
on the target server. The --exclude-path
option allows you to keep
the configuration files intact.
To prepare for cloning/synchronizing a PostgreSQL instance, set up the source server as follows:
Configure the database cluster for the instance to copy.
To copy from a remote server, configure the remote mode.
To use the PTRACK catchup mode, set up PTRACK backups.
Before cloning/synchronizing a PostgreSQL instance, ensure that the source server is running and accepting connections. To clone/sync a PostgreSQL instance, on the server with the destination instance, you can run the catchup command as follows:
pg_probackup catchup -bcatchup_mode
--source-pgdata=path_to_pgdata_on_remote_server
--destination-pgdata=path_to_local_dir
--stream [connection_options
] [remote_options
]
Where catchup_mode
can take one of the
following values:
FULL
— creates a full copy of the PostgreSQL instance.
The data directory of the destination instance must be empty for this mode.
DELTA
— reads all data files in the data directory and
creates an incremental copy for pages that have changed
since the destination instance was shut down.
PTRACK
— tracking page changes on the fly,
only reads and copies pages that have changed since the point of divergence
of the source and destination instances.
PTRACK catchup mode requires PTRACK not earlier than 2.0 and hence, PostgreSQL not earlier than 11.
By specifying the --stream
option, you can set
STREAM WAL delivery mode
of copying, which will include all the necessary WAL files by streaming them from
the server via replication protocol.
You can use connection_options to specify the connection to the source database cluster. If it is located on a different server, also specify remote_options.
If the source database cluster contains tablespaces that must be located in
a different directory, additionally specify the --tablespace-mapping
option:
pg_probackup catchup -bcatchup_mode
--source-pgdata=path_to_pgdata_on_remote_server
--destination-pgdata=path_to_local_dir
--stream --tablespace-mapping=OLDDIR
=NEWDIR
To run the catchup command on parallel threads, specify the number
of threads with the --threads
option:
pg_probackup catchup -bcatchup_mode
--source-pgdata=path_to_pgdata_on_remote_server
--destination-pgdata=path_to_local_dir
--stream --threads=num_threads
Before cloning/synchronising a PostgreSQL instance, you can run the
catchup command with the --dry-run
flag
to estimate the size of data files to be transferred, but make no changes on disk:
pg_probackup catchup -bcatchup_mode
--source-pgdata=path_to_pgdata_on_remote_server
--destination-pgdata=path_to_local_dir
--stream --dry-run
For example, assume that a remote standby server with the PostgreSQL instance having /replica-pgdata
data directory has fallen behind. To sync this instance with the one in /master-pgdata
data directory, you can run
the catchup command in the PTRACK
mode on four parallel threads as follows:
pg_probackup catchup --source-pgdata=/master-pgdata --destination-pgdata=/replica-pgdata -p 5432 -d postgres -U remote-postgres-user --stream --backup-mode=PTRACK --remote-host=remote-hostname --remote-user=remote-unix-username -j 4 --exclude-path=postgresql.conf --exclude-path=postgresql.auto.conf --exclude-path=pg_hba.conf --exclude-path=pg_ident.conf
Note that in this example, the configuration files will not be overwritten during synchronization.
Another example shows how you can add a new remote standby server with the PostgreSQL data directory /replica-pgdata
by running the catchup command in the FULL
mode
on four parallel threads:
pg_probackup catchup --source-pgdata=/master-pgdata --destination-pgdata=/replica-pgdata -p 5432 -d postgres -U remote-postgres-user --stream --backup-mode=FULL --remote-host=remote-hostname --remote-user=remote-unix-username -j 4
This section describes pg_probackup commands. Optional parameters are enclosed in square brackets. For detailed parameter descriptions, see the section Options.
pg_probackup help [command
]
Displays the synopsis of pg_probackup commands. If one of the pg_probackup commands is specified, shows detailed information about the options that can be used with this command.
pg_probackup init -B backup_dir
[--help]
Initializes the backup catalog in
backup_dir
that will store backup copies,
WAL archive, and meta information for the backed up database
clusters. If the specified backup_dir
already exists, it must be empty. Otherwise, pg_probackup
displays a corresponding error message.
For details, see the section Initializing the Backup Catalog.
pg_probackup add-instance -Bbackup_dir
-Ddata_dir
--instance=instance_name
[--help]
Initializes a new backup instance inside the backup catalog
backup_dir
and generates the
pg_probackup.conf
configuration file that controls
pg_probackup settings for the cluster with the specified
data_dir
data directory.
For details, see the section Adding a New Backup Instance.
pg_probackup del-instance -Bbackup_dir
--instance=instance_name
[--help]
Deletes all backups and WAL files associated with the specified instance.
pg_probackup set-config -Bbackup_dir
--instance=instance_name
[--help] [--pgdata=pgdata-path
] [--retention-redundancy=redundancy
][--retention-window=window
][--wal-depth=wal_depth
] [--compress-algorithm=compression_algorithm
] [--compress-level=compression_level
] [-ddbname
] [-hhost
] [-pport
] [-Uusername
] [--archive-timeout=timeout
] [--external-dirs=external_directory_path
] [--restore-command=cmdline
] [remote_options
] [remote_wal_archive_options
] [logging_options
]
Adds the specified connection, compression, retention, logging,
and external directory settings into the pg_probackup.conf
configuration file, or modifies the previously defined values.
For all available settings, see the Options section.
It is not recommended to
edit pg_probackup.conf
manually.
pg_probackup set-backup -Bbackup_dir
--instance=instance_name
-ibackup_id
{--ttl=ttl
| --expire-time=time
} [--note=backup_note
] [--help]
Sets the provided backup-specific settings into the
backup.control
configuration file, or modifies the previously
defined values.
--note=backup_note
Sets the text note for backup copy.
If backup_note
contain newline characters,
then only substring before first newline character will be saved.
Max size of text note is 1 KB.
The 'none'
value removes current note.
For all available pinning settings, see the section Pinning Options.
pg_probackup show-config -Bbackup_dir
--instance=instance_name
[--format=plain|json] [--no-scale-units] [logging_options
]
Displays the contents of the pg_probackup.conf
configuration
file located in the
directory. You can specify the
backup_dir
/backups/instance_name
--format=json
option to get the result
in the JSON format. By default, configuration settings are
shown as plain text.
You can also specify the --no-scale-units
option to display time and memory configuration settings in their base (unscaled) units.
Otherwise, the values are scaled to larger units for optimal display.
For example, if archive-timeout
is 300, then
5min
is displayed, but if archive-timeout
is 301, then 301s
is displayed.
Also, if the --no-scale-units
option is specified, configuration
settings are displayed without units and for the JSON format,
numeric and boolean values are not enclosed in quotes. This facilitates parsing
the output.
To edit pg_probackup.conf
, use the
set-config command.
pg_probackup show -Bbackup_dir
[--help] [--instance=instance_name
[-ibackup_id
| --archive]] [--format=plain|json] [--no-color]
Shows the contents of the backup catalog. If
instance_name
and
backup_id
are specified, shows detailed
information about this backup. If the --archive
option is
specified, shows the contents of WAL archive of the backup
catalog.
By default, the contents of the backup catalog is shown as
plain text. You can specify the
--format=json
option to get the result
in the JSON format.
If --no-color
flag is used,
then the output is not colored.
For details on usage, see the sections Managing the Backup Catalog and Viewing WAL Archive Information.
pg_probackup backup -Bbackup_dir
-bbackup_mode
--instance=instance_name
[--help] [-jnum_threads
] [--progress] [-C] [--stream [-S slot_name] [--temp-slot]] [--backup-pg-log] [--no-validate] [--skip-block-validation] [-w --no-password] [-W --password] [--archive-timeout=timeout
] [--external-dirs=external_directory_path
] [--no-sync] [--note=backup_note
] [connection_options
] [compression_options
] [remote_options
] [retention_options
] [pinning_options
] [logging_options
]
Creates a backup copy of the PostgreSQL instance.
-b mode
, --backup-mode=mode
Specifies the backup mode to use. Possible values are:
FULL
,
DELTA
,
PAGE
, and
PTRACK
.
-C
, --smooth-checkpoint
Spreads out the checkpoint over a period of time. By default, pg_probackup tries to complete the checkpoint as soon as possible.
--stream
Makes a STREAM backup, which includes all the necessary WAL files by streaming them from the database server via replication protocol.
--temp-slot
Creates a temporary physical replication slot for streaming
WAL from the backed up PostgreSQL instance. It ensures that
all the required WAL segments remain available if WAL is
rotated while the backup is in progress. This flag can only be
used together with the --stream
flag.
The default slot name is pg_probackup_slot
,
which can be changed using the --slot
/-S
option.
-S slot_name
, --slot=slot_name
Specifies the replication slot for WAL streaming. This option
can only be used together with the --stream
flag.
--backup-pg-log
Includes the log directory into the backup. This directory usually contains log messages. By default, log directory is excluded.
-E external_directory_path
, --external-dirs=external_directory_path
Includes the specified directory into the backup by recursively copying its contents into a separate subdirectory in the backup catalog. This option is useful to back up scripts, SQL dump files, and configuration files located outside of the data directory. If you would like to back up several external directories, separate their paths by a colon on Unix and a semicolon on Windows.
--archive-timeout=wait_time
Sets the timeout for WAL segment archiving and streaming, in seconds. By default, pg_probackup waits 300 seconds.
--skip-block-validation
Disables block-level checksum verification to speed up the backup process.
--no-validate
Skips automatic validation after the backup is taken. You can use this flag if you validate backups regularly and would like to save time when running backup operations.
--no-sync
Do not sync backed up files to disk. You can use this flag to speed up the backup process. Using this flag can result in data corruption in case of operating system or hardware crash. If you use this option, it is recommended to run the validate command once the backup is complete to detect possible issues.
--note=backup_note
Sets the text note for backup copy.
If backup_note
contain newline characters,
then only substring before first newline character will be saved.
Max size of text note is 1 KB.
The 'none'
value removes current note.
Additionally, connection options, retention options, pinning options, remote mode options, compression options, logging options, and common options can be used.
For details on usage, see the section Creating a Backup.
pg_probackup restore -Bbackup_dir
--instance=instance_name
[--help] [-Ddata_dir
] [-ibackup_id
] [-jnum_threads
] [--progress] [-TOLDDIR
=NEWDIR
] [--external-mapping=OLDDIR
=NEWDIR
] [--skip-external-dirs] [-R | --restore-as-replica] [--no-validate] [--skip-block-validation] [--force] [--no-sync] [--restore-command=cmdline
] [--primary-conninfo=primary_conninfo
] [-S | --primary-slot-name=slot_name
] [-Xwal_dir
| --waldir=wal_dir
] [recovery_target_options
] [logging_options
] [remote_options
] [partial_restore_options
] [remote_wal_archive_options
]
Restores the PostgreSQL instance from a backup copy located in
the backup_dir
backup catalog. If you
specify a recovery
target option, pg_probackup finds the closest
backup and restores it to the specified recovery target.
If neither the backup ID nor recovery target options are provided,
pg_probackup uses the most recent backup
to perform the recovery.
-R
, --restore-as-replica
Creates a minimal recovery configuration file to facilitate setting up a
standby server. If the replication connection requires a password,
you must specify the password manually in the
primary_conninfo
parameter as it is not included.
For PostgreSQL 11 or lower,
recovery settings are written into the recovery.conf
file. Starting from PostgreSQL 12,
pg_probackup writes these settings into
the probackup_recovery.conf
file in the data
directory, and then includes them into the
postgresql.auto.conf
when the cluster is
is started.
--primary-conninfo=primary_conninfo
Sets the
primary_conninfo
parameter to the specified value.
This option will be ignored unless the -R
flag is specified.
Example: --primary-conninfo="host=192.168.1.50 port=5432 user=foo password=foopass"
-S
, --primary-slot-name=slot_name
Sets the
primary_slot_name
parameter to the specified value.
This option will be ignored unless the -R
flag is specified.
-T OLDDIR
=NEWDIR
, --tablespace-mapping=OLDDIR
=NEWDIR
Relocates the tablespace from the OLDDIR
to the NEWDIR
directory at the time of recovery. Both OLDDIR
and NEWDIR
must
be absolute paths. If the path contains the equals sign (=
),
escape it with a backslash. This option can be specified
multiple times for multiple tablespaces.
--external-mapping=OLDDIR
=NEWDIR
Relocates an external directory included into the backup from
the OLDDIR
to the NEWDIR
directory at the time of recovery.
Both OLDDIR
and NEWDIR
must be absolute paths. If the path
contains the equals sign (=
), escape it with a backslash. This
option can be specified multiple times for multiple
directories.
--skip-external-dirs
Skip external directories included into the backup with the
--external-dirs
option. The contents of
these directories will not be restored.
--skip-block-validation
Disables block-level checksum verification to speed up validation. During automatic validation before the restore only file-level checksums will be verified.
--no-validate
Skips backup validation. You can use this flag if you validate backups regularly and would like to save time when running restore operations.
--restore-command=cmdline
Sets the
restore_command
parameter to the specified command. For example:
--restore-command='cp /mnt/server/archivedir/%f "%p"'
--force
Allows to ignore an invalid status of the backup. You can use
this flag if you need to restore the
PostgreSQL cluster from a corrupt or an invalid backup.
Use with caution.
If PGDATA
contains a non-empty directory with system ID different from that
of the backup being restored, incremental restore
with this flag overwrites the directory contents (while an error occurs without the flag). If tablespaces
are remapped through the --tablespace-mapping
option into non-empty directories,
the contents of such directories will be deleted.
--no-sync
Do not sync restored files to disk. You can use this flag to speed up restore process. Using this flag can result in data corruption in case of operating system or hardware crash. If it happens, you have to run the restore command again.
-X wal_dir
, --waldir=wal_dir
Specifies the directory where WAL should be stored.
Additionally, recovery target options, remote mode options, remote WAL archive options, logging options, partial restore options, and common options can be used.
For details on usage, see the section Restoring a Cluster.
pg_probackup checkdb [-Bbackup_dir
] [--instance=instance_name
] [-Ddata_dir
] [--help] [-jnum_threads
] [--progress] [--amcheck [--skip-block-validation] [--checkunique] [--heapallindexed]] [connection_options
] [logging_options
]
Verifies the PostgreSQL database cluster correctness by detecting physical and logical corruption.
--amcheck
Performs logical verification of indexes for the specified
PostgreSQL instance if no corruption was found while checking
data files. You must have the amcheck
extension or the amcheck_next extension
installed in the database to check its indexes. For databases
without amcheck, index verification will be skipped.
Additional options --checkunique
and --heapallindexed
are effective depending on the version of amcheck installed.
--checkunique
Verifies unique constraints during logical verification of indexes.
You can use this flag only together with the --amcheck
flag when
the amcheck extension is
installed in the database.
The verification of unique constraints is only possible if in the version of the
amcheck extension you are using, the
bt_index_check
function takes the
checkunique
parameter.
--heapallindexed
Checks that all heap tuples that should be indexed are
actually indexed. You can use this flag only together with the
--amcheck
flag.
This check is only possible if in the version of the
amcheck/amcheck_next extension
you are using, the bt_index_check
function takes the heapallindexed
parameter.
--skip-block-validation
Skip validation of data files. You can use this flag only
together with the --amcheck
flag, so that only logical
verification of indexes is performed.
Additionally, connection options and logging options can be used.
For details on usage, see the section Verifying a Cluster.
pg_probackup validate -Bbackup_dir
[--help] [--instance=instance_name
] [-ibackup_id
] [-jnum_threads
] [--progress] [--skip-block-validation] [recovery_target_options
] [logging_options
]
Verifies that all the files required to restore the cluster
are present and are not corrupt. If
instance_name
is not specified,
pg_probackup validates all backups available in the backup
catalog. If you specify the instance_name
without any additional options, pg_probackup validates all the
backups available for this backup instance. If you specify the
instance_name
with a
recovery target
option and/or a backup_id
,
pg_probackup checks whether it is possible to restore the
cluster using these options.
For details, see the section Validating a Backup.
pg_probackup merge -Bbackup_dir
--instance=instance_name
-ibackup_id
[--help] [-jnum_threads
] [--progress] [--no-validate] [--no-sync] [logging_options
]
Merges backups that belong to a common incremental backup chain. If you specify a full backup, it will be merged with its first incremental backup. If you specify an incremental backup, it will be merged to its parent full backup, together with all incremental backups between them. Once the merge is complete, the full backup takes in all the merged data, and the incremental backups are removed as redundant.
--no-validate
Skips automatic validation before and after merge.
--no-sync
Do not sync merged files to disk. You can use this flag to speed up the merge process. Using this flag can result in data corruption in case of operating system or hardware crash.
For details, see the section Merging Backups.
pg_probackup delete -Bbackup_dir
--instance=instance_name
[--help] [-jnum_threads
] [--progress] [--retention-redundancy=redundancy
][--retention-window=window
][--wal-depth=wal_depth
] [--delete-wal] {-ibackup_id
| --delete-expired [--merge-expired] | --merge-expired | --status=backup_status} [--dry-run] [--no-validate] [--no-sync] [logging_options
]
Deletes backup with specified backup_id
or launches the retention purge of backups and archived WAL
that do not satisfy the current retention policies.
--no-validate
Skips automatic validation before and after retention merge.
--no-sync
Do not sync merged files to disk. You can use this flag to speed up the retention merge process. Using this flag can result in data corruption in case of operating system or hardware crash.
For details, see the sections Deleting Backups, Retention Options, and Configuring Retention Policy.
pg_probackup archive-push -Bbackup_dir
--instance=instance_name
--wal-file-name=wal_file_name
[--wal-file-path=wal_file_path
] [--help] [--no-sync] [--compress] [--no-ready-rename] [--overwrite] [-jnum_threads
] [--batch-size=batch_size
] [--archive-timeout=timeout
] [--compress-algorithm=compression_algorithm
] [--compress-level=compression_level
] [remote_options
] [logging_options
]
Copies WAL files into the corresponding subdirectory of the
backup catalog and validates the backup instance by
instance_name
and
system-identifier
. If parameters of the
backup instance and the cluster do not match, this command
fails with the following error message: Refuse to push WAL
segment segment_name into archive. Instance parameters
mismatch.
If the files to be copied already exists in the backup catalog,
pg_probackup computes and compares their checksums. If the
checksums match, archive-push skips the corresponding file and
returns a successful execution code. Otherwise, archive-push
fails with an error. If you would like to replace WAL files in
the case of checksum mismatch, run the archive-push command
with the --overwrite
flag.
Each file is copied to a temporary file with the
.part
suffix. If the temporary file already
exists, pg_probackup will wait
archive_timeout
seconds before discarding it.
After the copy is done, atomic rename is performed.
This algorithm ensures that a failed archive-push
will not stall continuous archiving and that concurrent archiving from
multiple sources into a single WAL archive has no risk of archive
corruption.
To speed up archiving, you can specify the --batch-size
option
to copy WAL segments in batches of the specified size.
If --batch-size
option is used, then you can also specify
the -j
option to copy the batch of WAL segments on multiple threads.
WAL segments copied to the archive are synced to disk unless
the --no-sync
flag is used.
You can use archive-push in the archive_command PostgreSQL parameter to set up continuous WAL archiving.
For details, see sections Archiving Options and Compression Options.
pg_probackup archive-get -Bbackup_dir
--instance=instance_name
--wal-file-path=wal_file_path
--wal-file-name=wal_file_name
[-jnum_threads
] [--batch-size=batch_size
] [--prefetch-dir=prefetch_dir_path
] [--no-validate-wal] [--help] [remote_options
] [logging_options
]
Copies WAL files from the corresponding subdirectory of the backup catalog to the cluster's write-ahead log location. This command is automatically set by pg_probackup as part of the restore_command when restoring backups using a WAL archive. You do not need to set it manually.
To speed up recovery, you can specify the --batch-size
option
to copy WAL segments in batches of the specified size.
If --batch-size
option is used, then you can also specify
the -j
option to copy the batch of WAL segments on multiple threads.
For details, see section Archiving Options.
pg_probackup catchup -bcatchup_mode
--source-pgdata=path_to_pgdata_on_remote_server
--destination-pgdata=path_to_local_dir
[--help] [-j | --threads=num_threads
] [--stream] [--dry-run] [--temp-slot] [-P | --perm-slot] [-S | --slot=slot_name
] [--exclude-path=PATHNAME
] [-TOLDDIR
=NEWDIR
] [connection_options
] [remote_options
]
Creates a copy of a PostgreSQL instance without using the backup catalog.
-b catchup_mode
, --backup-mode=catchup_mode
Specifies the catchup mode to use. Possible values are:
FULL
,
DELTA
, and
PTRACK
.
--source-pgdata=path_to_pgdata_on_remote_server
Specifies the path to the data directory of the instance to be copied. The path can be local or remote.
--destination-pgdata=path_to_local_dir
Specifies the path to the local data directory to copy to.
-j num_threads
, --threads=num_threads
Sets the number of parallel threads for catchup process.
--stream
Copies the instance in STREAM WAL delivery mode, including all the necessary WAL files by streaming them from the server via replication protocol.
--dry-run
Displays the total size of the files to be transferred by catchup.
This flag initiates a trial run of catchup, which does
not actually create, delete or move files on disk. WAL streaming is skipped with --dry-run
.
This flag also allows you to check that
all the options are correct and cloning/synchronising is ready to run.
-x
=path_prefix
, --exclude-path
=path_prefix
Specifies a prefix for files to exclude from the synchronization of PostgreSQL instances during copying. The prefix must contain a path relative to the data directory of an instance. If the prefix specifies a directory, all files in this directory will not be synchronized.
This option is dangerous since excluding files from synchronization can result in incomplete synchronization; use with care.
--temp-slot
Creates a temporary physical replication slot for streaming
WAL from the PostgreSQL instance being copied. It ensures that
all the required WAL segments remain available if WAL is
rotated while the backup is in progress. This flag can only be
used together with the --stream
flag and
cannot be used together with the --perm-slot
flag.
The default slot name is pg_probackup_slot
,
which can be changed using the --slot
/-S
option.
-P
, --perm-slot
Creates a permanent physical replication slot for streaming
WAL from the PostgreSQL instance being copied. This flag can only be
used together with the --stream
flag and
cannot be used together with the --temp-slot
flag.
The default slot name is pg_probackup_perm_slot
,
which can be changed using the --slot
/-S
option.
-S slot_name
, --slot=slot_name
Specifies the replication slot for WAL streaming. This option
can only be used together with the --stream
flag.
-T OLDDIR
=NEWDIR
, --tablespace-mapping=OLDDIR
=NEWDIR
Relocates the tablespace from the OLDDIR
to the NEWDIR
directory at the time of recovery. Both OLDDIR
and NEWDIR
must
be absolute paths. If the path contains the equals sign (=
),
escape it with a backslash. This option can be specified
multiple times for multiple tablespaces.
Additionally, connection options, remote mode options can be used.
For details on usage, see the section Cloning and Synchronizing PostgreSQL Instance.
This section describes command-line options for pg_probackup
commands. If the option value can be derived from an environment
variable, this variable is specified below the command-line
option, in the uppercase. Some values can be taken from the
pg_probackup.conf
configuration file located in the backup
catalog.
For details, see the section called “Configuring pg_probackup”.
If an option is specified using more than one method,
command-line input has the highest priority, while the
pg_probackup.conf
settings have the lowest priority.
The list of general options.
-B directory
, --backup-path=directory
, BACKUP_PATH
Specifies the absolute path to the backup catalog. Backup
catalog is a directory where all backup files and meta
information are stored. Since this option is required for most
of the pg_probackup commands, you are recommended to specify
it once in the BACKUP_PATH
environment variable. In this case,
you do not need to use this option each time on the command
line.
-D directory
, --pgdata=directory
, PGDATA
Specifies the absolute path to the data directory of the
database cluster. This option is mandatory only for the
add-instance command.
Other commands can take its value from the PGDATA
environment
variable, or from the pg_probackup.conf
configuration file.
-i backup_id
, --backup-id=backup_id
Specifies the unique identifier of the backup.
-j num_threads
, --threads=num_threads
Sets the number of parallel threads for backup, restore, merge, validate, checkdb, and archive-push processes.
--progress
Shows the progress of operations.
--help
Shows detailed information about the options that can be used with this command.
If continuous WAL archiving is configured, you can use one of these options together with restore or validate commands to specify the moment up to which the database cluster must be restored or validated.
--recovery-target=immediate|latest
Defines when to stop the recovery:
The immediate
value stops the recovery
after reaching the consistent state of the specified
backup, or the latest available backup if the
-i
/--backup-id
option is omitted.
This is the default behavior for STREAM backups.
The latest
value continues the recovery
until all WAL segments available in the archive are
applied. This is the default behavior for ARCHIVE backups.
--recovery-target-timeline=timeline
Specifies a particular timeline to be used for recovery. By default, the timeline of the specified backup is used.
--recovery-target-lsn=lsn
Specifies the LSN of the write-ahead log location up to which recovery will proceed.
--recovery-target-name=recovery_target_name
Specifies a named savepoint up to which to restore the cluster.
--recovery-target-time=time
Specifies the timestamp up to which recovery will proceed. If the time zone offset is not specified, the local time zone is used.
Example: --recovery-target-time="2027-05-02 11:21:00+00"
--recovery-target-xid=xid
Specifies the transaction ID up to which recovery will proceed.
--recovery-target-inclusive=boolean
,
Specifies whether to stop just after the specified recovery
target (true
), or just before the recovery target (false
).
This option can only be used together with
--recovery-target-name
,
--recovery-target-time
,
--recovery-target-lsn
or
--recovery-target-xid
options. The default
depends on the
recovery_target_inclusive
parameter.
--recovery-target-action=pause|promote|shutdown
Specifies the action the server should take when the recovery target is reached.
Default: pause
You can use these options together with backup and delete commands.
For details on configuring retention policy, see the section Configuring Retention Policy.
--retention-redundancy=redundancy
,
Specifies the number of full backup copies to keep in the data directory. Must be a non-negative integer. The zero value disables this setting.
Default: 0
--retention-window=window
,
Number of days of recoverability. Must be a non-negative integer. The zero value disables this setting.
Default: 0
--wal-depth=wal_depth
Number of latest valid backups on every timeline that must retain the ability to perform PITR. Must be a non-negative integer. The zero value disables this setting.
Default: 0
--delete-wal
Deletes WAL files that are no longer required to restore the cluster from any of the existing backups.
--delete-expired
Deletes backups that do not conform to the retention policy
defined in the pg_probackup.conf
configuration file.
--merge-expired
,
Merges the oldest incremental backup that satisfies the requirements of retention policy with its parent backups that have already expired.
--dry-run
Displays the current status of all the available backups, without deleting or merging expired backups, if any.
You can use these options together with backup and set-backup commands.
For details on backup pinning, see the section Backup Pinning.
--ttl=ttl
Specifies the amount of time the backup should be pinned. Must be a non-negative integer. The zero value unpins the already pinned backup. Supported units: ms, s, min, h, d (s by default).
Example: --ttl=30d
--expire-time=time
Specifies the timestamp up to which the backup will stay pinned. Must be an ISO-8601 complaint timestamp. If the time zone offset is not specified, the local time zone is used.
Example: --expire-time="2027-05-02 11:21:00+00"
You can use these options with any command.
--no-color
Disable coloring for console log messages of warning
and error
levels.
--log-level-console=log_level
Controls which message levels are sent to the console log.
Valid values are verbose
,
log
, info
,
warning
, error
and
off
. Each level includes all the levels
that follow it. The later the level, the fewer messages are
sent. The off
level disables console
logging.
Default: info
All console log messages are going to stderr
, so
the output of show and
show-config commands does
not mingle with log messages.
--log-level-file=log_level
Controls which message levels are sent to a log file. Valid
values are verbose
, log
,
info
, warning
,
error
, and off
. Each
level includes all the levels that follow it. The later the
level, the fewer messages are sent. The off
level disables file logging.
Default: off
--log-filename=log_filename
Defines the filenames of the created log files. The filenames
are treated as a strftime
pattern, so you can use %-escapes to
specify time-varying filenames.
Default: pg_probackup.log
For example, if you specify the pg_probackup-%u.log
pattern,
pg_probackup generates a separate log file for each day of the
week, with %u
replaced by the corresponding decimal number:
pg_probackup-1.log
for Monday, pg_probackup-2.log
for Tuesday,
and so on.
This option takes effect if file logging is enabled by the
--log-level-file
option.
--error-log-filename=error_log_filename
Defines the filenames of log files for error messages only.
The filenames are treated as a strftime
pattern, so you can
use %-escapes to specify time-varying filenames.
Default: none
For example, if you specify the error-pg_probackup-%u.log
pattern, pg_probackup generates a separate log file for each
day of the week, with %u
replaced by the corresponding decimal
number: error-pg_probackup-1.log
for Monday,
error-pg_probackup-2.log
for Tuesday, and so on.
This option is useful for troubleshooting and monitoring.
--log-directory=log_directory
Defines the directory in which log files will be created. You must specify the absolute path. This directory is created lazily, when the first log message is written.
Default: $BACKUP_PATH/log/
--log-format-console=log_format
Defines the format of the console log. Only set from the command line. Note that you cannot
specify this option in the pg_probackup.conf
configuration file through
the set-config command and that the backup
command also treats this option specified in the configuration file as an error.
Possible values are:
plain
— sets the plain-text format of the console log.
json
— sets the JSON format of the console log.
Default: plain
--log-format-file=log_format
Defines the format of log files used. Possible values are:
plain
— sets the plain-text format of log files.
json
— sets the JSON format of log files.
Default: plain
--log-rotation-size=log_rotation_size
Maximum size of an individual log file. If this value is reached, the log file is rotated once a pg_probackup command is launched, except help and version commands. The zero value disables size-based rotation. Supported units: kB, MB, GB, TB (kB by default).
Default: 0
--log-rotation-age=log_rotation_age
Maximum lifetime of an individual log file. If this value is
reached, the log file is rotated once a pg_probackup command
is launched, except help and version commands. The time of the
last log file creation is stored in
$BACKUP_PATH/log/log_rotation
. The zero value disables
time-based rotation. Supported units: ms, s, min, h, d (min by
default).
Default: 0
You can use these options together with backup, catchup, and checkdb commands.
All libpq environment variables are supported.
-d dbname
, --pgdatabase=dbname
, PGDATABASE
Specifies the name of the database to connect to. The
connection is used only for managing backup process, so you
can connect to any existing database. If this option is not
provided on the command line, PGDATABASE
environment variable,
or the pg_probackup.conf
configuration file, pg_probackup
tries to take this value from the PGUSER
environment variable,
or from the current user name if PGUSER
variable is not set.
-h host
, --pghost=host
, PGHOST
Specifies the host name of the system on which the server is running. If the value begins with a slash, it is used as a directory for the Unix domain socket.
Default: localhost
-p port
, --pgport=port
, PGPORT
Specifies the TCP port or the local Unix domain socket file extension on which the server is listening for connections.
Default: 5432
-U username
, --pguser=username
, PGUSER
User name to connect as.
-w
, --no-password
Disables a password prompt. If the server requires password
authentication and a password is not available by other means
such as a
.pgpass
file or PGPASSWORD
environment variable, the connection
attempt will fail. This flag can be useful in batch jobs and
scripts where no user is present to enter a password.
-W
, --password
Forces a password prompt. (Deprecated)
You can use these options together with backup and archive-push commands.
--compress-algorithm=compression_algorithm
Defines the algorithm to use for compressing data files.
Possible values are zlib
,
pglz
, and none
. If set
to zlib
or pglz
, this option enables compression. By default,
compression is disabled. For the
archive-push command, the
pglz
compression algorithm is not supported.
Default: none
--compress-level=compression_level
Defines compression level (0 through 9, 0 being no compression
and 9 being best compression). This option can be used
together with the --compress-algorithm
option.
Default: 1
--compress
Alias for --compress-algorithm=zlib
and
--compress-level=1
.
These options can be used with the archive-push command in the archive_command setting and the archive-get command in the restore_command setting.
Additionally, remote mode options and logging options can be used.
--wal-file-path=wal_file_path
Provides the path to the WAL file in
archive_command
and
restore_command
. Use the %p
variable as the value for this option or explicitly specify the path to a file
outside of the data directory. If you skip this option, the path
specified in pg_probackup.conf
will be used.
--wal-file-name=wal_file_name
Provides the name of the WAL file in
archive_command
and
restore_command
. Use the %f
variable as the value for this option for correct processing.
If the value of --wal-file-path
is a path
outside of the data directory, explicitly specify the filename.
--overwrite
Overwrites archived WAL file. Use this flag together with the
archive-push command if
the specified subdirectory of the backup catalog already
contains this WAL file and it needs to be replaced with its
newer copy. Otherwise, archive-push reports that a WAL segment
already exists, and aborts the operation. If the file to
replace has not changed, archive-push skips this file
regardless of the --overwrite
flag.
--batch-size=batch_size
Sets the maximum number of files that can be copied into the archive by a single archive-push process, or from the archive by a single archive-get process.
--archive-timeout=wait_time
Sets the timeout for considering existing .part
files to be stale. By default, pg_probackup
waits 300 seconds.
This option can be used only with archive-push command.
--no-ready-rename
Do not rename status files in the archive_status
directory.
This option should be used only if archive_command
contains multiple commands.
This option can be used only with archive-push command.
--no-sync
Do not sync copied WAL files to disk. You can use this flag to speed up archiving process. Using this flag can result in WAL archive corruption in case of operating system or hardware crash. This option can be used only with archive-push command.
--prefetch-dir=path
Directory used to store prefetched WAL segments if --batch-size
option is used.
Directory must be located on the same filesystem and on the same mountpoint the
PGDATA/pg_wal
is located.
By default files are stored in PGDATA/pg_wal/pbk_prefetch
directory.
This option can be used only with archive-get command.
--no-validate-wal
Do not validate prefetched WAL file before using it. Use this option if you want to increase the speed of recovery. This option can be used only with archive-get command.
This section describes the options related to running pg_probackup operations remotely via SSH. These options can be used with add-instance, set-config, backup, catchup, restore, archive-push, and archive-get commands.
For details on configuring and using the remote mode, see the section called “Configuring the Remote Mode” and the section called “Using pg_probackup in the Remote Mode”.
--remote-proto=proto
Specifies the protocol to use for remote operations. Currently only the SSH protocol is supported. Possible values are:
ssh
enables the remote mode via
SSH. This is the default value.
none
explicitly disables the remote
mode.
You can omit this option if the
--remote-host
option is specified.
--remote-host=destination
,
Specifies the remote host IP address or hostname to connect to.
--remote-port=port
Specifies the remote host port to connect to.
Default: 22
--remote-user=username
Specifies remote host user for SSH connection. If you omit this option, the current user initiating the SSH connection is used.
--remote-path=path
Specifies pg_probackup installation directory on the remote system.
--ssh-options=ssh_options
Provides a string of SSH command-line options. For example,
the following options can be used to set keep-alive
for SSH
connections opened by pg_probackup:
--ssh-options="-o ServerAliveCountMax=5 -o ServerAliveInterval=60"
.
For the full list of possible options, see
ssh_config
manual page.
This section describes the options used to provide the arguments for remote mode options in archive-get used in the restore_command command when restoring ARCHIVE backups or performing PITR.
--archive-host=destination
Provides the argument for the --remote-host
option in the archive-get command.
--archive-port=port
Provides the argument for the --remote-port
option in the archive-get command.
Default: 22
--archive-user=username
Provides the argument for the --remote-user
option in the archive-get command. If you omit
this option, the user that has started the PostgreSQL cluster is used.
Default: PostgreSQL user
This section describes the options for incremental cluster restore. These options can be used with the restore command.
-I incremental_mode
, --incremental-mode=incremental_mode
Specifies the incremental mode to be used. Possible values are:
CHECKSUM
— replace only pages with mismatched checksum and LSN.
LSN
— replace only pages with LSN greater than point of divergence.
NONE
— regular restore.
This section describes the options for partial cluster restore. These options can be used with the restore command.
--db-exclude=dbname
Specifies the name of the database to exclude from restore. All other
databases in the cluster will be restored as usual, including
template0
and template1
.
This option can be specified multiple times for multiple
databases.
--db-include=dbname
Specifies the name of the database to restore from a backup. All other
databases in the cluster will not be restored, with the exception
of template0
and
template1
. This option can be specified
multiple times for multiple databases.