# Checking port 49277
# Found port 49277
Name: primary
Data directory: /home/vagrant/postgresql/src/test/recovery_13/tmp_check/t_035_standby_logical_decoding_primary_data/pgdata
Backup directory: /home/vagrant/postgresql/src/test/recovery_13/tmp_check/t_035_standby_logical_decoding_primary_data/backup
Archive directory: /home/vagrant/postgresql/src/test/recovery_13/tmp_check/t_035_standby_logical_decoding_primary_data/archives
Connection string: port=49277 host=/tmp/x7RIMHIkA0
Log file: /home/vagrant/postgresql/src/test/recovery_13/tmp_check/log/035_standby_logical_decoding_primary.log
# Checking port 49278
# Found port 49278
Name: standby
Data directory: /home/vagrant/postgresql/src/test/recovery_13/tmp_check/t_035_standby_logical_decoding_standby_data/pgdata
Backup directory: /home/vagrant/postgresql/src/test/recovery_13/tmp_check/t_035_standby_logical_decoding_standby_data/backup
Archive directory: /home/vagrant/postgresql/src/test/recovery_13/tmp_check/t_035_standby_logical_decoding_standby_data/archives
Connection string: port=49278 host=/tmp/x7RIMHIkA0
Log file: /home/vagrant/postgresql/src/test/recovery_13/tmp_check/log/035_standby_logical_decoding_standby.log
# Checking port 49279
# Found port 49279
Name: cascading_standby
Data directory: /home/vagrant/postgresql/src/test/recovery_13/tmp_check/t_035_standby_logical_decoding_cascading_standby_data/pgdata
Backup directory: /home/vagrant/postgresql/src/test/recovery_13/tmp_check/t_035_standby_logical_decoding_cascading_standby_data/backup
Archive directory: /home/vagrant/postgresql/src/test/recovery_13/tmp_check/t_035_standby_logical_decoding_cascading_standby_data/archives
Connection string: port=49279 host=/tmp/x7RIMHIkA0
Log file: /home/vagrant/postgresql/src/test/recovery_13/tmp_check/log/035_standby_logical_decoding_cascading_standby.log
# Checking port 49280
# Found port 49280
Name: subscriber
Data directory: /home/vagrant/postgresql/src/test/recovery_13/tmp_check/t_035_standby_logical_decoding_subscriber_data/pgdata
Backup directory: /home/vagrant/postgresql/src/test/recovery_13/tmp_check/t_035_standby_logical_decoding_subscriber_data/backup
Archive directory: /home/vagrant/postgresql/src/test/recovery_13/tmp_check/t_035_standby_logical_decoding_subscriber_data/archives
Connection string: port=49280 host=/tmp/x7RIMHIkA0
Log file: /home/vagrant/postgresql/src/test/recovery_13/tmp_check/log/035_standby_logical_decoding_subscriber.log
[13:34:33.360](1.029s) # initializing database system by copying initdb template
# Running: cp -RPp /home/vagrant/postgresql/tmp_install/initdb-template /home/vagrant/postgresql/src/test/recovery_13/tmp_check/t_035_standby_logical_decoding_primary_data/pgdata
# Running: /home/vagrant/postgresql/src/test/recovery_13/../../../src/test/regress/pg_regress --config-auth /home/vagrant/postgresql/src/test/recovery_13/tmp_check/t_035_standby_logical_decoding_primary_data/pgdata
### Enabling WAL archiving for node "primary"
Name: primary
Version: 17devel
Data directory: /home/vagrant/postgresql/src/test/recovery_13/tmp_check/t_035_standby_logical_decoding_primary_data/pgdata
Backup directory: /home/vagrant/postgresql/src/test/recovery_13/tmp_check/t_035_standby_logical_decoding_primary_data/backup
Archive directory: /home/vagrant/postgresql/src/test/recovery_13/tmp_check/t_035_standby_logical_decoding_primary_data/archives
Connection string: port=49277 host=/tmp/x7RIMHIkA0
Log file: /home/vagrant/postgresql/src/test/recovery_13/tmp_check/log/035_standby_logical_decoding_primary.log
### Starting node "primary"
# Running: pg_ctl -w -D /home/vagrant/postgresql/src/test/recovery_13/tmp_check/t_035_standby_logical_decoding_primary_data/pgdata -l /home/vagrant/postgresql/src/test/recovery_13/tmp_check/log/035_standby_logical_decoding_primary.log -o --cluster-name=primary start
waiting for server to start.... done
server started
# Postmaster PID for node "primary" is 188972
[13:35:26.431](53.071s) ok 1 - Physical slot reports conflict_reason as NULL
# Taking pg_basebackup b1 from node "primary"
# Running: pg_basebackup -D /home/vagrant/postgresql/src/test/recovery_13/tmp_check/t_035_standby_logical_decoding_primary_data/backup/b1 -h /tmp/x7RIMHIkA0 -p 49277 --checkpoint fast --no-sync
# Backup finished
# Initializing node "standby" from backup "b1" of node "primary"
### Enabling streaming replication for node "standby"
### Enabling WAL restore for node "standby"
### Starting node "standby"
# Running: pg_ctl -w -D /home/vagrant/postgresql/src/test/recovery_13/tmp_check/t_035_standby_logical_decoding_standby_data/pgdata -l /home/vagrant/postgresql/src/test/recovery_13/tmp_check/log/035_standby_logical_decoding_standby.log -o --cluster-name=standby start
waiting for server to start..... done
server started
# Postmaster PID for node "standby" is 190139
Waiting for replication conn standby's replay_lsn to pass 0/3019998 on primary
done
[13:38:03.360](156.929s) # initializing database system by copying initdb template
# Running: cp -RPp /home/vagrant/postgresql/tmp_install/initdb-template /home/vagrant/postgresql/src/test/recovery_13/tmp_check/t_035_standby_logical_decoding_subscriber_data/pgdata
# Running: /home/vagrant/postgresql/src/test/recovery_13/../../../src/test/regress/pg_regress --config-auth /home/vagrant/postgresql/src/test/recovery_13/tmp_check/t_035_standby_logical_decoding_subscriber_data/pgdata
### Starting node "subscriber"
# Running: pg_ctl -w -D /home/vagrant/postgresql/src/test/recovery_13/tmp_check/t_035_standby_logical_decoding_subscriber_data/pgdata -l /home/vagrant/postgresql/src/test/recovery_13/tmp_check/log/035_standby_logical_decoding_subscriber.log -o --cluster-name=subscriber start
waiting for server to start...... done
server started
# Postmaster PID for node "subscriber" is 192142
[13:39:03.356](59.996s) ok 2 - behaves_ok_inactiveslot on standby created
[13:39:06.513](3.157s) ok 3 - behaves_ok_activeslot on standby created
Waiting for replication conn standby's replay_lsn to pass 0/3025970 on primary
done
[13:39:09.304](2.791s) ok 4 - Decoding produced 14 rows (2 BEGIN/COMMIT and 10 rows)
Waiting for replication conn standby's replay_lsn to pass 0/3025AA0 on primary
done
[13:39:10.660](1.357s) ok 5 - got expected output from SQL decoding session
Waiting for replication conn standby's replay_lsn to pass 0/3026668 on primary
done
[13:39:12.503](1.843s) ok 6 - got same expected output from pg_recvlogical decoding session
[13:39:13.175](0.672s) ok 7 - pg_recvlogical acknowledged changes
Waiting for replication conn standby's replay_lsn to pass 0/3458670 on primary
done
[13:39:39.692](26.517s) ok 8 - replaying logical slot from another database fails
Waiting for replication conn standby's replay_lsn to pass 0/347F860 on primary
done
Waiting for all subscriptions in "subscriber" to synchronize data
Waiting for replication conn tap_sub's replay_lsn to pass 0/347F8D0 on standby
done
done
Waiting for replication conn standby's replay_lsn to pass 0/347FE60 on primary
done
Waiting for replication conn tap_sub's replay_lsn to pass 0/347FE60 on standby
done
[13:40:04.676](24.984s) ok 9 - check replicated inserts after subscription on standby
#### Begin standard error
psql:<stdin>:1: NOTICE:  dropped replication slot "tap_sub" on publisher
#### End standard error
### Stopping node "subscriber" using mode fast
# Running: pg_ctl -D /home/vagrant/postgresql/src/test/recovery_13/tmp_check/t_035_standby_logical_decoding_subscriber_data/pgdata -m fast stop
waiting for server to shut down.... done
server stopped
# No postmaster PID for node "subscriber"


[13:40:18.362](13.686s) ok 10 - vacuum_full_inactiveslot on standby created
[13:40:19.605](1.243s) ok 11 - vacuum_full_activeslot on standby created
### Reloading node "standby"
# Running: pg_ctl -D /home/vagrant/postgresql/src/test/recovery_13/tmp_check/t_035_standby_logical_decoding_standby_data/pgdata reload
server signaled

#### Begin standard error
psql:<stdin>:1: INFO:  vacuuming "pg_catalog.pg_class"
psql:<stdin>:1: INFO:  "pg_catalog.pg_class": found 5 removable, 420 nonremovable row versions in 14 pages
DETAIL:  0 dead row versions cannot be removed yet.
CPU: user: 0.00 s, system: 0.00 s, elapsed: 0.00 s.
#### End standard error
Waiting for replication conn standby's replay_lsn to pass 0/34B1068 on primary
done
[13:40:23.799](4.193s) ok 12 - inactiveslot slot invalidation is logged with vacuum FULL on pg_class
[13:40:23.800](0.001s) ok 13 - activeslot slot invalidation is logged with vacuum FULL on pg_class
[13:40:25.134](1.334s) ok 14 - confl_active_logicalslot updated
[13:40:25.392](0.258s) ok 15 - vacuum_full_activeslot conflict_reason is rows_removed
[13:40:25.703](0.312s) ok 16 - vacuum_full_inactiveslot conflict_reason is rows_removed
[13:40:27.359](1.656s) ok 17 - pg_recvlogical exited non-zero
[13:40:27.359](0.000s) ok 18 - slot has been invalidated
### Reloading node "standby"
# Running: pg_ctl -D /home/vagrant/postgresql/src/test/recovery_13/tmp_check/t_035_standby_logical_decoding_standby_data/pgdata reload
server signaled
### Restarting node "standby"
# Running: pg_ctl -w -D /home/vagrant/postgresql/src/test/recovery_13/tmp_check/t_035_standby_logical_decoding_standby_data/pgdata -l /home/vagrant/postgresql/src/test/recovery_13/tmp_check/log/035_standby_logical_decoding_standby.log restart
waiting for server to shut down.... done
server stopped
waiting for server to start.... done
server started
# Postmaster PID for node "standby" is 195820
[13:40:31.049](3.690s) ok 19 - vacuum_full_activeslot conflict_reason is rows_removed
[13:40:31.691](0.642s) ok 20 - vacuum_full_inactiveslot conflict_reason is rows_removed
Waiting for replication conn standby's replay_lsn to pass 0/40000F8 on primary
done


[13:40:40.438](8.747s) ok 21 - row_removal_inactiveslot on standby created
[13:40:41.797](1.359s) ok 22 - row_removal_activeslot on standby created
### Reloading node "standby"
# Running: pg_ctl -D /home/vagrant/postgresql/src/test/recovery_13/tmp_check/t_035_standby_logical_decoding_standby_data/pgdata reload
server signaled

#### Begin standard error
psql:<stdin>:1: INFO:  vacuuming "testdb.pg_catalog.pg_class"
psql:<stdin>:1: INFO:  finished vacuuming "testdb.pg_catalog.pg_class": index scans: 1
pages: 0 removed, 11 remain, 11 scanned (100.00% of total)
tuples: 4 removed, 419 remain, 0 are dead but not yet removable
removable cutoff: 754, which was 0 XIDs old when operation ended
frozen: 0 pages from table (0.00% of total) had 0 tuples frozen
index scan needed: 3 pages from table (27.27% of total) had 5 dead item identifiers removed
index "pg_class_oid_index": pages: 4 in total, 0 newly deleted, 0 currently deleted, 0 reusable
index "pg_class_relname_nsp_index": pages: 5 in total, 0 newly deleted, 0 currently deleted, 0 reusable
index "pg_class_tblspc_relfilenode_index": pages: 2 in total, 0 newly deleted, 0 currently deleted, 0 reusable
avg read rate: 0.025 MB/s, avg write rate: 0.053 MB/s
buffer usage: 78 hits, 7 misses, 15 dirtied
WAL usage: 28 records, 15 full page images, 122841 bytes
system usage: CPU: user: 0.00 s, system: 0.05 s, elapsed: 2.20 s
#### End standard error
Waiting for replication conn standby's replay_lsn to pass 0/403E6F8 on primary
done
[13:41:02.076](20.279s) ok 23 - inactiveslot slot invalidation is logged with vacuum on pg_class
[13:41:02.076](0.000s) not ok 24 - activeslot slot invalidation is logged with vacuum on pg_class
[13:41:02.077](0.000s) 
[13:41:02.077](0.000s) #   Failed test 'activeslot slot invalidation is logged with vacuum on pg_class'
#   at t/035_standby_logical_decoding.pl line 227.
