Thread: Question on session_replication_role
Our set up:
* Db version: postgres 9.3.4
* OS: CentOS 5.6
* kernel Version - Linux 2.6.18-238.19.1.el5 #1 SMP Fri Jul 15 07:31:24 EDT 2011 x86_64 x86_64 x86_64 GNU/Linux
* memory - 256 GB
* We have slony replicating to this DB from the primary server
* Some of the DB parameters:
max_connections - 1500
shared_buffers - 8GB
work_mem = 256MB
maintenance_work_mem = 1GB
Problem statement:
On Jan 27th 2015:
We inserted 77k records to a table with 780 million records by disabling triggers using SET session_replication_role = replica;
Table definition:
site=# \d reg_email_subscriptions
Table "emailsubscription.reg_email_subscriptions"
Column | Type | Modifiers
------------------+--------------------------------+-----------
registration_id | bigint | not null
subscription_id | bigint | not null
creation_date | timestamp(0) without time zone | not null
last_update_date | timestamp(0) without time zone | not null
Number of child tables: 8 (Use \d+ to list them.)
Tablespace: "emailsubscription"
Indexes/constraints on each of the 8 partitions:
"reg_email_subscriptions_p00_pkey" PRIMARY KEY, btree (registration_id, subscription_id), tablespace "emailsubscription"
"reg_email_subscriptions_p00_n1" btree (subscription_id), tablespace "emailsubscription"
"reg_email_subscriptions_p00_n20" btree (last_update_date), tablespace "emailsubscription"
"reg_email_subscriptions_p00_n3" btree (creation_date), tablespace “emailsubscription"
Soon after learning that using this setting also disables PK on the table, we manually deleted 117 duplicate records that got inserted.
On Jan 28th 2015:
The query that was taking 7 secs started taking over 30 seconds.
SELECT DISTINCT it.recipient_id FROM iru.iru_tags it WHERE it.recipient_id BETWEEN 758587587 and 968696896 AND it.status = 0 AND it.last_update_date >= date_trunc('day', now() - interval '90 days') AND EXISTS (SELECT DISTINCT res.registration_id FROM emailsubscription.reg_email_subscriptions res WHERE res.registration_id = it.recipient_id AND res.subscription_id = 200400);
Table definition of the other table involved in the query (iru.iru_tags): - No manual DML was done on this table.
site=# \d iru_tags
Table "iru.iru_tags"
Column | Type | Modifiers
------------------+-----------------------------+-----------
recipient_id | bigint | not null
tagger_id | bigint | not null
tag_id | integer | not null
batch_id | integer | not null
status | integer |
creation_date | timestamp without time zone | not null
last_update_date | timestamp without time zone | not null
Indexes:
"iru_tags_pk" PRIMARY KEY, btree (recipient_id, tagger_id, batch_id, tag_id)
"iru_tags_n1" btree (recipient_id, tag_id)
"iru_tags_n2" btree (last_update_date)
"iru_tags_n3" btree (creation_date)
"iru_tags_n31" btree (status, recipient_id, last_update_date)
"iru_tags_n4" btree (tagger_id)
Observation:
1) The query was taking 2-7 seconds consistently until the day before we inserted records by disabling the triggers.
2) When we run the selects on each table separately, the query runs really fast. The moment we introduce the join (AND EXISTS), the sql takes over 30 seconds.
3) The explain plan of this query shows that Primary key on reg_email_subscriptions and unique index on iru_tags table is being used.
site=# explain(analyze on, verbose on, costs on, buffers on, timing on) SELECT DISTINCT it.recipient_id FROM iru.iru_tags it WHERE it.recipient_id BETWEEN 758587587 and 968696896 AND it.status = 0 AND it.last_update_date >= date_trunc('day', now() - interval '90 days') AND EXISTS (SELECT DISTINCT res.registration_id FROM emailsubscription.reg_email_subscriptions res WHERE res.registration_id = it.recipient_id AND res.subscription_id = 200400);
QUERY PLAN
--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------
Unique (cost=0.57..290191.08 rows=45 width=8) (actual time=89.536..89.536 rows=0 loops=1)
Output: it.recipient_id
Buffers: shared hit=9 read=1
I/O Timings: read=47.097
-> Nested Loop Semi Join (cost=0.57..290187.87 rows=1286 width=8) (actual time=89.534..89.534 rows=0 loops=1)
Output: it.recipient_id
Buffers: shared hit=9 read=1
I/O Timings: read=47.097
-> Index Only Scan using iru_tags_n31 on iru.iru_tags it (cost=0.57..319.84 rows=1286 width=8) (actual time=89.532..89.532 rows=0 loops=1)
Output: it.status, it.recipient_id, it.last_update_date
Index Cond: ((it.status = 0) AND (it.recipient_id >= 758587587) AND (it.recipient_id <= 968696896) AND (it.last_update_date >= date_trunc('day'::text, (now() - '90 days'::interval))))
Heap Fetches: 0
Buffers: shared hit=9 read=1
I/O Timings: read=47.097
-> Append (cost=0.00..1780.03 rows=17 width=8) (never executed)
-> Seq Scan on emailsubscription.reg_email_subscriptions res (cost=0.00..1728.07 rows=1 width=8) (never executed)
Output: res.registration_id
Filter: ((res.subscription_id = 200400) AND (it.recipient_id = res.registration_id))
-> Index Only Scan using reg_email_subscriptions_p00_pkey on emailsubscription.reg_email_subscriptions_p00 res_1 (cost=0.57..6.46 rows=2 width=8) (never executed)
Output: res_1.registration_id
Index Cond: ((res_1.registration_id = it.recipient_id) AND (res_1.subscription_id = 200400))
Heap Fetches: 0
-> Index Only Scan using reg_email_subscriptions_p01_pkey on emailsubscription.reg_email_subscriptions_p01 res_2 (cost=0.57..6.53 rows=2 width=8) (never executed)
Output: res_2.registration_id
Index Cond: ((res_2.registration_id = it.recipient_id) AND (res_2.subscription_id = 200400))
Heap Fetches: 0
-> Index Only Scan using reg_email_subscriptions_p02_pkey on emailsubscription.reg_email_subscriptions_p02 res_3 (cost=0.57..6.46 rows=2 width=8) (never executed)
Output: res_3.registration_id
Index Cond: ((res_3.registration_id = it.recipient_id) AND (res_3.subscription_id = 200400))
Heap Fetches: 0
-> Index Only Scan using reg_email_subscriptions_p03_pkey on emailsubscription.reg_email_subscriptions_p03 res_4 (cost=0.57..6.53 rows=2 width=8) (never executed)
Output: res_4.registration_id
Index Cond: ((res_4.registration_id = it.recipient_id) AND (res_4.subscription_id = 200400))
Heap Fetches: 0
-> Index Only Scan using reg_email_subscriptions_p04_pkey on emailsubscription.reg_email_subscriptions_p04 res_5 (cost=0.57..6.46 rows=2 width=8) (never executed)
Output: res_5.registration_id
Index Cond: ((res_5.registration_id = it.recipient_id) AND (res_5.subscription_id = 200400))
Heap Fetches: 0
-> Index Only Scan using reg_email_subscriptions_p05_pkey on emailsubscription.reg_email_subscriptions_p05 res_6 (cost=0.57..6.53 rows=2 width=8) (never executed)
Output: res_6.registration_id
Index Cond: ((res_6.registration_id = it.recipient_id) AND (res_6.subscription_id = 200400))
Heap Fetches: 0
-> Index Only Scan using reg_email_subscriptions_p06_pkey on emailsubscription.reg_email_subscriptions_p06 res_7 (cost=0.57..6.46 rows=2 width=8) (never executed)
Output: res_7.registration_id
Index Cond: ((res_7.registration_id = it.recipient_id) AND (res_7.subscription_id = 200400))
Heap Fetches: 0
-> Index Only Scan using reg_email_subscriptions_p07_pkey on emailsubscription.reg_email_subscriptions_p07 res_8 (cost=0.57..6.53 rows=2 width=8) (never executed)
Output: res_8.registration_id
Index Cond: ((res_8.registration_id = it.recipient_id) AND (res_8.subscription_id = 200400))
Heap Fetches: 0
Total runtime: 90.314 ms
(51 rows)
4) We have done updates on various tables after setting session_replication_role to replica without any performance issues in the past.
What was done so far:
1) Full Vacuum on reg_email_subsciptions table
2) Full vacuum on iru_tags table
3) Manual reindexing on all 8 primary keys on reg_email_subsciptions_p0* tables
Questions:
1) Did setting session_replication_role to replica before inserting (duplicate) records corrupt the primary key or any of the indexes on reg_email_subscriptions table?
2) If so, should a reindex on the corrupt PK or index identify the corruption and report/fix it?
3) Is this even a problem with index/data corruption?
Thank You for your help
On 2/2/15 8:07 PM, Anand Kumar, Karthik wrote: > 2)When we run the selects on each table separately, the query runs > really fast. The moment we introduce the join (AND EXISTS), the sql > takes over 30 seconds. > > 3)The explain plan of this query shows that Primary key on > reg_email_subscriptions and unique index on iru_tags table is being used. > > site=# explain(analyze on, verbose on, costs on, buffers on, timing on) > SELECT DISTINCT it.recipient_id FROM iru.iru_tags it WHERE > it.recipient_id BETWEEN 758587587 and 968696896 AND it.status = 0 > AND it.last_update_date >= date_trunc('day', now() - interval '90 > days') AND EXISTS (SELECT DISTINCT res.registration_id FROM > emailsubscription.reg_email_subscriptions res WHERE res.registration_id > = it.recipient_id AND res.subscription_id = 200400); > > QUERY PLAN > > -------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- > > Unique (cost=0.57..290191.08 rows=45 width=8) (actual > time=89.536..89.536 rows=0 loops=1) Your EXPLAIN ANALYZE output doesn't show this taking 30 seconds, it shows it taking 90ms. Please provide an EXPLAIN ANALYZE That actually demonstrates the problem. > 1)Did setting session_replication_role to replica before inserting > (duplicate) records corrupt the primary key or any of the indexes on > reg_email_subscriptions table? I don't believe that session_replication_role does anything to affect indexing, nor do the docs indicate that. I also see nothing indicating that in the source code (search for SessionReplicationRole). So if you suddenly started seeing dupes then I think your index is corrupted. > 2)If so, should a reindex on the corrupt PK or index identify the > corruption and report/fix it? It won't report it, but it would fix it. (Note that it would throw an error if the data in the table isn't actually unique.) > 3)Is this even a problem with index/data corruption? I doubt it, though you certainly could have corruption. -- Jim Nasby, Data Architect, Blue Treble Consulting Data in Trouble? Get it in Treble! http://BlueTreble.com
Jim, I am OP (Karthik)'s colleague. Please see the responses below. > Your EXPLAIN ANALYZE output doesn't show this taking 30 seconds, it shows it taking 90ms. Please provide an EXPLAIN ANALYZEThat actually demonstrates the problem. SELECT DISTINCT it.recipient_id FROM iru.iru_tags it WHERE it.recipient_id BETWEEN 7257057171 AND 7257704235 AND it.status= 0 AND it.last_update_date >= date_trunc('day', now() - interval '90 days') AND EXISTS (SELECT DISTINCT res.registration_idFROM emailsubscription.reg_email_subscriptions res WHERE res.registration_id = it.recipient_id AND res.subscription_id= 200400); QUERY PLAN ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- Unique (cost=0.57..296573.59 rows=46 width=8) (actual time=24.613..21194.635 rows=904 loops=1) Output: it.recipient_id Buffers: shared hit=1094265 read=718 I/O Timings: read=28.674 -> Nested Loop Semi Join (cost=0.57..296570.25 rows=1335 width=8) (actual time=24.608..21190.382 rows=1264 loops=1) Output: it.recipient_id Buffers: shared hit=1094265 read=718 I/O Timings: read=28.674 -> Index Only Scan using iru_tags_n31 on iru.iru_tags it (cost=0.57..337.19 rows=1335 width=8) (actual time=0.184..25.875rows=1847 loops=1) Output: it.status, it.recipient_id, it.last_update_date Index Cond: ((it.status = 0) AND (it.recipient_id >= 7257057171::bigint) AND (it.recipient_id <= 7257704235::bigint)AND (it.last_update_date >= date_trunc('day'::text, (now() - '90 days'::interval)))) Heap Fetches: 103 Buffers: shared hit=820 read=180 I/O Timings: read=7.614 -> Append (cost=0.00..1748.87 rows=17 width=8) (actual time=11.454..11.454 rows=1 loops=1847) Buffers: shared hit=1093445 read=538 I/O Timings: read=21.060 -> Seq Scan on emailsubscription.reg_email_subscriptions res (cost=0.00..1728.07 rows=1 width=8) (actualtime=11.316..11.316 rows=0 loops=1847) Output: res.registration_id Filter: ((res.subscription_id = 200400) AND (it.recipient_id = res.registration_id)) Rows Removed by Filter: 77271 Buffers: shared hit=1050943 -> Index Only Scan using reg_email_subscriptions_p00_pkey on emailsubscription.reg_email_subscriptions_p00res_1 (cost=0.57..2.60 rows=2 width=8) (actual time=0.033..0.033 rows=0 loops=1847) Output: res_1.registration_id Index Cond: ((res_1.registration_id = it.recipient_id) AND (res_1.subscription_id = 200400)) Heap Fetches: 0 Buffers: shared hit=7415 read=65 I/O Timings: read=2.802 -> Index Only Scan using reg_email_subscriptions_p01_pkey on emailsubscription.reg_email_subscriptions_p01res_2 (cost=0.57..2.60 rows=2 width=8) (actual time=0.017..0.017 rows=0 loops=1710) Output: res_2.registration_id Index Cond: ((res_2.registration_id = it.recipient_id) AND (res_2.subscription_id = 200400)) Heap Fetches: 0 Buffers: shared hit=6866 read=67 I/O Timings: read=3.053 -> Index Only Scan using reg_email_subscriptions_p02_pkey on emailsubscription.reg_email_subscriptions_p02res_3 (cost=0.57..2.60 rows=2 width=8) (actual time=0.017..0.017 rows=0 loops=1567) Output: res_3.registration_id Index Cond: ((res_3.registration_id = it.recipient_id) AND (res_3.subscription_id = 200400)) Heap Fetches: 0 Buffers: shared hit=6296 read=68 I/O Timings: read=2.812 -> Index Only Scan using reg_email_subscriptions_p03_pkey on emailsubscription.reg_email_subscriptions_p03res_4 (cost=0.57..2.60 rows=2 width=8) (actual time=0.016..0.016 rows=0 loops=1406) Output: res_4.registration_id Index Cond: ((res_4.registration_id = it.recipient_id) AND (res_4.subscription_id = 200400)) Heap Fetches: 0 Buffers: shared hit=5678 read=68 I/O Timings: read=2.645 -> Index Only Scan using reg_email_subscriptions_p04_pkey on emailsubscription.reg_email_subscriptions_p04res_5 (cost=0.57..2.60 rows=2 width=8) (actual time=0.019..0.019 rows=0 loops=1246) Output: res_5.registration_id Index Cond: ((res_5.registration_id = it.recipient_id) AND (res_5.subscription_id = 200400)) Heap Fetches: 0 Buffers: shared hit=5016 read=67 I/O Timings: read=2.647 -> Index Only Scan using reg_email_subscriptions_p05_pkey on emailsubscription.reg_email_subscriptions_p05res_6 (cost=0.57..2.60 rows=2 width=8) (actual time=0.017..0.017 rows=0 loops=1082) Output: res_6.registration_id Index Cond: ((res_6.registration_id = it.recipient_id) AND (res_6.subscription_id = 200400)) Heap Fetches: 0 Buffers: shared hit=4355 read=67 I/O Timings: read=2.471 -> Index Only Scan using reg_email_subscriptions_p06_pkey on emailsubscription.reg_email_subscriptions_p06res_7 (cost=0.57..2.60 rows=2 width=8) (actual time=0.019..0.019 rows=0 loops=946) Output: res_7.registration_id Index Cond: ((res_7.registration_id = it.recipient_id) AND (res_7.subscription_id = 200400)) Heap Fetches: 2 Buffers: shared hit=3828 read=69 I/O Timings: read=2.363 -> Index Only Scan using reg_email_subscriptions_p07_pkey on emailsubscription.reg_email_subscriptions_p07res_8 (cost=0.57..2.60 rows=2 width=8) (actual time=0.021..0.021 rows=0 loops=752) Output: res_8.registration_id Index Cond: ((res_8.registration_id = it.recipient_id) AND (res_8.subscription_id = 200400)) Heap Fetches: 0 Buffers: shared hit=3048 read=67 I/O Timings: read=2.267 Total runtime: 21195.648 ms (71 rows) > I don't believe that session_replication_role does anything to affect indexing, nor do the docs indicate that. I also seenothing indicating that in the source code (search for SessionReplicationRole). > So if you suddenly started seeing dupes then I think your index is corrupted. How can we tell if any index is corrupted or not? If vacuum full re-indexes all the indexes in the table, would that have fixed the corruption, if we had any? > It won't report it, but it would fix it. (Note that it would throw an error if the data in the table isn't actually unique.) We did not get any error during re-indexing. So our understanding that there are no more duplicates in the table is correct! Thank You for your time Ramya
On 2/3/15 4:57 PM, Vasudevan, Ramya wrote: > -> Append (cost=0.00..1748.87 rows=17 width=8) (actual time=11.454..11.454 rows=1 loops=1847) > Buffers: shared hit=1093445 read=538 > I/O Timings: read=21.060 > -> Seq Scan on emailsubscription.reg_email_subscriptions res (cost=0.00..1728.07 rows=1 width=8) (actualtime=11.316..11.316 rows=0 loops=1847) > Output: res.registration_id > Filter: ((res.subscription_id = 200400) AND (it.recipient_id = res.registration_id)) > Rows Removed by Filter: 77271 > Buffers: shared hit=1050943 > -> Index Only Scan using reg_email_subscriptions_p00_pkey on emailsubscription.reg_email_subscriptions_p00res_1 (cost=0.57..2.60 rows=2 width=8) (actual time=0.033..0.033 rows=0 loops=1847) > Output: res_1.registration_id > Index Cond: ((res_1.registration_id = it.recipient_id) AND (res_1.subscription_id = 200400)) > Heap Fetches: 0 > Buffers: shared hit=7415 read=65 > I/O Timings: read=2.802 ... Here's the part that's slow. The index scan on each partition is taking ~2.5ms, and is being repeated 1847 times *for each partition*. What is the table partitioned on? >> I don't believe that session_replication_role does anything to affect indexing, nor do the docs indicate that. I alsosee nothing indicating that in the source code (search for SessionReplicationRole). >> So if you suddenly started seeing dupes then I think your index is corrupted. > How can we tell if any index is corrupted or not? > If vacuum full re-indexes all the indexes in the table, would that have fixed the corruption, if we had any? Index corruption, yes. >> It won't report it, but it would fix it. (Note that it would throw an error if the data in the table isn't actually unique.) > We did not get any error during re-indexing. So our understanding that there are no more duplicates in the table is correct! That's good. Now the question is: why was the index corrupted? It's *extremely* unlikely that it's Postgres. In my experience, corruption is caused be bad hardware, or a misconfiguration (specifically, fsync not doing what it's supposed to). -- Jim Nasby, Data Architect, Blue Treble Consulting Data in Trouble? Get it in Treble! http://BlueTreble.com
> Here's the part that's slow. The index scan on each partition is taking ~2.5ms, and is being repeated 1847 times *for each partition*.
> What is the table partitioned on?
The table is partitioned on registration_id.
CREATE TABLE emailsubscription.reg_email_subscriptions_p00
(
CONSTRAINT reg_email_subscriptions_p00_pkey PRIMARY KEY (registration_id, subscription_id),
CONSTRAINT reg_email_subscriptions_p00_registration_id_check CHECK ((abs(registration_id) % 8::bigint) = 0)
)
INHERITS (emailsubscription.reg_email_subscriptions)
WITH (
OIDS=FALSE
);
Update:
We changed the query from
SELECT DISTINCT it.recipient_id
FROM iru.iru_tags it
WHERE it.recipient_id BETWEEN 758587587 and 968696896
AND it.status = 0
AND it.last_update_date >= date_trunc('day', now() - interval '90 days')
AND EXISTS (SELECT DISTINCT res.registration_id
FROM emailsubscription.reg_email_subscriptions res
WHERE res.registration_id = it.recipient_id
AND res.subscription_id = 200400);
to:
SELECT DISTINCT it.recipient_id
FROM iru.iru_tags it
where it.STATUS = 0
AND it.last_update_date >= date_trunc('day', now() - interval '90 days')
AND EXISTS (SELECT res.registration_id
FROM emailsubscription.reg_email_subscriptions res
WHERE res.registration_id = it.recipient_id
and res.registration_id BETWEEN 8706059856 AND 8706077435
AND res.subscription_id = 200400);
And it slightly improved the performance.
QUERY PLAN
--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------
HashAggregate (cost=733840.24..734045.35 rows=6837 width=8) (actual time=14208.223..14208.414 rows=891 loops=1)
Output: it.recipient_id
Buffers: shared hit=73563 read=18189
I/O Timings: read=12490.324
-> Nested Loop (cost=30901.28..733823.14 rows=6837 width=8) (actual time=6445.168..14203.967 rows=2547 loops=1)
Output: it.recipient_id
Buffers: shared hit=73563 read=18189
I/O Timings: read=12490.324
-> HashAggregate (cost=30900.72..31284.18 rows=12782 width=8) (actual time=6394.062..6413.045 rows=14452 loops=1)
Output: res.registration_id
Buffers: shared hit=14158 read=914
I/O Timings: read=5771.423
-> Append (cost=0.00..30868.76 rows=12782 width=8) (actual time=85.811..6361.937 rows=14452 loops=1)
Buffers: shared hit=14158 read=914
I/O Timings: read=5771.423
-> Seq Scan on emailsubscription.reg_email_subscriptions res (cost=0.00..3470.45 rows=1 width=8) (actual time=75.713..75.713 rows=0 loops=1)
Output: res.registration_id
Filter: ((res.registration_id >= 8706059856::bigint) AND (res.registration_id <= 8706077435::bigint) AND (res.subscription_id = 200400))
Rows Removed by Filter: 77239
Buffers: shared hit=569
-> Index Only Scan using reg_email_subscriptions_p00_pkey on emailsubscription.reg_email_subscriptions_p00 res_1 (cost=0.57..3406.75 rows=1612 width=8) (actual time=10.095..611.086 rows=1802 loops=1)
Output: res_1.registration_id
Index Cond: ((res_1.registration_id >= 8706059856::bigint) AND (res_1.registration_id <= 8706077435::bigint) AND (res_1.subscription_id = 200400))
Heap Fetches: 1806
Buffers: shared hit=1695 read=110
I/O Timings: read=562.961
-> Index Only Scan using reg_email_subscriptions_p01_pkey on emailsubscription.reg_email_subscriptions_p01 res_2 (cost=0.57..3061.12 rows=1401 width=8) (actual time=19.052..849.618 rows=1794 loops=1)
Output: res_2.registration_id
Index Cond: ((res_2.registration_id >= 8706059856::bigint) AND (res_2.registration_id <= 8706077435::bigint) AND (res_2.subscription_id = 200400))
Heap Fetches: 1794
Buffers: shared hit=1674 read=120
I/O Timings: read=739.525
-> Index Only Scan using reg_email_subscriptions_p02_pkey on emailsubscription.reg_email_subscriptions_p02 res_3 (cost=0.57..3495.50 rows=1662 width=8) (actual time=19.480..1037.415 rows=1806 loops=1)
Output: res_3.registration_id
Index Cond: ((res_3.registration_id >= 8706059856::bigint) AND (res_3.registration_id <= 8706077435::bigint) AND (res_3.subscription_id = 200400))
Heap Fetches: 1807
Buffers: shared hit=1687 read=117
I/O Timings: read=946.189
-> Index Only Scan using reg_email_subscriptions_p03_pkey on emailsubscription.reg_email_subscriptions_p03 res_4 (cost=0.57..3349.44 rows=1532 width=8) (actual time=15.859..776.363 rows=1819 loops=1)
Output: res_4.registration_id
Index Cond: ((res_4.registration_id >= 8706059856::bigint) AND (res_4.registration_id <= 8706077435::bigint) AND (res_4.subscription_id = 200400))
Heap Fetches: 1821
Buffers: shared hit=1710 read=120
I/O Timings: read=718.126
-> Index Only Scan using reg_email_subscriptions_p04_pkey on emailsubscription.reg_email_subscriptions_p04 res_5 (cost=0.57..3600.05 rows=1713 width=8) (actual time=37.922..817.469 rows=1806 loops=1)
Output: res_5.registration_id
Index Cond: ((res_5.registration_id >= 8706059856::bigint) AND (res_5.registration_id <= 8706077435::bigint) AND (res_5.subscription_id = 200400))
Heap Fetches: 1809
Buffers: shared hit=1705 read=110
I/O Timings: read=740.888
-> Index Only Scan using reg_email_subscriptions_p05_pkey on emailsubscription.reg_email_subscriptions_p05 res_6 (cost=0.57..3156.12 rows=1445 width=8) (actual time=26.584..645.433 rows=1789 loops=1)
Output: res_6.registration_id
Index Cond: ((res_6.registration_id >= 8706059856::bigint) AND (res_6.registration_id <= 8706077435::bigint) AND (res_6.subscription_id = 200400))
Heap Fetches: 1792
Buffers: shared hit=1691 read=112
I/O Timings: read=627.311
-> Index Only Scan using reg_email_subscriptions_p06_pkey on emailsubscription.reg_email_subscriptions_p06 res_7 (cost=0.57..3369.90 rows=1605 width=8) (actual time=18.880..803.968 rows=1820 loops=1)
Output: res_7.registration_id
Index Cond: ((res_7.registration_id >= 8706059856::bigint) AND (res_7.registration_id <= 8706077435::bigint) AND (res_7.subscription_id = 200400))
Heap Fetches: 1821
Buffers: shared hit=1718 read=110
I/O Timings: read=742.403
-> Index Only Scan using reg_email_subscriptions_p07_pkey on emailsubscription.reg_email_subscriptions_p07 res_8 (cost=0.57..3959.44 rows=1811 width=8) (actual time=15.284..738.786 rows=1816 loops=1)
Output: res_8.registration_id
Index Cond: ((res_8.registration_id >= 8706059856::bigint) AND (res_8.registration_id <= 8706077435::bigint) AND (res_8.subscription_id = 200400))
Heap Fetches: 1819
Buffers: shared hit=1709 read=115
I/O Timings: read=694.020
-> Index Scan using iru_tags_n1 on iru.iru_tags it (cost=0.57..54.93 rows=1 width=8) (actual time=0.423..0.537 rows=0 loops=14452)
Output: it.recipient_id, it.tagger_id, it.tag_id, it.batch_id, it.status, it.creation_date, it.last_update_date
Index Cond: (it.recipient_id = res.registration_id)
Filter: ((it.status = 0) AND (it.last_update_date >= date_trunc('day'::text, (now() - '90 days'::interval))))
Rows Removed by Filter: 1
Buffers: shared hit=59405 read=17275
I/O Timings: read=6718.901
Total runtime: 14209.137 ms
> That's good. Now the question is: why was the index corrupted? It's
> *extremely* unlikely that it's Postgres. In my experience, corruption is caused be bad hardware, or a misconfiguration (specifically, fsync not doing what it's supposed to).
I am not sure if there was a corruption in the first place. Since we suddenly started having problem with a query that was running much faster before, we were wondering if there was a corruption. But, vacuum did not indicate that we had one. Details on the maintenance we did the night before is in the original post.
Thank You for your time
Ramya