Re: Use of inefficient index in the presence of dead tuples - Mailing list pgsql-general

From Alexander Staubo
Subject Re: Use of inefficient index in the presence of dead tuples
Date
Msg-id 2D6F40AA-1385-48AA-9F9F-DA8AA2BF30BC@purefiction.net
Whole thread Raw
In response to Re: Use of inefficient index in the presence of dead tuples  (Laurenz Albe <laurenz.albe@cybertec.at>)
Responses Re: Use of inefficient index in the presence of dead tuples
List pgsql-general
On 28 May 2024, at 13:02, Laurenz Albe <laurenz.albe@cybertec.at> wrote:
> ANALYZE considers only the live rows, so PostgreSQL knows that the query will
> return only few results.  So it chooses the smaller index rather than the one
> that matches the WHERE condition perfectly.
>
> Unfortunately, it has to wade through all the deleted rows, which is slow.

Sounds like the planner _should_ take the dead tuples into account. I’m surprised there are no parameters to tweak to
makethe planner understand that one index is more selective even though it is technically larger. 

> But try to execute the query a second time, and it will be much faster.
> PostgreSQL marks the index entries as "dead" during the first execution, so the
> second execution won't have to look at the heap any more.

Of course. It’s still not _free_; it’s still trawling through many megabytes of dead data, and going through the shared
buffercache and therefore competing with other queries that need shared buffers.  

> I understand your pain, but your use case is somewhat unusual.

I don’t think rapidly updated tables is an unusual use of Postgres, nor is the problem of long-running transaction
preventingdead tuple vacuuming. 

> What I would consider in your place is
> a) running an explicit VACUUM after you delete lots of rows or

The rows are deleted individually. It’s just that there are many transactions doing it concurrently.

> b) using partitioning to get rid of old data

Partitioning will generate dead tuples in the original partition when tuples are moved to the other partition, so I’m
notsure how that would help? 

I did explore a solution which is my “plan B” — adding a “done” column, then using “UPDATE … SET done = true” rather
thandeleting the rows. This causes dead tuples, of course, but then adding a new index with a “… WHERE NOT done” filter
fixesthe problem by forcing the query to use the right index. However, with this solution, rows will still have to be
deleted*sometime*, so this just delays the problem. But it would allow a “batch cleanup”: “DELETE … WHERE done; VACUUM”
inone fell swoop. 




pgsql-general by date:

Previous
From: Laurenz Albe
Date:
Subject: Re: Use of inefficient index in the presence of dead tuples
Next
From: "David G. Johnston"
Date:
Subject: Re: Use of inefficient index in the presence of dead tuples