Re: Multiple FPI_FOR_HINT for the same block during killing btreeindex items - Mailing list pgsql-hackers

From Ranier Vilela
Subject Re: Multiple FPI_FOR_HINT for the same block during killing btreeindex items
Date
Msg-id CAEudQAoTCPcZRgtxxfVarOqT+fj+RmVrSCCz2DDi7tm7Z4hnmA@mail.gmail.com
Whole thread Raw
In response to Re: Multiple FPI_FOR_HINT for the same block during killing btreeindex items  (Alvaro Herrera <alvherre@2ndquadrant.com>)
List pgsql-hackers
Em sex., 15 de mai. de 2020 às 18:53, Alvaro Herrera <alvherre@2ndquadrant.com> escreveu:
On 2020-Apr-10, Masahiko Sawada wrote:

> Okay. I think only adding the check would also help with reducing the
> likelihood. How about the changes for the current HEAD I've attached?

Pushed this to all branches.  (Branches 12 and older obviously needed an
adjustment.)  Thanks!

> Related to this behavior on btree indexes, this can happen even on
> heaps during searching heap tuples. To reduce the likelihood of that
> more generally I wonder if we can acquire a lock on buffer descriptor
> right before XLogSaveBufferForHint() and set a flag to the buffer
> descriptor that indicates that we're about to log FPI for hint bit so
> that concurrent process can be aware of that.

I'm not sure how that helps; the other process would have to go back and
redo their whole operation from scratch in order to find out whether
there's still something alive that needs killing.

I think you need to acquire the exclusive lock sooner: if, when scanning
the page, you find a killable item, *then* upgrade the lock to exclusive
and restart the scan.  This means that we'll have to wait for any other
process that's doing the scan, and they will all give up their share
lock to wait for the exclusive lock they need.  So the one that gets it
first will do all the killing, log the page, then release the lock.  At
that point the other processes will wake up and see that items have been
killed, so they will return having done nothing.

Like the attached.  I didn't verify that it works well or that it
actually improves performance ...
This is not related to your latest patch.
But I believe I can improve the performance.

So:
1. If killedsomething is false
2. Any killtuple is true and (not ItemIdIsDead(iid)) is false
3. Nothing to be done.

So why do all the work and then discard it.
We can eliminate the current item much earlier, testing if it is already dead.
 
regards,
Ranier VIlela
Attachment

pgsql-hackers by date:

Previous
From: Tom Lane
Date:
Subject: Re: pgindent && weirdness
Next
From: Euler Taveira
Date:
Subject: Re: [PATCH] Fix pg_dump --no-tablespaces for the custom format