Re: BTree vacuum before page splitting - Mailing list pgsql-patches

From Simon Riggs
Subject Re: BTree vacuum before page splitting
Date
Msg-id 1138777928.3096.75.camel@localhost.localdomain
Whole thread Raw
In response to Re: BTree vacuum before page splitting  (Junji TERAMOTO <teramoto.junji@lab.ntt.co.jp>)
Responses Re: BTree vacuum before page splitting
List pgsql-patches
On Tue, 2006-01-31 at 15:18 +0900, Junji TERAMOTO wrote:
> Tom Lane wrote:
> > I think this is quite likely to break things :-(.  What sort of
> > conditions have you tested it under?  (If this were safe, we'd
> > not have invented the LP_DELETE flag to begin with, but just have
> > deleted known-dead items immediately.)
>
> I apologize for my insufficient description(and bad english).
> I explain the operation of this patch in detail again.
>
> In _bt_insertonpg(), we insert the tupple by the following methods.
>
>  (1) Finds the right place(page) to insert the tuple.
>  (2) If free space is insufficient, we operate as follows.
> + (2.1) Confirm whether the lock to the found page is considerable
>         super-exclusive lock in BufferSuperLockHeldByMe()[new]. We
>         check bufHdr->refcount(=pin), and if pin == 1, we know this
>         lock is super-exclusive lock.
>         If our lock is not super-exclusive lock, goto (2.4).
> + (2.2) If our lock is super-exclusive lock, and the found page is
>         leaf, we look for tupples marked as "LP_DELETE" from the found
>         page, and remove found tuples in _bt_vacuum_page()[new].
> + (2.3) If free space is sufficient after removal, goto (4).
>   (2.4) Step right to next non-dead page.
>   (2.5) (2) is repeated until an enough free space is found or it
>         reaches a right edge at last.
>  (3) When an enough free space is not found by processing (2), splits
>      the target page (making sure that the split is equitable as far as
>      post-insert free space goes).
>  (4) Inserts the tuple.
>
> The numbers from 2.1) to 2.3) are new processing.
>
> The pointer's shifting by removing "LP_DELETE" tuples as shown in the
> above-mentioned becomes a problem, when we resume a stopped indexscan.
> Therefore, I am having it confirm the state of the lock by (2.1), and
> process only at super-exclucive lock, same as btbulkdelete().
>
> However, because own indexscan might be lost because of this removal,
> I also modify _bt_restscan() as follows.
>
>  (1) Check for index tuple we stopped on.
>  (2) If index tuple is moved, first of all, we search in this page
>      right.
> +(3) If not found, we try to look for from the head of this page
>      because index tuple may moved left due to removal.
>  (4) If still not found, we look for index tuple right.
>
> The number (3) is new processing.
>
> We do not delete the empty page. Therefore, there is no necessity for
> tracing a left page.
>
> I think that no problem occurs by this logic.
> Please point it out if there is a wrong point. Thanks.

Please read comments in nbtree.c p.557-566 which describe the required
locking protocol for btree pages.

Just because you hold the lock does *not* mean you are allowed to delete
tuples marked as deleted. This patch doesn't work in all cases, which is
what is required, since the failure cases don't raise an ERROR - they
just miss data - which is unacceptable. So your patch works 99.9% of the
time, but sometimes gives wrong answers to queries without you knowing.
So, patch rejected, but good thinking all the same.

You will need to argue for a change in the locking protocol before it
could be accepted. That would speed up VACUUM also, so if it were
possible it would be gratefully accepted.

The only help I can give is this: Why does an index scan ever want to
stop on a deleted tuple? Currently, the scan only remembers one tuple
its seen. If the tuple was dead, but wasn't yet marked as dead the index
scan will have read it and stopped there. Some while later, the index
scan will return and begin scanning right for the tuple. If you delete
it, the index scan *may* not be able to work out where to restart --
ERROR. So, if you can work out a way of not restarting an index scan on
a deleted tuple (and yet still marking them as deleted when it sees
them) then you may be able to solve the problem.

I'm looking at the same problem domain also (table bloat), but focusing
on a different approach that avoids this issue.

Best Regards, Simon Riggs


pgsql-patches by date:

Previous
From: David Fetter
Date:
Subject: Re: [BUGS] BUG #2221: Bad delimiters allowed in COPY ...
Next
From: Junji TERAMOTO
Date:
Subject: Re: BTree vacuum before page splitting